2026-03-20T17:53:21.766 INFO:root:teuthology version: 1.2.4.dev6+g1c580df7a 2026-03-20T17:53:21.771 DEBUG:teuthology.report:Pushing job info to http://localhost:8080 2026-03-20T17:53:21.790 INFO:teuthology.run:Config: archive_path: /archive/kyr-2026-03-20_17:25:16-rgw-tentacle-none-default-vps/2229 branch: tentacle description: rgw/multisite/{clusters frontend/beast ignore-pg-availability notify omap_limits overrides realms/two-zonegroup supported-random-distro$/{ubuntu_latest} tasks/test_multi} email: null first_in_suite: false flavor: default job_id: '2229' ktype: distro last_in_suite: false machine_type: vps name: kyr-2026-03-20_17:25:16-rgw-tentacle-none-default-vps no_nested_subset: false os_type: ubuntu os_version: '22.04' overrides: admin_socket: branch: tentacle ansible.cephlab: branch: main repo: https://github.com/kshtsk/ceph-cm-ansible.git skip_tags: nagios,monitoring-scripts,hostname,pubkeys,zap,sudoers,kerberos,ntp-client,resolvconf,cpan,nfs vars: logical_volumes: lv_1: scratch_dev: true size: 25%VG vg: vg_nvme lv_2: scratch_dev: true size: 25%VG vg: vg_nvme lv_3: scratch_dev: true size: 25%VG vg: vg_nvme lv_4: scratch_dev: true size: 25%VG vg: vg_nvme timezone: UTC volume_groups: vg_nvme: pvs: /dev/vdb,/dev/vdc,/dev/vdd,/dev/vde ceph: conf: client: debug rgw: 20 debug rgw lifecycle: 20 debug rgw notification: 20 debug rgw sync: 20 rgw crypt require ssl: false rgw crypt s3 kms backend: testing rgw crypt s3 kms encryption keys: testkey-1=YmluCmJvb3N0CmJvb3N0LWJ1aWxkCmNlcGguY29uZgo= rgw curl low speed time: 300 rgw data log num shards: 4 rgw data sync poll interval: 5 rgw md log max shards: 4 rgw meta sync poll interval: 5 rgw s3 auth use sts: true rgw sts key: abcdefghijklmnoq rgw sync data inject err probability: 0 rgw sync log trim interval: 0 rgw sync meta inject err probability: 0 rgw sync obj etag verify: true setgroup: ceph setuser: ceph client.0: rgw data notify interval msec: 0 mgr: debug mgr: 20 debug ms: 1 mon: debug mon: 20 debug ms: 1 debug paxos: 20 osd: debug ms: 1 debug osd: 20 osd mclock iops capacity threshold hdd: 49000 rocksdb delete range threshold: 0 osd.0: osd_max_omap_entries_per_request: 10 osd.1: osd_max_omap_entries_per_request: 1000 osd.2: osd_max_omap_entries_per_request: 10000 flavor: default log-ignorelist: - \(MDS_ALL_DOWN\) - \(MDS_UP_LESS_THAN_MAX\) - \(PG_AVAILABILITY\) - \(PG_DEGRADED\) - \(POOL_APP_NOT_ENABLED\) - not have an application enabled sha1: 70f8415b300f041766fa27faf7d5472699e32388 ceph-deploy: conf: client: log file: /var/log/ceph/ceph-$name.$pid.log mon: {} cephadm: cephadm_binary_url: https://download.ceph.com/rpm-20.2.0/el9/noarch/cephadm install: ceph: flavor: default sha1: 70f8415b300f041766fa27faf7d5472699e32388 extra_system_packages: deb: - python3-jmespath - python3-xmltodict - s3cmd rpm: - bzip2 - perl-Test-Harness - python3-jmespath - python3-xmltodict - s3cmd rgw: compression type: random frontend: beast rgw-multisite: realm: is_default: true name: test-realm zonegroups: - enabled_features: - resharding - notification_v2 endpoints: - c1.client.0 is_default: true is_master: true name: a zones: - endpoints: - c1.client.0 is_default: true is_master: true name: a1 - endpoints: - c1.client.1 name: a2 - enabled_features: - resharding - notification_v2 endpoints: - c2.client.0 is_default: true name: b zones: - endpoints: - c2.client.0 is_default: true is_master: true name: b1 - endpoints: - c2.client.1 name: b2 rgw-multisite-tests: args: - tests.py - -a - '!fails_with_rgw' workunit: branch: tt-tentacle sha1: c04fae7bd18063e6f8f9e1cf40c23f810c1f6ca4 owner: kyr priority: 1000 repo: https://github.com/ceph/ceph.git roles: - - c1.mon.a - c1.osd.0 - c1.osd.1 - c1.osd.2 - c1.client.0 - - c1.mgr.x - c1.osd.3 - c1.osd.4 - c1.osd.5 - c1.client.1 - - c2.mon.a - c2.osd.0 - c2.osd.1 - c2.osd.2 - c2.client.0 - - c2.mgr.x - c2.osd.3 - c2.osd.4 - c2.osd.5 - c2.client.1 seed: 5336 sha1: 70f8415b300f041766fa27faf7d5472699e32388 sleep_before_teardown: 0 suite: rgw suite_branch: tt-tentacle suite_path: /home/teuthos/src/github.com_kshtsk_ceph_c04fae7bd18063e6f8f9e1cf40c23f810c1f6ca4/qa suite_relpath: qa suite_repo: https://github.com/kshtsk/ceph.git suite_sha1: c04fae7bd18063e6f8f9e1cf40c23f810c1f6ca4 targets: vm02.local: ecdsa-sha2-nistp256 AAAAE2VjZHNhLXNoYTItbmlzdHAyNTYAAAAIbmlzdHAyNTYAAABBBLUlbeCQemEFIGAhOsbaSaC9cn/mRW0uqcv6nNbZb53B4Wibb+XPFovkYm/zpOAZ4wEdxyZ/PUnoWpm0Py9ocQE= vm05.local: ecdsa-sha2-nistp256 AAAAE2VjZHNhLXNoYTItbmlzdHAyNTYAAAAIbmlzdHAyNTYAAABBBMoug7TNAoa7mJjh9cncEud7jrvuXdSIQZHZLMdsjGiNzszM0eqBraWpNHWznD8Hn4dQwU6ldQdhO/fM1PnoaIE= vm06.local: ecdsa-sha2-nistp256 AAAAE2VjZHNhLXNoYTItbmlzdHAyNTYAAAAIbmlzdHAyNTYAAABBBAr4G6JTVJMADXrDYrEOUi/7hRfFSeJoHsHrYT4rHLq31Qu7OJgh6KH0wNL/WripVH20sVMfbmCvXmnnPcODwbA= vm08.local: ecdsa-sha2-nistp256 AAAAE2VjZHNhLXNoYTItbmlzdHAyNTYAAAAIbmlzdHAyNTYAAABBBOUGJylvumZGYQnc6Cm2BiuFCvdgt79isI8bVHVCdZ3r8NnfMvYpiSD3SPVaM+Ua03wQvWKlAWO93A22s3C5WOk= tasks: - install: null - ceph: cluster: c1 - ceph: cluster: c2 - rgw: c1.client.0: port: 8000 c1.client.1: port: 8001 c2.client.0: port: 8000 c2.client.1: port: 8001 - rgw-multisite: null - rgw-multisite-tests: config: reconfigure_delay: 90 teuthology: fragments_dropped: [] meta: {} postmerge: [] teuthology_branch: clyso-debian-13 teuthology_repo: https://github.com/clyso/teuthology teuthology_sha1: 1c580df7a9c7c2aadc272da296344fd99f27c444 timestamp: 2026-03-20_17:25:16 tube: vps user: kyr verbose: false worker_log: /home/teuthos/.teuthology/dispatcher/dispatcher.vps.4188345 2026-03-20T17:53:21.790 INFO:teuthology.run:suite_path is set to /home/teuthos/src/github.com_kshtsk_ceph_c04fae7bd18063e6f8f9e1cf40c23f810c1f6ca4/qa; will attempt to use it 2026-03-20T17:53:21.790 INFO:teuthology.run:Found tasks at /home/teuthos/src/github.com_kshtsk_ceph_c04fae7bd18063e6f8f9e1cf40c23f810c1f6ca4/qa/tasks 2026-03-20T17:53:21.790 INFO:teuthology.run_tasks:Running task internal.check_packages... 2026-03-20T17:53:21.790 INFO:teuthology.task.internal:Checking packages... 2026-03-20T17:53:21.791 INFO:teuthology.task.internal:Checking packages for os_type 'ubuntu', flavor 'default' and ceph hash '70f8415b300f041766fa27faf7d5472699e32388' 2026-03-20T17:53:21.791 WARNING:teuthology.packaging:More than one of ref, tag, branch, or sha1 supplied; using branch 2026-03-20T17:53:21.791 INFO:teuthology.packaging:ref: None 2026-03-20T17:53:21.791 INFO:teuthology.packaging:tag: None 2026-03-20T17:53:21.791 INFO:teuthology.packaging:branch: tentacle 2026-03-20T17:53:21.791 INFO:teuthology.packaging:sha1: 70f8415b300f041766fa27faf7d5472699e32388 2026-03-20T17:53:21.791 DEBUG:teuthology.packaging:Querying https://shaman.ceph.com/api/search?status=ready&project=ceph&flavor=default&distros=ubuntu%2F22.04%2Fx86_64&ref=tentacle 2026-03-20T17:53:22.589 INFO:teuthology.task.internal:Found packages for ceph version 20.2.0-714-g147f7c6a-1jammy 2026-03-20T17:53:22.590 INFO:teuthology.run_tasks:Running task internal.buildpackages_prep... 2026-03-20T17:53:22.591 INFO:teuthology.task.internal:no buildpackages task found 2026-03-20T17:53:22.591 INFO:teuthology.run_tasks:Running task internal.save_config... 2026-03-20T17:53:22.591 INFO:teuthology.task.internal:Saving configuration 2026-03-20T17:53:22.598 INFO:teuthology.run_tasks:Running task internal.check_lock... 2026-03-20T17:53:22.599 INFO:teuthology.task.internal.check_lock:Checking locks... 2026-03-20T17:53:22.605 DEBUG:teuthology.task.internal.check_lock:machine status is {'name': 'vm02.local', 'description': '/archive/kyr-2026-03-20_17:25:16-rgw-tentacle-none-default-vps/2229', 'up': True, 'machine_type': 'vps', 'is_vm': True, 'vm_host': {'name': 'localhost', 'description': None, 'up': True, 'machine_type': 'libvirt', 'is_vm': False, 'vm_host': None, 'os_type': None, 'os_version': None, 'arch': None, 'locked': True, 'locked_since': None, 'locked_by': None, 'mac_address': None, 'ssh_pub_key': None}, 'os_type': 'ubuntu', 'os_version': '22.04', 'arch': 'x86_64', 'locked': True, 'locked_since': '2026-03-20 17:51:14.131937', 'locked_by': 'kyr', 'mac_address': '52:55:00:00:00:02', 'ssh_pub_key': 'ecdsa-sha2-nistp256 AAAAE2VjZHNhLXNoYTItbmlzdHAyNTYAAAAIbmlzdHAyNTYAAABBBLUlbeCQemEFIGAhOsbaSaC9cn/mRW0uqcv6nNbZb53B4Wibb+XPFovkYm/zpOAZ4wEdxyZ/PUnoWpm0Py9ocQE='} 2026-03-20T17:53:22.610 DEBUG:teuthology.task.internal.check_lock:machine status is {'name': 'vm05.local', 'description': '/archive/kyr-2026-03-20_17:25:16-rgw-tentacle-none-default-vps/2229', 'up': True, 'machine_type': 'vps', 'is_vm': True, 'vm_host': {'name': 'localhost', 'description': None, 'up': True, 'machine_type': 'libvirt', 'is_vm': False, 'vm_host': None, 'os_type': None, 'os_version': None, 'arch': None, 'locked': True, 'locked_since': None, 'locked_by': None, 'mac_address': None, 'ssh_pub_key': None}, 'os_type': 'ubuntu', 'os_version': '22.04', 'arch': 'x86_64', 'locked': True, 'locked_since': '2026-03-20 17:51:14.132120', 'locked_by': 'kyr', 'mac_address': '52:55:00:00:00:05', 'ssh_pub_key': 'ecdsa-sha2-nistp256 AAAAE2VjZHNhLXNoYTItbmlzdHAyNTYAAAAIbmlzdHAyNTYAAABBBMoug7TNAoa7mJjh9cncEud7jrvuXdSIQZHZLMdsjGiNzszM0eqBraWpNHWznD8Hn4dQwU6ldQdhO/fM1PnoaIE='} 2026-03-20T17:53:22.614 DEBUG:teuthology.task.internal.check_lock:machine status is {'name': 'vm06.local', 'description': '/archive/kyr-2026-03-20_17:25:16-rgw-tentacle-none-default-vps/2229', 'up': True, 'machine_type': 'vps', 'is_vm': True, 'vm_host': {'name': 'localhost', 'description': None, 'up': True, 'machine_type': 'libvirt', 'is_vm': False, 'vm_host': None, 'os_type': None, 'os_version': None, 'arch': None, 'locked': True, 'locked_since': None, 'locked_by': None, 'mac_address': None, 'ssh_pub_key': None}, 'os_type': 'ubuntu', 'os_version': '22.04', 'arch': 'x86_64', 'locked': True, 'locked_since': '2026-03-20 17:51:14.131726', 'locked_by': 'kyr', 'mac_address': '52:55:00:00:00:06', 'ssh_pub_key': 'ecdsa-sha2-nistp256 AAAAE2VjZHNhLXNoYTItbmlzdHAyNTYAAAAIbmlzdHAyNTYAAABBBAr4G6JTVJMADXrDYrEOUi/7hRfFSeJoHsHrYT4rHLq31Qu7OJgh6KH0wNL/WripVH20sVMfbmCvXmnnPcODwbA='} 2026-03-20T17:53:22.618 DEBUG:teuthology.task.internal.check_lock:machine status is {'name': 'vm08.local', 'description': '/archive/kyr-2026-03-20_17:25:16-rgw-tentacle-none-default-vps/2229', 'up': True, 'machine_type': 'vps', 'is_vm': True, 'vm_host': {'name': 'localhost', 'description': None, 'up': True, 'machine_type': 'libvirt', 'is_vm': False, 'vm_host': None, 'os_type': None, 'os_version': None, 'arch': None, 'locked': True, 'locked_since': None, 'locked_by': None, 'mac_address': None, 'ssh_pub_key': None}, 'os_type': 'ubuntu', 'os_version': '22.04', 'arch': 'x86_64', 'locked': True, 'locked_since': '2026-03-20 17:51:14.131075', 'locked_by': 'kyr', 'mac_address': '52:55:00:00:00:08', 'ssh_pub_key': 'ecdsa-sha2-nistp256 AAAAE2VjZHNhLXNoYTItbmlzdHAyNTYAAAAIbmlzdHAyNTYAAABBBOUGJylvumZGYQnc6Cm2BiuFCvdgt79isI8bVHVCdZ3r8NnfMvYpiSD3SPVaM+Ua03wQvWKlAWO93A22s3C5WOk='} 2026-03-20T17:53:22.618 INFO:teuthology.run_tasks:Running task internal.add_remotes... 2026-03-20T17:53:22.619 INFO:teuthology.task.internal:roles: ubuntu@vm02.local - ['c1.mon.a', 'c1.osd.0', 'c1.osd.1', 'c1.osd.2', 'c1.client.0'] 2026-03-20T17:53:22.619 INFO:teuthology.task.internal:roles: ubuntu@vm05.local - ['c1.mgr.x', 'c1.osd.3', 'c1.osd.4', 'c1.osd.5', 'c1.client.1'] 2026-03-20T17:53:22.619 INFO:teuthology.task.internal:roles: ubuntu@vm06.local - ['c2.mon.a', 'c2.osd.0', 'c2.osd.1', 'c2.osd.2', 'c2.client.0'] 2026-03-20T17:53:22.619 INFO:teuthology.task.internal:roles: ubuntu@vm08.local - ['c2.mgr.x', 'c2.osd.3', 'c2.osd.4', 'c2.osd.5', 'c2.client.1'] 2026-03-20T17:53:22.619 INFO:teuthology.run_tasks:Running task console_log... 2026-03-20T17:53:22.625 DEBUG:teuthology.task.console_log:vm02 does not support IPMI; excluding 2026-03-20T17:53:22.630 DEBUG:teuthology.task.console_log:vm05 does not support IPMI; excluding 2026-03-20T17:53:22.634 DEBUG:teuthology.task.console_log:vm06 does not support IPMI; excluding 2026-03-20T17:53:22.639 DEBUG:teuthology.task.console_log:vm08 does not support IPMI; excluding 2026-03-20T17:53:22.639 DEBUG:teuthology.exit:Installing handler: Handler(exiter=, func=.kill_console_loggers at 0x7f3dba1f48b0>, signals=[15]) 2026-03-20T17:53:22.639 INFO:teuthology.run_tasks:Running task internal.connect... 2026-03-20T17:53:22.640 INFO:teuthology.task.internal:Opening connections... 2026-03-20T17:53:22.640 DEBUG:teuthology.task.internal:connecting to ubuntu@vm02.local 2026-03-20T17:53:22.640 DEBUG:teuthology.orchestra.connection:{'hostname': 'vm02.local', 'username': 'ubuntu', 'timeout': 60} 2026-03-20T17:53:22.700 DEBUG:teuthology.task.internal:connecting to ubuntu@vm05.local 2026-03-20T17:53:22.700 DEBUG:teuthology.orchestra.connection:{'hostname': 'vm05.local', 'username': 'ubuntu', 'timeout': 60} 2026-03-20T17:53:22.758 DEBUG:teuthology.task.internal:connecting to ubuntu@vm06.local 2026-03-20T17:53:22.758 DEBUG:teuthology.orchestra.connection:{'hostname': 'vm06.local', 'username': 'ubuntu', 'timeout': 60} 2026-03-20T17:53:22.817 DEBUG:teuthology.task.internal:connecting to ubuntu@vm08.local 2026-03-20T17:53:22.817 DEBUG:teuthology.orchestra.connection:{'hostname': 'vm08.local', 'username': 'ubuntu', 'timeout': 60} 2026-03-20T17:53:22.879 INFO:teuthology.run_tasks:Running task internal.push_inventory... 2026-03-20T17:53:22.880 DEBUG:teuthology.orchestra.run.vm02:> uname -m 2026-03-20T17:53:22.884 INFO:teuthology.orchestra.run.vm02.stdout:x86_64 2026-03-20T17:53:22.884 DEBUG:teuthology.orchestra.run.vm02:> cat /etc/os-release 2026-03-20T17:53:22.930 INFO:teuthology.orchestra.run.vm02.stdout:PRETTY_NAME="Ubuntu 22.04.5 LTS" 2026-03-20T17:53:22.931 INFO:teuthology.orchestra.run.vm02.stdout:NAME="Ubuntu" 2026-03-20T17:53:22.931 INFO:teuthology.orchestra.run.vm02.stdout:VERSION_ID="22.04" 2026-03-20T17:53:22.931 INFO:teuthology.orchestra.run.vm02.stdout:VERSION="22.04.5 LTS (Jammy Jellyfish)" 2026-03-20T17:53:22.931 INFO:teuthology.orchestra.run.vm02.stdout:VERSION_CODENAME=jammy 2026-03-20T17:53:22.931 INFO:teuthology.orchestra.run.vm02.stdout:ID=ubuntu 2026-03-20T17:53:22.931 INFO:teuthology.orchestra.run.vm02.stdout:ID_LIKE=debian 2026-03-20T17:53:22.931 INFO:teuthology.orchestra.run.vm02.stdout:HOME_URL="https://www.ubuntu.com/" 2026-03-20T17:53:22.931 INFO:teuthology.orchestra.run.vm02.stdout:SUPPORT_URL="https://help.ubuntu.com/" 2026-03-20T17:53:22.931 INFO:teuthology.orchestra.run.vm02.stdout:BUG_REPORT_URL="https://bugs.launchpad.net/ubuntu/" 2026-03-20T17:53:22.931 INFO:teuthology.orchestra.run.vm02.stdout:PRIVACY_POLICY_URL="https://www.ubuntu.com/legal/terms-and-policies/privacy-policy" 2026-03-20T17:53:22.931 INFO:teuthology.orchestra.run.vm02.stdout:UBUNTU_CODENAME=jammy 2026-03-20T17:53:22.931 INFO:teuthology.lock.ops:Updating vm02.local on lock server 2026-03-20T17:53:22.935 DEBUG:teuthology.orchestra.run.vm05:> uname -m 2026-03-20T17:53:22.939 INFO:teuthology.orchestra.run.vm05.stdout:x86_64 2026-03-20T17:53:22.939 DEBUG:teuthology.orchestra.run.vm05:> cat /etc/os-release 2026-03-20T17:53:22.981 INFO:teuthology.orchestra.run.vm05.stdout:PRETTY_NAME="Ubuntu 22.04.5 LTS" 2026-03-20T17:53:22.981 INFO:teuthology.orchestra.run.vm05.stdout:NAME="Ubuntu" 2026-03-20T17:53:22.981 INFO:teuthology.orchestra.run.vm05.stdout:VERSION_ID="22.04" 2026-03-20T17:53:22.981 INFO:teuthology.orchestra.run.vm05.stdout:VERSION="22.04.5 LTS (Jammy Jellyfish)" 2026-03-20T17:53:22.981 INFO:teuthology.orchestra.run.vm05.stdout:VERSION_CODENAME=jammy 2026-03-20T17:53:22.981 INFO:teuthology.orchestra.run.vm05.stdout:ID=ubuntu 2026-03-20T17:53:22.981 INFO:teuthology.orchestra.run.vm05.stdout:ID_LIKE=debian 2026-03-20T17:53:22.981 INFO:teuthology.orchestra.run.vm05.stdout:HOME_URL="https://www.ubuntu.com/" 2026-03-20T17:53:22.981 INFO:teuthology.orchestra.run.vm05.stdout:SUPPORT_URL="https://help.ubuntu.com/" 2026-03-20T17:53:22.981 INFO:teuthology.orchestra.run.vm05.stdout:BUG_REPORT_URL="https://bugs.launchpad.net/ubuntu/" 2026-03-20T17:53:22.981 INFO:teuthology.orchestra.run.vm05.stdout:PRIVACY_POLICY_URL="https://www.ubuntu.com/legal/terms-and-policies/privacy-policy" 2026-03-20T17:53:22.981 INFO:teuthology.orchestra.run.vm05.stdout:UBUNTU_CODENAME=jammy 2026-03-20T17:53:22.982 INFO:teuthology.lock.ops:Updating vm05.local on lock server 2026-03-20T17:53:22.985 DEBUG:teuthology.orchestra.run.vm06:> uname -m 2026-03-20T17:53:22.989 INFO:teuthology.orchestra.run.vm06.stdout:x86_64 2026-03-20T17:53:22.989 DEBUG:teuthology.orchestra.run.vm06:> cat /etc/os-release 2026-03-20T17:53:23.033 INFO:teuthology.orchestra.run.vm06.stdout:PRETTY_NAME="Ubuntu 22.04.5 LTS" 2026-03-20T17:53:23.033 INFO:teuthology.orchestra.run.vm06.stdout:NAME="Ubuntu" 2026-03-20T17:53:23.034 INFO:teuthology.orchestra.run.vm06.stdout:VERSION_ID="22.04" 2026-03-20T17:53:23.034 INFO:teuthology.orchestra.run.vm06.stdout:VERSION="22.04.5 LTS (Jammy Jellyfish)" 2026-03-20T17:53:23.034 INFO:teuthology.orchestra.run.vm06.stdout:VERSION_CODENAME=jammy 2026-03-20T17:53:23.034 INFO:teuthology.orchestra.run.vm06.stdout:ID=ubuntu 2026-03-20T17:53:23.034 INFO:teuthology.orchestra.run.vm06.stdout:ID_LIKE=debian 2026-03-20T17:53:23.034 INFO:teuthology.orchestra.run.vm06.stdout:HOME_URL="https://www.ubuntu.com/" 2026-03-20T17:53:23.034 INFO:teuthology.orchestra.run.vm06.stdout:SUPPORT_URL="https://help.ubuntu.com/" 2026-03-20T17:53:23.034 INFO:teuthology.orchestra.run.vm06.stdout:BUG_REPORT_URL="https://bugs.launchpad.net/ubuntu/" 2026-03-20T17:53:23.034 INFO:teuthology.orchestra.run.vm06.stdout:PRIVACY_POLICY_URL="https://www.ubuntu.com/legal/terms-and-policies/privacy-policy" 2026-03-20T17:53:23.034 INFO:teuthology.orchestra.run.vm06.stdout:UBUNTU_CODENAME=jammy 2026-03-20T17:53:23.034 INFO:teuthology.lock.ops:Updating vm06.local on lock server 2026-03-20T17:53:23.037 DEBUG:teuthology.orchestra.run.vm08:> uname -m 2026-03-20T17:53:23.040 INFO:teuthology.orchestra.run.vm08.stdout:x86_64 2026-03-20T17:53:23.040 DEBUG:teuthology.orchestra.run.vm08:> cat /etc/os-release 2026-03-20T17:53:23.084 INFO:teuthology.orchestra.run.vm08.stdout:PRETTY_NAME="Ubuntu 22.04.5 LTS" 2026-03-20T17:53:23.084 INFO:teuthology.orchestra.run.vm08.stdout:NAME="Ubuntu" 2026-03-20T17:53:23.084 INFO:teuthology.orchestra.run.vm08.stdout:VERSION_ID="22.04" 2026-03-20T17:53:23.084 INFO:teuthology.orchestra.run.vm08.stdout:VERSION="22.04.5 LTS (Jammy Jellyfish)" 2026-03-20T17:53:23.084 INFO:teuthology.orchestra.run.vm08.stdout:VERSION_CODENAME=jammy 2026-03-20T17:53:23.084 INFO:teuthology.orchestra.run.vm08.stdout:ID=ubuntu 2026-03-20T17:53:23.084 INFO:teuthology.orchestra.run.vm08.stdout:ID_LIKE=debian 2026-03-20T17:53:23.084 INFO:teuthology.orchestra.run.vm08.stdout:HOME_URL="https://www.ubuntu.com/" 2026-03-20T17:53:23.084 INFO:teuthology.orchestra.run.vm08.stdout:SUPPORT_URL="https://help.ubuntu.com/" 2026-03-20T17:53:23.084 INFO:teuthology.orchestra.run.vm08.stdout:BUG_REPORT_URL="https://bugs.launchpad.net/ubuntu/" 2026-03-20T17:53:23.084 INFO:teuthology.orchestra.run.vm08.stdout:PRIVACY_POLICY_URL="https://www.ubuntu.com/legal/terms-and-policies/privacy-policy" 2026-03-20T17:53:23.084 INFO:teuthology.orchestra.run.vm08.stdout:UBUNTU_CODENAME=jammy 2026-03-20T17:53:23.084 INFO:teuthology.lock.ops:Updating vm08.local on lock server 2026-03-20T17:53:23.088 INFO:teuthology.run_tasks:Running task internal.serialize_remote_roles... 2026-03-20T17:53:23.090 INFO:teuthology.run_tasks:Running task internal.check_conflict... 2026-03-20T17:53:23.091 INFO:teuthology.task.internal:Checking for old test directory... 2026-03-20T17:53:23.091 DEBUG:teuthology.orchestra.run.vm02:> test '!' -e /home/ubuntu/cephtest 2026-03-20T17:53:23.093 DEBUG:teuthology.orchestra.run.vm05:> test '!' -e /home/ubuntu/cephtest 2026-03-20T17:53:23.094 DEBUG:teuthology.orchestra.run.vm06:> test '!' -e /home/ubuntu/cephtest 2026-03-20T17:53:23.095 DEBUG:teuthology.orchestra.run.vm08:> test '!' -e /home/ubuntu/cephtest 2026-03-20T17:53:23.128 INFO:teuthology.run_tasks:Running task internal.check_ceph_data... 2026-03-20T17:53:23.129 INFO:teuthology.task.internal:Checking for non-empty /var/lib/ceph... 2026-03-20T17:53:23.129 DEBUG:teuthology.orchestra.run.vm02:> test -z $(ls -A /var/lib/ceph) 2026-03-20T17:53:23.136 DEBUG:teuthology.orchestra.run.vm05:> test -z $(ls -A /var/lib/ceph) 2026-03-20T17:53:23.138 INFO:teuthology.orchestra.run.vm02.stderr:ls: cannot access '/var/lib/ceph': No such file or directory 2026-03-20T17:53:23.139 DEBUG:teuthology.orchestra.run.vm06:> test -z $(ls -A /var/lib/ceph) 2026-03-20T17:53:23.140 DEBUG:teuthology.orchestra.run.vm08:> test -z $(ls -A /var/lib/ceph) 2026-03-20T17:53:23.141 INFO:teuthology.orchestra.run.vm05.stderr:ls: cannot access '/var/lib/ceph': No such file or directory 2026-03-20T17:53:23.143 INFO:teuthology.orchestra.run.vm06.stderr:ls: cannot access '/var/lib/ceph': No such file or directory 2026-03-20T17:53:23.172 INFO:teuthology.orchestra.run.vm08.stderr:ls: cannot access '/var/lib/ceph': No such file or directory 2026-03-20T17:53:23.173 INFO:teuthology.run_tasks:Running task internal.vm_setup... 2026-03-20T17:53:23.179 DEBUG:teuthology.orchestra.run.vm02:> test -e /ceph-qa-ready 2026-03-20T17:53:23.182 DEBUG:teuthology.orchestra.run:got remote process result: 1 2026-03-20T17:53:23.409 DEBUG:teuthology.orchestra.run.vm05:> test -e /ceph-qa-ready 2026-03-20T17:53:23.412 DEBUG:teuthology.orchestra.run:got remote process result: 1 2026-03-20T17:53:23.707 DEBUG:teuthology.orchestra.run.vm06:> test -e /ceph-qa-ready 2026-03-20T17:53:23.709 DEBUG:teuthology.orchestra.run:got remote process result: 1 2026-03-20T17:53:24.160 DEBUG:teuthology.orchestra.run.vm08:> test -e /ceph-qa-ready 2026-03-20T17:53:24.163 DEBUG:teuthology.orchestra.run:got remote process result: 1 2026-03-20T17:53:24.394 INFO:teuthology.run_tasks:Running task internal.base... 2026-03-20T17:53:24.396 INFO:teuthology.task.internal:Creating test directory... 2026-03-20T17:53:24.396 DEBUG:teuthology.orchestra.run.vm02:> mkdir -p -m0755 -- /home/ubuntu/cephtest 2026-03-20T17:53:24.397 DEBUG:teuthology.orchestra.run.vm05:> mkdir -p -m0755 -- /home/ubuntu/cephtest 2026-03-20T17:53:24.398 DEBUG:teuthology.orchestra.run.vm06:> mkdir -p -m0755 -- /home/ubuntu/cephtest 2026-03-20T17:53:24.399 DEBUG:teuthology.orchestra.run.vm08:> mkdir -p -m0755 -- /home/ubuntu/cephtest 2026-03-20T17:53:24.402 INFO:teuthology.run_tasks:Running task internal.archive_upload... 2026-03-20T17:53:24.403 INFO:teuthology.run_tasks:Running task internal.archive... 2026-03-20T17:53:24.404 INFO:teuthology.task.internal:Creating archive directory... 2026-03-20T17:53:24.404 DEBUG:teuthology.orchestra.run.vm02:> install -d -m0755 -- /home/ubuntu/cephtest/archive 2026-03-20T17:53:24.441 DEBUG:teuthology.orchestra.run.vm05:> install -d -m0755 -- /home/ubuntu/cephtest/archive 2026-03-20T17:53:24.443 DEBUG:teuthology.orchestra.run.vm06:> install -d -m0755 -- /home/ubuntu/cephtest/archive 2026-03-20T17:53:24.444 DEBUG:teuthology.orchestra.run.vm08:> install -d -m0755 -- /home/ubuntu/cephtest/archive 2026-03-20T17:53:24.450 INFO:teuthology.run_tasks:Running task internal.coredump... 2026-03-20T17:53:24.452 INFO:teuthology.task.internal:Enabling coredump saving... 2026-03-20T17:53:24.452 DEBUG:teuthology.orchestra.run.vm02:> test -f /run/.containerenv -o -f /.dockerenv 2026-03-20T17:53:24.486 DEBUG:teuthology.orchestra.run:got remote process result: 1 2026-03-20T17:53:24.486 DEBUG:teuthology.orchestra.run.vm05:> test -f /run/.containerenv -o -f /.dockerenv 2026-03-20T17:53:24.489 DEBUG:teuthology.orchestra.run:got remote process result: 1 2026-03-20T17:53:24.489 DEBUG:teuthology.orchestra.run.vm06:> test -f /run/.containerenv -o -f /.dockerenv 2026-03-20T17:53:24.491 DEBUG:teuthology.orchestra.run:got remote process result: 1 2026-03-20T17:53:24.491 DEBUG:teuthology.orchestra.run.vm08:> test -f /run/.containerenv -o -f /.dockerenv 2026-03-20T17:53:24.494 DEBUG:teuthology.orchestra.run:got remote process result: 1 2026-03-20T17:53:24.494 DEBUG:teuthology.orchestra.run.vm02:> install -d -m0755 -- /home/ubuntu/cephtest/archive/coredump && sudo sysctl -w kernel.core_pattern=/home/ubuntu/cephtest/archive/coredump/%t.%p.core && echo kernel.core_pattern=/home/ubuntu/cephtest/archive/coredump/%t.%p.core | sudo tee -a /etc/sysctl.conf 2026-03-20T17:53:24.529 DEBUG:teuthology.orchestra.run.vm05:> install -d -m0755 -- /home/ubuntu/cephtest/archive/coredump && sudo sysctl -w kernel.core_pattern=/home/ubuntu/cephtest/archive/coredump/%t.%p.core && echo kernel.core_pattern=/home/ubuntu/cephtest/archive/coredump/%t.%p.core | sudo tee -a /etc/sysctl.conf 2026-03-20T17:53:24.532 DEBUG:teuthology.orchestra.run.vm06:> install -d -m0755 -- /home/ubuntu/cephtest/archive/coredump && sudo sysctl -w kernel.core_pattern=/home/ubuntu/cephtest/archive/coredump/%t.%p.core && echo kernel.core_pattern=/home/ubuntu/cephtest/archive/coredump/%t.%p.core | sudo tee -a /etc/sysctl.conf 2026-03-20T17:53:24.536 DEBUG:teuthology.orchestra.run.vm08:> install -d -m0755 -- /home/ubuntu/cephtest/archive/coredump && sudo sysctl -w kernel.core_pattern=/home/ubuntu/cephtest/archive/coredump/%t.%p.core && echo kernel.core_pattern=/home/ubuntu/cephtest/archive/coredump/%t.%p.core | sudo tee -a /etc/sysctl.conf 2026-03-20T17:53:24.537 INFO:teuthology.orchestra.run.vm02.stdout:kernel.core_pattern = /home/ubuntu/cephtest/archive/coredump/%t.%p.core 2026-03-20T17:53:24.539 INFO:teuthology.orchestra.run.vm05.stdout:kernel.core_pattern = /home/ubuntu/cephtest/archive/coredump/%t.%p.core 2026-03-20T17:53:24.541 INFO:teuthology.orchestra.run.vm02.stdout:kernel.core_pattern=/home/ubuntu/cephtest/archive/coredump/%t.%p.core 2026-03-20T17:53:24.543 INFO:teuthology.orchestra.run.vm06.stdout:kernel.core_pattern = /home/ubuntu/cephtest/archive/coredump/%t.%p.core 2026-03-20T17:53:24.544 INFO:teuthology.orchestra.run.vm05.stdout:kernel.core_pattern=/home/ubuntu/cephtest/archive/coredump/%t.%p.core 2026-03-20T17:53:24.545 INFO:teuthology.orchestra.run.vm08.stdout:kernel.core_pattern = /home/ubuntu/cephtest/archive/coredump/%t.%p.core 2026-03-20T17:53:24.548 INFO:teuthology.orchestra.run.vm06.stdout:kernel.core_pattern=/home/ubuntu/cephtest/archive/coredump/%t.%p.core 2026-03-20T17:53:24.550 INFO:teuthology.orchestra.run.vm08.stdout:kernel.core_pattern=/home/ubuntu/cephtest/archive/coredump/%t.%p.core 2026-03-20T17:53:24.551 INFO:teuthology.run_tasks:Running task internal.sudo... 2026-03-20T17:53:24.552 INFO:teuthology.task.internal:Configuring sudo... 2026-03-20T17:53:24.552 DEBUG:teuthology.orchestra.run.vm02:> sudo sed -i.orig.teuthology -e 's/^\([^#]*\) \(requiretty\)/\1 !\2/g' -e 's/^\([^#]*\) !\(visiblepw\)/\1 \2/g' /etc/sudoers 2026-03-20T17:53:24.585 DEBUG:teuthology.orchestra.run.vm05:> sudo sed -i.orig.teuthology -e 's/^\([^#]*\) \(requiretty\)/\1 !\2/g' -e 's/^\([^#]*\) !\(visiblepw\)/\1 \2/g' /etc/sudoers 2026-03-20T17:53:24.588 DEBUG:teuthology.orchestra.run.vm06:> sudo sed -i.orig.teuthology -e 's/^\([^#]*\) \(requiretty\)/\1 !\2/g' -e 's/^\([^#]*\) !\(visiblepw\)/\1 \2/g' /etc/sudoers 2026-03-20T17:53:24.592 DEBUG:teuthology.orchestra.run.vm08:> sudo sed -i.orig.teuthology -e 's/^\([^#]*\) \(requiretty\)/\1 !\2/g' -e 's/^\([^#]*\) !\(visiblepw\)/\1 \2/g' /etc/sudoers 2026-03-20T17:53:24.601 INFO:teuthology.run_tasks:Running task internal.syslog... 2026-03-20T17:53:24.603 INFO:teuthology.task.internal.syslog:Starting syslog monitoring... 2026-03-20T17:53:24.603 DEBUG:teuthology.orchestra.run.vm02:> mkdir -p -m0755 -- /home/ubuntu/cephtest/archive/syslog 2026-03-20T17:53:24.633 DEBUG:teuthology.orchestra.run.vm05:> mkdir -p -m0755 -- /home/ubuntu/cephtest/archive/syslog 2026-03-20T17:53:24.636 DEBUG:teuthology.orchestra.run.vm06:> mkdir -p -m0755 -- /home/ubuntu/cephtest/archive/syslog 2026-03-20T17:53:24.640 DEBUG:teuthology.orchestra.run.vm08:> mkdir -p -m0755 -- /home/ubuntu/cephtest/archive/syslog 2026-03-20T17:53:24.644 DEBUG:teuthology.orchestra.run.vm02:> install -m 666 /dev/null /home/ubuntu/cephtest/archive/syslog/kern.log 2026-03-20T17:53:24.679 DEBUG:teuthology.orchestra.run.vm02:> install -m 666 /dev/null /home/ubuntu/cephtest/archive/syslog/misc.log 2026-03-20T17:53:24.723 DEBUG:teuthology.orchestra.run.vm02:> set -ex 2026-03-20T17:53:24.723 DEBUG:teuthology.orchestra.run.vm02:> sudo dd of=/etc/rsyslog.d/80-cephtest.conf 2026-03-20T17:53:24.773 DEBUG:teuthology.orchestra.run.vm05:> install -m 666 /dev/null /home/ubuntu/cephtest/archive/syslog/kern.log 2026-03-20T17:53:24.776 DEBUG:teuthology.orchestra.run.vm05:> install -m 666 /dev/null /home/ubuntu/cephtest/archive/syslog/misc.log 2026-03-20T17:53:24.822 DEBUG:teuthology.orchestra.run.vm05:> set -ex 2026-03-20T17:53:24.822 DEBUG:teuthology.orchestra.run.vm05:> sudo dd of=/etc/rsyslog.d/80-cephtest.conf 2026-03-20T17:53:24.870 DEBUG:teuthology.orchestra.run.vm06:> install -m 666 /dev/null /home/ubuntu/cephtest/archive/syslog/kern.log 2026-03-20T17:53:24.874 DEBUG:teuthology.orchestra.run.vm06:> install -m 666 /dev/null /home/ubuntu/cephtest/archive/syslog/misc.log 2026-03-20T17:53:24.918 DEBUG:teuthology.orchestra.run.vm06:> set -ex 2026-03-20T17:53:24.918 DEBUG:teuthology.orchestra.run.vm06:> sudo dd of=/etc/rsyslog.d/80-cephtest.conf 2026-03-20T17:53:24.966 DEBUG:teuthology.orchestra.run.vm08:> install -m 666 /dev/null /home/ubuntu/cephtest/archive/syslog/kern.log 2026-03-20T17:53:24.970 DEBUG:teuthology.orchestra.run.vm08:> install -m 666 /dev/null /home/ubuntu/cephtest/archive/syslog/misc.log 2026-03-20T17:53:25.012 DEBUG:teuthology.orchestra.run.vm08:> set -ex 2026-03-20T17:53:25.012 DEBUG:teuthology.orchestra.run.vm08:> sudo dd of=/etc/rsyslog.d/80-cephtest.conf 2026-03-20T17:53:25.060 DEBUG:teuthology.orchestra.run.vm02:> sudo service rsyslog restart 2026-03-20T17:53:25.062 DEBUG:teuthology.orchestra.run.vm05:> sudo service rsyslog restart 2026-03-20T17:53:25.063 DEBUG:teuthology.orchestra.run.vm06:> sudo service rsyslog restart 2026-03-20T17:53:25.064 DEBUG:teuthology.orchestra.run.vm08:> sudo service rsyslog restart 2026-03-20T17:53:25.117 INFO:teuthology.run_tasks:Running task internal.timer... 2026-03-20T17:53:25.119 INFO:teuthology.task.internal:Starting timer... 2026-03-20T17:53:25.119 INFO:teuthology.run_tasks:Running task pcp... 2026-03-20T17:53:25.122 INFO:teuthology.run_tasks:Running task selinux... 2026-03-20T17:53:25.124 INFO:teuthology.task.selinux:Excluding vm02: VMs are not yet supported 2026-03-20T17:53:25.124 INFO:teuthology.task.selinux:Excluding vm05: VMs are not yet supported 2026-03-20T17:53:25.124 INFO:teuthology.task.selinux:Excluding vm06: VMs are not yet supported 2026-03-20T17:53:25.124 INFO:teuthology.task.selinux:Excluding vm08: VMs are not yet supported 2026-03-20T17:53:25.124 DEBUG:teuthology.task.selinux:Getting current SELinux state 2026-03-20T17:53:25.124 DEBUG:teuthology.task.selinux:Existing SELinux modes: {} 2026-03-20T17:53:25.124 INFO:teuthology.task.selinux:Putting SELinux into permissive mode 2026-03-20T17:53:25.124 INFO:teuthology.run_tasks:Running task ansible.cephlab... 2026-03-20T17:53:25.126 DEBUG:teuthology.task:Applying overrides for task ansible.cephlab: {'branch': 'main', 'repo': 'https://github.com/kshtsk/ceph-cm-ansible.git', 'skip_tags': 'nagios,monitoring-scripts,hostname,pubkeys,zap,sudoers,kerberos,ntp-client,resolvconf,cpan,nfs', 'vars': {'logical_volumes': {'lv_1': {'scratch_dev': True, 'size': '25%VG', 'vg': 'vg_nvme'}, 'lv_2': {'scratch_dev': True, 'size': '25%VG', 'vg': 'vg_nvme'}, 'lv_3': {'scratch_dev': True, 'size': '25%VG', 'vg': 'vg_nvme'}, 'lv_4': {'scratch_dev': True, 'size': '25%VG', 'vg': 'vg_nvme'}}, 'timezone': 'UTC', 'volume_groups': {'vg_nvme': {'pvs': '/dev/vdb,/dev/vdc,/dev/vdd,/dev/vde'}}}} 2026-03-20T17:53:25.126 DEBUG:teuthology.repo_utils:Setting repo remote to https://github.com/kshtsk/ceph-cm-ansible.git 2026-03-20T17:53:25.131 INFO:teuthology.repo_utils:Fetching github.com_kshtsk_ceph-cm-ansible_main from origin 2026-03-20T17:53:26.039 DEBUG:teuthology.repo_utils:Resetting repo at /home/teuthos/src/github.com_kshtsk_ceph-cm-ansible_main to origin/main 2026-03-20T17:53:26.045 INFO:teuthology.task.ansible:Playbook: [{'import_playbook': 'ansible_managed.yml'}, {'import_playbook': 'teuthology.yml'}, {'hosts': 'testnodes', 'tasks': [{'set_fact': {'ran_from_cephlab_playbook': True}}]}, {'import_playbook': 'testnodes.yml'}, {'import_playbook': 'container-host.yml'}, {'import_playbook': 'cobbler.yml'}, {'import_playbook': 'paddles.yml'}, {'import_playbook': 'pulpito.yml'}, {'hosts': 'testnodes', 'become': True, 'tasks': [{'name': 'Touch /ceph-qa-ready', 'file': {'path': '/ceph-qa-ready', 'state': 'touch'}, 'when': 'ran_from_cephlab_playbook|bool'}]}] 2026-03-20T17:53:26.045 DEBUG:teuthology.task.ansible:Running ansible-playbook -v --extra-vars '{"ansible_ssh_user": "ubuntu", "logical_volumes": {"lv_1": {"scratch_dev": true, "size": "25%VG", "vg": "vg_nvme"}, "lv_2": {"scratch_dev": true, "size": "25%VG", "vg": "vg_nvme"}, "lv_3": {"scratch_dev": true, "size": "25%VG", "vg": "vg_nvme"}, "lv_4": {"scratch_dev": true, "size": "25%VG", "vg": "vg_nvme"}}, "timezone": "UTC", "volume_groups": {"vg_nvme": {"pvs": "/dev/vdb,/dev/vdc,/dev/vdd,/dev/vde"}}}' -i /tmp/teuth_ansible_inventoryca1rcbbo --limit vm02.local,vm05.local,vm06.local,vm08.local /home/teuthos/src/github.com_kshtsk_ceph-cm-ansible_main/cephlab.yml --skip-tags nagios,monitoring-scripts,hostname,pubkeys,zap,sudoers,kerberos,ntp-client,resolvconf,cpan,nfs 2026-03-20T17:55:59.819 DEBUG:teuthology.task.ansible:Reconnecting to [Remote(name='ubuntu@vm02.local'), Remote(name='ubuntu@vm05.local'), Remote(name='ubuntu@vm06.local'), Remote(name='ubuntu@vm08.local')] 2026-03-20T17:55:59.820 INFO:teuthology.orchestra.remote:Trying to reconnect to host 'ubuntu@vm02.local' 2026-03-20T17:55:59.820 DEBUG:teuthology.orchestra.connection:{'hostname': 'vm02.local', 'username': 'ubuntu', 'timeout': 60} 2026-03-20T17:55:59.887 DEBUG:teuthology.orchestra.run.vm02:> true 2026-03-20T17:56:00.105 INFO:teuthology.orchestra.remote:Successfully reconnected to host 'ubuntu@vm02.local' 2026-03-20T17:56:00.105 INFO:teuthology.orchestra.remote:Trying to reconnect to host 'ubuntu@vm05.local' 2026-03-20T17:56:00.105 DEBUG:teuthology.orchestra.connection:{'hostname': 'vm05.local', 'username': 'ubuntu', 'timeout': 60} 2026-03-20T17:56:00.162 DEBUG:teuthology.orchestra.run.vm05:> true 2026-03-20T17:56:00.389 INFO:teuthology.orchestra.remote:Successfully reconnected to host 'ubuntu@vm05.local' 2026-03-20T17:56:00.389 INFO:teuthology.orchestra.remote:Trying to reconnect to host 'ubuntu@vm06.local' 2026-03-20T17:56:00.390 DEBUG:teuthology.orchestra.connection:{'hostname': 'vm06.local', 'username': 'ubuntu', 'timeout': 60} 2026-03-20T17:56:00.453 DEBUG:teuthology.orchestra.run.vm06:> true 2026-03-20T17:56:00.685 INFO:teuthology.orchestra.remote:Successfully reconnected to host 'ubuntu@vm06.local' 2026-03-20T17:56:00.685 INFO:teuthology.orchestra.remote:Trying to reconnect to host 'ubuntu@vm08.local' 2026-03-20T17:56:00.685 DEBUG:teuthology.orchestra.connection:{'hostname': 'vm08.local', 'username': 'ubuntu', 'timeout': 60} 2026-03-20T17:56:00.748 DEBUG:teuthology.orchestra.run.vm08:> true 2026-03-20T17:56:00.973 INFO:teuthology.orchestra.remote:Successfully reconnected to host 'ubuntu@vm08.local' 2026-03-20T17:56:00.973 INFO:teuthology.run_tasks:Running task clock... 2026-03-20T17:56:00.976 INFO:teuthology.task.clock:Syncing clocks and checking initial clock skew... 2026-03-20T17:56:00.976 INFO:teuthology.orchestra.run:Running command with timeout 360 2026-03-20T17:56:00.976 DEBUG:teuthology.orchestra.run.vm02:> sudo systemctl stop ntp.service || sudo systemctl stop ntpd.service || sudo systemctl stop chronyd.service ; sudo ntpd -gq || sudo chronyc makestep ; sudo systemctl start ntp.service || sudo systemctl start ntpd.service || sudo systemctl start chronyd.service ; PATH=/usr/bin:/usr/sbin ntpq -p || PATH=/usr/bin:/usr/sbin chronyc sources || true 2026-03-20T17:56:00.977 INFO:teuthology.orchestra.run:Running command with timeout 360 2026-03-20T17:56:00.977 DEBUG:teuthology.orchestra.run.vm05:> sudo systemctl stop ntp.service || sudo systemctl stop ntpd.service || sudo systemctl stop chronyd.service ; sudo ntpd -gq || sudo chronyc makestep ; sudo systemctl start ntp.service || sudo systemctl start ntpd.service || sudo systemctl start chronyd.service ; PATH=/usr/bin:/usr/sbin ntpq -p || PATH=/usr/bin:/usr/sbin chronyc sources || true 2026-03-20T17:56:00.978 INFO:teuthology.orchestra.run:Running command with timeout 360 2026-03-20T17:56:00.978 DEBUG:teuthology.orchestra.run.vm06:> sudo systemctl stop ntp.service || sudo systemctl stop ntpd.service || sudo systemctl stop chronyd.service ; sudo ntpd -gq || sudo chronyc makestep ; sudo systemctl start ntp.service || sudo systemctl start ntpd.service || sudo systemctl start chronyd.service ; PATH=/usr/bin:/usr/sbin ntpq -p || PATH=/usr/bin:/usr/sbin chronyc sources || true 2026-03-20T17:56:00.980 INFO:teuthology.orchestra.run:Running command with timeout 360 2026-03-20T17:56:00.980 DEBUG:teuthology.orchestra.run.vm08:> sudo systemctl stop ntp.service || sudo systemctl stop ntpd.service || sudo systemctl stop chronyd.service ; sudo ntpd -gq || sudo chronyc makestep ; sudo systemctl start ntp.service || sudo systemctl start ntpd.service || sudo systemctl start chronyd.service ; PATH=/usr/bin:/usr/sbin ntpq -p || PATH=/usr/bin:/usr/sbin chronyc sources || true 2026-03-20T17:56:00.994 INFO:teuthology.orchestra.run.vm02.stdout:20 Mar 17:56:00 ntpd[16237]: ntpd 4.2.8p15@1.3728-o Wed Feb 16 17:13:02 UTC 2022 (1): Starting 2026-03-20T17:56:00.994 INFO:teuthology.orchestra.run.vm02.stdout:20 Mar 17:56:00 ntpd[16237]: Command line: ntpd -gq 2026-03-20T17:56:00.994 INFO:teuthology.orchestra.run.vm02.stdout:20 Mar 17:56:00 ntpd[16237]: ---------------------------------------------------- 2026-03-20T17:56:00.994 INFO:teuthology.orchestra.run.vm02.stdout:20 Mar 17:56:00 ntpd[16237]: ntp-4 is maintained by Network Time Foundation, 2026-03-20T17:56:00.994 INFO:teuthology.orchestra.run.vm02.stdout:20 Mar 17:56:00 ntpd[16237]: Inc. (NTF), a non-profit 501(c)(3) public-benefit 2026-03-20T17:56:00.994 INFO:teuthology.orchestra.run.vm02.stdout:20 Mar 17:56:00 ntpd[16237]: corporation. Support and training for ntp-4 are 2026-03-20T17:56:00.994 INFO:teuthology.orchestra.run.vm02.stdout:20 Mar 17:56:00 ntpd[16237]: available at https://www.nwtime.org/support 2026-03-20T17:56:00.994 INFO:teuthology.orchestra.run.vm02.stdout:20 Mar 17:56:00 ntpd[16237]: ---------------------------------------------------- 2026-03-20T17:56:00.995 INFO:teuthology.orchestra.run.vm02.stdout:20 Mar 17:56:00 ntpd[16237]: proto: precision = 0.030 usec (-25) 2026-03-20T17:56:00.995 INFO:teuthology.orchestra.run.vm05.stdout:20 Mar 17:56:00 ntpd[16231]: ntpd 4.2.8p15@1.3728-o Wed Feb 16 17:13:02 UTC 2022 (1): Starting 2026-03-20T17:56:00.995 INFO:teuthology.orchestra.run.vm05.stdout:20 Mar 17:56:00 ntpd[16231]: Command line: ntpd -gq 2026-03-20T17:56:00.995 INFO:teuthology.orchestra.run.vm05.stdout:20 Mar 17:56:00 ntpd[16231]: ---------------------------------------------------- 2026-03-20T17:56:00.995 INFO:teuthology.orchestra.run.vm05.stdout:20 Mar 17:56:00 ntpd[16231]: ntp-4 is maintained by Network Time Foundation, 2026-03-20T17:56:00.995 INFO:teuthology.orchestra.run.vm05.stdout:20 Mar 17:56:00 ntpd[16231]: Inc. (NTF), a non-profit 501(c)(3) public-benefit 2026-03-20T17:56:00.995 INFO:teuthology.orchestra.run.vm05.stdout:20 Mar 17:56:00 ntpd[16231]: corporation. Support and training for ntp-4 are 2026-03-20T17:56:00.995 INFO:teuthology.orchestra.run.vm05.stdout:20 Mar 17:56:00 ntpd[16231]: available at https://www.nwtime.org/support 2026-03-20T17:56:00.995 INFO:teuthology.orchestra.run.vm05.stdout:20 Mar 17:56:00 ntpd[16231]: ---------------------------------------------------- 2026-03-20T17:56:00.995 INFO:teuthology.orchestra.run.vm05.stdout:20 Mar 17:56:00 ntpd[16231]: proto: precision = 0.030 usec (-25) 2026-03-20T17:56:00.995 INFO:teuthology.orchestra.run.vm05.stdout:20 Mar 17:56:00 ntpd[16231]: basedate set to 2022-02-04 2026-03-20T17:56:00.995 INFO:teuthology.orchestra.run.vm05.stdout:20 Mar 17:56:00 ntpd[16231]: gps base set to 2022-02-06 (week 2196) 2026-03-20T17:56:00.996 INFO:teuthology.orchestra.run.vm05.stdout:20 Mar 17:56:00 ntpd[16231]: leapsecond file ('/usr/share/zoneinfo/leap-seconds.list'): good hash signature 2026-03-20T17:56:00.996 INFO:teuthology.orchestra.run.vm05.stdout:20 Mar 17:56:00 ntpd[16231]: leapsecond file ('/usr/share/zoneinfo/leap-seconds.list'): loaded, expire=2025-12-28T00:00:00Z last=2017-01-01T00:00:00Z ofs=37 2026-03-20T17:56:00.996 INFO:teuthology.orchestra.run.vm05.stderr:20 Mar 17:56:00 ntpd[16231]: leapsecond file ('/usr/share/zoneinfo/leap-seconds.list'): expired 83 days ago 2026-03-20T17:56:00.996 INFO:teuthology.orchestra.run.vm02.stdout:20 Mar 17:56:00 ntpd[16237]: basedate set to 2022-02-04 2026-03-20T17:56:00.996 INFO:teuthology.orchestra.run.vm02.stdout:20 Mar 17:56:00 ntpd[16237]: gps base set to 2022-02-06 (week 2196) 2026-03-20T17:56:00.996 INFO:teuthology.orchestra.run.vm02.stdout:20 Mar 17:56:00 ntpd[16237]: leapsecond file ('/usr/share/zoneinfo/leap-seconds.list'): good hash signature 2026-03-20T17:56:00.996 INFO:teuthology.orchestra.run.vm02.stdout:20 Mar 17:56:00 ntpd[16237]: leapsecond file ('/usr/share/zoneinfo/leap-seconds.list'): loaded, expire=2025-12-28T00:00:00Z last=2017-01-01T00:00:00Z ofs=37 2026-03-20T17:56:00.996 INFO:teuthology.orchestra.run.vm02.stderr:20 Mar 17:56:00 ntpd[16237]: leapsecond file ('/usr/share/zoneinfo/leap-seconds.list'): expired 83 days ago 2026-03-20T17:56:00.996 INFO:teuthology.orchestra.run.vm05.stdout:20 Mar 17:56:00 ntpd[16231]: Listen and drop on 0 v6wildcard [::]:123 2026-03-20T17:56:00.996 INFO:teuthology.orchestra.run.vm05.stdout:20 Mar 17:56:00 ntpd[16231]: Listen and drop on 1 v4wildcard 0.0.0.0:123 2026-03-20T17:56:00.996 INFO:teuthology.orchestra.run.vm05.stdout:20 Mar 17:56:00 ntpd[16231]: Listen normally on 2 lo 127.0.0.1:123 2026-03-20T17:56:00.996 INFO:teuthology.orchestra.run.vm05.stdout:20 Mar 17:56:00 ntpd[16231]: Listen normally on 3 ens3 192.168.123.105:123 2026-03-20T17:56:00.996 INFO:teuthology.orchestra.run.vm05.stdout:20 Mar 17:56:00 ntpd[16231]: Listen normally on 4 lo [::1]:123 2026-03-20T17:56:00.996 INFO:teuthology.orchestra.run.vm05.stdout:20 Mar 17:56:00 ntpd[16231]: Listen normally on 5 ens3 [fe80::5055:ff:fe00:5%2]:123 2026-03-20T17:56:00.996 INFO:teuthology.orchestra.run.vm05.stdout:20 Mar 17:56:00 ntpd[16231]: Listening on routing socket on fd #22 for interface updates 2026-03-20T17:56:00.997 INFO:teuthology.orchestra.run.vm02.stdout:20 Mar 17:56:00 ntpd[16237]: Listen and drop on 0 v6wildcard [::]:123 2026-03-20T17:56:00.997 INFO:teuthology.orchestra.run.vm02.stdout:20 Mar 17:56:00 ntpd[16237]: Listen and drop on 1 v4wildcard 0.0.0.0:123 2026-03-20T17:56:00.997 INFO:teuthology.orchestra.run.vm02.stdout:20 Mar 17:56:00 ntpd[16237]: Listen normally on 2 lo 127.0.0.1:123 2026-03-20T17:56:00.997 INFO:teuthology.orchestra.run.vm02.stdout:20 Mar 17:56:00 ntpd[16237]: Listen normally on 3 ens3 192.168.123.102:123 2026-03-20T17:56:00.997 INFO:teuthology.orchestra.run.vm02.stdout:20 Mar 17:56:00 ntpd[16237]: Listen normally on 4 lo [::1]:123 2026-03-20T17:56:00.997 INFO:teuthology.orchestra.run.vm02.stdout:20 Mar 17:56:00 ntpd[16237]: Listen normally on 5 ens3 [fe80::5055:ff:fe00:2%2]:123 2026-03-20T17:56:00.997 INFO:teuthology.orchestra.run.vm02.stdout:20 Mar 17:56:00 ntpd[16237]: Listening on routing socket on fd #22 for interface updates 2026-03-20T17:56:00.997 INFO:teuthology.orchestra.run.vm06.stdout:20 Mar 17:56:00 ntpd[16247]: ntpd 4.2.8p15@1.3728-o Wed Feb 16 17:13:02 UTC 2022 (1): Starting 2026-03-20T17:56:00.998 INFO:teuthology.orchestra.run.vm06.stdout:20 Mar 17:56:00 ntpd[16247]: Command line: ntpd -gq 2026-03-20T17:56:00.998 INFO:teuthology.orchestra.run.vm06.stdout:20 Mar 17:56:00 ntpd[16247]: ---------------------------------------------------- 2026-03-20T17:56:00.998 INFO:teuthology.orchestra.run.vm06.stdout:20 Mar 17:56:00 ntpd[16247]: ntp-4 is maintained by Network Time Foundation, 2026-03-20T17:56:00.998 INFO:teuthology.orchestra.run.vm06.stdout:20 Mar 17:56:00 ntpd[16247]: Inc. (NTF), a non-profit 501(c)(3) public-benefit 2026-03-20T17:56:00.998 INFO:teuthology.orchestra.run.vm06.stdout:20 Mar 17:56:00 ntpd[16247]: corporation. Support and training for ntp-4 are 2026-03-20T17:56:00.998 INFO:teuthology.orchestra.run.vm06.stdout:20 Mar 17:56:00 ntpd[16247]: available at https://www.nwtime.org/support 2026-03-20T17:56:00.998 INFO:teuthology.orchestra.run.vm06.stdout:20 Mar 17:56:00 ntpd[16247]: ---------------------------------------------------- 2026-03-20T17:56:00.998 INFO:teuthology.orchestra.run.vm06.stdout:20 Mar 17:56:00 ntpd[16247]: proto: precision = 0.029 usec (-25) 2026-03-20T17:56:00.998 INFO:teuthology.orchestra.run.vm06.stdout:20 Mar 17:56:00 ntpd[16247]: basedate set to 2022-02-04 2026-03-20T17:56:00.998 INFO:teuthology.orchestra.run.vm06.stdout:20 Mar 17:56:00 ntpd[16247]: gps base set to 2022-02-06 (week 2196) 2026-03-20T17:56:00.998 INFO:teuthology.orchestra.run.vm06.stdout:20 Mar 17:56:00 ntpd[16247]: leapsecond file ('/usr/share/zoneinfo/leap-seconds.list'): good hash signature 2026-03-20T17:56:00.998 INFO:teuthology.orchestra.run.vm06.stdout:20 Mar 17:56:00 ntpd[16247]: leapsecond file ('/usr/share/zoneinfo/leap-seconds.list'): loaded, expire=2025-12-28T00:00:00Z last=2017-01-01T00:00:00Z ofs=37 2026-03-20T17:56:00.998 INFO:teuthology.orchestra.run.vm06.stderr:20 Mar 17:56:00 ntpd[16247]: leapsecond file ('/usr/share/zoneinfo/leap-seconds.list'): expired 83 days ago 2026-03-20T17:56:00.999 INFO:teuthology.orchestra.run.vm06.stdout:20 Mar 17:56:00 ntpd[16247]: Listen and drop on 0 v6wildcard [::]:123 2026-03-20T17:56:00.999 INFO:teuthology.orchestra.run.vm06.stdout:20 Mar 17:56:00 ntpd[16247]: Listen and drop on 1 v4wildcard 0.0.0.0:123 2026-03-20T17:56:00.999 INFO:teuthology.orchestra.run.vm06.stdout:20 Mar 17:56:00 ntpd[16247]: Listen normally on 2 lo 127.0.0.1:123 2026-03-20T17:56:00.999 INFO:teuthology.orchestra.run.vm06.stdout:20 Mar 17:56:00 ntpd[16247]: Listen normally on 3 ens3 192.168.123.106:123 2026-03-20T17:56:00.999 INFO:teuthology.orchestra.run.vm06.stdout:20 Mar 17:56:00 ntpd[16247]: Listen normally on 4 lo [::1]:123 2026-03-20T17:56:00.999 INFO:teuthology.orchestra.run.vm06.stdout:20 Mar 17:56:00 ntpd[16247]: Listen normally on 5 ens3 [fe80::5055:ff:fe00:6%2]:123 2026-03-20T17:56:00.999 INFO:teuthology.orchestra.run.vm06.stdout:20 Mar 17:56:00 ntpd[16247]: Listening on routing socket on fd #22 for interface updates 2026-03-20T17:56:01.029 INFO:teuthology.orchestra.run.vm08.stdout:20 Mar 17:56:01 ntpd[16255]: ntpd 4.2.8p15@1.3728-o Wed Feb 16 17:13:02 UTC 2022 (1): Starting 2026-03-20T17:56:01.029 INFO:teuthology.orchestra.run.vm08.stdout:20 Mar 17:56:01 ntpd[16255]: Command line: ntpd -gq 2026-03-20T17:56:01.030 INFO:teuthology.orchestra.run.vm08.stdout:20 Mar 17:56:01 ntpd[16255]: ---------------------------------------------------- 2026-03-20T17:56:01.030 INFO:teuthology.orchestra.run.vm08.stdout:20 Mar 17:56:01 ntpd[16255]: ntp-4 is maintained by Network Time Foundation, 2026-03-20T17:56:01.030 INFO:teuthology.orchestra.run.vm08.stdout:20 Mar 17:56:01 ntpd[16255]: Inc. (NTF), a non-profit 501(c)(3) public-benefit 2026-03-20T17:56:01.030 INFO:teuthology.orchestra.run.vm08.stdout:20 Mar 17:56:01 ntpd[16255]: corporation. Support and training for ntp-4 are 2026-03-20T17:56:01.030 INFO:teuthology.orchestra.run.vm08.stdout:20 Mar 17:56:01 ntpd[16255]: available at https://www.nwtime.org/support 2026-03-20T17:56:01.030 INFO:teuthology.orchestra.run.vm08.stdout:20 Mar 17:56:01 ntpd[16255]: ---------------------------------------------------- 2026-03-20T17:56:01.030 INFO:teuthology.orchestra.run.vm08.stdout:20 Mar 17:56:01 ntpd[16255]: proto: precision = 0.029 usec (-25) 2026-03-20T17:56:01.030 INFO:teuthology.orchestra.run.vm08.stdout:20 Mar 17:56:01 ntpd[16255]: basedate set to 2022-02-04 2026-03-20T17:56:01.030 INFO:teuthology.orchestra.run.vm08.stdout:20 Mar 17:56:01 ntpd[16255]: gps base set to 2022-02-06 (week 2196) 2026-03-20T17:56:01.030 INFO:teuthology.orchestra.run.vm08.stdout:20 Mar 17:56:01 ntpd[16255]: leapsecond file ('/usr/share/zoneinfo/leap-seconds.list'): good hash signature 2026-03-20T17:56:01.030 INFO:teuthology.orchestra.run.vm08.stdout:20 Mar 17:56:01 ntpd[16255]: leapsecond file ('/usr/share/zoneinfo/leap-seconds.list'): loaded, expire=2025-12-28T00:00:00Z last=2017-01-01T00:00:00Z ofs=37 2026-03-20T17:56:01.030 INFO:teuthology.orchestra.run.vm08.stderr:20 Mar 17:56:01 ntpd[16255]: leapsecond file ('/usr/share/zoneinfo/leap-seconds.list'): expired 83 days ago 2026-03-20T17:56:01.031 INFO:teuthology.orchestra.run.vm08.stdout:20 Mar 17:56:01 ntpd[16255]: Listen and drop on 0 v6wildcard [::]:123 2026-03-20T17:56:01.031 INFO:teuthology.orchestra.run.vm08.stdout:20 Mar 17:56:01 ntpd[16255]: Listen and drop on 1 v4wildcard 0.0.0.0:123 2026-03-20T17:56:01.031 INFO:teuthology.orchestra.run.vm08.stdout:20 Mar 17:56:01 ntpd[16255]: Listen normally on 2 lo 127.0.0.1:123 2026-03-20T17:56:01.031 INFO:teuthology.orchestra.run.vm08.stdout:20 Mar 17:56:01 ntpd[16255]: Listen normally on 3 ens3 192.168.123.108:123 2026-03-20T17:56:01.031 INFO:teuthology.orchestra.run.vm08.stdout:20 Mar 17:56:01 ntpd[16255]: Listen normally on 4 lo [::1]:123 2026-03-20T17:56:01.031 INFO:teuthology.orchestra.run.vm08.stdout:20 Mar 17:56:01 ntpd[16255]: Listen normally on 5 ens3 [fe80::5055:ff:fe00:8%2]:123 2026-03-20T17:56:01.031 INFO:teuthology.orchestra.run.vm08.stdout:20 Mar 17:56:01 ntpd[16255]: Listening on routing socket on fd #22 for interface updates 2026-03-20T17:56:01.996 INFO:teuthology.orchestra.run.vm05.stdout:20 Mar 17:56:01 ntpd[16231]: Soliciting pool server 213.172.105.106 2026-03-20T17:56:01.996 INFO:teuthology.orchestra.run.vm02.stdout:20 Mar 17:56:01 ntpd[16237]: Soliciting pool server 213.172.105.106 2026-03-20T17:56:01.998 INFO:teuthology.orchestra.run.vm06.stdout:20 Mar 17:56:01 ntpd[16247]: Soliciting pool server 213.172.105.106 2026-03-20T17:56:02.030 INFO:teuthology.orchestra.run.vm08.stdout:20 Mar 17:56:02 ntpd[16255]: Soliciting pool server 129.70.132.35 2026-03-20T17:56:02.995 INFO:teuthology.orchestra.run.vm02.stdout:20 Mar 17:56:02 ntpd[16237]: Soliciting pool server 217.145.111.106 2026-03-20T17:56:02.995 INFO:teuthology.orchestra.run.vm05.stdout:20 Mar 17:56:02 ntpd[16231]: Soliciting pool server 217.145.111.106 2026-03-20T17:56:02.995 INFO:teuthology.orchestra.run.vm05.stdout:20 Mar 17:56:02 ntpd[16231]: Soliciting pool server 18.192.244.117 2026-03-20T17:56:02.996 INFO:teuthology.orchestra.run.vm02.stdout:20 Mar 17:56:02 ntpd[16237]: Soliciting pool server 18.192.244.117 2026-03-20T17:56:02.997 INFO:teuthology.orchestra.run.vm06.stdout:20 Mar 17:56:02 ntpd[16247]: Soliciting pool server 217.145.111.106 2026-03-20T17:56:02.998 INFO:teuthology.orchestra.run.vm06.stdout:20 Mar 17:56:02 ntpd[16247]: Soliciting pool server 18.192.244.117 2026-03-20T17:56:03.030 INFO:teuthology.orchestra.run.vm08.stdout:20 Mar 17:56:03 ntpd[16255]: Soliciting pool server 213.172.105.106 2026-03-20T17:56:03.030 INFO:teuthology.orchestra.run.vm08.stdout:20 Mar 17:56:03 ntpd[16255]: Soliciting pool server 78.46.238.113 2026-03-20T17:56:03.995 INFO:teuthology.orchestra.run.vm05.stdout:20 Mar 17:56:03 ntpd[16231]: Soliciting pool server 5.45.97.204 2026-03-20T17:56:03.995 INFO:teuthology.orchestra.run.vm05.stdout:20 Mar 17:56:03 ntpd[16231]: Soliciting pool server 131.188.3.223 2026-03-20T17:56:03.996 INFO:teuthology.orchestra.run.vm02.stdout:20 Mar 17:56:03 ntpd[16237]: Soliciting pool server 5.45.97.204 2026-03-20T17:56:03.996 INFO:teuthology.orchestra.run.vm02.stdout:20 Mar 17:56:03 ntpd[16237]: Soliciting pool server 131.188.3.223 2026-03-20T17:56:03.996 INFO:teuthology.orchestra.run.vm05.stdout:20 Mar 17:56:03 ntpd[16231]: Soliciting pool server 172.104.149.161 2026-03-20T17:56:03.996 INFO:teuthology.orchestra.run.vm02.stdout:20 Mar 17:56:03 ntpd[16237]: Soliciting pool server 172.104.149.161 2026-03-20T17:56:03.998 INFO:teuthology.orchestra.run.vm06.stdout:20 Mar 17:56:03 ntpd[16247]: Soliciting pool server 5.45.97.204 2026-03-20T17:56:03.998 INFO:teuthology.orchestra.run.vm06.stdout:20 Mar 17:56:03 ntpd[16247]: Soliciting pool server 131.188.3.223 2026-03-20T17:56:03.998 INFO:teuthology.orchestra.run.vm06.stdout:20 Mar 17:56:03 ntpd[16247]: Soliciting pool server 49.12.35.6 2026-03-20T17:56:04.030 INFO:teuthology.orchestra.run.vm08.stdout:20 Mar 17:56:04 ntpd[16255]: Soliciting pool server 18.192.244.117 2026-03-20T17:56:04.030 INFO:teuthology.orchestra.run.vm08.stdout:20 Mar 17:56:04 ntpd[16255]: Soliciting pool server 217.145.111.106 2026-03-20T17:56:04.030 INFO:teuthology.orchestra.run.vm08.stdout:20 Mar 17:56:04 ntpd[16255]: Soliciting pool server 94.16.122.152 2026-03-20T17:56:04.995 INFO:teuthology.orchestra.run.vm05.stdout:20 Mar 17:56:04 ntpd[16231]: Soliciting pool server 46.224.156.215 2026-03-20T17:56:04.995 INFO:teuthology.orchestra.run.vm05.stdout:20 Mar 17:56:04 ntpd[16231]: Soliciting pool server 5.75.181.179 2026-03-20T17:56:04.995 INFO:teuthology.orchestra.run.vm05.stdout:20 Mar 17:56:04 ntpd[16231]: Soliciting pool server 129.70.132.35 2026-03-20T17:56:04.995 INFO:teuthology.orchestra.run.vm02.stdout:20 Mar 17:56:04 ntpd[16237]: Soliciting pool server 46.224.156.215 2026-03-20T17:56:04.996 INFO:teuthology.orchestra.run.vm02.stdout:20 Mar 17:56:04 ntpd[16237]: Soliciting pool server 5.75.181.179 2026-03-20T17:56:04.996 INFO:teuthology.orchestra.run.vm02.stdout:20 Mar 17:56:04 ntpd[16237]: Soliciting pool server 129.70.132.35 2026-03-20T17:56:04.996 INFO:teuthology.orchestra.run.vm02.stdout:20 Mar 17:56:04 ntpd[16237]: Soliciting pool server 5.189.151.39 2026-03-20T17:56:04.996 INFO:teuthology.orchestra.run.vm05.stdout:20 Mar 17:56:04 ntpd[16231]: Soliciting pool server 5.189.151.39 2026-03-20T17:56:04.998 INFO:teuthology.orchestra.run.vm06.stdout:20 Mar 17:56:04 ntpd[16247]: Soliciting pool server 172.104.149.161 2026-03-20T17:56:04.998 INFO:teuthology.orchestra.run.vm06.stdout:20 Mar 17:56:04 ntpd[16247]: Soliciting pool server 5.75.181.179 2026-03-20T17:56:04.998 INFO:teuthology.orchestra.run.vm06.stdout:20 Mar 17:56:04 ntpd[16247]: Soliciting pool server 129.70.132.35 2026-03-20T17:56:04.998 INFO:teuthology.orchestra.run.vm06.stdout:20 Mar 17:56:04 ntpd[16247]: Soliciting pool server 51.75.67.47 2026-03-20T17:56:05.030 INFO:teuthology.orchestra.run.vm08.stdout:20 Mar 17:56:05 ntpd[16255]: Soliciting pool server 49.12.35.6 2026-03-20T17:56:05.030 INFO:teuthology.orchestra.run.vm08.stdout:20 Mar 17:56:05 ntpd[16255]: Soliciting pool server 5.45.97.204 2026-03-20T17:56:05.030 INFO:teuthology.orchestra.run.vm08.stdout:20 Mar 17:56:05 ntpd[16255]: Soliciting pool server 131.188.3.223 2026-03-20T17:56:05.030 INFO:teuthology.orchestra.run.vm08.stdout:20 Mar 17:56:05 ntpd[16255]: Soliciting pool server 37.114.42.119 2026-03-20T17:56:05.995 INFO:teuthology.orchestra.run.vm02.stdout:20 Mar 17:56:05 ntpd[16237]: Soliciting pool server 217.91.44.17 2026-03-20T17:56:05.995 INFO:teuthology.orchestra.run.vm02.stdout:20 Mar 17:56:05 ntpd[16237]: Soliciting pool server 94.16.122.152 2026-03-20T17:56:05.996 INFO:teuthology.orchestra.run.vm02.stdout:20 Mar 17:56:05 ntpd[16237]: Soliciting pool server 78.46.238.113 2026-03-20T17:56:05.996 INFO:teuthology.orchestra.run.vm02.stdout:20 Mar 17:56:05 ntpd[16237]: Soliciting pool server 185.125.190.58 2026-03-20T17:56:05.996 INFO:teuthology.orchestra.run.vm05.stdout:20 Mar 17:56:05 ntpd[16231]: Soliciting pool server 217.91.44.17 2026-03-20T17:56:05.996 INFO:teuthology.orchestra.run.vm05.stdout:20 Mar 17:56:05 ntpd[16231]: Soliciting pool server 94.16.122.152 2026-03-20T17:56:05.996 INFO:teuthology.orchestra.run.vm05.stdout:20 Mar 17:56:05 ntpd[16231]: Soliciting pool server 78.46.238.113 2026-03-20T17:56:05.996 INFO:teuthology.orchestra.run.vm05.stdout:20 Mar 17:56:05 ntpd[16231]: Soliciting pool server 185.125.190.58 2026-03-20T17:56:05.998 INFO:teuthology.orchestra.run.vm06.stdout:20 Mar 17:56:05 ntpd[16247]: Soliciting pool server 5.189.151.39 2026-03-20T17:56:05.998 INFO:teuthology.orchestra.run.vm06.stdout:20 Mar 17:56:05 ntpd[16247]: Soliciting pool server 46.224.156.215 2026-03-20T17:56:05.998 INFO:teuthology.orchestra.run.vm06.stdout:20 Mar 17:56:05 ntpd[16247]: Soliciting pool server 78.46.238.113 2026-03-20T17:56:05.999 INFO:teuthology.orchestra.run.vm06.stdout:20 Mar 17:56:05 ntpd[16247]: Soliciting pool server 91.189.91.157 2026-03-20T17:56:06.030 INFO:teuthology.orchestra.run.vm08.stdout:20 Mar 17:56:06 ntpd[16255]: Soliciting pool server 51.75.67.47 2026-03-20T17:56:06.030 INFO:teuthology.orchestra.run.vm08.stdout:20 Mar 17:56:06 ntpd[16255]: Soliciting pool server 172.104.149.161 2026-03-20T17:56:06.030 INFO:teuthology.orchestra.run.vm08.stdout:20 Mar 17:56:06 ntpd[16255]: Soliciting pool server 5.75.181.179 2026-03-20T17:56:06.030 INFO:teuthology.orchestra.run.vm08.stdout:20 Mar 17:56:06 ntpd[16255]: Soliciting pool server 185.125.190.56 2026-03-20T17:56:06.995 INFO:teuthology.orchestra.run.vm02.stdout:20 Mar 17:56:06 ntpd[16237]: Soliciting pool server 185.125.190.57 2026-03-20T17:56:06.996 INFO:teuthology.orchestra.run.vm02.stdout:20 Mar 17:56:06 ntpd[16237]: Soliciting pool server 37.114.42.119 2026-03-20T17:56:06.996 INFO:teuthology.orchestra.run.vm02.stdout:20 Mar 17:56:06 ntpd[16237]: Soliciting pool server 49.12.35.6 2026-03-20T17:56:06.996 INFO:teuthology.orchestra.run.vm05.stdout:20 Mar 17:56:06 ntpd[16231]: Soliciting pool server 185.125.190.57 2026-03-20T17:56:06.996 INFO:teuthology.orchestra.run.vm05.stdout:20 Mar 17:56:06 ntpd[16231]: Soliciting pool server 37.114.42.119 2026-03-20T17:56:06.996 INFO:teuthology.orchestra.run.vm05.stdout:20 Mar 17:56:06 ntpd[16231]: Soliciting pool server 49.12.35.6 2026-03-20T17:56:06.998 INFO:teuthology.orchestra.run.vm06.stdout:20 Mar 17:56:06 ntpd[16247]: Soliciting pool server 185.125.190.58 2026-03-20T17:56:06.998 INFO:teuthology.orchestra.run.vm06.stdout:20 Mar 17:56:06 ntpd[16247]: Soliciting pool server 217.91.44.17 2026-03-20T17:56:06.998 INFO:teuthology.orchestra.run.vm06.stdout:20 Mar 17:56:06 ntpd[16247]: Soliciting pool server 94.16.122.152 2026-03-20T17:56:07.030 INFO:teuthology.orchestra.run.vm08.stdout:20 Mar 17:56:07 ntpd[16255]: Soliciting pool server 91.189.91.157 2026-03-20T17:56:07.030 INFO:teuthology.orchestra.run.vm08.stdout:20 Mar 17:56:07 ntpd[16255]: Soliciting pool server 5.189.151.39 2026-03-20T17:56:07.030 INFO:teuthology.orchestra.run.vm08.stdout:20 Mar 17:56:07 ntpd[16255]: Soliciting pool server 46.224.156.215 2026-03-20T17:56:09.027 INFO:teuthology.orchestra.run.vm02.stdout:20 Mar 17:56:09 ntpd[16237]: ntpd: time slew +0.000175 s 2026-03-20T17:56:09.027 INFO:teuthology.orchestra.run.vm02.stdout:ntpd: time slew +0.000175s 2026-03-20T17:56:09.027 INFO:teuthology.orchestra.run.vm05.stdout:20 Mar 17:56:09 ntpd[16231]: ntpd: time slew -0.000590 s 2026-03-20T17:56:09.028 INFO:teuthology.orchestra.run.vm05.stdout:ntpd: time slew -0.000590s 2026-03-20T17:56:09.047 INFO:teuthology.orchestra.run.vm02.stdout: remote refid st t when poll reach delay offset jitter 2026-03-20T17:56:09.047 INFO:teuthology.orchestra.run.vm02.stdout:============================================================================== 2026-03-20T17:56:09.047 INFO:teuthology.orchestra.run.vm02.stdout: 0.ubuntu.pool.n .POOL. 16 p - 64 0 0.000 +0.000 0.000 2026-03-20T17:56:09.047 INFO:teuthology.orchestra.run.vm02.stdout: 1.ubuntu.pool.n .POOL. 16 p - 64 0 0.000 +0.000 0.000 2026-03-20T17:56:09.047 INFO:teuthology.orchestra.run.vm02.stdout: 2.ubuntu.pool.n .POOL. 16 p - 64 0 0.000 +0.000 0.000 2026-03-20T17:56:09.047 INFO:teuthology.orchestra.run.vm02.stdout: 3.ubuntu.pool.n .POOL. 16 p - 64 0 0.000 +0.000 0.000 2026-03-20T17:56:09.047 INFO:teuthology.orchestra.run.vm02.stdout: ntp.ubuntu.com .POOL. 16 p - 64 0 0.000 +0.000 0.000 2026-03-20T17:56:09.048 INFO:teuthology.orchestra.run.vm05.stdout: remote refid st t when poll reach delay offset jitter 2026-03-20T17:56:09.048 INFO:teuthology.orchestra.run.vm05.stdout:============================================================================== 2026-03-20T17:56:09.048 INFO:teuthology.orchestra.run.vm05.stdout: 0.ubuntu.pool.n .POOL. 16 p - 64 0 0.000 +0.000 0.000 2026-03-20T17:56:09.048 INFO:teuthology.orchestra.run.vm05.stdout: 1.ubuntu.pool.n .POOL. 16 p - 64 0 0.000 +0.000 0.000 2026-03-20T17:56:09.048 INFO:teuthology.orchestra.run.vm05.stdout: 2.ubuntu.pool.n .POOL. 16 p - 64 0 0.000 +0.000 0.000 2026-03-20T17:56:09.048 INFO:teuthology.orchestra.run.vm05.stdout: 3.ubuntu.pool.n .POOL. 16 p - 64 0 0.000 +0.000 0.000 2026-03-20T17:56:09.048 INFO:teuthology.orchestra.run.vm05.stdout: ntp.ubuntu.com .POOL. 16 p - 64 0 0.000 +0.000 0.000 2026-03-20T17:56:10.022 INFO:teuthology.orchestra.run.vm06.stdout:20 Mar 17:56:10 ntpd[16247]: ntpd: time slew -0.000325 s 2026-03-20T17:56:10.022 INFO:teuthology.orchestra.run.vm06.stdout:ntpd: time slew -0.000325s 2026-03-20T17:56:10.042 INFO:teuthology.orchestra.run.vm06.stdout: remote refid st t when poll reach delay offset jitter 2026-03-20T17:56:10.042 INFO:teuthology.orchestra.run.vm06.stdout:============================================================================== 2026-03-20T17:56:10.042 INFO:teuthology.orchestra.run.vm06.stdout: 0.ubuntu.pool.n .POOL. 16 p - 64 0 0.000 +0.000 0.000 2026-03-20T17:56:10.042 INFO:teuthology.orchestra.run.vm06.stdout: 1.ubuntu.pool.n .POOL. 16 p - 64 0 0.000 +0.000 0.000 2026-03-20T17:56:10.042 INFO:teuthology.orchestra.run.vm06.stdout: 2.ubuntu.pool.n .POOL. 16 p - 64 0 0.000 +0.000 0.000 2026-03-20T17:56:10.042 INFO:teuthology.orchestra.run.vm06.stdout: 3.ubuntu.pool.n .POOL. 16 p - 64 0 0.000 +0.000 0.000 2026-03-20T17:56:10.042 INFO:teuthology.orchestra.run.vm06.stdout: ntp.ubuntu.com .POOL. 16 p - 64 0 0.000 +0.000 0.000 2026-03-20T17:56:11.062 INFO:teuthology.orchestra.run.vm08.stdout:20 Mar 17:56:11 ntpd[16255]: ntpd: time slew +0.004043 s 2026-03-20T17:56:11.062 INFO:teuthology.orchestra.run.vm08.stdout:ntpd: time slew +0.004043s 2026-03-20T17:56:11.082 INFO:teuthology.orchestra.run.vm08.stdout: remote refid st t when poll reach delay offset jitter 2026-03-20T17:56:11.082 INFO:teuthology.orchestra.run.vm08.stdout:============================================================================== 2026-03-20T17:56:11.082 INFO:teuthology.orchestra.run.vm08.stdout: 0.ubuntu.pool.n .POOL. 16 p - 64 0 0.000 +0.000 0.000 2026-03-20T17:56:11.082 INFO:teuthology.orchestra.run.vm08.stdout: 1.ubuntu.pool.n .POOL. 16 p - 64 0 0.000 +0.000 0.000 2026-03-20T17:56:11.082 INFO:teuthology.orchestra.run.vm08.stdout: 2.ubuntu.pool.n .POOL. 16 p - 64 0 0.000 +0.000 0.000 2026-03-20T17:56:11.082 INFO:teuthology.orchestra.run.vm08.stdout: 3.ubuntu.pool.n .POOL. 16 p - 64 0 0.000 +0.000 0.000 2026-03-20T17:56:11.082 INFO:teuthology.orchestra.run.vm08.stdout: ntp.ubuntu.com .POOL. 16 p - 64 0 0.000 +0.000 0.000 2026-03-20T17:56:11.083 INFO:teuthology.run_tasks:Running task install... 2026-03-20T17:56:11.084 DEBUG:teuthology.task.install:project ceph 2026-03-20T17:56:11.084 DEBUG:teuthology.task.install:INSTALL overrides: {'ceph': {'flavor': 'default', 'sha1': '70f8415b300f041766fa27faf7d5472699e32388'}, 'extra_system_packages': {'deb': ['python3-jmespath', 'python3-xmltodict', 's3cmd'], 'rpm': ['bzip2', 'perl-Test-Harness', 'python3-jmespath', 'python3-xmltodict', 's3cmd']}} 2026-03-20T17:56:11.084 DEBUG:teuthology.task.install:config {'flavor': 'default', 'sha1': '70f8415b300f041766fa27faf7d5472699e32388', 'extra_system_packages': {'deb': ['python3-jmespath', 'python3-xmltodict', 's3cmd'], 'rpm': ['bzip2', 'perl-Test-Harness', 'python3-jmespath', 'python3-xmltodict', 's3cmd']}} 2026-03-20T17:56:11.084 INFO:teuthology.task.install:Using flavor: default 2026-03-20T17:56:11.087 DEBUG:teuthology.task.install:Package list is: {'deb': ['ceph', 'cephadm', 'ceph-mds', 'ceph-mgr', 'ceph-common', 'ceph-fuse', 'ceph-test', 'ceph-volume', 'radosgw', 'python3-rados', 'python3-rgw', 'python3-cephfs', 'python3-rbd', 'libcephfs2', 'libcephfs-dev', 'librados2', 'librbd1', 'rbd-fuse'], 'rpm': ['ceph-radosgw', 'ceph-test', 'ceph', 'ceph-base', 'cephadm', 'ceph-immutable-object-cache', 'ceph-mgr', 'ceph-mgr-dashboard', 'ceph-mgr-diskprediction-local', 'ceph-mgr-rook', 'ceph-mgr-cephadm', 'ceph-fuse', 'ceph-volume', 'librados-devel', 'libcephfs2', 'libcephfs-devel', 'librados2', 'librbd1', 'python3-rados', 'python3-rgw', 'python3-cephfs', 'python3-rbd', 'rbd-fuse', 'rbd-mirror', 'rbd-nbd']} 2026-03-20T17:56:11.087 INFO:teuthology.task.install:extra packages: [] 2026-03-20T17:56:11.087 DEBUG:teuthology.orchestra.run.vm02:> sudo apt-key list | grep Ceph 2026-03-20T17:56:11.087 DEBUG:teuthology.orchestra.run.vm05:> sudo apt-key list | grep Ceph 2026-03-20T17:56:11.087 DEBUG:teuthology.orchestra.run.vm06:> sudo apt-key list | grep Ceph 2026-03-20T17:56:11.087 DEBUG:teuthology.orchestra.run.vm08:> sudo apt-key list | grep Ceph 2026-03-20T17:56:11.124 INFO:teuthology.orchestra.run.vm02.stderr:Warning: apt-key is deprecated. Manage keyring files in trusted.gpg.d instead (see apt-key(8)). 2026-03-20T17:56:11.125 INFO:teuthology.orchestra.run.vm06.stderr:Warning: apt-key is deprecated. Manage keyring files in trusted.gpg.d instead (see apt-key(8)). 2026-03-20T17:56:11.127 INFO:teuthology.orchestra.run.vm05.stderr:Warning: apt-key is deprecated. Manage keyring files in trusted.gpg.d instead (see apt-key(8)). 2026-03-20T17:56:11.144 INFO:teuthology.orchestra.run.vm02.stdout:uid [ unknown] Ceph automated package build (Ceph automated package build) 2026-03-20T17:56:11.145 INFO:teuthology.orchestra.run.vm02.stdout:uid [ unknown] Ceph.com (release key) 2026-03-20T17:56:11.145 INFO:teuthology.task.install.deb:Installing packages: ceph, cephadm, ceph-mds, ceph-mgr, ceph-common, ceph-fuse, ceph-test, ceph-volume, radosgw, python3-rados, python3-rgw, python3-cephfs, python3-rbd, libcephfs2, libcephfs-dev, librados2, librbd1, rbd-fuse on remote deb x86_64 2026-03-20T17:56:11.145 INFO:teuthology.task.install.deb:Installing system (non-project) packages: python3-jmespath, python3-xmltodict, s3cmd on remote deb x86_64 2026-03-20T17:56:11.145 DEBUG:teuthology.packaging:Querying https://shaman.ceph.com/api/search?status=ready&project=ceph&flavor=default&distros=ubuntu%2F22.04%2Fx86_64&sha1=70f8415b300f041766fa27faf7d5472699e32388 2026-03-20T17:56:11.206 INFO:teuthology.orchestra.run.vm08.stderr:Warning: apt-key is deprecated. Manage keyring files in trusted.gpg.d instead (see apt-key(8)). 2026-03-20T17:56:11.207 INFO:teuthology.orchestra.run.vm05.stdout:uid [ unknown] Ceph automated package build (Ceph automated package build) 2026-03-20T17:56:11.207 INFO:teuthology.orchestra.run.vm05.stdout:uid [ unknown] Ceph.com (release key) 2026-03-20T17:56:11.207 INFO:teuthology.orchestra.run.vm06.stdout:uid [ unknown] Ceph automated package build (Ceph automated package build) 2026-03-20T17:56:11.207 INFO:teuthology.orchestra.run.vm06.stdout:uid [ unknown] Ceph.com (release key) 2026-03-20T17:56:11.207 INFO:teuthology.orchestra.run.vm08.stdout:uid [ unknown] Ceph automated package build (Ceph automated package build) 2026-03-20T17:56:11.207 INFO:teuthology.orchestra.run.vm08.stdout:uid [ unknown] Ceph.com (release key) 2026-03-20T17:56:11.207 INFO:teuthology.task.install.deb:Installing packages: ceph, cephadm, ceph-mds, ceph-mgr, ceph-common, ceph-fuse, ceph-test, ceph-volume, radosgw, python3-rados, python3-rgw, python3-cephfs, python3-rbd, libcephfs2, libcephfs-dev, librados2, librbd1, rbd-fuse on remote deb x86_64 2026-03-20T17:56:11.207 INFO:teuthology.task.install.deb:Installing system (non-project) packages: python3-jmespath, python3-xmltodict, s3cmd on remote deb x86_64 2026-03-20T17:56:11.207 DEBUG:teuthology.packaging:Querying https://shaman.ceph.com/api/search?status=ready&project=ceph&flavor=default&distros=ubuntu%2F22.04%2Fx86_64&sha1=70f8415b300f041766fa27faf7d5472699e32388 2026-03-20T17:56:11.208 INFO:teuthology.task.install.deb:Installing packages: ceph, cephadm, ceph-mds, ceph-mgr, ceph-common, ceph-fuse, ceph-test, ceph-volume, radosgw, python3-rados, python3-rgw, python3-cephfs, python3-rbd, libcephfs2, libcephfs-dev, librados2, librbd1, rbd-fuse on remote deb x86_64 2026-03-20T17:56:11.208 INFO:teuthology.task.install.deb:Installing system (non-project) packages: python3-jmespath, python3-xmltodict, s3cmd on remote deb x86_64 2026-03-20T17:56:11.208 DEBUG:teuthology.packaging:Querying https://shaman.ceph.com/api/search?status=ready&project=ceph&flavor=default&distros=ubuntu%2F22.04%2Fx86_64&sha1=70f8415b300f041766fa27faf7d5472699e32388 2026-03-20T17:56:11.208 INFO:teuthology.task.install.deb:Installing packages: ceph, cephadm, ceph-mds, ceph-mgr, ceph-common, ceph-fuse, ceph-test, ceph-volume, radosgw, python3-rados, python3-rgw, python3-cephfs, python3-rbd, libcephfs2, libcephfs-dev, librados2, librbd1, rbd-fuse on remote deb x86_64 2026-03-20T17:56:11.208 INFO:teuthology.task.install.deb:Installing system (non-project) packages: python3-jmespath, python3-xmltodict, s3cmd on remote deb x86_64 2026-03-20T17:56:11.208 DEBUG:teuthology.packaging:Querying https://shaman.ceph.com/api/search?status=ready&project=ceph&flavor=default&distros=ubuntu%2F22.04%2Fx86_64&sha1=70f8415b300f041766fa27faf7d5472699e32388 2026-03-20T17:56:11.828 INFO:teuthology.task.install.deb:Pulling from https://1.chacra.ceph.com/r/ceph/tentacle/70f8415b300f041766fa27faf7d5472699e32388/ubuntu/jammy/flavors/default/ 2026-03-20T17:56:11.828 INFO:teuthology.task.install.deb:Package version is 20.2.0-712-g70f8415b-1jammy 2026-03-20T17:56:11.861 INFO:teuthology.task.install.deb:Pulling from https://1.chacra.ceph.com/r/ceph/tentacle/70f8415b300f041766fa27faf7d5472699e32388/ubuntu/jammy/flavors/default/ 2026-03-20T17:56:11.861 INFO:teuthology.task.install.deb:Package version is 20.2.0-712-g70f8415b-1jammy 2026-03-20T17:56:11.948 INFO:teuthology.task.install.deb:Pulling from https://1.chacra.ceph.com/r/ceph/tentacle/70f8415b300f041766fa27faf7d5472699e32388/ubuntu/jammy/flavors/default/ 2026-03-20T17:56:11.948 INFO:teuthology.task.install.deb:Package version is 20.2.0-712-g70f8415b-1jammy 2026-03-20T17:56:12.015 INFO:teuthology.task.install.deb:Pulling from https://1.chacra.ceph.com/r/ceph/tentacle/70f8415b300f041766fa27faf7d5472699e32388/ubuntu/jammy/flavors/default/ 2026-03-20T17:56:12.015 INFO:teuthology.task.install.deb:Package version is 20.2.0-712-g70f8415b-1jammy 2026-03-20T17:56:12.327 DEBUG:teuthology.orchestra.run.vm05:> set -ex 2026-03-20T17:56:12.327 DEBUG:teuthology.orchestra.run.vm05:> sudo dd of=/etc/apt/sources.list.d/ceph.list 2026-03-20T17:56:12.336 DEBUG:teuthology.orchestra.run.vm05:> sudo apt-get update 2026-03-20T17:56:12.351 DEBUG:teuthology.orchestra.run.vm02:> set -ex 2026-03-20T17:56:12.351 DEBUG:teuthology.orchestra.run.vm02:> sudo dd of=/etc/apt/sources.list.d/ceph.list 2026-03-20T17:56:12.359 DEBUG:teuthology.orchestra.run.vm02:> sudo apt-get update 2026-03-20T17:56:12.435 DEBUG:teuthology.orchestra.run.vm06:> set -ex 2026-03-20T17:56:12.435 DEBUG:teuthology.orchestra.run.vm06:> sudo dd of=/etc/apt/sources.list.d/ceph.list 2026-03-20T17:56:12.444 DEBUG:teuthology.orchestra.run.vm06:> sudo apt-get update 2026-03-20T17:56:12.460 INFO:teuthology.orchestra.run.vm05.stdout:Hit:1 http://security.ubuntu.com/ubuntu jammy-security InRelease 2026-03-20T17:56:12.471 INFO:teuthology.orchestra.run.vm02.stdout:Hit:1 http://archive.ubuntu.com/ubuntu jammy InRelease 2026-03-20T17:56:12.475 INFO:teuthology.orchestra.run.vm02.stdout:Hit:2 http://archive.ubuntu.com/ubuntu jammy-updates InRelease 2026-03-20T17:56:12.482 INFO:teuthology.orchestra.run.vm02.stdout:Hit:3 http://archive.ubuntu.com/ubuntu jammy-backports InRelease 2026-03-20T17:56:12.487 DEBUG:teuthology.orchestra.run.vm08:> set -ex 2026-03-20T17:56:12.487 DEBUG:teuthology.orchestra.run.vm08:> sudo dd of=/etc/apt/sources.list.d/ceph.list 2026-03-20T17:56:12.497 DEBUG:teuthology.orchestra.run.vm08:> sudo apt-get update 2026-03-20T17:56:12.526 INFO:teuthology.orchestra.run.vm02.stdout:Hit:4 http://security.ubuntu.com/ubuntu jammy-security InRelease 2026-03-20T17:56:12.620 INFO:teuthology.orchestra.run.vm06.stdout:Hit:1 http://archive.ubuntu.com/ubuntu jammy InRelease 2026-03-20T17:56:12.621 INFO:teuthology.orchestra.run.vm08.stdout:Hit:1 http://archive.ubuntu.com/ubuntu jammy InRelease 2026-03-20T17:56:12.624 INFO:teuthology.orchestra.run.vm08.stdout:Hit:2 http://archive.ubuntu.com/ubuntu jammy-updates InRelease 2026-03-20T17:56:12.631 INFO:teuthology.orchestra.run.vm08.stdout:Hit:3 http://archive.ubuntu.com/ubuntu jammy-backports InRelease 2026-03-20T17:56:12.650 INFO:teuthology.orchestra.run.vm06.stdout:Hit:2 http://archive.ubuntu.com/ubuntu jammy-updates InRelease 2026-03-20T17:56:12.674 INFO:teuthology.orchestra.run.vm08.stdout:Hit:4 http://security.ubuntu.com/ubuntu jammy-security InRelease 2026-03-20T17:56:12.686 INFO:teuthology.orchestra.run.vm06.stdout:Hit:3 http://archive.ubuntu.com/ubuntu jammy-backports InRelease 2026-03-20T17:56:12.699 INFO:teuthology.orchestra.run.vm05.stdout:Hit:2 http://archive.ubuntu.com/ubuntu jammy InRelease 2026-03-20T17:56:12.736 INFO:teuthology.orchestra.run.vm06.stdout:Hit:4 http://security.ubuntu.com/ubuntu jammy-security InRelease 2026-03-20T17:56:12.812 INFO:teuthology.orchestra.run.vm05.stdout:Hit:3 http://archive.ubuntu.com/ubuntu jammy-updates InRelease 2026-03-20T17:56:12.925 INFO:teuthology.orchestra.run.vm05.stdout:Hit:4 http://archive.ubuntu.com/ubuntu jammy-backports InRelease 2026-03-20T17:56:13.030 INFO:teuthology.orchestra.run.vm05.stdout:Ign:5 https://1.chacra.ceph.com/r/ceph/tentacle/70f8415b300f041766fa27faf7d5472699e32388/ubuntu/jammy/flavors/default jammy InRelease 2026-03-20T17:56:13.031 INFO:teuthology.orchestra.run.vm06.stdout:Ign:5 https://1.chacra.ceph.com/r/ceph/tentacle/70f8415b300f041766fa27faf7d5472699e32388/ubuntu/jammy/flavors/default jammy InRelease 2026-03-20T17:56:13.037 INFO:teuthology.orchestra.run.vm02.stdout:Ign:5 https://1.chacra.ceph.com/r/ceph/tentacle/70f8415b300f041766fa27faf7d5472699e32388/ubuntu/jammy/flavors/default jammy InRelease 2026-03-20T17:56:13.062 INFO:teuthology.orchestra.run.vm08.stdout:Ign:5 https://1.chacra.ceph.com/r/ceph/tentacle/70f8415b300f041766fa27faf7d5472699e32388/ubuntu/jammy/flavors/default jammy InRelease 2026-03-20T17:56:13.146 INFO:teuthology.orchestra.run.vm05.stdout:Get:6 https://1.chacra.ceph.com/r/ceph/tentacle/70f8415b300f041766fa27faf7d5472699e32388/ubuntu/jammy/flavors/default jammy Release [7680 B] 2026-03-20T17:56:13.146 INFO:teuthology.orchestra.run.vm06.stdout:Get:6 https://1.chacra.ceph.com/r/ceph/tentacle/70f8415b300f041766fa27faf7d5472699e32388/ubuntu/jammy/flavors/default jammy Release [7680 B] 2026-03-20T17:56:13.155 INFO:teuthology.orchestra.run.vm02.stdout:Get:6 https://1.chacra.ceph.com/r/ceph/tentacle/70f8415b300f041766fa27faf7d5472699e32388/ubuntu/jammy/flavors/default jammy Release [7680 B] 2026-03-20T17:56:13.176 INFO:teuthology.orchestra.run.vm08.stdout:Get:6 https://1.chacra.ceph.com/r/ceph/tentacle/70f8415b300f041766fa27faf7d5472699e32388/ubuntu/jammy/flavors/default jammy Release [7680 B] 2026-03-20T17:56:13.261 INFO:teuthology.orchestra.run.vm06.stdout:Ign:7 https://1.chacra.ceph.com/r/ceph/tentacle/70f8415b300f041766fa27faf7d5472699e32388/ubuntu/jammy/flavors/default jammy Release.gpg 2026-03-20T17:56:13.262 INFO:teuthology.orchestra.run.vm05.stdout:Ign:7 https://1.chacra.ceph.com/r/ceph/tentacle/70f8415b300f041766fa27faf7d5472699e32388/ubuntu/jammy/flavors/default jammy Release.gpg 2026-03-20T17:56:13.272 INFO:teuthology.orchestra.run.vm02.stdout:Ign:7 https://1.chacra.ceph.com/r/ceph/tentacle/70f8415b300f041766fa27faf7d5472699e32388/ubuntu/jammy/flavors/default jammy Release.gpg 2026-03-20T17:56:13.291 INFO:teuthology.orchestra.run.vm08.stdout:Ign:7 https://1.chacra.ceph.com/r/ceph/tentacle/70f8415b300f041766fa27faf7d5472699e32388/ubuntu/jammy/flavors/default jammy Release.gpg 2026-03-20T17:56:13.376 INFO:teuthology.orchestra.run.vm06.stdout:Get:8 https://1.chacra.ceph.com/r/ceph/tentacle/70f8415b300f041766fa27faf7d5472699e32388/ubuntu/jammy/flavors/default jammy/main amd64 Packages [18.8 kB] 2026-03-20T17:56:13.377 INFO:teuthology.orchestra.run.vm05.stdout:Get:8 https://1.chacra.ceph.com/r/ceph/tentacle/70f8415b300f041766fa27faf7d5472699e32388/ubuntu/jammy/flavors/default jammy/main amd64 Packages [18.8 kB] 2026-03-20T17:56:13.389 INFO:teuthology.orchestra.run.vm02.stdout:Get:8 https://1.chacra.ceph.com/r/ceph/tentacle/70f8415b300f041766fa27faf7d5472699e32388/ubuntu/jammy/flavors/default jammy/main amd64 Packages [18.8 kB] 2026-03-20T17:56:13.405 INFO:teuthology.orchestra.run.vm08.stdout:Get:8 https://1.chacra.ceph.com/r/ceph/tentacle/70f8415b300f041766fa27faf7d5472699e32388/ubuntu/jammy/flavors/default jammy/main amd64 Packages [18.8 kB] 2026-03-20T17:56:13.449 INFO:teuthology.orchestra.run.vm06.stdout:Fetched 26.5 kB in 1s (31.2 kB/s) 2026-03-20T17:56:13.465 INFO:teuthology.orchestra.run.vm02.stdout:Fetched 26.5 kB in 1s (27.8 kB/s) 2026-03-20T17:56:13.477 INFO:teuthology.orchestra.run.vm08.stdout:Fetched 26.5 kB in 1s (32.2 kB/s) 2026-03-20T17:56:13.568 INFO:teuthology.orchestra.run.vm05.stdout:Fetched 26.5 kB in 1s (24.8 kB/s) 2026-03-20T17:56:14.078 INFO:teuthology.orchestra.run.vm06.stdout:Reading package lists... 2026-03-20T17:56:14.090 DEBUG:teuthology.orchestra.run.vm06:> sudo DEBIAN_FRONTEND=noninteractive apt-get -y --force-yes -o Dpkg::Options::="--force-confdef" -o Dpkg::Options::="--force-confold" install ceph=20.2.0-712-g70f8415b-1jammy cephadm=20.2.0-712-g70f8415b-1jammy ceph-mds=20.2.0-712-g70f8415b-1jammy ceph-mgr=20.2.0-712-g70f8415b-1jammy ceph-common=20.2.0-712-g70f8415b-1jammy ceph-fuse=20.2.0-712-g70f8415b-1jammy ceph-test=20.2.0-712-g70f8415b-1jammy ceph-volume=20.2.0-712-g70f8415b-1jammy radosgw=20.2.0-712-g70f8415b-1jammy python3-rados=20.2.0-712-g70f8415b-1jammy python3-rgw=20.2.0-712-g70f8415b-1jammy python3-cephfs=20.2.0-712-g70f8415b-1jammy python3-rbd=20.2.0-712-g70f8415b-1jammy libcephfs2=20.2.0-712-g70f8415b-1jammy libcephfs-dev=20.2.0-712-g70f8415b-1jammy librados2=20.2.0-712-g70f8415b-1jammy librbd1=20.2.0-712-g70f8415b-1jammy rbd-fuse=20.2.0-712-g70f8415b-1jammy 2026-03-20T17:56:14.097 INFO:teuthology.orchestra.run.vm02.stdout:Reading package lists... 2026-03-20T17:56:14.112 DEBUG:teuthology.orchestra.run.vm02:> sudo DEBIAN_FRONTEND=noninteractive apt-get -y --force-yes -o Dpkg::Options::="--force-confdef" -o Dpkg::Options::="--force-confold" install ceph=20.2.0-712-g70f8415b-1jammy cephadm=20.2.0-712-g70f8415b-1jammy ceph-mds=20.2.0-712-g70f8415b-1jammy ceph-mgr=20.2.0-712-g70f8415b-1jammy ceph-common=20.2.0-712-g70f8415b-1jammy ceph-fuse=20.2.0-712-g70f8415b-1jammy ceph-test=20.2.0-712-g70f8415b-1jammy ceph-volume=20.2.0-712-g70f8415b-1jammy radosgw=20.2.0-712-g70f8415b-1jammy python3-rados=20.2.0-712-g70f8415b-1jammy python3-rgw=20.2.0-712-g70f8415b-1jammy python3-cephfs=20.2.0-712-g70f8415b-1jammy python3-rbd=20.2.0-712-g70f8415b-1jammy libcephfs2=20.2.0-712-g70f8415b-1jammy libcephfs-dev=20.2.0-712-g70f8415b-1jammy librados2=20.2.0-712-g70f8415b-1jammy librbd1=20.2.0-712-g70f8415b-1jammy rbd-fuse=20.2.0-712-g70f8415b-1jammy 2026-03-20T17:56:14.117 INFO:teuthology.orchestra.run.vm08.stdout:Reading package lists... 2026-03-20T17:56:14.126 INFO:teuthology.orchestra.run.vm06.stdout:Reading package lists... 2026-03-20T17:56:14.129 DEBUG:teuthology.orchestra.run.vm08:> sudo DEBIAN_FRONTEND=noninteractive apt-get -y --force-yes -o Dpkg::Options::="--force-confdef" -o Dpkg::Options::="--force-confold" install ceph=20.2.0-712-g70f8415b-1jammy cephadm=20.2.0-712-g70f8415b-1jammy ceph-mds=20.2.0-712-g70f8415b-1jammy ceph-mgr=20.2.0-712-g70f8415b-1jammy ceph-common=20.2.0-712-g70f8415b-1jammy ceph-fuse=20.2.0-712-g70f8415b-1jammy ceph-test=20.2.0-712-g70f8415b-1jammy ceph-volume=20.2.0-712-g70f8415b-1jammy radosgw=20.2.0-712-g70f8415b-1jammy python3-rados=20.2.0-712-g70f8415b-1jammy python3-rgw=20.2.0-712-g70f8415b-1jammy python3-cephfs=20.2.0-712-g70f8415b-1jammy python3-rbd=20.2.0-712-g70f8415b-1jammy libcephfs2=20.2.0-712-g70f8415b-1jammy libcephfs-dev=20.2.0-712-g70f8415b-1jammy librados2=20.2.0-712-g70f8415b-1jammy librbd1=20.2.0-712-g70f8415b-1jammy rbd-fuse=20.2.0-712-g70f8415b-1jammy 2026-03-20T17:56:14.147 INFO:teuthology.orchestra.run.vm02.stdout:Reading package lists... 2026-03-20T17:56:14.162 INFO:teuthology.orchestra.run.vm08.stdout:Reading package lists... 2026-03-20T17:56:14.240 INFO:teuthology.orchestra.run.vm05.stdout:Reading package lists... 2026-03-20T17:56:14.254 DEBUG:teuthology.orchestra.run.vm05:> sudo DEBIAN_FRONTEND=noninteractive apt-get -y --force-yes -o Dpkg::Options::="--force-confdef" -o Dpkg::Options::="--force-confold" install ceph=20.2.0-712-g70f8415b-1jammy cephadm=20.2.0-712-g70f8415b-1jammy ceph-mds=20.2.0-712-g70f8415b-1jammy ceph-mgr=20.2.0-712-g70f8415b-1jammy ceph-common=20.2.0-712-g70f8415b-1jammy ceph-fuse=20.2.0-712-g70f8415b-1jammy ceph-test=20.2.0-712-g70f8415b-1jammy ceph-volume=20.2.0-712-g70f8415b-1jammy radosgw=20.2.0-712-g70f8415b-1jammy python3-rados=20.2.0-712-g70f8415b-1jammy python3-rgw=20.2.0-712-g70f8415b-1jammy python3-cephfs=20.2.0-712-g70f8415b-1jammy python3-rbd=20.2.0-712-g70f8415b-1jammy libcephfs2=20.2.0-712-g70f8415b-1jammy libcephfs-dev=20.2.0-712-g70f8415b-1jammy librados2=20.2.0-712-g70f8415b-1jammy librbd1=20.2.0-712-g70f8415b-1jammy rbd-fuse=20.2.0-712-g70f8415b-1jammy 2026-03-20T17:56:14.288 INFO:teuthology.orchestra.run.vm05.stdout:Reading package lists... 2026-03-20T17:56:14.312 INFO:teuthology.orchestra.run.vm06.stdout:Building dependency tree... 2026-03-20T17:56:14.312 INFO:teuthology.orchestra.run.vm06.stdout:Reading state information... 2026-03-20T17:56:14.334 INFO:teuthology.orchestra.run.vm02.stdout:Building dependency tree... 2026-03-20T17:56:14.334 INFO:teuthology.orchestra.run.vm02.stdout:Reading state information... 2026-03-20T17:56:14.349 INFO:teuthology.orchestra.run.vm08.stdout:Building dependency tree... 2026-03-20T17:56:14.349 INFO:teuthology.orchestra.run.vm08.stdout:Reading state information... 2026-03-20T17:56:14.467 INFO:teuthology.orchestra.run.vm06.stdout:The following packages were automatically installed and are no longer required: 2026-03-20T17:56:14.467 INFO:teuthology.orchestra.run.vm06.stdout: kpartx libboost-iostreams1.74.0 libboost-thread1.74.0 libpmemobj1 2026-03-20T17:56:14.467 INFO:teuthology.orchestra.run.vm06.stdout: libsgutils2-2 sg3-utils sg3-utils-udev 2026-03-20T17:56:14.467 INFO:teuthology.orchestra.run.vm06.stdout:Use 'sudo apt autoremove' to remove them. 2026-03-20T17:56:14.467 INFO:teuthology.orchestra.run.vm06.stdout:The following additional packages will be installed: 2026-03-20T17:56:14.467 INFO:teuthology.orchestra.run.vm06.stdout: ceph-base ceph-mgr-cephadm ceph-mgr-dashboard ceph-mgr-diskprediction-local 2026-03-20T17:56:14.468 INFO:teuthology.orchestra.run.vm06.stdout: ceph-mgr-k8sevents ceph-mgr-modules-core ceph-mon ceph-osd jq 2026-03-20T17:56:14.468 INFO:teuthology.orchestra.run.vm06.stdout: libcephfs-daemon libcephfs-proxy2 libdouble-conversion3 libfuse2 libjq1 2026-03-20T17:56:14.468 INFO:teuthology.orchestra.run.vm06.stdout: liblttng-ust1 libnbd0 liboath0 libonig5 libpcre2-16-0 libqt5core5a 2026-03-20T17:56:14.468 INFO:teuthology.orchestra.run.vm06.stdout: libqt5dbus5 libqt5network5 libradosstriper1 librdkafka1 librgw2 2026-03-20T17:56:14.468 INFO:teuthology.orchestra.run.vm06.stdout: libsqlite3-mod-ceph libthrift-0.16.0 nvme-cli python-asyncssh-doc 2026-03-20T17:56:14.468 INFO:teuthology.orchestra.run.vm06.stdout: python3-asyncssh python3-cachetools python3-ceph-argparse 2026-03-20T17:56:14.468 INFO:teuthology.orchestra.run.vm06.stdout: python3-ceph-common python3-cheroot python3-cherrypy3 python3-google-auth 2026-03-20T17:56:14.468 INFO:teuthology.orchestra.run.vm06.stdout: python3-iniconfig python3-jaraco.classes python3-jaraco.collections 2026-03-20T17:56:14.468 INFO:teuthology.orchestra.run.vm06.stdout: python3-jaraco.functools python3-jaraco.text python3-joblib 2026-03-20T17:56:14.468 INFO:teuthology.orchestra.run.vm06.stdout: python3-kubernetes python3-natsort python3-pluggy python3-portend 2026-03-20T17:56:14.468 INFO:teuthology.orchestra.run.vm06.stdout: python3-prettytable python3-psutil python3-py python3-pygments 2026-03-20T17:56:14.468 INFO:teuthology.orchestra.run.vm06.stdout: python3-pytest python3-repoze.lru python3-requests-oauthlib python3-routes 2026-03-20T17:56:14.468 INFO:teuthology.orchestra.run.vm06.stdout: python3-rsa python3-simplejson python3-sklearn python3-sklearn-lib 2026-03-20T17:56:14.468 INFO:teuthology.orchestra.run.vm06.stdout: python3-tempora python3-threadpoolctl python3-toml python3-wcwidth 2026-03-20T17:56:14.468 INFO:teuthology.orchestra.run.vm06.stdout: python3-webob python3-websocket python3-zc.lockfile qttranslations5-l10n 2026-03-20T17:56:14.468 INFO:teuthology.orchestra.run.vm06.stdout: smartmontools socat xmlstarlet 2026-03-20T17:56:14.469 INFO:teuthology.orchestra.run.vm06.stdout:Suggested packages: 2026-03-20T17:56:14.469 INFO:teuthology.orchestra.run.vm06.stdout: python3-influxdb liblua5.3-dev luarocks python-natsort-doc python-psutil-doc 2026-03-20T17:56:14.469 INFO:teuthology.orchestra.run.vm06.stdout: subversion python-pygments-doc ttf-bitstream-vera python3-paste python3-dap 2026-03-20T17:56:14.469 INFO:teuthology.orchestra.run.vm06.stdout: python-sklearn-doc ipython3 python-webob-doc gsmartcontrol smart-notifier 2026-03-20T17:56:14.469 INFO:teuthology.orchestra.run.vm06.stdout: mailx | mailutils 2026-03-20T17:56:14.469 INFO:teuthology.orchestra.run.vm06.stdout:Recommended packages: 2026-03-20T17:56:14.469 INFO:teuthology.orchestra.run.vm06.stdout: btrfs-tools 2026-03-20T17:56:14.495 INFO:teuthology.orchestra.run.vm05.stdout:Building dependency tree... 2026-03-20T17:56:14.496 INFO:teuthology.orchestra.run.vm05.stdout:Reading state information... 2026-03-20T17:56:14.513 INFO:teuthology.orchestra.run.vm06.stdout:The following NEW packages will be installed: 2026-03-20T17:56:14.513 INFO:teuthology.orchestra.run.vm06.stdout: ceph ceph-base ceph-common ceph-fuse ceph-mds ceph-mgr ceph-mgr-cephadm 2026-03-20T17:56:14.513 INFO:teuthology.orchestra.run.vm06.stdout: ceph-mgr-dashboard ceph-mgr-diskprediction-local ceph-mgr-k8sevents 2026-03-20T17:56:14.513 INFO:teuthology.orchestra.run.vm06.stdout: ceph-mgr-modules-core ceph-mon ceph-osd ceph-test ceph-volume cephadm jq 2026-03-20T17:56:14.513 INFO:teuthology.orchestra.run.vm06.stdout: libcephfs-daemon libcephfs-dev libcephfs-proxy2 libcephfs2 2026-03-20T17:56:14.513 INFO:teuthology.orchestra.run.vm06.stdout: libdouble-conversion3 libfuse2 libjq1 liblttng-ust1 libnbd0 liboath0 2026-03-20T17:56:14.513 INFO:teuthology.orchestra.run.vm06.stdout: libonig5 libpcre2-16-0 libqt5core5a libqt5dbus5 libqt5network5 2026-03-20T17:56:14.513 INFO:teuthology.orchestra.run.vm06.stdout: libradosstriper1 librdkafka1 librgw2 libsqlite3-mod-ceph libthrift-0.16.0 2026-03-20T17:56:14.514 INFO:teuthology.orchestra.run.vm06.stdout: nvme-cli python-asyncssh-doc python3-asyncssh python3-cachetools 2026-03-20T17:56:14.514 INFO:teuthology.orchestra.run.vm06.stdout: python3-ceph-argparse python3-ceph-common python3-cephfs python3-cheroot 2026-03-20T17:56:14.514 INFO:teuthology.orchestra.run.vm06.stdout: python3-cherrypy3 python3-google-auth python3-iniconfig 2026-03-20T17:56:14.514 INFO:teuthology.orchestra.run.vm06.stdout: python3-jaraco.classes python3-jaraco.collections python3-jaraco.functools 2026-03-20T17:56:14.514 INFO:teuthology.orchestra.run.vm06.stdout: python3-jaraco.text python3-joblib python3-kubernetes python3-natsort 2026-03-20T17:56:14.514 INFO:teuthology.orchestra.run.vm06.stdout: python3-pluggy python3-portend python3-prettytable python3-psutil python3-py 2026-03-20T17:56:14.514 INFO:teuthology.orchestra.run.vm06.stdout: python3-pygments python3-pytest python3-rados python3-rbd python3-repoze.lru 2026-03-20T17:56:14.514 INFO:teuthology.orchestra.run.vm06.stdout: python3-requests-oauthlib python3-rgw python3-routes python3-rsa 2026-03-20T17:56:14.514 INFO:teuthology.orchestra.run.vm06.stdout: python3-simplejson python3-sklearn python3-sklearn-lib python3-tempora 2026-03-20T17:56:14.514 INFO:teuthology.orchestra.run.vm06.stdout: python3-threadpoolctl python3-toml python3-wcwidth python3-webob 2026-03-20T17:56:14.514 INFO:teuthology.orchestra.run.vm06.stdout: python3-websocket python3-zc.lockfile qttranslations5-l10n radosgw rbd-fuse 2026-03-20T17:56:14.514 INFO:teuthology.orchestra.run.vm06.stdout: smartmontools socat xmlstarlet 2026-03-20T17:56:14.515 INFO:teuthology.orchestra.run.vm06.stdout:The following packages will be upgraded: 2026-03-20T17:56:14.515 INFO:teuthology.orchestra.run.vm06.stdout: librados2 librbd1 2026-03-20T17:56:14.532 INFO:teuthology.orchestra.run.vm02.stdout:The following packages were automatically installed and are no longer required: 2026-03-20T17:56:14.532 INFO:teuthology.orchestra.run.vm02.stdout: kpartx libboost-iostreams1.74.0 libboost-thread1.74.0 libpmemobj1 2026-03-20T17:56:14.532 INFO:teuthology.orchestra.run.vm02.stdout: libsgutils2-2 sg3-utils sg3-utils-udev 2026-03-20T17:56:14.532 INFO:teuthology.orchestra.run.vm02.stdout:Use 'sudo apt autoremove' to remove them. 2026-03-20T17:56:14.533 INFO:teuthology.orchestra.run.vm02.stdout:The following additional packages will be installed: 2026-03-20T17:56:14.533 INFO:teuthology.orchestra.run.vm02.stdout: ceph-base ceph-mgr-cephadm ceph-mgr-dashboard ceph-mgr-diskprediction-local 2026-03-20T17:56:14.533 INFO:teuthology.orchestra.run.vm02.stdout: ceph-mgr-k8sevents ceph-mgr-modules-core ceph-mon ceph-osd jq 2026-03-20T17:56:14.533 INFO:teuthology.orchestra.run.vm02.stdout: libcephfs-daemon libcephfs-proxy2 libdouble-conversion3 libfuse2 libjq1 2026-03-20T17:56:14.533 INFO:teuthology.orchestra.run.vm02.stdout: liblttng-ust1 libnbd0 liboath0 libonig5 libpcre2-16-0 libqt5core5a 2026-03-20T17:56:14.533 INFO:teuthology.orchestra.run.vm02.stdout: libqt5dbus5 libqt5network5 libradosstriper1 librdkafka1 librgw2 2026-03-20T17:56:14.534 INFO:teuthology.orchestra.run.vm02.stdout: libsqlite3-mod-ceph libthrift-0.16.0 nvme-cli python-asyncssh-doc 2026-03-20T17:56:14.534 INFO:teuthology.orchestra.run.vm02.stdout: python3-asyncssh python3-cachetools python3-ceph-argparse 2026-03-20T17:56:14.534 INFO:teuthology.orchestra.run.vm02.stdout: python3-ceph-common python3-cheroot python3-cherrypy3 python3-google-auth 2026-03-20T17:56:14.534 INFO:teuthology.orchestra.run.vm02.stdout: python3-iniconfig python3-jaraco.classes python3-jaraco.collections 2026-03-20T17:56:14.534 INFO:teuthology.orchestra.run.vm02.stdout: python3-jaraco.functools python3-jaraco.text python3-joblib 2026-03-20T17:56:14.534 INFO:teuthology.orchestra.run.vm02.stdout: python3-kubernetes python3-natsort python3-pluggy python3-portend 2026-03-20T17:56:14.534 INFO:teuthology.orchestra.run.vm02.stdout: python3-prettytable python3-psutil python3-py python3-pygments 2026-03-20T17:56:14.534 INFO:teuthology.orchestra.run.vm02.stdout: python3-pytest python3-repoze.lru python3-requests-oauthlib python3-routes 2026-03-20T17:56:14.534 INFO:teuthology.orchestra.run.vm02.stdout: python3-rsa python3-simplejson python3-sklearn python3-sklearn-lib 2026-03-20T17:56:14.534 INFO:teuthology.orchestra.run.vm02.stdout: python3-tempora python3-threadpoolctl python3-toml python3-wcwidth 2026-03-20T17:56:14.534 INFO:teuthology.orchestra.run.vm02.stdout: python3-webob python3-websocket python3-zc.lockfile qttranslations5-l10n 2026-03-20T17:56:14.534 INFO:teuthology.orchestra.run.vm02.stdout: smartmontools socat xmlstarlet 2026-03-20T17:56:14.534 INFO:teuthology.orchestra.run.vm02.stdout:Suggested packages: 2026-03-20T17:56:14.534 INFO:teuthology.orchestra.run.vm02.stdout: python3-influxdb liblua5.3-dev luarocks python-natsort-doc python-psutil-doc 2026-03-20T17:56:14.534 INFO:teuthology.orchestra.run.vm02.stdout: subversion python-pygments-doc ttf-bitstream-vera python3-paste python3-dap 2026-03-20T17:56:14.534 INFO:teuthology.orchestra.run.vm02.stdout: python-sklearn-doc ipython3 python-webob-doc gsmartcontrol smart-notifier 2026-03-20T17:56:14.534 INFO:teuthology.orchestra.run.vm02.stdout: mailx | mailutils 2026-03-20T17:56:14.534 INFO:teuthology.orchestra.run.vm02.stdout:Recommended packages: 2026-03-20T17:56:14.534 INFO:teuthology.orchestra.run.vm02.stdout: btrfs-tools 2026-03-20T17:56:14.540 INFO:teuthology.orchestra.run.vm08.stdout:The following packages were automatically installed and are no longer required: 2026-03-20T17:56:14.540 INFO:teuthology.orchestra.run.vm08.stdout: kpartx libboost-iostreams1.74.0 libboost-thread1.74.0 libpmemobj1 2026-03-20T17:56:14.540 INFO:teuthology.orchestra.run.vm08.stdout: libsgutils2-2 sg3-utils sg3-utils-udev 2026-03-20T17:56:14.540 INFO:teuthology.orchestra.run.vm08.stdout:Use 'sudo apt autoremove' to remove them. 2026-03-20T17:56:14.541 INFO:teuthology.orchestra.run.vm08.stdout:The following additional packages will be installed: 2026-03-20T17:56:14.541 INFO:teuthology.orchestra.run.vm08.stdout: ceph-base ceph-mgr-cephadm ceph-mgr-dashboard ceph-mgr-diskprediction-local 2026-03-20T17:56:14.541 INFO:teuthology.orchestra.run.vm08.stdout: ceph-mgr-k8sevents ceph-mgr-modules-core ceph-mon ceph-osd jq 2026-03-20T17:56:14.541 INFO:teuthology.orchestra.run.vm08.stdout: libcephfs-daemon libcephfs-proxy2 libdouble-conversion3 libfuse2 libjq1 2026-03-20T17:56:14.541 INFO:teuthology.orchestra.run.vm08.stdout: liblttng-ust1 libnbd0 liboath0 libonig5 libpcre2-16-0 libqt5core5a 2026-03-20T17:56:14.541 INFO:teuthology.orchestra.run.vm08.stdout: libqt5dbus5 libqt5network5 libradosstriper1 librdkafka1 librgw2 2026-03-20T17:56:14.541 INFO:teuthology.orchestra.run.vm08.stdout: libsqlite3-mod-ceph libthrift-0.16.0 nvme-cli python-asyncssh-doc 2026-03-20T17:56:14.541 INFO:teuthology.orchestra.run.vm08.stdout: python3-asyncssh python3-cachetools python3-ceph-argparse 2026-03-20T17:56:14.541 INFO:teuthology.orchestra.run.vm08.stdout: python3-ceph-common python3-cheroot python3-cherrypy3 python3-google-auth 2026-03-20T17:56:14.541 INFO:teuthology.orchestra.run.vm08.stdout: python3-iniconfig python3-jaraco.classes python3-jaraco.collections 2026-03-20T17:56:14.541 INFO:teuthology.orchestra.run.vm08.stdout: python3-jaraco.functools python3-jaraco.text python3-joblib 2026-03-20T17:56:14.541 INFO:teuthology.orchestra.run.vm08.stdout: python3-kubernetes python3-natsort python3-pluggy python3-portend 2026-03-20T17:56:14.541 INFO:teuthology.orchestra.run.vm08.stdout: python3-prettytable python3-psutil python3-py python3-pygments 2026-03-20T17:56:14.541 INFO:teuthology.orchestra.run.vm08.stdout: python3-pytest python3-repoze.lru python3-requests-oauthlib python3-routes 2026-03-20T17:56:14.542 INFO:teuthology.orchestra.run.vm08.stdout: python3-rsa python3-simplejson python3-sklearn python3-sklearn-lib 2026-03-20T17:56:14.542 INFO:teuthology.orchestra.run.vm08.stdout: python3-tempora python3-threadpoolctl python3-toml python3-wcwidth 2026-03-20T17:56:14.542 INFO:teuthology.orchestra.run.vm08.stdout: python3-webob python3-websocket python3-zc.lockfile qttranslations5-l10n 2026-03-20T17:56:14.542 INFO:teuthology.orchestra.run.vm08.stdout: smartmontools socat xmlstarlet 2026-03-20T17:56:14.542 INFO:teuthology.orchestra.run.vm08.stdout:Suggested packages: 2026-03-20T17:56:14.542 INFO:teuthology.orchestra.run.vm08.stdout: python3-influxdb liblua5.3-dev luarocks python-natsort-doc python-psutil-doc 2026-03-20T17:56:14.542 INFO:teuthology.orchestra.run.vm08.stdout: subversion python-pygments-doc ttf-bitstream-vera python3-paste python3-dap 2026-03-20T17:56:14.542 INFO:teuthology.orchestra.run.vm08.stdout: python-sklearn-doc ipython3 python-webob-doc gsmartcontrol smart-notifier 2026-03-20T17:56:14.542 INFO:teuthology.orchestra.run.vm08.stdout: mailx | mailutils 2026-03-20T17:56:14.542 INFO:teuthology.orchestra.run.vm08.stdout:Recommended packages: 2026-03-20T17:56:14.542 INFO:teuthology.orchestra.run.vm08.stdout: btrfs-tools 2026-03-20T17:56:14.547 INFO:teuthology.orchestra.run.vm06.stdout:2 upgraded, 85 newly installed, 0 to remove and 36 not upgraded. 2026-03-20T17:56:14.547 INFO:teuthology.orchestra.run.vm06.stdout:Need to get 281 MB of archives. 2026-03-20T17:56:14.547 INFO:teuthology.orchestra.run.vm06.stdout:After this operation, 1092 MB of additional disk space will be used. 2026-03-20T17:56:14.547 INFO:teuthology.orchestra.run.vm06.stdout:Get:1 http://archive.ubuntu.com/ubuntu jammy/main amd64 liblttng-ust1 amd64 2.13.1-1ubuntu1 [190 kB] 2026-03-20T17:56:14.577 INFO:teuthology.orchestra.run.vm02.stdout:The following NEW packages will be installed: 2026-03-20T17:56:14.577 INFO:teuthology.orchestra.run.vm02.stdout: ceph ceph-base ceph-common ceph-fuse ceph-mds ceph-mgr ceph-mgr-cephadm 2026-03-20T17:56:14.577 INFO:teuthology.orchestra.run.vm02.stdout: ceph-mgr-dashboard ceph-mgr-diskprediction-local ceph-mgr-k8sevents 2026-03-20T17:56:14.577 INFO:teuthology.orchestra.run.vm02.stdout: ceph-mgr-modules-core ceph-mon ceph-osd ceph-test ceph-volume cephadm jq 2026-03-20T17:56:14.577 INFO:teuthology.orchestra.run.vm02.stdout: libcephfs-daemon libcephfs-dev libcephfs-proxy2 libcephfs2 2026-03-20T17:56:14.577 INFO:teuthology.orchestra.run.vm02.stdout: libdouble-conversion3 libfuse2 libjq1 liblttng-ust1 libnbd0 liboath0 2026-03-20T17:56:14.577 INFO:teuthology.orchestra.run.vm02.stdout: libonig5 libpcre2-16-0 libqt5core5a libqt5dbus5 libqt5network5 2026-03-20T17:56:14.578 INFO:teuthology.orchestra.run.vm02.stdout: libradosstriper1 librdkafka1 librgw2 libsqlite3-mod-ceph libthrift-0.16.0 2026-03-20T17:56:14.578 INFO:teuthology.orchestra.run.vm02.stdout: nvme-cli python-asyncssh-doc python3-asyncssh python3-cachetools 2026-03-20T17:56:14.578 INFO:teuthology.orchestra.run.vm02.stdout: python3-ceph-argparse python3-ceph-common python3-cephfs python3-cheroot 2026-03-20T17:56:14.578 INFO:teuthology.orchestra.run.vm02.stdout: python3-cherrypy3 python3-google-auth python3-iniconfig 2026-03-20T17:56:14.578 INFO:teuthology.orchestra.run.vm02.stdout: python3-jaraco.classes python3-jaraco.collections python3-jaraco.functools 2026-03-20T17:56:14.578 INFO:teuthology.orchestra.run.vm02.stdout: python3-jaraco.text python3-joblib python3-kubernetes python3-natsort 2026-03-20T17:56:14.578 INFO:teuthology.orchestra.run.vm02.stdout: python3-pluggy python3-portend python3-prettytable python3-psutil python3-py 2026-03-20T17:56:14.578 INFO:teuthology.orchestra.run.vm02.stdout: python3-pygments python3-pytest python3-rados python3-rbd python3-repoze.lru 2026-03-20T17:56:14.578 INFO:teuthology.orchestra.run.vm02.stdout: python3-requests-oauthlib python3-rgw python3-routes python3-rsa 2026-03-20T17:56:14.578 INFO:teuthology.orchestra.run.vm02.stdout: python3-simplejson python3-sklearn python3-sklearn-lib python3-tempora 2026-03-20T17:56:14.578 INFO:teuthology.orchestra.run.vm02.stdout: python3-threadpoolctl python3-toml python3-wcwidth python3-webob 2026-03-20T17:56:14.578 INFO:teuthology.orchestra.run.vm02.stdout: python3-websocket python3-zc.lockfile qttranslations5-l10n radosgw rbd-fuse 2026-03-20T17:56:14.578 INFO:teuthology.orchestra.run.vm02.stdout: smartmontools socat xmlstarlet 2026-03-20T17:56:14.578 INFO:teuthology.orchestra.run.vm02.stdout:The following packages will be upgraded: 2026-03-20T17:56:14.579 INFO:teuthology.orchestra.run.vm02.stdout: librados2 librbd1 2026-03-20T17:56:14.582 INFO:teuthology.orchestra.run.vm06.stdout:Get:2 http://archive.ubuntu.com/ubuntu jammy/universe amd64 libdouble-conversion3 amd64 3.1.7-4 [39.0 kB] 2026-03-20T17:56:14.583 INFO:teuthology.orchestra.run.vm06.stdout:Get:3 http://archive.ubuntu.com/ubuntu jammy-updates/main amd64 libpcre2-16-0 amd64 10.39-3ubuntu0.1 [203 kB] 2026-03-20T17:56:14.584 INFO:teuthology.orchestra.run.vm08.stdout:The following NEW packages will be installed: 2026-03-20T17:56:14.584 INFO:teuthology.orchestra.run.vm08.stdout: ceph ceph-base ceph-common ceph-fuse ceph-mds ceph-mgr ceph-mgr-cephadm 2026-03-20T17:56:14.584 INFO:teuthology.orchestra.run.vm08.stdout: ceph-mgr-dashboard ceph-mgr-diskprediction-local ceph-mgr-k8sevents 2026-03-20T17:56:14.584 INFO:teuthology.orchestra.run.vm08.stdout: ceph-mgr-modules-core ceph-mon ceph-osd ceph-test ceph-volume cephadm jq 2026-03-20T17:56:14.584 INFO:teuthology.orchestra.run.vm08.stdout: libcephfs-daemon libcephfs-dev libcephfs-proxy2 libcephfs2 2026-03-20T17:56:14.584 INFO:teuthology.orchestra.run.vm08.stdout: libdouble-conversion3 libfuse2 libjq1 liblttng-ust1 libnbd0 liboath0 2026-03-20T17:56:14.584 INFO:teuthology.orchestra.run.vm08.stdout: libonig5 libpcre2-16-0 libqt5core5a libqt5dbus5 libqt5network5 2026-03-20T17:56:14.584 INFO:teuthology.orchestra.run.vm08.stdout: libradosstriper1 librdkafka1 librgw2 libsqlite3-mod-ceph libthrift-0.16.0 2026-03-20T17:56:14.584 INFO:teuthology.orchestra.run.vm08.stdout: nvme-cli python-asyncssh-doc python3-asyncssh python3-cachetools 2026-03-20T17:56:14.584 INFO:teuthology.orchestra.run.vm08.stdout: python3-ceph-argparse python3-ceph-common python3-cephfs python3-cheroot 2026-03-20T17:56:14.585 INFO:teuthology.orchestra.run.vm08.stdout: python3-cherrypy3 python3-google-auth python3-iniconfig 2026-03-20T17:56:14.585 INFO:teuthology.orchestra.run.vm08.stdout: python3-jaraco.classes python3-jaraco.collections python3-jaraco.functools 2026-03-20T17:56:14.585 INFO:teuthology.orchestra.run.vm08.stdout: python3-jaraco.text python3-joblib python3-kubernetes python3-natsort 2026-03-20T17:56:14.585 INFO:teuthology.orchestra.run.vm08.stdout: python3-pluggy python3-portend python3-prettytable python3-psutil python3-py 2026-03-20T17:56:14.585 INFO:teuthology.orchestra.run.vm08.stdout: python3-pygments python3-pytest python3-rados python3-rbd python3-repoze.lru 2026-03-20T17:56:14.585 INFO:teuthology.orchestra.run.vm08.stdout: python3-requests-oauthlib python3-rgw python3-routes python3-rsa 2026-03-20T17:56:14.585 INFO:teuthology.orchestra.run.vm08.stdout: python3-simplejson python3-sklearn python3-sklearn-lib python3-tempora 2026-03-20T17:56:14.585 INFO:teuthology.orchestra.run.vm08.stdout: python3-threadpoolctl python3-toml python3-wcwidth python3-webob 2026-03-20T17:56:14.585 INFO:teuthology.orchestra.run.vm08.stdout: python3-websocket python3-zc.lockfile qttranslations5-l10n radosgw rbd-fuse 2026-03-20T17:56:14.585 INFO:teuthology.orchestra.run.vm08.stdout: smartmontools socat xmlstarlet 2026-03-20T17:56:14.585 INFO:teuthology.orchestra.run.vm08.stdout:The following packages will be upgraded: 2026-03-20T17:56:14.585 INFO:teuthology.orchestra.run.vm08.stdout: librados2 librbd1 2026-03-20T17:56:14.590 INFO:teuthology.orchestra.run.vm06.stdout:Get:4 http://archive.ubuntu.com/ubuntu jammy-updates/universe amd64 libqt5core5a amd64 5.15.3+dfsg-2ubuntu0.2 [2006 kB] 2026-03-20T17:56:14.615 INFO:teuthology.orchestra.run.vm06.stdout:Get:5 http://archive.ubuntu.com/ubuntu jammy-updates/universe amd64 libqt5dbus5 amd64 5.15.3+dfsg-2ubuntu0.2 [222 kB] 2026-03-20T17:56:14.616 INFO:teuthology.orchestra.run.vm06.stdout:Get:6 http://archive.ubuntu.com/ubuntu jammy-updates/universe amd64 libqt5network5 amd64 5.15.3+dfsg-2ubuntu0.2 [731 kB] 2026-03-20T17:56:14.616 INFO:teuthology.orchestra.run.vm08.stdout:2 upgraded, 85 newly installed, 0 to remove and 36 not upgraded. 2026-03-20T17:56:14.617 INFO:teuthology.orchestra.run.vm08.stdout:Need to get 281 MB of archives. 2026-03-20T17:56:14.617 INFO:teuthology.orchestra.run.vm08.stdout:After this operation, 1092 MB of additional disk space will be used. 2026-03-20T17:56:14.617 INFO:teuthology.orchestra.run.vm08.stdout:Get:1 http://archive.ubuntu.com/ubuntu jammy/main amd64 liblttng-ust1 amd64 2.13.1-1ubuntu1 [190 kB] 2026-03-20T17:56:14.630 INFO:teuthology.orchestra.run.vm06.stdout:Get:7 http://archive.ubuntu.com/ubuntu jammy/universe amd64 libthrift-0.16.0 amd64 0.16.0-2 [267 kB] 2026-03-20T17:56:14.631 INFO:teuthology.orchestra.run.vm06.stdout:Get:8 http://archive.ubuntu.com/ubuntu jammy/universe amd64 libnbd0 amd64 1.10.5-1 [71.3 kB] 2026-03-20T17:56:14.632 INFO:teuthology.orchestra.run.vm06.stdout:Get:9 http://archive.ubuntu.com/ubuntu jammy/main amd64 python3-wcwidth all 0.2.5+dfsg1-1 [21.9 kB] 2026-03-20T17:56:14.632 INFO:teuthology.orchestra.run.vm06.stdout:Get:10 http://archive.ubuntu.com/ubuntu jammy/main amd64 python3-prettytable all 2.5.0-2 [31.3 kB] 2026-03-20T17:56:14.632 INFO:teuthology.orchestra.run.vm06.stdout:Get:11 http://archive.ubuntu.com/ubuntu jammy/universe amd64 librdkafka1 amd64 1.8.0-1build1 [633 kB] 2026-03-20T17:56:14.638 INFO:teuthology.orchestra.run.vm06.stdout:Get:12 http://archive.ubuntu.com/ubuntu jammy-updates/main amd64 liboath0 amd64 2.6.7-3ubuntu0.1 [41.3 kB] 2026-03-20T17:56:14.638 INFO:teuthology.orchestra.run.vm06.stdout:Get:13 http://archive.ubuntu.com/ubuntu jammy/main amd64 python3-jaraco.functools all 3.4.0-2 [9030 B] 2026-03-20T17:56:14.638 INFO:teuthology.orchestra.run.vm06.stdout:Get:14 http://archive.ubuntu.com/ubuntu jammy-updates/main amd64 python3-cheroot all 8.5.2+ds1-1ubuntu3.1 [71.1 kB] 2026-03-20T17:56:14.638 INFO:teuthology.orchestra.run.vm06.stdout:Get:15 http://archive.ubuntu.com/ubuntu jammy/main amd64 python3-jaraco.classes all 3.2.1-3 [6452 B] 2026-03-20T17:56:14.641 INFO:teuthology.orchestra.run.vm06.stdout:Get:16 http://archive.ubuntu.com/ubuntu jammy/main amd64 python3-jaraco.text all 3.6.0-2 [8716 B] 2026-03-20T17:56:14.641 INFO:teuthology.orchestra.run.vm06.stdout:Get:17 http://archive.ubuntu.com/ubuntu jammy/main amd64 python3-jaraco.collections all 3.4.0-2 [11.4 kB] 2026-03-20T17:56:14.641 INFO:teuthology.orchestra.run.vm06.stdout:Get:18 http://archive.ubuntu.com/ubuntu jammy/main amd64 python3-tempora all 4.1.2-1 [14.8 kB] 2026-03-20T17:56:14.641 INFO:teuthology.orchestra.run.vm06.stdout:Get:19 http://archive.ubuntu.com/ubuntu jammy/main amd64 python3-portend all 3.0.0-1 [7240 B] 2026-03-20T17:56:14.641 INFO:teuthology.orchestra.run.vm06.stdout:Get:20 http://archive.ubuntu.com/ubuntu jammy/main amd64 python3-zc.lockfile all 2.0-1 [8980 B] 2026-03-20T17:56:14.642 INFO:teuthology.orchestra.run.vm05.stdout:The following packages were automatically installed and are no longer required: 2026-03-20T17:56:14.642 INFO:teuthology.orchestra.run.vm05.stdout: kpartx libboost-iostreams1.74.0 libboost-thread1.74.0 libpmemobj1 2026-03-20T17:56:14.642 INFO:teuthology.orchestra.run.vm05.stdout: libsgutils2-2 sg3-utils sg3-utils-udev 2026-03-20T17:56:14.642 INFO:teuthology.orchestra.run.vm05.stdout:Use 'sudo apt autoremove' to remove them. 2026-03-20T17:56:14.643 INFO:teuthology.orchestra.run.vm05.stdout:The following additional packages will be installed: 2026-03-20T17:56:14.643 INFO:teuthology.orchestra.run.vm05.stdout: ceph-base ceph-mgr-cephadm ceph-mgr-dashboard ceph-mgr-diskprediction-local 2026-03-20T17:56:14.643 INFO:teuthology.orchestra.run.vm05.stdout: ceph-mgr-k8sevents ceph-mgr-modules-core ceph-mon ceph-osd jq 2026-03-20T17:56:14.643 INFO:teuthology.orchestra.run.vm05.stdout: libcephfs-daemon libcephfs-proxy2 libdouble-conversion3 libfuse2 libjq1 2026-03-20T17:56:14.643 INFO:teuthology.orchestra.run.vm05.stdout: liblttng-ust1 libnbd0 liboath0 libonig5 libpcre2-16-0 libqt5core5a 2026-03-20T17:56:14.643 INFO:teuthology.orchestra.run.vm05.stdout: libqt5dbus5 libqt5network5 libradosstriper1 librdkafka1 librgw2 2026-03-20T17:56:14.643 INFO:teuthology.orchestra.run.vm05.stdout: libsqlite3-mod-ceph libthrift-0.16.0 nvme-cli python-asyncssh-doc 2026-03-20T17:56:14.643 INFO:teuthology.orchestra.run.vm05.stdout: python3-asyncssh python3-cachetools python3-ceph-argparse 2026-03-20T17:56:14.643 INFO:teuthology.orchestra.run.vm05.stdout: python3-ceph-common python3-cheroot python3-cherrypy3 python3-google-auth 2026-03-20T17:56:14.643 INFO:teuthology.orchestra.run.vm05.stdout: python3-iniconfig python3-jaraco.classes python3-jaraco.collections 2026-03-20T17:56:14.643 INFO:teuthology.orchestra.run.vm05.stdout: python3-jaraco.functools python3-jaraco.text python3-joblib 2026-03-20T17:56:14.643 INFO:teuthology.orchestra.run.vm05.stdout: python3-kubernetes python3-natsort python3-pluggy python3-portend 2026-03-20T17:56:14.643 INFO:teuthology.orchestra.run.vm05.stdout: python3-prettytable python3-psutil python3-py python3-pygments 2026-03-20T17:56:14.643 INFO:teuthology.orchestra.run.vm05.stdout: python3-pytest python3-repoze.lru python3-requests-oauthlib python3-routes 2026-03-20T17:56:14.643 INFO:teuthology.orchestra.run.vm05.stdout: python3-rsa python3-simplejson python3-sklearn python3-sklearn-lib 2026-03-20T17:56:14.643 INFO:teuthology.orchestra.run.vm05.stdout: python3-tempora python3-threadpoolctl python3-toml python3-wcwidth 2026-03-20T17:56:14.643 INFO:teuthology.orchestra.run.vm05.stdout: python3-webob python3-websocket python3-zc.lockfile qttranslations5-l10n 2026-03-20T17:56:14.643 INFO:teuthology.orchestra.run.vm05.stdout: smartmontools socat xmlstarlet 2026-03-20T17:56:14.643 INFO:teuthology.orchestra.run.vm05.stdout:Suggested packages: 2026-03-20T17:56:14.643 INFO:teuthology.orchestra.run.vm05.stdout: python3-influxdb liblua5.3-dev luarocks python-natsort-doc python-psutil-doc 2026-03-20T17:56:14.643 INFO:teuthology.orchestra.run.vm05.stdout: subversion python-pygments-doc ttf-bitstream-vera python3-paste python3-dap 2026-03-20T17:56:14.643 INFO:teuthology.orchestra.run.vm05.stdout: python-sklearn-doc ipython3 python-webob-doc gsmartcontrol smart-notifier 2026-03-20T17:56:14.643 INFO:teuthology.orchestra.run.vm05.stdout: mailx | mailutils 2026-03-20T17:56:14.643 INFO:teuthology.orchestra.run.vm05.stdout:Recommended packages: 2026-03-20T17:56:14.643 INFO:teuthology.orchestra.run.vm05.stdout: btrfs-tools 2026-03-20T17:56:14.648 INFO:teuthology.orchestra.run.vm06.stdout:Get:21 http://archive.ubuntu.com/ubuntu jammy/main amd64 python3-cherrypy3 all 18.6.1-4 [208 kB] 2026-03-20T17:56:14.650 INFO:teuthology.orchestra.run.vm06.stdout:Get:22 http://archive.ubuntu.com/ubuntu jammy/universe amd64 python3-natsort all 8.0.2-1 [35.3 kB] 2026-03-20T17:56:14.650 INFO:teuthology.orchestra.run.vm06.stdout:Get:23 http://archive.ubuntu.com/ubuntu jammy/universe amd64 libfuse2 amd64 2.9.9-5ubuntu3 [90.3 kB] 2026-03-20T17:56:14.651 INFO:teuthology.orchestra.run.vm06.stdout:Get:24 http://archive.ubuntu.com/ubuntu jammy-updates/universe amd64 python3-asyncssh all 2.5.0-1ubuntu0.1 [189 kB] 2026-03-20T17:56:14.653 INFO:teuthology.orchestra.run.vm06.stdout:Get:25 http://archive.ubuntu.com/ubuntu jammy/main amd64 python3-repoze.lru all 0.7-2 [12.1 kB] 2026-03-20T17:56:14.655 INFO:teuthology.orchestra.run.vm06.stdout:Get:26 http://archive.ubuntu.com/ubuntu jammy/main amd64 python3-routes all 2.5.1-1ubuntu1 [89.0 kB] 2026-03-20T17:56:14.656 INFO:teuthology.orchestra.run.vm06.stdout:Get:27 http://archive.ubuntu.com/ubuntu jammy/universe amd64 python3-sklearn-lib amd64 0.23.2-5ubuntu6 [2058 kB] 2026-03-20T17:56:14.668 INFO:teuthology.orchestra.run.vm08.stdout:Get:2 http://archive.ubuntu.com/ubuntu jammy/universe amd64 libdouble-conversion3 amd64 3.1.7-4 [39.0 kB] 2026-03-20T17:56:14.681 INFO:teuthology.orchestra.run.vm05.stdout:The following NEW packages will be installed: 2026-03-20T17:56:14.682 INFO:teuthology.orchestra.run.vm05.stdout: ceph ceph-base ceph-common ceph-fuse ceph-mds ceph-mgr ceph-mgr-cephadm 2026-03-20T17:56:14.682 INFO:teuthology.orchestra.run.vm05.stdout: ceph-mgr-dashboard ceph-mgr-diskprediction-local ceph-mgr-k8sevents 2026-03-20T17:56:14.682 INFO:teuthology.orchestra.run.vm05.stdout: ceph-mgr-modules-core ceph-mon ceph-osd ceph-test ceph-volume cephadm jq 2026-03-20T17:56:14.682 INFO:teuthology.orchestra.run.vm05.stdout: libcephfs-daemon libcephfs-dev libcephfs-proxy2 libcephfs2 2026-03-20T17:56:14.682 INFO:teuthology.orchestra.run.vm05.stdout: libdouble-conversion3 libfuse2 libjq1 liblttng-ust1 libnbd0 liboath0 2026-03-20T17:56:14.682 INFO:teuthology.orchestra.run.vm05.stdout: libonig5 libpcre2-16-0 libqt5core5a libqt5dbus5 libqt5network5 2026-03-20T17:56:14.682 INFO:teuthology.orchestra.run.vm05.stdout: libradosstriper1 librdkafka1 librgw2 libsqlite3-mod-ceph libthrift-0.16.0 2026-03-20T17:56:14.682 INFO:teuthology.orchestra.run.vm05.stdout: nvme-cli python-asyncssh-doc python3-asyncssh python3-cachetools 2026-03-20T17:56:14.682 INFO:teuthology.orchestra.run.vm05.stdout: python3-ceph-argparse python3-ceph-common python3-cephfs python3-cheroot 2026-03-20T17:56:14.682 INFO:teuthology.orchestra.run.vm05.stdout: python3-cherrypy3 python3-google-auth python3-iniconfig 2026-03-20T17:56:14.682 INFO:teuthology.orchestra.run.vm05.stdout: python3-jaraco.classes python3-jaraco.collections python3-jaraco.functools 2026-03-20T17:56:14.682 INFO:teuthology.orchestra.run.vm05.stdout: python3-jaraco.text python3-joblib python3-kubernetes python3-natsort 2026-03-20T17:56:14.682 INFO:teuthology.orchestra.run.vm05.stdout: python3-pluggy python3-portend python3-prettytable python3-psutil python3-py 2026-03-20T17:56:14.682 INFO:teuthology.orchestra.run.vm05.stdout: python3-pygments python3-pytest python3-rados python3-rbd python3-repoze.lru 2026-03-20T17:56:14.682 INFO:teuthology.orchestra.run.vm05.stdout: python3-requests-oauthlib python3-rgw python3-routes python3-rsa 2026-03-20T17:56:14.682 INFO:teuthology.orchestra.run.vm05.stdout: python3-simplejson python3-sklearn python3-sklearn-lib python3-tempora 2026-03-20T17:56:14.682 INFO:teuthology.orchestra.run.vm05.stdout: python3-threadpoolctl python3-toml python3-wcwidth python3-webob 2026-03-20T17:56:14.682 INFO:teuthology.orchestra.run.vm05.stdout: python3-websocket python3-zc.lockfile qttranslations5-l10n radosgw rbd-fuse 2026-03-20T17:56:14.682 INFO:teuthology.orchestra.run.vm05.stdout: smartmontools socat xmlstarlet 2026-03-20T17:56:14.683 INFO:teuthology.orchestra.run.vm05.stdout:The following packages will be upgraded: 2026-03-20T17:56:14.683 INFO:teuthology.orchestra.run.vm05.stdout: librados2 librbd1 2026-03-20T17:56:14.684 INFO:teuthology.orchestra.run.vm06.stdout:Get:28 http://archive.ubuntu.com/ubuntu jammy/universe amd64 python3-joblib all 0.17.0-4ubuntu1 [204 kB] 2026-03-20T17:56:14.685 INFO:teuthology.orchestra.run.vm08.stdout:Get:3 http://archive.ubuntu.com/ubuntu jammy-updates/main amd64 libpcre2-16-0 amd64 10.39-3ubuntu0.1 [203 kB] 2026-03-20T17:56:14.685 INFO:teuthology.orchestra.run.vm06.stdout:Get:29 http://archive.ubuntu.com/ubuntu jammy/universe amd64 python3-threadpoolctl all 3.1.0-1 [21.3 kB] 2026-03-20T17:56:14.686 INFO:teuthology.orchestra.run.vm06.stdout:Get:30 http://archive.ubuntu.com/ubuntu jammy/universe amd64 python3-sklearn all 0.23.2-5ubuntu6 [1829 kB] 2026-03-20T17:56:14.688 INFO:teuthology.orchestra.run.vm08.stdout:Get:4 http://archive.ubuntu.com/ubuntu jammy-updates/universe amd64 libqt5core5a amd64 5.15.3+dfsg-2ubuntu0.2 [2006 kB] 2026-03-20T17:56:14.702 INFO:teuthology.orchestra.run.vm06.stdout:Get:31 http://archive.ubuntu.com/ubuntu jammy/main amd64 python3-cachetools all 5.0.0-1 [9722 B] 2026-03-20T17:56:14.702 INFO:teuthology.orchestra.run.vm06.stdout:Get:32 http://archive.ubuntu.com/ubuntu jammy/universe amd64 python3-rsa all 4.8-1 [28.4 kB] 2026-03-20T17:56:14.702 INFO:teuthology.orchestra.run.vm06.stdout:Get:33 http://archive.ubuntu.com/ubuntu jammy/universe amd64 python3-google-auth all 1.5.1-3 [35.7 kB] 2026-03-20T17:56:14.702 INFO:teuthology.orchestra.run.vm06.stdout:Get:34 http://archive.ubuntu.com/ubuntu jammy/universe amd64 python3-requests-oauthlib all 1.3.0+ds-0.1 [18.7 kB] 2026-03-20T17:56:14.703 INFO:teuthology.orchestra.run.vm06.stdout:Get:35 http://archive.ubuntu.com/ubuntu jammy/universe amd64 python3-websocket all 1.2.3-1 [34.7 kB] 2026-03-20T17:56:14.703 INFO:teuthology.orchestra.run.vm06.stdout:Get:36 http://archive.ubuntu.com/ubuntu jammy/universe amd64 python3-kubernetes all 12.0.1-1ubuntu1 [353 kB] 2026-03-20T17:56:14.704 INFO:teuthology.orchestra.run.vm06.stdout:Get:37 http://archive.ubuntu.com/ubuntu jammy/main amd64 libonig5 amd64 6.9.7.1-2build1 [172 kB] 2026-03-20T17:56:14.705 INFO:teuthology.orchestra.run.vm06.stdout:Get:38 http://archive.ubuntu.com/ubuntu jammy-updates/main amd64 libjq1 amd64 1.6-2.1ubuntu3.1 [133 kB] 2026-03-20T17:56:14.706 INFO:teuthology.orchestra.run.vm06.stdout:Get:39 http://archive.ubuntu.com/ubuntu jammy-updates/main amd64 jq amd64 1.6-2.1ubuntu3.1 [52.5 kB] 2026-03-20T17:56:14.711 INFO:teuthology.orchestra.run.vm06.stdout:Get:40 http://archive.ubuntu.com/ubuntu jammy/main amd64 socat amd64 1.7.4.1-3ubuntu4 [349 kB] 2026-03-20T17:56:14.715 INFO:teuthology.orchestra.run.vm06.stdout:Get:41 http://archive.ubuntu.com/ubuntu jammy/universe amd64 xmlstarlet amd64 1.6.1-2.1 [265 kB] 2026-03-20T17:56:14.717 INFO:teuthology.orchestra.run.vm06.stdout:Get:42 http://archive.ubuntu.com/ubuntu jammy-updates/main amd64 nvme-cli amd64 1.16-3ubuntu0.3 [474 kB] 2026-03-20T17:56:14.731 INFO:teuthology.orchestra.run.vm06.stdout:Get:43 http://archive.ubuntu.com/ubuntu jammy-updates/universe amd64 python-asyncssh-doc all 2.5.0-1ubuntu0.1 [309 kB] 2026-03-20T17:56:14.735 INFO:teuthology.orchestra.run.vm06.stdout:Get:44 http://archive.ubuntu.com/ubuntu jammy/universe amd64 python3-iniconfig all 1.1.1-2 [6024 B] 2026-03-20T17:56:14.735 INFO:teuthology.orchestra.run.vm06.stdout:Get:45 http://archive.ubuntu.com/ubuntu jammy/universe amd64 python3-pluggy all 0.13.0-7.1 [19.0 kB] 2026-03-20T17:56:14.735 INFO:teuthology.orchestra.run.vm06.stdout:Get:46 http://archive.ubuntu.com/ubuntu jammy/main amd64 python3-psutil amd64 5.9.0-1build1 [158 kB] 2026-03-20T17:56:14.736 INFO:teuthology.orchestra.run.vm06.stdout:Get:47 http://archive.ubuntu.com/ubuntu jammy/universe amd64 python3-py all 1.10.0-1 [71.9 kB] 2026-03-20T17:56:14.736 INFO:teuthology.orchestra.run.vm06.stdout:Get:48 http://archive.ubuntu.com/ubuntu jammy-updates/main amd64 python3-pygments all 2.11.2+dfsg-2ubuntu0.1 [750 kB] 2026-03-20T17:56:14.743 INFO:teuthology.orchestra.run.vm06.stdout:Get:49 http://archive.ubuntu.com/ubuntu jammy/universe amd64 python3-toml all 0.10.2-1 [16.5 kB] 2026-03-20T17:56:14.743 INFO:teuthology.orchestra.run.vm06.stdout:Get:50 http://archive.ubuntu.com/ubuntu jammy/universe amd64 python3-pytest all 6.2.5-1ubuntu2 [214 kB] 2026-03-20T17:56:14.744 INFO:teuthology.orchestra.run.vm08.stdout:Get:5 http://archive.ubuntu.com/ubuntu jammy-updates/universe amd64 libqt5dbus5 amd64 5.15.3+dfsg-2ubuntu0.2 [222 kB] 2026-03-20T17:56:14.747 INFO:teuthology.orchestra.run.vm08.stdout:Get:6 http://archive.ubuntu.com/ubuntu jammy-updates/universe amd64 libqt5network5 amd64 5.15.3+dfsg-2ubuntu0.2 [731 kB] 2026-03-20T17:56:14.747 INFO:teuthology.orchestra.run.vm06.stdout:Get:51 http://archive.ubuntu.com/ubuntu jammy/main amd64 python3-simplejson amd64 3.17.6-1build1 [54.7 kB] 2026-03-20T17:56:14.762 INFO:teuthology.orchestra.run.vm06.stdout:Get:52 http://archive.ubuntu.com/ubuntu jammy-updates/main amd64 python3-webob all 1:1.8.6-1.1ubuntu0.1 [86.7 kB] 2026-03-20T17:56:14.765 INFO:teuthology.orchestra.run.vm08.stdout:Get:7 http://archive.ubuntu.com/ubuntu jammy/universe amd64 libthrift-0.16.0 amd64 0.16.0-2 [267 kB] 2026-03-20T17:56:14.766 INFO:teuthology.orchestra.run.vm08.stdout:Get:8 http://archive.ubuntu.com/ubuntu jammy/universe amd64 libnbd0 amd64 1.10.5-1 [71.3 kB] 2026-03-20T17:56:14.766 INFO:teuthology.orchestra.run.vm08.stdout:Get:9 http://archive.ubuntu.com/ubuntu jammy/main amd64 python3-wcwidth all 0.2.5+dfsg1-1 [21.9 kB] 2026-03-20T17:56:14.767 INFO:teuthology.orchestra.run.vm08.stdout:Get:10 http://archive.ubuntu.com/ubuntu jammy/main amd64 python3-prettytable all 2.5.0-2 [31.3 kB] 2026-03-20T17:56:14.767 INFO:teuthology.orchestra.run.vm08.stdout:Get:11 http://archive.ubuntu.com/ubuntu jammy/universe amd64 librdkafka1 amd64 1.8.0-1build1 [633 kB] 2026-03-20T17:56:14.768 INFO:teuthology.orchestra.run.vm06.stdout:Get:53 http://archive.ubuntu.com/ubuntu jammy/universe amd64 qttranslations5-l10n all 5.15.3-1 [1983 kB] 2026-03-20T17:56:14.775 INFO:teuthology.orchestra.run.vm08.stdout:Get:12 http://archive.ubuntu.com/ubuntu jammy-updates/main amd64 liboath0 amd64 2.6.7-3ubuntu0.1 [41.3 kB] 2026-03-20T17:56:14.776 INFO:teuthology.orchestra.run.vm08.stdout:Get:13 http://archive.ubuntu.com/ubuntu jammy/main amd64 python3-jaraco.functools all 3.4.0-2 [9030 B] 2026-03-20T17:56:14.776 INFO:teuthology.orchestra.run.vm08.stdout:Get:14 http://archive.ubuntu.com/ubuntu jammy-updates/main amd64 python3-cheroot all 8.5.2+ds1-1ubuntu3.1 [71.1 kB] 2026-03-20T17:56:14.776 INFO:teuthology.orchestra.run.vm08.stdout:Get:15 http://archive.ubuntu.com/ubuntu jammy/main amd64 python3-jaraco.classes all 3.2.1-3 [6452 B] 2026-03-20T17:56:14.777 INFO:teuthology.orchestra.run.vm08.stdout:Get:16 http://archive.ubuntu.com/ubuntu jammy/main amd64 python3-jaraco.text all 3.6.0-2 [8716 B] 2026-03-20T17:56:14.777 INFO:teuthology.orchestra.run.vm08.stdout:Get:17 http://archive.ubuntu.com/ubuntu jammy/main amd64 python3-jaraco.collections all 3.4.0-2 [11.4 kB] 2026-03-20T17:56:14.778 INFO:teuthology.orchestra.run.vm08.stdout:Get:18 http://archive.ubuntu.com/ubuntu jammy/main amd64 python3-tempora all 4.1.2-1 [14.8 kB] 2026-03-20T17:56:14.778 INFO:teuthology.orchestra.run.vm08.stdout:Get:19 http://archive.ubuntu.com/ubuntu jammy/main amd64 python3-portend all 3.0.0-1 [7240 B] 2026-03-20T17:56:14.778 INFO:teuthology.orchestra.run.vm08.stdout:Get:20 http://archive.ubuntu.com/ubuntu jammy/main amd64 python3-zc.lockfile all 2.0-1 [8980 B] 2026-03-20T17:56:14.782 INFO:teuthology.orchestra.run.vm06.stdout:Get:54 http://archive.ubuntu.com/ubuntu jammy-updates/main amd64 smartmontools amd64 7.2-1ubuntu0.1 [583 kB] 2026-03-20T17:56:14.786 INFO:teuthology.orchestra.run.vm08.stdout:Get:21 http://archive.ubuntu.com/ubuntu jammy/main amd64 python3-cherrypy3 all 18.6.1-4 [208 kB] 2026-03-20T17:56:14.788 INFO:teuthology.orchestra.run.vm08.stdout:Get:22 http://archive.ubuntu.com/ubuntu jammy/universe amd64 python3-natsort all 8.0.2-1 [35.3 kB] 2026-03-20T17:56:14.789 INFO:teuthology.orchestra.run.vm08.stdout:Get:23 http://archive.ubuntu.com/ubuntu jammy/universe amd64 libfuse2 amd64 2.9.9-5ubuntu3 [90.3 kB] 2026-03-20T17:56:14.789 INFO:teuthology.orchestra.run.vm08.stdout:Get:24 http://archive.ubuntu.com/ubuntu jammy-updates/universe amd64 python3-asyncssh all 2.5.0-1ubuntu0.1 [189 kB] 2026-03-20T17:56:14.791 INFO:teuthology.orchestra.run.vm08.stdout:Get:25 http://archive.ubuntu.com/ubuntu jammy/main amd64 python3-repoze.lru all 0.7-2 [12.1 kB] 2026-03-20T17:56:14.794 INFO:teuthology.orchestra.run.vm08.stdout:Get:26 http://archive.ubuntu.com/ubuntu jammy/main amd64 python3-routes all 2.5.1-1ubuntu1 [89.0 kB] 2026-03-20T17:56:14.795 INFO:teuthology.orchestra.run.vm08.stdout:Get:27 http://archive.ubuntu.com/ubuntu jammy/universe amd64 python3-sklearn-lib amd64 0.23.2-5ubuntu6 [2058 kB] 2026-03-20T17:56:14.824 INFO:teuthology.orchestra.run.vm08.stdout:Get:28 http://archive.ubuntu.com/ubuntu jammy/universe amd64 python3-joblib all 0.17.0-4ubuntu1 [204 kB] 2026-03-20T17:56:14.825 INFO:teuthology.orchestra.run.vm08.stdout:Get:29 http://archive.ubuntu.com/ubuntu jammy/universe amd64 python3-threadpoolctl all 3.1.0-1 [21.3 kB] 2026-03-20T17:56:14.825 INFO:teuthology.orchestra.run.vm08.stdout:Get:30 http://archive.ubuntu.com/ubuntu jammy/universe amd64 python3-sklearn all 0.23.2-5ubuntu6 [1829 kB] 2026-03-20T17:56:14.834 INFO:teuthology.orchestra.run.vm08.stdout:Get:31 http://archive.ubuntu.com/ubuntu jammy/main amd64 python3-cachetools all 5.0.0-1 [9722 B] 2026-03-20T17:56:14.834 INFO:teuthology.orchestra.run.vm08.stdout:Get:32 http://archive.ubuntu.com/ubuntu jammy/universe amd64 python3-rsa all 4.8-1 [28.4 kB] 2026-03-20T17:56:14.834 INFO:teuthology.orchestra.run.vm08.stdout:Get:33 http://archive.ubuntu.com/ubuntu jammy/universe amd64 python3-google-auth all 1.5.1-3 [35.7 kB] 2026-03-20T17:56:14.834 INFO:teuthology.orchestra.run.vm08.stdout:Get:34 http://archive.ubuntu.com/ubuntu jammy/universe amd64 python3-requests-oauthlib all 1.3.0+ds-0.1 [18.7 kB] 2026-03-20T17:56:14.835 INFO:teuthology.orchestra.run.vm08.stdout:Get:35 http://archive.ubuntu.com/ubuntu jammy/universe amd64 python3-websocket all 1.2.3-1 [34.7 kB] 2026-03-20T17:56:14.835 INFO:teuthology.orchestra.run.vm08.stdout:Get:36 http://archive.ubuntu.com/ubuntu jammy/universe amd64 python3-kubernetes all 12.0.1-1ubuntu1 [353 kB] 2026-03-20T17:56:14.836 INFO:teuthology.orchestra.run.vm08.stdout:Get:37 http://archive.ubuntu.com/ubuntu jammy/main amd64 libonig5 amd64 6.9.7.1-2build1 [172 kB] 2026-03-20T17:56:14.842 INFO:teuthology.orchestra.run.vm08.stdout:Get:38 http://archive.ubuntu.com/ubuntu jammy-updates/main amd64 libjq1 amd64 1.6-2.1ubuntu3.1 [133 kB] 2026-03-20T17:56:14.843 INFO:teuthology.orchestra.run.vm08.stdout:Get:39 http://archive.ubuntu.com/ubuntu jammy-updates/main amd64 jq amd64 1.6-2.1ubuntu3.1 [52.5 kB] 2026-03-20T17:56:14.847 INFO:teuthology.orchestra.run.vm02.stdout:2 upgraded, 85 newly installed, 0 to remove and 36 not upgraded. 2026-03-20T17:56:14.847 INFO:teuthology.orchestra.run.vm02.stdout:Need to get 281 MB of archives. 2026-03-20T17:56:14.847 INFO:teuthology.orchestra.run.vm02.stdout:After this operation, 1092 MB of additional disk space will be used. 2026-03-20T17:56:14.847 INFO:teuthology.orchestra.run.vm02.stdout:Get:1 http://archive.ubuntu.com/ubuntu jammy/main amd64 liblttng-ust1 amd64 2.13.1-1ubuntu1 [190 kB] 2026-03-20T17:56:14.850 INFO:teuthology.orchestra.run.vm08.stdout:Get:40 http://archive.ubuntu.com/ubuntu jammy/main amd64 socat amd64 1.7.4.1-3ubuntu4 [349 kB] 2026-03-20T17:56:14.853 INFO:teuthology.orchestra.run.vm08.stdout:Get:41 http://archive.ubuntu.com/ubuntu jammy/universe amd64 xmlstarlet amd64 1.6.1-2.1 [265 kB] 2026-03-20T17:56:14.855 INFO:teuthology.orchestra.run.vm08.stdout:Get:42 http://archive.ubuntu.com/ubuntu jammy-updates/main amd64 nvme-cli amd64 1.16-3ubuntu0.3 [474 kB] 2026-03-20T17:56:14.859 INFO:teuthology.orchestra.run.vm08.stdout:Get:43 http://archive.ubuntu.com/ubuntu jammy-updates/universe amd64 python-asyncssh-doc all 2.5.0-1ubuntu0.1 [309 kB] 2026-03-20T17:56:14.871 INFO:teuthology.orchestra.run.vm08.stdout:Get:44 http://archive.ubuntu.com/ubuntu jammy/universe amd64 python3-iniconfig all 1.1.1-2 [6024 B] 2026-03-20T17:56:14.871 INFO:teuthology.orchestra.run.vm08.stdout:Get:45 http://archive.ubuntu.com/ubuntu jammy/universe amd64 python3-pluggy all 0.13.0-7.1 [19.0 kB] 2026-03-20T17:56:14.871 INFO:teuthology.orchestra.run.vm08.stdout:Get:46 http://archive.ubuntu.com/ubuntu jammy/main amd64 python3-psutil amd64 5.9.0-1build1 [158 kB] 2026-03-20T17:56:14.872 INFO:teuthology.orchestra.run.vm08.stdout:Get:47 http://archive.ubuntu.com/ubuntu jammy/universe amd64 python3-py all 1.10.0-1 [71.9 kB] 2026-03-20T17:56:14.872 INFO:teuthology.orchestra.run.vm08.stdout:Get:48 http://archive.ubuntu.com/ubuntu jammy-updates/main amd64 python3-pygments all 2.11.2+dfsg-2ubuntu0.1 [750 kB] 2026-03-20T17:56:14.878 INFO:teuthology.orchestra.run.vm08.stdout:Get:49 http://archive.ubuntu.com/ubuntu jammy/universe amd64 python3-toml all 0.10.2-1 [16.5 kB] 2026-03-20T17:56:14.878 INFO:teuthology.orchestra.run.vm08.stdout:Get:50 http://archive.ubuntu.com/ubuntu jammy/universe amd64 python3-pytest all 6.2.5-1ubuntu2 [214 kB] 2026-03-20T17:56:14.879 INFO:teuthology.orchestra.run.vm08.stdout:Get:51 http://archive.ubuntu.com/ubuntu jammy/main amd64 python3-simplejson amd64 3.17.6-1build1 [54.7 kB] 2026-03-20T17:56:14.879 INFO:teuthology.orchestra.run.vm08.stdout:Get:52 http://archive.ubuntu.com/ubuntu jammy-updates/main amd64 python3-webob all 1:1.8.6-1.1ubuntu0.1 [86.7 kB] 2026-03-20T17:56:14.880 INFO:teuthology.orchestra.run.vm08.stdout:Get:53 http://archive.ubuntu.com/ubuntu jammy/universe amd64 qttranslations5-l10n all 5.15.3-1 [1983 kB] 2026-03-20T17:56:14.911 INFO:teuthology.orchestra.run.vm08.stdout:Get:54 http://archive.ubuntu.com/ubuntu jammy-updates/main amd64 smartmontools amd64 7.2-1ubuntu0.1 [583 kB] 2026-03-20T17:56:14.955 INFO:teuthology.orchestra.run.vm05.stdout:2 upgraded, 85 newly installed, 0 to remove and 36 not upgraded. 2026-03-20T17:56:14.955 INFO:teuthology.orchestra.run.vm05.stdout:Need to get 281 MB of archives. 2026-03-20T17:56:14.955 INFO:teuthology.orchestra.run.vm05.stdout:After this operation, 1092 MB of additional disk space will be used. 2026-03-20T17:56:14.955 INFO:teuthology.orchestra.run.vm05.stdout:Get:1 http://archive.ubuntu.com/ubuntu jammy/main amd64 liblttng-ust1 amd64 2.13.1-1ubuntu1 [190 kB] 2026-03-20T17:56:15.148 INFO:teuthology.orchestra.run.vm06.stdout:Get:55 https://1.chacra.ceph.com/r/ceph/tentacle/70f8415b300f041766fa27faf7d5472699e32388/ubuntu/jammy/flavors/default jammy/main amd64 librbd1 amd64 20.2.0-712-g70f8415b-1jammy [2867 kB] 2026-03-20T17:56:15.167 INFO:teuthology.orchestra.run.vm08.stdout:Get:55 https://1.chacra.ceph.com/r/ceph/tentacle/70f8415b300f041766fa27faf7d5472699e32388/ubuntu/jammy/flavors/default jammy/main amd64 librbd1 amd64 20.2.0-712-g70f8415b-1jammy [2867 kB] 2026-03-20T17:56:15.185 INFO:teuthology.orchestra.run.vm02.stdout:Get:2 https://1.chacra.ceph.com/r/ceph/tentacle/70f8415b300f041766fa27faf7d5472699e32388/ubuntu/jammy/flavors/default jammy/main amd64 librbd1 amd64 20.2.0-712-g70f8415b-1jammy [2867 kB] 2026-03-20T17:56:15.383 INFO:teuthology.orchestra.run.vm02.stdout:Get:3 http://archive.ubuntu.com/ubuntu jammy/universe amd64 libdouble-conversion3 amd64 3.1.7-4 [39.0 kB] 2026-03-20T17:56:15.399 INFO:teuthology.orchestra.run.vm02.stdout:Get:4 http://archive.ubuntu.com/ubuntu jammy-updates/main amd64 libpcre2-16-0 amd64 10.39-3ubuntu0.1 [203 kB] 2026-03-20T17:56:15.501 INFO:teuthology.orchestra.run.vm02.stdout:Get:5 http://archive.ubuntu.com/ubuntu jammy-updates/universe amd64 libqt5core5a amd64 5.15.3+dfsg-2ubuntu0.2 [2006 kB] 2026-03-20T17:56:15.618 INFO:teuthology.orchestra.run.vm05.stdout:Get:2 http://archive.ubuntu.com/ubuntu jammy/universe amd64 libdouble-conversion3 amd64 3.1.7-4 [39.0 kB] 2026-03-20T17:56:15.636 INFO:teuthology.orchestra.run.vm05.stdout:Get:3 http://archive.ubuntu.com/ubuntu jammy-updates/main amd64 libpcre2-16-0 amd64 10.39-3ubuntu0.1 [203 kB] 2026-03-20T17:56:15.782 INFO:teuthology.orchestra.run.vm05.stdout:Get:4 http://archive.ubuntu.com/ubuntu jammy-updates/universe amd64 libqt5core5a amd64 5.15.3+dfsg-2ubuntu0.2 [2006 kB] 2026-03-20T17:56:15.846 INFO:teuthology.orchestra.run.vm02.stdout:Get:6 http://archive.ubuntu.com/ubuntu jammy-updates/universe amd64 libqt5dbus5 amd64 5.15.3+dfsg-2ubuntu0.2 [222 kB] 2026-03-20T17:56:15.860 INFO:teuthology.orchestra.run.vm02.stdout:Get:7 http://archive.ubuntu.com/ubuntu jammy-updates/universe amd64 libqt5network5 amd64 5.15.3+dfsg-2ubuntu0.2 [731 kB] 2026-03-20T17:56:15.901 INFO:teuthology.orchestra.run.vm02.stdout:Get:8 http://archive.ubuntu.com/ubuntu jammy/universe amd64 libthrift-0.16.0 amd64 0.16.0-2 [267 kB] 2026-03-20T17:56:15.917 INFO:teuthology.orchestra.run.vm02.stdout:Get:9 http://archive.ubuntu.com/ubuntu jammy/universe amd64 libnbd0 amd64 1.10.5-1 [71.3 kB] 2026-03-20T17:56:15.919 INFO:teuthology.orchestra.run.vm02.stdout:Get:10 http://archive.ubuntu.com/ubuntu jammy/main amd64 python3-wcwidth all 0.2.5+dfsg1-1 [21.9 kB] 2026-03-20T17:56:15.919 INFO:teuthology.orchestra.run.vm02.stdout:Get:11 http://archive.ubuntu.com/ubuntu jammy/main amd64 python3-prettytable all 2.5.0-2 [31.3 kB] 2026-03-20T17:56:15.920 INFO:teuthology.orchestra.run.vm02.stdout:Get:12 http://archive.ubuntu.com/ubuntu jammy/universe amd64 librdkafka1 amd64 1.8.0-1build1 [633 kB] 2026-03-20T17:56:15.941 INFO:teuthology.orchestra.run.vm02.stdout:Get:13 http://archive.ubuntu.com/ubuntu jammy-updates/main amd64 liboath0 amd64 2.6.7-3ubuntu0.1 [41.3 kB] 2026-03-20T17:56:15.942 INFO:teuthology.orchestra.run.vm02.stdout:Get:14 http://archive.ubuntu.com/ubuntu jammy/main amd64 python3-jaraco.functools all 3.4.0-2 [9030 B] 2026-03-20T17:56:15.975 INFO:teuthology.orchestra.run.vm02.stdout:Get:15 http://archive.ubuntu.com/ubuntu jammy-updates/main amd64 python3-cheroot all 8.5.2+ds1-1ubuntu3.1 [71.1 kB] 2026-03-20T17:56:15.992 INFO:teuthology.orchestra.run.vm08.stdout:Get:56 https://1.chacra.ceph.com/r/ceph/tentacle/70f8415b300f041766fa27faf7d5472699e32388/ubuntu/jammy/flavors/default jammy/main amd64 librados2 amd64 20.2.0-712-g70f8415b-1jammy [3583 kB] 2026-03-20T17:56:16.062 INFO:teuthology.orchestra.run.vm06.stdout:Get:56 https://1.chacra.ceph.com/r/ceph/tentacle/70f8415b300f041766fa27faf7d5472699e32388/ubuntu/jammy/flavors/default jammy/main amd64 librados2 amd64 20.2.0-712-g70f8415b-1jammy [3583 kB] 2026-03-20T17:56:16.070 INFO:teuthology.orchestra.run.vm02.stdout:Get:16 http://archive.ubuntu.com/ubuntu jammy/main amd64 python3-jaraco.classes all 3.2.1-3 [6452 B] 2026-03-20T17:56:16.070 INFO:teuthology.orchestra.run.vm02.stdout:Get:17 http://archive.ubuntu.com/ubuntu jammy/main amd64 python3-jaraco.text all 3.6.0-2 [8716 B] 2026-03-20T17:56:16.070 INFO:teuthology.orchestra.run.vm02.stdout:Get:18 http://archive.ubuntu.com/ubuntu jammy/main amd64 python3-jaraco.collections all 3.4.0-2 [11.4 kB] 2026-03-20T17:56:16.070 INFO:teuthology.orchestra.run.vm02.stdout:Get:19 http://archive.ubuntu.com/ubuntu jammy/main amd64 python3-tempora all 4.1.2-1 [14.8 kB] 2026-03-20T17:56:16.070 INFO:teuthology.orchestra.run.vm02.stdout:Get:20 http://archive.ubuntu.com/ubuntu jammy/main amd64 python3-portend all 3.0.0-1 [7240 B] 2026-03-20T17:56:16.070 INFO:teuthology.orchestra.run.vm02.stdout:Get:21 http://archive.ubuntu.com/ubuntu jammy/main amd64 python3-zc.lockfile all 2.0-1 [8980 B] 2026-03-20T17:56:16.071 INFO:teuthology.orchestra.run.vm02.stdout:Get:22 http://archive.ubuntu.com/ubuntu jammy/main amd64 python3-cherrypy3 all 18.6.1-4 [208 kB] 2026-03-20T17:56:16.071 INFO:teuthology.orchestra.run.vm02.stdout:Get:23 https://1.chacra.ceph.com/r/ceph/tentacle/70f8415b300f041766fa27faf7d5472699e32388/ubuntu/jammy/flavors/default jammy/main amd64 librados2 amd64 20.2.0-712-g70f8415b-1jammy [3583 kB] 2026-03-20T17:56:16.075 INFO:teuthology.orchestra.run.vm02.stdout:Get:24 http://archive.ubuntu.com/ubuntu jammy/universe amd64 python3-natsort all 8.0.2-1 [35.3 kB] 2026-03-20T17:56:16.103 INFO:teuthology.orchestra.run.vm02.stdout:Get:25 http://archive.ubuntu.com/ubuntu jammy/universe amd64 libfuse2 amd64 2.9.9-5ubuntu3 [90.3 kB] 2026-03-20T17:56:16.129 INFO:teuthology.orchestra.run.vm08.stdout:Get:57 https://1.chacra.ceph.com/r/ceph/tentacle/70f8415b300f041766fa27faf7d5472699e32388/ubuntu/jammy/flavors/default jammy/main amd64 libcephfs2 amd64 20.2.0-712-g70f8415b-1jammy [829 kB] 2026-03-20T17:56:16.142 INFO:teuthology.orchestra.run.vm08.stdout:Get:58 https://1.chacra.ceph.com/r/ceph/tentacle/70f8415b300f041766fa27faf7d5472699e32388/ubuntu/jammy/flavors/default jammy/main amd64 python3-rados amd64 20.2.0-712-g70f8415b-1jammy [364 kB] 2026-03-20T17:56:16.148 INFO:teuthology.orchestra.run.vm08.stdout:Get:59 https://1.chacra.ceph.com/r/ceph/tentacle/70f8415b300f041766fa27faf7d5472699e32388/ubuntu/jammy/flavors/default jammy/main amd64 python3-ceph-argparse all 20.2.0-712-g70f8415b-1jammy [32.8 kB] 2026-03-20T17:56:16.149 INFO:teuthology.orchestra.run.vm08.stdout:Get:60 https://1.chacra.ceph.com/r/ceph/tentacle/70f8415b300f041766fa27faf7d5472699e32388/ubuntu/jammy/flavors/default jammy/main amd64 python3-cephfs amd64 20.2.0-712-g70f8415b-1jammy [184 kB] 2026-03-20T17:56:16.153 INFO:teuthology.orchestra.run.vm08.stdout:Get:61 https://1.chacra.ceph.com/r/ceph/tentacle/70f8415b300f041766fa27faf7d5472699e32388/ubuntu/jammy/flavors/default jammy/main amd64 python3-ceph-common all 20.2.0-712-g70f8415b-1jammy [83.8 kB] 2026-03-20T17:56:16.154 INFO:teuthology.orchestra.run.vm08.stdout:Get:62 https://1.chacra.ceph.com/r/ceph/tentacle/70f8415b300f041766fa27faf7d5472699e32388/ubuntu/jammy/flavors/default jammy/main amd64 python3-rbd amd64 20.2.0-712-g70f8415b-1jammy [341 kB] 2026-03-20T17:56:16.158 INFO:teuthology.orchestra.run.vm05.stdout:Get:5 http://archive.ubuntu.com/ubuntu jammy-updates/universe amd64 libqt5dbus5 amd64 5.15.3+dfsg-2ubuntu0.2 [222 kB] 2026-03-20T17:56:16.159 INFO:teuthology.orchestra.run.vm08.stdout:Get:63 https://1.chacra.ceph.com/r/ceph/tentacle/70f8415b300f041766fa27faf7d5472699e32388/ubuntu/jammy/flavors/default jammy/main amd64 librgw2 amd64 20.2.0-712-g70f8415b-1jammy [8697 kB] 2026-03-20T17:56:16.174 INFO:teuthology.orchestra.run.vm05.stdout:Get:6 http://archive.ubuntu.com/ubuntu jammy-updates/universe amd64 libqt5network5 amd64 5.15.3+dfsg-2ubuntu0.2 [731 kB] 2026-03-20T17:56:16.178 INFO:teuthology.orchestra.run.vm02.stdout:Get:26 http://archive.ubuntu.com/ubuntu jammy-updates/universe amd64 python3-asyncssh all 2.5.0-1ubuntu0.1 [189 kB] 2026-03-20T17:56:16.181 INFO:teuthology.orchestra.run.vm02.stdout:Get:27 http://archive.ubuntu.com/ubuntu jammy/main amd64 python3-repoze.lru all 0.7-2 [12.1 kB] 2026-03-20T17:56:16.181 INFO:teuthology.orchestra.run.vm02.stdout:Get:28 http://archive.ubuntu.com/ubuntu jammy/main amd64 python3-routes all 2.5.1-1ubuntu1 [89.0 kB] 2026-03-20T17:56:16.183 INFO:teuthology.orchestra.run.vm02.stdout:Get:29 http://archive.ubuntu.com/ubuntu jammy/universe amd64 python3-sklearn-lib amd64 0.23.2-5ubuntu6 [2058 kB] 2026-03-20T17:56:16.192 INFO:teuthology.orchestra.run.vm06.stdout:Get:57 https://1.chacra.ceph.com/r/ceph/tentacle/70f8415b300f041766fa27faf7d5472699e32388/ubuntu/jammy/flavors/default jammy/main amd64 libcephfs2 amd64 20.2.0-712-g70f8415b-1jammy [829 kB] 2026-03-20T17:56:16.340 INFO:teuthology.orchestra.run.vm05.stdout:Get:7 http://archive.ubuntu.com/ubuntu jammy/universe amd64 libthrift-0.16.0 amd64 0.16.0-2 [267 kB] 2026-03-20T17:56:16.347 INFO:teuthology.orchestra.run.vm05.stdout:Get:8 http://archive.ubuntu.com/ubuntu jammy/universe amd64 libnbd0 amd64 1.10.5-1 [71.3 kB] 2026-03-20T17:56:16.347 INFO:teuthology.orchestra.run.vm05.stdout:Get:9 http://archive.ubuntu.com/ubuntu jammy/main amd64 python3-wcwidth all 0.2.5+dfsg1-1 [21.9 kB] 2026-03-20T17:56:16.347 INFO:teuthology.orchestra.run.vm05.stdout:Get:10 http://archive.ubuntu.com/ubuntu jammy/main amd64 python3-prettytable all 2.5.0-2 [31.3 kB] 2026-03-20T17:56:16.348 INFO:teuthology.orchestra.run.vm05.stdout:Get:11 http://archive.ubuntu.com/ubuntu jammy/universe amd64 librdkafka1 amd64 1.8.0-1build1 [633 kB] 2026-03-20T17:56:16.350 INFO:teuthology.orchestra.run.vm05.stdout:Get:12 http://archive.ubuntu.com/ubuntu jammy-updates/main amd64 liboath0 amd64 2.6.7-3ubuntu0.1 [41.3 kB] 2026-03-20T17:56:16.350 INFO:teuthology.orchestra.run.vm05.stdout:Get:13 http://archive.ubuntu.com/ubuntu jammy/main amd64 python3-jaraco.functools all 3.4.0-2 [9030 B] 2026-03-20T17:56:16.351 INFO:teuthology.orchestra.run.vm05.stdout:Get:14 http://archive.ubuntu.com/ubuntu jammy-updates/main amd64 python3-cheroot all 8.5.2+ds1-1ubuntu3.1 [71.1 kB] 2026-03-20T17:56:16.355 INFO:teuthology.orchestra.run.vm02.stdout:Get:30 http://archive.ubuntu.com/ubuntu jammy/universe amd64 python3-joblib all 0.17.0-4ubuntu1 [204 kB] 2026-03-20T17:56:16.359 INFO:teuthology.orchestra.run.vm02.stdout:Get:31 http://archive.ubuntu.com/ubuntu jammy/universe amd64 python3-threadpoolctl all 3.1.0-1 [21.3 kB] 2026-03-20T17:56:16.359 INFO:teuthology.orchestra.run.vm02.stdout:Get:32 http://archive.ubuntu.com/ubuntu jammy/universe amd64 python3-sklearn all 0.23.2-5ubuntu6 [1829 kB] 2026-03-20T17:56:16.366 INFO:teuthology.orchestra.run.vm02.stdout:Get:33 http://archive.ubuntu.com/ubuntu jammy/main amd64 python3-cachetools all 5.0.0-1 [9722 B] 2026-03-20T17:56:16.366 INFO:teuthology.orchestra.run.vm02.stdout:Get:34 http://archive.ubuntu.com/ubuntu jammy/universe amd64 python3-rsa all 4.8-1 [28.4 kB] 2026-03-20T17:56:16.366 INFO:teuthology.orchestra.run.vm02.stdout:Get:35 http://archive.ubuntu.com/ubuntu jammy/universe amd64 python3-google-auth all 1.5.1-3 [35.7 kB] 2026-03-20T17:56:16.376 INFO:teuthology.orchestra.run.vm02.stdout:Get:36 https://1.chacra.ceph.com/r/ceph/tentacle/70f8415b300f041766fa27faf7d5472699e32388/ubuntu/jammy/flavors/default jammy/main amd64 libcephfs2 amd64 20.2.0-712-g70f8415b-1jammy [829 kB] 2026-03-20T17:56:16.379 INFO:teuthology.orchestra.run.vm02.stdout:Get:37 https://1.chacra.ceph.com/r/ceph/tentacle/70f8415b300f041766fa27faf7d5472699e32388/ubuntu/jammy/flavors/default jammy/main amd64 python3-rados amd64 20.2.0-712-g70f8415b-1jammy [364 kB] 2026-03-20T17:56:16.380 INFO:teuthology.orchestra.run.vm02.stdout:Get:38 https://1.chacra.ceph.com/r/ceph/tentacle/70f8415b300f041766fa27faf7d5472699e32388/ubuntu/jammy/flavors/default jammy/main amd64 python3-ceph-argparse all 20.2.0-712-g70f8415b-1jammy [32.8 kB] 2026-03-20T17:56:16.380 INFO:teuthology.orchestra.run.vm02.stdout:Get:39 https://1.chacra.ceph.com/r/ceph/tentacle/70f8415b300f041766fa27faf7d5472699e32388/ubuntu/jammy/flavors/default jammy/main amd64 python3-cephfs amd64 20.2.0-712-g70f8415b-1jammy [184 kB] 2026-03-20T17:56:16.381 INFO:teuthology.orchestra.run.vm02.stdout:Get:40 https://1.chacra.ceph.com/r/ceph/tentacle/70f8415b300f041766fa27faf7d5472699e32388/ubuntu/jammy/flavors/default jammy/main amd64 python3-ceph-common all 20.2.0-712-g70f8415b-1jammy [83.8 kB] 2026-03-20T17:56:16.381 INFO:teuthology.orchestra.run.vm02.stdout:Get:41 https://1.chacra.ceph.com/r/ceph/tentacle/70f8415b300f041766fa27faf7d5472699e32388/ubuntu/jammy/flavors/default jammy/main amd64 python3-rbd amd64 20.2.0-712-g70f8415b-1jammy [341 kB] 2026-03-20T17:56:16.383 INFO:teuthology.orchestra.run.vm02.stdout:Get:42 https://1.chacra.ceph.com/r/ceph/tentacle/70f8415b300f041766fa27faf7d5472699e32388/ubuntu/jammy/flavors/default jammy/main amd64 librgw2 amd64 20.2.0-712-g70f8415b-1jammy [8697 kB] 2026-03-20T17:56:16.394 INFO:teuthology.orchestra.run.vm05.stdout:Get:15 http://archive.ubuntu.com/ubuntu jammy/main amd64 python3-jaraco.classes all 3.2.1-3 [6452 B] 2026-03-20T17:56:16.442 INFO:teuthology.orchestra.run.vm06.stdout:Get:58 https://1.chacra.ceph.com/r/ceph/tentacle/70f8415b300f041766fa27faf7d5472699e32388/ubuntu/jammy/flavors/default jammy/main amd64 python3-rados amd64 20.2.0-712-g70f8415b-1jammy [364 kB] 2026-03-20T17:56:16.443 INFO:teuthology.orchestra.run.vm06.stdout:Get:59 https://1.chacra.ceph.com/r/ceph/tentacle/70f8415b300f041766fa27faf7d5472699e32388/ubuntu/jammy/flavors/default jammy/main amd64 python3-ceph-argparse all 20.2.0-712-g70f8415b-1jammy [32.8 kB] 2026-03-20T17:56:16.443 INFO:teuthology.orchestra.run.vm06.stdout:Get:60 https://1.chacra.ceph.com/r/ceph/tentacle/70f8415b300f041766fa27faf7d5472699e32388/ubuntu/jammy/flavors/default jammy/main amd64 python3-cephfs amd64 20.2.0-712-g70f8415b-1jammy [184 kB] 2026-03-20T17:56:16.444 INFO:teuthology.orchestra.run.vm06.stdout:Get:61 https://1.chacra.ceph.com/r/ceph/tentacle/70f8415b300f041766fa27faf7d5472699e32388/ubuntu/jammy/flavors/default jammy/main amd64 python3-ceph-common all 20.2.0-712-g70f8415b-1jammy [83.8 kB] 2026-03-20T17:56:16.444 INFO:teuthology.orchestra.run.vm06.stdout:Get:62 https://1.chacra.ceph.com/r/ceph/tentacle/70f8415b300f041766fa27faf7d5472699e32388/ubuntu/jammy/flavors/default jammy/main amd64 python3-rbd amd64 20.2.0-712-g70f8415b-1jammy [341 kB] 2026-03-20T17:56:16.445 INFO:teuthology.orchestra.run.vm06.stdout:Get:63 https://1.chacra.ceph.com/r/ceph/tentacle/70f8415b300f041766fa27faf7d5472699e32388/ubuntu/jammy/flavors/default jammy/main amd64 librgw2 amd64 20.2.0-712-g70f8415b-1jammy [8697 kB] 2026-03-20T17:56:16.461 INFO:teuthology.orchestra.run.vm02.stdout:Get:43 http://archive.ubuntu.com/ubuntu jammy/universe amd64 python3-requests-oauthlib all 1.3.0+ds-0.1 [18.7 kB] 2026-03-20T17:56:16.468 INFO:teuthology.orchestra.run.vm05.stdout:Get:16 http://archive.ubuntu.com/ubuntu jammy/main amd64 python3-jaraco.text all 3.6.0-2 [8716 B] 2026-03-20T17:56:16.468 INFO:teuthology.orchestra.run.vm05.stdout:Get:17 http://archive.ubuntu.com/ubuntu jammy/main amd64 python3-jaraco.collections all 3.4.0-2 [11.4 kB] 2026-03-20T17:56:16.468 INFO:teuthology.orchestra.run.vm05.stdout:Get:18 http://archive.ubuntu.com/ubuntu jammy/main amd64 python3-tempora all 4.1.2-1 [14.8 kB] 2026-03-20T17:56:16.469 INFO:teuthology.orchestra.run.vm05.stdout:Get:19 http://archive.ubuntu.com/ubuntu jammy/main amd64 python3-portend all 3.0.0-1 [7240 B] 2026-03-20T17:56:16.469 INFO:teuthology.orchestra.run.vm05.stdout:Get:20 http://archive.ubuntu.com/ubuntu jammy/main amd64 python3-zc.lockfile all 2.0-1 [8980 B] 2026-03-20T17:56:16.469 INFO:teuthology.orchestra.run.vm05.stdout:Get:21 http://archive.ubuntu.com/ubuntu jammy/main amd64 python3-cherrypy3 all 18.6.1-4 [208 kB] 2026-03-20T17:56:16.473 INFO:teuthology.orchestra.run.vm05.stdout:Get:22 http://archive.ubuntu.com/ubuntu jammy/universe amd64 python3-natsort all 8.0.2-1 [35.3 kB] 2026-03-20T17:56:16.474 INFO:teuthology.orchestra.run.vm05.stdout:Get:23 http://archive.ubuntu.com/ubuntu jammy/universe amd64 libfuse2 amd64 2.9.9-5ubuntu3 [90.3 kB] 2026-03-20T17:56:16.495 INFO:teuthology.orchestra.run.vm02.stdout:Get:44 http://archive.ubuntu.com/ubuntu jammy/universe amd64 python3-websocket all 1.2.3-1 [34.7 kB] 2026-03-20T17:56:16.497 INFO:teuthology.orchestra.run.vm02.stdout:Get:45 http://archive.ubuntu.com/ubuntu jammy/universe amd64 python3-kubernetes all 12.0.1-1ubuntu1 [353 kB] 2026-03-20T17:56:16.503 INFO:teuthology.orchestra.run.vm02.stdout:Get:46 http://archive.ubuntu.com/ubuntu jammy/main amd64 libonig5 amd64 6.9.7.1-2build1 [172 kB] 2026-03-20T17:56:16.512 INFO:teuthology.orchestra.run.vm05.stdout:Get:24 http://archive.ubuntu.com/ubuntu jammy-updates/universe amd64 python3-asyncssh all 2.5.0-1ubuntu0.1 [189 kB] 2026-03-20T17:56:16.524 INFO:teuthology.orchestra.run.vm05.stdout:Get:25 https://1.chacra.ceph.com/r/ceph/tentacle/70f8415b300f041766fa27faf7d5472699e32388/ubuntu/jammy/flavors/default jammy/main amd64 librbd1 amd64 20.2.0-712-g70f8415b-1jammy [2867 kB] 2026-03-20T17:56:16.551 INFO:teuthology.orchestra.run.vm08.stdout:Get:64 https://1.chacra.ceph.com/r/ceph/tentacle/70f8415b300f041766fa27faf7d5472699e32388/ubuntu/jammy/flavors/default jammy/main amd64 python3-rgw amd64 20.2.0-712-g70f8415b-1jammy [112 kB] 2026-03-20T17:56:16.552 INFO:teuthology.orchestra.run.vm08.stdout:Get:65 https://1.chacra.ceph.com/r/ceph/tentacle/70f8415b300f041766fa27faf7d5472699e32388/ubuntu/jammy/flavors/default jammy/main amd64 libradosstriper1 amd64 20.2.0-712-g70f8415b-1jammy [261 kB] 2026-03-20T17:56:16.557 INFO:teuthology.orchestra.run.vm08.stdout:Get:66 https://1.chacra.ceph.com/r/ceph/tentacle/70f8415b300f041766fa27faf7d5472699e32388/ubuntu/jammy/flavors/default jammy/main amd64 ceph-common amd64 20.2.0-712-g70f8415b-1jammy [29.3 MB] 2026-03-20T17:56:16.624 INFO:teuthology.orchestra.run.vm02.stdout:Get:47 http://archive.ubuntu.com/ubuntu jammy-updates/main amd64 libjq1 amd64 1.6-2.1ubuntu3.1 [133 kB] 2026-03-20T17:56:16.627 INFO:teuthology.orchestra.run.vm02.stdout:Get:48 http://archive.ubuntu.com/ubuntu jammy-updates/main amd64 jq amd64 1.6-2.1ubuntu3.1 [52.5 kB] 2026-03-20T17:56:16.627 INFO:teuthology.orchestra.run.vm02.stdout:Get:49 http://archive.ubuntu.com/ubuntu jammy/main amd64 socat amd64 1.7.4.1-3ubuntu4 [349 kB] 2026-03-20T17:56:16.631 INFO:teuthology.orchestra.run.vm05.stdout:Get:26 http://archive.ubuntu.com/ubuntu jammy/main amd64 python3-repoze.lru all 0.7-2 [12.1 kB] 2026-03-20T17:56:16.631 INFO:teuthology.orchestra.run.vm05.stdout:Get:27 http://archive.ubuntu.com/ubuntu jammy/main amd64 python3-routes all 2.5.1-1ubuntu1 [89.0 kB] 2026-03-20T17:56:16.634 INFO:teuthology.orchestra.run.vm05.stdout:Get:28 http://archive.ubuntu.com/ubuntu jammy/universe amd64 python3-sklearn-lib amd64 0.23.2-5ubuntu6 [2058 kB] 2026-03-20T17:56:16.636 INFO:teuthology.orchestra.run.vm02.stdout:Get:50 http://archive.ubuntu.com/ubuntu jammy/universe amd64 xmlstarlet amd64 1.6.1-2.1 [265 kB] 2026-03-20T17:56:16.801 INFO:teuthology.orchestra.run.vm05.stdout:Get:29 http://archive.ubuntu.com/ubuntu jammy/universe amd64 python3-joblib all 0.17.0-4ubuntu1 [204 kB] 2026-03-20T17:56:16.801 INFO:teuthology.orchestra.run.vm05.stdout:Get:30 http://archive.ubuntu.com/ubuntu jammy/universe amd64 python3-threadpoolctl all 3.1.0-1 [21.3 kB] 2026-03-20T17:56:16.802 INFO:teuthology.orchestra.run.vm05.stdout:Get:31 http://archive.ubuntu.com/ubuntu jammy/universe amd64 python3-sklearn all 0.23.2-5ubuntu6 [1829 kB] 2026-03-20T17:56:16.830 INFO:teuthology.orchestra.run.vm02.stdout:Get:51 http://archive.ubuntu.com/ubuntu jammy-updates/main amd64 nvme-cli amd64 1.16-3ubuntu0.3 [474 kB] 2026-03-20T17:56:16.832 INFO:teuthology.orchestra.run.vm02.stdout:Get:52 http://archive.ubuntu.com/ubuntu jammy-updates/universe amd64 python-asyncssh-doc all 2.5.0-1ubuntu0.1 [309 kB] 2026-03-20T17:56:16.833 INFO:teuthology.orchestra.run.vm02.stdout:Get:53 http://archive.ubuntu.com/ubuntu jammy/universe amd64 python3-iniconfig all 1.1.1-2 [6024 B] 2026-03-20T17:56:16.833 INFO:teuthology.orchestra.run.vm02.stdout:Get:54 http://archive.ubuntu.com/ubuntu jammy/universe amd64 python3-pluggy all 0.13.0-7.1 [19.0 kB] 2026-03-20T17:56:16.833 INFO:teuthology.orchestra.run.vm02.stdout:Get:55 http://archive.ubuntu.com/ubuntu jammy/main amd64 python3-psutil amd64 5.9.0-1build1 [158 kB] 2026-03-20T17:56:16.834 INFO:teuthology.orchestra.run.vm02.stdout:Get:56 http://archive.ubuntu.com/ubuntu jammy/universe amd64 python3-py all 1.10.0-1 [71.9 kB] 2026-03-20T17:56:16.835 INFO:teuthology.orchestra.run.vm02.stdout:Get:57 http://archive.ubuntu.com/ubuntu jammy-updates/main amd64 python3-pygments all 2.11.2+dfsg-2ubuntu0.1 [750 kB] 2026-03-20T17:56:16.840 INFO:teuthology.orchestra.run.vm06.stdout:Get:64 https://1.chacra.ceph.com/r/ceph/tentacle/70f8415b300f041766fa27faf7d5472699e32388/ubuntu/jammy/flavors/default jammy/main amd64 python3-rgw amd64 20.2.0-712-g70f8415b-1jammy [112 kB] 2026-03-20T17:56:16.846 INFO:teuthology.orchestra.run.vm06.stdout:Get:65 https://1.chacra.ceph.com/r/ceph/tentacle/70f8415b300f041766fa27faf7d5472699e32388/ubuntu/jammy/flavors/default jammy/main amd64 libradosstriper1 amd64 20.2.0-712-g70f8415b-1jammy [261 kB] 2026-03-20T17:56:16.849 INFO:teuthology.orchestra.run.vm06.stdout:Get:66 https://1.chacra.ceph.com/r/ceph/tentacle/70f8415b300f041766fa27faf7d5472699e32388/ubuntu/jammy/flavors/default jammy/main amd64 ceph-common amd64 20.2.0-712-g70f8415b-1jammy [29.3 MB] 2026-03-20T17:56:16.960 INFO:teuthology.orchestra.run.vm05.stdout:Get:32 http://archive.ubuntu.com/ubuntu jammy/main amd64 python3-cachetools all 5.0.0-1 [9722 B] 2026-03-20T17:56:16.960 INFO:teuthology.orchestra.run.vm05.stdout:Get:33 http://archive.ubuntu.com/ubuntu jammy/universe amd64 python3-rsa all 4.8-1 [28.4 kB] 2026-03-20T17:56:16.961 INFO:teuthology.orchestra.run.vm05.stdout:Get:34 http://archive.ubuntu.com/ubuntu jammy/universe amd64 python3-google-auth all 1.5.1-3 [35.7 kB] 2026-03-20T17:56:16.961 INFO:teuthology.orchestra.run.vm05.stdout:Get:35 http://archive.ubuntu.com/ubuntu jammy/universe amd64 python3-requests-oauthlib all 1.3.0+ds-0.1 [18.7 kB] 2026-03-20T17:56:16.961 INFO:teuthology.orchestra.run.vm05.stdout:Get:36 http://archive.ubuntu.com/ubuntu jammy/universe amd64 python3-websocket all 1.2.3-1 [34.7 kB] 2026-03-20T17:56:16.962 INFO:teuthology.orchestra.run.vm05.stdout:Get:37 http://archive.ubuntu.com/ubuntu jammy/universe amd64 python3-kubernetes all 12.0.1-1ubuntu1 [353 kB] 2026-03-20T17:56:16.976 INFO:teuthology.orchestra.run.vm02.stdout:Get:58 http://archive.ubuntu.com/ubuntu jammy/universe amd64 python3-toml all 0.10.2-1 [16.5 kB] 2026-03-20T17:56:16.977 INFO:teuthology.orchestra.run.vm02.stdout:Get:59 http://archive.ubuntu.com/ubuntu jammy/universe amd64 python3-pytest all 6.2.5-1ubuntu2 [214 kB] 2026-03-20T17:56:16.979 INFO:teuthology.orchestra.run.vm02.stdout:Get:60 http://archive.ubuntu.com/ubuntu jammy/main amd64 python3-simplejson amd64 3.17.6-1build1 [54.7 kB] 2026-03-20T17:56:16.980 INFO:teuthology.orchestra.run.vm02.stdout:Get:61 http://archive.ubuntu.com/ubuntu jammy-updates/main amd64 python3-webob all 1:1.8.6-1.1ubuntu0.1 [86.7 kB] 2026-03-20T17:56:17.113 INFO:teuthology.orchestra.run.vm02.stdout:Get:62 http://archive.ubuntu.com/ubuntu jammy/universe amd64 qttranslations5-l10n all 5.15.3-1 [1983 kB] 2026-03-20T17:56:17.114 INFO:teuthology.orchestra.run.vm02.stdout:Get:63 https://1.chacra.ceph.com/r/ceph/tentacle/70f8415b300f041766fa27faf7d5472699e32388/ubuntu/jammy/flavors/default jammy/main amd64 python3-rgw amd64 20.2.0-712-g70f8415b-1jammy [112 kB] 2026-03-20T17:56:17.116 INFO:teuthology.orchestra.run.vm05.stdout:Get:38 http://archive.ubuntu.com/ubuntu jammy/main amd64 libonig5 amd64 6.9.7.1-2build1 [172 kB] 2026-03-20T17:56:17.117 INFO:teuthology.orchestra.run.vm02.stdout:Get:64 https://1.chacra.ceph.com/r/ceph/tentacle/70f8415b300f041766fa27faf7d5472699e32388/ubuntu/jammy/flavors/default jammy/main amd64 libradosstriper1 amd64 20.2.0-712-g70f8415b-1jammy [261 kB] 2026-03-20T17:56:17.117 INFO:teuthology.orchestra.run.vm05.stdout:Get:39 http://archive.ubuntu.com/ubuntu jammy-updates/main amd64 libjq1 amd64 1.6-2.1ubuntu3.1 [133 kB] 2026-03-20T17:56:17.118 INFO:teuthology.orchestra.run.vm05.stdout:Get:40 http://archive.ubuntu.com/ubuntu jammy-updates/main amd64 jq amd64 1.6-2.1ubuntu3.1 [52.5 kB] 2026-03-20T17:56:17.146 INFO:teuthology.orchestra.run.vm02.stdout:Get:65 http://archive.ubuntu.com/ubuntu jammy-updates/main amd64 smartmontools amd64 7.2-1ubuntu0.1 [583 kB] 2026-03-20T17:56:17.147 INFO:teuthology.orchestra.run.vm05.stdout:Get:41 http://archive.ubuntu.com/ubuntu jammy/main amd64 socat amd64 1.7.4.1-3ubuntu4 [349 kB] 2026-03-20T17:56:17.208 INFO:teuthology.orchestra.run.vm02.stdout:Get:66 https://1.chacra.ceph.com/r/ceph/tentacle/70f8415b300f041766fa27faf7d5472699e32388/ubuntu/jammy/flavors/default jammy/main amd64 ceph-common amd64 20.2.0-712-g70f8415b-1jammy [29.3 MB] 2026-03-20T17:56:17.305 INFO:teuthology.orchestra.run.vm05.stdout:Get:42 http://archive.ubuntu.com/ubuntu jammy/universe amd64 xmlstarlet amd64 1.6.1-2.1 [265 kB] 2026-03-20T17:56:17.306 INFO:teuthology.orchestra.run.vm05.stdout:Get:43 http://archive.ubuntu.com/ubuntu jammy-updates/main amd64 nvme-cli amd64 1.16-3ubuntu0.3 [474 kB] 2026-03-20T17:56:17.308 INFO:teuthology.orchestra.run.vm05.stdout:Get:44 http://archive.ubuntu.com/ubuntu jammy-updates/universe amd64 python-asyncssh-doc all 2.5.0-1ubuntu0.1 [309 kB] 2026-03-20T17:56:17.309 INFO:teuthology.orchestra.run.vm05.stdout:Get:45 http://archive.ubuntu.com/ubuntu jammy/universe amd64 python3-iniconfig all 1.1.1-2 [6024 B] 2026-03-20T17:56:17.310 INFO:teuthology.orchestra.run.vm05.stdout:Get:46 http://archive.ubuntu.com/ubuntu jammy/universe amd64 python3-pluggy all 0.13.0-7.1 [19.0 kB] 2026-03-20T17:56:17.310 INFO:teuthology.orchestra.run.vm05.stdout:Get:47 http://archive.ubuntu.com/ubuntu jammy/main amd64 python3-psutil amd64 5.9.0-1build1 [158 kB] 2026-03-20T17:56:17.311 INFO:teuthology.orchestra.run.vm05.stdout:Get:48 http://archive.ubuntu.com/ubuntu jammy/universe amd64 python3-py all 1.10.0-1 [71.9 kB] 2026-03-20T17:56:17.311 INFO:teuthology.orchestra.run.vm05.stdout:Get:49 http://archive.ubuntu.com/ubuntu jammy-updates/main amd64 python3-pygments all 2.11.2+dfsg-2ubuntu0.1 [750 kB] 2026-03-20T17:56:17.322 INFO:teuthology.orchestra.run.vm05.stdout:Get:50 http://archive.ubuntu.com/ubuntu jammy/universe amd64 python3-toml all 0.10.2-1 [16.5 kB] 2026-03-20T17:56:17.425 INFO:teuthology.orchestra.run.vm05.stdout:Get:51 http://archive.ubuntu.com/ubuntu jammy/universe amd64 python3-pytest all 6.2.5-1ubuntu2 [214 kB] 2026-03-20T17:56:17.427 INFO:teuthology.orchestra.run.vm05.stdout:Get:52 http://archive.ubuntu.com/ubuntu jammy/main amd64 python3-simplejson amd64 3.17.6-1build1 [54.7 kB] 2026-03-20T17:56:17.541 INFO:teuthology.orchestra.run.vm05.stdout:Get:53 http://archive.ubuntu.com/ubuntu jammy-updates/main amd64 python3-webob all 1:1.8.6-1.1ubuntu0.1 [86.7 kB] 2026-03-20T17:56:17.542 INFO:teuthology.orchestra.run.vm05.stdout:Get:54 http://archive.ubuntu.com/ubuntu jammy/universe amd64 qttranslations5-l10n all 5.15.3-1 [1983 kB] 2026-03-20T17:56:17.551 INFO:teuthology.orchestra.run.vm05.stdout:Get:55 https://1.chacra.ceph.com/r/ceph/tentacle/70f8415b300f041766fa27faf7d5472699e32388/ubuntu/jammy/flavors/default jammy/main amd64 librados2 amd64 20.2.0-712-g70f8415b-1jammy [3583 kB] 2026-03-20T17:56:17.733 INFO:teuthology.orchestra.run.vm05.stdout:Get:56 http://archive.ubuntu.com/ubuntu jammy-updates/main amd64 smartmontools amd64 7.2-1ubuntu0.1 [583 kB] 2026-03-20T17:56:17.866 INFO:teuthology.orchestra.run.vm05.stdout:Get:57 https://1.chacra.ceph.com/r/ceph/tentacle/70f8415b300f041766fa27faf7d5472699e32388/ubuntu/jammy/flavors/default jammy/main amd64 libcephfs2 amd64 20.2.0-712-g70f8415b-1jammy [829 kB] 2026-03-20T17:56:17.988 INFO:teuthology.orchestra.run.vm05.stdout:Get:58 https://1.chacra.ceph.com/r/ceph/tentacle/70f8415b300f041766fa27faf7d5472699e32388/ubuntu/jammy/flavors/default jammy/main amd64 python3-rados amd64 20.2.0-712-g70f8415b-1jammy [364 kB] 2026-03-20T17:56:17.998 INFO:teuthology.orchestra.run.vm05.stdout:Get:59 https://1.chacra.ceph.com/r/ceph/tentacle/70f8415b300f041766fa27faf7d5472699e32388/ubuntu/jammy/flavors/default jammy/main amd64 python3-ceph-argparse all 20.2.0-712-g70f8415b-1jammy [32.8 kB] 2026-03-20T17:56:17.999 INFO:teuthology.orchestra.run.vm05.stdout:Get:60 https://1.chacra.ceph.com/r/ceph/tentacle/70f8415b300f041766fa27faf7d5472699e32388/ubuntu/jammy/flavors/default jammy/main amd64 python3-cephfs amd64 20.2.0-712-g70f8415b-1jammy [184 kB] 2026-03-20T17:56:18.005 INFO:teuthology.orchestra.run.vm05.stdout:Get:61 https://1.chacra.ceph.com/r/ceph/tentacle/70f8415b300f041766fa27faf7d5472699e32388/ubuntu/jammy/flavors/default jammy/main amd64 python3-ceph-common all 20.2.0-712-g70f8415b-1jammy [83.8 kB] 2026-03-20T17:56:18.007 INFO:teuthology.orchestra.run.vm05.stdout:Get:62 https://1.chacra.ceph.com/r/ceph/tentacle/70f8415b300f041766fa27faf7d5472699e32388/ubuntu/jammy/flavors/default jammy/main amd64 python3-rbd amd64 20.2.0-712-g70f8415b-1jammy [341 kB] 2026-03-20T17:56:18.013 INFO:teuthology.orchestra.run.vm05.stdout:Get:63 https://1.chacra.ceph.com/r/ceph/tentacle/70f8415b300f041766fa27faf7d5472699e32388/ubuntu/jammy/flavors/default jammy/main amd64 librgw2 amd64 20.2.0-712-g70f8415b-1jammy [8697 kB] 2026-03-20T17:56:18.219 INFO:teuthology.orchestra.run.vm08.stdout:Get:67 https://1.chacra.ceph.com/r/ceph/tentacle/70f8415b300f041766fa27faf7d5472699e32388/ubuntu/jammy/flavors/default jammy/main amd64 ceph-base amd64 20.2.0-712-g70f8415b-1jammy [5415 kB] 2026-03-20T17:56:18.347 INFO:teuthology.orchestra.run.vm06.stdout:Get:67 https://1.chacra.ceph.com/r/ceph/tentacle/70f8415b300f041766fa27faf7d5472699e32388/ubuntu/jammy/flavors/default jammy/main amd64 ceph-base amd64 20.2.0-712-g70f8415b-1jammy [5415 kB] 2026-03-20T17:56:18.532 INFO:teuthology.orchestra.run.vm08.stdout:Get:68 https://1.chacra.ceph.com/r/ceph/tentacle/70f8415b300f041766fa27faf7d5472699e32388/ubuntu/jammy/flavors/default jammy/main amd64 ceph-mgr-modules-core all 20.2.0-712-g70f8415b-1jammy [246 kB] 2026-03-20T17:56:18.535 INFO:teuthology.orchestra.run.vm08.stdout:Get:69 https://1.chacra.ceph.com/r/ceph/tentacle/70f8415b300f041766fa27faf7d5472699e32388/ubuntu/jammy/flavors/default jammy/main amd64 libsqlite3-mod-ceph amd64 20.2.0-712-g70f8415b-1jammy [124 kB] 2026-03-20T17:56:18.537 INFO:teuthology.orchestra.run.vm08.stdout:Get:70 https://1.chacra.ceph.com/r/ceph/tentacle/70f8415b300f041766fa27faf7d5472699e32388/ubuntu/jammy/flavors/default jammy/main amd64 ceph-mgr amd64 20.2.0-712-g70f8415b-1jammy [906 kB] 2026-03-20T17:56:18.552 INFO:teuthology.orchestra.run.vm08.stdout:Get:71 https://1.chacra.ceph.com/r/ceph/tentacle/70f8415b300f041766fa27faf7d5472699e32388/ubuntu/jammy/flavors/default jammy/main amd64 ceph-mon amd64 20.2.0-712-g70f8415b-1jammy [6399 kB] 2026-03-20T17:56:18.652 INFO:teuthology.orchestra.run.vm06.stdout:Get:68 https://1.chacra.ceph.com/r/ceph/tentacle/70f8415b300f041766fa27faf7d5472699e32388/ubuntu/jammy/flavors/default jammy/main amd64 ceph-mgr-modules-core all 20.2.0-712-g70f8415b-1jammy [246 kB] 2026-03-20T17:56:18.659 INFO:teuthology.orchestra.run.vm05.stdout:Get:64 https://1.chacra.ceph.com/r/ceph/tentacle/70f8415b300f041766fa27faf7d5472699e32388/ubuntu/jammy/flavors/default jammy/main amd64 python3-rgw amd64 20.2.0-712-g70f8415b-1jammy [112 kB] 2026-03-20T17:56:18.662 INFO:teuthology.orchestra.run.vm05.stdout:Get:65 https://1.chacra.ceph.com/r/ceph/tentacle/70f8415b300f041766fa27faf7d5472699e32388/ubuntu/jammy/flavors/default jammy/main amd64 libradosstriper1 amd64 20.2.0-712-g70f8415b-1jammy [261 kB] 2026-03-20T17:56:18.663 INFO:teuthology.orchestra.run.vm06.stdout:Get:69 https://1.chacra.ceph.com/r/ceph/tentacle/70f8415b300f041766fa27faf7d5472699e32388/ubuntu/jammy/flavors/default jammy/main amd64 libsqlite3-mod-ceph amd64 20.2.0-712-g70f8415b-1jammy [124 kB] 2026-03-20T17:56:18.668 INFO:teuthology.orchestra.run.vm06.stdout:Get:70 https://1.chacra.ceph.com/r/ceph/tentacle/70f8415b300f041766fa27faf7d5472699e32388/ubuntu/jammy/flavors/default jammy/main amd64 ceph-mgr amd64 20.2.0-712-g70f8415b-1jammy [906 kB] 2026-03-20T17:56:18.668 INFO:teuthology.orchestra.run.vm05.stdout:Get:66 https://1.chacra.ceph.com/r/ceph/tentacle/70f8415b300f041766fa27faf7d5472699e32388/ubuntu/jammy/flavors/default jammy/main amd64 ceph-common amd64 20.2.0-712-g70f8415b-1jammy [29.3 MB] 2026-03-20T17:56:18.777 INFO:teuthology.orchestra.run.vm06.stdout:Get:71 https://1.chacra.ceph.com/r/ceph/tentacle/70f8415b300f041766fa27faf7d5472699e32388/ubuntu/jammy/flavors/default jammy/main amd64 ceph-mon amd64 20.2.0-712-g70f8415b-1jammy [6399 kB] 2026-03-20T17:56:18.900 INFO:teuthology.orchestra.run.vm08.stdout:Get:72 https://1.chacra.ceph.com/r/ceph/tentacle/70f8415b300f041766fa27faf7d5472699e32388/ubuntu/jammy/flavors/default jammy/main amd64 ceph-osd amd64 20.2.0-712-g70f8415b-1jammy [21.7 MB] 2026-03-20T17:56:19.125 INFO:teuthology.orchestra.run.vm06.stdout:Get:72 https://1.chacra.ceph.com/r/ceph/tentacle/70f8415b300f041766fa27faf7d5472699e32388/ubuntu/jammy/flavors/default jammy/main amd64 ceph-osd amd64 20.2.0-712-g70f8415b-1jammy [21.7 MB] 2026-03-20T17:56:19.880 INFO:teuthology.orchestra.run.vm02.stdout:Get:67 https://1.chacra.ceph.com/r/ceph/tentacle/70f8415b300f041766fa27faf7d5472699e32388/ubuntu/jammy/flavors/default jammy/main amd64 ceph-base amd64 20.2.0-712-g70f8415b-1jammy [5415 kB] 2026-03-20T17:56:20.314 INFO:teuthology.orchestra.run.vm06.stdout:Get:73 https://1.chacra.ceph.com/r/ceph/tentacle/70f8415b300f041766fa27faf7d5472699e32388/ubuntu/jammy/flavors/default jammy/main amd64 ceph amd64 20.2.0-712-g70f8415b-1jammy [14.1 kB] 2026-03-20T17:56:20.314 INFO:teuthology.orchestra.run.vm06.stdout:Get:74 https://1.chacra.ceph.com/r/ceph/tentacle/70f8415b300f041766fa27faf7d5472699e32388/ubuntu/jammy/flavors/default jammy/main amd64 ceph-fuse amd64 20.2.0-712-g70f8415b-1jammy [955 kB] 2026-03-20T17:56:20.343 INFO:teuthology.orchestra.run.vm06.stdout:Get:75 https://1.chacra.ceph.com/r/ceph/tentacle/70f8415b300f041766fa27faf7d5472699e32388/ubuntu/jammy/flavors/default jammy/main amd64 ceph-mds amd64 20.2.0-712-g70f8415b-1jammy [2341 kB] 2026-03-20T17:56:20.359 INFO:teuthology.orchestra.run.vm02.stdout:Get:68 https://1.chacra.ceph.com/r/ceph/tentacle/70f8415b300f041766fa27faf7d5472699e32388/ubuntu/jammy/flavors/default jammy/main amd64 ceph-mgr-modules-core all 20.2.0-712-g70f8415b-1jammy [246 kB] 2026-03-20T17:56:20.365 INFO:teuthology.orchestra.run.vm08.stdout:Get:73 https://1.chacra.ceph.com/r/ceph/tentacle/70f8415b300f041766fa27faf7d5472699e32388/ubuntu/jammy/flavors/default jammy/main amd64 ceph amd64 20.2.0-712-g70f8415b-1jammy [14.1 kB] 2026-03-20T17:56:20.365 INFO:teuthology.orchestra.run.vm08.stdout:Get:74 https://1.chacra.ceph.com/r/ceph/tentacle/70f8415b300f041766fa27faf7d5472699e32388/ubuntu/jammy/flavors/default jammy/main amd64 ceph-fuse amd64 20.2.0-712-g70f8415b-1jammy [955 kB] 2026-03-20T17:56:20.372 INFO:teuthology.orchestra.run.vm08.stdout:Get:75 https://1.chacra.ceph.com/r/ceph/tentacle/70f8415b300f041766fa27faf7d5472699e32388/ubuntu/jammy/flavors/default jammy/main amd64 ceph-mds amd64 20.2.0-712-g70f8415b-1jammy [2341 kB] 2026-03-20T17:56:20.392 INFO:teuthology.orchestra.run.vm02.stdout:Get:69 https://1.chacra.ceph.com/r/ceph/tentacle/70f8415b300f041766fa27faf7d5472699e32388/ubuntu/jammy/flavors/default jammy/main amd64 libsqlite3-mod-ceph amd64 20.2.0-712-g70f8415b-1jammy [124 kB] 2026-03-20T17:56:20.397 INFO:teuthology.orchestra.run.vm02.stdout:Get:70 https://1.chacra.ceph.com/r/ceph/tentacle/70f8415b300f041766fa27faf7d5472699e32388/ubuntu/jammy/flavors/default jammy/main amd64 ceph-mgr amd64 20.2.0-712-g70f8415b-1jammy [906 kB] 2026-03-20T17:56:20.483 INFO:teuthology.orchestra.run.vm06.stdout:Get:76 https://1.chacra.ceph.com/r/ceph/tentacle/70f8415b300f041766fa27faf7d5472699e32388/ubuntu/jammy/flavors/default jammy/main amd64 cephadm amd64 20.2.0-712-g70f8415b-1jammy [1049 kB] 2026-03-20T17:56:20.495 INFO:teuthology.orchestra.run.vm02.stdout:Get:71 https://1.chacra.ceph.com/r/ceph/tentacle/70f8415b300f041766fa27faf7d5472699e32388/ubuntu/jammy/flavors/default jammy/main amd64 ceph-mon amd64 20.2.0-712-g70f8415b-1jammy [6399 kB] 2026-03-20T17:56:20.529 INFO:teuthology.orchestra.run.vm06.stdout:Get:77 https://1.chacra.ceph.com/r/ceph/tentacle/70f8415b300f041766fa27faf7d5472699e32388/ubuntu/jammy/flavors/default jammy/main amd64 ceph-mgr-cephadm all 20.2.0-712-g70f8415b-1jammy [179 kB] 2026-03-20T17:56:20.537 INFO:teuthology.orchestra.run.vm08.stdout:Get:76 https://1.chacra.ceph.com/r/ceph/tentacle/70f8415b300f041766fa27faf7d5472699e32388/ubuntu/jammy/flavors/default jammy/main amd64 cephadm amd64 20.2.0-712-g70f8415b-1jammy [1049 kB] 2026-03-20T17:56:20.569 INFO:teuthology.orchestra.run.vm08.stdout:Get:77 https://1.chacra.ceph.com/r/ceph/tentacle/70f8415b300f041766fa27faf7d5472699e32388/ubuntu/jammy/flavors/default jammy/main amd64 ceph-mgr-cephadm all 20.2.0-712-g70f8415b-1jammy [179 kB] 2026-03-20T17:56:20.596 INFO:teuthology.orchestra.run.vm06.stdout:Get:78 https://1.chacra.ceph.com/r/ceph/tentacle/70f8415b300f041766fa27faf7d5472699e32388/ubuntu/jammy/flavors/default jammy/main amd64 ceph-mgr-dashboard all 20.2.0-712-g70f8415b-1jammy [45.5 MB] 2026-03-20T17:56:20.597 INFO:teuthology.orchestra.run.vm05.stdout:Get:67 https://1.chacra.ceph.com/r/ceph/tentacle/70f8415b300f041766fa27faf7d5472699e32388/ubuntu/jammy/flavors/default jammy/main amd64 ceph-base amd64 20.2.0-712-g70f8415b-1jammy [5415 kB] 2026-03-20T17:56:20.659 INFO:teuthology.orchestra.run.vm08.stdout:Get:78 https://1.chacra.ceph.com/r/ceph/tentacle/70f8415b300f041766fa27faf7d5472699e32388/ubuntu/jammy/flavors/default jammy/main amd64 ceph-mgr-dashboard all 20.2.0-712-g70f8415b-1jammy [45.5 MB] 2026-03-20T17:56:20.920 INFO:teuthology.orchestra.run.vm05.stdout:Get:68 https://1.chacra.ceph.com/r/ceph/tentacle/70f8415b300f041766fa27faf7d5472699e32388/ubuntu/jammy/flavors/default jammy/main amd64 ceph-mgr-modules-core all 20.2.0-712-g70f8415b-1jammy [246 kB] 2026-03-20T17:56:20.925 INFO:teuthology.orchestra.run.vm05.stdout:Get:69 https://1.chacra.ceph.com/r/ceph/tentacle/70f8415b300f041766fa27faf7d5472699e32388/ubuntu/jammy/flavors/default jammy/main amd64 libsqlite3-mod-ceph amd64 20.2.0-712-g70f8415b-1jammy [124 kB] 2026-03-20T17:56:20.930 INFO:teuthology.orchestra.run.vm05.stdout:Get:70 https://1.chacra.ceph.com/r/ceph/tentacle/70f8415b300f041766fa27faf7d5472699e32388/ubuntu/jammy/flavors/default jammy/main amd64 ceph-mgr amd64 20.2.0-712-g70f8415b-1jammy [906 kB] 2026-03-20T17:56:21.042 INFO:teuthology.orchestra.run.vm05.stdout:Get:71 https://1.chacra.ceph.com/r/ceph/tentacle/70f8415b300f041766fa27faf7d5472699e32388/ubuntu/jammy/flavors/default jammy/main amd64 ceph-mon amd64 20.2.0-712-g70f8415b-1jammy [6399 kB] 2026-03-20T17:56:21.067 INFO:teuthology.orchestra.run.vm02.stdout:Get:72 https://1.chacra.ceph.com/r/ceph/tentacle/70f8415b300f041766fa27faf7d5472699e32388/ubuntu/jammy/flavors/default jammy/main amd64 ceph-osd amd64 20.2.0-712-g70f8415b-1jammy [21.7 MB] 2026-03-20T17:56:21.409 INFO:teuthology.orchestra.run.vm05.stdout:Get:72 https://1.chacra.ceph.com/r/ceph/tentacle/70f8415b300f041766fa27faf7d5472699e32388/ubuntu/jammy/flavors/default jammy/main amd64 ceph-osd amd64 20.2.0-712-g70f8415b-1jammy [21.7 MB] 2026-03-20T17:56:22.748 INFO:teuthology.orchestra.run.vm05.stdout:Get:73 https://1.chacra.ceph.com/r/ceph/tentacle/70f8415b300f041766fa27faf7d5472699e32388/ubuntu/jammy/flavors/default jammy/main amd64 ceph amd64 20.2.0-712-g70f8415b-1jammy [14.1 kB] 2026-03-20T17:56:22.748 INFO:teuthology.orchestra.run.vm05.stdout:Get:74 https://1.chacra.ceph.com/r/ceph/tentacle/70f8415b300f041766fa27faf7d5472699e32388/ubuntu/jammy/flavors/default jammy/main amd64 ceph-fuse amd64 20.2.0-712-g70f8415b-1jammy [955 kB] 2026-03-20T17:56:22.762 INFO:teuthology.orchestra.run.vm05.stdout:Get:75 https://1.chacra.ceph.com/r/ceph/tentacle/70f8415b300f041766fa27faf7d5472699e32388/ubuntu/jammy/flavors/default jammy/main amd64 ceph-mds amd64 20.2.0-712-g70f8415b-1jammy [2341 kB] 2026-03-20T17:56:22.926 INFO:teuthology.orchestra.run.vm05.stdout:Get:76 https://1.chacra.ceph.com/r/ceph/tentacle/70f8415b300f041766fa27faf7d5472699e32388/ubuntu/jammy/flavors/default jammy/main amd64 cephadm amd64 20.2.0-712-g70f8415b-1jammy [1049 kB] 2026-03-20T17:56:22.966 INFO:teuthology.orchestra.run.vm05.stdout:Get:77 https://1.chacra.ceph.com/r/ceph/tentacle/70f8415b300f041766fa27faf7d5472699e32388/ubuntu/jammy/flavors/default jammy/main amd64 ceph-mgr-cephadm all 20.2.0-712-g70f8415b-1jammy [179 kB] 2026-03-20T17:56:22.974 INFO:teuthology.orchestra.run.vm05.stdout:Get:78 https://1.chacra.ceph.com/r/ceph/tentacle/70f8415b300f041766fa27faf7d5472699e32388/ubuntu/jammy/flavors/default jammy/main amd64 ceph-mgr-dashboard all 20.2.0-712-g70f8415b-1jammy [45.5 MB] 2026-03-20T17:56:23.039 INFO:teuthology.orchestra.run.vm02.stdout:Get:73 https://1.chacra.ceph.com/r/ceph/tentacle/70f8415b300f041766fa27faf7d5472699e32388/ubuntu/jammy/flavors/default jammy/main amd64 ceph amd64 20.2.0-712-g70f8415b-1jammy [14.1 kB] 2026-03-20T17:56:23.041 INFO:teuthology.orchestra.run.vm02.stdout:Get:74 https://1.chacra.ceph.com/r/ceph/tentacle/70f8415b300f041766fa27faf7d5472699e32388/ubuntu/jammy/flavors/default jammy/main amd64 ceph-fuse amd64 20.2.0-712-g70f8415b-1jammy [955 kB] 2026-03-20T17:56:23.146 INFO:teuthology.orchestra.run.vm06.stdout:Get:79 https://1.chacra.ceph.com/r/ceph/tentacle/70f8415b300f041766fa27faf7d5472699e32388/ubuntu/jammy/flavors/default jammy/main amd64 ceph-mgr-diskprediction-local all 20.2.0-712-g70f8415b-1jammy [8625 kB] 2026-03-20T17:56:23.160 INFO:teuthology.orchestra.run.vm02.stdout:Get:75 https://1.chacra.ceph.com/r/ceph/tentacle/70f8415b300f041766fa27faf7d5472699e32388/ubuntu/jammy/flavors/default jammy/main amd64 ceph-mds amd64 20.2.0-712-g70f8415b-1jammy [2341 kB] 2026-03-20T17:56:23.390 INFO:teuthology.orchestra.run.vm02.stdout:Get:76 https://1.chacra.ceph.com/r/ceph/tentacle/70f8415b300f041766fa27faf7d5472699e32388/ubuntu/jammy/flavors/default jammy/main amd64 cephadm amd64 20.2.0-712-g70f8415b-1jammy [1049 kB] 2026-03-20T17:56:23.461 INFO:teuthology.orchestra.run.vm02.stdout:Get:77 https://1.chacra.ceph.com/r/ceph/tentacle/70f8415b300f041766fa27faf7d5472699e32388/ubuntu/jammy/flavors/default jammy/main amd64 ceph-mgr-cephadm all 20.2.0-712-g70f8415b-1jammy [179 kB] 2026-03-20T17:56:23.474 INFO:teuthology.orchestra.run.vm02.stdout:Get:78 https://1.chacra.ceph.com/r/ceph/tentacle/70f8415b300f041766fa27faf7d5472699e32388/ubuntu/jammy/flavors/default jammy/main amd64 ceph-mgr-dashboard all 20.2.0-712-g70f8415b-1jammy [45.5 MB] 2026-03-20T17:56:23.576 INFO:teuthology.orchestra.run.vm06.stdout:Get:80 https://1.chacra.ceph.com/r/ceph/tentacle/70f8415b300f041766fa27faf7d5472699e32388/ubuntu/jammy/flavors/default jammy/main amd64 ceph-mgr-k8sevents all 20.2.0-712-g70f8415b-1jammy [14.2 kB] 2026-03-20T17:56:23.576 INFO:teuthology.orchestra.run.vm06.stdout:Get:81 https://1.chacra.ceph.com/r/ceph/tentacle/70f8415b300f041766fa27faf7d5472699e32388/ubuntu/jammy/flavors/default jammy/main amd64 ceph-test amd64 20.2.0-712-g70f8415b-1jammy [99.5 MB] 2026-03-20T17:56:23.706 INFO:teuthology.orchestra.run.vm08.stdout:Get:79 https://1.chacra.ceph.com/r/ceph/tentacle/70f8415b300f041766fa27faf7d5472699e32388/ubuntu/jammy/flavors/default jammy/main amd64 ceph-mgr-diskprediction-local all 20.2.0-712-g70f8415b-1jammy [8625 kB] 2026-03-20T17:56:24.330 INFO:teuthology.orchestra.run.vm08.stdout:Get:80 https://1.chacra.ceph.com/r/ceph/tentacle/70f8415b300f041766fa27faf7d5472699e32388/ubuntu/jammy/flavors/default jammy/main amd64 ceph-mgr-k8sevents all 20.2.0-712-g70f8415b-1jammy [14.2 kB] 2026-03-20T17:56:24.331 INFO:teuthology.orchestra.run.vm08.stdout:Get:81 https://1.chacra.ceph.com/r/ceph/tentacle/70f8415b300f041766fa27faf7d5472699e32388/ubuntu/jammy/flavors/default jammy/main amd64 ceph-test amd64 20.2.0-712-g70f8415b-1jammy [99.5 MB] 2026-03-20T17:56:26.290 INFO:teuthology.orchestra.run.vm05.stdout:Get:79 https://1.chacra.ceph.com/r/ceph/tentacle/70f8415b300f041766fa27faf7d5472699e32388/ubuntu/jammy/flavors/default jammy/main amd64 ceph-mgr-diskprediction-local all 20.2.0-712-g70f8415b-1jammy [8625 kB] 2026-03-20T17:56:26.759 INFO:teuthology.orchestra.run.vm05.stdout:Get:80 https://1.chacra.ceph.com/r/ceph/tentacle/70f8415b300f041766fa27faf7d5472699e32388/ubuntu/jammy/flavors/default jammy/main amd64 ceph-mgr-k8sevents all 20.2.0-712-g70f8415b-1jammy [14.2 kB] 2026-03-20T17:56:26.760 INFO:teuthology.orchestra.run.vm05.stdout:Get:81 https://1.chacra.ceph.com/r/ceph/tentacle/70f8415b300f041766fa27faf7d5472699e32388/ubuntu/jammy/flavors/default jammy/main amd64 ceph-test amd64 20.2.0-712-g70f8415b-1jammy [99.5 MB] 2026-03-20T17:56:28.221 INFO:teuthology.orchestra.run.vm02.stdout:Get:79 https://1.chacra.ceph.com/r/ceph/tentacle/70f8415b300f041766fa27faf7d5472699e32388/ubuntu/jammy/flavors/default jammy/main amd64 ceph-mgr-diskprediction-local all 20.2.0-712-g70f8415b-1jammy [8625 kB] 2026-03-20T17:56:28.660 INFO:teuthology.orchestra.run.vm06.stdout:Get:82 https://1.chacra.ceph.com/r/ceph/tentacle/70f8415b300f041766fa27faf7d5472699e32388/ubuntu/jammy/flavors/default jammy/main amd64 ceph-volume all 20.2.0-712-g70f8415b-1jammy [135 kB] 2026-03-20T17:56:28.661 INFO:teuthology.orchestra.run.vm06.stdout:Get:83 https://1.chacra.ceph.com/r/ceph/tentacle/70f8415b300f041766fa27faf7d5472699e32388/ubuntu/jammy/flavors/default jammy/main amd64 libcephfs-daemon amd64 20.2.0-712-g70f8415b-1jammy [43.3 kB] 2026-03-20T17:56:28.661 INFO:teuthology.orchestra.run.vm06.stdout:Get:84 https://1.chacra.ceph.com/r/ceph/tentacle/70f8415b300f041766fa27faf7d5472699e32388/ubuntu/jammy/flavors/default jammy/main amd64 libcephfs-proxy2 amd64 20.2.0-712-g70f8415b-1jammy [30.7 kB] 2026-03-20T17:56:28.661 INFO:teuthology.orchestra.run.vm06.stdout:Get:85 https://1.chacra.ceph.com/r/ceph/tentacle/70f8415b300f041766fa27faf7d5472699e32388/ubuntu/jammy/flavors/default jammy/main amd64 libcephfs-dev amd64 20.2.0-712-g70f8415b-1jammy [41.5 kB] 2026-03-20T17:56:28.662 INFO:teuthology.orchestra.run.vm06.stdout:Get:86 https://1.chacra.ceph.com/r/ceph/tentacle/70f8415b300f041766fa27faf7d5472699e32388/ubuntu/jammy/flavors/default jammy/main amd64 radosgw amd64 20.2.0-712-g70f8415b-1jammy [25.1 MB] 2026-03-20T17:56:29.051 INFO:teuthology.orchestra.run.vm02.stdout:Get:80 https://1.chacra.ceph.com/r/ceph/tentacle/70f8415b300f041766fa27faf7d5472699e32388/ubuntu/jammy/flavors/default jammy/main amd64 ceph-mgr-k8sevents all 20.2.0-712-g70f8415b-1jammy [14.2 kB] 2026-03-20T17:56:29.051 INFO:teuthology.orchestra.run.vm02.stdout:Get:81 https://1.chacra.ceph.com/r/ceph/tentacle/70f8415b300f041766fa27faf7d5472699e32388/ubuntu/jammy/flavors/default jammy/main amd64 ceph-test amd64 20.2.0-712-g70f8415b-1jammy [99.5 MB] 2026-03-20T17:56:29.922 INFO:teuthology.orchestra.run.vm06.stdout:Get:87 https://1.chacra.ceph.com/r/ceph/tentacle/70f8415b300f041766fa27faf7d5472699e32388/ubuntu/jammy/flavors/default jammy/main amd64 rbd-fuse amd64 20.2.0-712-g70f8415b-1jammy [97.9 kB] 2026-03-20T17:56:30.173 INFO:teuthology.orchestra.run.vm06.stdout:Fetched 281 MB in 15s (18.2 MB/s) 2026-03-20T17:56:30.314 INFO:teuthology.orchestra.run.vm06.stdout:Selecting previously unselected package liblttng-ust1:amd64. 2026-03-20T17:56:30.342 INFO:teuthology.orchestra.run.vm06.stdout:(Reading database ... (Reading database ... 5% (Reading database ... 10% (Reading database ... 15% (Reading database ... 20% (Reading database ... 25% (Reading database ... 30% (Reading database ... 35% (Reading database ... 40% (Reading database ... 45% (Reading database ... 50% (Reading database ... 55% (Reading database ... 60% (Reading database ... 65% (Reading database ... 70% (Reading database ... 75% (Reading database ... 80% (Reading database ... 85% (Reading database ... 90% (Reading database ... 95% (Reading database ... 100% (Reading database ... 119262 files and directories currently installed.) 2026-03-20T17:56:30.344 INFO:teuthology.orchestra.run.vm06.stdout:Preparing to unpack .../00-liblttng-ust1_2.13.1-1ubuntu1_amd64.deb ... 2026-03-20T17:56:30.346 INFO:teuthology.orchestra.run.vm06.stdout:Unpacking liblttng-ust1:amd64 (2.13.1-1ubuntu1) ... 2026-03-20T17:56:30.365 INFO:teuthology.orchestra.run.vm06.stdout:Selecting previously unselected package libdouble-conversion3:amd64. 2026-03-20T17:56:30.370 INFO:teuthology.orchestra.run.vm06.stdout:Preparing to unpack .../01-libdouble-conversion3_3.1.7-4_amd64.deb ... 2026-03-20T17:56:30.371 INFO:teuthology.orchestra.run.vm06.stdout:Unpacking libdouble-conversion3:amd64 (3.1.7-4) ... 2026-03-20T17:56:30.386 INFO:teuthology.orchestra.run.vm06.stdout:Selecting previously unselected package libpcre2-16-0:amd64. 2026-03-20T17:56:30.391 INFO:teuthology.orchestra.run.vm06.stdout:Preparing to unpack .../02-libpcre2-16-0_10.39-3ubuntu0.1_amd64.deb ... 2026-03-20T17:56:30.392 INFO:teuthology.orchestra.run.vm06.stdout:Unpacking libpcre2-16-0:amd64 (10.39-3ubuntu0.1) ... 2026-03-20T17:56:30.411 INFO:teuthology.orchestra.run.vm06.stdout:Selecting previously unselected package libqt5core5a:amd64. 2026-03-20T17:56:30.417 INFO:teuthology.orchestra.run.vm06.stdout:Preparing to unpack .../03-libqt5core5a_5.15.3+dfsg-2ubuntu0.2_amd64.deb ... 2026-03-20T17:56:30.421 INFO:teuthology.orchestra.run.vm06.stdout:Unpacking libqt5core5a:amd64 (5.15.3+dfsg-2ubuntu0.2) ... 2026-03-20T17:56:30.460 INFO:teuthology.orchestra.run.vm06.stdout:Selecting previously unselected package libqt5dbus5:amd64. 2026-03-20T17:56:30.465 INFO:teuthology.orchestra.run.vm06.stdout:Preparing to unpack .../04-libqt5dbus5_5.15.3+dfsg-2ubuntu0.2_amd64.deb ... 2026-03-20T17:56:30.466 INFO:teuthology.orchestra.run.vm06.stdout:Unpacking libqt5dbus5:amd64 (5.15.3+dfsg-2ubuntu0.2) ... 2026-03-20T17:56:30.483 INFO:teuthology.orchestra.run.vm06.stdout:Selecting previously unselected package libqt5network5:amd64. 2026-03-20T17:56:30.489 INFO:teuthology.orchestra.run.vm06.stdout:Preparing to unpack .../05-libqt5network5_5.15.3+dfsg-2ubuntu0.2_amd64.deb ... 2026-03-20T17:56:30.489 INFO:teuthology.orchestra.run.vm06.stdout:Unpacking libqt5network5:amd64 (5.15.3+dfsg-2ubuntu0.2) ... 2026-03-20T17:56:30.513 INFO:teuthology.orchestra.run.vm06.stdout:Selecting previously unselected package libthrift-0.16.0:amd64. 2026-03-20T17:56:30.518 INFO:teuthology.orchestra.run.vm06.stdout:Preparing to unpack .../06-libthrift-0.16.0_0.16.0-2_amd64.deb ... 2026-03-20T17:56:30.519 INFO:teuthology.orchestra.run.vm06.stdout:Unpacking libthrift-0.16.0:amd64 (0.16.0-2) ... 2026-03-20T17:56:30.544 INFO:teuthology.orchestra.run.vm06.stdout:Preparing to unpack .../07-librbd1_20.2.0-712-g70f8415b-1jammy_amd64.deb ... 2026-03-20T17:56:30.546 INFO:teuthology.orchestra.run.vm06.stdout:Unpacking librbd1 (20.2.0-712-g70f8415b-1jammy) over (17.2.9-0ubuntu0.22.04.2) ... 2026-03-20T17:56:30.607 INFO:teuthology.orchestra.run.vm06.stdout:Preparing to unpack .../08-librados2_20.2.0-712-g70f8415b-1jammy_amd64.deb ... 2026-03-20T17:56:30.609 INFO:teuthology.orchestra.run.vm06.stdout:Unpacking librados2 (20.2.0-712-g70f8415b-1jammy) over (17.2.9-0ubuntu0.22.04.2) ... 2026-03-20T17:56:30.671 INFO:teuthology.orchestra.run.vm06.stdout:Selecting previously unselected package libnbd0. 2026-03-20T17:56:30.677 INFO:teuthology.orchestra.run.vm06.stdout:Preparing to unpack .../09-libnbd0_1.10.5-1_amd64.deb ... 2026-03-20T17:56:30.678 INFO:teuthology.orchestra.run.vm06.stdout:Unpacking libnbd0 (1.10.5-1) ... 2026-03-20T17:56:30.694 INFO:teuthology.orchestra.run.vm06.stdout:Selecting previously unselected package libcephfs2. 2026-03-20T17:56:30.701 INFO:teuthology.orchestra.run.vm06.stdout:Preparing to unpack .../10-libcephfs2_20.2.0-712-g70f8415b-1jammy_amd64.deb ... 2026-03-20T17:56:30.702 INFO:teuthology.orchestra.run.vm06.stdout:Unpacking libcephfs2 (20.2.0-712-g70f8415b-1jammy) ... 2026-03-20T17:56:30.725 INFO:teuthology.orchestra.run.vm06.stdout:Selecting previously unselected package python3-rados. 2026-03-20T17:56:30.732 INFO:teuthology.orchestra.run.vm06.stdout:Preparing to unpack .../11-python3-rados_20.2.0-712-g70f8415b-1jammy_amd64.deb ... 2026-03-20T17:56:30.732 INFO:teuthology.orchestra.run.vm06.stdout:Unpacking python3-rados (20.2.0-712-g70f8415b-1jammy) ... 2026-03-20T17:56:30.751 INFO:teuthology.orchestra.run.vm06.stdout:Selecting previously unselected package python3-ceph-argparse. 2026-03-20T17:56:30.757 INFO:teuthology.orchestra.run.vm06.stdout:Preparing to unpack .../12-python3-ceph-argparse_20.2.0-712-g70f8415b-1jammy_all.deb ... 2026-03-20T17:56:30.757 INFO:teuthology.orchestra.run.vm06.stdout:Unpacking python3-ceph-argparse (20.2.0-712-g70f8415b-1jammy) ... 2026-03-20T17:56:30.771 INFO:teuthology.orchestra.run.vm06.stdout:Selecting previously unselected package python3-cephfs. 2026-03-20T17:56:30.777 INFO:teuthology.orchestra.run.vm06.stdout:Preparing to unpack .../13-python3-cephfs_20.2.0-712-g70f8415b-1jammy_amd64.deb ... 2026-03-20T17:56:30.778 INFO:teuthology.orchestra.run.vm06.stdout:Unpacking python3-cephfs (20.2.0-712-g70f8415b-1jammy) ... 2026-03-20T17:56:30.794 INFO:teuthology.orchestra.run.vm06.stdout:Selecting previously unselected package python3-ceph-common. 2026-03-20T17:56:30.800 INFO:teuthology.orchestra.run.vm06.stdout:Preparing to unpack .../14-python3-ceph-common_20.2.0-712-g70f8415b-1jammy_all.deb ... 2026-03-20T17:56:30.801 INFO:teuthology.orchestra.run.vm06.stdout:Unpacking python3-ceph-common (20.2.0-712-g70f8415b-1jammy) ... 2026-03-20T17:56:30.822 INFO:teuthology.orchestra.run.vm06.stdout:Selecting previously unselected package python3-wcwidth. 2026-03-20T17:56:30.829 INFO:teuthology.orchestra.run.vm06.stdout:Preparing to unpack .../15-python3-wcwidth_0.2.5+dfsg1-1_all.deb ... 2026-03-20T17:56:30.829 INFO:teuthology.orchestra.run.vm06.stdout:Unpacking python3-wcwidth (0.2.5+dfsg1-1) ... 2026-03-20T17:56:30.847 INFO:teuthology.orchestra.run.vm06.stdout:Selecting previously unselected package python3-prettytable. 2026-03-20T17:56:30.854 INFO:teuthology.orchestra.run.vm06.stdout:Preparing to unpack .../16-python3-prettytable_2.5.0-2_all.deb ... 2026-03-20T17:56:30.854 INFO:teuthology.orchestra.run.vm06.stdout:Unpacking python3-prettytable (2.5.0-2) ... 2026-03-20T17:56:30.870 INFO:teuthology.orchestra.run.vm06.stdout:Selecting previously unselected package python3-rbd. 2026-03-20T17:56:30.875 INFO:teuthology.orchestra.run.vm06.stdout:Preparing to unpack .../17-python3-rbd_20.2.0-712-g70f8415b-1jammy_amd64.deb ... 2026-03-20T17:56:30.876 INFO:teuthology.orchestra.run.vm06.stdout:Unpacking python3-rbd (20.2.0-712-g70f8415b-1jammy) ... 2026-03-20T17:56:30.896 INFO:teuthology.orchestra.run.vm06.stdout:Selecting previously unselected package librdkafka1:amd64. 2026-03-20T17:56:30.902 INFO:teuthology.orchestra.run.vm06.stdout:Preparing to unpack .../18-librdkafka1_1.8.0-1build1_amd64.deb ... 2026-03-20T17:56:30.902 INFO:teuthology.orchestra.run.vm06.stdout:Unpacking librdkafka1:amd64 (1.8.0-1build1) ... 2026-03-20T17:56:30.926 INFO:teuthology.orchestra.run.vm06.stdout:Selecting previously unselected package librgw2. 2026-03-20T17:56:30.933 INFO:teuthology.orchestra.run.vm06.stdout:Preparing to unpack .../19-librgw2_20.2.0-712-g70f8415b-1jammy_amd64.deb ... 2026-03-20T17:56:30.934 INFO:teuthology.orchestra.run.vm06.stdout:Unpacking librgw2 (20.2.0-712-g70f8415b-1jammy) ... 2026-03-20T17:56:30.946 INFO:teuthology.orchestra.run.vm08.stdout:Get:82 https://1.chacra.ceph.com/r/ceph/tentacle/70f8415b300f041766fa27faf7d5472699e32388/ubuntu/jammy/flavors/default jammy/main amd64 ceph-volume all 20.2.0-712-g70f8415b-1jammy [135 kB] 2026-03-20T17:56:30.947 INFO:teuthology.orchestra.run.vm08.stdout:Get:83 https://1.chacra.ceph.com/r/ceph/tentacle/70f8415b300f041766fa27faf7d5472699e32388/ubuntu/jammy/flavors/default jammy/main amd64 libcephfs-daemon amd64 20.2.0-712-g70f8415b-1jammy [43.3 kB] 2026-03-20T17:56:30.947 INFO:teuthology.orchestra.run.vm08.stdout:Get:84 https://1.chacra.ceph.com/r/ceph/tentacle/70f8415b300f041766fa27faf7d5472699e32388/ubuntu/jammy/flavors/default jammy/main amd64 libcephfs-proxy2 amd64 20.2.0-712-g70f8415b-1jammy [30.7 kB] 2026-03-20T17:56:30.947 INFO:teuthology.orchestra.run.vm08.stdout:Get:85 https://1.chacra.ceph.com/r/ceph/tentacle/70f8415b300f041766fa27faf7d5472699e32388/ubuntu/jammy/flavors/default jammy/main amd64 libcephfs-dev amd64 20.2.0-712-g70f8415b-1jammy [41.5 kB] 2026-03-20T17:56:30.947 INFO:teuthology.orchestra.run.vm08.stdout:Get:86 https://1.chacra.ceph.com/r/ceph/tentacle/70f8415b300f041766fa27faf7d5472699e32388/ubuntu/jammy/flavors/default jammy/main amd64 radosgw amd64 20.2.0-712-g70f8415b-1jammy [25.1 MB] 2026-03-20T17:56:31.066 INFO:teuthology.orchestra.run.vm06.stdout:Selecting previously unselected package python3-rgw. 2026-03-20T17:56:31.073 INFO:teuthology.orchestra.run.vm06.stdout:Preparing to unpack .../20-python3-rgw_20.2.0-712-g70f8415b-1jammy_amd64.deb ... 2026-03-20T17:56:31.074 INFO:teuthology.orchestra.run.vm06.stdout:Unpacking python3-rgw (20.2.0-712-g70f8415b-1jammy) ... 2026-03-20T17:56:31.089 INFO:teuthology.orchestra.run.vm06.stdout:Selecting previously unselected package liboath0:amd64. 2026-03-20T17:56:31.095 INFO:teuthology.orchestra.run.vm06.stdout:Preparing to unpack .../21-liboath0_2.6.7-3ubuntu0.1_amd64.deb ... 2026-03-20T17:56:31.096 INFO:teuthology.orchestra.run.vm06.stdout:Unpacking liboath0:amd64 (2.6.7-3ubuntu0.1) ... 2026-03-20T17:56:31.112 INFO:teuthology.orchestra.run.vm06.stdout:Selecting previously unselected package libradosstriper1. 2026-03-20T17:56:31.119 INFO:teuthology.orchestra.run.vm06.stdout:Preparing to unpack .../22-libradosstriper1_20.2.0-712-g70f8415b-1jammy_amd64.deb ... 2026-03-20T17:56:31.119 INFO:teuthology.orchestra.run.vm06.stdout:Unpacking libradosstriper1 (20.2.0-712-g70f8415b-1jammy) ... 2026-03-20T17:56:31.140 INFO:teuthology.orchestra.run.vm06.stdout:Selecting previously unselected package ceph-common. 2026-03-20T17:56:31.144 INFO:teuthology.orchestra.run.vm06.stdout:Preparing to unpack .../23-ceph-common_20.2.0-712-g70f8415b-1jammy_amd64.deb ... 2026-03-20T17:56:31.146 INFO:teuthology.orchestra.run.vm06.stdout:Unpacking ceph-common (20.2.0-712-g70f8415b-1jammy) ... 2026-03-20T17:56:31.530 INFO:teuthology.orchestra.run.vm06.stdout:Selecting previously unselected package ceph-base. 2026-03-20T17:56:31.537 INFO:teuthology.orchestra.run.vm06.stdout:Preparing to unpack .../24-ceph-base_20.2.0-712-g70f8415b-1jammy_amd64.deb ... 2026-03-20T17:56:31.542 INFO:teuthology.orchestra.run.vm06.stdout:Unpacking ceph-base (20.2.0-712-g70f8415b-1jammy) ... 2026-03-20T17:56:31.646 INFO:teuthology.orchestra.run.vm06.stdout:Selecting previously unselected package python3-jaraco.functools. 2026-03-20T17:56:31.652 INFO:teuthology.orchestra.run.vm06.stdout:Preparing to unpack .../25-python3-jaraco.functools_3.4.0-2_all.deb ... 2026-03-20T17:56:31.653 INFO:teuthology.orchestra.run.vm06.stdout:Unpacking python3-jaraco.functools (3.4.0-2) ... 2026-03-20T17:56:31.668 INFO:teuthology.orchestra.run.vm06.stdout:Selecting previously unselected package python3-cheroot. 2026-03-20T17:56:31.674 INFO:teuthology.orchestra.run.vm06.stdout:Preparing to unpack .../26-python3-cheroot_8.5.2+ds1-1ubuntu3.1_all.deb ... 2026-03-20T17:56:31.674 INFO:teuthology.orchestra.run.vm06.stdout:Unpacking python3-cheroot (8.5.2+ds1-1ubuntu3.1) ... 2026-03-20T17:56:31.692 INFO:teuthology.orchestra.run.vm06.stdout:Selecting previously unselected package python3-jaraco.classes. 2026-03-20T17:56:31.698 INFO:teuthology.orchestra.run.vm06.stdout:Preparing to unpack .../27-python3-jaraco.classes_3.2.1-3_all.deb ... 2026-03-20T17:56:31.698 INFO:teuthology.orchestra.run.vm06.stdout:Unpacking python3-jaraco.classes (3.2.1-3) ... 2026-03-20T17:56:31.714 INFO:teuthology.orchestra.run.vm06.stdout:Selecting previously unselected package python3-jaraco.text. 2026-03-20T17:56:31.720 INFO:teuthology.orchestra.run.vm06.stdout:Preparing to unpack .../28-python3-jaraco.text_3.6.0-2_all.deb ... 2026-03-20T17:56:31.720 INFO:teuthology.orchestra.run.vm06.stdout:Unpacking python3-jaraco.text (3.6.0-2) ... 2026-03-20T17:56:31.736 INFO:teuthology.orchestra.run.vm06.stdout:Selecting previously unselected package python3-jaraco.collections. 2026-03-20T17:56:31.742 INFO:teuthology.orchestra.run.vm06.stdout:Preparing to unpack .../29-python3-jaraco.collections_3.4.0-2_all.deb ... 2026-03-20T17:56:31.743 INFO:teuthology.orchestra.run.vm06.stdout:Unpacking python3-jaraco.collections (3.4.0-2) ... 2026-03-20T17:56:31.758 INFO:teuthology.orchestra.run.vm06.stdout:Selecting previously unselected package python3-tempora. 2026-03-20T17:56:31.764 INFO:teuthology.orchestra.run.vm06.stdout:Preparing to unpack .../30-python3-tempora_4.1.2-1_all.deb ... 2026-03-20T17:56:31.765 INFO:teuthology.orchestra.run.vm06.stdout:Unpacking python3-tempora (4.1.2-1) ... 2026-03-20T17:56:31.782 INFO:teuthology.orchestra.run.vm06.stdout:Selecting previously unselected package python3-portend. 2026-03-20T17:56:31.788 INFO:teuthology.orchestra.run.vm06.stdout:Preparing to unpack .../31-python3-portend_3.0.0-1_all.deb ... 2026-03-20T17:56:31.789 INFO:teuthology.orchestra.run.vm06.stdout:Unpacking python3-portend (3.0.0-1) ... 2026-03-20T17:56:31.804 INFO:teuthology.orchestra.run.vm06.stdout:Selecting previously unselected package python3-zc.lockfile. 2026-03-20T17:56:31.810 INFO:teuthology.orchestra.run.vm06.stdout:Preparing to unpack .../32-python3-zc.lockfile_2.0-1_all.deb ... 2026-03-20T17:56:31.810 INFO:teuthology.orchestra.run.vm06.stdout:Unpacking python3-zc.lockfile (2.0-1) ... 2026-03-20T17:56:31.826 INFO:teuthology.orchestra.run.vm06.stdout:Selecting previously unselected package python3-cherrypy3. 2026-03-20T17:56:31.833 INFO:teuthology.orchestra.run.vm06.stdout:Preparing to unpack .../33-python3-cherrypy3_18.6.1-4_all.deb ... 2026-03-20T17:56:31.834 INFO:teuthology.orchestra.run.vm06.stdout:Unpacking python3-cherrypy3 (18.6.1-4) ... 2026-03-20T17:56:31.863 INFO:teuthology.orchestra.run.vm06.stdout:Selecting previously unselected package python3-natsort. 2026-03-20T17:56:31.869 INFO:teuthology.orchestra.run.vm06.stdout:Preparing to unpack .../34-python3-natsort_8.0.2-1_all.deb ... 2026-03-20T17:56:31.869 INFO:teuthology.orchestra.run.vm06.stdout:Unpacking python3-natsort (8.0.2-1) ... 2026-03-20T17:56:31.885 INFO:teuthology.orchestra.run.vm06.stdout:Selecting previously unselected package ceph-mgr-modules-core. 2026-03-20T17:56:31.890 INFO:teuthology.orchestra.run.vm06.stdout:Preparing to unpack .../35-ceph-mgr-modules-core_20.2.0-712-g70f8415b-1jammy_all.deb ... 2026-03-20T17:56:31.892 INFO:teuthology.orchestra.run.vm06.stdout:Unpacking ceph-mgr-modules-core (20.2.0-712-g70f8415b-1jammy) ... 2026-03-20T17:56:31.923 INFO:teuthology.orchestra.run.vm06.stdout:Selecting previously unselected package libsqlite3-mod-ceph. 2026-03-20T17:56:31.929 INFO:teuthology.orchestra.run.vm06.stdout:Preparing to unpack .../36-libsqlite3-mod-ceph_20.2.0-712-g70f8415b-1jammy_amd64.deb ... 2026-03-20T17:56:31.930 INFO:teuthology.orchestra.run.vm06.stdout:Unpacking libsqlite3-mod-ceph (20.2.0-712-g70f8415b-1jammy) ... 2026-03-20T17:56:31.944 INFO:teuthology.orchestra.run.vm06.stdout:Selecting previously unselected package ceph-mgr. 2026-03-20T17:56:31.950 INFO:teuthology.orchestra.run.vm06.stdout:Preparing to unpack .../37-ceph-mgr_20.2.0-712-g70f8415b-1jammy_amd64.deb ... 2026-03-20T17:56:31.951 INFO:teuthology.orchestra.run.vm06.stdout:Unpacking ceph-mgr (20.2.0-712-g70f8415b-1jammy) ... 2026-03-20T17:56:31.976 INFO:teuthology.orchestra.run.vm06.stdout:Selecting previously unselected package ceph-mon. 2026-03-20T17:56:31.982 INFO:teuthology.orchestra.run.vm06.stdout:Preparing to unpack .../38-ceph-mon_20.2.0-712-g70f8415b-1jammy_amd64.deb ... 2026-03-20T17:56:31.983 INFO:teuthology.orchestra.run.vm06.stdout:Unpacking ceph-mon (20.2.0-712-g70f8415b-1jammy) ... 2026-03-20T17:56:32.124 INFO:teuthology.orchestra.run.vm06.stdout:Selecting previously unselected package libfuse2:amd64. 2026-03-20T17:56:32.130 INFO:teuthology.orchestra.run.vm06.stdout:Preparing to unpack .../39-libfuse2_2.9.9-5ubuntu3_amd64.deb ... 2026-03-20T17:56:32.131 INFO:teuthology.orchestra.run.vm06.stdout:Unpacking libfuse2:amd64 (2.9.9-5ubuntu3) ... 2026-03-20T17:56:32.149 INFO:teuthology.orchestra.run.vm06.stdout:Selecting previously unselected package ceph-osd. 2026-03-20T17:56:32.155 INFO:teuthology.orchestra.run.vm06.stdout:Preparing to unpack .../40-ceph-osd_20.2.0-712-g70f8415b-1jammy_amd64.deb ... 2026-03-20T17:56:32.156 INFO:teuthology.orchestra.run.vm06.stdout:Unpacking ceph-osd (20.2.0-712-g70f8415b-1jammy) ... 2026-03-20T17:56:32.159 INFO:teuthology.orchestra.run.vm05.stdout:Get:82 https://1.chacra.ceph.com/r/ceph/tentacle/70f8415b300f041766fa27faf7d5472699e32388/ubuntu/jammy/flavors/default jammy/main amd64 ceph-volume all 20.2.0-712-g70f8415b-1jammy [135 kB] 2026-03-20T17:56:32.160 INFO:teuthology.orchestra.run.vm05.stdout:Get:83 https://1.chacra.ceph.com/r/ceph/tentacle/70f8415b300f041766fa27faf7d5472699e32388/ubuntu/jammy/flavors/default jammy/main amd64 libcephfs-daemon amd64 20.2.0-712-g70f8415b-1jammy [43.3 kB] 2026-03-20T17:56:32.160 INFO:teuthology.orchestra.run.vm05.stdout:Get:84 https://1.chacra.ceph.com/r/ceph/tentacle/70f8415b300f041766fa27faf7d5472699e32388/ubuntu/jammy/flavors/default jammy/main amd64 libcephfs-proxy2 amd64 20.2.0-712-g70f8415b-1jammy [30.7 kB] 2026-03-20T17:56:32.160 INFO:teuthology.orchestra.run.vm05.stdout:Get:85 https://1.chacra.ceph.com/r/ceph/tentacle/70f8415b300f041766fa27faf7d5472699e32388/ubuntu/jammy/flavors/default jammy/main amd64 libcephfs-dev amd64 20.2.0-712-g70f8415b-1jammy [41.5 kB] 2026-03-20T17:56:32.160 INFO:teuthology.orchestra.run.vm05.stdout:Get:86 https://1.chacra.ceph.com/r/ceph/tentacle/70f8415b300f041766fa27faf7d5472699e32388/ubuntu/jammy/flavors/default jammy/main amd64 radosgw amd64 20.2.0-712-g70f8415b-1jammy [25.1 MB] 2026-03-20T17:56:32.167 INFO:teuthology.orchestra.run.vm08.stdout:Get:87 https://1.chacra.ceph.com/r/ceph/tentacle/70f8415b300f041766fa27faf7d5472699e32388/ubuntu/jammy/flavors/default jammy/main amd64 rbd-fuse amd64 20.2.0-712-g70f8415b-1jammy [97.9 kB] 2026-03-20T17:56:32.399 INFO:teuthology.orchestra.run.vm06.stdout:Selecting previously unselected package ceph. 2026-03-20T17:56:32.405 INFO:teuthology.orchestra.run.vm06.stdout:Preparing to unpack .../41-ceph_20.2.0-712-g70f8415b-1jammy_amd64.deb ... 2026-03-20T17:56:32.406 INFO:teuthology.orchestra.run.vm06.stdout:Unpacking ceph (20.2.0-712-g70f8415b-1jammy) ... 2026-03-20T17:56:32.421 INFO:teuthology.orchestra.run.vm06.stdout:Selecting previously unselected package ceph-fuse. 2026-03-20T17:56:32.427 INFO:teuthology.orchestra.run.vm06.stdout:Preparing to unpack .../42-ceph-fuse_20.2.0-712-g70f8415b-1jammy_amd64.deb ... 2026-03-20T17:56:32.427 INFO:teuthology.orchestra.run.vm06.stdout:Unpacking ceph-fuse (20.2.0-712-g70f8415b-1jammy) ... 2026-03-20T17:56:32.434 INFO:teuthology.orchestra.run.vm08.stdout:Fetched 281 MB in 18s (16.0 MB/s) 2026-03-20T17:56:32.580 INFO:teuthology.orchestra.run.vm08.stdout:Selecting previously unselected package liblttng-ust1:amd64. 2026-03-20T17:56:32.585 INFO:teuthology.orchestra.run.vm06.stdout:Selecting previously unselected package ceph-mds. 2026-03-20T17:56:32.591 INFO:teuthology.orchestra.run.vm06.stdout:Preparing to unpack .../43-ceph-mds_20.2.0-712-g70f8415b-1jammy_amd64.deb ... 2026-03-20T17:56:32.592 INFO:teuthology.orchestra.run.vm06.stdout:Unpacking ceph-mds (20.2.0-712-g70f8415b-1jammy) ... 2026-03-20T17:56:32.609 INFO:teuthology.orchestra.run.vm08.stdout:(Reading database ... (Reading database ... 5% (Reading database ... 10% (Reading database ... 15% (Reading database ... 20% (Reading database ... 25% (Reading database ... 30% (Reading database ... 35% (Reading database ... 40% (Reading database ... 45% (Reading database ... 50% (Reading database ... 55% (Reading database ... 60% (Reading database ... 65% (Reading database ... 70% (Reading database ... 75% (Reading database ... 80% (Reading database ... 85% (Reading database ... 90% (Reading database ... 95% (Reading database ... 100% (Reading database ... 119262 files and directories currently installed.) 2026-03-20T17:56:32.611 INFO:teuthology.orchestra.run.vm08.stdout:Preparing to unpack .../00-liblttng-ust1_2.13.1-1ubuntu1_amd64.deb ... 2026-03-20T17:56:32.613 INFO:teuthology.orchestra.run.vm08.stdout:Unpacking liblttng-ust1:amd64 (2.13.1-1ubuntu1) ... 2026-03-20T17:56:32.636 INFO:teuthology.orchestra.run.vm06.stdout:Selecting previously unselected package cephadm. 2026-03-20T17:56:32.637 INFO:teuthology.orchestra.run.vm08.stdout:Selecting previously unselected package libdouble-conversion3:amd64. 2026-03-20T17:56:32.642 INFO:teuthology.orchestra.run.vm06.stdout:Preparing to unpack .../44-cephadm_20.2.0-712-g70f8415b-1jammy_amd64.deb ... 2026-03-20T17:56:32.643 INFO:teuthology.orchestra.run.vm06.stdout:Unpacking cephadm (20.2.0-712-g70f8415b-1jammy) ... 2026-03-20T17:56:32.643 INFO:teuthology.orchestra.run.vm08.stdout:Preparing to unpack .../01-libdouble-conversion3_3.1.7-4_amd64.deb ... 2026-03-20T17:56:32.644 INFO:teuthology.orchestra.run.vm08.stdout:Unpacking libdouble-conversion3:amd64 (3.1.7-4) ... 2026-03-20T17:56:32.660 INFO:teuthology.orchestra.run.vm08.stdout:Selecting previously unselected package libpcre2-16-0:amd64. 2026-03-20T17:56:32.661 INFO:teuthology.orchestra.run.vm06.stdout:Selecting previously unselected package python3-asyncssh. 2026-03-20T17:56:32.666 INFO:teuthology.orchestra.run.vm08.stdout:Preparing to unpack .../02-libpcre2-16-0_10.39-3ubuntu0.1_amd64.deb ... 2026-03-20T17:56:32.667 INFO:teuthology.orchestra.run.vm08.stdout:Unpacking libpcre2-16-0:amd64 (10.39-3ubuntu0.1) ... 2026-03-20T17:56:32.668 INFO:teuthology.orchestra.run.vm06.stdout:Preparing to unpack .../45-python3-asyncssh_2.5.0-1ubuntu0.1_all.deb ... 2026-03-20T17:56:32.668 INFO:teuthology.orchestra.run.vm06.stdout:Unpacking python3-asyncssh (2.5.0-1ubuntu0.1) ... 2026-03-20T17:56:32.686 INFO:teuthology.orchestra.run.vm08.stdout:Selecting previously unselected package libqt5core5a:amd64. 2026-03-20T17:56:32.692 INFO:teuthology.orchestra.run.vm08.stdout:Preparing to unpack .../03-libqt5core5a_5.15.3+dfsg-2ubuntu0.2_amd64.deb ... 2026-03-20T17:56:32.695 INFO:teuthology.orchestra.run.vm06.stdout:Selecting previously unselected package ceph-mgr-cephadm. 2026-03-20T17:56:32.696 INFO:teuthology.orchestra.run.vm08.stdout:Unpacking libqt5core5a:amd64 (5.15.3+dfsg-2ubuntu0.2) ... 2026-03-20T17:56:32.701 INFO:teuthology.orchestra.run.vm06.stdout:Preparing to unpack .../46-ceph-mgr-cephadm_20.2.0-712-g70f8415b-1jammy_all.deb ... 2026-03-20T17:56:32.702 INFO:teuthology.orchestra.run.vm06.stdout:Unpacking ceph-mgr-cephadm (20.2.0-712-g70f8415b-1jammy) ... 2026-03-20T17:56:32.728 INFO:teuthology.orchestra.run.vm06.stdout:Selecting previously unselected package python3-repoze.lru. 2026-03-20T17:56:32.734 INFO:teuthology.orchestra.run.vm06.stdout:Preparing to unpack .../47-python3-repoze.lru_0.7-2_all.deb ... 2026-03-20T17:56:32.735 INFO:teuthology.orchestra.run.vm06.stdout:Unpacking python3-repoze.lru (0.7-2) ... 2026-03-20T17:56:32.739 INFO:teuthology.orchestra.run.vm08.stdout:Selecting previously unselected package libqt5dbus5:amd64. 2026-03-20T17:56:32.745 INFO:teuthology.orchestra.run.vm08.stdout:Preparing to unpack .../04-libqt5dbus5_5.15.3+dfsg-2ubuntu0.2_amd64.deb ... 2026-03-20T17:56:32.746 INFO:teuthology.orchestra.run.vm08.stdout:Unpacking libqt5dbus5:amd64 (5.15.3+dfsg-2ubuntu0.2) ... 2026-03-20T17:56:32.752 INFO:teuthology.orchestra.run.vm06.stdout:Selecting previously unselected package python3-routes. 2026-03-20T17:56:32.758 INFO:teuthology.orchestra.run.vm06.stdout:Preparing to unpack .../48-python3-routes_2.5.1-1ubuntu1_all.deb ... 2026-03-20T17:56:32.759 INFO:teuthology.orchestra.run.vm06.stdout:Unpacking python3-routes (2.5.1-1ubuntu1) ... 2026-03-20T17:56:32.763 INFO:teuthology.orchestra.run.vm08.stdout:Selecting previously unselected package libqt5network5:amd64. 2026-03-20T17:56:32.769 INFO:teuthology.orchestra.run.vm08.stdout:Preparing to unpack .../05-libqt5network5_5.15.3+dfsg-2ubuntu0.2_amd64.deb ... 2026-03-20T17:56:32.770 INFO:teuthology.orchestra.run.vm08.stdout:Unpacking libqt5network5:amd64 (5.15.3+dfsg-2ubuntu0.2) ... 2026-03-20T17:56:32.784 INFO:teuthology.orchestra.run.vm06.stdout:Selecting previously unselected package ceph-mgr-dashboard. 2026-03-20T17:56:32.790 INFO:teuthology.orchestra.run.vm06.stdout:Preparing to unpack .../49-ceph-mgr-dashboard_20.2.0-712-g70f8415b-1jammy_all.deb ... 2026-03-20T17:56:32.791 INFO:teuthology.orchestra.run.vm06.stdout:Unpacking ceph-mgr-dashboard (20.2.0-712-g70f8415b-1jammy) ... 2026-03-20T17:56:32.793 INFO:teuthology.orchestra.run.vm08.stdout:Selecting previously unselected package libthrift-0.16.0:amd64. 2026-03-20T17:56:32.798 INFO:teuthology.orchestra.run.vm08.stdout:Preparing to unpack .../06-libthrift-0.16.0_0.16.0-2_amd64.deb ... 2026-03-20T17:56:32.799 INFO:teuthology.orchestra.run.vm08.stdout:Unpacking libthrift-0.16.0:amd64 (0.16.0-2) ... 2026-03-20T17:56:32.822 INFO:teuthology.orchestra.run.vm08.stdout:Preparing to unpack .../07-librbd1_20.2.0-712-g70f8415b-1jammy_amd64.deb ... 2026-03-20T17:56:32.824 INFO:teuthology.orchestra.run.vm08.stdout:Unpacking librbd1 (20.2.0-712-g70f8415b-1jammy) over (17.2.9-0ubuntu0.22.04.2) ... 2026-03-20T17:56:32.888 INFO:teuthology.orchestra.run.vm08.stdout:Preparing to unpack .../08-librados2_20.2.0-712-g70f8415b-1jammy_amd64.deb ... 2026-03-20T17:56:32.890 INFO:teuthology.orchestra.run.vm08.stdout:Unpacking librados2 (20.2.0-712-g70f8415b-1jammy) over (17.2.9-0ubuntu0.22.04.2) ... 2026-03-20T17:56:32.957 INFO:teuthology.orchestra.run.vm08.stdout:Selecting previously unselected package libnbd0. 2026-03-20T17:56:32.963 INFO:teuthology.orchestra.run.vm08.stdout:Preparing to unpack .../09-libnbd0_1.10.5-1_amd64.deb ... 2026-03-20T17:56:32.964 INFO:teuthology.orchestra.run.vm08.stdout:Unpacking libnbd0 (1.10.5-1) ... 2026-03-20T17:56:32.981 INFO:teuthology.orchestra.run.vm08.stdout:Selecting previously unselected package libcephfs2. 2026-03-20T17:56:32.988 INFO:teuthology.orchestra.run.vm08.stdout:Preparing to unpack .../10-libcephfs2_20.2.0-712-g70f8415b-1jammy_amd64.deb ... 2026-03-20T17:56:32.988 INFO:teuthology.orchestra.run.vm08.stdout:Unpacking libcephfs2 (20.2.0-712-g70f8415b-1jammy) ... 2026-03-20T17:56:33.014 INFO:teuthology.orchestra.run.vm08.stdout:Selecting previously unselected package python3-rados. 2026-03-20T17:56:33.020 INFO:teuthology.orchestra.run.vm08.stdout:Preparing to unpack .../11-python3-rados_20.2.0-712-g70f8415b-1jammy_amd64.deb ... 2026-03-20T17:56:33.021 INFO:teuthology.orchestra.run.vm08.stdout:Unpacking python3-rados (20.2.0-712-g70f8415b-1jammy) ... 2026-03-20T17:56:33.041 INFO:teuthology.orchestra.run.vm08.stdout:Selecting previously unselected package python3-ceph-argparse. 2026-03-20T17:56:33.047 INFO:teuthology.orchestra.run.vm08.stdout:Preparing to unpack .../12-python3-ceph-argparse_20.2.0-712-g70f8415b-1jammy_all.deb ... 2026-03-20T17:56:33.048 INFO:teuthology.orchestra.run.vm08.stdout:Unpacking python3-ceph-argparse (20.2.0-712-g70f8415b-1jammy) ... 2026-03-20T17:56:33.062 INFO:teuthology.orchestra.run.vm08.stdout:Selecting previously unselected package python3-cephfs. 2026-03-20T17:56:33.068 INFO:teuthology.orchestra.run.vm08.stdout:Preparing to unpack .../13-python3-cephfs_20.2.0-712-g70f8415b-1jammy_amd64.deb ... 2026-03-20T17:56:33.068 INFO:teuthology.orchestra.run.vm08.stdout:Unpacking python3-cephfs (20.2.0-712-g70f8415b-1jammy) ... 2026-03-20T17:56:33.087 INFO:teuthology.orchestra.run.vm08.stdout:Selecting previously unselected package python3-ceph-common. 2026-03-20T17:56:33.093 INFO:teuthology.orchestra.run.vm08.stdout:Preparing to unpack .../14-python3-ceph-common_20.2.0-712-g70f8415b-1jammy_all.deb ... 2026-03-20T17:56:33.094 INFO:teuthology.orchestra.run.vm08.stdout:Unpacking python3-ceph-common (20.2.0-712-g70f8415b-1jammy) ... 2026-03-20T17:56:33.116 INFO:teuthology.orchestra.run.vm08.stdout:Selecting previously unselected package python3-wcwidth. 2026-03-20T17:56:33.122 INFO:teuthology.orchestra.run.vm08.stdout:Preparing to unpack .../15-python3-wcwidth_0.2.5+dfsg1-1_all.deb ... 2026-03-20T17:56:33.123 INFO:teuthology.orchestra.run.vm08.stdout:Unpacking python3-wcwidth (0.2.5+dfsg1-1) ... 2026-03-20T17:56:33.140 INFO:teuthology.orchestra.run.vm08.stdout:Selecting previously unselected package python3-prettytable. 2026-03-20T17:56:33.146 INFO:teuthology.orchestra.run.vm08.stdout:Preparing to unpack .../16-python3-prettytable_2.5.0-2_all.deb ... 2026-03-20T17:56:33.146 INFO:teuthology.orchestra.run.vm08.stdout:Unpacking python3-prettytable (2.5.0-2) ... 2026-03-20T17:56:33.161 INFO:teuthology.orchestra.run.vm08.stdout:Selecting previously unselected package python3-rbd. 2026-03-20T17:56:33.168 INFO:teuthology.orchestra.run.vm08.stdout:Preparing to unpack .../17-python3-rbd_20.2.0-712-g70f8415b-1jammy_amd64.deb ... 2026-03-20T17:56:33.169 INFO:teuthology.orchestra.run.vm08.stdout:Unpacking python3-rbd (20.2.0-712-g70f8415b-1jammy) ... 2026-03-20T17:56:33.188 INFO:teuthology.orchestra.run.vm08.stdout:Selecting previously unselected package librdkafka1:amd64. 2026-03-20T17:56:33.194 INFO:teuthology.orchestra.run.vm08.stdout:Preparing to unpack .../18-librdkafka1_1.8.0-1build1_amd64.deb ... 2026-03-20T17:56:33.195 INFO:teuthology.orchestra.run.vm08.stdout:Unpacking librdkafka1:amd64 (1.8.0-1build1) ... 2026-03-20T17:56:33.305 INFO:teuthology.orchestra.run.vm05.stdout:Get:87 https://1.chacra.ceph.com/r/ceph/tentacle/70f8415b300f041766fa27faf7d5472699e32388/ubuntu/jammy/flavors/default jammy/main amd64 rbd-fuse amd64 20.2.0-712-g70f8415b-1jammy [97.9 kB] 2026-03-20T17:56:33.359 INFO:teuthology.orchestra.run.vm08.stdout:Selecting previously unselected package librgw2. 2026-03-20T17:56:33.365 INFO:teuthology.orchestra.run.vm08.stdout:Preparing to unpack .../19-librgw2_20.2.0-712-g70f8415b-1jammy_amd64.deb ... 2026-03-20T17:56:33.366 INFO:teuthology.orchestra.run.vm08.stdout:Unpacking librgw2 (20.2.0-712-g70f8415b-1jammy) ... 2026-03-20T17:56:33.421 INFO:teuthology.orchestra.run.vm06.stdout:Selecting previously unselected package python3-sklearn-lib:amd64. 2026-03-20T17:56:33.428 INFO:teuthology.orchestra.run.vm06.stdout:Preparing to unpack .../50-python3-sklearn-lib_0.23.2-5ubuntu6_amd64.deb ... 2026-03-20T17:56:33.429 INFO:teuthology.orchestra.run.vm06.stdout:Unpacking python3-sklearn-lib:amd64 (0.23.2-5ubuntu6) ... 2026-03-20T17:56:33.518 INFO:teuthology.orchestra.run.vm08.stdout:Selecting previously unselected package python3-rgw. 2026-03-20T17:56:33.520 INFO:teuthology.orchestra.run.vm06.stdout:Selecting previously unselected package python3-joblib. 2026-03-20T17:56:33.524 INFO:teuthology.orchestra.run.vm08.stdout:Preparing to unpack .../20-python3-rgw_20.2.0-712-g70f8415b-1jammy_amd64.deb ... 2026-03-20T17:56:33.525 INFO:teuthology.orchestra.run.vm08.stdout:Unpacking python3-rgw (20.2.0-712-g70f8415b-1jammy) ... 2026-03-20T17:56:33.527 INFO:teuthology.orchestra.run.vm06.stdout:Preparing to unpack .../51-python3-joblib_0.17.0-4ubuntu1_all.deb ... 2026-03-20T17:56:33.528 INFO:teuthology.orchestra.run.vm06.stdout:Unpacking python3-joblib (0.17.0-4ubuntu1) ... 2026-03-20T17:56:33.542 INFO:teuthology.orchestra.run.vm08.stdout:Selecting previously unselected package liboath0:amd64. 2026-03-20T17:56:33.547 INFO:teuthology.orchestra.run.vm08.stdout:Preparing to unpack .../21-liboath0_2.6.7-3ubuntu0.1_amd64.deb ... 2026-03-20T17:56:33.548 INFO:teuthology.orchestra.run.vm08.stdout:Unpacking liboath0:amd64 (2.6.7-3ubuntu0.1) ... 2026-03-20T17:56:33.561 INFO:teuthology.orchestra.run.vm06.stdout:Selecting previously unselected package python3-threadpoolctl. 2026-03-20T17:56:33.563 INFO:teuthology.orchestra.run.vm08.stdout:Selecting previously unselected package libradosstriper1. 2026-03-20T17:56:33.566 INFO:teuthology.orchestra.run.vm05.stdout:Fetched 281 MB in 19s (15.1 MB/s) 2026-03-20T17:56:33.567 INFO:teuthology.orchestra.run.vm06.stdout:Preparing to unpack .../52-python3-threadpoolctl_3.1.0-1_all.deb ... 2026-03-20T17:56:33.568 INFO:teuthology.orchestra.run.vm06.stdout:Unpacking python3-threadpoolctl (3.1.0-1) ... 2026-03-20T17:56:33.569 INFO:teuthology.orchestra.run.vm08.stdout:Preparing to unpack .../22-libradosstriper1_20.2.0-712-g70f8415b-1jammy_amd64.deb ... 2026-03-20T17:56:33.570 INFO:teuthology.orchestra.run.vm08.stdout:Unpacking libradosstriper1 (20.2.0-712-g70f8415b-1jammy) ... 2026-03-20T17:56:33.711 INFO:teuthology.orchestra.run.vm05.stdout:Selecting previously unselected package liblttng-ust1:amd64. 2026-03-20T17:56:33.711 INFO:teuthology.orchestra.run.vm06.stdout:Selecting previously unselected package python3-sklearn. 2026-03-20T17:56:33.713 INFO:teuthology.orchestra.run.vm08.stdout:Selecting previously unselected package ceph-common. 2026-03-20T17:56:33.718 INFO:teuthology.orchestra.run.vm06.stdout:Preparing to unpack .../53-python3-sklearn_0.23.2-5ubuntu6_all.deb ... 2026-03-20T17:56:33.719 INFO:teuthology.orchestra.run.vm06.stdout:Unpacking python3-sklearn (0.23.2-5ubuntu6) ... 2026-03-20T17:56:33.719 INFO:teuthology.orchestra.run.vm08.stdout:Preparing to unpack .../23-ceph-common_20.2.0-712-g70f8415b-1jammy_amd64.deb ... 2026-03-20T17:56:33.720 INFO:teuthology.orchestra.run.vm08.stdout:Unpacking ceph-common (20.2.0-712-g70f8415b-1jammy) ... 2026-03-20T17:56:33.744 INFO:teuthology.orchestra.run.vm05.stdout:(Reading database ... (Reading database ... 5% (Reading database ... 10% (Reading database ... 15% (Reading database ... 20% (Reading database ... 25% (Reading database ... 30% (Reading database ... 35% (Reading database ... 40% (Reading database ... 45% (Reading database ... 50% (Reading database ... 55% (Reading database ... 60% (Reading database ... 65% (Reading database ... 70% (Reading database ... 75% (Reading database ... 80% (Reading database ... 85% (Reading database ... 90% (Reading database ... 95% (Reading database ... 100% (Reading database ... 119262 files and directories currently installed.) 2026-03-20T17:56:33.746 INFO:teuthology.orchestra.run.vm05.stdout:Preparing to unpack .../00-liblttng-ust1_2.13.1-1ubuntu1_amd64.deb ... 2026-03-20T17:56:33.748 INFO:teuthology.orchestra.run.vm05.stdout:Unpacking liblttng-ust1:amd64 (2.13.1-1ubuntu1) ... 2026-03-20T17:56:33.769 INFO:teuthology.orchestra.run.vm05.stdout:Selecting previously unselected package libdouble-conversion3:amd64. 2026-03-20T17:56:33.775 INFO:teuthology.orchestra.run.vm05.stdout:Preparing to unpack .../01-libdouble-conversion3_3.1.7-4_amd64.deb ... 2026-03-20T17:56:33.775 INFO:teuthology.orchestra.run.vm05.stdout:Unpacking libdouble-conversion3:amd64 (3.1.7-4) ... 2026-03-20T17:56:33.803 INFO:teuthology.orchestra.run.vm05.stdout:Selecting previously unselected package libpcre2-16-0:amd64. 2026-03-20T17:56:33.809 INFO:teuthology.orchestra.run.vm05.stdout:Preparing to unpack .../02-libpcre2-16-0_10.39-3ubuntu0.1_amd64.deb ... 2026-03-20T17:56:33.810 INFO:teuthology.orchestra.run.vm05.stdout:Unpacking libpcre2-16-0:amd64 (10.39-3ubuntu0.1) ... 2026-03-20T17:56:33.830 INFO:teuthology.orchestra.run.vm05.stdout:Selecting previously unselected package libqt5core5a:amd64. 2026-03-20T17:56:33.836 INFO:teuthology.orchestra.run.vm05.stdout:Preparing to unpack .../03-libqt5core5a_5.15.3+dfsg-2ubuntu0.2_amd64.deb ... 2026-03-20T17:56:33.837 INFO:teuthology.orchestra.run.vm06.stdout:Selecting previously unselected package ceph-mgr-diskprediction-local. 2026-03-20T17:56:33.841 INFO:teuthology.orchestra.run.vm05.stdout:Unpacking libqt5core5a:amd64 (5.15.3+dfsg-2ubuntu0.2) ... 2026-03-20T17:56:33.843 INFO:teuthology.orchestra.run.vm06.stdout:Preparing to unpack .../54-ceph-mgr-diskprediction-local_20.2.0-712-g70f8415b-1jammy_all.deb ... 2026-03-20T17:56:33.844 INFO:teuthology.orchestra.run.vm06.stdout:Unpacking ceph-mgr-diskprediction-local (20.2.0-712-g70f8415b-1jammy) ... 2026-03-20T17:56:33.933 INFO:teuthology.orchestra.run.vm05.stdout:Selecting previously unselected package libqt5dbus5:amd64. 2026-03-20T17:56:33.939 INFO:teuthology.orchestra.run.vm05.stdout:Preparing to unpack .../04-libqt5dbus5_5.15.3+dfsg-2ubuntu0.2_amd64.deb ... 2026-03-20T17:56:33.940 INFO:teuthology.orchestra.run.vm05.stdout:Unpacking libqt5dbus5:amd64 (5.15.3+dfsg-2ubuntu0.2) ... 2026-03-20T17:56:33.959 INFO:teuthology.orchestra.run.vm05.stdout:Selecting previously unselected package libqt5network5:amd64. 2026-03-20T17:56:33.966 INFO:teuthology.orchestra.run.vm05.stdout:Preparing to unpack .../05-libqt5network5_5.15.3+dfsg-2ubuntu0.2_amd64.deb ... 2026-03-20T17:56:33.967 INFO:teuthology.orchestra.run.vm05.stdout:Unpacking libqt5network5:amd64 (5.15.3+dfsg-2ubuntu0.2) ... 2026-03-20T17:56:33.995 INFO:teuthology.orchestra.run.vm05.stdout:Selecting previously unselected package libthrift-0.16.0:amd64. 2026-03-20T17:56:34.001 INFO:teuthology.orchestra.run.vm05.stdout:Preparing to unpack .../06-libthrift-0.16.0_0.16.0-2_amd64.deb ... 2026-03-20T17:56:34.093 INFO:teuthology.orchestra.run.vm05.stdout:Unpacking libthrift-0.16.0:amd64 (0.16.0-2) ... 2026-03-20T17:56:34.147 INFO:teuthology.orchestra.run.vm05.stdout:Preparing to unpack .../07-librbd1_20.2.0-712-g70f8415b-1jammy_amd64.deb ... 2026-03-20T17:56:34.148 INFO:teuthology.orchestra.run.vm08.stdout:Selecting previously unselected package ceph-base. 2026-03-20T17:56:34.149 INFO:teuthology.orchestra.run.vm05.stdout:Unpacking librbd1 (20.2.0-712-g70f8415b-1jammy) over (17.2.9-0ubuntu0.22.04.2) ... 2026-03-20T17:56:34.155 INFO:teuthology.orchestra.run.vm06.stdout:Selecting previously unselected package python3-cachetools. 2026-03-20T17:56:34.155 INFO:teuthology.orchestra.run.vm08.stdout:Preparing to unpack .../24-ceph-base_20.2.0-712-g70f8415b-1jammy_amd64.deb ... 2026-03-20T17:56:34.160 INFO:teuthology.orchestra.run.vm08.stdout:Unpacking ceph-base (20.2.0-712-g70f8415b-1jammy) ... 2026-03-20T17:56:34.161 INFO:teuthology.orchestra.run.vm06.stdout:Preparing to unpack .../55-python3-cachetools_5.0.0-1_all.deb ... 2026-03-20T17:56:34.161 INFO:teuthology.orchestra.run.vm06.stdout:Unpacking python3-cachetools (5.0.0-1) ... 2026-03-20T17:56:34.177 INFO:teuthology.orchestra.run.vm06.stdout:Selecting previously unselected package python3-rsa. 2026-03-20T17:56:34.183 INFO:teuthology.orchestra.run.vm06.stdout:Preparing to unpack .../56-python3-rsa_4.8-1_all.deb ... 2026-03-20T17:56:34.191 INFO:teuthology.orchestra.run.vm06.stdout:Unpacking python3-rsa (4.8-1) ... 2026-03-20T17:56:34.210 INFO:teuthology.orchestra.run.vm06.stdout:Selecting previously unselected package python3-google-auth. 2026-03-20T17:56:34.212 INFO:teuthology.orchestra.run.vm05.stdout:Preparing to unpack .../08-librados2_20.2.0-712-g70f8415b-1jammy_amd64.deb ... 2026-03-20T17:56:34.214 INFO:teuthology.orchestra.run.vm05.stdout:Unpacking librados2 (20.2.0-712-g70f8415b-1jammy) over (17.2.9-0ubuntu0.22.04.2) ... 2026-03-20T17:56:34.216 INFO:teuthology.orchestra.run.vm06.stdout:Preparing to unpack .../57-python3-google-auth_1.5.1-3_all.deb ... 2026-03-20T17:56:34.217 INFO:teuthology.orchestra.run.vm06.stdout:Unpacking python3-google-auth (1.5.1-3) ... 2026-03-20T17:56:34.271 INFO:teuthology.orchestra.run.vm06.stdout:Selecting previously unselected package python3-requests-oauthlib. 2026-03-20T17:56:34.277 INFO:teuthology.orchestra.run.vm08.stdout:Selecting previously unselected package python3-jaraco.functools. 2026-03-20T17:56:34.278 INFO:teuthology.orchestra.run.vm06.stdout:Preparing to unpack .../58-python3-requests-oauthlib_1.3.0+ds-0.1_all.deb ... 2026-03-20T17:56:34.278 INFO:teuthology.orchestra.run.vm05.stdout:Selecting previously unselected package libnbd0. 2026-03-20T17:56:34.279 INFO:teuthology.orchestra.run.vm06.stdout:Unpacking python3-requests-oauthlib (1.3.0+ds-0.1) ... 2026-03-20T17:56:34.283 INFO:teuthology.orchestra.run.vm08.stdout:Preparing to unpack .../25-python3-jaraco.functools_3.4.0-2_all.deb ... 2026-03-20T17:56:34.284 INFO:teuthology.orchestra.run.vm08.stdout:Unpacking python3-jaraco.functools (3.4.0-2) ... 2026-03-20T17:56:34.284 INFO:teuthology.orchestra.run.vm05.stdout:Preparing to unpack .../09-libnbd0_1.10.5-1_amd64.deb ... 2026-03-20T17:56:34.285 INFO:teuthology.orchestra.run.vm05.stdout:Unpacking libnbd0 (1.10.5-1) ... 2026-03-20T17:56:34.295 INFO:teuthology.orchestra.run.vm06.stdout:Selecting previously unselected package python3-websocket. 2026-03-20T17:56:34.299 INFO:teuthology.orchestra.run.vm08.stdout:Selecting previously unselected package python3-cheroot. 2026-03-20T17:56:34.300 INFO:teuthology.orchestra.run.vm05.stdout:Selecting previously unselected package libcephfs2. 2026-03-20T17:56:34.303 INFO:teuthology.orchestra.run.vm06.stdout:Preparing to unpack .../59-python3-websocket_1.2.3-1_all.deb ... 2026-03-20T17:56:34.303 INFO:teuthology.orchestra.run.vm06.stdout:Unpacking python3-websocket (1.2.3-1) ... 2026-03-20T17:56:34.305 INFO:teuthology.orchestra.run.vm08.stdout:Preparing to unpack .../26-python3-cheroot_8.5.2+ds1-1ubuntu3.1_all.deb ... 2026-03-20T17:56:34.306 INFO:teuthology.orchestra.run.vm08.stdout:Unpacking python3-cheroot (8.5.2+ds1-1ubuntu3.1) ... 2026-03-20T17:56:34.306 INFO:teuthology.orchestra.run.vm05.stdout:Preparing to unpack .../10-libcephfs2_20.2.0-712-g70f8415b-1jammy_amd64.deb ... 2026-03-20T17:56:34.307 INFO:teuthology.orchestra.run.vm05.stdout:Unpacking libcephfs2 (20.2.0-712-g70f8415b-1jammy) ... 2026-03-20T17:56:34.323 INFO:teuthology.orchestra.run.vm06.stdout:Selecting previously unselected package python3-kubernetes. 2026-03-20T17:56:34.327 INFO:teuthology.orchestra.run.vm08.stdout:Selecting previously unselected package python3-jaraco.classes. 2026-03-20T17:56:34.329 INFO:teuthology.orchestra.run.vm05.stdout:Selecting previously unselected package python3-rados. 2026-03-20T17:56:34.330 INFO:teuthology.orchestra.run.vm06.stdout:Preparing to unpack .../60-python3-kubernetes_12.0.1-1ubuntu1_all.deb ... 2026-03-20T17:56:34.331 INFO:teuthology.orchestra.run.vm06.stdout:Unpacking python3-kubernetes (12.0.1-1ubuntu1) ... 2026-03-20T17:56:34.333 INFO:teuthology.orchestra.run.vm08.stdout:Preparing to unpack .../27-python3-jaraco.classes_3.2.1-3_all.deb ... 2026-03-20T17:56:34.334 INFO:teuthology.orchestra.run.vm08.stdout:Unpacking python3-jaraco.classes (3.2.1-3) ... 2026-03-20T17:56:34.336 INFO:teuthology.orchestra.run.vm05.stdout:Preparing to unpack .../11-python3-rados_20.2.0-712-g70f8415b-1jammy_amd64.deb ... 2026-03-20T17:56:34.337 INFO:teuthology.orchestra.run.vm05.stdout:Unpacking python3-rados (20.2.0-712-g70f8415b-1jammy) ... 2026-03-20T17:56:34.351 INFO:teuthology.orchestra.run.vm08.stdout:Selecting previously unselected package python3-jaraco.text. 2026-03-20T17:56:34.356 INFO:teuthology.orchestra.run.vm05.stdout:Selecting previously unselected package python3-ceph-argparse. 2026-03-20T17:56:34.357 INFO:teuthology.orchestra.run.vm08.stdout:Preparing to unpack .../28-python3-jaraco.text_3.6.0-2_all.deb ... 2026-03-20T17:56:34.358 INFO:teuthology.orchestra.run.vm08.stdout:Unpacking python3-jaraco.text (3.6.0-2) ... 2026-03-20T17:56:34.362 INFO:teuthology.orchestra.run.vm05.stdout:Preparing to unpack .../12-python3-ceph-argparse_20.2.0-712-g70f8415b-1jammy_all.deb ... 2026-03-20T17:56:34.363 INFO:teuthology.orchestra.run.vm05.stdout:Unpacking python3-ceph-argparse (20.2.0-712-g70f8415b-1jammy) ... 2026-03-20T17:56:34.373 INFO:teuthology.orchestra.run.vm08.stdout:Selecting previously unselected package python3-jaraco.collections. 2026-03-20T17:56:34.378 INFO:teuthology.orchestra.run.vm05.stdout:Selecting previously unselected package python3-cephfs. 2026-03-20T17:56:34.379 INFO:teuthology.orchestra.run.vm08.stdout:Preparing to unpack .../29-python3-jaraco.collections_3.4.0-2_all.deb ... 2026-03-20T17:56:34.380 INFO:teuthology.orchestra.run.vm08.stdout:Unpacking python3-jaraco.collections (3.4.0-2) ... 2026-03-20T17:56:34.383 INFO:teuthology.orchestra.run.vm05.stdout:Preparing to unpack .../13-python3-cephfs_20.2.0-712-g70f8415b-1jammy_amd64.deb ... 2026-03-20T17:56:34.384 INFO:teuthology.orchestra.run.vm05.stdout:Unpacking python3-cephfs (20.2.0-712-g70f8415b-1jammy) ... 2026-03-20T17:56:34.394 INFO:teuthology.orchestra.run.vm08.stdout:Selecting previously unselected package python3-tempora. 2026-03-20T17:56:34.400 INFO:teuthology.orchestra.run.vm08.stdout:Preparing to unpack .../30-python3-tempora_4.1.2-1_all.deb ... 2026-03-20T17:56:34.401 INFO:teuthology.orchestra.run.vm08.stdout:Unpacking python3-tempora (4.1.2-1) ... 2026-03-20T17:56:34.402 INFO:teuthology.orchestra.run.vm05.stdout:Selecting previously unselected package python3-ceph-common. 2026-03-20T17:56:34.408 INFO:teuthology.orchestra.run.vm05.stdout:Preparing to unpack .../14-python3-ceph-common_20.2.0-712-g70f8415b-1jammy_all.deb ... 2026-03-20T17:56:34.409 INFO:teuthology.orchestra.run.vm05.stdout:Unpacking python3-ceph-common (20.2.0-712-g70f8415b-1jammy) ... 2026-03-20T17:56:34.432 INFO:teuthology.orchestra.run.vm08.stdout:Selecting previously unselected package python3-portend. 2026-03-20T17:56:34.438 INFO:teuthology.orchestra.run.vm08.stdout:Preparing to unpack .../31-python3-portend_3.0.0-1_all.deb ... 2026-03-20T17:56:34.439 INFO:teuthology.orchestra.run.vm08.stdout:Unpacking python3-portend (3.0.0-1) ... 2026-03-20T17:56:34.445 INFO:teuthology.orchestra.run.vm05.stdout:Selecting previously unselected package python3-wcwidth. 2026-03-20T17:56:34.451 INFO:teuthology.orchestra.run.vm05.stdout:Preparing to unpack .../15-python3-wcwidth_0.2.5+dfsg1-1_all.deb ... 2026-03-20T17:56:34.452 INFO:teuthology.orchestra.run.vm05.stdout:Unpacking python3-wcwidth (0.2.5+dfsg1-1) ... 2026-03-20T17:56:34.454 INFO:teuthology.orchestra.run.vm08.stdout:Selecting previously unselected package python3-zc.lockfile. 2026-03-20T17:56:34.459 INFO:teuthology.orchestra.run.vm08.stdout:Preparing to unpack .../32-python3-zc.lockfile_2.0-1_all.deb ... 2026-03-20T17:56:34.460 INFO:teuthology.orchestra.run.vm08.stdout:Unpacking python3-zc.lockfile (2.0-1) ... 2026-03-20T17:56:34.467 INFO:teuthology.orchestra.run.vm06.stdout:Selecting previously unselected package ceph-mgr-k8sevents. 2026-03-20T17:56:34.469 INFO:teuthology.orchestra.run.vm05.stdout:Selecting previously unselected package python3-prettytable. 2026-03-20T17:56:34.474 INFO:teuthology.orchestra.run.vm06.stdout:Preparing to unpack .../61-ceph-mgr-k8sevents_20.2.0-712-g70f8415b-1jammy_all.deb ... 2026-03-20T17:56:34.475 INFO:teuthology.orchestra.run.vm06.stdout:Unpacking ceph-mgr-k8sevents (20.2.0-712-g70f8415b-1jammy) ... 2026-03-20T17:56:34.476 INFO:teuthology.orchestra.run.vm08.stdout:Selecting previously unselected package python3-cherrypy3. 2026-03-20T17:56:34.476 INFO:teuthology.orchestra.run.vm05.stdout:Preparing to unpack .../16-python3-prettytable_2.5.0-2_all.deb ... 2026-03-20T17:56:34.477 INFO:teuthology.orchestra.run.vm05.stdout:Unpacking python3-prettytable (2.5.0-2) ... 2026-03-20T17:56:34.482 INFO:teuthology.orchestra.run.vm08.stdout:Preparing to unpack .../33-python3-cherrypy3_18.6.1-4_all.deb ... 2026-03-20T17:56:34.482 INFO:teuthology.orchestra.run.vm08.stdout:Unpacking python3-cherrypy3 (18.6.1-4) ... 2026-03-20T17:56:34.490 INFO:teuthology.orchestra.run.vm06.stdout:Selecting previously unselected package libonig5:amd64. 2026-03-20T17:56:34.492 INFO:teuthology.orchestra.run.vm05.stdout:Selecting previously unselected package python3-rbd. 2026-03-20T17:56:34.496 INFO:teuthology.orchestra.run.vm06.stdout:Preparing to unpack .../62-libonig5_6.9.7.1-2build1_amd64.deb ... 2026-03-20T17:56:34.497 INFO:teuthology.orchestra.run.vm06.stdout:Unpacking libonig5:amd64 (6.9.7.1-2build1) ... 2026-03-20T17:56:34.498 INFO:teuthology.orchestra.run.vm05.stdout:Preparing to unpack .../17-python3-rbd_20.2.0-712-g70f8415b-1jammy_amd64.deb ... 2026-03-20T17:56:34.499 INFO:teuthology.orchestra.run.vm05.stdout:Unpacking python3-rbd (20.2.0-712-g70f8415b-1jammy) ... 2026-03-20T17:56:34.513 INFO:teuthology.orchestra.run.vm08.stdout:Selecting previously unselected package python3-natsort. 2026-03-20T17:56:34.516 INFO:teuthology.orchestra.run.vm06.stdout:Selecting previously unselected package libjq1:amd64. 2026-03-20T17:56:34.520 INFO:teuthology.orchestra.run.vm05.stdout:Selecting previously unselected package librdkafka1:amd64. 2026-03-20T17:56:34.520 INFO:teuthology.orchestra.run.vm08.stdout:Preparing to unpack .../34-python3-natsort_8.0.2-1_all.deb ... 2026-03-20T17:56:34.521 INFO:teuthology.orchestra.run.vm08.stdout:Unpacking python3-natsort (8.0.2-1) ... 2026-03-20T17:56:34.522 INFO:teuthology.orchestra.run.vm06.stdout:Preparing to unpack .../63-libjq1_1.6-2.1ubuntu3.1_amd64.deb ... 2026-03-20T17:56:34.523 INFO:teuthology.orchestra.run.vm06.stdout:Unpacking libjq1:amd64 (1.6-2.1ubuntu3.1) ... 2026-03-20T17:56:34.525 INFO:teuthology.orchestra.run.vm05.stdout:Preparing to unpack .../18-librdkafka1_1.8.0-1build1_amd64.deb ... 2026-03-20T17:56:34.526 INFO:teuthology.orchestra.run.vm05.stdout:Unpacking librdkafka1:amd64 (1.8.0-1build1) ... 2026-03-20T17:56:34.537 INFO:teuthology.orchestra.run.vm08.stdout:Selecting previously unselected package ceph-mgr-modules-core. 2026-03-20T17:56:34.540 INFO:teuthology.orchestra.run.vm06.stdout:Selecting previously unselected package jq. 2026-03-20T17:56:34.543 INFO:teuthology.orchestra.run.vm08.stdout:Preparing to unpack .../35-ceph-mgr-modules-core_20.2.0-712-g70f8415b-1jammy_all.deb ... 2026-03-20T17:56:34.543 INFO:teuthology.orchestra.run.vm08.stdout:Unpacking ceph-mgr-modules-core (20.2.0-712-g70f8415b-1jammy) ... 2026-03-20T17:56:34.544 INFO:teuthology.orchestra.run.vm06.stdout:Preparing to unpack .../64-jq_1.6-2.1ubuntu3.1_amd64.deb ... 2026-03-20T17:56:34.545 INFO:teuthology.orchestra.run.vm06.stdout:Unpacking jq (1.6-2.1ubuntu3.1) ... 2026-03-20T17:56:34.549 INFO:teuthology.orchestra.run.vm05.stdout:Selecting previously unselected package librgw2. 2026-03-20T17:56:34.555 INFO:teuthology.orchestra.run.vm05.stdout:Preparing to unpack .../19-librgw2_20.2.0-712-g70f8415b-1jammy_amd64.deb ... 2026-03-20T17:56:34.556 INFO:teuthology.orchestra.run.vm05.stdout:Unpacking librgw2 (20.2.0-712-g70f8415b-1jammy) ... 2026-03-20T17:56:34.562 INFO:teuthology.orchestra.run.vm06.stdout:Selecting previously unselected package socat. 2026-03-20T17:56:34.568 INFO:teuthology.orchestra.run.vm06.stdout:Preparing to unpack .../65-socat_1.7.4.1-3ubuntu4_amd64.deb ... 2026-03-20T17:56:34.568 INFO:teuthology.orchestra.run.vm06.stdout:Unpacking socat (1.7.4.1-3ubuntu4) ... 2026-03-20T17:56:34.579 INFO:teuthology.orchestra.run.vm08.stdout:Selecting previously unselected package libsqlite3-mod-ceph. 2026-03-20T17:56:34.585 INFO:teuthology.orchestra.run.vm08.stdout:Preparing to unpack .../36-libsqlite3-mod-ceph_20.2.0-712-g70f8415b-1jammy_amd64.deb ... 2026-03-20T17:56:34.585 INFO:teuthology.orchestra.run.vm08.stdout:Unpacking libsqlite3-mod-ceph (20.2.0-712-g70f8415b-1jammy) ... 2026-03-20T17:56:34.592 INFO:teuthology.orchestra.run.vm06.stdout:Selecting previously unselected package xmlstarlet. 2026-03-20T17:56:34.599 INFO:teuthology.orchestra.run.vm06.stdout:Preparing to unpack .../66-xmlstarlet_1.6.1-2.1_amd64.deb ... 2026-03-20T17:56:34.600 INFO:teuthology.orchestra.run.vm06.stdout:Unpacking xmlstarlet (1.6.1-2.1) ... 2026-03-20T17:56:34.604 INFO:teuthology.orchestra.run.vm08.stdout:Selecting previously unselected package ceph-mgr. 2026-03-20T17:56:34.611 INFO:teuthology.orchestra.run.vm08.stdout:Preparing to unpack .../37-ceph-mgr_20.2.0-712-g70f8415b-1jammy_amd64.deb ... 2026-03-20T17:56:34.612 INFO:teuthology.orchestra.run.vm08.stdout:Unpacking ceph-mgr (20.2.0-712-g70f8415b-1jammy) ... 2026-03-20T17:56:34.648 INFO:teuthology.orchestra.run.vm08.stdout:Selecting previously unselected package ceph-mon. 2026-03-20T17:56:34.654 INFO:teuthology.orchestra.run.vm08.stdout:Preparing to unpack .../38-ceph-mon_20.2.0-712-g70f8415b-1jammy_amd64.deb ... 2026-03-20T17:56:34.655 INFO:teuthology.orchestra.run.vm08.stdout:Unpacking ceph-mon (20.2.0-712-g70f8415b-1jammy) ... 2026-03-20T17:56:34.656 INFO:teuthology.orchestra.run.vm06.stdout:Selecting previously unselected package ceph-test. 2026-03-20T17:56:34.663 INFO:teuthology.orchestra.run.vm06.stdout:Preparing to unpack .../67-ceph-test_20.2.0-712-g70f8415b-1jammy_amd64.deb ... 2026-03-20T17:56:34.689 INFO:teuthology.orchestra.run.vm06.stdout:Unpacking ceph-test (20.2.0-712-g70f8415b-1jammy) ... 2026-03-20T17:56:34.701 INFO:teuthology.orchestra.run.vm05.stdout:Selecting previously unselected package python3-rgw. 2026-03-20T17:56:34.707 INFO:teuthology.orchestra.run.vm05.stdout:Preparing to unpack .../20-python3-rgw_20.2.0-712-g70f8415b-1jammy_amd64.deb ... 2026-03-20T17:56:34.708 INFO:teuthology.orchestra.run.vm05.stdout:Unpacking python3-rgw (20.2.0-712-g70f8415b-1jammy) ... 2026-03-20T17:56:34.743 INFO:teuthology.orchestra.run.vm05.stdout:Selecting previously unselected package liboath0:amd64. 2026-03-20T17:56:34.748 INFO:teuthology.orchestra.run.vm08.stdout:Selecting previously unselected package libfuse2:amd64. 2026-03-20T17:56:34.749 INFO:teuthology.orchestra.run.vm05.stdout:Preparing to unpack .../21-liboath0_2.6.7-3ubuntu0.1_amd64.deb ... 2026-03-20T17:56:34.750 INFO:teuthology.orchestra.run.vm05.stdout:Unpacking liboath0:amd64 (2.6.7-3ubuntu0.1) ... 2026-03-20T17:56:34.754 INFO:teuthology.orchestra.run.vm08.stdout:Preparing to unpack .../39-libfuse2_2.9.9-5ubuntu3_amd64.deb ... 2026-03-20T17:56:34.754 INFO:teuthology.orchestra.run.vm08.stdout:Unpacking libfuse2:amd64 (2.9.9-5ubuntu3) ... 2026-03-20T17:56:34.765 INFO:teuthology.orchestra.run.vm05.stdout:Selecting previously unselected package libradosstriper1. 2026-03-20T17:56:34.771 INFO:teuthology.orchestra.run.vm05.stdout:Preparing to unpack .../22-libradosstriper1_20.2.0-712-g70f8415b-1jammy_amd64.deb ... 2026-03-20T17:56:34.771 INFO:teuthology.orchestra.run.vm05.stdout:Unpacking libradosstriper1 (20.2.0-712-g70f8415b-1jammy) ... 2026-03-20T17:56:34.772 INFO:teuthology.orchestra.run.vm08.stdout:Selecting previously unselected package ceph-osd. 2026-03-20T17:56:34.778 INFO:teuthology.orchestra.run.vm08.stdout:Preparing to unpack .../40-ceph-osd_20.2.0-712-g70f8415b-1jammy_amd64.deb ... 2026-03-20T17:56:34.779 INFO:teuthology.orchestra.run.vm08.stdout:Unpacking ceph-osd (20.2.0-712-g70f8415b-1jammy) ... 2026-03-20T17:56:34.792 INFO:teuthology.orchestra.run.vm05.stdout:Selecting previously unselected package ceph-common. 2026-03-20T17:56:34.798 INFO:teuthology.orchestra.run.vm05.stdout:Preparing to unpack .../23-ceph-common_20.2.0-712-g70f8415b-1jammy_amd64.deb ... 2026-03-20T17:56:34.799 INFO:teuthology.orchestra.run.vm05.stdout:Unpacking ceph-common (20.2.0-712-g70f8415b-1jammy) ... 2026-03-20T17:56:35.036 INFO:teuthology.orchestra.run.vm08.stdout:Selecting previously unselected package ceph. 2026-03-20T17:56:35.042 INFO:teuthology.orchestra.run.vm08.stdout:Preparing to unpack .../41-ceph_20.2.0-712-g70f8415b-1jammy_amd64.deb ... 2026-03-20T17:56:35.043 INFO:teuthology.orchestra.run.vm08.stdout:Unpacking ceph (20.2.0-712-g70f8415b-1jammy) ... 2026-03-20T17:56:35.059 INFO:teuthology.orchestra.run.vm08.stdout:Selecting previously unselected package ceph-fuse. 2026-03-20T17:56:35.065 INFO:teuthology.orchestra.run.vm08.stdout:Preparing to unpack .../42-ceph-fuse_20.2.0-712-g70f8415b-1jammy_amd64.deb ... 2026-03-20T17:56:35.066 INFO:teuthology.orchestra.run.vm08.stdout:Unpacking ceph-fuse (20.2.0-712-g70f8415b-1jammy) ... 2026-03-20T17:56:35.204 INFO:teuthology.orchestra.run.vm08.stdout:Selecting previously unselected package ceph-mds. 2026-03-20T17:56:35.210 INFO:teuthology.orchestra.run.vm08.stdout:Preparing to unpack .../43-ceph-mds_20.2.0-712-g70f8415b-1jammy_amd64.deb ... 2026-03-20T17:56:35.212 INFO:teuthology.orchestra.run.vm08.stdout:Unpacking ceph-mds (20.2.0-712-g70f8415b-1jammy) ... 2026-03-20T17:56:35.219 INFO:teuthology.orchestra.run.vm05.stdout:Selecting previously unselected package ceph-base. 2026-03-20T17:56:35.225 INFO:teuthology.orchestra.run.vm05.stdout:Preparing to unpack .../24-ceph-base_20.2.0-712-g70f8415b-1jammy_amd64.deb ... 2026-03-20T17:56:35.230 INFO:teuthology.orchestra.run.vm05.stdout:Unpacking ceph-base (20.2.0-712-g70f8415b-1jammy) ... 2026-03-20T17:56:35.257 INFO:teuthology.orchestra.run.vm08.stdout:Selecting previously unselected package cephadm. 2026-03-20T17:56:35.264 INFO:teuthology.orchestra.run.vm08.stdout:Preparing to unpack .../44-cephadm_20.2.0-712-g70f8415b-1jammy_amd64.deb ... 2026-03-20T17:56:35.265 INFO:teuthology.orchestra.run.vm08.stdout:Unpacking cephadm (20.2.0-712-g70f8415b-1jammy) ... 2026-03-20T17:56:35.285 INFO:teuthology.orchestra.run.vm08.stdout:Selecting previously unselected package python3-asyncssh. 2026-03-20T17:56:35.291 INFO:teuthology.orchestra.run.vm08.stdout:Preparing to unpack .../45-python3-asyncssh_2.5.0-1ubuntu0.1_all.deb ... 2026-03-20T17:56:35.292 INFO:teuthology.orchestra.run.vm08.stdout:Unpacking python3-asyncssh (2.5.0-1ubuntu0.1) ... 2026-03-20T17:56:35.336 INFO:teuthology.orchestra.run.vm08.stdout:Selecting previously unselected package ceph-mgr-cephadm. 2026-03-20T17:56:35.338 INFO:teuthology.orchestra.run.vm05.stdout:Selecting previously unselected package python3-jaraco.functools. 2026-03-20T17:56:35.340 INFO:teuthology.orchestra.run.vm08.stdout:Preparing to unpack .../46-ceph-mgr-cephadm_20.2.0-712-g70f8415b-1jammy_all.deb ... 2026-03-20T17:56:35.341 INFO:teuthology.orchestra.run.vm08.stdout:Unpacking ceph-mgr-cephadm (20.2.0-712-g70f8415b-1jammy) ... 2026-03-20T17:56:35.346 INFO:teuthology.orchestra.run.vm05.stdout:Preparing to unpack .../25-python3-jaraco.functools_3.4.0-2_all.deb ... 2026-03-20T17:56:35.347 INFO:teuthology.orchestra.run.vm05.stdout:Unpacking python3-jaraco.functools (3.4.0-2) ... 2026-03-20T17:56:35.365 INFO:teuthology.orchestra.run.vm08.stdout:Selecting previously unselected package python3-repoze.lru. 2026-03-20T17:56:35.366 INFO:teuthology.orchestra.run.vm05.stdout:Selecting previously unselected package python3-cheroot. 2026-03-20T17:56:35.370 INFO:teuthology.orchestra.run.vm08.stdout:Preparing to unpack .../47-python3-repoze.lru_0.7-2_all.deb ... 2026-03-20T17:56:35.371 INFO:teuthology.orchestra.run.vm08.stdout:Unpacking python3-repoze.lru (0.7-2) ... 2026-03-20T17:56:35.373 INFO:teuthology.orchestra.run.vm05.stdout:Preparing to unpack .../26-python3-cheroot_8.5.2+ds1-1ubuntu3.1_all.deb ... 2026-03-20T17:56:35.374 INFO:teuthology.orchestra.run.vm05.stdout:Unpacking python3-cheroot (8.5.2+ds1-1ubuntu3.1) ... 2026-03-20T17:56:35.389 INFO:teuthology.orchestra.run.vm08.stdout:Selecting previously unselected package python3-routes. 2026-03-20T17:56:35.395 INFO:teuthology.orchestra.run.vm05.stdout:Selecting previously unselected package python3-jaraco.classes. 2026-03-20T17:56:35.397 INFO:teuthology.orchestra.run.vm08.stdout:Preparing to unpack .../48-python3-routes_2.5.1-1ubuntu1_all.deb ... 2026-03-20T17:56:35.398 INFO:teuthology.orchestra.run.vm08.stdout:Unpacking python3-routes (2.5.1-1ubuntu1) ... 2026-03-20T17:56:35.401 INFO:teuthology.orchestra.run.vm05.stdout:Preparing to unpack .../27-python3-jaraco.classes_3.2.1-3_all.deb ... 2026-03-20T17:56:35.403 INFO:teuthology.orchestra.run.vm05.stdout:Unpacking python3-jaraco.classes (3.2.1-3) ... 2026-03-20T17:56:35.421 INFO:teuthology.orchestra.run.vm05.stdout:Selecting previously unselected package python3-jaraco.text. 2026-03-20T17:56:35.423 INFO:teuthology.orchestra.run.vm08.stdout:Selecting previously unselected package ceph-mgr-dashboard. 2026-03-20T17:56:35.427 INFO:teuthology.orchestra.run.vm05.stdout:Preparing to unpack .../28-python3-jaraco.text_3.6.0-2_all.deb ... 2026-03-20T17:56:35.428 INFO:teuthology.orchestra.run.vm05.stdout:Unpacking python3-jaraco.text (3.6.0-2) ... 2026-03-20T17:56:35.429 INFO:teuthology.orchestra.run.vm08.stdout:Preparing to unpack .../49-ceph-mgr-dashboard_20.2.0-712-g70f8415b-1jammy_all.deb ... 2026-03-20T17:56:35.430 INFO:teuthology.orchestra.run.vm08.stdout:Unpacking ceph-mgr-dashboard (20.2.0-712-g70f8415b-1jammy) ... 2026-03-20T17:56:35.445 INFO:teuthology.orchestra.run.vm05.stdout:Selecting previously unselected package python3-jaraco.collections. 2026-03-20T17:56:35.452 INFO:teuthology.orchestra.run.vm05.stdout:Preparing to unpack .../29-python3-jaraco.collections_3.4.0-2_all.deb ... 2026-03-20T17:56:35.453 INFO:teuthology.orchestra.run.vm05.stdout:Unpacking python3-jaraco.collections (3.4.0-2) ... 2026-03-20T17:56:35.469 INFO:teuthology.orchestra.run.vm05.stdout:Selecting previously unselected package python3-tempora. 2026-03-20T17:56:35.475 INFO:teuthology.orchestra.run.vm05.stdout:Preparing to unpack .../30-python3-tempora_4.1.2-1_all.deb ... 2026-03-20T17:56:35.476 INFO:teuthology.orchestra.run.vm05.stdout:Unpacking python3-tempora (4.1.2-1) ... 2026-03-20T17:56:35.494 INFO:teuthology.orchestra.run.vm05.stdout:Selecting previously unselected package python3-portend. 2026-03-20T17:56:35.501 INFO:teuthology.orchestra.run.vm05.stdout:Preparing to unpack .../31-python3-portend_3.0.0-1_all.deb ... 2026-03-20T17:56:35.502 INFO:teuthology.orchestra.run.vm05.stdout:Unpacking python3-portend (3.0.0-1) ... 2026-03-20T17:56:35.518 INFO:teuthology.orchestra.run.vm05.stdout:Selecting previously unselected package python3-zc.lockfile. 2026-03-20T17:56:35.525 INFO:teuthology.orchestra.run.vm05.stdout:Preparing to unpack .../32-python3-zc.lockfile_2.0-1_all.deb ... 2026-03-20T17:56:35.526 INFO:teuthology.orchestra.run.vm05.stdout:Unpacking python3-zc.lockfile (2.0-1) ... 2026-03-20T17:56:35.543 INFO:teuthology.orchestra.run.vm05.stdout:Selecting previously unselected package python3-cherrypy3. 2026-03-20T17:56:35.549 INFO:teuthology.orchestra.run.vm05.stdout:Preparing to unpack .../33-python3-cherrypy3_18.6.1-4_all.deb ... 2026-03-20T17:56:35.550 INFO:teuthology.orchestra.run.vm05.stdout:Unpacking python3-cherrypy3 (18.6.1-4) ... 2026-03-20T17:56:35.763 INFO:teuthology.orchestra.run.vm05.stdout:Selecting previously unselected package python3-natsort. 2026-03-20T17:56:35.769 INFO:teuthology.orchestra.run.vm05.stdout:Preparing to unpack .../34-python3-natsort_8.0.2-1_all.deb ... 2026-03-20T17:56:35.882 INFO:teuthology.orchestra.run.vm05.stdout:Unpacking python3-natsort (8.0.2-1) ... 2026-03-20T17:56:36.153 INFO:teuthology.orchestra.run.vm05.stdout:Selecting previously unselected package ceph-mgr-modules-core. 2026-03-20T17:56:36.155 INFO:teuthology.orchestra.run.vm06.stdout:Selecting previously unselected package ceph-volume. 2026-03-20T17:56:36.158 INFO:teuthology.orchestra.run.vm05.stdout:Preparing to unpack .../35-ceph-mgr-modules-core_20.2.0-712-g70f8415b-1jammy_all.deb ... 2026-03-20T17:56:36.160 INFO:teuthology.orchestra.run.vm05.stdout:Unpacking ceph-mgr-modules-core (20.2.0-712-g70f8415b-1jammy) ... 2026-03-20T17:56:36.161 INFO:teuthology.orchestra.run.vm06.stdout:Preparing to unpack .../68-ceph-volume_20.2.0-712-g70f8415b-1jammy_all.deb ... 2026-03-20T17:56:36.162 INFO:teuthology.orchestra.run.vm06.stdout:Unpacking ceph-volume (20.2.0-712-g70f8415b-1jammy) ... 2026-03-20T17:56:36.189 INFO:teuthology.orchestra.run.vm06.stdout:Selecting previously unselected package libcephfs-daemon. 2026-03-20T17:56:36.195 INFO:teuthology.orchestra.run.vm06.stdout:Preparing to unpack .../69-libcephfs-daemon_20.2.0-712-g70f8415b-1jammy_amd64.deb ... 2026-03-20T17:56:36.196 INFO:teuthology.orchestra.run.vm06.stdout:Unpacking libcephfs-daemon (20.2.0-712-g70f8415b-1jammy) ... 2026-03-20T17:56:36.200 INFO:teuthology.orchestra.run.vm05.stdout:Selecting previously unselected package libsqlite3-mod-ceph. 2026-03-20T17:56:36.206 INFO:teuthology.orchestra.run.vm05.stdout:Preparing to unpack .../36-libsqlite3-mod-ceph_20.2.0-712-g70f8415b-1jammy_amd64.deb ... 2026-03-20T17:56:36.207 INFO:teuthology.orchestra.run.vm05.stdout:Unpacking libsqlite3-mod-ceph (20.2.0-712-g70f8415b-1jammy) ... 2026-03-20T17:56:36.210 INFO:teuthology.orchestra.run.vm08.stdout:Selecting previously unselected package python3-sklearn-lib:amd64. 2026-03-20T17:56:36.211 INFO:teuthology.orchestra.run.vm06.stdout:Selecting previously unselected package libcephfs-proxy2. 2026-03-20T17:56:36.215 INFO:teuthology.orchestra.run.vm08.stdout:Preparing to unpack .../50-python3-sklearn-lib_0.23.2-5ubuntu6_amd64.deb ... 2026-03-20T17:56:36.216 INFO:teuthology.orchestra.run.vm08.stdout:Unpacking python3-sklearn-lib:amd64 (0.23.2-5ubuntu6) ... 2026-03-20T17:56:36.217 INFO:teuthology.orchestra.run.vm06.stdout:Preparing to unpack .../70-libcephfs-proxy2_20.2.0-712-g70f8415b-1jammy_amd64.deb ... 2026-03-20T17:56:36.217 INFO:teuthology.orchestra.run.vm06.stdout:Unpacking libcephfs-proxy2 (20.2.0-712-g70f8415b-1jammy) ... 2026-03-20T17:56:36.226 INFO:teuthology.orchestra.run.vm05.stdout:Selecting previously unselected package ceph-mgr. 2026-03-20T17:56:36.230 INFO:teuthology.orchestra.run.vm06.stdout:Selecting previously unselected package libcephfs-dev. 2026-03-20T17:56:36.233 INFO:teuthology.orchestra.run.vm05.stdout:Preparing to unpack .../37-ceph-mgr_20.2.0-712-g70f8415b-1jammy_amd64.deb ... 2026-03-20T17:56:36.233 INFO:teuthology.orchestra.run.vm05.stdout:Unpacking ceph-mgr (20.2.0-712-g70f8415b-1jammy) ... 2026-03-20T17:56:36.236 INFO:teuthology.orchestra.run.vm06.stdout:Preparing to unpack .../71-libcephfs-dev_20.2.0-712-g70f8415b-1jammy_amd64.deb ... 2026-03-20T17:56:36.237 INFO:teuthology.orchestra.run.vm06.stdout:Unpacking libcephfs-dev (20.2.0-712-g70f8415b-1jammy) ... 2026-03-20T17:56:36.266 INFO:teuthology.orchestra.run.vm06.stdout:Selecting previously unselected package nvme-cli. 2026-03-20T17:56:36.272 INFO:teuthology.orchestra.run.vm06.stdout:Preparing to unpack .../72-nvme-cli_1.16-3ubuntu0.3_amd64.deb ... 2026-03-20T17:56:36.273 INFO:teuthology.orchestra.run.vm05.stdout:Selecting previously unselected package ceph-mon. 2026-03-20T17:56:36.273 INFO:teuthology.orchestra.run.vm06.stdout:Unpacking nvme-cli (1.16-3ubuntu0.3) ... 2026-03-20T17:56:36.275 INFO:teuthology.orchestra.run.vm08.stdout:Selecting previously unselected package python3-joblib. 2026-03-20T17:56:36.279 INFO:teuthology.orchestra.run.vm05.stdout:Preparing to unpack .../38-ceph-mon_20.2.0-712-g70f8415b-1jammy_amd64.deb ... 2026-03-20T17:56:36.280 INFO:teuthology.orchestra.run.vm05.stdout:Unpacking ceph-mon (20.2.0-712-g70f8415b-1jammy) ... 2026-03-20T17:56:36.281 INFO:teuthology.orchestra.run.vm08.stdout:Preparing to unpack .../51-python3-joblib_0.17.0-4ubuntu1_all.deb ... 2026-03-20T17:56:36.282 INFO:teuthology.orchestra.run.vm08.stdout:Unpacking python3-joblib (0.17.0-4ubuntu1) ... 2026-03-20T17:56:36.311 INFO:teuthology.orchestra.run.vm06.stdout:Selecting previously unselected package python-asyncssh-doc. 2026-03-20T17:56:36.317 INFO:teuthology.orchestra.run.vm08.stdout:Selecting previously unselected package python3-threadpoolctl. 2026-03-20T17:56:36.317 INFO:teuthology.orchestra.run.vm06.stdout:Preparing to unpack .../73-python-asyncssh-doc_2.5.0-1ubuntu0.1_all.deb ... 2026-03-20T17:56:36.318 INFO:teuthology.orchestra.run.vm06.stdout:Unpacking python-asyncssh-doc (2.5.0-1ubuntu0.1) ... 2026-03-20T17:56:36.324 INFO:teuthology.orchestra.run.vm08.stdout:Preparing to unpack .../52-python3-threadpoolctl_3.1.0-1_all.deb ... 2026-03-20T17:56:36.325 INFO:teuthology.orchestra.run.vm08.stdout:Unpacking python3-threadpoolctl (3.1.0-1) ... 2026-03-20T17:56:36.367 INFO:teuthology.orchestra.run.vm08.stdout:Selecting previously unselected package python3-sklearn. 2026-03-20T17:56:36.373 INFO:teuthology.orchestra.run.vm08.stdout:Preparing to unpack .../53-python3-sklearn_0.23.2-5ubuntu6_all.deb ... 2026-03-20T17:56:36.374 INFO:teuthology.orchestra.run.vm08.stdout:Unpacking python3-sklearn (0.23.2-5ubuntu6) ... 2026-03-20T17:56:36.379 INFO:teuthology.orchestra.run.vm06.stdout:Selecting previously unselected package python3-iniconfig. 2026-03-20T17:56:36.380 INFO:teuthology.orchestra.run.vm05.stdout:Selecting previously unselected package libfuse2:amd64. 2026-03-20T17:56:36.384 INFO:teuthology.orchestra.run.vm05.stdout:Preparing to unpack .../39-libfuse2_2.9.9-5ubuntu3_amd64.deb ... 2026-03-20T17:56:36.385 INFO:teuthology.orchestra.run.vm05.stdout:Unpacking libfuse2:amd64 (2.9.9-5ubuntu3) ... 2026-03-20T17:56:36.387 INFO:teuthology.orchestra.run.vm06.stdout:Preparing to unpack .../74-python3-iniconfig_1.1.1-2_all.deb ... 2026-03-20T17:56:36.388 INFO:teuthology.orchestra.run.vm06.stdout:Unpacking python3-iniconfig (1.1.1-2) ... 2026-03-20T17:56:36.402 INFO:teuthology.orchestra.run.vm05.stdout:Selecting previously unselected package ceph-osd. 2026-03-20T17:56:36.405 INFO:teuthology.orchestra.run.vm06.stdout:Selecting previously unselected package python3-pluggy. 2026-03-20T17:56:36.407 INFO:teuthology.orchestra.run.vm05.stdout:Preparing to unpack .../40-ceph-osd_20.2.0-712-g70f8415b-1jammy_amd64.deb ... 2026-03-20T17:56:36.408 INFO:teuthology.orchestra.run.vm05.stdout:Unpacking ceph-osd (20.2.0-712-g70f8415b-1jammy) ... 2026-03-20T17:56:36.412 INFO:teuthology.orchestra.run.vm06.stdout:Preparing to unpack .../75-python3-pluggy_0.13.0-7.1_all.deb ... 2026-03-20T17:56:36.413 INFO:teuthology.orchestra.run.vm06.stdout:Unpacking python3-pluggy (0.13.0-7.1) ... 2026-03-20T17:56:36.432 INFO:teuthology.orchestra.run.vm06.stdout:Selecting previously unselected package python3-psutil. 2026-03-20T17:56:36.441 INFO:teuthology.orchestra.run.vm06.stdout:Preparing to unpack .../76-python3-psutil_5.9.0-1build1_amd64.deb ... 2026-03-20T17:56:36.441 INFO:teuthology.orchestra.run.vm06.stdout:Unpacking python3-psutil (5.9.0-1build1) ... 2026-03-20T17:56:36.475 INFO:teuthology.orchestra.run.vm06.stdout:Selecting previously unselected package python3-py. 2026-03-20T17:56:36.482 INFO:teuthology.orchestra.run.vm06.stdout:Preparing to unpack .../77-python3-py_1.10.0-1_all.deb ... 2026-03-20T17:56:36.483 INFO:teuthology.orchestra.run.vm06.stdout:Unpacking python3-py (1.10.0-1) ... 2026-03-20T17:56:36.502 INFO:teuthology.orchestra.run.vm08.stdout:Selecting previously unselected package ceph-mgr-diskprediction-local. 2026-03-20T17:56:36.508 INFO:teuthology.orchestra.run.vm06.stdout:Selecting previously unselected package python3-pygments. 2026-03-20T17:56:36.509 INFO:teuthology.orchestra.run.vm08.stdout:Preparing to unpack .../54-ceph-mgr-diskprediction-local_20.2.0-712-g70f8415b-1jammy_all.deb ... 2026-03-20T17:56:36.511 INFO:teuthology.orchestra.run.vm08.stdout:Unpacking ceph-mgr-diskprediction-local (20.2.0-712-g70f8415b-1jammy) ... 2026-03-20T17:56:36.515 INFO:teuthology.orchestra.run.vm06.stdout:Preparing to unpack .../78-python3-pygments_2.11.2+dfsg-2ubuntu0.1_all.deb ... 2026-03-20T17:56:36.516 INFO:teuthology.orchestra.run.vm06.stdout:Unpacking python3-pygments (2.11.2+dfsg-2ubuntu0.1) ... 2026-03-20T17:56:36.597 INFO:teuthology.orchestra.run.vm06.stdout:Selecting previously unselected package python3-toml. 2026-03-20T17:56:36.605 INFO:teuthology.orchestra.run.vm06.stdout:Preparing to unpack .../79-python3-toml_0.10.2-1_all.deb ... 2026-03-20T17:56:36.606 INFO:teuthology.orchestra.run.vm06.stdout:Unpacking python3-toml (0.10.2-1) ... 2026-03-20T17:56:36.743 INFO:teuthology.orchestra.run.vm05.stdout:Selecting previously unselected package ceph. 2026-03-20T17:56:36.744 INFO:teuthology.orchestra.run.vm06.stdout:Selecting previously unselected package python3-pytest. 2026-03-20T17:56:36.749 INFO:teuthology.orchestra.run.vm05.stdout:Preparing to unpack .../41-ceph_20.2.0-712-g70f8415b-1jammy_amd64.deb ... 2026-03-20T17:56:36.750 INFO:teuthology.orchestra.run.vm05.stdout:Unpacking ceph (20.2.0-712-g70f8415b-1jammy) ... 2026-03-20T17:56:36.751 INFO:teuthology.orchestra.run.vm06.stdout:Preparing to unpack .../80-python3-pytest_6.2.5-1ubuntu2_all.deb ... 2026-03-20T17:56:36.752 INFO:teuthology.orchestra.run.vm08.stdout:Selecting previously unselected package python3-cachetools. 2026-03-20T17:56:36.752 INFO:teuthology.orchestra.run.vm06.stdout:Unpacking python3-pytest (6.2.5-1ubuntu2) ... 2026-03-20T17:56:36.760 INFO:teuthology.orchestra.run.vm08.stdout:Preparing to unpack .../55-python3-cachetools_5.0.0-1_all.deb ... 2026-03-20T17:56:36.761 INFO:teuthology.orchestra.run.vm08.stdout:Unpacking python3-cachetools (5.0.0-1) ... 2026-03-20T17:56:36.767 INFO:teuthology.orchestra.run.vm05.stdout:Selecting previously unselected package ceph-fuse. 2026-03-20T17:56:36.773 INFO:teuthology.orchestra.run.vm05.stdout:Preparing to unpack .../42-ceph-fuse_20.2.0-712-g70f8415b-1jammy_amd64.deb ... 2026-03-20T17:56:36.774 INFO:teuthology.orchestra.run.vm05.stdout:Unpacking ceph-fuse (20.2.0-712-g70f8415b-1jammy) ... 2026-03-20T17:56:36.781 INFO:teuthology.orchestra.run.vm08.stdout:Selecting previously unselected package python3-rsa. 2026-03-20T17:56:36.788 INFO:teuthology.orchestra.run.vm08.stdout:Preparing to unpack .../56-python3-rsa_4.8-1_all.deb ... 2026-03-20T17:56:36.789 INFO:teuthology.orchestra.run.vm08.stdout:Unpacking python3-rsa (4.8-1) ... 2026-03-20T17:56:36.794 INFO:teuthology.orchestra.run.vm06.stdout:Selecting previously unselected package python3-simplejson. 2026-03-20T17:56:36.800 INFO:teuthology.orchestra.run.vm06.stdout:Preparing to unpack .../81-python3-simplejson_3.17.6-1build1_amd64.deb ... 2026-03-20T17:56:36.801 INFO:teuthology.orchestra.run.vm06.stdout:Unpacking python3-simplejson (3.17.6-1build1) ... 2026-03-20T17:56:36.802 INFO:teuthology.orchestra.run.vm05.stdout:Selecting previously unselected package ceph-mds. 2026-03-20T17:56:36.809 INFO:teuthology.orchestra.run.vm05.stdout:Preparing to unpack .../43-ceph-mds_20.2.0-712-g70f8415b-1jammy_amd64.deb ... 2026-03-20T17:56:36.809 INFO:teuthology.orchestra.run.vm08.stdout:Selecting previously unselected package python3-google-auth. 2026-03-20T17:56:36.810 INFO:teuthology.orchestra.run.vm05.stdout:Unpacking ceph-mds (20.2.0-712-g70f8415b-1jammy) ... 2026-03-20T17:56:36.817 INFO:teuthology.orchestra.run.vm08.stdout:Preparing to unpack .../57-python3-google-auth_1.5.1-3_all.deb ... 2026-03-20T17:56:36.818 INFO:teuthology.orchestra.run.vm08.stdout:Unpacking python3-google-auth (1.5.1-3) ... 2026-03-20T17:56:36.822 INFO:teuthology.orchestra.run.vm06.stdout:Selecting previously unselected package python3-webob. 2026-03-20T17:56:36.828 INFO:teuthology.orchestra.run.vm06.stdout:Preparing to unpack .../82-python3-webob_1%3a1.8.6-1.1ubuntu0.1_all.deb ... 2026-03-20T17:56:36.829 INFO:teuthology.orchestra.run.vm06.stdout:Unpacking python3-webob (1:1.8.6-1.1ubuntu0.1) ... 2026-03-20T17:56:36.844 INFO:teuthology.orchestra.run.vm08.stdout:Selecting previously unselected package python3-requests-oauthlib. 2026-03-20T17:56:36.850 INFO:teuthology.orchestra.run.vm08.stdout:Preparing to unpack .../58-python3-requests-oauthlib_1.3.0+ds-0.1_all.deb ... 2026-03-20T17:56:36.851 INFO:teuthology.orchestra.run.vm08.stdout:Unpacking python3-requests-oauthlib (1.3.0+ds-0.1) ... 2026-03-20T17:56:36.855 INFO:teuthology.orchestra.run.vm06.stdout:Selecting previously unselected package qttranslations5-l10n. 2026-03-20T17:56:36.857 INFO:teuthology.orchestra.run.vm05.stdout:Selecting previously unselected package cephadm. 2026-03-20T17:56:36.861 INFO:teuthology.orchestra.run.vm06.stdout:Preparing to unpack .../83-qttranslations5-l10n_5.15.3-1_all.deb ... 2026-03-20T17:56:36.862 INFO:teuthology.orchestra.run.vm06.stdout:Unpacking qttranslations5-l10n (5.15.3-1) ... 2026-03-20T17:56:36.864 INFO:teuthology.orchestra.run.vm05.stdout:Preparing to unpack .../44-cephadm_20.2.0-712-g70f8415b-1jammy_amd64.deb ... 2026-03-20T17:56:36.865 INFO:teuthology.orchestra.run.vm05.stdout:Unpacking cephadm (20.2.0-712-g70f8415b-1jammy) ... 2026-03-20T17:56:36.870 INFO:teuthology.orchestra.run.vm08.stdout:Selecting previously unselected package python3-websocket. 2026-03-20T17:56:36.877 INFO:teuthology.orchestra.run.vm08.stdout:Preparing to unpack .../59-python3-websocket_1.2.3-1_all.deb ... 2026-03-20T17:56:36.878 INFO:teuthology.orchestra.run.vm08.stdout:Unpacking python3-websocket (1.2.3-1) ... 2026-03-20T17:56:36.885 INFO:teuthology.orchestra.run.vm05.stdout:Selecting previously unselected package python3-asyncssh. 2026-03-20T17:56:36.891 INFO:teuthology.orchestra.run.vm05.stdout:Preparing to unpack .../45-python3-asyncssh_2.5.0-1ubuntu0.1_all.deb ... 2026-03-20T17:56:36.892 INFO:teuthology.orchestra.run.vm05.stdout:Unpacking python3-asyncssh (2.5.0-1ubuntu0.1) ... 2026-03-20T17:56:36.899 INFO:teuthology.orchestra.run.vm08.stdout:Selecting previously unselected package python3-kubernetes. 2026-03-20T17:56:36.906 INFO:teuthology.orchestra.run.vm08.stdout:Preparing to unpack .../60-python3-kubernetes_12.0.1-1ubuntu1_all.deb ... 2026-03-20T17:56:36.907 INFO:teuthology.orchestra.run.vm08.stdout:Unpacking python3-kubernetes (12.0.1-1ubuntu1) ... 2026-03-20T17:56:36.919 INFO:teuthology.orchestra.run.vm05.stdout:Selecting previously unselected package ceph-mgr-cephadm. 2026-03-20T17:56:36.925 INFO:teuthology.orchestra.run.vm05.stdout:Preparing to unpack .../46-ceph-mgr-cephadm_20.2.0-712-g70f8415b-1jammy_all.deb ... 2026-03-20T17:56:36.942 INFO:teuthology.orchestra.run.vm05.stdout:Unpacking ceph-mgr-cephadm (20.2.0-712-g70f8415b-1jammy) ... 2026-03-20T17:56:36.968 INFO:teuthology.orchestra.run.vm05.stdout:Selecting previously unselected package python3-repoze.lru. 2026-03-20T17:56:36.971 INFO:teuthology.orchestra.run.vm06.stdout:Selecting previously unselected package radosgw. 2026-03-20T17:56:36.973 INFO:teuthology.orchestra.run.vm05.stdout:Preparing to unpack .../47-python3-repoze.lru_0.7-2_all.deb ... 2026-03-20T17:56:36.974 INFO:teuthology.orchestra.run.vm05.stdout:Unpacking python3-repoze.lru (0.7-2) ... 2026-03-20T17:56:36.978 INFO:teuthology.orchestra.run.vm06.stdout:Preparing to unpack .../84-radosgw_20.2.0-712-g70f8415b-1jammy_amd64.deb ... 2026-03-20T17:56:36.979 INFO:teuthology.orchestra.run.vm06.stdout:Unpacking radosgw (20.2.0-712-g70f8415b-1jammy) ... 2026-03-20T17:56:36.990 INFO:teuthology.orchestra.run.vm05.stdout:Selecting previously unselected package python3-routes. 2026-03-20T17:56:36.996 INFO:teuthology.orchestra.run.vm05.stdout:Preparing to unpack .../48-python3-routes_2.5.1-1ubuntu1_all.deb ... 2026-03-20T17:56:36.997 INFO:teuthology.orchestra.run.vm05.stdout:Unpacking python3-routes (2.5.1-1ubuntu1) ... 2026-03-20T17:56:37.032 INFO:teuthology.orchestra.run.vm05.stdout:Selecting previously unselected package ceph-mgr-dashboard. 2026-03-20T17:56:37.038 INFO:teuthology.orchestra.run.vm05.stdout:Preparing to unpack .../49-ceph-mgr-dashboard_20.2.0-712-g70f8415b-1jammy_all.deb ... 2026-03-20T17:56:37.039 INFO:teuthology.orchestra.run.vm05.stdout:Unpacking ceph-mgr-dashboard (20.2.0-712-g70f8415b-1jammy) ... 2026-03-20T17:56:37.056 INFO:teuthology.orchestra.run.vm08.stdout:Selecting previously unselected package ceph-mgr-k8sevents. 2026-03-20T17:56:37.062 INFO:teuthology.orchestra.run.vm08.stdout:Preparing to unpack .../61-ceph-mgr-k8sevents_20.2.0-712-g70f8415b-1jammy_all.deb ... 2026-03-20T17:56:37.069 INFO:teuthology.orchestra.run.vm08.stdout:Unpacking ceph-mgr-k8sevents (20.2.0-712-g70f8415b-1jammy) ... 2026-03-20T17:56:37.087 INFO:teuthology.orchestra.run.vm08.stdout:Selecting previously unselected package libonig5:amd64. 2026-03-20T17:56:37.093 INFO:teuthology.orchestra.run.vm08.stdout:Preparing to unpack .../62-libonig5_6.9.7.1-2build1_amd64.deb ... 2026-03-20T17:56:37.094 INFO:teuthology.orchestra.run.vm08.stdout:Unpacking libonig5:amd64 (6.9.7.1-2build1) ... 2026-03-20T17:56:37.114 INFO:teuthology.orchestra.run.vm08.stdout:Selecting previously unselected package libjq1:amd64. 2026-03-20T17:56:37.121 INFO:teuthology.orchestra.run.vm08.stdout:Preparing to unpack .../63-libjq1_1.6-2.1ubuntu3.1_amd64.deb ... 2026-03-20T17:56:37.122 INFO:teuthology.orchestra.run.vm08.stdout:Unpacking libjq1:amd64 (1.6-2.1ubuntu3.1) ... 2026-03-20T17:56:37.142 INFO:teuthology.orchestra.run.vm08.stdout:Selecting previously unselected package jq. 2026-03-20T17:56:37.147 INFO:teuthology.orchestra.run.vm08.stdout:Preparing to unpack .../64-jq_1.6-2.1ubuntu3.1_amd64.deb ... 2026-03-20T17:56:37.148 INFO:teuthology.orchestra.run.vm08.stdout:Unpacking jq (1.6-2.1ubuntu3.1) ... 2026-03-20T17:56:37.166 INFO:teuthology.orchestra.run.vm08.stdout:Selecting previously unselected package socat. 2026-03-20T17:56:37.171 INFO:teuthology.orchestra.run.vm08.stdout:Preparing to unpack .../65-socat_1.7.4.1-3ubuntu4_amd64.deb ... 2026-03-20T17:56:37.172 INFO:teuthology.orchestra.run.vm08.stdout:Unpacking socat (1.7.4.1-3ubuntu4) ... 2026-03-20T17:56:37.187 INFO:teuthology.orchestra.run.vm02.stdout:Get:82 https://1.chacra.ceph.com/r/ceph/tentacle/70f8415b300f041766fa27faf7d5472699e32388/ubuntu/jammy/flavors/default jammy/main amd64 ceph-volume all 20.2.0-712-g70f8415b-1jammy [135 kB] 2026-03-20T17:56:37.187 INFO:teuthology.orchestra.run.vm02.stdout:Get:83 https://1.chacra.ceph.com/r/ceph/tentacle/70f8415b300f041766fa27faf7d5472699e32388/ubuntu/jammy/flavors/default jammy/main amd64 libcephfs-daemon amd64 20.2.0-712-g70f8415b-1jammy [43.3 kB] 2026-03-20T17:56:37.188 INFO:teuthology.orchestra.run.vm02.stdout:Get:84 https://1.chacra.ceph.com/r/ceph/tentacle/70f8415b300f041766fa27faf7d5472699e32388/ubuntu/jammy/flavors/default jammy/main amd64 libcephfs-proxy2 amd64 20.2.0-712-g70f8415b-1jammy [30.7 kB] 2026-03-20T17:56:37.188 INFO:teuthology.orchestra.run.vm02.stdout:Get:85 https://1.chacra.ceph.com/r/ceph/tentacle/70f8415b300f041766fa27faf7d5472699e32388/ubuntu/jammy/flavors/default jammy/main amd64 libcephfs-dev amd64 20.2.0-712-g70f8415b-1jammy [41.5 kB] 2026-03-20T17:56:37.188 INFO:teuthology.orchestra.run.vm02.stdout:Get:86 https://1.chacra.ceph.com/r/ceph/tentacle/70f8415b300f041766fa27faf7d5472699e32388/ubuntu/jammy/flavors/default jammy/main amd64 radosgw amd64 20.2.0-712-g70f8415b-1jammy [25.1 MB] 2026-03-20T17:56:37.199 INFO:teuthology.orchestra.run.vm08.stdout:Selecting previously unselected package xmlstarlet. 2026-03-20T17:56:37.206 INFO:teuthology.orchestra.run.vm08.stdout:Preparing to unpack .../66-xmlstarlet_1.6.1-2.1_amd64.deb ... 2026-03-20T17:56:37.207 INFO:teuthology.orchestra.run.vm08.stdout:Unpacking xmlstarlet (1.6.1-2.1) ... 2026-03-20T17:56:37.326 INFO:teuthology.orchestra.run.vm06.stdout:Selecting previously unselected package rbd-fuse. 2026-03-20T17:56:37.333 INFO:teuthology.orchestra.run.vm06.stdout:Preparing to unpack .../85-rbd-fuse_20.2.0-712-g70f8415b-1jammy_amd64.deb ... 2026-03-20T17:56:37.333 INFO:teuthology.orchestra.run.vm06.stdout:Unpacking rbd-fuse (20.2.0-712-g70f8415b-1jammy) ... 2026-03-20T17:56:37.334 INFO:teuthology.orchestra.run.vm08.stdout:Selecting previously unselected package ceph-test. 2026-03-20T17:56:37.338 INFO:teuthology.orchestra.run.vm08.stdout:Preparing to unpack .../67-ceph-test_20.2.0-712-g70f8415b-1jammy_amd64.deb ... 2026-03-20T17:56:37.339 INFO:teuthology.orchestra.run.vm08.stdout:Unpacking ceph-test (20.2.0-712-g70f8415b-1jammy) ... 2026-03-20T17:56:37.350 INFO:teuthology.orchestra.run.vm06.stdout:Selecting previously unselected package smartmontools. 2026-03-20T17:56:37.356 INFO:teuthology.orchestra.run.vm06.stdout:Preparing to unpack .../86-smartmontools_7.2-1ubuntu0.1_amd64.deb ... 2026-03-20T17:56:37.365 INFO:teuthology.orchestra.run.vm06.stdout:Unpacking smartmontools (7.2-1ubuntu0.1) ... 2026-03-20T17:56:37.406 INFO:teuthology.orchestra.run.vm06.stdout:Setting up smartmontools (7.2-1ubuntu0.1) ... 2026-03-20T17:56:37.683 INFO:teuthology.orchestra.run.vm05.stdout:Selecting previously unselected package python3-sklearn-lib:amd64. 2026-03-20T17:56:37.689 INFO:teuthology.orchestra.run.vm05.stdout:Preparing to unpack .../50-python3-sklearn-lib_0.23.2-5ubuntu6_amd64.deb ... 2026-03-20T17:56:37.690 INFO:teuthology.orchestra.run.vm05.stdout:Unpacking python3-sklearn-lib:amd64 (0.23.2-5ubuntu6) ... 2026-03-20T17:56:37.746 INFO:teuthology.orchestra.run.vm05.stdout:Selecting previously unselected package python3-joblib. 2026-03-20T17:56:37.752 INFO:teuthology.orchestra.run.vm05.stdout:Preparing to unpack .../51-python3-joblib_0.17.0-4ubuntu1_all.deb ... 2026-03-20T17:56:37.753 INFO:teuthology.orchestra.run.vm05.stdout:Unpacking python3-joblib (0.17.0-4ubuntu1) ... 2026-03-20T17:56:37.761 INFO:teuthology.orchestra.run.vm06.stdout:Created symlink /etc/systemd/system/smartd.service → /lib/systemd/system/smartmontools.service. 2026-03-20T17:56:37.761 INFO:teuthology.orchestra.run.vm06.stdout:Created symlink /etc/systemd/system/multi-user.target.wants/smartmontools.service → /lib/systemd/system/smartmontools.service. 2026-03-20T17:56:37.791 INFO:teuthology.orchestra.run.vm05.stdout:Selecting previously unselected package python3-threadpoolctl. 2026-03-20T17:56:37.797 INFO:teuthology.orchestra.run.vm05.stdout:Preparing to unpack .../52-python3-threadpoolctl_3.1.0-1_all.deb ... 2026-03-20T17:56:37.799 INFO:teuthology.orchestra.run.vm05.stdout:Unpacking python3-threadpoolctl (3.1.0-1) ... 2026-03-20T17:56:37.816 INFO:teuthology.orchestra.run.vm05.stdout:Selecting previously unselected package python3-sklearn. 2026-03-20T17:56:37.823 INFO:teuthology.orchestra.run.vm05.stdout:Preparing to unpack .../53-python3-sklearn_0.23.2-5ubuntu6_all.deb ... 2026-03-20T17:56:37.824 INFO:teuthology.orchestra.run.vm05.stdout:Unpacking python3-sklearn (0.23.2-5ubuntu6) ... 2026-03-20T17:56:37.953 INFO:teuthology.orchestra.run.vm05.stdout:Selecting previously unselected package ceph-mgr-diskprediction-local. 2026-03-20T17:56:37.959 INFO:teuthology.orchestra.run.vm05.stdout:Preparing to unpack .../54-ceph-mgr-diskprediction-local_20.2.0-712-g70f8415b-1jammy_all.deb ... 2026-03-20T17:56:37.960 INFO:teuthology.orchestra.run.vm05.stdout:Unpacking ceph-mgr-diskprediction-local (20.2.0-712-g70f8415b-1jammy) ... 2026-03-20T17:56:38.206 INFO:teuthology.orchestra.run.vm06.stdout:Setting up python3-iniconfig (1.1.1-2) ... 2026-03-20T17:56:38.228 INFO:teuthology.orchestra.run.vm05.stdout:Selecting previously unselected package python3-cachetools. 2026-03-20T17:56:38.234 INFO:teuthology.orchestra.run.vm05.stdout:Preparing to unpack .../55-python3-cachetools_5.0.0-1_all.deb ... 2026-03-20T17:56:38.235 INFO:teuthology.orchestra.run.vm05.stdout:Unpacking python3-cachetools (5.0.0-1) ... 2026-03-20T17:56:38.307 INFO:teuthology.orchestra.run.vm06.stdout:Setting up libdouble-conversion3:amd64 (3.1.7-4) ... 2026-03-20T17:56:38.516 INFO:teuthology.orchestra.run.vm06.stdout:Setting up nvme-cli (1.16-3ubuntu0.3) ... 2026-03-20T17:56:38.650 INFO:teuthology.orchestra.run.vm05.stdout:Selecting previously unselected package python3-rsa. 2026-03-20T17:56:38.657 INFO:teuthology.orchestra.run.vm05.stdout:Preparing to unpack .../56-python3-rsa_4.8-1_all.deb ... 2026-03-20T17:56:38.667 INFO:teuthology.orchestra.run.vm05.stdout:Unpacking python3-rsa (4.8-1) ... 2026-03-20T17:56:38.686 INFO:teuthology.orchestra.run.vm08.stdout:Selecting previously unselected package ceph-volume. 2026-03-20T17:56:38.686 INFO:teuthology.orchestra.run.vm05.stdout:Selecting previously unselected package python3-google-auth. 2026-03-20T17:56:38.691 INFO:teuthology.orchestra.run.vm08.stdout:Preparing to unpack .../68-ceph-volume_20.2.0-712-g70f8415b-1jammy_all.deb ... 2026-03-20T17:56:38.692 INFO:teuthology.orchestra.run.vm08.stdout:Unpacking ceph-volume (20.2.0-712-g70f8415b-1jammy) ... 2026-03-20T17:56:38.693 INFO:teuthology.orchestra.run.vm05.stdout:Preparing to unpack .../57-python3-google-auth_1.5.1-3_all.deb ... 2026-03-20T17:56:38.694 INFO:teuthology.orchestra.run.vm05.stdout:Unpacking python3-google-auth (1.5.1-3) ... 2026-03-20T17:56:38.714 INFO:teuthology.orchestra.run.vm05.stdout:Selecting previously unselected package python3-requests-oauthlib. 2026-03-20T17:56:38.719 INFO:teuthology.orchestra.run.vm06.stdout:Created symlink /etc/systemd/system/default.target.wants/nvmefc-boot-connections.service → /lib/systemd/system/nvmefc-boot-connections.service. 2026-03-20T17:56:38.721 INFO:teuthology.orchestra.run.vm05.stdout:Preparing to unpack .../58-python3-requests-oauthlib_1.3.0+ds-0.1_all.deb ... 2026-03-20T17:56:38.721 INFO:teuthology.orchestra.run.vm05.stdout:Unpacking python3-requests-oauthlib (1.3.0+ds-0.1) ... 2026-03-20T17:56:38.722 INFO:teuthology.orchestra.run.vm08.stdout:Selecting previously unselected package libcephfs-daemon. 2026-03-20T17:56:38.727 INFO:teuthology.orchestra.run.vm08.stdout:Preparing to unpack .../69-libcephfs-daemon_20.2.0-712-g70f8415b-1jammy_amd64.deb ... 2026-03-20T17:56:38.728 INFO:teuthology.orchestra.run.vm08.stdout:Unpacking libcephfs-daemon (20.2.0-712-g70f8415b-1jammy) ... 2026-03-20T17:56:38.740 INFO:teuthology.orchestra.run.vm05.stdout:Selecting previously unselected package python3-websocket. 2026-03-20T17:56:38.746 INFO:teuthology.orchestra.run.vm08.stdout:Selecting previously unselected package libcephfs-proxy2. 2026-03-20T17:56:38.747 INFO:teuthology.orchestra.run.vm05.stdout:Preparing to unpack .../59-python3-websocket_1.2.3-1_all.deb ... 2026-03-20T17:56:38.748 INFO:teuthology.orchestra.run.vm05.stdout:Unpacking python3-websocket (1.2.3-1) ... 2026-03-20T17:56:38.750 INFO:teuthology.orchestra.run.vm08.stdout:Preparing to unpack .../70-libcephfs-proxy2_20.2.0-712-g70f8415b-1jammy_amd64.deb ... 2026-03-20T17:56:38.751 INFO:teuthology.orchestra.run.vm08.stdout:Unpacking libcephfs-proxy2 (20.2.0-712-g70f8415b-1jammy) ... 2026-03-20T17:56:38.766 INFO:teuthology.orchestra.run.vm08.stdout:Selecting previously unselected package libcephfs-dev. 2026-03-20T17:56:38.769 INFO:teuthology.orchestra.run.vm05.stdout:Selecting previously unselected package python3-kubernetes. 2026-03-20T17:56:38.771 INFO:teuthology.orchestra.run.vm08.stdout:Preparing to unpack .../71-libcephfs-dev_20.2.0-712-g70f8415b-1jammy_amd64.deb ... 2026-03-20T17:56:38.772 INFO:teuthology.orchestra.run.vm08.stdout:Unpacking libcephfs-dev (20.2.0-712-g70f8415b-1jammy) ... 2026-03-20T17:56:38.776 INFO:teuthology.orchestra.run.vm05.stdout:Preparing to unpack .../60-python3-kubernetes_12.0.1-1ubuntu1_all.deb ... 2026-03-20T17:56:38.777 INFO:teuthology.orchestra.run.vm05.stdout:Unpacking python3-kubernetes (12.0.1-1ubuntu1) ... 2026-03-20T17:56:38.794 INFO:teuthology.orchestra.run.vm08.stdout:Selecting previously unselected package nvme-cli. 2026-03-20T17:56:38.800 INFO:teuthology.orchestra.run.vm08.stdout:Preparing to unpack .../72-nvme-cli_1.16-3ubuntu0.3_amd64.deb ... 2026-03-20T17:56:38.801 INFO:teuthology.orchestra.run.vm08.stdout:Unpacking nvme-cli (1.16-3ubuntu0.3) ... 2026-03-20T17:56:38.824 INFO:teuthology.orchestra.run.vm02.stdout:Get:87 https://1.chacra.ceph.com/r/ceph/tentacle/70f8415b300f041766fa27faf7d5472699e32388/ubuntu/jammy/flavors/default jammy/main amd64 rbd-fuse amd64 20.2.0-712-g70f8415b-1jammy [97.9 kB] 2026-03-20T17:56:38.842 INFO:teuthology.orchestra.run.vm08.stdout:Selecting previously unselected package python-asyncssh-doc. 2026-03-20T17:56:38.850 INFO:teuthology.orchestra.run.vm08.stdout:Preparing to unpack .../73-python-asyncssh-doc_2.5.0-1ubuntu0.1_all.deb ... 2026-03-20T17:56:38.851 INFO:teuthology.orchestra.run.vm08.stdout:Unpacking python-asyncssh-doc (2.5.0-1ubuntu0.1) ... 2026-03-20T17:56:38.906 INFO:teuthology.orchestra.run.vm08.stdout:Selecting previously unselected package python3-iniconfig. 2026-03-20T17:56:38.908 INFO:teuthology.orchestra.run.vm08.stdout:Preparing to unpack .../74-python3-iniconfig_1.1.1-2_all.deb ... 2026-03-20T17:56:38.909 INFO:teuthology.orchestra.run.vm08.stdout:Unpacking python3-iniconfig (1.1.1-2) ... 2026-03-20T17:56:38.926 INFO:teuthology.orchestra.run.vm05.stdout:Selecting previously unselected package ceph-mgr-k8sevents. 2026-03-20T17:56:38.930 INFO:teuthology.orchestra.run.vm08.stdout:Selecting previously unselected package python3-pluggy. 2026-03-20T17:56:38.932 INFO:teuthology.orchestra.run.vm08.stdout:Preparing to unpack .../75-python3-pluggy_0.13.0-7.1_all.deb ... 2026-03-20T17:56:38.933 INFO:teuthology.orchestra.run.vm08.stdout:Unpacking python3-pluggy (0.13.0-7.1) ... 2026-03-20T17:56:38.933 INFO:teuthology.orchestra.run.vm05.stdout:Preparing to unpack .../61-ceph-mgr-k8sevents_20.2.0-712-g70f8415b-1jammy_all.deb ... 2026-03-20T17:56:38.934 INFO:teuthology.orchestra.run.vm05.stdout:Unpacking ceph-mgr-k8sevents (20.2.0-712-g70f8415b-1jammy) ... 2026-03-20T17:56:38.951 INFO:teuthology.orchestra.run.vm05.stdout:Selecting previously unselected package libonig5:amd64. 2026-03-20T17:56:38.954 INFO:teuthology.orchestra.run.vm08.stdout:Selecting previously unselected package python3-psutil. 2026-03-20T17:56:38.958 INFO:teuthology.orchestra.run.vm08.stdout:Preparing to unpack .../76-python3-psutil_5.9.0-1build1_amd64.deb ... 2026-03-20T17:56:38.958 INFO:teuthology.orchestra.run.vm05.stdout:Preparing to unpack .../62-libonig5_6.9.7.1-2build1_amd64.deb ... 2026-03-20T17:56:38.958 INFO:teuthology.orchestra.run.vm06.stdout:Created symlink /etc/systemd/system/default.target.wants/nvmf-autoconnect.service → /lib/systemd/system/nvmf-autoconnect.service. 2026-03-20T17:56:38.959 INFO:teuthology.orchestra.run.vm08.stdout:Unpacking python3-psutil (5.9.0-1build1) ... 2026-03-20T17:56:38.959 INFO:teuthology.orchestra.run.vm05.stdout:Unpacking libonig5:amd64 (6.9.7.1-2build1) ... 2026-03-20T17:56:38.977 INFO:teuthology.orchestra.run.vm05.stdout:Selecting previously unselected package libjq1:amd64. 2026-03-20T17:56:38.982 INFO:teuthology.orchestra.run.vm08.stdout:Selecting previously unselected package python3-py. 2026-03-20T17:56:38.984 INFO:teuthology.orchestra.run.vm05.stdout:Preparing to unpack .../63-libjq1_1.6-2.1ubuntu3.1_amd64.deb ... 2026-03-20T17:56:38.985 INFO:teuthology.orchestra.run.vm08.stdout:Preparing to unpack .../77-python3-py_1.10.0-1_all.deb ... 2026-03-20T17:56:38.985 INFO:teuthology.orchestra.run.vm05.stdout:Unpacking libjq1:amd64 (1.6-2.1ubuntu3.1) ... 2026-03-20T17:56:38.986 INFO:teuthology.orchestra.run.vm08.stdout:Unpacking python3-py (1.10.0-1) ... 2026-03-20T17:56:39.003 INFO:teuthology.orchestra.run.vm05.stdout:Selecting previously unselected package jq. 2026-03-20T17:56:39.010 INFO:teuthology.orchestra.run.vm08.stdout:Selecting previously unselected package python3-pygments. 2026-03-20T17:56:39.011 INFO:teuthology.orchestra.run.vm05.stdout:Preparing to unpack .../64-jq_1.6-2.1ubuntu3.1_amd64.deb ... 2026-03-20T17:56:39.012 INFO:teuthology.orchestra.run.vm05.stdout:Unpacking jq (1.6-2.1ubuntu3.1) ... 2026-03-20T17:56:39.014 INFO:teuthology.orchestra.run.vm08.stdout:Preparing to unpack .../78-python3-pygments_2.11.2+dfsg-2ubuntu0.1_all.deb ... 2026-03-20T17:56:39.015 INFO:teuthology.orchestra.run.vm08.stdout:Unpacking python3-pygments (2.11.2+dfsg-2ubuntu0.1) ... 2026-03-20T17:56:39.029 INFO:teuthology.orchestra.run.vm05.stdout:Selecting previously unselected package socat. 2026-03-20T17:56:39.034 INFO:teuthology.orchestra.run.vm05.stdout:Preparing to unpack .../65-socat_1.7.4.1-3ubuntu4_amd64.deb ... 2026-03-20T17:56:39.035 INFO:teuthology.orchestra.run.vm05.stdout:Unpacking socat (1.7.4.1-3ubuntu4) ... 2026-03-20T17:56:39.062 INFO:teuthology.orchestra.run.vm05.stdout:Selecting previously unselected package xmlstarlet. 2026-03-20T17:56:39.068 INFO:teuthology.orchestra.run.vm05.stdout:Preparing to unpack .../66-xmlstarlet_1.6.1-2.1_amd64.deb ... 2026-03-20T17:56:39.069 INFO:teuthology.orchestra.run.vm05.stdout:Unpacking xmlstarlet (1.6.1-2.1) ... 2026-03-20T17:56:39.078 INFO:teuthology.orchestra.run.vm08.stdout:Selecting previously unselected package python3-toml. 2026-03-20T17:56:39.084 INFO:teuthology.orchestra.run.vm08.stdout:Preparing to unpack .../79-python3-toml_0.10.2-1_all.deb ... 2026-03-20T17:56:39.085 INFO:teuthology.orchestra.run.vm08.stdout:Unpacking python3-toml (0.10.2-1) ... 2026-03-20T17:56:39.102 INFO:teuthology.orchestra.run.vm08.stdout:Selecting previously unselected package python3-pytest. 2026-03-20T17:56:39.104 INFO:teuthology.orchestra.run.vm02.stdout:Fetched 281 MB in 24s (11.6 MB/s) 2026-03-20T17:56:39.106 INFO:teuthology.orchestra.run.vm08.stdout:Preparing to unpack .../80-python3-pytest_6.2.5-1ubuntu2_all.deb ... 2026-03-20T17:56:39.107 INFO:teuthology.orchestra.run.vm08.stdout:Unpacking python3-pytest (6.2.5-1ubuntu2) ... 2026-03-20T17:56:39.114 INFO:teuthology.orchestra.run.vm05.stdout:Selecting previously unselected package ceph-test. 2026-03-20T17:56:39.119 INFO:teuthology.orchestra.run.vm05.stdout:Preparing to unpack .../67-ceph-test_20.2.0-712-g70f8415b-1jammy_amd64.deb ... 2026-03-20T17:56:39.204 INFO:teuthology.orchestra.run.vm05.stdout:Unpacking ceph-test (20.2.0-712-g70f8415b-1jammy) ... 2026-03-20T17:56:39.252 INFO:teuthology.orchestra.run.vm02.stdout:Selecting previously unselected package liblttng-ust1:amd64. 2026-03-20T17:56:39.274 INFO:teuthology.orchestra.run.vm08.stdout:Selecting previously unselected package python3-simplejson. 2026-03-20T17:56:39.278 INFO:teuthology.orchestra.run.vm08.stdout:Preparing to unpack .../81-python3-simplejson_3.17.6-1build1_amd64.deb ... 2026-03-20T17:56:39.279 INFO:teuthology.orchestra.run.vm08.stdout:Unpacking python3-simplejson (3.17.6-1build1) ... 2026-03-20T17:56:39.286 INFO:teuthology.orchestra.run.vm02.stdout:(Reading database ... (Reading database ... 5% (Reading database ... 10% (Reading database ... 15% (Reading database ... 20% (Reading database ... 25% (Reading database ... 30% (Reading database ... 35% (Reading database ... 40% (Reading database ... 45% (Reading database ... 50% (Reading database ... 55% (Reading database ... 60% (Reading database ... 65% (Reading database ... 70% (Reading database ... 75% (Reading database ... 80% (Reading database ... 85% (Reading database ... 90% (Reading database ... 95% (Reading database ... 100% (Reading database ... 119262 files and directories currently installed.) 2026-03-20T17:56:39.288 INFO:teuthology.orchestra.run.vm02.stdout:Preparing to unpack .../00-liblttng-ust1_2.13.1-1ubuntu1_amd64.deb ... 2026-03-20T17:56:39.290 INFO:teuthology.orchestra.run.vm02.stdout:Unpacking liblttng-ust1:amd64 (2.13.1-1ubuntu1) ... 2026-03-20T17:56:39.302 INFO:teuthology.orchestra.run.vm08.stdout:Selecting previously unselected package python3-webob. 2026-03-20T17:56:39.307 INFO:teuthology.orchestra.run.vm08.stdout:Preparing to unpack .../82-python3-webob_1%3a1.8.6-1.1ubuntu0.1_all.deb ... 2026-03-20T17:56:39.308 INFO:teuthology.orchestra.run.vm08.stdout:Unpacking python3-webob (1:1.8.6-1.1ubuntu0.1) ... 2026-03-20T17:56:39.311 INFO:teuthology.orchestra.run.vm02.stdout:Selecting previously unselected package libdouble-conversion3:amd64. 2026-03-20T17:56:39.317 INFO:teuthology.orchestra.run.vm02.stdout:Preparing to unpack .../01-libdouble-conversion3_3.1.7-4_amd64.deb ... 2026-03-20T17:56:39.318 INFO:teuthology.orchestra.run.vm02.stdout:Unpacking libdouble-conversion3:amd64 (3.1.7-4) ... 2026-03-20T17:56:39.330 INFO:teuthology.orchestra.run.vm08.stdout:Selecting previously unselected package qttranslations5-l10n. 2026-03-20T17:56:39.332 INFO:teuthology.orchestra.run.vm08.stdout:Preparing to unpack .../83-qttranslations5-l10n_5.15.3-1_all.deb ... 2026-03-20T17:56:39.333 INFO:teuthology.orchestra.run.vm08.stdout:Unpacking qttranslations5-l10n (5.15.3-1) ... 2026-03-20T17:56:39.333 INFO:teuthology.orchestra.run.vm02.stdout:Selecting previously unselected package libpcre2-16-0:amd64. 2026-03-20T17:56:39.339 INFO:teuthology.orchestra.run.vm02.stdout:Preparing to unpack .../02-libpcre2-16-0_10.39-3ubuntu0.1_amd64.deb ... 2026-03-20T17:56:39.340 INFO:teuthology.orchestra.run.vm02.stdout:Unpacking libpcre2-16-0:amd64 (10.39-3ubuntu0.1) ... 2026-03-20T17:56:39.361 INFO:teuthology.orchestra.run.vm02.stdout:Selecting previously unselected package libqt5core5a:amd64. 2026-03-20T17:56:39.367 INFO:teuthology.orchestra.run.vm02.stdout:Preparing to unpack .../03-libqt5core5a_5.15.3+dfsg-2ubuntu0.2_amd64.deb ... 2026-03-20T17:56:39.371 INFO:teuthology.orchestra.run.vm02.stdout:Unpacking libqt5core5a:amd64 (5.15.3+dfsg-2ubuntu0.2) ... 2026-03-20T17:56:39.374 INFO:teuthology.orchestra.run.vm06.stdout:nvmf-connect.target is a disabled or a static unit, not starting it. 2026-03-20T17:56:39.417 INFO:teuthology.orchestra.run.vm06.stdout:Setting up cephadm (20.2.0-712-g70f8415b-1jammy) ... 2026-03-20T17:56:39.431 INFO:teuthology.orchestra.run.vm02.stdout:Selecting previously unselected package libqt5dbus5:amd64. 2026-03-20T17:56:39.437 INFO:teuthology.orchestra.run.vm02.stdout:Preparing to unpack .../04-libqt5dbus5_5.15.3+dfsg-2ubuntu0.2_amd64.deb ... 2026-03-20T17:56:39.438 INFO:teuthology.orchestra.run.vm02.stdout:Unpacking libqt5dbus5:amd64 (5.15.3+dfsg-2ubuntu0.2) ... 2026-03-20T17:56:39.445 INFO:teuthology.orchestra.run.vm08.stdout:Selecting previously unselected package radosgw. 2026-03-20T17:56:39.451 INFO:teuthology.orchestra.run.vm08.stdout:Preparing to unpack .../84-radosgw_20.2.0-712-g70f8415b-1jammy_amd64.deb ... 2026-03-20T17:56:39.452 INFO:teuthology.orchestra.run.vm08.stdout:Unpacking radosgw (20.2.0-712-g70f8415b-1jammy) ... 2026-03-20T17:56:39.456 INFO:teuthology.orchestra.run.vm02.stdout:Selecting previously unselected package libqt5network5:amd64. 2026-03-20T17:56:39.461 INFO:teuthology.orchestra.run.vm06.stdout:Adding system user cephadm....done 2026-03-20T17:56:39.462 INFO:teuthology.orchestra.run.vm02.stdout:Preparing to unpack .../05-libqt5network5_5.15.3+dfsg-2ubuntu0.2_amd64.deb ... 2026-03-20T17:56:39.463 INFO:teuthology.orchestra.run.vm02.stdout:Unpacking libqt5network5:amd64 (5.15.3+dfsg-2ubuntu0.2) ... 2026-03-20T17:56:39.471 INFO:teuthology.orchestra.run.vm06.stdout:Setting up python3-jaraco.classes (3.2.1-3) ... 2026-03-20T17:56:39.487 INFO:teuthology.orchestra.run.vm02.stdout:Selecting previously unselected package libthrift-0.16.0:amd64. 2026-03-20T17:56:39.493 INFO:teuthology.orchestra.run.vm02.stdout:Preparing to unpack .../06-libthrift-0.16.0_0.16.0-2_amd64.deb ... 2026-03-20T17:56:39.494 INFO:teuthology.orchestra.run.vm02.stdout:Unpacking libthrift-0.16.0:amd64 (0.16.0-2) ... 2026-03-20T17:56:39.523 INFO:teuthology.orchestra.run.vm02.stdout:Preparing to unpack .../07-librbd1_20.2.0-712-g70f8415b-1jammy_amd64.deb ... 2026-03-20T17:56:39.525 INFO:teuthology.orchestra.run.vm02.stdout:Unpacking librbd1 (20.2.0-712-g70f8415b-1jammy) over (17.2.9-0ubuntu0.22.04.2) ... 2026-03-20T17:56:39.542 INFO:teuthology.orchestra.run.vm06.stdout:Setting up python-asyncssh-doc (2.5.0-1ubuntu0.1) ... 2026-03-20T17:56:39.545 INFO:teuthology.orchestra.run.vm06.stdout:Setting up python3-jaraco.functools (3.4.0-2) ... 2026-03-20T17:56:39.616 INFO:teuthology.orchestra.run.vm02.stdout:Preparing to unpack .../08-librados2_20.2.0-712-g70f8415b-1jammy_amd64.deb ... 2026-03-20T17:56:39.618 INFO:teuthology.orchestra.run.vm02.stdout:Unpacking librados2 (20.2.0-712-g70f8415b-1jammy) over (17.2.9-0ubuntu0.22.04.2) ... 2026-03-20T17:56:39.620 INFO:teuthology.orchestra.run.vm06.stdout:Setting up python3-repoze.lru (0.7-2) ... 2026-03-20T17:56:39.688 INFO:teuthology.orchestra.run.vm02.stdout:Selecting previously unselected package libnbd0. 2026-03-20T17:56:39.692 INFO:teuthology.orchestra.run.vm06.stdout:Setting up liboath0:amd64 (2.6.7-3ubuntu0.1) ... 2026-03-20T17:56:39.694 INFO:teuthology.orchestra.run.vm02.stdout:Preparing to unpack .../09-libnbd0_1.10.5-1_amd64.deb ... 2026-03-20T17:56:39.695 INFO:teuthology.orchestra.run.vm06.stdout:Setting up python3-py (1.10.0-1) ... 2026-03-20T17:56:39.695 INFO:teuthology.orchestra.run.vm02.stdout:Unpacking libnbd0 (1.10.5-1) ... 2026-03-20T17:56:39.805 INFO:teuthology.orchestra.run.vm06.stdout:Setting up python3-joblib (0.17.0-4ubuntu1) ... 2026-03-20T17:56:39.813 INFO:teuthology.orchestra.run.vm02.stdout:Selecting previously unselected package libcephfs2. 2026-03-20T17:56:39.818 INFO:teuthology.orchestra.run.vm08.stdout:Selecting previously unselected package rbd-fuse. 2026-03-20T17:56:39.819 INFO:teuthology.orchestra.run.vm02.stdout:Preparing to unpack .../10-libcephfs2_20.2.0-712-g70f8415b-1jammy_amd64.deb ... 2026-03-20T17:56:39.820 INFO:teuthology.orchestra.run.vm02.stdout:Unpacking libcephfs2 (20.2.0-712-g70f8415b-1jammy) ... 2026-03-20T17:56:39.825 INFO:teuthology.orchestra.run.vm08.stdout:Preparing to unpack .../85-rbd-fuse_20.2.0-712-g70f8415b-1jammy_amd64.deb ... 2026-03-20T17:56:39.826 INFO:teuthology.orchestra.run.vm08.stdout:Unpacking rbd-fuse (20.2.0-712-g70f8415b-1jammy) ... 2026-03-20T17:56:39.846 INFO:teuthology.orchestra.run.vm02.stdout:Selecting previously unselected package python3-rados. 2026-03-20T17:56:39.848 INFO:teuthology.orchestra.run.vm08.stdout:Selecting previously unselected package smartmontools. 2026-03-20T17:56:39.852 INFO:teuthology.orchestra.run.vm02.stdout:Preparing to unpack .../11-python3-rados_20.2.0-712-g70f8415b-1jammy_amd64.deb ... 2026-03-20T17:56:39.853 INFO:teuthology.orchestra.run.vm02.stdout:Unpacking python3-rados (20.2.0-712-g70f8415b-1jammy) ... 2026-03-20T17:56:39.854 INFO:teuthology.orchestra.run.vm08.stdout:Preparing to unpack .../86-smartmontools_7.2-1ubuntu0.1_amd64.deb ... 2026-03-20T17:56:39.863 INFO:teuthology.orchestra.run.vm08.stdout:Unpacking smartmontools (7.2-1ubuntu0.1) ... 2026-03-20T17:56:39.875 INFO:teuthology.orchestra.run.vm02.stdout:Selecting previously unselected package python3-ceph-argparse. 2026-03-20T17:56:39.880 INFO:teuthology.orchestra.run.vm02.stdout:Preparing to unpack .../12-python3-ceph-argparse_20.2.0-712-g70f8415b-1jammy_all.deb ... 2026-03-20T17:56:39.881 INFO:teuthology.orchestra.run.vm02.stdout:Unpacking python3-ceph-argparse (20.2.0-712-g70f8415b-1jammy) ... 2026-03-20T17:56:39.899 INFO:teuthology.orchestra.run.vm02.stdout:Selecting previously unselected package python3-cephfs. 2026-03-20T17:56:39.905 INFO:teuthology.orchestra.run.vm08.stdout:Setting up smartmontools (7.2-1ubuntu0.1) ... 2026-03-20T17:56:39.906 INFO:teuthology.orchestra.run.vm02.stdout:Preparing to unpack .../13-python3-cephfs_20.2.0-712-g70f8415b-1jammy_amd64.deb ... 2026-03-20T17:56:39.907 INFO:teuthology.orchestra.run.vm02.stdout:Unpacking python3-cephfs (20.2.0-712-g70f8415b-1jammy) ... 2026-03-20T17:56:39.929 INFO:teuthology.orchestra.run.vm02.stdout:Selecting previously unselected package python3-ceph-common. 2026-03-20T17:56:39.931 INFO:teuthology.orchestra.run.vm06.stdout:Setting up python3-cachetools (5.0.0-1) ... 2026-03-20T17:56:39.934 INFO:teuthology.orchestra.run.vm02.stdout:Preparing to unpack .../14-python3-ceph-common_20.2.0-712-g70f8415b-1jammy_all.deb ... 2026-03-20T17:56:39.935 INFO:teuthology.orchestra.run.vm02.stdout:Unpacking python3-ceph-common (20.2.0-712-g70f8415b-1jammy) ... 2026-03-20T17:56:39.957 INFO:teuthology.orchestra.run.vm02.stdout:Selecting previously unselected package python3-wcwidth. 2026-03-20T17:56:39.963 INFO:teuthology.orchestra.run.vm02.stdout:Preparing to unpack .../15-python3-wcwidth_0.2.5+dfsg1-1_all.deb ... 2026-03-20T17:56:39.964 INFO:teuthology.orchestra.run.vm02.stdout:Unpacking python3-wcwidth (0.2.5+dfsg1-1) ... 2026-03-20T17:56:40.048 INFO:teuthology.orchestra.run.vm06.stdout:Setting up python3-threadpoolctl (3.1.0-1) ... 2026-03-20T17:56:40.172 INFO:teuthology.orchestra.run.vm08.stdout:Created symlink /etc/systemd/system/smartd.service → /lib/systemd/system/smartmontools.service. 2026-03-20T17:56:40.172 INFO:teuthology.orchestra.run.vm08.stdout:Created symlink /etc/systemd/system/multi-user.target.wants/smartmontools.service → /lib/systemd/system/smartmontools.service. 2026-03-20T17:56:40.275 INFO:teuthology.orchestra.run.vm02.stdout:Selecting previously unselected package python3-prettytable. 2026-03-20T17:56:40.281 INFO:teuthology.orchestra.run.vm02.stdout:Preparing to unpack .../16-python3-prettytable_2.5.0-2_all.deb ... 2026-03-20T17:56:40.295 INFO:teuthology.orchestra.run.vm06.stdout:Setting up python3-ceph-argparse (20.2.0-712-g70f8415b-1jammy) ... 2026-03-20T17:56:40.298 INFO:teuthology.orchestra.run.vm02.stdout:Unpacking python3-prettytable (2.5.0-2) ... 2026-03-20T17:56:40.316 INFO:teuthology.orchestra.run.vm02.stdout:Selecting previously unselected package python3-rbd. 2026-03-20T17:56:40.322 INFO:teuthology.orchestra.run.vm02.stdout:Preparing to unpack .../17-python3-rbd_20.2.0-712-g70f8415b-1jammy_amd64.deb ... 2026-03-20T17:56:40.323 INFO:teuthology.orchestra.run.vm02.stdout:Unpacking python3-rbd (20.2.0-712-g70f8415b-1jammy) ... 2026-03-20T17:56:40.432 INFO:teuthology.orchestra.run.vm06.stdout:Setting up python3-sklearn-lib:amd64 (0.23.2-5ubuntu6) ... 2026-03-20T17:56:40.434 INFO:teuthology.orchestra.run.vm02.stdout:Selecting previously unselected package librdkafka1:amd64. 2026-03-20T17:56:40.437 INFO:teuthology.orchestra.run.vm06.stdout:Setting up libnbd0 (1.10.5-1) ... 2026-03-20T17:56:40.440 INFO:teuthology.orchestra.run.vm06.stdout:Setting up libfuse2:amd64 (2.9.9-5ubuntu3) ... 2026-03-20T17:56:40.441 INFO:teuthology.orchestra.run.vm02.stdout:Preparing to unpack .../18-librdkafka1_1.8.0-1build1_amd64.deb ... 2026-03-20T17:56:40.441 INFO:teuthology.orchestra.run.vm02.stdout:Unpacking librdkafka1:amd64 (1.8.0-1build1) ... 2026-03-20T17:56:40.443 INFO:teuthology.orchestra.run.vm06.stdout:Setting up libpcre2-16-0:amd64 (10.39-3ubuntu0.1) ... 2026-03-20T17:56:40.445 INFO:teuthology.orchestra.run.vm06.stdout:Setting up python3-psutil (5.9.0-1build1) ... 2026-03-20T17:56:40.457 INFO:teuthology.orchestra.run.vm05.stdout:Selecting previously unselected package ceph-volume. 2026-03-20T17:56:40.464 INFO:teuthology.orchestra.run.vm05.stdout:Preparing to unpack .../68-ceph-volume_20.2.0-712-g70f8415b-1jammy_all.deb ... 2026-03-20T17:56:40.464 INFO:teuthology.orchestra.run.vm05.stdout:Unpacking ceph-volume (20.2.0-712-g70f8415b-1jammy) ... 2026-03-20T17:56:40.466 INFO:teuthology.orchestra.run.vm02.stdout:Selecting previously unselected package librgw2. 2026-03-20T17:56:40.471 INFO:teuthology.orchestra.run.vm02.stdout:Preparing to unpack .../19-librgw2_20.2.0-712-g70f8415b-1jammy_amd64.deb ... 2026-03-20T17:56:40.472 INFO:teuthology.orchestra.run.vm02.stdout:Unpacking librgw2 (20.2.0-712-g70f8415b-1jammy) ... 2026-03-20T17:56:40.492 INFO:teuthology.orchestra.run.vm05.stdout:Selecting previously unselected package libcephfs-daemon. 2026-03-20T17:56:40.498 INFO:teuthology.orchestra.run.vm05.stdout:Preparing to unpack .../69-libcephfs-daemon_20.2.0-712-g70f8415b-1jammy_amd64.deb ... 2026-03-20T17:56:40.499 INFO:teuthology.orchestra.run.vm05.stdout:Unpacking libcephfs-daemon (20.2.0-712-g70f8415b-1jammy) ... 2026-03-20T17:56:40.516 INFO:teuthology.orchestra.run.vm05.stdout:Selecting previously unselected package libcephfs-proxy2. 2026-03-20T17:56:40.520 INFO:teuthology.orchestra.run.vm08.stdout:Setting up python3-iniconfig (1.1.1-2) ... 2026-03-20T17:56:40.521 INFO:teuthology.orchestra.run.vm05.stdout:Preparing to unpack .../70-libcephfs-proxy2_20.2.0-712-g70f8415b-1jammy_amd64.deb ... 2026-03-20T17:56:40.522 INFO:teuthology.orchestra.run.vm05.stdout:Unpacking libcephfs-proxy2 (20.2.0-712-g70f8415b-1jammy) ... 2026-03-20T17:56:40.537 INFO:teuthology.orchestra.run.vm05.stdout:Selecting previously unselected package libcephfs-dev. 2026-03-20T17:56:40.543 INFO:teuthology.orchestra.run.vm05.stdout:Preparing to unpack .../71-libcephfs-dev_20.2.0-712-g70f8415b-1jammy_amd64.deb ... 2026-03-20T17:56:40.544 INFO:teuthology.orchestra.run.vm05.stdout:Unpacking libcephfs-dev (20.2.0-712-g70f8415b-1jammy) ... 2026-03-20T17:56:40.564 INFO:teuthology.orchestra.run.vm05.stdout:Selecting previously unselected package nvme-cli. 2026-03-20T17:56:40.569 INFO:teuthology.orchestra.run.vm05.stdout:Preparing to unpack .../72-nvme-cli_1.16-3ubuntu0.3_amd64.deb ... 2026-03-20T17:56:40.603 INFO:teuthology.orchestra.run.vm05.stdout:Unpacking nvme-cli (1.16-3ubuntu0.3) ... 2026-03-20T17:56:40.603 INFO:teuthology.orchestra.run.vm08.stdout:Setting up libdouble-conversion3:amd64 (3.1.7-4) ... 2026-03-20T17:56:40.604 INFO:teuthology.orchestra.run.vm06.stdout:Setting up python3-natsort (8.0.2-1) ... 2026-03-20T17:56:40.606 INFO:teuthology.orchestra.run.vm08.stdout:Setting up nvme-cli (1.16-3ubuntu0.3) ... 2026-03-20T17:56:40.616 INFO:teuthology.orchestra.run.vm02.stdout:Selecting previously unselected package python3-rgw. 2026-03-20T17:56:40.622 INFO:teuthology.orchestra.run.vm02.stdout:Preparing to unpack .../20-python3-rgw_20.2.0-712-g70f8415b-1jammy_amd64.deb ... 2026-03-20T17:56:40.623 INFO:teuthology.orchestra.run.vm02.stdout:Unpacking python3-rgw (20.2.0-712-g70f8415b-1jammy) ... 2026-03-20T17:56:40.639 INFO:teuthology.orchestra.run.vm05.stdout:Selecting previously unselected package python-asyncssh-doc. 2026-03-20T17:56:40.640 INFO:teuthology.orchestra.run.vm02.stdout:Selecting previously unselected package liboath0:amd64. 2026-03-20T17:56:40.646 INFO:teuthology.orchestra.run.vm05.stdout:Preparing to unpack .../73-python-asyncssh-doc_2.5.0-1ubuntu0.1_all.deb ... 2026-03-20T17:56:40.647 INFO:teuthology.orchestra.run.vm02.stdout:Preparing to unpack .../21-liboath0_2.6.7-3ubuntu0.1_amd64.deb ... 2026-03-20T17:56:40.647 INFO:teuthology.orchestra.run.vm05.stdout:Unpacking python-asyncssh-doc (2.5.0-1ubuntu0.1) ... 2026-03-20T17:56:40.648 INFO:teuthology.orchestra.run.vm02.stdout:Unpacking liboath0:amd64 (2.6.7-3ubuntu0.1) ... 2026-03-20T17:56:40.664 INFO:teuthology.orchestra.run.vm02.stdout:Selecting previously unselected package libradosstriper1. 2026-03-20T17:56:40.671 INFO:teuthology.orchestra.run.vm02.stdout:Preparing to unpack .../22-libradosstriper1_20.2.0-712-g70f8415b-1jammy_amd64.deb ... 2026-03-20T17:56:40.673 INFO:teuthology.orchestra.run.vm02.stdout:Unpacking libradosstriper1 (20.2.0-712-g70f8415b-1jammy) ... 2026-03-20T17:56:40.677 INFO:teuthology.orchestra.run.vm08.stdout:Created symlink /etc/systemd/system/default.target.wants/nvmefc-boot-connections.service → /lib/systemd/system/nvmefc-boot-connections.service. 2026-03-20T17:56:40.683 INFO:teuthology.orchestra.run.vm06.stdout:Setting up libcephfs-proxy2 (20.2.0-712-g70f8415b-1jammy) ... 2026-03-20T17:56:40.686 INFO:teuthology.orchestra.run.vm06.stdout:Setting up python3-routes (2.5.1-1ubuntu1) ... 2026-03-20T17:56:40.690 INFO:teuthology.orchestra.run.vm05.stdout:Selecting previously unselected package python3-iniconfig. 2026-03-20T17:56:40.696 INFO:teuthology.orchestra.run.vm05.stdout:Preparing to unpack .../74-python3-iniconfig_1.1.1-2_all.deb ... 2026-03-20T17:56:40.696 INFO:teuthology.orchestra.run.vm02.stdout:Selecting previously unselected package ceph-common. 2026-03-20T17:56:40.697 INFO:teuthology.orchestra.run.vm05.stdout:Unpacking python3-iniconfig (1.1.1-2) ... 2026-03-20T17:56:40.703 INFO:teuthology.orchestra.run.vm02.stdout:Preparing to unpack .../23-ceph-common_20.2.0-712-g70f8415b-1jammy_amd64.deb ... 2026-03-20T17:56:40.704 INFO:teuthology.orchestra.run.vm02.stdout:Unpacking ceph-common (20.2.0-712-g70f8415b-1jammy) ... 2026-03-20T17:56:40.712 INFO:teuthology.orchestra.run.vm05.stdout:Selecting previously unselected package python3-pluggy. 2026-03-20T17:56:40.719 INFO:teuthology.orchestra.run.vm05.stdout:Preparing to unpack .../75-python3-pluggy_0.13.0-7.1_all.deb ... 2026-03-20T17:56:40.719 INFO:teuthology.orchestra.run.vm05.stdout:Unpacking python3-pluggy (0.13.0-7.1) ... 2026-03-20T17:56:40.735 INFO:teuthology.orchestra.run.vm05.stdout:Selecting previously unselected package python3-psutil. 2026-03-20T17:56:40.742 INFO:teuthology.orchestra.run.vm05.stdout:Preparing to unpack .../76-python3-psutil_5.9.0-1build1_amd64.deb ... 2026-03-20T17:56:40.742 INFO:teuthology.orchestra.run.vm05.stdout:Unpacking python3-psutil (5.9.0-1build1) ... 2026-03-20T17:56:40.759 INFO:teuthology.orchestra.run.vm06.stdout:Setting up python3-simplejson (3.17.6-1build1) ... 2026-03-20T17:56:40.765 INFO:teuthology.orchestra.run.vm05.stdout:Selecting previously unselected package python3-py. 2026-03-20T17:56:40.772 INFO:teuthology.orchestra.run.vm05.stdout:Preparing to unpack .../77-python3-py_1.10.0-1_all.deb ... 2026-03-20T17:56:40.773 INFO:teuthology.orchestra.run.vm05.stdout:Unpacking python3-py (1.10.0-1) ... 2026-03-20T17:56:40.798 INFO:teuthology.orchestra.run.vm05.stdout:Selecting previously unselected package python3-pygments. 2026-03-20T17:56:40.804 INFO:teuthology.orchestra.run.vm05.stdout:Preparing to unpack .../78-python3-pygments_2.11.2+dfsg-2ubuntu0.1_all.deb ... 2026-03-20T17:56:40.805 INFO:teuthology.orchestra.run.vm05.stdout:Unpacking python3-pygments (2.11.2+dfsg-2ubuntu0.1) ... 2026-03-20T17:56:40.847 INFO:teuthology.orchestra.run.vm06.stdout:Setting up python3-pygments (2.11.2+dfsg-2ubuntu0.1) ... 2026-03-20T17:56:40.864 INFO:teuthology.orchestra.run.vm05.stdout:Selecting previously unselected package python3-toml. 2026-03-20T17:56:40.871 INFO:teuthology.orchestra.run.vm05.stdout:Preparing to unpack .../79-python3-toml_0.10.2-1_all.deb ... 2026-03-20T17:56:40.871 INFO:teuthology.orchestra.run.vm05.stdout:Unpacking python3-toml (0.10.2-1) ... 2026-03-20T17:56:40.889 INFO:teuthology.orchestra.run.vm05.stdout:Selecting previously unselected package python3-pytest. 2026-03-20T17:56:40.895 INFO:teuthology.orchestra.run.vm05.stdout:Preparing to unpack .../80-python3-pytest_6.2.5-1ubuntu2_all.deb ... 2026-03-20T17:56:40.897 INFO:teuthology.orchestra.run.vm05.stdout:Unpacking python3-pytest (6.2.5-1ubuntu2) ... 2026-03-20T17:56:40.937 INFO:teuthology.orchestra.run.vm08.stdout:Created symlink /etc/systemd/system/default.target.wants/nvmf-autoconnect.service → /lib/systemd/system/nvmf-autoconnect.service. 2026-03-20T17:56:40.940 INFO:teuthology.orchestra.run.vm05.stdout:Selecting previously unselected package python3-simplejson. 2026-03-20T17:56:40.943 INFO:teuthology.orchestra.run.vm05.stdout:Preparing to unpack .../81-python3-simplejson_3.17.6-1build1_amd64.deb ... 2026-03-20T17:56:40.944 INFO:teuthology.orchestra.run.vm05.stdout:Unpacking python3-simplejson (3.17.6-1build1) ... 2026-03-20T17:56:40.968 INFO:teuthology.orchestra.run.vm05.stdout:Selecting previously unselected package python3-webob. 2026-03-20T17:56:40.973 INFO:teuthology.orchestra.run.vm05.stdout:Preparing to unpack .../82-python3-webob_1%3a1.8.6-1.1ubuntu0.1_all.deb ... 2026-03-20T17:56:40.973 INFO:teuthology.orchestra.run.vm05.stdout:Unpacking python3-webob (1:1.8.6-1.1ubuntu0.1) ... 2026-03-20T17:56:40.996 INFO:teuthology.orchestra.run.vm05.stdout:Selecting previously unselected package qttranslations5-l10n. 2026-03-20T17:56:40.998 INFO:teuthology.orchestra.run.vm05.stdout:Preparing to unpack .../83-qttranslations5-l10n_5.15.3-1_all.deb ... 2026-03-20T17:56:40.999 INFO:teuthology.orchestra.run.vm05.stdout:Unpacking qttranslations5-l10n (5.15.3-1) ... 2026-03-20T17:56:41.134 INFO:teuthology.orchestra.run.vm06.stdout:Setting up qttranslations5-l10n (5.15.3-1) ... 2026-03-20T17:56:41.137 INFO:teuthology.orchestra.run.vm06.stdout:Setting up python3-wcwidth (0.2.5+dfsg1-1) ... 2026-03-20T17:56:41.160 INFO:teuthology.orchestra.run.vm05.stdout:Selecting previously unselected package radosgw. 2026-03-20T17:56:41.161 INFO:teuthology.orchestra.run.vm02.stdout:Selecting previously unselected package ceph-base. 2026-03-20T17:56:41.166 INFO:teuthology.orchestra.run.vm05.stdout:Preparing to unpack .../84-radosgw_20.2.0-712-g70f8415b-1jammy_amd64.deb ... 2026-03-20T17:56:41.167 INFO:teuthology.orchestra.run.vm05.stdout:Unpacking radosgw (20.2.0-712-g70f8415b-1jammy) ... 2026-03-20T17:56:41.167 INFO:teuthology.orchestra.run.vm02.stdout:Preparing to unpack .../24-ceph-base_20.2.0-712-g70f8415b-1jammy_amd64.deb ... 2026-03-20T17:56:41.172 INFO:teuthology.orchestra.run.vm02.stdout:Unpacking ceph-base (20.2.0-712-g70f8415b-1jammy) ... 2026-03-20T17:56:41.255 INFO:teuthology.orchestra.run.vm06.stdout:Setting up python3-asyncssh (2.5.0-1ubuntu0.1) ... 2026-03-20T17:56:41.271 INFO:teuthology.orchestra.run.vm02.stdout:Selecting previously unselected package python3-jaraco.functools. 2026-03-20T17:56:41.278 INFO:teuthology.orchestra.run.vm02.stdout:Preparing to unpack .../25-python3-jaraco.functools_3.4.0-2_all.deb ... 2026-03-20T17:56:41.279 INFO:teuthology.orchestra.run.vm02.stdout:Unpacking python3-jaraco.functools (3.4.0-2) ... 2026-03-20T17:56:41.295 INFO:teuthology.orchestra.run.vm08.stdout:nvmf-connect.target is a disabled or a static unit, not starting it. 2026-03-20T17:56:41.296 INFO:teuthology.orchestra.run.vm02.stdout:Selecting previously unselected package python3-cheroot. 2026-03-20T17:56:41.304 INFO:teuthology.orchestra.run.vm02.stdout:Preparing to unpack .../26-python3-cheroot_8.5.2+ds1-1ubuntu3.1_all.deb ... 2026-03-20T17:56:41.305 INFO:teuthology.orchestra.run.vm02.stdout:Unpacking python3-cheroot (8.5.2+ds1-1ubuntu3.1) ... 2026-03-20T17:56:41.310 INFO:teuthology.orchestra.run.vm08.stdout:Setting up cephadm (20.2.0-712-g70f8415b-1jammy) ... 2026-03-20T17:56:41.327 INFO:teuthology.orchestra.run.vm02.stdout:Selecting previously unselected package python3-jaraco.classes. 2026-03-20T17:56:41.334 INFO:teuthology.orchestra.run.vm02.stdout:Preparing to unpack .../27-python3-jaraco.classes_3.2.1-3_all.deb ... 2026-03-20T17:56:41.335 INFO:teuthology.orchestra.run.vm02.stdout:Unpacking python3-jaraco.classes (3.2.1-3) ... 2026-03-20T17:56:41.351 INFO:teuthology.orchestra.run.vm02.stdout:Selecting previously unselected package python3-jaraco.text. 2026-03-20T17:56:41.357 INFO:teuthology.orchestra.run.vm08.stdout:Adding system user cephadm....done 2026-03-20T17:56:41.358 INFO:teuthology.orchestra.run.vm02.stdout:Preparing to unpack .../28-python3-jaraco.text_3.6.0-2_all.deb ... 2026-03-20T17:56:41.359 INFO:teuthology.orchestra.run.vm02.stdout:Unpacking python3-jaraco.text (3.6.0-2) ... 2026-03-20T17:56:41.371 INFO:teuthology.orchestra.run.vm08.stdout:Setting up python3-jaraco.classes (3.2.1-3) ... 2026-03-20T17:56:41.377 INFO:teuthology.orchestra.run.vm02.stdout:Selecting previously unselected package python3-jaraco.collections. 2026-03-20T17:56:41.386 INFO:teuthology.orchestra.run.vm02.stdout:Preparing to unpack .../29-python3-jaraco.collections_3.4.0-2_all.deb ... 2026-03-20T17:56:41.387 INFO:teuthology.orchestra.run.vm02.stdout:Unpacking python3-jaraco.collections (3.4.0-2) ... 2026-03-20T17:56:41.401 INFO:teuthology.orchestra.run.vm06.stdout:Setting up python3-cheroot (8.5.2+ds1-1ubuntu3.1) ... 2026-03-20T17:56:41.403 INFO:teuthology.orchestra.run.vm02.stdout:Selecting previously unselected package python3-tempora. 2026-03-20T17:56:41.410 INFO:teuthology.orchestra.run.vm02.stdout:Preparing to unpack .../30-python3-tempora_4.1.2-1_all.deb ... 2026-03-20T17:56:41.435 INFO:teuthology.orchestra.run.vm02.stdout:Unpacking python3-tempora (4.1.2-1) ... 2026-03-20T17:56:41.511 INFO:teuthology.orchestra.run.vm08.stdout:Setting up python-asyncssh-doc (2.5.0-1ubuntu0.1) ... 2026-03-20T17:56:41.512 INFO:teuthology.orchestra.run.vm06.stdout:Setting up python3-jaraco.text (3.6.0-2) ... 2026-03-20T17:56:41.514 INFO:teuthology.orchestra.run.vm08.stdout:Setting up python3-jaraco.functools (3.4.0-2) ... 2026-03-20T17:56:41.523 INFO:teuthology.orchestra.run.vm02.stdout:Selecting previously unselected package python3-portend. 2026-03-20T17:56:41.524 INFO:teuthology.orchestra.run.vm05.stdout:Selecting previously unselected package rbd-fuse. 2026-03-20T17:56:41.530 INFO:teuthology.orchestra.run.vm05.stdout:Preparing to unpack .../85-rbd-fuse_20.2.0-712-g70f8415b-1jammy_amd64.deb ... 2026-03-20T17:56:41.530 INFO:teuthology.orchestra.run.vm02.stdout:Preparing to unpack .../31-python3-portend_3.0.0-1_all.deb ... 2026-03-20T17:56:41.531 INFO:teuthology.orchestra.run.vm05.stdout:Unpacking rbd-fuse (20.2.0-712-g70f8415b-1jammy) ... 2026-03-20T17:56:41.531 INFO:teuthology.orchestra.run.vm02.stdout:Unpacking python3-portend (3.0.0-1) ... 2026-03-20T17:56:41.547 INFO:teuthology.orchestra.run.vm02.stdout:Selecting previously unselected package python3-zc.lockfile. 2026-03-20T17:56:41.552 INFO:teuthology.orchestra.run.vm05.stdout:Selecting previously unselected package smartmontools. 2026-03-20T17:56:41.554 INFO:teuthology.orchestra.run.vm02.stdout:Preparing to unpack .../32-python3-zc.lockfile_2.0-1_all.deb ... 2026-03-20T17:56:41.555 INFO:teuthology.orchestra.run.vm02.stdout:Unpacking python3-zc.lockfile (2.0-1) ... 2026-03-20T17:56:41.557 INFO:teuthology.orchestra.run.vm05.stdout:Preparing to unpack .../86-smartmontools_7.2-1ubuntu0.1_amd64.deb ... 2026-03-20T17:56:41.565 INFO:teuthology.orchestra.run.vm05.stdout:Unpacking smartmontools (7.2-1ubuntu0.1) ... 2026-03-20T17:56:41.571 INFO:teuthology.orchestra.run.vm02.stdout:Selecting previously unselected package python3-cherrypy3. 2026-03-20T17:56:41.578 INFO:teuthology.orchestra.run.vm02.stdout:Preparing to unpack .../33-python3-cherrypy3_18.6.1-4_all.deb ... 2026-03-20T17:56:41.579 INFO:teuthology.orchestra.run.vm02.stdout:Unpacking python3-cherrypy3 (18.6.1-4) ... 2026-03-20T17:56:41.582 INFO:teuthology.orchestra.run.vm06.stdout:Setting up socat (1.7.4.1-3ubuntu4) ... 2026-03-20T17:56:41.583 INFO:teuthology.orchestra.run.vm08.stdout:Setting up python3-repoze.lru (0.7-2) ... 2026-03-20T17:56:41.584 INFO:teuthology.orchestra.run.vm06.stdout:Setting up python3-ceph-common (20.2.0-712-g70f8415b-1jammy) ... 2026-03-20T17:56:41.610 INFO:teuthology.orchestra.run.vm02.stdout:Selecting previously unselected package python3-natsort. 2026-03-20T17:56:41.616 INFO:teuthology.orchestra.run.vm05.stdout:Setting up smartmontools (7.2-1ubuntu0.1) ... 2026-03-20T17:56:41.622 INFO:teuthology.orchestra.run.vm02.stdout:Preparing to unpack .../34-python3-natsort_8.0.2-1_all.deb ... 2026-03-20T17:56:41.623 INFO:teuthology.orchestra.run.vm02.stdout:Unpacking python3-natsort (8.0.2-1) ... 2026-03-20T17:56:41.641 INFO:teuthology.orchestra.run.vm02.stdout:Selecting previously unselected package ceph-mgr-modules-core. 2026-03-20T17:56:41.648 INFO:teuthology.orchestra.run.vm02.stdout:Preparing to unpack .../35-ceph-mgr-modules-core_20.2.0-712-g70f8415b-1jammy_all.deb ... 2026-03-20T17:56:41.648 INFO:teuthology.orchestra.run.vm02.stdout:Unpacking ceph-mgr-modules-core (20.2.0-712-g70f8415b-1jammy) ... 2026-03-20T17:56:41.660 INFO:teuthology.orchestra.run.vm08.stdout:Setting up liboath0:amd64 (2.6.7-3ubuntu0.1) ... 2026-03-20T17:56:41.663 INFO:teuthology.orchestra.run.vm08.stdout:Setting up python3-py (1.10.0-1) ... 2026-03-20T17:56:41.681 INFO:teuthology.orchestra.run.vm06.stdout:Setting up python3-sklearn (0.23.2-5ubuntu6) ... 2026-03-20T17:56:41.684 INFO:teuthology.orchestra.run.vm02.stdout:Selecting previously unselected package libsqlite3-mod-ceph. 2026-03-20T17:56:41.690 INFO:teuthology.orchestra.run.vm02.stdout:Preparing to unpack .../36-libsqlite3-mod-ceph_20.2.0-712-g70f8415b-1jammy_amd64.deb ... 2026-03-20T17:56:41.691 INFO:teuthology.orchestra.run.vm02.stdout:Unpacking libsqlite3-mod-ceph (20.2.0-712-g70f8415b-1jammy) ... 2026-03-20T17:56:41.708 INFO:teuthology.orchestra.run.vm02.stdout:Selecting previously unselected package ceph-mgr. 2026-03-20T17:56:41.714 INFO:teuthology.orchestra.run.vm02.stdout:Preparing to unpack .../37-ceph-mgr_20.2.0-712-g70f8415b-1jammy_amd64.deb ... 2026-03-20T17:56:41.715 INFO:teuthology.orchestra.run.vm02.stdout:Unpacking ceph-mgr (20.2.0-712-g70f8415b-1jammy) ... 2026-03-20T17:56:41.745 INFO:teuthology.orchestra.run.vm02.stdout:Selecting previously unselected package ceph-mon. 2026-03-20T17:56:41.751 INFO:teuthology.orchestra.run.vm02.stdout:Preparing to unpack .../38-ceph-mon_20.2.0-712-g70f8415b-1jammy_amd64.deb ... 2026-03-20T17:56:41.752 INFO:teuthology.orchestra.run.vm02.stdout:Unpacking ceph-mon (20.2.0-712-g70f8415b-1jammy) ... 2026-03-20T17:56:41.757 INFO:teuthology.orchestra.run.vm08.stdout:Setting up python3-joblib (0.17.0-4ubuntu1) ... 2026-03-20T17:56:41.843 INFO:teuthology.orchestra.run.vm02.stdout:Selecting previously unselected package libfuse2:amd64. 2026-03-20T17:56:41.849 INFO:teuthology.orchestra.run.vm02.stdout:Preparing to unpack .../39-libfuse2_2.9.9-5ubuntu3_amd64.deb ... 2026-03-20T17:56:41.850 INFO:teuthology.orchestra.run.vm02.stdout:Unpacking libfuse2:amd64 (2.9.9-5ubuntu3) ... 2026-03-20T17:56:41.870 INFO:teuthology.orchestra.run.vm02.stdout:Selecting previously unselected package ceph-osd. 2026-03-20T17:56:41.876 INFO:teuthology.orchestra.run.vm02.stdout:Preparing to unpack .../40-ceph-osd_20.2.0-712-g70f8415b-1jammy_amd64.deb ... 2026-03-20T17:56:41.877 INFO:teuthology.orchestra.run.vm02.stdout:Unpacking ceph-osd (20.2.0-712-g70f8415b-1jammy) ... 2026-03-20T17:56:41.879 INFO:teuthology.orchestra.run.vm05.stdout:Created symlink /etc/systemd/system/smartd.service → /lib/systemd/system/smartmontools.service. 2026-03-20T17:56:41.879 INFO:teuthology.orchestra.run.vm05.stdout:Created symlink /etc/systemd/system/multi-user.target.wants/smartmontools.service → /lib/systemd/system/smartmontools.service. 2026-03-20T17:56:41.882 INFO:teuthology.orchestra.run.vm08.stdout:Setting up python3-cachetools (5.0.0-1) ... 2026-03-20T17:56:41.949 INFO:teuthology.orchestra.run.vm08.stdout:Setting up python3-threadpoolctl (3.1.0-1) ... 2026-03-20T17:56:42.018 INFO:teuthology.orchestra.run.vm08.stdout:Setting up python3-ceph-argparse (20.2.0-712-g70f8415b-1jammy) ... 2026-03-20T17:56:42.112 INFO:teuthology.orchestra.run.vm08.stdout:Setting up python3-sklearn-lib:amd64 (0.23.2-5ubuntu6) ... 2026-03-20T17:56:42.114 INFO:teuthology.orchestra.run.vm08.stdout:Setting up libnbd0 (1.10.5-1) ... 2026-03-20T17:56:42.117 INFO:teuthology.orchestra.run.vm08.stdout:Setting up libfuse2:amd64 (2.9.9-5ubuntu3) ... 2026-03-20T17:56:42.119 INFO:teuthology.orchestra.run.vm08.stdout:Setting up libpcre2-16-0:amd64 (10.39-3ubuntu0.1) ... 2026-03-20T17:56:42.121 INFO:teuthology.orchestra.run.vm08.stdout:Setting up python3-psutil (5.9.0-1build1) ... 2026-03-20T17:56:42.126 INFO:teuthology.orchestra.run.vm02.stdout:Selecting previously unselected package ceph. 2026-03-20T17:56:42.132 INFO:teuthology.orchestra.run.vm02.stdout:Preparing to unpack .../41-ceph_20.2.0-712-g70f8415b-1jammy_amd64.deb ... 2026-03-20T17:56:42.133 INFO:teuthology.orchestra.run.vm02.stdout:Unpacking ceph (20.2.0-712-g70f8415b-1jammy) ... 2026-03-20T17:56:42.148 INFO:teuthology.orchestra.run.vm02.stdout:Selecting previously unselected package ceph-fuse. 2026-03-20T17:56:42.155 INFO:teuthology.orchestra.run.vm02.stdout:Preparing to unpack .../42-ceph-fuse_20.2.0-712-g70f8415b-1jammy_amd64.deb ... 2026-03-20T17:56:42.156 INFO:teuthology.orchestra.run.vm02.stdout:Unpacking ceph-fuse (20.2.0-712-g70f8415b-1jammy) ... 2026-03-20T17:56:42.185 INFO:teuthology.orchestra.run.vm02.stdout:Selecting previously unselected package ceph-mds. 2026-03-20T17:56:42.191 INFO:teuthology.orchestra.run.vm02.stdout:Preparing to unpack .../43-ceph-mds_20.2.0-712-g70f8415b-1jammy_amd64.deb ... 2026-03-20T17:56:42.191 INFO:teuthology.orchestra.run.vm02.stdout:Unpacking ceph-mds (20.2.0-712-g70f8415b-1jammy) ... 2026-03-20T17:56:42.233 INFO:teuthology.orchestra.run.vm02.stdout:Selecting previously unselected package cephadm. 2026-03-20T17:56:42.239 INFO:teuthology.orchestra.run.vm02.stdout:Preparing to unpack .../44-cephadm_20.2.0-712-g70f8415b-1jammy_amd64.deb ... 2026-03-20T17:56:42.240 INFO:teuthology.orchestra.run.vm02.stdout:Unpacking cephadm (20.2.0-712-g70f8415b-1jammy) ... 2026-03-20T17:56:42.244 INFO:teuthology.orchestra.run.vm08.stdout:Setting up python3-natsort (8.0.2-1) ... 2026-03-20T17:56:42.250 INFO:teuthology.orchestra.run.vm06.stdout:Setting up libqt5core5a:amd64 (5.15.3+dfsg-2ubuntu0.2) ... 2026-03-20T17:56:42.256 INFO:teuthology.orchestra.run.vm06.stdout:Setting up python3-toml (0.10.2-1) ... 2026-03-20T17:56:42.260 INFO:teuthology.orchestra.run.vm02.stdout:Selecting previously unselected package python3-asyncssh. 2026-03-20T17:56:42.266 INFO:teuthology.orchestra.run.vm02.stdout:Preparing to unpack .../45-python3-asyncssh_2.5.0-1ubuntu0.1_all.deb ... 2026-03-20T17:56:42.267 INFO:teuthology.orchestra.run.vm02.stdout:Unpacking python3-asyncssh (2.5.0-1ubuntu0.1) ... 2026-03-20T17:56:42.276 INFO:teuthology.orchestra.run.vm05.stdout:Setting up python3-iniconfig (1.1.1-2) ... 2026-03-20T17:56:42.300 INFO:teuthology.orchestra.run.vm02.stdout:Selecting previously unselected package ceph-mgr-cephadm. 2026-03-20T17:56:42.306 INFO:teuthology.orchestra.run.vm02.stdout:Preparing to unpack .../46-ceph-mgr-cephadm_20.2.0-712-g70f8415b-1jammy_all.deb ... 2026-03-20T17:56:42.307 INFO:teuthology.orchestra.run.vm02.stdout:Unpacking ceph-mgr-cephadm (20.2.0-712-g70f8415b-1jammy) ... 2026-03-20T17:56:42.321 INFO:teuthology.orchestra.run.vm08.stdout:Setting up libcephfs-proxy2 (20.2.0-712-g70f8415b-1jammy) ... 2026-03-20T17:56:42.323 INFO:teuthology.orchestra.run.vm08.stdout:Setting up python3-routes (2.5.1-1ubuntu1) ... 2026-03-20T17:56:42.330 INFO:teuthology.orchestra.run.vm06.stdout:Setting up librdkafka1:amd64 (1.8.0-1build1) ... 2026-03-20T17:56:42.333 INFO:teuthology.orchestra.run.vm06.stdout:Setting up xmlstarlet (1.6.1-2.1) ... 2026-03-20T17:56:42.335 INFO:teuthology.orchestra.run.vm02.stdout:Selecting previously unselected package python3-repoze.lru. 2026-03-20T17:56:42.335 INFO:teuthology.orchestra.run.vm06.stdout:Setting up python3-pluggy (0.13.0-7.1) ... 2026-03-20T17:56:42.343 INFO:teuthology.orchestra.run.vm02.stdout:Preparing to unpack .../47-python3-repoze.lru_0.7-2_all.deb ... 2026-03-20T17:56:42.344 INFO:teuthology.orchestra.run.vm02.stdout:Unpacking python3-repoze.lru (0.7-2) ... 2026-03-20T17:56:42.345 INFO:teuthology.orchestra.run.vm05.stdout:Setting up libdouble-conversion3:amd64 (3.1.7-4) ... 2026-03-20T17:56:42.348 INFO:teuthology.orchestra.run.vm05.stdout:Setting up nvme-cli (1.16-3ubuntu0.3) ... 2026-03-20T17:56:42.361 INFO:teuthology.orchestra.run.vm02.stdout:Selecting previously unselected package python3-routes. 2026-03-20T17:56:42.371 INFO:teuthology.orchestra.run.vm02.stdout:Preparing to unpack .../48-python3-routes_2.5.1-1ubuntu1_all.deb ... 2026-03-20T17:56:42.372 INFO:teuthology.orchestra.run.vm02.stdout:Unpacking python3-routes (2.5.1-1ubuntu1) ... 2026-03-20T17:56:42.397 INFO:teuthology.orchestra.run.vm02.stdout:Selecting previously unselected package ceph-mgr-dashboard. 2026-03-20T17:56:42.400 INFO:teuthology.orchestra.run.vm08.stdout:Setting up python3-simplejson (3.17.6-1build1) ... 2026-03-20T17:56:42.402 INFO:teuthology.orchestra.run.vm02.stdout:Preparing to unpack .../49-ceph-mgr-dashboard_20.2.0-712-g70f8415b-1jammy_all.deb ... 2026-03-20T17:56:42.403 INFO:teuthology.orchestra.run.vm02.stdout:Unpacking ceph-mgr-dashboard (20.2.0-712-g70f8415b-1jammy) ... 2026-03-20T17:56:42.410 INFO:teuthology.orchestra.run.vm06.stdout:Setting up python3-zc.lockfile (2.0-1) ... 2026-03-20T17:56:42.416 INFO:teuthology.orchestra.run.vm05.stdout:Created symlink /etc/systemd/system/default.target.wants/nvmefc-boot-connections.service → /lib/systemd/system/nvmefc-boot-connections.service. 2026-03-20T17:56:42.476 INFO:teuthology.orchestra.run.vm06.stdout:Setting up libqt5dbus5:amd64 (5.15.3+dfsg-2ubuntu0.2) ... 2026-03-20T17:56:42.478 INFO:teuthology.orchestra.run.vm06.stdout:Setting up python3-rsa (4.8-1) ... 2026-03-20T17:56:42.480 INFO:teuthology.orchestra.run.vm08.stdout:Setting up python3-pygments (2.11.2+dfsg-2ubuntu0.1) ... 2026-03-20T17:56:42.553 INFO:teuthology.orchestra.run.vm06.stdout:Setting up python3-tempora (4.1.2-1) ... 2026-03-20T17:56:42.620 INFO:teuthology.orchestra.run.vm06.stdout:Setting up python3-prettytable (2.5.0-2) ... 2026-03-20T17:56:42.647 INFO:teuthology.orchestra.run.vm05.stdout:Created symlink /etc/systemd/system/default.target.wants/nvmf-autoconnect.service → /lib/systemd/system/nvmf-autoconnect.service. 2026-03-20T17:56:42.700 INFO:teuthology.orchestra.run.vm06.stdout:Setting up liblttng-ust1:amd64 (2.13.1-1ubuntu1) ... 2026-03-20T17:56:42.702 INFO:teuthology.orchestra.run.vm06.stdout:Setting up python3-websocket (1.2.3-1) ... 2026-03-20T17:56:42.764 INFO:teuthology.orchestra.run.vm08.stdout:Setting up qttranslations5-l10n (5.15.3-1) ... 2026-03-20T17:56:42.766 INFO:teuthology.orchestra.run.vm08.stdout:Setting up python3-wcwidth (0.2.5+dfsg1-1) ... 2026-03-20T17:56:42.784 INFO:teuthology.orchestra.run.vm06.stdout:Setting up libonig5:amd64 (6.9.7.1-2build1) ... 2026-03-20T17:56:42.786 INFO:teuthology.orchestra.run.vm06.stdout:Setting up python3-requests-oauthlib (1.3.0+ds-0.1) ... 2026-03-20T17:56:42.856 INFO:teuthology.orchestra.run.vm08.stdout:Setting up python3-asyncssh (2.5.0-1ubuntu0.1) ... 2026-03-20T17:56:42.863 INFO:teuthology.orchestra.run.vm06.stdout:Setting up python3-webob (1:1.8.6-1.1ubuntu0.1) ... 2026-03-20T17:56:43.014 INFO:teuthology.orchestra.run.vm05.stdout:nvmf-connect.target is a disabled or a static unit, not starting it. 2026-03-20T17:56:43.020 INFO:teuthology.orchestra.run.vm06.stdout:Setting up python3-jaraco.collections (3.4.0-2) ... 2026-03-20T17:56:43.028 INFO:teuthology.orchestra.run.vm08.stdout:Setting up python3-cheroot (8.5.2+ds1-1ubuntu3.1) ... 2026-03-20T17:56:43.032 INFO:teuthology.orchestra.run.vm05.stdout:Setting up cephadm (20.2.0-712-g70f8415b-1jammy) ... 2026-03-20T17:56:43.074 INFO:teuthology.orchestra.run.vm05.stdout:Adding system user cephadm....done 2026-03-20T17:56:43.084 INFO:teuthology.orchestra.run.vm05.stdout:Setting up python3-jaraco.classes (3.2.1-3) ... 2026-03-20T17:56:43.098 INFO:teuthology.orchestra.run.vm06.stdout:Setting up libjq1:amd64 (1.6-2.1ubuntu3.1) ... 2026-03-20T17:56:43.098 INFO:teuthology.orchestra.run.vm02.stdout:Selecting previously unselected package python3-sklearn-lib:amd64. 2026-03-20T17:56:43.100 INFO:teuthology.orchestra.run.vm06.stdout:Setting up python3-pytest (6.2.5-1ubuntu2) ... 2026-03-20T17:56:43.105 INFO:teuthology.orchestra.run.vm02.stdout:Preparing to unpack .../50-python3-sklearn-lib_0.23.2-5ubuntu6_amd64.deb ... 2026-03-20T17:56:43.107 INFO:teuthology.orchestra.run.vm02.stdout:Unpacking python3-sklearn-lib:amd64 (0.23.2-5ubuntu6) ... 2026-03-20T17:56:43.118 INFO:teuthology.orchestra.run.vm08.stdout:Setting up python3-jaraco.text (3.6.0-2) ... 2026-03-20T17:56:43.156 INFO:teuthology.orchestra.run.vm05.stdout:Setting up python-asyncssh-doc (2.5.0-1ubuntu0.1) ... 2026-03-20T17:56:43.158 INFO:teuthology.orchestra.run.vm05.stdout:Setting up python3-jaraco.functools (3.4.0-2) ... 2026-03-20T17:56:43.169 INFO:teuthology.orchestra.run.vm02.stdout:Selecting previously unselected package python3-joblib. 2026-03-20T17:56:43.175 INFO:teuthology.orchestra.run.vm02.stdout:Preparing to unpack .../51-python3-joblib_0.17.0-4ubuntu1_all.deb ... 2026-03-20T17:56:43.176 INFO:teuthology.orchestra.run.vm02.stdout:Unpacking python3-joblib (0.17.0-4ubuntu1) ... 2026-03-20T17:56:43.187 INFO:teuthology.orchestra.run.vm08.stdout:Setting up socat (1.7.4.1-3ubuntu4) ... 2026-03-20T17:56:43.190 INFO:teuthology.orchestra.run.vm08.stdout:Setting up python3-ceph-common (20.2.0-712-g70f8415b-1jammy) ... 2026-03-20T17:56:43.221 INFO:teuthology.orchestra.run.vm02.stdout:Selecting previously unselected package python3-threadpoolctl. 2026-03-20T17:56:43.227 INFO:teuthology.orchestra.run.vm05.stdout:Setting up python3-repoze.lru (0.7-2) ... 2026-03-20T17:56:43.228 INFO:teuthology.orchestra.run.vm02.stdout:Preparing to unpack .../52-python3-threadpoolctl_3.1.0-1_all.deb ... 2026-03-20T17:56:43.229 INFO:teuthology.orchestra.run.vm02.stdout:Unpacking python3-threadpoolctl (3.1.0-1) ... 2026-03-20T17:56:43.250 INFO:teuthology.orchestra.run.vm02.stdout:Selecting previously unselected package python3-sklearn. 2026-03-20T17:56:43.251 INFO:teuthology.orchestra.run.vm06.stdout:Setting up python3-portend (3.0.0-1) ... 2026-03-20T17:56:43.254 INFO:teuthology.orchestra.run.vm02.stdout:Preparing to unpack .../53-python3-sklearn_0.23.2-5ubuntu6_all.deb ... 2026-03-20T17:56:43.255 INFO:teuthology.orchestra.run.vm02.stdout:Unpacking python3-sklearn (0.23.2-5ubuntu6) ... 2026-03-20T17:56:43.288 INFO:teuthology.orchestra.run.vm08.stdout:Setting up python3-sklearn (0.23.2-5ubuntu6) ... 2026-03-20T17:56:43.298 INFO:teuthology.orchestra.run.vm05.stdout:Setting up liboath0:amd64 (2.6.7-3ubuntu0.1) ... 2026-03-20T17:56:43.300 INFO:teuthology.orchestra.run.vm05.stdout:Setting up python3-py (1.10.0-1) ... 2026-03-20T17:56:43.323 INFO:teuthology.orchestra.run.vm06.stdout:Setting up libqt5network5:amd64 (5.15.3+dfsg-2ubuntu0.2) ... 2026-03-20T17:56:43.325 INFO:teuthology.orchestra.run.vm06.stdout:Setting up python3-google-auth (1.5.1-3) ... 2026-03-20T17:56:43.380 INFO:teuthology.orchestra.run.vm02.stdout:Selecting previously unselected package ceph-mgr-diskprediction-local. 2026-03-20T17:56:43.387 INFO:teuthology.orchestra.run.vm02.stdout:Preparing to unpack .../54-ceph-mgr-diskprediction-local_20.2.0-712-g70f8415b-1jammy_all.deb ... 2026-03-20T17:56:43.388 INFO:teuthology.orchestra.run.vm02.stdout:Unpacking ceph-mgr-diskprediction-local (20.2.0-712-g70f8415b-1jammy) ... 2026-03-20T17:56:43.396 INFO:teuthology.orchestra.run.vm05.stdout:Setting up python3-joblib (0.17.0-4ubuntu1) ... 2026-03-20T17:56:43.409 INFO:teuthology.orchestra.run.vm06.stdout:Setting up jq (1.6-2.1ubuntu3.1) ... 2026-03-20T17:56:43.413 INFO:teuthology.orchestra.run.vm06.stdout:Setting up python3-cherrypy3 (18.6.1-4) ... 2026-03-20T17:56:43.520 INFO:teuthology.orchestra.run.vm05.stdout:Setting up python3-cachetools (5.0.0-1) ... 2026-03-20T17:56:43.583 INFO:teuthology.orchestra.run.vm06.stdout:Setting up libthrift-0.16.0:amd64 (0.16.0-2) ... 2026-03-20T17:56:43.602 INFO:teuthology.orchestra.run.vm05.stdout:Setting up python3-threadpoolctl (3.1.0-1) ... 2026-03-20T17:56:43.602 INFO:teuthology.orchestra.run.vm06.stdout:Setting up librados2 (20.2.0-712-g70f8415b-1jammy) ... 2026-03-20T17:56:43.605 INFO:teuthology.orchestra.run.vm06.stdout:Setting up librgw2 (20.2.0-712-g70f8415b-1jammy) ... 2026-03-20T17:56:43.607 INFO:teuthology.orchestra.run.vm06.stdout:Setting up libsqlite3-mod-ceph (20.2.0-712-g70f8415b-1jammy) ... 2026-03-20T17:56:43.622 INFO:teuthology.orchestra.run.vm06.stdout:Setting up python3-kubernetes (12.0.1-1ubuntu1) ... 2026-03-20T17:56:43.640 INFO:teuthology.orchestra.run.vm02.stdout:Selecting previously unselected package python3-cachetools. 2026-03-20T17:56:43.647 INFO:teuthology.orchestra.run.vm02.stdout:Preparing to unpack .../55-python3-cachetools_5.0.0-1_all.deb ... 2026-03-20T17:56:43.648 INFO:teuthology.orchestra.run.vm02.stdout:Unpacking python3-cachetools (5.0.0-1) ... 2026-03-20T17:56:43.665 INFO:teuthology.orchestra.run.vm02.stdout:Selecting previously unselected package python3-rsa. 2026-03-20T17:56:43.670 INFO:teuthology.orchestra.run.vm05.stdout:Setting up python3-ceph-argparse (20.2.0-712-g70f8415b-1jammy) ... 2026-03-20T17:56:43.671 INFO:teuthology.orchestra.run.vm02.stdout:Preparing to unpack .../56-python3-rsa_4.8-1_all.deb ... 2026-03-20T17:56:43.673 INFO:teuthology.orchestra.run.vm02.stdout:Unpacking python3-rsa (4.8-1) ... 2026-03-20T17:56:43.691 INFO:teuthology.orchestra.run.vm02.stdout:Selecting previously unselected package python3-google-auth. 2026-03-20T17:56:43.699 INFO:teuthology.orchestra.run.vm02.stdout:Preparing to unpack .../57-python3-google-auth_1.5.1-3_all.deb ... 2026-03-20T17:56:43.700 INFO:teuthology.orchestra.run.vm02.stdout:Unpacking python3-google-auth (1.5.1-3) ... 2026-03-20T17:56:43.719 INFO:teuthology.orchestra.run.vm02.stdout:Selecting previously unselected package python3-requests-oauthlib. 2026-03-20T17:56:43.725 INFO:teuthology.orchestra.run.vm02.stdout:Preparing to unpack .../58-python3-requests-oauthlib_1.3.0+ds-0.1_all.deb ... 2026-03-20T17:56:43.726 INFO:teuthology.orchestra.run.vm02.stdout:Unpacking python3-requests-oauthlib (1.3.0+ds-0.1) ... 2026-03-20T17:56:43.743 INFO:teuthology.orchestra.run.vm05.stdout:Setting up python3-sklearn-lib:amd64 (0.23.2-5ubuntu6) ... 2026-03-20T17:56:43.745 INFO:teuthology.orchestra.run.vm02.stdout:Selecting previously unselected package python3-websocket. 2026-03-20T17:56:43.745 INFO:teuthology.orchestra.run.vm05.stdout:Setting up libnbd0 (1.10.5-1) ... 2026-03-20T17:56:43.748 INFO:teuthology.orchestra.run.vm05.stdout:Setting up libfuse2:amd64 (2.9.9-5ubuntu3) ... 2026-03-20T17:56:43.750 INFO:teuthology.orchestra.run.vm05.stdout:Setting up libpcre2-16-0:amd64 (10.39-3ubuntu0.1) ... 2026-03-20T17:56:43.751 INFO:teuthology.orchestra.run.vm02.stdout:Preparing to unpack .../59-python3-websocket_1.2.3-1_all.deb ... 2026-03-20T17:56:43.752 INFO:teuthology.orchestra.run.vm05.stdout:Setting up python3-psutil (5.9.0-1build1) ... 2026-03-20T17:56:43.753 INFO:teuthology.orchestra.run.vm02.stdout:Unpacking python3-websocket (1.2.3-1) ... 2026-03-20T17:56:43.772 INFO:teuthology.orchestra.run.vm02.stdout:Selecting previously unselected package python3-kubernetes. 2026-03-20T17:56:43.779 INFO:teuthology.orchestra.run.vm02.stdout:Preparing to unpack .../60-python3-kubernetes_12.0.1-1ubuntu1_all.deb ... 2026-03-20T17:56:43.780 INFO:teuthology.orchestra.run.vm02.stdout:Unpacking python3-kubernetes (12.0.1-1ubuntu1) ... 2026-03-20T17:56:43.869 INFO:teuthology.orchestra.run.vm08.stdout:Setting up libqt5core5a:amd64 (5.15.3+dfsg-2ubuntu0.2) ... 2026-03-20T17:56:43.889 INFO:teuthology.orchestra.run.vm08.stdout:Setting up python3-toml (0.10.2-1) ... 2026-03-20T17:56:43.890 INFO:teuthology.orchestra.run.vm05.stdout:Setting up python3-natsort (8.0.2-1) ... 2026-03-20T17:56:43.931 INFO:teuthology.orchestra.run.vm02.stdout:Selecting previously unselected package ceph-mgr-k8sevents. 2026-03-20T17:56:43.938 INFO:teuthology.orchestra.run.vm02.stdout:Preparing to unpack .../61-ceph-mgr-k8sevents_20.2.0-712-g70f8415b-1jammy_all.deb ... 2026-03-20T17:56:43.939 INFO:teuthology.orchestra.run.vm02.stdout:Unpacking ceph-mgr-k8sevents (20.2.0-712-g70f8415b-1jammy) ... 2026-03-20T17:56:43.956 INFO:teuthology.orchestra.run.vm02.stdout:Selecting previously unselected package libonig5:amd64. 2026-03-20T17:56:43.961 INFO:teuthology.orchestra.run.vm08.stdout:Setting up librdkafka1:amd64 (1.8.0-1build1) ... 2026-03-20T17:56:43.962 INFO:teuthology.orchestra.run.vm02.stdout:Preparing to unpack .../62-libonig5_6.9.7.1-2build1_amd64.deb ... 2026-03-20T17:56:43.963 INFO:teuthology.orchestra.run.vm02.stdout:Unpacking libonig5:amd64 (6.9.7.1-2build1) ... 2026-03-20T17:56:43.964 INFO:teuthology.orchestra.run.vm08.stdout:Setting up xmlstarlet (1.6.1-2.1) ... 2026-03-20T17:56:43.966 INFO:teuthology.orchestra.run.vm08.stdout:Setting up python3-pluggy (0.13.0-7.1) ... 2026-03-20T17:56:43.969 INFO:teuthology.orchestra.run.vm05.stdout:Setting up libcephfs-proxy2 (20.2.0-712-g70f8415b-1jammy) ... 2026-03-20T17:56:43.971 INFO:teuthology.orchestra.run.vm05.stdout:Setting up python3-routes (2.5.1-1ubuntu1) ... 2026-03-20T17:56:43.986 INFO:teuthology.orchestra.run.vm02.stdout:Selecting previously unselected package libjq1:amd64. 2026-03-20T17:56:43.992 INFO:teuthology.orchestra.run.vm02.stdout:Preparing to unpack .../63-libjq1_1.6-2.1ubuntu3.1_amd64.deb ... 2026-03-20T17:56:43.993 INFO:teuthology.orchestra.run.vm02.stdout:Unpacking libjq1:amd64 (1.6-2.1ubuntu3.1) ... 2026-03-20T17:56:44.018 INFO:teuthology.orchestra.run.vm02.stdout:Selecting previously unselected package jq. 2026-03-20T17:56:44.023 INFO:teuthology.orchestra.run.vm02.stdout:Preparing to unpack .../64-jq_1.6-2.1ubuntu3.1_amd64.deb ... 2026-03-20T17:56:44.024 INFO:teuthology.orchestra.run.vm02.stdout:Unpacking jq (1.6-2.1ubuntu3.1) ... 2026-03-20T17:56:44.042 INFO:teuthology.orchestra.run.vm02.stdout:Selecting previously unselected package socat. 2026-03-20T17:56:44.043 INFO:teuthology.orchestra.run.vm08.stdout:Setting up python3-zc.lockfile (2.0-1) ... 2026-03-20T17:56:44.048 INFO:teuthology.orchestra.run.vm02.stdout:Preparing to unpack .../65-socat_1.7.4.1-3ubuntu4_amd64.deb ... 2026-03-20T17:56:44.049 INFO:teuthology.orchestra.run.vm02.stdout:Unpacking socat (1.7.4.1-3ubuntu4) ... 2026-03-20T17:56:44.054 INFO:teuthology.orchestra.run.vm05.stdout:Setting up python3-simplejson (3.17.6-1build1) ... 2026-03-20T17:56:44.080 INFO:teuthology.orchestra.run.vm02.stdout:Selecting previously unselected package xmlstarlet. 2026-03-20T17:56:44.087 INFO:teuthology.orchestra.run.vm02.stdout:Preparing to unpack .../66-xmlstarlet_1.6.1-2.1_amd64.deb ... 2026-03-20T17:56:44.088 INFO:teuthology.orchestra.run.vm02.stdout:Unpacking xmlstarlet (1.6.1-2.1) ... 2026-03-20T17:56:44.115 INFO:teuthology.orchestra.run.vm08.stdout:Setting up libqt5dbus5:amd64 (5.15.3+dfsg-2ubuntu0.2) ... 2026-03-20T17:56:44.118 INFO:teuthology.orchestra.run.vm08.stdout:Setting up python3-rsa (4.8-1) ... 2026-03-20T17:56:44.132 INFO:teuthology.orchestra.run.vm02.stdout:Selecting previously unselected package ceph-test. 2026-03-20T17:56:44.134 INFO:teuthology.orchestra.run.vm05.stdout:Setting up python3-pygments (2.11.2+dfsg-2ubuntu0.1) ... 2026-03-20T17:56:44.139 INFO:teuthology.orchestra.run.vm02.stdout:Preparing to unpack .../67-ceph-test_20.2.0-712-g70f8415b-1jammy_amd64.deb ... 2026-03-20T17:56:44.140 INFO:teuthology.orchestra.run.vm02.stdout:Unpacking ceph-test (20.2.0-712-g70f8415b-1jammy) ... 2026-03-20T17:56:44.194 INFO:teuthology.orchestra.run.vm08.stdout:Setting up python3-tempora (4.1.2-1) ... 2026-03-20T17:56:44.220 INFO:teuthology.orchestra.run.vm06.stdout:Setting up libcephfs2 (20.2.0-712-g70f8415b-1jammy) ... 2026-03-20T17:56:44.226 INFO:teuthology.orchestra.run.vm06.stdout:Setting up libradosstriper1 (20.2.0-712-g70f8415b-1jammy) ... 2026-03-20T17:56:44.228 INFO:teuthology.orchestra.run.vm06.stdout:Setting up librbd1 (20.2.0-712-g70f8415b-1jammy) ... 2026-03-20T17:56:44.231 INFO:teuthology.orchestra.run.vm06.stdout:Setting up ceph-mgr-modules-core (20.2.0-712-g70f8415b-1jammy) ... 2026-03-20T17:56:44.233 INFO:teuthology.orchestra.run.vm06.stdout:Setting up ceph-fuse (20.2.0-712-g70f8415b-1jammy) ... 2026-03-20T17:56:44.263 INFO:teuthology.orchestra.run.vm08.stdout:Setting up python3-prettytable (2.5.0-2) ... 2026-03-20T17:56:44.301 INFO:teuthology.orchestra.run.vm06.stdout:Created symlink /etc/systemd/system/remote-fs.target.wants/ceph-fuse.target → /lib/systemd/system/ceph-fuse.target. 2026-03-20T17:56:44.301 INFO:teuthology.orchestra.run.vm06.stdout:Created symlink /etc/systemd/system/ceph.target.wants/ceph-fuse.target → /lib/systemd/system/ceph-fuse.target. 2026-03-20T17:56:44.339 INFO:teuthology.orchestra.run.vm08.stdout:Setting up liblttng-ust1:amd64 (2.13.1-1ubuntu1) ... 2026-03-20T17:56:44.342 INFO:teuthology.orchestra.run.vm08.stdout:Setting up python3-websocket (1.2.3-1) ... 2026-03-20T17:56:44.411 INFO:teuthology.orchestra.run.vm05.stdout:Setting up qttranslations5-l10n (5.15.3-1) ... 2026-03-20T17:56:44.414 INFO:teuthology.orchestra.run.vm05.stdout:Setting up python3-wcwidth (0.2.5+dfsg1-1) ... 2026-03-20T17:56:44.424 INFO:teuthology.orchestra.run.vm08.stdout:Setting up libonig5:amd64 (6.9.7.1-2build1) ... 2026-03-20T17:56:44.426 INFO:teuthology.orchestra.run.vm08.stdout:Setting up python3-requests-oauthlib (1.3.0+ds-0.1) ... 2026-03-20T17:56:44.503 INFO:teuthology.orchestra.run.vm08.stdout:Setting up python3-webob (1:1.8.6-1.1ubuntu0.1) ... 2026-03-20T17:56:44.506 INFO:teuthology.orchestra.run.vm05.stdout:Setting up python3-asyncssh (2.5.0-1ubuntu0.1) ... 2026-03-20T17:56:44.597 INFO:teuthology.orchestra.run.vm08.stdout:Setting up python3-jaraco.collections (3.4.0-2) ... 2026-03-20T17:56:44.655 INFO:teuthology.orchestra.run.vm05.stdout:Setting up python3-cheroot (8.5.2+ds1-1ubuntu3.1) ... 2026-03-20T17:56:44.666 INFO:teuthology.orchestra.run.vm06.stdout:Setting up libcephfs-dev (20.2.0-712-g70f8415b-1jammy) ... 2026-03-20T17:56:44.668 INFO:teuthology.orchestra.run.vm06.stdout:Setting up python3-rados (20.2.0-712-g70f8415b-1jammy) ... 2026-03-20T17:56:44.671 INFO:teuthology.orchestra.run.vm06.stdout:Setting up libcephfs-daemon (20.2.0-712-g70f8415b-1jammy) ... 2026-03-20T17:56:44.672 INFO:teuthology.orchestra.run.vm08.stdout:Setting up libjq1:amd64 (1.6-2.1ubuntu3.1) ... 2026-03-20T17:56:44.673 INFO:teuthology.orchestra.run.vm06.stdout:Setting up python3-rbd (20.2.0-712-g70f8415b-1jammy) ... 2026-03-20T17:56:44.674 INFO:teuthology.orchestra.run.vm08.stdout:Setting up python3-pytest (6.2.5-1ubuntu2) ... 2026-03-20T17:56:44.676 INFO:teuthology.orchestra.run.vm06.stdout:Setting up rbd-fuse (20.2.0-712-g70f8415b-1jammy) ... 2026-03-20T17:56:44.679 INFO:teuthology.orchestra.run.vm06.stdout:Setting up python3-rgw (20.2.0-712-g70f8415b-1jammy) ... 2026-03-20T17:56:44.681 INFO:teuthology.orchestra.run.vm06.stdout:Setting up python3-cephfs (20.2.0-712-g70f8415b-1jammy) ... 2026-03-20T17:56:44.683 INFO:teuthology.orchestra.run.vm06.stdout:Setting up ceph-common (20.2.0-712-g70f8415b-1jammy) ... 2026-03-20T17:56:44.721 INFO:teuthology.orchestra.run.vm06.stdout:Adding group ceph....done 2026-03-20T17:56:44.746 INFO:teuthology.orchestra.run.vm05.stdout:Setting up python3-jaraco.text (3.6.0-2) ... 2026-03-20T17:56:44.762 INFO:teuthology.orchestra.run.vm06.stdout:Adding system user ceph....done 2026-03-20T17:56:44.772 INFO:teuthology.orchestra.run.vm06.stdout:Setting system user ceph properties....done 2026-03-20T17:56:44.777 INFO:teuthology.orchestra.run.vm06.stdout:chown: cannot access '/var/log/ceph/*.log*': No such file or directory 2026-03-20T17:56:44.813 INFO:teuthology.orchestra.run.vm05.stdout:Setting up socat (1.7.4.1-3ubuntu4) ... 2026-03-20T17:56:44.815 INFO:teuthology.orchestra.run.vm05.stdout:Setting up python3-ceph-common (20.2.0-712-g70f8415b-1jammy) ... 2026-03-20T17:56:44.818 INFO:teuthology.orchestra.run.vm08.stdout:Setting up python3-portend (3.0.0-1) ... 2026-03-20T17:56:44.853 INFO:teuthology.orchestra.run.vm06.stdout:Created symlink /etc/systemd/system/multi-user.target.wants/ceph.target → /lib/systemd/system/ceph.target. 2026-03-20T17:56:44.910 INFO:teuthology.orchestra.run.vm08.stdout:Setting up libqt5network5:amd64 (5.15.3+dfsg-2ubuntu0.2) ... 2026-03-20T17:56:44.912 INFO:teuthology.orchestra.run.vm08.stdout:Setting up python3-google-auth (1.5.1-3) ... 2026-03-20T17:56:44.918 INFO:teuthology.orchestra.run.vm05.stdout:Setting up python3-sklearn (0.23.2-5ubuntu6) ... 2026-03-20T17:56:44.994 INFO:teuthology.orchestra.run.vm08.stdout:Setting up jq (1.6-2.1ubuntu3.1) ... 2026-03-20T17:56:44.996 INFO:teuthology.orchestra.run.vm08.stdout:Setting up python3-cherrypy3 (18.6.1-4) ... 2026-03-20T17:56:45.098 INFO:teuthology.orchestra.run.vm06.stdout:Created symlink /etc/systemd/system/multi-user.target.wants/rbdmap.service → /lib/systemd/system/rbdmap.service. 2026-03-20T17:56:45.175 INFO:teuthology.orchestra.run.vm08.stdout:Setting up libthrift-0.16.0:amd64 (0.16.0-2) ... 2026-03-20T17:56:45.334 INFO:teuthology.orchestra.run.vm08.stdout:Setting up librados2 (20.2.0-712-g70f8415b-1jammy) ... 2026-03-20T17:56:45.471 INFO:teuthology.orchestra.run.vm08.stdout:Setting up librgw2 (20.2.0-712-g70f8415b-1jammy) ... 2026-03-20T17:56:45.475 INFO:teuthology.orchestra.run.vm08.stdout:Setting up libsqlite3-mod-ceph (20.2.0-712-g70f8415b-1jammy) ... 2026-03-20T17:56:45.477 INFO:teuthology.orchestra.run.vm08.stdout:Setting up python3-kubernetes (12.0.1-1ubuntu1) ... 2026-03-20T17:56:45.488 INFO:teuthology.orchestra.run.vm05.stdout:Setting up libqt5core5a:amd64 (5.15.3+dfsg-2ubuntu0.2) ... 2026-03-20T17:56:45.489 INFO:teuthology.orchestra.run.vm02.stdout:Selecting previously unselected package ceph-volume. 2026-03-20T17:56:45.493 INFO:teuthology.orchestra.run.vm05.stdout:Setting up python3-toml (0.10.2-1) ... 2026-03-20T17:56:45.494 INFO:teuthology.orchestra.run.vm06.stdout:Setting up ceph-test (20.2.0-712-g70f8415b-1jammy) ... 2026-03-20T17:56:45.496 INFO:teuthology.orchestra.run.vm02.stdout:Preparing to unpack .../68-ceph-volume_20.2.0-712-g70f8415b-1jammy_all.deb ... 2026-03-20T17:56:45.496 INFO:teuthology.orchestra.run.vm06.stdout:Setting up radosgw (20.2.0-712-g70f8415b-1jammy) ... 2026-03-20T17:56:45.497 INFO:teuthology.orchestra.run.vm02.stdout:Unpacking ceph-volume (20.2.0-712-g70f8415b-1jammy) ... 2026-03-20T17:56:45.527 INFO:teuthology.orchestra.run.vm02.stdout:Selecting previously unselected package libcephfs-daemon. 2026-03-20T17:56:45.533 INFO:teuthology.orchestra.run.vm02.stdout:Preparing to unpack .../69-libcephfs-daemon_20.2.0-712-g70f8415b-1jammy_amd64.deb ... 2026-03-20T17:56:45.533 INFO:teuthology.orchestra.run.vm02.stdout:Unpacking libcephfs-daemon (20.2.0-712-g70f8415b-1jammy) ... 2026-03-20T17:56:45.551 INFO:teuthology.orchestra.run.vm02.stdout:Selecting previously unselected package libcephfs-proxy2. 2026-03-20T17:56:45.559 INFO:teuthology.orchestra.run.vm02.stdout:Preparing to unpack .../70-libcephfs-proxy2_20.2.0-712-g70f8415b-1jammy_amd64.deb ... 2026-03-20T17:56:45.560 INFO:teuthology.orchestra.run.vm02.stdout:Unpacking libcephfs-proxy2 (20.2.0-712-g70f8415b-1jammy) ... 2026-03-20T17:56:45.569 INFO:teuthology.orchestra.run.vm05.stdout:Setting up librdkafka1:amd64 (1.8.0-1build1) ... 2026-03-20T17:56:45.571 INFO:teuthology.orchestra.run.vm05.stdout:Setting up xmlstarlet (1.6.1-2.1) ... 2026-03-20T17:56:45.574 INFO:teuthology.orchestra.run.vm05.stdout:Setting up python3-pluggy (0.13.0-7.1) ... 2026-03-20T17:56:45.577 INFO:teuthology.orchestra.run.vm02.stdout:Selecting previously unselected package libcephfs-dev. 2026-03-20T17:56:45.583 INFO:teuthology.orchestra.run.vm02.stdout:Preparing to unpack .../71-libcephfs-dev_20.2.0-712-g70f8415b-1jammy_amd64.deb ... 2026-03-20T17:56:45.584 INFO:teuthology.orchestra.run.vm02.stdout:Unpacking libcephfs-dev (20.2.0-712-g70f8415b-1jammy) ... 2026-03-20T17:56:45.603 INFO:teuthology.orchestra.run.vm02.stdout:Selecting previously unselected package nvme-cli. 2026-03-20T17:56:45.607 INFO:teuthology.orchestra.run.vm02.stdout:Preparing to unpack .../72-nvme-cli_1.16-3ubuntu0.3_amd64.deb ... 2026-03-20T17:56:45.608 INFO:teuthology.orchestra.run.vm02.stdout:Unpacking nvme-cli (1.16-3ubuntu0.3) ... 2026-03-20T17:56:45.646 INFO:teuthology.orchestra.run.vm02.stdout:Selecting previously unselected package python-asyncssh-doc. 2026-03-20T17:56:45.646 INFO:teuthology.orchestra.run.vm05.stdout:Setting up python3-zc.lockfile (2.0-1) ... 2026-03-20T17:56:45.652 INFO:teuthology.orchestra.run.vm02.stdout:Preparing to unpack .../73-python-asyncssh-doc_2.5.0-1ubuntu0.1_all.deb ... 2026-03-20T17:56:45.653 INFO:teuthology.orchestra.run.vm02.stdout:Unpacking python-asyncssh-doc (2.5.0-1ubuntu0.1) ... 2026-03-20T17:56:45.695 INFO:teuthology.orchestra.run.vm02.stdout:Selecting previously unselected package python3-iniconfig. 2026-03-20T17:56:45.702 INFO:teuthology.orchestra.run.vm02.stdout:Preparing to unpack .../74-python3-iniconfig_1.1.1-2_all.deb ... 2026-03-20T17:56:45.703 INFO:teuthology.orchestra.run.vm02.stdout:Unpacking python3-iniconfig (1.1.1-2) ... 2026-03-20T17:56:45.715 INFO:teuthology.orchestra.run.vm05.stdout:Setting up libqt5dbus5:amd64 (5.15.3+dfsg-2ubuntu0.2) ... 2026-03-20T17:56:45.717 INFO:teuthology.orchestra.run.vm05.stdout:Setting up python3-rsa (4.8-1) ... 2026-03-20T17:56:45.719 INFO:teuthology.orchestra.run.vm02.stdout:Selecting previously unselected package python3-pluggy. 2026-03-20T17:56:45.726 INFO:teuthology.orchestra.run.vm02.stdout:Preparing to unpack .../75-python3-pluggy_0.13.0-7.1_all.deb ... 2026-03-20T17:56:45.727 INFO:teuthology.orchestra.run.vm02.stdout:Unpacking python3-pluggy (0.13.0-7.1) ... 2026-03-20T17:56:45.744 INFO:teuthology.orchestra.run.vm02.stdout:Selecting previously unselected package python3-psutil. 2026-03-20T17:56:45.748 INFO:teuthology.orchestra.run.vm06.stdout:Created symlink /etc/systemd/system/multi-user.target.wants/ceph-radosgw.target → /lib/systemd/system/ceph-radosgw.target. 2026-03-20T17:56:45.748 INFO:teuthology.orchestra.run.vm06.stdout:Created symlink /etc/systemd/system/ceph.target.wants/ceph-radosgw.target → /lib/systemd/system/ceph-radosgw.target. 2026-03-20T17:56:45.751 INFO:teuthology.orchestra.run.vm02.stdout:Preparing to unpack .../76-python3-psutil_5.9.0-1build1_amd64.deb ... 2026-03-20T17:56:45.752 INFO:teuthology.orchestra.run.vm02.stdout:Unpacking python3-psutil (5.9.0-1build1) ... 2026-03-20T17:56:45.775 INFO:teuthology.orchestra.run.vm02.stdout:Selecting previously unselected package python3-py. 2026-03-20T17:56:45.781 INFO:teuthology.orchestra.run.vm02.stdout:Preparing to unpack .../77-python3-py_1.10.0-1_all.deb ... 2026-03-20T17:56:45.782 INFO:teuthology.orchestra.run.vm02.stdout:Unpacking python3-py (1.10.0-1) ... 2026-03-20T17:56:45.795 INFO:teuthology.orchestra.run.vm05.stdout:Setting up python3-tempora (4.1.2-1) ... 2026-03-20T17:56:45.805 INFO:teuthology.orchestra.run.vm02.stdout:Selecting previously unselected package python3-pygments. 2026-03-20T17:56:45.811 INFO:teuthology.orchestra.run.vm02.stdout:Preparing to unpack .../78-python3-pygments_2.11.2+dfsg-2ubuntu0.1_all.deb ... 2026-03-20T17:56:45.812 INFO:teuthology.orchestra.run.vm02.stdout:Unpacking python3-pygments (2.11.2+dfsg-2ubuntu0.1) ... 2026-03-20T17:56:45.868 INFO:teuthology.orchestra.run.vm05.stdout:Setting up python3-prettytable (2.5.0-2) ... 2026-03-20T17:56:45.870 INFO:teuthology.orchestra.run.vm02.stdout:Selecting previously unselected package python3-toml. 2026-03-20T17:56:45.876 INFO:teuthology.orchestra.run.vm02.stdout:Preparing to unpack .../79-python3-toml_0.10.2-1_all.deb ... 2026-03-20T17:56:45.877 INFO:teuthology.orchestra.run.vm02.stdout:Unpacking python3-toml (0.10.2-1) ... 2026-03-20T17:56:45.893 INFO:teuthology.orchestra.run.vm02.stdout:Selecting previously unselected package python3-pytest. 2026-03-20T17:56:45.900 INFO:teuthology.orchestra.run.vm02.stdout:Preparing to unpack .../80-python3-pytest_6.2.5-1ubuntu2_all.deb ... 2026-03-20T17:56:45.901 INFO:teuthology.orchestra.run.vm02.stdout:Unpacking python3-pytest (6.2.5-1ubuntu2) ... 2026-03-20T17:56:45.940 INFO:teuthology.orchestra.run.vm05.stdout:Setting up liblttng-ust1:amd64 (2.13.1-1ubuntu1) ... 2026-03-20T17:56:45.943 INFO:teuthology.orchestra.run.vm05.stdout:Setting up python3-websocket (1.2.3-1) ... 2026-03-20T17:56:45.943 INFO:teuthology.orchestra.run.vm02.stdout:Selecting previously unselected package python3-simplejson. 2026-03-20T17:56:45.950 INFO:teuthology.orchestra.run.vm02.stdout:Preparing to unpack .../81-python3-simplejson_3.17.6-1build1_amd64.deb ... 2026-03-20T17:56:45.951 INFO:teuthology.orchestra.run.vm02.stdout:Unpacking python3-simplejson (3.17.6-1build1) ... 2026-03-20T17:56:45.972 INFO:teuthology.orchestra.run.vm02.stdout:Selecting previously unselected package python3-webob. 2026-03-20T17:56:45.979 INFO:teuthology.orchestra.run.vm02.stdout:Preparing to unpack .../82-python3-webob_1%3a1.8.6-1.1ubuntu0.1_all.deb ... 2026-03-20T17:56:45.980 INFO:teuthology.orchestra.run.vm02.stdout:Unpacking python3-webob (1:1.8.6-1.1ubuntu0.1) ... 2026-03-20T17:56:45.999 INFO:teuthology.orchestra.run.vm02.stdout:Selecting previously unselected package qttranslations5-l10n. 2026-03-20T17:56:46.006 INFO:teuthology.orchestra.run.vm02.stdout:Preparing to unpack .../83-qttranslations5-l10n_5.15.3-1_all.deb ... 2026-03-20T17:56:46.007 INFO:teuthology.orchestra.run.vm02.stdout:Unpacking qttranslations5-l10n (5.15.3-1) ... 2026-03-20T17:56:46.025 INFO:teuthology.orchestra.run.vm05.stdout:Setting up libonig5:amd64 (6.9.7.1-2build1) ... 2026-03-20T17:56:46.027 INFO:teuthology.orchestra.run.vm05.stdout:Setting up python3-requests-oauthlib (1.3.0+ds-0.1) ... 2026-03-20T17:56:46.091 INFO:teuthology.orchestra.run.vm08.stdout:Setting up libcephfs2 (20.2.0-712-g70f8415b-1jammy) ... 2026-03-20T17:56:46.093 INFO:teuthology.orchestra.run.vm08.stdout:Setting up libradosstriper1 (20.2.0-712-g70f8415b-1jammy) ... 2026-03-20T17:56:46.096 INFO:teuthology.orchestra.run.vm08.stdout:Setting up librbd1 (20.2.0-712-g70f8415b-1jammy) ... 2026-03-20T17:56:46.098 INFO:teuthology.orchestra.run.vm08.stdout:Setting up ceph-mgr-modules-core (20.2.0-712-g70f8415b-1jammy) ... 2026-03-20T17:56:46.101 INFO:teuthology.orchestra.run.vm08.stdout:Setting up ceph-fuse (20.2.0-712-g70f8415b-1jammy) ... 2026-03-20T17:56:46.102 INFO:teuthology.orchestra.run.vm05.stdout:Setting up python3-webob (1:1.8.6-1.1ubuntu0.1) ... 2026-03-20T17:56:46.118 INFO:teuthology.orchestra.run.vm02.stdout:Selecting previously unselected package radosgw. 2026-03-20T17:56:46.121 INFO:teuthology.orchestra.run.vm02.stdout:Preparing to unpack .../84-radosgw_20.2.0-712-g70f8415b-1jammy_amd64.deb ... 2026-03-20T17:56:46.122 INFO:teuthology.orchestra.run.vm02.stdout:Unpacking radosgw (20.2.0-712-g70f8415b-1jammy) ... 2026-03-20T17:56:46.139 INFO:teuthology.orchestra.run.vm06.stdout:Setting up ceph-base (20.2.0-712-g70f8415b-1jammy) ... 2026-03-20T17:56:46.164 INFO:teuthology.orchestra.run.vm08.stdout:Created symlink /etc/systemd/system/remote-fs.target.wants/ceph-fuse.target → /lib/systemd/system/ceph-fuse.target. 2026-03-20T17:56:46.164 INFO:teuthology.orchestra.run.vm08.stdout:Created symlink /etc/systemd/system/ceph.target.wants/ceph-fuse.target → /lib/systemd/system/ceph-fuse.target. 2026-03-20T17:56:46.196 INFO:teuthology.orchestra.run.vm05.stdout:Setting up python3-jaraco.collections (3.4.0-2) ... 2026-03-20T17:56:46.235 INFO:teuthology.orchestra.run.vm06.stdout:Created symlink /etc/systemd/system/ceph.target.wants/ceph-crash.service → /lib/systemd/system/ceph-crash.service. 2026-03-20T17:56:46.266 INFO:teuthology.orchestra.run.vm05.stdout:Setting up libjq1:amd64 (1.6-2.1ubuntu3.1) ... 2026-03-20T17:56:46.268 INFO:teuthology.orchestra.run.vm05.stdout:Setting up python3-pytest (6.2.5-1ubuntu2) ... 2026-03-20T17:56:46.453 INFO:teuthology.orchestra.run.vm05.stdout:Setting up python3-portend (3.0.0-1) ... 2026-03-20T17:56:46.470 INFO:teuthology.orchestra.run.vm02.stdout:Selecting previously unselected package rbd-fuse. 2026-03-20T17:56:46.474 INFO:teuthology.orchestra.run.vm02.stdout:Preparing to unpack .../85-rbd-fuse_20.2.0-712-g70f8415b-1jammy_amd64.deb ... 2026-03-20T17:56:46.475 INFO:teuthology.orchestra.run.vm02.stdout:Unpacking rbd-fuse (20.2.0-712-g70f8415b-1jammy) ... 2026-03-20T17:56:46.495 INFO:teuthology.orchestra.run.vm02.stdout:Selecting previously unselected package smartmontools. 2026-03-20T17:56:46.502 INFO:teuthology.orchestra.run.vm02.stdout:Preparing to unpack .../86-smartmontools_7.2-1ubuntu0.1_amd64.deb ... 2026-03-20T17:56:46.512 INFO:teuthology.orchestra.run.vm02.stdout:Unpacking smartmontools (7.2-1ubuntu0.1) ... 2026-03-20T17:56:46.520 INFO:teuthology.orchestra.run.vm05.stdout:Setting up libqt5network5:amd64 (5.15.3+dfsg-2ubuntu0.2) ... 2026-03-20T17:56:46.523 INFO:teuthology.orchestra.run.vm05.stdout:Setting up python3-google-auth (1.5.1-3) ... 2026-03-20T17:56:46.541 INFO:teuthology.orchestra.run.vm08.stdout:Setting up libcephfs-dev (20.2.0-712-g70f8415b-1jammy) ... 2026-03-20T17:56:46.544 INFO:teuthology.orchestra.run.vm08.stdout:Setting up python3-rados (20.2.0-712-g70f8415b-1jammy) ... 2026-03-20T17:56:46.546 INFO:teuthology.orchestra.run.vm08.stdout:Setting up libcephfs-daemon (20.2.0-712-g70f8415b-1jammy) ... 2026-03-20T17:56:46.548 INFO:teuthology.orchestra.run.vm08.stdout:Setting up python3-rbd (20.2.0-712-g70f8415b-1jammy) ... 2026-03-20T17:56:46.551 INFO:teuthology.orchestra.run.vm08.stdout:Setting up rbd-fuse (20.2.0-712-g70f8415b-1jammy) ... 2026-03-20T17:56:46.554 INFO:teuthology.orchestra.run.vm08.stdout:Setting up python3-rgw (20.2.0-712-g70f8415b-1jammy) ... 2026-03-20T17:56:46.555 INFO:teuthology.orchestra.run.vm02.stdout:Setting up smartmontools (7.2-1ubuntu0.1) ... 2026-03-20T17:56:46.557 INFO:teuthology.orchestra.run.vm08.stdout:Setting up python3-cephfs (20.2.0-712-g70f8415b-1jammy) ... 2026-03-20T17:56:46.560 INFO:teuthology.orchestra.run.vm08.stdout:Setting up ceph-common (20.2.0-712-g70f8415b-1jammy) ... 2026-03-20T17:56:46.594 INFO:teuthology.orchestra.run.vm08.stdout:Adding group ceph....done 2026-03-20T17:56:46.602 INFO:teuthology.orchestra.run.vm05.stdout:Setting up jq (1.6-2.1ubuntu3.1) ... 2026-03-20T17:56:46.605 INFO:teuthology.orchestra.run.vm05.stdout:Setting up python3-cherrypy3 (18.6.1-4) ... 2026-03-20T17:56:46.622 INFO:teuthology.orchestra.run.vm06.stdout:Setting up ceph-mds (20.2.0-712-g70f8415b-1jammy) ... 2026-03-20T17:56:46.633 INFO:teuthology.orchestra.run.vm08.stdout:Adding system user ceph....done 2026-03-20T17:56:46.643 INFO:teuthology.orchestra.run.vm08.stdout:Setting system user ceph properties....done 2026-03-20T17:56:46.647 INFO:teuthology.orchestra.run.vm08.stdout:chown: cannot access '/var/log/ceph/*.log*': No such file or directory 2026-03-20T17:56:46.692 INFO:teuthology.orchestra.run.vm06.stdout:Created symlink /etc/systemd/system/multi-user.target.wants/ceph-mds.target → /lib/systemd/system/ceph-mds.target. 2026-03-20T17:56:46.692 INFO:teuthology.orchestra.run.vm06.stdout:Created symlink /etc/systemd/system/ceph.target.wants/ceph-mds.target → /lib/systemd/system/ceph-mds.target. 2026-03-20T17:56:46.715 INFO:teuthology.orchestra.run.vm08.stdout:Created symlink /etc/systemd/system/multi-user.target.wants/ceph.target → /lib/systemd/system/ceph.target. 2026-03-20T17:56:46.744 INFO:teuthology.orchestra.run.vm05.stdout:Setting up libthrift-0.16.0:amd64 (0.16.0-2) ... 2026-03-20T17:56:46.746 INFO:teuthology.orchestra.run.vm05.stdout:Setting up librados2 (20.2.0-712-g70f8415b-1jammy) ... 2026-03-20T17:56:46.749 INFO:teuthology.orchestra.run.vm05.stdout:Setting up librgw2 (20.2.0-712-g70f8415b-1jammy) ... 2026-03-20T17:56:46.751 INFO:teuthology.orchestra.run.vm05.stdout:Setting up libsqlite3-mod-ceph (20.2.0-712-g70f8415b-1jammy) ... 2026-03-20T17:56:46.754 INFO:teuthology.orchestra.run.vm05.stdout:Setting up python3-kubernetes (12.0.1-1ubuntu1) ... 2026-03-20T17:56:46.814 INFO:teuthology.orchestra.run.vm02.stdout:Created symlink /etc/systemd/system/smartd.service → /lib/systemd/system/smartmontools.service. 2026-03-20T17:56:46.814 INFO:teuthology.orchestra.run.vm02.stdout:Created symlink /etc/systemd/system/multi-user.target.wants/smartmontools.service → /lib/systemd/system/smartmontools.service. 2026-03-20T17:56:46.962 INFO:teuthology.orchestra.run.vm08.stdout:Created symlink /etc/systemd/system/multi-user.target.wants/rbdmap.service → /lib/systemd/system/rbdmap.service. 2026-03-20T17:56:47.091 INFO:teuthology.orchestra.run.vm06.stdout:Setting up ceph-mgr (20.2.0-712-g70f8415b-1jammy) ... 2026-03-20T17:56:47.166 INFO:teuthology.orchestra.run.vm06.stdout:Created symlink /etc/systemd/system/multi-user.target.wants/ceph-mgr.target → /lib/systemd/system/ceph-mgr.target. 2026-03-20T17:56:47.166 INFO:teuthology.orchestra.run.vm06.stdout:Created symlink /etc/systemd/system/ceph.target.wants/ceph-mgr.target → /lib/systemd/system/ceph-mgr.target. 2026-03-20T17:56:47.357 INFO:teuthology.orchestra.run.vm05.stdout:Setting up libcephfs2 (20.2.0-712-g70f8415b-1jammy) ... 2026-03-20T17:56:47.357 INFO:teuthology.orchestra.run.vm02.stdout:Setting up python3-iniconfig (1.1.1-2) ... 2026-03-20T17:56:47.359 INFO:teuthology.orchestra.run.vm05.stdout:Setting up libradosstriper1 (20.2.0-712-g70f8415b-1jammy) ... 2026-03-20T17:56:47.362 INFO:teuthology.orchestra.run.vm05.stdout:Setting up librbd1 (20.2.0-712-g70f8415b-1jammy) ... 2026-03-20T17:56:47.364 INFO:teuthology.orchestra.run.vm05.stdout:Setting up ceph-mgr-modules-core (20.2.0-712-g70f8415b-1jammy) ... 2026-03-20T17:56:47.367 INFO:teuthology.orchestra.run.vm05.stdout:Setting up ceph-fuse (20.2.0-712-g70f8415b-1jammy) ... 2026-03-20T17:56:47.369 INFO:teuthology.orchestra.run.vm08.stdout:Setting up ceph-test (20.2.0-712-g70f8415b-1jammy) ... 2026-03-20T17:56:47.371 INFO:teuthology.orchestra.run.vm08.stdout:Setting up radosgw (20.2.0-712-g70f8415b-1jammy) ... 2026-03-20T17:56:47.427 INFO:teuthology.orchestra.run.vm02.stdout:Setting up libdouble-conversion3:amd64 (3.1.7-4) ... 2026-03-20T17:56:47.428 INFO:teuthology.orchestra.run.vm05.stdout:Created symlink /etc/systemd/system/remote-fs.target.wants/ceph-fuse.target → /lib/systemd/system/ceph-fuse.target. 2026-03-20T17:56:47.428 INFO:teuthology.orchestra.run.vm05.stdout:Created symlink /etc/systemd/system/ceph.target.wants/ceph-fuse.target → /lib/systemd/system/ceph-fuse.target. 2026-03-20T17:56:47.430 INFO:teuthology.orchestra.run.vm02.stdout:Setting up nvme-cli (1.16-3ubuntu0.3) ... 2026-03-20T17:56:47.495 INFO:teuthology.orchestra.run.vm02.stdout:Created symlink /etc/systemd/system/default.target.wants/nvmefc-boot-connections.service → /lib/systemd/system/nvmefc-boot-connections.service. 2026-03-20T17:56:47.533 INFO:teuthology.orchestra.run.vm06.stdout:Setting up ceph-osd (20.2.0-712-g70f8415b-1jammy) ... 2026-03-20T17:56:47.617 INFO:teuthology.orchestra.run.vm06.stdout:Created symlink /etc/systemd/system/multi-user.target.wants/ceph-osd.target → /lib/systemd/system/ceph-osd.target. 2026-03-20T17:56:47.617 INFO:teuthology.orchestra.run.vm06.stdout:Created symlink /etc/systemd/system/ceph.target.wants/ceph-osd.target → /lib/systemd/system/ceph-osd.target. 2026-03-20T17:56:47.646 INFO:teuthology.orchestra.run.vm08.stdout:Created symlink /etc/systemd/system/multi-user.target.wants/ceph-radosgw.target → /lib/systemd/system/ceph-radosgw.target. 2026-03-20T17:56:47.646 INFO:teuthology.orchestra.run.vm08.stdout:Created symlink /etc/systemd/system/ceph.target.wants/ceph-radosgw.target → /lib/systemd/system/ceph-radosgw.target. 2026-03-20T17:56:47.728 INFO:teuthology.orchestra.run.vm02.stdout:Created symlink /etc/systemd/system/default.target.wants/nvmf-autoconnect.service → /lib/systemd/system/nvmf-autoconnect.service. 2026-03-20T17:56:47.872 INFO:teuthology.orchestra.run.vm05.stdout:Setting up libcephfs-dev (20.2.0-712-g70f8415b-1jammy) ... 2026-03-20T17:56:47.874 INFO:teuthology.orchestra.run.vm05.stdout:Setting up python3-rados (20.2.0-712-g70f8415b-1jammy) ... 2026-03-20T17:56:47.877 INFO:teuthology.orchestra.run.vm05.stdout:Setting up libcephfs-daemon (20.2.0-712-g70f8415b-1jammy) ... 2026-03-20T17:56:47.880 INFO:teuthology.orchestra.run.vm05.stdout:Setting up python3-rbd (20.2.0-712-g70f8415b-1jammy) ... 2026-03-20T17:56:47.883 INFO:teuthology.orchestra.run.vm05.stdout:Setting up rbd-fuse (20.2.0-712-g70f8415b-1jammy) ... 2026-03-20T17:56:47.885 INFO:teuthology.orchestra.run.vm05.stdout:Setting up python3-rgw (20.2.0-712-g70f8415b-1jammy) ... 2026-03-20T17:56:47.888 INFO:teuthology.orchestra.run.vm05.stdout:Setting up python3-cephfs (20.2.0-712-g70f8415b-1jammy) ... 2026-03-20T17:56:47.891 INFO:teuthology.orchestra.run.vm05.stdout:Setting up ceph-common (20.2.0-712-g70f8415b-1jammy) ... 2026-03-20T17:56:47.928 INFO:teuthology.orchestra.run.vm05.stdout:Adding group ceph....done 2026-03-20T17:56:47.963 INFO:teuthology.orchestra.run.vm05.stdout:Adding system user ceph....done 2026-03-20T17:56:47.972 INFO:teuthology.orchestra.run.vm05.stdout:Setting system user ceph properties....done 2026-03-20T17:56:47.976 INFO:teuthology.orchestra.run.vm05.stdout:chown: cannot access '/var/log/ceph/*.log*': No such file or directory 2026-03-20T17:56:48.005 INFO:teuthology.orchestra.run.vm06.stdout:Setting up ceph-mgr-k8sevents (20.2.0-712-g70f8415b-1jammy) ... 2026-03-20T17:56:48.007 INFO:teuthology.orchestra.run.vm06.stdout:Setting up ceph-mgr-diskprediction-local (20.2.0-712-g70f8415b-1jammy) ... 2026-03-20T17:56:48.017 INFO:teuthology.orchestra.run.vm08.stdout:Setting up ceph-base (20.2.0-712-g70f8415b-1jammy) ... 2026-03-20T17:56:48.020 INFO:teuthology.orchestra.run.vm06.stdout:Setting up ceph-mon (20.2.0-712-g70f8415b-1jammy) ... 2026-03-20T17:56:48.044 INFO:teuthology.orchestra.run.vm05.stdout:Created symlink /etc/systemd/system/multi-user.target.wants/ceph.target → /lib/systemd/system/ceph.target. 2026-03-20T17:56:48.086 INFO:teuthology.orchestra.run.vm06.stdout:Created symlink /etc/systemd/system/multi-user.target.wants/ceph-mon.target → /lib/systemd/system/ceph-mon.target. 2026-03-20T17:56:48.086 INFO:teuthology.orchestra.run.vm06.stdout:Created symlink /etc/systemd/system/ceph.target.wants/ceph-mon.target → /lib/systemd/system/ceph-mon.target. 2026-03-20T17:56:48.111 INFO:teuthology.orchestra.run.vm08.stdout:Created symlink /etc/systemd/system/ceph.target.wants/ceph-crash.service → /lib/systemd/system/ceph-crash.service. 2026-03-20T17:56:48.133 INFO:teuthology.orchestra.run.vm02.stdout:nvmf-connect.target is a disabled or a static unit, not starting it. 2026-03-20T17:56:48.153 INFO:teuthology.orchestra.run.vm02.stdout:Setting up cephadm (20.2.0-712-g70f8415b-1jammy) ... 2026-03-20T17:56:48.205 INFO:teuthology.orchestra.run.vm02.stdout:Adding system user cephadm....done 2026-03-20T17:56:48.214 INFO:teuthology.orchestra.run.vm02.stdout:Setting up python3-jaraco.classes (3.2.1-3) ... 2026-03-20T17:56:48.289 INFO:teuthology.orchestra.run.vm02.stdout:Setting up python-asyncssh-doc (2.5.0-1ubuntu0.1) ... 2026-03-20T17:56:48.292 INFO:teuthology.orchestra.run.vm02.stdout:Setting up python3-jaraco.functools (3.4.0-2) ... 2026-03-20T17:56:48.298 INFO:teuthology.orchestra.run.vm05.stdout:Created symlink /etc/systemd/system/multi-user.target.wants/rbdmap.service → /lib/systemd/system/rbdmap.service. 2026-03-20T17:56:48.358 INFO:teuthology.orchestra.run.vm02.stdout:Setting up python3-repoze.lru (0.7-2) ... 2026-03-20T17:56:48.430 INFO:teuthology.orchestra.run.vm02.stdout:Setting up liboath0:amd64 (2.6.7-3ubuntu0.1) ... 2026-03-20T17:56:48.432 INFO:teuthology.orchestra.run.vm02.stdout:Setting up python3-py (1.10.0-1) ... 2026-03-20T17:56:48.480 INFO:teuthology.orchestra.run.vm06.stdout:Setting up ceph-mgr-cephadm (20.2.0-712-g70f8415b-1jammy) ... 2026-03-20T17:56:48.493 INFO:teuthology.orchestra.run.vm06.stdout:Setting up ceph (20.2.0-712-g70f8415b-1jammy) ... 2026-03-20T17:56:48.495 INFO:teuthology.orchestra.run.vm06.stdout:Setting up ceph-mgr-dashboard (20.2.0-712-g70f8415b-1jammy) ... 2026-03-20T17:56:48.507 INFO:teuthology.orchestra.run.vm06.stdout:Setting up ceph-volume (20.2.0-712-g70f8415b-1jammy) ... 2026-03-20T17:56:48.520 INFO:teuthology.orchestra.run.vm08.stdout:Setting up ceph-mds (20.2.0-712-g70f8415b-1jammy) ... 2026-03-20T17:56:48.527 INFO:teuthology.orchestra.run.vm02.stdout:Setting up python3-joblib (0.17.0-4ubuntu1) ... 2026-03-20T17:56:48.587 INFO:teuthology.orchestra.run.vm08.stdout:Created symlink /etc/systemd/system/multi-user.target.wants/ceph-mds.target → /lib/systemd/system/ceph-mds.target. 2026-03-20T17:56:48.588 INFO:teuthology.orchestra.run.vm08.stdout:Created symlink /etc/systemd/system/ceph.target.wants/ceph-mds.target → /lib/systemd/system/ceph-mds.target. 2026-03-20T17:56:48.627 INFO:teuthology.orchestra.run.vm06.stdout:Processing triggers for man-db (2.10.2-1) ... 2026-03-20T17:56:48.655 INFO:teuthology.orchestra.run.vm02.stdout:Setting up python3-cachetools (5.0.0-1) ... 2026-03-20T17:56:48.709 INFO:teuthology.orchestra.run.vm05.stdout:Setting up ceph-test (20.2.0-712-g70f8415b-1jammy) ... 2026-03-20T17:56:48.709 INFO:teuthology.orchestra.run.vm06.stdout:Processing triggers for libc-bin (2.35-0ubuntu3.13) ... 2026-03-20T17:56:48.711 INFO:teuthology.orchestra.run.vm05.stdout:Setting up radosgw (20.2.0-712-g70f8415b-1jammy) ... 2026-03-20T17:56:48.731 INFO:teuthology.orchestra.run.vm02.stdout:Setting up python3-threadpoolctl (3.1.0-1) ... 2026-03-20T17:56:48.804 INFO:teuthology.orchestra.run.vm02.stdout:Setting up python3-ceph-argparse (20.2.0-712-g70f8415b-1jammy) ... 2026-03-20T17:56:48.877 INFO:teuthology.orchestra.run.vm02.stdout:Setting up python3-sklearn-lib:amd64 (0.23.2-5ubuntu6) ... 2026-03-20T17:56:48.879 INFO:teuthology.orchestra.run.vm02.stdout:Setting up libnbd0 (1.10.5-1) ... 2026-03-20T17:56:48.881 INFO:teuthology.orchestra.run.vm02.stdout:Setting up libfuse2:amd64 (2.9.9-5ubuntu3) ... 2026-03-20T17:56:48.884 INFO:teuthology.orchestra.run.vm02.stdout:Setting up libpcre2-16-0:amd64 (10.39-3ubuntu0.1) ... 2026-03-20T17:56:48.887 INFO:teuthology.orchestra.run.vm02.stdout:Setting up python3-psutil (5.9.0-1build1) ... 2026-03-20T17:56:48.966 INFO:teuthology.orchestra.run.vm08.stdout:Setting up ceph-mgr (20.2.0-712-g70f8415b-1jammy) ... 2026-03-20T17:56:48.983 INFO:teuthology.orchestra.run.vm05.stdout:Created symlink /etc/systemd/system/multi-user.target.wants/ceph-radosgw.target → /lib/systemd/system/ceph-radosgw.target. 2026-03-20T17:56:48.983 INFO:teuthology.orchestra.run.vm05.stdout:Created symlink /etc/systemd/system/ceph.target.wants/ceph-radosgw.target → /lib/systemd/system/ceph-radosgw.target. 2026-03-20T17:56:49.009 INFO:teuthology.orchestra.run.vm02.stdout:Setting up python3-natsort (8.0.2-1) ... 2026-03-20T17:56:49.023 INFO:teuthology.orchestra.run.vm06.stdout: 2026-03-20T17:56:49.023 INFO:teuthology.orchestra.run.vm06.stdout:Running kernel seems to be up-to-date. 2026-03-20T17:56:49.023 INFO:teuthology.orchestra.run.vm06.stdout: 2026-03-20T17:56:49.023 INFO:teuthology.orchestra.run.vm06.stdout:Services to be restarted: 2026-03-20T17:56:49.028 INFO:teuthology.orchestra.run.vm06.stdout: systemctl restart apache-htcacheclean.service 2026-03-20T17:56:49.034 INFO:teuthology.orchestra.run.vm06.stdout: systemctl restart rsyslog.service 2026-03-20T17:56:49.037 INFO:teuthology.orchestra.run.vm06.stdout: 2026-03-20T17:56:49.037 INFO:teuthology.orchestra.run.vm06.stdout:Service restarts being deferred: 2026-03-20T17:56:49.037 INFO:teuthology.orchestra.run.vm06.stdout: systemctl restart networkd-dispatcher.service 2026-03-20T17:56:49.037 INFO:teuthology.orchestra.run.vm06.stdout: systemctl restart unattended-upgrades.service 2026-03-20T17:56:49.037 INFO:teuthology.orchestra.run.vm06.stdout: 2026-03-20T17:56:49.037 INFO:teuthology.orchestra.run.vm06.stdout:No containers need to be restarted. 2026-03-20T17:56:49.037 INFO:teuthology.orchestra.run.vm06.stdout: 2026-03-20T17:56:49.037 INFO:teuthology.orchestra.run.vm06.stdout:No user sessions are running outdated binaries. 2026-03-20T17:56:49.037 INFO:teuthology.orchestra.run.vm06.stdout: 2026-03-20T17:56:49.037 INFO:teuthology.orchestra.run.vm06.stdout:No VM guests are running outdated hypervisor (qemu) binaries on this host. 2026-03-20T17:56:49.055 INFO:teuthology.orchestra.run.vm08.stdout:Created symlink /etc/systemd/system/multi-user.target.wants/ceph-mgr.target → /lib/systemd/system/ceph-mgr.target. 2026-03-20T17:56:49.055 INFO:teuthology.orchestra.run.vm08.stdout:Created symlink /etc/systemd/system/ceph.target.wants/ceph-mgr.target → /lib/systemd/system/ceph-mgr.target. 2026-03-20T17:56:49.085 INFO:teuthology.orchestra.run.vm02.stdout:Setting up libcephfs-proxy2 (20.2.0-712-g70f8415b-1jammy) ... 2026-03-20T17:56:49.087 INFO:teuthology.orchestra.run.vm02.stdout:Setting up python3-routes (2.5.1-1ubuntu1) ... 2026-03-20T17:56:49.162 INFO:teuthology.orchestra.run.vm02.stdout:Setting up python3-simplejson (3.17.6-1build1) ... 2026-03-20T17:56:49.247 INFO:teuthology.orchestra.run.vm02.stdout:Setting up python3-pygments (2.11.2+dfsg-2ubuntu0.1) ... 2026-03-20T17:56:49.409 INFO:teuthology.orchestra.run.vm05.stdout:Setting up ceph-base (20.2.0-712-g70f8415b-1jammy) ... 2026-03-20T17:56:49.444 INFO:teuthology.orchestra.run.vm08.stdout:Setting up ceph-osd (20.2.0-712-g70f8415b-1jammy) ... 2026-03-20T17:56:49.499 INFO:teuthology.orchestra.run.vm05.stdout:Created symlink /etc/systemd/system/ceph.target.wants/ceph-crash.service → /lib/systemd/system/ceph-crash.service. 2026-03-20T17:56:49.521 INFO:teuthology.orchestra.run.vm08.stdout:Created symlink /etc/systemd/system/multi-user.target.wants/ceph-osd.target → /lib/systemd/system/ceph-osd.target. 2026-03-20T17:56:49.521 INFO:teuthology.orchestra.run.vm08.stdout:Created symlink /etc/systemd/system/ceph.target.wants/ceph-osd.target → /lib/systemd/system/ceph-osd.target. 2026-03-20T17:56:49.524 INFO:teuthology.orchestra.run.vm02.stdout:Setting up qttranslations5-l10n (5.15.3-1) ... 2026-03-20T17:56:49.526 INFO:teuthology.orchestra.run.vm02.stdout:Setting up python3-wcwidth (0.2.5+dfsg1-1) ... 2026-03-20T17:56:49.619 INFO:teuthology.orchestra.run.vm02.stdout:Setting up python3-asyncssh (2.5.0-1ubuntu0.1) ... 2026-03-20T17:56:49.761 INFO:teuthology.orchestra.run.vm02.stdout:Setting up python3-cheroot (8.5.2+ds1-1ubuntu3.1) ... 2026-03-20T17:56:49.851 INFO:teuthology.orchestra.run.vm02.stdout:Setting up python3-jaraco.text (3.6.0-2) ... 2026-03-20T17:56:49.863 INFO:teuthology.orchestra.run.vm06.stderr:W: --force-yes is deprecated, use one of the options starting with --allow instead. 2026-03-20T17:56:49.866 DEBUG:teuthology.orchestra.run.vm06:> sudo DEBIAN_FRONTEND=noninteractive apt-get -y --force-yes -o Dpkg::Options::="--force-confdef" -o Dpkg::Options::="--force-confold" install python3-jmespath python3-xmltodict s3cmd 2026-03-20T17:56:49.942 INFO:teuthology.orchestra.run.vm06.stdout:Reading package lists... 2026-03-20T17:56:49.996 INFO:teuthology.orchestra.run.vm08.stdout:Setting up ceph-mgr-k8sevents (20.2.0-712-g70f8415b-1jammy) ... 2026-03-20T17:56:49.996 INFO:teuthology.orchestra.run.vm05.stdout:Setting up ceph-mds (20.2.0-712-g70f8415b-1jammy) ... 2026-03-20T17:56:49.997 INFO:teuthology.orchestra.run.vm02.stdout:Setting up socat (1.7.4.1-3ubuntu4) ... 2026-03-20T17:56:49.999 INFO:teuthology.orchestra.run.vm02.stdout:Setting up python3-ceph-common (20.2.0-712-g70f8415b-1jammy) ... 2026-03-20T17:56:49.999 INFO:teuthology.orchestra.run.vm08.stdout:Setting up ceph-mgr-diskprediction-local (20.2.0-712-g70f8415b-1jammy) ... 2026-03-20T17:56:50.013 INFO:teuthology.orchestra.run.vm08.stdout:Setting up ceph-mon (20.2.0-712-g70f8415b-1jammy) ... 2026-03-20T17:56:50.062 INFO:teuthology.orchestra.run.vm05.stdout:Created symlink /etc/systemd/system/multi-user.target.wants/ceph-mds.target → /lib/systemd/system/ceph-mds.target. 2026-03-20T17:56:50.062 INFO:teuthology.orchestra.run.vm05.stdout:Created symlink /etc/systemd/system/ceph.target.wants/ceph-mds.target → /lib/systemd/system/ceph-mds.target. 2026-03-20T17:56:50.074 INFO:teuthology.orchestra.run.vm08.stdout:Created symlink /etc/systemd/system/multi-user.target.wants/ceph-mon.target → /lib/systemd/system/ceph-mon.target. 2026-03-20T17:56:50.074 INFO:teuthology.orchestra.run.vm08.stdout:Created symlink /etc/systemd/system/ceph.target.wants/ceph-mon.target → /lib/systemd/system/ceph-mon.target. 2026-03-20T17:56:50.094 INFO:teuthology.orchestra.run.vm06.stdout:Building dependency tree... 2026-03-20T17:56:50.095 INFO:teuthology.orchestra.run.vm06.stdout:Reading state information... 2026-03-20T17:56:50.098 INFO:teuthology.orchestra.run.vm02.stdout:Setting up python3-sklearn (0.23.2-5ubuntu6) ... 2026-03-20T17:56:50.229 INFO:teuthology.orchestra.run.vm06.stdout:The following packages were automatically installed and are no longer required: 2026-03-20T17:56:50.229 INFO:teuthology.orchestra.run.vm06.stdout: kpartx libboost-iostreams1.74.0 libboost-thread1.74.0 libpmemobj1 2026-03-20T17:56:50.229 INFO:teuthology.orchestra.run.vm06.stdout: libsgutils2-2 sg3-utils sg3-utils-udev 2026-03-20T17:56:50.229 INFO:teuthology.orchestra.run.vm06.stdout:Use 'sudo apt autoremove' to remove them. 2026-03-20T17:56:50.240 INFO:teuthology.orchestra.run.vm06.stdout:The following NEW packages will be installed: 2026-03-20T17:56:50.240 INFO:teuthology.orchestra.run.vm06.stdout: python3-jmespath python3-xmltodict s3cmd 2026-03-20T17:56:50.263 INFO:teuthology.orchestra.run.vm06.stdout:0 upgraded, 3 newly installed, 0 to remove and 36 not upgraded. 2026-03-20T17:56:50.263 INFO:teuthology.orchestra.run.vm06.stdout:Need to get 155 kB of archives. 2026-03-20T17:56:50.263 INFO:teuthology.orchestra.run.vm06.stdout:After this operation, 678 kB of additional disk space will be used. 2026-03-20T17:56:50.263 INFO:teuthology.orchestra.run.vm06.stdout:Get:1 http://archive.ubuntu.com/ubuntu jammy/main amd64 python3-jmespath all 0.10.0-1 [21.7 kB] 2026-03-20T17:56:50.280 INFO:teuthology.orchestra.run.vm06.stdout:Get:2 http://archive.ubuntu.com/ubuntu jammy/universe amd64 python3-xmltodict all 0.12.0-2 [12.6 kB] 2026-03-20T17:56:50.281 INFO:teuthology.orchestra.run.vm06.stdout:Get:3 http://archive.ubuntu.com/ubuntu jammy/universe amd64 s3cmd all 2.2.0-1 [120 kB] 2026-03-20T17:56:50.429 INFO:teuthology.orchestra.run.vm08.stdout:Setting up ceph-mgr-cephadm (20.2.0-712-g70f8415b-1jammy) ... 2026-03-20T17:56:50.442 INFO:teuthology.orchestra.run.vm08.stdout:Setting up ceph (20.2.0-712-g70f8415b-1jammy) ... 2026-03-20T17:56:50.445 INFO:teuthology.orchestra.run.vm08.stdout:Setting up ceph-mgr-dashboard (20.2.0-712-g70f8415b-1jammy) ... 2026-03-20T17:56:50.458 INFO:teuthology.orchestra.run.vm08.stdout:Setting up ceph-volume (20.2.0-712-g70f8415b-1jammy) ... 2026-03-20T17:56:50.471 INFO:teuthology.orchestra.run.vm05.stdout:Setting up ceph-mgr (20.2.0-712-g70f8415b-1jammy) ... 2026-03-20T17:56:50.491 INFO:teuthology.orchestra.run.vm06.stdout:Fetched 155 kB in 0s (2812 kB/s) 2026-03-20T17:56:50.506 INFO:teuthology.orchestra.run.vm06.stdout:Selecting previously unselected package python3-jmespath. 2026-03-20T17:56:50.537 INFO:teuthology.orchestra.run.vm06.stdout:(Reading database ... (Reading database ... 5% (Reading database ... 10% (Reading database ... 15% (Reading database ... 20% (Reading database ... 25% (Reading database ... 30% (Reading database ... 35% (Reading database ... 40% (Reading database ... 45% (Reading database ... 50% (Reading database ... 55% (Reading database ... 60% (Reading database ... 65% (Reading database ... 70% (Reading database ... 75% (Reading database ... 80% (Reading database ... 85% (Reading database ... 90% (Reading database ... 95% (Reading database ... 100% (Reading database ... 126082 files and directories currently installed.) 2026-03-20T17:56:50.540 INFO:teuthology.orchestra.run.vm06.stdout:Preparing to unpack .../python3-jmespath_0.10.0-1_all.deb ... 2026-03-20T17:56:50.541 INFO:teuthology.orchestra.run.vm06.stdout:Unpacking python3-jmespath (0.10.0-1) ... 2026-03-20T17:56:50.550 INFO:teuthology.orchestra.run.vm05.stdout:Created symlink /etc/systemd/system/multi-user.target.wants/ceph-mgr.target → /lib/systemd/system/ceph-mgr.target. 2026-03-20T17:56:50.550 INFO:teuthology.orchestra.run.vm05.stdout:Created symlink /etc/systemd/system/ceph.target.wants/ceph-mgr.target → /lib/systemd/system/ceph-mgr.target. 2026-03-20T17:56:50.558 INFO:teuthology.orchestra.run.vm06.stdout:Selecting previously unselected package python3-xmltodict. 2026-03-20T17:56:50.565 INFO:teuthology.orchestra.run.vm06.stdout:Preparing to unpack .../python3-xmltodict_0.12.0-2_all.deb ... 2026-03-20T17:56:50.566 INFO:teuthology.orchestra.run.vm06.stdout:Unpacking python3-xmltodict (0.12.0-2) ... 2026-03-20T17:56:50.580 INFO:teuthology.orchestra.run.vm08.stdout:Processing triggers for man-db (2.10.2-1) ... 2026-03-20T17:56:50.583 INFO:teuthology.orchestra.run.vm06.stdout:Selecting previously unselected package s3cmd. 2026-03-20T17:56:50.590 INFO:teuthology.orchestra.run.vm06.stdout:Preparing to unpack .../archives/s3cmd_2.2.0-1_all.deb ... 2026-03-20T17:56:50.591 INFO:teuthology.orchestra.run.vm06.stdout:Unpacking s3cmd (2.2.0-1) ... 2026-03-20T17:56:50.626 INFO:teuthology.orchestra.run.vm06.stdout:Setting up s3cmd (2.2.0-1) ... 2026-03-20T17:56:50.654 INFO:teuthology.orchestra.run.vm02.stdout:Setting up libqt5core5a:amd64 (5.15.3+dfsg-2ubuntu0.2) ... 2026-03-20T17:56:50.658 INFO:teuthology.orchestra.run.vm08.stdout:Processing triggers for libc-bin (2.35-0ubuntu3.13) ... 2026-03-20T17:56:50.660 INFO:teuthology.orchestra.run.vm02.stdout:Setting up python3-toml (0.10.2-1) ... 2026-03-20T17:56:50.718 INFO:teuthology.orchestra.run.vm06.stdout:Setting up python3-xmltodict (0.12.0-2) ... 2026-03-20T17:56:50.731 INFO:teuthology.orchestra.run.vm02.stdout:Setting up librdkafka1:amd64 (1.8.0-1build1) ... 2026-03-20T17:56:50.734 INFO:teuthology.orchestra.run.vm02.stdout:Setting up xmlstarlet (1.6.1-2.1) ... 2026-03-20T17:56:50.736 INFO:teuthology.orchestra.run.vm02.stdout:Setting up python3-pluggy (0.13.0-7.1) ... 2026-03-20T17:56:50.784 INFO:teuthology.orchestra.run.vm06.stdout:Setting up python3-jmespath (0.10.0-1) ... 2026-03-20T17:56:50.805 INFO:teuthology.orchestra.run.vm02.stdout:Setting up python3-zc.lockfile (2.0-1) ... 2026-03-20T17:56:50.857 INFO:teuthology.orchestra.run.vm06.stdout:Processing triggers for man-db (2.10.2-1) ... 2026-03-20T17:56:50.875 INFO:teuthology.orchestra.run.vm02.stdout:Setting up libqt5dbus5:amd64 (5.15.3+dfsg-2ubuntu0.2) ... 2026-03-20T17:56:50.877 INFO:teuthology.orchestra.run.vm02.stdout:Setting up python3-rsa (4.8-1) ... 2026-03-20T17:56:50.952 INFO:teuthology.orchestra.run.vm02.stdout:Setting up python3-tempora (4.1.2-1) ... 2026-03-20T17:56:50.971 INFO:teuthology.orchestra.run.vm08.stdout: 2026-03-20T17:56:50.971 INFO:teuthology.orchestra.run.vm08.stdout:Running kernel seems to be up-to-date. 2026-03-20T17:56:50.971 INFO:teuthology.orchestra.run.vm08.stdout: 2026-03-20T17:56:50.971 INFO:teuthology.orchestra.run.vm08.stdout:Services to be restarted: 2026-03-20T17:56:50.974 INFO:teuthology.orchestra.run.vm08.stdout: systemctl restart apache-htcacheclean.service 2026-03-20T17:56:50.981 INFO:teuthology.orchestra.run.vm08.stdout: systemctl restart rsyslog.service 2026-03-20T17:56:50.984 INFO:teuthology.orchestra.run.vm08.stdout: 2026-03-20T17:56:50.984 INFO:teuthology.orchestra.run.vm08.stdout:Service restarts being deferred: 2026-03-20T17:56:50.984 INFO:teuthology.orchestra.run.vm08.stdout: systemctl restart networkd-dispatcher.service 2026-03-20T17:56:50.985 INFO:teuthology.orchestra.run.vm08.stdout: systemctl restart unattended-upgrades.service 2026-03-20T17:56:50.985 INFO:teuthology.orchestra.run.vm08.stdout: 2026-03-20T17:56:50.985 INFO:teuthology.orchestra.run.vm08.stdout:No containers need to be restarted. 2026-03-20T17:56:50.985 INFO:teuthology.orchestra.run.vm08.stdout: 2026-03-20T17:56:50.985 INFO:teuthology.orchestra.run.vm08.stdout:No user sessions are running outdated binaries. 2026-03-20T17:56:50.985 INFO:teuthology.orchestra.run.vm08.stdout: 2026-03-20T17:56:50.985 INFO:teuthology.orchestra.run.vm08.stdout:No VM guests are running outdated hypervisor (qemu) binaries on this host. 2026-03-20T17:56:50.993 INFO:teuthology.orchestra.run.vm05.stdout:Setting up ceph-osd (20.2.0-712-g70f8415b-1jammy) ... 2026-03-20T17:56:51.025 INFO:teuthology.orchestra.run.vm02.stdout:Setting up python3-prettytable (2.5.0-2) ... 2026-03-20T17:56:51.076 INFO:teuthology.orchestra.run.vm05.stdout:Created symlink /etc/systemd/system/multi-user.target.wants/ceph-osd.target → /lib/systemd/system/ceph-osd.target. 2026-03-20T17:56:51.076 INFO:teuthology.orchestra.run.vm05.stdout:Created symlink /etc/systemd/system/ceph.target.wants/ceph-osd.target → /lib/systemd/system/ceph-osd.target. 2026-03-20T17:56:51.101 INFO:teuthology.orchestra.run.vm02.stdout:Setting up liblttng-ust1:amd64 (2.13.1-1ubuntu1) ... 2026-03-20T17:56:51.103 INFO:teuthology.orchestra.run.vm02.stdout:Setting up python3-websocket (1.2.3-1) ... 2026-03-20T17:56:51.186 INFO:teuthology.orchestra.run.vm02.stdout:Setting up libonig5:amd64 (6.9.7.1-2build1) ... 2026-03-20T17:56:51.188 INFO:teuthology.orchestra.run.vm02.stdout:Setting up python3-requests-oauthlib (1.3.0+ds-0.1) ... 2026-03-20T17:56:51.190 INFO:teuthology.orchestra.run.vm06.stdout: 2026-03-20T17:56:51.190 INFO:teuthology.orchestra.run.vm06.stdout:Running kernel seems to be up-to-date. 2026-03-20T17:56:51.190 INFO:teuthology.orchestra.run.vm06.stdout: 2026-03-20T17:56:51.191 INFO:teuthology.orchestra.run.vm06.stdout:Services to be restarted: 2026-03-20T17:56:51.193 INFO:teuthology.orchestra.run.vm06.stdout: systemctl restart apache-htcacheclean.service 2026-03-20T17:56:51.200 INFO:teuthology.orchestra.run.vm06.stdout: systemctl restart rsyslog.service 2026-03-20T17:56:51.203 INFO:teuthology.orchestra.run.vm06.stdout: 2026-03-20T17:56:51.203 INFO:teuthology.orchestra.run.vm06.stdout:Service restarts being deferred: 2026-03-20T17:56:51.203 INFO:teuthology.orchestra.run.vm06.stdout: systemctl restart networkd-dispatcher.service 2026-03-20T17:56:51.203 INFO:teuthology.orchestra.run.vm06.stdout: systemctl restart unattended-upgrades.service 2026-03-20T17:56:51.203 INFO:teuthology.orchestra.run.vm06.stdout: 2026-03-20T17:56:51.203 INFO:teuthology.orchestra.run.vm06.stdout:No containers need to be restarted. 2026-03-20T17:56:51.203 INFO:teuthology.orchestra.run.vm06.stdout: 2026-03-20T17:56:51.203 INFO:teuthology.orchestra.run.vm06.stdout:No user sessions are running outdated binaries. 2026-03-20T17:56:51.203 INFO:teuthology.orchestra.run.vm06.stdout: 2026-03-20T17:56:51.203 INFO:teuthology.orchestra.run.vm06.stdout:No VM guests are running outdated hypervisor (qemu) binaries on this host. 2026-03-20T17:56:51.260 INFO:teuthology.orchestra.run.vm02.stdout:Setting up python3-webob (1:1.8.6-1.1ubuntu0.1) ... 2026-03-20T17:56:51.353 INFO:teuthology.orchestra.run.vm02.stdout:Setting up python3-jaraco.collections (3.4.0-2) ... 2026-03-20T17:56:51.423 INFO:teuthology.orchestra.run.vm02.stdout:Setting up libjq1:amd64 (1.6-2.1ubuntu3.1) ... 2026-03-20T17:56:51.425 INFO:teuthology.orchestra.run.vm02.stdout:Setting up python3-pytest (6.2.5-1ubuntu2) ... 2026-03-20T17:56:51.465 INFO:teuthology.orchestra.run.vm05.stdout:Setting up ceph-mgr-k8sevents (20.2.0-712-g70f8415b-1jammy) ... 2026-03-20T17:56:51.468 INFO:teuthology.orchestra.run.vm05.stdout:Setting up ceph-mgr-diskprediction-local (20.2.0-712-g70f8415b-1jammy) ... 2026-03-20T17:56:51.481 INFO:teuthology.orchestra.run.vm05.stdout:Setting up ceph-mon (20.2.0-712-g70f8415b-1jammy) ... 2026-03-20T17:56:51.544 INFO:teuthology.orchestra.run.vm05.stdout:Created symlink /etc/systemd/system/multi-user.target.wants/ceph-mon.target → /lib/systemd/system/ceph-mon.target. 2026-03-20T17:56:51.544 INFO:teuthology.orchestra.run.vm05.stdout:Created symlink /etc/systemd/system/ceph.target.wants/ceph-mon.target → /lib/systemd/system/ceph-mon.target. 2026-03-20T17:56:51.571 INFO:teuthology.orchestra.run.vm02.stdout:Setting up python3-portend (3.0.0-1) ... 2026-03-20T17:56:51.637 INFO:teuthology.orchestra.run.vm02.stdout:Setting up libqt5network5:amd64 (5.15.3+dfsg-2ubuntu0.2) ... 2026-03-20T17:56:51.640 INFO:teuthology.orchestra.run.vm02.stdout:Setting up python3-google-auth (1.5.1-3) ... 2026-03-20T17:56:51.723 INFO:teuthology.orchestra.run.vm02.stdout:Setting up jq (1.6-2.1ubuntu3.1) ... 2026-03-20T17:56:51.725 INFO:teuthology.orchestra.run.vm02.stdout:Setting up python3-cherrypy3 (18.6.1-4) ... 2026-03-20T17:56:51.879 INFO:teuthology.orchestra.run.vm02.stdout:Setting up libthrift-0.16.0:amd64 (0.16.0-2) ... 2026-03-20T17:56:51.881 INFO:teuthology.orchestra.run.vm02.stdout:Setting up librados2 (20.2.0-712-g70f8415b-1jammy) ... 2026-03-20T17:56:51.883 INFO:teuthology.orchestra.run.vm02.stdout:Setting up librgw2 (20.2.0-712-g70f8415b-1jammy) ... 2026-03-20T17:56:51.886 INFO:teuthology.orchestra.run.vm02.stdout:Setting up libsqlite3-mod-ceph (20.2.0-712-g70f8415b-1jammy) ... 2026-03-20T17:56:51.888 INFO:teuthology.orchestra.run.vm02.stdout:Setting up python3-kubernetes (12.0.1-1ubuntu1) ... 2026-03-20T17:56:51.903 INFO:teuthology.orchestra.run.vm08.stderr:W: --force-yes is deprecated, use one of the options starting with --allow instead. 2026-03-20T17:56:51.906 DEBUG:teuthology.orchestra.run.vm08:> sudo DEBIAN_FRONTEND=noninteractive apt-get -y --force-yes -o Dpkg::Options::="--force-confdef" -o Dpkg::Options::="--force-confold" install python3-jmespath python3-xmltodict s3cmd 2026-03-20T17:56:51.981 INFO:teuthology.orchestra.run.vm08.stdout:Reading package lists... 2026-03-20T17:56:52.005 INFO:teuthology.orchestra.run.vm06.stderr:W: --force-yes is deprecated, use one of the options starting with --allow instead. 2026-03-20T17:56:52.007 DEBUG:teuthology.parallel:result is None 2026-03-20T17:56:52.037 INFO:teuthology.orchestra.run.vm05.stdout:Setting up ceph-mgr-cephadm (20.2.0-712-g70f8415b-1jammy) ... 2026-03-20T17:56:52.051 INFO:teuthology.orchestra.run.vm05.stdout:Setting up ceph (20.2.0-712-g70f8415b-1jammy) ... 2026-03-20T17:56:52.053 INFO:teuthology.orchestra.run.vm05.stdout:Setting up ceph-mgr-dashboard (20.2.0-712-g70f8415b-1jammy) ... 2026-03-20T17:56:52.066 INFO:teuthology.orchestra.run.vm05.stdout:Setting up ceph-volume (20.2.0-712-g70f8415b-1jammy) ... 2026-03-20T17:56:52.169 INFO:teuthology.orchestra.run.vm08.stdout:Building dependency tree... 2026-03-20T17:56:52.169 INFO:teuthology.orchestra.run.vm08.stdout:Reading state information... 2026-03-20T17:56:52.196 INFO:teuthology.orchestra.run.vm05.stdout:Processing triggers for man-db (2.10.2-1) ... 2026-03-20T17:56:52.278 INFO:teuthology.orchestra.run.vm05.stdout:Processing triggers for libc-bin (2.35-0ubuntu3.13) ... 2026-03-20T17:56:52.328 INFO:teuthology.orchestra.run.vm08.stdout:The following packages were automatically installed and are no longer required: 2026-03-20T17:56:52.328 INFO:teuthology.orchestra.run.vm08.stdout: kpartx libboost-iostreams1.74.0 libboost-thread1.74.0 libpmemobj1 2026-03-20T17:56:52.329 INFO:teuthology.orchestra.run.vm08.stdout: libsgutils2-2 sg3-utils sg3-utils-udev 2026-03-20T17:56:52.329 INFO:teuthology.orchestra.run.vm08.stdout:Use 'sudo apt autoremove' to remove them. 2026-03-20T17:56:52.343 INFO:teuthology.orchestra.run.vm08.stdout:The following NEW packages will be installed: 2026-03-20T17:56:52.343 INFO:teuthology.orchestra.run.vm08.stdout: python3-jmespath python3-xmltodict s3cmd 2026-03-20T17:56:52.421 INFO:teuthology.orchestra.run.vm08.stdout:0 upgraded, 3 newly installed, 0 to remove and 36 not upgraded. 2026-03-20T17:56:52.421 INFO:teuthology.orchestra.run.vm08.stdout:Need to get 155 kB of archives. 2026-03-20T17:56:52.421 INFO:teuthology.orchestra.run.vm08.stdout:After this operation, 678 kB of additional disk space will be used. 2026-03-20T17:56:52.421 INFO:teuthology.orchestra.run.vm08.stdout:Get:1 http://archive.ubuntu.com/ubuntu jammy/main amd64 python3-jmespath all 0.10.0-1 [21.7 kB] 2026-03-20T17:56:52.492 INFO:teuthology.orchestra.run.vm02.stdout:Setting up libcephfs2 (20.2.0-712-g70f8415b-1jammy) ... 2026-03-20T17:56:52.494 INFO:teuthology.orchestra.run.vm02.stdout:Setting up libradosstriper1 (20.2.0-712-g70f8415b-1jammy) ... 2026-03-20T17:56:52.497 INFO:teuthology.orchestra.run.vm02.stdout:Setting up librbd1 (20.2.0-712-g70f8415b-1jammy) ... 2026-03-20T17:56:52.498 INFO:teuthology.orchestra.run.vm08.stdout:Get:2 http://archive.ubuntu.com/ubuntu jammy/universe amd64 python3-xmltodict all 0.12.0-2 [12.6 kB] 2026-03-20T17:56:52.499 INFO:teuthology.orchestra.run.vm02.stdout:Setting up ceph-mgr-modules-core (20.2.0-712-g70f8415b-1jammy) ... 2026-03-20T17:56:52.502 INFO:teuthology.orchestra.run.vm02.stdout:Setting up ceph-fuse (20.2.0-712-g70f8415b-1jammy) ... 2026-03-20T17:56:52.507 INFO:teuthology.orchestra.run.vm08.stdout:Get:3 http://archive.ubuntu.com/ubuntu jammy/universe amd64 s3cmd all 2.2.0-1 [120 kB] 2026-03-20T17:56:52.567 INFO:teuthology.orchestra.run.vm02.stdout:Created symlink /etc/systemd/system/remote-fs.target.wants/ceph-fuse.target → /lib/systemd/system/ceph-fuse.target. 2026-03-20T17:56:52.567 INFO:teuthology.orchestra.run.vm02.stdout:Created symlink /etc/systemd/system/ceph.target.wants/ceph-fuse.target → /lib/systemd/system/ceph-fuse.target. 2026-03-20T17:56:52.591 INFO:teuthology.orchestra.run.vm05.stdout: 2026-03-20T17:56:52.591 INFO:teuthology.orchestra.run.vm05.stdout:Running kernel seems to be up-to-date. 2026-03-20T17:56:52.591 INFO:teuthology.orchestra.run.vm05.stdout: 2026-03-20T17:56:52.591 INFO:teuthology.orchestra.run.vm05.stdout:Services to be restarted: 2026-03-20T17:56:52.594 INFO:teuthology.orchestra.run.vm05.stdout: systemctl restart apache-htcacheclean.service 2026-03-20T17:56:52.600 INFO:teuthology.orchestra.run.vm05.stdout: systemctl restart rsyslog.service 2026-03-20T17:56:52.603 INFO:teuthology.orchestra.run.vm05.stdout: 2026-03-20T17:56:52.603 INFO:teuthology.orchestra.run.vm05.stdout:Service restarts being deferred: 2026-03-20T17:56:52.603 INFO:teuthology.orchestra.run.vm05.stdout: systemctl restart networkd-dispatcher.service 2026-03-20T17:56:52.603 INFO:teuthology.orchestra.run.vm05.stdout: systemctl restart unattended-upgrades.service 2026-03-20T17:56:52.603 INFO:teuthology.orchestra.run.vm05.stdout: 2026-03-20T17:56:52.603 INFO:teuthology.orchestra.run.vm05.stdout:No containers need to be restarted. 2026-03-20T17:56:52.603 INFO:teuthology.orchestra.run.vm05.stdout: 2026-03-20T17:56:52.603 INFO:teuthology.orchestra.run.vm05.stdout:No user sessions are running outdated binaries. 2026-03-20T17:56:52.603 INFO:teuthology.orchestra.run.vm05.stdout: 2026-03-20T17:56:52.603 INFO:teuthology.orchestra.run.vm05.stdout:No VM guests are running outdated hypervisor (qemu) binaries on this host. 2026-03-20T17:56:52.778 INFO:teuthology.orchestra.run.vm08.stdout:Fetched 155 kB in 0s (623 kB/s) 2026-03-20T17:56:52.791 INFO:teuthology.orchestra.run.vm08.stdout:Selecting previously unselected package python3-jmespath. 2026-03-20T17:56:52.820 INFO:teuthology.orchestra.run.vm08.stdout:(Reading database ... (Reading database ... 5% (Reading database ... 10% (Reading database ... 15% (Reading database ... 20% (Reading database ... 25% (Reading database ... 30% (Reading database ... 35% (Reading database ... 40% (Reading database ... 45% (Reading database ... 50% (Reading database ... 55% (Reading database ... 60% (Reading database ... 65% (Reading database ... 70% (Reading database ... 75% (Reading database ... 80% (Reading database ... 85% (Reading database ... 90% (Reading database ... 95% (Reading database ... 100% (Reading database ... 126082 files and directories currently installed.) 2026-03-20T17:56:52.822 INFO:teuthology.orchestra.run.vm08.stdout:Preparing to unpack .../python3-jmespath_0.10.0-1_all.deb ... 2026-03-20T17:56:52.823 INFO:teuthology.orchestra.run.vm08.stdout:Unpacking python3-jmespath (0.10.0-1) ... 2026-03-20T17:56:52.840 INFO:teuthology.orchestra.run.vm08.stdout:Selecting previously unselected package python3-xmltodict. 2026-03-20T17:56:52.848 INFO:teuthology.orchestra.run.vm08.stdout:Preparing to unpack .../python3-xmltodict_0.12.0-2_all.deb ... 2026-03-20T17:56:52.849 INFO:teuthology.orchestra.run.vm08.stdout:Unpacking python3-xmltodict (0.12.0-2) ... 2026-03-20T17:56:52.865 INFO:teuthology.orchestra.run.vm08.stdout:Selecting previously unselected package s3cmd. 2026-03-20T17:56:52.871 INFO:teuthology.orchestra.run.vm08.stdout:Preparing to unpack .../archives/s3cmd_2.2.0-1_all.deb ... 2026-03-20T17:56:52.871 INFO:teuthology.orchestra.run.vm08.stdout:Unpacking s3cmd (2.2.0-1) ... 2026-03-20T17:56:52.904 INFO:teuthology.orchestra.run.vm08.stdout:Setting up s3cmd (2.2.0-1) ... 2026-03-20T17:56:52.955 INFO:teuthology.orchestra.run.vm02.stdout:Setting up libcephfs-dev (20.2.0-712-g70f8415b-1jammy) ... 2026-03-20T17:56:52.957 INFO:teuthology.orchestra.run.vm02.stdout:Setting up python3-rados (20.2.0-712-g70f8415b-1jammy) ... 2026-03-20T17:56:52.960 INFO:teuthology.orchestra.run.vm02.stdout:Setting up libcephfs-daemon (20.2.0-712-g70f8415b-1jammy) ... 2026-03-20T17:56:52.962 INFO:teuthology.orchestra.run.vm02.stdout:Setting up python3-rbd (20.2.0-712-g70f8415b-1jammy) ... 2026-03-20T17:56:52.964 INFO:teuthology.orchestra.run.vm02.stdout:Setting up rbd-fuse (20.2.0-712-g70f8415b-1jammy) ... 2026-03-20T17:56:52.967 INFO:teuthology.orchestra.run.vm02.stdout:Setting up python3-rgw (20.2.0-712-g70f8415b-1jammy) ... 2026-03-20T17:56:52.969 INFO:teuthology.orchestra.run.vm02.stdout:Setting up python3-cephfs (20.2.0-712-g70f8415b-1jammy) ... 2026-03-20T17:56:52.972 INFO:teuthology.orchestra.run.vm02.stdout:Setting up ceph-common (20.2.0-712-g70f8415b-1jammy) ... 2026-03-20T17:56:52.992 INFO:teuthology.orchestra.run.vm08.stdout:Setting up python3-xmltodict (0.12.0-2) ... 2026-03-20T17:56:53.006 INFO:teuthology.orchestra.run.vm02.stdout:Adding group ceph....done 2026-03-20T17:56:53.044 INFO:teuthology.orchestra.run.vm02.stdout:Adding system user ceph....done 2026-03-20T17:56:53.054 INFO:teuthology.orchestra.run.vm02.stdout:Setting system user ceph properties....done 2026-03-20T17:56:53.058 INFO:teuthology.orchestra.run.vm08.stdout:Setting up python3-jmespath (0.10.0-1) ... 2026-03-20T17:56:53.059 INFO:teuthology.orchestra.run.vm02.stdout:chown: cannot access '/var/log/ceph/*.log*': No such file or directory 2026-03-20T17:56:53.125 INFO:teuthology.orchestra.run.vm02.stdout:Created symlink /etc/systemd/system/multi-user.target.wants/ceph.target → /lib/systemd/system/ceph.target. 2026-03-20T17:56:53.127 INFO:teuthology.orchestra.run.vm08.stdout:Processing triggers for man-db (2.10.2-1) ... 2026-03-20T17:56:53.356 INFO:teuthology.orchestra.run.vm02.stdout:Created symlink /etc/systemd/system/multi-user.target.wants/rbdmap.service → /lib/systemd/system/rbdmap.service. 2026-03-20T17:56:53.443 INFO:teuthology.orchestra.run.vm08.stdout: 2026-03-20T17:56:53.443 INFO:teuthology.orchestra.run.vm08.stdout:Running kernel seems to be up-to-date. 2026-03-20T17:56:53.443 INFO:teuthology.orchestra.run.vm08.stdout: 2026-03-20T17:56:53.443 INFO:teuthology.orchestra.run.vm08.stdout:Services to be restarted: 2026-03-20T17:56:53.445 INFO:teuthology.orchestra.run.vm08.stdout: systemctl restart apache-htcacheclean.service 2026-03-20T17:56:53.451 INFO:teuthology.orchestra.run.vm08.stdout: systemctl restart rsyslog.service 2026-03-20T17:56:53.453 INFO:teuthology.orchestra.run.vm08.stdout: 2026-03-20T17:56:53.453 INFO:teuthology.orchestra.run.vm08.stdout:Service restarts being deferred: 2026-03-20T17:56:53.453 INFO:teuthology.orchestra.run.vm08.stdout: systemctl restart networkd-dispatcher.service 2026-03-20T17:56:53.453 INFO:teuthology.orchestra.run.vm08.stdout: systemctl restart unattended-upgrades.service 2026-03-20T17:56:53.453 INFO:teuthology.orchestra.run.vm08.stdout: 2026-03-20T17:56:53.453 INFO:teuthology.orchestra.run.vm08.stdout:No containers need to be restarted. 2026-03-20T17:56:53.453 INFO:teuthology.orchestra.run.vm08.stdout: 2026-03-20T17:56:53.453 INFO:teuthology.orchestra.run.vm08.stdout:No user sessions are running outdated binaries. 2026-03-20T17:56:53.453 INFO:teuthology.orchestra.run.vm08.stdout: 2026-03-20T17:56:53.453 INFO:teuthology.orchestra.run.vm08.stdout:No VM guests are running outdated hypervisor (qemu) binaries on this host. 2026-03-20T17:56:53.479 INFO:teuthology.orchestra.run.vm05.stderr:W: --force-yes is deprecated, use one of the options starting with --allow instead. 2026-03-20T17:56:53.482 DEBUG:teuthology.orchestra.run.vm05:> sudo DEBIAN_FRONTEND=noninteractive apt-get -y --force-yes -o Dpkg::Options::="--force-confdef" -o Dpkg::Options::="--force-confold" install python3-jmespath python3-xmltodict s3cmd 2026-03-20T17:56:53.561 INFO:teuthology.orchestra.run.vm05.stdout:Reading package lists... 2026-03-20T17:56:53.733 INFO:teuthology.orchestra.run.vm05.stdout:Building dependency tree... 2026-03-20T17:56:53.733 INFO:teuthology.orchestra.run.vm05.stdout:Reading state information... 2026-03-20T17:56:53.745 INFO:teuthology.orchestra.run.vm02.stdout:Setting up ceph-test (20.2.0-712-g70f8415b-1jammy) ... 2026-03-20T17:56:53.747 INFO:teuthology.orchestra.run.vm02.stdout:Setting up radosgw (20.2.0-712-g70f8415b-1jammy) ... 2026-03-20T17:56:53.864 INFO:teuthology.orchestra.run.vm05.stdout:The following packages were automatically installed and are no longer required: 2026-03-20T17:56:53.864 INFO:teuthology.orchestra.run.vm05.stdout: kpartx libboost-iostreams1.74.0 libboost-thread1.74.0 libpmemobj1 2026-03-20T17:56:53.865 INFO:teuthology.orchestra.run.vm05.stdout: libsgutils2-2 sg3-utils sg3-utils-udev 2026-03-20T17:56:53.865 INFO:teuthology.orchestra.run.vm05.stdout:Use 'sudo apt autoremove' to remove them. 2026-03-20T17:56:53.879 INFO:teuthology.orchestra.run.vm05.stdout:The following NEW packages will be installed: 2026-03-20T17:56:53.879 INFO:teuthology.orchestra.run.vm05.stdout: python3-jmespath python3-xmltodict s3cmd 2026-03-20T17:56:53.903 INFO:teuthology.orchestra.run.vm05.stdout:0 upgraded, 3 newly installed, 0 to remove and 36 not upgraded. 2026-03-20T17:56:53.903 INFO:teuthology.orchestra.run.vm05.stdout:Need to get 155 kB of archives. 2026-03-20T17:56:53.903 INFO:teuthology.orchestra.run.vm05.stdout:After this operation, 678 kB of additional disk space will be used. 2026-03-20T17:56:53.903 INFO:teuthology.orchestra.run.vm05.stdout:Get:1 http://archive.ubuntu.com/ubuntu jammy/main amd64 python3-jmespath all 0.10.0-1 [21.7 kB] 2026-03-20T17:56:53.919 INFO:teuthology.orchestra.run.vm05.stdout:Get:2 http://archive.ubuntu.com/ubuntu jammy/universe amd64 python3-xmltodict all 0.12.0-2 [12.6 kB] 2026-03-20T17:56:53.920 INFO:teuthology.orchestra.run.vm05.stdout:Get:3 http://archive.ubuntu.com/ubuntu jammy/universe amd64 s3cmd all 2.2.0-1 [120 kB] 2026-03-20T17:56:53.999 INFO:teuthology.orchestra.run.vm02.stdout:Created symlink /etc/systemd/system/multi-user.target.wants/ceph-radosgw.target → /lib/systemd/system/ceph-radosgw.target. 2026-03-20T17:56:53.999 INFO:teuthology.orchestra.run.vm02.stdout:Created symlink /etc/systemd/system/ceph.target.wants/ceph-radosgw.target → /lib/systemd/system/ceph-radosgw.target. 2026-03-20T17:56:54.123 INFO:teuthology.orchestra.run.vm05.stdout:Fetched 155 kB in 0s (2819 kB/s) 2026-03-20T17:56:54.138 INFO:teuthology.orchestra.run.vm05.stdout:Selecting previously unselected package python3-jmespath. 2026-03-20T17:56:54.167 INFO:teuthology.orchestra.run.vm05.stdout:(Reading database ... (Reading database ... 5% (Reading database ... 10% (Reading database ... 15% (Reading database ... 20% (Reading database ... 25% (Reading database ... 30% (Reading database ... 35% (Reading database ... 40% (Reading database ... 45% (Reading database ... 50% (Reading database ... 55% (Reading database ... 60% (Reading database ... 65% (Reading database ... 70% (Reading database ... 75% (Reading database ... 80% (Reading database ... 85% (Reading database ... 90% (Reading database ... 95% (Reading database ... 100% (Reading database ... 126082 files and directories currently installed.) 2026-03-20T17:56:54.169 INFO:teuthology.orchestra.run.vm05.stdout:Preparing to unpack .../python3-jmespath_0.10.0-1_all.deb ... 2026-03-20T17:56:54.170 INFO:teuthology.orchestra.run.vm05.stdout:Unpacking python3-jmespath (0.10.0-1) ... 2026-03-20T17:56:54.187 INFO:teuthology.orchestra.run.vm05.stdout:Selecting previously unselected package python3-xmltodict. 2026-03-20T17:56:54.193 INFO:teuthology.orchestra.run.vm05.stdout:Preparing to unpack .../python3-xmltodict_0.12.0-2_all.deb ... 2026-03-20T17:56:54.193 INFO:teuthology.orchestra.run.vm05.stdout:Unpacking python3-xmltodict (0.12.0-2) ... 2026-03-20T17:56:54.208 INFO:teuthology.orchestra.run.vm05.stdout:Selecting previously unselected package s3cmd. 2026-03-20T17:56:54.214 INFO:teuthology.orchestra.run.vm05.stdout:Preparing to unpack .../archives/s3cmd_2.2.0-1_all.deb ... 2026-03-20T17:56:54.215 INFO:teuthology.orchestra.run.vm05.stdout:Unpacking s3cmd (2.2.0-1) ... 2026-03-20T17:56:54.245 INFO:teuthology.orchestra.run.vm05.stdout:Setting up s3cmd (2.2.0-1) ... 2026-03-20T17:56:54.330 INFO:teuthology.orchestra.run.vm05.stdout:Setting up python3-xmltodict (0.12.0-2) ... 2026-03-20T17:56:54.360 INFO:teuthology.orchestra.run.vm08.stderr:W: --force-yes is deprecated, use one of the options starting with --allow instead. 2026-03-20T17:56:54.364 DEBUG:teuthology.parallel:result is None 2026-03-20T17:56:54.374 INFO:teuthology.orchestra.run.vm02.stdout:Setting up ceph-base (20.2.0-712-g70f8415b-1jammy) ... 2026-03-20T17:56:54.394 INFO:teuthology.orchestra.run.vm05.stdout:Setting up python3-jmespath (0.10.0-1) ... 2026-03-20T17:56:54.459 INFO:teuthology.orchestra.run.vm02.stdout:Created symlink /etc/systemd/system/ceph.target.wants/ceph-crash.service → /lib/systemd/system/ceph-crash.service. 2026-03-20T17:56:54.461 INFO:teuthology.orchestra.run.vm05.stdout:Processing triggers for man-db (2.10.2-1) ... 2026-03-20T17:56:54.759 INFO:teuthology.orchestra.run.vm05.stdout: 2026-03-20T17:56:54.759 INFO:teuthology.orchestra.run.vm05.stdout:Running kernel seems to be up-to-date. 2026-03-20T17:56:54.759 INFO:teuthology.orchestra.run.vm05.stdout: 2026-03-20T17:56:54.759 INFO:teuthology.orchestra.run.vm05.stdout:Services to be restarted: 2026-03-20T17:56:54.762 INFO:teuthology.orchestra.run.vm05.stdout: systemctl restart apache-htcacheclean.service 2026-03-20T17:56:54.767 INFO:teuthology.orchestra.run.vm05.stdout: systemctl restart rsyslog.service 2026-03-20T17:56:54.769 INFO:teuthology.orchestra.run.vm05.stdout: 2026-03-20T17:56:54.769 INFO:teuthology.orchestra.run.vm05.stdout:Service restarts being deferred: 2026-03-20T17:56:54.769 INFO:teuthology.orchestra.run.vm05.stdout: systemctl restart networkd-dispatcher.service 2026-03-20T17:56:54.769 INFO:teuthology.orchestra.run.vm05.stdout: systemctl restart unattended-upgrades.service 2026-03-20T17:56:54.769 INFO:teuthology.orchestra.run.vm05.stdout: 2026-03-20T17:56:54.769 INFO:teuthology.orchestra.run.vm05.stdout:No containers need to be restarted. 2026-03-20T17:56:54.769 INFO:teuthology.orchestra.run.vm05.stdout: 2026-03-20T17:56:54.769 INFO:teuthology.orchestra.run.vm05.stdout:No user sessions are running outdated binaries. 2026-03-20T17:56:54.770 INFO:teuthology.orchestra.run.vm05.stdout: 2026-03-20T17:56:54.770 INFO:teuthology.orchestra.run.vm05.stdout:No VM guests are running outdated hypervisor (qemu) binaries on this host. 2026-03-20T17:56:54.831 INFO:teuthology.orchestra.run.vm02.stdout:Setting up ceph-mds (20.2.0-712-g70f8415b-1jammy) ... 2026-03-20T17:56:54.892 INFO:teuthology.orchestra.run.vm02.stdout:Created symlink /etc/systemd/system/multi-user.target.wants/ceph-mds.target → /lib/systemd/system/ceph-mds.target. 2026-03-20T17:56:54.892 INFO:teuthology.orchestra.run.vm02.stdout:Created symlink /etc/systemd/system/ceph.target.wants/ceph-mds.target → /lib/systemd/system/ceph-mds.target. 2026-03-20T17:56:55.246 INFO:teuthology.orchestra.run.vm02.stdout:Setting up ceph-mgr (20.2.0-712-g70f8415b-1jammy) ... 2026-03-20T17:56:55.318 INFO:teuthology.orchestra.run.vm02.stdout:Created symlink /etc/systemd/system/multi-user.target.wants/ceph-mgr.target → /lib/systemd/system/ceph-mgr.target. 2026-03-20T17:56:55.318 INFO:teuthology.orchestra.run.vm02.stdout:Created symlink /etc/systemd/system/ceph.target.wants/ceph-mgr.target → /lib/systemd/system/ceph-mgr.target. 2026-03-20T17:56:55.538 INFO:teuthology.orchestra.run.vm05.stderr:W: --force-yes is deprecated, use one of the options starting with --allow instead. 2026-03-20T17:56:55.541 DEBUG:teuthology.parallel:result is None 2026-03-20T17:56:55.670 INFO:teuthology.orchestra.run.vm02.stdout:Setting up ceph-osd (20.2.0-712-g70f8415b-1jammy) ... 2026-03-20T17:56:55.744 INFO:teuthology.orchestra.run.vm02.stdout:Created symlink /etc/systemd/system/multi-user.target.wants/ceph-osd.target → /lib/systemd/system/ceph-osd.target. 2026-03-20T17:56:55.744 INFO:teuthology.orchestra.run.vm02.stdout:Created symlink /etc/systemd/system/ceph.target.wants/ceph-osd.target → /lib/systemd/system/ceph-osd.target. 2026-03-20T17:56:56.097 INFO:teuthology.orchestra.run.vm02.stdout:Setting up ceph-mgr-k8sevents (20.2.0-712-g70f8415b-1jammy) ... 2026-03-20T17:56:56.100 INFO:teuthology.orchestra.run.vm02.stdout:Setting up ceph-mgr-diskprediction-local (20.2.0-712-g70f8415b-1jammy) ... 2026-03-20T17:56:56.111 INFO:teuthology.orchestra.run.vm02.stdout:Setting up ceph-mon (20.2.0-712-g70f8415b-1jammy) ... 2026-03-20T17:56:56.166 INFO:teuthology.orchestra.run.vm02.stdout:Created symlink /etc/systemd/system/multi-user.target.wants/ceph-mon.target → /lib/systemd/system/ceph-mon.target. 2026-03-20T17:56:56.166 INFO:teuthology.orchestra.run.vm02.stdout:Created symlink /etc/systemd/system/ceph.target.wants/ceph-mon.target → /lib/systemd/system/ceph-mon.target. 2026-03-20T17:56:56.509 INFO:teuthology.orchestra.run.vm02.stdout:Setting up ceph-mgr-cephadm (20.2.0-712-g70f8415b-1jammy) ... 2026-03-20T17:56:56.521 INFO:teuthology.orchestra.run.vm02.stdout:Setting up ceph (20.2.0-712-g70f8415b-1jammy) ... 2026-03-20T17:56:56.523 INFO:teuthology.orchestra.run.vm02.stdout:Setting up ceph-mgr-dashboard (20.2.0-712-g70f8415b-1jammy) ... 2026-03-20T17:56:56.534 INFO:teuthology.orchestra.run.vm02.stdout:Setting up ceph-volume (20.2.0-712-g70f8415b-1jammy) ... 2026-03-20T17:56:56.642 INFO:teuthology.orchestra.run.vm02.stdout:Processing triggers for man-db (2.10.2-1) ... 2026-03-20T17:56:56.712 INFO:teuthology.orchestra.run.vm02.stdout:Processing triggers for libc-bin (2.35-0ubuntu3.13) ... 2026-03-20T17:56:56.994 INFO:teuthology.orchestra.run.vm02.stdout: 2026-03-20T17:56:56.995 INFO:teuthology.orchestra.run.vm02.stdout:Running kernel seems to be up-to-date. 2026-03-20T17:56:56.995 INFO:teuthology.orchestra.run.vm02.stdout: 2026-03-20T17:56:56.995 INFO:teuthology.orchestra.run.vm02.stdout:Services to be restarted: 2026-03-20T17:56:56.997 INFO:teuthology.orchestra.run.vm02.stdout: systemctl restart apache-htcacheclean.service 2026-03-20T17:56:57.002 INFO:teuthology.orchestra.run.vm02.stdout: systemctl restart rsyslog.service 2026-03-20T17:56:57.005 INFO:teuthology.orchestra.run.vm02.stdout: 2026-03-20T17:56:57.005 INFO:teuthology.orchestra.run.vm02.stdout:Service restarts being deferred: 2026-03-20T17:56:57.005 INFO:teuthology.orchestra.run.vm02.stdout: systemctl restart networkd-dispatcher.service 2026-03-20T17:56:57.005 INFO:teuthology.orchestra.run.vm02.stdout: systemctl restart unattended-upgrades.service 2026-03-20T17:56:57.005 INFO:teuthology.orchestra.run.vm02.stdout: 2026-03-20T17:56:57.005 INFO:teuthology.orchestra.run.vm02.stdout:No containers need to be restarted. 2026-03-20T17:56:57.005 INFO:teuthology.orchestra.run.vm02.stdout: 2026-03-20T17:56:57.005 INFO:teuthology.orchestra.run.vm02.stdout:No user sessions are running outdated binaries. 2026-03-20T17:56:57.005 INFO:teuthology.orchestra.run.vm02.stdout: 2026-03-20T17:56:57.005 INFO:teuthology.orchestra.run.vm02.stdout:No VM guests are running outdated hypervisor (qemu) binaries on this host. 2026-03-20T17:56:57.795 INFO:teuthology.orchestra.run.vm02.stderr:W: --force-yes is deprecated, use one of the options starting with --allow instead. 2026-03-20T17:56:57.799 DEBUG:teuthology.orchestra.run.vm02:> sudo DEBIAN_FRONTEND=noninteractive apt-get -y --force-yes -o Dpkg::Options::="--force-confdef" -o Dpkg::Options::="--force-confold" install python3-jmespath python3-xmltodict s3cmd 2026-03-20T17:56:57.875 INFO:teuthology.orchestra.run.vm02.stdout:Reading package lists... 2026-03-20T17:56:58.004 INFO:teuthology.orchestra.run.vm02.stdout:Building dependency tree... 2026-03-20T17:56:58.005 INFO:teuthology.orchestra.run.vm02.stdout:Reading state information... 2026-03-20T17:56:58.104 INFO:teuthology.orchestra.run.vm02.stdout:The following packages were automatically installed and are no longer required: 2026-03-20T17:56:58.104 INFO:teuthology.orchestra.run.vm02.stdout: kpartx libboost-iostreams1.74.0 libboost-thread1.74.0 libpmemobj1 2026-03-20T17:56:58.104 INFO:teuthology.orchestra.run.vm02.stdout: libsgutils2-2 sg3-utils sg3-utils-udev 2026-03-20T17:56:58.104 INFO:teuthology.orchestra.run.vm02.stdout:Use 'sudo apt autoremove' to remove them. 2026-03-20T17:56:58.115 INFO:teuthology.orchestra.run.vm02.stdout:The following NEW packages will be installed: 2026-03-20T17:56:58.115 INFO:teuthology.orchestra.run.vm02.stdout: python3-jmespath python3-xmltodict s3cmd 2026-03-20T17:56:58.343 INFO:teuthology.orchestra.run.vm02.stdout:0 upgraded, 3 newly installed, 0 to remove and 36 not upgraded. 2026-03-20T17:56:58.343 INFO:teuthology.orchestra.run.vm02.stdout:Need to get 155 kB of archives. 2026-03-20T17:56:58.343 INFO:teuthology.orchestra.run.vm02.stdout:After this operation, 678 kB of additional disk space will be used. 2026-03-20T17:56:58.343 INFO:teuthology.orchestra.run.vm02.stdout:Get:1 http://archive.ubuntu.com/ubuntu jammy/main amd64 python3-jmespath all 0.10.0-1 [21.7 kB] 2026-03-20T17:56:58.580 INFO:teuthology.orchestra.run.vm02.stdout:Get:2 http://archive.ubuntu.com/ubuntu jammy/universe amd64 python3-xmltodict all 0.12.0-2 [12.6 kB] 2026-03-20T17:56:58.604 INFO:teuthology.orchestra.run.vm02.stdout:Get:3 http://archive.ubuntu.com/ubuntu jammy/universe amd64 s3cmd all 2.2.0-1 [120 kB] 2026-03-20T17:56:59.015 INFO:teuthology.orchestra.run.vm02.stdout:Fetched 155 kB in 1s (213 kB/s) 2026-03-20T17:56:59.031 INFO:teuthology.orchestra.run.vm02.stdout:Selecting previously unselected package python3-jmespath. 2026-03-20T17:56:59.058 INFO:teuthology.orchestra.run.vm02.stdout:(Reading database ... (Reading database ... 5% (Reading database ... 10% (Reading database ... 15% (Reading database ... 20% (Reading database ... 25% (Reading database ... 30% (Reading database ... 35% (Reading database ... 40% (Reading database ... 45% (Reading database ... 50% (Reading database ... 55% (Reading database ... 60% (Reading database ... 65% (Reading database ... 70% (Reading database ... 75% (Reading database ... 80% (Reading database ... 85% (Reading database ... 90% (Reading database ... 95% (Reading database ... 100% (Reading database ... 126082 files and directories currently installed.) 2026-03-20T17:56:59.059 INFO:teuthology.orchestra.run.vm02.stdout:Preparing to unpack .../python3-jmespath_0.10.0-1_all.deb ... 2026-03-20T17:56:59.061 INFO:teuthology.orchestra.run.vm02.stdout:Unpacking python3-jmespath (0.10.0-1) ... 2026-03-20T17:56:59.076 INFO:teuthology.orchestra.run.vm02.stdout:Selecting previously unselected package python3-xmltodict. 2026-03-20T17:56:59.081 INFO:teuthology.orchestra.run.vm02.stdout:Preparing to unpack .../python3-xmltodict_0.12.0-2_all.deb ... 2026-03-20T17:56:59.082 INFO:teuthology.orchestra.run.vm02.stdout:Unpacking python3-xmltodict (0.12.0-2) ... 2026-03-20T17:56:59.097 INFO:teuthology.orchestra.run.vm02.stdout:Selecting previously unselected package s3cmd. 2026-03-20T17:56:59.102 INFO:teuthology.orchestra.run.vm02.stdout:Preparing to unpack .../archives/s3cmd_2.2.0-1_all.deb ... 2026-03-20T17:56:59.103 INFO:teuthology.orchestra.run.vm02.stdout:Unpacking s3cmd (2.2.0-1) ... 2026-03-20T17:56:59.132 INFO:teuthology.orchestra.run.vm02.stdout:Setting up s3cmd (2.2.0-1) ... 2026-03-20T17:56:59.214 INFO:teuthology.orchestra.run.vm02.stdout:Setting up python3-xmltodict (0.12.0-2) ... 2026-03-20T17:56:59.274 INFO:teuthology.orchestra.run.vm02.stdout:Setting up python3-jmespath (0.10.0-1) ... 2026-03-20T17:56:59.342 INFO:teuthology.orchestra.run.vm02.stdout:Processing triggers for man-db (2.10.2-1) ... 2026-03-20T17:56:59.648 INFO:teuthology.orchestra.run.vm02.stdout: 2026-03-20T17:56:59.648 INFO:teuthology.orchestra.run.vm02.stdout:Running kernel seems to be up-to-date. 2026-03-20T17:56:59.649 INFO:teuthology.orchestra.run.vm02.stdout: 2026-03-20T17:56:59.649 INFO:teuthology.orchestra.run.vm02.stdout:Services to be restarted: 2026-03-20T17:56:59.651 INFO:teuthology.orchestra.run.vm02.stdout: systemctl restart apache-htcacheclean.service 2026-03-20T17:56:59.657 INFO:teuthology.orchestra.run.vm02.stdout: systemctl restart rsyslog.service 2026-03-20T17:56:59.660 INFO:teuthology.orchestra.run.vm02.stdout: 2026-03-20T17:56:59.660 INFO:teuthology.orchestra.run.vm02.stdout:Service restarts being deferred: 2026-03-20T17:56:59.660 INFO:teuthology.orchestra.run.vm02.stdout: systemctl restart networkd-dispatcher.service 2026-03-20T17:56:59.660 INFO:teuthology.orchestra.run.vm02.stdout: systemctl restart unattended-upgrades.service 2026-03-20T17:56:59.660 INFO:teuthology.orchestra.run.vm02.stdout: 2026-03-20T17:56:59.660 INFO:teuthology.orchestra.run.vm02.stdout:No containers need to be restarted. 2026-03-20T17:56:59.660 INFO:teuthology.orchestra.run.vm02.stdout: 2026-03-20T17:56:59.660 INFO:teuthology.orchestra.run.vm02.stdout:No user sessions are running outdated binaries. 2026-03-20T17:56:59.660 INFO:teuthology.orchestra.run.vm02.stdout: 2026-03-20T17:56:59.660 INFO:teuthology.orchestra.run.vm02.stdout:No VM guests are running outdated hypervisor (qemu) binaries on this host. 2026-03-20T17:57:00.469 INFO:teuthology.orchestra.run.vm02.stderr:W: --force-yes is deprecated, use one of the options starting with --allow instead. 2026-03-20T17:57:00.472 DEBUG:teuthology.parallel:result is None 2026-03-20T17:57:00.472 DEBUG:teuthology.packaging:Querying https://shaman.ceph.com/api/search?status=ready&project=ceph&flavor=default&distros=ubuntu%2F22.04%2Fx86_64&sha1=70f8415b300f041766fa27faf7d5472699e32388 2026-03-20T17:57:01.054 DEBUG:teuthology.orchestra.run.vm02:> dpkg-query -W -f '${Version}' ceph 2026-03-20T17:57:01.063 INFO:teuthology.orchestra.run.vm02.stdout:20.2.0-712-g70f8415b-1jammy 2026-03-20T17:57:01.063 INFO:teuthology.packaging:The installed version of ceph is 20.2.0-712-g70f8415b-1jammy 2026-03-20T17:57:01.063 INFO:teuthology.task.install:The correct ceph version 20.2.0-712-g70f8415b-1jammy is installed. 2026-03-20T17:57:01.064 DEBUG:teuthology.packaging:Querying https://shaman.ceph.com/api/search?status=ready&project=ceph&flavor=default&distros=ubuntu%2F22.04%2Fx86_64&sha1=70f8415b300f041766fa27faf7d5472699e32388 2026-03-20T17:57:01.681 DEBUG:teuthology.orchestra.run.vm05:> dpkg-query -W -f '${Version}' ceph 2026-03-20T17:57:01.690 INFO:teuthology.orchestra.run.vm05.stdout:20.2.0-712-g70f8415b-1jammy 2026-03-20T17:57:01.690 INFO:teuthology.packaging:The installed version of ceph is 20.2.0-712-g70f8415b-1jammy 2026-03-20T17:57:01.690 INFO:teuthology.task.install:The correct ceph version 20.2.0-712-g70f8415b-1jammy is installed. 2026-03-20T17:57:01.691 DEBUG:teuthology.packaging:Querying https://shaman.ceph.com/api/search?status=ready&project=ceph&flavor=default&distros=ubuntu%2F22.04%2Fx86_64&sha1=70f8415b300f041766fa27faf7d5472699e32388 2026-03-20T17:57:02.312 DEBUG:teuthology.orchestra.run.vm06:> dpkg-query -W -f '${Version}' ceph 2026-03-20T17:57:02.320 INFO:teuthology.orchestra.run.vm06.stdout:20.2.0-712-g70f8415b-1jammy 2026-03-20T17:57:02.321 INFO:teuthology.packaging:The installed version of ceph is 20.2.0-712-g70f8415b-1jammy 2026-03-20T17:57:02.321 INFO:teuthology.task.install:The correct ceph version 20.2.0-712-g70f8415b-1jammy is installed. 2026-03-20T17:57:02.322 DEBUG:teuthology.packaging:Querying https://shaman.ceph.com/api/search?status=ready&project=ceph&flavor=default&distros=ubuntu%2F22.04%2Fx86_64&sha1=70f8415b300f041766fa27faf7d5472699e32388 2026-03-20T17:57:02.975 DEBUG:teuthology.orchestra.run.vm08:> dpkg-query -W -f '${Version}' ceph 2026-03-20T17:57:02.984 INFO:teuthology.orchestra.run.vm08.stdout:20.2.0-712-g70f8415b-1jammy 2026-03-20T17:57:02.984 INFO:teuthology.packaging:The installed version of ceph is 20.2.0-712-g70f8415b-1jammy 2026-03-20T17:57:02.984 INFO:teuthology.task.install:The correct ceph version 20.2.0-712-g70f8415b-1jammy is installed. 2026-03-20T17:57:02.985 INFO:teuthology.task.install.util:Shipping valgrind.supp... 2026-03-20T17:57:02.985 DEBUG:teuthology.orchestra.run.vm02:> set -ex 2026-03-20T17:57:02.985 DEBUG:teuthology.orchestra.run.vm02:> sudo dd of=/home/ubuntu/cephtest/valgrind.supp 2026-03-20T17:57:02.993 DEBUG:teuthology.orchestra.run.vm05:> set -ex 2026-03-20T17:57:02.993 DEBUG:teuthology.orchestra.run.vm05:> sudo dd of=/home/ubuntu/cephtest/valgrind.supp 2026-03-20T17:57:03.001 DEBUG:teuthology.orchestra.run.vm06:> set -ex 2026-03-20T17:57:03.002 DEBUG:teuthology.orchestra.run.vm06:> sudo dd of=/home/ubuntu/cephtest/valgrind.supp 2026-03-20T17:57:03.010 DEBUG:teuthology.orchestra.run.vm08:> set -ex 2026-03-20T17:57:03.010 DEBUG:teuthology.orchestra.run.vm08:> sudo dd of=/home/ubuntu/cephtest/valgrind.supp 2026-03-20T17:57:03.033 INFO:teuthology.task.install.util:Shipping 'daemon-helper'... 2026-03-20T17:57:03.034 DEBUG:teuthology.orchestra.run.vm02:> set -ex 2026-03-20T17:57:03.034 DEBUG:teuthology.orchestra.run.vm02:> sudo dd of=/usr/bin/daemon-helper 2026-03-20T17:57:03.042 DEBUG:teuthology.orchestra.run.vm02:> sudo chmod a=rx -- /usr/bin/daemon-helper 2026-03-20T17:57:03.090 DEBUG:teuthology.orchestra.run.vm05:> set -ex 2026-03-20T17:57:03.090 DEBUG:teuthology.orchestra.run.vm05:> sudo dd of=/usr/bin/daemon-helper 2026-03-20T17:57:03.098 DEBUG:teuthology.orchestra.run.vm05:> sudo chmod a=rx -- /usr/bin/daemon-helper 2026-03-20T17:57:03.147 DEBUG:teuthology.orchestra.run.vm06:> set -ex 2026-03-20T17:57:03.147 DEBUG:teuthology.orchestra.run.vm06:> sudo dd of=/usr/bin/daemon-helper 2026-03-20T17:57:03.155 DEBUG:teuthology.orchestra.run.vm06:> sudo chmod a=rx -- /usr/bin/daemon-helper 2026-03-20T17:57:03.202 DEBUG:teuthology.orchestra.run.vm08:> set -ex 2026-03-20T17:57:03.203 DEBUG:teuthology.orchestra.run.vm08:> sudo dd of=/usr/bin/daemon-helper 2026-03-20T17:57:03.210 DEBUG:teuthology.orchestra.run.vm08:> sudo chmod a=rx -- /usr/bin/daemon-helper 2026-03-20T17:57:03.261 INFO:teuthology.task.install.util:Shipping 'adjust-ulimits'... 2026-03-20T17:57:03.261 DEBUG:teuthology.orchestra.run.vm02:> set -ex 2026-03-20T17:57:03.261 DEBUG:teuthology.orchestra.run.vm02:> sudo dd of=/usr/bin/adjust-ulimits 2026-03-20T17:57:03.270 DEBUG:teuthology.orchestra.run.vm02:> sudo chmod a=rx -- /usr/bin/adjust-ulimits 2026-03-20T17:57:03.321 DEBUG:teuthology.orchestra.run.vm05:> set -ex 2026-03-20T17:57:03.321 DEBUG:teuthology.orchestra.run.vm05:> sudo dd of=/usr/bin/adjust-ulimits 2026-03-20T17:57:03.329 DEBUG:teuthology.orchestra.run.vm05:> sudo chmod a=rx -- /usr/bin/adjust-ulimits 2026-03-20T17:57:03.379 DEBUG:teuthology.orchestra.run.vm06:> set -ex 2026-03-20T17:57:03.379 DEBUG:teuthology.orchestra.run.vm06:> sudo dd of=/usr/bin/adjust-ulimits 2026-03-20T17:57:03.387 DEBUG:teuthology.orchestra.run.vm06:> sudo chmod a=rx -- /usr/bin/adjust-ulimits 2026-03-20T17:57:03.435 DEBUG:teuthology.orchestra.run.vm08:> set -ex 2026-03-20T17:57:03.435 DEBUG:teuthology.orchestra.run.vm08:> sudo dd of=/usr/bin/adjust-ulimits 2026-03-20T17:57:03.443 DEBUG:teuthology.orchestra.run.vm08:> sudo chmod a=rx -- /usr/bin/adjust-ulimits 2026-03-20T17:57:03.493 INFO:teuthology.task.install.util:Shipping 'stdin-killer'... 2026-03-20T17:57:03.493 DEBUG:teuthology.orchestra.run.vm02:> set -ex 2026-03-20T17:57:03.494 DEBUG:teuthology.orchestra.run.vm02:> sudo dd of=/usr/bin/stdin-killer 2026-03-20T17:57:03.502 DEBUG:teuthology.orchestra.run.vm02:> sudo chmod a=rx -- /usr/bin/stdin-killer 2026-03-20T17:57:03.554 DEBUG:teuthology.orchestra.run.vm05:> set -ex 2026-03-20T17:57:03.554 DEBUG:teuthology.orchestra.run.vm05:> sudo dd of=/usr/bin/stdin-killer 2026-03-20T17:57:03.563 DEBUG:teuthology.orchestra.run.vm05:> sudo chmod a=rx -- /usr/bin/stdin-killer 2026-03-20T17:57:03.610 DEBUG:teuthology.orchestra.run.vm06:> set -ex 2026-03-20T17:57:03.611 DEBUG:teuthology.orchestra.run.vm06:> sudo dd of=/usr/bin/stdin-killer 2026-03-20T17:57:03.619 DEBUG:teuthology.orchestra.run.vm06:> sudo chmod a=rx -- /usr/bin/stdin-killer 2026-03-20T17:57:03.668 DEBUG:teuthology.orchestra.run.vm08:> set -ex 2026-03-20T17:57:03.668 DEBUG:teuthology.orchestra.run.vm08:> sudo dd of=/usr/bin/stdin-killer 2026-03-20T17:57:03.676 DEBUG:teuthology.orchestra.run.vm08:> sudo chmod a=rx -- /usr/bin/stdin-killer 2026-03-20T17:57:03.725 INFO:teuthology.run_tasks:Running task ceph... 2026-03-20T17:57:03.764 INFO:tasks.ceph:Making ceph log dir writeable by non-root... 2026-03-20T17:57:03.764 DEBUG:teuthology.orchestra.run.vm02:> sudo chmod 777 /var/log/ceph 2026-03-20T17:57:03.765 DEBUG:teuthology.orchestra.run.vm05:> sudo chmod 777 /var/log/ceph 2026-03-20T17:57:03.766 DEBUG:teuthology.orchestra.run.vm06:> sudo chmod 777 /var/log/ceph 2026-03-20T17:57:03.767 DEBUG:teuthology.orchestra.run.vm08:> sudo chmod 777 /var/log/ceph 2026-03-20T17:57:03.775 INFO:tasks.ceph:Disabling ceph logrotate... 2026-03-20T17:57:03.775 DEBUG:teuthology.orchestra.run.vm02:> sudo rm -f -- /etc/logrotate.d/ceph 2026-03-20T17:57:03.815 DEBUG:teuthology.orchestra.run.vm05:> sudo rm -f -- /etc/logrotate.d/ceph 2026-03-20T17:57:03.820 DEBUG:teuthology.orchestra.run.vm06:> sudo rm -f -- /etc/logrotate.d/ceph 2026-03-20T17:57:03.822 DEBUG:teuthology.orchestra.run.vm08:> sudo rm -f -- /etc/logrotate.d/ceph 2026-03-20T17:57:03.830 INFO:tasks.ceph:Creating extra log directories... 2026-03-20T17:57:03.830 DEBUG:teuthology.orchestra.run.vm02:> sudo install -d -m0777 -- /var/log/ceph/valgrind /var/log/ceph/profiling-logger 2026-03-20T17:57:03.867 DEBUG:teuthology.orchestra.run.vm05:> sudo install -d -m0777 -- /var/log/ceph/valgrind /var/log/ceph/profiling-logger 2026-03-20T17:57:03.868 DEBUG:teuthology.orchestra.run.vm06:> sudo install -d -m0777 -- /var/log/ceph/valgrind /var/log/ceph/profiling-logger 2026-03-20T17:57:03.873 DEBUG:teuthology.orchestra.run.vm08:> sudo install -d -m0777 -- /var/log/ceph/valgrind /var/log/ceph/profiling-logger 2026-03-20T17:57:03.887 INFO:tasks.ceph:Creating ceph cluster c1... 2026-03-20T17:57:03.887 INFO:tasks.ceph:config {'conf': {'client': {'debug rgw': 20, 'debug rgw lifecycle': 20, 'debug rgw notification': 20, 'debug rgw sync': 20, 'rgw crypt require ssl': False, 'rgw crypt s3 kms backend': 'testing', 'rgw crypt s3 kms encryption keys': 'testkey-1=YmluCmJvb3N0CmJvb3N0LWJ1aWxkCmNlcGguY29uZgo=', 'rgw curl low speed time': 300, 'rgw data log num shards': 4, 'rgw data sync poll interval': 5, 'rgw md log max shards': 4, 'rgw meta sync poll interval': 5, 'rgw s3 auth use sts': True, 'rgw sts key': 'abcdefghijklmnoq', 'rgw sync data inject err probability': 0, 'rgw sync log trim interval': 0, 'rgw sync meta inject err probability': 0, 'rgw sync obj etag verify': True, 'setgroup': 'ceph', 'setuser': 'ceph'}, 'client.0': {'rgw data notify interval msec': 0}, 'mgr': {'debug mgr': 20, 'debug ms': 1}, 'mon': {'debug mon': 20, 'debug ms': 1, 'debug paxos': 20}, 'osd': {'debug ms': 1, 'debug osd': 20, 'osd mclock iops capacity threshold hdd': 49000, 'rocksdb delete range threshold': 0}, 'osd.0': {'osd_max_omap_entries_per_request': 10}, 'osd.1': {'osd_max_omap_entries_per_request': 1000}, 'osd.2': {'osd_max_omap_entries_per_request': 10000}}, 'fs': 'xfs', 'mkfs_options': None, 'mount_options': None, 'skip_mgr_daemons': False, 'log_ignorelist': ['\\(MDS_ALL_DOWN\\)', '\\(MDS_UP_LESS_THAN_MAX\\)', '\\(PG_AVAILABILITY\\)', '\\(PG_DEGRADED\\)', '\\(POOL_APP_NOT_ENABLED\\)', 'not have an application enabled'], 'cpu_profile': set(), 'cluster': 'c1', 'mon_bind_msgr2': True, 'mon_bind_addrvec': True} 2026-03-20T17:57:03.887 INFO:tasks.ceph:ctx.config {'archive_path': '/archive/kyr-2026-03-20_17:25:16-rgw-tentacle-none-default-vps/2229', 'branch': 'tentacle', 'description': 'rgw/multisite/{clusters frontend/beast ignore-pg-availability notify omap_limits overrides realms/two-zonegroup supported-random-distro$/{ubuntu_latest} tasks/test_multi}', 'email': None, 'first_in_suite': False, 'flavor': 'default', 'job_id': '2229', 'ktype': 'distro', 'last_in_suite': False, 'machine_type': 'vps', 'name': 'kyr-2026-03-20_17:25:16-rgw-tentacle-none-default-vps', 'no_nested_subset': False, 'os_type': 'ubuntu', 'os_version': '22.04', 'overrides': {'admin_socket': {'branch': 'tentacle'}, 'ansible.cephlab': {'branch': 'main', 'repo': 'https://github.com/kshtsk/ceph-cm-ansible.git', 'skip_tags': 'nagios,monitoring-scripts,hostname,pubkeys,zap,sudoers,kerberos,ntp-client,resolvconf,cpan,nfs', 'vars': {'logical_volumes': {'lv_1': {'scratch_dev': True, 'size': '25%VG', 'vg': 'vg_nvme'}, 'lv_2': {'scratch_dev': True, 'size': '25%VG', 'vg': 'vg_nvme'}, 'lv_3': {'scratch_dev': True, 'size': '25%VG', 'vg': 'vg_nvme'}, 'lv_4': {'scratch_dev': True, 'size': '25%VG', 'vg': 'vg_nvme'}}, 'timezone': 'UTC', 'volume_groups': {'vg_nvme': {'pvs': '/dev/vdb,/dev/vdc,/dev/vdd,/dev/vde'}}}}, 'ceph': {'conf': {'client': {'debug rgw': 20, 'debug rgw lifecycle': 20, 'debug rgw notification': 20, 'debug rgw sync': 20, 'rgw crypt require ssl': False, 'rgw crypt s3 kms backend': 'testing', 'rgw crypt s3 kms encryption keys': 'testkey-1=YmluCmJvb3N0CmJvb3N0LWJ1aWxkCmNlcGguY29uZgo=', 'rgw curl low speed time': 300, 'rgw data log num shards': 4, 'rgw data sync poll interval': 5, 'rgw md log max shards': 4, 'rgw meta sync poll interval': 5, 'rgw s3 auth use sts': True, 'rgw sts key': 'abcdefghijklmnoq', 'rgw sync data inject err probability': 0, 'rgw sync log trim interval': 0, 'rgw sync meta inject err probability': 0, 'rgw sync obj etag verify': True, 'setgroup': 'ceph', 'setuser': 'ceph'}, 'client.0': {'rgw data notify interval msec': 0}, 'mgr': {'debug mgr': 20, 'debug ms': 1}, 'mon': {'debug mon': 20, 'debug ms': 1, 'debug paxos': 20}, 'osd': {'debug ms': 1, 'debug osd': 20, 'osd mclock iops capacity threshold hdd': 49000, 'rocksdb delete range threshold': 0}, 'osd.0': {'osd_max_omap_entries_per_request': 10}, 'osd.1': {'osd_max_omap_entries_per_request': 1000}, 'osd.2': {'osd_max_omap_entries_per_request': 10000}}, 'flavor': 'default', 'log-ignorelist': ['\\(MDS_ALL_DOWN\\)', '\\(MDS_UP_LESS_THAN_MAX\\)', '\\(PG_AVAILABILITY\\)', '\\(PG_DEGRADED\\)', '\\(POOL_APP_NOT_ENABLED\\)', 'not have an application enabled'], 'sha1': '70f8415b300f041766fa27faf7d5472699e32388'}, 'ceph-deploy': {'conf': {'client': {'log file': '/var/log/ceph/ceph-$name.$pid.log'}, 'mon': {}}}, 'cephadm': {'cephadm_binary_url': 'https://download.ceph.com/rpm-20.2.0/el9/noarch/cephadm'}, 'install': {'ceph': {'flavor': 'default', 'sha1': '70f8415b300f041766fa27faf7d5472699e32388'}, 'extra_system_packages': {'deb': ['python3-jmespath', 'python3-xmltodict', 's3cmd'], 'rpm': ['bzip2', 'perl-Test-Harness', 'python3-jmespath', 'python3-xmltodict', 's3cmd']}}, 'rgw': {'compression type': 'random', 'frontend': 'beast'}, 'rgw-multisite': {'realm': {'is_default': True, 'name': 'test-realm'}, 'zonegroups': [{'enabled_features': ['resharding', 'notification_v2'], 'endpoints': ['c1.client.0'], 'is_default': True, 'is_master': True, 'name': 'a', 'zones': [{'endpoints': ['c1.client.0'], 'is_default': True, 'is_master': True, 'name': 'a1'}, {'endpoints': ['c1.client.1'], 'name': 'a2'}]}, {'enabled_features': ['resharding', 'notification_v2'], 'endpoints': ['c2.client.0'], 'is_default': True, 'name': 'b', 'zones': [{'endpoints': ['c2.client.0'], 'is_default': True, 'is_master': True, 'name': 'b1'}, {'endpoints': ['c2.client.1'], 'name': 'b2'}]}]}, 'rgw-multisite-tests': {'args': ['tests.py', '-a', '!fails_with_rgw']}, 'workunit': {'branch': 'tt-tentacle', 'sha1': 'c04fae7bd18063e6f8f9e1cf40c23f810c1f6ca4'}}, 'owner': 'kyr', 'priority': 1000, 'repo': 'https://github.com/ceph/ceph.git', 'roles': [['c1.mon.a', 'c1.osd.0', 'c1.osd.1', 'c1.osd.2', 'c1.client.0'], ['c1.mgr.x', 'c1.osd.3', 'c1.osd.4', 'c1.osd.5', 'c1.client.1'], ['c2.mon.a', 'c2.osd.0', 'c2.osd.1', 'c2.osd.2', 'c2.client.0'], ['c2.mgr.x', 'c2.osd.3', 'c2.osd.4', 'c2.osd.5', 'c2.client.1']], 'seed': 5336, 'sha1': '70f8415b300f041766fa27faf7d5472699e32388', 'sleep_before_teardown': 0, 'suite': 'rgw', 'suite_branch': 'tt-tentacle', 'suite_path': '/home/teuthos/src/github.com_kshtsk_ceph_c04fae7bd18063e6f8f9e1cf40c23f810c1f6ca4/qa', 'suite_relpath': 'qa', 'suite_repo': 'https://github.com/kshtsk/ceph.git', 'suite_sha1': 'c04fae7bd18063e6f8f9e1cf40c23f810c1f6ca4', 'targets': {'vm02.local': 'ecdsa-sha2-nistp256 AAAAE2VjZHNhLXNoYTItbmlzdHAyNTYAAAAIbmlzdHAyNTYAAABBBLUlbeCQemEFIGAhOsbaSaC9cn/mRW0uqcv6nNbZb53B4Wibb+XPFovkYm/zpOAZ4wEdxyZ/PUnoWpm0Py9ocQE=', 'vm05.local': 'ecdsa-sha2-nistp256 AAAAE2VjZHNhLXNoYTItbmlzdHAyNTYAAAAIbmlzdHAyNTYAAABBBMoug7TNAoa7mJjh9cncEud7jrvuXdSIQZHZLMdsjGiNzszM0eqBraWpNHWznD8Hn4dQwU6ldQdhO/fM1PnoaIE=', 'vm06.local': 'ecdsa-sha2-nistp256 AAAAE2VjZHNhLXNoYTItbmlzdHAyNTYAAAAIbmlzdHAyNTYAAABBBAr4G6JTVJMADXrDYrEOUi/7hRfFSeJoHsHrYT4rHLq31Qu7OJgh6KH0wNL/WripVH20sVMfbmCvXmnnPcODwbA=', 'vm08.local': 'ecdsa-sha2-nistp256 AAAAE2VjZHNhLXNoYTItbmlzdHAyNTYAAAAIbmlzdHAyNTYAAABBBOUGJylvumZGYQnc6Cm2BiuFCvdgt79isI8bVHVCdZ3r8NnfMvYpiSD3SPVaM+Ua03wQvWKlAWO93A22s3C5WOk='}, 'tasks': [{'internal.check_packages': None}, {'internal.buildpackages_prep': None}, {'internal.save_config': None}, {'internal.check_lock': None}, {'internal.add_remotes': None}, {'console_log': None}, {'internal.connect': None}, {'internal.push_inventory': None}, {'internal.serialize_remote_roles': None}, {'internal.check_conflict': None}, {'internal.check_ceph_data': None}, {'internal.vm_setup': None}, {'internal.base': None}, {'internal.archive_upload': None}, {'internal.archive': None}, {'internal.coredump': None}, {'internal.sudo': None}, {'internal.syslog': None}, {'internal.timer': None}, {'pcp': None}, {'selinux': None}, {'ansible.cephlab': None}, {'clock': None}, {'install': None}, {'ceph': {'cluster': 'c1', 'conf': {'client': {'debug rgw': 20, 'debug rgw lifecycle': 20, 'debug rgw notification': 20, 'debug rgw sync': 20, 'rgw crypt require ssl': False, 'rgw crypt s3 kms backend': 'testing', 'rgw crypt s3 kms encryption keys': 'testkey-1=YmluCmJvb3N0CmJvb3N0LWJ1aWxkCmNlcGguY29uZgo=', 'rgw curl low speed time': 300, 'rgw data log num shards': 4, 'rgw data sync poll interval': 5, 'rgw md log max shards': 4, 'rgw meta sync poll interval': 5, 'rgw s3 auth use sts': True, 'rgw sts key': 'abcdefghijklmnoq', 'rgw sync data inject err probability': 0, 'rgw sync log trim interval': 0, 'rgw sync meta inject err probability': 0, 'rgw sync obj etag verify': True, 'setgroup': 'ceph', 'setuser': 'ceph'}, 'client.0': {'rgw data notify interval msec': 0}, 'mgr': {'debug mgr': 20, 'debug ms': 1}, 'mon': {'debug mon': 20, 'debug ms': 1, 'debug paxos': 20}, 'osd': {'debug ms': 1, 'debug osd': 20, 'osd mclock iops capacity threshold hdd': 49000, 'rocksdb delete range threshold': 0}, 'osd.0': {'osd_max_omap_entries_per_request': 10}, 'osd.1': {'osd_max_omap_entries_per_request': 1000}, 'osd.2': {'osd_max_omap_entries_per_request': 10000}}, 'flavor': 'default', 'log-ignorelist': ['\\(MDS_ALL_DOWN\\)', '\\(MDS_UP_LESS_THAN_MAX\\)', '\\(PG_AVAILABILITY\\)', '\\(PG_DEGRADED\\)', '\\(POOL_APP_NOT_ENABLED\\)', 'not have an application enabled'], 'sha1': '70f8415b300f041766fa27faf7d5472699e32388'}}, {'ceph': {'cluster': 'c2'}}, {'rgw': {'c1.client.0': {'port': 8000}, 'c1.client.1': {'port': 8001}, 'c2.client.0': {'port': 8000}, 'c2.client.1': {'port': 8001}}}, {'rgw-multisite': None}, {'rgw-multisite-tests': {'config': {'reconfigure_delay': 90}}}], 'teuthology': {'fragments_dropped': [], 'meta': {}, 'postmerge': []}, 'teuthology_branch': 'clyso-debian-13', 'teuthology_repo': 'https://github.com/clyso/teuthology', 'teuthology_sha1': '1c580df7a9c7c2aadc272da296344fd99f27c444', 'timestamp': '2026-03-20_17:25:16', 'tube': 'vps', 'user': 'kyr', 'verbose': False, 'worker_log': '/home/teuthos/.teuthology/dispatcher/dispatcher.vps.4188345'} 2026-03-20T17:57:03.887 DEBUG:teuthology.orchestra.run.vm02:> install -d -m0755 -- /home/ubuntu/cephtest/c1.data 2026-03-20T17:57:03.919 DEBUG:teuthology.orchestra.run.vm05:> install -d -m0755 -- /home/ubuntu/cephtest/c1.data 2026-03-20T17:57:03.920 DEBUG:teuthology.orchestra.run.vm06:> install -d -m0755 -- /home/ubuntu/cephtest/c1.data 2026-03-20T17:57:03.928 DEBUG:teuthology.orchestra.run.vm08:> install -d -m0755 -- /home/ubuntu/cephtest/c1.data 2026-03-20T17:57:03.932 DEBUG:teuthology.orchestra.run.vm02:> sudo install -d -m0777 -- /var/run/ceph 2026-03-20T17:57:03.963 DEBUG:teuthology.orchestra.run.vm05:> sudo install -d -m0777 -- /var/run/ceph 2026-03-20T17:57:03.964 DEBUG:teuthology.orchestra.run.vm06:> sudo install -d -m0777 -- /var/run/ceph 2026-03-20T17:57:03.972 DEBUG:teuthology.orchestra.run.vm08:> sudo install -d -m0777 -- /var/run/ceph 2026-03-20T17:57:03.982 DEBUG:teuthology.orchestra.run.vm02:> set -ex 2026-03-20T17:57:03.982 DEBUG:teuthology.orchestra.run.vm02:> dd if=/scratch_devs of=/dev/stdout 2026-03-20T17:57:04.017 DEBUG:teuthology.misc:devs=['/dev/vg_nvme/lv_1', '/dev/vg_nvme/lv_2', '/dev/vg_nvme/lv_3', '/dev/vg_nvme/lv_4'] 2026-03-20T17:57:04.017 DEBUG:teuthology.orchestra.run.vm02:> stat /dev/vg_nvme/lv_1 2026-03-20T17:57:04.061 INFO:teuthology.orchestra.run.vm02.stdout: File: /dev/vg_nvme/lv_1 -> ../dm-0 2026-03-20T17:57:04.061 INFO:teuthology.orchestra.run.vm02.stdout: Size: 7 Blocks: 0 IO Block: 4096 symbolic link 2026-03-20T17:57:04.061 INFO:teuthology.orchestra.run.vm02.stdout:Device: 5h/5d Inode: 777 Links: 1 2026-03-20T17:57:04.061 INFO:teuthology.orchestra.run.vm02.stdout:Access: (0777/lrwxrwxrwx) Uid: ( 0/ root) Gid: ( 0/ root) 2026-03-20T17:57:04.061 INFO:teuthology.orchestra.run.vm02.stdout:Access: 2026-03-20 17:55:45.740601275 +0000 2026-03-20T17:57:04.061 INFO:teuthology.orchestra.run.vm02.stdout:Modify: 2026-03-20 17:55:45.592675275 +0000 2026-03-20T17:57:04.061 INFO:teuthology.orchestra.run.vm02.stdout:Change: 2026-03-20 17:55:45.592675275 +0000 2026-03-20T17:57:04.061 INFO:teuthology.orchestra.run.vm02.stdout: Birth: - 2026-03-20T17:57:04.061 DEBUG:teuthology.orchestra.run.vm02:> sudo dd if=/dev/vg_nvme/lv_1 of=/dev/null count=1 2026-03-20T17:57:04.109 INFO:teuthology.orchestra.run.vm02.stderr:1+0 records in 2026-03-20T17:57:04.109 INFO:teuthology.orchestra.run.vm02.stderr:1+0 records out 2026-03-20T17:57:04.109 INFO:teuthology.orchestra.run.vm02.stderr:512 bytes copied, 0.000206045 s, 2.5 MB/s 2026-03-20T17:57:04.110 DEBUG:teuthology.orchestra.run.vm02:> ! mount | grep -v devtmpfs | grep -q /dev/vg_nvme/lv_1 2026-03-20T17:57:04.154 DEBUG:teuthology.orchestra.run.vm02:> stat /dev/vg_nvme/lv_2 2026-03-20T17:57:04.201 INFO:teuthology.orchestra.run.vm02.stdout: File: /dev/vg_nvme/lv_2 -> ../dm-1 2026-03-20T17:57:04.201 INFO:teuthology.orchestra.run.vm02.stdout: Size: 7 Blocks: 0 IO Block: 4096 symbolic link 2026-03-20T17:57:04.201 INFO:teuthology.orchestra.run.vm02.stdout:Device: 5h/5d Inode: 807 Links: 1 2026-03-20T17:57:04.201 INFO:teuthology.orchestra.run.vm02.stdout:Access: (0777/lrwxrwxrwx) Uid: ( 0/ root) Gid: ( 0/ root) 2026-03-20T17:57:04.201 INFO:teuthology.orchestra.run.vm02.stdout:Access: 2026-03-20 17:55:46.036453275 +0000 2026-03-20T17:57:04.201 INFO:teuthology.orchestra.run.vm02.stdout:Modify: 2026-03-20 17:55:45.892525275 +0000 2026-03-20T17:57:04.201 INFO:teuthology.orchestra.run.vm02.stdout:Change: 2026-03-20 17:55:45.892525275 +0000 2026-03-20T17:57:04.201 INFO:teuthology.orchestra.run.vm02.stdout: Birth: - 2026-03-20T17:57:04.201 DEBUG:teuthology.orchestra.run.vm02:> sudo dd if=/dev/vg_nvme/lv_2 of=/dev/null count=1 2026-03-20T17:57:04.249 INFO:teuthology.orchestra.run.vm02.stderr:1+0 records in 2026-03-20T17:57:04.249 INFO:teuthology.orchestra.run.vm02.stderr:1+0 records out 2026-03-20T17:57:04.249 INFO:teuthology.orchestra.run.vm02.stderr:512 bytes copied, 0.000187151 s, 2.7 MB/s 2026-03-20T17:57:04.250 DEBUG:teuthology.orchestra.run.vm02:> ! mount | grep -v devtmpfs | grep -q /dev/vg_nvme/lv_2 2026-03-20T17:57:04.294 DEBUG:teuthology.orchestra.run.vm02:> stat /dev/vg_nvme/lv_3 2026-03-20T17:57:04.341 INFO:teuthology.orchestra.run.vm02.stdout: File: /dev/vg_nvme/lv_3 -> ../dm-2 2026-03-20T17:57:04.341 INFO:teuthology.orchestra.run.vm02.stdout: Size: 7 Blocks: 0 IO Block: 4096 symbolic link 2026-03-20T17:57:04.341 INFO:teuthology.orchestra.run.vm02.stdout:Device: 5h/5d Inode: 842 Links: 1 2026-03-20T17:57:04.341 INFO:teuthology.orchestra.run.vm02.stdout:Access: (0777/lrwxrwxrwx) Uid: ( 0/ root) Gid: ( 0/ root) 2026-03-20T17:57:04.341 INFO:teuthology.orchestra.run.vm02.stdout:Access: 2026-03-20 17:55:46.320311275 +0000 2026-03-20T17:57:04.341 INFO:teuthology.orchestra.run.vm02.stdout:Modify: 2026-03-20 17:55:46.192375275 +0000 2026-03-20T17:57:04.341 INFO:teuthology.orchestra.run.vm02.stdout:Change: 2026-03-20 17:55:46.192375275 +0000 2026-03-20T17:57:04.341 INFO:teuthology.orchestra.run.vm02.stdout: Birth: - 2026-03-20T17:57:04.341 DEBUG:teuthology.orchestra.run.vm02:> sudo dd if=/dev/vg_nvme/lv_3 of=/dev/null count=1 2026-03-20T17:57:04.390 INFO:teuthology.orchestra.run.vm02.stderr:1+0 records in 2026-03-20T17:57:04.390 INFO:teuthology.orchestra.run.vm02.stderr:1+0 records out 2026-03-20T17:57:04.390 INFO:teuthology.orchestra.run.vm02.stderr:512 bytes copied, 0.000162804 s, 3.1 MB/s 2026-03-20T17:57:04.390 DEBUG:teuthology.orchestra.run.vm02:> ! mount | grep -v devtmpfs | grep -q /dev/vg_nvme/lv_3 2026-03-20T17:57:04.438 DEBUG:teuthology.orchestra.run.vm02:> stat /dev/vg_nvme/lv_4 2026-03-20T17:57:04.485 INFO:teuthology.orchestra.run.vm02.stdout: File: /dev/vg_nvme/lv_4 -> ../dm-3 2026-03-20T17:57:04.485 INFO:teuthology.orchestra.run.vm02.stdout: Size: 7 Blocks: 0 IO Block: 4096 symbolic link 2026-03-20T17:57:04.485 INFO:teuthology.orchestra.run.vm02.stdout:Device: 5h/5d Inode: 871 Links: 1 2026-03-20T17:57:04.485 INFO:teuthology.orchestra.run.vm02.stdout:Access: (0777/lrwxrwxrwx) Uid: ( 0/ root) Gid: ( 0/ root) 2026-03-20T17:57:04.485 INFO:teuthology.orchestra.run.vm02.stdout:Access: 2026-03-20 17:55:56.655141273 +0000 2026-03-20T17:57:04.485 INFO:teuthology.orchestra.run.vm02.stdout:Modify: 2026-03-20 17:55:46.480231275 +0000 2026-03-20T17:57:04.485 INFO:teuthology.orchestra.run.vm02.stdout:Change: 2026-03-20 17:55:46.480231275 +0000 2026-03-20T17:57:04.485 INFO:teuthology.orchestra.run.vm02.stdout: Birth: - 2026-03-20T17:57:04.485 DEBUG:teuthology.orchestra.run.vm02:> sudo dd if=/dev/vg_nvme/lv_4 of=/dev/null count=1 2026-03-20T17:57:04.532 INFO:teuthology.orchestra.run.vm02.stderr:1+0 records in 2026-03-20T17:57:04.532 INFO:teuthology.orchestra.run.vm02.stderr:1+0 records out 2026-03-20T17:57:04.532 INFO:teuthology.orchestra.run.vm02.stderr:512 bytes copied, 0.000182732 s, 2.8 MB/s 2026-03-20T17:57:04.533 DEBUG:teuthology.orchestra.run.vm02:> ! mount | grep -v devtmpfs | grep -q /dev/vg_nvme/lv_4 2026-03-20T17:57:04.578 INFO:tasks.ceph:osd dev map: {'c1.osd.0': '/dev/vg_nvme/lv_1', 'c1.osd.1': '/dev/vg_nvme/lv_2', 'c1.osd.2': '/dev/vg_nvme/lv_3'} 2026-03-20T17:57:04.578 DEBUG:teuthology.orchestra.run.vm05:> set -ex 2026-03-20T17:57:04.578 DEBUG:teuthology.orchestra.run.vm05:> dd if=/scratch_devs of=/dev/stdout 2026-03-20T17:57:04.582 DEBUG:teuthology.misc:devs=['/dev/vg_nvme/lv_1', '/dev/vg_nvme/lv_2', '/dev/vg_nvme/lv_3', '/dev/vg_nvme/lv_4'] 2026-03-20T17:57:04.582 DEBUG:teuthology.orchestra.run.vm05:> stat /dev/vg_nvme/lv_1 2026-03-20T17:57:04.627 INFO:teuthology.orchestra.run.vm05.stdout: File: /dev/vg_nvme/lv_1 -> ../dm-0 2026-03-20T17:57:04.627 INFO:teuthology.orchestra.run.vm05.stdout: Size: 7 Blocks: 0 IO Block: 4096 symbolic link 2026-03-20T17:57:04.627 INFO:teuthology.orchestra.run.vm05.stdout:Device: 5h/5d Inode: 788 Links: 1 2026-03-20T17:57:04.627 INFO:teuthology.orchestra.run.vm05.stdout:Access: (0777/lrwxrwxrwx) Uid: ( 0/ root) Gid: ( 0/ root) 2026-03-20T17:57:04.627 INFO:teuthology.orchestra.run.vm05.stdout:Access: 2026-03-20 17:55:43.846387000 +0000 2026-03-20T17:57:04.627 INFO:teuthology.orchestra.run.vm05.stdout:Modify: 2026-03-20 17:55:43.718387000 +0000 2026-03-20T17:57:04.627 INFO:teuthology.orchestra.run.vm05.stdout:Change: 2026-03-20 17:55:43.718387000 +0000 2026-03-20T17:57:04.627 INFO:teuthology.orchestra.run.vm05.stdout: Birth: - 2026-03-20T17:57:04.627 DEBUG:teuthology.orchestra.run.vm05:> sudo dd if=/dev/vg_nvme/lv_1 of=/dev/null count=1 2026-03-20T17:57:04.674 INFO:teuthology.orchestra.run.vm05.stderr:1+0 records in 2026-03-20T17:57:04.674 INFO:teuthology.orchestra.run.vm05.stderr:1+0 records out 2026-03-20T17:57:04.674 INFO:teuthology.orchestra.run.vm05.stderr:512 bytes copied, 0.000191428 s, 2.7 MB/s 2026-03-20T17:57:04.675 DEBUG:teuthology.orchestra.run.vm05:> ! mount | grep -v devtmpfs | grep -q /dev/vg_nvme/lv_1 2026-03-20T17:57:04.720 DEBUG:teuthology.orchestra.run.vm05:> stat /dev/vg_nvme/lv_2 2026-03-20T17:57:04.766 INFO:teuthology.orchestra.run.vm05.stdout: File: /dev/vg_nvme/lv_2 -> ../dm-1 2026-03-20T17:57:04.766 INFO:teuthology.orchestra.run.vm05.stdout: Size: 7 Blocks: 0 IO Block: 4096 symbolic link 2026-03-20T17:57:04.767 INFO:teuthology.orchestra.run.vm05.stdout:Device: 5h/5d Inode: 821 Links: 1 2026-03-20T17:57:04.767 INFO:teuthology.orchestra.run.vm05.stdout:Access: (0777/lrwxrwxrwx) Uid: ( 0/ root) Gid: ( 0/ root) 2026-03-20T17:57:04.767 INFO:teuthology.orchestra.run.vm05.stdout:Access: 2026-03-20 17:55:44.194387000 +0000 2026-03-20T17:57:04.767 INFO:teuthology.orchestra.run.vm05.stdout:Modify: 2026-03-20 17:55:44.018387000 +0000 2026-03-20T17:57:04.767 INFO:teuthology.orchestra.run.vm05.stdout:Change: 2026-03-20 17:55:44.018387000 +0000 2026-03-20T17:57:04.767 INFO:teuthology.orchestra.run.vm05.stdout: Birth: - 2026-03-20T17:57:04.767 DEBUG:teuthology.orchestra.run.vm05:> sudo dd if=/dev/vg_nvme/lv_2 of=/dev/null count=1 2026-03-20T17:57:04.814 INFO:teuthology.orchestra.run.vm05.stderr:1+0 records in 2026-03-20T17:57:04.814 INFO:teuthology.orchestra.run.vm05.stderr:1+0 records out 2026-03-20T17:57:04.814 INFO:teuthology.orchestra.run.vm05.stderr:512 bytes copied, 0.000180758 s, 2.8 MB/s 2026-03-20T17:57:04.815 DEBUG:teuthology.orchestra.run.vm05:> ! mount | grep -v devtmpfs | grep -q /dev/vg_nvme/lv_2 2026-03-20T17:57:04.859 DEBUG:teuthology.orchestra.run.vm05:> stat /dev/vg_nvme/lv_3 2026-03-20T17:57:04.902 INFO:teuthology.orchestra.run.vm05.stdout: File: /dev/vg_nvme/lv_3 -> ../dm-2 2026-03-20T17:57:04.903 INFO:teuthology.orchestra.run.vm05.stdout: Size: 7 Blocks: 0 IO Block: 4096 symbolic link 2026-03-20T17:57:04.903 INFO:teuthology.orchestra.run.vm05.stdout:Device: 5h/5d Inode: 851 Links: 1 2026-03-20T17:57:04.903 INFO:teuthology.orchestra.run.vm05.stdout:Access: (0777/lrwxrwxrwx) Uid: ( 0/ root) Gid: ( 0/ root) 2026-03-20T17:57:04.903 INFO:teuthology.orchestra.run.vm05.stdout:Access: 2026-03-20 17:55:44.502387000 +0000 2026-03-20T17:57:04.903 INFO:teuthology.orchestra.run.vm05.stdout:Modify: 2026-03-20 17:55:44.366387000 +0000 2026-03-20T17:57:04.903 INFO:teuthology.orchestra.run.vm05.stdout:Change: 2026-03-20 17:55:44.366387000 +0000 2026-03-20T17:57:04.903 INFO:teuthology.orchestra.run.vm05.stdout: Birth: - 2026-03-20T17:57:04.903 DEBUG:teuthology.orchestra.run.vm05:> sudo dd if=/dev/vg_nvme/lv_3 of=/dev/null count=1 2026-03-20T17:57:04.950 INFO:teuthology.orchestra.run.vm05.stderr:1+0 records in 2026-03-20T17:57:04.950 INFO:teuthology.orchestra.run.vm05.stderr:1+0 records out 2026-03-20T17:57:04.950 INFO:teuthology.orchestra.run.vm05.stderr:512 bytes copied, 0.000216476 s, 2.4 MB/s 2026-03-20T17:57:04.951 DEBUG:teuthology.orchestra.run.vm05:> ! mount | grep -v devtmpfs | grep -q /dev/vg_nvme/lv_3 2026-03-20T17:57:04.995 DEBUG:teuthology.orchestra.run.vm05:> stat /dev/vg_nvme/lv_4 2026-03-20T17:57:05.038 INFO:teuthology.orchestra.run.vm05.stdout: File: /dev/vg_nvme/lv_4 -> ../dm-3 2026-03-20T17:57:05.038 INFO:teuthology.orchestra.run.vm05.stdout: Size: 7 Blocks: 0 IO Block: 4096 symbolic link 2026-03-20T17:57:05.038 INFO:teuthology.orchestra.run.vm05.stdout:Device: 5h/5d Inode: 882 Links: 1 2026-03-20T17:57:05.038 INFO:teuthology.orchestra.run.vm05.stdout:Access: (0777/lrwxrwxrwx) Uid: ( 0/ root) Gid: ( 0/ root) 2026-03-20T17:57:05.038 INFO:teuthology.orchestra.run.vm05.stdout:Access: 2026-03-20 17:55:56.674387000 +0000 2026-03-20T17:57:05.038 INFO:teuthology.orchestra.run.vm05.stdout:Modify: 2026-03-20 17:55:44.686387000 +0000 2026-03-20T17:57:05.038 INFO:teuthology.orchestra.run.vm05.stdout:Change: 2026-03-20 17:55:44.686387000 +0000 2026-03-20T17:57:05.038 INFO:teuthology.orchestra.run.vm05.stdout: Birth: - 2026-03-20T17:57:05.038 DEBUG:teuthology.orchestra.run.vm05:> sudo dd if=/dev/vg_nvme/lv_4 of=/dev/null count=1 2026-03-20T17:57:05.085 INFO:teuthology.orchestra.run.vm05.stderr:1+0 records in 2026-03-20T17:57:05.085 INFO:teuthology.orchestra.run.vm05.stderr:1+0 records out 2026-03-20T17:57:05.085 INFO:teuthology.orchestra.run.vm05.stderr:512 bytes copied, 0.000190016 s, 2.7 MB/s 2026-03-20T17:57:05.086 DEBUG:teuthology.orchestra.run.vm05:> ! mount | grep -v devtmpfs | grep -q /dev/vg_nvme/lv_4 2026-03-20T17:57:05.131 INFO:tasks.ceph:osd dev map: {'c1.osd.3': '/dev/vg_nvme/lv_1', 'c1.osd.4': '/dev/vg_nvme/lv_2', 'c1.osd.5': '/dev/vg_nvme/lv_3'} 2026-03-20T17:57:05.131 INFO:tasks.ceph:remote_to_roles_to_devs: {Remote(name='ubuntu@vm02.local'): {'c1.osd.0': '/dev/vg_nvme/lv_1', 'c1.osd.1': '/dev/vg_nvme/lv_2', 'c1.osd.2': '/dev/vg_nvme/lv_3'}, Remote(name='ubuntu@vm05.local'): {'c1.osd.3': '/dev/vg_nvme/lv_1', 'c1.osd.4': '/dev/vg_nvme/lv_2', 'c1.osd.5': '/dev/vg_nvme/lv_3'}} 2026-03-20T17:57:05.131 INFO:tasks.ceph:Generating config... 2026-03-20T17:57:05.131 INFO:tasks.ceph:[client] debug rgw = 20 2026-03-20T17:57:05.131 INFO:tasks.ceph:[client] debug rgw lifecycle = 20 2026-03-20T17:57:05.131 INFO:tasks.ceph:[client] debug rgw notification = 20 2026-03-20T17:57:05.131 INFO:tasks.ceph:[client] debug rgw sync = 20 2026-03-20T17:57:05.131 INFO:tasks.ceph:[client] rgw crypt require ssl = False 2026-03-20T17:57:05.131 INFO:tasks.ceph:[client] rgw crypt s3 kms backend = testing 2026-03-20T17:57:05.131 INFO:tasks.ceph:[client] rgw crypt s3 kms encryption keys = testkey-1=YmluCmJvb3N0CmJvb3N0LWJ1aWxkCmNlcGguY29uZgo= 2026-03-20T17:57:05.131 INFO:tasks.ceph:[client] rgw curl low speed time = 300 2026-03-20T17:57:05.131 INFO:tasks.ceph:[client] rgw data log num shards = 4 2026-03-20T17:57:05.132 INFO:tasks.ceph:[client] rgw data sync poll interval = 5 2026-03-20T17:57:05.132 INFO:tasks.ceph:[client] rgw md log max shards = 4 2026-03-20T17:57:05.132 INFO:tasks.ceph:[client] rgw meta sync poll interval = 5 2026-03-20T17:57:05.132 INFO:tasks.ceph:[client] rgw s3 auth use sts = True 2026-03-20T17:57:05.132 INFO:tasks.ceph:[client] rgw sts key = abcdefghijklmnoq 2026-03-20T17:57:05.132 INFO:tasks.ceph:[client] rgw sync data inject err probability = 0 2026-03-20T17:57:05.132 INFO:tasks.ceph:[client] rgw sync log trim interval = 0 2026-03-20T17:57:05.132 INFO:tasks.ceph:[client] rgw sync meta inject err probability = 0 2026-03-20T17:57:05.132 INFO:tasks.ceph:[client] rgw sync obj etag verify = True 2026-03-20T17:57:05.132 INFO:tasks.ceph:[client] setgroup = ceph 2026-03-20T17:57:05.132 INFO:tasks.ceph:[client] setuser = ceph 2026-03-20T17:57:05.132 INFO:tasks.ceph:[client.0] rgw data notify interval msec = 0 2026-03-20T17:57:05.132 INFO:tasks.ceph:[mgr] debug mgr = 20 2026-03-20T17:57:05.132 INFO:tasks.ceph:[mgr] debug ms = 1 2026-03-20T17:57:05.132 INFO:tasks.ceph:[mon] debug mon = 20 2026-03-20T17:57:05.132 INFO:tasks.ceph:[mon] debug ms = 1 2026-03-20T17:57:05.132 INFO:tasks.ceph:[mon] debug paxos = 20 2026-03-20T17:57:05.132 INFO:tasks.ceph:[osd] debug ms = 1 2026-03-20T17:57:05.132 INFO:tasks.ceph:[osd] debug osd = 20 2026-03-20T17:57:05.132 INFO:tasks.ceph:[osd] osd mclock iops capacity threshold hdd = 49000 2026-03-20T17:57:05.132 INFO:tasks.ceph:[osd] rocksdb delete range threshold = 0 2026-03-20T17:57:05.132 INFO:tasks.ceph:[osd.0] osd_max_omap_entries_per_request = 10 2026-03-20T17:57:05.132 INFO:tasks.ceph:[osd.1] osd_max_omap_entries_per_request = 1000 2026-03-20T17:57:05.132 INFO:tasks.ceph:[osd.2] osd_max_omap_entries_per_request = 10000 2026-03-20T17:57:05.132 INFO:tasks.ceph:Setting up c1.mon.a... 2026-03-20T17:57:05.132 DEBUG:teuthology.orchestra.run.vm02:> sudo adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage ceph-authtool --create-keyring /etc/ceph/c1.keyring 2026-03-20T17:57:05.148 INFO:teuthology.orchestra.run.vm02.stdout:creating /etc/ceph/c1.keyring 2026-03-20T17:57:05.150 DEBUG:teuthology.orchestra.run.vm02:> sudo adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage ceph-authtool --gen-key --name=mon. /etc/ceph/c1.keyring 2026-03-20T17:57:05.211 DEBUG:teuthology.orchestra.run.vm02:> sudo chmod 0644 /etc/ceph/c1.keyring 2026-03-20T17:57:05.260 DEBUG:tasks.ceph:Ceph mon addresses: [('c1.mon.a', '192.168.123.102')] 2026-03-20T17:57:05.260 DEBUG:tasks.ceph:writing out conf {'global': {'chdir': '', 'pid file': '/var/run/ceph/$cluster-$name.pid', 'auth supported': 'cephx', 'filestore xattr use omap': 'true', 'mon clock drift allowed': '1.000', 'osd crush chooseleaf type': '0', 'auth debug': 'true', 'ms die on old message': 'true', 'ms die on bug': 'true', 'mon max pg per osd': '10000', 'mon pg warn max object skew': '0', 'osd_pool_default_pg_autoscale_mode': 'off', 'osd pool default size': '2', 'mon osd allow primary affinity': 'true', 'mon osd allow pg remap': 'true', 'mon warn on legacy crush tunables': 'false', 'mon warn on crush straw calc version zero': 'false', 'mon warn on no sortbitwise': 'false', 'mon warn on osd down out interval zero': 'false', 'mon warn on too few osds': 'false', 'mon_warn_on_pool_pg_num_not_power_of_two': 'false', 'mon_warn_on_pool_no_redundancy': 'false', 'mon_allow_pool_size_one': 'true', 'osd pool default erasure code profile': 'plugin=isa technique=reed_sol_van k=2 m=1 crush-failure-domain=osd', 'osd default data pool replay window': '5', 'mon allow pool delete': 'true', 'mon cluster log file level': 'debug', 'debug asserts on shutdown': 'true', 'mon health detail to clog': 'false', 'mon host': '192.168.123.102'}, 'osd': {'osd journal size': '100', 'osd scrub load threshold': '5.0', 'osd scrub max interval': '600', 'osd mclock profile': 'high_recovery_ops', 'osd mclock skip benchmark': 'true', 'osd recover clone overlap': 'true', 'osd recovery max chunk': '1048576', 'osd debug shutdown': 'true', 'osd debug op order': 'true', 'osd debug verify stray on activate': 'true', 'osd debug trim objects': 'true', 'osd open classes on start': 'true', 'osd debug pg log writeout': 'true', 'osd deep scrub update digest min age': '30', 'osd map max advance': '10', 'journal zero on create': 'true', 'filestore ondisk finisher threads': '3', 'filestore apply finisher threads': '3', 'bdev debug aio': 'true', 'osd debug misdirected ops': 'true', 'debug ms': 1, 'debug osd': 20, 'osd mclock iops capacity threshold hdd': 49000, 'rocksdb delete range threshold': 0}, 'mgr': {'debug ms': 1, 'debug mgr': 20, 'debug mon': '20', 'debug auth': '20', 'mon reweight min pgs per osd': '4', 'mon reweight min bytes per osd': '10', 'mgr/telemetry/nag': 'false'}, 'mon': {'debug ms': 1, 'debug mon': 20, 'debug paxos': 20, 'debug auth': '20', 'mon data avail warn': '5', 'mon mgr mkfs grace': '240', 'mon reweight min pgs per osd': '4', 'mon osd reporter subtree level': 'osd', 'mon osd prime pg temp': 'true', 'mon reweight min bytes per osd': '10', 'auth mon ticket ttl': '660', 'auth service ticket ttl': '240', 'mon_warn_on_insecure_global_id_reclaim': 'false', 'mon_warn_on_insecure_global_id_reclaim_allowed': 'false', 'mon_down_mkfs_grace': '2m', 'mon_warn_on_filestore_osds': 'false'}, 'client': {'rgw cache enabled': 'true', 'rgw enable ops log': 'true', 'rgw enable usage log': 'true', 'log file': '/var/log/ceph/$cluster-$name.$pid.log', 'admin socket': '/var/run/ceph/$cluster-$name.$pid.asok', 'debug rgw': 20, 'debug rgw lifecycle': 20, 'debug rgw notification': 20, 'debug rgw sync': 20, 'rgw crypt require ssl': False, 'rgw crypt s3 kms backend': 'testing', 'rgw crypt s3 kms encryption keys': 'testkey-1=YmluCmJvb3N0CmJvb3N0LWJ1aWxkCmNlcGguY29uZgo=', 'rgw curl low speed time': 300, 'rgw data log num shards': 4, 'rgw data sync poll interval': 5, 'rgw md log max shards': 4, 'rgw meta sync poll interval': 5, 'rgw s3 auth use sts': True, 'rgw sts key': 'abcdefghijklmnoq', 'rgw sync data inject err probability': 0, 'rgw sync log trim interval': 0, 'rgw sync meta inject err probability': 0, 'rgw sync obj etag verify': True, 'setgroup': 'ceph', 'setuser': 'ceph'}, 'mon.a': {}, 'client.0': {'rgw data notify interval msec': 0}, 'osd.0': {'osd_max_omap_entries_per_request': 10}, 'osd.1': {'osd_max_omap_entries_per_request': 1000}, 'osd.2': {'osd_max_omap_entries_per_request': 10000}} 2026-03-20T17:57:05.261 DEBUG:teuthology.orchestra.run.vm02:> set -ex 2026-03-20T17:57:05.261 DEBUG:teuthology.orchestra.run.vm02:> dd of=/home/ubuntu/cephtest/ceph.tmp.conf 2026-03-20T17:57:05.305 DEBUG:teuthology.orchestra.run.vm02:> adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage monmaptool -c /home/ubuntu/cephtest/ceph.tmp.conf --create --clobber --enable-all-features --add a 192.168.123.102 --print /home/ubuntu/cephtest/c1.monmap 2026-03-20T17:57:05.362 INFO:teuthology.orchestra.run.vm02.stderr:ignoring --setuser ceph since I am not root 2026-03-20T17:57:05.362 INFO:teuthology.orchestra.run.vm02.stderr:ignoring --setgroup ceph since I am not root 2026-03-20T17:57:05.362 INFO:teuthology.orchestra.run.vm02.stdout:monmaptool: monmap file /home/ubuntu/cephtest/c1.monmap 2026-03-20T17:57:05.362 INFO:teuthology.orchestra.run.vm02.stdout:monmaptool: generated fsid 67b0dd48-80db-4ea8-b708-f5136e717957 2026-03-20T17:57:05.362 INFO:teuthology.orchestra.run.vm02.stdout:setting min_mon_release = tentacle 2026-03-20T17:57:05.362 INFO:teuthology.orchestra.run.vm02.stdout:epoch 0 2026-03-20T17:57:05.362 INFO:teuthology.orchestra.run.vm02.stdout:fsid 67b0dd48-80db-4ea8-b708-f5136e717957 2026-03-20T17:57:05.362 INFO:teuthology.orchestra.run.vm02.stdout:last_changed 2026-03-20T17:57:05.362254+0000 2026-03-20T17:57:05.362 INFO:teuthology.orchestra.run.vm02.stdout:created 2026-03-20T17:57:05.362254+0000 2026-03-20T17:57:05.362 INFO:teuthology.orchestra.run.vm02.stdout:min_mon_release 20 (tentacle) 2026-03-20T17:57:05.362 INFO:teuthology.orchestra.run.vm02.stdout:election_strategy: 1 2026-03-20T17:57:05.362 INFO:teuthology.orchestra.run.vm02.stdout:0: [v2:192.168.123.102:3300/0,v1:192.168.123.102:6789/0] mon.a 2026-03-20T17:57:05.362 INFO:teuthology.orchestra.run.vm02.stdout:monmaptool: writing epoch 0 to /home/ubuntu/cephtest/c1.monmap (1 monitors) 2026-03-20T17:57:05.363 DEBUG:teuthology.orchestra.run.vm02:> rm -- /home/ubuntu/cephtest/ceph.tmp.conf 2026-03-20T17:57:05.408 INFO:tasks.ceph:Writing /etc/ceph/c1.conf for FSID 67b0dd48-80db-4ea8-b708-f5136e717957... 2026-03-20T17:57:05.409 DEBUG:teuthology.orchestra.run.vm02:> sudo mkdir -p /etc/ceph && sudo chmod 0755 /etc/ceph && sudo tee /etc/ceph/c1.conf && sudo chmod 0644 /etc/ceph/c1.conf > /dev/null 2026-03-20T17:57:05.451 DEBUG:teuthology.orchestra.run.vm05:> sudo mkdir -p /etc/ceph && sudo chmod 0755 /etc/ceph && sudo tee /etc/ceph/c1.conf && sudo chmod 0644 /etc/ceph/c1.conf > /dev/null 2026-03-20T17:57:05.452 DEBUG:teuthology.orchestra.run.vm06:> sudo mkdir -p /etc/ceph && sudo chmod 0755 /etc/ceph && sudo tee /etc/ceph/c1.conf && sudo chmod 0644 /etc/ceph/c1.conf > /dev/null 2026-03-20T17:57:05.453 DEBUG:teuthology.orchestra.run.vm08:> sudo mkdir -p /etc/ceph && sudo chmod 0755 /etc/ceph && sudo tee /etc/ceph/c1.conf && sudo chmod 0644 /etc/ceph/c1.conf > /dev/null 2026-03-20T17:57:05.466 INFO:teuthology.orchestra.run.vm02.stdout:[global] 2026-03-20T17:57:05.466 INFO:teuthology.orchestra.run.vm02.stdout: chdir = "" 2026-03-20T17:57:05.466 INFO:teuthology.orchestra.run.vm02.stdout: pid file = /var/run/ceph/$cluster-$name.pid 2026-03-20T17:57:05.466 INFO:teuthology.orchestra.run.vm02.stdout: auth supported = cephx 2026-03-20T17:57:05.466 INFO:teuthology.orchestra.run.vm02.stdout: 2026-03-20T17:57:05.466 INFO:teuthology.orchestra.run.vm02.stdout: filestore xattr use omap = true 2026-03-20T17:57:05.466 INFO:teuthology.orchestra.run.vm02.stdout: 2026-03-20T17:57:05.466 INFO:teuthology.orchestra.run.vm02.stdout: mon clock drift allowed = 1.000 2026-03-20T17:57:05.466 INFO:teuthology.orchestra.run.vm02.stdout: 2026-03-20T17:57:05.466 INFO:teuthology.orchestra.run.vm02.stdout: osd crush chooseleaf type = 0 2026-03-20T17:57:05.466 INFO:teuthology.orchestra.run.vm02.stdout: auth debug = true 2026-03-20T17:57:05.466 INFO:teuthology.orchestra.run.vm02.stdout: 2026-03-20T17:57:05.466 INFO:teuthology.orchestra.run.vm02.stdout: ms die on old message = true 2026-03-20T17:57:05.466 INFO:teuthology.orchestra.run.vm02.stdout: ms die on bug = true 2026-03-20T17:57:05.466 INFO:teuthology.orchestra.run.vm02.stdout: 2026-03-20T17:57:05.466 INFO:teuthology.orchestra.run.vm02.stdout: mon max pg per osd = 10000 # >= luminous 2026-03-20T17:57:05.466 INFO:teuthology.orchestra.run.vm02.stdout: mon pg warn max object skew = 0 2026-03-20T17:57:05.466 INFO:teuthology.orchestra.run.vm02.stdout: 2026-03-20T17:57:05.466 INFO:teuthology.orchestra.run.vm02.stdout: # disable pg_autoscaler by default for new pools 2026-03-20T17:57:05.466 INFO:teuthology.orchestra.run.vm02.stdout: osd_pool_default_pg_autoscale_mode = off 2026-03-20T17:57:05.466 INFO:teuthology.orchestra.run.vm02.stdout: 2026-03-20T17:57:05.466 INFO:teuthology.orchestra.run.vm02.stdout: osd pool default size = 2 2026-03-20T17:57:05.466 INFO:teuthology.orchestra.run.vm02.stdout: 2026-03-20T17:57:05.466 INFO:teuthology.orchestra.run.vm02.stdout: mon osd allow primary affinity = true 2026-03-20T17:57:05.466 INFO:teuthology.orchestra.run.vm02.stdout: mon osd allow pg remap = true 2026-03-20T17:57:05.466 INFO:teuthology.orchestra.run.vm02.stdout: mon warn on legacy crush tunables = false 2026-03-20T17:57:05.467 INFO:teuthology.orchestra.run.vm02.stdout: mon warn on crush straw calc version zero = false 2026-03-20T17:57:05.467 INFO:teuthology.orchestra.run.vm02.stdout: mon warn on no sortbitwise = false 2026-03-20T17:57:05.467 INFO:teuthology.orchestra.run.vm02.stdout: mon warn on osd down out interval zero = false 2026-03-20T17:57:05.467 INFO:teuthology.orchestra.run.vm02.stdout: mon warn on too few osds = false 2026-03-20T17:57:05.467 INFO:teuthology.orchestra.run.vm02.stdout: mon_warn_on_pool_pg_num_not_power_of_two = false 2026-03-20T17:57:05.467 INFO:teuthology.orchestra.run.vm02.stdout: mon_warn_on_pool_no_redundancy = false 2026-03-20T17:57:05.467 INFO:teuthology.orchestra.run.vm02.stdout: mon_allow_pool_size_one = true 2026-03-20T17:57:05.467 INFO:teuthology.orchestra.run.vm02.stdout: 2026-03-20T17:57:05.467 INFO:teuthology.orchestra.run.vm02.stdout: osd pool default erasure code profile = plugin=isa technique=reed_sol_van k=2 m=1 crush-failure-domain=osd 2026-03-20T17:57:05.467 INFO:teuthology.orchestra.run.vm02.stdout: 2026-03-20T17:57:05.467 INFO:teuthology.orchestra.run.vm02.stdout: osd default data pool replay window = 5 2026-03-20T17:57:05.467 INFO:teuthology.orchestra.run.vm02.stdout: 2026-03-20T17:57:05.467 INFO:teuthology.orchestra.run.vm02.stdout: mon allow pool delete = true 2026-03-20T17:57:05.467 INFO:teuthology.orchestra.run.vm02.stdout: 2026-03-20T17:57:05.467 INFO:teuthology.orchestra.run.vm02.stdout: mon cluster log file level = debug 2026-03-20T17:57:05.467 INFO:teuthology.orchestra.run.vm02.stdout: debug asserts on shutdown = true 2026-03-20T17:57:05.467 INFO:teuthology.orchestra.run.vm02.stdout: mon health detail to clog = false 2026-03-20T17:57:05.467 INFO:teuthology.orchestra.run.vm02.stdout: mon host = 192.168.123.102 2026-03-20T17:57:05.467 INFO:teuthology.orchestra.run.vm02.stdout: fsid = 67b0dd48-80db-4ea8-b708-f5136e717957 2026-03-20T17:57:05.467 INFO:teuthology.orchestra.run.vm02.stdout: 2026-03-20T17:57:05.467 INFO:teuthology.orchestra.run.vm02.stdout:[osd] 2026-03-20T17:57:05.467 INFO:teuthology.orchestra.run.vm02.stdout: osd journal size = 100 2026-03-20T17:57:05.467 INFO:teuthology.orchestra.run.vm02.stdout: 2026-03-20T17:57:05.467 INFO:teuthology.orchestra.run.vm02.stdout: osd scrub load threshold = 5.0 2026-03-20T17:57:05.467 INFO:teuthology.orchestra.run.vm02.stdout: osd scrub max interval = 600 2026-03-20T17:57:05.467 INFO:teuthology.orchestra.run.vm02.stdout: osd mclock profile = high_recovery_ops 2026-03-20T17:57:05.467 INFO:teuthology.orchestra.run.vm02.stdout: osd mclock skip benchmark = true 2026-03-20T17:57:05.467 INFO:teuthology.orchestra.run.vm02.stdout: 2026-03-20T17:57:05.467 INFO:teuthology.orchestra.run.vm02.stdout: osd recover clone overlap = true 2026-03-20T17:57:05.467 INFO:teuthology.orchestra.run.vm02.stdout: osd recovery max chunk = 1048576 2026-03-20T17:57:05.467 INFO:teuthology.orchestra.run.vm02.stdout: 2026-03-20T17:57:05.467 INFO:teuthology.orchestra.run.vm02.stdout: osd debug shutdown = true 2026-03-20T17:57:05.467 INFO:teuthology.orchestra.run.vm02.stdout: osd debug op order = true 2026-03-20T17:57:05.467 INFO:teuthology.orchestra.run.vm02.stdout: osd debug verify stray on activate = true 2026-03-20T17:57:05.467 INFO:teuthology.orchestra.run.vm02.stdout: osd debug trim objects = true 2026-03-20T17:57:05.467 INFO:teuthology.orchestra.run.vm02.stdout: 2026-03-20T17:57:05.467 INFO:teuthology.orchestra.run.vm02.stdout: osd open classes on start = true 2026-03-20T17:57:05.467 INFO:teuthology.orchestra.run.vm02.stdout: osd debug pg log writeout = true 2026-03-20T17:57:05.467 INFO:teuthology.orchestra.run.vm02.stdout: 2026-03-20T17:57:05.467 INFO:teuthology.orchestra.run.vm02.stdout: osd deep scrub update digest min age = 30 2026-03-20T17:57:05.467 INFO:teuthology.orchestra.run.vm02.stdout: 2026-03-20T17:57:05.467 INFO:teuthology.orchestra.run.vm02.stdout: osd map max advance = 10 2026-03-20T17:57:05.467 INFO:teuthology.orchestra.run.vm02.stdout: 2026-03-20T17:57:05.467 INFO:teuthology.orchestra.run.vm02.stdout: journal zero on create = true 2026-03-20T17:57:05.467 INFO:teuthology.orchestra.run.vm02.stdout: 2026-03-20T17:57:05.467 INFO:teuthology.orchestra.run.vm02.stdout: filestore ondisk finisher threads = 3 2026-03-20T17:57:05.467 INFO:teuthology.orchestra.run.vm02.stdout: filestore apply finisher threads = 3 2026-03-20T17:57:05.467 INFO:teuthology.orchestra.run.vm02.stdout: 2026-03-20T17:57:05.467 INFO:teuthology.orchestra.run.vm02.stdout: bdev debug aio = true 2026-03-20T17:57:05.467 INFO:teuthology.orchestra.run.vm02.stdout: osd debug misdirected ops = true 2026-03-20T17:57:05.467 INFO:teuthology.orchestra.run.vm02.stdout: debug ms = 1 2026-03-20T17:57:05.467 INFO:teuthology.orchestra.run.vm02.stdout: debug osd = 20 2026-03-20T17:57:05.467 INFO:teuthology.orchestra.run.vm02.stdout: osd mclock iops capacity threshold hdd = 49000 2026-03-20T17:57:05.467 INFO:teuthology.orchestra.run.vm02.stdout: rocksdb delete range threshold = 0 2026-03-20T17:57:05.467 INFO:teuthology.orchestra.run.vm02.stdout: 2026-03-20T17:57:05.467 INFO:teuthology.orchestra.run.vm02.stdout:[mgr] 2026-03-20T17:57:05.467 INFO:teuthology.orchestra.run.vm02.stdout: debug ms = 1 2026-03-20T17:57:05.467 INFO:teuthology.orchestra.run.vm02.stdout: debug mgr = 20 2026-03-20T17:57:05.467 INFO:teuthology.orchestra.run.vm02.stdout: debug mon = 20 2026-03-20T17:57:05.467 INFO:teuthology.orchestra.run.vm02.stdout: debug auth = 20 2026-03-20T17:57:05.467 INFO:teuthology.orchestra.run.vm02.stdout: mon reweight min pgs per osd = 4 2026-03-20T17:57:05.467 INFO:teuthology.orchestra.run.vm02.stdout: mon reweight min bytes per osd = 10 2026-03-20T17:57:05.467 INFO:teuthology.orchestra.run.vm02.stdout: mgr/telemetry/nag = false 2026-03-20T17:57:05.467 INFO:teuthology.orchestra.run.vm02.stdout: 2026-03-20T17:57:05.467 INFO:teuthology.orchestra.run.vm02.stdout:[mon] 2026-03-20T17:57:05.467 INFO:teuthology.orchestra.run.vm02.stdout: debug ms = 1 2026-03-20T17:57:05.468 INFO:teuthology.orchestra.run.vm02.stdout: debug mon = 20 2026-03-20T17:57:05.468 INFO:teuthology.orchestra.run.vm02.stdout: debug paxos = 20 2026-03-20T17:57:05.468 INFO:teuthology.orchestra.run.vm02.stdout: debug auth = 20 2026-03-20T17:57:05.468 INFO:teuthology.orchestra.run.vm02.stdout: mon data avail warn = 5 2026-03-20T17:57:05.468 INFO:teuthology.orchestra.run.vm02.stdout: mon mgr mkfs grace = 240 2026-03-20T17:57:05.468 INFO:teuthology.orchestra.run.vm02.stdout: mon reweight min pgs per osd = 4 2026-03-20T17:57:05.468 INFO:teuthology.orchestra.run.vm02.stdout: mon osd reporter subtree level = osd 2026-03-20T17:57:05.468 INFO:teuthology.orchestra.run.vm02.stdout: mon osd prime pg temp = true 2026-03-20T17:57:05.468 INFO:teuthology.orchestra.run.vm02.stdout: mon reweight min bytes per osd = 10 2026-03-20T17:57:05.468 INFO:teuthology.orchestra.run.vm02.stdout: 2026-03-20T17:57:05.468 INFO:teuthology.orchestra.run.vm02.stdout: # rotate auth tickets quickly to exercise renewal paths 2026-03-20T17:57:05.468 INFO:teuthology.orchestra.run.vm02.stdout: auth mon ticket ttl = 660 # 11m 2026-03-20T17:57:05.468 INFO:teuthology.orchestra.run.vm02.stdout: auth service ticket ttl = 240 # 4m 2026-03-20T17:57:05.468 INFO:teuthology.orchestra.run.vm02.stdout: 2026-03-20T17:57:05.468 INFO:teuthology.orchestra.run.vm02.stdout: # don't complain about insecure global_id in the test suite 2026-03-20T17:57:05.468 INFO:teuthology.orchestra.run.vm02.stdout: mon_warn_on_insecure_global_id_reclaim = false 2026-03-20T17:57:05.468 INFO:teuthology.orchestra.run.vm02.stdout: mon_warn_on_insecure_global_id_reclaim_allowed = false 2026-03-20T17:57:05.468 INFO:teuthology.orchestra.run.vm02.stdout: 2026-03-20T17:57:05.468 INFO:teuthology.orchestra.run.vm02.stdout: # 1m isn't quite enough 2026-03-20T17:57:05.468 INFO:teuthology.orchestra.run.vm02.stdout: mon_down_mkfs_grace = 2m 2026-03-20T17:57:05.468 INFO:teuthology.orchestra.run.vm02.stdout: 2026-03-20T17:57:05.468 INFO:teuthology.orchestra.run.vm02.stdout: mon_warn_on_filestore_osds = false 2026-03-20T17:57:05.468 INFO:teuthology.orchestra.run.vm02.stdout: 2026-03-20T17:57:05.468 INFO:teuthology.orchestra.run.vm02.stdout:[client] 2026-03-20T17:57:05.468 INFO:teuthology.orchestra.run.vm02.stdout: rgw cache enabled = true 2026-03-20T17:57:05.468 INFO:teuthology.orchestra.run.vm02.stdout: rgw enable ops log = true 2026-03-20T17:57:05.468 INFO:teuthology.orchestra.run.vm02.stdout: rgw enable usage log = true 2026-03-20T17:57:05.468 INFO:teuthology.orchestra.run.vm02.stdout: log file = /var/log/ceph/$cluster-$name.$pid.log 2026-03-20T17:57:05.468 INFO:teuthology.orchestra.run.vm02.stdout: admin socket = /var/run/ceph/$cluster-$name.$pid.asok 2026-03-20T17:57:05.468 INFO:teuthology.orchestra.run.vm02.stdout: debug rgw = 20 2026-03-20T17:57:05.468 INFO:teuthology.orchestra.run.vm02.stdout: debug rgw lifecycle = 20 2026-03-20T17:57:05.468 INFO:teuthology.orchestra.run.vm02.stdout: debug rgw notification = 20 2026-03-20T17:57:05.468 INFO:teuthology.orchestra.run.vm02.stdout: debug rgw sync = 20 2026-03-20T17:57:05.468 INFO:teuthology.orchestra.run.vm02.stdout: rgw crypt require ssl = False 2026-03-20T17:57:05.468 INFO:teuthology.orchestra.run.vm02.stdout: rgw crypt s3 kms backend = testing 2026-03-20T17:57:05.468 INFO:teuthology.orchestra.run.vm02.stdout: rgw crypt s3 kms encryption keys = testkey-1=YmluCmJvb3N0CmJvb3N0LWJ1aWxkCmNlcGguY29uZgo= 2026-03-20T17:57:05.468 INFO:teuthology.orchestra.run.vm02.stdout: rgw curl low speed time = 300 2026-03-20T17:57:05.468 INFO:teuthology.orchestra.run.vm02.stdout: rgw data log num shards = 4 2026-03-20T17:57:05.468 INFO:teuthology.orchestra.run.vm02.stdout: rgw data sync poll interval = 5 2026-03-20T17:57:05.468 INFO:teuthology.orchestra.run.vm02.stdout: rgw md log max shards = 4 2026-03-20T17:57:05.468 INFO:teuthology.orchestra.run.vm02.stdout: rgw meta sync poll interval = 5 2026-03-20T17:57:05.468 INFO:teuthology.orchestra.run.vm02.stdout: rgw s3 auth use sts = True 2026-03-20T17:57:05.468 INFO:teuthology.orchestra.run.vm02.stdout: rgw sts key = abcdefghijklmnoq 2026-03-20T17:57:05.468 INFO:teuthology.orchestra.run.vm02.stdout: rgw sync data inject err probability = 0 2026-03-20T17:57:05.468 INFO:teuthology.orchestra.run.vm02.stdout: rgw sync log trim interval = 0 2026-03-20T17:57:05.468 INFO:teuthology.orchestra.run.vm02.stdout: rgw sync meta inject err probability = 0 2026-03-20T17:57:05.468 INFO:teuthology.orchestra.run.vm02.stdout: rgw sync obj etag verify = True 2026-03-20T17:57:05.468 INFO:teuthology.orchestra.run.vm02.stdout: setgroup = ceph 2026-03-20T17:57:05.468 INFO:teuthology.orchestra.run.vm02.stdout: setuser = ceph 2026-03-20T17:57:05.468 INFO:teuthology.orchestra.run.vm02.stdout:[mon.a] 2026-03-20T17:57:05.468 INFO:teuthology.orchestra.run.vm02.stdout:[client.0] 2026-03-20T17:57:05.468 INFO:teuthology.orchestra.run.vm02.stdout: rgw data notify interval msec = 0 2026-03-20T17:57:05.468 INFO:teuthology.orchestra.run.vm02.stdout:[osd.0] 2026-03-20T17:57:05.468 INFO:teuthology.orchestra.run.vm02.stdout: osd_max_omap_entries_per_request = 10 2026-03-20T17:57:05.468 INFO:teuthology.orchestra.run.vm02.stdout:[osd.1] 2026-03-20T17:57:05.468 INFO:teuthology.orchestra.run.vm02.stdout: osd_max_omap_entries_per_request = 1000 2026-03-20T17:57:05.468 INFO:teuthology.orchestra.run.vm02.stdout:[osd.2] 2026-03-20T17:57:05.468 INFO:teuthology.orchestra.run.vm02.stdout: osd_max_omap_entries_per_request = 10000 2026-03-20T17:57:05.469 INFO:teuthology.orchestra.run.vm05.stdout:[global] 2026-03-20T17:57:05.469 INFO:teuthology.orchestra.run.vm05.stdout: chdir = "" 2026-03-20T17:57:05.469 INFO:teuthology.orchestra.run.vm05.stdout: pid file = /var/run/ceph/$cluster-$name.pid 2026-03-20T17:57:05.469 INFO:teuthology.orchestra.run.vm05.stdout: auth supported = cephx 2026-03-20T17:57:05.469 INFO:teuthology.orchestra.run.vm05.stdout: 2026-03-20T17:57:05.469 INFO:teuthology.orchestra.run.vm05.stdout: filestore xattr use omap = true 2026-03-20T17:57:05.469 INFO:teuthology.orchestra.run.vm05.stdout: 2026-03-20T17:57:05.469 INFO:teuthology.orchestra.run.vm05.stdout: mon clock drift allowed = 1.000 2026-03-20T17:57:05.469 INFO:teuthology.orchestra.run.vm05.stdout: 2026-03-20T17:57:05.469 INFO:teuthology.orchestra.run.vm05.stdout: osd crush chooseleaf type = 0 2026-03-20T17:57:05.469 INFO:teuthology.orchestra.run.vm05.stdout: auth debug = true 2026-03-20T17:57:05.469 INFO:teuthology.orchestra.run.vm05.stdout: 2026-03-20T17:57:05.469 INFO:teuthology.orchestra.run.vm05.stdout: ms die on old message = true 2026-03-20T17:57:05.469 INFO:teuthology.orchestra.run.vm05.stdout: ms die on bug = true 2026-03-20T17:57:05.469 INFO:teuthology.orchestra.run.vm05.stdout: 2026-03-20T17:57:05.469 INFO:teuthology.orchestra.run.vm05.stdout: mon max pg per osd = 10000 # >= luminous 2026-03-20T17:57:05.469 INFO:teuthology.orchestra.run.vm05.stdout: mon pg warn max object skew = 0 2026-03-20T17:57:05.469 INFO:teuthology.orchestra.run.vm05.stdout: 2026-03-20T17:57:05.469 INFO:teuthology.orchestra.run.vm05.stdout: # disable pg_autoscaler by default for new pools 2026-03-20T17:57:05.469 INFO:teuthology.orchestra.run.vm05.stdout: osd_pool_default_pg_autoscale_mode = off 2026-03-20T17:57:05.469 INFO:teuthology.orchestra.run.vm05.stdout: 2026-03-20T17:57:05.469 INFO:teuthology.orchestra.run.vm05.stdout: osd pool default size = 2 2026-03-20T17:57:05.469 INFO:teuthology.orchestra.run.vm05.stdout: 2026-03-20T17:57:05.469 INFO:teuthology.orchestra.run.vm05.stdout: mon osd allow primary affinity = true 2026-03-20T17:57:05.469 INFO:teuthology.orchestra.run.vm05.stdout: mon osd allow pg remap = true 2026-03-20T17:57:05.469 INFO:teuthology.orchestra.run.vm05.stdout: mon warn on legacy crush tunables = false 2026-03-20T17:57:05.469 INFO:teuthology.orchestra.run.vm05.stdout: mon warn on crush straw calc version zero = false 2026-03-20T17:57:05.469 INFO:teuthology.orchestra.run.vm05.stdout: mon warn on no sortbitwise = false 2026-03-20T17:57:05.469 INFO:teuthology.orchestra.run.vm05.stdout: mon warn on osd down out interval zero = false 2026-03-20T17:57:05.469 INFO:teuthology.orchestra.run.vm05.stdout: mon warn on too few osds = false 2026-03-20T17:57:05.469 INFO:teuthology.orchestra.run.vm05.stdout: mon_warn_on_pool_pg_num_not_power_of_two = false 2026-03-20T17:57:05.469 INFO:teuthology.orchestra.run.vm05.stdout: mon_warn_on_pool_no_redundancy = false 2026-03-20T17:57:05.469 INFO:teuthology.orchestra.run.vm05.stdout: mon_allow_pool_size_one = true 2026-03-20T17:57:05.469 INFO:teuthology.orchestra.run.vm05.stdout: 2026-03-20T17:57:05.469 INFO:teuthology.orchestra.run.vm05.stdout: osd pool default erasure code profile = plugin=isa technique=reed_sol_van k=2 m=1 crush-failure-domain=osd 2026-03-20T17:57:05.469 INFO:teuthology.orchestra.run.vm05.stdout: 2026-03-20T17:57:05.469 INFO:teuthology.orchestra.run.vm05.stdout: osd default data pool replay window = 5 2026-03-20T17:57:05.469 INFO:teuthology.orchestra.run.vm05.stdout: 2026-03-20T17:57:05.469 INFO:teuthology.orchestra.run.vm05.stdout: mon allow pool delete = true 2026-03-20T17:57:05.469 INFO:teuthology.orchestra.run.vm05.stdout: 2026-03-20T17:57:05.469 INFO:teuthology.orchestra.run.vm05.stdout: mon cluster log file level = debug 2026-03-20T17:57:05.469 INFO:teuthology.orchestra.run.vm05.stdout: debug asserts on shutdown = true 2026-03-20T17:57:05.469 INFO:teuthology.orchestra.run.vm05.stdout: mon health detail to clog = false 2026-03-20T17:57:05.469 INFO:teuthology.orchestra.run.vm05.stdout: mon host = 192.168.123.102 2026-03-20T17:57:05.469 INFO:teuthology.orchestra.run.vm05.stdout: fsid = 67b0dd48-80db-4ea8-b708-f5136e717957 2026-03-20T17:57:05.469 INFO:teuthology.orchestra.run.vm05.stdout: 2026-03-20T17:57:05.469 INFO:teuthology.orchestra.run.vm05.stdout:[osd] 2026-03-20T17:57:05.469 INFO:teuthology.orchestra.run.vm05.stdout: osd journal size = 100 2026-03-20T17:57:05.469 INFO:teuthology.orchestra.run.vm05.stdout: 2026-03-20T17:57:05.469 INFO:teuthology.orchestra.run.vm05.stdout: osd scrub load threshold = 5.0 2026-03-20T17:57:05.469 INFO:teuthology.orchestra.run.vm05.stdout: osd scrub max interval = 600 2026-03-20T17:57:05.469 INFO:teuthology.orchestra.run.vm05.stdout: osd mclock profile = high_recovery_ops 2026-03-20T17:57:05.469 INFO:teuthology.orchestra.run.vm05.stdout: osd mclock skip benchmark = true 2026-03-20T17:57:05.469 INFO:teuthology.orchestra.run.vm05.stdout: 2026-03-20T17:57:05.470 INFO:teuthology.orchestra.run.vm05.stdout: osd recover clone overlap = true 2026-03-20T17:57:05.470 INFO:teuthology.orchestra.run.vm05.stdout: osd recovery max chunk = 1048576 2026-03-20T17:57:05.470 INFO:teuthology.orchestra.run.vm05.stdout: 2026-03-20T17:57:05.470 INFO:teuthology.orchestra.run.vm05.stdout: osd debug shutdown = true 2026-03-20T17:57:05.470 INFO:teuthology.orchestra.run.vm05.stdout: osd debug op order = true 2026-03-20T17:57:05.470 INFO:teuthology.orchestra.run.vm05.stdout: osd debug verify stray on activate = true 2026-03-20T17:57:05.470 INFO:teuthology.orchestra.run.vm05.stdout: osd debug trim objects = true 2026-03-20T17:57:05.470 INFO:teuthology.orchestra.run.vm05.stdout: 2026-03-20T17:57:05.470 INFO:teuthology.orchestra.run.vm05.stdout: osd open classes on start = true 2026-03-20T17:57:05.470 INFO:teuthology.orchestra.run.vm05.stdout: osd debug pg log writeout = true 2026-03-20T17:57:05.470 INFO:teuthology.orchestra.run.vm05.stdout: 2026-03-20T17:57:05.470 INFO:teuthology.orchestra.run.vm05.stdout: osd deep scrub update digest min age = 30 2026-03-20T17:57:05.470 INFO:teuthology.orchestra.run.vm05.stdout: 2026-03-20T17:57:05.470 INFO:teuthology.orchestra.run.vm05.stdout: osd map max advance = 10 2026-03-20T17:57:05.470 INFO:teuthology.orchestra.run.vm05.stdout: 2026-03-20T17:57:05.470 INFO:teuthology.orchestra.run.vm05.stdout: journal zero on create = true 2026-03-20T17:57:05.470 INFO:teuthology.orchestra.run.vm05.stdout: 2026-03-20T17:57:05.470 INFO:teuthology.orchestra.run.vm05.stdout: filestore ondisk finisher threads = 3 2026-03-20T17:57:05.470 INFO:teuthology.orchestra.run.vm05.stdout: filestore apply finisher threads = 3 2026-03-20T17:57:05.470 INFO:teuthology.orchestra.run.vm05.stdout: 2026-03-20T17:57:05.470 INFO:teuthology.orchestra.run.vm05.stdout: bdev debug aio = true 2026-03-20T17:57:05.470 INFO:teuthology.orchestra.run.vm05.stdout: osd debug misdirected ops = true 2026-03-20T17:57:05.470 INFO:teuthology.orchestra.run.vm05.stdout: debug ms = 1 2026-03-20T17:57:05.470 INFO:teuthology.orchestra.run.vm05.stdout: debug osd = 20 2026-03-20T17:57:05.470 INFO:teuthology.orchestra.run.vm05.stdout: osd mclock iops capacity threshold hdd = 49000 2026-03-20T17:57:05.470 INFO:teuthology.orchestra.run.vm05.stdout: rocksdb delete range threshold = 0 2026-03-20T17:57:05.470 INFO:teuthology.orchestra.run.vm05.stdout: 2026-03-20T17:57:05.470 INFO:teuthology.orchestra.run.vm05.stdout:[mgr] 2026-03-20T17:57:05.470 INFO:teuthology.orchestra.run.vm05.stdout: debug ms = 1 2026-03-20T17:57:05.470 INFO:teuthology.orchestra.run.vm05.stdout: debug mgr = 20 2026-03-20T17:57:05.470 INFO:teuthology.orchestra.run.vm05.stdout: debug mon = 20 2026-03-20T17:57:05.470 INFO:teuthology.orchestra.run.vm05.stdout: debug auth = 20 2026-03-20T17:57:05.470 INFO:teuthology.orchestra.run.vm05.stdout: mon reweight min pgs per osd = 4 2026-03-20T17:57:05.470 INFO:teuthology.orchestra.run.vm05.stdout: mon reweight min bytes per osd = 10 2026-03-20T17:57:05.470 INFO:teuthology.orchestra.run.vm05.stdout: mgr/telemetry/nag = false 2026-03-20T17:57:05.470 INFO:teuthology.orchestra.run.vm05.stdout: 2026-03-20T17:57:05.470 INFO:teuthology.orchestra.run.vm05.stdout:[mon] 2026-03-20T17:57:05.470 INFO:teuthology.orchestra.run.vm05.stdout: debug ms = 1 2026-03-20T17:57:05.470 INFO:teuthology.orchestra.run.vm05.stdout: debug mon = 20 2026-03-20T17:57:05.470 INFO:teuthology.orchestra.run.vm05.stdout: debug paxos = 20 2026-03-20T17:57:05.470 INFO:teuthology.orchestra.run.vm05.stdout: debug auth = 20 2026-03-20T17:57:05.470 INFO:teuthology.orchestra.run.vm05.stdout: mon data avail warn = 5 2026-03-20T17:57:05.470 INFO:teuthology.orchestra.run.vm05.stdout: mon mgr mkfs grace = 240 2026-03-20T17:57:05.470 INFO:teuthology.orchestra.run.vm05.stdout: mon reweight min pgs per osd = 4 2026-03-20T17:57:05.470 INFO:teuthology.orchestra.run.vm05.stdout: mon osd reporter subtree level = osd 2026-03-20T17:57:05.470 INFO:teuthology.orchestra.run.vm05.stdout: mon osd prime pg temp = true 2026-03-20T17:57:05.470 INFO:teuthology.orchestra.run.vm05.stdout: mon reweight min bytes per osd = 10 2026-03-20T17:57:05.470 INFO:teuthology.orchestra.run.vm05.stdout: 2026-03-20T17:57:05.470 INFO:teuthology.orchestra.run.vm05.stdout: # rotate auth tickets quickly to exercise renewal paths 2026-03-20T17:57:05.470 INFO:teuthology.orchestra.run.vm05.stdout: auth mon ticket ttl = 660 # 11m 2026-03-20T17:57:05.470 INFO:teuthology.orchestra.run.vm05.stdout: auth service ticket ttl = 240 # 4m 2026-03-20T17:57:05.470 INFO:teuthology.orchestra.run.vm05.stdout: 2026-03-20T17:57:05.470 INFO:teuthology.orchestra.run.vm05.stdout: # don't complain about insecure global_id in the test suite 2026-03-20T17:57:05.470 INFO:teuthology.orchestra.run.vm05.stdout: mon_warn_on_insecure_global_id_reclaim = false 2026-03-20T17:57:05.470 INFO:teuthology.orchestra.run.vm05.stdout: mon_warn_on_insecure_global_id_reclaim_allowed = false 2026-03-20T17:57:05.470 INFO:teuthology.orchestra.run.vm05.stdout: 2026-03-20T17:57:05.470 INFO:teuthology.orchestra.run.vm05.stdout: # 1m isn't quite enough 2026-03-20T17:57:05.470 INFO:teuthology.orchestra.run.vm05.stdout: mon_down_mkfs_grace = 2m 2026-03-20T17:57:05.470 INFO:teuthology.orchestra.run.vm05.stdout: 2026-03-20T17:57:05.470 INFO:teuthology.orchestra.run.vm05.stdout: mon_warn_on_filestore_osds = false 2026-03-20T17:57:05.470 INFO:teuthology.orchestra.run.vm05.stdout: 2026-03-20T17:57:05.470 INFO:teuthology.orchestra.run.vm05.stdout:[client] 2026-03-20T17:57:05.470 INFO:teuthology.orchestra.run.vm05.stdout: rgw cache enabled = true 2026-03-20T17:57:05.470 INFO:teuthology.orchestra.run.vm05.stdout: rgw enable ops log = true 2026-03-20T17:57:05.470 INFO:teuthology.orchestra.run.vm05.stdout: rgw enable usage log = true 2026-03-20T17:57:05.471 INFO:teuthology.orchestra.run.vm05.stdout: log file = /var/log/ceph/$cluster-$name.$pid.log 2026-03-20T17:57:05.471 INFO:teuthology.orchestra.run.vm05.stdout: admin socket = /var/run/ceph/$cluster-$name.$pid.asok 2026-03-20T17:57:05.471 INFO:teuthology.orchestra.run.vm05.stdout: debug rgw = 20 2026-03-20T17:57:05.471 INFO:teuthology.orchestra.run.vm05.stdout: debug rgw lifecycle = 20 2026-03-20T17:57:05.471 INFO:teuthology.orchestra.run.vm05.stdout: debug rgw notification = 20 2026-03-20T17:57:05.471 INFO:teuthology.orchestra.run.vm05.stdout: debug rgw sync = 20 2026-03-20T17:57:05.471 INFO:teuthology.orchestra.run.vm05.stdout: rgw crypt require ssl = False 2026-03-20T17:57:05.471 INFO:teuthology.orchestra.run.vm05.stdout: rgw crypt s3 kms backend = testing 2026-03-20T17:57:05.471 INFO:teuthology.orchestra.run.vm05.stdout: rgw crypt s3 kms encryption keys = testkey-1=YmluCmJvb3N0CmJvb3N0LWJ1aWxkCmNlcGguY29uZgo= 2026-03-20T17:57:05.471 INFO:teuthology.orchestra.run.vm05.stdout: rgw curl low speed time = 300 2026-03-20T17:57:05.471 INFO:teuthology.orchestra.run.vm05.stdout: rgw data log num shards = 4 2026-03-20T17:57:05.471 INFO:teuthology.orchestra.run.vm05.stdout: rgw data sync poll interval = 5 2026-03-20T17:57:05.471 INFO:teuthology.orchestra.run.vm05.stdout: rgw md log max shards = 4 2026-03-20T17:57:05.471 INFO:teuthology.orchestra.run.vm05.stdout: rgw meta sync poll interval = 5 2026-03-20T17:57:05.471 INFO:teuthology.orchestra.run.vm05.stdout: rgw s3 auth use sts = True 2026-03-20T17:57:05.471 INFO:teuthology.orchestra.run.vm05.stdout: rgw sts key = abcdefghijklmnoq 2026-03-20T17:57:05.471 INFO:teuthology.orchestra.run.vm05.stdout: rgw sync data inject err probability = 0 2026-03-20T17:57:05.471 INFO:teuthology.orchestra.run.vm05.stdout: rgw sync log trim interval = 0 2026-03-20T17:57:05.471 INFO:teuthology.orchestra.run.vm05.stdout: rgw sync meta inject err probability = 0 2026-03-20T17:57:05.471 INFO:teuthology.orchestra.run.vm05.stdout: rgw sync obj etag verify = True 2026-03-20T17:57:05.471 INFO:teuthology.orchestra.run.vm05.stdout: setgroup = ceph 2026-03-20T17:57:05.471 INFO:teuthology.orchestra.run.vm05.stdout: setuser = ceph 2026-03-20T17:57:05.471 INFO:teuthology.orchestra.run.vm05.stdout:[mon.a] 2026-03-20T17:57:05.471 INFO:teuthology.orchestra.run.vm05.stdout:[client.0] 2026-03-20T17:57:05.471 INFO:teuthology.orchestra.run.vm05.stdout: rgw data notify interval msec = 0 2026-03-20T17:57:05.471 INFO:teuthology.orchestra.run.vm05.stdout:[osd.0] 2026-03-20T17:57:05.471 INFO:teuthology.orchestra.run.vm05.stdout: osd_max_omap_entries_per_request = 10 2026-03-20T17:57:05.471 INFO:teuthology.orchestra.run.vm05.stdout:[osd.1] 2026-03-20T17:57:05.471 INFO:teuthology.orchestra.run.vm05.stdout: osd_max_omap_entries_per_request = 1000 2026-03-20T17:57:05.471 INFO:teuthology.orchestra.run.vm05.stdout:[osd.2] 2026-03-20T17:57:05.471 INFO:teuthology.orchestra.run.vm05.stdout: osd_max_omap_entries_per_request = 10000 2026-03-20T17:57:05.471 INFO:teuthology.orchestra.run.vm06.stdout:[global] 2026-03-20T17:57:05.471 INFO:teuthology.orchestra.run.vm06.stdout: chdir = "" 2026-03-20T17:57:05.471 INFO:teuthology.orchestra.run.vm06.stdout: pid file = /var/run/ceph/$cluster-$name.pid 2026-03-20T17:57:05.471 INFO:teuthology.orchestra.run.vm06.stdout: auth supported = cephx 2026-03-20T17:57:05.471 INFO:teuthology.orchestra.run.vm06.stdout: 2026-03-20T17:57:05.472 INFO:teuthology.orchestra.run.vm06.stdout: filestore xattr use omap = true 2026-03-20T17:57:05.472 INFO:teuthology.orchestra.run.vm06.stdout: 2026-03-20T17:57:05.472 INFO:teuthology.orchestra.run.vm06.stdout: mon clock drift allowed = 1.000 2026-03-20T17:57:05.472 INFO:teuthology.orchestra.run.vm06.stdout: 2026-03-20T17:57:05.472 INFO:teuthology.orchestra.run.vm06.stdout: osd crush chooseleaf type = 0 2026-03-20T17:57:05.472 INFO:teuthology.orchestra.run.vm06.stdout: auth debug = true 2026-03-20T17:57:05.472 INFO:teuthology.orchestra.run.vm06.stdout: 2026-03-20T17:57:05.472 INFO:teuthology.orchestra.run.vm06.stdout: ms die on old message = true 2026-03-20T17:57:05.472 INFO:teuthology.orchestra.run.vm06.stdout: ms die on bug = true 2026-03-20T17:57:05.472 INFO:teuthology.orchestra.run.vm06.stdout: 2026-03-20T17:57:05.472 INFO:teuthology.orchestra.run.vm06.stdout: mon max pg per osd = 10000 # >= luminous 2026-03-20T17:57:05.472 INFO:teuthology.orchestra.run.vm06.stdout: mon pg warn max object skew = 0 2026-03-20T17:57:05.472 INFO:teuthology.orchestra.run.vm06.stdout: 2026-03-20T17:57:05.472 INFO:teuthology.orchestra.run.vm06.stdout: # disable pg_autoscaler by default for new pools 2026-03-20T17:57:05.472 INFO:teuthology.orchestra.run.vm06.stdout: osd_pool_default_pg_autoscale_mode = off 2026-03-20T17:57:05.472 INFO:teuthology.orchestra.run.vm06.stdout: 2026-03-20T17:57:05.472 INFO:teuthology.orchestra.run.vm06.stdout: osd pool default size = 2 2026-03-20T17:57:05.472 INFO:teuthology.orchestra.run.vm06.stdout: 2026-03-20T17:57:05.472 INFO:teuthology.orchestra.run.vm06.stdout: mon osd allow primary affinity = true 2026-03-20T17:57:05.472 INFO:teuthology.orchestra.run.vm06.stdout: mon osd allow pg remap = true 2026-03-20T17:57:05.472 INFO:teuthology.orchestra.run.vm06.stdout: mon warn on legacy crush tunables = false 2026-03-20T17:57:05.472 INFO:teuthology.orchestra.run.vm06.stdout: mon warn on crush straw calc version zero = false 2026-03-20T17:57:05.472 INFO:teuthology.orchestra.run.vm06.stdout: mon warn on no sortbitwise = false 2026-03-20T17:57:05.472 INFO:teuthology.orchestra.run.vm06.stdout: mon warn on osd down out interval zero = false 2026-03-20T17:57:05.472 INFO:teuthology.orchestra.run.vm06.stdout: mon warn on too few osds = false 2026-03-20T17:57:05.472 INFO:teuthology.orchestra.run.vm06.stdout: mon_warn_on_pool_pg_num_not_power_of_two = false 2026-03-20T17:57:05.472 INFO:teuthology.orchestra.run.vm06.stdout: mon_warn_on_pool_no_redundancy = false 2026-03-20T17:57:05.472 INFO:teuthology.orchestra.run.vm06.stdout: mon_allow_pool_size_one = true 2026-03-20T17:57:05.472 INFO:teuthology.orchestra.run.vm06.stdout: 2026-03-20T17:57:05.472 INFO:teuthology.orchestra.run.vm06.stdout: osd pool default erasure code profile = plugin=isa technique=reed_sol_van k=2 m=1 crush-failure-domain=osd 2026-03-20T17:57:05.472 INFO:teuthology.orchestra.run.vm06.stdout: 2026-03-20T17:57:05.472 INFO:teuthology.orchestra.run.vm06.stdout: osd default data pool replay window = 5 2026-03-20T17:57:05.472 INFO:teuthology.orchestra.run.vm06.stdout: 2026-03-20T17:57:05.472 INFO:teuthology.orchestra.run.vm06.stdout: mon allow pool delete = true 2026-03-20T17:57:05.472 INFO:teuthology.orchestra.run.vm06.stdout: 2026-03-20T17:57:05.472 INFO:teuthology.orchestra.run.vm06.stdout: mon cluster log file level = debug 2026-03-20T17:57:05.472 INFO:teuthology.orchestra.run.vm06.stdout: debug asserts on shutdown = true 2026-03-20T17:57:05.472 INFO:teuthology.orchestra.run.vm06.stdout: mon health detail to clog = false 2026-03-20T17:57:05.472 INFO:teuthology.orchestra.run.vm06.stdout: mon host = 192.168.123.102 2026-03-20T17:57:05.472 INFO:teuthology.orchestra.run.vm06.stdout: fsid = 67b0dd48-80db-4ea8-b708-f5136e717957 2026-03-20T17:57:05.472 INFO:teuthology.orchestra.run.vm06.stdout: 2026-03-20T17:57:05.472 INFO:teuthology.orchestra.run.vm06.stdout:[osd] 2026-03-20T17:57:05.472 INFO:teuthology.orchestra.run.vm06.stdout: osd journal size = 100 2026-03-20T17:57:05.472 INFO:teuthology.orchestra.run.vm06.stdout: 2026-03-20T17:57:05.472 INFO:teuthology.orchestra.run.vm06.stdout: osd scrub load threshold = 5.0 2026-03-20T17:57:05.472 INFO:teuthology.orchestra.run.vm06.stdout: osd scrub max interval = 600 2026-03-20T17:57:05.472 INFO:teuthology.orchestra.run.vm06.stdout: osd mclock profile = high_recovery_ops 2026-03-20T17:57:05.472 INFO:teuthology.orchestra.run.vm06.stdout: osd mclock skip benchmark = true 2026-03-20T17:57:05.472 INFO:teuthology.orchestra.run.vm06.stdout: 2026-03-20T17:57:05.472 INFO:teuthology.orchestra.run.vm06.stdout: osd recover clone overlap = true 2026-03-20T17:57:05.472 INFO:teuthology.orchestra.run.vm06.stdout: osd recovery max chunk = 1048576 2026-03-20T17:57:05.472 INFO:teuthology.orchestra.run.vm06.stdout: 2026-03-20T17:57:05.472 INFO:teuthology.orchestra.run.vm06.stdout: osd debug shutdown = true 2026-03-20T17:57:05.472 INFO:teuthology.orchestra.run.vm06.stdout: osd debug op order = true 2026-03-20T17:57:05.472 INFO:teuthology.orchestra.run.vm06.stdout: osd debug verify stray on activate = true 2026-03-20T17:57:05.472 INFO:teuthology.orchestra.run.vm06.stdout: osd debug trim objects = true 2026-03-20T17:57:05.472 INFO:teuthology.orchestra.run.vm06.stdout: 2026-03-20T17:57:05.472 INFO:teuthology.orchestra.run.vm06.stdout: osd open classes on start = true 2026-03-20T17:57:05.472 INFO:teuthology.orchestra.run.vm06.stdout: osd debug pg log writeout = true 2026-03-20T17:57:05.472 INFO:teuthology.orchestra.run.vm06.stdout: 2026-03-20T17:57:05.472 INFO:teuthology.orchestra.run.vm06.stdout: osd deep scrub update digest min age = 30 2026-03-20T17:57:05.472 INFO:teuthology.orchestra.run.vm06.stdout: 2026-03-20T17:57:05.472 INFO:teuthology.orchestra.run.vm06.stdout: osd map max advance = 10 2026-03-20T17:57:05.472 INFO:teuthology.orchestra.run.vm06.stdout: 2026-03-20T17:57:05.472 INFO:teuthology.orchestra.run.vm06.stdout: journal zero on create = true 2026-03-20T17:57:05.473 INFO:teuthology.orchestra.run.vm06.stdout: 2026-03-20T17:57:05.473 INFO:teuthology.orchestra.run.vm06.stdout: filestore ondisk finisher threads = 3 2026-03-20T17:57:05.473 INFO:teuthology.orchestra.run.vm06.stdout: filestore apply finisher threads = 3 2026-03-20T17:57:05.473 INFO:teuthology.orchestra.run.vm06.stdout: 2026-03-20T17:57:05.473 INFO:teuthology.orchestra.run.vm06.stdout: bdev debug aio = true 2026-03-20T17:57:05.473 INFO:teuthology.orchestra.run.vm06.stdout: osd debug misdirected ops = true 2026-03-20T17:57:05.473 INFO:teuthology.orchestra.run.vm06.stdout: debug ms = 1 2026-03-20T17:57:05.473 INFO:teuthology.orchestra.run.vm06.stdout: debug osd = 20 2026-03-20T17:57:05.473 INFO:teuthology.orchestra.run.vm06.stdout: osd mclock iops capacity threshold hdd = 49000 2026-03-20T17:57:05.473 INFO:teuthology.orchestra.run.vm06.stdout: rocksdb delete range threshold = 0 2026-03-20T17:57:05.473 INFO:teuthology.orchestra.run.vm06.stdout: 2026-03-20T17:57:05.473 INFO:teuthology.orchestra.run.vm06.stdout:[mgr] 2026-03-20T17:57:05.473 INFO:teuthology.orchestra.run.vm06.stdout: debug ms = 1 2026-03-20T17:57:05.473 INFO:teuthology.orchestra.run.vm06.stdout: debug mgr = 20 2026-03-20T17:57:05.473 INFO:teuthology.orchestra.run.vm06.stdout: debug mon = 20 2026-03-20T17:57:05.473 INFO:teuthology.orchestra.run.vm06.stdout: debug auth = 20 2026-03-20T17:57:05.473 INFO:teuthology.orchestra.run.vm06.stdout: mon reweight min pgs per osd = 4 2026-03-20T17:57:05.473 INFO:teuthology.orchestra.run.vm06.stdout: mon reweight min bytes per osd = 10 2026-03-20T17:57:05.473 INFO:teuthology.orchestra.run.vm06.stdout: mgr/telemetry/nag = false 2026-03-20T17:57:05.473 INFO:teuthology.orchestra.run.vm06.stdout: 2026-03-20T17:57:05.473 INFO:teuthology.orchestra.run.vm06.stdout:[mon] 2026-03-20T17:57:05.473 INFO:teuthology.orchestra.run.vm06.stdout: debug ms = 1 2026-03-20T17:57:05.473 INFO:teuthology.orchestra.run.vm06.stdout: debug mon = 20 2026-03-20T17:57:05.473 INFO:teuthology.orchestra.run.vm06.stdout: debug paxos = 20 2026-03-20T17:57:05.473 INFO:teuthology.orchestra.run.vm06.stdout: debug auth = 20 2026-03-20T17:57:05.473 INFO:teuthology.orchestra.run.vm06.stdout: mon data avail warn = 5 2026-03-20T17:57:05.473 INFO:teuthology.orchestra.run.vm06.stdout: mon mgr mkfs grace = 240 2026-03-20T17:57:05.473 INFO:teuthology.orchestra.run.vm06.stdout: mon reweight min pgs per osd = 4 2026-03-20T17:57:05.473 INFO:teuthology.orchestra.run.vm06.stdout: mon osd reporter subtree level = osd 2026-03-20T17:57:05.473 INFO:teuthology.orchestra.run.vm06.stdout: mon osd prime pg temp = true 2026-03-20T17:57:05.473 INFO:teuthology.orchestra.run.vm06.stdout: mon reweight min bytes per osd = 10 2026-03-20T17:57:05.473 INFO:teuthology.orchestra.run.vm06.stdout: 2026-03-20T17:57:05.473 INFO:teuthology.orchestra.run.vm06.stdout: # rotate auth tickets quickly to exercise renewal paths 2026-03-20T17:57:05.473 INFO:teuthology.orchestra.run.vm06.stdout: auth mon ticket ttl = 660 # 11m 2026-03-20T17:57:05.473 INFO:teuthology.orchestra.run.vm06.stdout: auth service ticket ttl = 240 # 4m 2026-03-20T17:57:05.473 INFO:teuthology.orchestra.run.vm06.stdout: 2026-03-20T17:57:05.473 INFO:teuthology.orchestra.run.vm06.stdout: # don't complain about insecure global_id in the test suite 2026-03-20T17:57:05.473 INFO:teuthology.orchestra.run.vm06.stdout: mon_warn_on_insecure_global_id_reclaim = false 2026-03-20T17:57:05.473 INFO:teuthology.orchestra.run.vm06.stdout: mon_warn_on_insecure_global_id_reclaim_allowed = false 2026-03-20T17:57:05.473 INFO:teuthology.orchestra.run.vm06.stdout: 2026-03-20T17:57:05.473 INFO:teuthology.orchestra.run.vm06.stdout: # 1m isn't quite enough 2026-03-20T17:57:05.473 INFO:teuthology.orchestra.run.vm06.stdout: mon_down_mkfs_grace = 2m 2026-03-20T17:57:05.473 INFO:teuthology.orchestra.run.vm06.stdout: 2026-03-20T17:57:05.473 INFO:teuthology.orchestra.run.vm06.stdout: mon_warn_on_filestore_osds = false 2026-03-20T17:57:05.473 INFO:teuthology.orchestra.run.vm06.stdout: 2026-03-20T17:57:05.473 INFO:teuthology.orchestra.run.vm06.stdout:[client] 2026-03-20T17:57:05.473 INFO:teuthology.orchestra.run.vm06.stdout: rgw cache enabled = true 2026-03-20T17:57:05.473 INFO:teuthology.orchestra.run.vm06.stdout: rgw enable ops log = true 2026-03-20T17:57:05.473 INFO:teuthology.orchestra.run.vm06.stdout: rgw enable usage log = true 2026-03-20T17:57:05.473 INFO:teuthology.orchestra.run.vm06.stdout: log file = /var/log/ceph/$cluster-$name.$pid.log 2026-03-20T17:57:05.473 INFO:teuthology.orchestra.run.vm06.stdout: admin socket = /var/run/ceph/$cluster-$name.$pid.asok 2026-03-20T17:57:05.473 INFO:teuthology.orchestra.run.vm06.stdout: debug rgw = 20 2026-03-20T17:57:05.473 INFO:teuthology.orchestra.run.vm06.stdout: debug rgw lifecycle = 20 2026-03-20T17:57:05.473 INFO:teuthology.orchestra.run.vm06.stdout: debug rgw notification = 20 2026-03-20T17:57:05.473 INFO:teuthology.orchestra.run.vm06.stdout: debug rgw sync = 20 2026-03-20T17:57:05.473 INFO:teuthology.orchestra.run.vm06.stdout: rgw crypt require ssl = False 2026-03-20T17:57:05.473 INFO:teuthology.orchestra.run.vm06.stdout: rgw crypt s3 kms backend = testing 2026-03-20T17:57:05.473 INFO:teuthology.orchestra.run.vm06.stdout: rgw crypt s3 kms encryption keys = testkey-1=YmluCmJvb3N0CmJvb3N0LWJ1aWxkCmNlcGguY29uZgo= 2026-03-20T17:57:05.473 INFO:teuthology.orchestra.run.vm06.stdout: rgw curl low speed time = 300 2026-03-20T17:57:05.473 INFO:teuthology.orchestra.run.vm06.stdout: rgw data log num shards = 4 2026-03-20T17:57:05.473 INFO:teuthology.orchestra.run.vm06.stdout: rgw data sync poll interval = 5 2026-03-20T17:57:05.474 INFO:teuthology.orchestra.run.vm06.stdout: rgw md log max shards = 4 2026-03-20T17:57:05.474 INFO:teuthology.orchestra.run.vm06.stdout: rgw meta sync poll interval = 5 2026-03-20T17:57:05.474 INFO:teuthology.orchestra.run.vm06.stdout: rgw s3 auth use sts = True 2026-03-20T17:57:05.474 INFO:teuthology.orchestra.run.vm06.stdout: rgw sts key = abcdefghijklmnoq 2026-03-20T17:57:05.474 INFO:teuthology.orchestra.run.vm06.stdout: rgw sync data inject err probability = 0 2026-03-20T17:57:05.474 INFO:teuthology.orchestra.run.vm06.stdout: rgw sync log trim interval = 0 2026-03-20T17:57:05.474 INFO:teuthology.orchestra.run.vm06.stdout: rgw sync meta inject err probability = 0 2026-03-20T17:57:05.474 INFO:teuthology.orchestra.run.vm06.stdout: rgw sync obj etag verify = True 2026-03-20T17:57:05.474 INFO:teuthology.orchestra.run.vm06.stdout: setgroup = ceph 2026-03-20T17:57:05.474 INFO:teuthology.orchestra.run.vm06.stdout: setuser = ceph 2026-03-20T17:57:05.474 INFO:teuthology.orchestra.run.vm06.stdout:[mon.a] 2026-03-20T17:57:05.474 INFO:teuthology.orchestra.run.vm06.stdout:[client.0] 2026-03-20T17:57:05.474 INFO:teuthology.orchestra.run.vm06.stdout: rgw data notify interval msec = 0 2026-03-20T17:57:05.474 INFO:teuthology.orchestra.run.vm06.stdout:[osd.0] 2026-03-20T17:57:05.474 INFO:teuthology.orchestra.run.vm06.stdout: osd_max_omap_entries_per_request = 10 2026-03-20T17:57:05.474 INFO:teuthology.orchestra.run.vm06.stdout:[osd.1] 2026-03-20T17:57:05.474 INFO:teuthology.orchestra.run.vm06.stdout: osd_max_omap_entries_per_request = 1000 2026-03-20T17:57:05.474 INFO:teuthology.orchestra.run.vm06.stdout:[osd.2] 2026-03-20T17:57:05.474 INFO:teuthology.orchestra.run.vm06.stdout: osd_max_omap_entries_per_request = 10000 2026-03-20T17:57:05.474 INFO:teuthology.orchestra.run.vm08.stdout:[global] 2026-03-20T17:57:05.474 INFO:teuthology.orchestra.run.vm08.stdout: chdir = "" 2026-03-20T17:57:05.474 INFO:teuthology.orchestra.run.vm08.stdout: pid file = /var/run/ceph/$cluster-$name.pid 2026-03-20T17:57:05.474 INFO:teuthology.orchestra.run.vm08.stdout: auth supported = cephx 2026-03-20T17:57:05.474 INFO:teuthology.orchestra.run.vm08.stdout: 2026-03-20T17:57:05.474 INFO:teuthology.orchestra.run.vm08.stdout: filestore xattr use omap = true 2026-03-20T17:57:05.474 INFO:teuthology.orchestra.run.vm08.stdout: 2026-03-20T17:57:05.474 INFO:teuthology.orchestra.run.vm08.stdout: mon clock drift allowed = 1.000 2026-03-20T17:57:05.474 INFO:teuthology.orchestra.run.vm08.stdout: 2026-03-20T17:57:05.474 INFO:teuthology.orchestra.run.vm08.stdout: osd crush chooseleaf type = 0 2026-03-20T17:57:05.474 INFO:teuthology.orchestra.run.vm08.stdout: auth debug = true 2026-03-20T17:57:05.474 INFO:teuthology.orchestra.run.vm08.stdout: 2026-03-20T17:57:05.474 INFO:teuthology.orchestra.run.vm08.stdout: ms die on old message = true 2026-03-20T17:57:05.474 INFO:teuthology.orchestra.run.vm08.stdout: ms die on bug = true 2026-03-20T17:57:05.474 INFO:teuthology.orchestra.run.vm08.stdout: 2026-03-20T17:57:05.474 INFO:teuthology.orchestra.run.vm08.stdout: mon max pg per osd = 10000 # >= luminous 2026-03-20T17:57:05.474 INFO:teuthology.orchestra.run.vm08.stdout: mon pg warn max object skew = 0 2026-03-20T17:57:05.474 INFO:teuthology.orchestra.run.vm08.stdout: 2026-03-20T17:57:05.474 INFO:teuthology.orchestra.run.vm08.stdout: # disable pg_autoscaler by default for new pools 2026-03-20T17:57:05.474 INFO:teuthology.orchestra.run.vm08.stdout: osd_pool_default_pg_autoscale_mode = off 2026-03-20T17:57:05.474 INFO:teuthology.orchestra.run.vm08.stdout: 2026-03-20T17:57:05.474 INFO:teuthology.orchestra.run.vm08.stdout: osd pool default size = 2 2026-03-20T17:57:05.474 INFO:teuthology.orchestra.run.vm08.stdout: 2026-03-20T17:57:05.474 INFO:teuthology.orchestra.run.vm08.stdout: mon osd allow primary affinity = true 2026-03-20T17:57:05.474 INFO:teuthology.orchestra.run.vm08.stdout: mon osd allow pg remap = true 2026-03-20T17:57:05.474 INFO:teuthology.orchestra.run.vm08.stdout: mon warn on legacy crush tunables = false 2026-03-20T17:57:05.474 INFO:teuthology.orchestra.run.vm08.stdout: mon warn on crush straw calc version zero = false 2026-03-20T17:57:05.474 INFO:teuthology.orchestra.run.vm08.stdout: mon warn on no sortbitwise = false 2026-03-20T17:57:05.474 INFO:teuthology.orchestra.run.vm08.stdout: mon warn on osd down out interval zero = false 2026-03-20T17:57:05.475 INFO:teuthology.orchestra.run.vm08.stdout: mon warn on too few osds = false 2026-03-20T17:57:05.475 INFO:teuthology.orchestra.run.vm08.stdout: mon_warn_on_pool_pg_num_not_power_of_two = false 2026-03-20T17:57:05.475 INFO:teuthology.orchestra.run.vm08.stdout: mon_warn_on_pool_no_redundancy = false 2026-03-20T17:57:05.475 INFO:teuthology.orchestra.run.vm08.stdout: mon_allow_pool_size_one = true 2026-03-20T17:57:05.475 INFO:teuthology.orchestra.run.vm08.stdout: 2026-03-20T17:57:05.475 INFO:teuthology.orchestra.run.vm08.stdout: osd pool default erasure code profile = plugin=isa technique=reed_sol_van k=2 m=1 crush-failure-domain=osd 2026-03-20T17:57:05.475 INFO:teuthology.orchestra.run.vm08.stdout: 2026-03-20T17:57:05.475 INFO:teuthology.orchestra.run.vm08.stdout: osd default data pool replay window = 5 2026-03-20T17:57:05.475 INFO:teuthology.orchestra.run.vm08.stdout: 2026-03-20T17:57:05.475 INFO:teuthology.orchestra.run.vm08.stdout: mon allow pool delete = true 2026-03-20T17:57:05.475 INFO:teuthology.orchestra.run.vm08.stdout: 2026-03-20T17:57:05.475 INFO:teuthology.orchestra.run.vm08.stdout: mon cluster log file level = debug 2026-03-20T17:57:05.475 INFO:teuthology.orchestra.run.vm08.stdout: debug asserts on shutdown = true 2026-03-20T17:57:05.475 INFO:teuthology.orchestra.run.vm08.stdout: mon health detail to clog = false 2026-03-20T17:57:05.475 INFO:teuthology.orchestra.run.vm08.stdout: mon host = 192.168.123.102 2026-03-20T17:57:05.475 INFO:teuthology.orchestra.run.vm08.stdout: fsid = 67b0dd48-80db-4ea8-b708-f5136e717957 2026-03-20T17:57:05.475 INFO:teuthology.orchestra.run.vm08.stdout: 2026-03-20T17:57:05.475 INFO:teuthology.orchestra.run.vm08.stdout:[osd] 2026-03-20T17:57:05.475 INFO:teuthology.orchestra.run.vm08.stdout: osd journal size = 100 2026-03-20T17:57:05.475 INFO:teuthology.orchestra.run.vm08.stdout: 2026-03-20T17:57:05.475 INFO:teuthology.orchestra.run.vm08.stdout: osd scrub load threshold = 5.0 2026-03-20T17:57:05.475 INFO:teuthology.orchestra.run.vm08.stdout: osd scrub max interval = 600 2026-03-20T17:57:05.475 INFO:teuthology.orchestra.run.vm08.stdout: osd mclock profile = high_recovery_ops 2026-03-20T17:57:05.475 INFO:teuthology.orchestra.run.vm08.stdout: osd mclock skip benchmark = true 2026-03-20T17:57:05.475 INFO:teuthology.orchestra.run.vm08.stdout: 2026-03-20T17:57:05.475 INFO:teuthology.orchestra.run.vm08.stdout: osd recover clone overlap = true 2026-03-20T17:57:05.475 INFO:teuthology.orchestra.run.vm08.stdout: osd recovery max chunk = 1048576 2026-03-20T17:57:05.475 INFO:teuthology.orchestra.run.vm08.stdout: 2026-03-20T17:57:05.475 INFO:teuthology.orchestra.run.vm08.stdout: osd debug shutdown = true 2026-03-20T17:57:05.475 INFO:teuthology.orchestra.run.vm08.stdout: osd debug op order = true 2026-03-20T17:57:05.475 INFO:teuthology.orchestra.run.vm08.stdout: osd debug verify stray on activate = true 2026-03-20T17:57:05.475 INFO:teuthology.orchestra.run.vm08.stdout: osd debug trim objects = true 2026-03-20T17:57:05.475 INFO:teuthology.orchestra.run.vm08.stdout: 2026-03-20T17:57:05.475 INFO:teuthology.orchestra.run.vm08.stdout: osd open classes on start = true 2026-03-20T17:57:05.475 INFO:teuthology.orchestra.run.vm08.stdout: osd debug pg log writeout = true 2026-03-20T17:57:05.475 INFO:teuthology.orchestra.run.vm08.stdout: 2026-03-20T17:57:05.475 INFO:teuthology.orchestra.run.vm08.stdout: osd deep scrub update digest min age = 30 2026-03-20T17:57:05.475 INFO:teuthology.orchestra.run.vm08.stdout: 2026-03-20T17:57:05.475 INFO:teuthology.orchestra.run.vm08.stdout: osd map max advance = 10 2026-03-20T17:57:05.475 INFO:teuthology.orchestra.run.vm08.stdout: 2026-03-20T17:57:05.475 INFO:teuthology.orchestra.run.vm08.stdout: journal zero on create = true 2026-03-20T17:57:05.475 INFO:teuthology.orchestra.run.vm08.stdout: 2026-03-20T17:57:05.475 INFO:teuthology.orchestra.run.vm08.stdout: filestore ondisk finisher threads = 3 2026-03-20T17:57:05.475 INFO:teuthology.orchestra.run.vm08.stdout: filestore apply finisher threads = 3 2026-03-20T17:57:05.475 INFO:teuthology.orchestra.run.vm08.stdout: 2026-03-20T17:57:05.475 INFO:teuthology.orchestra.run.vm08.stdout: bdev debug aio = true 2026-03-20T17:57:05.475 INFO:teuthology.orchestra.run.vm08.stdout: osd debug misdirected ops = true 2026-03-20T17:57:05.475 INFO:teuthology.orchestra.run.vm08.stdout: debug ms = 1 2026-03-20T17:57:05.475 INFO:teuthology.orchestra.run.vm08.stdout: debug osd = 20 2026-03-20T17:57:05.475 INFO:teuthology.orchestra.run.vm08.stdout: osd mclock iops capacity threshold hdd = 49000 2026-03-20T17:57:05.475 INFO:teuthology.orchestra.run.vm08.stdout: rocksdb delete range threshold = 0 2026-03-20T17:57:05.475 INFO:teuthology.orchestra.run.vm08.stdout: 2026-03-20T17:57:05.475 INFO:teuthology.orchestra.run.vm08.stdout:[mgr] 2026-03-20T17:57:05.475 INFO:teuthology.orchestra.run.vm08.stdout: debug ms = 1 2026-03-20T17:57:05.475 INFO:teuthology.orchestra.run.vm08.stdout: debug mgr = 20 2026-03-20T17:57:05.475 INFO:teuthology.orchestra.run.vm08.stdout: debug mon = 20 2026-03-20T17:57:05.475 INFO:teuthology.orchestra.run.vm08.stdout: debug auth = 20 2026-03-20T17:57:05.475 INFO:teuthology.orchestra.run.vm08.stdout: mon reweight min pgs per osd = 4 2026-03-20T17:57:05.475 INFO:teuthology.orchestra.run.vm08.stdout: mon reweight min bytes per osd = 10 2026-03-20T17:57:05.475 INFO:teuthology.orchestra.run.vm08.stdout: mgr/telemetry/nag = false 2026-03-20T17:57:05.475 INFO:teuthology.orchestra.run.vm08.stdout: 2026-03-20T17:57:05.475 INFO:teuthology.orchestra.run.vm08.stdout:[mon] 2026-03-20T17:57:05.475 INFO:teuthology.orchestra.run.vm08.stdout: debug ms = 1 2026-03-20T17:57:05.475 INFO:teuthology.orchestra.run.vm08.stdout: debug mon = 20 2026-03-20T17:57:05.475 INFO:teuthology.orchestra.run.vm08.stdout: debug paxos = 20 2026-03-20T17:57:05.475 INFO:teuthology.orchestra.run.vm08.stdout: debug auth = 20 2026-03-20T17:57:05.475 INFO:teuthology.orchestra.run.vm08.stdout: mon data avail warn = 5 2026-03-20T17:57:05.476 INFO:teuthology.orchestra.run.vm08.stdout: mon mgr mkfs grace = 240 2026-03-20T17:57:05.476 INFO:teuthology.orchestra.run.vm08.stdout: mon reweight min pgs per osd = 4 2026-03-20T17:57:05.476 INFO:teuthology.orchestra.run.vm08.stdout: mon osd reporter subtree level = osd 2026-03-20T17:57:05.476 INFO:teuthology.orchestra.run.vm08.stdout: mon osd prime pg temp = true 2026-03-20T17:57:05.476 INFO:teuthology.orchestra.run.vm08.stdout: mon reweight min bytes per osd = 10 2026-03-20T17:57:05.476 INFO:teuthology.orchestra.run.vm08.stdout: 2026-03-20T17:57:05.476 INFO:teuthology.orchestra.run.vm08.stdout: # rotate auth tickets quickly to exercise renewal paths 2026-03-20T17:57:05.476 INFO:teuthology.orchestra.run.vm08.stdout: auth mon ticket ttl = 660 # 11m 2026-03-20T17:57:05.476 INFO:teuthology.orchestra.run.vm08.stdout: auth service ticket ttl = 240 # 4m 2026-03-20T17:57:05.476 INFO:teuthology.orchestra.run.vm08.stdout: 2026-03-20T17:57:05.476 INFO:teuthology.orchestra.run.vm08.stdout: # don't complain about insecure global_id in the test suite 2026-03-20T17:57:05.476 INFO:teuthology.orchestra.run.vm08.stdout: mon_warn_on_insecure_global_id_reclaim = false 2026-03-20T17:57:05.476 INFO:teuthology.orchestra.run.vm08.stdout: mon_warn_on_insecure_global_id_reclaim_allowed = false 2026-03-20T17:57:05.476 INFO:teuthology.orchestra.run.vm08.stdout: 2026-03-20T17:57:05.476 INFO:teuthology.orchestra.run.vm08.stdout: # 1m isn't quite enough 2026-03-20T17:57:05.476 INFO:teuthology.orchestra.run.vm08.stdout: mon_down_mkfs_grace = 2m 2026-03-20T17:57:05.476 INFO:teuthology.orchestra.run.vm08.stdout: 2026-03-20T17:57:05.476 INFO:teuthology.orchestra.run.vm08.stdout: mon_warn_on_filestore_osds = false 2026-03-20T17:57:05.476 INFO:teuthology.orchestra.run.vm08.stdout: 2026-03-20T17:57:05.476 INFO:teuthology.orchestra.run.vm08.stdout:[client] 2026-03-20T17:57:05.476 INFO:teuthology.orchestra.run.vm08.stdout: rgw cache enabled = true 2026-03-20T17:57:05.476 INFO:teuthology.orchestra.run.vm08.stdout: rgw enable ops log = true 2026-03-20T17:57:05.476 INFO:teuthology.orchestra.run.vm08.stdout: rgw enable usage log = true 2026-03-20T17:57:05.476 INFO:teuthology.orchestra.run.vm08.stdout: log file = /var/log/ceph/$cluster-$name.$pid.log 2026-03-20T17:57:05.476 INFO:teuthology.orchestra.run.vm08.stdout: admin socket = /var/run/ceph/$cluster-$name.$pid.asok 2026-03-20T17:57:05.476 INFO:teuthology.orchestra.run.vm08.stdout: debug rgw = 20 2026-03-20T17:57:05.476 INFO:teuthology.orchestra.run.vm08.stdout: debug rgw lifecycle = 20 2026-03-20T17:57:05.476 INFO:teuthology.orchestra.run.vm08.stdout: debug rgw notification = 20 2026-03-20T17:57:05.476 INFO:teuthology.orchestra.run.vm08.stdout: debug rgw sync = 20 2026-03-20T17:57:05.476 INFO:teuthology.orchestra.run.vm08.stdout: rgw crypt require ssl = False 2026-03-20T17:57:05.476 INFO:teuthology.orchestra.run.vm08.stdout: rgw crypt s3 kms backend = testing 2026-03-20T17:57:05.476 INFO:teuthology.orchestra.run.vm08.stdout: rgw crypt s3 kms encryption keys = testkey-1=YmluCmJvb3N0CmJvb3N0LWJ1aWxkCmNlcGguY29uZgo= 2026-03-20T17:57:05.476 INFO:teuthology.orchestra.run.vm08.stdout: rgw curl low speed time = 300 2026-03-20T17:57:05.476 INFO:teuthology.orchestra.run.vm08.stdout: rgw data log num shards = 4 2026-03-20T17:57:05.476 INFO:teuthology.orchestra.run.vm08.stdout: rgw data sync poll interval = 5 2026-03-20T17:57:05.476 INFO:teuthology.orchestra.run.vm08.stdout: rgw md log max shards = 4 2026-03-20T17:57:05.476 INFO:teuthology.orchestra.run.vm08.stdout: rgw meta sync poll interval = 5 2026-03-20T17:57:05.476 INFO:teuthology.orchestra.run.vm08.stdout: rgw s3 auth use sts = True 2026-03-20T17:57:05.476 INFO:teuthology.orchestra.run.vm08.stdout: rgw sts key = abcdefghijklmnoq 2026-03-20T17:57:05.476 INFO:teuthology.orchestra.run.vm08.stdout: rgw sync data inject err probability = 0 2026-03-20T17:57:05.476 INFO:teuthology.orchestra.run.vm08.stdout: rgw sync log trim interval = 0 2026-03-20T17:57:05.476 INFO:teuthology.orchestra.run.vm08.stdout: rgw sync meta inject err probability = 0 2026-03-20T17:57:05.476 INFO:teuthology.orchestra.run.vm08.stdout: rgw sync obj etag verify = True 2026-03-20T17:57:05.476 INFO:teuthology.orchestra.run.vm08.stdout: setgroup = ceph 2026-03-20T17:57:05.476 INFO:teuthology.orchestra.run.vm08.stdout: setuser = ceph 2026-03-20T17:57:05.476 INFO:teuthology.orchestra.run.vm08.stdout:[mon.a] 2026-03-20T17:57:05.476 INFO:teuthology.orchestra.run.vm08.stdout:[client.0] 2026-03-20T17:57:05.476 INFO:teuthology.orchestra.run.vm08.stdout: rgw data notify interval msec = 0 2026-03-20T17:57:05.476 INFO:teuthology.orchestra.run.vm08.stdout:[osd.0] 2026-03-20T17:57:05.476 INFO:teuthology.orchestra.run.vm08.stdout: osd_max_omap_entries_per_request = 10 2026-03-20T17:57:05.476 INFO:teuthology.orchestra.run.vm08.stdout:[osd.1] 2026-03-20T17:57:05.476 INFO:teuthology.orchestra.run.vm08.stdout: osd_max_omap_entries_per_request = 1000 2026-03-20T17:57:05.476 INFO:teuthology.orchestra.run.vm08.stdout:[osd.2] 2026-03-20T17:57:05.476 INFO:teuthology.orchestra.run.vm08.stdout: osd_max_omap_entries_per_request = 10000 2026-03-20T17:57:05.478 INFO:tasks.ceph:Creating admin key on c1.mon.a... 2026-03-20T17:57:05.478 DEBUG:teuthology.orchestra.run.vm02:> sudo adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage ceph-authtool --gen-key --name=client.admin --cap mon 'allow *' --cap osd 'allow *' --cap mds 'allow *' --cap mgr 'allow *' /etc/ceph/c1.keyring 2026-03-20T17:57:05.536 INFO:tasks.ceph:Copying monmap to all nodes... 2026-03-20T17:57:05.536 DEBUG:teuthology.orchestra.run.vm02:> set -ex 2026-03-20T17:57:05.537 DEBUG:teuthology.orchestra.run.vm02:> dd if=/etc/ceph/c1.keyring of=/dev/stdout 2026-03-20T17:57:05.584 DEBUG:teuthology.orchestra.run.vm02:> set -ex 2026-03-20T17:57:05.584 DEBUG:teuthology.orchestra.run.vm02:> dd if=/home/ubuntu/cephtest/c1.monmap of=/dev/stdout 2026-03-20T17:57:05.628 INFO:tasks.ceph:Sending monmap to node ubuntu@vm02.local 2026-03-20T17:57:05.628 DEBUG:teuthology.orchestra.run.vm02:> set -ex 2026-03-20T17:57:05.628 DEBUG:teuthology.orchestra.run.vm02:> sudo dd of=/etc/ceph/c1.keyring 2026-03-20T17:57:05.628 DEBUG:teuthology.orchestra.run.vm02:> sudo chmod 0644 /etc/ceph/c1.keyring 2026-03-20T17:57:05.681 DEBUG:teuthology.orchestra.run.vm02:> set -ex 2026-03-20T17:57:05.681 DEBUG:teuthology.orchestra.run.vm02:> dd of=/home/ubuntu/cephtest/c1.monmap 2026-03-20T17:57:05.724 INFO:tasks.ceph:Sending monmap to node ubuntu@vm05.local 2026-03-20T17:57:05.724 DEBUG:teuthology.orchestra.run.vm05:> set -ex 2026-03-20T17:57:05.724 DEBUG:teuthology.orchestra.run.vm05:> sudo dd of=/etc/ceph/c1.keyring 2026-03-20T17:57:05.724 DEBUG:teuthology.orchestra.run.vm05:> sudo chmod 0644 /etc/ceph/c1.keyring 2026-03-20T17:57:05.738 DEBUG:teuthology.orchestra.run.vm05:> set -ex 2026-03-20T17:57:05.738 DEBUG:teuthology.orchestra.run.vm05:> dd of=/home/ubuntu/cephtest/c1.monmap 2026-03-20T17:57:05.782 INFO:tasks.ceph:Sending monmap to node ubuntu@vm06.local 2026-03-20T17:57:05.782 DEBUG:teuthology.orchestra.run.vm06:> set -ex 2026-03-20T17:57:05.782 DEBUG:teuthology.orchestra.run.vm06:> sudo dd of=/etc/ceph/c1.keyring 2026-03-20T17:57:05.782 DEBUG:teuthology.orchestra.run.vm06:> sudo chmod 0644 /etc/ceph/c1.keyring 2026-03-20T17:57:05.796 DEBUG:teuthology.orchestra.run.vm06:> set -ex 2026-03-20T17:57:05.796 DEBUG:teuthology.orchestra.run.vm06:> dd of=/home/ubuntu/cephtest/c1.monmap 2026-03-20T17:57:05.843 INFO:tasks.ceph:Sending monmap to node ubuntu@vm08.local 2026-03-20T17:57:05.843 DEBUG:teuthology.orchestra.run.vm08:> set -ex 2026-03-20T17:57:05.843 DEBUG:teuthology.orchestra.run.vm08:> sudo dd of=/etc/ceph/c1.keyring 2026-03-20T17:57:05.843 DEBUG:teuthology.orchestra.run.vm08:> sudo chmod 0644 /etc/ceph/c1.keyring 2026-03-20T17:57:05.855 DEBUG:teuthology.orchestra.run.vm08:> set -ex 2026-03-20T17:57:05.856 DEBUG:teuthology.orchestra.run.vm08:> dd of=/home/ubuntu/cephtest/c1.monmap 2026-03-20T17:57:05.901 INFO:tasks.ceph:Setting up mon nodes... 2026-03-20T17:57:05.901 INFO:tasks.ceph:Setting up mgr nodes... 2026-03-20T17:57:05.901 DEBUG:teuthology.orchestra.run.vm05:> sudo mkdir -p /var/lib/ceph/mgr/c1-x && sudo adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage ceph-authtool --create-keyring --gen-key --name=mgr.x /var/lib/ceph/mgr/c1-x/keyring 2026-03-20T17:57:05.923 INFO:teuthology.orchestra.run.vm05.stdout:creating /var/lib/ceph/mgr/c1-x/keyring 2026-03-20T17:57:05.925 INFO:tasks.ceph:Setting up mds nodes... 2026-03-20T17:57:05.925 INFO:tasks.ceph_client:Setting up client nodes... 2026-03-20T17:57:05.925 DEBUG:teuthology.orchestra.run.vm02:> sudo adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage ceph-authtool --create-keyring --gen-key --name=client.0 /etc/ceph/c1.client.0.keyring && sudo chmod 0644 /etc/ceph/c1.client.0.keyring 2026-03-20T17:57:05.942 INFO:teuthology.orchestra.run.vm02.stdout:creating /etc/ceph/c1.client.0.keyring 2026-03-20T17:57:05.948 DEBUG:teuthology.orchestra.run.vm05:> sudo adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage ceph-authtool --create-keyring --gen-key --name=client.1 /etc/ceph/c1.client.1.keyring && sudo chmod 0644 /etc/ceph/c1.client.1.keyring 2026-03-20T17:57:05.985 INFO:teuthology.orchestra.run.vm05.stdout:creating /etc/ceph/c1.client.1.keyring 2026-03-20T17:57:05.993 INFO:tasks.ceph:Running mkfs on osd nodes... 2026-03-20T17:57:05.993 INFO:tasks.ceph:ctx.disk_config.remote_to_roles_to_dev: {Remote(name='ubuntu@vm02.local'): {'c1.osd.0': '/dev/vg_nvme/lv_1', 'c1.osd.1': '/dev/vg_nvme/lv_2', 'c1.osd.2': '/dev/vg_nvme/lv_3'}, Remote(name='ubuntu@vm05.local'): {'c1.osd.3': '/dev/vg_nvme/lv_1', 'c1.osd.4': '/dev/vg_nvme/lv_2', 'c1.osd.5': '/dev/vg_nvme/lv_3'}} 2026-03-20T17:57:05.993 DEBUG:teuthology.orchestra.run.vm02:> sudo mkdir -p /var/lib/ceph/osd/c1-0 2026-03-20T17:57:06.000 INFO:tasks.ceph:roles_to_devs: {'c1.osd.0': '/dev/vg_nvme/lv_1', 'c1.osd.1': '/dev/vg_nvme/lv_2', 'c1.osd.2': '/dev/vg_nvme/lv_3'} 2026-03-20T17:57:06.000 INFO:tasks.ceph:role: c1.osd.0 2026-03-20T17:57:06.000 INFO:tasks.ceph:['mkfs.xfs', '-f', '-i', 'size=2048'] on /dev/vg_nvme/lv_1 on ubuntu@vm02.local 2026-03-20T17:57:06.000 DEBUG:teuthology.orchestra.run.vm02:> yes | sudo mkfs.xfs -f -i size=2048 /dev/vg_nvme/lv_1 2026-03-20T17:57:06.050 INFO:teuthology.orchestra.run.vm02.stdout:meta-data=/dev/vg_nvme/lv_1 isize=2048 agcount=4, agsize=1310464 blks 2026-03-20T17:57:06.050 INFO:teuthology.orchestra.run.vm02.stdout: = sectsz=512 attr=2, projid32bit=1 2026-03-20T17:57:06.050 INFO:teuthology.orchestra.run.vm02.stdout: = crc=1 finobt=1, sparse=1, rmapbt=0 2026-03-20T17:57:06.050 INFO:teuthology.orchestra.run.vm02.stdout: = reflink=1 bigtime=0 inobtcount=0 2026-03-20T17:57:06.050 INFO:teuthology.orchestra.run.vm02.stdout:data = bsize=4096 blocks=5241856, imaxpct=25 2026-03-20T17:57:06.050 INFO:teuthology.orchestra.run.vm02.stdout: = sunit=0 swidth=0 blks 2026-03-20T17:57:06.050 INFO:teuthology.orchestra.run.vm02.stdout:naming =version 2 bsize=4096 ascii-ci=0, ftype=1 2026-03-20T17:57:06.050 INFO:teuthology.orchestra.run.vm02.stdout:log =internal log bsize=4096 blocks=2560, version=2 2026-03-20T17:57:06.050 INFO:teuthology.orchestra.run.vm02.stdout: = sectsz=512 sunit=0 blks, lazy-count=1 2026-03-20T17:57:06.050 INFO:teuthology.orchestra.run.vm02.stdout:realtime =none extsz=4096 blocks=0, rtextents=0 2026-03-20T17:57:06.054 INFO:teuthology.orchestra.run.vm02.stdout:Discarding blocks...Done. 2026-03-20T17:57:06.059 INFO:tasks.ceph:mount /dev/vg_nvme/lv_1 on ubuntu@vm02.local -o noatime 2026-03-20T17:57:06.059 DEBUG:teuthology.orchestra.run.vm02:> sudo mount -t xfs -o noatime /dev/vg_nvme/lv_1 /var/lib/ceph/osd/c1-0 2026-03-20T17:57:06.145 DEBUG:teuthology.orchestra.run.vm02:> sudo /sbin/restorecon /var/lib/ceph/osd/c1-0 2026-03-20T17:57:06.191 INFO:teuthology.orchestra.run.vm02.stderr:sudo: /sbin/restorecon: command not found 2026-03-20T17:57:06.192 DEBUG:teuthology.orchestra.run:got remote process result: 1 2026-03-20T17:57:06.192 DEBUG:teuthology.orchestra.run.vm02:> sudo mkdir -p /var/lib/ceph/osd/c1-1 2026-03-20T17:57:06.240 INFO:tasks.ceph:roles_to_devs: {'c1.osd.0': '/dev/vg_nvme/lv_1', 'c1.osd.1': '/dev/vg_nvme/lv_2', 'c1.osd.2': '/dev/vg_nvme/lv_3'} 2026-03-20T17:57:06.240 INFO:tasks.ceph:role: c1.osd.1 2026-03-20T17:57:06.240 INFO:tasks.ceph:['mkfs.xfs', '-f', '-i', 'size=2048'] on /dev/vg_nvme/lv_2 on ubuntu@vm02.local 2026-03-20T17:57:06.240 DEBUG:teuthology.orchestra.run.vm02:> yes | sudo mkfs.xfs -f -i size=2048 /dev/vg_nvme/lv_2 2026-03-20T17:57:06.288 INFO:teuthology.orchestra.run.vm02.stdout:meta-data=/dev/vg_nvme/lv_2 isize=2048 agcount=4, agsize=1310464 blks 2026-03-20T17:57:06.288 INFO:teuthology.orchestra.run.vm02.stdout: = sectsz=512 attr=2, projid32bit=1 2026-03-20T17:57:06.288 INFO:teuthology.orchestra.run.vm02.stdout: = crc=1 finobt=1, sparse=1, rmapbt=0 2026-03-20T17:57:06.288 INFO:teuthology.orchestra.run.vm02.stdout: = reflink=1 bigtime=0 inobtcount=0 2026-03-20T17:57:06.288 INFO:teuthology.orchestra.run.vm02.stdout:data = bsize=4096 blocks=5241856, imaxpct=25 2026-03-20T17:57:06.288 INFO:teuthology.orchestra.run.vm02.stdout: = sunit=0 swidth=0 blks 2026-03-20T17:57:06.288 INFO:teuthology.orchestra.run.vm02.stdout:naming =version 2 bsize=4096 ascii-ci=0, ftype=1 2026-03-20T17:57:06.288 INFO:teuthology.orchestra.run.vm02.stdout:log =internal log bsize=4096 blocks=2560, version=2 2026-03-20T17:57:06.288 INFO:teuthology.orchestra.run.vm02.stdout: = sectsz=512 sunit=0 blks, lazy-count=1 2026-03-20T17:57:06.288 INFO:teuthology.orchestra.run.vm02.stdout:realtime =none extsz=4096 blocks=0, rtextents=0 2026-03-20T17:57:06.293 INFO:teuthology.orchestra.run.vm02.stdout:Discarding blocks...Done. 2026-03-20T17:57:06.294 INFO:tasks.ceph:mount /dev/vg_nvme/lv_2 on ubuntu@vm02.local -o noatime 2026-03-20T17:57:06.294 DEBUG:teuthology.orchestra.run.vm02:> sudo mount -t xfs -o noatime /dev/vg_nvme/lv_2 /var/lib/ceph/osd/c1-1 2026-03-20T17:57:06.348 DEBUG:teuthology.orchestra.run.vm02:> sudo /sbin/restorecon /var/lib/ceph/osd/c1-1 2026-03-20T17:57:06.395 INFO:teuthology.orchestra.run.vm02.stderr:sudo: /sbin/restorecon: command not found 2026-03-20T17:57:06.395 DEBUG:teuthology.orchestra.run:got remote process result: 1 2026-03-20T17:57:06.395 DEBUG:teuthology.orchestra.run.vm02:> sudo mkdir -p /var/lib/ceph/osd/c1-2 2026-03-20T17:57:06.445 INFO:tasks.ceph:roles_to_devs: {'c1.osd.0': '/dev/vg_nvme/lv_1', 'c1.osd.1': '/dev/vg_nvme/lv_2', 'c1.osd.2': '/dev/vg_nvme/lv_3'} 2026-03-20T17:57:06.445 INFO:tasks.ceph:role: c1.osd.2 2026-03-20T17:57:06.445 INFO:tasks.ceph:['mkfs.xfs', '-f', '-i', 'size=2048'] on /dev/vg_nvme/lv_3 on ubuntu@vm02.local 2026-03-20T17:57:06.445 DEBUG:teuthology.orchestra.run.vm02:> yes | sudo mkfs.xfs -f -i size=2048 /dev/vg_nvme/lv_3 2026-03-20T17:57:06.492 INFO:teuthology.orchestra.run.vm02.stdout:meta-data=/dev/vg_nvme/lv_3 isize=2048 agcount=4, agsize=1310464 blks 2026-03-20T17:57:06.492 INFO:teuthology.orchestra.run.vm02.stdout: = sectsz=512 attr=2, projid32bit=1 2026-03-20T17:57:06.492 INFO:teuthology.orchestra.run.vm02.stdout: = crc=1 finobt=1, sparse=1, rmapbt=0 2026-03-20T17:57:06.492 INFO:teuthology.orchestra.run.vm02.stdout: = reflink=1 bigtime=0 inobtcount=0 2026-03-20T17:57:06.492 INFO:teuthology.orchestra.run.vm02.stdout:data = bsize=4096 blocks=5241856, imaxpct=25 2026-03-20T17:57:06.492 INFO:teuthology.orchestra.run.vm02.stdout: = sunit=0 swidth=0 blks 2026-03-20T17:57:06.492 INFO:teuthology.orchestra.run.vm02.stdout:naming =version 2 bsize=4096 ascii-ci=0, ftype=1 2026-03-20T17:57:06.492 INFO:teuthology.orchestra.run.vm02.stdout:log =internal log bsize=4096 blocks=2560, version=2 2026-03-20T17:57:06.492 INFO:teuthology.orchestra.run.vm02.stdout: = sectsz=512 sunit=0 blks, lazy-count=1 2026-03-20T17:57:06.492 INFO:teuthology.orchestra.run.vm02.stdout:realtime =none extsz=4096 blocks=0, rtextents=0 2026-03-20T17:57:06.497 INFO:teuthology.orchestra.run.vm02.stdout:Discarding blocks...Done. 2026-03-20T17:57:06.498 INFO:tasks.ceph:mount /dev/vg_nvme/lv_3 on ubuntu@vm02.local -o noatime 2026-03-20T17:57:06.498 DEBUG:teuthology.orchestra.run.vm02:> sudo mount -t xfs -o noatime /dev/vg_nvme/lv_3 /var/lib/ceph/osd/c1-2 2026-03-20T17:57:06.555 DEBUG:teuthology.orchestra.run.vm02:> sudo /sbin/restorecon /var/lib/ceph/osd/c1-2 2026-03-20T17:57:06.604 INFO:teuthology.orchestra.run.vm02.stderr:sudo: /sbin/restorecon: command not found 2026-03-20T17:57:06.605 DEBUG:teuthology.orchestra.run:got remote process result: 1 2026-03-20T17:57:06.605 DEBUG:teuthology.orchestra.run.vm02:> sudo MALLOC_CHECK_=3 adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage ceph-osd --no-mon-config --cluster c1 --mkfs --mkkey -i 0 --monmap /home/ubuntu/cephtest/c1.monmap 2026-03-20T17:57:06.666 INFO:teuthology.orchestra.run.vm02.stderr:2026-03-20T17:57:06.665+0000 7fd735f5da40 -1 auth: error reading file: /var/lib/ceph/osd/c1-0/keyring: can't open /var/lib/ceph/osd/c1-0/keyring: (2) No such file or directory 2026-03-20T17:57:06.667 INFO:teuthology.orchestra.run.vm02.stderr:2026-03-20T17:57:06.665+0000 7fd735f5da40 -1 created new key in keyring /var/lib/ceph/osd/c1-0/keyring 2026-03-20T17:57:06.667 INFO:teuthology.orchestra.run.vm02.stderr:2026-03-20T17:57:06.665+0000 7fd735f5da40 -1 bdev(0x562f8acef800 /var/lib/ceph/osd/c1-0/block) open stat got: (1) Operation not permitted 2026-03-20T17:57:06.667 INFO:teuthology.orchestra.run.vm02.stderr:2026-03-20T17:57:06.665+0000 7fd735f5da40 -1 bluestore(/var/lib/ceph/osd/c1-0) _read_fsid unparsable uuid 2026-03-20T17:57:07.290 DEBUG:teuthology.orchestra.run.vm02:> sudo chown -R ceph:ceph /var/lib/ceph/osd/c1-0 2026-03-20T17:57:07.342 DEBUG:teuthology.orchestra.run.vm02:> sudo MALLOC_CHECK_=3 adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage ceph-osd --no-mon-config --cluster c1 --mkfs --mkkey -i 1 --monmap /home/ubuntu/cephtest/c1.monmap 2026-03-20T17:57:07.406 INFO:teuthology.orchestra.run.vm02.stderr:2026-03-20T17:57:07.401+0000 7fac81330a40 -1 auth: error reading file: /var/lib/ceph/osd/c1-1/keyring: can't open /var/lib/ceph/osd/c1-1/keyring: (2) No such file or directory 2026-03-20T17:57:07.406 INFO:teuthology.orchestra.run.vm02.stderr:2026-03-20T17:57:07.405+0000 7fac81330a40 -1 created new key in keyring /var/lib/ceph/osd/c1-1/keyring 2026-03-20T17:57:07.406 INFO:teuthology.orchestra.run.vm02.stderr:2026-03-20T17:57:07.405+0000 7fac81330a40 -1 bdev(0x5616c3517800 /var/lib/ceph/osd/c1-1/block) open stat got: (1) Operation not permitted 2026-03-20T17:57:07.406 INFO:teuthology.orchestra.run.vm02.stderr:2026-03-20T17:57:07.405+0000 7fac81330a40 -1 bluestore(/var/lib/ceph/osd/c1-1) _read_fsid unparsable uuid 2026-03-20T17:57:08.016 DEBUG:teuthology.orchestra.run.vm02:> sudo chown -R ceph:ceph /var/lib/ceph/osd/c1-1 2026-03-20T17:57:08.066 DEBUG:teuthology.orchestra.run.vm02:> sudo MALLOC_CHECK_=3 adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage ceph-osd --no-mon-config --cluster c1 --mkfs --mkkey -i 2 --monmap /home/ubuntu/cephtest/c1.monmap 2026-03-20T17:57:08.127 INFO:teuthology.orchestra.run.vm02.stderr:2026-03-20T17:57:08.125+0000 7fde46450a40 -1 auth: error reading file: /var/lib/ceph/osd/c1-2/keyring: can't open /var/lib/ceph/osd/c1-2/keyring: (2) No such file or directory 2026-03-20T17:57:08.127 INFO:teuthology.orchestra.run.vm02.stderr:2026-03-20T17:57:08.125+0000 7fde46450a40 -1 created new key in keyring /var/lib/ceph/osd/c1-2/keyring 2026-03-20T17:57:08.127 INFO:teuthology.orchestra.run.vm02.stderr:2026-03-20T17:57:08.125+0000 7fde46450a40 -1 bdev(0x55e3de3b1800 /var/lib/ceph/osd/c1-2/block) open stat got: (1) Operation not permitted 2026-03-20T17:57:08.127 INFO:teuthology.orchestra.run.vm02.stderr:2026-03-20T17:57:08.125+0000 7fde46450a40 -1 bluestore(/var/lib/ceph/osd/c1-2) _read_fsid unparsable uuid 2026-03-20T17:57:08.756 DEBUG:teuthology.orchestra.run.vm02:> sudo chown -R ceph:ceph /var/lib/ceph/osd/c1-2 2026-03-20T17:57:08.805 DEBUG:teuthology.orchestra.run.vm05:> sudo mkdir -p /var/lib/ceph/osd/c1-3 2026-03-20T17:57:08.813 INFO:tasks.ceph:roles_to_devs: {'c1.osd.3': '/dev/vg_nvme/lv_1', 'c1.osd.4': '/dev/vg_nvme/lv_2', 'c1.osd.5': '/dev/vg_nvme/lv_3'} 2026-03-20T17:57:08.813 INFO:tasks.ceph:role: c1.osd.3 2026-03-20T17:57:08.813 INFO:tasks.ceph:['mkfs.xfs', '-f', '-i', 'size=2048'] on /dev/vg_nvme/lv_1 on ubuntu@vm05.local 2026-03-20T17:57:08.813 DEBUG:teuthology.orchestra.run.vm05:> yes | sudo mkfs.xfs -f -i size=2048 /dev/vg_nvme/lv_1 2026-03-20T17:57:08.863 INFO:teuthology.orchestra.run.vm05.stdout:meta-data=/dev/vg_nvme/lv_1 isize=2048 agcount=4, agsize=1310464 blks 2026-03-20T17:57:08.863 INFO:teuthology.orchestra.run.vm05.stdout: = sectsz=512 attr=2, projid32bit=1 2026-03-20T17:57:08.864 INFO:teuthology.orchestra.run.vm05.stdout: = crc=1 finobt=1, sparse=1, rmapbt=0 2026-03-20T17:57:08.864 INFO:teuthology.orchestra.run.vm05.stdout: = reflink=1 bigtime=0 inobtcount=0 2026-03-20T17:57:08.864 INFO:teuthology.orchestra.run.vm05.stdout:data = bsize=4096 blocks=5241856, imaxpct=25 2026-03-20T17:57:08.864 INFO:teuthology.orchestra.run.vm05.stdout: = sunit=0 swidth=0 blks 2026-03-20T17:57:08.864 INFO:teuthology.orchestra.run.vm05.stdout:naming =version 2 bsize=4096 ascii-ci=0, ftype=1 2026-03-20T17:57:08.864 INFO:teuthology.orchestra.run.vm05.stdout:log =internal log bsize=4096 blocks=2560, version=2 2026-03-20T17:57:08.864 INFO:teuthology.orchestra.run.vm05.stdout: = sectsz=512 sunit=0 blks, lazy-count=1 2026-03-20T17:57:08.864 INFO:teuthology.orchestra.run.vm05.stdout:realtime =none extsz=4096 blocks=0, rtextents=0 2026-03-20T17:57:08.872 INFO:teuthology.orchestra.run.vm05.stdout:Discarding blocks...Done. 2026-03-20T17:57:08.873 INFO:tasks.ceph:mount /dev/vg_nvme/lv_1 on ubuntu@vm05.local -o noatime 2026-03-20T17:57:08.873 DEBUG:teuthology.orchestra.run.vm05:> sudo mount -t xfs -o noatime /dev/vg_nvme/lv_1 /var/lib/ceph/osd/c1-3 2026-03-20T17:57:08.964 DEBUG:teuthology.orchestra.run.vm05:> sudo /sbin/restorecon /var/lib/ceph/osd/c1-3 2026-03-20T17:57:09.018 INFO:teuthology.orchestra.run.vm05.stderr:sudo: /sbin/restorecon: command not found 2026-03-20T17:57:09.018 DEBUG:teuthology.orchestra.run:got remote process result: 1 2026-03-20T17:57:09.019 DEBUG:teuthology.orchestra.run.vm05:> sudo mkdir -p /var/lib/ceph/osd/c1-4 2026-03-20T17:57:09.068 INFO:tasks.ceph:roles_to_devs: {'c1.osd.3': '/dev/vg_nvme/lv_1', 'c1.osd.4': '/dev/vg_nvme/lv_2', 'c1.osd.5': '/dev/vg_nvme/lv_3'} 2026-03-20T17:57:09.068 INFO:tasks.ceph:role: c1.osd.4 2026-03-20T17:57:09.068 INFO:tasks.ceph:['mkfs.xfs', '-f', '-i', 'size=2048'] on /dev/vg_nvme/lv_2 on ubuntu@vm05.local 2026-03-20T17:57:09.068 DEBUG:teuthology.orchestra.run.vm05:> yes | sudo mkfs.xfs -f -i size=2048 /dev/vg_nvme/lv_2 2026-03-20T17:57:09.120 INFO:teuthology.orchestra.run.vm05.stdout:meta-data=/dev/vg_nvme/lv_2 isize=2048 agcount=4, agsize=1310464 blks 2026-03-20T17:57:09.120 INFO:teuthology.orchestra.run.vm05.stdout: = sectsz=512 attr=2, projid32bit=1 2026-03-20T17:57:09.120 INFO:teuthology.orchestra.run.vm05.stdout: = crc=1 finobt=1, sparse=1, rmapbt=0 2026-03-20T17:57:09.120 INFO:teuthology.orchestra.run.vm05.stdout: = reflink=1 bigtime=0 inobtcount=0 2026-03-20T17:57:09.120 INFO:teuthology.orchestra.run.vm05.stdout:data = bsize=4096 blocks=5241856, imaxpct=25 2026-03-20T17:57:09.120 INFO:teuthology.orchestra.run.vm05.stdout: = sunit=0 swidth=0 blks 2026-03-20T17:57:09.120 INFO:teuthology.orchestra.run.vm05.stdout:naming =version 2 bsize=4096 ascii-ci=0, ftype=1 2026-03-20T17:57:09.120 INFO:teuthology.orchestra.run.vm05.stdout:log =internal log bsize=4096 blocks=2560, version=2 2026-03-20T17:57:09.120 INFO:teuthology.orchestra.run.vm05.stdout: = sectsz=512 sunit=0 blks, lazy-count=1 2026-03-20T17:57:09.120 INFO:teuthology.orchestra.run.vm05.stdout:realtime =none extsz=4096 blocks=0, rtextents=0 2026-03-20T17:57:09.124 INFO:teuthology.orchestra.run.vm05.stdout:Discarding blocks...Done. 2026-03-20T17:57:09.126 INFO:tasks.ceph:mount /dev/vg_nvme/lv_2 on ubuntu@vm05.local -o noatime 2026-03-20T17:57:09.126 DEBUG:teuthology.orchestra.run.vm05:> sudo mount -t xfs -o noatime /dev/vg_nvme/lv_2 /var/lib/ceph/osd/c1-4 2026-03-20T17:57:09.183 DEBUG:teuthology.orchestra.run.vm05:> sudo /sbin/restorecon /var/lib/ceph/osd/c1-4 2026-03-20T17:57:09.230 INFO:teuthology.orchestra.run.vm05.stderr:sudo: /sbin/restorecon: command not found 2026-03-20T17:57:09.230 DEBUG:teuthology.orchestra.run:got remote process result: 1 2026-03-20T17:57:09.230 DEBUG:teuthology.orchestra.run.vm05:> sudo mkdir -p /var/lib/ceph/osd/c1-5 2026-03-20T17:57:09.280 INFO:tasks.ceph:roles_to_devs: {'c1.osd.3': '/dev/vg_nvme/lv_1', 'c1.osd.4': '/dev/vg_nvme/lv_2', 'c1.osd.5': '/dev/vg_nvme/lv_3'} 2026-03-20T17:57:09.280 INFO:tasks.ceph:role: c1.osd.5 2026-03-20T17:57:09.280 INFO:tasks.ceph:['mkfs.xfs', '-f', '-i', 'size=2048'] on /dev/vg_nvme/lv_3 on ubuntu@vm05.local 2026-03-20T17:57:09.280 DEBUG:teuthology.orchestra.run.vm05:> yes | sudo mkfs.xfs -f -i size=2048 /dev/vg_nvme/lv_3 2026-03-20T17:57:09.331 INFO:teuthology.orchestra.run.vm05.stdout:meta-data=/dev/vg_nvme/lv_3 isize=2048 agcount=4, agsize=1310464 blks 2026-03-20T17:57:09.331 INFO:teuthology.orchestra.run.vm05.stdout: = sectsz=512 attr=2, projid32bit=1 2026-03-20T17:57:09.331 INFO:teuthology.orchestra.run.vm05.stdout: = crc=1 finobt=1, sparse=1, rmapbt=0 2026-03-20T17:57:09.331 INFO:teuthology.orchestra.run.vm05.stdout: = reflink=1 bigtime=0 inobtcount=0 2026-03-20T17:57:09.331 INFO:teuthology.orchestra.run.vm05.stdout:data = bsize=4096 blocks=5241856, imaxpct=25 2026-03-20T17:57:09.331 INFO:teuthology.orchestra.run.vm05.stdout: = sunit=0 swidth=0 blks 2026-03-20T17:57:09.331 INFO:teuthology.orchestra.run.vm05.stdout:naming =version 2 bsize=4096 ascii-ci=0, ftype=1 2026-03-20T17:57:09.331 INFO:teuthology.orchestra.run.vm05.stdout:log =internal log bsize=4096 blocks=2560, version=2 2026-03-20T17:57:09.331 INFO:teuthology.orchestra.run.vm05.stdout: = sectsz=512 sunit=0 blks, lazy-count=1 2026-03-20T17:57:09.331 INFO:teuthology.orchestra.run.vm05.stdout:realtime =none extsz=4096 blocks=0, rtextents=0 2026-03-20T17:57:09.336 INFO:teuthology.orchestra.run.vm05.stdout:Discarding blocks...Done. 2026-03-20T17:57:09.337 INFO:tasks.ceph:mount /dev/vg_nvme/lv_3 on ubuntu@vm05.local -o noatime 2026-03-20T17:57:09.337 DEBUG:teuthology.orchestra.run.vm05:> sudo mount -t xfs -o noatime /dev/vg_nvme/lv_3 /var/lib/ceph/osd/c1-5 2026-03-20T17:57:09.391 DEBUG:teuthology.orchestra.run.vm05:> sudo /sbin/restorecon /var/lib/ceph/osd/c1-5 2026-03-20T17:57:09.437 INFO:teuthology.orchestra.run.vm05.stderr:sudo: /sbin/restorecon: command not found 2026-03-20T17:57:09.437 DEBUG:teuthology.orchestra.run:got remote process result: 1 2026-03-20T17:57:09.437 DEBUG:teuthology.orchestra.run.vm05:> sudo MALLOC_CHECK_=3 adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage ceph-osd --no-mon-config --cluster c1 --mkfs --mkkey -i 3 --monmap /home/ubuntu/cephtest/c1.monmap 2026-03-20T17:57:09.499 INFO:teuthology.orchestra.run.vm05.stderr:2026-03-20T17:57:09.493+0000 7f520ffa8a40 -1 auth: error reading file: /var/lib/ceph/osd/c1-3/keyring: can't open /var/lib/ceph/osd/c1-3/keyring: (2) No such file or directory 2026-03-20T17:57:09.500 INFO:teuthology.orchestra.run.vm05.stderr:2026-03-20T17:57:09.497+0000 7f520ffa8a40 -1 created new key in keyring /var/lib/ceph/osd/c1-3/keyring 2026-03-20T17:57:09.500 INFO:teuthology.orchestra.run.vm05.stderr:2026-03-20T17:57:09.497+0000 7f520ffa8a40 -1 bdev(0x561a38167800 /var/lib/ceph/osd/c1-3/block) open stat got: (1) Operation not permitted 2026-03-20T17:57:09.500 INFO:teuthology.orchestra.run.vm05.stderr:2026-03-20T17:57:09.497+0000 7f520ffa8a40 -1 bluestore(/var/lib/ceph/osd/c1-3) _read_fsid unparsable uuid 2026-03-20T17:57:10.118 DEBUG:teuthology.orchestra.run.vm05:> sudo chown -R ceph:ceph /var/lib/ceph/osd/c1-3 2026-03-20T17:57:10.167 DEBUG:teuthology.orchestra.run.vm05:> sudo MALLOC_CHECK_=3 adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage ceph-osd --no-mon-config --cluster c1 --mkfs --mkkey -i 4 --monmap /home/ubuntu/cephtest/c1.monmap 2026-03-20T17:57:10.227 INFO:teuthology.orchestra.run.vm05.stderr:2026-03-20T17:57:10.221+0000 7f5da2e93a40 -1 auth: error reading file: /var/lib/ceph/osd/c1-4/keyring: can't open /var/lib/ceph/osd/c1-4/keyring: (2) No such file or directory 2026-03-20T17:57:10.227 INFO:teuthology.orchestra.run.vm05.stderr:2026-03-20T17:57:10.221+0000 7f5da2e93a40 -1 created new key in keyring /var/lib/ceph/osd/c1-4/keyring 2026-03-20T17:57:10.227 INFO:teuthology.orchestra.run.vm05.stderr:2026-03-20T17:57:10.221+0000 7f5da2e93a40 -1 bdev(0x556805e27800 /var/lib/ceph/osd/c1-4/block) open stat got: (1) Operation not permitted 2026-03-20T17:57:10.227 INFO:teuthology.orchestra.run.vm05.stderr:2026-03-20T17:57:10.221+0000 7f5da2e93a40 -1 bluestore(/var/lib/ceph/osd/c1-4) _read_fsid unparsable uuid 2026-03-20T17:57:10.838 DEBUG:teuthology.orchestra.run.vm05:> sudo chown -R ceph:ceph /var/lib/ceph/osd/c1-4 2026-03-20T17:57:10.887 DEBUG:teuthology.orchestra.run.vm05:> sudo MALLOC_CHECK_=3 adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage ceph-osd --no-mon-config --cluster c1 --mkfs --mkkey -i 5 --monmap /home/ubuntu/cephtest/c1.monmap 2026-03-20T17:57:10.947 INFO:teuthology.orchestra.run.vm05.stderr:2026-03-20T17:57:10.941+0000 7fb3cc610a40 -1 auth: error reading file: /var/lib/ceph/osd/c1-5/keyring: can't open /var/lib/ceph/osd/c1-5/keyring: (2) No such file or directory 2026-03-20T17:57:10.948 INFO:teuthology.orchestra.run.vm05.stderr:2026-03-20T17:57:10.945+0000 7fb3cc610a40 -1 created new key in keyring /var/lib/ceph/osd/c1-5/keyring 2026-03-20T17:57:10.948 INFO:teuthology.orchestra.run.vm05.stderr:2026-03-20T17:57:10.945+0000 7fb3cc610a40 -1 bdev(0x5588e3d37800 /var/lib/ceph/osd/c1-5/block) open stat got: (1) Operation not permitted 2026-03-20T17:57:10.948 INFO:teuthology.orchestra.run.vm05.stderr:2026-03-20T17:57:10.945+0000 7fb3cc610a40 -1 bluestore(/var/lib/ceph/osd/c1-5) _read_fsid unparsable uuid 2026-03-20T17:57:11.566 DEBUG:teuthology.orchestra.run.vm05:> sudo chown -R ceph:ceph /var/lib/ceph/osd/c1-5 2026-03-20T17:57:11.615 INFO:tasks.ceph:Reading keys from all nodes... 2026-03-20T17:57:11.615 DEBUG:teuthology.orchestra.run.vm02:> set -ex 2026-03-20T17:57:11.615 DEBUG:teuthology.orchestra.run.vm02:> sudo dd if=/var/lib/ceph/osd/c1-0/keyring of=/dev/stdout 2026-03-20T17:57:11.623 DEBUG:teuthology.orchestra.run.vm02:> set -ex 2026-03-20T17:57:11.623 DEBUG:teuthology.orchestra.run.vm02:> sudo dd if=/var/lib/ceph/osd/c1-1/keyring of=/dev/stdout 2026-03-20T17:57:11.673 DEBUG:teuthology.orchestra.run.vm02:> set -ex 2026-03-20T17:57:11.673 DEBUG:teuthology.orchestra.run.vm02:> sudo dd if=/var/lib/ceph/osd/c1-2/keyring of=/dev/stdout 2026-03-20T17:57:11.721 DEBUG:teuthology.orchestra.run.vm05:> set -ex 2026-03-20T17:57:11.721 DEBUG:teuthology.orchestra.run.vm05:> sudo dd if=/var/lib/ceph/mgr/c1-x/keyring of=/dev/stdout 2026-03-20T17:57:11.729 DEBUG:teuthology.orchestra.run.vm05:> set -ex 2026-03-20T17:57:11.729 DEBUG:teuthology.orchestra.run.vm05:> sudo dd if=/var/lib/ceph/osd/c1-3/keyring of=/dev/stdout 2026-03-20T17:57:11.778 DEBUG:teuthology.orchestra.run.vm05:> set -ex 2026-03-20T17:57:11.778 DEBUG:teuthology.orchestra.run.vm05:> sudo dd if=/var/lib/ceph/osd/c1-4/keyring of=/dev/stdout 2026-03-20T17:57:11.827 DEBUG:teuthology.orchestra.run.vm05:> set -ex 2026-03-20T17:57:11.827 DEBUG:teuthology.orchestra.run.vm05:> sudo dd if=/var/lib/ceph/osd/c1-5/keyring of=/dev/stdout 2026-03-20T17:57:11.876 DEBUG:teuthology.orchestra.run.vm02:> set -ex 2026-03-20T17:57:11.876 DEBUG:teuthology.orchestra.run.vm02:> dd if=/etc/ceph/c1.client.0.keyring of=/dev/stdout 2026-03-20T17:57:11.879 DEBUG:teuthology.orchestra.run.vm05:> set -ex 2026-03-20T17:57:11.879 DEBUG:teuthology.orchestra.run.vm05:> dd if=/etc/ceph/c1.client.1.keyring of=/dev/stdout 2026-03-20T17:57:11.926 INFO:tasks.ceph:Adding keys to all mons... 2026-03-20T17:57:11.926 DEBUG:teuthology.orchestra.run.vm02:> sudo tee -a /etc/ceph/c1.keyring 2026-03-20T17:57:11.934 INFO:teuthology.orchestra.run.vm02.stdout:[osd.0] 2026-03-20T17:57:11.934 INFO:teuthology.orchestra.run.vm02.stdout: key = AQByir1p/aK8JxAAAJCydY6clmtIiUDpr5RDXg== 2026-03-20T17:57:11.934 INFO:teuthology.orchestra.run.vm02.stdout:[osd.1] 2026-03-20T17:57:11.934 INFO:teuthology.orchestra.run.vm02.stdout: key = AQBzir1prII7GBAAPmgd8M6nLQr0vTQzQFT25w== 2026-03-20T17:57:11.934 INFO:teuthology.orchestra.run.vm02.stdout:[osd.2] 2026-03-20T17:57:11.934 INFO:teuthology.orchestra.run.vm02.stdout: key = AQB0ir1pJRyUBxAAoq81BsFAf4En/MoOxl5kQg== 2026-03-20T17:57:11.934 INFO:teuthology.orchestra.run.vm02.stdout:[mgr.x] 2026-03-20T17:57:11.934 INFO:teuthology.orchestra.run.vm02.stdout: key = AQBxir1pZo8TNxAAi0rvM7zJGAa+iThoYOp3Zw== 2026-03-20T17:57:11.934 INFO:teuthology.orchestra.run.vm02.stdout:[osd.3] 2026-03-20T17:57:11.934 INFO:teuthology.orchestra.run.vm02.stdout: key = AQB1ir1poG3QHRAA3P1qW547ZznY2JTSLHqoyQ== 2026-03-20T17:57:11.934 INFO:teuthology.orchestra.run.vm02.stdout:[osd.4] 2026-03-20T17:57:11.934 INFO:teuthology.orchestra.run.vm02.stdout: key = AQB2ir1pq8aNDRAAMgpJJGk/lptV5ItCo3a4WQ== 2026-03-20T17:57:11.934 INFO:teuthology.orchestra.run.vm02.stdout:[osd.5] 2026-03-20T17:57:11.934 INFO:teuthology.orchestra.run.vm02.stdout: key = AQB2ir1pBhyDOBAA8/MOHy4NAqRuuvrz0qZ96g== 2026-03-20T17:57:11.934 INFO:teuthology.orchestra.run.vm02.stdout:[client.0] 2026-03-20T17:57:11.934 INFO:teuthology.orchestra.run.vm02.stdout: key = AQBxir1pr0w0OBAABLea/iOlhvINBHBgJSnNtA== 2026-03-20T17:57:11.934 INFO:teuthology.orchestra.run.vm02.stdout:[client.1] 2026-03-20T17:57:11.934 INFO:teuthology.orchestra.run.vm02.stdout: key = AQBxir1pqN3EOhAAQ1w12tlv7TUtYlW6FtBElA== 2026-03-20T17:57:11.935 DEBUG:teuthology.orchestra.run.vm02:> sudo adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage ceph-authtool /etc/ceph/c1.keyring --name=osd.0 --cap mon 'allow profile osd' --cap mgr 'allow profile osd' --cap osd 'allow *' 2026-03-20T17:57:12.001 DEBUG:teuthology.orchestra.run.vm02:> sudo adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage ceph-authtool /etc/ceph/c1.keyring --name=osd.1 --cap mon 'allow profile osd' --cap mgr 'allow profile osd' --cap osd 'allow *' 2026-03-20T17:57:12.064 DEBUG:teuthology.orchestra.run.vm02:> sudo adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage ceph-authtool /etc/ceph/c1.keyring --name=osd.2 --cap mon 'allow profile osd' --cap mgr 'allow profile osd' --cap osd 'allow *' 2026-03-20T17:57:12.128 DEBUG:teuthology.orchestra.run.vm02:> sudo adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage ceph-authtool /etc/ceph/c1.keyring --name=mgr.x --cap mon 'allow profile mgr' --cap osd 'allow *' --cap mds 'allow *' 2026-03-20T17:57:12.193 DEBUG:teuthology.orchestra.run.vm02:> sudo adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage ceph-authtool /etc/ceph/c1.keyring --name=osd.3 --cap mon 'allow profile osd' --cap mgr 'allow profile osd' --cap osd 'allow *' 2026-03-20T17:57:12.262 DEBUG:teuthology.orchestra.run.vm02:> sudo adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage ceph-authtool /etc/ceph/c1.keyring --name=osd.4 --cap mon 'allow profile osd' --cap mgr 'allow profile osd' --cap osd 'allow *' 2026-03-20T17:57:12.328 DEBUG:teuthology.orchestra.run.vm02:> sudo adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage ceph-authtool /etc/ceph/c1.keyring --name=osd.5 --cap mon 'allow profile osd' --cap mgr 'allow profile osd' --cap osd 'allow *' 2026-03-20T17:57:12.393 DEBUG:teuthology.orchestra.run.vm02:> sudo adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage ceph-authtool /etc/ceph/c1.keyring --name=client.0 --cap mon 'allow rw' --cap mgr 'allow r' --cap osd 'allow rwx' --cap mds allow 2026-03-20T17:57:12.457 DEBUG:teuthology.orchestra.run.vm02:> sudo adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage ceph-authtool /etc/ceph/c1.keyring --name=client.1 --cap mon 'allow rw' --cap mgr 'allow r' --cap osd 'allow rwx' --cap mds allow 2026-03-20T17:57:12.523 INFO:tasks.ceph:Running mkfs on mon nodes... 2026-03-20T17:57:12.523 DEBUG:teuthology.orchestra.run.vm02:> sudo mkdir -p /var/lib/ceph/mon/c1-a 2026-03-20T17:57:12.574 DEBUG:teuthology.orchestra.run.vm02:> sudo adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage ceph-mon --cluster c1 --mkfs -i a --monmap /home/ubuntu/cephtest/c1.monmap --keyring /etc/ceph/c1.keyring 2026-03-20T17:57:12.654 DEBUG:teuthology.orchestra.run.vm02:> sudo chown -R ceph:ceph /var/lib/ceph/mon/c1-a 2026-03-20T17:57:12.707 DEBUG:teuthology.orchestra.run.vm02:> rm -- /home/ubuntu/cephtest/c1.monmap 2026-03-20T17:57:12.753 INFO:tasks.ceph:Starting mon daemons in cluster c1... 2026-03-20T17:57:12.753 INFO:tasks.ceph.c1.mon.a:Restarting daemon 2026-03-20T17:57:12.753 DEBUG:teuthology.orchestra.run.vm02:> sudo adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage daemon-helper kill ceph-mon -f --cluster c1 -i a 2026-03-20T17:57:12.795 INFO:tasks.ceph.c1.mon.a:Started 2026-03-20T17:57:12.795 INFO:tasks.ceph:Starting mgr daemons in cluster c1... 2026-03-20T17:57:12.795 INFO:tasks.ceph.c1.mgr.x:Restarting daemon 2026-03-20T17:57:12.795 DEBUG:teuthology.orchestra.run.vm05:> sudo adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage daemon-helper kill ceph-mgr -f --cluster c1 -i x 2026-03-20T17:57:12.797 INFO:tasks.ceph.c1.mgr.x:Started 2026-03-20T17:57:12.797 DEBUG:tasks.ceph:set 0 configs 2026-03-20T17:57:12.797 DEBUG:teuthology.orchestra.run.vm02:> sudo ceph --cluster c1 config dump 2026-03-20T17:57:12.893 INFO:teuthology.orchestra.run.vm02.stdout:WHO MASK LEVEL OPTION VALUE RO 2026-03-20T17:57:12.905 INFO:tasks.ceph:Setting crush tunables to default 2026-03-20T17:57:12.905 DEBUG:teuthology.orchestra.run.vm02:> sudo ceph --cluster c1 osd crush tunables default 2026-03-20T17:57:13.007 INFO:teuthology.orchestra.run.vm02.stderr:adjusted tunables profile to default 2026-03-20T17:57:13.019 INFO:tasks.ceph:check_enable_crimson: False 2026-03-20T17:57:13.019 INFO:tasks.ceph:Starting osd daemons in cluster c1... 2026-03-20T17:57:13.019 DEBUG:teuthology.orchestra.run.vm02:> set -ex 2026-03-20T17:57:13.019 DEBUG:teuthology.orchestra.run.vm02:> sudo dd if=/var/lib/ceph/osd/c1-0/fsid of=/dev/stdout 2026-03-20T17:57:13.027 DEBUG:teuthology.orchestra.run.vm02:> set -ex 2026-03-20T17:57:13.027 DEBUG:teuthology.orchestra.run.vm02:> sudo dd if=/var/lib/ceph/osd/c1-1/fsid of=/dev/stdout 2026-03-20T17:57:13.078 DEBUG:teuthology.orchestra.run.vm02:> set -ex 2026-03-20T17:57:13.078 DEBUG:teuthology.orchestra.run.vm02:> sudo dd if=/var/lib/ceph/osd/c1-2/fsid of=/dev/stdout 2026-03-20T17:57:13.130 DEBUG:teuthology.orchestra.run.vm05:> set -ex 2026-03-20T17:57:13.130 DEBUG:teuthology.orchestra.run.vm05:> sudo dd if=/var/lib/ceph/osd/c1-3/fsid of=/dev/stdout 2026-03-20T17:57:13.138 DEBUG:teuthology.orchestra.run.vm05:> set -ex 2026-03-20T17:57:13.138 DEBUG:teuthology.orchestra.run.vm05:> sudo dd if=/var/lib/ceph/osd/c1-4/fsid of=/dev/stdout 2026-03-20T17:57:13.186 DEBUG:teuthology.orchestra.run.vm05:> set -ex 2026-03-20T17:57:13.186 DEBUG:teuthology.orchestra.run.vm05:> sudo dd if=/var/lib/ceph/osd/c1-5/fsid of=/dev/stdout 2026-03-20T17:57:13.240 DEBUG:teuthology.orchestra.run.vm05:> sudo ceph --cluster c1 osd new d6bdc1d9-2e5a-421c-82d0-a5a63e4cce9f 0 2026-03-20T17:57:13.395 INFO:teuthology.orchestra.run.vm05.stdout:0 2026-03-20T17:57:13.407 DEBUG:teuthology.orchestra.run.vm05:> sudo ceph --cluster c1 osd new c0ab8e1d-5750-4187-a689-bd970c3d9755 1 2026-03-20T17:57:13.514 INFO:teuthology.orchestra.run.vm05.stdout:1 2026-03-20T17:57:13.527 DEBUG:teuthology.orchestra.run.vm05:> sudo ceph --cluster c1 osd new b585a68a-8110-4ac0-8c2f-7336eda5d34f 2 2026-03-20T17:57:13.634 INFO:teuthology.orchestra.run.vm05.stdout:2 2026-03-20T17:57:13.646 DEBUG:teuthology.orchestra.run.vm05:> sudo ceph --cluster c1 osd new bae3b9ef-03f0-49fb-916d-44717bb6debe 3 2026-03-20T17:57:13.750 INFO:teuthology.orchestra.run.vm05.stdout:3 2026-03-20T17:57:13.763 DEBUG:teuthology.orchestra.run.vm05:> sudo ceph --cluster c1 osd new e40a79a4-66fe-4b2a-9956-feed9d871626 4 2026-03-20T17:57:13.868 INFO:teuthology.orchestra.run.vm05.stdout:4 2026-03-20T17:57:13.880 DEBUG:teuthology.orchestra.run.vm05:> sudo ceph --cluster c1 osd new b4b92b50-8b59-4f2c-8dc8-23d74ca97904 5 2026-03-20T17:57:13.982 INFO:teuthology.orchestra.run.vm05.stdout:5 2026-03-20T17:57:13.994 INFO:tasks.ceph.c1.osd.0:Restarting daemon 2026-03-20T17:57:13.994 DEBUG:teuthology.orchestra.run.vm02:> sudo adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage daemon-helper kill ceph-osd -f --cluster c1 -i 0 2026-03-20T17:57:13.995 INFO:tasks.ceph.c1.osd.0:Started 2026-03-20T17:57:13.995 INFO:tasks.ceph.c1.osd.1:Restarting daemon 2026-03-20T17:57:13.995 DEBUG:teuthology.orchestra.run.vm02:> sudo adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage daemon-helper kill ceph-osd -f --cluster c1 -i 1 2026-03-20T17:57:13.996 INFO:tasks.ceph.c1.osd.1:Started 2026-03-20T17:57:13.996 INFO:tasks.ceph.c1.osd.2:Restarting daemon 2026-03-20T17:57:13.996 DEBUG:teuthology.orchestra.run.vm02:> sudo adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage daemon-helper kill ceph-osd -f --cluster c1 -i 2 2026-03-20T17:57:13.996 INFO:tasks.ceph.c1.osd.2:Started 2026-03-20T17:57:13.996 INFO:tasks.ceph.c1.osd.3:Restarting daemon 2026-03-20T17:57:13.996 DEBUG:teuthology.orchestra.run.vm05:> sudo adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage daemon-helper kill ceph-osd -f --cluster c1 -i 3 2026-03-20T17:57:13.997 INFO:tasks.ceph.c1.osd.3:Started 2026-03-20T17:57:13.997 INFO:tasks.ceph.c1.osd.4:Restarting daemon 2026-03-20T17:57:13.997 DEBUG:teuthology.orchestra.run.vm05:> sudo adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage daemon-helper kill ceph-osd -f --cluster c1 -i 4 2026-03-20T17:57:13.998 INFO:tasks.ceph.c1.osd.4:Started 2026-03-20T17:57:13.998 INFO:tasks.ceph.c1.osd.5:Restarting daemon 2026-03-20T17:57:13.998 DEBUG:teuthology.orchestra.run.vm05:> sudo adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage daemon-helper kill ceph-osd -f --cluster c1 -i 5 2026-03-20T17:57:13.999 INFO:tasks.ceph.c1.osd.5:Started 2026-03-20T17:57:13.999 DEBUG:teuthology.orchestra.run.vm02:> sudo adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage timeout 120 ceph --cluster c1 osd dump --format=json 2026-03-20T17:57:14.100 INFO:teuthology.orchestra.run.vm02.stdout: 2026-03-20T17:57:14.100 INFO:teuthology.orchestra.run.vm02.stdout:{"epoch":8,"fsid":"67b0dd48-80db-4ea8-b708-f5136e717957","created":"2026-03-20T17:57:12.845954+0000","modified":"2026-03-20T17:57:13.980717+0000","last_up_change":"0.000000","last_in_change":"2026-03-20T17:57:13.980717+0000","flags":"sortbitwise,recovery_deletes,purged_snapdirs,pglog_hardlimit","flags_num":5799936,"flags_set":["pglog_hardlimit","purged_snapdirs","recovery_deletes","sortbitwise"],"crush_version":2,"full_ratio":0.94999998807907104,"backfillfull_ratio":0.89999997615814209,"nearfull_ratio":0.85000002384185791,"cluster_snapshot":"","pool_max":0,"max_osd":6,"require_min_compat_client":"luminous","min_compat_client":"jewel","require_osd_release":"tentacle","allow_crimson":false,"pools":[],"osds":[{"osd":0,"uuid":"d6bdc1d9-2e5a-421c-82d0-a5a63e4cce9f","up":0,"in":1,"weight":1,"primary_affinity":1,"last_clean_begin":0,"last_clean_end":0,"up_from":0,"up_thru":0,"down_at":0,"lost_at":0,"public_addrs":{"addrvec":[]},"cluster_addrs":{"addrvec":[]},"heartbeat_back_addrs":{"addrvec":[]},"heartbeat_front_addrs":{"addrvec":[]},"public_addr":"(unrecognized address family 0)/0","cluster_addr":"(unrecognized address family 0)/0","heartbeat_back_addr":"(unrecognized address family 0)/0","heartbeat_front_addr":"(unrecognized address family 0)/0","state":["exists","new"]},{"osd":1,"uuid":"c0ab8e1d-5750-4187-a689-bd970c3d9755","up":0,"in":1,"weight":1,"primary_affinity":1,"last_clean_begin":0,"last_clean_end":0,"up_from":0,"up_thru":0,"down_at":0,"lost_at":0,"public_addrs":{"addrvec":[]},"cluster_addrs":{"addrvec":[]},"heartbeat_back_addrs":{"addrvec":[]},"heartbeat_front_addrs":{"addrvec":[]},"public_addr":"(unrecognized address family 0)/0","cluster_addr":"(unrecognized address family 0)/0","heartbeat_back_addr":"(unrecognized address family 0)/0","heartbeat_front_addr":"(unrecognized address family 0)/0","state":["exists","new"]},{"osd":2,"uuid":"b585a68a-8110-4ac0-8c2f-7336eda5d34f","up":0,"in":1,"weight":1,"primary_affinity":1,"last_clean_begin":0,"last_clean_end":0,"up_from":0,"up_thru":0,"down_at":0,"lost_at":0,"public_addrs":{"addrvec":[]},"cluster_addrs":{"addrvec":[]},"heartbeat_back_addrs":{"addrvec":[]},"heartbeat_front_addrs":{"addrvec":[]},"public_addr":"(unrecognized address family 0)/0","cluster_addr":"(unrecognized address family 0)/0","heartbeat_back_addr":"(unrecognized address family 0)/0","heartbeat_front_addr":"(unrecognized address family 0)/0","state":["exists","new"]},{"osd":3,"uuid":"bae3b9ef-03f0-49fb-916d-44717bb6debe","up":0,"in":1,"weight":1,"primary_affinity":1,"last_clean_begin":0,"last_clean_end":0,"up_from":0,"up_thru":0,"down_at":0,"lost_at":0,"public_addrs":{"addrvec":[]},"cluster_addrs":{"addrvec":[]},"heartbeat_back_addrs":{"addrvec":[]},"heartbeat_front_addrs":{"addrvec":[]},"public_addr":"(unrecognized address family 0)/0","cluster_addr":"(unrecognized address family 0)/0","heartbeat_back_addr":"(unrecognized address family 0)/0","heartbeat_front_addr":"(unrecognized address family 0)/0","state":["exists","new"]},{"osd":4,"uuid":"e40a79a4-66fe-4b2a-9956-feed9d871626","up":0,"in":1,"weight":1,"primary_affinity":1,"last_clean_begin":0,"last_clean_end":0,"up_from":0,"up_thru":0,"down_at":0,"lost_at":0,"public_addrs":{"addrvec":[]},"cluster_addrs":{"addrvec":[]},"heartbeat_back_addrs":{"addrvec":[]},"heartbeat_front_addrs":{"addrvec":[]},"public_addr":"(unrecognized address family 0)/0","cluster_addr":"(unrecognized address family 0)/0","heartbeat_back_addr":"(unrecognized address family 0)/0","heartbeat_front_addr":"(unrecognized address family 0)/0","state":["exists","new"]},{"osd":5,"uuid":"b4b92b50-8b59-4f2c-8dc8-23d74ca97904","up":0,"in":1,"weight":1,"primary_affinity":1,"last_clean_begin":0,"last_clean_end":0,"up_from":0,"up_thru":0,"down_at":0,"lost_at":0,"public_addrs":{"addrvec":[]},"cluster_addrs":{"addrvec":[]},"heartbeat_back_addrs":{"addrvec":[]},"heartbeat_front_addrs":{"addrvec":[]},"public_addr":"(unrecognized address family 0)/0","cluster_addr":"(unrecognized address family 0)/0","heartbeat_back_addr":"(unrecognized address family 0)/0","heartbeat_front_addr":"(unrecognized address family 0)/0","state":["exists","new"]}],"osd_xinfo":[{"osd":0,"down_stamp":"0.000000","laggy_probability":0,"laggy_interval":0,"features":0,"old_weight":0,"last_purged_snaps_scrub":"0.000000","dead_epoch":0},{"osd":1,"down_stamp":"0.000000","laggy_probability":0,"laggy_interval":0,"features":0,"old_weight":0,"last_purged_snaps_scrub":"0.000000","dead_epoch":0},{"osd":2,"down_stamp":"0.000000","laggy_probability":0,"laggy_interval":0,"features":0,"old_weight":0,"last_purged_snaps_scrub":"0.000000","dead_epoch":0},{"osd":3,"down_stamp":"0.000000","laggy_probability":0,"laggy_interval":0,"features":0,"old_weight":0,"last_purged_snaps_scrub":"0.000000","dead_epoch":0},{"osd":4,"down_stamp":"0.000000","laggy_probability":0,"laggy_interval":0,"features":0,"old_weight":0,"last_purged_snaps_scrub":"0.000000","dead_epoch":0},{"osd":5,"down_stamp":"0.000000","laggy_probability":0,"laggy_interval":0,"features":0,"old_weight":0,"last_purged_snaps_scrub":"0.000000","dead_epoch":0}],"pg_upmap":[],"pg_upmap_items":[],"pg_upmap_primaries":[],"pg_temp":[],"primary_temp":[],"blocklist":{},"range_blocklist":{},"erasure_code_profiles":{"default":{"crush-failure-domain":"osd","k":"2","m":"1","plugin":"isa","technique":"reed_sol_van"}},"removed_snaps_queue":[],"new_removed_snaps":[],"new_purged_snaps":[],"crush_node_flags":{},"device_class_flags":{},"stretch_mode":{"stretch_mode_enabled":false,"stretch_bucket_count":0,"degraded_stretch_mode":0,"recovering_stretch_mode":0,"stretch_mode_bucket":0}} 2026-03-20T17:57:14.107 INFO:tasks.ceph.c1.mgr.x.vm05.stderr:/usr/lib/python3/dist-packages/scipy/__init__.py:67: UserWarning: NumPy was imported from a Python sub-interpreter but NumPy does not properly support sub-interpreters. This will likely work for most users but might cause hard to track down issues or subtle bugs. A common user of the rare sub-interpreter feature is wsgi which also allows single-interpreter mode. 2026-03-20T17:57:14.107 INFO:tasks.ceph.c1.mgr.x.vm05.stderr:Improvements in the case of bugs are welcome, but is not on the NumPy roadmap, and full support may require significant effort to achieve. 2026-03-20T17:57:14.107 INFO:tasks.ceph.c1.mgr.x.vm05.stderr: from numpy import show_config as show_numpy_config 2026-03-20T17:57:14.112 INFO:tasks.ceph.ceph_manager.c1:[] 2026-03-20T17:57:14.112 INFO:tasks.ceph:Waiting for OSDs to come up 2026-03-20T17:57:14.292 INFO:tasks.ceph.c1.osd.4.vm05.stderr:2026-03-20T17:57:14.289+0000 7f582e033a40 -1 Falling back to public interface 2026-03-20T17:57:14.348 INFO:tasks.ceph.c1.osd.5.vm05.stderr:2026-03-20T17:57:14.345+0000 7f9f4f9e4a40 -1 Falling back to public interface 2026-03-20T17:57:14.356 INFO:tasks.ceph.c1.osd.3.vm05.stderr:2026-03-20T17:57:14.353+0000 7fe484aeaa40 -1 Falling back to public interface 2026-03-20T17:57:14.359 INFO:tasks.ceph.c1.osd.0.vm02.stderr:2026-03-20T17:57:14.357+0000 7f7f2933fa40 -1 Falling back to public interface 2026-03-20T17:57:14.359 INFO:tasks.ceph.c1.osd.2.vm02.stderr:2026-03-20T17:57:14.357+0000 7fe7815b5a40 -1 Falling back to public interface 2026-03-20T17:57:14.359 INFO:tasks.ceph.c1.osd.2.vm02.stderr:2026-03-20T17:57:14.357+0000 7fe780575640 -1 PosixStack listen unable to listen on v2:0.0.0.0:6800/0: (98) Address already in use 2026-03-20T17:57:14.379 INFO:tasks.ceph.c1.osd.1.vm02.stderr:2026-03-20T17:57:14.377+0000 7fd7e2287a40 -1 Falling back to public interface 2026-03-20T17:57:14.526 INFO:tasks.ceph.c1.osd.4.vm05.stderr:2026-03-20T17:57:14.521+0000 7f582e033a40 -1 osd.4 0 log_to_monitors true 2026-03-20T17:57:14.640 INFO:tasks.ceph.c1.osd.5.vm05.stderr:2026-03-20T17:57:14.633+0000 7f9f4f9e4a40 -1 osd.5 0 log_to_monitors true 2026-03-20T17:57:14.644 INFO:tasks.ceph.c1.osd.1.vm02.stderr:2026-03-20T17:57:14.641+0000 7fd7e2287a40 -1 osd.1 0 log_to_monitors true 2026-03-20T17:57:14.646 INFO:tasks.ceph.c1.osd.3.vm05.stderr:2026-03-20T17:57:14.641+0000 7fe484aeaa40 -1 osd.3 0 log_to_monitors true 2026-03-20T17:57:14.670 INFO:tasks.ceph.c1.osd.0.vm02.stderr:2026-03-20T17:57:14.665+0000 7f7f2933fa40 -1 osd.0 0 log_to_monitors true 2026-03-20T17:57:14.686 INFO:tasks.ceph.c1.osd.2.vm02.stderr:2026-03-20T17:57:14.681+0000 7fe7815b5a40 -1 osd.2 0 log_to_monitors true 2026-03-20T17:57:14.701 INFO:tasks.ceph.c1.mgr.x.vm05.stderr:Failed to import NVMeoFClient and related components: cannot import name 'NVMeoFClient' from 'dashboard.services.nvmeof_client' (/usr/share/ceph/mgr/dashboard/services/nvmeof_client.py) 2026-03-20T17:57:14.714 DEBUG:teuthology.orchestra.run.vm02:> adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage ceph --cluster c1 osd dump --format=json 2026-03-20T17:57:14.815 INFO:teuthology.misc.health.vm02.stdout: 2026-03-20T17:57:14.815 INFO:teuthology.misc.health.vm02.stdout:{"epoch":8,"fsid":"67b0dd48-80db-4ea8-b708-f5136e717957","created":"2026-03-20T17:57:12.845954+0000","modified":"2026-03-20T17:57:13.980717+0000","last_up_change":"0.000000","last_in_change":"2026-03-20T17:57:13.980717+0000","flags":"sortbitwise,recovery_deletes,purged_snapdirs,pglog_hardlimit","flags_num":5799936,"flags_set":["pglog_hardlimit","purged_snapdirs","recovery_deletes","sortbitwise"],"crush_version":2,"full_ratio":0.94999998807907104,"backfillfull_ratio":0.89999997615814209,"nearfull_ratio":0.85000002384185791,"cluster_snapshot":"","pool_max":0,"max_osd":6,"require_min_compat_client":"luminous","min_compat_client":"jewel","require_osd_release":"tentacle","allow_crimson":false,"pools":[],"osds":[{"osd":0,"uuid":"d6bdc1d9-2e5a-421c-82d0-a5a63e4cce9f","up":0,"in":1,"weight":1,"primary_affinity":1,"last_clean_begin":0,"last_clean_end":0,"up_from":0,"up_thru":0,"down_at":0,"lost_at":0,"public_addrs":{"addrvec":[]},"cluster_addrs":{"addrvec":[]},"heartbeat_back_addrs":{"addrvec":[]},"heartbeat_front_addrs":{"addrvec":[]},"public_addr":"(unrecognized address family 0)/0","cluster_addr":"(unrecognized address family 0)/0","heartbeat_back_addr":"(unrecognized address family 0)/0","heartbeat_front_addr":"(unrecognized address family 0)/0","state":["exists","new"]},{"osd":1,"uuid":"c0ab8e1d-5750-4187-a689-bd970c3d9755","up":0,"in":1,"weight":1,"primary_affinity":1,"last_clean_begin":0,"last_clean_end":0,"up_from":0,"up_thru":0,"down_at":0,"lost_at":0,"public_addrs":{"addrvec":[]},"cluster_addrs":{"addrvec":[]},"heartbeat_back_addrs":{"addrvec":[]},"heartbeat_front_addrs":{"addrvec":[]},"public_addr":"(unrecognized address family 0)/0","cluster_addr":"(unrecognized address family 0)/0","heartbeat_back_addr":"(unrecognized address family 0)/0","heartbeat_front_addr":"(unrecognized address family 0)/0","state":["exists","new"]},{"osd":2,"uuid":"b585a68a-8110-4ac0-8c2f-7336eda5d34f","up":0,"in":1,"weight":1,"primary_affinity":1,"last_clean_begin":0,"last_clean_end":0,"up_from":0,"up_thru":0,"down_at":0,"lost_at":0,"public_addrs":{"addrvec":[]},"cluster_addrs":{"addrvec":[]},"heartbeat_back_addrs":{"addrvec":[]},"heartbeat_front_addrs":{"addrvec":[]},"public_addr":"(unrecognized address family 0)/0","cluster_addr":"(unrecognized address family 0)/0","heartbeat_back_addr":"(unrecognized address family 0)/0","heartbeat_front_addr":"(unrecognized address family 0)/0","state":["exists","new"]},{"osd":3,"uuid":"bae3b9ef-03f0-49fb-916d-44717bb6debe","up":0,"in":1,"weight":1,"primary_affinity":1,"last_clean_begin":0,"last_clean_end":0,"up_from":0,"up_thru":0,"down_at":0,"lost_at":0,"public_addrs":{"addrvec":[]},"cluster_addrs":{"addrvec":[]},"heartbeat_back_addrs":{"addrvec":[]},"heartbeat_front_addrs":{"addrvec":[]},"public_addr":"(unrecognized address family 0)/0","cluster_addr":"(unrecognized address family 0)/0","heartbeat_back_addr":"(unrecognized address family 0)/0","heartbeat_front_addr":"(unrecognized address family 0)/0","state":["exists","new"]},{"osd":4,"uuid":"e40a79a4-66fe-4b2a-9956-feed9d871626","up":0,"in":1,"weight":1,"primary_affinity":1,"last_clean_begin":0,"last_clean_end":0,"up_from":0,"up_thru":0,"down_at":0,"lost_at":0,"public_addrs":{"addrvec":[]},"cluster_addrs":{"addrvec":[]},"heartbeat_back_addrs":{"addrvec":[]},"heartbeat_front_addrs":{"addrvec":[]},"public_addr":"(unrecognized address family 0)/0","cluster_addr":"(unrecognized address family 0)/0","heartbeat_back_addr":"(unrecognized address family 0)/0","heartbeat_front_addr":"(unrecognized address family 0)/0","state":["exists","new"]},{"osd":5,"uuid":"b4b92b50-8b59-4f2c-8dc8-23d74ca97904","up":0,"in":1,"weight":1,"primary_affinity":1,"last_clean_begin":0,"last_clean_end":0,"up_from":0,"up_thru":0,"down_at":0,"lost_at":0,"public_addrs":{"addrvec":[]},"cluster_addrs":{"addrvec":[]},"heartbeat_back_addrs":{"addrvec":[]},"heartbeat_front_addrs":{"addrvec":[]},"public_addr":"(unrecognized address family 0)/0","cluster_addr":"(unrecognized address family 0)/0","heartbeat_back_addr":"(unrecognized address family 0)/0","heartbeat_front_addr":"(unrecognized address family 0)/0","state":["exists","new"]}],"osd_xinfo":[{"osd":0,"down_stamp":"0.000000","laggy_probability":0,"laggy_interval":0,"features":0,"old_weight":0,"last_purged_snaps_scrub":"0.000000","dead_epoch":0},{"osd":1,"down_stamp":"0.000000","laggy_probability":0,"laggy_interval":0,"features":0,"old_weight":0,"last_purged_snaps_scrub":"0.000000","dead_epoch":0},{"osd":2,"down_stamp":"0.000000","laggy_probability":0,"laggy_interval":0,"features":0,"old_weight":0,"last_purged_snaps_scrub":"0.000000","dead_epoch":0},{"osd":3,"down_stamp":"0.000000","laggy_probability":0,"laggy_interval":0,"features":0,"old_weight":0,"last_purged_snaps_scrub":"0.000000","dead_epoch":0},{"osd":4,"down_stamp":"0.000000","laggy_probability":0,"laggy_interval":0,"features":0,"old_weight":0,"last_purged_snaps_scrub":"0.000000","dead_epoch":0},{"osd":5,"down_stamp":"0.000000","laggy_probability":0,"laggy_interval":0,"features":0,"old_weight":0,"last_purged_snaps_scrub":"0.000000","dead_epoch":0}],"pg_upmap":[],"pg_upmap_items":[],"pg_upmap_primaries":[],"pg_temp":[],"primary_temp":[],"blocklist":{},"range_blocklist":{},"erasure_code_profiles":{"default":{"crush-failure-domain":"osd","k":"2","m":"1","plugin":"isa","technique":"reed_sol_van"}},"removed_snaps_queue":[],"new_removed_snaps":[],"new_purged_snaps":[],"crush_node_flags":{},"device_class_flags":{},"stretch_mode":{"stretch_mode_enabled":false,"stretch_bucket_count":0,"degraded_stretch_mode":0,"recovering_stretch_mode":0,"stretch_mode_bucket":0}} 2026-03-20T17:57:14.826 DEBUG:teuthology.misc:0 of 6 OSDs are up 2026-03-20T17:57:16.004 INFO:tasks.ceph.c1.osd.3.vm05.stderr:2026-03-20T17:57:16.001+0000 7fe480a93640 -1 osd.3 0 waiting for initial osdmap 2026-03-20T17:57:16.006 INFO:tasks.ceph.c1.osd.3.vm05.stderr:2026-03-20T17:57:16.001+0000 7fe47b8a1640 -1 osd.3 10 set_numa_affinity unable to identify public interface '' numa node: (2) No such file or directory 2026-03-20T17:57:16.010 INFO:tasks.ceph.c1.osd.0.vm02.stderr:2026-03-20T17:57:16.009+0000 7f7f252e8640 -1 osd.0 0 waiting for initial osdmap 2026-03-20T17:57:16.011 INFO:tasks.ceph.c1.osd.1.vm02.stderr:2026-03-20T17:57:16.009+0000 7fd7de230640 -1 osd.1 0 waiting for initial osdmap 2026-03-20T17:57:16.011 INFO:tasks.ceph.c1.osd.2.vm02.stderr:2026-03-20T17:57:16.009+0000 7fe77d55e640 -1 osd.2 0 waiting for initial osdmap 2026-03-20T17:57:16.012 INFO:tasks.ceph.c1.osd.4.vm05.stderr:2026-03-20T17:57:16.009+0000 7f5829fdc640 -1 osd.4 0 waiting for initial osdmap 2026-03-20T17:57:16.012 INFO:tasks.ceph.c1.osd.5.vm05.stderr:2026-03-20T17:57:16.009+0000 7f9f4b98d640 -1 osd.5 0 waiting for initial osdmap 2026-03-20T17:57:16.014 INFO:tasks.ceph.c1.osd.0.vm02.stderr:2026-03-20T17:57:16.009+0000 7f7f200f6640 -1 osd.0 10 set_numa_affinity unable to identify public interface '' numa node: (2) No such file or directory 2026-03-20T17:57:16.014 INFO:tasks.ceph.c1.osd.1.vm02.stderr:2026-03-20T17:57:16.009+0000 7fd7d903e640 -1 osd.1 10 set_numa_affinity unable to identify public interface '' numa node: (2) No such file or directory 2026-03-20T17:57:16.014 INFO:tasks.ceph.c1.osd.2.vm02.stderr:2026-03-20T17:57:16.009+0000 7fe77836c640 -1 osd.2 10 set_numa_affinity unable to identify public interface '' numa node: (2) No such file or directory 2026-03-20T17:57:16.015 INFO:tasks.ceph.c1.osd.4.vm05.stderr:2026-03-20T17:57:16.009+0000 7f5824dea640 -1 osd.4 10 set_numa_affinity unable to identify public interface '' numa node: (2) No such file or directory 2026-03-20T17:57:16.016 INFO:tasks.ceph.c1.osd.5.vm05.stderr:2026-03-20T17:57:16.009+0000 7f9f4679b640 -1 osd.5 10 set_numa_affinity unable to identify public interface '' numa node: (2) No such file or directory 2026-03-20T17:57:16.174 INFO:tasks.ceph.c1.mgr.x.vm05.stderr:2026-03-20T17:57:16.169+0000 7f1df05bd640 -1 mgr.server handle_report got status from non-daemon mon.a 2026-03-20T17:57:21.429 DEBUG:teuthology.orchestra.run.vm02:> adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage ceph --cluster c1 osd dump --format=json 2026-03-20T17:57:21.589 INFO:teuthology.misc.health.vm02.stdout: 2026-03-20T17:57:21.589 INFO:teuthology.misc.health.vm02.stdout:{"epoch":14,"fsid":"67b0dd48-80db-4ea8-b708-f5136e717957","created":"2026-03-20T17:57:12.845954+0000","modified":"2026-03-20T17:57:20.179663+0000","last_up_change":"2026-03-20T17:57:16.988698+0000","last_in_change":"2026-03-20T17:57:13.980717+0000","flags":"sortbitwise,recovery_deletes,purged_snapdirs,pglog_hardlimit","flags_num":5799936,"flags_set":["pglog_hardlimit","purged_snapdirs","recovery_deletes","sortbitwise"],"crush_version":4,"full_ratio":0.94999998807907104,"backfillfull_ratio":0.89999997615814209,"nearfull_ratio":0.85000002384185791,"cluster_snapshot":"","pool_max":1,"max_osd":6,"require_min_compat_client":"luminous","min_compat_client":"jewel","require_osd_release":"tentacle","allow_crimson":false,"pools":[{"pool":1,"pool_name":".mgr","create_time":"2026-03-20T17:57:17.182191+0000","flags":1,"flags_names":"hashpspool","type":1,"size":2,"min_size":1,"crush_rule":0,"peering_crush_bucket_count":0,"peering_crush_bucket_target":0,"peering_crush_bucket_barrier":0,"peering_crush_bucket_mandatory_member":2147483647,"is_stretch_pool":false,"object_hash":2,"pg_autoscale_mode":"off","pg_num":1,"pg_placement_num":1,"pg_placement_num_target":1,"pg_num_target":1,"pg_num_pending":1,"last_pg_merge_meta":{"source_pgid":"0.0","ready_epoch":0,"last_epoch_started":0,"last_epoch_clean":0,"source_version":"0'0","target_version":"0'0"},"last_change":"14","last_force_op_resend":"0","last_force_op_resend_prenautilus":"0","last_force_op_resend_preluminous":"0","auid":0,"snap_mode":"selfmanaged","snap_seq":0,"snap_epoch":0,"pool_snaps":[],"removed_snaps":"[]","quota_max_bytes":0,"quota_max_objects":0,"tiers":[],"tier_of":-1,"read_tier":-1,"write_tier":-1,"cache_mode":"none","target_max_bytes":0,"target_max_objects":0,"cache_target_dirty_ratio_micro":400000,"cache_target_dirty_high_ratio_micro":600000,"cache_target_full_ratio_micro":800000,"cache_min_flush_age":0,"cache_min_evict_age":0,"erasure_code_profile":"","hit_set_params":{"type":"none"},"hit_set_period":0,"hit_set_count":0,"use_gmt_hitset":true,"min_read_recency_for_promote":0,"min_write_recency_for_promote":0,"hit_set_grade_decay_rate":0,"hit_set_search_last_n":0,"grade_table":[],"stripe_width":0,"expected_num_objects":0,"fast_read":false,"nonprimary_shards":"{}","options":{"pg_num_max":32,"pg_num_min":1},"application_metadata":{"mgr":{}},"read_balance":{"score_type":"Fair distribution","score_acting":6.059999942779541,"score_stable":6.059999942779541,"optimal_score":0.33000001311302185,"raw_score_acting":2,"raw_score_stable":2,"primary_affinity_weighted":1,"average_primary_affinity":1,"average_primary_affinity_weighted":1}}],"osds":[{"osd":0,"uuid":"d6bdc1d9-2e5a-421c-82d0-a5a63e4cce9f","up":1,"in":1,"weight":1,"primary_affinity":1,"last_clean_begin":0,"last_clean_end":0,"up_from":11,"up_thru":0,"down_at":0,"lost_at":0,"public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.102:6800","nonce":4063991565},{"type":"v1","addr":"192.168.123.102:6802","nonce":4063991565}]},"cluster_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.102:6804","nonce":4063991565},{"type":"v1","addr":"192.168.123.102:6806","nonce":4063991565}]},"heartbeat_back_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.102:6812","nonce":4063991565},{"type":"v1","addr":"192.168.123.102:6814","nonce":4063991565}]},"heartbeat_front_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.102:6808","nonce":4063991565},{"type":"v1","addr":"192.168.123.102:6810","nonce":4063991565}]},"public_addr":"192.168.123.102:6802/4063991565","cluster_addr":"192.168.123.102:6806/4063991565","heartbeat_back_addr":"192.168.123.102:6814/4063991565","heartbeat_front_addr":"192.168.123.102:6810/4063991565","state":["exists","up"]},{"osd":1,"uuid":"c0ab8e1d-5750-4187-a689-bd970c3d9755","up":1,"in":1,"weight":1,"primary_affinity":1,"last_clean_begin":0,"last_clean_end":0,"up_from":11,"up_thru":0,"down_at":0,"lost_at":0,"public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.102:6816","nonce":1097301766},{"type":"v1","addr":"192.168.123.102:6817","nonce":1097301766}]},"cluster_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.102:6818","nonce":1097301766},{"type":"v1","addr":"192.168.123.102:6819","nonce":1097301766}]},"heartbeat_back_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.102:6822","nonce":1097301766},{"type":"v1","addr":"192.168.123.102:6823","nonce":1097301766}]},"heartbeat_front_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.102:6820","nonce":1097301766},{"type":"v1","addr":"192.168.123.102:6821","nonce":1097301766}]},"public_addr":"192.168.123.102:6817/1097301766","cluster_addr":"192.168.123.102:6819/1097301766","heartbeat_back_addr":"192.168.123.102:6823/1097301766","heartbeat_front_addr":"192.168.123.102:6821/1097301766","state":["exists","up"]},{"osd":2,"uuid":"b585a68a-8110-4ac0-8c2f-7336eda5d34f","up":1,"in":1,"weight":1,"primary_affinity":1,"last_clean_begin":0,"last_clean_end":0,"up_from":11,"up_thru":0,"down_at":0,"lost_at":0,"public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.102:6801","nonce":2611319604},{"type":"v1","addr":"192.168.123.102:6803","nonce":2611319604}]},"cluster_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.102:6805","nonce":2611319604},{"type":"v1","addr":"192.168.123.102:6807","nonce":2611319604}]},"heartbeat_back_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.102:6813","nonce":2611319604},{"type":"v1","addr":"192.168.123.102:6815","nonce":2611319604}]},"heartbeat_front_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.102:6809","nonce":2611319604},{"type":"v1","addr":"192.168.123.102:6811","nonce":2611319604}]},"public_addr":"192.168.123.102:6803/2611319604","cluster_addr":"192.168.123.102:6807/2611319604","heartbeat_back_addr":"192.168.123.102:6815/2611319604","heartbeat_front_addr":"192.168.123.102:6811/2611319604","state":["exists","up"]},{"osd":3,"uuid":"bae3b9ef-03f0-49fb-916d-44717bb6debe","up":1,"in":1,"weight":1,"primary_affinity":1,"last_clean_begin":0,"last_clean_end":0,"up_from":11,"up_thru":12,"down_at":0,"lost_at":0,"public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.105:6816","nonce":573174972},{"type":"v1","addr":"192.168.123.105:6817","nonce":573174972}]},"cluster_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.105:6818","nonce":573174972},{"type":"v1","addr":"192.168.123.105:6819","nonce":573174972}]},"heartbeat_back_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.105:6822","nonce":573174972},{"type":"v1","addr":"192.168.123.105:6823","nonce":573174972}]},"heartbeat_front_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.105:6820","nonce":573174972},{"type":"v1","addr":"192.168.123.105:6821","nonce":573174972}]},"public_addr":"192.168.123.105:6817/573174972","cluster_addr":"192.168.123.105:6819/573174972","heartbeat_back_addr":"192.168.123.105:6823/573174972","heartbeat_front_addr":"192.168.123.105:6821/573174972","state":["exists","up"]},{"osd":4,"uuid":"e40a79a4-66fe-4b2a-9956-feed9d871626","up":1,"in":1,"weight":1,"primary_affinity":1,"last_clean_begin":0,"last_clean_end":0,"up_from":11,"up_thru":0,"down_at":0,"lost_at":0,"public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.105:6800","nonce":2795356309},{"type":"v1","addr":"192.168.123.105:6801","nonce":2795356309}]},"cluster_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.105:6802","nonce":2795356309},{"type":"v1","addr":"192.168.123.105:6803","nonce":2795356309}]},"heartbeat_back_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.105:6806","nonce":2795356309},{"type":"v1","addr":"192.168.123.105:6807","nonce":2795356309}]},"heartbeat_front_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.105:6804","nonce":2795356309},{"type":"v1","addr":"192.168.123.105:6805","nonce":2795356309}]},"public_addr":"192.168.123.105:6801/2795356309","cluster_addr":"192.168.123.105:6803/2795356309","heartbeat_back_addr":"192.168.123.105:6807/2795356309","heartbeat_front_addr":"192.168.123.105:6805/2795356309","state":["exists","up"]},{"osd":5,"uuid":"b4b92b50-8b59-4f2c-8dc8-23d74ca97904","up":1,"in":1,"weight":1,"primary_affinity":1,"last_clean_begin":0,"last_clean_end":0,"up_from":11,"up_thru":0,"down_at":0,"lost_at":0,"public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.105:6808","nonce":3989640680},{"type":"v1","addr":"192.168.123.105:6809","nonce":3989640680}]},"cluster_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.105:6810","nonce":3989640680},{"type":"v1","addr":"192.168.123.105:6811","nonce":3989640680}]},"heartbeat_back_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.105:6814","nonce":3989640680},{"type":"v1","addr":"192.168.123.105:6815","nonce":3989640680}]},"heartbeat_front_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.105:6812","nonce":3989640680},{"type":"v1","addr":"192.168.123.105:6813","nonce":3989640680}]},"public_addr":"192.168.123.105:6809/3989640680","cluster_addr":"192.168.123.105:6811/3989640680","heartbeat_back_addr":"192.168.123.105:6815/3989640680","heartbeat_front_addr":"192.168.123.105:6813/3989640680","state":["exists","up"]}],"osd_xinfo":[{"osd":0,"down_stamp":"0.000000","laggy_probability":0,"laggy_interval":0,"features":4544132024016699391,"old_weight":0,"last_purged_snaps_scrub":"2026-03-20T17:57:15.637775+0000","dead_epoch":0},{"osd":1,"down_stamp":"0.000000","laggy_probability":0,"laggy_interval":0,"features":4544132024016699391,"old_weight":0,"last_purged_snaps_scrub":"2026-03-20T17:57:15.678150+0000","dead_epoch":0},{"osd":2,"down_stamp":"0.000000","laggy_probability":0,"laggy_interval":0,"features":4544132024016699391,"old_weight":0,"last_purged_snaps_scrub":"2026-03-20T17:57:15.737220+0000","dead_epoch":0},{"osd":3,"down_stamp":"0.000000","laggy_probability":0,"laggy_interval":0,"features":4544132024016699391,"old_weight":0,"last_purged_snaps_scrub":"0.000000","dead_epoch":0},{"osd":4,"down_stamp":"0.000000","laggy_probability":0,"laggy_interval":0,"features":4544132024016699391,"old_weight":0,"last_purged_snaps_scrub":"0.000000","dead_epoch":0},{"osd":5,"down_stamp":"0.000000","laggy_probability":0,"laggy_interval":0,"features":4544132024016699391,"old_weight":0,"last_purged_snaps_scrub":"0.000000","dead_epoch":0}],"pg_upmap":[],"pg_upmap_items":[],"pg_upmap_primaries":[],"pg_temp":[],"primary_temp":[],"blocklist":{},"range_blocklist":{},"erasure_code_profiles":{"default":{"crush-failure-domain":"osd","k":"2","m":"1","plugin":"isa","technique":"reed_sol_van"}},"removed_snaps_queue":[],"new_removed_snaps":[],"new_purged_snaps":[],"crush_node_flags":{},"device_class_flags":{},"stretch_mode":{"stretch_mode_enabled":false,"stretch_bucket_count":0,"degraded_stretch_mode":0,"recovering_stretch_mode":0,"stretch_mode_bucket":0}} 2026-03-20T17:57:21.606 DEBUG:teuthology.misc:6 of 6 OSDs are up 2026-03-20T17:57:21.606 INFO:tasks.ceph:Creating RBD pool 2026-03-20T17:57:21.606 DEBUG:teuthology.orchestra.run.vm02:> sudo ceph --cluster c1 osd pool create rbd 8 2026-03-20T17:57:22.194 INFO:teuthology.orchestra.run.vm02.stderr:pool 'rbd' created 2026-03-20T17:57:22.210 DEBUG:teuthology.orchestra.run.vm02:> rbd --cluster c1 pool init rbd 2026-03-20T17:57:22.225 INFO:teuthology.orchestra.run.vm02.stderr:ignoring --setuser ceph since I am not root 2026-03-20T17:57:22.225 INFO:teuthology.orchestra.run.vm02.stderr:ignoring --setgroup ceph since I am not root 2026-03-20T17:57:25.257 INFO:tasks.ceph:Starting mds daemons in cluster c1... 2026-03-20T17:57:25.257 DEBUG:teuthology.orchestra.run.vm02:> sudo adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage timeout 120 ceph --cluster c1 config log 1 --format=json 2026-03-20T17:57:25.257 INFO:tasks.daemonwatchdog.daemon_watchdog:watchdog starting 2026-03-20T17:57:25.463 INFO:teuthology.orchestra.run.vm02.stdout: 2026-03-20T17:57:25.477 INFO:teuthology.orchestra.run.vm02.stdout:[{"version":1,"timestamp":"0.000000","name":"","changes":[]}] 2026-03-20T17:57:25.477 INFO:tasks.ceph_manager:config epoch is 1 2026-03-20T17:57:25.478 INFO:tasks.ceph:Waiting until c1 daemons up and pgs clean... 2026-03-20T17:57:25.478 INFO:tasks.ceph.ceph_manager.c1:waiting for mgr available 2026-03-20T17:57:25.478 DEBUG:teuthology.orchestra.run.vm02:> sudo adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage timeout 120 ceph --cluster c1 mgr dump --format=json 2026-03-20T17:57:25.655 INFO:teuthology.orchestra.run.vm02.stdout: 2026-03-20T17:57:25.668 INFO:teuthology.orchestra.run.vm02.stdout:{"epoch":5,"flags":0,"active_gid":4105,"active_name":"x","active_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.105:6824","nonce":2893263527},{"type":"v1","addr":"192.168.123.105:6825","nonce":2893263527}]},"active_addr":"192.168.123.105:6825/2893263527","active_change":"2026-03-20T17:57:15.162159+0000","active_mgr_features":4544132024016699391,"available":true,"standbys":[],"modules":["iostat","nfs"],"available_modules":[{"name":"alerts","can_run":true,"error_string":"","module_options":{"interval":{"name":"interval","type":"secs","level":"advanced","flags":1,"default_value":"60","min":"","max":"","enum_allowed":[],"desc":"How frequently to reexamine health status","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"smtp_destination":{"name":"smtp_destination","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"Email address to send alerts to, use commas to separate multiple","long_desc":"","tags":[],"see_also":[]},"smtp_from_name":{"name":"smtp_from_name","type":"str","level":"advanced","flags":1,"default_value":"Ceph","min":"","max":"","enum_allowed":[],"desc":"Email From: name","long_desc":"","tags":[],"see_also":[]},"smtp_host":{"name":"smtp_host","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"SMTP server","long_desc":"","tags":[],"see_also":[]},"smtp_password":{"name":"smtp_password","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"Password to authenticate with","long_desc":"","tags":[],"see_also":[]},"smtp_port":{"name":"smtp_port","type":"int","level":"advanced","flags":1,"default_value":"465","min":"","max":"","enum_allowed":[],"desc":"SMTP port","long_desc":"","tags":[],"see_also":[]},"smtp_sender":{"name":"smtp_sender","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"SMTP envelope sender","long_desc":"","tags":[],"see_also":[]},"smtp_ssl":{"name":"smtp_ssl","type":"bool","level":"advanced","flags":1,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"Use SSL to connect to SMTP server","long_desc":"","tags":[],"see_also":[]},"smtp_user":{"name":"smtp_user","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"User to authenticate as","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"balancer","can_run":true,"error_string":"","module_options":{"active":{"name":"active","type":"bool","level":"advanced","flags":1,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"automatically balance PGs across cluster","long_desc":"","tags":[],"see_also":[]},"begin_time":{"name":"begin_time","type":"str","level":"advanced","flags":1,"default_value":"0000","min":"","max":"","enum_allowed":[],"desc":"beginning time of day to automatically balance","long_desc":"This is a time of day in the format HHMM.","tags":[],"see_also":[]},"begin_weekday":{"name":"begin_weekday","type":"uint","level":"advanced","flags":1,"default_value":"0","min":"0","max":"6","enum_allowed":[],"desc":"Restrict automatic balancing to this day of the week or later","long_desc":"0 = Sunday, 1 = Monday, etc.","tags":[],"see_also":[]},"crush_compat_max_iterations":{"name":"crush_compat_max_iterations","type":"uint","level":"advanced","flags":1,"default_value":"25","min":"1","max":"250","enum_allowed":[],"desc":"maximum number of iterations to attempt optimization","long_desc":"","tags":[],"see_also":[]},"crush_compat_metrics":{"name":"crush_compat_metrics","type":"str","level":"advanced","flags":1,"default_value":"pgs,objects,bytes","min":"","max":"","enum_allowed":[],"desc":"metrics with which to calculate OSD utilization","long_desc":"Value is a list of one or more of \"pgs\", \"objects\", or \"bytes\", and indicates which metrics to use to balance utilization.","tags":[],"see_also":[]},"crush_compat_step":{"name":"crush_compat_step","type":"float","level":"advanced","flags":1,"default_value":"0.5","min":"0.001","max":"0.999","enum_allowed":[],"desc":"aggressiveness of optimization","long_desc":".99 is very aggressive, .01 is less aggressive","tags":[],"see_also":[]},"end_time":{"name":"end_time","type":"str","level":"advanced","flags":1,"default_value":"2359","min":"","max":"","enum_allowed":[],"desc":"ending time of day to automatically balance","long_desc":"This is a time of day in the format HHMM.","tags":[],"see_also":[]},"end_weekday":{"name":"end_weekday","type":"uint","level":"advanced","flags":1,"default_value":"0","min":"0","max":"6","enum_allowed":[],"desc":"Restrict automatic balancing to days of the week earlier than this","long_desc":"0 = Sunday, 1 = Monday, etc.","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"min_score":{"name":"min_score","type":"float","level":"advanced","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"minimum score, below which no optimization is attempted","long_desc":"","tags":[],"see_also":[]},"mode":{"name":"mode","type":"str","level":"advanced","flags":1,"default_value":"upmap","min":"","max":"","enum_allowed":["crush-compat","none","read","upmap","upmap-read"],"desc":"Balancer mode","long_desc":"","tags":[],"see_also":[]},"pool_ids":{"name":"pool_ids","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"pools which the automatic balancing will be limited to","long_desc":"","tags":[],"see_also":[]},"sleep_interval":{"name":"sleep_interval","type":"secs","level":"advanced","flags":1,"default_value":"60","min":"","max":"","enum_allowed":[],"desc":"how frequently to wake up and attempt optimization","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"update_pg_upmap_activity":{"name":"update_pg_upmap_activity","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Updates pg_upmap activity stats to be used in `balancer status detail`","long_desc":"","tags":[],"see_also":[]},"upmap_max_deviation":{"name":"upmap_max_deviation","type":"int","level":"advanced","flags":1,"default_value":"5","min":"1","max":"","enum_allowed":[],"desc":"deviation below which no optimization is attempted","long_desc":"If the number of PGs are within this count then no optimization is attempted","tags":[],"see_also":[]},"upmap_max_optimizations":{"name":"upmap_max_optimizations","type":"uint","level":"advanced","flags":1,"default_value":"10","min":"","max":"","enum_allowed":[],"desc":"maximum upmap optimizations to make per attempt","long_desc":"","tags":[],"see_also":[]}}},{"name":"cephadm","can_run":true,"error_string":"","module_options":{"agent_down_multiplier":{"name":"agent_down_multiplier","type":"float","level":"advanced","flags":0,"default_value":"3.0","min":"","max":"","enum_allowed":[],"desc":"Multiplied by agent refresh rate to calculate how long agent must not report before being marked down","long_desc":"","tags":[],"see_also":[]},"agent_refresh_rate":{"name":"agent_refresh_rate","type":"secs","level":"advanced","flags":0,"default_value":"20","min":"","max":"","enum_allowed":[],"desc":"How often agent on each host will try to gather and send metadata","long_desc":"","tags":[],"see_also":[]},"agent_starting_port":{"name":"agent_starting_port","type":"int","level":"advanced","flags":0,"default_value":"4721","min":"","max":"","enum_allowed":[],"desc":"First port agent will try to bind to (will also try up to next 1000 subsequent ports if blocked)","long_desc":"","tags":[],"see_also":[]},"allow_ptrace":{"name":"allow_ptrace","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"allow SYS_PTRACE capability on ceph containers","long_desc":"The SYS_PTRACE capability is needed to attach to a process with gdb or strace. Enabling this options can allow debugging daemons that encounter problems at runtime.","tags":[],"see_also":[]},"autotune_interval":{"name":"autotune_interval","type":"secs","level":"advanced","flags":0,"default_value":"600","min":"","max":"","enum_allowed":[],"desc":"how frequently to autotune daemon memory","long_desc":"","tags":[],"see_also":[]},"autotune_memory_target_ratio":{"name":"autotune_memory_target_ratio","type":"float","level":"advanced","flags":0,"default_value":"0.7","min":"","max":"","enum_allowed":[],"desc":"ratio of total system memory to divide amongst autotuned daemons","long_desc":"","tags":[],"see_also":[]},"cephadm_log_destination":{"name":"cephadm_log_destination","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":["file","file,syslog","syslog"],"desc":"Destination for cephadm command's persistent logging","long_desc":"","tags":[],"see_also":[]},"certificate_automated_rotation_enabled":{"name":"certificate_automated_rotation_enabled","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"This flag controls whether cephadm automatically rotates certificates upon expiration.","long_desc":"","tags":[],"see_also":[]},"certificate_check_debug_mode":{"name":"certificate_check_debug_mode","type":"bool","level":"dev","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"FOR TESTING ONLY: This flag forces the certificate check instead of waiting for certificate_check_period.","long_desc":"","tags":[],"see_also":[]},"certificate_check_period":{"name":"certificate_check_period","type":"int","level":"advanced","flags":0,"default_value":"1","min":"0","max":"30","enum_allowed":[],"desc":"Specifies how often (in days) the certificate should be checked for validity.","long_desc":"","tags":[],"see_also":[]},"certificate_duration_days":{"name":"certificate_duration_days","type":"int","level":"advanced","flags":0,"default_value":"1095","min":"90","max":"3650","enum_allowed":[],"desc":"Specifies the duration of self certificates generated and signed by cephadm root CA","long_desc":"","tags":[],"see_also":[]},"certificate_renewal_threshold_days":{"name":"certificate_renewal_threshold_days","type":"int","level":"advanced","flags":0,"default_value":"30","min":"10","max":"90","enum_allowed":[],"desc":"Specifies the lead time in days to initiate certificate renewal before expiration.","long_desc":"","tags":[],"see_also":[]},"cgroups_split":{"name":"cgroups_split","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"Pass --cgroups=split when cephadm creates containers (currently podman only)","long_desc":"","tags":[],"see_also":[]},"config_checks_enabled":{"name":"config_checks_enabled","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Enable or disable the cephadm configuration analysis","long_desc":"","tags":[],"see_also":[]},"config_dashboard":{"name":"config_dashboard","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"manage configs like API endpoints in Dashboard.","long_desc":"","tags":[],"see_also":[]},"container_image_alertmanager":{"name":"container_image_alertmanager","type":"str","level":"advanced","flags":0,"default_value":"quay.io/prometheus/alertmanager:v0.28.1","min":"","max":"","enum_allowed":[],"desc":"Alertmanager container image","long_desc":"","tags":[],"see_also":[]},"container_image_base":{"name":"container_image_base","type":"str","level":"advanced","flags":1,"default_value":"quay.io/ceph/ceph","min":"","max":"","enum_allowed":[],"desc":"Container image name, without the tag","long_desc":"","tags":[],"see_also":[]},"container_image_elasticsearch":{"name":"container_image_elasticsearch","type":"str","level":"advanced","flags":0,"default_value":"quay.io/omrizeneva/elasticsearch:6.8.23","min":"","max":"","enum_allowed":[],"desc":"Elasticsearch container image","long_desc":"","tags":[],"see_also":[]},"container_image_grafana":{"name":"container_image_grafana","type":"str","level":"advanced","flags":0,"default_value":"quay.io/ceph/grafana:12.3.1","min":"","max":"","enum_allowed":[],"desc":"Grafana container image","long_desc":"","tags":[],"see_also":[]},"container_image_haproxy":{"name":"container_image_haproxy","type":"str","level":"advanced","flags":0,"default_value":"quay.io/ceph/haproxy:2.3","min":"","max":"","enum_allowed":[],"desc":"Haproxy container image","long_desc":"","tags":[],"see_also":[]},"container_image_jaeger_agent":{"name":"container_image_jaeger_agent","type":"str","level":"advanced","flags":0,"default_value":"quay.io/jaegertracing/jaeger-agent:1.29","min":"","max":"","enum_allowed":[],"desc":"Jaeger agent container image","long_desc":"","tags":[],"see_also":[]},"container_image_jaeger_collector":{"name":"container_image_jaeger_collector","type":"str","level":"advanced","flags":0,"default_value":"quay.io/jaegertracing/jaeger-collector:1.29","min":"","max":"","enum_allowed":[],"desc":"Jaeger collector container image","long_desc":"","tags":[],"see_also":[]},"container_image_jaeger_query":{"name":"container_image_jaeger_query","type":"str","level":"advanced","flags":0,"default_value":"quay.io/jaegertracing/jaeger-query:1.29","min":"","max":"","enum_allowed":[],"desc":"Jaeger query container image","long_desc":"","tags":[],"see_also":[]},"container_image_keepalived":{"name":"container_image_keepalived","type":"str","level":"advanced","flags":0,"default_value":"quay.io/ceph/keepalived:2.2.4","min":"","max":"","enum_allowed":[],"desc":"Keepalived container image","long_desc":"","tags":[],"see_also":[]},"container_image_loki":{"name":"container_image_loki","type":"str","level":"advanced","flags":0,"default_value":"docker.io/grafana/loki:3.0.0","min":"","max":"","enum_allowed":[],"desc":"Loki container image","long_desc":"","tags":[],"see_also":[]},"container_image_nginx":{"name":"container_image_nginx","type":"str","level":"advanced","flags":0,"default_value":"quay.io/ceph/nginx:sclorg-nginx-126","min":"","max":"","enum_allowed":[],"desc":"Nginx container image","long_desc":"","tags":[],"see_also":[]},"container_image_node_exporter":{"name":"container_image_node_exporter","type":"str","level":"advanced","flags":0,"default_value":"quay.io/prometheus/node-exporter:v1.9.1","min":"","max":"","enum_allowed":[],"desc":"Node exporter container image","long_desc":"","tags":[],"see_also":[]},"container_image_nvmeof":{"name":"container_image_nvmeof","type":"str","level":"advanced","flags":0,"default_value":"quay.io/ceph/nvmeof:1.5","min":"","max":"","enum_allowed":[],"desc":"Nvmeof container image","long_desc":"","tags":[],"see_also":[]},"container_image_oauth2_proxy":{"name":"container_image_oauth2_proxy","type":"str","level":"advanced","flags":0,"default_value":"quay.io/oauth2-proxy/oauth2-proxy:v7.6.0","min":"","max":"","enum_allowed":[],"desc":"Oauth2 proxy container image","long_desc":"","tags":[],"see_also":[]},"container_image_prometheus":{"name":"container_image_prometheus","type":"str","level":"advanced","flags":0,"default_value":"quay.io/prometheus/prometheus:v3.6.0","min":"","max":"","enum_allowed":[],"desc":"Prometheus container image","long_desc":"","tags":[],"see_also":[]},"container_image_promtail":{"name":"container_image_promtail","type":"str","level":"advanced","flags":0,"default_value":"docker.io/grafana/promtail:3.0.0","min":"","max":"","enum_allowed":[],"desc":"Promtail container image","long_desc":"","tags":[],"see_also":[]},"container_image_samba":{"name":"container_image_samba","type":"str","level":"advanced","flags":0,"default_value":"quay.io/samba.org/samba-server:ceph20-centos-amd64","min":"","max":"","enum_allowed":[],"desc":"Samba container image","long_desc":"","tags":[],"see_also":[]},"container_image_samba_metrics":{"name":"container_image_samba_metrics","type":"str","level":"advanced","flags":0,"default_value":"quay.io/samba.org/samba-metrics:ceph20-centos-amd64","min":"","max":"","enum_allowed":[],"desc":"Samba metrics container image","long_desc":"","tags":[],"see_also":[]},"container_image_snmp_gateway":{"name":"container_image_snmp_gateway","type":"str","level":"advanced","flags":0,"default_value":"docker.io/maxwo/snmp-notifier:v1.2.1","min":"","max":"","enum_allowed":[],"desc":"Snmp gateway container image","long_desc":"","tags":[],"see_also":[]},"container_init":{"name":"container_init","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"Run podman/docker with `--init`","long_desc":"","tags":[],"see_also":[]},"daemon_cache_timeout":{"name":"daemon_cache_timeout","type":"secs","level":"advanced","flags":0,"default_value":"600","min":"","max":"","enum_allowed":[],"desc":"seconds to cache service (daemon) inventory","long_desc":"","tags":[],"see_also":[]},"default_cephadm_command_timeout":{"name":"default_cephadm_command_timeout","type":"int","level":"advanced","flags":0,"default_value":"900","min":"","max":"","enum_allowed":[],"desc":"Default timeout applied to cephadm commands run directly on the host (in seconds)","long_desc":"","tags":[],"see_also":[]},"default_registry":{"name":"default_registry","type":"str","level":"advanced","flags":0,"default_value":"quay.io","min":"","max":"","enum_allowed":[],"desc":"Search-registry to which we should normalize unqualified image names. This is not the default registry","long_desc":"","tags":[],"see_also":[]},"device_cache_timeout":{"name":"device_cache_timeout","type":"secs","level":"advanced","flags":0,"default_value":"1800","min":"","max":"","enum_allowed":[],"desc":"seconds to cache device inventory","long_desc":"","tags":[],"see_also":[]},"device_enhanced_scan":{"name":"device_enhanced_scan","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Use libstoragemgmt during device scans","long_desc":"","tags":[],"see_also":[]},"facts_cache_timeout":{"name":"facts_cache_timeout","type":"secs","level":"advanced","flags":0,"default_value":"60","min":"","max":"","enum_allowed":[],"desc":"seconds to cache host facts data","long_desc":"","tags":[],"see_also":[]},"grafana_dashboards_path":{"name":"grafana_dashboards_path","type":"str","level":"advanced","flags":0,"default_value":"/etc/grafana/dashboards/ceph-dashboard/","min":"","max":"","enum_allowed":[],"desc":"location of dashboards to include in grafana deployments","long_desc":"","tags":[],"see_also":[]},"host_check_interval":{"name":"host_check_interval","type":"secs","level":"advanced","flags":0,"default_value":"600","min":"","max":"","enum_allowed":[],"desc":"how frequently to perform a host check","long_desc":"","tags":[],"see_also":[]},"hw_monitoring":{"name":"hw_monitoring","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Deploy hw monitoring daemon on every host.","long_desc":"","tags":[],"see_also":[]},"inventory_list_all":{"name":"inventory_list_all","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Whether ceph-volume inventory should report more devices (mostly mappers (LVs / mpaths), partitions...)","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_refresh_metadata":{"name":"log_refresh_metadata","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Log all refresh metadata. Includes daemon, device, and host info collected regularly. Only has effect if logging at debug level","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"log to the \"cephadm\" cluster log channel\"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"manage_etc_ceph_ceph_conf":{"name":"manage_etc_ceph_ceph_conf","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Manage and own /etc/ceph/ceph.conf on the hosts.","long_desc":"","tags":[],"see_also":[]},"manage_etc_ceph_ceph_conf_hosts":{"name":"manage_etc_ceph_ceph_conf_hosts","type":"str","level":"advanced","flags":0,"default_value":"*","min":"","max":"","enum_allowed":[],"desc":"PlacementSpec describing on which hosts to manage /etc/ceph/ceph.conf","long_desc":"","tags":[],"see_also":[]},"max_count_per_host":{"name":"max_count_per_host","type":"int","level":"advanced","flags":0,"default_value":"10","min":"","max":"","enum_allowed":[],"desc":"max number of daemons per service per host","long_desc":"","tags":[],"see_also":[]},"max_osd_draining_count":{"name":"max_osd_draining_count","type":"int","level":"advanced","flags":0,"default_value":"10","min":"","max":"","enum_allowed":[],"desc":"max number of osds that will be drained simultaneously when osds are removed","long_desc":"","tags":[],"see_also":[]},"migration_current":{"name":"migration_current","type":"int","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"internal - do not modify","long_desc":"","tags":[],"see_also":[]},"mode":{"name":"mode","type":"str","level":"advanced","flags":0,"default_value":"root","min":"","max":"","enum_allowed":["cephadm-package","root"],"desc":"mode for remote execution of cephadm","long_desc":"","tags":[],"see_also":[]},"oob_default_addr":{"name":"oob_default_addr","type":"str","level":"advanced","flags":0,"default_value":"169.254.1.1","min":"","max":"","enum_allowed":[],"desc":"Default address for RedFish API (oob management).","long_desc":"","tags":[],"see_also":[]},"prometheus_alerts_path":{"name":"prometheus_alerts_path","type":"str","level":"advanced","flags":0,"default_value":"/etc/prometheus/ceph/ceph_default_alerts.yml","min":"","max":"","enum_allowed":[],"desc":"location of alerts to include in prometheus deployments","long_desc":"","tags":[],"see_also":[]},"registry_insecure":{"name":"registry_insecure","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Registry is to be considered insecure (no TLS available). Only for development purposes.","long_desc":"","tags":[],"see_also":[]},"registry_password":{"name":"registry_password","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"Custom repository password. Only used for logging into a registry.","long_desc":"","tags":[],"see_also":[]},"registry_url":{"name":"registry_url","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"Registry url for login purposes. This is not the default registry","long_desc":"","tags":[],"see_also":[]},"registry_username":{"name":"registry_username","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"Custom repository username. Only used for logging into a registry.","long_desc":"","tags":[],"see_also":[]},"secure_monitoring_stack":{"name":"secure_monitoring_stack","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Enable TLS security for all the monitoring stack daemons","long_desc":"","tags":[],"see_also":[]},"service_discovery_port":{"name":"service_discovery_port","type":"int","level":"advanced","flags":0,"default_value":"8765","min":"","max":"","enum_allowed":[],"desc":"cephadm service discovery port","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"ssh_config_file":{"name":"ssh_config_file","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"customized SSH config file to connect to managed hosts","long_desc":"","tags":[],"see_also":[]},"ssh_keepalive_count_max":{"name":"ssh_keepalive_count_max","type":"int","level":"advanced","flags":0,"default_value":"3","min":"","max":"","enum_allowed":[],"desc":"How many times ssh connections can fail liveness checks before the host is marked offline","long_desc":"","tags":[],"see_also":[]},"ssh_keepalive_interval":{"name":"ssh_keepalive_interval","type":"int","level":"advanced","flags":0,"default_value":"7","min":"","max":"","enum_allowed":[],"desc":"How often ssh connections are checked for liveness","long_desc":"","tags":[],"see_also":[]},"stray_daemon_check_interval":{"name":"stray_daemon_check_interval","type":"secs","level":"advanced","flags":0,"default_value":"1800","min":"","max":"","enum_allowed":[],"desc":"how frequently cephadm should check for the presence of stray daemons","long_desc":"","tags":[],"see_also":[]},"use_agent":{"name":"use_agent","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Use cephadm agent on each host to gather and send metadata","long_desc":"","tags":[],"see_also":[]},"use_repo_digest":{"name":"use_repo_digest","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"Automatically convert image tags to image digest. Make sure all daemons use the same image","long_desc":"","tags":[],"see_also":[]},"warn_on_failed_host_check":{"name":"warn_on_failed_host_check","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"raise a health warning if the host check fails","long_desc":"","tags":[],"see_also":[]},"warn_on_stray_daemons":{"name":"warn_on_stray_daemons","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"raise a health warning if daemons are detected that are not managed by cephadm","long_desc":"","tags":[],"see_also":[]},"warn_on_stray_hosts":{"name":"warn_on_stray_hosts","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"raise a health warning if daemons are detected on a host that is not managed by cephadm","long_desc":"","tags":[],"see_also":[]}}},{"name":"crash","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"retain_interval":{"name":"retain_interval","type":"secs","level":"advanced","flags":1,"default_value":"31536000","min":"","max":"","enum_allowed":[],"desc":"how long to retain crashes before pruning them","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"warn_recent_interval":{"name":"warn_recent_interval","type":"secs","level":"advanced","flags":1,"default_value":"1209600","min":"","max":"","enum_allowed":[],"desc":"time interval in which to warn about recent crashes","long_desc":"","tags":[],"see_also":[]}}},{"name":"dashboard","can_run":true,"error_string":"","module_options":{"ACCOUNT_LOCKOUT_ATTEMPTS":{"name":"ACCOUNT_LOCKOUT_ATTEMPTS","type":"int","level":"advanced","flags":0,"default_value":"10","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"ALERTMANAGER_API_HOST":{"name":"ALERTMANAGER_API_HOST","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"ALERTMANAGER_API_SSL_VERIFY":{"name":"ALERTMANAGER_API_SSL_VERIFY","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"AUDIT_API_ENABLED":{"name":"AUDIT_API_ENABLED","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"AUDIT_API_LOG_PAYLOAD":{"name":"AUDIT_API_LOG_PAYLOAD","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"ENABLE_BROWSABLE_API":{"name":"ENABLE_BROWSABLE_API","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"FEATURE_TOGGLE_CEPHFS":{"name":"FEATURE_TOGGLE_CEPHFS","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"FEATURE_TOGGLE_DASHBOARD":{"name":"FEATURE_TOGGLE_DASHBOARD","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"FEATURE_TOGGLE_ISCSI":{"name":"FEATURE_TOGGLE_ISCSI","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"FEATURE_TOGGLE_MIRRORING":{"name":"FEATURE_TOGGLE_MIRRORING","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"FEATURE_TOGGLE_NFS":{"name":"FEATURE_TOGGLE_NFS","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"FEATURE_TOGGLE_RBD":{"name":"FEATURE_TOGGLE_RBD","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"FEATURE_TOGGLE_RGW":{"name":"FEATURE_TOGGLE_RGW","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"GANESHA_CLUSTERS_RADOS_POOL_NAMESPACE":{"name":"GANESHA_CLUSTERS_RADOS_POOL_NAMESPACE","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"GRAFANA_API_PASSWORD":{"name":"GRAFANA_API_PASSWORD","type":"str","level":"advanced","flags":0,"default_value":"admin","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"GRAFANA_API_SSL_VERIFY":{"name":"GRAFANA_API_SSL_VERIFY","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"GRAFANA_API_URL":{"name":"GRAFANA_API_URL","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"GRAFANA_API_USERNAME":{"name":"GRAFANA_API_USERNAME","type":"str","level":"advanced","flags":0,"default_value":"admin","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"GRAFANA_FRONTEND_API_URL":{"name":"GRAFANA_FRONTEND_API_URL","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"GRAFANA_UPDATE_DASHBOARDS":{"name":"GRAFANA_UPDATE_DASHBOARDS","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"ISCSI_API_SSL_VERIFICATION":{"name":"ISCSI_API_SSL_VERIFICATION","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"ISSUE_TRACKER_API_KEY":{"name":"ISSUE_TRACKER_API_KEY","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"MANAGED_BY_CLUSTERS":{"name":"MANAGED_BY_CLUSTERS","type":"str","level":"advanced","flags":0,"default_value":"[]","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"MULTICLUSTER_CONFIG":{"name":"MULTICLUSTER_CONFIG","type":"str","level":"advanced","flags":0,"default_value":"{}","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PROMETHEUS_API_HOST":{"name":"PROMETHEUS_API_HOST","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PROMETHEUS_API_SSL_VERIFY":{"name":"PROMETHEUS_API_SSL_VERIFY","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PROM_ALERT_CREDENTIAL_CACHE_TTL":{"name":"PROM_ALERT_CREDENTIAL_CACHE_TTL","type":"int","level":"advanced","flags":0,"default_value":"60","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PWD_POLICY_CHECK_COMPLEXITY_ENABLED":{"name":"PWD_POLICY_CHECK_COMPLEXITY_ENABLED","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PWD_POLICY_CHECK_EXCLUSION_LIST_ENABLED":{"name":"PWD_POLICY_CHECK_EXCLUSION_LIST_ENABLED","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PWD_POLICY_CHECK_LENGTH_ENABLED":{"name":"PWD_POLICY_CHECK_LENGTH_ENABLED","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PWD_POLICY_CHECK_OLDPWD_ENABLED":{"name":"PWD_POLICY_CHECK_OLDPWD_ENABLED","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PWD_POLICY_CHECK_REPETITIVE_CHARS_ENABLED":{"name":"PWD_POLICY_CHECK_REPETITIVE_CHARS_ENABLED","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PWD_POLICY_CHECK_SEQUENTIAL_CHARS_ENABLED":{"name":"PWD_POLICY_CHECK_SEQUENTIAL_CHARS_ENABLED","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PWD_POLICY_CHECK_USERNAME_ENABLED":{"name":"PWD_POLICY_CHECK_USERNAME_ENABLED","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PWD_POLICY_ENABLED":{"name":"PWD_POLICY_ENABLED","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PWD_POLICY_EXCLUSION_LIST":{"name":"PWD_POLICY_EXCLUSION_LIST","type":"str","level":"advanced","flags":0,"default_value":"osd,host,dashboard,pool,block,nfs,ceph,monitors,gateway,logs,crush,maps","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PWD_POLICY_MIN_COMPLEXITY":{"name":"PWD_POLICY_MIN_COMPLEXITY","type":"int","level":"advanced","flags":0,"default_value":"10","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PWD_POLICY_MIN_LENGTH":{"name":"PWD_POLICY_MIN_LENGTH","type":"int","level":"advanced","flags":0,"default_value":"8","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"REST_REQUESTS_TIMEOUT":{"name":"REST_REQUESTS_TIMEOUT","type":"int","level":"advanced","flags":0,"default_value":"45","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"RGW_API_ACCESS_KEY":{"name":"RGW_API_ACCESS_KEY","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"RGW_API_ADMIN_RESOURCE":{"name":"RGW_API_ADMIN_RESOURCE","type":"str","level":"advanced","flags":0,"default_value":"admin","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"RGW_API_SECRET_KEY":{"name":"RGW_API_SECRET_KEY","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"RGW_API_SSL_VERIFY":{"name":"RGW_API_SSL_VERIFY","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"RGW_HOSTNAME_PER_DAEMON":{"name":"RGW_HOSTNAME_PER_DAEMON","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"UNSAFE_TLS_v1_2":{"name":"UNSAFE_TLS_v1_2","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"USER_PWD_EXPIRATION_SPAN":{"name":"USER_PWD_EXPIRATION_SPAN","type":"int","level":"advanced","flags":0,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"USER_PWD_EXPIRATION_WARNING_1":{"name":"USER_PWD_EXPIRATION_WARNING_1","type":"int","level":"advanced","flags":0,"default_value":"10","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"USER_PWD_EXPIRATION_WARNING_2":{"name":"USER_PWD_EXPIRATION_WARNING_2","type":"int","level":"advanced","flags":0,"default_value":"5","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"cross_origin_url":{"name":"cross_origin_url","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"crt_file":{"name":"crt_file","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"crypto_caller":{"name":"crypto_caller","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"debug":{"name":"debug","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Enable/disable debug options","long_desc":"","tags":[],"see_also":[]},"jwt_token_ttl":{"name":"jwt_token_ttl","type":"int","level":"advanced","flags":0,"default_value":"28800","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"key_file":{"name":"key_file","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"motd":{"name":"motd","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"The message of the day","long_desc":"","tags":[],"see_also":[]},"redirect_resolve_ip_addr":{"name":"redirect_resolve_ip_addr","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"server_addr":{"name":"server_addr","type":"str","level":"advanced","flags":0,"default_value":"::","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"server_port":{"name":"server_port","type":"int","level":"advanced","flags":0,"default_value":"8080","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"ssl":{"name":"ssl","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"ssl_server_port":{"name":"ssl_server_port","type":"int","level":"advanced","flags":0,"default_value":"8443","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sso_oauth2":{"name":"sso_oauth2","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"standby_behaviour":{"name":"standby_behaviour","type":"str","level":"advanced","flags":0,"default_value":"redirect","min":"","max":"","enum_allowed":["error","redirect"],"desc":"","long_desc":"","tags":[],"see_also":[]},"standby_error_status_code":{"name":"standby_error_status_code","type":"int","level":"advanced","flags":0,"default_value":"500","min":"400","max":"599","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"url_prefix":{"name":"url_prefix","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"devicehealth","can_run":true,"error_string":"","module_options":{"enable_monitoring":{"name":"enable_monitoring","type":"bool","level":"advanced","flags":1,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"monitor device health metrics","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"mark_out_threshold":{"name":"mark_out_threshold","type":"secs","level":"advanced","flags":1,"default_value":"2419200","min":"","max":"","enum_allowed":[],"desc":"automatically mark OSD if it may fail before this long","long_desc":"","tags":[],"see_also":[]},"pool_name":{"name":"pool_name","type":"str","level":"advanced","flags":1,"default_value":"device_health_metrics","min":"","max":"","enum_allowed":[],"desc":"name of pool in which to store device health metrics","long_desc":"","tags":[],"see_also":[]},"retention_period":{"name":"retention_period","type":"secs","level":"advanced","flags":1,"default_value":"15552000","min":"","max":"","enum_allowed":[],"desc":"how long to retain device health metrics","long_desc":"","tags":[],"see_also":[]},"scrape_frequency":{"name":"scrape_frequency","type":"secs","level":"advanced","flags":1,"default_value":"86400","min":"","max":"","enum_allowed":[],"desc":"how frequently to scrape device health metrics","long_desc":"","tags":[],"see_also":[]},"self_heal":{"name":"self_heal","type":"bool","level":"advanced","flags":1,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"preemptively heal cluster around devices that may fail","long_desc":"","tags":[],"see_also":[]},"sleep_interval":{"name":"sleep_interval","type":"secs","level":"advanced","flags":1,"default_value":"600","min":"","max":"","enum_allowed":[],"desc":"how frequently to wake up and check device health","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"warn_threshold":{"name":"warn_threshold","type":"secs","level":"advanced","flags":1,"default_value":"7257600","min":"","max":"","enum_allowed":[],"desc":"raise health warning if OSD may fail before this long","long_desc":"","tags":[],"see_also":[]}}},{"name":"diskprediction_local","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"predict_interval":{"name":"predict_interval","type":"str","level":"advanced","flags":0,"default_value":"86400","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"predictor_model":{"name":"predictor_model","type":"str","level":"advanced","flags":0,"default_value":"prophetstor","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sleep_interval":{"name":"sleep_interval","type":"str","level":"advanced","flags":0,"default_value":"600","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"influx","can_run":false,"error_string":"influxdb python module not found","module_options":{"batch_size":{"name":"batch_size","type":"int","level":"advanced","flags":0,"default_value":"5000","min":"","max":"","enum_allowed":[],"desc":"How big batches of data points should be when sending to InfluxDB.","long_desc":"","tags":[],"see_also":[]},"database":{"name":"database","type":"str","level":"advanced","flags":0,"default_value":"ceph","min":"","max":"","enum_allowed":[],"desc":"InfluxDB database name. You will need to create this database and grant write privileges to the configured username or the username must have admin privileges to create it.","long_desc":"","tags":[],"see_also":[]},"hostname":{"name":"hostname","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"InfluxDB server hostname","long_desc":"","tags":[],"see_also":[]},"interval":{"name":"interval","type":"secs","level":"advanced","flags":0,"default_value":"30","min":"5","max":"","enum_allowed":[],"desc":"Time between reports to InfluxDB. Default 30 seconds.","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"password":{"name":"password","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"password of InfluxDB server user","long_desc":"","tags":[],"see_also":[]},"port":{"name":"port","type":"int","level":"advanced","flags":0,"default_value":"8086","min":"","max":"","enum_allowed":[],"desc":"InfluxDB server port","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"ssl":{"name":"ssl","type":"str","level":"advanced","flags":0,"default_value":"false","min":"","max":"","enum_allowed":[],"desc":"Use https connection for InfluxDB server. Use \"true\" or \"false\".","long_desc":"","tags":[],"see_also":[]},"threads":{"name":"threads","type":"int","level":"advanced","flags":0,"default_value":"5","min":"1","max":"32","enum_allowed":[],"desc":"How many worker threads should be spawned for sending data to InfluxDB.","long_desc":"","tags":[],"see_also":[]},"username":{"name":"username","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"username of InfluxDB server user","long_desc":"","tags":[],"see_also":[]},"verify_ssl":{"name":"verify_ssl","type":"str","level":"advanced","flags":0,"default_value":"true","min":"","max":"","enum_allowed":[],"desc":"Verify https cert for InfluxDB server. Use \"true\" or \"false\".","long_desc":"","tags":[],"see_also":[]}}},{"name":"insights","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"iostat","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"localpool","can_run":true,"error_string":"","module_options":{"failure_domain":{"name":"failure_domain","type":"str","level":"advanced","flags":1,"default_value":"host","min":"","max":"","enum_allowed":[],"desc":"failure domain for any created local pool","long_desc":"what failure domain we should separate data replicas across.","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"min_size":{"name":"min_size","type":"int","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"default min_size for any created local pool","long_desc":"value to set min_size to (unchanged from Ceph's default if this option is not set)","tags":[],"see_also":[]},"num_rep":{"name":"num_rep","type":"int","level":"advanced","flags":1,"default_value":"3","min":"","max":"","enum_allowed":[],"desc":"default replica count for any created local pool","long_desc":"","tags":[],"see_also":[]},"pg_num":{"name":"pg_num","type":"int","level":"advanced","flags":1,"default_value":"128","min":"","max":"","enum_allowed":[],"desc":"default pg_num for any created local pool","long_desc":"","tags":[],"see_also":[]},"prefix":{"name":"prefix","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"name prefix for any created local pool","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"subtree":{"name":"subtree","type":"str","level":"advanced","flags":1,"default_value":"rack","min":"","max":"","enum_allowed":[],"desc":"CRUSH level for which to create a local pool","long_desc":"which CRUSH subtree type the module should create a pool for.","tags":[],"see_also":[]}}},{"name":"mirroring","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"nfs","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"orchestrator","can_run":true,"error_string":"","module_options":{"fail_fs":{"name":"fail_fs","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Fail filesystem for rapid multi-rank mds upgrade","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"orchestrator":{"name":"orchestrator","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["cephadm","rook","test_orchestrator"],"desc":"Orchestrator backend","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"osd_perf_query","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"osd_support","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"pg_autoscaler","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sleep_interval":{"name":"sleep_interval","type":"secs","level":"advanced","flags":0,"default_value":"60","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"threshold":{"name":"threshold","type":"float","level":"advanced","flags":0,"default_value":"3.0","min":"1.0","max":"","enum_allowed":[],"desc":"scaling threshold","long_desc":"The factor by which the `NEW PG_NUM` must vary from the current`PG_NUM` before being accepted. Cannot be less than 1.0","tags":[],"see_also":[]}}},{"name":"progress","can_run":true,"error_string":"","module_options":{"allow_pg_recovery_event":{"name":"allow_pg_recovery_event","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"allow the module to show pg recovery progress","long_desc":"","tags":[],"see_also":[]},"enabled":{"name":"enabled","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"max_completed_events":{"name":"max_completed_events","type":"int","level":"advanced","flags":1,"default_value":"50","min":"","max":"","enum_allowed":[],"desc":"number of past completed events to remember","long_desc":"","tags":[],"see_also":[]},"sleep_interval":{"name":"sleep_interval","type":"secs","level":"advanced","flags":1,"default_value":"5","min":"","max":"","enum_allowed":[],"desc":"how long the module is going to sleep","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"prometheus","can_run":true,"error_string":"","module_options":{"cache":{"name":"cache","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"exclude_perf_counters":{"name":"exclude_perf_counters","type":"bool","level":"advanced","flags":1,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"Do not include perf-counters in the metrics output","long_desc":"Gathering perf-counters from a single Prometheus exporter can degrade ceph-mgr performance, especially in large clusters. Instead, Ceph-exporter daemons are now used by default for perf-counter gathering. This should only be disabled when no ceph-exporters are deployed.","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"rbd_stats_pools":{"name":"rbd_stats_pools","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"rbd_stats_pools_refresh_interval":{"name":"rbd_stats_pools_refresh_interval","type":"int","level":"advanced","flags":0,"default_value":"300","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"scrape_interval":{"name":"scrape_interval","type":"float","level":"advanced","flags":0,"default_value":"15.0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"server_addr":{"name":"server_addr","type":"str","level":"advanced","flags":0,"default_value":"::","min":"","max":"","enum_allowed":[],"desc":"the IPv4 or IPv6 address on which the module listens for HTTP requests","long_desc":"","tags":[],"see_also":[]},"server_port":{"name":"server_port","type":"int","level":"advanced","flags":1,"default_value":"9283","min":"","max":"","enum_allowed":[],"desc":"the port on which the module listens for HTTP requests","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"stale_cache_strategy":{"name":"stale_cache_strategy","type":"str","level":"advanced","flags":0,"default_value":"log","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"standby_behaviour":{"name":"standby_behaviour","type":"str","level":"advanced","flags":1,"default_value":"default","min":"","max":"","enum_allowed":["default","error"],"desc":"","long_desc":"","tags":[],"see_also":[]},"standby_error_status_code":{"name":"standby_error_status_code","type":"int","level":"advanced","flags":1,"default_value":"500","min":"400","max":"599","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"rbd_support","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"max_concurrent_snap_create":{"name":"max_concurrent_snap_create","type":"int","level":"advanced","flags":0,"default_value":"10","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"mirror_snapshot_schedule":{"name":"mirror_snapshot_schedule","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"trash_purge_schedule":{"name":"trash_purge_schedule","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"rgw","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"secondary_zone_period_retry_limit":{"name":"secondary_zone_period_retry_limit","type":"int","level":"advanced","flags":0,"default_value":"5","min":"","max":"","enum_allowed":[],"desc":"RGW module period update retry limit for secondary site","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"selftest","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"roption1":{"name":"roption1","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"roption2":{"name":"roption2","type":"str","level":"advanced","flags":0,"default_value":"xyz","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"rwoption1":{"name":"rwoption1","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"rwoption2":{"name":"rwoption2","type":"int","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"rwoption3":{"name":"rwoption3","type":"float","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"rwoption4":{"name":"rwoption4","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"rwoption5":{"name":"rwoption5","type":"bool","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"rwoption6":{"name":"rwoption6","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"rwoption7":{"name":"rwoption7","type":"int","level":"advanced","flags":0,"default_value":"","min":"1","max":"42","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"testkey":{"name":"testkey","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"testlkey":{"name":"testlkey","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"testnewline":{"name":"testnewline","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"snap_schedule","can_run":true,"error_string":"","module_options":{"allow_m_granularity":{"name":"allow_m_granularity","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"allow minute scheduled snapshots","long_desc":"","tags":[],"see_also":[]},"dump_on_update":{"name":"dump_on_update","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"dump database to debug log on update","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"stats","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"status","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"telegraf","can_run":true,"error_string":"","module_options":{"address":{"name":"address","type":"str","level":"advanced","flags":0,"default_value":"unixgram:///tmp/telegraf.sock","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"interval":{"name":"interval","type":"secs","level":"advanced","flags":0,"default_value":"15","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"telemetry","can_run":true,"error_string":"","module_options":{"channel_basic":{"name":"channel_basic","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"Share basic cluster information (size, version)","long_desc":"","tags":[],"see_also":[]},"channel_crash":{"name":"channel_crash","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"Share metadata about Ceph daemon crashes (version, stack straces, etc)","long_desc":"","tags":[],"see_also":[]},"channel_device":{"name":"channel_device","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"Share device health metrics (e.g., SMART data, minus potentially identifying info like serial numbers)","long_desc":"","tags":[],"see_also":[]},"channel_ident":{"name":"channel_ident","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Share a user-provided description and/or contact email for the cluster","long_desc":"","tags":[],"see_also":[]},"channel_perf":{"name":"channel_perf","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Share various performance metrics of a cluster","long_desc":"","tags":[],"see_also":[]},"contact":{"name":"contact","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"description":{"name":"description","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"device_url":{"name":"device_url","type":"str","level":"advanced","flags":0,"default_value":"https://telemetry.ceph.com/device","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"enabled":{"name":"enabled","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"interval":{"name":"interval","type":"int","level":"advanced","flags":0,"default_value":"24","min":"8","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"last_opt_revision":{"name":"last_opt_revision","type":"int","level":"advanced","flags":0,"default_value":"1","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"leaderboard":{"name":"leaderboard","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"leaderboard_description":{"name":"leaderboard_description","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"organization":{"name":"organization","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"proxy":{"name":"proxy","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"url":{"name":"url","type":"str","level":"advanced","flags":0,"default_value":"https://telemetry.ceph.com/report","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"test_orchestrator","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"volumes","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"max_concurrent_clones":{"name":"max_concurrent_clones","type":"int","level":"advanced","flags":0,"default_value":"4","min":"","max":"","enum_allowed":[],"desc":"Number of asynchronous cloner threads","long_desc":"","tags":[],"see_also":[]},"pause_cloning":{"name":"pause_cloning","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Pause asynchronous cloner threads","long_desc":"","tags":[],"see_also":[]},"pause_purging":{"name":"pause_purging","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Pause asynchronous subvolume purge threads","long_desc":"","tags":[],"see_also":[]},"periodic_async_work":{"name":"periodic_async_work","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Periodically check for async work","long_desc":"","tags":[],"see_also":[]},"snapshot_clone_delay":{"name":"snapshot_clone_delay","type":"int","level":"advanced","flags":0,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"Delay clone begin operation by snapshot_clone_delay seconds","long_desc":"","tags":[],"see_also":[]},"snapshot_clone_no_wait":{"name":"snapshot_clone_no_wait","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"Reject subvolume clone request when cloner threads are busy","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}}],"services":{},"always_on_modules":{"octopus":["balancer","crash","devicehealth","orchestrator","pg_autoscaler","progress","rbd_support","status","telemetry","volumes"],"pacific":["balancer","crash","devicehealth","orchestrator","pg_autoscaler","progress","rbd_support","status","telemetry","volumes"],"quincy":["balancer","crash","devicehealth","orchestrator","pg_autoscaler","progress","rbd_support","status","telemetry","volumes"],"reef":["balancer","crash","devicehealth","orchestrator","pg_autoscaler","progress","rbd_support","status","telemetry","volumes"],"squid":["balancer","crash","devicehealth","orchestrator","pg_autoscaler","progress","rbd_support","status","telemetry","volumes"],"tentacle":["balancer","crash","devicehealth","orchestrator","pg_autoscaler","progress","rbd_support","status","telemetry","volumes"]},"force_disabled_modules":{},"last_failure_osd_epoch":0,"active_clients":[{"name":"devicehealth","addrvec":[{"type":"v2","addr":"192.168.123.105:0","nonce":4080340097}]},{"name":"libcephsqlite","addrvec":[{"type":"v2","addr":"192.168.123.105:0","nonce":635890982}]},{"name":"rbd_support","addrvec":[{"type":"v2","addr":"192.168.123.105:0","nonce":4117358404}]},{"name":"volumes","addrvec":[{"type":"v2","addr":"192.168.123.105:0","nonce":2448688516}]}]} 2026-03-20T17:57:25.668 INFO:tasks.ceph.ceph_manager.c1:mgr available! 2026-03-20T17:57:25.669 INFO:tasks.ceph.ceph_manager.c1:waiting for all up 2026-03-20T17:57:25.669 DEBUG:teuthology.orchestra.run.vm02:> sudo adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage timeout 120 ceph --cluster c1 osd dump --format=json 2026-03-20T17:57:25.828 INFO:teuthology.orchestra.run.vm02.stdout: 2026-03-20T17:57:25.829 INFO:teuthology.orchestra.run.vm02.stdout:{"epoch":18,"fsid":"67b0dd48-80db-4ea8-b708-f5136e717957","created":"2026-03-20T17:57:12.845954+0000","modified":"2026-03-20T17:57:25.243326+0000","last_up_change":"2026-03-20T17:57:16.988698+0000","last_in_change":"2026-03-20T17:57:13.980717+0000","flags":"sortbitwise,recovery_deletes,purged_snapdirs,pglog_hardlimit","flags_num":5799936,"flags_set":["pglog_hardlimit","purged_snapdirs","recovery_deletes","sortbitwise"],"crush_version":4,"full_ratio":0.94999998807907104,"backfillfull_ratio":0.89999997615814209,"nearfull_ratio":0.85000002384185791,"cluster_snapshot":"","pool_max":2,"max_osd":6,"require_min_compat_client":"luminous","min_compat_client":"jewel","require_osd_release":"tentacle","allow_crimson":false,"pools":[{"pool":1,"pool_name":".mgr","create_time":"2026-03-20T17:57:17.182191+0000","flags":1,"flags_names":"hashpspool","type":1,"size":2,"min_size":1,"crush_rule":0,"peering_crush_bucket_count":0,"peering_crush_bucket_target":0,"peering_crush_bucket_barrier":0,"peering_crush_bucket_mandatory_member":2147483647,"is_stretch_pool":false,"object_hash":2,"pg_autoscale_mode":"off","pg_num":1,"pg_placement_num":1,"pg_placement_num_target":1,"pg_num_target":1,"pg_num_pending":1,"last_pg_merge_meta":{"source_pgid":"0.0","ready_epoch":0,"last_epoch_started":0,"last_epoch_clean":0,"source_version":"0'0","target_version":"0'0"},"last_change":"14","last_force_op_resend":"0","last_force_op_resend_prenautilus":"0","last_force_op_resend_preluminous":"0","auid":0,"snap_mode":"selfmanaged","snap_seq":0,"snap_epoch":0,"pool_snaps":[],"removed_snaps":"[]","quota_max_bytes":0,"quota_max_objects":0,"tiers":[],"tier_of":-1,"read_tier":-1,"write_tier":-1,"cache_mode":"none","target_max_bytes":0,"target_max_objects":0,"cache_target_dirty_ratio_micro":400000,"cache_target_dirty_high_ratio_micro":600000,"cache_target_full_ratio_micro":800000,"cache_min_flush_age":0,"cache_min_evict_age":0,"erasure_code_profile":"","hit_set_params":{"type":"none"},"hit_set_period":0,"hit_set_count":0,"use_gmt_hitset":true,"min_read_recency_for_promote":0,"min_write_recency_for_promote":0,"hit_set_grade_decay_rate":0,"hit_set_search_last_n":0,"grade_table":[],"stripe_width":0,"expected_num_objects":0,"fast_read":false,"nonprimary_shards":"{}","options":{"pg_num_max":32,"pg_num_min":1},"application_metadata":{"mgr":{}},"read_balance":{"score_type":"Fair distribution","score_acting":6.059999942779541,"score_stable":6.059999942779541,"optimal_score":0.33000001311302185,"raw_score_acting":2,"raw_score_stable":2,"primary_affinity_weighted":1,"average_primary_affinity":1,"average_primary_affinity_weighted":1}},{"pool":2,"pool_name":"rbd","create_time":"2026-03-20T17:57:21.776478+0000","flags":8193,"flags_names":"hashpspool,selfmanaged_snaps","type":1,"size":2,"min_size":1,"crush_rule":0,"peering_crush_bucket_count":0,"peering_crush_bucket_target":0,"peering_crush_bucket_barrier":0,"peering_crush_bucket_mandatory_member":2147483647,"is_stretch_pool":false,"object_hash":2,"pg_autoscale_mode":"off","pg_num":8,"pg_placement_num":8,"pg_placement_num_target":8,"pg_num_target":8,"pg_num_pending":8,"last_pg_merge_meta":{"source_pgid":"0.0","ready_epoch":0,"last_epoch_started":0,"last_epoch_clean":0,"source_version":"0'0","target_version":"0'0"},"last_change":"18","last_force_op_resend":"0","last_force_op_resend_prenautilus":"0","last_force_op_resend_preluminous":"0","auid":0,"snap_mode":"selfmanaged","snap_seq":2,"snap_epoch":18,"pool_snaps":[],"removed_snaps":"[]","quota_max_bytes":0,"quota_max_objects":0,"tiers":[],"tier_of":-1,"read_tier":-1,"write_tier":-1,"cache_mode":"none","target_max_bytes":0,"target_max_objects":0,"cache_target_dirty_ratio_micro":400000,"cache_target_dirty_high_ratio_micro":600000,"cache_target_full_ratio_micro":800000,"cache_min_flush_age":0,"cache_min_evict_age":0,"erasure_code_profile":"","hit_set_params":{"type":"none"},"hit_set_period":0,"hit_set_count":0,"use_gmt_hitset":true,"min_read_recency_for_promote":0,"min_write_recency_for_promote":0,"hit_set_grade_decay_rate":0,"hit_set_search_last_n":0,"grade_table":[],"stripe_width":0,"expected_num_objects":0,"fast_read":false,"nonprimary_shards":"{}","options":{},"application_metadata":{"rbd":{}},"read_balance":{"score_type":"Fair distribution","score_acting":2.25,"score_stable":2.25,"optimal_score":1,"raw_score_acting":2.25,"raw_score_stable":2.25,"primary_affinity_weighted":1,"average_primary_affinity":1,"average_primary_affinity_weighted":1}}],"osds":[{"osd":0,"uuid":"d6bdc1d9-2e5a-421c-82d0-a5a63e4cce9f","up":1,"in":1,"weight":1,"primary_affinity":1,"last_clean_begin":0,"last_clean_end":0,"up_from":11,"up_thru":0,"down_at":0,"lost_at":0,"public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.102:6800","nonce":4063991565},{"type":"v1","addr":"192.168.123.102:6802","nonce":4063991565}]},"cluster_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.102:6804","nonce":4063991565},{"type":"v1","addr":"192.168.123.102:6806","nonce":4063991565}]},"heartbeat_back_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.102:6812","nonce":4063991565},{"type":"v1","addr":"192.168.123.102:6814","nonce":4063991565}]},"heartbeat_front_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.102:6808","nonce":4063991565},{"type":"v1","addr":"192.168.123.102:6810","nonce":4063991565}]},"public_addr":"192.168.123.102:6802/4063991565","cluster_addr":"192.168.123.102:6806/4063991565","heartbeat_back_addr":"192.168.123.102:6814/4063991565","heartbeat_front_addr":"192.168.123.102:6810/4063991565","state":["exists","up"]},{"osd":1,"uuid":"c0ab8e1d-5750-4187-a689-bd970c3d9755","up":1,"in":1,"weight":1,"primary_affinity":1,"last_clean_begin":0,"last_clean_end":0,"up_from":11,"up_thru":15,"down_at":0,"lost_at":0,"public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.102:6816","nonce":1097301766},{"type":"v1","addr":"192.168.123.102:6817","nonce":1097301766}]},"cluster_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.102:6818","nonce":1097301766},{"type":"v1","addr":"192.168.123.102:6819","nonce":1097301766}]},"heartbeat_back_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.102:6822","nonce":1097301766},{"type":"v1","addr":"192.168.123.102:6823","nonce":1097301766}]},"heartbeat_front_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.102:6820","nonce":1097301766},{"type":"v1","addr":"192.168.123.102:6821","nonce":1097301766}]},"public_addr":"192.168.123.102:6817/1097301766","cluster_addr":"192.168.123.102:6819/1097301766","heartbeat_back_addr":"192.168.123.102:6823/1097301766","heartbeat_front_addr":"192.168.123.102:6821/1097301766","state":["exists","up"]},{"osd":2,"uuid":"b585a68a-8110-4ac0-8c2f-7336eda5d34f","up":1,"in":1,"weight":1,"primary_affinity":1,"last_clean_begin":0,"last_clean_end":0,"up_from":11,"up_thru":15,"down_at":0,"lost_at":0,"public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.102:6801","nonce":2611319604},{"type":"v1","addr":"192.168.123.102:6803","nonce":2611319604}]},"cluster_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.102:6805","nonce":2611319604},{"type":"v1","addr":"192.168.123.102:6807","nonce":2611319604}]},"heartbeat_back_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.102:6813","nonce":2611319604},{"type":"v1","addr":"192.168.123.102:6815","nonce":2611319604}]},"heartbeat_front_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.102:6809","nonce":2611319604},{"type":"v1","addr":"192.168.123.102:6811","nonce":2611319604}]},"public_addr":"192.168.123.102:6803/2611319604","cluster_addr":"192.168.123.102:6807/2611319604","heartbeat_back_addr":"192.168.123.102:6815/2611319604","heartbeat_front_addr":"192.168.123.102:6811/2611319604","state":["exists","up"]},{"osd":3,"uuid":"bae3b9ef-03f0-49fb-916d-44717bb6debe","up":1,"in":1,"weight":1,"primary_affinity":1,"last_clean_begin":0,"last_clean_end":0,"up_from":11,"up_thru":15,"down_at":0,"lost_at":0,"public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.105:6816","nonce":573174972},{"type":"v1","addr":"192.168.123.105:6817","nonce":573174972}]},"cluster_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.105:6818","nonce":573174972},{"type":"v1","addr":"192.168.123.105:6819","nonce":573174972}]},"heartbeat_back_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.105:6822","nonce":573174972},{"type":"v1","addr":"192.168.123.105:6823","nonce":573174972}]},"heartbeat_front_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.105:6820","nonce":573174972},{"type":"v1","addr":"192.168.123.105:6821","nonce":573174972}]},"public_addr":"192.168.123.105:6817/573174972","cluster_addr":"192.168.123.105:6819/573174972","heartbeat_back_addr":"192.168.123.105:6823/573174972","heartbeat_front_addr":"192.168.123.105:6821/573174972","state":["exists","up"]},{"osd":4,"uuid":"e40a79a4-66fe-4b2a-9956-feed9d871626","up":1,"in":1,"weight":1,"primary_affinity":1,"last_clean_begin":0,"last_clean_end":0,"up_from":11,"up_thru":0,"down_at":0,"lost_at":0,"public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.105:6800","nonce":2795356309},{"type":"v1","addr":"192.168.123.105:6801","nonce":2795356309}]},"cluster_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.105:6802","nonce":2795356309},{"type":"v1","addr":"192.168.123.105:6803","nonce":2795356309}]},"heartbeat_back_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.105:6806","nonce":2795356309},{"type":"v1","addr":"192.168.123.105:6807","nonce":2795356309}]},"heartbeat_front_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.105:6804","nonce":2795356309},{"type":"v1","addr":"192.168.123.105:6805","nonce":2795356309}]},"public_addr":"192.168.123.105:6801/2795356309","cluster_addr":"192.168.123.105:6803/2795356309","heartbeat_back_addr":"192.168.123.105:6807/2795356309","heartbeat_front_addr":"192.168.123.105:6805/2795356309","state":["exists","up"]},{"osd":5,"uuid":"b4b92b50-8b59-4f2c-8dc8-23d74ca97904","up":1,"in":1,"weight":1,"primary_affinity":1,"last_clean_begin":0,"last_clean_end":0,"up_from":11,"up_thru":15,"down_at":0,"lost_at":0,"public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.105:6808","nonce":3989640680},{"type":"v1","addr":"192.168.123.105:6809","nonce":3989640680}]},"cluster_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.105:6810","nonce":3989640680},{"type":"v1","addr":"192.168.123.105:6811","nonce":3989640680}]},"heartbeat_back_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.105:6814","nonce":3989640680},{"type":"v1","addr":"192.168.123.105:6815","nonce":3989640680}]},"heartbeat_front_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.105:6812","nonce":3989640680},{"type":"v1","addr":"192.168.123.105:6813","nonce":3989640680}]},"public_addr":"192.168.123.105:6809/3989640680","cluster_addr":"192.168.123.105:6811/3989640680","heartbeat_back_addr":"192.168.123.105:6815/3989640680","heartbeat_front_addr":"192.168.123.105:6813/3989640680","state":["exists","up"]}],"osd_xinfo":[{"osd":0,"down_stamp":"0.000000","laggy_probability":0,"laggy_interval":0,"features":4544132024016699391,"old_weight":0,"last_purged_snaps_scrub":"2026-03-20T17:57:15.637775+0000","dead_epoch":0},{"osd":1,"down_stamp":"0.000000","laggy_probability":0,"laggy_interval":0,"features":4544132024016699391,"old_weight":0,"last_purged_snaps_scrub":"2026-03-20T17:57:15.678150+0000","dead_epoch":0},{"osd":2,"down_stamp":"0.000000","laggy_probability":0,"laggy_interval":0,"features":4544132024016699391,"old_weight":0,"last_purged_snaps_scrub":"2026-03-20T17:57:15.737220+0000","dead_epoch":0},{"osd":3,"down_stamp":"0.000000","laggy_probability":0,"laggy_interval":0,"features":4544132024016699391,"old_weight":0,"last_purged_snaps_scrub":"0.000000","dead_epoch":0},{"osd":4,"down_stamp":"0.000000","laggy_probability":0,"laggy_interval":0,"features":4544132024016699391,"old_weight":0,"last_purged_snaps_scrub":"0.000000","dead_epoch":0},{"osd":5,"down_stamp":"0.000000","laggy_probability":0,"laggy_interval":0,"features":4544132024016699391,"old_weight":0,"last_purged_snaps_scrub":"0.000000","dead_epoch":0}],"pg_upmap":[],"pg_upmap_items":[],"pg_upmap_primaries":[],"pg_temp":[],"primary_temp":[],"blocklist":{},"range_blocklist":{},"erasure_code_profiles":{"default":{"crush-failure-domain":"osd","k":"2","m":"1","plugin":"isa","technique":"reed_sol_van"}},"removed_snaps_queue":[{"pool":2,"snaps":[{"begin":2,"length":1}]}],"new_removed_snaps":[{"pool":2,"snaps":[{"begin":2,"length":1}]}],"new_purged_snaps":[],"crush_node_flags":{},"device_class_flags":{},"stretch_mode":{"stretch_mode_enabled":false,"stretch_bucket_count":0,"degraded_stretch_mode":0,"recovering_stretch_mode":0,"stretch_mode_bucket":0}} 2026-03-20T17:57:25.842 INFO:tasks.ceph.ceph_manager.c1:all up! 2026-03-20T17:57:25.842 DEBUG:teuthology.orchestra.run.vm02:> sudo adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage timeout 120 ceph --cluster c1 osd dump --format=json 2026-03-20T17:57:26.000 INFO:teuthology.orchestra.run.vm02.stdout: 2026-03-20T17:57:26.000 INFO:teuthology.orchestra.run.vm02.stdout:{"epoch":18,"fsid":"67b0dd48-80db-4ea8-b708-f5136e717957","created":"2026-03-20T17:57:12.845954+0000","modified":"2026-03-20T17:57:25.243326+0000","last_up_change":"2026-03-20T17:57:16.988698+0000","last_in_change":"2026-03-20T17:57:13.980717+0000","flags":"sortbitwise,recovery_deletes,purged_snapdirs,pglog_hardlimit","flags_num":5799936,"flags_set":["pglog_hardlimit","purged_snapdirs","recovery_deletes","sortbitwise"],"crush_version":4,"full_ratio":0.94999998807907104,"backfillfull_ratio":0.89999997615814209,"nearfull_ratio":0.85000002384185791,"cluster_snapshot":"","pool_max":2,"max_osd":6,"require_min_compat_client":"luminous","min_compat_client":"jewel","require_osd_release":"tentacle","allow_crimson":false,"pools":[{"pool":1,"pool_name":".mgr","create_time":"2026-03-20T17:57:17.182191+0000","flags":1,"flags_names":"hashpspool","type":1,"size":2,"min_size":1,"crush_rule":0,"peering_crush_bucket_count":0,"peering_crush_bucket_target":0,"peering_crush_bucket_barrier":0,"peering_crush_bucket_mandatory_member":2147483647,"is_stretch_pool":false,"object_hash":2,"pg_autoscale_mode":"off","pg_num":1,"pg_placement_num":1,"pg_placement_num_target":1,"pg_num_target":1,"pg_num_pending":1,"last_pg_merge_meta":{"source_pgid":"0.0","ready_epoch":0,"last_epoch_started":0,"last_epoch_clean":0,"source_version":"0'0","target_version":"0'0"},"last_change":"14","last_force_op_resend":"0","last_force_op_resend_prenautilus":"0","last_force_op_resend_preluminous":"0","auid":0,"snap_mode":"selfmanaged","snap_seq":0,"snap_epoch":0,"pool_snaps":[],"removed_snaps":"[]","quota_max_bytes":0,"quota_max_objects":0,"tiers":[],"tier_of":-1,"read_tier":-1,"write_tier":-1,"cache_mode":"none","target_max_bytes":0,"target_max_objects":0,"cache_target_dirty_ratio_micro":400000,"cache_target_dirty_high_ratio_micro":600000,"cache_target_full_ratio_micro":800000,"cache_min_flush_age":0,"cache_min_evict_age":0,"erasure_code_profile":"","hit_set_params":{"type":"none"},"hit_set_period":0,"hit_set_count":0,"use_gmt_hitset":true,"min_read_recency_for_promote":0,"min_write_recency_for_promote":0,"hit_set_grade_decay_rate":0,"hit_set_search_last_n":0,"grade_table":[],"stripe_width":0,"expected_num_objects":0,"fast_read":false,"nonprimary_shards":"{}","options":{"pg_num_max":32,"pg_num_min":1},"application_metadata":{"mgr":{}},"read_balance":{"score_type":"Fair distribution","score_acting":6.059999942779541,"score_stable":6.059999942779541,"optimal_score":0.33000001311302185,"raw_score_acting":2,"raw_score_stable":2,"primary_affinity_weighted":1,"average_primary_affinity":1,"average_primary_affinity_weighted":1}},{"pool":2,"pool_name":"rbd","create_time":"2026-03-20T17:57:21.776478+0000","flags":8193,"flags_names":"hashpspool,selfmanaged_snaps","type":1,"size":2,"min_size":1,"crush_rule":0,"peering_crush_bucket_count":0,"peering_crush_bucket_target":0,"peering_crush_bucket_barrier":0,"peering_crush_bucket_mandatory_member":2147483647,"is_stretch_pool":false,"object_hash":2,"pg_autoscale_mode":"off","pg_num":8,"pg_placement_num":8,"pg_placement_num_target":8,"pg_num_target":8,"pg_num_pending":8,"last_pg_merge_meta":{"source_pgid":"0.0","ready_epoch":0,"last_epoch_started":0,"last_epoch_clean":0,"source_version":"0'0","target_version":"0'0"},"last_change":"18","last_force_op_resend":"0","last_force_op_resend_prenautilus":"0","last_force_op_resend_preluminous":"0","auid":0,"snap_mode":"selfmanaged","snap_seq":2,"snap_epoch":18,"pool_snaps":[],"removed_snaps":"[]","quota_max_bytes":0,"quota_max_objects":0,"tiers":[],"tier_of":-1,"read_tier":-1,"write_tier":-1,"cache_mode":"none","target_max_bytes":0,"target_max_objects":0,"cache_target_dirty_ratio_micro":400000,"cache_target_dirty_high_ratio_micro":600000,"cache_target_full_ratio_micro":800000,"cache_min_flush_age":0,"cache_min_evict_age":0,"erasure_code_profile":"","hit_set_params":{"type":"none"},"hit_set_period":0,"hit_set_count":0,"use_gmt_hitset":true,"min_read_recency_for_promote":0,"min_write_recency_for_promote":0,"hit_set_grade_decay_rate":0,"hit_set_search_last_n":0,"grade_table":[],"stripe_width":0,"expected_num_objects":0,"fast_read":false,"nonprimary_shards":"{}","options":{},"application_metadata":{"rbd":{}},"read_balance":{"score_type":"Fair distribution","score_acting":2.25,"score_stable":2.25,"optimal_score":1,"raw_score_acting":2.25,"raw_score_stable":2.25,"primary_affinity_weighted":1,"average_primary_affinity":1,"average_primary_affinity_weighted":1}}],"osds":[{"osd":0,"uuid":"d6bdc1d9-2e5a-421c-82d0-a5a63e4cce9f","up":1,"in":1,"weight":1,"primary_affinity":1,"last_clean_begin":0,"last_clean_end":0,"up_from":11,"up_thru":0,"down_at":0,"lost_at":0,"public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.102:6800","nonce":4063991565},{"type":"v1","addr":"192.168.123.102:6802","nonce":4063991565}]},"cluster_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.102:6804","nonce":4063991565},{"type":"v1","addr":"192.168.123.102:6806","nonce":4063991565}]},"heartbeat_back_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.102:6812","nonce":4063991565},{"type":"v1","addr":"192.168.123.102:6814","nonce":4063991565}]},"heartbeat_front_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.102:6808","nonce":4063991565},{"type":"v1","addr":"192.168.123.102:6810","nonce":4063991565}]},"public_addr":"192.168.123.102:6802/4063991565","cluster_addr":"192.168.123.102:6806/4063991565","heartbeat_back_addr":"192.168.123.102:6814/4063991565","heartbeat_front_addr":"192.168.123.102:6810/4063991565","state":["exists","up"]},{"osd":1,"uuid":"c0ab8e1d-5750-4187-a689-bd970c3d9755","up":1,"in":1,"weight":1,"primary_affinity":1,"last_clean_begin":0,"last_clean_end":0,"up_from":11,"up_thru":15,"down_at":0,"lost_at":0,"public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.102:6816","nonce":1097301766},{"type":"v1","addr":"192.168.123.102:6817","nonce":1097301766}]},"cluster_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.102:6818","nonce":1097301766},{"type":"v1","addr":"192.168.123.102:6819","nonce":1097301766}]},"heartbeat_back_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.102:6822","nonce":1097301766},{"type":"v1","addr":"192.168.123.102:6823","nonce":1097301766}]},"heartbeat_front_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.102:6820","nonce":1097301766},{"type":"v1","addr":"192.168.123.102:6821","nonce":1097301766}]},"public_addr":"192.168.123.102:6817/1097301766","cluster_addr":"192.168.123.102:6819/1097301766","heartbeat_back_addr":"192.168.123.102:6823/1097301766","heartbeat_front_addr":"192.168.123.102:6821/1097301766","state":["exists","up"]},{"osd":2,"uuid":"b585a68a-8110-4ac0-8c2f-7336eda5d34f","up":1,"in":1,"weight":1,"primary_affinity":1,"last_clean_begin":0,"last_clean_end":0,"up_from":11,"up_thru":15,"down_at":0,"lost_at":0,"public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.102:6801","nonce":2611319604},{"type":"v1","addr":"192.168.123.102:6803","nonce":2611319604}]},"cluster_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.102:6805","nonce":2611319604},{"type":"v1","addr":"192.168.123.102:6807","nonce":2611319604}]},"heartbeat_back_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.102:6813","nonce":2611319604},{"type":"v1","addr":"192.168.123.102:6815","nonce":2611319604}]},"heartbeat_front_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.102:6809","nonce":2611319604},{"type":"v1","addr":"192.168.123.102:6811","nonce":2611319604}]},"public_addr":"192.168.123.102:6803/2611319604","cluster_addr":"192.168.123.102:6807/2611319604","heartbeat_back_addr":"192.168.123.102:6815/2611319604","heartbeat_front_addr":"192.168.123.102:6811/2611319604","state":["exists","up"]},{"osd":3,"uuid":"bae3b9ef-03f0-49fb-916d-44717bb6debe","up":1,"in":1,"weight":1,"primary_affinity":1,"last_clean_begin":0,"last_clean_end":0,"up_from":11,"up_thru":15,"down_at":0,"lost_at":0,"public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.105:6816","nonce":573174972},{"type":"v1","addr":"192.168.123.105:6817","nonce":573174972}]},"cluster_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.105:6818","nonce":573174972},{"type":"v1","addr":"192.168.123.105:6819","nonce":573174972}]},"heartbeat_back_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.105:6822","nonce":573174972},{"type":"v1","addr":"192.168.123.105:6823","nonce":573174972}]},"heartbeat_front_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.105:6820","nonce":573174972},{"type":"v1","addr":"192.168.123.105:6821","nonce":573174972}]},"public_addr":"192.168.123.105:6817/573174972","cluster_addr":"192.168.123.105:6819/573174972","heartbeat_back_addr":"192.168.123.105:6823/573174972","heartbeat_front_addr":"192.168.123.105:6821/573174972","state":["exists","up"]},{"osd":4,"uuid":"e40a79a4-66fe-4b2a-9956-feed9d871626","up":1,"in":1,"weight":1,"primary_affinity":1,"last_clean_begin":0,"last_clean_end":0,"up_from":11,"up_thru":0,"down_at":0,"lost_at":0,"public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.105:6800","nonce":2795356309},{"type":"v1","addr":"192.168.123.105:6801","nonce":2795356309}]},"cluster_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.105:6802","nonce":2795356309},{"type":"v1","addr":"192.168.123.105:6803","nonce":2795356309}]},"heartbeat_back_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.105:6806","nonce":2795356309},{"type":"v1","addr":"192.168.123.105:6807","nonce":2795356309}]},"heartbeat_front_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.105:6804","nonce":2795356309},{"type":"v1","addr":"192.168.123.105:6805","nonce":2795356309}]},"public_addr":"192.168.123.105:6801/2795356309","cluster_addr":"192.168.123.105:6803/2795356309","heartbeat_back_addr":"192.168.123.105:6807/2795356309","heartbeat_front_addr":"192.168.123.105:6805/2795356309","state":["exists","up"]},{"osd":5,"uuid":"b4b92b50-8b59-4f2c-8dc8-23d74ca97904","up":1,"in":1,"weight":1,"primary_affinity":1,"last_clean_begin":0,"last_clean_end":0,"up_from":11,"up_thru":15,"down_at":0,"lost_at":0,"public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.105:6808","nonce":3989640680},{"type":"v1","addr":"192.168.123.105:6809","nonce":3989640680}]},"cluster_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.105:6810","nonce":3989640680},{"type":"v1","addr":"192.168.123.105:6811","nonce":3989640680}]},"heartbeat_back_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.105:6814","nonce":3989640680},{"type":"v1","addr":"192.168.123.105:6815","nonce":3989640680}]},"heartbeat_front_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.105:6812","nonce":3989640680},{"type":"v1","addr":"192.168.123.105:6813","nonce":3989640680}]},"public_addr":"192.168.123.105:6809/3989640680","cluster_addr":"192.168.123.105:6811/3989640680","heartbeat_back_addr":"192.168.123.105:6815/3989640680","heartbeat_front_addr":"192.168.123.105:6813/3989640680","state":["exists","up"]}],"osd_xinfo":[{"osd":0,"down_stamp":"0.000000","laggy_probability":0,"laggy_interval":0,"features":4544132024016699391,"old_weight":0,"last_purged_snaps_scrub":"2026-03-20T17:57:15.637775+0000","dead_epoch":0},{"osd":1,"down_stamp":"0.000000","laggy_probability":0,"laggy_interval":0,"features":4544132024016699391,"old_weight":0,"last_purged_snaps_scrub":"2026-03-20T17:57:15.678150+0000","dead_epoch":0},{"osd":2,"down_stamp":"0.000000","laggy_probability":0,"laggy_interval":0,"features":4544132024016699391,"old_weight":0,"last_purged_snaps_scrub":"2026-03-20T17:57:15.737220+0000","dead_epoch":0},{"osd":3,"down_stamp":"0.000000","laggy_probability":0,"laggy_interval":0,"features":4544132024016699391,"old_weight":0,"last_purged_snaps_scrub":"0.000000","dead_epoch":0},{"osd":4,"down_stamp":"0.000000","laggy_probability":0,"laggy_interval":0,"features":4544132024016699391,"old_weight":0,"last_purged_snaps_scrub":"0.000000","dead_epoch":0},{"osd":5,"down_stamp":"0.000000","laggy_probability":0,"laggy_interval":0,"features":4544132024016699391,"old_weight":0,"last_purged_snaps_scrub":"0.000000","dead_epoch":0}],"pg_upmap":[],"pg_upmap_items":[],"pg_upmap_primaries":[],"pg_temp":[],"primary_temp":[],"blocklist":{},"range_blocklist":{},"erasure_code_profiles":{"default":{"crush-failure-domain":"osd","k":"2","m":"1","plugin":"isa","technique":"reed_sol_van"}},"removed_snaps_queue":[{"pool":2,"snaps":[{"begin":2,"length":1}]}],"new_removed_snaps":[{"pool":2,"snaps":[{"begin":2,"length":1}]}],"new_purged_snaps":[],"crush_node_flags":{},"device_class_flags":{},"stretch_mode":{"stretch_mode_enabled":false,"stretch_bucket_count":0,"degraded_stretch_mode":0,"recovering_stretch_mode":0,"stretch_mode_bucket":0}} 2026-03-20T17:57:26.013 DEBUG:teuthology.orchestra.run.vm02:> sudo adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage timeout 120 ceph --cluster c1 tell osd.0 flush_pg_stats 2026-03-20T17:57:26.013 DEBUG:teuthology.orchestra.run.vm02:> sudo adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage timeout 120 ceph --cluster c1 tell osd.1 flush_pg_stats 2026-03-20T17:57:26.013 DEBUG:teuthology.orchestra.run.vm02:> sudo adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage timeout 120 ceph --cluster c1 tell osd.2 flush_pg_stats 2026-03-20T17:57:26.013 DEBUG:teuthology.orchestra.run.vm02:> sudo adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage timeout 120 ceph --cluster c1 tell osd.3 flush_pg_stats 2026-03-20T17:57:26.014 DEBUG:teuthology.orchestra.run.vm02:> sudo adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage timeout 120 ceph --cluster c1 tell osd.4 flush_pg_stats 2026-03-20T17:57:26.014 DEBUG:teuthology.orchestra.run.vm02:> sudo adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage timeout 120 ceph --cluster c1 tell osd.5 flush_pg_stats 2026-03-20T17:57:26.135 INFO:teuthology.orchestra.run.vm02.stdout:47244640259 2026-03-20T17:57:26.135 DEBUG:teuthology.orchestra.run.vm02:> sudo adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage timeout 120 ceph --cluster c1 osd last-stat-seq osd.0 2026-03-20T17:57:26.155 INFO:teuthology.orchestra.run.vm02.stdout:47244640259 2026-03-20T17:57:26.155 DEBUG:teuthology.orchestra.run.vm02:> sudo adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage timeout 120 ceph --cluster c1 osd last-stat-seq osd.3 2026-03-20T17:57:26.170 INFO:teuthology.orchestra.run.vm02.stdout:47244640259 2026-03-20T17:57:26.170 DEBUG:teuthology.orchestra.run.vm02:> sudo adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage timeout 120 ceph --cluster c1 osd last-stat-seq osd.4 2026-03-20T17:57:26.173 INFO:teuthology.orchestra.run.vm02.stdout:47244640259 2026-03-20T17:57:26.173 DEBUG:teuthology.orchestra.run.vm02:> sudo adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage timeout 120 ceph --cluster c1 osd last-stat-seq osd.1 2026-03-20T17:57:26.177 INFO:teuthology.orchestra.run.vm02.stdout:47244640259 2026-03-20T17:57:26.177 DEBUG:teuthology.orchestra.run.vm02:> sudo adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage timeout 120 ceph --cluster c1 osd last-stat-seq osd.5 2026-03-20T17:57:26.229 INFO:teuthology.orchestra.run.vm02.stdout:47244640259 2026-03-20T17:57:26.230 DEBUG:teuthology.orchestra.run.vm02:> sudo adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage timeout 120 ceph --cluster c1 osd last-stat-seq osd.2 2026-03-20T17:57:26.377 INFO:teuthology.orchestra.run.vm02.stdout:47244640258 2026-03-20T17:57:26.396 INFO:tasks.ceph.ceph_manager.c1:need seq 47244640259 got 47244640258 for osd.3 2026-03-20T17:57:26.416 INFO:teuthology.orchestra.run.vm02.stdout:47244640258 2026-03-20T17:57:26.445 INFO:teuthology.orchestra.run.vm02.stdout:47244640258 2026-03-20T17:57:26.447 INFO:teuthology.orchestra.run.vm02.stdout:47244640258 2026-03-20T17:57:26.449 INFO:tasks.ceph.ceph_manager.c1:need seq 47244640259 got 47244640258 for osd.4 2026-03-20T17:57:26.467 INFO:tasks.ceph.ceph_manager.c1:need seq 47244640259 got 47244640258 for osd.0 2026-03-20T17:57:26.469 INFO:tasks.ceph.ceph_manager.c1:need seq 47244640259 got 47244640258 for osd.1 2026-03-20T17:57:26.492 INFO:teuthology.orchestra.run.vm02.stdout:47244640258 2026-03-20T17:57:26.501 INFO:teuthology.orchestra.run.vm02.stdout:47244640258 2026-03-20T17:57:26.506 INFO:tasks.ceph.ceph_manager.c1:need seq 47244640259 got 47244640258 for osd.2 2026-03-20T17:57:26.517 INFO:tasks.ceph.ceph_manager.c1:need seq 47244640259 got 47244640258 for osd.5 2026-03-20T17:57:27.397 DEBUG:teuthology.orchestra.run.vm02:> sudo adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage timeout 120 ceph --cluster c1 osd last-stat-seq osd.3 2026-03-20T17:57:27.450 DEBUG:teuthology.orchestra.run.vm02:> sudo adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage timeout 120 ceph --cluster c1 osd last-stat-seq osd.4 2026-03-20T17:57:27.468 DEBUG:teuthology.orchestra.run.vm02:> sudo adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage timeout 120 ceph --cluster c1 osd last-stat-seq osd.0 2026-03-20T17:57:27.470 DEBUG:teuthology.orchestra.run.vm02:> sudo adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage timeout 120 ceph --cluster c1 osd last-stat-seq osd.1 2026-03-20T17:57:27.508 DEBUG:teuthology.orchestra.run.vm02:> sudo adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage timeout 120 ceph --cluster c1 osd last-stat-seq osd.2 2026-03-20T17:57:27.518 DEBUG:teuthology.orchestra.run.vm02:> sudo adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage timeout 120 ceph --cluster c1 osd last-stat-seq osd.5 2026-03-20T17:57:27.583 INFO:teuthology.orchestra.run.vm02.stdout:47244640259 2026-03-20T17:57:27.602 INFO:tasks.ceph.ceph_manager.c1:need seq 47244640259 got 47244640259 for osd.3 2026-03-20T17:57:27.602 DEBUG:teuthology.parallel:result is None 2026-03-20T17:57:27.636 INFO:teuthology.orchestra.run.vm02.stdout:47244640259 2026-03-20T17:57:27.651 INFO:tasks.ceph.ceph_manager.c1:need seq 47244640259 got 47244640259 for osd.4 2026-03-20T17:57:27.651 DEBUG:teuthology.parallel:result is None 2026-03-20T17:57:27.731 INFO:teuthology.orchestra.run.vm02.stdout:47244640259 2026-03-20T17:57:27.747 INFO:tasks.ceph.ceph_manager.c1:need seq 47244640259 got 47244640259 for osd.0 2026-03-20T17:57:27.747 INFO:teuthology.orchestra.run.vm02.stdout:47244640259 2026-03-20T17:57:27.747 DEBUG:teuthology.parallel:result is None 2026-03-20T17:57:27.754 INFO:teuthology.orchestra.run.vm02.stdout:47244640259 2026-03-20T17:57:27.764 INFO:tasks.ceph.ceph_manager.c1:need seq 47244640259 got 47244640259 for osd.1 2026-03-20T17:57:27.764 DEBUG:teuthology.parallel:result is None 2026-03-20T17:57:27.770 INFO:tasks.ceph.ceph_manager.c1:need seq 47244640259 got 47244640259 for osd.2 2026-03-20T17:57:27.770 DEBUG:teuthology.parallel:result is None 2026-03-20T17:57:27.790 INFO:teuthology.orchestra.run.vm02.stdout:47244640259 2026-03-20T17:57:27.805 INFO:tasks.ceph.ceph_manager.c1:need seq 47244640259 got 47244640259 for osd.5 2026-03-20T17:57:27.805 DEBUG:teuthology.parallel:result is None 2026-03-20T17:57:27.805 INFO:tasks.ceph.ceph_manager.c1:waiting for clean 2026-03-20T17:57:27.805 DEBUG:teuthology.orchestra.run.vm02:> sudo adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage timeout 120 ceph --cluster c1 pg dump --format=json 2026-03-20T17:57:28.005 INFO:teuthology.orchestra.run.vm02.stdout: 2026-03-20T17:57:28.006 INFO:teuthology.orchestra.run.vm02.stderr:dumped all 2026-03-20T17:57:28.018 INFO:teuthology.orchestra.run.vm02.stdout:{"pg_ready":true,"pg_map":{"version":18,"stamp":"2026-03-20T17:57:27.843751+0000","last_osdmap_epoch":0,"last_pg_scan":0,"pg_stats_sum":{"stat_sum":{"num_bytes":459299,"num_objects":4,"num_object_clones":0,"num_object_copies":8,"num_objects_missing_on_primary":0,"num_objects_missing":0,"num_objects_degraded":0,"num_objects_misplaced":0,"num_objects_unfound":0,"num_objects_dirty":4,"num_whiteouts":0,"num_read":46,"num_read_kb":37,"num_write":59,"num_write_kb":586,"num_scrub_errors":0,"num_shallow_scrub_errors":0,"num_deep_scrub_errors":0,"num_objects_recovered":0,"num_bytes_recovered":0,"num_keys_recovered":0,"num_objects_omap":0,"num_objects_hit_set_archive":0,"num_bytes_hit_set_archive":0,"num_flush":0,"num_flush_kb":0,"num_evict":0,"num_evict_kb":0,"num_promote":0,"num_flush_mode_high":0,"num_flush_mode_low":0,"num_evict_mode_some":0,"num_evict_mode_full":0,"num_objects_pinned":0,"num_legacy_snapsets":0,"num_large_omap_objects":0,"num_objects_manifest":0,"num_omap_bytes":0,"num_omap_keys":0,"num_objects_repaired":0},"store_stats":{"total":0,"available":0,"internally_reserved":0,"allocated":0,"data_stored":0,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":0},"log_size":35,"ondisk_log_size":35,"up":18,"acting":18,"num_store_stats":0},"osd_stats_sum":{"up_from":0,"seq":0,"num_pgs":17,"num_osds":6,"num_per_pool_osds":6,"num_per_pool_omap_osds":6,"kb":629145600,"kb_used":163016,"kb_used_data":1920,"kb_used_omap":50,"kb_used_meta":160845,"kb_avail":628982584,"statfs":{"total":644245094400,"available":644078166016,"internally_reserved":0,"allocated":1966080,"data_stored":1256986,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":51973,"internal_metadata":164705531},"hb_peers":[],"snap_trim_queue_len":0,"num_snap_trimming":0,"num_shards_repaired":0,"op_queue_age_hist":{"histogram":[],"upper_bound":1},"perf_stat":{"commit_latency_ms":0,"apply_latency_ms":0,"commit_latency_ns":0,"apply_latency_ns":0},"alerts":[],"network_ping_times":[]},"pg_stats_delta":{"stat_sum":{"num_bytes":19,"num_objects":2,"num_object_clones":0,"num_object_copies":4,"num_objects_missing_on_primary":0,"num_objects_missing":0,"num_objects_degraded":0,"num_objects_misplaced":0,"num_objects_unfound":0,"num_objects_dirty":2,"num_whiteouts":0,"num_read":0,"num_read_kb":0,"num_write":2,"num_write_kb":2,"num_scrub_errors":0,"num_shallow_scrub_errors":0,"num_deep_scrub_errors":0,"num_objects_recovered":0,"num_bytes_recovered":0,"num_keys_recovered":0,"num_objects_omap":0,"num_objects_hit_set_archive":0,"num_bytes_hit_set_archive":0,"num_flush":0,"num_flush_kb":0,"num_evict":0,"num_evict_kb":0,"num_promote":0,"num_flush_mode_high":0,"num_flush_mode_low":0,"num_evict_mode_some":0,"num_evict_mode_full":0,"num_objects_pinned":0,"num_legacy_snapsets":0,"num_large_omap_objects":0,"num_objects_manifest":0,"num_omap_bytes":0,"num_omap_keys":0,"num_objects_repaired":0},"store_stats":{"total":0,"available":0,"internally_reserved":0,"allocated":0,"data_stored":0,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":0},"log_size":0,"ondisk_log_size":0,"up":0,"acting":0,"num_store_stats":0,"stamp_delta":"4.674425"},"pg_stats":[{"pgid":"2.7","version":"0'0","reported_seq":20,"reported_epoch":18,"state":"active+clean","last_fresh":"2026-03-20T17:57:25.678438+0000","last_change":"2026-03-20T17:57:25.678438+0000","last_active":"2026-03-20T17:57:25.678438+0000","last_peered":"2026-03-20T17:57:25.678438+0000","last_clean":"2026-03-20T17:57:25.678438+0000","last_became_active":"2026-03-20T17:57:23.677052+0000","last_became_peered":"2026-03-20T17:57:23.677052+0000","last_unstale":"2026-03-20T17:57:25.678438+0000","last_undegraded":"2026-03-20T17:57:25.678438+0000","last_fullsized":"2026-03-20T17:57:25.678438+0000","mapping_epoch":15,"log_start":"0'0","ondisk_log_start":"0'0","created":15,"last_epoch_clean":16,"parent":"0.0","parent_split_bits":0,"last_scrub":"0'0","last_scrub_stamp":"2026-03-20T17:57:22.189850+0000","last_deep_scrub":"0'0","last_deep_scrub_stamp":"2026-03-20T17:57:22.189850+0000","last_clean_scrub_stamp":"2026-03-20T17:57:22.189850+0000","objects_scrubbed":0,"log_size":0,"log_dups_size":0,"ondisk_log_size":0,"stats_invalid":false,"dirty_stats_invalid":false,"omap_stats_invalid":false,"hitset_stats_invalid":false,"hitset_bytes_stats_invalid":false,"pin_stats_invalid":false,"manifest_stats_invalid":false,"snaptrimq_len":0,"last_scrub_duration":0,"scrub_schedule":"periodic scrub scheduled @ 2026-03-22T01:42:28.030112+0000","scrub_duration":0,"objects_trimmed":0,"snaptrim_duration":0.00024659200000000001,"stat_sum":{"num_bytes":0,"num_objects":0,"num_object_clones":0,"num_object_copies":0,"num_objects_missing_on_primary":0,"num_objects_missing":0,"num_objects_degraded":0,"num_objects_misplaced":0,"num_objects_unfound":0,"num_objects_dirty":0,"num_whiteouts":0,"num_read":0,"num_read_kb":0,"num_write":0,"num_write_kb":0,"num_scrub_errors":0,"num_shallow_scrub_errors":0,"num_deep_scrub_errors":0,"num_objects_recovered":0,"num_bytes_recovered":0,"num_keys_recovered":0,"num_objects_omap":0,"num_objects_hit_set_archive":0,"num_bytes_hit_set_archive":0,"num_flush":0,"num_flush_kb":0,"num_evict":0,"num_evict_kb":0,"num_promote":0,"num_flush_mode_high":0,"num_flush_mode_low":0,"num_evict_mode_some":0,"num_evict_mode_full":0,"num_objects_pinned":0,"num_legacy_snapsets":0,"num_large_omap_objects":0,"num_objects_manifest":0,"num_omap_bytes":0,"num_omap_keys":0,"num_objects_repaired":0},"up":[3,4],"acting":[3,4],"avail_no_missing":[],"object_location_counts":[],"blocked_by":[],"up_primary":3,"acting_primary":3,"purged_snaps":[{"start":"2","length":"1"}]},{"pgid":"2.6","version":"0'0","reported_seq":20,"reported_epoch":18,"state":"active+clean","last_fresh":"2026-03-20T17:57:25.252369+0000","last_change":"2026-03-20T17:57:25.252369+0000","last_active":"2026-03-20T17:57:25.252369+0000","last_peered":"2026-03-20T17:57:25.252369+0000","last_clean":"2026-03-20T17:57:25.252369+0000","last_became_active":"2026-03-20T17:57:23.203017+0000","last_became_peered":"2026-03-20T17:57:23.203017+0000","last_unstale":"2026-03-20T17:57:25.252369+0000","last_undegraded":"2026-03-20T17:57:25.252369+0000","last_fullsized":"2026-03-20T17:57:25.252369+0000","mapping_epoch":15,"log_start":"0'0","ondisk_log_start":"0'0","created":15,"last_epoch_clean":16,"parent":"0.0","parent_split_bits":0,"last_scrub":"0'0","last_scrub_stamp":"2026-03-20T17:57:22.189850+0000","last_deep_scrub":"0'0","last_deep_scrub_stamp":"2026-03-20T17:57:22.189850+0000","last_clean_scrub_stamp":"2026-03-20T17:57:22.189850+0000","objects_scrubbed":0,"log_size":0,"log_dups_size":0,"ondisk_log_size":0,"stats_invalid":false,"dirty_stats_invalid":false,"omap_stats_invalid":false,"hitset_stats_invalid":false,"hitset_bytes_stats_invalid":false,"pin_stats_invalid":false,"manifest_stats_invalid":false,"snaptrimq_len":0,"last_scrub_duration":0,"scrub_schedule":"periodic scrub scheduled @ 2026-03-21T18:16:06.167963+0000","scrub_duration":0,"objects_trimmed":0,"snaptrim_duration":0.00024865500000000002,"stat_sum":{"num_bytes":0,"num_objects":0,"num_object_clones":0,"num_object_copies":0,"num_objects_missing_on_primary":0,"num_objects_missing":0,"num_objects_degraded":0,"num_objects_misplaced":0,"num_objects_unfound":0,"num_objects_dirty":0,"num_whiteouts":0,"num_read":0,"num_read_kb":0,"num_write":0,"num_write_kb":0,"num_scrub_errors":0,"num_shallow_scrub_errors":0,"num_deep_scrub_errors":0,"num_objects_recovered":0,"num_bytes_recovered":0,"num_keys_recovered":0,"num_objects_omap":0,"num_objects_hit_set_archive":0,"num_bytes_hit_set_archive":0,"num_flush":0,"num_flush_kb":0,"num_evict":0,"num_evict_kb":0,"num_promote":0,"num_flush_mode_high":0,"num_flush_mode_low":0,"num_evict_mode_some":0,"num_evict_mode_full":0,"num_objects_pinned":0,"num_legacy_snapsets":0,"num_large_omap_objects":0,"num_objects_manifest":0,"num_omap_bytes":0,"num_omap_keys":0,"num_objects_repaired":0},"up":[1,3],"acting":[1,3],"avail_no_missing":[],"object_location_counts":[],"blocked_by":[],"up_primary":1,"acting_primary":1,"purged_snaps":[{"start":"2","length":"1"}]},{"pgid":"2.5","version":"0'0","reported_seq":20,"reported_epoch":18,"state":"active+clean","last_fresh":"2026-03-20T17:57:25.678482+0000","last_change":"2026-03-20T17:57:25.678482+0000","last_active":"2026-03-20T17:57:25.678482+0000","last_peered":"2026-03-20T17:57:25.678482+0000","last_clean":"2026-03-20T17:57:25.678482+0000","last_became_active":"2026-03-20T17:57:23.578680+0000","last_became_peered":"2026-03-20T17:57:23.578680+0000","last_unstale":"2026-03-20T17:57:25.678482+0000","last_undegraded":"2026-03-20T17:57:25.678482+0000","last_fullsized":"2026-03-20T17:57:25.678482+0000","mapping_epoch":15,"log_start":"0'0","ondisk_log_start":"0'0","created":15,"last_epoch_clean":16,"parent":"0.0","parent_split_bits":0,"last_scrub":"0'0","last_scrub_stamp":"2026-03-20T17:57:22.189850+0000","last_deep_scrub":"0'0","last_deep_scrub_stamp":"2026-03-20T17:57:22.189850+0000","last_clean_scrub_stamp":"2026-03-20T17:57:22.189850+0000","objects_scrubbed":0,"log_size":0,"log_dups_size":0,"ondisk_log_size":0,"stats_invalid":false,"dirty_stats_invalid":false,"omap_stats_invalid":false,"hitset_stats_invalid":false,"hitset_bytes_stats_invalid":false,"pin_stats_invalid":false,"manifest_stats_invalid":false,"snaptrimq_len":0,"last_scrub_duration":0,"scrub_schedule":"periodic scrub scheduled @ 2026-03-22T04:40:49.859879+0000","scrub_duration":0,"objects_trimmed":0,"snaptrim_duration":0.00021413000000000001,"stat_sum":{"num_bytes":0,"num_objects":0,"num_object_clones":0,"num_object_copies":0,"num_objects_missing_on_primary":0,"num_objects_missing":0,"num_objects_degraded":0,"num_objects_misplaced":0,"num_objects_unfound":0,"num_objects_dirty":0,"num_whiteouts":0,"num_read":0,"num_read_kb":0,"num_write":0,"num_write_kb":0,"num_scrub_errors":0,"num_shallow_scrub_errors":0,"num_deep_scrub_errors":0,"num_objects_recovered":0,"num_bytes_recovered":0,"num_keys_recovered":0,"num_objects_omap":0,"num_objects_hit_set_archive":0,"num_bytes_hit_set_archive":0,"num_flush":0,"num_flush_kb":0,"num_evict":0,"num_evict_kb":0,"num_promote":0,"num_flush_mode_high":0,"num_flush_mode_low":0,"num_evict_mode_some":0,"num_evict_mode_full":0,"num_objects_pinned":0,"num_legacy_snapsets":0,"num_large_omap_objects":0,"num_objects_manifest":0,"num_omap_bytes":0,"num_omap_keys":0,"num_objects_repaired":0},"up":[3,0],"acting":[3,0],"avail_no_missing":[],"object_location_counts":[],"blocked_by":[],"up_primary":3,"acting_primary":3,"purged_snaps":[{"start":"2","length":"1"}]},{"pgid":"2.4","version":"0'0","reported_seq":20,"reported_epoch":18,"state":"active+clean","last_fresh":"2026-03-20T17:57:25.252428+0000","last_change":"2026-03-20T17:57:25.252428+0000","last_active":"2026-03-20T17:57:25.252428+0000","last_peered":"2026-03-20T17:57:25.252428+0000","last_clean":"2026-03-20T17:57:25.252428+0000","last_became_active":"2026-03-20T17:57:23.577676+0000","last_became_peered":"2026-03-20T17:57:23.577676+0000","last_unstale":"2026-03-20T17:57:25.252428+0000","last_undegraded":"2026-03-20T17:57:25.252428+0000","last_fullsized":"2026-03-20T17:57:25.252428+0000","mapping_epoch":15,"log_start":"0'0","ondisk_log_start":"0'0","created":15,"last_epoch_clean":16,"parent":"0.0","parent_split_bits":0,"last_scrub":"0'0","last_scrub_stamp":"2026-03-20T17:57:22.189850+0000","last_deep_scrub":"0'0","last_deep_scrub_stamp":"2026-03-20T17:57:22.189850+0000","last_clean_scrub_stamp":"2026-03-20T17:57:22.189850+0000","objects_scrubbed":0,"log_size":0,"log_dups_size":0,"ondisk_log_size":0,"stats_invalid":false,"dirty_stats_invalid":false,"omap_stats_invalid":false,"hitset_stats_invalid":false,"hitset_bytes_stats_invalid":false,"pin_stats_invalid":false,"manifest_stats_invalid":false,"snaptrimq_len":0,"last_scrub_duration":0,"scrub_schedule":"periodic scrub scheduled @ 2026-03-22T04:16:13.040208+0000","scrub_duration":0,"objects_trimmed":0,"snaptrim_duration":0.000238576,"stat_sum":{"num_bytes":0,"num_objects":0,"num_object_clones":0,"num_object_copies":0,"num_objects_missing_on_primary":0,"num_objects_missing":0,"num_objects_degraded":0,"num_objects_misplaced":0,"num_objects_unfound":0,"num_objects_dirty":0,"num_whiteouts":0,"num_read":0,"num_read_kb":0,"num_write":0,"num_write_kb":0,"num_scrub_errors":0,"num_shallow_scrub_errors":0,"num_deep_scrub_errors":0,"num_objects_recovered":0,"num_bytes_recovered":0,"num_keys_recovered":0,"num_objects_omap":0,"num_objects_hit_set_archive":0,"num_bytes_hit_set_archive":0,"num_flush":0,"num_flush_kb":0,"num_evict":0,"num_evict_kb":0,"num_promote":0,"num_flush_mode_high":0,"num_flush_mode_low":0,"num_evict_mode_some":0,"num_evict_mode_full":0,"num_objects_pinned":0,"num_legacy_snapsets":0,"num_large_omap_objects":0,"num_objects_manifest":0,"num_omap_bytes":0,"num_omap_keys":0,"num_objects_repaired":0},"up":[1,0],"acting":[1,0],"avail_no_missing":[],"object_location_counts":[],"blocked_by":[],"up_primary":1,"acting_primary":1,"purged_snaps":[{"start":"2","length":"1"}]},{"pgid":"2.2","version":"18'2","reported_seq":22,"reported_epoch":18,"state":"active+clean","last_fresh":"2026-03-20T17:57:25.252876+0000","last_change":"2026-03-20T17:57:25.251301+0000","last_active":"2026-03-20T17:57:25.252876+0000","last_peered":"2026-03-20T17:57:25.252876+0000","last_clean":"2026-03-20T17:57:25.252876+0000","last_became_active":"2026-03-20T17:57:23.203631+0000","last_became_peered":"2026-03-20T17:57:23.203631+0000","last_unstale":"2026-03-20T17:57:25.252876+0000","last_undegraded":"2026-03-20T17:57:25.252876+0000","last_fullsized":"2026-03-20T17:57:25.252876+0000","mapping_epoch":15,"log_start":"0'0","ondisk_log_start":"0'0","created":15,"last_epoch_clean":16,"parent":"0.0","parent_split_bits":0,"last_scrub":"0'0","last_scrub_stamp":"2026-03-20T17:57:22.189850+0000","last_deep_scrub":"0'0","last_deep_scrub_stamp":"2026-03-20T17:57:22.189850+0000","last_clean_scrub_stamp":"2026-03-20T17:57:22.189850+0000","objects_scrubbed":0,"log_size":2,"log_dups_size":0,"ondisk_log_size":2,"stats_invalid":false,"dirty_stats_invalid":false,"omap_stats_invalid":false,"hitset_stats_invalid":false,"hitset_bytes_stats_invalid":false,"pin_stats_invalid":false,"manifest_stats_invalid":false,"snaptrimq_len":0,"last_scrub_duration":0,"scrub_schedule":"periodic scrub scheduled @ 2026-03-21T23:45:13.592323+0000","scrub_duration":0,"objects_trimmed":0,"snaptrim_duration":0.00050427299999999996,"stat_sum":{"num_bytes":19,"num_objects":1,"num_object_clones":0,"num_object_copies":2,"num_objects_missing_on_primary":0,"num_objects_missing":0,"num_objects_degraded":0,"num_objects_misplaced":0,"num_objects_unfound":0,"num_objects_dirty":1,"num_whiteouts":0,"num_read":0,"num_read_kb":0,"num_write":2,"num_write_kb":2,"num_scrub_errors":0,"num_shallow_scrub_errors":0,"num_deep_scrub_errors":0,"num_objects_recovered":0,"num_bytes_recovered":0,"num_keys_recovered":0,"num_objects_omap":0,"num_objects_hit_set_archive":0,"num_bytes_hit_set_archive":0,"num_flush":0,"num_flush_kb":0,"num_evict":0,"num_evict_kb":0,"num_promote":0,"num_flush_mode_high":0,"num_flush_mode_low":0,"num_evict_mode_some":0,"num_evict_mode_full":0,"num_objects_pinned":0,"num_legacy_snapsets":0,"num_large_omap_objects":0,"num_objects_manifest":0,"num_omap_bytes":0,"num_omap_keys":0,"num_objects_repaired":0},"up":[5,1],"acting":[5,1],"avail_no_missing":[],"object_location_counts":[],"blocked_by":[],"up_primary":5,"acting_primary":5,"purged_snaps":[{"start":"2","length":"1"}]},{"pgid":"2.1","version":"0'0","reported_seq":20,"reported_epoch":18,"state":"active+clean","last_fresh":"2026-03-20T17:57:25.714907+0000","last_change":"2026-03-20T17:57:25.714907+0000","last_active":"2026-03-20T17:57:25.714907+0000","last_peered":"2026-03-20T17:57:25.714907+0000","last_clean":"2026-03-20T17:57:25.714907+0000","last_became_active":"2026-03-20T17:57:23.203291+0000","last_became_peered":"2026-03-20T17:57:23.203291+0000","last_unstale":"2026-03-20T17:57:25.714907+0000","last_undegraded":"2026-03-20T17:57:25.714907+0000","last_fullsized":"2026-03-20T17:57:25.714907+0000","mapping_epoch":15,"log_start":"0'0","ondisk_log_start":"0'0","created":15,"last_epoch_clean":16,"parent":"0.0","parent_split_bits":0,"last_scrub":"0'0","last_scrub_stamp":"2026-03-20T17:57:22.189850+0000","last_deep_scrub":"0'0","last_deep_scrub_stamp":"2026-03-20T17:57:22.189850+0000","last_clean_scrub_stamp":"2026-03-20T17:57:22.189850+0000","objects_scrubbed":0,"log_size":0,"log_dups_size":0,"ondisk_log_size":0,"stats_invalid":false,"dirty_stats_invalid":false,"omap_stats_invalid":false,"hitset_stats_invalid":false,"hitset_bytes_stats_invalid":false,"pin_stats_invalid":false,"manifest_stats_invalid":false,"snaptrimq_len":0,"last_scrub_duration":0,"scrub_schedule":"periodic scrub scheduled @ 2026-03-21T18:37:05.593601+0000","scrub_duration":0,"objects_trimmed":0,"snaptrim_duration":0.000234458,"stat_sum":{"num_bytes":0,"num_objects":0,"num_object_clones":0,"num_object_copies":0,"num_objects_missing_on_primary":0,"num_objects_missing":0,"num_objects_degraded":0,"num_objects_misplaced":0,"num_objects_unfound":0,"num_objects_dirty":0,"num_whiteouts":0,"num_read":0,"num_read_kb":0,"num_write":0,"num_write_kb":0,"num_scrub_errors":0,"num_shallow_scrub_errors":0,"num_deep_scrub_errors":0,"num_objects_recovered":0,"num_bytes_recovered":0,"num_keys_recovered":0,"num_objects_omap":0,"num_objects_hit_set_archive":0,"num_bytes_hit_set_archive":0,"num_flush":0,"num_flush_kb":0,"num_evict":0,"num_evict_kb":0,"num_promote":0,"num_flush_mode_high":0,"num_flush_mode_low":0,"num_evict_mode_some":0,"num_evict_mode_full":0,"num_objects_pinned":0,"num_legacy_snapsets":0,"num_large_omap_objects":0,"num_objects_manifest":0,"num_omap_bytes":0,"num_omap_keys":0,"num_objects_repaired":0},"up":[2,1],"acting":[2,1],"avail_no_missing":[],"object_location_counts":[],"blocked_by":[],"up_primary":2,"acting_primary":2,"purged_snaps":[{"start":"2","length":"1"}]},{"pgid":"2.0","version":"0'0","reported_seq":20,"reported_epoch":18,"state":"active+clean","last_fresh":"2026-03-20T17:57:25.678549+0000","last_change":"2026-03-20T17:57:25.678549+0000","last_active":"2026-03-20T17:57:25.678549+0000","last_peered":"2026-03-20T17:57:25.678549+0000","last_clean":"2026-03-20T17:57:25.678549+0000","last_became_active":"2026-03-20T17:57:23.203323+0000","last_became_peered":"2026-03-20T17:57:23.203323+0000","last_unstale":"2026-03-20T17:57:25.678549+0000","last_undegraded":"2026-03-20T17:57:25.678549+0000","last_fullsized":"2026-03-20T17:57:25.678549+0000","mapping_epoch":15,"log_start":"0'0","ondisk_log_start":"0'0","created":15,"last_epoch_clean":16,"parent":"0.0","parent_split_bits":0,"last_scrub":"0'0","last_scrub_stamp":"2026-03-20T17:57:22.189850+0000","last_deep_scrub":"0'0","last_deep_scrub_stamp":"2026-03-20T17:57:22.189850+0000","last_clean_scrub_stamp":"2026-03-20T17:57:22.189850+0000","objects_scrubbed":0,"log_size":0,"log_dups_size":0,"ondisk_log_size":0,"stats_invalid":false,"dirty_stats_invalid":false,"omap_stats_invalid":false,"hitset_stats_invalid":false,"hitset_bytes_stats_invalid":false,"pin_stats_invalid":false,"manifest_stats_invalid":false,"snaptrimq_len":0,"last_scrub_duration":0,"scrub_schedule":"periodic scrub scheduled @ 2026-03-21T23:07:56.359597+0000","scrub_duration":0,"objects_trimmed":0,"snaptrim_duration":0.00022052199999999999,"stat_sum":{"num_bytes":0,"num_objects":0,"num_object_clones":0,"num_object_copies":0,"num_objects_missing_on_primary":0,"num_objects_missing":0,"num_objects_degraded":0,"num_objects_misplaced":0,"num_objects_unfound":0,"num_objects_dirty":0,"num_whiteouts":0,"num_read":0,"num_read_kb":0,"num_write":0,"num_write_kb":0,"num_scrub_errors":0,"num_shallow_scrub_errors":0,"num_deep_scrub_errors":0,"num_objects_recovered":0,"num_bytes_recovered":0,"num_keys_recovered":0,"num_objects_omap":0,"num_objects_hit_set_archive":0,"num_bytes_hit_set_archive":0,"num_flush":0,"num_flush_kb":0,"num_evict":0,"num_evict_kb":0,"num_promote":0,"num_flush_mode_high":0,"num_flush_mode_low":0,"num_evict_mode_some":0,"num_evict_mode_full":0,"num_objects_pinned":0,"num_legacy_snapsets":0,"num_large_omap_objects":0,"num_objects_manifest":0,"num_omap_bytes":0,"num_omap_keys":0,"num_objects_repaired":0},"up":[3,1],"acting":[3,1],"avail_no_missing":[],"object_location_counts":[],"blocked_by":[],"up_primary":3,"acting_primary":3,"purged_snaps":[{"start":"2","length":"1"}]},{"pgid":"2.3","version":"16'1","reported_seq":21,"reported_epoch":18,"state":"active+clean","last_fresh":"2026-03-20T17:57:25.251045+0000","last_change":"2026-03-20T17:57:25.251045+0000","last_active":"2026-03-20T17:57:25.251045+0000","last_peered":"2026-03-20T17:57:25.251045+0000","last_clean":"2026-03-20T17:57:25.251045+0000","last_became_active":"2026-03-20T17:57:23.202693+0000","last_became_peered":"2026-03-20T17:57:23.202693+0000","last_unstale":"2026-03-20T17:57:25.251045+0000","last_undegraded":"2026-03-20T17:57:25.251045+0000","last_fullsized":"2026-03-20T17:57:25.251045+0000","mapping_epoch":15,"log_start":"0'0","ondisk_log_start":"0'0","created":15,"last_epoch_clean":16,"parent":"0.0","parent_split_bits":0,"last_scrub":"0'0","last_scrub_stamp":"2026-03-20T17:57:22.189850+0000","last_deep_scrub":"0'0","last_deep_scrub_stamp":"2026-03-20T17:57:22.189850+0000","last_clean_scrub_stamp":"2026-03-20T17:57:22.189850+0000","objects_scrubbed":0,"log_size":1,"log_dups_size":0,"ondisk_log_size":1,"stats_invalid":false,"dirty_stats_invalid":false,"omap_stats_invalid":false,"hitset_stats_invalid":false,"hitset_bytes_stats_invalid":false,"pin_stats_invalid":false,"manifest_stats_invalid":false,"snaptrimq_len":0,"last_scrub_duration":0,"scrub_schedule":"periodic scrub scheduled @ 2026-03-21T21:06:48.407505+0000","scrub_duration":0,"objects_trimmed":0,"snaptrim_duration":0.0002408,"stat_sum":{"num_bytes":0,"num_objects":1,"num_object_clones":0,"num_object_copies":2,"num_objects_missing_on_primary":0,"num_objects_missing":0,"num_objects_degraded":0,"num_objects_misplaced":0,"num_objects_unfound":0,"num_objects_dirty":1,"num_whiteouts":0,"num_read":0,"num_read_kb":0,"num_write":0,"num_write_kb":0,"num_scrub_errors":0,"num_shallow_scrub_errors":0,"num_deep_scrub_errors":0,"num_objects_recovered":0,"num_bytes_recovered":0,"num_keys_recovered":0,"num_objects_omap":0,"num_objects_hit_set_archive":0,"num_bytes_hit_set_archive":0,"num_flush":0,"num_flush_kb":0,"num_evict":0,"num_evict_kb":0,"num_promote":0,"num_flush_mode_high":0,"num_flush_mode_low":0,"num_evict_mode_some":0,"num_evict_mode_full":0,"num_objects_pinned":0,"num_legacy_snapsets":0,"num_large_omap_objects":0,"num_objects_manifest":0,"num_omap_bytes":0,"num_omap_keys":0,"num_objects_repaired":0},"up":[5,2],"acting":[5,2],"avail_no_missing":[],"object_location_counts":[],"blocked_by":[],"up_primary":5,"acting_primary":5,"purged_snaps":[{"start":"2","length":"1"}]},{"pgid":"1.0","version":"13'32","reported_seq":65,"reported_epoch":18,"state":"active+clean","last_fresh":"2026-03-20T17:57:25.678363+0000","last_change":"2026-03-20T17:57:19.630505+0000","last_active":"2026-03-20T17:57:25.678363+0000","last_peered":"2026-03-20T17:57:25.678363+0000","last_clean":"2026-03-20T17:57:25.678363+0000","last_became_active":"2026-03-20T17:57:19.630379+0000","last_became_peered":"2026-03-20T17:57:19.630379+0000","last_unstale":"2026-03-20T17:57:25.678363+0000","last_undegraded":"2026-03-20T17:57:25.678363+0000","last_fullsized":"2026-03-20T17:57:25.678363+0000","mapping_epoch":12,"log_start":"0'0","ondisk_log_start":"0'0","created":12,"last_epoch_clean":13,"parent":"0.0","parent_split_bits":0,"last_scrub":"0'0","last_scrub_stamp":"2026-03-20T17:57:18.175579+0000","last_deep_scrub":"0'0","last_deep_scrub_stamp":"2026-03-20T17:57:18.175579+0000","last_clean_scrub_stamp":"2026-03-20T17:57:18.175579+0000","objects_scrubbed":0,"log_size":32,"log_dups_size":0,"ondisk_log_size":32,"stats_invalid":false,"dirty_stats_invalid":false,"omap_stats_invalid":false,"hitset_stats_invalid":false,"hitset_bytes_stats_invalid":false,"pin_stats_invalid":false,"manifest_stats_invalid":false,"snaptrimq_len":0,"last_scrub_duration":0,"scrub_schedule":"periodic scrub scheduled @ 2026-03-22T03:25:29.659058+0000","scrub_duration":0,"objects_trimmed":0,"snaptrim_duration":0,"stat_sum":{"num_bytes":459280,"num_objects":2,"num_object_clones":0,"num_object_copies":4,"num_objects_missing_on_primary":0,"num_objects_missing":0,"num_objects_degraded":0,"num_objects_misplaced":0,"num_objects_unfound":0,"num_objects_dirty":2,"num_whiteouts":0,"num_read":46,"num_read_kb":37,"num_write":57,"num_write_kb":584,"num_scrub_errors":0,"num_shallow_scrub_errors":0,"num_deep_scrub_errors":0,"num_objects_recovered":0,"num_bytes_recovered":0,"num_keys_recovered":0,"num_objects_omap":0,"num_objects_hit_set_archive":0,"num_bytes_hit_set_archive":0,"num_flush":0,"num_flush_kb":0,"num_evict":0,"num_evict_kb":0,"num_promote":0,"num_flush_mode_high":0,"num_flush_mode_low":0,"num_evict_mode_some":0,"num_evict_mode_full":0,"num_objects_pinned":0,"num_legacy_snapsets":0,"num_large_omap_objects":0,"num_objects_manifest":0,"num_omap_bytes":0,"num_omap_keys":0,"num_objects_repaired":0},"up":[3,0],"acting":[3,0],"avail_no_missing":[],"object_location_counts":[],"blocked_by":[],"up_primary":3,"acting_primary":3,"purged_snaps":[]}],"pool_stats":[{"poolid":2,"num_pg":8,"stat_sum":{"num_bytes":19,"num_objects":2,"num_object_clones":0,"num_object_copies":4,"num_objects_missing_on_primary":0,"num_objects_missing":0,"num_objects_degraded":0,"num_objects_misplaced":0,"num_objects_unfound":0,"num_objects_dirty":2,"num_whiteouts":0,"num_read":0,"num_read_kb":0,"num_write":2,"num_write_kb":2,"num_scrub_errors":0,"num_shallow_scrub_errors":0,"num_deep_scrub_errors":0,"num_objects_recovered":0,"num_bytes_recovered":0,"num_keys_recovered":0,"num_objects_omap":0,"num_objects_hit_set_archive":0,"num_bytes_hit_set_archive":0,"num_flush":0,"num_flush_kb":0,"num_evict":0,"num_evict_kb":0,"num_promote":0,"num_flush_mode_high":0,"num_flush_mode_low":0,"num_evict_mode_some":0,"num_evict_mode_full":0,"num_objects_pinned":0,"num_legacy_snapsets":0,"num_large_omap_objects":0,"num_objects_manifest":0,"num_omap_bytes":0,"num_omap_keys":0,"num_objects_repaired":0},"store_stats":{"total":0,"available":0,"internally_reserved":0,"allocated":8192,"data_stored":38,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":0},"log_size":3,"ondisk_log_size":3,"up":16,"acting":16,"num_store_stats":6},{"poolid":1,"num_pg":1,"stat_sum":{"num_bytes":459280,"num_objects":2,"num_object_clones":0,"num_object_copies":4,"num_objects_missing_on_primary":0,"num_objects_missing":0,"num_objects_degraded":0,"num_objects_misplaced":0,"num_objects_unfound":0,"num_objects_dirty":2,"num_whiteouts":0,"num_read":46,"num_read_kb":37,"num_write":57,"num_write_kb":584,"num_scrub_errors":0,"num_shallow_scrub_errors":0,"num_deep_scrub_errors":0,"num_objects_recovered":0,"num_bytes_recovered":0,"num_keys_recovered":0,"num_objects_omap":0,"num_objects_hit_set_archive":0,"num_bytes_hit_set_archive":0,"num_flush":0,"num_flush_kb":0,"num_evict":0,"num_evict_kb":0,"num_promote":0,"num_flush_mode_high":0,"num_flush_mode_low":0,"num_evict_mode_some":0,"num_evict_mode_full":0,"num_objects_pinned":0,"num_legacy_snapsets":0,"num_large_omap_objects":0,"num_objects_manifest":0,"num_omap_bytes":0,"num_omap_keys":0,"num_objects_repaired":0},"store_stats":{"total":0,"available":0,"internally_reserved":0,"allocated":925696,"data_stored":918560,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":0},"log_size":32,"ondisk_log_size":32,"up":2,"acting":2,"num_store_stats":2}],"osd_stats":[{"osd":5,"up_from":11,"seq":47244640260,"num_pgs":2,"num_osds":1,"num_per_pool_osds":1,"num_per_pool_omap_osds":1,"kb":104857600,"kb_used":27020,"kb_used_data":172,"kb_used_omap":8,"kb_used_meta":26807,"kb_avail":104830580,"statfs":{"total":107374182400,"available":107346513920,"internally_reserved":0,"allocated":176128,"data_stored":56417,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":8770,"internal_metadata":27450814},"hb_peers":[0,1,2,3,4],"snap_trim_queue_len":0,"num_snap_trimming":0,"num_shards_repaired":0,"op_queue_age_hist":{"histogram":[],"upper_bound":1},"perf_stat":{"commit_latency_ms":0,"apply_latency_ms":0,"commit_latency_ns":0,"apply_latency_ns":0},"alerts":[]},{"osd":4,"up_from":11,"seq":47244640260,"num_pgs":1,"num_osds":1,"num_per_pool_osds":1,"num_per_pool_omap_osds":1,"kb":104857600,"kb_used":27024,"kb_used_data":168,"kb_used_omap":8,"kb_used_meta":26807,"kb_avail":104830576,"statfs":{"total":107374182400,"available":107346509824,"internally_reserved":0,"allocated":172032,"data_stored":56398,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":8771,"internal_metadata":27450813},"hb_peers":[0,1,2,3,5],"snap_trim_queue_len":0,"num_snap_trimming":0,"num_shards_repaired":0,"op_queue_age_hist":{"histogram":[],"upper_bound":1},"perf_stat":{"commit_latency_ms":0,"apply_latency_ms":0,"commit_latency_ns":0,"apply_latency_ns":0},"alerts":[]},{"osd":3,"up_from":11,"seq":47244640260,"num_pgs":4,"num_osds":1,"num_per_pool_osds":1,"num_per_pool_omap_osds":1,"kb":104857600,"kb_used":27468,"kb_used_data":620,"kb_used_omap":8,"kb_used_meta":26807,"kb_avail":104830132,"statfs":{"total":107374182400,"available":107346055168,"internally_reserved":0,"allocated":634880,"data_stored":515678,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":8769,"internal_metadata":27450815},"hb_peers":[0,1,2,4,5],"snap_trim_queue_len":0,"num_snap_trimming":0,"num_shards_repaired":0,"op_queue_age_hist":{"histogram":[],"upper_bound":1},"perf_stat":{"commit_latency_ms":0,"apply_latency_ms":0,"commit_latency_ns":0,"apply_latency_ns":0},"alerts":[]},{"osd":2,"up_from":11,"seq":47244640260,"num_pgs":2,"num_osds":1,"num_per_pool_osds":1,"num_per_pool_omap_osds":1,"kb":104857600,"kb_used":27016,"kb_used_data":168,"kb_used_omap":8,"kb_used_meta":26807,"kb_avail":104830584,"statfs":{"total":107374182400,"available":107346518016,"internally_reserved":0,"allocated":172032,"data_stored":56398,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":8771,"internal_metadata":27450813},"hb_peers":[0,1,3,4,5],"snap_trim_queue_len":0,"num_snap_trimming":0,"num_shards_repaired":0,"op_queue_age_hist":{"histogram":[],"upper_bound":1},"perf_stat":{"commit_latency_ms":0,"apply_latency_ms":0,"commit_latency_ns":0,"apply_latency_ns":0},"alerts":[]},{"osd":1,"up_from":11,"seq":47244640260,"num_pgs":5,"num_osds":1,"num_per_pool_osds":1,"num_per_pool_omap_osds":1,"kb":104857600,"kb_used":27020,"kb_used_data":172,"kb_used_omap":7,"kb_used_meta":26808,"kb_avail":104830580,"statfs":{"total":107374182400,"available":107346513920,"internally_reserved":0,"allocated":176128,"data_stored":56417,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":8121,"internal_metadata":27451463},"hb_peers":[0,2,3,4,5],"snap_trim_queue_len":0,"num_snap_trimming":0,"num_shards_repaired":0,"op_queue_age_hist":{"histogram":[],"upper_bound":1},"perf_stat":{"commit_latency_ms":0,"apply_latency_ms":0,"commit_latency_ns":0,"apply_latency_ns":0},"alerts":[]},{"osd":0,"up_from":11,"seq":47244640260,"num_pgs":3,"num_osds":1,"num_per_pool_osds":1,"num_per_pool_omap_osds":1,"kb":104857600,"kb_used":27468,"kb_used_data":620,"kb_used_omap":8,"kb_used_meta":26807,"kb_avail":104830132,"statfs":{"total":107374182400,"available":107346055168,"internally_reserved":0,"allocated":634880,"data_stored":515678,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":8771,"internal_metadata":27450813},"hb_peers":[1,2,3,4,5],"snap_trim_queue_len":0,"num_snap_trimming":0,"num_shards_repaired":0,"op_queue_age_hist":{"histogram":[],"upper_bound":1},"perf_stat":{"commit_latency_ms":0,"apply_latency_ms":0,"commit_latency_ns":0,"apply_latency_ns":0},"alerts":[]}],"pool_statfs":[{"poolid":1,"osd":0,"total":0,"available":0,"internally_reserved":0,"allocated":462848,"data_stored":459280,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":0},{"poolid":1,"osd":3,"total":0,"available":0,"internally_reserved":0,"allocated":462848,"data_stored":459280,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":0},{"poolid":2,"osd":0,"total":0,"available":0,"internally_reserved":0,"allocated":0,"data_stored":0,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":0},{"poolid":2,"osd":1,"total":0,"available":0,"internally_reserved":0,"allocated":4096,"data_stored":19,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":0},{"poolid":2,"osd":2,"total":0,"available":0,"internally_reserved":0,"allocated":0,"data_stored":0,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":0},{"poolid":2,"osd":3,"total":0,"available":0,"internally_reserved":0,"allocated":0,"data_stored":0,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":0},{"poolid":2,"osd":4,"total":0,"available":0,"internally_reserved":0,"allocated":0,"data_stored":0,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":0},{"poolid":2,"osd":5,"total":0,"available":0,"internally_reserved":0,"allocated":4096,"data_stored":19,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":0}]}} 2026-03-20T17:57:28.019 DEBUG:teuthology.orchestra.run.vm02:> sudo adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage timeout 120 ceph --cluster c1 pg dump --format=json 2026-03-20T17:57:28.179 INFO:teuthology.orchestra.run.vm02.stdout: 2026-03-20T17:57:28.180 INFO:teuthology.orchestra.run.vm02.stderr:dumped all 2026-03-20T17:57:28.195 INFO:teuthology.orchestra.run.vm02.stdout:{"pg_ready":true,"pg_map":{"version":18,"stamp":"2026-03-20T17:57:27.843751+0000","last_osdmap_epoch":0,"last_pg_scan":0,"pg_stats_sum":{"stat_sum":{"num_bytes":459299,"num_objects":4,"num_object_clones":0,"num_object_copies":8,"num_objects_missing_on_primary":0,"num_objects_missing":0,"num_objects_degraded":0,"num_objects_misplaced":0,"num_objects_unfound":0,"num_objects_dirty":4,"num_whiteouts":0,"num_read":46,"num_read_kb":37,"num_write":59,"num_write_kb":586,"num_scrub_errors":0,"num_shallow_scrub_errors":0,"num_deep_scrub_errors":0,"num_objects_recovered":0,"num_bytes_recovered":0,"num_keys_recovered":0,"num_objects_omap":0,"num_objects_hit_set_archive":0,"num_bytes_hit_set_archive":0,"num_flush":0,"num_flush_kb":0,"num_evict":0,"num_evict_kb":0,"num_promote":0,"num_flush_mode_high":0,"num_flush_mode_low":0,"num_evict_mode_some":0,"num_evict_mode_full":0,"num_objects_pinned":0,"num_legacy_snapsets":0,"num_large_omap_objects":0,"num_objects_manifest":0,"num_omap_bytes":0,"num_omap_keys":0,"num_objects_repaired":0},"store_stats":{"total":0,"available":0,"internally_reserved":0,"allocated":0,"data_stored":0,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":0},"log_size":35,"ondisk_log_size":35,"up":18,"acting":18,"num_store_stats":0},"osd_stats_sum":{"up_from":0,"seq":0,"num_pgs":17,"num_osds":6,"num_per_pool_osds":6,"num_per_pool_omap_osds":6,"kb":629145600,"kb_used":163016,"kb_used_data":1920,"kb_used_omap":50,"kb_used_meta":160845,"kb_avail":628982584,"statfs":{"total":644245094400,"available":644078166016,"internally_reserved":0,"allocated":1966080,"data_stored":1256986,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":51973,"internal_metadata":164705531},"hb_peers":[],"snap_trim_queue_len":0,"num_snap_trimming":0,"num_shards_repaired":0,"op_queue_age_hist":{"histogram":[],"upper_bound":1},"perf_stat":{"commit_latency_ms":0,"apply_latency_ms":0,"commit_latency_ns":0,"apply_latency_ns":0},"alerts":[],"network_ping_times":[]},"pg_stats_delta":{"stat_sum":{"num_bytes":19,"num_objects":2,"num_object_clones":0,"num_object_copies":4,"num_objects_missing_on_primary":0,"num_objects_missing":0,"num_objects_degraded":0,"num_objects_misplaced":0,"num_objects_unfound":0,"num_objects_dirty":2,"num_whiteouts":0,"num_read":0,"num_read_kb":0,"num_write":2,"num_write_kb":2,"num_scrub_errors":0,"num_shallow_scrub_errors":0,"num_deep_scrub_errors":0,"num_objects_recovered":0,"num_bytes_recovered":0,"num_keys_recovered":0,"num_objects_omap":0,"num_objects_hit_set_archive":0,"num_bytes_hit_set_archive":0,"num_flush":0,"num_flush_kb":0,"num_evict":0,"num_evict_kb":0,"num_promote":0,"num_flush_mode_high":0,"num_flush_mode_low":0,"num_evict_mode_some":0,"num_evict_mode_full":0,"num_objects_pinned":0,"num_legacy_snapsets":0,"num_large_omap_objects":0,"num_objects_manifest":0,"num_omap_bytes":0,"num_omap_keys":0,"num_objects_repaired":0},"store_stats":{"total":0,"available":0,"internally_reserved":0,"allocated":0,"data_stored":0,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":0},"log_size":0,"ondisk_log_size":0,"up":0,"acting":0,"num_store_stats":0,"stamp_delta":"4.674425"},"pg_stats":[{"pgid":"2.7","version":"0'0","reported_seq":20,"reported_epoch":18,"state":"active+clean","last_fresh":"2026-03-20T17:57:25.678438+0000","last_change":"2026-03-20T17:57:25.678438+0000","last_active":"2026-03-20T17:57:25.678438+0000","last_peered":"2026-03-20T17:57:25.678438+0000","last_clean":"2026-03-20T17:57:25.678438+0000","last_became_active":"2026-03-20T17:57:23.677052+0000","last_became_peered":"2026-03-20T17:57:23.677052+0000","last_unstale":"2026-03-20T17:57:25.678438+0000","last_undegraded":"2026-03-20T17:57:25.678438+0000","last_fullsized":"2026-03-20T17:57:25.678438+0000","mapping_epoch":15,"log_start":"0'0","ondisk_log_start":"0'0","created":15,"last_epoch_clean":16,"parent":"0.0","parent_split_bits":0,"last_scrub":"0'0","last_scrub_stamp":"2026-03-20T17:57:22.189850+0000","last_deep_scrub":"0'0","last_deep_scrub_stamp":"2026-03-20T17:57:22.189850+0000","last_clean_scrub_stamp":"2026-03-20T17:57:22.189850+0000","objects_scrubbed":0,"log_size":0,"log_dups_size":0,"ondisk_log_size":0,"stats_invalid":false,"dirty_stats_invalid":false,"omap_stats_invalid":false,"hitset_stats_invalid":false,"hitset_bytes_stats_invalid":false,"pin_stats_invalid":false,"manifest_stats_invalid":false,"snaptrimq_len":0,"last_scrub_duration":0,"scrub_schedule":"periodic scrub scheduled @ 2026-03-22T01:42:28.030112+0000","scrub_duration":0,"objects_trimmed":0,"snaptrim_duration":0.00024659200000000001,"stat_sum":{"num_bytes":0,"num_objects":0,"num_object_clones":0,"num_object_copies":0,"num_objects_missing_on_primary":0,"num_objects_missing":0,"num_objects_degraded":0,"num_objects_misplaced":0,"num_objects_unfound":0,"num_objects_dirty":0,"num_whiteouts":0,"num_read":0,"num_read_kb":0,"num_write":0,"num_write_kb":0,"num_scrub_errors":0,"num_shallow_scrub_errors":0,"num_deep_scrub_errors":0,"num_objects_recovered":0,"num_bytes_recovered":0,"num_keys_recovered":0,"num_objects_omap":0,"num_objects_hit_set_archive":0,"num_bytes_hit_set_archive":0,"num_flush":0,"num_flush_kb":0,"num_evict":0,"num_evict_kb":0,"num_promote":0,"num_flush_mode_high":0,"num_flush_mode_low":0,"num_evict_mode_some":0,"num_evict_mode_full":0,"num_objects_pinned":0,"num_legacy_snapsets":0,"num_large_omap_objects":0,"num_objects_manifest":0,"num_omap_bytes":0,"num_omap_keys":0,"num_objects_repaired":0},"up":[3,4],"acting":[3,4],"avail_no_missing":[],"object_location_counts":[],"blocked_by":[],"up_primary":3,"acting_primary":3,"purged_snaps":[{"start":"2","length":"1"}]},{"pgid":"2.6","version":"0'0","reported_seq":20,"reported_epoch":18,"state":"active+clean","last_fresh":"2026-03-20T17:57:25.252369+0000","last_change":"2026-03-20T17:57:25.252369+0000","last_active":"2026-03-20T17:57:25.252369+0000","last_peered":"2026-03-20T17:57:25.252369+0000","last_clean":"2026-03-20T17:57:25.252369+0000","last_became_active":"2026-03-20T17:57:23.203017+0000","last_became_peered":"2026-03-20T17:57:23.203017+0000","last_unstale":"2026-03-20T17:57:25.252369+0000","last_undegraded":"2026-03-20T17:57:25.252369+0000","last_fullsized":"2026-03-20T17:57:25.252369+0000","mapping_epoch":15,"log_start":"0'0","ondisk_log_start":"0'0","created":15,"last_epoch_clean":16,"parent":"0.0","parent_split_bits":0,"last_scrub":"0'0","last_scrub_stamp":"2026-03-20T17:57:22.189850+0000","last_deep_scrub":"0'0","last_deep_scrub_stamp":"2026-03-20T17:57:22.189850+0000","last_clean_scrub_stamp":"2026-03-20T17:57:22.189850+0000","objects_scrubbed":0,"log_size":0,"log_dups_size":0,"ondisk_log_size":0,"stats_invalid":false,"dirty_stats_invalid":false,"omap_stats_invalid":false,"hitset_stats_invalid":false,"hitset_bytes_stats_invalid":false,"pin_stats_invalid":false,"manifest_stats_invalid":false,"snaptrimq_len":0,"last_scrub_duration":0,"scrub_schedule":"periodic scrub scheduled @ 2026-03-21T18:16:06.167963+0000","scrub_duration":0,"objects_trimmed":0,"snaptrim_duration":0.00024865500000000002,"stat_sum":{"num_bytes":0,"num_objects":0,"num_object_clones":0,"num_object_copies":0,"num_objects_missing_on_primary":0,"num_objects_missing":0,"num_objects_degraded":0,"num_objects_misplaced":0,"num_objects_unfound":0,"num_objects_dirty":0,"num_whiteouts":0,"num_read":0,"num_read_kb":0,"num_write":0,"num_write_kb":0,"num_scrub_errors":0,"num_shallow_scrub_errors":0,"num_deep_scrub_errors":0,"num_objects_recovered":0,"num_bytes_recovered":0,"num_keys_recovered":0,"num_objects_omap":0,"num_objects_hit_set_archive":0,"num_bytes_hit_set_archive":0,"num_flush":0,"num_flush_kb":0,"num_evict":0,"num_evict_kb":0,"num_promote":0,"num_flush_mode_high":0,"num_flush_mode_low":0,"num_evict_mode_some":0,"num_evict_mode_full":0,"num_objects_pinned":0,"num_legacy_snapsets":0,"num_large_omap_objects":0,"num_objects_manifest":0,"num_omap_bytes":0,"num_omap_keys":0,"num_objects_repaired":0},"up":[1,3],"acting":[1,3],"avail_no_missing":[],"object_location_counts":[],"blocked_by":[],"up_primary":1,"acting_primary":1,"purged_snaps":[{"start":"2","length":"1"}]},{"pgid":"2.5","version":"0'0","reported_seq":20,"reported_epoch":18,"state":"active+clean","last_fresh":"2026-03-20T17:57:25.678482+0000","last_change":"2026-03-20T17:57:25.678482+0000","last_active":"2026-03-20T17:57:25.678482+0000","last_peered":"2026-03-20T17:57:25.678482+0000","last_clean":"2026-03-20T17:57:25.678482+0000","last_became_active":"2026-03-20T17:57:23.578680+0000","last_became_peered":"2026-03-20T17:57:23.578680+0000","last_unstale":"2026-03-20T17:57:25.678482+0000","last_undegraded":"2026-03-20T17:57:25.678482+0000","last_fullsized":"2026-03-20T17:57:25.678482+0000","mapping_epoch":15,"log_start":"0'0","ondisk_log_start":"0'0","created":15,"last_epoch_clean":16,"parent":"0.0","parent_split_bits":0,"last_scrub":"0'0","last_scrub_stamp":"2026-03-20T17:57:22.189850+0000","last_deep_scrub":"0'0","last_deep_scrub_stamp":"2026-03-20T17:57:22.189850+0000","last_clean_scrub_stamp":"2026-03-20T17:57:22.189850+0000","objects_scrubbed":0,"log_size":0,"log_dups_size":0,"ondisk_log_size":0,"stats_invalid":false,"dirty_stats_invalid":false,"omap_stats_invalid":false,"hitset_stats_invalid":false,"hitset_bytes_stats_invalid":false,"pin_stats_invalid":false,"manifest_stats_invalid":false,"snaptrimq_len":0,"last_scrub_duration":0,"scrub_schedule":"periodic scrub scheduled @ 2026-03-22T04:40:49.859879+0000","scrub_duration":0,"objects_trimmed":0,"snaptrim_duration":0.00021413000000000001,"stat_sum":{"num_bytes":0,"num_objects":0,"num_object_clones":0,"num_object_copies":0,"num_objects_missing_on_primary":0,"num_objects_missing":0,"num_objects_degraded":0,"num_objects_misplaced":0,"num_objects_unfound":0,"num_objects_dirty":0,"num_whiteouts":0,"num_read":0,"num_read_kb":0,"num_write":0,"num_write_kb":0,"num_scrub_errors":0,"num_shallow_scrub_errors":0,"num_deep_scrub_errors":0,"num_objects_recovered":0,"num_bytes_recovered":0,"num_keys_recovered":0,"num_objects_omap":0,"num_objects_hit_set_archive":0,"num_bytes_hit_set_archive":0,"num_flush":0,"num_flush_kb":0,"num_evict":0,"num_evict_kb":0,"num_promote":0,"num_flush_mode_high":0,"num_flush_mode_low":0,"num_evict_mode_some":0,"num_evict_mode_full":0,"num_objects_pinned":0,"num_legacy_snapsets":0,"num_large_omap_objects":0,"num_objects_manifest":0,"num_omap_bytes":0,"num_omap_keys":0,"num_objects_repaired":0},"up":[3,0],"acting":[3,0],"avail_no_missing":[],"object_location_counts":[],"blocked_by":[],"up_primary":3,"acting_primary":3,"purged_snaps":[{"start":"2","length":"1"}]},{"pgid":"2.4","version":"0'0","reported_seq":20,"reported_epoch":18,"state":"active+clean","last_fresh":"2026-03-20T17:57:25.252428+0000","last_change":"2026-03-20T17:57:25.252428+0000","last_active":"2026-03-20T17:57:25.252428+0000","last_peered":"2026-03-20T17:57:25.252428+0000","last_clean":"2026-03-20T17:57:25.252428+0000","last_became_active":"2026-03-20T17:57:23.577676+0000","last_became_peered":"2026-03-20T17:57:23.577676+0000","last_unstale":"2026-03-20T17:57:25.252428+0000","last_undegraded":"2026-03-20T17:57:25.252428+0000","last_fullsized":"2026-03-20T17:57:25.252428+0000","mapping_epoch":15,"log_start":"0'0","ondisk_log_start":"0'0","created":15,"last_epoch_clean":16,"parent":"0.0","parent_split_bits":0,"last_scrub":"0'0","last_scrub_stamp":"2026-03-20T17:57:22.189850+0000","last_deep_scrub":"0'0","last_deep_scrub_stamp":"2026-03-20T17:57:22.189850+0000","last_clean_scrub_stamp":"2026-03-20T17:57:22.189850+0000","objects_scrubbed":0,"log_size":0,"log_dups_size":0,"ondisk_log_size":0,"stats_invalid":false,"dirty_stats_invalid":false,"omap_stats_invalid":false,"hitset_stats_invalid":false,"hitset_bytes_stats_invalid":false,"pin_stats_invalid":false,"manifest_stats_invalid":false,"snaptrimq_len":0,"last_scrub_duration":0,"scrub_schedule":"periodic scrub scheduled @ 2026-03-22T04:16:13.040208+0000","scrub_duration":0,"objects_trimmed":0,"snaptrim_duration":0.000238576,"stat_sum":{"num_bytes":0,"num_objects":0,"num_object_clones":0,"num_object_copies":0,"num_objects_missing_on_primary":0,"num_objects_missing":0,"num_objects_degraded":0,"num_objects_misplaced":0,"num_objects_unfound":0,"num_objects_dirty":0,"num_whiteouts":0,"num_read":0,"num_read_kb":0,"num_write":0,"num_write_kb":0,"num_scrub_errors":0,"num_shallow_scrub_errors":0,"num_deep_scrub_errors":0,"num_objects_recovered":0,"num_bytes_recovered":0,"num_keys_recovered":0,"num_objects_omap":0,"num_objects_hit_set_archive":0,"num_bytes_hit_set_archive":0,"num_flush":0,"num_flush_kb":0,"num_evict":0,"num_evict_kb":0,"num_promote":0,"num_flush_mode_high":0,"num_flush_mode_low":0,"num_evict_mode_some":0,"num_evict_mode_full":0,"num_objects_pinned":0,"num_legacy_snapsets":0,"num_large_omap_objects":0,"num_objects_manifest":0,"num_omap_bytes":0,"num_omap_keys":0,"num_objects_repaired":0},"up":[1,0],"acting":[1,0],"avail_no_missing":[],"object_location_counts":[],"blocked_by":[],"up_primary":1,"acting_primary":1,"purged_snaps":[{"start":"2","length":"1"}]},{"pgid":"2.2","version":"18'2","reported_seq":22,"reported_epoch":18,"state":"active+clean","last_fresh":"2026-03-20T17:57:25.252876+0000","last_change":"2026-03-20T17:57:25.251301+0000","last_active":"2026-03-20T17:57:25.252876+0000","last_peered":"2026-03-20T17:57:25.252876+0000","last_clean":"2026-03-20T17:57:25.252876+0000","last_became_active":"2026-03-20T17:57:23.203631+0000","last_became_peered":"2026-03-20T17:57:23.203631+0000","last_unstale":"2026-03-20T17:57:25.252876+0000","last_undegraded":"2026-03-20T17:57:25.252876+0000","last_fullsized":"2026-03-20T17:57:25.252876+0000","mapping_epoch":15,"log_start":"0'0","ondisk_log_start":"0'0","created":15,"last_epoch_clean":16,"parent":"0.0","parent_split_bits":0,"last_scrub":"0'0","last_scrub_stamp":"2026-03-20T17:57:22.189850+0000","last_deep_scrub":"0'0","last_deep_scrub_stamp":"2026-03-20T17:57:22.189850+0000","last_clean_scrub_stamp":"2026-03-20T17:57:22.189850+0000","objects_scrubbed":0,"log_size":2,"log_dups_size":0,"ondisk_log_size":2,"stats_invalid":false,"dirty_stats_invalid":false,"omap_stats_invalid":false,"hitset_stats_invalid":false,"hitset_bytes_stats_invalid":false,"pin_stats_invalid":false,"manifest_stats_invalid":false,"snaptrimq_len":0,"last_scrub_duration":0,"scrub_schedule":"periodic scrub scheduled @ 2026-03-21T23:45:13.592323+0000","scrub_duration":0,"objects_trimmed":0,"snaptrim_duration":0.00050427299999999996,"stat_sum":{"num_bytes":19,"num_objects":1,"num_object_clones":0,"num_object_copies":2,"num_objects_missing_on_primary":0,"num_objects_missing":0,"num_objects_degraded":0,"num_objects_misplaced":0,"num_objects_unfound":0,"num_objects_dirty":1,"num_whiteouts":0,"num_read":0,"num_read_kb":0,"num_write":2,"num_write_kb":2,"num_scrub_errors":0,"num_shallow_scrub_errors":0,"num_deep_scrub_errors":0,"num_objects_recovered":0,"num_bytes_recovered":0,"num_keys_recovered":0,"num_objects_omap":0,"num_objects_hit_set_archive":0,"num_bytes_hit_set_archive":0,"num_flush":0,"num_flush_kb":0,"num_evict":0,"num_evict_kb":0,"num_promote":0,"num_flush_mode_high":0,"num_flush_mode_low":0,"num_evict_mode_some":0,"num_evict_mode_full":0,"num_objects_pinned":0,"num_legacy_snapsets":0,"num_large_omap_objects":0,"num_objects_manifest":0,"num_omap_bytes":0,"num_omap_keys":0,"num_objects_repaired":0},"up":[5,1],"acting":[5,1],"avail_no_missing":[],"object_location_counts":[],"blocked_by":[],"up_primary":5,"acting_primary":5,"purged_snaps":[{"start":"2","length":"1"}]},{"pgid":"2.1","version":"0'0","reported_seq":20,"reported_epoch":18,"state":"active+clean","last_fresh":"2026-03-20T17:57:25.714907+0000","last_change":"2026-03-20T17:57:25.714907+0000","last_active":"2026-03-20T17:57:25.714907+0000","last_peered":"2026-03-20T17:57:25.714907+0000","last_clean":"2026-03-20T17:57:25.714907+0000","last_became_active":"2026-03-20T17:57:23.203291+0000","last_became_peered":"2026-03-20T17:57:23.203291+0000","last_unstale":"2026-03-20T17:57:25.714907+0000","last_undegraded":"2026-03-20T17:57:25.714907+0000","last_fullsized":"2026-03-20T17:57:25.714907+0000","mapping_epoch":15,"log_start":"0'0","ondisk_log_start":"0'0","created":15,"last_epoch_clean":16,"parent":"0.0","parent_split_bits":0,"last_scrub":"0'0","last_scrub_stamp":"2026-03-20T17:57:22.189850+0000","last_deep_scrub":"0'0","last_deep_scrub_stamp":"2026-03-20T17:57:22.189850+0000","last_clean_scrub_stamp":"2026-03-20T17:57:22.189850+0000","objects_scrubbed":0,"log_size":0,"log_dups_size":0,"ondisk_log_size":0,"stats_invalid":false,"dirty_stats_invalid":false,"omap_stats_invalid":false,"hitset_stats_invalid":false,"hitset_bytes_stats_invalid":false,"pin_stats_invalid":false,"manifest_stats_invalid":false,"snaptrimq_len":0,"last_scrub_duration":0,"scrub_schedule":"periodic scrub scheduled @ 2026-03-21T18:37:05.593601+0000","scrub_duration":0,"objects_trimmed":0,"snaptrim_duration":0.000234458,"stat_sum":{"num_bytes":0,"num_objects":0,"num_object_clones":0,"num_object_copies":0,"num_objects_missing_on_primary":0,"num_objects_missing":0,"num_objects_degraded":0,"num_objects_misplaced":0,"num_objects_unfound":0,"num_objects_dirty":0,"num_whiteouts":0,"num_read":0,"num_read_kb":0,"num_write":0,"num_write_kb":0,"num_scrub_errors":0,"num_shallow_scrub_errors":0,"num_deep_scrub_errors":0,"num_objects_recovered":0,"num_bytes_recovered":0,"num_keys_recovered":0,"num_objects_omap":0,"num_objects_hit_set_archive":0,"num_bytes_hit_set_archive":0,"num_flush":0,"num_flush_kb":0,"num_evict":0,"num_evict_kb":0,"num_promote":0,"num_flush_mode_high":0,"num_flush_mode_low":0,"num_evict_mode_some":0,"num_evict_mode_full":0,"num_objects_pinned":0,"num_legacy_snapsets":0,"num_large_omap_objects":0,"num_objects_manifest":0,"num_omap_bytes":0,"num_omap_keys":0,"num_objects_repaired":0},"up":[2,1],"acting":[2,1],"avail_no_missing":[],"object_location_counts":[],"blocked_by":[],"up_primary":2,"acting_primary":2,"purged_snaps":[{"start":"2","length":"1"}]},{"pgid":"2.0","version":"0'0","reported_seq":20,"reported_epoch":18,"state":"active+clean","last_fresh":"2026-03-20T17:57:25.678549+0000","last_change":"2026-03-20T17:57:25.678549+0000","last_active":"2026-03-20T17:57:25.678549+0000","last_peered":"2026-03-20T17:57:25.678549+0000","last_clean":"2026-03-20T17:57:25.678549+0000","last_became_active":"2026-03-20T17:57:23.203323+0000","last_became_peered":"2026-03-20T17:57:23.203323+0000","last_unstale":"2026-03-20T17:57:25.678549+0000","last_undegraded":"2026-03-20T17:57:25.678549+0000","last_fullsized":"2026-03-20T17:57:25.678549+0000","mapping_epoch":15,"log_start":"0'0","ondisk_log_start":"0'0","created":15,"last_epoch_clean":16,"parent":"0.0","parent_split_bits":0,"last_scrub":"0'0","last_scrub_stamp":"2026-03-20T17:57:22.189850+0000","last_deep_scrub":"0'0","last_deep_scrub_stamp":"2026-03-20T17:57:22.189850+0000","last_clean_scrub_stamp":"2026-03-20T17:57:22.189850+0000","objects_scrubbed":0,"log_size":0,"log_dups_size":0,"ondisk_log_size":0,"stats_invalid":false,"dirty_stats_invalid":false,"omap_stats_invalid":false,"hitset_stats_invalid":false,"hitset_bytes_stats_invalid":false,"pin_stats_invalid":false,"manifest_stats_invalid":false,"snaptrimq_len":0,"last_scrub_duration":0,"scrub_schedule":"periodic scrub scheduled @ 2026-03-21T23:07:56.359597+0000","scrub_duration":0,"objects_trimmed":0,"snaptrim_duration":0.00022052199999999999,"stat_sum":{"num_bytes":0,"num_objects":0,"num_object_clones":0,"num_object_copies":0,"num_objects_missing_on_primary":0,"num_objects_missing":0,"num_objects_degraded":0,"num_objects_misplaced":0,"num_objects_unfound":0,"num_objects_dirty":0,"num_whiteouts":0,"num_read":0,"num_read_kb":0,"num_write":0,"num_write_kb":0,"num_scrub_errors":0,"num_shallow_scrub_errors":0,"num_deep_scrub_errors":0,"num_objects_recovered":0,"num_bytes_recovered":0,"num_keys_recovered":0,"num_objects_omap":0,"num_objects_hit_set_archive":0,"num_bytes_hit_set_archive":0,"num_flush":0,"num_flush_kb":0,"num_evict":0,"num_evict_kb":0,"num_promote":0,"num_flush_mode_high":0,"num_flush_mode_low":0,"num_evict_mode_some":0,"num_evict_mode_full":0,"num_objects_pinned":0,"num_legacy_snapsets":0,"num_large_omap_objects":0,"num_objects_manifest":0,"num_omap_bytes":0,"num_omap_keys":0,"num_objects_repaired":0},"up":[3,1],"acting":[3,1],"avail_no_missing":[],"object_location_counts":[],"blocked_by":[],"up_primary":3,"acting_primary":3,"purged_snaps":[{"start":"2","length":"1"}]},{"pgid":"2.3","version":"16'1","reported_seq":21,"reported_epoch":18,"state":"active+clean","last_fresh":"2026-03-20T17:57:25.251045+0000","last_change":"2026-03-20T17:57:25.251045+0000","last_active":"2026-03-20T17:57:25.251045+0000","last_peered":"2026-03-20T17:57:25.251045+0000","last_clean":"2026-03-20T17:57:25.251045+0000","last_became_active":"2026-03-20T17:57:23.202693+0000","last_became_peered":"2026-03-20T17:57:23.202693+0000","last_unstale":"2026-03-20T17:57:25.251045+0000","last_undegraded":"2026-03-20T17:57:25.251045+0000","last_fullsized":"2026-03-20T17:57:25.251045+0000","mapping_epoch":15,"log_start":"0'0","ondisk_log_start":"0'0","created":15,"last_epoch_clean":16,"parent":"0.0","parent_split_bits":0,"last_scrub":"0'0","last_scrub_stamp":"2026-03-20T17:57:22.189850+0000","last_deep_scrub":"0'0","last_deep_scrub_stamp":"2026-03-20T17:57:22.189850+0000","last_clean_scrub_stamp":"2026-03-20T17:57:22.189850+0000","objects_scrubbed":0,"log_size":1,"log_dups_size":0,"ondisk_log_size":1,"stats_invalid":false,"dirty_stats_invalid":false,"omap_stats_invalid":false,"hitset_stats_invalid":false,"hitset_bytes_stats_invalid":false,"pin_stats_invalid":false,"manifest_stats_invalid":false,"snaptrimq_len":0,"last_scrub_duration":0,"scrub_schedule":"periodic scrub scheduled @ 2026-03-21T21:06:48.407505+0000","scrub_duration":0,"objects_trimmed":0,"snaptrim_duration":0.0002408,"stat_sum":{"num_bytes":0,"num_objects":1,"num_object_clones":0,"num_object_copies":2,"num_objects_missing_on_primary":0,"num_objects_missing":0,"num_objects_degraded":0,"num_objects_misplaced":0,"num_objects_unfound":0,"num_objects_dirty":1,"num_whiteouts":0,"num_read":0,"num_read_kb":0,"num_write":0,"num_write_kb":0,"num_scrub_errors":0,"num_shallow_scrub_errors":0,"num_deep_scrub_errors":0,"num_objects_recovered":0,"num_bytes_recovered":0,"num_keys_recovered":0,"num_objects_omap":0,"num_objects_hit_set_archive":0,"num_bytes_hit_set_archive":0,"num_flush":0,"num_flush_kb":0,"num_evict":0,"num_evict_kb":0,"num_promote":0,"num_flush_mode_high":0,"num_flush_mode_low":0,"num_evict_mode_some":0,"num_evict_mode_full":0,"num_objects_pinned":0,"num_legacy_snapsets":0,"num_large_omap_objects":0,"num_objects_manifest":0,"num_omap_bytes":0,"num_omap_keys":0,"num_objects_repaired":0},"up":[5,2],"acting":[5,2],"avail_no_missing":[],"object_location_counts":[],"blocked_by":[],"up_primary":5,"acting_primary":5,"purged_snaps":[{"start":"2","length":"1"}]},{"pgid":"1.0","version":"13'32","reported_seq":65,"reported_epoch":18,"state":"active+clean","last_fresh":"2026-03-20T17:57:25.678363+0000","last_change":"2026-03-20T17:57:19.630505+0000","last_active":"2026-03-20T17:57:25.678363+0000","last_peered":"2026-03-20T17:57:25.678363+0000","last_clean":"2026-03-20T17:57:25.678363+0000","last_became_active":"2026-03-20T17:57:19.630379+0000","last_became_peered":"2026-03-20T17:57:19.630379+0000","last_unstale":"2026-03-20T17:57:25.678363+0000","last_undegraded":"2026-03-20T17:57:25.678363+0000","last_fullsized":"2026-03-20T17:57:25.678363+0000","mapping_epoch":12,"log_start":"0'0","ondisk_log_start":"0'0","created":12,"last_epoch_clean":13,"parent":"0.0","parent_split_bits":0,"last_scrub":"0'0","last_scrub_stamp":"2026-03-20T17:57:18.175579+0000","last_deep_scrub":"0'0","last_deep_scrub_stamp":"2026-03-20T17:57:18.175579+0000","last_clean_scrub_stamp":"2026-03-20T17:57:18.175579+0000","objects_scrubbed":0,"log_size":32,"log_dups_size":0,"ondisk_log_size":32,"stats_invalid":false,"dirty_stats_invalid":false,"omap_stats_invalid":false,"hitset_stats_invalid":false,"hitset_bytes_stats_invalid":false,"pin_stats_invalid":false,"manifest_stats_invalid":false,"snaptrimq_len":0,"last_scrub_duration":0,"scrub_schedule":"periodic scrub scheduled @ 2026-03-22T03:25:29.659058+0000","scrub_duration":0,"objects_trimmed":0,"snaptrim_duration":0,"stat_sum":{"num_bytes":459280,"num_objects":2,"num_object_clones":0,"num_object_copies":4,"num_objects_missing_on_primary":0,"num_objects_missing":0,"num_objects_degraded":0,"num_objects_misplaced":0,"num_objects_unfound":0,"num_objects_dirty":2,"num_whiteouts":0,"num_read":46,"num_read_kb":37,"num_write":57,"num_write_kb":584,"num_scrub_errors":0,"num_shallow_scrub_errors":0,"num_deep_scrub_errors":0,"num_objects_recovered":0,"num_bytes_recovered":0,"num_keys_recovered":0,"num_objects_omap":0,"num_objects_hit_set_archive":0,"num_bytes_hit_set_archive":0,"num_flush":0,"num_flush_kb":0,"num_evict":0,"num_evict_kb":0,"num_promote":0,"num_flush_mode_high":0,"num_flush_mode_low":0,"num_evict_mode_some":0,"num_evict_mode_full":0,"num_objects_pinned":0,"num_legacy_snapsets":0,"num_large_omap_objects":0,"num_objects_manifest":0,"num_omap_bytes":0,"num_omap_keys":0,"num_objects_repaired":0},"up":[3,0],"acting":[3,0],"avail_no_missing":[],"object_location_counts":[],"blocked_by":[],"up_primary":3,"acting_primary":3,"purged_snaps":[]}],"pool_stats":[{"poolid":2,"num_pg":8,"stat_sum":{"num_bytes":19,"num_objects":2,"num_object_clones":0,"num_object_copies":4,"num_objects_missing_on_primary":0,"num_objects_missing":0,"num_objects_degraded":0,"num_objects_misplaced":0,"num_objects_unfound":0,"num_objects_dirty":2,"num_whiteouts":0,"num_read":0,"num_read_kb":0,"num_write":2,"num_write_kb":2,"num_scrub_errors":0,"num_shallow_scrub_errors":0,"num_deep_scrub_errors":0,"num_objects_recovered":0,"num_bytes_recovered":0,"num_keys_recovered":0,"num_objects_omap":0,"num_objects_hit_set_archive":0,"num_bytes_hit_set_archive":0,"num_flush":0,"num_flush_kb":0,"num_evict":0,"num_evict_kb":0,"num_promote":0,"num_flush_mode_high":0,"num_flush_mode_low":0,"num_evict_mode_some":0,"num_evict_mode_full":0,"num_objects_pinned":0,"num_legacy_snapsets":0,"num_large_omap_objects":0,"num_objects_manifest":0,"num_omap_bytes":0,"num_omap_keys":0,"num_objects_repaired":0},"store_stats":{"total":0,"available":0,"internally_reserved":0,"allocated":8192,"data_stored":38,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":0},"log_size":3,"ondisk_log_size":3,"up":16,"acting":16,"num_store_stats":6},{"poolid":1,"num_pg":1,"stat_sum":{"num_bytes":459280,"num_objects":2,"num_object_clones":0,"num_object_copies":4,"num_objects_missing_on_primary":0,"num_objects_missing":0,"num_objects_degraded":0,"num_objects_misplaced":0,"num_objects_unfound":0,"num_objects_dirty":2,"num_whiteouts":0,"num_read":46,"num_read_kb":37,"num_write":57,"num_write_kb":584,"num_scrub_errors":0,"num_shallow_scrub_errors":0,"num_deep_scrub_errors":0,"num_objects_recovered":0,"num_bytes_recovered":0,"num_keys_recovered":0,"num_objects_omap":0,"num_objects_hit_set_archive":0,"num_bytes_hit_set_archive":0,"num_flush":0,"num_flush_kb":0,"num_evict":0,"num_evict_kb":0,"num_promote":0,"num_flush_mode_high":0,"num_flush_mode_low":0,"num_evict_mode_some":0,"num_evict_mode_full":0,"num_objects_pinned":0,"num_legacy_snapsets":0,"num_large_omap_objects":0,"num_objects_manifest":0,"num_omap_bytes":0,"num_omap_keys":0,"num_objects_repaired":0},"store_stats":{"total":0,"available":0,"internally_reserved":0,"allocated":925696,"data_stored":918560,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":0},"log_size":32,"ondisk_log_size":32,"up":2,"acting":2,"num_store_stats":2}],"osd_stats":[{"osd":5,"up_from":11,"seq":47244640260,"num_pgs":2,"num_osds":1,"num_per_pool_osds":1,"num_per_pool_omap_osds":1,"kb":104857600,"kb_used":27020,"kb_used_data":172,"kb_used_omap":8,"kb_used_meta":26807,"kb_avail":104830580,"statfs":{"total":107374182400,"available":107346513920,"internally_reserved":0,"allocated":176128,"data_stored":56417,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":8770,"internal_metadata":27450814},"hb_peers":[0,1,2,3,4],"snap_trim_queue_len":0,"num_snap_trimming":0,"num_shards_repaired":0,"op_queue_age_hist":{"histogram":[],"upper_bound":1},"perf_stat":{"commit_latency_ms":0,"apply_latency_ms":0,"commit_latency_ns":0,"apply_latency_ns":0},"alerts":[]},{"osd":4,"up_from":11,"seq":47244640260,"num_pgs":1,"num_osds":1,"num_per_pool_osds":1,"num_per_pool_omap_osds":1,"kb":104857600,"kb_used":27024,"kb_used_data":168,"kb_used_omap":8,"kb_used_meta":26807,"kb_avail":104830576,"statfs":{"total":107374182400,"available":107346509824,"internally_reserved":0,"allocated":172032,"data_stored":56398,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":8771,"internal_metadata":27450813},"hb_peers":[0,1,2,3,5],"snap_trim_queue_len":0,"num_snap_trimming":0,"num_shards_repaired":0,"op_queue_age_hist":{"histogram":[],"upper_bound":1},"perf_stat":{"commit_latency_ms":0,"apply_latency_ms":0,"commit_latency_ns":0,"apply_latency_ns":0},"alerts":[]},{"osd":3,"up_from":11,"seq":47244640260,"num_pgs":4,"num_osds":1,"num_per_pool_osds":1,"num_per_pool_omap_osds":1,"kb":104857600,"kb_used":27468,"kb_used_data":620,"kb_used_omap":8,"kb_used_meta":26807,"kb_avail":104830132,"statfs":{"total":107374182400,"available":107346055168,"internally_reserved":0,"allocated":634880,"data_stored":515678,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":8769,"internal_metadata":27450815},"hb_peers":[0,1,2,4,5],"snap_trim_queue_len":0,"num_snap_trimming":0,"num_shards_repaired":0,"op_queue_age_hist":{"histogram":[],"upper_bound":1},"perf_stat":{"commit_latency_ms":0,"apply_latency_ms":0,"commit_latency_ns":0,"apply_latency_ns":0},"alerts":[]},{"osd":2,"up_from":11,"seq":47244640260,"num_pgs":2,"num_osds":1,"num_per_pool_osds":1,"num_per_pool_omap_osds":1,"kb":104857600,"kb_used":27016,"kb_used_data":168,"kb_used_omap":8,"kb_used_meta":26807,"kb_avail":104830584,"statfs":{"total":107374182400,"available":107346518016,"internally_reserved":0,"allocated":172032,"data_stored":56398,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":8771,"internal_metadata":27450813},"hb_peers":[0,1,3,4,5],"snap_trim_queue_len":0,"num_snap_trimming":0,"num_shards_repaired":0,"op_queue_age_hist":{"histogram":[],"upper_bound":1},"perf_stat":{"commit_latency_ms":0,"apply_latency_ms":0,"commit_latency_ns":0,"apply_latency_ns":0},"alerts":[]},{"osd":1,"up_from":11,"seq":47244640260,"num_pgs":5,"num_osds":1,"num_per_pool_osds":1,"num_per_pool_omap_osds":1,"kb":104857600,"kb_used":27020,"kb_used_data":172,"kb_used_omap":7,"kb_used_meta":26808,"kb_avail":104830580,"statfs":{"total":107374182400,"available":107346513920,"internally_reserved":0,"allocated":176128,"data_stored":56417,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":8121,"internal_metadata":27451463},"hb_peers":[0,2,3,4,5],"snap_trim_queue_len":0,"num_snap_trimming":0,"num_shards_repaired":0,"op_queue_age_hist":{"histogram":[],"upper_bound":1},"perf_stat":{"commit_latency_ms":0,"apply_latency_ms":0,"commit_latency_ns":0,"apply_latency_ns":0},"alerts":[]},{"osd":0,"up_from":11,"seq":47244640260,"num_pgs":3,"num_osds":1,"num_per_pool_osds":1,"num_per_pool_omap_osds":1,"kb":104857600,"kb_used":27468,"kb_used_data":620,"kb_used_omap":8,"kb_used_meta":26807,"kb_avail":104830132,"statfs":{"total":107374182400,"available":107346055168,"internally_reserved":0,"allocated":634880,"data_stored":515678,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":8771,"internal_metadata":27450813},"hb_peers":[1,2,3,4,5],"snap_trim_queue_len":0,"num_snap_trimming":0,"num_shards_repaired":0,"op_queue_age_hist":{"histogram":[],"upper_bound":1},"perf_stat":{"commit_latency_ms":0,"apply_latency_ms":0,"commit_latency_ns":0,"apply_latency_ns":0},"alerts":[]}],"pool_statfs":[{"poolid":1,"osd":0,"total":0,"available":0,"internally_reserved":0,"allocated":462848,"data_stored":459280,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":0},{"poolid":1,"osd":3,"total":0,"available":0,"internally_reserved":0,"allocated":462848,"data_stored":459280,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":0},{"poolid":2,"osd":0,"total":0,"available":0,"internally_reserved":0,"allocated":0,"data_stored":0,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":0},{"poolid":2,"osd":1,"total":0,"available":0,"internally_reserved":0,"allocated":4096,"data_stored":19,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":0},{"poolid":2,"osd":2,"total":0,"available":0,"internally_reserved":0,"allocated":0,"data_stored":0,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":0},{"poolid":2,"osd":3,"total":0,"available":0,"internally_reserved":0,"allocated":0,"data_stored":0,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":0},{"poolid":2,"osd":4,"total":0,"available":0,"internally_reserved":0,"allocated":0,"data_stored":0,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":0},{"poolid":2,"osd":5,"total":0,"available":0,"internally_reserved":0,"allocated":4096,"data_stored":19,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":0}]}} 2026-03-20T17:57:28.195 INFO:tasks.ceph.ceph_manager.c1:clean! 2026-03-20T17:57:28.195 INFO:tasks.ceph:Waiting until ceph cluster c1 is healthy... 2026-03-20T17:57:28.195 INFO:tasks.ceph.ceph_manager.c1:wait_until_healthy 2026-03-20T17:57:28.195 DEBUG:teuthology.orchestra.run.vm02:> sudo adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage timeout 120 ceph --cluster c1 health --format=json 2026-03-20T17:57:28.378 INFO:teuthology.orchestra.run.vm02.stdout: 2026-03-20T17:57:28.378 INFO:teuthology.orchestra.run.vm02.stdout:{"status":"HEALTH_OK","checks":{},"mutes":[]} 2026-03-20T17:57:28.393 INFO:tasks.ceph.ceph_manager.c1:wait_until_healthy done 2026-03-20T17:57:28.393 INFO:teuthology.run_tasks:Running task ceph... 2026-03-20T17:57:28.396 INFO:tasks.ceph:Creating ceph cluster c2... 2026-03-20T17:57:28.396 INFO:tasks.ceph:config {'conf': {'client': {'debug rgw': 20, 'debug rgw lifecycle': 20, 'debug rgw notification': 20, 'debug rgw sync': 20, 'rgw crypt require ssl': False, 'rgw crypt s3 kms backend': 'testing', 'rgw crypt s3 kms encryption keys': 'testkey-1=YmluCmJvb3N0CmJvb3N0LWJ1aWxkCmNlcGguY29uZgo=', 'rgw curl low speed time': 300, 'rgw data log num shards': 4, 'rgw data sync poll interval': 5, 'rgw md log max shards': 4, 'rgw meta sync poll interval': 5, 'rgw s3 auth use sts': True, 'rgw sts key': 'abcdefghijklmnoq', 'rgw sync data inject err probability': 0, 'rgw sync log trim interval': 0, 'rgw sync meta inject err probability': 0, 'rgw sync obj etag verify': True, 'setgroup': 'ceph', 'setuser': 'ceph'}, 'client.0': {'rgw data notify interval msec': 0}, 'mgr': {'debug mgr': 20, 'debug ms': 1}, 'mon': {'debug mon': 20, 'debug ms': 1, 'debug paxos': 20}, 'osd': {'debug ms': 1, 'debug osd': 20, 'osd mclock iops capacity threshold hdd': 49000, 'rocksdb delete range threshold': 0}, 'osd.0': {'osd_max_omap_entries_per_request': 10}, 'osd.1': {'osd_max_omap_entries_per_request': 1000}, 'osd.2': {'osd_max_omap_entries_per_request': 10000}}, 'fs': 'xfs', 'mkfs_options': None, 'mount_options': None, 'skip_mgr_daemons': False, 'log_ignorelist': ['\\(MDS_ALL_DOWN\\)', '\\(MDS_UP_LESS_THAN_MAX\\)', '\\(PG_AVAILABILITY\\)', '\\(PG_DEGRADED\\)', '\\(POOL_APP_NOT_ENABLED\\)', 'not have an application enabled'], 'cpu_profile': set(), 'cluster': 'c2', 'mon_bind_msgr2': True, 'mon_bind_addrvec': True} 2026-03-20T17:57:28.396 INFO:tasks.ceph:ctx.config {'archive_path': '/archive/kyr-2026-03-20_17:25:16-rgw-tentacle-none-default-vps/2229', 'branch': 'tentacle', 'description': 'rgw/multisite/{clusters frontend/beast ignore-pg-availability notify omap_limits overrides realms/two-zonegroup supported-random-distro$/{ubuntu_latest} tasks/test_multi}', 'email': None, 'first_in_suite': False, 'flavor': 'default', 'job_id': '2229', 'ktype': 'distro', 'last_in_suite': False, 'machine_type': 'vps', 'name': 'kyr-2026-03-20_17:25:16-rgw-tentacle-none-default-vps', 'no_nested_subset': False, 'os_type': 'ubuntu', 'os_version': '22.04', 'overrides': {'admin_socket': {'branch': 'tentacle'}, 'ansible.cephlab': {'branch': 'main', 'repo': 'https://github.com/kshtsk/ceph-cm-ansible.git', 'skip_tags': 'nagios,monitoring-scripts,hostname,pubkeys,zap,sudoers,kerberos,ntp-client,resolvconf,cpan,nfs', 'vars': {'logical_volumes': {'lv_1': {'scratch_dev': True, 'size': '25%VG', 'vg': 'vg_nvme'}, 'lv_2': {'scratch_dev': True, 'size': '25%VG', 'vg': 'vg_nvme'}, 'lv_3': {'scratch_dev': True, 'size': '25%VG', 'vg': 'vg_nvme'}, 'lv_4': {'scratch_dev': True, 'size': '25%VG', 'vg': 'vg_nvme'}}, 'timezone': 'UTC', 'volume_groups': {'vg_nvme': {'pvs': '/dev/vdb,/dev/vdc,/dev/vdd,/dev/vde'}}}}, 'ceph': {'conf': {'client': {'debug rgw': 20, 'debug rgw lifecycle': 20, 'debug rgw notification': 20, 'debug rgw sync': 20, 'rgw crypt require ssl': False, 'rgw crypt s3 kms backend': 'testing', 'rgw crypt s3 kms encryption keys': 'testkey-1=YmluCmJvb3N0CmJvb3N0LWJ1aWxkCmNlcGguY29uZgo=', 'rgw curl low speed time': 300, 'rgw data log num shards': 4, 'rgw data sync poll interval': 5, 'rgw md log max shards': 4, 'rgw meta sync poll interval': 5, 'rgw s3 auth use sts': True, 'rgw sts key': 'abcdefghijklmnoq', 'rgw sync data inject err probability': 0, 'rgw sync log trim interval': 0, 'rgw sync meta inject err probability': 0, 'rgw sync obj etag verify': True, 'setgroup': 'ceph', 'setuser': 'ceph'}, 'client.0': {'rgw data notify interval msec': 0}, 'mgr': {'debug mgr': 20, 'debug ms': 1}, 'mon': {'debug mon': 20, 'debug ms': 1, 'debug paxos': 20}, 'osd': {'debug ms': 1, 'debug osd': 20, 'osd mclock iops capacity threshold hdd': 49000, 'rocksdb delete range threshold': 0}, 'osd.0': {'osd_max_omap_entries_per_request': 10}, 'osd.1': {'osd_max_omap_entries_per_request': 1000}, 'osd.2': {'osd_max_omap_entries_per_request': 10000}}, 'flavor': 'default', 'log-ignorelist': ['\\(MDS_ALL_DOWN\\)', '\\(MDS_UP_LESS_THAN_MAX\\)', '\\(PG_AVAILABILITY\\)', '\\(PG_DEGRADED\\)', '\\(POOL_APP_NOT_ENABLED\\)', 'not have an application enabled'], 'sha1': '70f8415b300f041766fa27faf7d5472699e32388'}, 'ceph-deploy': {'conf': {'client': {'log file': '/var/log/ceph/ceph-$name.$pid.log'}, 'mon': {}}}, 'cephadm': {'cephadm_binary_url': 'https://download.ceph.com/rpm-20.2.0/el9/noarch/cephadm'}, 'install': {'ceph': {'flavor': 'default', 'sha1': '70f8415b300f041766fa27faf7d5472699e32388'}, 'extra_system_packages': {'deb': ['python3-jmespath', 'python3-xmltodict', 's3cmd'], 'rpm': ['bzip2', 'perl-Test-Harness', 'python3-jmespath', 'python3-xmltodict', 's3cmd']}}, 'rgw': {'compression type': 'random', 'frontend': 'beast'}, 'rgw-multisite': {'realm': {'is_default': True, 'name': 'test-realm'}, 'zonegroups': [{'enabled_features': ['resharding', 'notification_v2'], 'endpoints': ['c1.client.0'], 'is_default': True, 'is_master': True, 'name': 'a', 'zones': [{'endpoints': ['c1.client.0'], 'is_default': True, 'is_master': True, 'name': 'a1'}, {'endpoints': ['c1.client.1'], 'name': 'a2'}]}, {'enabled_features': ['resharding', 'notification_v2'], 'endpoints': ['c2.client.0'], 'is_default': True, 'name': 'b', 'zones': [{'endpoints': ['c2.client.0'], 'is_default': True, 'is_master': True, 'name': 'b1'}, {'endpoints': ['c2.client.1'], 'name': 'b2'}]}]}, 'rgw-multisite-tests': {'args': ['tests.py', '-a', '!fails_with_rgw']}, 'workunit': {'branch': 'tt-tentacle', 'sha1': 'c04fae7bd18063e6f8f9e1cf40c23f810c1f6ca4'}}, 'owner': 'kyr', 'priority': 1000, 'repo': 'https://github.com/ceph/ceph.git', 'roles': [['c1.mon.a', 'c1.osd.0', 'c1.osd.1', 'c1.osd.2', 'c1.client.0'], ['c1.mgr.x', 'c1.osd.3', 'c1.osd.4', 'c1.osd.5', 'c1.client.1'], ['c2.mon.a', 'c2.osd.0', 'c2.osd.1', 'c2.osd.2', 'c2.client.0'], ['c2.mgr.x', 'c2.osd.3', 'c2.osd.4', 'c2.osd.5', 'c2.client.1']], 'seed': 5336, 'sha1': '70f8415b300f041766fa27faf7d5472699e32388', 'sleep_before_teardown': 0, 'suite': 'rgw', 'suite_branch': 'tt-tentacle', 'suite_path': '/home/teuthos/src/github.com_kshtsk_ceph_c04fae7bd18063e6f8f9e1cf40c23f810c1f6ca4/qa', 'suite_relpath': 'qa', 'suite_repo': 'https://github.com/kshtsk/ceph.git', 'suite_sha1': 'c04fae7bd18063e6f8f9e1cf40c23f810c1f6ca4', 'targets': {'vm02.local': 'ecdsa-sha2-nistp256 AAAAE2VjZHNhLXNoYTItbmlzdHAyNTYAAAAIbmlzdHAyNTYAAABBBLUlbeCQemEFIGAhOsbaSaC9cn/mRW0uqcv6nNbZb53B4Wibb+XPFovkYm/zpOAZ4wEdxyZ/PUnoWpm0Py9ocQE=', 'vm05.local': 'ecdsa-sha2-nistp256 AAAAE2VjZHNhLXNoYTItbmlzdHAyNTYAAAAIbmlzdHAyNTYAAABBBMoug7TNAoa7mJjh9cncEud7jrvuXdSIQZHZLMdsjGiNzszM0eqBraWpNHWznD8Hn4dQwU6ldQdhO/fM1PnoaIE=', 'vm06.local': 'ecdsa-sha2-nistp256 AAAAE2VjZHNhLXNoYTItbmlzdHAyNTYAAAAIbmlzdHAyNTYAAABBBAr4G6JTVJMADXrDYrEOUi/7hRfFSeJoHsHrYT4rHLq31Qu7OJgh6KH0wNL/WripVH20sVMfbmCvXmnnPcODwbA=', 'vm08.local': 'ecdsa-sha2-nistp256 AAAAE2VjZHNhLXNoYTItbmlzdHAyNTYAAAAIbmlzdHAyNTYAAABBBOUGJylvumZGYQnc6Cm2BiuFCvdgt79isI8bVHVCdZ3r8NnfMvYpiSD3SPVaM+Ua03wQvWKlAWO93A22s3C5WOk='}, 'tasks': [{'internal.check_packages': None}, {'internal.buildpackages_prep': None}, {'internal.save_config': None}, {'internal.check_lock': None}, {'internal.add_remotes': None}, {'console_log': None}, {'internal.connect': None}, {'internal.push_inventory': None}, {'internal.serialize_remote_roles': None}, {'internal.check_conflict': None}, {'internal.check_ceph_data': None}, {'internal.vm_setup': None}, {'internal.base': None}, {'internal.archive_upload': None}, {'internal.archive': None}, {'internal.coredump': None}, {'internal.sudo': None}, {'internal.syslog': None}, {'internal.timer': None}, {'pcp': None}, {'selinux': None}, {'ansible.cephlab': None}, {'clock': None}, {'install': None}, {'ceph': {'cluster': 'c1', 'conf': {'client': {'debug rgw': 20, 'debug rgw lifecycle': 20, 'debug rgw notification': 20, 'debug rgw sync': 20, 'rgw crypt require ssl': False, 'rgw crypt s3 kms backend': 'testing', 'rgw crypt s3 kms encryption keys': 'testkey-1=YmluCmJvb3N0CmJvb3N0LWJ1aWxkCmNlcGguY29uZgo=', 'rgw curl low speed time': 300, 'rgw data log num shards': 4, 'rgw data sync poll interval': 5, 'rgw md log max shards': 4, 'rgw meta sync poll interval': 5, 'rgw s3 auth use sts': True, 'rgw sts key': 'abcdefghijklmnoq', 'rgw sync data inject err probability': 0, 'rgw sync log trim interval': 0, 'rgw sync meta inject err probability': 0, 'rgw sync obj etag verify': True, 'setgroup': 'ceph', 'setuser': 'ceph'}, 'client.0': {'rgw data notify interval msec': 0}, 'mgr': {'debug mgr': 20, 'debug ms': 1}, 'mon': {'debug mon': 20, 'debug ms': 1, 'debug paxos': 20}, 'osd': {'debug ms': 1, 'debug osd': 20, 'osd mclock iops capacity threshold hdd': 49000, 'rocksdb delete range threshold': 0}, 'osd.0': {'osd_max_omap_entries_per_request': 10}, 'osd.1': {'osd_max_omap_entries_per_request': 1000}, 'osd.2': {'osd_max_omap_entries_per_request': 10000}}, 'flavor': 'default', 'log-ignorelist': ['\\(MDS_ALL_DOWN\\)', '\\(MDS_UP_LESS_THAN_MAX\\)', '\\(PG_AVAILABILITY\\)', '\\(PG_DEGRADED\\)', '\\(POOL_APP_NOT_ENABLED\\)', 'not have an application enabled'], 'sha1': '70f8415b300f041766fa27faf7d5472699e32388'}}, {'ceph': {'cluster': 'c2', 'conf': {'client': {'debug rgw': 20, 'debug rgw lifecycle': 20, 'debug rgw notification': 20, 'debug rgw sync': 20, 'rgw crypt require ssl': False, 'rgw crypt s3 kms backend': 'testing', 'rgw crypt s3 kms encryption keys': 'testkey-1=YmluCmJvb3N0CmJvb3N0LWJ1aWxkCmNlcGguY29uZgo=', 'rgw curl low speed time': 300, 'rgw data log num shards': 4, 'rgw data sync poll interval': 5, 'rgw md log max shards': 4, 'rgw meta sync poll interval': 5, 'rgw s3 auth use sts': True, 'rgw sts key': 'abcdefghijklmnoq', 'rgw sync data inject err probability': 0, 'rgw sync log trim interval': 0, 'rgw sync meta inject err probability': 0, 'rgw sync obj etag verify': True, 'setgroup': 'ceph', 'setuser': 'ceph'}, 'client.0': {'rgw data notify interval msec': 0}, 'mgr': {'debug mgr': 20, 'debug ms': 1}, 'mon': {'debug mon': 20, 'debug ms': 1, 'debug paxos': 20}, 'osd': {'debug ms': 1, 'debug osd': 20, 'osd mclock iops capacity threshold hdd': 49000, 'rocksdb delete range threshold': 0}, 'osd.0': {'osd_max_omap_entries_per_request': 10}, 'osd.1': {'osd_max_omap_entries_per_request': 1000}, 'osd.2': {'osd_max_omap_entries_per_request': 10000}}, 'flavor': 'default', 'log-ignorelist': ['\\(MDS_ALL_DOWN\\)', '\\(MDS_UP_LESS_THAN_MAX\\)', '\\(PG_AVAILABILITY\\)', '\\(PG_DEGRADED\\)', '\\(POOL_APP_NOT_ENABLED\\)', 'not have an application enabled'], 'sha1': '70f8415b300f041766fa27faf7d5472699e32388'}}, {'rgw': {'c1.client.0': {'port': 8000}, 'c1.client.1': {'port': 8001}, 'c2.client.0': {'port': 8000}, 'c2.client.1': {'port': 8001}}}, {'rgw-multisite': None}, {'rgw-multisite-tests': {'config': {'reconfigure_delay': 90}}}], 'teuthology': {'fragments_dropped': [], 'meta': {}, 'postmerge': []}, 'teuthology_branch': 'clyso-debian-13', 'teuthology_repo': 'https://github.com/clyso/teuthology', 'teuthology_sha1': '1c580df7a9c7c2aadc272da296344fd99f27c444', 'timestamp': '2026-03-20_17:25:16', 'tube': 'vps', 'user': 'kyr', 'verbose': False, 'worker_log': '/home/teuthos/.teuthology/dispatcher/dispatcher.vps.4188345'} 2026-03-20T17:57:28.396 DEBUG:teuthology.orchestra.run.vm02:> install -d -m0755 -- /home/ubuntu/cephtest/c2.data 2026-03-20T17:57:28.397 DEBUG:teuthology.orchestra.run.vm05:> install -d -m0755 -- /home/ubuntu/cephtest/c2.data 2026-03-20T17:57:28.399 DEBUG:teuthology.orchestra.run.vm06:> install -d -m0755 -- /home/ubuntu/cephtest/c2.data 2026-03-20T17:57:28.400 DEBUG:teuthology.orchestra.run.vm08:> install -d -m0755 -- /home/ubuntu/cephtest/c2.data 2026-03-20T17:57:28.403 DEBUG:teuthology.orchestra.run.vm02:> sudo install -d -m0777 -- /var/run/ceph 2026-03-20T17:57:28.443 DEBUG:teuthology.orchestra.run.vm05:> sudo install -d -m0777 -- /var/run/ceph 2026-03-20T17:57:28.445 DEBUG:teuthology.orchestra.run.vm06:> sudo install -d -m0777 -- /var/run/ceph 2026-03-20T17:57:28.445 DEBUG:teuthology.orchestra.run.vm08:> sudo install -d -m0777 -- /var/run/ceph 2026-03-20T17:57:28.455 DEBUG:teuthology.orchestra.run.vm06:> set -ex 2026-03-20T17:57:28.455 DEBUG:teuthology.orchestra.run.vm06:> dd if=/scratch_devs of=/dev/stdout 2026-03-20T17:57:28.498 DEBUG:teuthology.misc:devs=['/dev/vg_nvme/lv_1', '/dev/vg_nvme/lv_2', '/dev/vg_nvme/lv_3', '/dev/vg_nvme/lv_4'] 2026-03-20T17:57:28.499 DEBUG:teuthology.orchestra.run.vm06:> stat /dev/vg_nvme/lv_1 2026-03-20T17:57:28.543 INFO:teuthology.orchestra.run.vm06.stdout: File: /dev/vg_nvme/lv_1 -> ../dm-0 2026-03-20T17:57:28.543 INFO:teuthology.orchestra.run.vm06.stdout: Size: 7 Blocks: 0 IO Block: 4096 symbolic link 2026-03-20T17:57:28.543 INFO:teuthology.orchestra.run.vm06.stdout:Device: 5h/5d Inode: 778 Links: 1 2026-03-20T17:57:28.543 INFO:teuthology.orchestra.run.vm06.stdout:Access: (0777/lrwxrwxrwx) Uid: ( 0/ root) Gid: ( 0/ root) 2026-03-20T17:57:28.543 INFO:teuthology.orchestra.run.vm06.stdout:Access: 2026-03-20 17:55:52.108158000 +0000 2026-03-20T17:57:28.543 INFO:teuthology.orchestra.run.vm06.stdout:Modify: 2026-03-20 17:55:51.980158000 +0000 2026-03-20T17:57:28.543 INFO:teuthology.orchestra.run.vm06.stdout:Change: 2026-03-20 17:55:51.980158000 +0000 2026-03-20T17:57:28.543 INFO:teuthology.orchestra.run.vm06.stdout: Birth: - 2026-03-20T17:57:28.543 DEBUG:teuthology.orchestra.run.vm06:> sudo dd if=/dev/vg_nvme/lv_1 of=/dev/null count=1 2026-03-20T17:57:28.591 INFO:teuthology.orchestra.run.vm06.stderr:1+0 records in 2026-03-20T17:57:28.591 INFO:teuthology.orchestra.run.vm06.stderr:1+0 records out 2026-03-20T17:57:28.591 INFO:teuthology.orchestra.run.vm06.stderr:512 bytes copied, 0.000162464 s, 3.2 MB/s 2026-03-20T17:57:28.592 DEBUG:teuthology.orchestra.run.vm06:> ! mount | grep -v devtmpfs | grep -q /dev/vg_nvme/lv_1 2026-03-20T17:57:28.639 DEBUG:teuthology.orchestra.run.vm06:> stat /dev/vg_nvme/lv_2 2026-03-20T17:57:28.682 INFO:teuthology.orchestra.run.vm06.stdout: File: /dev/vg_nvme/lv_2 -> ../dm-1 2026-03-20T17:57:28.682 INFO:teuthology.orchestra.run.vm06.stdout: Size: 7 Blocks: 0 IO Block: 4096 symbolic link 2026-03-20T17:57:28.682 INFO:teuthology.orchestra.run.vm06.stdout:Device: 5h/5d Inode: 808 Links: 1 2026-03-20T17:57:28.682 INFO:teuthology.orchestra.run.vm06.stdout:Access: (0777/lrwxrwxrwx) Uid: ( 0/ root) Gid: ( 0/ root) 2026-03-20T17:57:28.682 INFO:teuthology.orchestra.run.vm06.stdout:Access: 2026-03-20 17:55:52.396158000 +0000 2026-03-20T17:57:28.682 INFO:teuthology.orchestra.run.vm06.stdout:Modify: 2026-03-20 17:55:52.260158000 +0000 2026-03-20T17:57:28.682 INFO:teuthology.orchestra.run.vm06.stdout:Change: 2026-03-20 17:55:52.260158000 +0000 2026-03-20T17:57:28.682 INFO:teuthology.orchestra.run.vm06.stdout: Birth: - 2026-03-20T17:57:28.682 DEBUG:teuthology.orchestra.run.vm06:> sudo dd if=/dev/vg_nvme/lv_2 of=/dev/null count=1 2026-03-20T17:57:28.731 INFO:teuthology.orchestra.run.vm06.stderr:1+0 records in 2026-03-20T17:57:28.731 INFO:teuthology.orchestra.run.vm06.stderr:1+0 records out 2026-03-20T17:57:28.731 INFO:teuthology.orchestra.run.vm06.stderr:512 bytes copied, 0.000161312 s, 3.2 MB/s 2026-03-20T17:57:28.732 DEBUG:teuthology.orchestra.run.vm06:> ! mount | grep -v devtmpfs | grep -q /dev/vg_nvme/lv_2 2026-03-20T17:57:28.775 DEBUG:teuthology.orchestra.run.vm06:> stat /dev/vg_nvme/lv_3 2026-03-20T17:57:28.818 INFO:teuthology.orchestra.run.vm06.stdout: File: /dev/vg_nvme/lv_3 -> ../dm-2 2026-03-20T17:57:28.818 INFO:teuthology.orchestra.run.vm06.stdout: Size: 7 Blocks: 0 IO Block: 4096 symbolic link 2026-03-20T17:57:28.818 INFO:teuthology.orchestra.run.vm06.stdout:Device: 5h/5d Inode: 837 Links: 1 2026-03-20T17:57:28.818 INFO:teuthology.orchestra.run.vm06.stdout:Access: (0777/lrwxrwxrwx) Uid: ( 0/ root) Gid: ( 0/ root) 2026-03-20T17:57:28.818 INFO:teuthology.orchestra.run.vm06.stdout:Access: 2026-03-20 17:55:52.540158000 +0000 2026-03-20T17:57:28.818 INFO:teuthology.orchestra.run.vm06.stdout:Modify: 2026-03-20 17:55:52.536158000 +0000 2026-03-20T17:57:28.818 INFO:teuthology.orchestra.run.vm06.stdout:Change: 2026-03-20 17:55:52.536158000 +0000 2026-03-20T17:57:28.818 INFO:teuthology.orchestra.run.vm06.stdout: Birth: - 2026-03-20T17:57:28.818 DEBUG:teuthology.orchestra.run.vm06:> sudo dd if=/dev/vg_nvme/lv_3 of=/dev/null count=1 2026-03-20T17:57:28.867 INFO:teuthology.orchestra.run.vm06.stderr:1+0 records in 2026-03-20T17:57:28.867 INFO:teuthology.orchestra.run.vm06.stderr:1+0 records out 2026-03-20T17:57:28.867 INFO:teuthology.orchestra.run.vm06.stderr:512 bytes copied, 0.000154749 s, 3.3 MB/s 2026-03-20T17:57:28.868 DEBUG:teuthology.orchestra.run.vm06:> ! mount | grep -v devtmpfs | grep -q /dev/vg_nvme/lv_3 2026-03-20T17:57:28.916 DEBUG:teuthology.orchestra.run.vm06:> stat /dev/vg_nvme/lv_4 2026-03-20T17:57:28.962 INFO:teuthology.orchestra.run.vm06.stdout: File: /dev/vg_nvme/lv_4 -> ../dm-3 2026-03-20T17:57:28.963 INFO:teuthology.orchestra.run.vm06.stdout: Size: 7 Blocks: 0 IO Block: 4096 symbolic link 2026-03-20T17:57:28.963 INFO:teuthology.orchestra.run.vm06.stdout:Device: 5h/5d Inode: 874 Links: 1 2026-03-20T17:57:28.963 INFO:teuthology.orchestra.run.vm06.stdout:Access: (0777/lrwxrwxrwx) Uid: ( 0/ root) Gid: ( 0/ root) 2026-03-20T17:57:28.963 INFO:teuthology.orchestra.run.vm06.stdout:Access: 2026-03-20 17:55:56.612158000 +0000 2026-03-20T17:57:28.963 INFO:teuthology.orchestra.run.vm06.stdout:Modify: 2026-03-20 17:55:52.844158000 +0000 2026-03-20T17:57:28.963 INFO:teuthology.orchestra.run.vm06.stdout:Change: 2026-03-20 17:55:52.844158000 +0000 2026-03-20T17:57:28.963 INFO:teuthology.orchestra.run.vm06.stdout: Birth: - 2026-03-20T17:57:28.963 DEBUG:teuthology.orchestra.run.vm06:> sudo dd if=/dev/vg_nvme/lv_4 of=/dev/null count=1 2026-03-20T17:57:29.010 INFO:teuthology.orchestra.run.vm06.stderr:1+0 records in 2026-03-20T17:57:29.018 INFO:teuthology.orchestra.run.vm06.stderr:1+0 records out 2026-03-20T17:57:29.018 INFO:teuthology.orchestra.run.vm06.stderr:512 bytes copied, 0.000172412 s, 3.0 MB/s 2026-03-20T17:57:29.019 DEBUG:teuthology.orchestra.run.vm06:> ! mount | grep -v devtmpfs | grep -q /dev/vg_nvme/lv_4 2026-03-20T17:57:29.064 INFO:tasks.ceph:osd dev map: {'c2.osd.0': '/dev/vg_nvme/lv_1', 'c2.osd.1': '/dev/vg_nvme/lv_2', 'c2.osd.2': '/dev/vg_nvme/lv_3'} 2026-03-20T17:57:29.064 DEBUG:teuthology.orchestra.run.vm08:> set -ex 2026-03-20T17:57:29.064 DEBUG:teuthology.orchestra.run.vm08:> dd if=/scratch_devs of=/dev/stdout 2026-03-20T17:57:29.068 DEBUG:teuthology.misc:devs=['/dev/vg_nvme/lv_1', '/dev/vg_nvme/lv_2', '/dev/vg_nvme/lv_3', '/dev/vg_nvme/lv_4'] 2026-03-20T17:57:29.068 DEBUG:teuthology.orchestra.run.vm08:> stat /dev/vg_nvme/lv_1 2026-03-20T17:57:29.113 INFO:teuthology.orchestra.run.vm08.stdout: File: /dev/vg_nvme/lv_1 -> ../dm-0 2026-03-20T17:57:29.113 INFO:teuthology.orchestra.run.vm08.stdout: Size: 7 Blocks: 0 IO Block: 4096 symbolic link 2026-03-20T17:57:29.113 INFO:teuthology.orchestra.run.vm08.stdout:Device: 5h/5d Inode: 781 Links: 1 2026-03-20T17:57:29.113 INFO:teuthology.orchestra.run.vm08.stdout:Access: (0777/lrwxrwxrwx) Uid: ( 0/ root) Gid: ( 0/ root) 2026-03-20T17:57:29.113 INFO:teuthology.orchestra.run.vm08.stdout:Access: 2026-03-20 17:55:44.849870000 +0000 2026-03-20T17:57:29.113 INFO:teuthology.orchestra.run.vm08.stdout:Modify: 2026-03-20 17:55:44.845870000 +0000 2026-03-20T17:57:29.113 INFO:teuthology.orchestra.run.vm08.stdout:Change: 2026-03-20 17:55:44.845870000 +0000 2026-03-20T17:57:29.113 INFO:teuthology.orchestra.run.vm08.stdout: Birth: - 2026-03-20T17:57:29.113 DEBUG:teuthology.orchestra.run.vm08:> sudo dd if=/dev/vg_nvme/lv_1 of=/dev/null count=1 2026-03-20T17:57:29.161 INFO:teuthology.orchestra.run.vm08.stderr:1+0 records in 2026-03-20T17:57:29.161 INFO:teuthology.orchestra.run.vm08.stderr:1+0 records out 2026-03-20T17:57:29.161 INFO:teuthology.orchestra.run.vm08.stderr:512 bytes copied, 0.00021365 s, 2.4 MB/s 2026-03-20T17:57:29.162 DEBUG:teuthology.orchestra.run.vm08:> ! mount | grep -v devtmpfs | grep -q /dev/vg_nvme/lv_1 2026-03-20T17:57:29.206 DEBUG:teuthology.orchestra.run.vm08:> stat /dev/vg_nvme/lv_2 2026-03-20T17:57:29.249 INFO:teuthology.orchestra.run.vm08.stdout: File: /dev/vg_nvme/lv_2 -> ../dm-1 2026-03-20T17:57:29.249 INFO:teuthology.orchestra.run.vm08.stdout: Size: 7 Blocks: 0 IO Block: 4096 symbolic link 2026-03-20T17:57:29.249 INFO:teuthology.orchestra.run.vm08.stdout:Device: 5h/5d Inode: 820 Links: 1 2026-03-20T17:57:29.249 INFO:teuthology.orchestra.run.vm08.stdout:Access: (0777/lrwxrwxrwx) Uid: ( 0/ root) Gid: ( 0/ root) 2026-03-20T17:57:29.249 INFO:teuthology.orchestra.run.vm08.stdout:Access: 2026-03-20 17:55:45.281870000 +0000 2026-03-20T17:57:29.249 INFO:teuthology.orchestra.run.vm08.stdout:Modify: 2026-03-20 17:55:45.145870000 +0000 2026-03-20T17:57:29.249 INFO:teuthology.orchestra.run.vm08.stdout:Change: 2026-03-20 17:55:45.145870000 +0000 2026-03-20T17:57:29.249 INFO:teuthology.orchestra.run.vm08.stdout: Birth: - 2026-03-20T17:57:29.249 DEBUG:teuthology.orchestra.run.vm08:> sudo dd if=/dev/vg_nvme/lv_2 of=/dev/null count=1 2026-03-20T17:57:29.297 INFO:teuthology.orchestra.run.vm08.stderr:1+0 records in 2026-03-20T17:57:29.297 INFO:teuthology.orchestra.run.vm08.stderr:1+0 records out 2026-03-20T17:57:29.297 INFO:teuthology.orchestra.run.vm08.stderr:512 bytes copied, 0.000200755 s, 2.6 MB/s 2026-03-20T17:57:29.297 DEBUG:teuthology.orchestra.run.vm08:> ! mount | grep -v devtmpfs | grep -q /dev/vg_nvme/lv_2 2026-03-20T17:57:29.342 DEBUG:teuthology.orchestra.run.vm08:> stat /dev/vg_nvme/lv_3 2026-03-20T17:57:29.389 INFO:teuthology.orchestra.run.vm08.stdout: File: /dev/vg_nvme/lv_3 -> ../dm-2 2026-03-20T17:57:29.389 INFO:teuthology.orchestra.run.vm08.stdout: Size: 7 Blocks: 0 IO Block: 4096 symbolic link 2026-03-20T17:57:29.389 INFO:teuthology.orchestra.run.vm08.stdout:Device: 5h/5d Inode: 849 Links: 1 2026-03-20T17:57:29.389 INFO:teuthology.orchestra.run.vm08.stdout:Access: (0777/lrwxrwxrwx) Uid: ( 0/ root) Gid: ( 0/ root) 2026-03-20T17:57:29.389 INFO:teuthology.orchestra.run.vm08.stdout:Access: 2026-03-20 17:55:45.433870000 +0000 2026-03-20T17:57:29.389 INFO:teuthology.orchestra.run.vm08.stdout:Modify: 2026-03-20 17:55:45.429870000 +0000 2026-03-20T17:57:29.389 INFO:teuthology.orchestra.run.vm08.stdout:Change: 2026-03-20 17:55:45.429870000 +0000 2026-03-20T17:57:29.389 INFO:teuthology.orchestra.run.vm08.stdout: Birth: - 2026-03-20T17:57:29.389 DEBUG:teuthology.orchestra.run.vm08:> sudo dd if=/dev/vg_nvme/lv_3 of=/dev/null count=1 2026-03-20T17:57:29.436 INFO:teuthology.orchestra.run.vm08.stderr:1+0 records in 2026-03-20T17:57:29.436 INFO:teuthology.orchestra.run.vm08.stderr:1+0 records out 2026-03-20T17:57:29.436 INFO:teuthology.orchestra.run.vm08.stderr:512 bytes copied, 0.000178534 s, 2.9 MB/s 2026-03-20T17:57:29.437 DEBUG:teuthology.orchestra.run.vm08:> ! mount | grep -v devtmpfs | grep -q /dev/vg_nvme/lv_3 2026-03-20T17:57:29.482 DEBUG:teuthology.orchestra.run.vm08:> stat /dev/vg_nvme/lv_4 2026-03-20T17:57:29.529 INFO:teuthology.orchestra.run.vm08.stdout: File: /dev/vg_nvme/lv_4 -> ../dm-3 2026-03-20T17:57:29.529 INFO:teuthology.orchestra.run.vm08.stdout: Size: 7 Blocks: 0 IO Block: 4096 symbolic link 2026-03-20T17:57:29.529 INFO:teuthology.orchestra.run.vm08.stdout:Device: 5h/5d Inode: 884 Links: 1 2026-03-20T17:57:29.529 INFO:teuthology.orchestra.run.vm08.stdout:Access: (0777/lrwxrwxrwx) Uid: ( 0/ root) Gid: ( 0/ root) 2026-03-20T17:57:29.529 INFO:teuthology.orchestra.run.vm08.stdout:Access: 2026-03-20 17:55:56.649870000 +0000 2026-03-20T17:57:29.529 INFO:teuthology.orchestra.run.vm08.stdout:Modify: 2026-03-20 17:55:45.729870000 +0000 2026-03-20T17:57:29.529 INFO:teuthology.orchestra.run.vm08.stdout:Change: 2026-03-20 17:55:45.729870000 +0000 2026-03-20T17:57:29.529 INFO:teuthology.orchestra.run.vm08.stdout: Birth: - 2026-03-20T17:57:29.530 DEBUG:teuthology.orchestra.run.vm08:> sudo dd if=/dev/vg_nvme/lv_4 of=/dev/null count=1 2026-03-20T17:57:29.578 INFO:teuthology.orchestra.run.vm08.stderr:1+0 records in 2026-03-20T17:57:29.578 INFO:teuthology.orchestra.run.vm08.stderr:1+0 records out 2026-03-20T17:57:29.578 INFO:teuthology.orchestra.run.vm08.stderr:512 bytes copied, 0.000211095 s, 2.4 MB/s 2026-03-20T17:57:29.579 DEBUG:teuthology.orchestra.run.vm08:> ! mount | grep -v devtmpfs | grep -q /dev/vg_nvme/lv_4 2026-03-20T17:57:29.626 INFO:tasks.ceph:osd dev map: {'c2.osd.3': '/dev/vg_nvme/lv_1', 'c2.osd.4': '/dev/vg_nvme/lv_2', 'c2.osd.5': '/dev/vg_nvme/lv_3'} 2026-03-20T17:57:29.626 INFO:tasks.ceph:remote_to_roles_to_devs: {Remote(name='ubuntu@vm06.local'): {'c2.osd.0': '/dev/vg_nvme/lv_1', 'c2.osd.1': '/dev/vg_nvme/lv_2', 'c2.osd.2': '/dev/vg_nvme/lv_3'}, Remote(name='ubuntu@vm08.local'): {'c2.osd.3': '/dev/vg_nvme/lv_1', 'c2.osd.4': '/dev/vg_nvme/lv_2', 'c2.osd.5': '/dev/vg_nvme/lv_3'}} 2026-03-20T17:57:29.627 INFO:tasks.ceph:Generating config... 2026-03-20T17:57:29.627 INFO:tasks.ceph:[client] debug rgw = 20 2026-03-20T17:57:29.627 INFO:tasks.ceph:[client] debug rgw lifecycle = 20 2026-03-20T17:57:29.627 INFO:tasks.ceph:[client] debug rgw notification = 20 2026-03-20T17:57:29.627 INFO:tasks.ceph:[client] debug rgw sync = 20 2026-03-20T17:57:29.627 INFO:tasks.ceph:[client] rgw crypt require ssl = False 2026-03-20T17:57:29.627 INFO:tasks.ceph:[client] rgw crypt s3 kms backend = testing 2026-03-20T17:57:29.627 INFO:tasks.ceph:[client] rgw crypt s3 kms encryption keys = testkey-1=YmluCmJvb3N0CmJvb3N0LWJ1aWxkCmNlcGguY29uZgo= 2026-03-20T17:57:29.627 INFO:tasks.ceph:[client] rgw curl low speed time = 300 2026-03-20T17:57:29.627 INFO:tasks.ceph:[client] rgw data log num shards = 4 2026-03-20T17:57:29.627 INFO:tasks.ceph:[client] rgw data sync poll interval = 5 2026-03-20T17:57:29.627 INFO:tasks.ceph:[client] rgw md log max shards = 4 2026-03-20T17:57:29.627 INFO:tasks.ceph:[client] rgw meta sync poll interval = 5 2026-03-20T17:57:29.627 INFO:tasks.ceph:[client] rgw s3 auth use sts = True 2026-03-20T17:57:29.627 INFO:tasks.ceph:[client] rgw sts key = abcdefghijklmnoq 2026-03-20T17:57:29.627 INFO:tasks.ceph:[client] rgw sync data inject err probability = 0 2026-03-20T17:57:29.627 INFO:tasks.ceph:[client] rgw sync log trim interval = 0 2026-03-20T17:57:29.627 INFO:tasks.ceph:[client] rgw sync meta inject err probability = 0 2026-03-20T17:57:29.627 INFO:tasks.ceph:[client] rgw sync obj etag verify = True 2026-03-20T17:57:29.627 INFO:tasks.ceph:[client] setgroup = ceph 2026-03-20T17:57:29.627 INFO:tasks.ceph:[client] setuser = ceph 2026-03-20T17:57:29.627 INFO:tasks.ceph:[client.0] rgw data notify interval msec = 0 2026-03-20T17:57:29.627 INFO:tasks.ceph:[mgr] debug mgr = 20 2026-03-20T17:57:29.627 INFO:tasks.ceph:[mgr] debug ms = 1 2026-03-20T17:57:29.627 INFO:tasks.ceph:[mon] debug mon = 20 2026-03-20T17:57:29.627 INFO:tasks.ceph:[mon] debug ms = 1 2026-03-20T17:57:29.627 INFO:tasks.ceph:[mon] debug paxos = 20 2026-03-20T17:57:29.627 INFO:tasks.ceph:[osd] debug ms = 1 2026-03-20T17:57:29.627 INFO:tasks.ceph:[osd] debug osd = 20 2026-03-20T17:57:29.627 INFO:tasks.ceph:[osd] osd mclock iops capacity threshold hdd = 49000 2026-03-20T17:57:29.627 INFO:tasks.ceph:[osd] rocksdb delete range threshold = 0 2026-03-20T17:57:29.627 INFO:tasks.ceph:[osd.0] osd_max_omap_entries_per_request = 10 2026-03-20T17:57:29.627 INFO:tasks.ceph:[osd.1] osd_max_omap_entries_per_request = 1000 2026-03-20T17:57:29.627 INFO:tasks.ceph:[osd.2] osd_max_omap_entries_per_request = 10000 2026-03-20T17:57:29.628 INFO:tasks.ceph:Setting up c2.mon.a... 2026-03-20T17:57:29.628 DEBUG:teuthology.orchestra.run.vm06:> sudo adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage ceph-authtool --create-keyring /etc/ceph/c2.keyring 2026-03-20T17:57:29.645 INFO:teuthology.orchestra.run.vm06.stdout:creating /etc/ceph/c2.keyring 2026-03-20T17:57:29.647 DEBUG:teuthology.orchestra.run.vm06:> sudo adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage ceph-authtool --gen-key --name=mon. /etc/ceph/c2.keyring 2026-03-20T17:57:29.707 DEBUG:teuthology.orchestra.run.vm06:> sudo chmod 0644 /etc/ceph/c2.keyring 2026-03-20T17:57:29.755 DEBUG:tasks.ceph:Ceph mon addresses: [('c2.mon.a', '192.168.123.106')] 2026-03-20T17:57:29.755 DEBUG:tasks.ceph:writing out conf {'global': {'chdir': '', 'pid file': '/var/run/ceph/$cluster-$name.pid', 'auth supported': 'cephx', 'filestore xattr use omap': 'true', 'mon clock drift allowed': '1.000', 'osd crush chooseleaf type': '0', 'auth debug': 'true', 'ms die on old message': 'true', 'ms die on bug': 'true', 'mon max pg per osd': '10000', 'mon pg warn max object skew': '0', 'osd_pool_default_pg_autoscale_mode': 'off', 'osd pool default size': '2', 'mon osd allow primary affinity': 'true', 'mon osd allow pg remap': 'true', 'mon warn on legacy crush tunables': 'false', 'mon warn on crush straw calc version zero': 'false', 'mon warn on no sortbitwise': 'false', 'mon warn on osd down out interval zero': 'false', 'mon warn on too few osds': 'false', 'mon_warn_on_pool_pg_num_not_power_of_two': 'false', 'mon_warn_on_pool_no_redundancy': 'false', 'mon_allow_pool_size_one': 'true', 'osd pool default erasure code profile': 'plugin=isa technique=reed_sol_van k=2 m=1 crush-failure-domain=osd', 'osd default data pool replay window': '5', 'mon allow pool delete': 'true', 'mon cluster log file level': 'debug', 'debug asserts on shutdown': 'true', 'mon health detail to clog': 'false', 'mon host': '192.168.123.106'}, 'osd': {'osd journal size': '100', 'osd scrub load threshold': '5.0', 'osd scrub max interval': '600', 'osd mclock profile': 'high_recovery_ops', 'osd mclock skip benchmark': 'true', 'osd recover clone overlap': 'true', 'osd recovery max chunk': '1048576', 'osd debug shutdown': 'true', 'osd debug op order': 'true', 'osd debug verify stray on activate': 'true', 'osd debug trim objects': 'true', 'osd open classes on start': 'true', 'osd debug pg log writeout': 'true', 'osd deep scrub update digest min age': '30', 'osd map max advance': '10', 'journal zero on create': 'true', 'filestore ondisk finisher threads': '3', 'filestore apply finisher threads': '3', 'bdev debug aio': 'true', 'osd debug misdirected ops': 'true', 'debug ms': 1, 'debug osd': 20, 'osd mclock iops capacity threshold hdd': 49000, 'rocksdb delete range threshold': 0}, 'mgr': {'debug ms': 1, 'debug mgr': 20, 'debug mon': '20', 'debug auth': '20', 'mon reweight min pgs per osd': '4', 'mon reweight min bytes per osd': '10', 'mgr/telemetry/nag': 'false'}, 'mon': {'debug ms': 1, 'debug mon': 20, 'debug paxos': 20, 'debug auth': '20', 'mon data avail warn': '5', 'mon mgr mkfs grace': '240', 'mon reweight min pgs per osd': '4', 'mon osd reporter subtree level': 'osd', 'mon osd prime pg temp': 'true', 'mon reweight min bytes per osd': '10', 'auth mon ticket ttl': '660', 'auth service ticket ttl': '240', 'mon_warn_on_insecure_global_id_reclaim': 'false', 'mon_warn_on_insecure_global_id_reclaim_allowed': 'false', 'mon_down_mkfs_grace': '2m', 'mon_warn_on_filestore_osds': 'false'}, 'client': {'rgw cache enabled': 'true', 'rgw enable ops log': 'true', 'rgw enable usage log': 'true', 'log file': '/var/log/ceph/$cluster-$name.$pid.log', 'admin socket': '/var/run/ceph/$cluster-$name.$pid.asok', 'debug rgw': 20, 'debug rgw lifecycle': 20, 'debug rgw notification': 20, 'debug rgw sync': 20, 'rgw crypt require ssl': False, 'rgw crypt s3 kms backend': 'testing', 'rgw crypt s3 kms encryption keys': 'testkey-1=YmluCmJvb3N0CmJvb3N0LWJ1aWxkCmNlcGguY29uZgo=', 'rgw curl low speed time': 300, 'rgw data log num shards': 4, 'rgw data sync poll interval': 5, 'rgw md log max shards': 4, 'rgw meta sync poll interval': 5, 'rgw s3 auth use sts': True, 'rgw sts key': 'abcdefghijklmnoq', 'rgw sync data inject err probability': 0, 'rgw sync log trim interval': 0, 'rgw sync meta inject err probability': 0, 'rgw sync obj etag verify': True, 'setgroup': 'ceph', 'setuser': 'ceph'}, 'mon.a': {}, 'client.0': {'rgw data notify interval msec': 0}, 'osd.0': {'osd_max_omap_entries_per_request': 10}, 'osd.1': {'osd_max_omap_entries_per_request': 1000}, 'osd.2': {'osd_max_omap_entries_per_request': 10000}} 2026-03-20T17:57:29.756 DEBUG:teuthology.orchestra.run.vm06:> set -ex 2026-03-20T17:57:29.756 DEBUG:teuthology.orchestra.run.vm06:> dd of=/home/ubuntu/cephtest/ceph.tmp.conf 2026-03-20T17:57:29.799 DEBUG:teuthology.orchestra.run.vm06:> adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage monmaptool -c /home/ubuntu/cephtest/ceph.tmp.conf --create --clobber --enable-all-features --add a 192.168.123.106 --print /home/ubuntu/cephtest/c2.monmap 2026-03-20T17:57:29.856 INFO:teuthology.orchestra.run.vm06.stderr:ignoring --setuser ceph since I am not root 2026-03-20T17:57:29.856 INFO:teuthology.orchestra.run.vm06.stderr:ignoring --setgroup ceph since I am not root 2026-03-20T17:57:29.856 INFO:teuthology.orchestra.run.vm06.stdout:monmaptool: monmap file /home/ubuntu/cephtest/c2.monmap 2026-03-20T17:57:29.856 INFO:teuthology.orchestra.run.vm06.stdout:monmaptool: generated fsid 02f24f96-dd4e-4d53-915a-6d7df6cbe6f6 2026-03-20T17:57:29.856 INFO:teuthology.orchestra.run.vm06.stdout:setting min_mon_release = tentacle 2026-03-20T17:57:29.856 INFO:teuthology.orchestra.run.vm06.stdout:epoch 0 2026-03-20T17:57:29.857 INFO:teuthology.orchestra.run.vm06.stdout:fsid 02f24f96-dd4e-4d53-915a-6d7df6cbe6f6 2026-03-20T17:57:29.857 INFO:teuthology.orchestra.run.vm06.stdout:last_changed 2026-03-20T17:57:29.855690+0000 2026-03-20T17:57:29.857 INFO:teuthology.orchestra.run.vm06.stdout:created 2026-03-20T17:57:29.855690+0000 2026-03-20T17:57:29.857 INFO:teuthology.orchestra.run.vm06.stdout:min_mon_release 20 (tentacle) 2026-03-20T17:57:29.857 INFO:teuthology.orchestra.run.vm06.stdout:election_strategy: 1 2026-03-20T17:57:29.857 INFO:teuthology.orchestra.run.vm06.stdout:0: [v2:192.168.123.106:3300/0,v1:192.168.123.106:6789/0] mon.a 2026-03-20T17:57:29.857 INFO:teuthology.orchestra.run.vm06.stdout:monmaptool: writing epoch 0 to /home/ubuntu/cephtest/c2.monmap (1 monitors) 2026-03-20T17:57:29.858 DEBUG:teuthology.orchestra.run.vm06:> rm -- /home/ubuntu/cephtest/ceph.tmp.conf 2026-03-20T17:57:29.903 INFO:tasks.ceph:Writing /etc/ceph/c2.conf for FSID 02f24f96-dd4e-4d53-915a-6d7df6cbe6f6... 2026-03-20T17:57:29.903 DEBUG:teuthology.orchestra.run.vm02:> sudo mkdir -p /etc/ceph && sudo chmod 0755 /etc/ceph && sudo tee /etc/ceph/c2.conf && sudo chmod 0644 /etc/ceph/c2.conf > /dev/null 2026-03-20T17:57:29.905 DEBUG:teuthology.orchestra.run.vm05:> sudo mkdir -p /etc/ceph && sudo chmod 0755 /etc/ceph && sudo tee /etc/ceph/c2.conf && sudo chmod 0644 /etc/ceph/c2.conf > /dev/null 2026-03-20T17:57:29.906 DEBUG:teuthology.orchestra.run.vm06:> sudo mkdir -p /etc/ceph && sudo chmod 0755 /etc/ceph && sudo tee /etc/ceph/c2.conf && sudo chmod 0644 /etc/ceph/c2.conf > /dev/null 2026-03-20T17:57:29.945 DEBUG:teuthology.orchestra.run.vm08:> sudo mkdir -p /etc/ceph && sudo chmod 0755 /etc/ceph && sudo tee /etc/ceph/c2.conf && sudo chmod 0644 /etc/ceph/c2.conf > /dev/null 2026-03-20T17:57:29.946 INFO:teuthology.orchestra.run.vm05.stdout:[global] 2026-03-20T17:57:29.947 INFO:teuthology.orchestra.run.vm05.stdout: chdir = "" 2026-03-20T17:57:29.947 INFO:teuthology.orchestra.run.vm05.stdout: pid file = /var/run/ceph/$cluster-$name.pid 2026-03-20T17:57:29.947 INFO:teuthology.orchestra.run.vm05.stdout: auth supported = cephx 2026-03-20T17:57:29.947 INFO:teuthology.orchestra.run.vm05.stdout: 2026-03-20T17:57:29.947 INFO:teuthology.orchestra.run.vm05.stdout: filestore xattr use omap = true 2026-03-20T17:57:29.947 INFO:teuthology.orchestra.run.vm05.stdout: 2026-03-20T17:57:29.947 INFO:teuthology.orchestra.run.vm05.stdout: mon clock drift allowed = 1.000 2026-03-20T17:57:29.947 INFO:teuthology.orchestra.run.vm05.stdout: 2026-03-20T17:57:29.947 INFO:teuthology.orchestra.run.vm05.stdout: osd crush chooseleaf type = 0 2026-03-20T17:57:29.947 INFO:teuthology.orchestra.run.vm05.stdout: auth debug = true 2026-03-20T17:57:29.947 INFO:teuthology.orchestra.run.vm05.stdout: 2026-03-20T17:57:29.947 INFO:teuthology.orchestra.run.vm05.stdout: ms die on old message = true 2026-03-20T17:57:29.947 INFO:teuthology.orchestra.run.vm05.stdout: ms die on bug = true 2026-03-20T17:57:29.947 INFO:teuthology.orchestra.run.vm05.stdout: 2026-03-20T17:57:29.947 INFO:teuthology.orchestra.run.vm05.stdout: mon max pg per osd = 10000 # >= luminous 2026-03-20T17:57:29.947 INFO:teuthology.orchestra.run.vm05.stdout: mon pg warn max object skew = 0 2026-03-20T17:57:29.947 INFO:teuthology.orchestra.run.vm05.stdout: 2026-03-20T17:57:29.947 INFO:teuthology.orchestra.run.vm05.stdout: # disable pg_autoscaler by default for new pools 2026-03-20T17:57:29.947 INFO:teuthology.orchestra.run.vm05.stdout: osd_pool_default_pg_autoscale_mode = off 2026-03-20T17:57:29.947 INFO:teuthology.orchestra.run.vm05.stdout: 2026-03-20T17:57:29.947 INFO:teuthology.orchestra.run.vm05.stdout: osd pool default size = 2 2026-03-20T17:57:29.947 INFO:teuthology.orchestra.run.vm05.stdout: 2026-03-20T17:57:29.947 INFO:teuthology.orchestra.run.vm05.stdout: mon osd allow primary affinity = true 2026-03-20T17:57:29.947 INFO:teuthology.orchestra.run.vm05.stdout: mon osd allow pg remap = true 2026-03-20T17:57:29.947 INFO:teuthology.orchestra.run.vm05.stdout: mon warn on legacy crush tunables = false 2026-03-20T17:57:29.947 INFO:teuthology.orchestra.run.vm05.stdout: mon warn on crush straw calc version zero = false 2026-03-20T17:57:29.947 INFO:teuthology.orchestra.run.vm05.stdout: mon warn on no sortbitwise = false 2026-03-20T17:57:29.947 INFO:teuthology.orchestra.run.vm05.stdout: mon warn on osd down out interval zero = false 2026-03-20T17:57:29.947 INFO:teuthology.orchestra.run.vm05.stdout: mon warn on too few osds = false 2026-03-20T17:57:29.947 INFO:teuthology.orchestra.run.vm05.stdout: mon_warn_on_pool_pg_num_not_power_of_two = false 2026-03-20T17:57:29.947 INFO:teuthology.orchestra.run.vm05.stdout: mon_warn_on_pool_no_redundancy = false 2026-03-20T17:57:29.947 INFO:teuthology.orchestra.run.vm05.stdout: mon_allow_pool_size_one = true 2026-03-20T17:57:29.947 INFO:teuthology.orchestra.run.vm05.stdout: 2026-03-20T17:57:29.947 INFO:teuthology.orchestra.run.vm05.stdout: osd pool default erasure code profile = plugin=isa technique=reed_sol_van k=2 m=1 crush-failure-domain=osd 2026-03-20T17:57:29.947 INFO:teuthology.orchestra.run.vm05.stdout: 2026-03-20T17:57:29.947 INFO:teuthology.orchestra.run.vm05.stdout: osd default data pool replay window = 5 2026-03-20T17:57:29.947 INFO:teuthology.orchestra.run.vm05.stdout: 2026-03-20T17:57:29.947 INFO:teuthology.orchestra.run.vm05.stdout: mon allow pool delete = true 2026-03-20T17:57:29.947 INFO:teuthology.orchestra.run.vm05.stdout: 2026-03-20T17:57:29.947 INFO:teuthology.orchestra.run.vm05.stdout: mon cluster log file level = debug 2026-03-20T17:57:29.947 INFO:teuthology.orchestra.run.vm05.stdout: debug asserts on shutdown = true 2026-03-20T17:57:29.947 INFO:teuthology.orchestra.run.vm05.stdout: mon health detail to clog = false 2026-03-20T17:57:29.947 INFO:teuthology.orchestra.run.vm05.stdout: mon host = 192.168.123.106 2026-03-20T17:57:29.947 INFO:teuthology.orchestra.run.vm05.stdout: fsid = 02f24f96-dd4e-4d53-915a-6d7df6cbe6f6 2026-03-20T17:57:29.947 INFO:teuthology.orchestra.run.vm05.stdout: 2026-03-20T17:57:29.947 INFO:teuthology.orchestra.run.vm05.stdout:[osd] 2026-03-20T17:57:29.947 INFO:teuthology.orchestra.run.vm05.stdout: osd journal size = 100 2026-03-20T17:57:29.947 INFO:teuthology.orchestra.run.vm05.stdout: 2026-03-20T17:57:29.947 INFO:teuthology.orchestra.run.vm05.stdout: osd scrub load threshold = 5.0 2026-03-20T17:57:29.947 INFO:teuthology.orchestra.run.vm05.stdout: osd scrub max interval = 600 2026-03-20T17:57:29.947 INFO:teuthology.orchestra.run.vm05.stdout: osd mclock profile = high_recovery_ops 2026-03-20T17:57:29.947 INFO:teuthology.orchestra.run.vm05.stdout: osd mclock skip benchmark = true 2026-03-20T17:57:29.947 INFO:teuthology.orchestra.run.vm05.stdout: 2026-03-20T17:57:29.948 INFO:teuthology.orchestra.run.vm05.stdout: osd recover clone overlap = true 2026-03-20T17:57:29.948 INFO:teuthology.orchestra.run.vm05.stdout: osd recovery max chunk = 1048576 2026-03-20T17:57:29.948 INFO:teuthology.orchestra.run.vm05.stdout: 2026-03-20T17:57:29.948 INFO:teuthology.orchestra.run.vm05.stdout: osd debug shutdown = true 2026-03-20T17:57:29.948 INFO:teuthology.orchestra.run.vm05.stdout: osd debug op order = true 2026-03-20T17:57:29.948 INFO:teuthology.orchestra.run.vm05.stdout: osd debug verify stray on activate = true 2026-03-20T17:57:29.948 INFO:teuthology.orchestra.run.vm05.stdout: osd debug trim objects = true 2026-03-20T17:57:29.948 INFO:teuthology.orchestra.run.vm05.stdout: 2026-03-20T17:57:29.948 INFO:teuthology.orchestra.run.vm05.stdout: osd open classes on start = true 2026-03-20T17:57:29.948 INFO:teuthology.orchestra.run.vm05.stdout: osd debug pg log writeout = true 2026-03-20T17:57:29.948 INFO:teuthology.orchestra.run.vm05.stdout: 2026-03-20T17:57:29.948 INFO:teuthology.orchestra.run.vm05.stdout: osd deep scrub update digest min age = 30 2026-03-20T17:57:29.948 INFO:teuthology.orchestra.run.vm05.stdout: 2026-03-20T17:57:29.948 INFO:teuthology.orchestra.run.vm05.stdout: osd map max advance = 10 2026-03-20T17:57:29.948 INFO:teuthology.orchestra.run.vm05.stdout: 2026-03-20T17:57:29.948 INFO:teuthology.orchestra.run.vm05.stdout: journal zero on create = true 2026-03-20T17:57:29.948 INFO:teuthology.orchestra.run.vm05.stdout: 2026-03-20T17:57:29.948 INFO:teuthology.orchestra.run.vm05.stdout: filestore ondisk finisher threads = 3 2026-03-20T17:57:29.948 INFO:teuthology.orchestra.run.vm05.stdout: filestore apply finisher threads = 3 2026-03-20T17:57:29.948 INFO:teuthology.orchestra.run.vm05.stdout: 2026-03-20T17:57:29.948 INFO:teuthology.orchestra.run.vm05.stdout: bdev debug aio = true 2026-03-20T17:57:29.948 INFO:teuthology.orchestra.run.vm05.stdout: osd debug misdirected ops = true 2026-03-20T17:57:29.948 INFO:teuthology.orchestra.run.vm05.stdout: debug ms = 1 2026-03-20T17:57:29.948 INFO:teuthology.orchestra.run.vm05.stdout: debug osd = 20 2026-03-20T17:57:29.948 INFO:teuthology.orchestra.run.vm05.stdout: osd mclock iops capacity threshold hdd = 49000 2026-03-20T17:57:29.948 INFO:teuthology.orchestra.run.vm05.stdout: rocksdb delete range threshold = 0 2026-03-20T17:57:29.948 INFO:teuthology.orchestra.run.vm05.stdout: 2026-03-20T17:57:29.948 INFO:teuthology.orchestra.run.vm05.stdout:[mgr] 2026-03-20T17:57:29.948 INFO:teuthology.orchestra.run.vm05.stdout: debug ms = 1 2026-03-20T17:57:29.948 INFO:teuthology.orchestra.run.vm05.stdout: debug mgr = 20 2026-03-20T17:57:29.948 INFO:teuthology.orchestra.run.vm05.stdout: debug mon = 20 2026-03-20T17:57:29.948 INFO:teuthology.orchestra.run.vm05.stdout: debug auth = 20 2026-03-20T17:57:29.948 INFO:teuthology.orchestra.run.vm05.stdout: mon reweight min pgs per osd = 4 2026-03-20T17:57:29.948 INFO:teuthology.orchestra.run.vm05.stdout: mon reweight min bytes per osd = 10 2026-03-20T17:57:29.948 INFO:teuthology.orchestra.run.vm05.stdout: mgr/telemetry/nag = false 2026-03-20T17:57:29.948 INFO:teuthology.orchestra.run.vm05.stdout: 2026-03-20T17:57:29.948 INFO:teuthology.orchestra.run.vm05.stdout:[mon] 2026-03-20T17:57:29.948 INFO:teuthology.orchestra.run.vm05.stdout: debug ms = 1 2026-03-20T17:57:29.948 INFO:teuthology.orchestra.run.vm05.stdout: debug mon = 20 2026-03-20T17:57:29.948 INFO:teuthology.orchestra.run.vm05.stdout: debug paxos = 20 2026-03-20T17:57:29.948 INFO:teuthology.orchestra.run.vm05.stdout: debug auth = 20 2026-03-20T17:57:29.948 INFO:teuthology.orchestra.run.vm05.stdout: mon data avail warn = 5 2026-03-20T17:57:29.948 INFO:teuthology.orchestra.run.vm05.stdout: mon mgr mkfs grace = 240 2026-03-20T17:57:29.948 INFO:teuthology.orchestra.run.vm05.stdout: mon reweight min pgs per osd = 4 2026-03-20T17:57:29.948 INFO:teuthology.orchestra.run.vm05.stdout: mon osd reporter subtree level = osd 2026-03-20T17:57:29.948 INFO:teuthology.orchestra.run.vm05.stdout: mon osd prime pg temp = true 2026-03-20T17:57:29.948 INFO:teuthology.orchestra.run.vm05.stdout: mon reweight min bytes per osd = 10 2026-03-20T17:57:29.948 INFO:teuthology.orchestra.run.vm05.stdout: 2026-03-20T17:57:29.948 INFO:teuthology.orchestra.run.vm05.stdout: # rotate auth tickets quickly to exercise renewal paths 2026-03-20T17:57:29.948 INFO:teuthology.orchestra.run.vm05.stdout: auth mon ticket ttl = 660 # 11m 2026-03-20T17:57:29.948 INFO:teuthology.orchestra.run.vm05.stdout: auth service ticket ttl = 240 # 4m 2026-03-20T17:57:29.948 INFO:teuthology.orchestra.run.vm05.stdout: 2026-03-20T17:57:29.949 INFO:teuthology.orchestra.run.vm05.stdout: # don't complain about insecure global_id in the test suite 2026-03-20T17:57:29.949 INFO:teuthology.orchestra.run.vm05.stdout: mon_warn_on_insecure_global_id_reclaim = false 2026-03-20T17:57:29.949 INFO:teuthology.orchestra.run.vm05.stdout: mon_warn_on_insecure_global_id_reclaim_allowed = false 2026-03-20T17:57:29.949 INFO:teuthology.orchestra.run.vm05.stdout: 2026-03-20T17:57:29.949 INFO:teuthology.orchestra.run.vm05.stdout: # 1m isn't quite enough 2026-03-20T17:57:29.949 INFO:teuthology.orchestra.run.vm05.stdout: mon_down_mkfs_grace = 2m 2026-03-20T17:57:29.949 INFO:teuthology.orchestra.run.vm05.stdout: 2026-03-20T17:57:29.949 INFO:teuthology.orchestra.run.vm05.stdout: mon_warn_on_filestore_osds = false 2026-03-20T17:57:29.949 INFO:teuthology.orchestra.run.vm05.stdout: 2026-03-20T17:57:29.949 INFO:teuthology.orchestra.run.vm05.stdout:[client] 2026-03-20T17:57:29.949 INFO:teuthology.orchestra.run.vm05.stdout: rgw cache enabled = true 2026-03-20T17:57:29.949 INFO:teuthology.orchestra.run.vm05.stdout: rgw enable ops log = true 2026-03-20T17:57:29.949 INFO:teuthology.orchestra.run.vm05.stdout: rgw enable usage log = true 2026-03-20T17:57:29.949 INFO:teuthology.orchestra.run.vm05.stdout: log file = /var/log/ceph/$cluster-$name.$pid.log 2026-03-20T17:57:29.949 INFO:teuthology.orchestra.run.vm05.stdout: admin socket = /var/run/ceph/$cluster-$name.$pid.asok 2026-03-20T17:57:29.949 INFO:teuthology.orchestra.run.vm05.stdout: debug rgw = 20 2026-03-20T17:57:29.949 INFO:teuthology.orchestra.run.vm05.stdout: debug rgw lifecycle = 20 2026-03-20T17:57:29.949 INFO:teuthology.orchestra.run.vm05.stdout: debug rgw notification = 20 2026-03-20T17:57:29.949 INFO:teuthology.orchestra.run.vm05.stdout: debug rgw sync = 20 2026-03-20T17:57:29.949 INFO:teuthology.orchestra.run.vm05.stdout: rgw crypt require ssl = False 2026-03-20T17:57:29.949 INFO:teuthology.orchestra.run.vm05.stdout: rgw crypt s3 kms backend = testing 2026-03-20T17:57:29.949 INFO:teuthology.orchestra.run.vm05.stdout: rgw crypt s3 kms encryption keys = testkey-1=YmluCmJvb3N0CmJvb3N0LWJ1aWxkCmNlcGguY29uZgo= 2026-03-20T17:57:29.949 INFO:teuthology.orchestra.run.vm05.stdout: rgw curl low speed time = 300 2026-03-20T17:57:29.949 INFO:teuthology.orchestra.run.vm05.stdout: rgw data log num shards = 4 2026-03-20T17:57:29.949 INFO:teuthology.orchestra.run.vm05.stdout: rgw data sync poll interval = 5 2026-03-20T17:57:29.949 INFO:teuthology.orchestra.run.vm05.stdout: rgw md log max shards = 4 2026-03-20T17:57:29.949 INFO:teuthology.orchestra.run.vm05.stdout: rgw meta sync poll interval = 5 2026-03-20T17:57:29.949 INFO:teuthology.orchestra.run.vm05.stdout: rgw s3 auth use sts = True 2026-03-20T17:57:29.949 INFO:teuthology.orchestra.run.vm05.stdout: rgw sts key = abcdefghijklmnoq 2026-03-20T17:57:29.949 INFO:teuthology.orchestra.run.vm05.stdout: rgw sync data inject err probability = 0 2026-03-20T17:57:29.949 INFO:teuthology.orchestra.run.vm05.stdout: rgw sync log trim interval = 0 2026-03-20T17:57:29.949 INFO:teuthology.orchestra.run.vm05.stdout: rgw sync meta inject err probability = 0 2026-03-20T17:57:29.949 INFO:teuthology.orchestra.run.vm05.stdout: rgw sync obj etag verify = True 2026-03-20T17:57:29.949 INFO:teuthology.orchestra.run.vm05.stdout: setgroup = ceph 2026-03-20T17:57:29.949 INFO:teuthology.orchestra.run.vm05.stdout: setuser = ceph 2026-03-20T17:57:29.949 INFO:teuthology.orchestra.run.vm05.stdout:[mon.a] 2026-03-20T17:57:29.949 INFO:teuthology.orchestra.run.vm05.stdout:[client.0] 2026-03-20T17:57:29.949 INFO:teuthology.orchestra.run.vm05.stdout: rgw data notify interval msec = 0 2026-03-20T17:57:29.949 INFO:teuthology.orchestra.run.vm05.stdout:[osd.0] 2026-03-20T17:57:29.949 INFO:teuthology.orchestra.run.vm05.stdout: osd_max_omap_entries_per_request = 10 2026-03-20T17:57:29.949 INFO:teuthology.orchestra.run.vm05.stdout:[osd.1] 2026-03-20T17:57:29.949 INFO:teuthology.orchestra.run.vm05.stdout: osd_max_omap_entries_per_request = 1000 2026-03-20T17:57:29.949 INFO:teuthology.orchestra.run.vm05.stdout:[osd.2] 2026-03-20T17:57:29.949 INFO:teuthology.orchestra.run.vm05.stdout: osd_max_omap_entries_per_request = 10000 2026-03-20T17:57:29.949 INFO:teuthology.orchestra.run.vm02.stdout:[global] 2026-03-20T17:57:29.949 INFO:teuthology.orchestra.run.vm02.stdout: chdir = "" 2026-03-20T17:57:29.949 INFO:teuthology.orchestra.run.vm02.stdout: pid file = /var/run/ceph/$cluster-$name.pid 2026-03-20T17:57:29.949 INFO:teuthology.orchestra.run.vm02.stdout: auth supported = cephx 2026-03-20T17:57:29.949 INFO:teuthology.orchestra.run.vm02.stdout: 2026-03-20T17:57:29.950 INFO:teuthology.orchestra.run.vm02.stdout: filestore xattr use omap = true 2026-03-20T17:57:29.950 INFO:teuthology.orchestra.run.vm02.stdout: 2026-03-20T17:57:29.950 INFO:teuthology.orchestra.run.vm02.stdout: mon clock drift allowed = 1.000 2026-03-20T17:57:29.950 INFO:teuthology.orchestra.run.vm02.stdout: 2026-03-20T17:57:29.950 INFO:teuthology.orchestra.run.vm02.stdout: osd crush chooseleaf type = 0 2026-03-20T17:57:29.950 INFO:teuthology.orchestra.run.vm02.stdout: auth debug = true 2026-03-20T17:57:29.950 INFO:teuthology.orchestra.run.vm02.stdout: 2026-03-20T17:57:29.950 INFO:teuthology.orchestra.run.vm02.stdout: ms die on old message = true 2026-03-20T17:57:29.950 INFO:teuthology.orchestra.run.vm02.stdout: ms die on bug = true 2026-03-20T17:57:29.950 INFO:teuthology.orchestra.run.vm02.stdout: 2026-03-20T17:57:29.950 INFO:teuthology.orchestra.run.vm02.stdout: mon max pg per osd = 10000 # >= luminous 2026-03-20T17:57:29.950 INFO:teuthology.orchestra.run.vm02.stdout: mon pg warn max object skew = 0 2026-03-20T17:57:29.950 INFO:teuthology.orchestra.run.vm02.stdout: 2026-03-20T17:57:29.950 INFO:teuthology.orchestra.run.vm02.stdout: # disable pg_autoscaler by default for new pools 2026-03-20T17:57:29.950 INFO:teuthology.orchestra.run.vm02.stdout: osd_pool_default_pg_autoscale_mode = off 2026-03-20T17:57:29.950 INFO:teuthology.orchestra.run.vm02.stdout: 2026-03-20T17:57:29.950 INFO:teuthology.orchestra.run.vm02.stdout: osd pool default size = 2 2026-03-20T17:57:29.950 INFO:teuthology.orchestra.run.vm02.stdout: 2026-03-20T17:57:29.950 INFO:teuthology.orchestra.run.vm02.stdout: mon osd allow primary affinity = true 2026-03-20T17:57:29.950 INFO:teuthology.orchestra.run.vm02.stdout: mon osd allow pg remap = true 2026-03-20T17:57:29.950 INFO:teuthology.orchestra.run.vm02.stdout: mon warn on legacy crush tunables = false 2026-03-20T17:57:29.950 INFO:teuthology.orchestra.run.vm02.stdout: mon warn on crush straw calc version zero = false 2026-03-20T17:57:29.950 INFO:teuthology.orchestra.run.vm02.stdout: mon warn on no sortbitwise = false 2026-03-20T17:57:29.950 INFO:teuthology.orchestra.run.vm02.stdout: mon warn on osd down out interval zero = false 2026-03-20T17:57:29.950 INFO:teuthology.orchestra.run.vm02.stdout: mon warn on too few osds = false 2026-03-20T17:57:29.950 INFO:teuthology.orchestra.run.vm02.stdout: mon_warn_on_pool_pg_num_not_power_of_two = false 2026-03-20T17:57:29.950 INFO:teuthology.orchestra.run.vm02.stdout: mon_warn_on_pool_no_redundancy = false 2026-03-20T17:57:29.950 INFO:teuthology.orchestra.run.vm02.stdout: mon_allow_pool_size_one = true 2026-03-20T17:57:29.950 INFO:teuthology.orchestra.run.vm02.stdout: 2026-03-20T17:57:29.950 INFO:teuthology.orchestra.run.vm02.stdout: osd pool default erasure code profile = plugin=isa technique=reed_sol_van k=2 m=1 crush-failure-domain=osd 2026-03-20T17:57:29.950 INFO:teuthology.orchestra.run.vm02.stdout: 2026-03-20T17:57:29.950 INFO:teuthology.orchestra.run.vm02.stdout: osd default data pool replay window = 5 2026-03-20T17:57:29.950 INFO:teuthology.orchestra.run.vm02.stdout: 2026-03-20T17:57:29.950 INFO:teuthology.orchestra.run.vm02.stdout: mon allow pool delete = true 2026-03-20T17:57:29.950 INFO:teuthology.orchestra.run.vm02.stdout: 2026-03-20T17:57:29.950 INFO:teuthology.orchestra.run.vm02.stdout: mon cluster log file level = debug 2026-03-20T17:57:29.950 INFO:teuthology.orchestra.run.vm02.stdout: debug asserts on shutdown = true 2026-03-20T17:57:29.950 INFO:teuthology.orchestra.run.vm02.stdout: mon health detail to clog = false 2026-03-20T17:57:29.950 INFO:teuthology.orchestra.run.vm02.stdout: mon host = 192.168.123.106 2026-03-20T17:57:29.950 INFO:teuthology.orchestra.run.vm02.stdout: fsid = 02f24f96-dd4e-4d53-915a-6d7df6cbe6f6 2026-03-20T17:57:29.950 INFO:teuthology.orchestra.run.vm02.stdout: 2026-03-20T17:57:29.950 INFO:teuthology.orchestra.run.vm02.stdout:[osd] 2026-03-20T17:57:29.950 INFO:teuthology.orchestra.run.vm02.stdout: osd journal size = 100 2026-03-20T17:57:29.950 INFO:teuthology.orchestra.run.vm02.stdout: 2026-03-20T17:57:29.950 INFO:teuthology.orchestra.run.vm02.stdout: osd scrub load threshold = 5.0 2026-03-20T17:57:29.950 INFO:teuthology.orchestra.run.vm02.stdout: osd scrub max interval = 600 2026-03-20T17:57:29.950 INFO:teuthology.orchestra.run.vm02.stdout: osd mclock profile = high_recovery_ops 2026-03-20T17:57:29.950 INFO:teuthology.orchestra.run.vm02.stdout: osd mclock skip benchmark = true 2026-03-20T17:57:29.950 INFO:teuthology.orchestra.run.vm02.stdout: 2026-03-20T17:57:29.950 INFO:teuthology.orchestra.run.vm02.stdout: osd recover clone overlap = true 2026-03-20T17:57:29.950 INFO:teuthology.orchestra.run.vm02.stdout: osd recovery max chunk = 1048576 2026-03-20T17:57:29.950 INFO:teuthology.orchestra.run.vm02.stdout: 2026-03-20T17:57:29.951 INFO:teuthology.orchestra.run.vm02.stdout: osd debug shutdown = true 2026-03-20T17:57:29.951 INFO:teuthology.orchestra.run.vm02.stdout: osd debug op order = true 2026-03-20T17:57:29.951 INFO:teuthology.orchestra.run.vm02.stdout: osd debug verify stray on activate = true 2026-03-20T17:57:29.951 INFO:teuthology.orchestra.run.vm02.stdout: osd debug trim objects = true 2026-03-20T17:57:29.951 INFO:teuthology.orchestra.run.vm02.stdout: 2026-03-20T17:57:29.951 INFO:teuthology.orchestra.run.vm02.stdout: osd open classes on start = true 2026-03-20T17:57:29.951 INFO:teuthology.orchestra.run.vm02.stdout: osd debug pg log writeout = true 2026-03-20T17:57:29.951 INFO:teuthology.orchestra.run.vm02.stdout: 2026-03-20T17:57:29.951 INFO:teuthology.orchestra.run.vm02.stdout: osd deep scrub update digest min age = 30 2026-03-20T17:57:29.951 INFO:teuthology.orchestra.run.vm02.stdout: 2026-03-20T17:57:29.951 INFO:teuthology.orchestra.run.vm02.stdout: osd map max advance = 10 2026-03-20T17:57:29.951 INFO:teuthology.orchestra.run.vm02.stdout: 2026-03-20T17:57:29.951 INFO:teuthology.orchestra.run.vm02.stdout: journal zero on create = true 2026-03-20T17:57:29.951 INFO:teuthology.orchestra.run.vm02.stdout: 2026-03-20T17:57:29.951 INFO:teuthology.orchestra.run.vm02.stdout: filestore ondisk finisher threads = 3 2026-03-20T17:57:29.951 INFO:teuthology.orchestra.run.vm02.stdout: filestore apply finisher threads = 3 2026-03-20T17:57:29.951 INFO:teuthology.orchestra.run.vm02.stdout: 2026-03-20T17:57:29.951 INFO:teuthology.orchestra.run.vm02.stdout: bdev debug aio = true 2026-03-20T17:57:29.951 INFO:teuthology.orchestra.run.vm02.stdout: osd debug misdirected ops = true 2026-03-20T17:57:29.951 INFO:teuthology.orchestra.run.vm02.stdout: debug ms = 1 2026-03-20T17:57:29.951 INFO:teuthology.orchestra.run.vm02.stdout: debug osd = 20 2026-03-20T17:57:29.951 INFO:teuthology.orchestra.run.vm02.stdout: osd mclock iops capacity threshold hdd = 49000 2026-03-20T17:57:29.951 INFO:teuthology.orchestra.run.vm02.stdout: rocksdb delete range threshold = 0 2026-03-20T17:57:29.951 INFO:teuthology.orchestra.run.vm02.stdout: 2026-03-20T17:57:29.951 INFO:teuthology.orchestra.run.vm02.stdout:[mgr] 2026-03-20T17:57:29.951 INFO:teuthology.orchestra.run.vm02.stdout: debug ms = 1 2026-03-20T17:57:29.951 INFO:teuthology.orchestra.run.vm02.stdout: debug mgr = 20 2026-03-20T17:57:29.951 INFO:teuthology.orchestra.run.vm02.stdout: debug mon = 20 2026-03-20T17:57:29.951 INFO:teuthology.orchestra.run.vm02.stdout: debug auth = 20 2026-03-20T17:57:29.951 INFO:teuthology.orchestra.run.vm02.stdout: mon reweight min pgs per osd = 4 2026-03-20T17:57:29.951 INFO:teuthology.orchestra.run.vm02.stdout: mon reweight min bytes per osd = 10 2026-03-20T17:57:29.951 INFO:teuthology.orchestra.run.vm02.stdout: mgr/telemetry/nag = false 2026-03-20T17:57:29.951 INFO:teuthology.orchestra.run.vm02.stdout: 2026-03-20T17:57:29.951 INFO:teuthology.orchestra.run.vm02.stdout:[mon] 2026-03-20T17:57:29.951 INFO:teuthology.orchestra.run.vm02.stdout: debug ms = 1 2026-03-20T17:57:29.951 INFO:teuthology.orchestra.run.vm02.stdout: debug mon = 20 2026-03-20T17:57:29.951 INFO:teuthology.orchestra.run.vm02.stdout: debug paxos = 20 2026-03-20T17:57:29.951 INFO:teuthology.orchestra.run.vm02.stdout: debug auth = 20 2026-03-20T17:57:29.951 INFO:teuthology.orchestra.run.vm02.stdout: mon data avail warn = 5 2026-03-20T17:57:29.951 INFO:teuthology.orchestra.run.vm02.stdout: mon mgr mkfs grace = 240 2026-03-20T17:57:29.951 INFO:teuthology.orchestra.run.vm02.stdout: mon reweight min pgs per osd = 4 2026-03-20T17:57:29.951 INFO:teuthology.orchestra.run.vm02.stdout: mon osd reporter subtree level = osd 2026-03-20T17:57:29.951 INFO:teuthology.orchestra.run.vm02.stdout: mon osd prime pg temp = true 2026-03-20T17:57:29.951 INFO:teuthology.orchestra.run.vm02.stdout: mon reweight min bytes per osd = 10 2026-03-20T17:57:29.951 INFO:teuthology.orchestra.run.vm02.stdout: 2026-03-20T17:57:29.951 INFO:teuthology.orchestra.run.vm02.stdout: # rotate auth tickets quickly to exercise renewal paths 2026-03-20T17:57:29.951 INFO:teuthology.orchestra.run.vm02.stdout: auth mon ticket ttl = 660 # 11m 2026-03-20T17:57:29.951 INFO:teuthology.orchestra.run.vm02.stdout: auth service ticket ttl = 240 # 4m 2026-03-20T17:57:29.951 INFO:teuthology.orchestra.run.vm02.stdout: 2026-03-20T17:57:29.951 INFO:teuthology.orchestra.run.vm02.stdout: # don't complain about insecure global_id in the test suite 2026-03-20T17:57:29.951 INFO:teuthology.orchestra.run.vm02.stdout: mon_warn_on_insecure_global_id_reclaim = false 2026-03-20T17:57:29.951 INFO:teuthology.orchestra.run.vm02.stdout: mon_warn_on_insecure_global_id_reclaim_allowed = false 2026-03-20T17:57:29.951 INFO:teuthology.orchestra.run.vm02.stdout: 2026-03-20T17:57:29.951 INFO:teuthology.orchestra.run.vm02.stdout: # 1m isn't quite enough 2026-03-20T17:57:29.951 INFO:teuthology.orchestra.run.vm02.stdout: mon_down_mkfs_grace = 2m 2026-03-20T17:57:29.951 INFO:teuthology.orchestra.run.vm02.stdout: 2026-03-20T17:57:29.951 INFO:teuthology.orchestra.run.vm02.stdout: mon_warn_on_filestore_osds = false 2026-03-20T17:57:29.951 INFO:teuthology.orchestra.run.vm02.stdout: 2026-03-20T17:57:29.951 INFO:teuthology.orchestra.run.vm02.stdout:[client] 2026-03-20T17:57:29.951 INFO:teuthology.orchestra.run.vm02.stdout: rgw cache enabled = true 2026-03-20T17:57:29.951 INFO:teuthology.orchestra.run.vm02.stdout: rgw enable ops log = true 2026-03-20T17:57:29.951 INFO:teuthology.orchestra.run.vm02.stdout: rgw enable usage log = true 2026-03-20T17:57:29.952 INFO:teuthology.orchestra.run.vm02.stdout: log file = /var/log/ceph/$cluster-$name.$pid.log 2026-03-20T17:57:29.952 INFO:teuthology.orchestra.run.vm02.stdout: admin socket = /var/run/ceph/$cluster-$name.$pid.asok 2026-03-20T17:57:29.952 INFO:teuthology.orchestra.run.vm02.stdout: debug rgw = 20 2026-03-20T17:57:29.952 INFO:teuthology.orchestra.run.vm02.stdout: debug rgw lifecycle = 20 2026-03-20T17:57:29.952 INFO:teuthology.orchestra.run.vm02.stdout: debug rgw notification = 20 2026-03-20T17:57:29.952 INFO:teuthology.orchestra.run.vm02.stdout: debug rgw sync = 20 2026-03-20T17:57:29.952 INFO:teuthology.orchestra.run.vm02.stdout: rgw crypt require ssl = False 2026-03-20T17:57:29.952 INFO:teuthology.orchestra.run.vm02.stdout: rgw crypt s3 kms backend = testing 2026-03-20T17:57:29.952 INFO:teuthology.orchestra.run.vm02.stdout: rgw crypt s3 kms encryption keys = testkey-1=YmluCmJvb3N0CmJvb3N0LWJ1aWxkCmNlcGguY29uZgo= 2026-03-20T17:57:29.952 INFO:teuthology.orchestra.run.vm02.stdout: rgw curl low speed time = 300 2026-03-20T17:57:29.952 INFO:teuthology.orchestra.run.vm02.stdout: rgw data log num shards = 4 2026-03-20T17:57:29.952 INFO:teuthology.orchestra.run.vm02.stdout: rgw data sync poll interval = 5 2026-03-20T17:57:29.952 INFO:teuthology.orchestra.run.vm02.stdout: rgw md log max shards = 4 2026-03-20T17:57:29.952 INFO:teuthology.orchestra.run.vm02.stdout: rgw meta sync poll interval = 5 2026-03-20T17:57:29.952 INFO:teuthology.orchestra.run.vm02.stdout: rgw s3 auth use sts = True 2026-03-20T17:57:29.952 INFO:teuthology.orchestra.run.vm02.stdout: rgw sts key = abcdefghijklmnoq 2026-03-20T17:57:29.952 INFO:teuthology.orchestra.run.vm02.stdout: rgw sync data inject err probability = 0 2026-03-20T17:57:29.952 INFO:teuthology.orchestra.run.vm02.stdout: rgw sync log trim interval = 0 2026-03-20T17:57:29.952 INFO:teuthology.orchestra.run.vm02.stdout: rgw sync meta inject err probability = 0 2026-03-20T17:57:29.952 INFO:teuthology.orchestra.run.vm02.stdout: rgw sync obj etag verify = True 2026-03-20T17:57:29.952 INFO:teuthology.orchestra.run.vm02.stdout: setgroup = ceph 2026-03-20T17:57:29.952 INFO:teuthology.orchestra.run.vm02.stdout: setuser = ceph 2026-03-20T17:57:29.952 INFO:teuthology.orchestra.run.vm02.stdout:[mon.a] 2026-03-20T17:57:29.952 INFO:teuthology.orchestra.run.vm02.stdout:[client.0] 2026-03-20T17:57:29.952 INFO:teuthology.orchestra.run.vm02.stdout: rgw data notify interval msec = 0 2026-03-20T17:57:29.952 INFO:teuthology.orchestra.run.vm02.stdout:[osd.0] 2026-03-20T17:57:29.952 INFO:teuthology.orchestra.run.vm02.stdout: osd_max_omap_entries_per_request = 10 2026-03-20T17:57:29.952 INFO:teuthology.orchestra.run.vm02.stdout:[osd.1] 2026-03-20T17:57:29.952 INFO:teuthology.orchestra.run.vm02.stdout: osd_max_omap_entries_per_request = 1000 2026-03-20T17:57:29.952 INFO:teuthology.orchestra.run.vm02.stdout:[osd.2] 2026-03-20T17:57:29.952 INFO:teuthology.orchestra.run.vm02.stdout: osd_max_omap_entries_per_request = 10000 2026-03-20T17:57:29.963 INFO:teuthology.orchestra.run.vm08.stdout:[global] 2026-03-20T17:57:29.963 INFO:teuthology.orchestra.run.vm08.stdout: chdir = "" 2026-03-20T17:57:29.963 INFO:teuthology.orchestra.run.vm08.stdout: pid file = /var/run/ceph/$cluster-$name.pid 2026-03-20T17:57:29.963 INFO:teuthology.orchestra.run.vm08.stdout: auth supported = cephx 2026-03-20T17:57:29.963 INFO:teuthology.orchestra.run.vm08.stdout: 2026-03-20T17:57:29.963 INFO:teuthology.orchestra.run.vm08.stdout: filestore xattr use omap = true 2026-03-20T17:57:29.963 INFO:teuthology.orchestra.run.vm08.stdout: 2026-03-20T17:57:29.963 INFO:teuthology.orchestra.run.vm08.stdout: mon clock drift allowed = 1.000 2026-03-20T17:57:29.963 INFO:teuthology.orchestra.run.vm08.stdout: 2026-03-20T17:57:29.963 INFO:teuthology.orchestra.run.vm08.stdout: osd crush chooseleaf type = 0 2026-03-20T17:57:29.963 INFO:teuthology.orchestra.run.vm08.stdout: auth debug = true 2026-03-20T17:57:29.963 INFO:teuthology.orchestra.run.vm08.stdout: 2026-03-20T17:57:29.963 INFO:teuthology.orchestra.run.vm08.stdout: ms die on old message = true 2026-03-20T17:57:29.963 INFO:teuthology.orchestra.run.vm08.stdout: ms die on bug = true 2026-03-20T17:57:29.963 INFO:teuthology.orchestra.run.vm08.stdout: 2026-03-20T17:57:29.963 INFO:teuthology.orchestra.run.vm08.stdout: mon max pg per osd = 10000 # >= luminous 2026-03-20T17:57:29.963 INFO:teuthology.orchestra.run.vm08.stdout: mon pg warn max object skew = 0 2026-03-20T17:57:29.963 INFO:teuthology.orchestra.run.vm08.stdout: 2026-03-20T17:57:29.964 INFO:teuthology.orchestra.run.vm08.stdout: # disable pg_autoscaler by default for new pools 2026-03-20T17:57:29.964 INFO:teuthology.orchestra.run.vm08.stdout: osd_pool_default_pg_autoscale_mode = off 2026-03-20T17:57:29.964 INFO:teuthology.orchestra.run.vm08.stdout: 2026-03-20T17:57:29.964 INFO:teuthology.orchestra.run.vm08.stdout: osd pool default size = 2 2026-03-20T17:57:29.964 INFO:teuthology.orchestra.run.vm08.stdout: 2026-03-20T17:57:29.964 INFO:teuthology.orchestra.run.vm08.stdout: mon osd allow primary affinity = true 2026-03-20T17:57:29.964 INFO:teuthology.orchestra.run.vm08.stdout: mon osd allow pg remap = true 2026-03-20T17:57:29.964 INFO:teuthology.orchestra.run.vm08.stdout: mon warn on legacy crush tunables = false 2026-03-20T17:57:29.964 INFO:teuthology.orchestra.run.vm08.stdout: mon warn on crush straw calc version zero = false 2026-03-20T17:57:29.964 INFO:teuthology.orchestra.run.vm08.stdout: mon warn on no sortbitwise = false 2026-03-20T17:57:29.964 INFO:teuthology.orchestra.run.vm08.stdout: mon warn on osd down out interval zero = false 2026-03-20T17:57:29.964 INFO:teuthology.orchestra.run.vm08.stdout: mon warn on too few osds = false 2026-03-20T17:57:29.964 INFO:teuthology.orchestra.run.vm08.stdout: mon_warn_on_pool_pg_num_not_power_of_two = false 2026-03-20T17:57:29.964 INFO:teuthology.orchestra.run.vm08.stdout: mon_warn_on_pool_no_redundancy = false 2026-03-20T17:57:29.964 INFO:teuthology.orchestra.run.vm08.stdout: mon_allow_pool_size_one = true 2026-03-20T17:57:29.964 INFO:teuthology.orchestra.run.vm08.stdout: 2026-03-20T17:57:29.964 INFO:teuthology.orchestra.run.vm08.stdout: osd pool default erasure code profile = plugin=isa technique=reed_sol_van k=2 m=1 crush-failure-domain=osd 2026-03-20T17:57:29.964 INFO:teuthology.orchestra.run.vm08.stdout: 2026-03-20T17:57:29.964 INFO:teuthology.orchestra.run.vm08.stdout: osd default data pool replay window = 5 2026-03-20T17:57:29.964 INFO:teuthology.orchestra.run.vm08.stdout: 2026-03-20T17:57:29.964 INFO:teuthology.orchestra.run.vm08.stdout: mon allow pool delete = true 2026-03-20T17:57:29.964 INFO:teuthology.orchestra.run.vm08.stdout: 2026-03-20T17:57:29.964 INFO:teuthology.orchestra.run.vm08.stdout: mon cluster log file level = debug 2026-03-20T17:57:29.964 INFO:teuthology.orchestra.run.vm08.stdout: debug asserts on shutdown = true 2026-03-20T17:57:29.964 INFO:teuthology.orchestra.run.vm08.stdout: mon health detail to clog = false 2026-03-20T17:57:29.964 INFO:teuthology.orchestra.run.vm08.stdout: mon host = 192.168.123.106 2026-03-20T17:57:29.964 INFO:teuthology.orchestra.run.vm08.stdout: fsid = 02f24f96-dd4e-4d53-915a-6d7df6cbe6f6 2026-03-20T17:57:29.964 INFO:teuthology.orchestra.run.vm08.stdout: 2026-03-20T17:57:29.964 INFO:teuthology.orchestra.run.vm08.stdout:[osd] 2026-03-20T17:57:29.964 INFO:teuthology.orchestra.run.vm08.stdout: osd journal size = 100 2026-03-20T17:57:29.964 INFO:teuthology.orchestra.run.vm08.stdout: 2026-03-20T17:57:29.964 INFO:teuthology.orchestra.run.vm08.stdout: osd scrub load threshold = 5.0 2026-03-20T17:57:29.964 INFO:teuthology.orchestra.run.vm08.stdout: osd scrub max interval = 600 2026-03-20T17:57:29.964 INFO:teuthology.orchestra.run.vm08.stdout: osd mclock profile = high_recovery_ops 2026-03-20T17:57:29.964 INFO:teuthology.orchestra.run.vm08.stdout: osd mclock skip benchmark = true 2026-03-20T17:57:29.964 INFO:teuthology.orchestra.run.vm08.stdout: 2026-03-20T17:57:29.964 INFO:teuthology.orchestra.run.vm08.stdout: osd recover clone overlap = true 2026-03-20T17:57:29.964 INFO:teuthology.orchestra.run.vm08.stdout: osd recovery max chunk = 1048576 2026-03-20T17:57:29.964 INFO:teuthology.orchestra.run.vm08.stdout: 2026-03-20T17:57:29.964 INFO:teuthology.orchestra.run.vm08.stdout: osd debug shutdown = true 2026-03-20T17:57:29.964 INFO:teuthology.orchestra.run.vm08.stdout: osd debug op order = true 2026-03-20T17:57:29.964 INFO:teuthology.orchestra.run.vm08.stdout: osd debug verify stray on activate = true 2026-03-20T17:57:29.964 INFO:teuthology.orchestra.run.vm08.stdout: osd debug trim objects = true 2026-03-20T17:57:29.964 INFO:teuthology.orchestra.run.vm08.stdout: 2026-03-20T17:57:29.964 INFO:teuthology.orchestra.run.vm08.stdout: osd open classes on start = true 2026-03-20T17:57:29.964 INFO:teuthology.orchestra.run.vm08.stdout: osd debug pg log writeout = true 2026-03-20T17:57:29.964 INFO:teuthology.orchestra.run.vm08.stdout: 2026-03-20T17:57:29.964 INFO:teuthology.orchestra.run.vm08.stdout: osd deep scrub update digest min age = 30 2026-03-20T17:57:29.964 INFO:teuthology.orchestra.run.vm08.stdout: 2026-03-20T17:57:29.964 INFO:teuthology.orchestra.run.vm08.stdout: osd map max advance = 10 2026-03-20T17:57:29.964 INFO:teuthology.orchestra.run.vm08.stdout: 2026-03-20T17:57:29.964 INFO:teuthology.orchestra.run.vm08.stdout: journal zero on create = true 2026-03-20T17:57:29.964 INFO:teuthology.orchestra.run.vm08.stdout: 2026-03-20T17:57:29.964 INFO:teuthology.orchestra.run.vm08.stdout: filestore ondisk finisher threads = 3 2026-03-20T17:57:29.964 INFO:teuthology.orchestra.run.vm08.stdout: filestore apply finisher threads = 3 2026-03-20T17:57:29.964 INFO:teuthology.orchestra.run.vm08.stdout: 2026-03-20T17:57:29.964 INFO:teuthology.orchestra.run.vm08.stdout: bdev debug aio = true 2026-03-20T17:57:29.964 INFO:teuthology.orchestra.run.vm08.stdout: osd debug misdirected ops = true 2026-03-20T17:57:29.964 INFO:teuthology.orchestra.run.vm08.stdout: debug ms = 1 2026-03-20T17:57:29.964 INFO:teuthology.orchestra.run.vm08.stdout: debug osd = 20 2026-03-20T17:57:29.964 INFO:teuthology.orchestra.run.vm08.stdout: osd mclock iops capacity threshold hdd = 49000 2026-03-20T17:57:29.964 INFO:teuthology.orchestra.run.vm08.stdout: rocksdb delete range threshold = 0 2026-03-20T17:57:29.964 INFO:teuthology.orchestra.run.vm08.stdout: 2026-03-20T17:57:29.965 INFO:teuthology.orchestra.run.vm08.stdout:[mgr] 2026-03-20T17:57:29.965 INFO:teuthology.orchestra.run.vm08.stdout: debug ms = 1 2026-03-20T17:57:29.965 INFO:teuthology.orchestra.run.vm08.stdout: debug mgr = 20 2026-03-20T17:57:29.965 INFO:teuthology.orchestra.run.vm08.stdout: debug mon = 20 2026-03-20T17:57:29.965 INFO:teuthology.orchestra.run.vm08.stdout: debug auth = 20 2026-03-20T17:57:29.965 INFO:teuthology.orchestra.run.vm08.stdout: mon reweight min pgs per osd = 4 2026-03-20T17:57:29.965 INFO:teuthology.orchestra.run.vm08.stdout: mon reweight min bytes per osd = 10 2026-03-20T17:57:29.965 INFO:teuthology.orchestra.run.vm08.stdout: mgr/telemetry/nag = false 2026-03-20T17:57:29.965 INFO:teuthology.orchestra.run.vm08.stdout: 2026-03-20T17:57:29.965 INFO:teuthology.orchestra.run.vm08.stdout:[mon] 2026-03-20T17:57:29.965 INFO:teuthology.orchestra.run.vm08.stdout: debug ms = 1 2026-03-20T17:57:29.965 INFO:teuthology.orchestra.run.vm08.stdout: debug mon = 20 2026-03-20T17:57:29.965 INFO:teuthology.orchestra.run.vm08.stdout: debug paxos = 20 2026-03-20T17:57:29.965 INFO:teuthology.orchestra.run.vm08.stdout: debug auth = 20 2026-03-20T17:57:29.965 INFO:teuthology.orchestra.run.vm08.stdout: mon data avail warn = 5 2026-03-20T17:57:29.965 INFO:teuthology.orchestra.run.vm08.stdout: mon mgr mkfs grace = 240 2026-03-20T17:57:29.965 INFO:teuthology.orchestra.run.vm08.stdout: mon reweight min pgs per osd = 4 2026-03-20T17:57:29.965 INFO:teuthology.orchestra.run.vm08.stdout: mon osd reporter subtree level = osd 2026-03-20T17:57:29.965 INFO:teuthology.orchestra.run.vm08.stdout: mon osd prime pg temp = true 2026-03-20T17:57:29.965 INFO:teuthology.orchestra.run.vm08.stdout: mon reweight min bytes per osd = 10 2026-03-20T17:57:29.965 INFO:teuthology.orchestra.run.vm08.stdout: 2026-03-20T17:57:29.965 INFO:teuthology.orchestra.run.vm08.stdout: # rotate auth tickets quickly to exercise renewal paths 2026-03-20T17:57:29.965 INFO:teuthology.orchestra.run.vm08.stdout: auth mon ticket ttl = 660 # 11m 2026-03-20T17:57:29.965 INFO:teuthology.orchestra.run.vm08.stdout: auth service ticket ttl = 240 # 4m 2026-03-20T17:57:29.965 INFO:teuthology.orchestra.run.vm08.stdout: 2026-03-20T17:57:29.965 INFO:teuthology.orchestra.run.vm08.stdout: # don't complain about insecure global_id in the test suite 2026-03-20T17:57:29.965 INFO:teuthology.orchestra.run.vm08.stdout: mon_warn_on_insecure_global_id_reclaim = false 2026-03-20T17:57:29.965 INFO:teuthology.orchestra.run.vm08.stdout: mon_warn_on_insecure_global_id_reclaim_allowed = false 2026-03-20T17:57:29.965 INFO:teuthology.orchestra.run.vm08.stdout: 2026-03-20T17:57:29.965 INFO:teuthology.orchestra.run.vm08.stdout: # 1m isn't quite enough 2026-03-20T17:57:29.965 INFO:teuthology.orchestra.run.vm08.stdout: mon_down_mkfs_grace = 2m 2026-03-20T17:57:29.965 INFO:teuthology.orchestra.run.vm08.stdout: 2026-03-20T17:57:29.965 INFO:teuthology.orchestra.run.vm08.stdout: mon_warn_on_filestore_osds = false 2026-03-20T17:57:29.965 INFO:teuthology.orchestra.run.vm08.stdout: 2026-03-20T17:57:29.965 INFO:teuthology.orchestra.run.vm08.stdout:[client] 2026-03-20T17:57:29.965 INFO:teuthology.orchestra.run.vm08.stdout: rgw cache enabled = true 2026-03-20T17:57:29.965 INFO:teuthology.orchestra.run.vm08.stdout: rgw enable ops log = true 2026-03-20T17:57:29.965 INFO:teuthology.orchestra.run.vm08.stdout: rgw enable usage log = true 2026-03-20T17:57:29.965 INFO:teuthology.orchestra.run.vm08.stdout: log file = /var/log/ceph/$cluster-$name.$pid.log 2026-03-20T17:57:29.965 INFO:teuthology.orchestra.run.vm08.stdout: admin socket = /var/run/ceph/$cluster-$name.$pid.asok 2026-03-20T17:57:29.965 INFO:teuthology.orchestra.run.vm08.stdout: debug rgw = 20 2026-03-20T17:57:29.965 INFO:teuthology.orchestra.run.vm08.stdout: debug rgw lifecycle = 20 2026-03-20T17:57:29.965 INFO:teuthology.orchestra.run.vm08.stdout: debug rgw notification = 20 2026-03-20T17:57:29.965 INFO:teuthology.orchestra.run.vm08.stdout: debug rgw sync = 20 2026-03-20T17:57:29.965 INFO:teuthology.orchestra.run.vm08.stdout: rgw crypt require ssl = False 2026-03-20T17:57:29.965 INFO:teuthology.orchestra.run.vm08.stdout: rgw crypt s3 kms backend = testing 2026-03-20T17:57:29.965 INFO:teuthology.orchestra.run.vm08.stdout: rgw crypt s3 kms encryption keys = testkey-1=YmluCmJvb3N0CmJvb3N0LWJ1aWxkCmNlcGguY29uZgo= 2026-03-20T17:57:29.965 INFO:teuthology.orchestra.run.vm08.stdout: rgw curl low speed time = 300 2026-03-20T17:57:29.965 INFO:teuthology.orchestra.run.vm08.stdout: rgw data log num shards = 4 2026-03-20T17:57:29.965 INFO:teuthology.orchestra.run.vm08.stdout: rgw data sync poll interval = 5 2026-03-20T17:57:29.965 INFO:teuthology.orchestra.run.vm08.stdout: rgw md log max shards = 4 2026-03-20T17:57:29.965 INFO:teuthology.orchestra.run.vm08.stdout: rgw meta sync poll interval = 5 2026-03-20T17:57:29.965 INFO:teuthology.orchestra.run.vm08.stdout: rgw s3 auth use sts = True 2026-03-20T17:57:29.965 INFO:teuthology.orchestra.run.vm08.stdout: rgw sts key = abcdefghijklmnoq 2026-03-20T17:57:29.965 INFO:teuthology.orchestra.run.vm08.stdout: rgw sync data inject err probability = 0 2026-03-20T17:57:29.965 INFO:teuthology.orchestra.run.vm08.stdout: rgw sync log trim interval = 0 2026-03-20T17:57:29.965 INFO:teuthology.orchestra.run.vm08.stdout: rgw sync meta inject err probability = 0 2026-03-20T17:57:29.965 INFO:teuthology.orchestra.run.vm08.stdout: rgw sync obj etag verify = True 2026-03-20T17:57:29.965 INFO:teuthology.orchestra.run.vm08.stdout: setgroup = ceph 2026-03-20T17:57:29.966 INFO:teuthology.orchestra.run.vm08.stdout: setuser = ceph 2026-03-20T17:57:29.966 INFO:teuthology.orchestra.run.vm08.stdout:[mon.a] 2026-03-20T17:57:29.966 INFO:teuthology.orchestra.run.vm08.stdout:[client.0] 2026-03-20T17:57:29.966 INFO:teuthology.orchestra.run.vm08.stdout: rgw data notify interval msec = 0 2026-03-20T17:57:29.966 INFO:teuthology.orchestra.run.vm08.stdout:[osd.0] 2026-03-20T17:57:29.966 INFO:teuthology.orchestra.run.vm08.stdout: osd_max_omap_entries_per_request = 10 2026-03-20T17:57:29.966 INFO:teuthology.orchestra.run.vm08.stdout:[osd.1] 2026-03-20T17:57:29.966 INFO:teuthology.orchestra.run.vm08.stdout: osd_max_omap_entries_per_request = 1000 2026-03-20T17:57:29.966 INFO:teuthology.orchestra.run.vm08.stdout:[osd.2] 2026-03-20T17:57:29.966 INFO:teuthology.orchestra.run.vm08.stdout: osd_max_omap_entries_per_request = 10000 2026-03-20T17:57:29.966 INFO:teuthology.orchestra.run.vm06.stdout:[global] 2026-03-20T17:57:29.966 INFO:teuthology.orchestra.run.vm06.stdout: chdir = "" 2026-03-20T17:57:29.966 INFO:teuthology.orchestra.run.vm06.stdout: pid file = /var/run/ceph/$cluster-$name.pid 2026-03-20T17:57:29.966 INFO:teuthology.orchestra.run.vm06.stdout: auth supported = cephx 2026-03-20T17:57:29.966 INFO:teuthology.orchestra.run.vm06.stdout: 2026-03-20T17:57:29.966 INFO:teuthology.orchestra.run.vm06.stdout: filestore xattr use omap = true 2026-03-20T17:57:29.966 INFO:teuthology.orchestra.run.vm06.stdout: 2026-03-20T17:57:29.966 INFO:teuthology.orchestra.run.vm06.stdout: mon clock drift allowed = 1.000 2026-03-20T17:57:29.966 INFO:teuthology.orchestra.run.vm06.stdout: 2026-03-20T17:57:29.966 INFO:teuthology.orchestra.run.vm06.stdout: osd crush chooseleaf type = 0 2026-03-20T17:57:29.966 INFO:teuthology.orchestra.run.vm06.stdout: auth debug = true 2026-03-20T17:57:29.966 INFO:teuthology.orchestra.run.vm06.stdout: 2026-03-20T17:57:29.966 INFO:teuthology.orchestra.run.vm06.stdout: ms die on old message = true 2026-03-20T17:57:29.966 INFO:teuthology.orchestra.run.vm06.stdout: ms die on bug = true 2026-03-20T17:57:29.966 INFO:teuthology.orchestra.run.vm06.stdout: 2026-03-20T17:57:29.966 INFO:teuthology.orchestra.run.vm06.stdout: mon max pg per osd = 10000 # >= luminous 2026-03-20T17:57:29.966 INFO:teuthology.orchestra.run.vm06.stdout: mon pg warn max object skew = 0 2026-03-20T17:57:29.966 INFO:teuthology.orchestra.run.vm06.stdout: 2026-03-20T17:57:29.966 INFO:teuthology.orchestra.run.vm06.stdout: # disable pg_autoscaler by default for new pools 2026-03-20T17:57:29.966 INFO:teuthology.orchestra.run.vm06.stdout: osd_pool_default_pg_autoscale_mode = off 2026-03-20T17:57:29.966 INFO:teuthology.orchestra.run.vm06.stdout: 2026-03-20T17:57:29.966 INFO:teuthology.orchestra.run.vm06.stdout: osd pool default size = 2 2026-03-20T17:57:29.966 INFO:teuthology.orchestra.run.vm06.stdout: 2026-03-20T17:57:29.966 INFO:teuthology.orchestra.run.vm06.stdout: mon osd allow primary affinity = true 2026-03-20T17:57:29.966 INFO:teuthology.orchestra.run.vm06.stdout: mon osd allow pg remap = true 2026-03-20T17:57:29.966 INFO:teuthology.orchestra.run.vm06.stdout: mon warn on legacy crush tunables = false 2026-03-20T17:57:29.966 INFO:teuthology.orchestra.run.vm06.stdout: mon warn on crush straw calc version zero = false 2026-03-20T17:57:29.966 INFO:teuthology.orchestra.run.vm06.stdout: mon warn on no sortbitwise = false 2026-03-20T17:57:29.966 INFO:teuthology.orchestra.run.vm06.stdout: mon warn on osd down out interval zero = false 2026-03-20T17:57:29.966 INFO:teuthology.orchestra.run.vm06.stdout: mon warn on too few osds = false 2026-03-20T17:57:29.966 INFO:teuthology.orchestra.run.vm06.stdout: mon_warn_on_pool_pg_num_not_power_of_two = false 2026-03-20T17:57:29.966 INFO:teuthology.orchestra.run.vm06.stdout: mon_warn_on_pool_no_redundancy = false 2026-03-20T17:57:29.966 INFO:teuthology.orchestra.run.vm06.stdout: mon_allow_pool_size_one = true 2026-03-20T17:57:29.966 INFO:teuthology.orchestra.run.vm06.stdout: 2026-03-20T17:57:29.966 INFO:teuthology.orchestra.run.vm06.stdout: osd pool default erasure code profile = plugin=isa technique=reed_sol_van k=2 m=1 crush-failure-domain=osd 2026-03-20T17:57:29.966 INFO:teuthology.orchestra.run.vm06.stdout: 2026-03-20T17:57:29.966 INFO:teuthology.orchestra.run.vm06.stdout: osd default data pool replay window = 5 2026-03-20T17:57:29.966 INFO:teuthology.orchestra.run.vm06.stdout: 2026-03-20T17:57:29.966 INFO:teuthology.orchestra.run.vm06.stdout: mon allow pool delete = true 2026-03-20T17:57:29.967 INFO:teuthology.orchestra.run.vm06.stdout: 2026-03-20T17:57:29.967 INFO:teuthology.orchestra.run.vm06.stdout: mon cluster log file level = debug 2026-03-20T17:57:29.967 INFO:teuthology.orchestra.run.vm06.stdout: debug asserts on shutdown = true 2026-03-20T17:57:29.967 INFO:teuthology.orchestra.run.vm06.stdout: mon health detail to clog = false 2026-03-20T17:57:29.967 INFO:teuthology.orchestra.run.vm06.stdout: mon host = 192.168.123.106 2026-03-20T17:57:29.967 INFO:teuthology.orchestra.run.vm06.stdout: fsid = 02f24f96-dd4e-4d53-915a-6d7df6cbe6f6 2026-03-20T17:57:29.967 INFO:teuthology.orchestra.run.vm06.stdout: 2026-03-20T17:57:29.967 INFO:teuthology.orchestra.run.vm06.stdout:[osd] 2026-03-20T17:57:29.967 INFO:teuthology.orchestra.run.vm06.stdout: osd journal size = 100 2026-03-20T17:57:29.967 INFO:teuthology.orchestra.run.vm06.stdout: 2026-03-20T17:57:29.967 INFO:teuthology.orchestra.run.vm06.stdout: osd scrub load threshold = 5.0 2026-03-20T17:57:29.967 INFO:teuthology.orchestra.run.vm06.stdout: osd scrub max interval = 600 2026-03-20T17:57:29.967 INFO:teuthology.orchestra.run.vm06.stdout: osd mclock profile = high_recovery_ops 2026-03-20T17:57:29.967 INFO:teuthology.orchestra.run.vm06.stdout: osd mclock skip benchmark = true 2026-03-20T17:57:29.967 INFO:teuthology.orchestra.run.vm06.stdout: 2026-03-20T17:57:29.967 INFO:teuthology.orchestra.run.vm06.stdout: osd recover clone overlap = true 2026-03-20T17:57:29.967 INFO:teuthology.orchestra.run.vm06.stdout: osd recovery max chunk = 1048576 2026-03-20T17:57:29.967 INFO:teuthology.orchestra.run.vm06.stdout: 2026-03-20T17:57:29.967 INFO:teuthology.orchestra.run.vm06.stdout: osd debug shutdown = true 2026-03-20T17:57:29.967 INFO:teuthology.orchestra.run.vm06.stdout: osd debug op order = true 2026-03-20T17:57:29.967 INFO:teuthology.orchestra.run.vm06.stdout: osd debug verify stray on activate = true 2026-03-20T17:57:29.967 INFO:teuthology.orchestra.run.vm06.stdout: osd debug trim objects = true 2026-03-20T17:57:29.967 INFO:teuthology.orchestra.run.vm06.stdout: 2026-03-20T17:57:29.967 INFO:teuthology.orchestra.run.vm06.stdout: osd open classes on start = true 2026-03-20T17:57:29.967 INFO:teuthology.orchestra.run.vm06.stdout: osd debug pg log writeout = true 2026-03-20T17:57:29.967 INFO:teuthology.orchestra.run.vm06.stdout: 2026-03-20T17:57:29.967 INFO:teuthology.orchestra.run.vm06.stdout: osd deep scrub update digest min age = 30 2026-03-20T17:57:29.967 INFO:teuthology.orchestra.run.vm06.stdout: 2026-03-20T17:57:29.967 INFO:teuthology.orchestra.run.vm06.stdout: osd map max advance = 10 2026-03-20T17:57:29.967 INFO:teuthology.orchestra.run.vm06.stdout: 2026-03-20T17:57:29.967 INFO:teuthology.orchestra.run.vm06.stdout: journal zero on create = true 2026-03-20T17:57:29.967 INFO:teuthology.orchestra.run.vm06.stdout: 2026-03-20T17:57:29.967 INFO:teuthology.orchestra.run.vm06.stdout: filestore ondisk finisher threads = 3 2026-03-20T17:57:29.967 INFO:teuthology.orchestra.run.vm06.stdout: filestore apply finisher threads = 3 2026-03-20T17:57:29.967 INFO:teuthology.orchestra.run.vm06.stdout: 2026-03-20T17:57:29.967 INFO:teuthology.orchestra.run.vm06.stdout: bdev debug aio = true 2026-03-20T17:57:29.967 INFO:teuthology.orchestra.run.vm06.stdout: osd debug misdirected ops = true 2026-03-20T17:57:29.967 INFO:teuthology.orchestra.run.vm06.stdout: debug ms = 1 2026-03-20T17:57:29.967 INFO:teuthology.orchestra.run.vm06.stdout: debug osd = 20 2026-03-20T17:57:29.967 INFO:teuthology.orchestra.run.vm06.stdout: osd mclock iops capacity threshold hdd = 49000 2026-03-20T17:57:29.967 INFO:teuthology.orchestra.run.vm06.stdout: rocksdb delete range threshold = 0 2026-03-20T17:57:29.967 INFO:teuthology.orchestra.run.vm06.stdout: 2026-03-20T17:57:29.967 INFO:teuthology.orchestra.run.vm06.stdout:[mgr] 2026-03-20T17:57:29.967 INFO:teuthology.orchestra.run.vm06.stdout: debug ms = 1 2026-03-20T17:57:29.967 INFO:teuthology.orchestra.run.vm06.stdout: debug mgr = 20 2026-03-20T17:57:29.967 INFO:teuthology.orchestra.run.vm06.stdout: debug mon = 20 2026-03-20T17:57:29.967 INFO:teuthology.orchestra.run.vm06.stdout: debug auth = 20 2026-03-20T17:57:29.967 INFO:teuthology.orchestra.run.vm06.stdout: mon reweight min pgs per osd = 4 2026-03-20T17:57:29.967 INFO:teuthology.orchestra.run.vm06.stdout: mon reweight min bytes per osd = 10 2026-03-20T17:57:29.967 INFO:teuthology.orchestra.run.vm06.stdout: mgr/telemetry/nag = false 2026-03-20T17:57:29.967 INFO:teuthology.orchestra.run.vm06.stdout: 2026-03-20T17:57:29.967 INFO:teuthology.orchestra.run.vm06.stdout:[mon] 2026-03-20T17:57:29.967 INFO:teuthology.orchestra.run.vm06.stdout: debug ms = 1 2026-03-20T17:57:29.967 INFO:teuthology.orchestra.run.vm06.stdout: debug mon = 20 2026-03-20T17:57:29.967 INFO:teuthology.orchestra.run.vm06.stdout: debug paxos = 20 2026-03-20T17:57:29.967 INFO:teuthology.orchestra.run.vm06.stdout: debug auth = 20 2026-03-20T17:57:29.967 INFO:teuthology.orchestra.run.vm06.stdout: mon data avail warn = 5 2026-03-20T17:57:29.967 INFO:teuthology.orchestra.run.vm06.stdout: mon mgr mkfs grace = 240 2026-03-20T17:57:29.967 INFO:teuthology.orchestra.run.vm06.stdout: mon reweight min pgs per osd = 4 2026-03-20T17:57:29.967 INFO:teuthology.orchestra.run.vm06.stdout: mon osd reporter subtree level = osd 2026-03-20T17:57:29.967 INFO:teuthology.orchestra.run.vm06.stdout: mon osd prime pg temp = true 2026-03-20T17:57:29.967 INFO:teuthology.orchestra.run.vm06.stdout: mon reweight min bytes per osd = 10 2026-03-20T17:57:29.967 INFO:teuthology.orchestra.run.vm06.stdout: 2026-03-20T17:57:29.967 INFO:teuthology.orchestra.run.vm06.stdout: # rotate auth tickets quickly to exercise renewal paths 2026-03-20T17:57:29.967 INFO:teuthology.orchestra.run.vm06.stdout: auth mon ticket ttl = 660 # 11m 2026-03-20T17:57:29.968 INFO:teuthology.orchestra.run.vm06.stdout: auth service ticket ttl = 240 # 4m 2026-03-20T17:57:29.968 INFO:teuthology.orchestra.run.vm06.stdout: 2026-03-20T17:57:29.968 INFO:teuthology.orchestra.run.vm06.stdout: # don't complain about insecure global_id in the test suite 2026-03-20T17:57:29.968 INFO:teuthology.orchestra.run.vm06.stdout: mon_warn_on_insecure_global_id_reclaim = false 2026-03-20T17:57:29.968 INFO:teuthology.orchestra.run.vm06.stdout: mon_warn_on_insecure_global_id_reclaim_allowed = false 2026-03-20T17:57:29.968 INFO:teuthology.orchestra.run.vm06.stdout: 2026-03-20T17:57:29.968 INFO:teuthology.orchestra.run.vm06.stdout: # 1m isn't quite enough 2026-03-20T17:57:29.968 INFO:teuthology.orchestra.run.vm06.stdout: mon_down_mkfs_grace = 2m 2026-03-20T17:57:29.968 INFO:teuthology.orchestra.run.vm06.stdout: 2026-03-20T17:57:29.968 INFO:teuthology.orchestra.run.vm06.stdout: mon_warn_on_filestore_osds = false 2026-03-20T17:57:29.968 INFO:teuthology.orchestra.run.vm06.stdout: 2026-03-20T17:57:29.968 INFO:teuthology.orchestra.run.vm06.stdout:[client] 2026-03-20T17:57:29.968 INFO:teuthology.orchestra.run.vm06.stdout: rgw cache enabled = true 2026-03-20T17:57:29.968 INFO:teuthology.orchestra.run.vm06.stdout: rgw enable ops log = true 2026-03-20T17:57:29.968 INFO:teuthology.orchestra.run.vm06.stdout: rgw enable usage log = true 2026-03-20T17:57:29.968 INFO:teuthology.orchestra.run.vm06.stdout: log file = /var/log/ceph/$cluster-$name.$pid.log 2026-03-20T17:57:29.968 INFO:teuthology.orchestra.run.vm06.stdout: admin socket = /var/run/ceph/$cluster-$name.$pid.asok 2026-03-20T17:57:29.968 INFO:teuthology.orchestra.run.vm06.stdout: debug rgw = 20 2026-03-20T17:57:29.968 INFO:teuthology.orchestra.run.vm06.stdout: debug rgw lifecycle = 20 2026-03-20T17:57:29.968 INFO:teuthology.orchestra.run.vm06.stdout: debug rgw notification = 20 2026-03-20T17:57:29.968 INFO:teuthology.orchestra.run.vm06.stdout: debug rgw sync = 20 2026-03-20T17:57:29.968 INFO:teuthology.orchestra.run.vm06.stdout: rgw crypt require ssl = False 2026-03-20T17:57:29.968 INFO:teuthology.orchestra.run.vm06.stdout: rgw crypt s3 kms backend = testing 2026-03-20T17:57:29.968 INFO:teuthology.orchestra.run.vm06.stdout: rgw crypt s3 kms encryption keys = testkey-1=YmluCmJvb3N0CmJvb3N0LWJ1aWxkCmNlcGguY29uZgo= 2026-03-20T17:57:29.968 INFO:teuthology.orchestra.run.vm06.stdout: rgw curl low speed time = 300 2026-03-20T17:57:29.968 INFO:teuthology.orchestra.run.vm06.stdout: rgw data log num shards = 4 2026-03-20T17:57:29.968 INFO:teuthology.orchestra.run.vm06.stdout: rgw data sync poll interval = 5 2026-03-20T17:57:29.968 INFO:teuthology.orchestra.run.vm06.stdout: rgw md log max shards = 4 2026-03-20T17:57:29.968 INFO:teuthology.orchestra.run.vm06.stdout: rgw meta sync poll interval = 5 2026-03-20T17:57:29.968 INFO:teuthology.orchestra.run.vm06.stdout: rgw s3 auth use sts = True 2026-03-20T17:57:29.968 INFO:teuthology.orchestra.run.vm06.stdout: rgw sts key = abcdefghijklmnoq 2026-03-20T17:57:29.968 INFO:teuthology.orchestra.run.vm06.stdout: rgw sync data inject err probability = 0 2026-03-20T17:57:29.968 INFO:teuthology.orchestra.run.vm06.stdout: rgw sync log trim interval = 0 2026-03-20T17:57:29.968 INFO:teuthology.orchestra.run.vm06.stdout: rgw sync meta inject err probability = 0 2026-03-20T17:57:29.968 INFO:teuthology.orchestra.run.vm06.stdout: rgw sync obj etag verify = True 2026-03-20T17:57:29.968 INFO:teuthology.orchestra.run.vm06.stdout: setgroup = ceph 2026-03-20T17:57:29.968 INFO:teuthology.orchestra.run.vm06.stdout: setuser = ceph 2026-03-20T17:57:29.968 INFO:teuthology.orchestra.run.vm06.stdout:[mon.a] 2026-03-20T17:57:29.968 INFO:teuthology.orchestra.run.vm06.stdout:[client.0] 2026-03-20T17:57:29.968 INFO:teuthology.orchestra.run.vm06.stdout: rgw data notify interval msec = 0 2026-03-20T17:57:29.968 INFO:teuthology.orchestra.run.vm06.stdout:[osd.0] 2026-03-20T17:57:29.968 INFO:teuthology.orchestra.run.vm06.stdout: osd_max_omap_entries_per_request = 10 2026-03-20T17:57:29.968 INFO:teuthology.orchestra.run.vm06.stdout:[osd.1] 2026-03-20T17:57:29.968 INFO:teuthology.orchestra.run.vm06.stdout: osd_max_omap_entries_per_request = 1000 2026-03-20T17:57:29.968 INFO:teuthology.orchestra.run.vm06.stdout:[osd.2] 2026-03-20T17:57:29.968 INFO:teuthology.orchestra.run.vm06.stdout: osd_max_omap_entries_per_request = 10000 2026-03-20T17:57:29.970 INFO:tasks.ceph:Creating admin key on c2.mon.a... 2026-03-20T17:57:29.970 DEBUG:teuthology.orchestra.run.vm06:> sudo adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage ceph-authtool --gen-key --name=client.admin --cap mon 'allow *' --cap osd 'allow *' --cap mds 'allow *' --cap mgr 'allow *' /etc/ceph/c2.keyring 2026-03-20T17:57:30.038 INFO:tasks.ceph:Copying monmap to all nodes... 2026-03-20T17:57:30.039 DEBUG:teuthology.orchestra.run.vm06:> set -ex 2026-03-20T17:57:30.039 DEBUG:teuthology.orchestra.run.vm06:> dd if=/etc/ceph/c2.keyring of=/dev/stdout 2026-03-20T17:57:30.083 DEBUG:teuthology.orchestra.run.vm06:> set -ex 2026-03-20T17:57:30.083 DEBUG:teuthology.orchestra.run.vm06:> dd if=/home/ubuntu/cephtest/c2.monmap of=/dev/stdout 2026-03-20T17:57:30.127 INFO:tasks.ceph:Sending monmap to node ubuntu@vm02.local 2026-03-20T17:57:30.127 DEBUG:teuthology.orchestra.run.vm02:> set -ex 2026-03-20T17:57:30.127 DEBUG:teuthology.orchestra.run.vm02:> sudo dd of=/etc/ceph/c2.keyring 2026-03-20T17:57:30.127 DEBUG:teuthology.orchestra.run.vm02:> sudo chmod 0644 /etc/ceph/c2.keyring 2026-03-20T17:57:30.142 DEBUG:teuthology.orchestra.run.vm02:> set -ex 2026-03-20T17:57:30.142 DEBUG:teuthology.orchestra.run.vm02:> dd of=/home/ubuntu/cephtest/c2.monmap 2026-03-20T17:57:30.188 INFO:tasks.ceph:Sending monmap to node ubuntu@vm05.local 2026-03-20T17:57:30.188 DEBUG:teuthology.orchestra.run.vm05:> set -ex 2026-03-20T17:57:30.189 DEBUG:teuthology.orchestra.run.vm05:> sudo dd of=/etc/ceph/c2.keyring 2026-03-20T17:57:30.189 DEBUG:teuthology.orchestra.run.vm05:> sudo chmod 0644 /etc/ceph/c2.keyring 2026-03-20T17:57:30.203 DEBUG:teuthology.orchestra.run.vm05:> set -ex 2026-03-20T17:57:30.203 DEBUG:teuthology.orchestra.run.vm05:> dd of=/home/ubuntu/cephtest/c2.monmap 2026-03-20T17:57:30.247 INFO:tasks.ceph:Sending monmap to node ubuntu@vm06.local 2026-03-20T17:57:30.247 DEBUG:teuthology.orchestra.run.vm06:> set -ex 2026-03-20T17:57:30.247 DEBUG:teuthology.orchestra.run.vm06:> sudo dd of=/etc/ceph/c2.keyring 2026-03-20T17:57:30.247 DEBUG:teuthology.orchestra.run.vm06:> sudo chmod 0644 /etc/ceph/c2.keyring 2026-03-20T17:57:30.260 DEBUG:teuthology.orchestra.run.vm06:> set -ex 2026-03-20T17:57:30.260 DEBUG:teuthology.orchestra.run.vm06:> dd of=/home/ubuntu/cephtest/c2.monmap 2026-03-20T17:57:30.311 INFO:tasks.ceph:Sending monmap to node ubuntu@vm08.local 2026-03-20T17:57:30.311 DEBUG:teuthology.orchestra.run.vm08:> set -ex 2026-03-20T17:57:30.311 DEBUG:teuthology.orchestra.run.vm08:> sudo dd of=/etc/ceph/c2.keyring 2026-03-20T17:57:30.311 DEBUG:teuthology.orchestra.run.vm08:> sudo chmod 0644 /etc/ceph/c2.keyring 2026-03-20T17:57:30.324 DEBUG:teuthology.orchestra.run.vm08:> set -ex 2026-03-20T17:57:30.324 DEBUG:teuthology.orchestra.run.vm08:> dd of=/home/ubuntu/cephtest/c2.monmap 2026-03-20T17:57:30.373 INFO:tasks.ceph:Setting up mon nodes... 2026-03-20T17:57:30.373 INFO:tasks.ceph:Setting up mgr nodes... 2026-03-20T17:57:30.373 DEBUG:teuthology.orchestra.run.vm08:> sudo mkdir -p /var/lib/ceph/mgr/c2-x && sudo adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage ceph-authtool --create-keyring --gen-key --name=mgr.x /var/lib/ceph/mgr/c2-x/keyring 2026-03-20T17:57:30.435 INFO:teuthology.orchestra.run.vm08.stdout:creating /var/lib/ceph/mgr/c2-x/keyring 2026-03-20T17:57:30.437 INFO:tasks.ceph:Setting up mds nodes... 2026-03-20T17:57:30.437 INFO:tasks.ceph_client:Setting up client nodes... 2026-03-20T17:57:30.437 DEBUG:teuthology.orchestra.run.vm06:> sudo adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage ceph-authtool --create-keyring --gen-key --name=client.0 /etc/ceph/c2.client.0.keyring && sudo chmod 0644 /etc/ceph/c2.client.0.keyring 2026-03-20T17:57:30.455 INFO:teuthology.orchestra.run.vm06.stdout:creating /etc/ceph/c2.client.0.keyring 2026-03-20T17:57:30.462 DEBUG:teuthology.orchestra.run.vm08:> sudo adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage ceph-authtool --create-keyring --gen-key --name=client.1 /etc/ceph/c2.client.1.keyring && sudo chmod 0644 /etc/ceph/c2.client.1.keyring 2026-03-20T17:57:30.494 INFO:teuthology.orchestra.run.vm08.stdout:creating /etc/ceph/c2.client.1.keyring 2026-03-20T17:57:30.501 INFO:tasks.ceph:Running mkfs on osd nodes... 2026-03-20T17:57:30.501 INFO:tasks.ceph:ctx.disk_config.remote_to_roles_to_dev: {Remote(name='ubuntu@vm02.local'): {'c1.osd.0': '/dev/vg_nvme/lv_1', 'c1.osd.1': '/dev/vg_nvme/lv_2', 'c1.osd.2': '/dev/vg_nvme/lv_3'}, Remote(name='ubuntu@vm05.local'): {'c1.osd.3': '/dev/vg_nvme/lv_1', 'c1.osd.4': '/dev/vg_nvme/lv_2', 'c1.osd.5': '/dev/vg_nvme/lv_3'}, Remote(name='ubuntu@vm06.local'): {'c2.osd.0': '/dev/vg_nvme/lv_1', 'c2.osd.1': '/dev/vg_nvme/lv_2', 'c2.osd.2': '/dev/vg_nvme/lv_3'}, Remote(name='ubuntu@vm08.local'): {'c2.osd.3': '/dev/vg_nvme/lv_1', 'c2.osd.4': '/dev/vg_nvme/lv_2', 'c2.osd.5': '/dev/vg_nvme/lv_3'}} 2026-03-20T17:57:30.501 DEBUG:teuthology.orchestra.run.vm06:> sudo mkdir -p /var/lib/ceph/osd/c2-0 2026-03-20T17:57:30.511 INFO:tasks.ceph:roles_to_devs: {'c2.osd.0': '/dev/vg_nvme/lv_1', 'c2.osd.1': '/dev/vg_nvme/lv_2', 'c2.osd.2': '/dev/vg_nvme/lv_3'} 2026-03-20T17:57:30.511 INFO:tasks.ceph:role: c2.osd.0 2026-03-20T17:57:30.511 INFO:tasks.ceph:['mkfs.xfs', '-f', '-i', 'size=2048'] on /dev/vg_nvme/lv_1 on ubuntu@vm06.local 2026-03-20T17:57:30.511 DEBUG:teuthology.orchestra.run.vm06:> yes | sudo mkfs.xfs -f -i size=2048 /dev/vg_nvme/lv_1 2026-03-20T17:57:30.560 INFO:teuthology.orchestra.run.vm06.stdout:meta-data=/dev/vg_nvme/lv_1 isize=2048 agcount=4, agsize=1310464 blks 2026-03-20T17:57:30.560 INFO:teuthology.orchestra.run.vm06.stdout: = sectsz=512 attr=2, projid32bit=1 2026-03-20T17:57:30.560 INFO:teuthology.orchestra.run.vm06.stdout: = crc=1 finobt=1, sparse=1, rmapbt=0 2026-03-20T17:57:30.560 INFO:teuthology.orchestra.run.vm06.stdout: = reflink=1 bigtime=0 inobtcount=0 2026-03-20T17:57:30.560 INFO:teuthology.orchestra.run.vm06.stdout:data = bsize=4096 blocks=5241856, imaxpct=25 2026-03-20T17:57:30.560 INFO:teuthology.orchestra.run.vm06.stdout: = sunit=0 swidth=0 blks 2026-03-20T17:57:30.561 INFO:teuthology.orchestra.run.vm06.stdout:naming =version 2 bsize=4096 ascii-ci=0, ftype=1 2026-03-20T17:57:30.561 INFO:teuthology.orchestra.run.vm06.stdout:log =internal log bsize=4096 blocks=2560, version=2 2026-03-20T17:57:30.561 INFO:teuthology.orchestra.run.vm06.stdout: = sectsz=512 sunit=0 blks, lazy-count=1 2026-03-20T17:57:30.561 INFO:teuthology.orchestra.run.vm06.stdout:realtime =none extsz=4096 blocks=0, rtextents=0 2026-03-20T17:57:30.565 INFO:teuthology.orchestra.run.vm06.stdout:Discarding blocks...Done. 2026-03-20T17:57:30.566 INFO:tasks.ceph:mount /dev/vg_nvme/lv_1 on ubuntu@vm06.local -o noatime 2026-03-20T17:57:30.566 DEBUG:teuthology.orchestra.run.vm06:> sudo mount -t xfs -o noatime /dev/vg_nvme/lv_1 /var/lib/ceph/osd/c2-0 2026-03-20T17:57:30.653 DEBUG:teuthology.orchestra.run.vm06:> sudo /sbin/restorecon /var/lib/ceph/osd/c2-0 2026-03-20T17:57:30.659 INFO:teuthology.orchestra.run.vm06.stderr:sudo: /sbin/restorecon: command not found 2026-03-20T17:57:30.659 DEBUG:teuthology.orchestra.run:got remote process result: 1 2026-03-20T17:57:30.659 DEBUG:teuthology.orchestra.run.vm06:> sudo mkdir -p /var/lib/ceph/osd/c2-1 2026-03-20T17:57:30.712 INFO:tasks.ceph:roles_to_devs: {'c2.osd.0': '/dev/vg_nvme/lv_1', 'c2.osd.1': '/dev/vg_nvme/lv_2', 'c2.osd.2': '/dev/vg_nvme/lv_3'} 2026-03-20T17:57:30.712 INFO:tasks.ceph:role: c2.osd.1 2026-03-20T17:57:30.712 INFO:tasks.ceph:['mkfs.xfs', '-f', '-i', 'size=2048'] on /dev/vg_nvme/lv_2 on ubuntu@vm06.local 2026-03-20T17:57:30.712 DEBUG:teuthology.orchestra.run.vm06:> yes | sudo mkfs.xfs -f -i size=2048 /dev/vg_nvme/lv_2 2026-03-20T17:57:30.767 INFO:teuthology.orchestra.run.vm06.stdout:meta-data=/dev/vg_nvme/lv_2 isize=2048 agcount=4, agsize=1310464 blks 2026-03-20T17:57:30.768 INFO:teuthology.orchestra.run.vm06.stdout: = sectsz=512 attr=2, projid32bit=1 2026-03-20T17:57:30.768 INFO:teuthology.orchestra.run.vm06.stdout: = crc=1 finobt=1, sparse=1, rmapbt=0 2026-03-20T17:57:30.768 INFO:teuthology.orchestra.run.vm06.stdout: = reflink=1 bigtime=0 inobtcount=0 2026-03-20T17:57:30.768 INFO:teuthology.orchestra.run.vm06.stdout:data = bsize=4096 blocks=5241856, imaxpct=25 2026-03-20T17:57:30.768 INFO:teuthology.orchestra.run.vm06.stdout: = sunit=0 swidth=0 blks 2026-03-20T17:57:30.768 INFO:teuthology.orchestra.run.vm06.stdout:naming =version 2 bsize=4096 ascii-ci=0, ftype=1 2026-03-20T17:57:30.768 INFO:teuthology.orchestra.run.vm06.stdout:log =internal log bsize=4096 blocks=2560, version=2 2026-03-20T17:57:30.768 INFO:teuthology.orchestra.run.vm06.stdout: = sectsz=512 sunit=0 blks, lazy-count=1 2026-03-20T17:57:30.768 INFO:teuthology.orchestra.run.vm06.stdout:realtime =none extsz=4096 blocks=0, rtextents=0 2026-03-20T17:57:30.772 INFO:teuthology.orchestra.run.vm06.stdout:Discarding blocks...Done. 2026-03-20T17:57:30.773 INFO:tasks.ceph:mount /dev/vg_nvme/lv_2 on ubuntu@vm06.local -o noatime 2026-03-20T17:57:30.774 DEBUG:teuthology.orchestra.run.vm06:> sudo mount -t xfs -o noatime /dev/vg_nvme/lv_2 /var/lib/ceph/osd/c2-1 2026-03-20T17:57:30.829 DEBUG:teuthology.orchestra.run.vm06:> sudo /sbin/restorecon /var/lib/ceph/osd/c2-1 2026-03-20T17:57:30.878 INFO:teuthology.orchestra.run.vm06.stderr:sudo: /sbin/restorecon: command not found 2026-03-20T17:57:30.878 DEBUG:teuthology.orchestra.run:got remote process result: 1 2026-03-20T17:57:30.878 DEBUG:teuthology.orchestra.run.vm06:> sudo mkdir -p /var/lib/ceph/osd/c2-2 2026-03-20T17:57:30.927 INFO:tasks.ceph:roles_to_devs: {'c2.osd.0': '/dev/vg_nvme/lv_1', 'c2.osd.1': '/dev/vg_nvme/lv_2', 'c2.osd.2': '/dev/vg_nvme/lv_3'} 2026-03-20T17:57:30.927 INFO:tasks.ceph:role: c2.osd.2 2026-03-20T17:57:30.927 INFO:tasks.ceph:['mkfs.xfs', '-f', '-i', 'size=2048'] on /dev/vg_nvme/lv_3 on ubuntu@vm06.local 2026-03-20T17:57:30.927 DEBUG:teuthology.orchestra.run.vm06:> yes | sudo mkfs.xfs -f -i size=2048 /dev/vg_nvme/lv_3 2026-03-20T17:57:30.976 INFO:teuthology.orchestra.run.vm06.stdout:meta-data=/dev/vg_nvme/lv_3 isize=2048 agcount=4, agsize=1310464 blks 2026-03-20T17:57:30.976 INFO:teuthology.orchestra.run.vm06.stdout: = sectsz=512 attr=2, projid32bit=1 2026-03-20T17:57:30.976 INFO:teuthology.orchestra.run.vm06.stdout: = crc=1 finobt=1, sparse=1, rmapbt=0 2026-03-20T17:57:30.976 INFO:teuthology.orchestra.run.vm06.stdout: = reflink=1 bigtime=0 inobtcount=0 2026-03-20T17:57:30.976 INFO:teuthology.orchestra.run.vm06.stdout:data = bsize=4096 blocks=5241856, imaxpct=25 2026-03-20T17:57:30.976 INFO:teuthology.orchestra.run.vm06.stdout: = sunit=0 swidth=0 blks 2026-03-20T17:57:30.976 INFO:teuthology.orchestra.run.vm06.stdout:naming =version 2 bsize=4096 ascii-ci=0, ftype=1 2026-03-20T17:57:30.976 INFO:teuthology.orchestra.run.vm06.stdout:log =internal log bsize=4096 blocks=2560, version=2 2026-03-20T17:57:30.976 INFO:teuthology.orchestra.run.vm06.stdout: = sectsz=512 sunit=0 blks, lazy-count=1 2026-03-20T17:57:30.976 INFO:teuthology.orchestra.run.vm06.stdout:realtime =none extsz=4096 blocks=0, rtextents=0 2026-03-20T17:57:30.980 INFO:teuthology.orchestra.run.vm06.stdout:Discarding blocks...Done. 2026-03-20T17:57:30.982 INFO:tasks.ceph:mount /dev/vg_nvme/lv_3 on ubuntu@vm06.local -o noatime 2026-03-20T17:57:30.982 DEBUG:teuthology.orchestra.run.vm06:> sudo mount -t xfs -o noatime /dev/vg_nvme/lv_3 /var/lib/ceph/osd/c2-2 2026-03-20T17:57:31.035 DEBUG:teuthology.orchestra.run.vm06:> sudo /sbin/restorecon /var/lib/ceph/osd/c2-2 2026-03-20T17:57:31.082 INFO:teuthology.orchestra.run.vm06.stderr:sudo: /sbin/restorecon: command not found 2026-03-20T17:57:31.082 DEBUG:teuthology.orchestra.run:got remote process result: 1 2026-03-20T17:57:31.083 DEBUG:teuthology.orchestra.run.vm06:> sudo MALLOC_CHECK_=3 adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage ceph-osd --no-mon-config --cluster c2 --mkfs --mkkey -i 0 --monmap /home/ubuntu/cephtest/c2.monmap 2026-03-20T17:57:31.149 INFO:teuthology.orchestra.run.vm06.stderr:2026-03-20T17:57:31.143+0000 7fc51e30ba40 -1 auth: error reading file: /var/lib/ceph/osd/c2-0/keyring: can't open /var/lib/ceph/osd/c2-0/keyring: (2) No such file or directory 2026-03-20T17:57:31.150 INFO:teuthology.orchestra.run.vm06.stderr:2026-03-20T17:57:31.143+0000 7fc51e30ba40 -1 created new key in keyring /var/lib/ceph/osd/c2-0/keyring 2026-03-20T17:57:31.150 INFO:teuthology.orchestra.run.vm06.stderr:2026-03-20T17:57:31.143+0000 7fc51e30ba40 -1 bdev(0x5569bcef3800 /var/lib/ceph/osd/c2-0/block) open stat got: (1) Operation not permitted 2026-03-20T17:57:31.150 INFO:teuthology.orchestra.run.vm06.stderr:2026-03-20T17:57:31.143+0000 7fc51e30ba40 -1 bluestore(/var/lib/ceph/osd/c2-0) _read_fsid unparsable uuid 2026-03-20T17:57:31.744 DEBUG:teuthology.orchestra.run.vm06:> sudo chown -R ceph:ceph /var/lib/ceph/osd/c2-0 2026-03-20T17:57:31.796 DEBUG:teuthology.orchestra.run.vm06:> sudo MALLOC_CHECK_=3 adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage ceph-osd --no-mon-config --cluster c2 --mkfs --mkkey -i 1 --monmap /home/ubuntu/cephtest/c2.monmap 2026-03-20T17:57:31.861 INFO:teuthology.orchestra.run.vm06.stderr:2026-03-20T17:57:31.855+0000 7f955cd31a40 -1 auth: error reading file: /var/lib/ceph/osd/c2-1/keyring: can't open /var/lib/ceph/osd/c2-1/keyring: (2) No such file or directory 2026-03-20T17:57:31.861 INFO:teuthology.orchestra.run.vm06.stderr:2026-03-20T17:57:31.855+0000 7f955cd31a40 -1 created new key in keyring /var/lib/ceph/osd/c2-1/keyring 2026-03-20T17:57:31.862 INFO:teuthology.orchestra.run.vm06.stderr:2026-03-20T17:57:31.855+0000 7f955cd31a40 -1 bdev(0x562b25fa5800 /var/lib/ceph/osd/c2-1/block) open stat got: (1) Operation not permitted 2026-03-20T17:57:31.862 INFO:teuthology.orchestra.run.vm06.stderr:2026-03-20T17:57:31.855+0000 7f955cd31a40 -1 bluestore(/var/lib/ceph/osd/c2-1) _read_fsid unparsable uuid 2026-03-20T17:57:32.496 DEBUG:teuthology.orchestra.run.vm06:> sudo chown -R ceph:ceph /var/lib/ceph/osd/c2-1 2026-03-20T17:57:32.548 DEBUG:teuthology.orchestra.run.vm06:> sudo MALLOC_CHECK_=3 adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage ceph-osd --no-mon-config --cluster c2 --mkfs --mkkey -i 2 --monmap /home/ubuntu/cephtest/c2.monmap 2026-03-20T17:57:32.614 INFO:teuthology.orchestra.run.vm06.stderr:2026-03-20T17:57:32.607+0000 7fba456b5a40 -1 auth: error reading file: /var/lib/ceph/osd/c2-2/keyring: can't open /var/lib/ceph/osd/c2-2/keyring: (2) No such file or directory 2026-03-20T17:57:32.614 INFO:teuthology.orchestra.run.vm06.stderr:2026-03-20T17:57:32.607+0000 7fba456b5a40 -1 created new key in keyring /var/lib/ceph/osd/c2-2/keyring 2026-03-20T17:57:32.614 INFO:teuthology.orchestra.run.vm06.stderr:2026-03-20T17:57:32.607+0000 7fba456b5a40 -1 bdev(0x55d034817800 /var/lib/ceph/osd/c2-2/block) open stat got: (1) Operation not permitted 2026-03-20T17:57:32.614 INFO:teuthology.orchestra.run.vm06.stderr:2026-03-20T17:57:32.607+0000 7fba456b5a40 -1 bluestore(/var/lib/ceph/osd/c2-2) _read_fsid unparsable uuid 2026-03-20T17:57:33.454 DEBUG:teuthology.orchestra.run.vm06:> sudo chown -R ceph:ceph /var/lib/ceph/osd/c2-2 2026-03-20T17:57:33.509 DEBUG:teuthology.orchestra.run.vm08:> sudo mkdir -p /var/lib/ceph/osd/c2-3 2026-03-20T17:57:33.517 INFO:tasks.ceph:roles_to_devs: {'c2.osd.3': '/dev/vg_nvme/lv_1', 'c2.osd.4': '/dev/vg_nvme/lv_2', 'c2.osd.5': '/dev/vg_nvme/lv_3'} 2026-03-20T17:57:33.517 INFO:tasks.ceph:role: c2.osd.3 2026-03-20T17:57:33.517 INFO:tasks.ceph:['mkfs.xfs', '-f', '-i', 'size=2048'] on /dev/vg_nvme/lv_1 on ubuntu@vm08.local 2026-03-20T17:57:33.517 DEBUG:teuthology.orchestra.run.vm08:> yes | sudo mkfs.xfs -f -i size=2048 /dev/vg_nvme/lv_1 2026-03-20T17:57:33.567 INFO:teuthology.orchestra.run.vm08.stdout:meta-data=/dev/vg_nvme/lv_1 isize=2048 agcount=4, agsize=1310464 blks 2026-03-20T17:57:33.568 INFO:teuthology.orchestra.run.vm08.stdout: = sectsz=512 attr=2, projid32bit=1 2026-03-20T17:57:33.568 INFO:teuthology.orchestra.run.vm08.stdout: = crc=1 finobt=1, sparse=1, rmapbt=0 2026-03-20T17:57:33.568 INFO:teuthology.orchestra.run.vm08.stdout: = reflink=1 bigtime=0 inobtcount=0 2026-03-20T17:57:33.568 INFO:teuthology.orchestra.run.vm08.stdout:data = bsize=4096 blocks=5241856, imaxpct=25 2026-03-20T17:57:33.568 INFO:teuthology.orchestra.run.vm08.stdout: = sunit=0 swidth=0 blks 2026-03-20T17:57:33.568 INFO:teuthology.orchestra.run.vm08.stdout:naming =version 2 bsize=4096 ascii-ci=0, ftype=1 2026-03-20T17:57:33.568 INFO:teuthology.orchestra.run.vm08.stdout:log =internal log bsize=4096 blocks=2560, version=2 2026-03-20T17:57:33.568 INFO:teuthology.orchestra.run.vm08.stdout: = sectsz=512 sunit=0 blks, lazy-count=1 2026-03-20T17:57:33.568 INFO:teuthology.orchestra.run.vm08.stdout:realtime =none extsz=4096 blocks=0, rtextents=0 2026-03-20T17:57:33.572 INFO:teuthology.orchestra.run.vm08.stdout:Discarding blocks...Done. 2026-03-20T17:57:33.575 INFO:tasks.ceph:mount /dev/vg_nvme/lv_1 on ubuntu@vm08.local -o noatime 2026-03-20T17:57:33.575 DEBUG:teuthology.orchestra.run.vm08:> sudo mount -t xfs -o noatime /dev/vg_nvme/lv_1 /var/lib/ceph/osd/c2-3 2026-03-20T17:57:33.665 DEBUG:teuthology.orchestra.run.vm08:> sudo /sbin/restorecon /var/lib/ceph/osd/c2-3 2026-03-20T17:57:33.671 INFO:teuthology.orchestra.run.vm08.stderr:sudo: /sbin/restorecon: command not found 2026-03-20T17:57:33.672 DEBUG:teuthology.orchestra.run:got remote process result: 1 2026-03-20T17:57:33.672 DEBUG:teuthology.orchestra.run.vm08:> sudo mkdir -p /var/lib/ceph/osd/c2-4 2026-03-20T17:57:33.722 INFO:tasks.ceph:roles_to_devs: {'c2.osd.3': '/dev/vg_nvme/lv_1', 'c2.osd.4': '/dev/vg_nvme/lv_2', 'c2.osd.5': '/dev/vg_nvme/lv_3'} 2026-03-20T17:57:33.722 INFO:tasks.ceph:role: c2.osd.4 2026-03-20T17:57:33.722 INFO:tasks.ceph:['mkfs.xfs', '-f', '-i', 'size=2048'] on /dev/vg_nvme/lv_2 on ubuntu@vm08.local 2026-03-20T17:57:33.722 DEBUG:teuthology.orchestra.run.vm08:> yes | sudo mkfs.xfs -f -i size=2048 /dev/vg_nvme/lv_2 2026-03-20T17:57:33.773 INFO:teuthology.orchestra.run.vm08.stdout:meta-data=/dev/vg_nvme/lv_2 isize=2048 agcount=4, agsize=1310464 blks 2026-03-20T17:57:33.773 INFO:teuthology.orchestra.run.vm08.stdout: = sectsz=512 attr=2, projid32bit=1 2026-03-20T17:57:33.773 INFO:teuthology.orchestra.run.vm08.stdout: = crc=1 finobt=1, sparse=1, rmapbt=0 2026-03-20T17:57:33.773 INFO:teuthology.orchestra.run.vm08.stdout: = reflink=1 bigtime=0 inobtcount=0 2026-03-20T17:57:33.773 INFO:teuthology.orchestra.run.vm08.stdout:data = bsize=4096 blocks=5241856, imaxpct=25 2026-03-20T17:57:33.773 INFO:teuthology.orchestra.run.vm08.stdout: = sunit=0 swidth=0 blks 2026-03-20T17:57:33.773 INFO:teuthology.orchestra.run.vm08.stdout:naming =version 2 bsize=4096 ascii-ci=0, ftype=1 2026-03-20T17:57:33.773 INFO:teuthology.orchestra.run.vm08.stdout:log =internal log bsize=4096 blocks=2560, version=2 2026-03-20T17:57:33.773 INFO:teuthology.orchestra.run.vm08.stdout: = sectsz=512 sunit=0 blks, lazy-count=1 2026-03-20T17:57:33.773 INFO:teuthology.orchestra.run.vm08.stdout:realtime =none extsz=4096 blocks=0, rtextents=0 2026-03-20T17:57:33.779 INFO:teuthology.orchestra.run.vm08.stdout:Discarding blocks...Done. 2026-03-20T17:57:33.780 INFO:tasks.ceph:mount /dev/vg_nvme/lv_2 on ubuntu@vm08.local -o noatime 2026-03-20T17:57:33.780 DEBUG:teuthology.orchestra.run.vm08:> sudo mount -t xfs -o noatime /dev/vg_nvme/lv_2 /var/lib/ceph/osd/c2-4 2026-03-20T17:57:33.833 DEBUG:teuthology.orchestra.run.vm08:> sudo /sbin/restorecon /var/lib/ceph/osd/c2-4 2026-03-20T17:57:33.879 INFO:teuthology.orchestra.run.vm08.stderr:sudo: /sbin/restorecon: command not found 2026-03-20T17:57:33.880 DEBUG:teuthology.orchestra.run:got remote process result: 1 2026-03-20T17:57:33.880 DEBUG:teuthology.orchestra.run.vm08:> sudo mkdir -p /var/lib/ceph/osd/c2-5 2026-03-20T17:57:33.929 INFO:tasks.ceph:roles_to_devs: {'c2.osd.3': '/dev/vg_nvme/lv_1', 'c2.osd.4': '/dev/vg_nvme/lv_2', 'c2.osd.5': '/dev/vg_nvme/lv_3'} 2026-03-20T17:57:33.929 INFO:tasks.ceph:role: c2.osd.5 2026-03-20T17:57:33.929 INFO:tasks.ceph:['mkfs.xfs', '-f', '-i', 'size=2048'] on /dev/vg_nvme/lv_3 on ubuntu@vm08.local 2026-03-20T17:57:33.929 DEBUG:teuthology.orchestra.run.vm08:> yes | sudo mkfs.xfs -f -i size=2048 /dev/vg_nvme/lv_3 2026-03-20T17:57:33.981 INFO:teuthology.orchestra.run.vm08.stdout:meta-data=/dev/vg_nvme/lv_3 isize=2048 agcount=4, agsize=1310464 blks 2026-03-20T17:57:33.981 INFO:teuthology.orchestra.run.vm08.stdout: = sectsz=512 attr=2, projid32bit=1 2026-03-20T17:57:33.981 INFO:teuthology.orchestra.run.vm08.stdout: = crc=1 finobt=1, sparse=1, rmapbt=0 2026-03-20T17:57:33.981 INFO:teuthology.orchestra.run.vm08.stdout: = reflink=1 bigtime=0 inobtcount=0 2026-03-20T17:57:33.981 INFO:teuthology.orchestra.run.vm08.stdout:data = bsize=4096 blocks=5241856, imaxpct=25 2026-03-20T17:57:33.981 INFO:teuthology.orchestra.run.vm08.stdout: = sunit=0 swidth=0 blks 2026-03-20T17:57:33.981 INFO:teuthology.orchestra.run.vm08.stdout:naming =version 2 bsize=4096 ascii-ci=0, ftype=1 2026-03-20T17:57:33.981 INFO:teuthology.orchestra.run.vm08.stdout:log =internal log bsize=4096 blocks=2560, version=2 2026-03-20T17:57:33.981 INFO:teuthology.orchestra.run.vm08.stdout: = sectsz=512 sunit=0 blks, lazy-count=1 2026-03-20T17:57:33.981 INFO:teuthology.orchestra.run.vm08.stdout:realtime =none extsz=4096 blocks=0, rtextents=0 2026-03-20T17:57:33.986 INFO:teuthology.orchestra.run.vm08.stdout:Discarding blocks...Done. 2026-03-20T17:57:33.988 INFO:tasks.ceph:mount /dev/vg_nvme/lv_3 on ubuntu@vm08.local -o noatime 2026-03-20T17:57:33.988 DEBUG:teuthology.orchestra.run.vm08:> sudo mount -t xfs -o noatime /dev/vg_nvme/lv_3 /var/lib/ceph/osd/c2-5 2026-03-20T17:57:34.045 DEBUG:teuthology.orchestra.run.vm08:> sudo /sbin/restorecon /var/lib/ceph/osd/c2-5 2026-03-20T17:57:34.092 INFO:teuthology.orchestra.run.vm08.stderr:sudo: /sbin/restorecon: command not found 2026-03-20T17:57:34.092 DEBUG:teuthology.orchestra.run:got remote process result: 1 2026-03-20T17:57:34.092 DEBUG:teuthology.orchestra.run.vm08:> sudo MALLOC_CHECK_=3 adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage ceph-osd --no-mon-config --cluster c2 --mkfs --mkkey -i 3 --monmap /home/ubuntu/cephtest/c2.monmap 2026-03-20T17:57:34.154 INFO:teuthology.orchestra.run.vm08.stderr:2026-03-20T17:57:34.153+0000 7f6178c63a40 -1 auth: error reading file: /var/lib/ceph/osd/c2-3/keyring: can't open /var/lib/ceph/osd/c2-3/keyring: (2) No such file or directory 2026-03-20T17:57:34.155 INFO:teuthology.orchestra.run.vm08.stderr:2026-03-20T17:57:34.153+0000 7f6178c63a40 -1 created new key in keyring /var/lib/ceph/osd/c2-3/keyring 2026-03-20T17:57:34.155 INFO:teuthology.orchestra.run.vm08.stderr:2026-03-20T17:57:34.153+0000 7f6178c63a40 -1 bdev(0x55e061001800 /var/lib/ceph/osd/c2-3/block) open stat got: (1) Operation not permitted 2026-03-20T17:57:34.155 INFO:teuthology.orchestra.run.vm08.stderr:2026-03-20T17:57:34.153+0000 7f6178c63a40 -1 bluestore(/var/lib/ceph/osd/c2-3) _read_fsid unparsable uuid 2026-03-20T17:57:34.769 DEBUG:teuthology.orchestra.run.vm08:> sudo chown -R ceph:ceph /var/lib/ceph/osd/c2-3 2026-03-20T17:57:34.817 DEBUG:teuthology.orchestra.run.vm08:> sudo MALLOC_CHECK_=3 adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage ceph-osd --no-mon-config --cluster c2 --mkfs --mkkey -i 4 --monmap /home/ubuntu/cephtest/c2.monmap 2026-03-20T17:57:34.877 INFO:teuthology.orchestra.run.vm08.stderr:2026-03-20T17:57:34.873+0000 7f4134339a40 -1 auth: error reading file: /var/lib/ceph/osd/c2-4/keyring: can't open /var/lib/ceph/osd/c2-4/keyring: (2) No such file or directory 2026-03-20T17:57:34.878 INFO:teuthology.orchestra.run.vm08.stderr:2026-03-20T17:57:34.873+0000 7f4134339a40 -1 created new key in keyring /var/lib/ceph/osd/c2-4/keyring 2026-03-20T17:57:34.878 INFO:teuthology.orchestra.run.vm08.stderr:2026-03-20T17:57:34.873+0000 7f4134339a40 -1 bdev(0x556f251f9800 /var/lib/ceph/osd/c2-4/block) open stat got: (1) Operation not permitted 2026-03-20T17:57:34.878 INFO:teuthology.orchestra.run.vm08.stderr:2026-03-20T17:57:34.873+0000 7f4134339a40 -1 bluestore(/var/lib/ceph/osd/c2-4) _read_fsid unparsable uuid 2026-03-20T17:57:35.525 DEBUG:teuthology.orchestra.run.vm08:> sudo chown -R ceph:ceph /var/lib/ceph/osd/c2-4 2026-03-20T17:57:35.573 DEBUG:teuthology.orchestra.run.vm08:> sudo MALLOC_CHECK_=3 adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage ceph-osd --no-mon-config --cluster c2 --mkfs --mkkey -i 5 --monmap /home/ubuntu/cephtest/c2.monmap 2026-03-20T17:57:35.634 INFO:teuthology.orchestra.run.vm08.stderr:2026-03-20T17:57:35.633+0000 7fddf54a0a40 -1 auth: error reading file: /var/lib/ceph/osd/c2-5/keyring: can't open /var/lib/ceph/osd/c2-5/keyring: (2) No such file or directory 2026-03-20T17:57:35.635 INFO:teuthology.orchestra.run.vm08.stderr:2026-03-20T17:57:35.633+0000 7fddf54a0a40 -1 created new key in keyring /var/lib/ceph/osd/c2-5/keyring 2026-03-20T17:57:35.635 INFO:teuthology.orchestra.run.vm08.stderr:2026-03-20T17:57:35.633+0000 7fddf54a0a40 -1 bdev(0x561f0b5b9800 /var/lib/ceph/osd/c2-5/block) open stat got: (1) Operation not permitted 2026-03-20T17:57:35.635 INFO:teuthology.orchestra.run.vm08.stderr:2026-03-20T17:57:35.633+0000 7fddf54a0a40 -1 bluestore(/var/lib/ceph/osd/c2-5) _read_fsid unparsable uuid 2026-03-20T17:57:36.272 DEBUG:teuthology.orchestra.run.vm08:> sudo chown -R ceph:ceph /var/lib/ceph/osd/c2-5 2026-03-20T17:57:36.321 INFO:tasks.ceph:Reading keys from all nodes... 2026-03-20T17:57:36.321 DEBUG:teuthology.orchestra.run.vm06:> set -ex 2026-03-20T17:57:36.322 DEBUG:teuthology.orchestra.run.vm06:> sudo dd if=/var/lib/ceph/osd/c2-0/keyring of=/dev/stdout 2026-03-20T17:57:36.329 DEBUG:teuthology.orchestra.run.vm06:> set -ex 2026-03-20T17:57:36.329 DEBUG:teuthology.orchestra.run.vm06:> sudo dd if=/var/lib/ceph/osd/c2-1/keyring of=/dev/stdout 2026-03-20T17:57:36.378 DEBUG:teuthology.orchestra.run.vm06:> set -ex 2026-03-20T17:57:36.378 DEBUG:teuthology.orchestra.run.vm06:> sudo dd if=/var/lib/ceph/osd/c2-2/keyring of=/dev/stdout 2026-03-20T17:57:36.426 DEBUG:teuthology.orchestra.run.vm08:> set -ex 2026-03-20T17:57:36.426 DEBUG:teuthology.orchestra.run.vm08:> sudo dd if=/var/lib/ceph/mgr/c2-x/keyring of=/dev/stdout 2026-03-20T17:57:36.433 DEBUG:teuthology.orchestra.run.vm08:> set -ex 2026-03-20T17:57:36.434 DEBUG:teuthology.orchestra.run.vm08:> sudo dd if=/var/lib/ceph/osd/c2-3/keyring of=/dev/stdout 2026-03-20T17:57:36.481 DEBUG:teuthology.orchestra.run.vm08:> set -ex 2026-03-20T17:57:36.481 DEBUG:teuthology.orchestra.run.vm08:> sudo dd if=/var/lib/ceph/osd/c2-4/keyring of=/dev/stdout 2026-03-20T17:57:36.528 DEBUG:teuthology.orchestra.run.vm08:> set -ex 2026-03-20T17:57:36.528 DEBUG:teuthology.orchestra.run.vm08:> sudo dd if=/var/lib/ceph/osd/c2-5/keyring of=/dev/stdout 2026-03-20T17:57:36.577 DEBUG:teuthology.orchestra.run.vm06:> set -ex 2026-03-20T17:57:36.577 DEBUG:teuthology.orchestra.run.vm06:> dd if=/etc/ceph/c2.client.0.keyring of=/dev/stdout 2026-03-20T17:57:36.580 DEBUG:teuthology.orchestra.run.vm08:> set -ex 2026-03-20T17:57:36.580 DEBUG:teuthology.orchestra.run.vm08:> dd if=/etc/ceph/c2.client.1.keyring of=/dev/stdout 2026-03-20T17:57:36.621 INFO:tasks.ceph:Adding keys to all mons... 2026-03-20T17:57:36.621 DEBUG:teuthology.orchestra.run.vm06:> sudo tee -a /etc/ceph/c2.keyring 2026-03-20T17:57:36.630 INFO:teuthology.orchestra.run.vm06.stdout:[osd.0] 2026-03-20T17:57:36.630 INFO:teuthology.orchestra.run.vm06.stdout: key = AQCLir1pn+PbCBAAl4Cz2JOzQm6NFZt2YYHo+Q== 2026-03-20T17:57:36.630 INFO:teuthology.orchestra.run.vm06.stdout:[osd.1] 2026-03-20T17:57:36.630 INFO:teuthology.orchestra.run.vm06.stdout: key = AQCLir1pDhpHMxAAz/HB94AvChbLDmHnRMRxng== 2026-03-20T17:57:36.630 INFO:teuthology.orchestra.run.vm06.stdout:[osd.2] 2026-03-20T17:57:36.630 INFO:teuthology.orchestra.run.vm06.stdout: key = AQCMir1pp/aBJBAAw7Ajvuf3fKEC8MO08/+6Tg== 2026-03-20T17:57:36.630 INFO:teuthology.orchestra.run.vm06.stdout:[mgr.x] 2026-03-20T17:57:36.630 INFO:teuthology.orchestra.run.vm06.stdout: key = AQCKir1p7BP5GRAAoQk1cYY7NFsUO4r6u5AKgA== 2026-03-20T17:57:36.630 INFO:teuthology.orchestra.run.vm06.stdout:[osd.3] 2026-03-20T17:57:36.630 INFO:teuthology.orchestra.run.vm06.stdout: key = AQCOir1pTKg4CRAAlUn/smgUyyJgkXX2pZeraQ== 2026-03-20T17:57:36.630 INFO:teuthology.orchestra.run.vm06.stdout:[osd.4] 2026-03-20T17:57:36.630 INFO:teuthology.orchestra.run.vm06.stdout: key = AQCOir1pe2dRNBAAWR1KR3naF3NRUw8Mr+bc3w== 2026-03-20T17:57:36.630 INFO:teuthology.orchestra.run.vm06.stdout:[osd.5] 2026-03-20T17:57:36.630 INFO:teuthology.orchestra.run.vm06.stdout: key = AQCPir1pn/nTJRAAeBeXVvJVIUqUO9JkPOR4mA== 2026-03-20T17:57:36.630 INFO:teuthology.orchestra.run.vm06.stdout:[client.0] 2026-03-20T17:57:36.630 INFO:teuthology.orchestra.run.vm06.stdout: key = AQCKir1pWp4XGxAAsKtmlFdq/+dESFjNSUf19A== 2026-03-20T17:57:36.630 INFO:teuthology.orchestra.run.vm06.stdout:[client.1] 2026-03-20T17:57:36.631 INFO:teuthology.orchestra.run.vm06.stdout: key = AQCKir1pzsiAHRAANSBQ/yTsDoq/sg6d2OnWyA== 2026-03-20T17:57:36.631 DEBUG:teuthology.orchestra.run.vm06:> sudo adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage ceph-authtool /etc/ceph/c2.keyring --name=osd.0 --cap mon 'allow profile osd' --cap mgr 'allow profile osd' --cap osd 'allow *' 2026-03-20T17:57:36.696 DEBUG:teuthology.orchestra.run.vm06:> sudo adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage ceph-authtool /etc/ceph/c2.keyring --name=osd.1 --cap mon 'allow profile osd' --cap mgr 'allow profile osd' --cap osd 'allow *' 2026-03-20T17:57:36.761 DEBUG:teuthology.orchestra.run.vm06:> sudo adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage ceph-authtool /etc/ceph/c2.keyring --name=osd.2 --cap mon 'allow profile osd' --cap mgr 'allow profile osd' --cap osd 'allow *' 2026-03-20T17:57:36.825 DEBUG:teuthology.orchestra.run.vm06:> sudo adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage ceph-authtool /etc/ceph/c2.keyring --name=mgr.x --cap mon 'allow profile mgr' --cap osd 'allow *' --cap mds 'allow *' 2026-03-20T17:57:36.893 DEBUG:teuthology.orchestra.run.vm06:> sudo adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage ceph-authtool /etc/ceph/c2.keyring --name=osd.3 --cap mon 'allow profile osd' --cap mgr 'allow profile osd' --cap osd 'allow *' 2026-03-20T17:57:36.962 DEBUG:teuthology.orchestra.run.vm06:> sudo adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage ceph-authtool /etc/ceph/c2.keyring --name=osd.4 --cap mon 'allow profile osd' --cap mgr 'allow profile osd' --cap osd 'allow *' 2026-03-20T17:57:37.025 DEBUG:teuthology.orchestra.run.vm06:> sudo adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage ceph-authtool /etc/ceph/c2.keyring --name=osd.5 --cap mon 'allow profile osd' --cap mgr 'allow profile osd' --cap osd 'allow *' 2026-03-20T17:57:37.090 DEBUG:teuthology.orchestra.run.vm06:> sudo adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage ceph-authtool /etc/ceph/c2.keyring --name=client.0 --cap mon 'allow rw' --cap mgr 'allow r' --cap osd 'allow rwx' --cap mds allow 2026-03-20T17:57:37.155 DEBUG:teuthology.orchestra.run.vm06:> sudo adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage ceph-authtool /etc/ceph/c2.keyring --name=client.1 --cap mon 'allow rw' --cap mgr 'allow r' --cap osd 'allow rwx' --cap mds allow 2026-03-20T17:57:37.219 INFO:tasks.ceph:Running mkfs on mon nodes... 2026-03-20T17:57:37.219 DEBUG:teuthology.orchestra.run.vm06:> sudo mkdir -p /var/lib/ceph/mon/c2-a 2026-03-20T17:57:37.267 DEBUG:teuthology.orchestra.run.vm06:> sudo adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage ceph-mon --cluster c2 --mkfs -i a --monmap /home/ubuntu/cephtest/c2.monmap --keyring /etc/ceph/c2.keyring 2026-03-20T17:57:37.345 DEBUG:teuthology.orchestra.run.vm06:> sudo chown -R ceph:ceph /var/lib/ceph/mon/c2-a 2026-03-20T17:57:37.395 DEBUG:teuthology.orchestra.run.vm06:> rm -- /home/ubuntu/cephtest/c2.monmap 2026-03-20T17:57:37.439 INFO:tasks.ceph:Starting mon daemons in cluster c2... 2026-03-20T17:57:37.439 INFO:tasks.ceph.c2.mon.a:Restarting daemon 2026-03-20T17:57:37.439 DEBUG:teuthology.orchestra.run.vm06:> sudo adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage daemon-helper kill ceph-mon -f --cluster c2 -i a 2026-03-20T17:57:37.481 INFO:tasks.ceph.c2.mon.a:Started 2026-03-20T17:57:37.481 INFO:tasks.ceph:Starting mgr daemons in cluster c2... 2026-03-20T17:57:37.481 INFO:tasks.ceph.c2.mgr.x:Restarting daemon 2026-03-20T17:57:37.481 DEBUG:teuthology.orchestra.run.vm08:> sudo adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage daemon-helper kill ceph-mgr -f --cluster c2 -i x 2026-03-20T17:57:37.482 INFO:tasks.ceph.c2.mgr.x:Started 2026-03-20T17:57:37.482 DEBUG:tasks.ceph:set 0 configs 2026-03-20T17:57:37.482 DEBUG:teuthology.orchestra.run.vm06:> sudo ceph --cluster c2 config dump 2026-03-20T17:57:37.588 INFO:teuthology.orchestra.run.vm06.stdout:WHO MASK LEVEL OPTION VALUE RO 2026-03-20T17:57:37.600 INFO:tasks.ceph:Setting crush tunables to default 2026-03-20T17:57:37.601 DEBUG:teuthology.orchestra.run.vm06:> sudo ceph --cluster c2 osd crush tunables default 2026-03-20T17:57:37.709 INFO:teuthology.orchestra.run.vm06.stderr:adjusted tunables profile to default 2026-03-20T17:57:37.722 INFO:tasks.ceph:check_enable_crimson: False 2026-03-20T17:57:37.722 INFO:tasks.ceph:Starting osd daemons in cluster c2... 2026-03-20T17:57:37.722 DEBUG:teuthology.orchestra.run.vm06:> set -ex 2026-03-20T17:57:37.722 DEBUG:teuthology.orchestra.run.vm06:> sudo dd if=/var/lib/ceph/osd/c2-0/fsid of=/dev/stdout 2026-03-20T17:57:37.731 DEBUG:teuthology.orchestra.run.vm06:> set -ex 2026-03-20T17:57:37.731 DEBUG:teuthology.orchestra.run.vm06:> sudo dd if=/var/lib/ceph/osd/c2-1/fsid of=/dev/stdout 2026-03-20T17:57:37.780 DEBUG:teuthology.orchestra.run.vm06:> set -ex 2026-03-20T17:57:37.780 DEBUG:teuthology.orchestra.run.vm06:> sudo dd if=/var/lib/ceph/osd/c2-2/fsid of=/dev/stdout 2026-03-20T17:57:37.828 DEBUG:teuthology.orchestra.run.vm08:> set -ex 2026-03-20T17:57:37.828 DEBUG:teuthology.orchestra.run.vm08:> sudo dd if=/var/lib/ceph/osd/c2-3/fsid of=/dev/stdout 2026-03-20T17:57:37.837 DEBUG:teuthology.orchestra.run.vm08:> set -ex 2026-03-20T17:57:37.837 DEBUG:teuthology.orchestra.run.vm08:> sudo dd if=/var/lib/ceph/osd/c2-4/fsid of=/dev/stdout 2026-03-20T17:57:37.886 DEBUG:teuthology.orchestra.run.vm08:> set -ex 2026-03-20T17:57:37.886 DEBUG:teuthology.orchestra.run.vm08:> sudo dd if=/var/lib/ceph/osd/c2-5/fsid of=/dev/stdout 2026-03-20T17:57:37.938 DEBUG:teuthology.orchestra.run.vm08:> sudo ceph --cluster c2 osd new 3425c007-e53b-452b-83a2-185100f81062 0 2026-03-20T17:57:38.090 INFO:teuthology.orchestra.run.vm08.stdout:0 2026-03-20T17:57:38.103 DEBUG:teuthology.orchestra.run.vm08:> sudo ceph --cluster c2 osd new 6b897cf8-b9b9-46da-a8e9-a4fa94a6bcff 1 2026-03-20T17:57:38.207 INFO:teuthology.orchestra.run.vm08.stdout:1 2026-03-20T17:57:38.220 DEBUG:teuthology.orchestra.run.vm08:> sudo ceph --cluster c2 osd new 16f37894-1ade-42af-a668-a7c2d98c3d31 2 2026-03-20T17:57:38.322 INFO:teuthology.orchestra.run.vm08.stdout:2 2026-03-20T17:57:38.334 DEBUG:teuthology.orchestra.run.vm08:> sudo ceph --cluster c2 osd new 0fa43dbc-6811-4426-9ff2-82a296a19efb 3 2026-03-20T17:57:38.437 INFO:teuthology.orchestra.run.vm08.stdout:3 2026-03-20T17:57:38.449 DEBUG:teuthology.orchestra.run.vm08:> sudo ceph --cluster c2 osd new e58e5154-9b17-4109-926c-b20bd25945b6 4 2026-03-20T17:57:38.553 INFO:teuthology.orchestra.run.vm08.stdout:4 2026-03-20T17:57:38.566 DEBUG:teuthology.orchestra.run.vm08:> sudo ceph --cluster c2 osd new 9d3c09ae-4c03-41c0-88b1-13e6e36d67bf 5 2026-03-20T17:57:38.672 INFO:teuthology.orchestra.run.vm08.stdout:5 2026-03-20T17:57:38.684 INFO:tasks.ceph.c2.osd.0:Restarting daemon 2026-03-20T17:57:38.684 DEBUG:teuthology.orchestra.run.vm06:> sudo adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage daemon-helper kill ceph-osd -f --cluster c2 -i 0 2026-03-20T17:57:38.685 INFO:tasks.ceph.c2.osd.0:Started 2026-03-20T17:57:38.685 INFO:tasks.ceph.c2.osd.1:Restarting daemon 2026-03-20T17:57:38.685 DEBUG:teuthology.orchestra.run.vm06:> sudo adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage daemon-helper kill ceph-osd -f --cluster c2 -i 1 2026-03-20T17:57:38.686 INFO:tasks.ceph.c2.osd.1:Started 2026-03-20T17:57:38.686 INFO:tasks.ceph.c2.osd.2:Restarting daemon 2026-03-20T17:57:38.686 DEBUG:teuthology.orchestra.run.vm06:> sudo adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage daemon-helper kill ceph-osd -f --cluster c2 -i 2 2026-03-20T17:57:38.687 INFO:tasks.ceph.c2.osd.2:Started 2026-03-20T17:57:38.687 INFO:tasks.ceph.c2.osd.3:Restarting daemon 2026-03-20T17:57:38.687 DEBUG:teuthology.orchestra.run.vm08:> sudo adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage daemon-helper kill ceph-osd -f --cluster c2 -i 3 2026-03-20T17:57:38.687 INFO:tasks.ceph.c2.osd.3:Started 2026-03-20T17:57:38.687 INFO:tasks.ceph.c2.osd.4:Restarting daemon 2026-03-20T17:57:38.688 DEBUG:teuthology.orchestra.run.vm08:> sudo adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage daemon-helper kill ceph-osd -f --cluster c2 -i 4 2026-03-20T17:57:38.688 INFO:tasks.ceph.c2.osd.4:Started 2026-03-20T17:57:38.688 INFO:tasks.ceph.c2.osd.5:Restarting daemon 2026-03-20T17:57:38.688 DEBUG:teuthology.orchestra.run.vm08:> sudo adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage daemon-helper kill ceph-osd -f --cluster c2 -i 5 2026-03-20T17:57:38.689 INFO:tasks.ceph.c2.osd.5:Started 2026-03-20T17:57:38.689 DEBUG:teuthology.orchestra.run.vm06:> sudo adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage timeout 120 ceph --cluster c2 osd dump --format=json 2026-03-20T17:57:38.806 INFO:teuthology.orchestra.run.vm06.stdout: 2026-03-20T17:57:38.806 INFO:teuthology.orchestra.run.vm06.stdout:{"epoch":8,"fsid":"02f24f96-dd4e-4d53-915a-6d7df6cbe6f6","created":"2026-03-20T17:57:37.534249+0000","modified":"2026-03-20T17:57:38.670004+0000","last_up_change":"0.000000","last_in_change":"2026-03-20T17:57:38.670004+0000","flags":"sortbitwise,recovery_deletes,purged_snapdirs,pglog_hardlimit","flags_num":5799936,"flags_set":["pglog_hardlimit","purged_snapdirs","recovery_deletes","sortbitwise"],"crush_version":2,"full_ratio":0.94999998807907104,"backfillfull_ratio":0.89999997615814209,"nearfull_ratio":0.85000002384185791,"cluster_snapshot":"","pool_max":0,"max_osd":6,"require_min_compat_client":"luminous","min_compat_client":"jewel","require_osd_release":"tentacle","allow_crimson":false,"pools":[],"osds":[{"osd":0,"uuid":"3425c007-e53b-452b-83a2-185100f81062","up":0,"in":1,"weight":1,"primary_affinity":1,"last_clean_begin":0,"last_clean_end":0,"up_from":0,"up_thru":0,"down_at":0,"lost_at":0,"public_addrs":{"addrvec":[]},"cluster_addrs":{"addrvec":[]},"heartbeat_back_addrs":{"addrvec":[]},"heartbeat_front_addrs":{"addrvec":[]},"public_addr":"(unrecognized address family 0)/0","cluster_addr":"(unrecognized address family 0)/0","heartbeat_back_addr":"(unrecognized address family 0)/0","heartbeat_front_addr":"(unrecognized address family 0)/0","state":["exists","new"]},{"osd":1,"uuid":"6b897cf8-b9b9-46da-a8e9-a4fa94a6bcff","up":0,"in":1,"weight":1,"primary_affinity":1,"last_clean_begin":0,"last_clean_end":0,"up_from":0,"up_thru":0,"down_at":0,"lost_at":0,"public_addrs":{"addrvec":[]},"cluster_addrs":{"addrvec":[]},"heartbeat_back_addrs":{"addrvec":[]},"heartbeat_front_addrs":{"addrvec":[]},"public_addr":"(unrecognized address family 0)/0","cluster_addr":"(unrecognized address family 0)/0","heartbeat_back_addr":"(unrecognized address family 0)/0","heartbeat_front_addr":"(unrecognized address family 0)/0","state":["exists","new"]},{"osd":2,"uuid":"16f37894-1ade-42af-a668-a7c2d98c3d31","up":0,"in":1,"weight":1,"primary_affinity":1,"last_clean_begin":0,"last_clean_end":0,"up_from":0,"up_thru":0,"down_at":0,"lost_at":0,"public_addrs":{"addrvec":[]},"cluster_addrs":{"addrvec":[]},"heartbeat_back_addrs":{"addrvec":[]},"heartbeat_front_addrs":{"addrvec":[]},"public_addr":"(unrecognized address family 0)/0","cluster_addr":"(unrecognized address family 0)/0","heartbeat_back_addr":"(unrecognized address family 0)/0","heartbeat_front_addr":"(unrecognized address family 0)/0","state":["exists","new"]},{"osd":3,"uuid":"0fa43dbc-6811-4426-9ff2-82a296a19efb","up":0,"in":1,"weight":1,"primary_affinity":1,"last_clean_begin":0,"last_clean_end":0,"up_from":0,"up_thru":0,"down_at":0,"lost_at":0,"public_addrs":{"addrvec":[]},"cluster_addrs":{"addrvec":[]},"heartbeat_back_addrs":{"addrvec":[]},"heartbeat_front_addrs":{"addrvec":[]},"public_addr":"(unrecognized address family 0)/0","cluster_addr":"(unrecognized address family 0)/0","heartbeat_back_addr":"(unrecognized address family 0)/0","heartbeat_front_addr":"(unrecognized address family 0)/0","state":["exists","new"]},{"osd":4,"uuid":"e58e5154-9b17-4109-926c-b20bd25945b6","up":0,"in":1,"weight":1,"primary_affinity":1,"last_clean_begin":0,"last_clean_end":0,"up_from":0,"up_thru":0,"down_at":0,"lost_at":0,"public_addrs":{"addrvec":[]},"cluster_addrs":{"addrvec":[]},"heartbeat_back_addrs":{"addrvec":[]},"heartbeat_front_addrs":{"addrvec":[]},"public_addr":"(unrecognized address family 0)/0","cluster_addr":"(unrecognized address family 0)/0","heartbeat_back_addr":"(unrecognized address family 0)/0","heartbeat_front_addr":"(unrecognized address family 0)/0","state":["exists","new"]},{"osd":5,"uuid":"9d3c09ae-4c03-41c0-88b1-13e6e36d67bf","up":0,"in":1,"weight":1,"primary_affinity":1,"last_clean_begin":0,"last_clean_end":0,"up_from":0,"up_thru":0,"down_at":0,"lost_at":0,"public_addrs":{"addrvec":[]},"cluster_addrs":{"addrvec":[]},"heartbeat_back_addrs":{"addrvec":[]},"heartbeat_front_addrs":{"addrvec":[]},"public_addr":"(unrecognized address family 0)/0","cluster_addr":"(unrecognized address family 0)/0","heartbeat_back_addr":"(unrecognized address family 0)/0","heartbeat_front_addr":"(unrecognized address family 0)/0","state":["exists","new"]}],"osd_xinfo":[{"osd":0,"down_stamp":"0.000000","laggy_probability":0,"laggy_interval":0,"features":0,"old_weight":0,"last_purged_snaps_scrub":"0.000000","dead_epoch":0},{"osd":1,"down_stamp":"0.000000","laggy_probability":0,"laggy_interval":0,"features":0,"old_weight":0,"last_purged_snaps_scrub":"0.000000","dead_epoch":0},{"osd":2,"down_stamp":"0.000000","laggy_probability":0,"laggy_interval":0,"features":0,"old_weight":0,"last_purged_snaps_scrub":"0.000000","dead_epoch":0},{"osd":3,"down_stamp":"0.000000","laggy_probability":0,"laggy_interval":0,"features":0,"old_weight":0,"last_purged_snaps_scrub":"0.000000","dead_epoch":0},{"osd":4,"down_stamp":"0.000000","laggy_probability":0,"laggy_interval":0,"features":0,"old_weight":0,"last_purged_snaps_scrub":"0.000000","dead_epoch":0},{"osd":5,"down_stamp":"0.000000","laggy_probability":0,"laggy_interval":0,"features":0,"old_weight":0,"last_purged_snaps_scrub":"0.000000","dead_epoch":0}],"pg_upmap":[],"pg_upmap_items":[],"pg_upmap_primaries":[],"pg_temp":[],"primary_temp":[],"blocklist":{},"range_blocklist":{},"erasure_code_profiles":{"default":{"crush-failure-domain":"osd","k":"2","m":"1","plugin":"isa","technique":"reed_sol_van"}},"removed_snaps_queue":[],"new_removed_snaps":[],"new_purged_snaps":[],"crush_node_flags":{},"device_class_flags":{},"stretch_mode":{"stretch_mode_enabled":false,"stretch_bucket_count":0,"degraded_stretch_mode":0,"recovering_stretch_mode":0,"stretch_mode_bucket":0}} 2026-03-20T17:57:38.808 INFO:tasks.ceph.c2.mgr.x.vm08.stderr:/usr/lib/python3/dist-packages/scipy/__init__.py:67: UserWarning: NumPy was imported from a Python sub-interpreter but NumPy does not properly support sub-interpreters. This will likely work for most users but might cause hard to track down issues or subtle bugs. A common user of the rare sub-interpreter feature is wsgi which also allows single-interpreter mode. 2026-03-20T17:57:38.808 INFO:tasks.ceph.c2.mgr.x.vm08.stderr:Improvements in the case of bugs are welcome, but is not on the NumPy roadmap, and full support may require significant effort to achieve. 2026-03-20T17:57:38.808 INFO:tasks.ceph.c2.mgr.x.vm08.stderr: from numpy import show_config as show_numpy_config 2026-03-20T17:57:38.818 INFO:tasks.ceph.ceph_manager.c2:[] 2026-03-20T17:57:38.819 INFO:tasks.ceph:Waiting for OSDs to come up 2026-03-20T17:57:38.975 INFO:tasks.ceph.c2.osd.3.vm08.stderr:2026-03-20T17:57:38.973+0000 7f280d2eba40 -1 Falling back to public interface 2026-03-20T17:57:39.009 INFO:tasks.ceph.c2.osd.0.vm06.stderr:2026-03-20T17:57:39.003+0000 7f1ee1c27a40 -1 Falling back to public interface 2026-03-20T17:57:39.024 INFO:tasks.ceph.c2.osd.1.vm06.stderr:2026-03-20T17:57:39.019+0000 7f2e00557a40 -1 Falling back to public interface 2026-03-20T17:57:39.035 INFO:tasks.ceph.c2.osd.5.vm08.stderr:2026-03-20T17:57:39.033+0000 7fe0888c2a40 -1 Falling back to public interface 2026-03-20T17:57:39.044 INFO:tasks.ceph.c2.osd.2.vm06.stderr:2026-03-20T17:57:39.039+0000 7fc8a0205a40 -1 Falling back to public interface 2026-03-20T17:57:39.051 INFO:tasks.ceph.c2.osd.4.vm08.stderr:2026-03-20T17:57:39.049+0000 7f81ed8a5a40 -1 Falling back to public interface 2026-03-20T17:57:39.223 INFO:tasks.ceph.c2.osd.3.vm08.stderr:2026-03-20T17:57:39.221+0000 7f280d2eba40 -1 osd.3 0 log_to_monitors true 2026-03-20T17:57:39.300 INFO:tasks.ceph.c2.osd.1.vm06.stderr:2026-03-20T17:57:39.291+0000 7f2e00557a40 -1 osd.1 0 log_to_monitors true 2026-03-20T17:57:39.313 INFO:tasks.ceph.c2.osd.5.vm08.stderr:2026-03-20T17:57:39.309+0000 7fe0888c2a40 -1 osd.5 0 log_to_monitors true 2026-03-20T17:57:39.321 INFO:tasks.ceph.c2.osd.0.vm06.stderr:2026-03-20T17:57:39.315+0000 7f1ee1c27a40 -1 osd.0 0 log_to_monitors true 2026-03-20T17:57:39.344 INFO:tasks.ceph.c2.osd.4.vm08.stderr:2026-03-20T17:57:39.341+0000 7f81ed8a5a40 -1 osd.4 0 log_to_monitors true 2026-03-20T17:57:39.374 INFO:tasks.ceph.c2.osd.2.vm06.stderr:2026-03-20T17:57:39.367+0000 7fc8a0205a40 -1 osd.2 0 log_to_monitors true 2026-03-20T17:57:39.394 INFO:tasks.ceph.c2.mgr.x.vm08.stderr:Failed to import NVMeoFClient and related components: cannot import name 'NVMeoFClient' from 'dashboard.services.nvmeof_client' (/usr/share/ceph/mgr/dashboard/services/nvmeof_client.py) 2026-03-20T17:57:39.421 DEBUG:teuthology.orchestra.run.vm06:> adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage ceph --cluster c2 osd dump --format=json 2026-03-20T17:57:39.513 INFO:teuthology.misc.health.vm06.stdout: 2026-03-20T17:57:39.513 INFO:teuthology.misc.health.vm06.stdout:{"epoch":8,"fsid":"02f24f96-dd4e-4d53-915a-6d7df6cbe6f6","created":"2026-03-20T17:57:37.534249+0000","modified":"2026-03-20T17:57:38.670004+0000","last_up_change":"0.000000","last_in_change":"2026-03-20T17:57:38.670004+0000","flags":"sortbitwise,recovery_deletes,purged_snapdirs,pglog_hardlimit","flags_num":5799936,"flags_set":["pglog_hardlimit","purged_snapdirs","recovery_deletes","sortbitwise"],"crush_version":2,"full_ratio":0.94999998807907104,"backfillfull_ratio":0.89999997615814209,"nearfull_ratio":0.85000002384185791,"cluster_snapshot":"","pool_max":0,"max_osd":6,"require_min_compat_client":"luminous","min_compat_client":"jewel","require_osd_release":"tentacle","allow_crimson":false,"pools":[],"osds":[{"osd":0,"uuid":"3425c007-e53b-452b-83a2-185100f81062","up":0,"in":1,"weight":1,"primary_affinity":1,"last_clean_begin":0,"last_clean_end":0,"up_from":0,"up_thru":0,"down_at":0,"lost_at":0,"public_addrs":{"addrvec":[]},"cluster_addrs":{"addrvec":[]},"heartbeat_back_addrs":{"addrvec":[]},"heartbeat_front_addrs":{"addrvec":[]},"public_addr":"(unrecognized address family 0)/0","cluster_addr":"(unrecognized address family 0)/0","heartbeat_back_addr":"(unrecognized address family 0)/0","heartbeat_front_addr":"(unrecognized address family 0)/0","state":["exists","new"]},{"osd":1,"uuid":"6b897cf8-b9b9-46da-a8e9-a4fa94a6bcff","up":0,"in":1,"weight":1,"primary_affinity":1,"last_clean_begin":0,"last_clean_end":0,"up_from":0,"up_thru":0,"down_at":0,"lost_at":0,"public_addrs":{"addrvec":[]},"cluster_addrs":{"addrvec":[]},"heartbeat_back_addrs":{"addrvec":[]},"heartbeat_front_addrs":{"addrvec":[]},"public_addr":"(unrecognized address family 0)/0","cluster_addr":"(unrecognized address family 0)/0","heartbeat_back_addr":"(unrecognized address family 0)/0","heartbeat_front_addr":"(unrecognized address family 0)/0","state":["exists","new"]},{"osd":2,"uuid":"16f37894-1ade-42af-a668-a7c2d98c3d31","up":0,"in":1,"weight":1,"primary_affinity":1,"last_clean_begin":0,"last_clean_end":0,"up_from":0,"up_thru":0,"down_at":0,"lost_at":0,"public_addrs":{"addrvec":[]},"cluster_addrs":{"addrvec":[]},"heartbeat_back_addrs":{"addrvec":[]},"heartbeat_front_addrs":{"addrvec":[]},"public_addr":"(unrecognized address family 0)/0","cluster_addr":"(unrecognized address family 0)/0","heartbeat_back_addr":"(unrecognized address family 0)/0","heartbeat_front_addr":"(unrecognized address family 0)/0","state":["exists","new"]},{"osd":3,"uuid":"0fa43dbc-6811-4426-9ff2-82a296a19efb","up":0,"in":1,"weight":1,"primary_affinity":1,"last_clean_begin":0,"last_clean_end":0,"up_from":0,"up_thru":0,"down_at":0,"lost_at":0,"public_addrs":{"addrvec":[]},"cluster_addrs":{"addrvec":[]},"heartbeat_back_addrs":{"addrvec":[]},"heartbeat_front_addrs":{"addrvec":[]},"public_addr":"(unrecognized address family 0)/0","cluster_addr":"(unrecognized address family 0)/0","heartbeat_back_addr":"(unrecognized address family 0)/0","heartbeat_front_addr":"(unrecognized address family 0)/0","state":["exists","new"]},{"osd":4,"uuid":"e58e5154-9b17-4109-926c-b20bd25945b6","up":0,"in":1,"weight":1,"primary_affinity":1,"last_clean_begin":0,"last_clean_end":0,"up_from":0,"up_thru":0,"down_at":0,"lost_at":0,"public_addrs":{"addrvec":[]},"cluster_addrs":{"addrvec":[]},"heartbeat_back_addrs":{"addrvec":[]},"heartbeat_front_addrs":{"addrvec":[]},"public_addr":"(unrecognized address family 0)/0","cluster_addr":"(unrecognized address family 0)/0","heartbeat_back_addr":"(unrecognized address family 0)/0","heartbeat_front_addr":"(unrecognized address family 0)/0","state":["exists","new"]},{"osd":5,"uuid":"9d3c09ae-4c03-41c0-88b1-13e6e36d67bf","up":0,"in":1,"weight":1,"primary_affinity":1,"last_clean_begin":0,"last_clean_end":0,"up_from":0,"up_thru":0,"down_at":0,"lost_at":0,"public_addrs":{"addrvec":[]},"cluster_addrs":{"addrvec":[]},"heartbeat_back_addrs":{"addrvec":[]},"heartbeat_front_addrs":{"addrvec":[]},"public_addr":"(unrecognized address family 0)/0","cluster_addr":"(unrecognized address family 0)/0","heartbeat_back_addr":"(unrecognized address family 0)/0","heartbeat_front_addr":"(unrecognized address family 0)/0","state":["exists","new"]}],"osd_xinfo":[{"osd":0,"down_stamp":"0.000000","laggy_probability":0,"laggy_interval":0,"features":0,"old_weight":0,"last_purged_snaps_scrub":"0.000000","dead_epoch":0},{"osd":1,"down_stamp":"0.000000","laggy_probability":0,"laggy_interval":0,"features":0,"old_weight":0,"last_purged_snaps_scrub":"0.000000","dead_epoch":0},{"osd":2,"down_stamp":"0.000000","laggy_probability":0,"laggy_interval":0,"features":0,"old_weight":0,"last_purged_snaps_scrub":"0.000000","dead_epoch":0},{"osd":3,"down_stamp":"0.000000","laggy_probability":0,"laggy_interval":0,"features":0,"old_weight":0,"last_purged_snaps_scrub":"0.000000","dead_epoch":0},{"osd":4,"down_stamp":"0.000000","laggy_probability":0,"laggy_interval":0,"features":0,"old_weight":0,"last_purged_snaps_scrub":"0.000000","dead_epoch":0},{"osd":5,"down_stamp":"0.000000","laggy_probability":0,"laggy_interval":0,"features":0,"old_weight":0,"last_purged_snaps_scrub":"0.000000","dead_epoch":0}],"pg_upmap":[],"pg_upmap_items":[],"pg_upmap_primaries":[],"pg_temp":[],"primary_temp":[],"blocklist":{},"range_blocklist":{},"erasure_code_profiles":{"default":{"crush-failure-domain":"osd","k":"2","m":"1","plugin":"isa","technique":"reed_sol_van"}},"removed_snaps_queue":[],"new_removed_snaps":[],"new_purged_snaps":[],"crush_node_flags":{},"device_class_flags":{},"stretch_mode":{"stretch_mode_enabled":false,"stretch_bucket_count":0,"degraded_stretch_mode":0,"recovering_stretch_mode":0,"stretch_mode_bucket":0}} 2026-03-20T17:57:39.523 DEBUG:teuthology.misc:0 of 6 OSDs are up 2026-03-20T17:57:40.692 INFO:tasks.ceph.c2.osd.2.vm06.stderr:2026-03-20T17:57:40.687+0000 7fc89c1ae640 -1 osd.2 0 waiting for initial osdmap 2026-03-20T17:57:40.694 INFO:tasks.ceph.c2.osd.4.vm08.stderr:2026-03-20T17:57:40.693+0000 7f81e984e640 -1 osd.4 0 waiting for initial osdmap 2026-03-20T17:57:40.694 INFO:tasks.ceph.c2.osd.5.vm08.stderr:2026-03-20T17:57:40.693+0000 7fe08486b640 -1 osd.5 0 waiting for initial osdmap 2026-03-20T17:57:40.694 INFO:tasks.ceph.c2.osd.3.vm08.stderr:2026-03-20T17:57:40.693+0000 7f2809294640 -1 osd.3 0 waiting for initial osdmap 2026-03-20T17:57:40.695 INFO:tasks.ceph.c2.osd.2.vm06.stderr:2026-03-20T17:57:40.687+0000 7fc896fbc640 -1 osd.2 10 set_numa_affinity unable to identify public interface '' numa node: (2) No such file or directory 2026-03-20T17:57:40.698 INFO:tasks.ceph.c2.osd.4.vm08.stderr:2026-03-20T17:57:40.693+0000 7f81e465c640 -1 osd.4 10 set_numa_affinity unable to identify public interface '' numa node: (2) No such file or directory 2026-03-20T17:57:40.698 INFO:tasks.ceph.c2.osd.5.vm08.stderr:2026-03-20T17:57:40.693+0000 7fe07f679640 -1 osd.5 10 set_numa_affinity unable to identify public interface '' numa node: (2) No such file or directory 2026-03-20T17:57:40.698 INFO:tasks.ceph.c2.osd.3.vm08.stderr:2026-03-20T17:57:40.693+0000 7f28040a2640 -1 osd.3 10 set_numa_affinity unable to identify public interface '' numa node: (2) No such file or directory 2026-03-20T17:57:40.700 INFO:tasks.ceph.c2.osd.0.vm06.stderr:2026-03-20T17:57:40.695+0000 7f1ede3e2640 -1 osd.0 0 waiting for initial osdmap 2026-03-20T17:57:40.700 INFO:tasks.ceph.c2.osd.1.vm06.stderr:2026-03-20T17:57:40.695+0000 7f2dfcd12640 -1 osd.1 0 waiting for initial osdmap 2026-03-20T17:57:40.702 INFO:tasks.ceph.c2.osd.0.vm06.stderr:2026-03-20T17:57:40.695+0000 7f1ed89de640 -1 osd.0 10 set_numa_affinity unable to identify public interface '' numa node: (2) No such file or directory 2026-03-20T17:57:40.703 INFO:tasks.ceph.c2.osd.1.vm06.stderr:2026-03-20T17:57:40.695+0000 7f2df730e640 -1 osd.1 10 set_numa_affinity unable to identify public interface '' numa node: (2) No such file or directory 2026-03-20T17:57:40.849 INFO:tasks.ceph.c2.mgr.x.vm08.stderr:2026-03-20T17:57:40.845+0000 7f6490ac5640 -1 mgr.server handle_report got status from non-daemon mon.a 2026-03-20T17:57:46.129 DEBUG:teuthology.orchestra.run.vm06:> adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage ceph --cluster c2 osd dump --format=json 2026-03-20T17:57:46.303 INFO:teuthology.misc.health.vm06.stdout: 2026-03-20T17:57:46.303 INFO:teuthology.misc.health.vm06.stdout:{"epoch":14,"fsid":"02f24f96-dd4e-4d53-915a-6d7df6cbe6f6","created":"2026-03-20T17:57:37.534249+0000","modified":"2026-03-20T17:57:44.862882+0000","last_up_change":"2026-03-20T17:57:41.678232+0000","last_in_change":"2026-03-20T17:57:38.670004+0000","flags":"sortbitwise,recovery_deletes,purged_snapdirs,pglog_hardlimit","flags_num":5799936,"flags_set":["pglog_hardlimit","purged_snapdirs","recovery_deletes","sortbitwise"],"crush_version":4,"full_ratio":0.94999998807907104,"backfillfull_ratio":0.89999997615814209,"nearfull_ratio":0.85000002384185791,"cluster_snapshot":"","pool_max":1,"max_osd":6,"require_min_compat_client":"luminous","min_compat_client":"jewel","require_osd_release":"tentacle","allow_crimson":false,"pools":[{"pool":1,"pool_name":".mgr","create_time":"2026-03-20T17:57:41.855010+0000","flags":1,"flags_names":"hashpspool","type":1,"size":2,"min_size":1,"crush_rule":0,"peering_crush_bucket_count":0,"peering_crush_bucket_target":0,"peering_crush_bucket_barrier":0,"peering_crush_bucket_mandatory_member":2147483647,"is_stretch_pool":false,"object_hash":2,"pg_autoscale_mode":"off","pg_num":1,"pg_placement_num":1,"pg_placement_num_target":1,"pg_num_target":1,"pg_num_pending":1,"last_pg_merge_meta":{"source_pgid":"0.0","ready_epoch":0,"last_epoch_started":0,"last_epoch_clean":0,"source_version":"0'0","target_version":"0'0"},"last_change":"14","last_force_op_resend":"0","last_force_op_resend_prenautilus":"0","last_force_op_resend_preluminous":"0","auid":0,"snap_mode":"selfmanaged","snap_seq":0,"snap_epoch":0,"pool_snaps":[],"removed_snaps":"[]","quota_max_bytes":0,"quota_max_objects":0,"tiers":[],"tier_of":-1,"read_tier":-1,"write_tier":-1,"cache_mode":"none","target_max_bytes":0,"target_max_objects":0,"cache_target_dirty_ratio_micro":400000,"cache_target_dirty_high_ratio_micro":600000,"cache_target_full_ratio_micro":800000,"cache_min_flush_age":0,"cache_min_evict_age":0,"erasure_code_profile":"","hit_set_params":{"type":"none"},"hit_set_period":0,"hit_set_count":0,"use_gmt_hitset":true,"min_read_recency_for_promote":0,"min_write_recency_for_promote":0,"hit_set_grade_decay_rate":0,"hit_set_search_last_n":0,"grade_table":[],"stripe_width":0,"expected_num_objects":0,"fast_read":false,"nonprimary_shards":"{}","options":{"pg_num_max":32,"pg_num_min":1},"application_metadata":{"mgr":{}},"read_balance":{"score_type":"Fair distribution","score_acting":6.059999942779541,"score_stable":6.059999942779541,"optimal_score":0.33000001311302185,"raw_score_acting":2,"raw_score_stable":2,"primary_affinity_weighted":1,"average_primary_affinity":1,"average_primary_affinity_weighted":1}}],"osds":[{"osd":0,"uuid":"3425c007-e53b-452b-83a2-185100f81062","up":1,"in":1,"weight":1,"primary_affinity":1,"last_clean_begin":0,"last_clean_end":0,"up_from":11,"up_thru":0,"down_at":0,"lost_at":0,"public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.106:6800","nonce":3200331076},{"type":"v1","addr":"192.168.123.106:6801","nonce":3200331076}]},"cluster_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.106:6802","nonce":3200331076},{"type":"v1","addr":"192.168.123.106:6803","nonce":3200331076}]},"heartbeat_back_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.106:6806","nonce":3200331076},{"type":"v1","addr":"192.168.123.106:6807","nonce":3200331076}]},"heartbeat_front_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.106:6804","nonce":3200331076},{"type":"v1","addr":"192.168.123.106:6805","nonce":3200331076}]},"public_addr":"192.168.123.106:6801/3200331076","cluster_addr":"192.168.123.106:6803/3200331076","heartbeat_back_addr":"192.168.123.106:6807/3200331076","heartbeat_front_addr":"192.168.123.106:6805/3200331076","state":["exists","up"]},{"osd":1,"uuid":"6b897cf8-b9b9-46da-a8e9-a4fa94a6bcff","up":1,"in":1,"weight":1,"primary_affinity":1,"last_clean_begin":0,"last_clean_end":0,"up_from":11,"up_thru":12,"down_at":0,"lost_at":0,"public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.106:6808","nonce":630280801},{"type":"v1","addr":"192.168.123.106:6809","nonce":630280801}]},"cluster_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.106:6810","nonce":630280801},{"type":"v1","addr":"192.168.123.106:6811","nonce":630280801}]},"heartbeat_back_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.106:6814","nonce":630280801},{"type":"v1","addr":"192.168.123.106:6815","nonce":630280801}]},"heartbeat_front_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.106:6812","nonce":630280801},{"type":"v1","addr":"192.168.123.106:6813","nonce":630280801}]},"public_addr":"192.168.123.106:6809/630280801","cluster_addr":"192.168.123.106:6811/630280801","heartbeat_back_addr":"192.168.123.106:6815/630280801","heartbeat_front_addr":"192.168.123.106:6813/630280801","state":["exists","up"]},{"osd":2,"uuid":"16f37894-1ade-42af-a668-a7c2d98c3d31","up":1,"in":1,"weight":1,"primary_affinity":1,"last_clean_begin":0,"last_clean_end":0,"up_from":11,"up_thru":0,"down_at":0,"lost_at":0,"public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.106:6816","nonce":905863174},{"type":"v1","addr":"192.168.123.106:6817","nonce":905863174}]},"cluster_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.106:6818","nonce":905863174},{"type":"v1","addr":"192.168.123.106:6819","nonce":905863174}]},"heartbeat_back_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.106:6822","nonce":905863174},{"type":"v1","addr":"192.168.123.106:6823","nonce":905863174}]},"heartbeat_front_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.106:6820","nonce":905863174},{"type":"v1","addr":"192.168.123.106:6821","nonce":905863174}]},"public_addr":"192.168.123.106:6817/905863174","cluster_addr":"192.168.123.106:6819/905863174","heartbeat_back_addr":"192.168.123.106:6823/905863174","heartbeat_front_addr":"192.168.123.106:6821/905863174","state":["exists","up"]},{"osd":3,"uuid":"0fa43dbc-6811-4426-9ff2-82a296a19efb","up":1,"in":1,"weight":1,"primary_affinity":1,"last_clean_begin":0,"last_clean_end":0,"up_from":11,"up_thru":0,"down_at":0,"lost_at":0,"public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.108:6800","nonce":1817589236},{"type":"v1","addr":"192.168.123.108:6801","nonce":1817589236}]},"cluster_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.108:6802","nonce":1817589236},{"type":"v1","addr":"192.168.123.108:6803","nonce":1817589236}]},"heartbeat_back_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.108:6806","nonce":1817589236},{"type":"v1","addr":"192.168.123.108:6807","nonce":1817589236}]},"heartbeat_front_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.108:6804","nonce":1817589236},{"type":"v1","addr":"192.168.123.108:6805","nonce":1817589236}]},"public_addr":"192.168.123.108:6801/1817589236","cluster_addr":"192.168.123.108:6803/1817589236","heartbeat_back_addr":"192.168.123.108:6807/1817589236","heartbeat_front_addr":"192.168.123.108:6805/1817589236","state":["exists","up"]},{"osd":4,"uuid":"e58e5154-9b17-4109-926c-b20bd25945b6","up":1,"in":1,"weight":1,"primary_affinity":1,"last_clean_begin":0,"last_clean_end":0,"up_from":11,"up_thru":0,"down_at":0,"lost_at":0,"public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.108:6816","nonce":4097418099},{"type":"v1","addr":"192.168.123.108:6817","nonce":4097418099}]},"cluster_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.108:6818","nonce":4097418099},{"type":"v1","addr":"192.168.123.108:6819","nonce":4097418099}]},"heartbeat_back_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.108:6822","nonce":4097418099},{"type":"v1","addr":"192.168.123.108:6823","nonce":4097418099}]},"heartbeat_front_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.108:6820","nonce":4097418099},{"type":"v1","addr":"192.168.123.108:6821","nonce":4097418099}]},"public_addr":"192.168.123.108:6817/4097418099","cluster_addr":"192.168.123.108:6819/4097418099","heartbeat_back_addr":"192.168.123.108:6823/4097418099","heartbeat_front_addr":"192.168.123.108:6821/4097418099","state":["exists","up"]},{"osd":5,"uuid":"9d3c09ae-4c03-41c0-88b1-13e6e36d67bf","up":1,"in":1,"weight":1,"primary_affinity":1,"last_clean_begin":0,"last_clean_end":0,"up_from":11,"up_thru":0,"down_at":0,"lost_at":0,"public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.108:6808","nonce":1082012952},{"type":"v1","addr":"192.168.123.108:6809","nonce":1082012952}]},"cluster_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.108:6810","nonce":1082012952},{"type":"v1","addr":"192.168.123.108:6811","nonce":1082012952}]},"heartbeat_back_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.108:6814","nonce":1082012952},{"type":"v1","addr":"192.168.123.108:6815","nonce":1082012952}]},"heartbeat_front_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.108:6812","nonce":1082012952},{"type":"v1","addr":"192.168.123.108:6813","nonce":1082012952}]},"public_addr":"192.168.123.108:6809/1082012952","cluster_addr":"192.168.123.108:6811/1082012952","heartbeat_back_addr":"192.168.123.108:6815/1082012952","heartbeat_front_addr":"192.168.123.108:6813/1082012952","state":["exists","up"]}],"osd_xinfo":[{"osd":0,"down_stamp":"0.000000","laggy_probability":0,"laggy_interval":0,"features":4544132024016699391,"old_weight":0,"last_purged_snaps_scrub":"2026-03-20T17:57:40.304164+0000","dead_epoch":0},{"osd":1,"down_stamp":"0.000000","laggy_probability":0,"laggy_interval":0,"features":4544132024016699391,"old_weight":0,"last_purged_snaps_scrub":"2026-03-20T17:57:40.311385+0000","dead_epoch":0},{"osd":2,"down_stamp":"0.000000","laggy_probability":0,"laggy_interval":0,"features":4544132024016699391,"old_weight":0,"last_purged_snaps_scrub":"2026-03-20T17:57:40.407763+0000","dead_epoch":0},{"osd":3,"down_stamp":"0.000000","laggy_probability":0,"laggy_interval":0,"features":4544132024016699391,"old_weight":0,"last_purged_snaps_scrub":"2026-03-20T17:57:40.210113+0000","dead_epoch":0},{"osd":4,"down_stamp":"0.000000","laggy_probability":0,"laggy_interval":0,"features":4544132024016699391,"old_weight":0,"last_purged_snaps_scrub":"2026-03-20T17:57:40.384187+0000","dead_epoch":0},{"osd":5,"down_stamp":"0.000000","laggy_probability":0,"laggy_interval":0,"features":4544132024016699391,"old_weight":0,"last_purged_snaps_scrub":"2026-03-20T17:57:40.282793+0000","dead_epoch":0}],"pg_upmap":[],"pg_upmap_items":[],"pg_upmap_primaries":[],"pg_temp":[],"primary_temp":[],"blocklist":{},"range_blocklist":{},"erasure_code_profiles":{"default":{"crush-failure-domain":"osd","k":"2","m":"1","plugin":"isa","technique":"reed_sol_van"}},"removed_snaps_queue":[],"new_removed_snaps":[],"new_purged_snaps":[],"crush_node_flags":{},"device_class_flags":{},"stretch_mode":{"stretch_mode_enabled":false,"stretch_bucket_count":0,"degraded_stretch_mode":0,"recovering_stretch_mode":0,"stretch_mode_bucket":0}} 2026-03-20T17:57:46.316 DEBUG:teuthology.misc:6 of 6 OSDs are up 2026-03-20T17:57:46.316 INFO:tasks.ceph:Creating RBD pool 2026-03-20T17:57:46.316 DEBUG:teuthology.orchestra.run.vm06:> sudo ceph --cluster c2 osd pool create rbd 8 2026-03-20T17:57:46.888 INFO:teuthology.orchestra.run.vm06.stderr:pool 'rbd' created 2026-03-20T17:57:46.903 DEBUG:teuthology.orchestra.run.vm06:> rbd --cluster c2 pool init rbd 2026-03-20T17:57:46.918 INFO:teuthology.orchestra.run.vm06.stderr:ignoring --setuser ceph since I am not root 2026-03-20T17:57:46.918 INFO:teuthology.orchestra.run.vm06.stderr:ignoring --setgroup ceph since I am not root 2026-03-20T17:57:49.911 INFO:tasks.ceph:Starting mds daemons in cluster c2... 2026-03-20T17:57:49.912 DEBUG:teuthology.orchestra.run.vm06:> sudo adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage timeout 120 ceph --cluster c2 config log 1 --format=json 2026-03-20T17:57:49.912 INFO:tasks.daemonwatchdog.daemon_watchdog:watchdog starting 2026-03-20T17:57:50.122 INFO:teuthology.orchestra.run.vm06.stdout: 2026-03-20T17:57:50.137 INFO:teuthology.orchestra.run.vm06.stdout:[{"version":1,"timestamp":"0.000000","name":"","changes":[]}] 2026-03-20T17:57:50.137 INFO:tasks.ceph_manager:config epoch is 1 2026-03-20T17:57:50.137 INFO:tasks.ceph:Waiting until c2 daemons up and pgs clean... 2026-03-20T17:57:50.137 INFO:tasks.ceph.ceph_manager.c2:waiting for mgr available 2026-03-20T17:57:50.137 DEBUG:teuthology.orchestra.run.vm06:> sudo adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage timeout 120 ceph --cluster c2 mgr dump --format=json 2026-03-20T17:57:50.336 INFO:teuthology.orchestra.run.vm06.stdout: 2026-03-20T17:57:50.350 INFO:teuthology.orchestra.run.vm06.stdout:{"epoch":5,"flags":0,"active_gid":4103,"active_name":"x","active_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.108:6824","nonce":1386472261},{"type":"v1","addr":"192.168.123.108:6825","nonce":1386472261}]},"active_addr":"192.168.123.108:6825/1386472261","active_change":"2026-03-20T17:57:39.836914+0000","active_mgr_features":4544132024016699391,"available":true,"standbys":[],"modules":["iostat","nfs"],"available_modules":[{"name":"alerts","can_run":true,"error_string":"","module_options":{"interval":{"name":"interval","type":"secs","level":"advanced","flags":1,"default_value":"60","min":"","max":"","enum_allowed":[],"desc":"How frequently to reexamine health status","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"smtp_destination":{"name":"smtp_destination","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"Email address to send alerts to, use commas to separate multiple","long_desc":"","tags":[],"see_also":[]},"smtp_from_name":{"name":"smtp_from_name","type":"str","level":"advanced","flags":1,"default_value":"Ceph","min":"","max":"","enum_allowed":[],"desc":"Email From: name","long_desc":"","tags":[],"see_also":[]},"smtp_host":{"name":"smtp_host","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"SMTP server","long_desc":"","tags":[],"see_also":[]},"smtp_password":{"name":"smtp_password","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"Password to authenticate with","long_desc":"","tags":[],"see_also":[]},"smtp_port":{"name":"smtp_port","type":"int","level":"advanced","flags":1,"default_value":"465","min":"","max":"","enum_allowed":[],"desc":"SMTP port","long_desc":"","tags":[],"see_also":[]},"smtp_sender":{"name":"smtp_sender","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"SMTP envelope sender","long_desc":"","tags":[],"see_also":[]},"smtp_ssl":{"name":"smtp_ssl","type":"bool","level":"advanced","flags":1,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"Use SSL to connect to SMTP server","long_desc":"","tags":[],"see_also":[]},"smtp_user":{"name":"smtp_user","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"User to authenticate as","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"balancer","can_run":true,"error_string":"","module_options":{"active":{"name":"active","type":"bool","level":"advanced","flags":1,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"automatically balance PGs across cluster","long_desc":"","tags":[],"see_also":[]},"begin_time":{"name":"begin_time","type":"str","level":"advanced","flags":1,"default_value":"0000","min":"","max":"","enum_allowed":[],"desc":"beginning time of day to automatically balance","long_desc":"This is a time of day in the format HHMM.","tags":[],"see_also":[]},"begin_weekday":{"name":"begin_weekday","type":"uint","level":"advanced","flags":1,"default_value":"0","min":"0","max":"6","enum_allowed":[],"desc":"Restrict automatic balancing to this day of the week or later","long_desc":"0 = Sunday, 1 = Monday, etc.","tags":[],"see_also":[]},"crush_compat_max_iterations":{"name":"crush_compat_max_iterations","type":"uint","level":"advanced","flags":1,"default_value":"25","min":"1","max":"250","enum_allowed":[],"desc":"maximum number of iterations to attempt optimization","long_desc":"","tags":[],"see_also":[]},"crush_compat_metrics":{"name":"crush_compat_metrics","type":"str","level":"advanced","flags":1,"default_value":"pgs,objects,bytes","min":"","max":"","enum_allowed":[],"desc":"metrics with which to calculate OSD utilization","long_desc":"Value is a list of one or more of \"pgs\", \"objects\", or \"bytes\", and indicates which metrics to use to balance utilization.","tags":[],"see_also":[]},"crush_compat_step":{"name":"crush_compat_step","type":"float","level":"advanced","flags":1,"default_value":"0.5","min":"0.001","max":"0.999","enum_allowed":[],"desc":"aggressiveness of optimization","long_desc":".99 is very aggressive, .01 is less aggressive","tags":[],"see_also":[]},"end_time":{"name":"end_time","type":"str","level":"advanced","flags":1,"default_value":"2359","min":"","max":"","enum_allowed":[],"desc":"ending time of day to automatically balance","long_desc":"This is a time of day in the format HHMM.","tags":[],"see_also":[]},"end_weekday":{"name":"end_weekday","type":"uint","level":"advanced","flags":1,"default_value":"0","min":"0","max":"6","enum_allowed":[],"desc":"Restrict automatic balancing to days of the week earlier than this","long_desc":"0 = Sunday, 1 = Monday, etc.","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"min_score":{"name":"min_score","type":"float","level":"advanced","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"minimum score, below which no optimization is attempted","long_desc":"","tags":[],"see_also":[]},"mode":{"name":"mode","type":"str","level":"advanced","flags":1,"default_value":"upmap","min":"","max":"","enum_allowed":["crush-compat","none","read","upmap","upmap-read"],"desc":"Balancer mode","long_desc":"","tags":[],"see_also":[]},"pool_ids":{"name":"pool_ids","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"pools which the automatic balancing will be limited to","long_desc":"","tags":[],"see_also":[]},"sleep_interval":{"name":"sleep_interval","type":"secs","level":"advanced","flags":1,"default_value":"60","min":"","max":"","enum_allowed":[],"desc":"how frequently to wake up and attempt optimization","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"update_pg_upmap_activity":{"name":"update_pg_upmap_activity","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Updates pg_upmap activity stats to be used in `balancer status detail`","long_desc":"","tags":[],"see_also":[]},"upmap_max_deviation":{"name":"upmap_max_deviation","type":"int","level":"advanced","flags":1,"default_value":"5","min":"1","max":"","enum_allowed":[],"desc":"deviation below which no optimization is attempted","long_desc":"If the number of PGs are within this count then no optimization is attempted","tags":[],"see_also":[]},"upmap_max_optimizations":{"name":"upmap_max_optimizations","type":"uint","level":"advanced","flags":1,"default_value":"10","min":"","max":"","enum_allowed":[],"desc":"maximum upmap optimizations to make per attempt","long_desc":"","tags":[],"see_also":[]}}},{"name":"cephadm","can_run":true,"error_string":"","module_options":{"agent_down_multiplier":{"name":"agent_down_multiplier","type":"float","level":"advanced","flags":0,"default_value":"3.0","min":"","max":"","enum_allowed":[],"desc":"Multiplied by agent refresh rate to calculate how long agent must not report before being marked down","long_desc":"","tags":[],"see_also":[]},"agent_refresh_rate":{"name":"agent_refresh_rate","type":"secs","level":"advanced","flags":0,"default_value":"20","min":"","max":"","enum_allowed":[],"desc":"How often agent on each host will try to gather and send metadata","long_desc":"","tags":[],"see_also":[]},"agent_starting_port":{"name":"agent_starting_port","type":"int","level":"advanced","flags":0,"default_value":"4721","min":"","max":"","enum_allowed":[],"desc":"First port agent will try to bind to (will also try up to next 1000 subsequent ports if blocked)","long_desc":"","tags":[],"see_also":[]},"allow_ptrace":{"name":"allow_ptrace","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"allow SYS_PTRACE capability on ceph containers","long_desc":"The SYS_PTRACE capability is needed to attach to a process with gdb or strace. Enabling this options can allow debugging daemons that encounter problems at runtime.","tags":[],"see_also":[]},"autotune_interval":{"name":"autotune_interval","type":"secs","level":"advanced","flags":0,"default_value":"600","min":"","max":"","enum_allowed":[],"desc":"how frequently to autotune daemon memory","long_desc":"","tags":[],"see_also":[]},"autotune_memory_target_ratio":{"name":"autotune_memory_target_ratio","type":"float","level":"advanced","flags":0,"default_value":"0.7","min":"","max":"","enum_allowed":[],"desc":"ratio of total system memory to divide amongst autotuned daemons","long_desc":"","tags":[],"see_also":[]},"cephadm_log_destination":{"name":"cephadm_log_destination","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":["file","file,syslog","syslog"],"desc":"Destination for cephadm command's persistent logging","long_desc":"","tags":[],"see_also":[]},"certificate_automated_rotation_enabled":{"name":"certificate_automated_rotation_enabled","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"This flag controls whether cephadm automatically rotates certificates upon expiration.","long_desc":"","tags":[],"see_also":[]},"certificate_check_debug_mode":{"name":"certificate_check_debug_mode","type":"bool","level":"dev","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"FOR TESTING ONLY: This flag forces the certificate check instead of waiting for certificate_check_period.","long_desc":"","tags":[],"see_also":[]},"certificate_check_period":{"name":"certificate_check_period","type":"int","level":"advanced","flags":0,"default_value":"1","min":"0","max":"30","enum_allowed":[],"desc":"Specifies how often (in days) the certificate should be checked for validity.","long_desc":"","tags":[],"see_also":[]},"certificate_duration_days":{"name":"certificate_duration_days","type":"int","level":"advanced","flags":0,"default_value":"1095","min":"90","max":"3650","enum_allowed":[],"desc":"Specifies the duration of self certificates generated and signed by cephadm root CA","long_desc":"","tags":[],"see_also":[]},"certificate_renewal_threshold_days":{"name":"certificate_renewal_threshold_days","type":"int","level":"advanced","flags":0,"default_value":"30","min":"10","max":"90","enum_allowed":[],"desc":"Specifies the lead time in days to initiate certificate renewal before expiration.","long_desc":"","tags":[],"see_also":[]},"cgroups_split":{"name":"cgroups_split","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"Pass --cgroups=split when cephadm creates containers (currently podman only)","long_desc":"","tags":[],"see_also":[]},"config_checks_enabled":{"name":"config_checks_enabled","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Enable or disable the cephadm configuration analysis","long_desc":"","tags":[],"see_also":[]},"config_dashboard":{"name":"config_dashboard","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"manage configs like API endpoints in Dashboard.","long_desc":"","tags":[],"see_also":[]},"container_image_alertmanager":{"name":"container_image_alertmanager","type":"str","level":"advanced","flags":0,"default_value":"quay.io/prometheus/alertmanager:v0.28.1","min":"","max":"","enum_allowed":[],"desc":"Alertmanager container image","long_desc":"","tags":[],"see_also":[]},"container_image_base":{"name":"container_image_base","type":"str","level":"advanced","flags":1,"default_value":"quay.io/ceph/ceph","min":"","max":"","enum_allowed":[],"desc":"Container image name, without the tag","long_desc":"","tags":[],"see_also":[]},"container_image_elasticsearch":{"name":"container_image_elasticsearch","type":"str","level":"advanced","flags":0,"default_value":"quay.io/omrizeneva/elasticsearch:6.8.23","min":"","max":"","enum_allowed":[],"desc":"Elasticsearch container image","long_desc":"","tags":[],"see_also":[]},"container_image_grafana":{"name":"container_image_grafana","type":"str","level":"advanced","flags":0,"default_value":"quay.io/ceph/grafana:12.3.1","min":"","max":"","enum_allowed":[],"desc":"Grafana container image","long_desc":"","tags":[],"see_also":[]},"container_image_haproxy":{"name":"container_image_haproxy","type":"str","level":"advanced","flags":0,"default_value":"quay.io/ceph/haproxy:2.3","min":"","max":"","enum_allowed":[],"desc":"Haproxy container image","long_desc":"","tags":[],"see_also":[]},"container_image_jaeger_agent":{"name":"container_image_jaeger_agent","type":"str","level":"advanced","flags":0,"default_value":"quay.io/jaegertracing/jaeger-agent:1.29","min":"","max":"","enum_allowed":[],"desc":"Jaeger agent container image","long_desc":"","tags":[],"see_also":[]},"container_image_jaeger_collector":{"name":"container_image_jaeger_collector","type":"str","level":"advanced","flags":0,"default_value":"quay.io/jaegertracing/jaeger-collector:1.29","min":"","max":"","enum_allowed":[],"desc":"Jaeger collector container image","long_desc":"","tags":[],"see_also":[]},"container_image_jaeger_query":{"name":"container_image_jaeger_query","type":"str","level":"advanced","flags":0,"default_value":"quay.io/jaegertracing/jaeger-query:1.29","min":"","max":"","enum_allowed":[],"desc":"Jaeger query container image","long_desc":"","tags":[],"see_also":[]},"container_image_keepalived":{"name":"container_image_keepalived","type":"str","level":"advanced","flags":0,"default_value":"quay.io/ceph/keepalived:2.2.4","min":"","max":"","enum_allowed":[],"desc":"Keepalived container image","long_desc":"","tags":[],"see_also":[]},"container_image_loki":{"name":"container_image_loki","type":"str","level":"advanced","flags":0,"default_value":"docker.io/grafana/loki:3.0.0","min":"","max":"","enum_allowed":[],"desc":"Loki container image","long_desc":"","tags":[],"see_also":[]},"container_image_nginx":{"name":"container_image_nginx","type":"str","level":"advanced","flags":0,"default_value":"quay.io/ceph/nginx:sclorg-nginx-126","min":"","max":"","enum_allowed":[],"desc":"Nginx container image","long_desc":"","tags":[],"see_also":[]},"container_image_node_exporter":{"name":"container_image_node_exporter","type":"str","level":"advanced","flags":0,"default_value":"quay.io/prometheus/node-exporter:v1.9.1","min":"","max":"","enum_allowed":[],"desc":"Node exporter container image","long_desc":"","tags":[],"see_also":[]},"container_image_nvmeof":{"name":"container_image_nvmeof","type":"str","level":"advanced","flags":0,"default_value":"quay.io/ceph/nvmeof:1.5","min":"","max":"","enum_allowed":[],"desc":"Nvmeof container image","long_desc":"","tags":[],"see_also":[]},"container_image_oauth2_proxy":{"name":"container_image_oauth2_proxy","type":"str","level":"advanced","flags":0,"default_value":"quay.io/oauth2-proxy/oauth2-proxy:v7.6.0","min":"","max":"","enum_allowed":[],"desc":"Oauth2 proxy container image","long_desc":"","tags":[],"see_also":[]},"container_image_prometheus":{"name":"container_image_prometheus","type":"str","level":"advanced","flags":0,"default_value":"quay.io/prometheus/prometheus:v3.6.0","min":"","max":"","enum_allowed":[],"desc":"Prometheus container image","long_desc":"","tags":[],"see_also":[]},"container_image_promtail":{"name":"container_image_promtail","type":"str","level":"advanced","flags":0,"default_value":"docker.io/grafana/promtail:3.0.0","min":"","max":"","enum_allowed":[],"desc":"Promtail container image","long_desc":"","tags":[],"see_also":[]},"container_image_samba":{"name":"container_image_samba","type":"str","level":"advanced","flags":0,"default_value":"quay.io/samba.org/samba-server:ceph20-centos-amd64","min":"","max":"","enum_allowed":[],"desc":"Samba container image","long_desc":"","tags":[],"see_also":[]},"container_image_samba_metrics":{"name":"container_image_samba_metrics","type":"str","level":"advanced","flags":0,"default_value":"quay.io/samba.org/samba-metrics:ceph20-centos-amd64","min":"","max":"","enum_allowed":[],"desc":"Samba metrics container image","long_desc":"","tags":[],"see_also":[]},"container_image_snmp_gateway":{"name":"container_image_snmp_gateway","type":"str","level":"advanced","flags":0,"default_value":"docker.io/maxwo/snmp-notifier:v1.2.1","min":"","max":"","enum_allowed":[],"desc":"Snmp gateway container image","long_desc":"","tags":[],"see_also":[]},"container_init":{"name":"container_init","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"Run podman/docker with `--init`","long_desc":"","tags":[],"see_also":[]},"daemon_cache_timeout":{"name":"daemon_cache_timeout","type":"secs","level":"advanced","flags":0,"default_value":"600","min":"","max":"","enum_allowed":[],"desc":"seconds to cache service (daemon) inventory","long_desc":"","tags":[],"see_also":[]},"default_cephadm_command_timeout":{"name":"default_cephadm_command_timeout","type":"int","level":"advanced","flags":0,"default_value":"900","min":"","max":"","enum_allowed":[],"desc":"Default timeout applied to cephadm commands run directly on the host (in seconds)","long_desc":"","tags":[],"see_also":[]},"default_registry":{"name":"default_registry","type":"str","level":"advanced","flags":0,"default_value":"quay.io","min":"","max":"","enum_allowed":[],"desc":"Search-registry to which we should normalize unqualified image names. This is not the default registry","long_desc":"","tags":[],"see_also":[]},"device_cache_timeout":{"name":"device_cache_timeout","type":"secs","level":"advanced","flags":0,"default_value":"1800","min":"","max":"","enum_allowed":[],"desc":"seconds to cache device inventory","long_desc":"","tags":[],"see_also":[]},"device_enhanced_scan":{"name":"device_enhanced_scan","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Use libstoragemgmt during device scans","long_desc":"","tags":[],"see_also":[]},"facts_cache_timeout":{"name":"facts_cache_timeout","type":"secs","level":"advanced","flags":0,"default_value":"60","min":"","max":"","enum_allowed":[],"desc":"seconds to cache host facts data","long_desc":"","tags":[],"see_also":[]},"grafana_dashboards_path":{"name":"grafana_dashboards_path","type":"str","level":"advanced","flags":0,"default_value":"/etc/grafana/dashboards/ceph-dashboard/","min":"","max":"","enum_allowed":[],"desc":"location of dashboards to include in grafana deployments","long_desc":"","tags":[],"see_also":[]},"host_check_interval":{"name":"host_check_interval","type":"secs","level":"advanced","flags":0,"default_value":"600","min":"","max":"","enum_allowed":[],"desc":"how frequently to perform a host check","long_desc":"","tags":[],"see_also":[]},"hw_monitoring":{"name":"hw_monitoring","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Deploy hw monitoring daemon on every host.","long_desc":"","tags":[],"see_also":[]},"inventory_list_all":{"name":"inventory_list_all","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Whether ceph-volume inventory should report more devices (mostly mappers (LVs / mpaths), partitions...)","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_refresh_metadata":{"name":"log_refresh_metadata","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Log all refresh metadata. Includes daemon, device, and host info collected regularly. Only has effect if logging at debug level","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"log to the \"cephadm\" cluster log channel\"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"manage_etc_ceph_ceph_conf":{"name":"manage_etc_ceph_ceph_conf","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Manage and own /etc/ceph/ceph.conf on the hosts.","long_desc":"","tags":[],"see_also":[]},"manage_etc_ceph_ceph_conf_hosts":{"name":"manage_etc_ceph_ceph_conf_hosts","type":"str","level":"advanced","flags":0,"default_value":"*","min":"","max":"","enum_allowed":[],"desc":"PlacementSpec describing on which hosts to manage /etc/ceph/ceph.conf","long_desc":"","tags":[],"see_also":[]},"max_count_per_host":{"name":"max_count_per_host","type":"int","level":"advanced","flags":0,"default_value":"10","min":"","max":"","enum_allowed":[],"desc":"max number of daemons per service per host","long_desc":"","tags":[],"see_also":[]},"max_osd_draining_count":{"name":"max_osd_draining_count","type":"int","level":"advanced","flags":0,"default_value":"10","min":"","max":"","enum_allowed":[],"desc":"max number of osds that will be drained simultaneously when osds are removed","long_desc":"","tags":[],"see_also":[]},"migration_current":{"name":"migration_current","type":"int","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"internal - do not modify","long_desc":"","tags":[],"see_also":[]},"mode":{"name":"mode","type":"str","level":"advanced","flags":0,"default_value":"root","min":"","max":"","enum_allowed":["cephadm-package","root"],"desc":"mode for remote execution of cephadm","long_desc":"","tags":[],"see_also":[]},"oob_default_addr":{"name":"oob_default_addr","type":"str","level":"advanced","flags":0,"default_value":"169.254.1.1","min":"","max":"","enum_allowed":[],"desc":"Default address for RedFish API (oob management).","long_desc":"","tags":[],"see_also":[]},"prometheus_alerts_path":{"name":"prometheus_alerts_path","type":"str","level":"advanced","flags":0,"default_value":"/etc/prometheus/ceph/ceph_default_alerts.yml","min":"","max":"","enum_allowed":[],"desc":"location of alerts to include in prometheus deployments","long_desc":"","tags":[],"see_also":[]},"registry_insecure":{"name":"registry_insecure","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Registry is to be considered insecure (no TLS available). Only for development purposes.","long_desc":"","tags":[],"see_also":[]},"registry_password":{"name":"registry_password","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"Custom repository password. Only used for logging into a registry.","long_desc":"","tags":[],"see_also":[]},"registry_url":{"name":"registry_url","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"Registry url for login purposes. This is not the default registry","long_desc":"","tags":[],"see_also":[]},"registry_username":{"name":"registry_username","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"Custom repository username. Only used for logging into a registry.","long_desc":"","tags":[],"see_also":[]},"secure_monitoring_stack":{"name":"secure_monitoring_stack","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Enable TLS security for all the monitoring stack daemons","long_desc":"","tags":[],"see_also":[]},"service_discovery_port":{"name":"service_discovery_port","type":"int","level":"advanced","flags":0,"default_value":"8765","min":"","max":"","enum_allowed":[],"desc":"cephadm service discovery port","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"ssh_config_file":{"name":"ssh_config_file","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"customized SSH config file to connect to managed hosts","long_desc":"","tags":[],"see_also":[]},"ssh_keepalive_count_max":{"name":"ssh_keepalive_count_max","type":"int","level":"advanced","flags":0,"default_value":"3","min":"","max":"","enum_allowed":[],"desc":"How many times ssh connections can fail liveness checks before the host is marked offline","long_desc":"","tags":[],"see_also":[]},"ssh_keepalive_interval":{"name":"ssh_keepalive_interval","type":"int","level":"advanced","flags":0,"default_value":"7","min":"","max":"","enum_allowed":[],"desc":"How often ssh connections are checked for liveness","long_desc":"","tags":[],"see_also":[]},"stray_daemon_check_interval":{"name":"stray_daemon_check_interval","type":"secs","level":"advanced","flags":0,"default_value":"1800","min":"","max":"","enum_allowed":[],"desc":"how frequently cephadm should check for the presence of stray daemons","long_desc":"","tags":[],"see_also":[]},"use_agent":{"name":"use_agent","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Use cephadm agent on each host to gather and send metadata","long_desc":"","tags":[],"see_also":[]},"use_repo_digest":{"name":"use_repo_digest","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"Automatically convert image tags to image digest. Make sure all daemons use the same image","long_desc":"","tags":[],"see_also":[]},"warn_on_failed_host_check":{"name":"warn_on_failed_host_check","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"raise a health warning if the host check fails","long_desc":"","tags":[],"see_also":[]},"warn_on_stray_daemons":{"name":"warn_on_stray_daemons","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"raise a health warning if daemons are detected that are not managed by cephadm","long_desc":"","tags":[],"see_also":[]},"warn_on_stray_hosts":{"name":"warn_on_stray_hosts","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"raise a health warning if daemons are detected on a host that is not managed by cephadm","long_desc":"","tags":[],"see_also":[]}}},{"name":"crash","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"retain_interval":{"name":"retain_interval","type":"secs","level":"advanced","flags":1,"default_value":"31536000","min":"","max":"","enum_allowed":[],"desc":"how long to retain crashes before pruning them","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"warn_recent_interval":{"name":"warn_recent_interval","type":"secs","level":"advanced","flags":1,"default_value":"1209600","min":"","max":"","enum_allowed":[],"desc":"time interval in which to warn about recent crashes","long_desc":"","tags":[],"see_also":[]}}},{"name":"dashboard","can_run":true,"error_string":"","module_options":{"ACCOUNT_LOCKOUT_ATTEMPTS":{"name":"ACCOUNT_LOCKOUT_ATTEMPTS","type":"int","level":"advanced","flags":0,"default_value":"10","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"ALERTMANAGER_API_HOST":{"name":"ALERTMANAGER_API_HOST","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"ALERTMANAGER_API_SSL_VERIFY":{"name":"ALERTMANAGER_API_SSL_VERIFY","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"AUDIT_API_ENABLED":{"name":"AUDIT_API_ENABLED","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"AUDIT_API_LOG_PAYLOAD":{"name":"AUDIT_API_LOG_PAYLOAD","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"ENABLE_BROWSABLE_API":{"name":"ENABLE_BROWSABLE_API","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"FEATURE_TOGGLE_CEPHFS":{"name":"FEATURE_TOGGLE_CEPHFS","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"FEATURE_TOGGLE_DASHBOARD":{"name":"FEATURE_TOGGLE_DASHBOARD","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"FEATURE_TOGGLE_ISCSI":{"name":"FEATURE_TOGGLE_ISCSI","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"FEATURE_TOGGLE_MIRRORING":{"name":"FEATURE_TOGGLE_MIRRORING","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"FEATURE_TOGGLE_NFS":{"name":"FEATURE_TOGGLE_NFS","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"FEATURE_TOGGLE_RBD":{"name":"FEATURE_TOGGLE_RBD","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"FEATURE_TOGGLE_RGW":{"name":"FEATURE_TOGGLE_RGW","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"GANESHA_CLUSTERS_RADOS_POOL_NAMESPACE":{"name":"GANESHA_CLUSTERS_RADOS_POOL_NAMESPACE","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"GRAFANA_API_PASSWORD":{"name":"GRAFANA_API_PASSWORD","type":"str","level":"advanced","flags":0,"default_value":"admin","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"GRAFANA_API_SSL_VERIFY":{"name":"GRAFANA_API_SSL_VERIFY","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"GRAFANA_API_URL":{"name":"GRAFANA_API_URL","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"GRAFANA_API_USERNAME":{"name":"GRAFANA_API_USERNAME","type":"str","level":"advanced","flags":0,"default_value":"admin","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"GRAFANA_FRONTEND_API_URL":{"name":"GRAFANA_FRONTEND_API_URL","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"GRAFANA_UPDATE_DASHBOARDS":{"name":"GRAFANA_UPDATE_DASHBOARDS","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"ISCSI_API_SSL_VERIFICATION":{"name":"ISCSI_API_SSL_VERIFICATION","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"ISSUE_TRACKER_API_KEY":{"name":"ISSUE_TRACKER_API_KEY","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"MANAGED_BY_CLUSTERS":{"name":"MANAGED_BY_CLUSTERS","type":"str","level":"advanced","flags":0,"default_value":"[]","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"MULTICLUSTER_CONFIG":{"name":"MULTICLUSTER_CONFIG","type":"str","level":"advanced","flags":0,"default_value":"{}","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PROMETHEUS_API_HOST":{"name":"PROMETHEUS_API_HOST","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PROMETHEUS_API_SSL_VERIFY":{"name":"PROMETHEUS_API_SSL_VERIFY","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PROM_ALERT_CREDENTIAL_CACHE_TTL":{"name":"PROM_ALERT_CREDENTIAL_CACHE_TTL","type":"int","level":"advanced","flags":0,"default_value":"60","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PWD_POLICY_CHECK_COMPLEXITY_ENABLED":{"name":"PWD_POLICY_CHECK_COMPLEXITY_ENABLED","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PWD_POLICY_CHECK_EXCLUSION_LIST_ENABLED":{"name":"PWD_POLICY_CHECK_EXCLUSION_LIST_ENABLED","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PWD_POLICY_CHECK_LENGTH_ENABLED":{"name":"PWD_POLICY_CHECK_LENGTH_ENABLED","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PWD_POLICY_CHECK_OLDPWD_ENABLED":{"name":"PWD_POLICY_CHECK_OLDPWD_ENABLED","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PWD_POLICY_CHECK_REPETITIVE_CHARS_ENABLED":{"name":"PWD_POLICY_CHECK_REPETITIVE_CHARS_ENABLED","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PWD_POLICY_CHECK_SEQUENTIAL_CHARS_ENABLED":{"name":"PWD_POLICY_CHECK_SEQUENTIAL_CHARS_ENABLED","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PWD_POLICY_CHECK_USERNAME_ENABLED":{"name":"PWD_POLICY_CHECK_USERNAME_ENABLED","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PWD_POLICY_ENABLED":{"name":"PWD_POLICY_ENABLED","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PWD_POLICY_EXCLUSION_LIST":{"name":"PWD_POLICY_EXCLUSION_LIST","type":"str","level":"advanced","flags":0,"default_value":"osd,host,dashboard,pool,block,nfs,ceph,monitors,gateway,logs,crush,maps","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PWD_POLICY_MIN_COMPLEXITY":{"name":"PWD_POLICY_MIN_COMPLEXITY","type":"int","level":"advanced","flags":0,"default_value":"10","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PWD_POLICY_MIN_LENGTH":{"name":"PWD_POLICY_MIN_LENGTH","type":"int","level":"advanced","flags":0,"default_value":"8","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"REST_REQUESTS_TIMEOUT":{"name":"REST_REQUESTS_TIMEOUT","type":"int","level":"advanced","flags":0,"default_value":"45","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"RGW_API_ACCESS_KEY":{"name":"RGW_API_ACCESS_KEY","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"RGW_API_ADMIN_RESOURCE":{"name":"RGW_API_ADMIN_RESOURCE","type":"str","level":"advanced","flags":0,"default_value":"admin","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"RGW_API_SECRET_KEY":{"name":"RGW_API_SECRET_KEY","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"RGW_API_SSL_VERIFY":{"name":"RGW_API_SSL_VERIFY","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"RGW_HOSTNAME_PER_DAEMON":{"name":"RGW_HOSTNAME_PER_DAEMON","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"UNSAFE_TLS_v1_2":{"name":"UNSAFE_TLS_v1_2","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"USER_PWD_EXPIRATION_SPAN":{"name":"USER_PWD_EXPIRATION_SPAN","type":"int","level":"advanced","flags":0,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"USER_PWD_EXPIRATION_WARNING_1":{"name":"USER_PWD_EXPIRATION_WARNING_1","type":"int","level":"advanced","flags":0,"default_value":"10","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"USER_PWD_EXPIRATION_WARNING_2":{"name":"USER_PWD_EXPIRATION_WARNING_2","type":"int","level":"advanced","flags":0,"default_value":"5","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"cross_origin_url":{"name":"cross_origin_url","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"crt_file":{"name":"crt_file","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"crypto_caller":{"name":"crypto_caller","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"debug":{"name":"debug","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Enable/disable debug options","long_desc":"","tags":[],"see_also":[]},"jwt_token_ttl":{"name":"jwt_token_ttl","type":"int","level":"advanced","flags":0,"default_value":"28800","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"key_file":{"name":"key_file","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"motd":{"name":"motd","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"The message of the day","long_desc":"","tags":[],"see_also":[]},"redirect_resolve_ip_addr":{"name":"redirect_resolve_ip_addr","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"server_addr":{"name":"server_addr","type":"str","level":"advanced","flags":0,"default_value":"::","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"server_port":{"name":"server_port","type":"int","level":"advanced","flags":0,"default_value":"8080","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"ssl":{"name":"ssl","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"ssl_server_port":{"name":"ssl_server_port","type":"int","level":"advanced","flags":0,"default_value":"8443","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sso_oauth2":{"name":"sso_oauth2","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"standby_behaviour":{"name":"standby_behaviour","type":"str","level":"advanced","flags":0,"default_value":"redirect","min":"","max":"","enum_allowed":["error","redirect"],"desc":"","long_desc":"","tags":[],"see_also":[]},"standby_error_status_code":{"name":"standby_error_status_code","type":"int","level":"advanced","flags":0,"default_value":"500","min":"400","max":"599","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"url_prefix":{"name":"url_prefix","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"devicehealth","can_run":true,"error_string":"","module_options":{"enable_monitoring":{"name":"enable_monitoring","type":"bool","level":"advanced","flags":1,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"monitor device health metrics","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"mark_out_threshold":{"name":"mark_out_threshold","type":"secs","level":"advanced","flags":1,"default_value":"2419200","min":"","max":"","enum_allowed":[],"desc":"automatically mark OSD if it may fail before this long","long_desc":"","tags":[],"see_also":[]},"pool_name":{"name":"pool_name","type":"str","level":"advanced","flags":1,"default_value":"device_health_metrics","min":"","max":"","enum_allowed":[],"desc":"name of pool in which to store device health metrics","long_desc":"","tags":[],"see_also":[]},"retention_period":{"name":"retention_period","type":"secs","level":"advanced","flags":1,"default_value":"15552000","min":"","max":"","enum_allowed":[],"desc":"how long to retain device health metrics","long_desc":"","tags":[],"see_also":[]},"scrape_frequency":{"name":"scrape_frequency","type":"secs","level":"advanced","flags":1,"default_value":"86400","min":"","max":"","enum_allowed":[],"desc":"how frequently to scrape device health metrics","long_desc":"","tags":[],"see_also":[]},"self_heal":{"name":"self_heal","type":"bool","level":"advanced","flags":1,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"preemptively heal cluster around devices that may fail","long_desc":"","tags":[],"see_also":[]},"sleep_interval":{"name":"sleep_interval","type":"secs","level":"advanced","flags":1,"default_value":"600","min":"","max":"","enum_allowed":[],"desc":"how frequently to wake up and check device health","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"warn_threshold":{"name":"warn_threshold","type":"secs","level":"advanced","flags":1,"default_value":"7257600","min":"","max":"","enum_allowed":[],"desc":"raise health warning if OSD may fail before this long","long_desc":"","tags":[],"see_also":[]}}},{"name":"diskprediction_local","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"predict_interval":{"name":"predict_interval","type":"str","level":"advanced","flags":0,"default_value":"86400","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"predictor_model":{"name":"predictor_model","type":"str","level":"advanced","flags":0,"default_value":"prophetstor","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sleep_interval":{"name":"sleep_interval","type":"str","level":"advanced","flags":0,"default_value":"600","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"influx","can_run":false,"error_string":"influxdb python module not found","module_options":{"batch_size":{"name":"batch_size","type":"int","level":"advanced","flags":0,"default_value":"5000","min":"","max":"","enum_allowed":[],"desc":"How big batches of data points should be when sending to InfluxDB.","long_desc":"","tags":[],"see_also":[]},"database":{"name":"database","type":"str","level":"advanced","flags":0,"default_value":"ceph","min":"","max":"","enum_allowed":[],"desc":"InfluxDB database name. You will need to create this database and grant write privileges to the configured username or the username must have admin privileges to create it.","long_desc":"","tags":[],"see_also":[]},"hostname":{"name":"hostname","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"InfluxDB server hostname","long_desc":"","tags":[],"see_also":[]},"interval":{"name":"interval","type":"secs","level":"advanced","flags":0,"default_value":"30","min":"5","max":"","enum_allowed":[],"desc":"Time between reports to InfluxDB. Default 30 seconds.","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"password":{"name":"password","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"password of InfluxDB server user","long_desc":"","tags":[],"see_also":[]},"port":{"name":"port","type":"int","level":"advanced","flags":0,"default_value":"8086","min":"","max":"","enum_allowed":[],"desc":"InfluxDB server port","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"ssl":{"name":"ssl","type":"str","level":"advanced","flags":0,"default_value":"false","min":"","max":"","enum_allowed":[],"desc":"Use https connection for InfluxDB server. Use \"true\" or \"false\".","long_desc":"","tags":[],"see_also":[]},"threads":{"name":"threads","type":"int","level":"advanced","flags":0,"default_value":"5","min":"1","max":"32","enum_allowed":[],"desc":"How many worker threads should be spawned for sending data to InfluxDB.","long_desc":"","tags":[],"see_also":[]},"username":{"name":"username","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"username of InfluxDB server user","long_desc":"","tags":[],"see_also":[]},"verify_ssl":{"name":"verify_ssl","type":"str","level":"advanced","flags":0,"default_value":"true","min":"","max":"","enum_allowed":[],"desc":"Verify https cert for InfluxDB server. Use \"true\" or \"false\".","long_desc":"","tags":[],"see_also":[]}}},{"name":"insights","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"iostat","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"localpool","can_run":true,"error_string":"","module_options":{"failure_domain":{"name":"failure_domain","type":"str","level":"advanced","flags":1,"default_value":"host","min":"","max":"","enum_allowed":[],"desc":"failure domain for any created local pool","long_desc":"what failure domain we should separate data replicas across.","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"min_size":{"name":"min_size","type":"int","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"default min_size for any created local pool","long_desc":"value to set min_size to (unchanged from Ceph's default if this option is not set)","tags":[],"see_also":[]},"num_rep":{"name":"num_rep","type":"int","level":"advanced","flags":1,"default_value":"3","min":"","max":"","enum_allowed":[],"desc":"default replica count for any created local pool","long_desc":"","tags":[],"see_also":[]},"pg_num":{"name":"pg_num","type":"int","level":"advanced","flags":1,"default_value":"128","min":"","max":"","enum_allowed":[],"desc":"default pg_num for any created local pool","long_desc":"","tags":[],"see_also":[]},"prefix":{"name":"prefix","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"name prefix for any created local pool","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"subtree":{"name":"subtree","type":"str","level":"advanced","flags":1,"default_value":"rack","min":"","max":"","enum_allowed":[],"desc":"CRUSH level for which to create a local pool","long_desc":"which CRUSH subtree type the module should create a pool for.","tags":[],"see_also":[]}}},{"name":"mirroring","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"nfs","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"orchestrator","can_run":true,"error_string":"","module_options":{"fail_fs":{"name":"fail_fs","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Fail filesystem for rapid multi-rank mds upgrade","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"orchestrator":{"name":"orchestrator","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["cephadm","rook","test_orchestrator"],"desc":"Orchestrator backend","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"osd_perf_query","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"osd_support","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"pg_autoscaler","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sleep_interval":{"name":"sleep_interval","type":"secs","level":"advanced","flags":0,"default_value":"60","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"threshold":{"name":"threshold","type":"float","level":"advanced","flags":0,"default_value":"3.0","min":"1.0","max":"","enum_allowed":[],"desc":"scaling threshold","long_desc":"The factor by which the `NEW PG_NUM` must vary from the current`PG_NUM` before being accepted. Cannot be less than 1.0","tags":[],"see_also":[]}}},{"name":"progress","can_run":true,"error_string":"","module_options":{"allow_pg_recovery_event":{"name":"allow_pg_recovery_event","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"allow the module to show pg recovery progress","long_desc":"","tags":[],"see_also":[]},"enabled":{"name":"enabled","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"max_completed_events":{"name":"max_completed_events","type":"int","level":"advanced","flags":1,"default_value":"50","min":"","max":"","enum_allowed":[],"desc":"number of past completed events to remember","long_desc":"","tags":[],"see_also":[]},"sleep_interval":{"name":"sleep_interval","type":"secs","level":"advanced","flags":1,"default_value":"5","min":"","max":"","enum_allowed":[],"desc":"how long the module is going to sleep","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"prometheus","can_run":true,"error_string":"","module_options":{"cache":{"name":"cache","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"exclude_perf_counters":{"name":"exclude_perf_counters","type":"bool","level":"advanced","flags":1,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"Do not include perf-counters in the metrics output","long_desc":"Gathering perf-counters from a single Prometheus exporter can degrade ceph-mgr performance, especially in large clusters. Instead, Ceph-exporter daemons are now used by default for perf-counter gathering. This should only be disabled when no ceph-exporters are deployed.","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"rbd_stats_pools":{"name":"rbd_stats_pools","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"rbd_stats_pools_refresh_interval":{"name":"rbd_stats_pools_refresh_interval","type":"int","level":"advanced","flags":0,"default_value":"300","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"scrape_interval":{"name":"scrape_interval","type":"float","level":"advanced","flags":0,"default_value":"15.0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"server_addr":{"name":"server_addr","type":"str","level":"advanced","flags":0,"default_value":"::","min":"","max":"","enum_allowed":[],"desc":"the IPv4 or IPv6 address on which the module listens for HTTP requests","long_desc":"","tags":[],"see_also":[]},"server_port":{"name":"server_port","type":"int","level":"advanced","flags":1,"default_value":"9283","min":"","max":"","enum_allowed":[],"desc":"the port on which the module listens for HTTP requests","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"stale_cache_strategy":{"name":"stale_cache_strategy","type":"str","level":"advanced","flags":0,"default_value":"log","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"standby_behaviour":{"name":"standby_behaviour","type":"str","level":"advanced","flags":1,"default_value":"default","min":"","max":"","enum_allowed":["default","error"],"desc":"","long_desc":"","tags":[],"see_also":[]},"standby_error_status_code":{"name":"standby_error_status_code","type":"int","level":"advanced","flags":1,"default_value":"500","min":"400","max":"599","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"rbd_support","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"max_concurrent_snap_create":{"name":"max_concurrent_snap_create","type":"int","level":"advanced","flags":0,"default_value":"10","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"mirror_snapshot_schedule":{"name":"mirror_snapshot_schedule","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"trash_purge_schedule":{"name":"trash_purge_schedule","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"rgw","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"secondary_zone_period_retry_limit":{"name":"secondary_zone_period_retry_limit","type":"int","level":"advanced","flags":0,"default_value":"5","min":"","max":"","enum_allowed":[],"desc":"RGW module period update retry limit for secondary site","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"selftest","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"roption1":{"name":"roption1","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"roption2":{"name":"roption2","type":"str","level":"advanced","flags":0,"default_value":"xyz","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"rwoption1":{"name":"rwoption1","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"rwoption2":{"name":"rwoption2","type":"int","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"rwoption3":{"name":"rwoption3","type":"float","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"rwoption4":{"name":"rwoption4","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"rwoption5":{"name":"rwoption5","type":"bool","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"rwoption6":{"name":"rwoption6","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"rwoption7":{"name":"rwoption7","type":"int","level":"advanced","flags":0,"default_value":"","min":"1","max":"42","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"testkey":{"name":"testkey","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"testlkey":{"name":"testlkey","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"testnewline":{"name":"testnewline","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"snap_schedule","can_run":true,"error_string":"","module_options":{"allow_m_granularity":{"name":"allow_m_granularity","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"allow minute scheduled snapshots","long_desc":"","tags":[],"see_also":[]},"dump_on_update":{"name":"dump_on_update","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"dump database to debug log on update","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"stats","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"status","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"telegraf","can_run":true,"error_string":"","module_options":{"address":{"name":"address","type":"str","level":"advanced","flags":0,"default_value":"unixgram:///tmp/telegraf.sock","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"interval":{"name":"interval","type":"secs","level":"advanced","flags":0,"default_value":"15","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"telemetry","can_run":true,"error_string":"","module_options":{"channel_basic":{"name":"channel_basic","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"Share basic cluster information (size, version)","long_desc":"","tags":[],"see_also":[]},"channel_crash":{"name":"channel_crash","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"Share metadata about Ceph daemon crashes (version, stack straces, etc)","long_desc":"","tags":[],"see_also":[]},"channel_device":{"name":"channel_device","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"Share device health metrics (e.g., SMART data, minus potentially identifying info like serial numbers)","long_desc":"","tags":[],"see_also":[]},"channel_ident":{"name":"channel_ident","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Share a user-provided description and/or contact email for the cluster","long_desc":"","tags":[],"see_also":[]},"channel_perf":{"name":"channel_perf","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Share various performance metrics of a cluster","long_desc":"","tags":[],"see_also":[]},"contact":{"name":"contact","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"description":{"name":"description","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"device_url":{"name":"device_url","type":"str","level":"advanced","flags":0,"default_value":"https://telemetry.ceph.com/device","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"enabled":{"name":"enabled","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"interval":{"name":"interval","type":"int","level":"advanced","flags":0,"default_value":"24","min":"8","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"last_opt_revision":{"name":"last_opt_revision","type":"int","level":"advanced","flags":0,"default_value":"1","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"leaderboard":{"name":"leaderboard","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"leaderboard_description":{"name":"leaderboard_description","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"organization":{"name":"organization","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"proxy":{"name":"proxy","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"url":{"name":"url","type":"str","level":"advanced","flags":0,"default_value":"https://telemetry.ceph.com/report","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"test_orchestrator","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"volumes","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"max_concurrent_clones":{"name":"max_concurrent_clones","type":"int","level":"advanced","flags":0,"default_value":"4","min":"","max":"","enum_allowed":[],"desc":"Number of asynchronous cloner threads","long_desc":"","tags":[],"see_also":[]},"pause_cloning":{"name":"pause_cloning","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Pause asynchronous cloner threads","long_desc":"","tags":[],"see_also":[]},"pause_purging":{"name":"pause_purging","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Pause asynchronous subvolume purge threads","long_desc":"","tags":[],"see_also":[]},"periodic_async_work":{"name":"periodic_async_work","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Periodically check for async work","long_desc":"","tags":[],"see_also":[]},"snapshot_clone_delay":{"name":"snapshot_clone_delay","type":"int","level":"advanced","flags":0,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"Delay clone begin operation by snapshot_clone_delay seconds","long_desc":"","tags":[],"see_also":[]},"snapshot_clone_no_wait":{"name":"snapshot_clone_no_wait","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"Reject subvolume clone request when cloner threads are busy","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}}],"services":{},"always_on_modules":{"octopus":["balancer","crash","devicehealth","orchestrator","pg_autoscaler","progress","rbd_support","status","telemetry","volumes"],"pacific":["balancer","crash","devicehealth","orchestrator","pg_autoscaler","progress","rbd_support","status","telemetry","volumes"],"quincy":["balancer","crash","devicehealth","orchestrator","pg_autoscaler","progress","rbd_support","status","telemetry","volumes"],"reef":["balancer","crash","devicehealth","orchestrator","pg_autoscaler","progress","rbd_support","status","telemetry","volumes"],"squid":["balancer","crash","devicehealth","orchestrator","pg_autoscaler","progress","rbd_support","status","telemetry","volumes"],"tentacle":["balancer","crash","devicehealth","orchestrator","pg_autoscaler","progress","rbd_support","status","telemetry","volumes"]},"force_disabled_modules":{},"last_failure_osd_epoch":0,"active_clients":[{"name":"devicehealth","addrvec":[{"type":"v2","addr":"192.168.123.108:0","nonce":2994162241}]},{"name":"libcephsqlite","addrvec":[{"type":"v2","addr":"192.168.123.108:0","nonce":1293458211}]},{"name":"rbd_support","addrvec":[{"type":"v2","addr":"192.168.123.108:0","nonce":2165837346}]},{"name":"volumes","addrvec":[{"type":"v2","addr":"192.168.123.108:0","nonce":1534941629}]}]} 2026-03-20T17:57:50.351 INFO:tasks.ceph.ceph_manager.c2:mgr available! 2026-03-20T17:57:50.351 INFO:tasks.ceph.ceph_manager.c2:waiting for all up 2026-03-20T17:57:50.351 DEBUG:teuthology.orchestra.run.vm06:> sudo adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage timeout 120 ceph --cluster c2 osd dump --format=json 2026-03-20T17:57:50.525 INFO:teuthology.orchestra.run.vm06.stdout: 2026-03-20T17:57:50.525 INFO:teuthology.orchestra.run.vm06.stdout:{"epoch":18,"fsid":"02f24f96-dd4e-4d53-915a-6d7df6cbe6f6","created":"2026-03-20T17:57:37.534249+0000","modified":"2026-03-20T17:57:49.899429+0000","last_up_change":"2026-03-20T17:57:41.678232+0000","last_in_change":"2026-03-20T17:57:38.670004+0000","flags":"sortbitwise,recovery_deletes,purged_snapdirs,pglog_hardlimit","flags_num":5799936,"flags_set":["pglog_hardlimit","purged_snapdirs","recovery_deletes","sortbitwise"],"crush_version":4,"full_ratio":0.94999998807907104,"backfillfull_ratio":0.89999997615814209,"nearfull_ratio":0.85000002384185791,"cluster_snapshot":"","pool_max":2,"max_osd":6,"require_min_compat_client":"luminous","min_compat_client":"jewel","require_osd_release":"tentacle","allow_crimson":false,"pools":[{"pool":1,"pool_name":".mgr","create_time":"2026-03-20T17:57:41.855010+0000","flags":1,"flags_names":"hashpspool","type":1,"size":2,"min_size":1,"crush_rule":0,"peering_crush_bucket_count":0,"peering_crush_bucket_target":0,"peering_crush_bucket_barrier":0,"peering_crush_bucket_mandatory_member":2147483647,"is_stretch_pool":false,"object_hash":2,"pg_autoscale_mode":"off","pg_num":1,"pg_placement_num":1,"pg_placement_num_target":1,"pg_num_target":1,"pg_num_pending":1,"last_pg_merge_meta":{"source_pgid":"0.0","ready_epoch":0,"last_epoch_started":0,"last_epoch_clean":0,"source_version":"0'0","target_version":"0'0"},"last_change":"14","last_force_op_resend":"0","last_force_op_resend_prenautilus":"0","last_force_op_resend_preluminous":"0","auid":0,"snap_mode":"selfmanaged","snap_seq":0,"snap_epoch":0,"pool_snaps":[],"removed_snaps":"[]","quota_max_bytes":0,"quota_max_objects":0,"tiers":[],"tier_of":-1,"read_tier":-1,"write_tier":-1,"cache_mode":"none","target_max_bytes":0,"target_max_objects":0,"cache_target_dirty_ratio_micro":400000,"cache_target_dirty_high_ratio_micro":600000,"cache_target_full_ratio_micro":800000,"cache_min_flush_age":0,"cache_min_evict_age":0,"erasure_code_profile":"","hit_set_params":{"type":"none"},"hit_set_period":0,"hit_set_count":0,"use_gmt_hitset":true,"min_read_recency_for_promote":0,"min_write_recency_for_promote":0,"hit_set_grade_decay_rate":0,"hit_set_search_last_n":0,"grade_table":[],"stripe_width":0,"expected_num_objects":0,"fast_read":false,"nonprimary_shards":"{}","options":{"pg_num_max":32,"pg_num_min":1},"application_metadata":{"mgr":{}},"read_balance":{"score_type":"Fair distribution","score_acting":6.059999942779541,"score_stable":6.059999942779541,"optimal_score":0.33000001311302185,"raw_score_acting":2,"raw_score_stable":2,"primary_affinity_weighted":1,"average_primary_affinity":1,"average_primary_affinity_weighted":1}},{"pool":2,"pool_name":"rbd","create_time":"2026-03-20T17:57:46.479692+0000","flags":8193,"flags_names":"hashpspool,selfmanaged_snaps","type":1,"size":2,"min_size":1,"crush_rule":0,"peering_crush_bucket_count":0,"peering_crush_bucket_target":0,"peering_crush_bucket_barrier":0,"peering_crush_bucket_mandatory_member":2147483647,"is_stretch_pool":false,"object_hash":2,"pg_autoscale_mode":"off","pg_num":8,"pg_placement_num":8,"pg_placement_num_target":8,"pg_num_target":8,"pg_num_pending":8,"last_pg_merge_meta":{"source_pgid":"0.0","ready_epoch":0,"last_epoch_started":0,"last_epoch_clean":0,"source_version":"0'0","target_version":"0'0"},"last_change":"18","last_force_op_resend":"0","last_force_op_resend_prenautilus":"0","last_force_op_resend_preluminous":"0","auid":0,"snap_mode":"selfmanaged","snap_seq":2,"snap_epoch":18,"pool_snaps":[],"removed_snaps":"[]","quota_max_bytes":0,"quota_max_objects":0,"tiers":[],"tier_of":-1,"read_tier":-1,"write_tier":-1,"cache_mode":"none","target_max_bytes":0,"target_max_objects":0,"cache_target_dirty_ratio_micro":400000,"cache_target_dirty_high_ratio_micro":600000,"cache_target_full_ratio_micro":800000,"cache_min_flush_age":0,"cache_min_evict_age":0,"erasure_code_profile":"","hit_set_params":{"type":"none"},"hit_set_period":0,"hit_set_count":0,"use_gmt_hitset":true,"min_read_recency_for_promote":0,"min_write_recency_for_promote":0,"hit_set_grade_decay_rate":0,"hit_set_search_last_n":0,"grade_table":[],"stripe_width":0,"expected_num_objects":0,"fast_read":false,"nonprimary_shards":"{}","options":{},"application_metadata":{"rbd":{}},"read_balance":{"score_type":"Fair distribution","score_acting":2.25,"score_stable":2.25,"optimal_score":1,"raw_score_acting":2.25,"raw_score_stable":2.25,"primary_affinity_weighted":1,"average_primary_affinity":1,"average_primary_affinity_weighted":1}}],"osds":[{"osd":0,"uuid":"3425c007-e53b-452b-83a2-185100f81062","up":1,"in":1,"weight":1,"primary_affinity":1,"last_clean_begin":0,"last_clean_end":0,"up_from":11,"up_thru":15,"down_at":0,"lost_at":0,"public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.106:6800","nonce":3200331076},{"type":"v1","addr":"192.168.123.106:6801","nonce":3200331076}]},"cluster_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.106:6802","nonce":3200331076},{"type":"v1","addr":"192.168.123.106:6803","nonce":3200331076}]},"heartbeat_back_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.106:6806","nonce":3200331076},{"type":"v1","addr":"192.168.123.106:6807","nonce":3200331076}]},"heartbeat_front_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.106:6804","nonce":3200331076},{"type":"v1","addr":"192.168.123.106:6805","nonce":3200331076}]},"public_addr":"192.168.123.106:6801/3200331076","cluster_addr":"192.168.123.106:6803/3200331076","heartbeat_back_addr":"192.168.123.106:6807/3200331076","heartbeat_front_addr":"192.168.123.106:6805/3200331076","state":["exists","up"]},{"osd":1,"uuid":"6b897cf8-b9b9-46da-a8e9-a4fa94a6bcff","up":1,"in":1,"weight":1,"primary_affinity":1,"last_clean_begin":0,"last_clean_end":0,"up_from":11,"up_thru":15,"down_at":0,"lost_at":0,"public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.106:6808","nonce":630280801},{"type":"v1","addr":"192.168.123.106:6809","nonce":630280801}]},"cluster_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.106:6810","nonce":630280801},{"type":"v1","addr":"192.168.123.106:6811","nonce":630280801}]},"heartbeat_back_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.106:6814","nonce":630280801},{"type":"v1","addr":"192.168.123.106:6815","nonce":630280801}]},"heartbeat_front_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.106:6812","nonce":630280801},{"type":"v1","addr":"192.168.123.106:6813","nonce":630280801}]},"public_addr":"192.168.123.106:6809/630280801","cluster_addr":"192.168.123.106:6811/630280801","heartbeat_back_addr":"192.168.123.106:6815/630280801","heartbeat_front_addr":"192.168.123.106:6813/630280801","state":["exists","up"]},{"osd":2,"uuid":"16f37894-1ade-42af-a668-a7c2d98c3d31","up":1,"in":1,"weight":1,"primary_affinity":1,"last_clean_begin":0,"last_clean_end":0,"up_from":11,"up_thru":15,"down_at":0,"lost_at":0,"public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.106:6816","nonce":905863174},{"type":"v1","addr":"192.168.123.106:6817","nonce":905863174}]},"cluster_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.106:6818","nonce":905863174},{"type":"v1","addr":"192.168.123.106:6819","nonce":905863174}]},"heartbeat_back_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.106:6822","nonce":905863174},{"type":"v1","addr":"192.168.123.106:6823","nonce":905863174}]},"heartbeat_front_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.106:6820","nonce":905863174},{"type":"v1","addr":"192.168.123.106:6821","nonce":905863174}]},"public_addr":"192.168.123.106:6817/905863174","cluster_addr":"192.168.123.106:6819/905863174","heartbeat_back_addr":"192.168.123.106:6823/905863174","heartbeat_front_addr":"192.168.123.106:6821/905863174","state":["exists","up"]},{"osd":3,"uuid":"0fa43dbc-6811-4426-9ff2-82a296a19efb","up":1,"in":1,"weight":1,"primary_affinity":1,"last_clean_begin":0,"last_clean_end":0,"up_from":11,"up_thru":15,"down_at":0,"lost_at":0,"public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.108:6800","nonce":1817589236},{"type":"v1","addr":"192.168.123.108:6801","nonce":1817589236}]},"cluster_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.108:6802","nonce":1817589236},{"type":"v1","addr":"192.168.123.108:6803","nonce":1817589236}]},"heartbeat_back_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.108:6806","nonce":1817589236},{"type":"v1","addr":"192.168.123.108:6807","nonce":1817589236}]},"heartbeat_front_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.108:6804","nonce":1817589236},{"type":"v1","addr":"192.168.123.108:6805","nonce":1817589236}]},"public_addr":"192.168.123.108:6801/1817589236","cluster_addr":"192.168.123.108:6803/1817589236","heartbeat_back_addr":"192.168.123.108:6807/1817589236","heartbeat_front_addr":"192.168.123.108:6805/1817589236","state":["exists","up"]},{"osd":4,"uuid":"e58e5154-9b17-4109-926c-b20bd25945b6","up":1,"in":1,"weight":1,"primary_affinity":1,"last_clean_begin":0,"last_clean_end":0,"up_from":11,"up_thru":15,"down_at":0,"lost_at":0,"public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.108:6816","nonce":4097418099},{"type":"v1","addr":"192.168.123.108:6817","nonce":4097418099}]},"cluster_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.108:6818","nonce":4097418099},{"type":"v1","addr":"192.168.123.108:6819","nonce":4097418099}]},"heartbeat_back_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.108:6822","nonce":4097418099},{"type":"v1","addr":"192.168.123.108:6823","nonce":4097418099}]},"heartbeat_front_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.108:6820","nonce":4097418099},{"type":"v1","addr":"192.168.123.108:6821","nonce":4097418099}]},"public_addr":"192.168.123.108:6817/4097418099","cluster_addr":"192.168.123.108:6819/4097418099","heartbeat_back_addr":"192.168.123.108:6823/4097418099","heartbeat_front_addr":"192.168.123.108:6821/4097418099","state":["exists","up"]},{"osd":5,"uuid":"9d3c09ae-4c03-41c0-88b1-13e6e36d67bf","up":1,"in":1,"weight":1,"primary_affinity":1,"last_clean_begin":0,"last_clean_end":0,"up_from":11,"up_thru":15,"down_at":0,"lost_at":0,"public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.108:6808","nonce":1082012952},{"type":"v1","addr":"192.168.123.108:6809","nonce":1082012952}]},"cluster_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.108:6810","nonce":1082012952},{"type":"v1","addr":"192.168.123.108:6811","nonce":1082012952}]},"heartbeat_back_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.108:6814","nonce":1082012952},{"type":"v1","addr":"192.168.123.108:6815","nonce":1082012952}]},"heartbeat_front_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.108:6812","nonce":1082012952},{"type":"v1","addr":"192.168.123.108:6813","nonce":1082012952}]},"public_addr":"192.168.123.108:6809/1082012952","cluster_addr":"192.168.123.108:6811/1082012952","heartbeat_back_addr":"192.168.123.108:6815/1082012952","heartbeat_front_addr":"192.168.123.108:6813/1082012952","state":["exists","up"]}],"osd_xinfo":[{"osd":0,"down_stamp":"0.000000","laggy_probability":0,"laggy_interval":0,"features":4544132024016699391,"old_weight":0,"last_purged_snaps_scrub":"2026-03-20T17:57:40.304164+0000","dead_epoch":0},{"osd":1,"down_stamp":"0.000000","laggy_probability":0,"laggy_interval":0,"features":4544132024016699391,"old_weight":0,"last_purged_snaps_scrub":"2026-03-20T17:57:40.311385+0000","dead_epoch":0},{"osd":2,"down_stamp":"0.000000","laggy_probability":0,"laggy_interval":0,"features":4544132024016699391,"old_weight":0,"last_purged_snaps_scrub":"2026-03-20T17:57:40.407763+0000","dead_epoch":0},{"osd":3,"down_stamp":"0.000000","laggy_probability":0,"laggy_interval":0,"features":4544132024016699391,"old_weight":0,"last_purged_snaps_scrub":"2026-03-20T17:57:40.210113+0000","dead_epoch":0},{"osd":4,"down_stamp":"0.000000","laggy_probability":0,"laggy_interval":0,"features":4544132024016699391,"old_weight":0,"last_purged_snaps_scrub":"2026-03-20T17:57:40.384187+0000","dead_epoch":0},{"osd":5,"down_stamp":"0.000000","laggy_probability":0,"laggy_interval":0,"features":4544132024016699391,"old_weight":0,"last_purged_snaps_scrub":"2026-03-20T17:57:40.282793+0000","dead_epoch":0}],"pg_upmap":[],"pg_upmap_items":[],"pg_upmap_primaries":[],"pg_temp":[],"primary_temp":[],"blocklist":{},"range_blocklist":{},"erasure_code_profiles":{"default":{"crush-failure-domain":"osd","k":"2","m":"1","plugin":"isa","technique":"reed_sol_van"}},"removed_snaps_queue":[{"pool":2,"snaps":[{"begin":2,"length":1}]}],"new_removed_snaps":[{"pool":2,"snaps":[{"begin":2,"length":1}]}],"new_purged_snaps":[],"crush_node_flags":{},"device_class_flags":{},"stretch_mode":{"stretch_mode_enabled":false,"stretch_bucket_count":0,"degraded_stretch_mode":0,"recovering_stretch_mode":0,"stretch_mode_bucket":0}} 2026-03-20T17:57:50.539 INFO:tasks.ceph.ceph_manager.c2:all up! 2026-03-20T17:57:50.539 DEBUG:teuthology.orchestra.run.vm06:> sudo adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage timeout 120 ceph --cluster c2 osd dump --format=json 2026-03-20T17:57:50.712 INFO:teuthology.orchestra.run.vm06.stdout: 2026-03-20T17:57:50.712 INFO:teuthology.orchestra.run.vm06.stdout:{"epoch":18,"fsid":"02f24f96-dd4e-4d53-915a-6d7df6cbe6f6","created":"2026-03-20T17:57:37.534249+0000","modified":"2026-03-20T17:57:49.899429+0000","last_up_change":"2026-03-20T17:57:41.678232+0000","last_in_change":"2026-03-20T17:57:38.670004+0000","flags":"sortbitwise,recovery_deletes,purged_snapdirs,pglog_hardlimit","flags_num":5799936,"flags_set":["pglog_hardlimit","purged_snapdirs","recovery_deletes","sortbitwise"],"crush_version":4,"full_ratio":0.94999998807907104,"backfillfull_ratio":0.89999997615814209,"nearfull_ratio":0.85000002384185791,"cluster_snapshot":"","pool_max":2,"max_osd":6,"require_min_compat_client":"luminous","min_compat_client":"jewel","require_osd_release":"tentacle","allow_crimson":false,"pools":[{"pool":1,"pool_name":".mgr","create_time":"2026-03-20T17:57:41.855010+0000","flags":1,"flags_names":"hashpspool","type":1,"size":2,"min_size":1,"crush_rule":0,"peering_crush_bucket_count":0,"peering_crush_bucket_target":0,"peering_crush_bucket_barrier":0,"peering_crush_bucket_mandatory_member":2147483647,"is_stretch_pool":false,"object_hash":2,"pg_autoscale_mode":"off","pg_num":1,"pg_placement_num":1,"pg_placement_num_target":1,"pg_num_target":1,"pg_num_pending":1,"last_pg_merge_meta":{"source_pgid":"0.0","ready_epoch":0,"last_epoch_started":0,"last_epoch_clean":0,"source_version":"0'0","target_version":"0'0"},"last_change":"14","last_force_op_resend":"0","last_force_op_resend_prenautilus":"0","last_force_op_resend_preluminous":"0","auid":0,"snap_mode":"selfmanaged","snap_seq":0,"snap_epoch":0,"pool_snaps":[],"removed_snaps":"[]","quota_max_bytes":0,"quota_max_objects":0,"tiers":[],"tier_of":-1,"read_tier":-1,"write_tier":-1,"cache_mode":"none","target_max_bytes":0,"target_max_objects":0,"cache_target_dirty_ratio_micro":400000,"cache_target_dirty_high_ratio_micro":600000,"cache_target_full_ratio_micro":800000,"cache_min_flush_age":0,"cache_min_evict_age":0,"erasure_code_profile":"","hit_set_params":{"type":"none"},"hit_set_period":0,"hit_set_count":0,"use_gmt_hitset":true,"min_read_recency_for_promote":0,"min_write_recency_for_promote":0,"hit_set_grade_decay_rate":0,"hit_set_search_last_n":0,"grade_table":[],"stripe_width":0,"expected_num_objects":0,"fast_read":false,"nonprimary_shards":"{}","options":{"pg_num_max":32,"pg_num_min":1},"application_metadata":{"mgr":{}},"read_balance":{"score_type":"Fair distribution","score_acting":6.059999942779541,"score_stable":6.059999942779541,"optimal_score":0.33000001311302185,"raw_score_acting":2,"raw_score_stable":2,"primary_affinity_weighted":1,"average_primary_affinity":1,"average_primary_affinity_weighted":1}},{"pool":2,"pool_name":"rbd","create_time":"2026-03-20T17:57:46.479692+0000","flags":8193,"flags_names":"hashpspool,selfmanaged_snaps","type":1,"size":2,"min_size":1,"crush_rule":0,"peering_crush_bucket_count":0,"peering_crush_bucket_target":0,"peering_crush_bucket_barrier":0,"peering_crush_bucket_mandatory_member":2147483647,"is_stretch_pool":false,"object_hash":2,"pg_autoscale_mode":"off","pg_num":8,"pg_placement_num":8,"pg_placement_num_target":8,"pg_num_target":8,"pg_num_pending":8,"last_pg_merge_meta":{"source_pgid":"0.0","ready_epoch":0,"last_epoch_started":0,"last_epoch_clean":0,"source_version":"0'0","target_version":"0'0"},"last_change":"18","last_force_op_resend":"0","last_force_op_resend_prenautilus":"0","last_force_op_resend_preluminous":"0","auid":0,"snap_mode":"selfmanaged","snap_seq":2,"snap_epoch":18,"pool_snaps":[],"removed_snaps":"[]","quota_max_bytes":0,"quota_max_objects":0,"tiers":[],"tier_of":-1,"read_tier":-1,"write_tier":-1,"cache_mode":"none","target_max_bytes":0,"target_max_objects":0,"cache_target_dirty_ratio_micro":400000,"cache_target_dirty_high_ratio_micro":600000,"cache_target_full_ratio_micro":800000,"cache_min_flush_age":0,"cache_min_evict_age":0,"erasure_code_profile":"","hit_set_params":{"type":"none"},"hit_set_period":0,"hit_set_count":0,"use_gmt_hitset":true,"min_read_recency_for_promote":0,"min_write_recency_for_promote":0,"hit_set_grade_decay_rate":0,"hit_set_search_last_n":0,"grade_table":[],"stripe_width":0,"expected_num_objects":0,"fast_read":false,"nonprimary_shards":"{}","options":{},"application_metadata":{"rbd":{}},"read_balance":{"score_type":"Fair distribution","score_acting":2.25,"score_stable":2.25,"optimal_score":1,"raw_score_acting":2.25,"raw_score_stable":2.25,"primary_affinity_weighted":1,"average_primary_affinity":1,"average_primary_affinity_weighted":1}}],"osds":[{"osd":0,"uuid":"3425c007-e53b-452b-83a2-185100f81062","up":1,"in":1,"weight":1,"primary_affinity":1,"last_clean_begin":0,"last_clean_end":0,"up_from":11,"up_thru":15,"down_at":0,"lost_at":0,"public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.106:6800","nonce":3200331076},{"type":"v1","addr":"192.168.123.106:6801","nonce":3200331076}]},"cluster_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.106:6802","nonce":3200331076},{"type":"v1","addr":"192.168.123.106:6803","nonce":3200331076}]},"heartbeat_back_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.106:6806","nonce":3200331076},{"type":"v1","addr":"192.168.123.106:6807","nonce":3200331076}]},"heartbeat_front_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.106:6804","nonce":3200331076},{"type":"v1","addr":"192.168.123.106:6805","nonce":3200331076}]},"public_addr":"192.168.123.106:6801/3200331076","cluster_addr":"192.168.123.106:6803/3200331076","heartbeat_back_addr":"192.168.123.106:6807/3200331076","heartbeat_front_addr":"192.168.123.106:6805/3200331076","state":["exists","up"]},{"osd":1,"uuid":"6b897cf8-b9b9-46da-a8e9-a4fa94a6bcff","up":1,"in":1,"weight":1,"primary_affinity":1,"last_clean_begin":0,"last_clean_end":0,"up_from":11,"up_thru":15,"down_at":0,"lost_at":0,"public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.106:6808","nonce":630280801},{"type":"v1","addr":"192.168.123.106:6809","nonce":630280801}]},"cluster_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.106:6810","nonce":630280801},{"type":"v1","addr":"192.168.123.106:6811","nonce":630280801}]},"heartbeat_back_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.106:6814","nonce":630280801},{"type":"v1","addr":"192.168.123.106:6815","nonce":630280801}]},"heartbeat_front_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.106:6812","nonce":630280801},{"type":"v1","addr":"192.168.123.106:6813","nonce":630280801}]},"public_addr":"192.168.123.106:6809/630280801","cluster_addr":"192.168.123.106:6811/630280801","heartbeat_back_addr":"192.168.123.106:6815/630280801","heartbeat_front_addr":"192.168.123.106:6813/630280801","state":["exists","up"]},{"osd":2,"uuid":"16f37894-1ade-42af-a668-a7c2d98c3d31","up":1,"in":1,"weight":1,"primary_affinity":1,"last_clean_begin":0,"last_clean_end":0,"up_from":11,"up_thru":15,"down_at":0,"lost_at":0,"public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.106:6816","nonce":905863174},{"type":"v1","addr":"192.168.123.106:6817","nonce":905863174}]},"cluster_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.106:6818","nonce":905863174},{"type":"v1","addr":"192.168.123.106:6819","nonce":905863174}]},"heartbeat_back_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.106:6822","nonce":905863174},{"type":"v1","addr":"192.168.123.106:6823","nonce":905863174}]},"heartbeat_front_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.106:6820","nonce":905863174},{"type":"v1","addr":"192.168.123.106:6821","nonce":905863174}]},"public_addr":"192.168.123.106:6817/905863174","cluster_addr":"192.168.123.106:6819/905863174","heartbeat_back_addr":"192.168.123.106:6823/905863174","heartbeat_front_addr":"192.168.123.106:6821/905863174","state":["exists","up"]},{"osd":3,"uuid":"0fa43dbc-6811-4426-9ff2-82a296a19efb","up":1,"in":1,"weight":1,"primary_affinity":1,"last_clean_begin":0,"last_clean_end":0,"up_from":11,"up_thru":15,"down_at":0,"lost_at":0,"public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.108:6800","nonce":1817589236},{"type":"v1","addr":"192.168.123.108:6801","nonce":1817589236}]},"cluster_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.108:6802","nonce":1817589236},{"type":"v1","addr":"192.168.123.108:6803","nonce":1817589236}]},"heartbeat_back_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.108:6806","nonce":1817589236},{"type":"v1","addr":"192.168.123.108:6807","nonce":1817589236}]},"heartbeat_front_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.108:6804","nonce":1817589236},{"type":"v1","addr":"192.168.123.108:6805","nonce":1817589236}]},"public_addr":"192.168.123.108:6801/1817589236","cluster_addr":"192.168.123.108:6803/1817589236","heartbeat_back_addr":"192.168.123.108:6807/1817589236","heartbeat_front_addr":"192.168.123.108:6805/1817589236","state":["exists","up"]},{"osd":4,"uuid":"e58e5154-9b17-4109-926c-b20bd25945b6","up":1,"in":1,"weight":1,"primary_affinity":1,"last_clean_begin":0,"last_clean_end":0,"up_from":11,"up_thru":15,"down_at":0,"lost_at":0,"public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.108:6816","nonce":4097418099},{"type":"v1","addr":"192.168.123.108:6817","nonce":4097418099}]},"cluster_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.108:6818","nonce":4097418099},{"type":"v1","addr":"192.168.123.108:6819","nonce":4097418099}]},"heartbeat_back_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.108:6822","nonce":4097418099},{"type":"v1","addr":"192.168.123.108:6823","nonce":4097418099}]},"heartbeat_front_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.108:6820","nonce":4097418099},{"type":"v1","addr":"192.168.123.108:6821","nonce":4097418099}]},"public_addr":"192.168.123.108:6817/4097418099","cluster_addr":"192.168.123.108:6819/4097418099","heartbeat_back_addr":"192.168.123.108:6823/4097418099","heartbeat_front_addr":"192.168.123.108:6821/4097418099","state":["exists","up"]},{"osd":5,"uuid":"9d3c09ae-4c03-41c0-88b1-13e6e36d67bf","up":1,"in":1,"weight":1,"primary_affinity":1,"last_clean_begin":0,"last_clean_end":0,"up_from":11,"up_thru":15,"down_at":0,"lost_at":0,"public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.108:6808","nonce":1082012952},{"type":"v1","addr":"192.168.123.108:6809","nonce":1082012952}]},"cluster_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.108:6810","nonce":1082012952},{"type":"v1","addr":"192.168.123.108:6811","nonce":1082012952}]},"heartbeat_back_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.108:6814","nonce":1082012952},{"type":"v1","addr":"192.168.123.108:6815","nonce":1082012952}]},"heartbeat_front_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.108:6812","nonce":1082012952},{"type":"v1","addr":"192.168.123.108:6813","nonce":1082012952}]},"public_addr":"192.168.123.108:6809/1082012952","cluster_addr":"192.168.123.108:6811/1082012952","heartbeat_back_addr":"192.168.123.108:6815/1082012952","heartbeat_front_addr":"192.168.123.108:6813/1082012952","state":["exists","up"]}],"osd_xinfo":[{"osd":0,"down_stamp":"0.000000","laggy_probability":0,"laggy_interval":0,"features":4544132024016699391,"old_weight":0,"last_purged_snaps_scrub":"2026-03-20T17:57:40.304164+0000","dead_epoch":0},{"osd":1,"down_stamp":"0.000000","laggy_probability":0,"laggy_interval":0,"features":4544132024016699391,"old_weight":0,"last_purged_snaps_scrub":"2026-03-20T17:57:40.311385+0000","dead_epoch":0},{"osd":2,"down_stamp":"0.000000","laggy_probability":0,"laggy_interval":0,"features":4544132024016699391,"old_weight":0,"last_purged_snaps_scrub":"2026-03-20T17:57:40.407763+0000","dead_epoch":0},{"osd":3,"down_stamp":"0.000000","laggy_probability":0,"laggy_interval":0,"features":4544132024016699391,"old_weight":0,"last_purged_snaps_scrub":"2026-03-20T17:57:40.210113+0000","dead_epoch":0},{"osd":4,"down_stamp":"0.000000","laggy_probability":0,"laggy_interval":0,"features":4544132024016699391,"old_weight":0,"last_purged_snaps_scrub":"2026-03-20T17:57:40.384187+0000","dead_epoch":0},{"osd":5,"down_stamp":"0.000000","laggy_probability":0,"laggy_interval":0,"features":4544132024016699391,"old_weight":0,"last_purged_snaps_scrub":"2026-03-20T17:57:40.282793+0000","dead_epoch":0}],"pg_upmap":[],"pg_upmap_items":[],"pg_upmap_primaries":[],"pg_temp":[],"primary_temp":[],"blocklist":{},"range_blocklist":{},"erasure_code_profiles":{"default":{"crush-failure-domain":"osd","k":"2","m":"1","plugin":"isa","technique":"reed_sol_van"}},"removed_snaps_queue":[{"pool":2,"snaps":[{"begin":2,"length":1}]}],"new_removed_snaps":[{"pool":2,"snaps":[{"begin":2,"length":1}]}],"new_purged_snaps":[],"crush_node_flags":{},"device_class_flags":{},"stretch_mode":{"stretch_mode_enabled":false,"stretch_bucket_count":0,"degraded_stretch_mode":0,"recovering_stretch_mode":0,"stretch_mode_bucket":0}} 2026-03-20T17:57:50.727 DEBUG:teuthology.orchestra.run.vm06:> sudo adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage timeout 120 ceph --cluster c2 tell osd.0 flush_pg_stats 2026-03-20T17:57:50.727 DEBUG:teuthology.orchestra.run.vm06:> sudo adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage timeout 120 ceph --cluster c2 tell osd.1 flush_pg_stats 2026-03-20T17:57:50.727 DEBUG:teuthology.orchestra.run.vm06:> sudo adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage timeout 120 ceph --cluster c2 tell osd.2 flush_pg_stats 2026-03-20T17:57:50.727 DEBUG:teuthology.orchestra.run.vm06:> sudo adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage timeout 120 ceph --cluster c2 tell osd.3 flush_pg_stats 2026-03-20T17:57:50.727 DEBUG:teuthology.orchestra.run.vm06:> sudo adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage timeout 120 ceph --cluster c2 tell osd.4 flush_pg_stats 2026-03-20T17:57:50.727 DEBUG:teuthology.orchestra.run.vm06:> sudo adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage timeout 120 ceph --cluster c2 tell osd.5 flush_pg_stats 2026-03-20T17:57:50.854 INFO:teuthology.orchestra.run.vm06.stdout:47244640260 2026-03-20T17:57:50.854 DEBUG:teuthology.orchestra.run.vm06:> sudo adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage timeout 120 ceph --cluster c2 osd last-stat-seq osd.0 2026-03-20T17:57:50.859 INFO:teuthology.orchestra.run.vm06.stdout:47244640259 2026-03-20T17:57:50.859 DEBUG:teuthology.orchestra.run.vm06:> sudo adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage timeout 120 ceph --cluster c2 osd last-stat-seq osd.1 2026-03-20T17:57:50.901 INFO:teuthology.orchestra.run.vm06.stdout:47244640260 2026-03-20T17:57:50.901 DEBUG:teuthology.orchestra.run.vm06:> sudo adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage timeout 120 ceph --cluster c2 osd last-stat-seq osd.3 2026-03-20T17:57:50.910 INFO:teuthology.orchestra.run.vm06.stdout:47244640259 2026-03-20T17:57:50.910 DEBUG:teuthology.orchestra.run.vm06:> sudo adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage timeout 120 ceph --cluster c2 osd last-stat-seq osd.2 2026-03-20T17:57:50.939 INFO:teuthology.orchestra.run.vm06.stdout:47244640259 2026-03-20T17:57:50.939 DEBUG:teuthology.orchestra.run.vm06:> sudo adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage timeout 120 ceph --cluster c2 osd last-stat-seq osd.4 2026-03-20T17:57:50.943 INFO:teuthology.orchestra.run.vm06.stdout:47244640259 2026-03-20T17:57:50.943 DEBUG:teuthology.orchestra.run.vm06:> sudo adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage timeout 120 ceph --cluster c2 osd last-stat-seq osd.5 2026-03-20T17:57:51.100 INFO:teuthology.orchestra.run.vm06.stdout:47244640259 2026-03-20T17:57:51.130 INFO:tasks.ceph.ceph_manager.c2:need seq 47244640260 got 47244640259 for osd.0 2026-03-20T17:57:51.146 INFO:teuthology.orchestra.run.vm06.stdout:47244640258 2026-03-20T17:57:51.161 INFO:tasks.ceph.ceph_manager.c2:need seq 47244640259 got 47244640258 for osd.1 2026-03-20T17:57:51.185 INFO:teuthology.orchestra.run.vm06.stdout:47244640259 2026-03-20T17:57:51.196 INFO:teuthology.orchestra.run.vm06.stdout:47244640258 2026-03-20T17:57:51.203 INFO:tasks.ceph.ceph_manager.c2:need seq 47244640260 got 47244640259 for osd.3 2026-03-20T17:57:51.218 INFO:tasks.ceph.ceph_manager.c2:need seq 47244640259 got 47244640258 for osd.2 2026-03-20T17:57:51.245 INFO:teuthology.orchestra.run.vm06.stdout:47244640258 2026-03-20T17:57:51.245 INFO:teuthology.orchestra.run.vm06.stdout:47244640258 2026-03-20T17:57:51.259 INFO:tasks.ceph.ceph_manager.c2:need seq 47244640259 got 47244640258 for osd.4 2026-03-20T17:57:51.261 INFO:tasks.ceph.ceph_manager.c2:need seq 47244640259 got 47244640258 for osd.5 2026-03-20T17:57:52.131 DEBUG:teuthology.orchestra.run.vm06:> sudo adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage timeout 120 ceph --cluster c2 osd last-stat-seq osd.0 2026-03-20T17:57:52.161 DEBUG:teuthology.orchestra.run.vm06:> sudo adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage timeout 120 ceph --cluster c2 osd last-stat-seq osd.1 2026-03-20T17:57:52.204 DEBUG:teuthology.orchestra.run.vm06:> sudo adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage timeout 120 ceph --cluster c2 osd last-stat-seq osd.3 2026-03-20T17:57:52.219 DEBUG:teuthology.orchestra.run.vm06:> sudo adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage timeout 120 ceph --cluster c2 osd last-stat-seq osd.2 2026-03-20T17:57:52.260 DEBUG:teuthology.orchestra.run.vm06:> sudo adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage timeout 120 ceph --cluster c2 osd last-stat-seq osd.4 2026-03-20T17:57:52.262 DEBUG:teuthology.orchestra.run.vm06:> sudo adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage timeout 120 ceph --cluster c2 osd last-stat-seq osd.5 2026-03-20T17:57:52.337 INFO:teuthology.orchestra.run.vm06.stdout:47244640260 2026-03-20T17:57:52.360 INFO:tasks.ceph.ceph_manager.c2:need seq 47244640260 got 47244640260 for osd.0 2026-03-20T17:57:52.360 DEBUG:teuthology.parallel:result is None 2026-03-20T17:57:52.372 INFO:teuthology.orchestra.run.vm06.stdout:47244640259 2026-03-20T17:57:52.410 INFO:tasks.ceph.ceph_manager.c2:need seq 47244640259 got 47244640259 for osd.1 2026-03-20T17:57:52.410 DEBUG:teuthology.parallel:result is None 2026-03-20T17:57:52.457 INFO:teuthology.orchestra.run.vm06.stdout:47244640259 2026-03-20T17:57:52.475 INFO:tasks.ceph.ceph_manager.c2:need seq 47244640259 got 47244640259 for osd.2 2026-03-20T17:57:52.475 DEBUG:teuthology.parallel:result is None 2026-03-20T17:57:52.482 INFO:teuthology.orchestra.run.vm06.stdout:47244640260 2026-03-20T17:57:52.497 INFO:tasks.ceph.ceph_manager.c2:need seq 47244640260 got 47244640260 for osd.3 2026-03-20T17:57:52.498 DEBUG:teuthology.parallel:result is None 2026-03-20T17:57:52.508 INFO:teuthology.orchestra.run.vm06.stdout:47244640259 2026-03-20T17:57:52.525 INFO:tasks.ceph.ceph_manager.c2:need seq 47244640259 got 47244640259 for osd.4 2026-03-20T17:57:52.525 DEBUG:teuthology.parallel:result is None 2026-03-20T17:57:52.526 INFO:teuthology.orchestra.run.vm06.stdout:47244640259 2026-03-20T17:57:52.540 INFO:tasks.ceph.ceph_manager.c2:need seq 47244640259 got 47244640259 for osd.5 2026-03-20T17:57:52.540 DEBUG:teuthology.parallel:result is None 2026-03-20T17:57:52.540 INFO:tasks.ceph.ceph_manager.c2:waiting for clean 2026-03-20T17:57:52.540 DEBUG:teuthology.orchestra.run.vm06:> sudo adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage timeout 120 ceph --cluster c2 pg dump --format=json 2026-03-20T17:57:52.754 INFO:teuthology.orchestra.run.vm06.stdout: 2026-03-20T17:57:52.754 INFO:teuthology.orchestra.run.vm06.stderr:dumped all 2026-03-20T17:57:52.768 INFO:teuthology.orchestra.run.vm06.stdout:{"pg_ready":true,"pg_map":{"version":17,"stamp":"2026-03-20T17:57:51.844997+0000","last_osdmap_epoch":0,"last_pg_scan":0,"pg_stats_sum":{"stat_sum":{"num_bytes":459299,"num_objects":4,"num_object_clones":0,"num_object_copies":8,"num_objects_missing_on_primary":0,"num_objects_missing":0,"num_objects_degraded":0,"num_objects_misplaced":0,"num_objects_unfound":0,"num_objects_dirty":4,"num_whiteouts":0,"num_read":46,"num_read_kb":37,"num_write":59,"num_write_kb":586,"num_scrub_errors":0,"num_shallow_scrub_errors":0,"num_deep_scrub_errors":0,"num_objects_recovered":0,"num_bytes_recovered":0,"num_keys_recovered":0,"num_objects_omap":0,"num_objects_hit_set_archive":0,"num_bytes_hit_set_archive":0,"num_flush":0,"num_flush_kb":0,"num_evict":0,"num_evict_kb":0,"num_promote":0,"num_flush_mode_high":0,"num_flush_mode_low":0,"num_evict_mode_some":0,"num_evict_mode_full":0,"num_objects_pinned":0,"num_legacy_snapsets":0,"num_large_omap_objects":0,"num_objects_manifest":0,"num_omap_bytes":0,"num_omap_keys":0,"num_objects_repaired":0},"store_stats":{"total":0,"available":0,"internally_reserved":0,"allocated":0,"data_stored":0,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":0},"log_size":35,"ondisk_log_size":35,"up":18,"acting":18,"num_store_stats":0},"osd_stats_sum":{"up_from":0,"seq":0,"num_pgs":16,"num_osds":6,"num_per_pool_osds":6,"num_per_pool_omap_osds":6,"kb":629145600,"kb_used":163000,"kb_used_data":1872,"kb_used_omap":45,"kb_used_meta":160850,"kb_avail":628982600,"statfs":{"total":644245094400,"available":644078182400,"internally_reserved":0,"allocated":1916928,"data_stored":1238174,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":46124,"internal_metadata":164711380},"hb_peers":[],"snap_trim_queue_len":0,"num_snap_trimming":0,"num_shards_repaired":0,"op_queue_age_hist":{"histogram":[],"upper_bound":1},"perf_stat":{"commit_latency_ms":1,"apply_latency_ms":1,"commit_latency_ns":1000000,"apply_latency_ns":1000000},"alerts":[],"network_ping_times":[]},"pg_stats_delta":{"stat_sum":{"num_bytes":19,"num_objects":2,"num_object_clones":0,"num_object_copies":4,"num_objects_missing_on_primary":0,"num_objects_missing":0,"num_objects_degraded":0,"num_objects_misplaced":0,"num_objects_unfound":0,"num_objects_dirty":2,"num_whiteouts":0,"num_read":0,"num_read_kb":0,"num_write":2,"num_write_kb":2,"num_scrub_errors":0,"num_shallow_scrub_errors":0,"num_deep_scrub_errors":0,"num_objects_recovered":0,"num_bytes_recovered":0,"num_keys_recovered":0,"num_objects_omap":0,"num_objects_hit_set_archive":0,"num_bytes_hit_set_archive":0,"num_flush":0,"num_flush_kb":0,"num_evict":0,"num_evict_kb":0,"num_promote":0,"num_flush_mode_high":0,"num_flush_mode_low":0,"num_evict_mode_some":0,"num_evict_mode_full":0,"num_objects_pinned":0,"num_legacy_snapsets":0,"num_large_omap_objects":0,"num_objects_manifest":0,"num_omap_bytes":0,"num_omap_keys":0,"num_objects_repaired":0},"store_stats":{"total":0,"available":0,"internally_reserved":0,"allocated":0,"data_stored":0,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":0},"log_size":0,"ondisk_log_size":0,"up":0,"acting":0,"num_store_stats":0,"stamp_delta":"4.956044"},"pg_stats":[{"pgid":"2.7","version":"0'0","reported_seq":13,"reported_epoch":17,"state":"active+clean","last_fresh":"2026-03-20T17:57:49.202484+0000","last_change":"2026-03-20T17:57:47.894639+0000","last_active":"2026-03-20T17:57:49.202484+0000","last_peered":"2026-03-20T17:57:49.202484+0000","last_clean":"2026-03-20T17:57:49.202484+0000","last_became_active":"2026-03-20T17:57:47.894483+0000","last_became_peered":"2026-03-20T17:57:47.894483+0000","last_unstale":"2026-03-20T17:57:49.202484+0000","last_undegraded":"2026-03-20T17:57:49.202484+0000","last_fullsized":"2026-03-20T17:57:49.202484+0000","mapping_epoch":15,"log_start":"0'0","ondisk_log_start":"0'0","created":15,"last_epoch_clean":16,"parent":"0.0","parent_split_bits":0,"last_scrub":"0'0","last_scrub_stamp":"2026-03-20T17:57:46.883576+0000","last_deep_scrub":"0'0","last_deep_scrub_stamp":"2026-03-20T17:57:46.883576+0000","last_clean_scrub_stamp":"2026-03-20T17:57:46.883576+0000","objects_scrubbed":0,"log_size":0,"log_dups_size":0,"ondisk_log_size":0,"stats_invalid":false,"dirty_stats_invalid":false,"omap_stats_invalid":false,"hitset_stats_invalid":false,"hitset_bytes_stats_invalid":false,"pin_stats_invalid":false,"manifest_stats_invalid":false,"snaptrimq_len":0,"last_scrub_duration":0,"scrub_schedule":"periodic scrub scheduled @ 2026-03-21T22:36:28.128151+0000","scrub_duration":0,"objects_trimmed":0,"snaptrim_duration":0,"stat_sum":{"num_bytes":0,"num_objects":0,"num_object_clones":0,"num_object_copies":0,"num_objects_missing_on_primary":0,"num_objects_missing":0,"num_objects_degraded":0,"num_objects_misplaced":0,"num_objects_unfound":0,"num_objects_dirty":0,"num_whiteouts":0,"num_read":0,"num_read_kb":0,"num_write":0,"num_write_kb":0,"num_scrub_errors":0,"num_shallow_scrub_errors":0,"num_deep_scrub_errors":0,"num_objects_recovered":0,"num_bytes_recovered":0,"num_keys_recovered":0,"num_objects_omap":0,"num_objects_hit_set_archive":0,"num_bytes_hit_set_archive":0,"num_flush":0,"num_flush_kb":0,"num_evict":0,"num_evict_kb":0,"num_promote":0,"num_flush_mode_high":0,"num_flush_mode_low":0,"num_evict_mode_some":0,"num_evict_mode_full":0,"num_objects_pinned":0,"num_legacy_snapsets":0,"num_large_omap_objects":0,"num_objects_manifest":0,"num_omap_bytes":0,"num_omap_keys":0,"num_objects_repaired":0},"up":[1,0],"acting":[1,0],"avail_no_missing":[],"object_location_counts":[],"blocked_by":[],"up_primary":1,"acting_primary":1,"purged_snaps":[]},{"pgid":"2.6","version":"0'0","reported_seq":20,"reported_epoch":18,"state":"active+clean","last_fresh":"2026-03-20T17:57:49.907815+0000","last_change":"2026-03-20T17:57:49.907815+0000","last_active":"2026-03-20T17:57:49.907815+0000","last_peered":"2026-03-20T17:57:49.907815+0000","last_clean":"2026-03-20T17:57:49.907815+0000","last_became_active":"2026-03-20T17:57:47.895290+0000","last_became_peered":"2026-03-20T17:57:47.895290+0000","last_unstale":"2026-03-20T17:57:49.907815+0000","last_undegraded":"2026-03-20T17:57:49.907815+0000","last_fullsized":"2026-03-20T17:57:49.907815+0000","mapping_epoch":15,"log_start":"0'0","ondisk_log_start":"0'0","created":15,"last_epoch_clean":16,"parent":"0.0","parent_split_bits":0,"last_scrub":"0'0","last_scrub_stamp":"2026-03-20T17:57:46.883576+0000","last_deep_scrub":"0'0","last_deep_scrub_stamp":"2026-03-20T17:57:46.883576+0000","last_clean_scrub_stamp":"2026-03-20T17:57:46.883576+0000","objects_scrubbed":0,"log_size":0,"log_dups_size":0,"ondisk_log_size":0,"stats_invalid":false,"dirty_stats_invalid":false,"omap_stats_invalid":false,"hitset_stats_invalid":false,"hitset_bytes_stats_invalid":false,"pin_stats_invalid":false,"manifest_stats_invalid":false,"snaptrimq_len":0,"last_scrub_duration":0,"scrub_schedule":"periodic scrub scheduled @ 2026-03-21T22:10:31.042370+0000","scrub_duration":0,"objects_trimmed":0,"snaptrim_duration":0.000290434,"stat_sum":{"num_bytes":0,"num_objects":0,"num_object_clones":0,"num_object_copies":0,"num_objects_missing_on_primary":0,"num_objects_missing":0,"num_objects_degraded":0,"num_objects_misplaced":0,"num_objects_unfound":0,"num_objects_dirty":0,"num_whiteouts":0,"num_read":0,"num_read_kb":0,"num_write":0,"num_write_kb":0,"num_scrub_errors":0,"num_shallow_scrub_errors":0,"num_deep_scrub_errors":0,"num_objects_recovered":0,"num_bytes_recovered":0,"num_keys_recovered":0,"num_objects_omap":0,"num_objects_hit_set_archive":0,"num_bytes_hit_set_archive":0,"num_flush":0,"num_flush_kb":0,"num_evict":0,"num_evict_kb":0,"num_promote":0,"num_flush_mode_high":0,"num_flush_mode_low":0,"num_evict_mode_some":0,"num_evict_mode_full":0,"num_objects_pinned":0,"num_legacy_snapsets":0,"num_large_omap_objects":0,"num_objects_manifest":0,"num_omap_bytes":0,"num_omap_keys":0,"num_objects_repaired":0},"up":[4,0],"acting":[4,0],"avail_no_missing":[],"object_location_counts":[],"blocked_by":[],"up_primary":4,"acting_primary":4,"purged_snaps":[{"start":"2","length":"1"}]},{"pgid":"2.5","version":"0'0","reported_seq":13,"reported_epoch":17,"state":"active+clean","last_fresh":"2026-03-20T17:57:49.202435+0000","last_change":"2026-03-20T17:57:47.894255+0000","last_active":"2026-03-20T17:57:49.202435+0000","last_peered":"2026-03-20T17:57:49.202435+0000","last_clean":"2026-03-20T17:57:49.202435+0000","last_became_active":"2026-03-20T17:57:47.894163+0000","last_became_peered":"2026-03-20T17:57:47.894163+0000","last_unstale":"2026-03-20T17:57:49.202435+0000","last_undegraded":"2026-03-20T17:57:49.202435+0000","last_fullsized":"2026-03-20T17:57:49.202435+0000","mapping_epoch":15,"log_start":"0'0","ondisk_log_start":"0'0","created":15,"last_epoch_clean":16,"parent":"0.0","parent_split_bits":0,"last_scrub":"0'0","last_scrub_stamp":"2026-03-20T17:57:46.883576+0000","last_deep_scrub":"0'0","last_deep_scrub_stamp":"2026-03-20T17:57:46.883576+0000","last_clean_scrub_stamp":"2026-03-20T17:57:46.883576+0000","objects_scrubbed":0,"log_size":0,"log_dups_size":0,"ondisk_log_size":0,"stats_invalid":false,"dirty_stats_invalid":false,"omap_stats_invalid":false,"hitset_stats_invalid":false,"hitset_bytes_stats_invalid":false,"pin_stats_invalid":false,"manifest_stats_invalid":false,"snaptrimq_len":0,"last_scrub_duration":0,"scrub_schedule":"periodic scrub scheduled @ 2026-03-21T19:37:25.305326+0000","scrub_duration":0,"objects_trimmed":0,"snaptrim_duration":0,"stat_sum":{"num_bytes":0,"num_objects":0,"num_object_clones":0,"num_object_copies":0,"num_objects_missing_on_primary":0,"num_objects_missing":0,"num_objects_degraded":0,"num_objects_misplaced":0,"num_objects_unfound":0,"num_objects_dirty":0,"num_whiteouts":0,"num_read":0,"num_read_kb":0,"num_write":0,"num_write_kb":0,"num_scrub_errors":0,"num_shallow_scrub_errors":0,"num_deep_scrub_errors":0,"num_objects_recovered":0,"num_bytes_recovered":0,"num_keys_recovered":0,"num_objects_omap":0,"num_objects_hit_set_archive":0,"num_bytes_hit_set_archive":0,"num_flush":0,"num_flush_kb":0,"num_evict":0,"num_evict_kb":0,"num_promote":0,"num_flush_mode_high":0,"num_flush_mode_low":0,"num_evict_mode_some":0,"num_evict_mode_full":0,"num_objects_pinned":0,"num_legacy_snapsets":0,"num_large_omap_objects":0,"num_objects_manifest":0,"num_omap_bytes":0,"num_omap_keys":0,"num_objects_repaired":0},"up":[1,4],"acting":[1,4],"avail_no_missing":[],"object_location_counts":[],"blocked_by":[],"up_primary":1,"acting_primary":1,"purged_snaps":[]},{"pgid":"2.4","version":"0'0","reported_seq":13,"reported_epoch":17,"state":"active+clean","last_fresh":"2026-03-20T17:57:49.526389+0000","last_change":"2026-03-20T17:57:47.895163+0000","last_active":"2026-03-20T17:57:49.526389+0000","last_peered":"2026-03-20T17:57:49.526389+0000","last_clean":"2026-03-20T17:57:49.526389+0000","last_became_active":"2026-03-20T17:57:47.894910+0000","last_became_peered":"2026-03-20T17:57:47.894910+0000","last_unstale":"2026-03-20T17:57:49.526389+0000","last_undegraded":"2026-03-20T17:57:49.526389+0000","last_fullsized":"2026-03-20T17:57:49.526389+0000","mapping_epoch":15,"log_start":"0'0","ondisk_log_start":"0'0","created":15,"last_epoch_clean":16,"parent":"0.0","parent_split_bits":0,"last_scrub":"0'0","last_scrub_stamp":"2026-03-20T17:57:46.883576+0000","last_deep_scrub":"0'0","last_deep_scrub_stamp":"2026-03-20T17:57:46.883576+0000","last_clean_scrub_stamp":"2026-03-20T17:57:46.883576+0000","objects_scrubbed":0,"log_size":0,"log_dups_size":0,"ondisk_log_size":0,"stats_invalid":false,"dirty_stats_invalid":false,"omap_stats_invalid":false,"hitset_stats_invalid":false,"hitset_bytes_stats_invalid":false,"pin_stats_invalid":false,"manifest_stats_invalid":false,"snaptrimq_len":0,"last_scrub_duration":0,"scrub_schedule":"periodic scrub scheduled @ 2026-03-22T03:49:16.918512+0000","scrub_duration":0,"objects_trimmed":0,"snaptrim_duration":0,"stat_sum":{"num_bytes":0,"num_objects":0,"num_object_clones":0,"num_object_copies":0,"num_objects_missing_on_primary":0,"num_objects_missing":0,"num_objects_degraded":0,"num_objects_misplaced":0,"num_objects_unfound":0,"num_objects_dirty":0,"num_whiteouts":0,"num_read":0,"num_read_kb":0,"num_write":0,"num_write_kb":0,"num_scrub_errors":0,"num_shallow_scrub_errors":0,"num_deep_scrub_errors":0,"num_objects_recovered":0,"num_bytes_recovered":0,"num_keys_recovered":0,"num_objects_omap":0,"num_objects_hit_set_archive":0,"num_bytes_hit_set_archive":0,"num_flush":0,"num_flush_kb":0,"num_evict":0,"num_evict_kb":0,"num_promote":0,"num_flush_mode_high":0,"num_flush_mode_low":0,"num_evict_mode_some":0,"num_evict_mode_full":0,"num_objects_pinned":0,"num_legacy_snapsets":0,"num_large_omap_objects":0,"num_objects_manifest":0,"num_omap_bytes":0,"num_omap_keys":0,"num_objects_repaired":0},"up":[5,3],"acting":[5,3],"avail_no_missing":[],"object_location_counts":[],"blocked_by":[],"up_primary":5,"acting_primary":5,"purged_snaps":[]},{"pgid":"2.2","version":"18'2","reported_seq":22,"reported_epoch":18,"state":"active+clean","last_fresh":"2026-03-20T17:57:49.905884+0000","last_change":"2026-03-20T17:57:49.904408+0000","last_active":"2026-03-20T17:57:49.905884+0000","last_peered":"2026-03-20T17:57:49.905884+0000","last_clean":"2026-03-20T17:57:49.905884+0000","last_became_active":"2026-03-20T17:57:47.894069+0000","last_became_peered":"2026-03-20T17:57:47.894069+0000","last_unstale":"2026-03-20T17:57:49.905884+0000","last_undegraded":"2026-03-20T17:57:49.905884+0000","last_fullsized":"2026-03-20T17:57:49.905884+0000","mapping_epoch":15,"log_start":"0'0","ondisk_log_start":"0'0","created":15,"last_epoch_clean":16,"parent":"0.0","parent_split_bits":0,"last_scrub":"0'0","last_scrub_stamp":"2026-03-20T17:57:46.883576+0000","last_deep_scrub":"0'0","last_deep_scrub_stamp":"2026-03-20T17:57:46.883576+0000","last_clean_scrub_stamp":"2026-03-20T17:57:46.883576+0000","objects_scrubbed":0,"log_size":2,"log_dups_size":0,"ondisk_log_size":2,"stats_invalid":false,"dirty_stats_invalid":false,"omap_stats_invalid":false,"hitset_stats_invalid":false,"hitset_bytes_stats_invalid":false,"pin_stats_invalid":false,"manifest_stats_invalid":false,"snaptrimq_len":0,"last_scrub_duration":0,"scrub_schedule":"periodic scrub scheduled @ 2026-03-21T21:38:07.100589+0000","scrub_duration":0,"objects_trimmed":0,"snaptrim_duration":0.00022488099999999999,"stat_sum":{"num_bytes":19,"num_objects":1,"num_object_clones":0,"num_object_copies":2,"num_objects_missing_on_primary":0,"num_objects_missing":0,"num_objects_degraded":0,"num_objects_misplaced":0,"num_objects_unfound":0,"num_objects_dirty":1,"num_whiteouts":0,"num_read":0,"num_read_kb":0,"num_write":2,"num_write_kb":2,"num_scrub_errors":0,"num_shallow_scrub_errors":0,"num_deep_scrub_errors":0,"num_objects_recovered":0,"num_bytes_recovered":0,"num_keys_recovered":0,"num_objects_omap":0,"num_objects_hit_set_archive":0,"num_bytes_hit_set_archive":0,"num_flush":0,"num_flush_kb":0,"num_evict":0,"num_evict_kb":0,"num_promote":0,"num_flush_mode_high":0,"num_flush_mode_low":0,"num_evict_mode_some":0,"num_evict_mode_full":0,"num_objects_pinned":0,"num_legacy_snapsets":0,"num_large_omap_objects":0,"num_objects_manifest":0,"num_omap_bytes":0,"num_omap_keys":0,"num_objects_repaired":0},"up":[0,4],"acting":[0,4],"avail_no_missing":[],"object_location_counts":[],"blocked_by":[],"up_primary":0,"acting_primary":0,"purged_snaps":[{"start":"2","length":"1"}]},{"pgid":"2.1","version":"0'0","reported_seq":13,"reported_epoch":17,"state":"active+clean","last_fresh":"2026-03-20T17:57:49.025873+0000","last_change":"2026-03-20T17:57:47.896102+0000","last_active":"2026-03-20T17:57:49.025873+0000","last_peered":"2026-03-20T17:57:49.025873+0000","last_clean":"2026-03-20T17:57:49.025873+0000","last_became_active":"2026-03-20T17:57:47.895906+0000","last_became_peered":"2026-03-20T17:57:47.895906+0000","last_unstale":"2026-03-20T17:57:49.025873+0000","last_undegraded":"2026-03-20T17:57:49.025873+0000","last_fullsized":"2026-03-20T17:57:49.025873+0000","mapping_epoch":15,"log_start":"0'0","ondisk_log_start":"0'0","created":15,"last_epoch_clean":16,"parent":"0.0","parent_split_bits":0,"last_scrub":"0'0","last_scrub_stamp":"2026-03-20T17:57:46.883576+0000","last_deep_scrub":"0'0","last_deep_scrub_stamp":"2026-03-20T17:57:46.883576+0000","last_clean_scrub_stamp":"2026-03-20T17:57:46.883576+0000","objects_scrubbed":0,"log_size":0,"log_dups_size":0,"ondisk_log_size":0,"stats_invalid":false,"dirty_stats_invalid":false,"omap_stats_invalid":false,"hitset_stats_invalid":false,"hitset_bytes_stats_invalid":false,"pin_stats_invalid":false,"manifest_stats_invalid":false,"snaptrimq_len":0,"last_scrub_duration":0,"scrub_schedule":"periodic scrub scheduled @ 2026-03-21T21:09:24.427454+0000","scrub_duration":0,"objects_trimmed":0,"snaptrim_duration":0,"stat_sum":{"num_bytes":0,"num_objects":0,"num_object_clones":0,"num_object_copies":0,"num_objects_missing_on_primary":0,"num_objects_missing":0,"num_objects_degraded":0,"num_objects_misplaced":0,"num_objects_unfound":0,"num_objects_dirty":0,"num_whiteouts":0,"num_read":0,"num_read_kb":0,"num_write":0,"num_write_kb":0,"num_scrub_errors":0,"num_shallow_scrub_errors":0,"num_deep_scrub_errors":0,"num_objects_recovered":0,"num_bytes_recovered":0,"num_keys_recovered":0,"num_objects_omap":0,"num_objects_hit_set_archive":0,"num_bytes_hit_set_archive":0,"num_flush":0,"num_flush_kb":0,"num_evict":0,"num_evict_kb":0,"num_promote":0,"num_flush_mode_high":0,"num_flush_mode_low":0,"num_evict_mode_some":0,"num_evict_mode_full":0,"num_objects_pinned":0,"num_legacy_snapsets":0,"num_large_omap_objects":0,"num_objects_manifest":0,"num_omap_bytes":0,"num_omap_keys":0,"num_objects_repaired":0},"up":[3,5],"acting":[3,5],"avail_no_missing":[],"object_location_counts":[],"blocked_by":[],"up_primary":3,"acting_primary":3,"purged_snaps":[]},{"pgid":"2.0","version":"0'0","reported_seq":13,"reported_epoch":17,"state":"active+clean","last_fresh":"2026-03-20T17:57:49.525624+0000","last_change":"2026-03-20T17:57:47.895021+0000","last_active":"2026-03-20T17:57:49.525624+0000","last_peered":"2026-03-20T17:57:49.525624+0000","last_clean":"2026-03-20T17:57:49.525624+0000","last_became_active":"2026-03-20T17:57:47.894775+0000","last_became_peered":"2026-03-20T17:57:47.894775+0000","last_unstale":"2026-03-20T17:57:49.525624+0000","last_undegraded":"2026-03-20T17:57:49.525624+0000","last_fullsized":"2026-03-20T17:57:49.525624+0000","mapping_epoch":15,"log_start":"0'0","ondisk_log_start":"0'0","created":15,"last_epoch_clean":16,"parent":"0.0","parent_split_bits":0,"last_scrub":"0'0","last_scrub_stamp":"2026-03-20T17:57:46.883576+0000","last_deep_scrub":"0'0","last_deep_scrub_stamp":"2026-03-20T17:57:46.883576+0000","last_clean_scrub_stamp":"2026-03-20T17:57:46.883576+0000","objects_scrubbed":0,"log_size":0,"log_dups_size":0,"ondisk_log_size":0,"stats_invalid":false,"dirty_stats_invalid":false,"omap_stats_invalid":false,"hitset_stats_invalid":false,"hitset_bytes_stats_invalid":false,"pin_stats_invalid":false,"manifest_stats_invalid":false,"snaptrimq_len":0,"last_scrub_duration":0,"scrub_schedule":"periodic scrub scheduled @ 2026-03-22T00:44:03.161568+0000","scrub_duration":0,"objects_trimmed":0,"snaptrim_duration":0,"stat_sum":{"num_bytes":0,"num_objects":0,"num_object_clones":0,"num_object_copies":0,"num_objects_missing_on_primary":0,"num_objects_missing":0,"num_objects_degraded":0,"num_objects_misplaced":0,"num_objects_unfound":0,"num_objects_dirty":0,"num_whiteouts":0,"num_read":0,"num_read_kb":0,"num_write":0,"num_write_kb":0,"num_scrub_errors":0,"num_shallow_scrub_errors":0,"num_deep_scrub_errors":0,"num_objects_recovered":0,"num_bytes_recovered":0,"num_keys_recovered":0,"num_objects_omap":0,"num_objects_hit_set_archive":0,"num_bytes_hit_set_archive":0,"num_flush":0,"num_flush_kb":0,"num_evict":0,"num_evict_kb":0,"num_promote":0,"num_flush_mode_high":0,"num_flush_mode_low":0,"num_evict_mode_some":0,"num_evict_mode_full":0,"num_objects_pinned":0,"num_legacy_snapsets":0,"num_large_omap_objects":0,"num_objects_manifest":0,"num_omap_bytes":0,"num_omap_keys":0,"num_objects_repaired":0},"up":[2,5],"acting":[2,5],"avail_no_missing":[],"object_location_counts":[],"blocked_by":[],"up_primary":2,"acting_primary":2,"purged_snaps":[]},{"pgid":"2.3","version":"16'1","reported_seq":14,"reported_epoch":17,"state":"active+clean","last_fresh":"2026-03-20T17:57:49.202453+0000","last_change":"2026-03-20T17:57:47.896525+0000","last_active":"2026-03-20T17:57:49.202453+0000","last_peered":"2026-03-20T17:57:49.202453+0000","last_clean":"2026-03-20T17:57:49.202453+0000","last_became_active":"2026-03-20T17:57:47.896163+0000","last_became_peered":"2026-03-20T17:57:47.896163+0000","last_unstale":"2026-03-20T17:57:49.202453+0000","last_undegraded":"2026-03-20T17:57:49.202453+0000","last_fullsized":"2026-03-20T17:57:49.202453+0000","mapping_epoch":15,"log_start":"0'0","ondisk_log_start":"0'0","created":15,"last_epoch_clean":16,"parent":"0.0","parent_split_bits":0,"last_scrub":"0'0","last_scrub_stamp":"2026-03-20T17:57:46.883576+0000","last_deep_scrub":"0'0","last_deep_scrub_stamp":"2026-03-20T17:57:46.883576+0000","last_clean_scrub_stamp":"2026-03-20T17:57:46.883576+0000","objects_scrubbed":0,"log_size":1,"log_dups_size":0,"ondisk_log_size":1,"stats_invalid":false,"dirty_stats_invalid":false,"omap_stats_invalid":false,"hitset_stats_invalid":false,"hitset_bytes_stats_invalid":false,"pin_stats_invalid":false,"manifest_stats_invalid":false,"snaptrimq_len":0,"last_scrub_duration":0,"scrub_schedule":"periodic scrub scheduled @ 2026-03-21T22:20:00.347952+0000","scrub_duration":0,"objects_trimmed":0,"snaptrim_duration":0,"stat_sum":{"num_bytes":0,"num_objects":1,"num_object_clones":0,"num_object_copies":2,"num_objects_missing_on_primary":0,"num_objects_missing":0,"num_objects_degraded":0,"num_objects_misplaced":0,"num_objects_unfound":0,"num_objects_dirty":1,"num_whiteouts":0,"num_read":0,"num_read_kb":0,"num_write":0,"num_write_kb":0,"num_scrub_errors":0,"num_shallow_scrub_errors":0,"num_deep_scrub_errors":0,"num_objects_recovered":0,"num_bytes_recovered":0,"num_keys_recovered":0,"num_objects_omap":0,"num_objects_hit_set_archive":0,"num_bytes_hit_set_archive":0,"num_flush":0,"num_flush_kb":0,"num_evict":0,"num_evict_kb":0,"num_promote":0,"num_flush_mode_high":0,"num_flush_mode_low":0,"num_evict_mode_some":0,"num_evict_mode_full":0,"num_objects_pinned":0,"num_legacy_snapsets":0,"num_large_omap_objects":0,"num_objects_manifest":0,"num_omap_bytes":0,"num_omap_keys":0,"num_objects_repaired":0},"up":[1,2],"acting":[1,2],"avail_no_missing":[],"object_location_counts":[],"blocked_by":[],"up_primary":1,"acting_primary":1,"purged_snaps":[]},{"pgid":"1.0","version":"13'32","reported_seq":63,"reported_epoch":17,"state":"active+clean","last_fresh":"2026-03-20T17:57:49.202372+0000","last_change":"2026-03-20T17:57:44.003553+0000","last_active":"2026-03-20T17:57:49.202372+0000","last_peered":"2026-03-20T17:57:49.202372+0000","last_clean":"2026-03-20T17:57:49.202372+0000","last_became_active":"2026-03-20T17:57:44.003422+0000","last_became_peered":"2026-03-20T17:57:44.003422+0000","last_unstale":"2026-03-20T17:57:49.202372+0000","last_undegraded":"2026-03-20T17:57:49.202372+0000","last_fullsized":"2026-03-20T17:57:49.202372+0000","mapping_epoch":12,"log_start":"0'0","ondisk_log_start":"0'0","created":12,"last_epoch_clean":13,"parent":"0.0","parent_split_bits":0,"last_scrub":"0'0","last_scrub_stamp":"2026-03-20T17:57:42.849727+0000","last_deep_scrub":"0'0","last_deep_scrub_stamp":"2026-03-20T17:57:42.849727+0000","last_clean_scrub_stamp":"2026-03-20T17:57:42.849727+0000","objects_scrubbed":0,"log_size":32,"log_dups_size":0,"ondisk_log_size":32,"stats_invalid":false,"dirty_stats_invalid":false,"omap_stats_invalid":false,"hitset_stats_invalid":false,"hitset_bytes_stats_invalid":false,"pin_stats_invalid":false,"manifest_stats_invalid":false,"snaptrimq_len":0,"last_scrub_duration":0,"scrub_schedule":"periodic scrub scheduled @ 2026-03-22T02:36:03.668295+0000","scrub_duration":0,"objects_trimmed":0,"snaptrim_duration":0,"stat_sum":{"num_bytes":459280,"num_objects":2,"num_object_clones":0,"num_object_copies":4,"num_objects_missing_on_primary":0,"num_objects_missing":0,"num_objects_degraded":0,"num_objects_misplaced":0,"num_objects_unfound":0,"num_objects_dirty":2,"num_whiteouts":0,"num_read":46,"num_read_kb":37,"num_write":57,"num_write_kb":584,"num_scrub_errors":0,"num_shallow_scrub_errors":0,"num_deep_scrub_errors":0,"num_objects_recovered":0,"num_bytes_recovered":0,"num_keys_recovered":0,"num_objects_omap":0,"num_objects_hit_set_archive":0,"num_bytes_hit_set_archive":0,"num_flush":0,"num_flush_kb":0,"num_evict":0,"num_evict_kb":0,"num_promote":0,"num_flush_mode_high":0,"num_flush_mode_low":0,"num_evict_mode_some":0,"num_evict_mode_full":0,"num_objects_pinned":0,"num_legacy_snapsets":0,"num_large_omap_objects":0,"num_objects_manifest":0,"num_omap_bytes":0,"num_omap_keys":0,"num_objects_repaired":0},"up":[1,5],"acting":[1,5],"avail_no_missing":[],"object_location_counts":[],"blocked_by":[],"up_primary":1,"acting_primary":1,"purged_snaps":[]}],"pool_stats":[{"poolid":2,"num_pg":8,"stat_sum":{"num_bytes":19,"num_objects":2,"num_object_clones":0,"num_object_copies":4,"num_objects_missing_on_primary":0,"num_objects_missing":0,"num_objects_degraded":0,"num_objects_misplaced":0,"num_objects_unfound":0,"num_objects_dirty":2,"num_whiteouts":0,"num_read":0,"num_read_kb":0,"num_write":2,"num_write_kb":2,"num_scrub_errors":0,"num_shallow_scrub_errors":0,"num_deep_scrub_errors":0,"num_objects_recovered":0,"num_bytes_recovered":0,"num_keys_recovered":0,"num_objects_omap":0,"num_objects_hit_set_archive":0,"num_bytes_hit_set_archive":0,"num_flush":0,"num_flush_kb":0,"num_evict":0,"num_evict_kb":0,"num_promote":0,"num_flush_mode_high":0,"num_flush_mode_low":0,"num_evict_mode_some":0,"num_evict_mode_full":0,"num_objects_pinned":0,"num_legacy_snapsets":0,"num_large_omap_objects":0,"num_objects_manifest":0,"num_omap_bytes":0,"num_omap_keys":0,"num_objects_repaired":0},"store_stats":{"total":0,"available":0,"internally_reserved":0,"allocated":8192,"data_stored":38,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":0},"log_size":3,"ondisk_log_size":3,"up":16,"acting":16,"num_store_stats":6},{"poolid":1,"num_pg":1,"stat_sum":{"num_bytes":459280,"num_objects":2,"num_object_clones":0,"num_object_copies":4,"num_objects_missing_on_primary":0,"num_objects_missing":0,"num_objects_degraded":0,"num_objects_misplaced":0,"num_objects_unfound":0,"num_objects_dirty":2,"num_whiteouts":0,"num_read":46,"num_read_kb":37,"num_write":57,"num_write_kb":584,"num_scrub_errors":0,"num_shallow_scrub_errors":0,"num_deep_scrub_errors":0,"num_objects_recovered":0,"num_bytes_recovered":0,"num_keys_recovered":0,"num_objects_omap":0,"num_objects_hit_set_archive":0,"num_bytes_hit_set_archive":0,"num_flush":0,"num_flush_kb":0,"num_evict":0,"num_evict_kb":0,"num_promote":0,"num_flush_mode_high":0,"num_flush_mode_low":0,"num_evict_mode_some":0,"num_evict_mode_full":0,"num_objects_pinned":0,"num_legacy_snapsets":0,"num_large_omap_objects":0,"num_objects_manifest":0,"num_omap_bytes":0,"num_omap_keys":0,"num_objects_repaired":0},"store_stats":{"total":0,"available":0,"internally_reserved":0,"allocated":925696,"data_stored":918560,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":0},"log_size":32,"ondisk_log_size":32,"up":2,"acting":2,"num_store_stats":2}],"osd_stats":[{"osd":5,"up_from":11,"seq":47244640259,"num_pgs":2,"num_osds":1,"num_per_pool_osds":1,"num_per_pool_omap_osds":1,"kb":104857600,"kb_used":27472,"kb_used_data":608,"kb_used_omap":7,"kb_used_meta":26808,"kb_avail":104830128,"statfs":{"total":107374182400,"available":107346051072,"internally_reserved":0,"allocated":622592,"data_stored":510892,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":7471,"internal_metadata":27452113},"hb_peers":[0,1,2,3,4],"snap_trim_queue_len":0,"num_snap_trimming":0,"num_shards_repaired":0,"op_queue_age_hist":{"histogram":[],"upper_bound":1},"perf_stat":{"commit_latency_ms":0,"apply_latency_ms":0,"commit_latency_ns":0,"apply_latency_ns":0},"alerts":[]},{"osd":4,"up_from":11,"seq":47244640259,"num_pgs":3,"num_osds":1,"num_per_pool_osds":1,"num_per_pool_omap_osds":1,"kb":104857600,"kb_used":27020,"kb_used_data":172,"kb_used_omap":7,"kb_used_meta":26808,"kb_avail":104830580,"statfs":{"total":107374182400,"available":107346513920,"internally_reserved":0,"allocated":176128,"data_stored":56583,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":8121,"internal_metadata":27451463},"hb_peers":[0,1,2,3,5],"snap_trim_queue_len":0,"num_snap_trimming":0,"num_shards_repaired":0,"op_queue_age_hist":{"histogram":[],"upper_bound":1},"perf_stat":{"commit_latency_ms":0,"apply_latency_ms":0,"commit_latency_ns":0,"apply_latency_ns":0},"alerts":[]},{"osd":3,"up_from":11,"seq":47244640260,"num_pgs":2,"num_osds":1,"num_per_pool_osds":1,"num_per_pool_omap_osds":1,"kb":104857600,"kb_used":27012,"kb_used_data":156,"kb_used_omap":6,"kb_used_meta":26809,"kb_avail":104830588,"statfs":{"total":107374182400,"available":107346522112,"internally_reserved":0,"allocated":159744,"data_stored":51612,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":6819,"internal_metadata":27452765},"hb_peers":[0,1,2,4,5],"snap_trim_queue_len":0,"num_snap_trimming":0,"num_shards_repaired":0,"op_queue_age_hist":{"histogram":[],"upper_bound":1},"perf_stat":{"commit_latency_ms":0,"apply_latency_ms":0,"commit_latency_ns":0,"apply_latency_ns":0},"alerts":[]},{"osd":2,"up_from":11,"seq":47244640259,"num_pgs":2,"num_osds":1,"num_per_pool_osds":1,"num_per_pool_omap_osds":1,"kb":104857600,"kb_used":27012,"kb_used_data":156,"kb_used_omap":7,"kb_used_meta":26808,"kb_avail":104830588,"statfs":{"total":107374182400,"available":107346522112,"internally_reserved":0,"allocated":159744,"data_stored":51612,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":7471,"internal_metadata":27452113},"hb_peers":[0,1,3,4,5],"snap_trim_queue_len":0,"num_snap_trimming":0,"num_shards_repaired":0,"op_queue_age_hist":{"histogram":[],"upper_bound":1},"perf_stat":{"commit_latency_ms":1,"apply_latency_ms":1,"commit_latency_ns":1000000,"apply_latency_ns":1000000},"alerts":[]},{"osd":1,"up_from":11,"seq":47244640259,"num_pgs":4,"num_osds":1,"num_per_pool_osds":1,"num_per_pool_omap_osds":1,"kb":104857600,"kb_used":27464,"kb_used_data":608,"kb_used_omap":7,"kb_used_meta":26808,"kb_avail":104830136,"statfs":{"total":107374182400,"available":107346059264,"internally_reserved":0,"allocated":622592,"data_stored":510892,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":8121,"internal_metadata":27451463},"hb_peers":[0,2,3,4,5],"snap_trim_queue_len":0,"num_snap_trimming":0,"num_shards_repaired":0,"op_queue_age_hist":{"histogram":[],"upper_bound":1},"perf_stat":{"commit_latency_ms":0,"apply_latency_ms":0,"commit_latency_ns":0,"apply_latency_ns":0},"alerts":[]},{"osd":0,"up_from":11,"seq":47244640260,"num_pgs":3,"num_osds":1,"num_per_pool_osds":1,"num_per_pool_omap_osds":1,"kb":104857600,"kb_used":27020,"kb_used_data":172,"kb_used_omap":7,"kb_used_meta":26808,"kb_avail":104830580,"statfs":{"total":107374182400,"available":107346513920,"internally_reserved":0,"allocated":176128,"data_stored":56583,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":8121,"internal_metadata":27451463},"hb_peers":[1,2,3,4,5],"snap_trim_queue_len":0,"num_snap_trimming":0,"num_shards_repaired":0,"op_queue_age_hist":{"histogram":[],"upper_bound":1},"perf_stat":{"commit_latency_ms":0,"apply_latency_ms":0,"commit_latency_ns":0,"apply_latency_ns":0},"alerts":[]}],"pool_statfs":[{"poolid":1,"osd":1,"total":0,"available":0,"internally_reserved":0,"allocated":462848,"data_stored":459280,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":0},{"poolid":1,"osd":5,"total":0,"available":0,"internally_reserved":0,"allocated":462848,"data_stored":459280,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":0},{"poolid":2,"osd":0,"total":0,"available":0,"internally_reserved":0,"allocated":4096,"data_stored":19,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":0},{"poolid":2,"osd":1,"total":0,"available":0,"internally_reserved":0,"allocated":0,"data_stored":0,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":0},{"poolid":2,"osd":2,"total":0,"available":0,"internally_reserved":0,"allocated":0,"data_stored":0,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":0},{"poolid":2,"osd":3,"total":0,"available":0,"internally_reserved":0,"allocated":0,"data_stored":0,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":0},{"poolid":2,"osd":4,"total":0,"available":0,"internally_reserved":0,"allocated":4096,"data_stored":19,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":0},{"poolid":2,"osd":5,"total":0,"available":0,"internally_reserved":0,"allocated":0,"data_stored":0,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":0}]}} 2026-03-20T17:57:52.769 DEBUG:teuthology.orchestra.run.vm06:> sudo adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage timeout 120 ceph --cluster c2 pg dump --format=json 2026-03-20T17:57:52.941 INFO:teuthology.orchestra.run.vm06.stdout: 2026-03-20T17:57:52.941 INFO:teuthology.orchestra.run.vm06.stderr:dumped all 2026-03-20T17:57:52.956 INFO:teuthology.orchestra.run.vm06.stdout:{"pg_ready":true,"pg_map":{"version":17,"stamp":"2026-03-20T17:57:51.844997+0000","last_osdmap_epoch":0,"last_pg_scan":0,"pg_stats_sum":{"stat_sum":{"num_bytes":459299,"num_objects":4,"num_object_clones":0,"num_object_copies":8,"num_objects_missing_on_primary":0,"num_objects_missing":0,"num_objects_degraded":0,"num_objects_misplaced":0,"num_objects_unfound":0,"num_objects_dirty":4,"num_whiteouts":0,"num_read":46,"num_read_kb":37,"num_write":59,"num_write_kb":586,"num_scrub_errors":0,"num_shallow_scrub_errors":0,"num_deep_scrub_errors":0,"num_objects_recovered":0,"num_bytes_recovered":0,"num_keys_recovered":0,"num_objects_omap":0,"num_objects_hit_set_archive":0,"num_bytes_hit_set_archive":0,"num_flush":0,"num_flush_kb":0,"num_evict":0,"num_evict_kb":0,"num_promote":0,"num_flush_mode_high":0,"num_flush_mode_low":0,"num_evict_mode_some":0,"num_evict_mode_full":0,"num_objects_pinned":0,"num_legacy_snapsets":0,"num_large_omap_objects":0,"num_objects_manifest":0,"num_omap_bytes":0,"num_omap_keys":0,"num_objects_repaired":0},"store_stats":{"total":0,"available":0,"internally_reserved":0,"allocated":0,"data_stored":0,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":0},"log_size":35,"ondisk_log_size":35,"up":18,"acting":18,"num_store_stats":0},"osd_stats_sum":{"up_from":0,"seq":0,"num_pgs":16,"num_osds":6,"num_per_pool_osds":6,"num_per_pool_omap_osds":6,"kb":629145600,"kb_used":163000,"kb_used_data":1872,"kb_used_omap":45,"kb_used_meta":160850,"kb_avail":628982600,"statfs":{"total":644245094400,"available":644078182400,"internally_reserved":0,"allocated":1916928,"data_stored":1238174,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":46124,"internal_metadata":164711380},"hb_peers":[],"snap_trim_queue_len":0,"num_snap_trimming":0,"num_shards_repaired":0,"op_queue_age_hist":{"histogram":[],"upper_bound":1},"perf_stat":{"commit_latency_ms":1,"apply_latency_ms":1,"commit_latency_ns":1000000,"apply_latency_ns":1000000},"alerts":[],"network_ping_times":[]},"pg_stats_delta":{"stat_sum":{"num_bytes":19,"num_objects":2,"num_object_clones":0,"num_object_copies":4,"num_objects_missing_on_primary":0,"num_objects_missing":0,"num_objects_degraded":0,"num_objects_misplaced":0,"num_objects_unfound":0,"num_objects_dirty":2,"num_whiteouts":0,"num_read":0,"num_read_kb":0,"num_write":2,"num_write_kb":2,"num_scrub_errors":0,"num_shallow_scrub_errors":0,"num_deep_scrub_errors":0,"num_objects_recovered":0,"num_bytes_recovered":0,"num_keys_recovered":0,"num_objects_omap":0,"num_objects_hit_set_archive":0,"num_bytes_hit_set_archive":0,"num_flush":0,"num_flush_kb":0,"num_evict":0,"num_evict_kb":0,"num_promote":0,"num_flush_mode_high":0,"num_flush_mode_low":0,"num_evict_mode_some":0,"num_evict_mode_full":0,"num_objects_pinned":0,"num_legacy_snapsets":0,"num_large_omap_objects":0,"num_objects_manifest":0,"num_omap_bytes":0,"num_omap_keys":0,"num_objects_repaired":0},"store_stats":{"total":0,"available":0,"internally_reserved":0,"allocated":0,"data_stored":0,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":0},"log_size":0,"ondisk_log_size":0,"up":0,"acting":0,"num_store_stats":0,"stamp_delta":"4.956044"},"pg_stats":[{"pgid":"2.7","version":"0'0","reported_seq":13,"reported_epoch":17,"state":"active+clean","last_fresh":"2026-03-20T17:57:49.202484+0000","last_change":"2026-03-20T17:57:47.894639+0000","last_active":"2026-03-20T17:57:49.202484+0000","last_peered":"2026-03-20T17:57:49.202484+0000","last_clean":"2026-03-20T17:57:49.202484+0000","last_became_active":"2026-03-20T17:57:47.894483+0000","last_became_peered":"2026-03-20T17:57:47.894483+0000","last_unstale":"2026-03-20T17:57:49.202484+0000","last_undegraded":"2026-03-20T17:57:49.202484+0000","last_fullsized":"2026-03-20T17:57:49.202484+0000","mapping_epoch":15,"log_start":"0'0","ondisk_log_start":"0'0","created":15,"last_epoch_clean":16,"parent":"0.0","parent_split_bits":0,"last_scrub":"0'0","last_scrub_stamp":"2026-03-20T17:57:46.883576+0000","last_deep_scrub":"0'0","last_deep_scrub_stamp":"2026-03-20T17:57:46.883576+0000","last_clean_scrub_stamp":"2026-03-20T17:57:46.883576+0000","objects_scrubbed":0,"log_size":0,"log_dups_size":0,"ondisk_log_size":0,"stats_invalid":false,"dirty_stats_invalid":false,"omap_stats_invalid":false,"hitset_stats_invalid":false,"hitset_bytes_stats_invalid":false,"pin_stats_invalid":false,"manifest_stats_invalid":false,"snaptrimq_len":0,"last_scrub_duration":0,"scrub_schedule":"periodic scrub scheduled @ 2026-03-21T22:36:28.128151+0000","scrub_duration":0,"objects_trimmed":0,"snaptrim_duration":0,"stat_sum":{"num_bytes":0,"num_objects":0,"num_object_clones":0,"num_object_copies":0,"num_objects_missing_on_primary":0,"num_objects_missing":0,"num_objects_degraded":0,"num_objects_misplaced":0,"num_objects_unfound":0,"num_objects_dirty":0,"num_whiteouts":0,"num_read":0,"num_read_kb":0,"num_write":0,"num_write_kb":0,"num_scrub_errors":0,"num_shallow_scrub_errors":0,"num_deep_scrub_errors":0,"num_objects_recovered":0,"num_bytes_recovered":0,"num_keys_recovered":0,"num_objects_omap":0,"num_objects_hit_set_archive":0,"num_bytes_hit_set_archive":0,"num_flush":0,"num_flush_kb":0,"num_evict":0,"num_evict_kb":0,"num_promote":0,"num_flush_mode_high":0,"num_flush_mode_low":0,"num_evict_mode_some":0,"num_evict_mode_full":0,"num_objects_pinned":0,"num_legacy_snapsets":0,"num_large_omap_objects":0,"num_objects_manifest":0,"num_omap_bytes":0,"num_omap_keys":0,"num_objects_repaired":0},"up":[1,0],"acting":[1,0],"avail_no_missing":[],"object_location_counts":[],"blocked_by":[],"up_primary":1,"acting_primary":1,"purged_snaps":[]},{"pgid":"2.6","version":"0'0","reported_seq":20,"reported_epoch":18,"state":"active+clean","last_fresh":"2026-03-20T17:57:49.907815+0000","last_change":"2026-03-20T17:57:49.907815+0000","last_active":"2026-03-20T17:57:49.907815+0000","last_peered":"2026-03-20T17:57:49.907815+0000","last_clean":"2026-03-20T17:57:49.907815+0000","last_became_active":"2026-03-20T17:57:47.895290+0000","last_became_peered":"2026-03-20T17:57:47.895290+0000","last_unstale":"2026-03-20T17:57:49.907815+0000","last_undegraded":"2026-03-20T17:57:49.907815+0000","last_fullsized":"2026-03-20T17:57:49.907815+0000","mapping_epoch":15,"log_start":"0'0","ondisk_log_start":"0'0","created":15,"last_epoch_clean":16,"parent":"0.0","parent_split_bits":0,"last_scrub":"0'0","last_scrub_stamp":"2026-03-20T17:57:46.883576+0000","last_deep_scrub":"0'0","last_deep_scrub_stamp":"2026-03-20T17:57:46.883576+0000","last_clean_scrub_stamp":"2026-03-20T17:57:46.883576+0000","objects_scrubbed":0,"log_size":0,"log_dups_size":0,"ondisk_log_size":0,"stats_invalid":false,"dirty_stats_invalid":false,"omap_stats_invalid":false,"hitset_stats_invalid":false,"hitset_bytes_stats_invalid":false,"pin_stats_invalid":false,"manifest_stats_invalid":false,"snaptrimq_len":0,"last_scrub_duration":0,"scrub_schedule":"periodic scrub scheduled @ 2026-03-21T22:10:31.042370+0000","scrub_duration":0,"objects_trimmed":0,"snaptrim_duration":0.000290434,"stat_sum":{"num_bytes":0,"num_objects":0,"num_object_clones":0,"num_object_copies":0,"num_objects_missing_on_primary":0,"num_objects_missing":0,"num_objects_degraded":0,"num_objects_misplaced":0,"num_objects_unfound":0,"num_objects_dirty":0,"num_whiteouts":0,"num_read":0,"num_read_kb":0,"num_write":0,"num_write_kb":0,"num_scrub_errors":0,"num_shallow_scrub_errors":0,"num_deep_scrub_errors":0,"num_objects_recovered":0,"num_bytes_recovered":0,"num_keys_recovered":0,"num_objects_omap":0,"num_objects_hit_set_archive":0,"num_bytes_hit_set_archive":0,"num_flush":0,"num_flush_kb":0,"num_evict":0,"num_evict_kb":0,"num_promote":0,"num_flush_mode_high":0,"num_flush_mode_low":0,"num_evict_mode_some":0,"num_evict_mode_full":0,"num_objects_pinned":0,"num_legacy_snapsets":0,"num_large_omap_objects":0,"num_objects_manifest":0,"num_omap_bytes":0,"num_omap_keys":0,"num_objects_repaired":0},"up":[4,0],"acting":[4,0],"avail_no_missing":[],"object_location_counts":[],"blocked_by":[],"up_primary":4,"acting_primary":4,"purged_snaps":[{"start":"2","length":"1"}]},{"pgid":"2.5","version":"0'0","reported_seq":13,"reported_epoch":17,"state":"active+clean","last_fresh":"2026-03-20T17:57:49.202435+0000","last_change":"2026-03-20T17:57:47.894255+0000","last_active":"2026-03-20T17:57:49.202435+0000","last_peered":"2026-03-20T17:57:49.202435+0000","last_clean":"2026-03-20T17:57:49.202435+0000","last_became_active":"2026-03-20T17:57:47.894163+0000","last_became_peered":"2026-03-20T17:57:47.894163+0000","last_unstale":"2026-03-20T17:57:49.202435+0000","last_undegraded":"2026-03-20T17:57:49.202435+0000","last_fullsized":"2026-03-20T17:57:49.202435+0000","mapping_epoch":15,"log_start":"0'0","ondisk_log_start":"0'0","created":15,"last_epoch_clean":16,"parent":"0.0","parent_split_bits":0,"last_scrub":"0'0","last_scrub_stamp":"2026-03-20T17:57:46.883576+0000","last_deep_scrub":"0'0","last_deep_scrub_stamp":"2026-03-20T17:57:46.883576+0000","last_clean_scrub_stamp":"2026-03-20T17:57:46.883576+0000","objects_scrubbed":0,"log_size":0,"log_dups_size":0,"ondisk_log_size":0,"stats_invalid":false,"dirty_stats_invalid":false,"omap_stats_invalid":false,"hitset_stats_invalid":false,"hitset_bytes_stats_invalid":false,"pin_stats_invalid":false,"manifest_stats_invalid":false,"snaptrimq_len":0,"last_scrub_duration":0,"scrub_schedule":"periodic scrub scheduled @ 2026-03-21T19:37:25.305326+0000","scrub_duration":0,"objects_trimmed":0,"snaptrim_duration":0,"stat_sum":{"num_bytes":0,"num_objects":0,"num_object_clones":0,"num_object_copies":0,"num_objects_missing_on_primary":0,"num_objects_missing":0,"num_objects_degraded":0,"num_objects_misplaced":0,"num_objects_unfound":0,"num_objects_dirty":0,"num_whiteouts":0,"num_read":0,"num_read_kb":0,"num_write":0,"num_write_kb":0,"num_scrub_errors":0,"num_shallow_scrub_errors":0,"num_deep_scrub_errors":0,"num_objects_recovered":0,"num_bytes_recovered":0,"num_keys_recovered":0,"num_objects_omap":0,"num_objects_hit_set_archive":0,"num_bytes_hit_set_archive":0,"num_flush":0,"num_flush_kb":0,"num_evict":0,"num_evict_kb":0,"num_promote":0,"num_flush_mode_high":0,"num_flush_mode_low":0,"num_evict_mode_some":0,"num_evict_mode_full":0,"num_objects_pinned":0,"num_legacy_snapsets":0,"num_large_omap_objects":0,"num_objects_manifest":0,"num_omap_bytes":0,"num_omap_keys":0,"num_objects_repaired":0},"up":[1,4],"acting":[1,4],"avail_no_missing":[],"object_location_counts":[],"blocked_by":[],"up_primary":1,"acting_primary":1,"purged_snaps":[]},{"pgid":"2.4","version":"0'0","reported_seq":13,"reported_epoch":17,"state":"active+clean","last_fresh":"2026-03-20T17:57:49.526389+0000","last_change":"2026-03-20T17:57:47.895163+0000","last_active":"2026-03-20T17:57:49.526389+0000","last_peered":"2026-03-20T17:57:49.526389+0000","last_clean":"2026-03-20T17:57:49.526389+0000","last_became_active":"2026-03-20T17:57:47.894910+0000","last_became_peered":"2026-03-20T17:57:47.894910+0000","last_unstale":"2026-03-20T17:57:49.526389+0000","last_undegraded":"2026-03-20T17:57:49.526389+0000","last_fullsized":"2026-03-20T17:57:49.526389+0000","mapping_epoch":15,"log_start":"0'0","ondisk_log_start":"0'0","created":15,"last_epoch_clean":16,"parent":"0.0","parent_split_bits":0,"last_scrub":"0'0","last_scrub_stamp":"2026-03-20T17:57:46.883576+0000","last_deep_scrub":"0'0","last_deep_scrub_stamp":"2026-03-20T17:57:46.883576+0000","last_clean_scrub_stamp":"2026-03-20T17:57:46.883576+0000","objects_scrubbed":0,"log_size":0,"log_dups_size":0,"ondisk_log_size":0,"stats_invalid":false,"dirty_stats_invalid":false,"omap_stats_invalid":false,"hitset_stats_invalid":false,"hitset_bytes_stats_invalid":false,"pin_stats_invalid":false,"manifest_stats_invalid":false,"snaptrimq_len":0,"last_scrub_duration":0,"scrub_schedule":"periodic scrub scheduled @ 2026-03-22T03:49:16.918512+0000","scrub_duration":0,"objects_trimmed":0,"snaptrim_duration":0,"stat_sum":{"num_bytes":0,"num_objects":0,"num_object_clones":0,"num_object_copies":0,"num_objects_missing_on_primary":0,"num_objects_missing":0,"num_objects_degraded":0,"num_objects_misplaced":0,"num_objects_unfound":0,"num_objects_dirty":0,"num_whiteouts":0,"num_read":0,"num_read_kb":0,"num_write":0,"num_write_kb":0,"num_scrub_errors":0,"num_shallow_scrub_errors":0,"num_deep_scrub_errors":0,"num_objects_recovered":0,"num_bytes_recovered":0,"num_keys_recovered":0,"num_objects_omap":0,"num_objects_hit_set_archive":0,"num_bytes_hit_set_archive":0,"num_flush":0,"num_flush_kb":0,"num_evict":0,"num_evict_kb":0,"num_promote":0,"num_flush_mode_high":0,"num_flush_mode_low":0,"num_evict_mode_some":0,"num_evict_mode_full":0,"num_objects_pinned":0,"num_legacy_snapsets":0,"num_large_omap_objects":0,"num_objects_manifest":0,"num_omap_bytes":0,"num_omap_keys":0,"num_objects_repaired":0},"up":[5,3],"acting":[5,3],"avail_no_missing":[],"object_location_counts":[],"blocked_by":[],"up_primary":5,"acting_primary":5,"purged_snaps":[]},{"pgid":"2.2","version":"18'2","reported_seq":22,"reported_epoch":18,"state":"active+clean","last_fresh":"2026-03-20T17:57:49.905884+0000","last_change":"2026-03-20T17:57:49.904408+0000","last_active":"2026-03-20T17:57:49.905884+0000","last_peered":"2026-03-20T17:57:49.905884+0000","last_clean":"2026-03-20T17:57:49.905884+0000","last_became_active":"2026-03-20T17:57:47.894069+0000","last_became_peered":"2026-03-20T17:57:47.894069+0000","last_unstale":"2026-03-20T17:57:49.905884+0000","last_undegraded":"2026-03-20T17:57:49.905884+0000","last_fullsized":"2026-03-20T17:57:49.905884+0000","mapping_epoch":15,"log_start":"0'0","ondisk_log_start":"0'0","created":15,"last_epoch_clean":16,"parent":"0.0","parent_split_bits":0,"last_scrub":"0'0","last_scrub_stamp":"2026-03-20T17:57:46.883576+0000","last_deep_scrub":"0'0","last_deep_scrub_stamp":"2026-03-20T17:57:46.883576+0000","last_clean_scrub_stamp":"2026-03-20T17:57:46.883576+0000","objects_scrubbed":0,"log_size":2,"log_dups_size":0,"ondisk_log_size":2,"stats_invalid":false,"dirty_stats_invalid":false,"omap_stats_invalid":false,"hitset_stats_invalid":false,"hitset_bytes_stats_invalid":false,"pin_stats_invalid":false,"manifest_stats_invalid":false,"snaptrimq_len":0,"last_scrub_duration":0,"scrub_schedule":"periodic scrub scheduled @ 2026-03-21T21:38:07.100589+0000","scrub_duration":0,"objects_trimmed":0,"snaptrim_duration":0.00022488099999999999,"stat_sum":{"num_bytes":19,"num_objects":1,"num_object_clones":0,"num_object_copies":2,"num_objects_missing_on_primary":0,"num_objects_missing":0,"num_objects_degraded":0,"num_objects_misplaced":0,"num_objects_unfound":0,"num_objects_dirty":1,"num_whiteouts":0,"num_read":0,"num_read_kb":0,"num_write":2,"num_write_kb":2,"num_scrub_errors":0,"num_shallow_scrub_errors":0,"num_deep_scrub_errors":0,"num_objects_recovered":0,"num_bytes_recovered":0,"num_keys_recovered":0,"num_objects_omap":0,"num_objects_hit_set_archive":0,"num_bytes_hit_set_archive":0,"num_flush":0,"num_flush_kb":0,"num_evict":0,"num_evict_kb":0,"num_promote":0,"num_flush_mode_high":0,"num_flush_mode_low":0,"num_evict_mode_some":0,"num_evict_mode_full":0,"num_objects_pinned":0,"num_legacy_snapsets":0,"num_large_omap_objects":0,"num_objects_manifest":0,"num_omap_bytes":0,"num_omap_keys":0,"num_objects_repaired":0},"up":[0,4],"acting":[0,4],"avail_no_missing":[],"object_location_counts":[],"blocked_by":[],"up_primary":0,"acting_primary":0,"purged_snaps":[{"start":"2","length":"1"}]},{"pgid":"2.1","version":"0'0","reported_seq":13,"reported_epoch":17,"state":"active+clean","last_fresh":"2026-03-20T17:57:49.025873+0000","last_change":"2026-03-20T17:57:47.896102+0000","last_active":"2026-03-20T17:57:49.025873+0000","last_peered":"2026-03-20T17:57:49.025873+0000","last_clean":"2026-03-20T17:57:49.025873+0000","last_became_active":"2026-03-20T17:57:47.895906+0000","last_became_peered":"2026-03-20T17:57:47.895906+0000","last_unstale":"2026-03-20T17:57:49.025873+0000","last_undegraded":"2026-03-20T17:57:49.025873+0000","last_fullsized":"2026-03-20T17:57:49.025873+0000","mapping_epoch":15,"log_start":"0'0","ondisk_log_start":"0'0","created":15,"last_epoch_clean":16,"parent":"0.0","parent_split_bits":0,"last_scrub":"0'0","last_scrub_stamp":"2026-03-20T17:57:46.883576+0000","last_deep_scrub":"0'0","last_deep_scrub_stamp":"2026-03-20T17:57:46.883576+0000","last_clean_scrub_stamp":"2026-03-20T17:57:46.883576+0000","objects_scrubbed":0,"log_size":0,"log_dups_size":0,"ondisk_log_size":0,"stats_invalid":false,"dirty_stats_invalid":false,"omap_stats_invalid":false,"hitset_stats_invalid":false,"hitset_bytes_stats_invalid":false,"pin_stats_invalid":false,"manifest_stats_invalid":false,"snaptrimq_len":0,"last_scrub_duration":0,"scrub_schedule":"periodic scrub scheduled @ 2026-03-21T21:09:24.427454+0000","scrub_duration":0,"objects_trimmed":0,"snaptrim_duration":0,"stat_sum":{"num_bytes":0,"num_objects":0,"num_object_clones":0,"num_object_copies":0,"num_objects_missing_on_primary":0,"num_objects_missing":0,"num_objects_degraded":0,"num_objects_misplaced":0,"num_objects_unfound":0,"num_objects_dirty":0,"num_whiteouts":0,"num_read":0,"num_read_kb":0,"num_write":0,"num_write_kb":0,"num_scrub_errors":0,"num_shallow_scrub_errors":0,"num_deep_scrub_errors":0,"num_objects_recovered":0,"num_bytes_recovered":0,"num_keys_recovered":0,"num_objects_omap":0,"num_objects_hit_set_archive":0,"num_bytes_hit_set_archive":0,"num_flush":0,"num_flush_kb":0,"num_evict":0,"num_evict_kb":0,"num_promote":0,"num_flush_mode_high":0,"num_flush_mode_low":0,"num_evict_mode_some":0,"num_evict_mode_full":0,"num_objects_pinned":0,"num_legacy_snapsets":0,"num_large_omap_objects":0,"num_objects_manifest":0,"num_omap_bytes":0,"num_omap_keys":0,"num_objects_repaired":0},"up":[3,5],"acting":[3,5],"avail_no_missing":[],"object_location_counts":[],"blocked_by":[],"up_primary":3,"acting_primary":3,"purged_snaps":[]},{"pgid":"2.0","version":"0'0","reported_seq":13,"reported_epoch":17,"state":"active+clean","last_fresh":"2026-03-20T17:57:49.525624+0000","last_change":"2026-03-20T17:57:47.895021+0000","last_active":"2026-03-20T17:57:49.525624+0000","last_peered":"2026-03-20T17:57:49.525624+0000","last_clean":"2026-03-20T17:57:49.525624+0000","last_became_active":"2026-03-20T17:57:47.894775+0000","last_became_peered":"2026-03-20T17:57:47.894775+0000","last_unstale":"2026-03-20T17:57:49.525624+0000","last_undegraded":"2026-03-20T17:57:49.525624+0000","last_fullsized":"2026-03-20T17:57:49.525624+0000","mapping_epoch":15,"log_start":"0'0","ondisk_log_start":"0'0","created":15,"last_epoch_clean":16,"parent":"0.0","parent_split_bits":0,"last_scrub":"0'0","last_scrub_stamp":"2026-03-20T17:57:46.883576+0000","last_deep_scrub":"0'0","last_deep_scrub_stamp":"2026-03-20T17:57:46.883576+0000","last_clean_scrub_stamp":"2026-03-20T17:57:46.883576+0000","objects_scrubbed":0,"log_size":0,"log_dups_size":0,"ondisk_log_size":0,"stats_invalid":false,"dirty_stats_invalid":false,"omap_stats_invalid":false,"hitset_stats_invalid":false,"hitset_bytes_stats_invalid":false,"pin_stats_invalid":false,"manifest_stats_invalid":false,"snaptrimq_len":0,"last_scrub_duration":0,"scrub_schedule":"periodic scrub scheduled @ 2026-03-22T00:44:03.161568+0000","scrub_duration":0,"objects_trimmed":0,"snaptrim_duration":0,"stat_sum":{"num_bytes":0,"num_objects":0,"num_object_clones":0,"num_object_copies":0,"num_objects_missing_on_primary":0,"num_objects_missing":0,"num_objects_degraded":0,"num_objects_misplaced":0,"num_objects_unfound":0,"num_objects_dirty":0,"num_whiteouts":0,"num_read":0,"num_read_kb":0,"num_write":0,"num_write_kb":0,"num_scrub_errors":0,"num_shallow_scrub_errors":0,"num_deep_scrub_errors":0,"num_objects_recovered":0,"num_bytes_recovered":0,"num_keys_recovered":0,"num_objects_omap":0,"num_objects_hit_set_archive":0,"num_bytes_hit_set_archive":0,"num_flush":0,"num_flush_kb":0,"num_evict":0,"num_evict_kb":0,"num_promote":0,"num_flush_mode_high":0,"num_flush_mode_low":0,"num_evict_mode_some":0,"num_evict_mode_full":0,"num_objects_pinned":0,"num_legacy_snapsets":0,"num_large_omap_objects":0,"num_objects_manifest":0,"num_omap_bytes":0,"num_omap_keys":0,"num_objects_repaired":0},"up":[2,5],"acting":[2,5],"avail_no_missing":[],"object_location_counts":[],"blocked_by":[],"up_primary":2,"acting_primary":2,"purged_snaps":[]},{"pgid":"2.3","version":"16'1","reported_seq":14,"reported_epoch":17,"state":"active+clean","last_fresh":"2026-03-20T17:57:49.202453+0000","last_change":"2026-03-20T17:57:47.896525+0000","last_active":"2026-03-20T17:57:49.202453+0000","last_peered":"2026-03-20T17:57:49.202453+0000","last_clean":"2026-03-20T17:57:49.202453+0000","last_became_active":"2026-03-20T17:57:47.896163+0000","last_became_peered":"2026-03-20T17:57:47.896163+0000","last_unstale":"2026-03-20T17:57:49.202453+0000","last_undegraded":"2026-03-20T17:57:49.202453+0000","last_fullsized":"2026-03-20T17:57:49.202453+0000","mapping_epoch":15,"log_start":"0'0","ondisk_log_start":"0'0","created":15,"last_epoch_clean":16,"parent":"0.0","parent_split_bits":0,"last_scrub":"0'0","last_scrub_stamp":"2026-03-20T17:57:46.883576+0000","last_deep_scrub":"0'0","last_deep_scrub_stamp":"2026-03-20T17:57:46.883576+0000","last_clean_scrub_stamp":"2026-03-20T17:57:46.883576+0000","objects_scrubbed":0,"log_size":1,"log_dups_size":0,"ondisk_log_size":1,"stats_invalid":false,"dirty_stats_invalid":false,"omap_stats_invalid":false,"hitset_stats_invalid":false,"hitset_bytes_stats_invalid":false,"pin_stats_invalid":false,"manifest_stats_invalid":false,"snaptrimq_len":0,"last_scrub_duration":0,"scrub_schedule":"periodic scrub scheduled @ 2026-03-21T22:20:00.347952+0000","scrub_duration":0,"objects_trimmed":0,"snaptrim_duration":0,"stat_sum":{"num_bytes":0,"num_objects":1,"num_object_clones":0,"num_object_copies":2,"num_objects_missing_on_primary":0,"num_objects_missing":0,"num_objects_degraded":0,"num_objects_misplaced":0,"num_objects_unfound":0,"num_objects_dirty":1,"num_whiteouts":0,"num_read":0,"num_read_kb":0,"num_write":0,"num_write_kb":0,"num_scrub_errors":0,"num_shallow_scrub_errors":0,"num_deep_scrub_errors":0,"num_objects_recovered":0,"num_bytes_recovered":0,"num_keys_recovered":0,"num_objects_omap":0,"num_objects_hit_set_archive":0,"num_bytes_hit_set_archive":0,"num_flush":0,"num_flush_kb":0,"num_evict":0,"num_evict_kb":0,"num_promote":0,"num_flush_mode_high":0,"num_flush_mode_low":0,"num_evict_mode_some":0,"num_evict_mode_full":0,"num_objects_pinned":0,"num_legacy_snapsets":0,"num_large_omap_objects":0,"num_objects_manifest":0,"num_omap_bytes":0,"num_omap_keys":0,"num_objects_repaired":0},"up":[1,2],"acting":[1,2],"avail_no_missing":[],"object_location_counts":[],"blocked_by":[],"up_primary":1,"acting_primary":1,"purged_snaps":[]},{"pgid":"1.0","version":"13'32","reported_seq":63,"reported_epoch":17,"state":"active+clean","last_fresh":"2026-03-20T17:57:49.202372+0000","last_change":"2026-03-20T17:57:44.003553+0000","last_active":"2026-03-20T17:57:49.202372+0000","last_peered":"2026-03-20T17:57:49.202372+0000","last_clean":"2026-03-20T17:57:49.202372+0000","last_became_active":"2026-03-20T17:57:44.003422+0000","last_became_peered":"2026-03-20T17:57:44.003422+0000","last_unstale":"2026-03-20T17:57:49.202372+0000","last_undegraded":"2026-03-20T17:57:49.202372+0000","last_fullsized":"2026-03-20T17:57:49.202372+0000","mapping_epoch":12,"log_start":"0'0","ondisk_log_start":"0'0","created":12,"last_epoch_clean":13,"parent":"0.0","parent_split_bits":0,"last_scrub":"0'0","last_scrub_stamp":"2026-03-20T17:57:42.849727+0000","last_deep_scrub":"0'0","last_deep_scrub_stamp":"2026-03-20T17:57:42.849727+0000","last_clean_scrub_stamp":"2026-03-20T17:57:42.849727+0000","objects_scrubbed":0,"log_size":32,"log_dups_size":0,"ondisk_log_size":32,"stats_invalid":false,"dirty_stats_invalid":false,"omap_stats_invalid":false,"hitset_stats_invalid":false,"hitset_bytes_stats_invalid":false,"pin_stats_invalid":false,"manifest_stats_invalid":false,"snaptrimq_len":0,"last_scrub_duration":0,"scrub_schedule":"periodic scrub scheduled @ 2026-03-22T02:36:03.668295+0000","scrub_duration":0,"objects_trimmed":0,"snaptrim_duration":0,"stat_sum":{"num_bytes":459280,"num_objects":2,"num_object_clones":0,"num_object_copies":4,"num_objects_missing_on_primary":0,"num_objects_missing":0,"num_objects_degraded":0,"num_objects_misplaced":0,"num_objects_unfound":0,"num_objects_dirty":2,"num_whiteouts":0,"num_read":46,"num_read_kb":37,"num_write":57,"num_write_kb":584,"num_scrub_errors":0,"num_shallow_scrub_errors":0,"num_deep_scrub_errors":0,"num_objects_recovered":0,"num_bytes_recovered":0,"num_keys_recovered":0,"num_objects_omap":0,"num_objects_hit_set_archive":0,"num_bytes_hit_set_archive":0,"num_flush":0,"num_flush_kb":0,"num_evict":0,"num_evict_kb":0,"num_promote":0,"num_flush_mode_high":0,"num_flush_mode_low":0,"num_evict_mode_some":0,"num_evict_mode_full":0,"num_objects_pinned":0,"num_legacy_snapsets":0,"num_large_omap_objects":0,"num_objects_manifest":0,"num_omap_bytes":0,"num_omap_keys":0,"num_objects_repaired":0},"up":[1,5],"acting":[1,5],"avail_no_missing":[],"object_location_counts":[],"blocked_by":[],"up_primary":1,"acting_primary":1,"purged_snaps":[]}],"pool_stats":[{"poolid":2,"num_pg":8,"stat_sum":{"num_bytes":19,"num_objects":2,"num_object_clones":0,"num_object_copies":4,"num_objects_missing_on_primary":0,"num_objects_missing":0,"num_objects_degraded":0,"num_objects_misplaced":0,"num_objects_unfound":0,"num_objects_dirty":2,"num_whiteouts":0,"num_read":0,"num_read_kb":0,"num_write":2,"num_write_kb":2,"num_scrub_errors":0,"num_shallow_scrub_errors":0,"num_deep_scrub_errors":0,"num_objects_recovered":0,"num_bytes_recovered":0,"num_keys_recovered":0,"num_objects_omap":0,"num_objects_hit_set_archive":0,"num_bytes_hit_set_archive":0,"num_flush":0,"num_flush_kb":0,"num_evict":0,"num_evict_kb":0,"num_promote":0,"num_flush_mode_high":0,"num_flush_mode_low":0,"num_evict_mode_some":0,"num_evict_mode_full":0,"num_objects_pinned":0,"num_legacy_snapsets":0,"num_large_omap_objects":0,"num_objects_manifest":0,"num_omap_bytes":0,"num_omap_keys":0,"num_objects_repaired":0},"store_stats":{"total":0,"available":0,"internally_reserved":0,"allocated":8192,"data_stored":38,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":0},"log_size":3,"ondisk_log_size":3,"up":16,"acting":16,"num_store_stats":6},{"poolid":1,"num_pg":1,"stat_sum":{"num_bytes":459280,"num_objects":2,"num_object_clones":0,"num_object_copies":4,"num_objects_missing_on_primary":0,"num_objects_missing":0,"num_objects_degraded":0,"num_objects_misplaced":0,"num_objects_unfound":0,"num_objects_dirty":2,"num_whiteouts":0,"num_read":46,"num_read_kb":37,"num_write":57,"num_write_kb":584,"num_scrub_errors":0,"num_shallow_scrub_errors":0,"num_deep_scrub_errors":0,"num_objects_recovered":0,"num_bytes_recovered":0,"num_keys_recovered":0,"num_objects_omap":0,"num_objects_hit_set_archive":0,"num_bytes_hit_set_archive":0,"num_flush":0,"num_flush_kb":0,"num_evict":0,"num_evict_kb":0,"num_promote":0,"num_flush_mode_high":0,"num_flush_mode_low":0,"num_evict_mode_some":0,"num_evict_mode_full":0,"num_objects_pinned":0,"num_legacy_snapsets":0,"num_large_omap_objects":0,"num_objects_manifest":0,"num_omap_bytes":0,"num_omap_keys":0,"num_objects_repaired":0},"store_stats":{"total":0,"available":0,"internally_reserved":0,"allocated":925696,"data_stored":918560,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":0},"log_size":32,"ondisk_log_size":32,"up":2,"acting":2,"num_store_stats":2}],"osd_stats":[{"osd":5,"up_from":11,"seq":47244640259,"num_pgs":2,"num_osds":1,"num_per_pool_osds":1,"num_per_pool_omap_osds":1,"kb":104857600,"kb_used":27472,"kb_used_data":608,"kb_used_omap":7,"kb_used_meta":26808,"kb_avail":104830128,"statfs":{"total":107374182400,"available":107346051072,"internally_reserved":0,"allocated":622592,"data_stored":510892,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":7471,"internal_metadata":27452113},"hb_peers":[0,1,2,3,4],"snap_trim_queue_len":0,"num_snap_trimming":0,"num_shards_repaired":0,"op_queue_age_hist":{"histogram":[],"upper_bound":1},"perf_stat":{"commit_latency_ms":0,"apply_latency_ms":0,"commit_latency_ns":0,"apply_latency_ns":0},"alerts":[]},{"osd":4,"up_from":11,"seq":47244640259,"num_pgs":3,"num_osds":1,"num_per_pool_osds":1,"num_per_pool_omap_osds":1,"kb":104857600,"kb_used":27020,"kb_used_data":172,"kb_used_omap":7,"kb_used_meta":26808,"kb_avail":104830580,"statfs":{"total":107374182400,"available":107346513920,"internally_reserved":0,"allocated":176128,"data_stored":56583,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":8121,"internal_metadata":27451463},"hb_peers":[0,1,2,3,5],"snap_trim_queue_len":0,"num_snap_trimming":0,"num_shards_repaired":0,"op_queue_age_hist":{"histogram":[],"upper_bound":1},"perf_stat":{"commit_latency_ms":0,"apply_latency_ms":0,"commit_latency_ns":0,"apply_latency_ns":0},"alerts":[]},{"osd":3,"up_from":11,"seq":47244640260,"num_pgs":2,"num_osds":1,"num_per_pool_osds":1,"num_per_pool_omap_osds":1,"kb":104857600,"kb_used":27012,"kb_used_data":156,"kb_used_omap":6,"kb_used_meta":26809,"kb_avail":104830588,"statfs":{"total":107374182400,"available":107346522112,"internally_reserved":0,"allocated":159744,"data_stored":51612,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":6819,"internal_metadata":27452765},"hb_peers":[0,1,2,4,5],"snap_trim_queue_len":0,"num_snap_trimming":0,"num_shards_repaired":0,"op_queue_age_hist":{"histogram":[],"upper_bound":1},"perf_stat":{"commit_latency_ms":0,"apply_latency_ms":0,"commit_latency_ns":0,"apply_latency_ns":0},"alerts":[]},{"osd":2,"up_from":11,"seq":47244640259,"num_pgs":2,"num_osds":1,"num_per_pool_osds":1,"num_per_pool_omap_osds":1,"kb":104857600,"kb_used":27012,"kb_used_data":156,"kb_used_omap":7,"kb_used_meta":26808,"kb_avail":104830588,"statfs":{"total":107374182400,"available":107346522112,"internally_reserved":0,"allocated":159744,"data_stored":51612,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":7471,"internal_metadata":27452113},"hb_peers":[0,1,3,4,5],"snap_trim_queue_len":0,"num_snap_trimming":0,"num_shards_repaired":0,"op_queue_age_hist":{"histogram":[],"upper_bound":1},"perf_stat":{"commit_latency_ms":1,"apply_latency_ms":1,"commit_latency_ns":1000000,"apply_latency_ns":1000000},"alerts":[]},{"osd":1,"up_from":11,"seq":47244640259,"num_pgs":4,"num_osds":1,"num_per_pool_osds":1,"num_per_pool_omap_osds":1,"kb":104857600,"kb_used":27464,"kb_used_data":608,"kb_used_omap":7,"kb_used_meta":26808,"kb_avail":104830136,"statfs":{"total":107374182400,"available":107346059264,"internally_reserved":0,"allocated":622592,"data_stored":510892,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":8121,"internal_metadata":27451463},"hb_peers":[0,2,3,4,5],"snap_trim_queue_len":0,"num_snap_trimming":0,"num_shards_repaired":0,"op_queue_age_hist":{"histogram":[],"upper_bound":1},"perf_stat":{"commit_latency_ms":0,"apply_latency_ms":0,"commit_latency_ns":0,"apply_latency_ns":0},"alerts":[]},{"osd":0,"up_from":11,"seq":47244640260,"num_pgs":3,"num_osds":1,"num_per_pool_osds":1,"num_per_pool_omap_osds":1,"kb":104857600,"kb_used":27020,"kb_used_data":172,"kb_used_omap":7,"kb_used_meta":26808,"kb_avail":104830580,"statfs":{"total":107374182400,"available":107346513920,"internally_reserved":0,"allocated":176128,"data_stored":56583,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":8121,"internal_metadata":27451463},"hb_peers":[1,2,3,4,5],"snap_trim_queue_len":0,"num_snap_trimming":0,"num_shards_repaired":0,"op_queue_age_hist":{"histogram":[],"upper_bound":1},"perf_stat":{"commit_latency_ms":0,"apply_latency_ms":0,"commit_latency_ns":0,"apply_latency_ns":0},"alerts":[]}],"pool_statfs":[{"poolid":1,"osd":1,"total":0,"available":0,"internally_reserved":0,"allocated":462848,"data_stored":459280,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":0},{"poolid":1,"osd":5,"total":0,"available":0,"internally_reserved":0,"allocated":462848,"data_stored":459280,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":0},{"poolid":2,"osd":0,"total":0,"available":0,"internally_reserved":0,"allocated":4096,"data_stored":19,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":0},{"poolid":2,"osd":1,"total":0,"available":0,"internally_reserved":0,"allocated":0,"data_stored":0,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":0},{"poolid":2,"osd":2,"total":0,"available":0,"internally_reserved":0,"allocated":0,"data_stored":0,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":0},{"poolid":2,"osd":3,"total":0,"available":0,"internally_reserved":0,"allocated":0,"data_stored":0,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":0},{"poolid":2,"osd":4,"total":0,"available":0,"internally_reserved":0,"allocated":4096,"data_stored":19,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":0},{"poolid":2,"osd":5,"total":0,"available":0,"internally_reserved":0,"allocated":0,"data_stored":0,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":0}]}} 2026-03-20T17:57:52.957 INFO:tasks.ceph.ceph_manager.c2:clean! 2026-03-20T17:57:52.957 INFO:tasks.ceph:Waiting until ceph cluster c2 is healthy... 2026-03-20T17:57:52.957 INFO:tasks.ceph.ceph_manager.c2:wait_until_healthy 2026-03-20T17:57:52.957 DEBUG:teuthology.orchestra.run.vm06:> sudo adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage timeout 120 ceph --cluster c2 health --format=json 2026-03-20T17:57:53.150 INFO:teuthology.orchestra.run.vm06.stdout: 2026-03-20T17:57:53.150 INFO:teuthology.orchestra.run.vm06.stdout:{"status":"HEALTH_OK","checks":{},"mutes":[]} 2026-03-20T17:57:53.164 INFO:tasks.ceph.ceph_manager.c2:wait_until_healthy done 2026-03-20T17:57:53.164 INFO:teuthology.run_tasks:Running task rgw... 2026-03-20T17:57:53.170 DEBUG:tasks.rgw:config is {'c1.client.0': {'port': 8000}, 'c1.client.1': {'port': 8001}, 'c2.client.0': {'port': 8000}, 'c2.client.1': {'port': 8001}} 2026-03-20T17:57:53.170 DEBUG:tasks.rgw:client list is dict_keys(['c1.client.0', 'c1.client.1', 'c2.client.0', 'c2.client.1']) 2026-03-20T17:57:53.170 INFO:tasks.rgw:Creating data pools 2026-03-20T17:57:53.170 DEBUG:tasks.rgw:Obtaining remote for client c1.client.0 2026-03-20T17:57:53.170 DEBUG:teuthology.orchestra.run.vm02:> sudo ceph osd pool create default.rgw.buckets.data 64 64 --cluster c1 2026-03-20T17:57:53.482 INFO:teuthology.orchestra.run.vm02.stderr:pool 'default.rgw.buckets.data' created 2026-03-20T17:57:53.504 DEBUG:teuthology.orchestra.run.vm02:> sudo ceph osd pool application enable default.rgw.buckets.data rgw --cluster c1 2026-03-20T17:57:54.484 INFO:teuthology.orchestra.run.vm02.stderr:enabled application 'rgw' on pool 'default.rgw.buckets.data' 2026-03-20T17:57:54.524 DEBUG:teuthology.orchestra.run.vm02:> sudo ceph osd pool create default.rgw.buckets.index 64 64 --cluster c1 2026-03-20T17:57:55.487 INFO:teuthology.orchestra.run.vm02.stderr:pool 'default.rgw.buckets.index' created 2026-03-20T17:57:55.514 DEBUG:teuthology.orchestra.run.vm02:> sudo ceph osd pool application enable default.rgw.buckets.index rgw --cluster c1 2026-03-20T17:57:56.489 INFO:teuthology.orchestra.run.vm02.stderr:enabled application 'rgw' on pool 'default.rgw.buckets.index' 2026-03-20T17:57:56.527 DEBUG:tasks.rgw:Obtaining remote for client c1.client.1 2026-03-20T17:57:56.527 DEBUG:teuthology.orchestra.run.vm05:> sudo ceph osd pool create default.rgw.buckets.data 64 64 --cluster c1 2026-03-20T17:57:56.712 INFO:teuthology.orchestra.run.vm05.stderr:pool 'default.rgw.buckets.data' already exists 2026-03-20T17:57:56.727 DEBUG:teuthology.orchestra.run.vm05:> sudo ceph osd pool application enable default.rgw.buckets.data rgw --cluster c1 2026-03-20T17:57:57.491 INFO:teuthology.orchestra.run.vm05.stderr:enabled application 'rgw' on pool 'default.rgw.buckets.data' 2026-03-20T17:57:57.505 DEBUG:teuthology.orchestra.run.vm05:> sudo ceph osd pool create default.rgw.buckets.index 64 64 --cluster c1 2026-03-20T17:57:57.692 INFO:teuthology.orchestra.run.vm05.stderr:pool 'default.rgw.buckets.index' already exists 2026-03-20T17:57:57.707 DEBUG:teuthology.orchestra.run.vm05:> sudo ceph osd pool application enable default.rgw.buckets.index rgw --cluster c1 2026-03-20T17:57:58.519 INFO:teuthology.orchestra.run.vm05.stderr:enabled application 'rgw' on pool 'default.rgw.buckets.index' 2026-03-20T17:57:58.537 DEBUG:tasks.rgw:Obtaining remote for client c2.client.0 2026-03-20T17:57:58.537 DEBUG:teuthology.orchestra.run.vm06:> sudo ceph osd pool create default.rgw.buckets.data 64 64 --cluster c2 2026-03-20T17:57:58.929 INFO:teuthology.orchestra.run.vm06.stderr:pool 'default.rgw.buckets.data' created 2026-03-20T17:57:58.957 DEBUG:teuthology.orchestra.run.vm06:> sudo ceph osd pool application enable default.rgw.buckets.data rgw --cluster c2 2026-03-20T17:57:59.934 INFO:teuthology.orchestra.run.vm06.stderr:enabled application 'rgw' on pool 'default.rgw.buckets.data' 2026-03-20T17:57:59.972 DEBUG:teuthology.orchestra.run.vm06:> sudo ceph osd pool create default.rgw.buckets.index 64 64 --cluster c2 2026-03-20T17:58:00.941 INFO:teuthology.orchestra.run.vm06.stderr:pool 'default.rgw.buckets.index' created 2026-03-20T17:58:00.964 DEBUG:teuthology.orchestra.run.vm06:> sudo ceph osd pool application enable default.rgw.buckets.index rgw --cluster c2 2026-03-20T17:58:01.950 INFO:teuthology.orchestra.run.vm06.stderr:enabled application 'rgw' on pool 'default.rgw.buckets.index' 2026-03-20T17:58:01.984 DEBUG:tasks.rgw:Obtaining remote for client c2.client.1 2026-03-20T17:58:01.984 DEBUG:teuthology.orchestra.run.vm08:> sudo ceph osd pool create default.rgw.buckets.data 64 64 --cluster c2 2026-03-20T17:58:02.159 INFO:teuthology.orchestra.run.vm08.stderr:pool 'default.rgw.buckets.data' already exists 2026-03-20T17:58:02.173 DEBUG:teuthology.orchestra.run.vm08:> sudo ceph osd pool application enable default.rgw.buckets.data rgw --cluster c2 2026-03-20T17:58:02.533 INFO:teuthology.orchestra.run.vm08.stderr:enabled application 'rgw' on pool 'default.rgw.buckets.data' 2026-03-20T17:58:02.547 DEBUG:teuthology.orchestra.run.vm08:> sudo ceph osd pool create default.rgw.buckets.index 64 64 --cluster c2 2026-03-20T17:58:02.726 INFO:teuthology.orchestra.run.vm08.stderr:pool 'default.rgw.buckets.index' already exists 2026-03-20T17:58:02.744 DEBUG:teuthology.orchestra.run.vm08:> sudo ceph osd pool application enable default.rgw.buckets.index rgw --cluster c2 2026-03-20T17:58:03.534 INFO:teuthology.orchestra.run.vm08.stderr:enabled application 'rgw' on pool 'default.rgw.buckets.index' 2026-03-20T17:58:03.549 DEBUG:tasks.rgw:Pools created 2026-03-20T17:58:03.549 INFO:tasks.rgw:Configuring compression type = random 2026-03-20T17:58:03.549 INFO:tasks.util.rgw:rgwadmin: c1.client.0 : ['user', 'list'] 2026-03-20T17:58:03.549 DEBUG:tasks.util.rgw:rgwadmin: cmd=['adjust-ulimits', 'ceph-coverage', '/home/ubuntu/cephtest/archive/coverage', 'radosgw-admin', '--log-to-stderr', '--format', 'json', '-n', 'client.0', '--cluster', 'c1', 'user', 'list'] 2026-03-20T17:58:03.549 DEBUG:teuthology.orchestra.run.vm02:> adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage radosgw-admin --log-to-stderr --format json -n client.0 --cluster c1 user list 2026-03-20T17:58:03.569 INFO:teuthology.orchestra.run.vm02.stderr:ignoring --setuser ceph since I am not root 2026-03-20T17:58:03.569 INFO:teuthology.orchestra.run.vm02.stderr:ignoring --setgroup ceph since I am not root 2026-03-20T17:58:05.612 INFO:teuthology.orchestra.run.vm02.stderr:2026-03-20T17:58:05.609+0000 7fb0b714aa80 20 rados->read ofs=0 len=0 2026-03-20T17:58:05.613 INFO:teuthology.orchestra.run.vm02.stderr:2026-03-20T17:58:05.609+0000 7fb0b714aa80 20 rados_obj.operate() r=-2 bl.length=0 2026-03-20T17:58:05.613 INFO:teuthology.orchestra.run.vm02.stderr:2026-03-20T17:58:05.609+0000 7fb0b714aa80 20 realm 2026-03-20T17:58:05.613 INFO:teuthology.orchestra.run.vm02.stderr:2026-03-20T17:58:05.609+0000 7fb0b714aa80 20 rados->read ofs=0 len=0 2026-03-20T17:58:05.613 INFO:teuthology.orchestra.run.vm02.stderr:2026-03-20T17:58:05.609+0000 7fb0b714aa80 20 rados_obj.operate() r=-2 bl.length=0 2026-03-20T17:58:05.613 INFO:teuthology.orchestra.run.vm02.stderr:2026-03-20T17:58:05.609+0000 7fb0b714aa80 4 RGWPeriod::init failed to init realm id : (2) No such file or directory 2026-03-20T17:58:05.613 INFO:teuthology.orchestra.run.vm02.stderr:2026-03-20T17:58:05.609+0000 7fb0b714aa80 20 rados->read ofs=0 len=0 2026-03-20T17:58:05.613 INFO:teuthology.orchestra.run.vm02.stderr:2026-03-20T17:58:05.609+0000 7fb0b714aa80 20 rados_obj.operate() r=-2 bl.length=0 2026-03-20T17:58:05.613 INFO:teuthology.orchestra.run.vm02.stderr:2026-03-20T17:58:05.609+0000 7fb0b714aa80 20 rados->read ofs=0 len=0 2026-03-20T17:58:05.613 INFO:teuthology.orchestra.run.vm02.stderr:2026-03-20T17:58:05.609+0000 7fb0b714aa80 20 rados_obj.operate() r=0 bl.length=46 2026-03-20T17:58:05.613 INFO:teuthology.orchestra.run.vm02.stderr:2026-03-20T17:58:05.609+0000 7fb0b714aa80 20 rados->read ofs=0 len=0 2026-03-20T17:58:05.614 INFO:teuthology.orchestra.run.vm02.stderr:2026-03-20T17:58:05.609+0000 7fb0b714aa80 20 rados_obj.operate() r=0 bl.length=1060 2026-03-20T17:58:05.614 INFO:teuthology.orchestra.run.vm02.stderr:2026-03-20T17:58:05.609+0000 7fb0b714aa80 20 searching for the correct realm 2026-03-20T17:58:05.621 INFO:teuthology.orchestra.run.vm02.stderr:2026-03-20T17:58:05.617+0000 7fb0b714aa80 20 RGWRados::pool_iterate: got zonegroup_info.b24151bb-5fb7-4fb5-a2d6-5d2063bc9ebe 2026-03-20T17:58:05.621 INFO:teuthology.orchestra.run.vm02.stderr:2026-03-20T17:58:05.617+0000 7fb0b714aa80 20 RGWRados::pool_iterate: got default.zonegroup. 2026-03-20T17:58:05.621 INFO:teuthology.orchestra.run.vm02.stderr:2026-03-20T17:58:05.617+0000 7fb0b714aa80 20 RGWRados::pool_iterate: got default.zone. 2026-03-20T17:58:05.621 INFO:teuthology.orchestra.run.vm02.stderr:2026-03-20T17:58:05.617+0000 7fb0b714aa80 20 RGWRados::pool_iterate: got zone_info.bd0a7e49-2751-4046-9c81-a69cd7a6f5e0 2026-03-20T17:58:05.621 INFO:teuthology.orchestra.run.vm02.stderr:2026-03-20T17:58:05.617+0000 7fb0b714aa80 20 RGWRados::pool_iterate: got zone_names.default 2026-03-20T17:58:05.621 INFO:teuthology.orchestra.run.vm02.stderr:2026-03-20T17:58:05.617+0000 7fb0b714aa80 20 RGWRados::pool_iterate: got zonegroups_names.default 2026-03-20T17:58:05.621 INFO:teuthology.orchestra.run.vm02.stderr:2026-03-20T17:58:05.617+0000 7fb0b714aa80 20 rados->read ofs=0 len=0 2026-03-20T17:58:05.621 INFO:teuthology.orchestra.run.vm02.stderr:2026-03-20T17:58:05.617+0000 7fb0b714aa80 20 rados_obj.operate() r=-2 bl.length=0 2026-03-20T17:58:05.621 INFO:teuthology.orchestra.run.vm02.stderr:2026-03-20T17:58:05.617+0000 7fb0b714aa80 20 rados->read ofs=0 len=0 2026-03-20T17:58:05.621 INFO:teuthology.orchestra.run.vm02.stderr:2026-03-20T17:58:05.617+0000 7fb0b714aa80 20 rados_obj.operate() r=0 bl.length=46 2026-03-20T17:58:05.621 INFO:teuthology.orchestra.run.vm02.stderr:2026-03-20T17:58:05.617+0000 7fb0b714aa80 20 rados->read ofs=0 len=0 2026-03-20T17:58:05.621 INFO:teuthology.orchestra.run.vm02.stderr:2026-03-20T17:58:05.617+0000 7fb0b714aa80 20 rados_obj.operate() r=0 bl.length=436 2026-03-20T17:58:05.621 INFO:teuthology.orchestra.run.vm02.stderr:2026-03-20T17:58:05.617+0000 7fb0b714aa80 20 zone default found 2026-03-20T17:58:05.621 INFO:teuthology.orchestra.run.vm02.stderr:2026-03-20T17:58:05.617+0000 7fb0b714aa80 4 Realm: () 2026-03-20T17:58:05.622 INFO:teuthology.orchestra.run.vm02.stderr:2026-03-20T17:58:05.617+0000 7fb0b714aa80 4 ZoneGroup: default (b24151bb-5fb7-4fb5-a2d6-5d2063bc9ebe) 2026-03-20T17:58:05.622 INFO:teuthology.orchestra.run.vm02.stderr:2026-03-20T17:58:05.617+0000 7fb0b714aa80 4 Zone: default (bd0a7e49-2751-4046-9c81-a69cd7a6f5e0) 2026-03-20T17:58:05.622 INFO:teuthology.orchestra.run.vm02.stderr:2026-03-20T17:58:05.617+0000 7fb0b714aa80 10 cannot find current period zonegroup using local zonegroup configuration 2026-03-20T17:58:05.622 INFO:teuthology.orchestra.run.vm02.stderr:2026-03-20T17:58:05.617+0000 7fb0b714aa80 20 zonegroup default 2026-03-20T17:58:05.622 INFO:teuthology.orchestra.run.vm02.stderr:2026-03-20T17:58:05.617+0000 7fb0b714aa80 20 rados->read ofs=0 len=0 2026-03-20T17:58:05.622 INFO:teuthology.orchestra.run.vm02.stderr:2026-03-20T17:58:05.617+0000 7fb0b714aa80 20 rados_obj.operate() r=-2 bl.length=0 2026-03-20T17:58:05.622 INFO:teuthology.orchestra.run.vm02.stderr:2026-03-20T17:58:05.617+0000 7fb0b714aa80 20 rados->read ofs=0 len=0 2026-03-20T17:58:07.599 INFO:teuthology.orchestra.run.vm02.stderr:2026-03-20T17:58:07.597+0000 7fb0b714aa80 20 rados_obj.operate() r=-2 bl.length=0 2026-03-20T17:58:07.599 INFO:teuthology.orchestra.run.vm02.stderr:2026-03-20T17:58:07.597+0000 7fb0b714aa80 20 rados->read ofs=0 len=0 2026-03-20T17:58:07.603 INFO:teuthology.orchestra.run.vm02.stderr:2026-03-20T17:58:07.601+0000 7fb0b714aa80 20 rados_obj.operate() r=-2 bl.length=0 2026-03-20T17:58:07.603 INFO:teuthology.orchestra.run.vm02.stderr:2026-03-20T17:58:07.601+0000 7fb0b714aa80 20 started sync module instance, tier type = 2026-03-20T17:58:07.603 INFO:teuthology.orchestra.run.vm02.stderr:2026-03-20T17:58:07.601+0000 7fb0b714aa80 20 started zone id=bd0a7e49-2751-4046-9c81-a69cd7a6f5e0 (name=default) with tier type = 2026-03-20T17:58:09.617 INFO:teuthology.orchestra.run.vm02.stderr:2026-03-20T17:58:09.613+0000 7fb0b714aa80 20 add_watcher() i=0 2026-03-20T17:58:09.617 INFO:teuthology.orchestra.run.vm02.stderr:2026-03-20T17:58:09.613+0000 7fb0b714aa80 20 add_watcher() i=3 2026-03-20T17:58:09.617 INFO:teuthology.orchestra.run.vm02.stderr:2026-03-20T17:58:09.613+0000 7fb0b714aa80 20 add_watcher() i=2 2026-03-20T17:58:09.617 INFO:teuthology.orchestra.run.vm02.stderr:2026-03-20T17:58:09.613+0000 7fb0b714aa80 20 add_watcher() i=1 2026-03-20T17:58:09.617 INFO:teuthology.orchestra.run.vm02.stderr:2026-03-20T17:58:09.613+0000 7fb0b714aa80 20 add_watcher() i=4 2026-03-20T17:58:09.617 INFO:teuthology.orchestra.run.vm02.stderr:2026-03-20T17:58:09.613+0000 7fb0b714aa80 20 add_watcher() i=6 2026-03-20T17:58:09.620 INFO:teuthology.orchestra.run.vm02.stderr:2026-03-20T17:58:09.617+0000 7fb0b714aa80 20 add_watcher() i=7 2026-03-20T17:58:09.620 INFO:teuthology.orchestra.run.vm02.stderr:2026-03-20T17:58:09.617+0000 7fb0b714aa80 20 add_watcher() i=5 2026-03-20T17:58:09.620 INFO:teuthology.orchestra.run.vm02.stderr:2026-03-20T17:58:09.617+0000 7fb0b714aa80 2 all 8 watchers are set, enabling cache 2026-03-20T17:58:09.621 INFO:teuthology.orchestra.run.vm02.stderr:2026-03-20T17:58:09.617+0000 7fb0a5ffb640 5 boost::asio::awaitable, obj_version> > logback_generations::read(const DoutPrefixProvider*):446: oid=data_loggenerations_metadata not found 2026-03-20T17:58:09.621 INFO:teuthology.orchestra.run.vm02.stderr:2026-03-20T17:58:09.617+0000 7fb0a5ffb640 20 boost::asio::awaitable {anonymous}::probe_shard(const DoutPrefixProvider*, neorados::RADOS, const neorados::Object&, const neorados::IOContext&, bool&):59 probing obj=data_log.0 2026-03-20T17:58:09.621 INFO:teuthology.orchestra.run.vm02.stderr:2026-03-20T17:58:09.617+0000 7fb0a5ffb640 20 do_open: entering 2026-03-20T17:58:09.622 INFO:teuthology.orchestra.run.vm02.stderr:2026-03-20T17:58:09.621+0000 7fb0a57fa640 20 boost::asio::awaitable {anonymous}::probe_shard(const DoutPrefixProvider*, neorados::RADOS, const neorados::Object&, const neorados::IOContext&, bool&):78: obj=data_log.0 does not exist 2026-03-20T17:58:09.622 INFO:teuthology.orchestra.run.vm02.stderr:2026-03-20T17:58:09.621+0000 7fb0a57fa640 20 boost::asio::awaitable {anonymous}::probe_shard(const DoutPrefixProvider*, neorados::RADOS, const neorados::Object&, const neorados::IOContext&, bool&):59 probing obj=data_log.1 2026-03-20T17:58:09.622 INFO:teuthology.orchestra.run.vm02.stderr:2026-03-20T17:58:09.621+0000 7fb0a57fa640 20 do_open: entering 2026-03-20T17:58:09.622 INFO:teuthology.orchestra.run.vm02.stderr:2026-03-20T17:58:09.621+0000 7fb0a4ff9640 20 boost::asio::awaitable {anonymous}::probe_shard(const DoutPrefixProvider*, neorados::RADOS, const neorados::Object&, const neorados::IOContext&, bool&):78: obj=data_log.1 does not exist 2026-03-20T17:58:09.622 INFO:teuthology.orchestra.run.vm02.stderr:2026-03-20T17:58:09.621+0000 7fb0a4ff9640 20 boost::asio::awaitable {anonymous}::probe_shard(const DoutPrefixProvider*, neorados::RADOS, const neorados::Object&, const neorados::IOContext&, bool&):59 probing obj=data_log.2 2026-03-20T17:58:09.622 INFO:teuthology.orchestra.run.vm02.stderr:2026-03-20T17:58:09.621+0000 7fb0a4ff9640 20 do_open: entering 2026-03-20T17:58:09.623 INFO:teuthology.orchestra.run.vm02.stderr:2026-03-20T17:58:09.621+0000 7fb0b5c1d640 20 boost::asio::awaitable {anonymous}::probe_shard(const DoutPrefixProvider*, neorados::RADOS, const neorados::Object&, const neorados::IOContext&, bool&):78: obj=data_log.2 does not exist 2026-03-20T17:58:09.623 INFO:teuthology.orchestra.run.vm02.stderr:2026-03-20T17:58:09.621+0000 7fb0b5c1d640 20 boost::asio::awaitable {anonymous}::probe_shard(const DoutPrefixProvider*, neorados::RADOS, const neorados::Object&, const neorados::IOContext&, bool&):59 probing obj=data_log.3 2026-03-20T17:58:09.623 INFO:teuthology.orchestra.run.vm02.stderr:2026-03-20T17:58:09.621+0000 7fb0b5c1d640 20 do_open: entering 2026-03-20T17:58:09.623 INFO:teuthology.orchestra.run.vm02.stderr:2026-03-20T17:58:09.621+0000 7fb0a77fe640 20 boost::asio::awaitable {anonymous}::probe_shard(const DoutPrefixProvider*, neorados::RADOS, const neorados::Object&, const neorados::IOContext&, bool&):78: obj=data_log.3 does not exist 2026-03-20T17:58:09.623 INFO:teuthology.orchestra.run.vm02.stderr:2026-03-20T17:58:09.621+0000 7fb0a77fe640 20 do_create: entering 2026-03-20T17:58:09.624 INFO:teuthology.orchestra.run.vm02.stderr:2026-03-20T17:58:09.621+0000 7fb0a6ffd640 20 do_open: entering 2026-03-20T17:58:09.627 INFO:teuthology.orchestra.run.vm02.stderr:2026-03-20T17:58:09.625+0000 7fb0b714aa80 20 rgw_check_secure_mon_conn(): auth registy supported: methods=[2] modes=[2,1] 2026-03-20T17:58:09.627 INFO:teuthology.orchestra.run.vm02.stderr:2026-03-20T17:58:09.625+0000 7fb0b714aa80 20 rgw_check_secure_mon_conn(): mode 1 is insecure 2026-03-20T17:58:12.629 INFO:teuthology.orchestra.run.vm02.stderr:2026-03-20T17:58:12.625+0000 7fb0b714aa80 10 rgw_init_ioctx warning: failed to set recovery_priority on default.rgw.meta 2026-03-20T17:58:12.629 INFO:teuthology.orchestra.run.vm02.stderr:2026-03-20T17:58:12.625+0000 7fb0b714aa80 5 note: GC not initialized 2026-03-20T17:58:12.629 INFO:teuthology.orchestra.run.vm02.stderr:2026-03-20T17:58:12.625+0000 7fb05efe5640 20 reqs_thread_entry: start 2026-03-20T17:58:12.686 INFO:teuthology.orchestra.run.vm02.stderr:2026-03-20T17:58:12.681+0000 7fb0b714aa80 20 init_complete bucket index max shards: 11 2026-03-20T17:58:12.686 INFO:teuthology.orchestra.run.vm02.stderr:2026-03-20T17:58:12.681+0000 7fb0b714aa80 20 Filter name: none 2026-03-20T17:58:12.686 INFO:teuthology.orchestra.run.vm02.stderr:2026-03-20T17:58:12.681+0000 7fb05cfe1640 20 reqs_thread_entry: start 2026-03-20T17:58:12.695 INFO:teuthology.orchestra.run.vm02.stderr:2026-03-20T17:58:12.693+0000 7fb0b714aa80 20 remove_watcher() i=0 2026-03-20T17:58:12.695 INFO:teuthology.orchestra.run.vm02.stderr:2026-03-20T17:58:12.693+0000 7fb0b714aa80 2 removed watcher, disabling cache 2026-03-20T17:58:12.696 INFO:teuthology.orchestra.run.vm02.stderr:2026-03-20T17:58:12.693+0000 7fb0b714aa80 20 remove_watcher() i=3 2026-03-20T17:58:12.696 INFO:teuthology.orchestra.run.vm02.stderr:2026-03-20T17:58:12.693+0000 7fb0b714aa80 20 remove_watcher() i=5 2026-03-20T17:58:12.696 INFO:teuthology.orchestra.run.vm02.stderr:2026-03-20T17:58:12.693+0000 7fb0b714aa80 20 remove_watcher() i=6 2026-03-20T17:58:12.696 INFO:teuthology.orchestra.run.vm02.stderr:2026-03-20T17:58:12.693+0000 7fb0b714aa80 20 remove_watcher() i=7 2026-03-20T17:58:12.696 INFO:teuthology.orchestra.run.vm02.stderr:2026-03-20T17:58:12.693+0000 7fb0b714aa80 20 remove_watcher() i=2 2026-03-20T17:58:12.696 INFO:teuthology.orchestra.run.vm02.stderr:2026-03-20T17:58:12.693+0000 7fb0b714aa80 20 remove_watcher() i=1 2026-03-20T17:58:12.696 INFO:teuthology.orchestra.run.vm02.stderr:2026-03-20T17:58:12.693+0000 7fb0b714aa80 20 remove_watcher() i=4 2026-03-20T17:58:12.702 INFO:teuthology.orchestra.run.vm02.stdout:[] 2026-03-20T17:58:12.702 DEBUG:tasks.util.rgw: json result: [] 2026-03-20T17:58:12.702 INFO:tasks.util.rgw:rgwadmin: c1.client.0 : ['zone', 'placement', 'modify', '--rgw-zone', 'default', '--placement-id', 'default-placement', '--compression', 'random'] 2026-03-20T17:58:12.702 DEBUG:tasks.util.rgw:rgwadmin: cmd=['adjust-ulimits', 'ceph-coverage', '/home/ubuntu/cephtest/archive/coverage', 'radosgw-admin', '--log-to-stderr', '--format', 'json', '-n', 'client.0', '--cluster', 'c1', 'zone', 'placement', 'modify', '--rgw-zone', 'default', '--placement-id', 'default-placement', '--compression', 'random'] 2026-03-20T17:58:12.702 DEBUG:teuthology.orchestra.run.vm02:> adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage radosgw-admin --log-to-stderr --format json -n client.0 --cluster c1 zone placement modify --rgw-zone default --placement-id default-placement --compression random 2026-03-20T17:58:12.765 INFO:teuthology.orchestra.run.vm02.stderr:ignoring --setuser ceph since I am not root 2026-03-20T17:58:12.765 INFO:teuthology.orchestra.run.vm02.stderr:ignoring --setgroup ceph since I am not root 2026-03-20T17:58:12.778 INFO:teuthology.orchestra.run.vm02.stderr:2026-03-20T17:58:12.777+0000 7f872c015a80 20 rgw_check_secure_mon_conn(): auth registy supported: methods=[2] modes=[2,1] 2026-03-20T17:58:12.779 INFO:teuthology.orchestra.run.vm02.stderr:2026-03-20T17:58:12.777+0000 7f872c015a80 20 rgw_check_secure_mon_conn(): mode 1 is insecure 2026-03-20T17:58:12.779 INFO:teuthology.orchestra.run.vm02.stderr:2026-03-20T17:58:12.777+0000 7f86d37fe640 20 reqs_thread_entry: start 2026-03-20T17:58:12.790 INFO:teuthology.orchestra.run.vm02.stdout:{"id":"bd0a7e49-2751-4046-9c81-a69cd7a6f5e0","name":"default","domain_root":"default.rgw.meta:root","control_pool":"default.rgw.control","dedup_pool":"default.rgw.dedup","gc_pool":"default.rgw.log:gc","lc_pool":"default.rgw.log:lc","log_pool":"default.rgw.log","intent_log_pool":"default.rgw.log:intent","usage_log_pool":"default.rgw.log:usage","roles_pool":"default.rgw.meta:roles","reshard_pool":"default.rgw.log:reshard","user_keys_pool":"default.rgw.meta:users.keys","user_email_pool":"default.rgw.meta:users.email","user_swift_pool":"default.rgw.meta:users.swift","user_uid_pool":"default.rgw.meta:users.uid","otp_pool":"default.rgw.otp","notif_pool":"default.rgw.log:notif","topics_pool":"default.rgw.meta:topics","account_pool":"default.rgw.meta:accounts","group_pool":"default.rgw.meta:groups","system_key":{"access_key":"","secret_key":""},"placement_pools":[{"key":"default-placement","val":{"index_pool":"default.rgw.buckets.index","storage_classes":{"STANDARD":{"data_pool":"default.rgw.buckets.data","compression_type":"random"}},"data_extra_pool":"default.rgw.buckets.non-ec","index_type":0,"inline_data":true}}],"realm_id":"","restore_pool":"default.rgw.log:restore"} 2026-03-20T17:58:12.791 DEBUG:tasks.util.rgw: json result: {'id': 'bd0a7e49-2751-4046-9c81-a69cd7a6f5e0', 'name': 'default', 'domain_root': 'default.rgw.meta:root', 'control_pool': 'default.rgw.control', 'dedup_pool': 'default.rgw.dedup', 'gc_pool': 'default.rgw.log:gc', 'lc_pool': 'default.rgw.log:lc', 'log_pool': 'default.rgw.log', 'intent_log_pool': 'default.rgw.log:intent', 'usage_log_pool': 'default.rgw.log:usage', 'roles_pool': 'default.rgw.meta:roles', 'reshard_pool': 'default.rgw.log:reshard', 'user_keys_pool': 'default.rgw.meta:users.keys', 'user_email_pool': 'default.rgw.meta:users.email', 'user_swift_pool': 'default.rgw.meta:users.swift', 'user_uid_pool': 'default.rgw.meta:users.uid', 'otp_pool': 'default.rgw.otp', 'notif_pool': 'default.rgw.log:notif', 'topics_pool': 'default.rgw.meta:topics', 'account_pool': 'default.rgw.meta:accounts', 'group_pool': 'default.rgw.meta:groups', 'system_key': {'access_key': '', 'secret_key': ''}, 'placement_pools': [{'key': 'default-placement', 'val': {'index_pool': 'default.rgw.buckets.index', 'storage_classes': {'STANDARD': {'data_pool': 'default.rgw.buckets.data', 'compression_type': 'random'}}, 'data_extra_pool': 'default.rgw.buckets.non-ec', 'index_type': 0, 'inline_data': True}}], 'realm_id': '', 'restore_pool': 'default.rgw.log:restore'} 2026-03-20T17:58:12.791 INFO:tasks.util.rgw:rgwadmin: c1.client.1 : ['user', 'list'] 2026-03-20T17:58:12.791 DEBUG:tasks.util.rgw:rgwadmin: cmd=['adjust-ulimits', 'ceph-coverage', '/home/ubuntu/cephtest/archive/coverage', 'radosgw-admin', '--log-to-stderr', '--format', 'json', '-n', 'client.1', '--cluster', 'c1', 'user', 'list'] 2026-03-20T17:58:12.791 DEBUG:teuthology.orchestra.run.vm05:> adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage radosgw-admin --log-to-stderr --format json -n client.1 --cluster c1 user list 2026-03-20T17:58:12.811 INFO:teuthology.orchestra.run.vm05.stderr:ignoring --setuser ceph since I am not root 2026-03-20T17:58:12.811 INFO:teuthology.orchestra.run.vm05.stderr:ignoring --setgroup ceph since I am not root 2026-03-20T17:58:12.829 INFO:teuthology.orchestra.run.vm05.stderr:2026-03-20T17:58:12.825+0000 7f6863796a80 20 rados->read ofs=0 len=0 2026-03-20T17:58:12.830 INFO:teuthology.orchestra.run.vm05.stderr:2026-03-20T17:58:12.825+0000 7f6863796a80 20 rados_obj.operate() r=-2 bl.length=0 2026-03-20T17:58:12.830 INFO:teuthology.orchestra.run.vm05.stderr:2026-03-20T17:58:12.825+0000 7f6863796a80 20 realm 2026-03-20T17:58:12.830 INFO:teuthology.orchestra.run.vm05.stderr:2026-03-20T17:58:12.825+0000 7f6863796a80 20 rados->read ofs=0 len=0 2026-03-20T17:58:12.830 INFO:teuthology.orchestra.run.vm05.stderr:2026-03-20T17:58:12.825+0000 7f6863796a80 20 rados_obj.operate() r=-2 bl.length=0 2026-03-20T17:58:12.830 INFO:teuthology.orchestra.run.vm05.stderr:2026-03-20T17:58:12.825+0000 7f6863796a80 4 RGWPeriod::init failed to init realm id : (2) No such file or directory 2026-03-20T17:58:12.830 INFO:teuthology.orchestra.run.vm05.stderr:2026-03-20T17:58:12.825+0000 7f6863796a80 20 rados->read ofs=0 len=0 2026-03-20T17:58:12.830 INFO:teuthology.orchestra.run.vm05.stderr:2026-03-20T17:58:12.825+0000 7f6863796a80 20 rados_obj.operate() r=-2 bl.length=0 2026-03-20T17:58:12.830 INFO:teuthology.orchestra.run.vm05.stderr:2026-03-20T17:58:12.825+0000 7f6863796a80 20 rados->read ofs=0 len=0 2026-03-20T17:58:12.831 INFO:teuthology.orchestra.run.vm05.stderr:2026-03-20T17:58:12.825+0000 7f6863796a80 20 rados_obj.operate() r=0 bl.length=46 2026-03-20T17:58:12.831 INFO:teuthology.orchestra.run.vm05.stderr:2026-03-20T17:58:12.825+0000 7f6863796a80 20 rados->read ofs=0 len=0 2026-03-20T17:58:12.832 INFO:teuthology.orchestra.run.vm05.stderr:2026-03-20T17:58:12.825+0000 7f6863796a80 20 rados_obj.operate() r=0 bl.length=1076 2026-03-20T17:58:12.832 INFO:teuthology.orchestra.run.vm05.stderr:2026-03-20T17:58:12.825+0000 7f6863796a80 20 searching for the correct realm 2026-03-20T17:58:12.839 INFO:teuthology.orchestra.run.vm05.stderr:2026-03-20T17:58:12.833+0000 7f6863796a80 20 RGWRados::pool_iterate: got zonegroup_info.b24151bb-5fb7-4fb5-a2d6-5d2063bc9ebe 2026-03-20T17:58:12.839 INFO:teuthology.orchestra.run.vm05.stderr:2026-03-20T17:58:12.833+0000 7f6863796a80 20 RGWRados::pool_iterate: got default.zonegroup. 2026-03-20T17:58:12.840 INFO:teuthology.orchestra.run.vm05.stderr:2026-03-20T17:58:12.833+0000 7f6863796a80 20 RGWRados::pool_iterate: got default.zone. 2026-03-20T17:58:12.840 INFO:teuthology.orchestra.run.vm05.stderr:2026-03-20T17:58:12.833+0000 7f6863796a80 20 RGWRados::pool_iterate: got zone_info.bd0a7e49-2751-4046-9c81-a69cd7a6f5e0 2026-03-20T17:58:12.840 INFO:teuthology.orchestra.run.vm05.stderr:2026-03-20T17:58:12.833+0000 7f6863796a80 20 RGWRados::pool_iterate: got zone_names.default 2026-03-20T17:58:12.840 INFO:teuthology.orchestra.run.vm05.stderr:2026-03-20T17:58:12.833+0000 7f6863796a80 20 RGWRados::pool_iterate: got zonegroups_names.default 2026-03-20T17:58:12.840 INFO:teuthology.orchestra.run.vm05.stderr:2026-03-20T17:58:12.833+0000 7f6863796a80 20 rados->read ofs=0 len=0 2026-03-20T17:58:12.840 INFO:teuthology.orchestra.run.vm05.stderr:2026-03-20T17:58:12.837+0000 7f6863796a80 20 rados_obj.operate() r=-2 bl.length=0 2026-03-20T17:58:12.840 INFO:teuthology.orchestra.run.vm05.stderr:2026-03-20T17:58:12.837+0000 7f6863796a80 20 rados->read ofs=0 len=0 2026-03-20T17:58:12.840 INFO:teuthology.orchestra.run.vm05.stderr:2026-03-20T17:58:12.837+0000 7f6863796a80 20 rados_obj.operate() r=0 bl.length=46 2026-03-20T17:58:12.840 INFO:teuthology.orchestra.run.vm05.stderr:2026-03-20T17:58:12.837+0000 7f6863796a80 20 rados->read ofs=0 len=0 2026-03-20T17:58:12.840 INFO:teuthology.orchestra.run.vm05.stderr:2026-03-20T17:58:12.837+0000 7f6863796a80 20 rados_obj.operate() r=0 bl.length=436 2026-03-20T17:58:12.840 INFO:teuthology.orchestra.run.vm05.stderr:2026-03-20T17:58:12.837+0000 7f6863796a80 20 zone default found 2026-03-20T17:58:12.840 INFO:teuthology.orchestra.run.vm05.stderr:2026-03-20T17:58:12.837+0000 7f6863796a80 4 Realm: () 2026-03-20T17:58:12.840 INFO:teuthology.orchestra.run.vm05.stderr:2026-03-20T17:58:12.837+0000 7f6863796a80 4 ZoneGroup: default (b24151bb-5fb7-4fb5-a2d6-5d2063bc9ebe) 2026-03-20T17:58:12.840 INFO:teuthology.orchestra.run.vm05.stderr:2026-03-20T17:58:12.837+0000 7f6863796a80 4 Zone: default (bd0a7e49-2751-4046-9c81-a69cd7a6f5e0) 2026-03-20T17:58:12.840 INFO:teuthology.orchestra.run.vm05.stderr:2026-03-20T17:58:12.837+0000 7f6863796a80 10 cannot find current period zonegroup using local zonegroup configuration 2026-03-20T17:58:12.840 INFO:teuthology.orchestra.run.vm05.stderr:2026-03-20T17:58:12.837+0000 7f6863796a80 20 zonegroup default 2026-03-20T17:58:12.840 INFO:teuthology.orchestra.run.vm05.stderr:2026-03-20T17:58:12.837+0000 7f6863796a80 20 rados->read ofs=0 len=0 2026-03-20T17:58:12.840 INFO:teuthology.orchestra.run.vm05.stderr:2026-03-20T17:58:12.837+0000 7f6863796a80 20 rados_obj.operate() r=-2 bl.length=0 2026-03-20T17:58:12.840 INFO:teuthology.orchestra.run.vm05.stderr:2026-03-20T17:58:12.837+0000 7f6863796a80 20 rados->read ofs=0 len=0 2026-03-20T17:58:12.841 INFO:teuthology.orchestra.run.vm05.stderr:2026-03-20T17:58:12.837+0000 7f6863796a80 20 rados_obj.operate() r=-2 bl.length=0 2026-03-20T17:58:12.841 INFO:teuthology.orchestra.run.vm05.stderr:2026-03-20T17:58:12.837+0000 7f6863796a80 20 rados->read ofs=0 len=0 2026-03-20T17:58:12.841 INFO:teuthology.orchestra.run.vm05.stderr:2026-03-20T17:58:12.837+0000 7f6863796a80 20 rados_obj.operate() r=-2 bl.length=0 2026-03-20T17:58:12.841 INFO:teuthology.orchestra.run.vm05.stderr:2026-03-20T17:58:12.837+0000 7f6863796a80 20 started sync module instance, tier type = 2026-03-20T17:58:12.841 INFO:teuthology.orchestra.run.vm05.stderr:2026-03-20T17:58:12.837+0000 7f6863796a80 20 started zone id=bd0a7e49-2751-4046-9c81-a69cd7a6f5e0 (name=default) with tier type = 2026-03-20T17:58:12.843 INFO:teuthology.orchestra.run.vm05.stderr:2026-03-20T17:58:12.837+0000 7f6863796a80 20 add_watcher() i=0 2026-03-20T17:58:12.844 INFO:teuthology.orchestra.run.vm05.stderr:2026-03-20T17:58:12.841+0000 7f6863796a80 20 add_watcher() i=2 2026-03-20T17:58:12.845 INFO:teuthology.orchestra.run.vm05.stderr:2026-03-20T17:58:12.841+0000 7f6863796a80 20 add_watcher() i=5 2026-03-20T17:58:12.845 INFO:teuthology.orchestra.run.vm05.stderr:2026-03-20T17:58:12.841+0000 7f6863796a80 20 add_watcher() i=3 2026-03-20T17:58:12.845 INFO:teuthology.orchestra.run.vm05.stderr:2026-03-20T17:58:12.841+0000 7f6863796a80 20 add_watcher() i=4 2026-03-20T17:58:12.845 INFO:teuthology.orchestra.run.vm05.stderr:2026-03-20T17:58:12.841+0000 7f6863796a80 20 add_watcher() i=1 2026-03-20T17:58:12.845 INFO:teuthology.orchestra.run.vm05.stderr:2026-03-20T17:58:12.841+0000 7f6863796a80 20 add_watcher() i=7 2026-03-20T17:58:12.845 INFO:teuthology.orchestra.run.vm05.stderr:2026-03-20T17:58:12.841+0000 7f6863796a80 20 add_watcher() i=6 2026-03-20T17:58:12.845 INFO:teuthology.orchestra.run.vm05.stderr:2026-03-20T17:58:12.841+0000 7f6863796a80 2 all 8 watchers are set, enabling cache 2026-03-20T17:58:12.847 INFO:teuthology.orchestra.run.vm05.stderr:2026-03-20T17:58:12.841+0000 7f6863796a80 20 rgw_check_secure_mon_conn(): auth registy supported: methods=[2] modes=[2,1] 2026-03-20T17:58:12.847 INFO:teuthology.orchestra.run.vm05.stderr:2026-03-20T17:58:12.841+0000 7f6863796a80 20 rgw_check_secure_mon_conn(): mode 1 is insecure 2026-03-20T17:58:12.847 INFO:teuthology.orchestra.run.vm05.stderr:2026-03-20T17:58:12.841+0000 7f6863796a80 5 note: GC not initialized 2026-03-20T17:58:12.847 INFO:teuthology.orchestra.run.vm05.stderr:2026-03-20T17:58:12.841+0000 7f680efe5640 20 reqs_thread_entry: start 2026-03-20T17:58:12.886 INFO:teuthology.orchestra.run.vm05.stderr:2026-03-20T17:58:12.881+0000 7f6863796a80 20 init_complete bucket index max shards: 11 2026-03-20T17:58:12.887 INFO:teuthology.orchestra.run.vm05.stderr:2026-03-20T17:58:12.881+0000 7f6863796a80 20 Filter name: none 2026-03-20T17:58:12.887 INFO:teuthology.orchestra.run.vm05.stderr:2026-03-20T17:58:12.881+0000 7f680cfe1640 20 reqs_thread_entry: start 2026-03-20T17:58:12.897 INFO:teuthology.orchestra.run.vm05.stderr:2026-03-20T17:58:12.893+0000 7f6863796a80 20 remove_watcher() i=0 2026-03-20T17:58:12.897 INFO:teuthology.orchestra.run.vm05.stderr:2026-03-20T17:58:12.893+0000 7f6863796a80 2 removed watcher, disabling cache 2026-03-20T17:58:12.897 INFO:teuthology.orchestra.run.vm05.stderr:2026-03-20T17:58:12.893+0000 7f6863796a80 20 remove_watcher() i=3 2026-03-20T17:58:12.897 INFO:teuthology.orchestra.run.vm05.stderr:2026-03-20T17:58:12.893+0000 7f6863796a80 20 remove_watcher() i=6 2026-03-20T17:58:12.897 INFO:teuthology.orchestra.run.vm05.stderr:2026-03-20T17:58:12.893+0000 7f6863796a80 20 remove_watcher() i=5 2026-03-20T17:58:12.897 INFO:teuthology.orchestra.run.vm05.stderr:2026-03-20T17:58:12.893+0000 7f6863796a80 20 remove_watcher() i=4 2026-03-20T17:58:12.897 INFO:teuthology.orchestra.run.vm05.stderr:2026-03-20T17:58:12.893+0000 7f6863796a80 20 remove_watcher() i=1 2026-03-20T17:58:12.897 INFO:teuthology.orchestra.run.vm05.stderr:2026-03-20T17:58:12.893+0000 7f6863796a80 20 remove_watcher() i=2 2026-03-20T17:58:12.897 INFO:teuthology.orchestra.run.vm05.stderr:2026-03-20T17:58:12.893+0000 7f6863796a80 20 remove_watcher() i=7 2026-03-20T17:58:12.903 INFO:teuthology.orchestra.run.vm05.stdout:[] 2026-03-20T17:58:12.903 DEBUG:tasks.util.rgw: json result: [] 2026-03-20T17:58:12.903 INFO:tasks.util.rgw:rgwadmin: c1.client.1 : ['zone', 'placement', 'modify', '--rgw-zone', 'default', '--placement-id', 'default-placement', '--compression', 'random'] 2026-03-20T17:58:12.903 DEBUG:tasks.util.rgw:rgwadmin: cmd=['adjust-ulimits', 'ceph-coverage', '/home/ubuntu/cephtest/archive/coverage', 'radosgw-admin', '--log-to-stderr', '--format', 'json', '-n', 'client.1', '--cluster', 'c1', 'zone', 'placement', 'modify', '--rgw-zone', 'default', '--placement-id', 'default-placement', '--compression', 'random'] 2026-03-20T17:58:12.904 DEBUG:teuthology.orchestra.run.vm05:> adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage radosgw-admin --log-to-stderr --format json -n client.1 --cluster c1 zone placement modify --rgw-zone default --placement-id default-placement --compression random 2026-03-20T17:58:12.966 INFO:teuthology.orchestra.run.vm05.stderr:ignoring --setuser ceph since I am not root 2026-03-20T17:58:12.966 INFO:teuthology.orchestra.run.vm05.stderr:ignoring --setgroup ceph since I am not root 2026-03-20T17:58:12.981 INFO:teuthology.orchestra.run.vm05.stderr:2026-03-20T17:58:12.977+0000 7fdc3d542a80 20 rgw_check_secure_mon_conn(): auth registy supported: methods=[2] modes=[2,1] 2026-03-20T17:58:12.981 INFO:teuthology.orchestra.run.vm05.stderr:2026-03-20T17:58:12.977+0000 7fdc3d542a80 20 rgw_check_secure_mon_conn(): mode 1 is insecure 2026-03-20T17:58:12.981 INFO:teuthology.orchestra.run.vm05.stderr:2026-03-20T17:58:12.977+0000 7fdbe97e2640 20 reqs_thread_entry: start 2026-03-20T17:58:12.993 INFO:teuthology.orchestra.run.vm05.stdout:{"id":"bd0a7e49-2751-4046-9c81-a69cd7a6f5e0","name":"default","domain_root":"default.rgw.meta:root","control_pool":"default.rgw.control","dedup_pool":"default.rgw.dedup","gc_pool":"default.rgw.log:gc","lc_pool":"default.rgw.log:lc","log_pool":"default.rgw.log","intent_log_pool":"default.rgw.log:intent","usage_log_pool":"default.rgw.log:usage","roles_pool":"default.rgw.meta:roles","reshard_pool":"default.rgw.log:reshard","user_keys_pool":"default.rgw.meta:users.keys","user_email_pool":"default.rgw.meta:users.email","user_swift_pool":"default.rgw.meta:users.swift","user_uid_pool":"default.rgw.meta:users.uid","otp_pool":"default.rgw.otp","notif_pool":"default.rgw.log:notif","topics_pool":"default.rgw.meta:topics","account_pool":"default.rgw.meta:accounts","group_pool":"default.rgw.meta:groups","system_key":{"access_key":"","secret_key":""},"placement_pools":[{"key":"default-placement","val":{"index_pool":"default.rgw.buckets.index","storage_classes":{"STANDARD":{"data_pool":"default.rgw.buckets.data","compression_type":"random"}},"data_extra_pool":"default.rgw.buckets.non-ec","index_type":0,"inline_data":true}}],"realm_id":"","restore_pool":"default.rgw.log:restore"} 2026-03-20T17:58:12.993 DEBUG:tasks.util.rgw: json result: {'id': 'bd0a7e49-2751-4046-9c81-a69cd7a6f5e0', 'name': 'default', 'domain_root': 'default.rgw.meta:root', 'control_pool': 'default.rgw.control', 'dedup_pool': 'default.rgw.dedup', 'gc_pool': 'default.rgw.log:gc', 'lc_pool': 'default.rgw.log:lc', 'log_pool': 'default.rgw.log', 'intent_log_pool': 'default.rgw.log:intent', 'usage_log_pool': 'default.rgw.log:usage', 'roles_pool': 'default.rgw.meta:roles', 'reshard_pool': 'default.rgw.log:reshard', 'user_keys_pool': 'default.rgw.meta:users.keys', 'user_email_pool': 'default.rgw.meta:users.email', 'user_swift_pool': 'default.rgw.meta:users.swift', 'user_uid_pool': 'default.rgw.meta:users.uid', 'otp_pool': 'default.rgw.otp', 'notif_pool': 'default.rgw.log:notif', 'topics_pool': 'default.rgw.meta:topics', 'account_pool': 'default.rgw.meta:accounts', 'group_pool': 'default.rgw.meta:groups', 'system_key': {'access_key': '', 'secret_key': ''}, 'placement_pools': [{'key': 'default-placement', 'val': {'index_pool': 'default.rgw.buckets.index', 'storage_classes': {'STANDARD': {'data_pool': 'default.rgw.buckets.data', 'compression_type': 'random'}}, 'data_extra_pool': 'default.rgw.buckets.non-ec', 'index_type': 0, 'inline_data': True}}], 'realm_id': '', 'restore_pool': 'default.rgw.log:restore'} 2026-03-20T17:58:12.993 INFO:tasks.util.rgw:rgwadmin: c2.client.0 : ['user', 'list'] 2026-03-20T17:58:12.993 DEBUG:tasks.util.rgw:rgwadmin: cmd=['adjust-ulimits', 'ceph-coverage', '/home/ubuntu/cephtest/archive/coverage', 'radosgw-admin', '--log-to-stderr', '--format', 'json', '-n', 'client.0', '--cluster', 'c2', 'user', 'list'] 2026-03-20T17:58:12.993 DEBUG:teuthology.orchestra.run.vm06:> adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage radosgw-admin --log-to-stderr --format json -n client.0 --cluster c2 user list 2026-03-20T17:58:13.013 INFO:teuthology.orchestra.run.vm06.stderr:ignoring --setuser ceph since I am not root 2026-03-20T17:58:13.013 INFO:teuthology.orchestra.run.vm06.stderr:ignoring --setgroup ceph since I am not root 2026-03-20T17:58:14.615 INFO:teuthology.orchestra.run.vm06.stderr:2026-03-20T17:58:14.607+0000 7f63a5e66a80 20 rados->read ofs=0 len=0 2026-03-20T17:58:14.616 INFO:teuthology.orchestra.run.vm06.stderr:2026-03-20T17:58:14.611+0000 7f63a5e66a80 20 rados_obj.operate() r=-2 bl.length=0 2026-03-20T17:58:14.616 INFO:teuthology.orchestra.run.vm06.stderr:2026-03-20T17:58:14.611+0000 7f63a5e66a80 20 realm 2026-03-20T17:58:14.616 INFO:teuthology.orchestra.run.vm06.stderr:2026-03-20T17:58:14.611+0000 7f63a5e66a80 20 rados->read ofs=0 len=0 2026-03-20T17:58:14.617 INFO:teuthology.orchestra.run.vm06.stderr:2026-03-20T17:58:14.611+0000 7f63a5e66a80 20 rados_obj.operate() r=-2 bl.length=0 2026-03-20T17:58:14.617 INFO:teuthology.orchestra.run.vm06.stderr:2026-03-20T17:58:14.611+0000 7f63a5e66a80 4 RGWPeriod::init failed to init realm id : (2) No such file or directory 2026-03-20T17:58:14.617 INFO:teuthology.orchestra.run.vm06.stderr:2026-03-20T17:58:14.611+0000 7f63a5e66a80 20 rados->read ofs=0 len=0 2026-03-20T17:58:14.617 INFO:teuthology.orchestra.run.vm06.stderr:2026-03-20T17:58:14.611+0000 7f63a5e66a80 20 rados_obj.operate() r=-2 bl.length=0 2026-03-20T17:58:14.617 INFO:teuthology.orchestra.run.vm06.stderr:2026-03-20T17:58:14.611+0000 7f63a5e66a80 20 rados->read ofs=0 len=0 2026-03-20T17:58:14.617 INFO:teuthology.orchestra.run.vm06.stderr:2026-03-20T17:58:14.611+0000 7f63a5e66a80 20 rados_obj.operate() r=0 bl.length=46 2026-03-20T17:58:14.617 INFO:teuthology.orchestra.run.vm06.stderr:2026-03-20T17:58:14.611+0000 7f63a5e66a80 20 rados->read ofs=0 len=0 2026-03-20T17:58:14.617 INFO:teuthology.orchestra.run.vm06.stderr:2026-03-20T17:58:14.611+0000 7f63a5e66a80 20 rados_obj.operate() r=0 bl.length=1060 2026-03-20T17:58:14.618 INFO:teuthology.orchestra.run.vm06.stderr:2026-03-20T17:58:14.611+0000 7f63a5e66a80 20 searching for the correct realm 2026-03-20T17:58:14.626 INFO:teuthology.orchestra.run.vm06.stderr:2026-03-20T17:58:14.619+0000 7f63a5e66a80 20 RGWRados::pool_iterate: got zonegroup_info.1d07d9a5-4697-4059-9fa7-b607322d3e51 2026-03-20T17:58:14.626 INFO:teuthology.orchestra.run.vm06.stderr:2026-03-20T17:58:14.619+0000 7f63a5e66a80 20 RGWRados::pool_iterate: got zone_info.40b52331-3f86-4f4d-a1e5-7de64d326aa1 2026-03-20T17:58:14.626 INFO:teuthology.orchestra.run.vm06.stderr:2026-03-20T17:58:14.619+0000 7f63a5e66a80 20 RGWRados::pool_iterate: got default.zonegroup. 2026-03-20T17:58:14.626 INFO:teuthology.orchestra.run.vm06.stderr:2026-03-20T17:58:14.619+0000 7f63a5e66a80 20 RGWRados::pool_iterate: got default.zone. 2026-03-20T17:58:14.626 INFO:teuthology.orchestra.run.vm06.stderr:2026-03-20T17:58:14.619+0000 7f63a5e66a80 20 RGWRados::pool_iterate: got zone_names.default 2026-03-20T17:58:14.626 INFO:teuthology.orchestra.run.vm06.stderr:2026-03-20T17:58:14.619+0000 7f63a5e66a80 20 RGWRados::pool_iterate: got zonegroups_names.default 2026-03-20T17:58:14.626 INFO:teuthology.orchestra.run.vm06.stderr:2026-03-20T17:58:14.619+0000 7f63a5e66a80 20 rados->read ofs=0 len=0 2026-03-20T17:58:14.626 INFO:teuthology.orchestra.run.vm06.stderr:2026-03-20T17:58:14.619+0000 7f63a5e66a80 20 rados_obj.operate() r=-2 bl.length=0 2026-03-20T17:58:14.626 INFO:teuthology.orchestra.run.vm06.stderr:2026-03-20T17:58:14.619+0000 7f63a5e66a80 20 rados->read ofs=0 len=0 2026-03-20T17:58:14.626 INFO:teuthology.orchestra.run.vm06.stderr:2026-03-20T17:58:14.619+0000 7f63a5e66a80 20 rados_obj.operate() r=0 bl.length=46 2026-03-20T17:58:14.626 INFO:teuthology.orchestra.run.vm06.stderr:2026-03-20T17:58:14.619+0000 7f63a5e66a80 20 rados->read ofs=0 len=0 2026-03-20T17:58:14.627 INFO:teuthology.orchestra.run.vm06.stderr:2026-03-20T17:58:14.619+0000 7f63a5e66a80 20 rados_obj.operate() r=0 bl.length=436 2026-03-20T17:58:14.627 INFO:teuthology.orchestra.run.vm06.stderr:2026-03-20T17:58:14.619+0000 7f63a5e66a80 20 zone default found 2026-03-20T17:58:14.627 INFO:teuthology.orchestra.run.vm06.stderr:2026-03-20T17:58:14.619+0000 7f63a5e66a80 4 Realm: () 2026-03-20T17:58:14.627 INFO:teuthology.orchestra.run.vm06.stderr:2026-03-20T17:58:14.619+0000 7f63a5e66a80 4 ZoneGroup: default (1d07d9a5-4697-4059-9fa7-b607322d3e51) 2026-03-20T17:58:14.627 INFO:teuthology.orchestra.run.vm06.stderr:2026-03-20T17:58:14.619+0000 7f63a5e66a80 4 Zone: default (40b52331-3f86-4f4d-a1e5-7de64d326aa1) 2026-03-20T17:58:14.627 INFO:teuthology.orchestra.run.vm06.stderr:2026-03-20T17:58:14.619+0000 7f63a5e66a80 10 cannot find current period zonegroup using local zonegroup configuration 2026-03-20T17:58:14.627 INFO:teuthology.orchestra.run.vm06.stderr:2026-03-20T17:58:14.619+0000 7f63a5e66a80 20 zonegroup default 2026-03-20T17:58:14.627 INFO:teuthology.orchestra.run.vm06.stderr:2026-03-20T17:58:14.619+0000 7f63a5e66a80 20 rados->read ofs=0 len=0 2026-03-20T17:58:14.627 INFO:teuthology.orchestra.run.vm06.stderr:2026-03-20T17:58:14.619+0000 7f63a5e66a80 20 rados_obj.operate() r=-2 bl.length=0 2026-03-20T17:58:14.627 INFO:teuthology.orchestra.run.vm06.stderr:2026-03-20T17:58:14.619+0000 7f63a5e66a80 20 rados->read ofs=0 len=0 2026-03-20T17:58:16.616 INFO:teuthology.orchestra.run.vm06.stderr:2026-03-20T17:58:16.611+0000 7f63a5e66a80 20 rados_obj.operate() r=-2 bl.length=0 2026-03-20T17:58:16.617 INFO:teuthology.orchestra.run.vm06.stderr:2026-03-20T17:58:16.611+0000 7f63a5e66a80 20 rados->read ofs=0 len=0 2026-03-20T17:58:16.617 INFO:teuthology.orchestra.run.vm06.stderr:2026-03-20T17:58:16.611+0000 7f63a5e66a80 20 rados_obj.operate() r=-2 bl.length=0 2026-03-20T17:58:16.617 INFO:teuthology.orchestra.run.vm06.stderr:2026-03-20T17:58:16.611+0000 7f63a5e66a80 20 started sync module instance, tier type = 2026-03-20T17:58:16.617 INFO:teuthology.orchestra.run.vm06.stderr:2026-03-20T17:58:16.611+0000 7f63a5e66a80 20 started zone id=40b52331-3f86-4f4d-a1e5-7de64d326aa1 (name=default) with tier type = 2026-03-20T17:58:18.599 INFO:teuthology.orchestra.run.vm06.stderr:2026-03-20T17:58:18.591+0000 7f63a5e66a80 20 add_watcher() i=6 2026-03-20T17:58:18.599 INFO:teuthology.orchestra.run.vm06.stderr:2026-03-20T17:58:18.591+0000 7f63a5e66a80 20 add_watcher() i=7 2026-03-20T17:58:18.600 INFO:teuthology.orchestra.run.vm06.stderr:2026-03-20T17:58:18.591+0000 7f63a5e66a80 20 add_watcher() i=2 2026-03-20T17:58:18.600 INFO:teuthology.orchestra.run.vm06.stderr:2026-03-20T17:58:18.595+0000 7f63a5e66a80 20 add_watcher() i=1 2026-03-20T17:58:18.600 INFO:teuthology.orchestra.run.vm06.stderr:2026-03-20T17:58:18.595+0000 7f63a5e66a80 20 add_watcher() i=4 2026-03-20T17:58:18.603 INFO:teuthology.orchestra.run.vm06.stderr:2026-03-20T17:58:18.595+0000 7f63a5e66a80 20 add_watcher() i=5 2026-03-20T17:58:18.604 INFO:teuthology.orchestra.run.vm06.stderr:2026-03-20T17:58:18.599+0000 7f63a5e66a80 20 add_watcher() i=3 2026-03-20T17:58:18.606 INFO:teuthology.orchestra.run.vm06.stderr:2026-03-20T17:58:18.599+0000 7f63a5e66a80 20 add_watcher() i=0 2026-03-20T17:58:18.606 INFO:teuthology.orchestra.run.vm06.stderr:2026-03-20T17:58:18.599+0000 7f63a5e66a80 2 all 8 watchers are set, enabling cache 2026-03-20T17:58:18.609 INFO:teuthology.orchestra.run.vm06.stderr:2026-03-20T17:58:18.603+0000 7f639d7fa640 5 boost::asio::awaitable, obj_version> > logback_generations::read(const DoutPrefixProvider*):446: oid=data_loggenerations_metadata not found 2026-03-20T17:58:18.609 INFO:teuthology.orchestra.run.vm06.stderr:2026-03-20T17:58:18.603+0000 7f639d7fa640 20 boost::asio::awaitable {anonymous}::probe_shard(const DoutPrefixProvider*, neorados::RADOS, const neorados::Object&, const neorados::IOContext&, bool&):59 probing obj=data_log.0 2026-03-20T17:58:18.610 INFO:teuthology.orchestra.run.vm06.stderr:2026-03-20T17:58:18.603+0000 7f639d7fa640 20 do_open: entering 2026-03-20T17:58:18.610 INFO:teuthology.orchestra.run.vm06.stderr:2026-03-20T17:58:18.603+0000 7f639cff9640 20 boost::asio::awaitable {anonymous}::probe_shard(const DoutPrefixProvider*, neorados::RADOS, const neorados::Object&, const neorados::IOContext&, bool&):78: obj=data_log.0 does not exist 2026-03-20T17:58:18.610 INFO:teuthology.orchestra.run.vm06.stderr:2026-03-20T17:58:18.603+0000 7f639cff9640 20 boost::asio::awaitable {anonymous}::probe_shard(const DoutPrefixProvider*, neorados::RADOS, const neorados::Object&, const neorados::IOContext&, bool&):59 probing obj=data_log.1 2026-03-20T17:58:18.610 INFO:teuthology.orchestra.run.vm06.stderr:2026-03-20T17:58:18.603+0000 7f639cff9640 20 do_open: entering 2026-03-20T17:58:18.610 INFO:teuthology.orchestra.run.vm06.stderr:2026-03-20T17:58:18.603+0000 7f63977fe640 20 boost::asio::awaitable {anonymous}::probe_shard(const DoutPrefixProvider*, neorados::RADOS, const neorados::Object&, const neorados::IOContext&, bool&):78: obj=data_log.1 does not exist 2026-03-20T17:58:18.610 INFO:teuthology.orchestra.run.vm06.stderr:2026-03-20T17:58:18.603+0000 7f63977fe640 20 boost::asio::awaitable {anonymous}::probe_shard(const DoutPrefixProvider*, neorados::RADOS, const neorados::Object&, const neorados::IOContext&, bool&):59 probing obj=data_log.2 2026-03-20T17:58:18.610 INFO:teuthology.orchestra.run.vm06.stderr:2026-03-20T17:58:18.603+0000 7f63977fe640 20 do_open: entering 2026-03-20T17:58:18.611 INFO:teuthology.orchestra.run.vm06.stderr:2026-03-20T17:58:18.603+0000 7f63a4939640 20 boost::asio::awaitable {anonymous}::probe_shard(const DoutPrefixProvider*, neorados::RADOS, const neorados::Object&, const neorados::IOContext&, bool&):78: obj=data_log.2 does not exist 2026-03-20T17:58:18.611 INFO:teuthology.orchestra.run.vm06.stderr:2026-03-20T17:58:18.603+0000 7f63a4939640 20 boost::asio::awaitable {anonymous}::probe_shard(const DoutPrefixProvider*, neorados::RADOS, const neorados::Object&, const neorados::IOContext&, bool&):59 probing obj=data_log.3 2026-03-20T17:58:18.611 INFO:teuthology.orchestra.run.vm06.stderr:2026-03-20T17:58:18.603+0000 7f63a4939640 20 do_open: entering 2026-03-20T17:58:18.612 INFO:teuthology.orchestra.run.vm06.stderr:2026-03-20T17:58:18.607+0000 7f639effd640 20 boost::asio::awaitable {anonymous}::probe_shard(const DoutPrefixProvider*, neorados::RADOS, const neorados::Object&, const neorados::IOContext&, bool&):78: obj=data_log.3 does not exist 2026-03-20T17:58:18.612 INFO:teuthology.orchestra.run.vm06.stderr:2026-03-20T17:58:18.607+0000 7f639effd640 20 do_create: entering 2026-03-20T17:58:18.613 INFO:teuthology.orchestra.run.vm06.stderr:2026-03-20T17:58:18.607+0000 7f639e7fc640 20 do_open: entering 2026-03-20T17:58:18.616 INFO:teuthology.orchestra.run.vm06.stderr:2026-03-20T17:58:18.611+0000 7f63a5e66a80 20 rgw_check_secure_mon_conn(): auth registy supported: methods=[2] modes=[2,1] 2026-03-20T17:58:18.616 INFO:teuthology.orchestra.run.vm06.stderr:2026-03-20T17:58:18.611+0000 7f63a5e66a80 20 rgw_check_secure_mon_conn(): mode 1 is insecure 2026-03-20T17:58:21.603 INFO:teuthology.orchestra.run.vm06.stderr:2026-03-20T17:58:21.595+0000 7f63a5e66a80 10 rgw_init_ioctx warning: failed to set recovery_priority on default.rgw.meta 2026-03-20T17:58:21.603 INFO:teuthology.orchestra.run.vm06.stderr:2026-03-20T17:58:21.595+0000 7f63a5e66a80 5 note: GC not initialized 2026-03-20T17:58:21.603 INFO:teuthology.orchestra.run.vm06.stderr:2026-03-20T17:58:21.595+0000 7f634dfe3640 20 reqs_thread_entry: start 2026-03-20T17:58:21.655 INFO:teuthology.orchestra.run.vm06.stderr:2026-03-20T17:58:21.647+0000 7f63a5e66a80 20 init_complete bucket index max shards: 11 2026-03-20T17:58:21.655 INFO:teuthology.orchestra.run.vm06.stderr:2026-03-20T17:58:21.647+0000 7f63a5e66a80 20 Filter name: none 2026-03-20T17:58:21.655 INFO:teuthology.orchestra.run.vm06.stderr:2026-03-20T17:58:21.647+0000 7f63477fe640 20 reqs_thread_entry: start 2026-03-20T17:58:21.666 INFO:teuthology.orchestra.run.vm06.stderr:2026-03-20T17:58:21.659+0000 7f63a5e66a80 20 remove_watcher() i=6 2026-03-20T17:58:21.666 INFO:teuthology.orchestra.run.vm06.stderr:2026-03-20T17:58:21.659+0000 7f63a5e66a80 2 removed watcher, disabling cache 2026-03-20T17:58:21.666 INFO:teuthology.orchestra.run.vm06.stderr:2026-03-20T17:58:21.659+0000 7f63a5e66a80 20 remove_watcher() i=0 2026-03-20T17:58:21.666 INFO:teuthology.orchestra.run.vm06.stderr:2026-03-20T17:58:21.659+0000 7f63a5e66a80 20 remove_watcher() i=3 2026-03-20T17:58:21.666 INFO:teuthology.orchestra.run.vm06.stderr:2026-03-20T17:58:21.659+0000 7f63a5e66a80 20 remove_watcher() i=1 2026-03-20T17:58:21.667 INFO:teuthology.orchestra.run.vm06.stderr:2026-03-20T17:58:21.659+0000 7f63a5e66a80 20 remove_watcher() i=2 2026-03-20T17:58:21.667 INFO:teuthology.orchestra.run.vm06.stderr:2026-03-20T17:58:21.659+0000 7f63a5e66a80 20 remove_watcher() i=5 2026-03-20T17:58:21.667 INFO:teuthology.orchestra.run.vm06.stderr:2026-03-20T17:58:21.659+0000 7f63a5e66a80 20 remove_watcher() i=4 2026-03-20T17:58:21.667 INFO:teuthology.orchestra.run.vm06.stderr:2026-03-20T17:58:21.659+0000 7f63a5e66a80 20 remove_watcher() i=7 2026-03-20T17:58:21.672 INFO:teuthology.orchestra.run.vm06.stdout:[] 2026-03-20T17:58:21.672 DEBUG:tasks.util.rgw: json result: [] 2026-03-20T17:58:21.672 INFO:tasks.util.rgw:rgwadmin: c2.client.0 : ['zone', 'placement', 'modify', '--rgw-zone', 'default', '--placement-id', 'default-placement', '--compression', 'random'] 2026-03-20T17:58:21.672 DEBUG:tasks.util.rgw:rgwadmin: cmd=['adjust-ulimits', 'ceph-coverage', '/home/ubuntu/cephtest/archive/coverage', 'radosgw-admin', '--log-to-stderr', '--format', 'json', '-n', 'client.0', '--cluster', 'c2', 'zone', 'placement', 'modify', '--rgw-zone', 'default', '--placement-id', 'default-placement', '--compression', 'random'] 2026-03-20T17:58:21.672 DEBUG:teuthology.orchestra.run.vm06:> adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage radosgw-admin --log-to-stderr --format json -n client.0 --cluster c2 zone placement modify --rgw-zone default --placement-id default-placement --compression random 2026-03-20T17:58:21.740 INFO:teuthology.orchestra.run.vm06.stderr:ignoring --setuser ceph since I am not root 2026-03-20T17:58:21.740 INFO:teuthology.orchestra.run.vm06.stderr:ignoring --setgroup ceph since I am not root 2026-03-20T17:58:21.753 INFO:teuthology.orchestra.run.vm06.stderr:2026-03-20T17:58:21.747+0000 7ffa15823a80 20 rgw_check_secure_mon_conn(): auth registy supported: methods=[2] modes=[2,1] 2026-03-20T17:58:21.753 INFO:teuthology.orchestra.run.vm06.stderr:2026-03-20T17:58:21.747+0000 7ffa15823a80 20 rgw_check_secure_mon_conn(): mode 1 is insecure 2026-03-20T17:58:21.753 INFO:teuthology.orchestra.run.vm06.stderr:2026-03-20T17:58:21.747+0000 7ff9c17e2640 20 reqs_thread_entry: start 2026-03-20T17:58:21.764 INFO:teuthology.orchestra.run.vm06.stdout:{"id":"40b52331-3f86-4f4d-a1e5-7de64d326aa1","name":"default","domain_root":"default.rgw.meta:root","control_pool":"default.rgw.control","dedup_pool":"default.rgw.dedup","gc_pool":"default.rgw.log:gc","lc_pool":"default.rgw.log:lc","log_pool":"default.rgw.log","intent_log_pool":"default.rgw.log:intent","usage_log_pool":"default.rgw.log:usage","roles_pool":"default.rgw.meta:roles","reshard_pool":"default.rgw.log:reshard","user_keys_pool":"default.rgw.meta:users.keys","user_email_pool":"default.rgw.meta:users.email","user_swift_pool":"default.rgw.meta:users.swift","user_uid_pool":"default.rgw.meta:users.uid","otp_pool":"default.rgw.otp","notif_pool":"default.rgw.log:notif","topics_pool":"default.rgw.meta:topics","account_pool":"default.rgw.meta:accounts","group_pool":"default.rgw.meta:groups","system_key":{"access_key":"","secret_key":""},"placement_pools":[{"key":"default-placement","val":{"index_pool":"default.rgw.buckets.index","storage_classes":{"STANDARD":{"data_pool":"default.rgw.buckets.data","compression_type":"random"}},"data_extra_pool":"default.rgw.buckets.non-ec","index_type":0,"inline_data":true}}],"realm_id":"","restore_pool":"default.rgw.log:restore"} 2026-03-20T17:58:21.764 DEBUG:tasks.util.rgw: json result: {'id': '40b52331-3f86-4f4d-a1e5-7de64d326aa1', 'name': 'default', 'domain_root': 'default.rgw.meta:root', 'control_pool': 'default.rgw.control', 'dedup_pool': 'default.rgw.dedup', 'gc_pool': 'default.rgw.log:gc', 'lc_pool': 'default.rgw.log:lc', 'log_pool': 'default.rgw.log', 'intent_log_pool': 'default.rgw.log:intent', 'usage_log_pool': 'default.rgw.log:usage', 'roles_pool': 'default.rgw.meta:roles', 'reshard_pool': 'default.rgw.log:reshard', 'user_keys_pool': 'default.rgw.meta:users.keys', 'user_email_pool': 'default.rgw.meta:users.email', 'user_swift_pool': 'default.rgw.meta:users.swift', 'user_uid_pool': 'default.rgw.meta:users.uid', 'otp_pool': 'default.rgw.otp', 'notif_pool': 'default.rgw.log:notif', 'topics_pool': 'default.rgw.meta:topics', 'account_pool': 'default.rgw.meta:accounts', 'group_pool': 'default.rgw.meta:groups', 'system_key': {'access_key': '', 'secret_key': ''}, 'placement_pools': [{'key': 'default-placement', 'val': {'index_pool': 'default.rgw.buckets.index', 'storage_classes': {'STANDARD': {'data_pool': 'default.rgw.buckets.data', 'compression_type': 'random'}}, 'data_extra_pool': 'default.rgw.buckets.non-ec', 'index_type': 0, 'inline_data': True}}], 'realm_id': '', 'restore_pool': 'default.rgw.log:restore'} 2026-03-20T17:58:21.764 INFO:tasks.util.rgw:rgwadmin: c2.client.1 : ['user', 'list'] 2026-03-20T17:58:21.764 DEBUG:tasks.util.rgw:rgwadmin: cmd=['adjust-ulimits', 'ceph-coverage', '/home/ubuntu/cephtest/archive/coverage', 'radosgw-admin', '--log-to-stderr', '--format', 'json', '-n', 'client.1', '--cluster', 'c2', 'user', 'list'] 2026-03-20T17:58:21.764 DEBUG:teuthology.orchestra.run.vm08:> adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage radosgw-admin --log-to-stderr --format json -n client.1 --cluster c2 user list 2026-03-20T17:58:21.784 INFO:teuthology.orchestra.run.vm08.stderr:ignoring --setuser ceph since I am not root 2026-03-20T17:58:21.784 INFO:teuthology.orchestra.run.vm08.stderr:ignoring --setgroup ceph since I am not root 2026-03-20T17:58:21.802 INFO:teuthology.orchestra.run.vm08.stderr:2026-03-20T17:58:21.801+0000 7f448aa18a80 20 rados->read ofs=0 len=0 2026-03-20T17:58:21.803 INFO:teuthology.orchestra.run.vm08.stderr:2026-03-20T17:58:21.801+0000 7f448aa18a80 20 rados_obj.operate() r=-2 bl.length=0 2026-03-20T17:58:21.803 INFO:teuthology.orchestra.run.vm08.stderr:2026-03-20T17:58:21.801+0000 7f448aa18a80 20 realm 2026-03-20T17:58:21.803 INFO:teuthology.orchestra.run.vm08.stderr:2026-03-20T17:58:21.801+0000 7f448aa18a80 20 rados->read ofs=0 len=0 2026-03-20T17:58:21.803 INFO:teuthology.orchestra.run.vm08.stderr:2026-03-20T17:58:21.801+0000 7f448aa18a80 20 rados_obj.operate() r=-2 bl.length=0 2026-03-20T17:58:21.803 INFO:teuthology.orchestra.run.vm08.stderr:2026-03-20T17:58:21.801+0000 7f448aa18a80 4 RGWPeriod::init failed to init realm id : (2) No such file or directory 2026-03-20T17:58:21.803 INFO:teuthology.orchestra.run.vm08.stderr:2026-03-20T17:58:21.801+0000 7f448aa18a80 20 rados->read ofs=0 len=0 2026-03-20T17:58:21.803 INFO:teuthology.orchestra.run.vm08.stderr:2026-03-20T17:58:21.801+0000 7f448aa18a80 20 rados_obj.operate() r=-2 bl.length=0 2026-03-20T17:58:21.803 INFO:teuthology.orchestra.run.vm08.stderr:2026-03-20T17:58:21.801+0000 7f448aa18a80 20 rados->read ofs=0 len=0 2026-03-20T17:58:21.804 INFO:teuthology.orchestra.run.vm08.stderr:2026-03-20T17:58:21.801+0000 7f448aa18a80 20 rados_obj.operate() r=0 bl.length=46 2026-03-20T17:58:21.804 INFO:teuthology.orchestra.run.vm08.stderr:2026-03-20T17:58:21.801+0000 7f448aa18a80 20 rados->read ofs=0 len=0 2026-03-20T17:58:21.804 INFO:teuthology.orchestra.run.vm08.stderr:2026-03-20T17:58:21.801+0000 7f448aa18a80 20 rados_obj.operate() r=0 bl.length=1076 2026-03-20T17:58:21.804 INFO:teuthology.orchestra.run.vm08.stderr:2026-03-20T17:58:21.801+0000 7f448aa18a80 20 searching for the correct realm 2026-03-20T17:58:21.812 INFO:teuthology.orchestra.run.vm08.stderr:2026-03-20T17:58:21.809+0000 7f448aa18a80 20 RGWRados::pool_iterate: got zonegroup_info.1d07d9a5-4697-4059-9fa7-b607322d3e51 2026-03-20T17:58:21.812 INFO:teuthology.orchestra.run.vm08.stderr:2026-03-20T17:58:21.809+0000 7f448aa18a80 20 RGWRados::pool_iterate: got zone_info.40b52331-3f86-4f4d-a1e5-7de64d326aa1 2026-03-20T17:58:21.812 INFO:teuthology.orchestra.run.vm08.stderr:2026-03-20T17:58:21.809+0000 7f448aa18a80 20 RGWRados::pool_iterate: got default.zonegroup. 2026-03-20T17:58:21.812 INFO:teuthology.orchestra.run.vm08.stderr:2026-03-20T17:58:21.809+0000 7f448aa18a80 20 RGWRados::pool_iterate: got default.zone. 2026-03-20T17:58:21.812 INFO:teuthology.orchestra.run.vm08.stderr:2026-03-20T17:58:21.809+0000 7f448aa18a80 20 RGWRados::pool_iterate: got zone_names.default 2026-03-20T17:58:21.812 INFO:teuthology.orchestra.run.vm08.stderr:2026-03-20T17:58:21.809+0000 7f448aa18a80 20 RGWRados::pool_iterate: got zonegroups_names.default 2026-03-20T17:58:21.812 INFO:teuthology.orchestra.run.vm08.stderr:2026-03-20T17:58:21.809+0000 7f448aa18a80 20 rados->read ofs=0 len=0 2026-03-20T17:58:21.812 INFO:teuthology.orchestra.run.vm08.stderr:2026-03-20T17:58:21.809+0000 7f448aa18a80 20 rados_obj.operate() r=-2 bl.length=0 2026-03-20T17:58:21.812 INFO:teuthology.orchestra.run.vm08.stderr:2026-03-20T17:58:21.809+0000 7f448aa18a80 20 rados->read ofs=0 len=0 2026-03-20T17:58:21.812 INFO:teuthology.orchestra.run.vm08.stderr:2026-03-20T17:58:21.809+0000 7f448aa18a80 20 rados_obj.operate() r=0 bl.length=46 2026-03-20T17:58:21.812 INFO:teuthology.orchestra.run.vm08.stderr:2026-03-20T17:58:21.809+0000 7f448aa18a80 20 rados->read ofs=0 len=0 2026-03-20T17:58:21.813 INFO:teuthology.orchestra.run.vm08.stderr:2026-03-20T17:58:21.809+0000 7f448aa18a80 20 rados_obj.operate() r=0 bl.length=436 2026-03-20T17:58:21.813 INFO:teuthology.orchestra.run.vm08.stderr:2026-03-20T17:58:21.809+0000 7f448aa18a80 20 zone default found 2026-03-20T17:58:21.813 INFO:teuthology.orchestra.run.vm08.stderr:2026-03-20T17:58:21.809+0000 7f448aa18a80 4 Realm: () 2026-03-20T17:58:21.813 INFO:teuthology.orchestra.run.vm08.stderr:2026-03-20T17:58:21.809+0000 7f448aa18a80 4 ZoneGroup: default (1d07d9a5-4697-4059-9fa7-b607322d3e51) 2026-03-20T17:58:21.813 INFO:teuthology.orchestra.run.vm08.stderr:2026-03-20T17:58:21.809+0000 7f448aa18a80 4 Zone: default (40b52331-3f86-4f4d-a1e5-7de64d326aa1) 2026-03-20T17:58:21.813 INFO:teuthology.orchestra.run.vm08.stderr:2026-03-20T17:58:21.809+0000 7f448aa18a80 10 cannot find current period zonegroup using local zonegroup configuration 2026-03-20T17:58:21.813 INFO:teuthology.orchestra.run.vm08.stderr:2026-03-20T17:58:21.809+0000 7f448aa18a80 20 zonegroup default 2026-03-20T17:58:21.813 INFO:teuthology.orchestra.run.vm08.stderr:2026-03-20T17:58:21.809+0000 7f448aa18a80 20 rados->read ofs=0 len=0 2026-03-20T17:58:21.813 INFO:teuthology.orchestra.run.vm08.stderr:2026-03-20T17:58:21.809+0000 7f448aa18a80 20 rados_obj.operate() r=-2 bl.length=0 2026-03-20T17:58:21.813 INFO:teuthology.orchestra.run.vm08.stderr:2026-03-20T17:58:21.809+0000 7f448aa18a80 20 rados->read ofs=0 len=0 2026-03-20T17:58:21.813 INFO:teuthology.orchestra.run.vm08.stderr:2026-03-20T17:58:21.809+0000 7f448aa18a80 20 rados_obj.operate() r=-2 bl.length=0 2026-03-20T17:58:21.813 INFO:teuthology.orchestra.run.vm08.stderr:2026-03-20T17:58:21.809+0000 7f448aa18a80 20 rados->read ofs=0 len=0 2026-03-20T17:58:21.813 INFO:teuthology.orchestra.run.vm08.stderr:2026-03-20T17:58:21.809+0000 7f448aa18a80 20 rados_obj.operate() r=-2 bl.length=0 2026-03-20T17:58:21.813 INFO:teuthology.orchestra.run.vm08.stderr:2026-03-20T17:58:21.809+0000 7f448aa18a80 20 started sync module instance, tier type = 2026-03-20T17:58:21.813 INFO:teuthology.orchestra.run.vm08.stderr:2026-03-20T17:58:21.809+0000 7f448aa18a80 20 started zone id=40b52331-3f86-4f4d-a1e5-7de64d326aa1 (name=default) with tier type = 2026-03-20T17:58:21.816 INFO:teuthology.orchestra.run.vm08.stderr:2026-03-20T17:58:21.813+0000 7f448aa18a80 20 add_watcher() i=7 2026-03-20T17:58:21.816 INFO:teuthology.orchestra.run.vm08.stderr:2026-03-20T17:58:21.813+0000 7f448aa18a80 20 add_watcher() i=6 2026-03-20T17:58:21.817 INFO:teuthology.orchestra.run.vm08.stderr:2026-03-20T17:58:21.813+0000 7f448aa18a80 20 add_watcher() i=2 2026-03-20T17:58:21.817 INFO:teuthology.orchestra.run.vm08.stderr:2026-03-20T17:58:21.813+0000 7f448aa18a80 20 add_watcher() i=3 2026-03-20T17:58:21.818 INFO:teuthology.orchestra.run.vm08.stderr:2026-03-20T17:58:21.813+0000 7f448aa18a80 20 add_watcher() i=5 2026-03-20T17:58:21.818 INFO:teuthology.orchestra.run.vm08.stderr:2026-03-20T17:58:21.813+0000 7f448aa18a80 20 add_watcher() i=4 2026-03-20T17:58:21.818 INFO:teuthology.orchestra.run.vm08.stderr:2026-03-20T17:58:21.813+0000 7f448aa18a80 20 add_watcher() i=1 2026-03-20T17:58:21.818 INFO:teuthology.orchestra.run.vm08.stderr:2026-03-20T17:58:21.813+0000 7f448aa18a80 20 add_watcher() i=0 2026-03-20T17:58:21.818 INFO:teuthology.orchestra.run.vm08.stderr:2026-03-20T17:58:21.813+0000 7f448aa18a80 2 all 8 watchers are set, enabling cache 2026-03-20T17:58:21.820 INFO:teuthology.orchestra.run.vm08.stderr:2026-03-20T17:58:21.817+0000 7f448aa18a80 20 rgw_check_secure_mon_conn(): auth registy supported: methods=[2] modes=[2,1] 2026-03-20T17:58:21.820 INFO:teuthology.orchestra.run.vm08.stderr:2026-03-20T17:58:21.817+0000 7f448aa18a80 20 rgw_check_secure_mon_conn(): mode 1 is insecure 2026-03-20T17:58:21.820 INFO:teuthology.orchestra.run.vm08.stderr:2026-03-20T17:58:21.817+0000 7f448aa18a80 5 note: GC not initialized 2026-03-20T17:58:21.820 INFO:teuthology.orchestra.run.vm08.stderr:2026-03-20T17:58:21.817+0000 7f44367dc640 20 reqs_thread_entry: start 2026-03-20T17:58:21.862 INFO:teuthology.orchestra.run.vm08.stderr:2026-03-20T17:58:21.857+0000 7f448aa18a80 20 init_complete bucket index max shards: 11 2026-03-20T17:58:21.862 INFO:teuthology.orchestra.run.vm08.stderr:2026-03-20T17:58:21.857+0000 7f448aa18a80 20 Filter name: none 2026-03-20T17:58:21.862 INFO:teuthology.orchestra.run.vm08.stderr:2026-03-20T17:58:21.857+0000 7f43e7fff640 20 reqs_thread_entry: start 2026-03-20T17:58:21.871 INFO:teuthology.orchestra.run.vm08.stderr:2026-03-20T17:58:21.869+0000 7f448aa18a80 20 remove_watcher() i=3 2026-03-20T17:58:21.871 INFO:teuthology.orchestra.run.vm08.stderr:2026-03-20T17:58:21.869+0000 7f448aa18a80 2 removed watcher, disabling cache 2026-03-20T17:58:21.871 INFO:teuthology.orchestra.run.vm08.stderr:2026-03-20T17:58:21.869+0000 7f448aa18a80 20 remove_watcher() i=4 2026-03-20T17:58:21.871 INFO:teuthology.orchestra.run.vm08.stderr:2026-03-20T17:58:21.869+0000 7f448aa18a80 20 remove_watcher() i=0 2026-03-20T17:58:21.871 INFO:teuthology.orchestra.run.vm08.stderr:2026-03-20T17:58:21.869+0000 7f448aa18a80 20 remove_watcher() i=6 2026-03-20T17:58:21.871 INFO:teuthology.orchestra.run.vm08.stderr:2026-03-20T17:58:21.869+0000 7f448aa18a80 20 remove_watcher() i=7 2026-03-20T17:58:21.871 INFO:teuthology.orchestra.run.vm08.stderr:2026-03-20T17:58:21.869+0000 7f448aa18a80 20 remove_watcher() i=2 2026-03-20T17:58:21.871 INFO:teuthology.orchestra.run.vm08.stderr:2026-03-20T17:58:21.869+0000 7f448aa18a80 20 remove_watcher() i=5 2026-03-20T17:58:21.871 INFO:teuthology.orchestra.run.vm08.stderr:2026-03-20T17:58:21.869+0000 7f448aa18a80 20 remove_watcher() i=1 2026-03-20T17:58:21.876 INFO:teuthology.orchestra.run.vm08.stdout:[] 2026-03-20T17:58:21.876 DEBUG:tasks.util.rgw: json result: [] 2026-03-20T17:58:21.876 INFO:tasks.util.rgw:rgwadmin: c2.client.1 : ['zone', 'placement', 'modify', '--rgw-zone', 'default', '--placement-id', 'default-placement', '--compression', 'random'] 2026-03-20T17:58:21.876 DEBUG:tasks.util.rgw:rgwadmin: cmd=['adjust-ulimits', 'ceph-coverage', '/home/ubuntu/cephtest/archive/coverage', 'radosgw-admin', '--log-to-stderr', '--format', 'json', '-n', 'client.1', '--cluster', 'c2', 'zone', 'placement', 'modify', '--rgw-zone', 'default', '--placement-id', 'default-placement', '--compression', 'random'] 2026-03-20T17:58:21.876 DEBUG:teuthology.orchestra.run.vm08:> adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage radosgw-admin --log-to-stderr --format json -n client.1 --cluster c2 zone placement modify --rgw-zone default --placement-id default-placement --compression random 2026-03-20T17:58:21.943 INFO:teuthology.orchestra.run.vm08.stderr:ignoring --setuser ceph since I am not root 2026-03-20T17:58:21.943 INFO:teuthology.orchestra.run.vm08.stderr:ignoring --setgroup ceph since I am not root 2026-03-20T17:58:21.958 INFO:teuthology.orchestra.run.vm08.stderr:2026-03-20T17:58:21.953+0000 7fd5b91cda80 20 rgw_check_secure_mon_conn(): auth registy supported: methods=[2] modes=[2,1] 2026-03-20T17:58:21.958 INFO:teuthology.orchestra.run.vm08.stderr:2026-03-20T17:58:21.953+0000 7fd5b91cda80 20 rgw_check_secure_mon_conn(): mode 1 is insecure 2026-03-20T17:58:21.959 INFO:teuthology.orchestra.run.vm08.stderr:2026-03-20T17:58:21.957+0000 7fd565fe3640 20 reqs_thread_entry: start 2026-03-20T17:58:21.969 INFO:teuthology.orchestra.run.vm08.stdout:{"id":"40b52331-3f86-4f4d-a1e5-7de64d326aa1","name":"default","domain_root":"default.rgw.meta:root","control_pool":"default.rgw.control","dedup_pool":"default.rgw.dedup","gc_pool":"default.rgw.log:gc","lc_pool":"default.rgw.log:lc","log_pool":"default.rgw.log","intent_log_pool":"default.rgw.log:intent","usage_log_pool":"default.rgw.log:usage","roles_pool":"default.rgw.meta:roles","reshard_pool":"default.rgw.log:reshard","user_keys_pool":"default.rgw.meta:users.keys","user_email_pool":"default.rgw.meta:users.email","user_swift_pool":"default.rgw.meta:users.swift","user_uid_pool":"default.rgw.meta:users.uid","otp_pool":"default.rgw.otp","notif_pool":"default.rgw.log:notif","topics_pool":"default.rgw.meta:topics","account_pool":"default.rgw.meta:accounts","group_pool":"default.rgw.meta:groups","system_key":{"access_key":"","secret_key":""},"placement_pools":[{"key":"default-placement","val":{"index_pool":"default.rgw.buckets.index","storage_classes":{"STANDARD":{"data_pool":"default.rgw.buckets.data","compression_type":"random"}},"data_extra_pool":"default.rgw.buckets.non-ec","index_type":0,"inline_data":true}}],"realm_id":"","restore_pool":"default.rgw.log:restore"} 2026-03-20T17:58:21.969 DEBUG:tasks.util.rgw: json result: {'id': '40b52331-3f86-4f4d-a1e5-7de64d326aa1', 'name': 'default', 'domain_root': 'default.rgw.meta:root', 'control_pool': 'default.rgw.control', 'dedup_pool': 'default.rgw.dedup', 'gc_pool': 'default.rgw.log:gc', 'lc_pool': 'default.rgw.log:lc', 'log_pool': 'default.rgw.log', 'intent_log_pool': 'default.rgw.log:intent', 'usage_log_pool': 'default.rgw.log:usage', 'roles_pool': 'default.rgw.meta:roles', 'reshard_pool': 'default.rgw.log:reshard', 'user_keys_pool': 'default.rgw.meta:users.keys', 'user_email_pool': 'default.rgw.meta:users.email', 'user_swift_pool': 'default.rgw.meta:users.swift', 'user_uid_pool': 'default.rgw.meta:users.uid', 'otp_pool': 'default.rgw.otp', 'notif_pool': 'default.rgw.log:notif', 'topics_pool': 'default.rgw.meta:topics', 'account_pool': 'default.rgw.meta:accounts', 'group_pool': 'default.rgw.meta:groups', 'system_key': {'access_key': '', 'secret_key': ''}, 'placement_pools': [{'key': 'default-placement', 'val': {'index_pool': 'default.rgw.buckets.index', 'storage_classes': {'STANDARD': {'data_pool': 'default.rgw.buckets.data', 'compression_type': 'random'}}, 'data_extra_pool': 'default.rgw.buckets.non-ec', 'index_type': 0, 'inline_data': True}}], 'realm_id': '', 'restore_pool': 'default.rgw.log:restore'} 2026-03-20T17:58:21.969 INFO:tasks.rgw:Starting rgw... 2026-03-20T17:58:21.969 INFO:tasks.rgw:rgw c1.client.0 config is {'port': 8000} 2026-03-20T17:58:21.969 INFO:tasks.rgw:Using beast as radosgw frontend 2026-03-20T17:58:21.969 DEBUG:teuthology.orchestra.run.vm02:> sudo echo -n http://vm02.local:8000 | sudo tee /home/ubuntu/cephtest/url_file 2026-03-20T17:58:21.980 INFO:teuthology.orchestra.run.vm02.stdout:http://vm02.local:8000 2026-03-20T17:58:21.980 DEBUG:teuthology.orchestra.run.vm02:> sudo chown ceph /home/ubuntu/cephtest/url_file 2026-03-20T17:58:22.034 INFO:tasks.rgw.c1.client.0:Restarting daemon 2026-03-20T17:58:22.034 DEBUG:teuthology.orchestra.run.vm02:> sudo adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage daemon-helper term radosgw --rgw-frontends 'beast port=8000' -n client.0 --cluster c1 -k /etc/ceph/c1.client.0.keyring --log-file /var/log/ceph/rgw.c1.client.0.log --rgw_ops_log_socket_path /home/ubuntu/cephtest/rgw.opslog.c1.client.0.sock --foreground | sudo tee /var/log/ceph/rgw.c1.client.0.stdout 2>&1 2026-03-20T17:58:22.075 INFO:tasks.rgw.c1.client.0:Started 2026-03-20T17:58:22.075 INFO:tasks.rgw:rgw c1.client.1 config is {'port': 8001} 2026-03-20T17:58:22.075 INFO:tasks.rgw:Using beast as radosgw frontend 2026-03-20T17:58:22.075 DEBUG:teuthology.orchestra.run.vm05:> sudo echo -n http://vm05.local:8001 | sudo tee /home/ubuntu/cephtest/url_file 2026-03-20T17:58:22.085 INFO:teuthology.orchestra.run.vm05.stdout:http://vm05.local:8001 2026-03-20T17:58:22.085 DEBUG:teuthology.orchestra.run.vm05:> sudo chown ceph /home/ubuntu/cephtest/url_file 2026-03-20T17:58:22.135 INFO:tasks.rgw.c1.client.1:Restarting daemon 2026-03-20T17:58:22.135 DEBUG:teuthology.orchestra.run.vm05:> sudo adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage daemon-helper term radosgw --rgw-frontends 'beast port=8001' -n client.1 --cluster c1 -k /etc/ceph/c1.client.1.keyring --log-file /var/log/ceph/rgw.c1.client.1.log --rgw_ops_log_socket_path /home/ubuntu/cephtest/rgw.opslog.c1.client.1.sock --foreground | sudo tee /var/log/ceph/rgw.c1.client.1.stdout 2>&1 2026-03-20T17:58:22.176 INFO:tasks.rgw.c1.client.1:Started 2026-03-20T17:58:22.176 INFO:tasks.rgw:rgw c2.client.0 config is {'port': 8000} 2026-03-20T17:58:22.176 INFO:tasks.rgw:Using beast as radosgw frontend 2026-03-20T17:58:22.176 DEBUG:teuthology.orchestra.run.vm06:> sudo echo -n http://vm06.local:8000 | sudo tee /home/ubuntu/cephtest/url_file 2026-03-20T17:58:22.187 INFO:teuthology.orchestra.run.vm06.stdout:http://vm06.local:8000 2026-03-20T17:58:22.187 DEBUG:teuthology.orchestra.run.vm06:> sudo chown ceph /home/ubuntu/cephtest/url_file 2026-03-20T17:58:22.237 INFO:tasks.rgw.c2.client.0:Restarting daemon 2026-03-20T17:58:22.237 DEBUG:teuthology.orchestra.run.vm06:> sudo adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage daemon-helper term radosgw --rgw-frontends 'beast port=8000' -n client.0 --cluster c2 -k /etc/ceph/c2.client.0.keyring --log-file /var/log/ceph/rgw.c2.client.0.log --rgw_ops_log_socket_path /home/ubuntu/cephtest/rgw.opslog.c2.client.0.sock --foreground | sudo tee /var/log/ceph/rgw.c2.client.0.stdout 2>&1 2026-03-20T17:58:22.286 INFO:tasks.rgw.c2.client.0:Started 2026-03-20T17:58:22.286 INFO:tasks.rgw:rgw c2.client.1 config is {'port': 8001} 2026-03-20T17:58:22.286 INFO:tasks.rgw:Using beast as radosgw frontend 2026-03-20T17:58:22.286 DEBUG:teuthology.orchestra.run.vm08:> sudo echo -n http://vm08.local:8001 | sudo tee /home/ubuntu/cephtest/url_file 2026-03-20T17:58:22.297 INFO:teuthology.orchestra.run.vm08.stdout:http://vm08.local:8001 2026-03-20T17:58:22.298 DEBUG:teuthology.orchestra.run.vm08:> sudo chown ceph /home/ubuntu/cephtest/url_file 2026-03-20T17:58:22.351 INFO:tasks.rgw.c2.client.1:Restarting daemon 2026-03-20T17:58:22.351 DEBUG:teuthology.orchestra.run.vm08:> sudo adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage daemon-helper term radosgw --rgw-frontends 'beast port=8001' -n client.1 --cluster c2 -k /etc/ceph/c2.client.1.keyring --log-file /var/log/ceph/rgw.c2.client.1.log --rgw_ops_log_socket_path /home/ubuntu/cephtest/rgw.opslog.c2.client.1.sock --foreground | sudo tee /var/log/ceph/rgw.c2.client.1.stdout 2>&1 2026-03-20T17:58:22.395 INFO:tasks.rgw.c2.client.1:Started 2026-03-20T17:58:22.395 INFO:tasks.rgw:Polling c1.client.0 until it starts accepting connections on http://vm02.local:8000/ 2026-03-20T17:58:22.395 DEBUG:teuthology.orchestra.run.vm02:> curl http://vm02.local:8000/ 2026-03-20T17:58:22.404 INFO:teuthology.orchestra.run.vm02.stderr: % Total % Received % Xferd Average Speed Time Time Time Current 2026-03-20T17:58:22.404 INFO:teuthology.orchestra.run.vm02.stderr: Dload Upload Total Spent Left Speed 2026-03-20T17:58:22.406 INFO:teuthology.orchestra.run.vm02.stderr: 0 0 0 0 0 0 0 0 --:--:-- --:--:-- --:--:-- 0 100 187 0 187 0 0 76576 0 --:--:-- --:--:-- --:--:-- 93500 2026-03-20T17:58:22.407 INFO:teuthology.orchestra.run.vm02.stdout:anonymous 2026-03-20T17:58:22.407 INFO:tasks.rgw:Polling c1.client.1 until it starts accepting connections on http://vm05.local:8001/ 2026-03-20T17:58:22.407 DEBUG:teuthology.orchestra.run.vm05:> curl http://vm05.local:8001/ 2026-03-20T17:58:22.432 INFO:teuthology.orchestra.run.vm05.stderr: % Total % Received % Xferd Average Speed Time Time Time Current 2026-03-20T17:58:22.432 INFO:teuthology.orchestra.run.vm05.stderr: Dload Upload Total Spent Left Speed 2026-03-20T17:58:22.434 INFO:teuthology.orchestra.run.vm05.stderr: 0 0 0 0 0 0 0 0 --:--:-- --:--:-- --:--:-- 0 0 0 0 0 0 0 0 0 --:--:-- --:--:-- --:--:-- 0 2026-03-20T17:58:22.434 INFO:teuthology.orchestra.run.vm05.stderr:curl: (7) Failed to connect to vm05.local port 8001 after 1 ms: Connection refused 2026-03-20T17:58:22.435 DEBUG:teuthology.orchestra.run:got remote process result: 7 2026-03-20T17:58:23.436 DEBUG:teuthology.orchestra.run.vm05:> curl http://vm05.local:8001/ 2026-03-20T17:58:23.442 INFO:teuthology.orchestra.run.vm05.stderr: % Total % Received % Xferd Average Speed Time Time Time Current 2026-03-20T17:58:23.442 INFO:teuthology.orchestra.run.vm05.stderr: Dload Upload Total Spent Left Speed 2026-03-20T17:58:23.445 INFO:teuthology.orchestra.run.vm05.stderr: 0 0 0 0 0 0 0 0 --:--:-- --:--:-- --:--:-- 0 100 187 0 187 0 0 92254 0 --:--:-- --:--:-- --:--:-- 182k 2026-03-20T17:58:23.445 INFO:teuthology.orchestra.run.vm05.stdout:anonymous 2026-03-20T17:58:23.445 INFO:tasks.rgw:Polling c2.client.0 until it starts accepting connections on http://vm06.local:8000/ 2026-03-20T17:58:23.445 DEBUG:teuthology.orchestra.run.vm06:> curl http://vm06.local:8000/ 2026-03-20T17:58:23.459 INFO:teuthology.orchestra.run.vm06.stderr: % Total % Received % Xferd Average Speed Time Time Time Current 2026-03-20T17:58:23.459 INFO:teuthology.orchestra.run.vm06.stderr: Dload Upload Total Spent Left Speed 2026-03-20T17:58:23.459 INFO:teuthology.orchestra.run.vm06.stderr: 0 0 0 0 0 0 0 0 --:--:-- --:--:-- --:--:-- 0 100 187 0 187 0 0 34642 0 --:--:-- --:--:-- --:--:-- 37400 2026-03-20T17:58:23.459 INFO:teuthology.orchestra.run.vm06.stdout:anonymous 2026-03-20T17:58:23.459 INFO:tasks.rgw:Polling c2.client.1 until it starts accepting connections on http://vm08.local:8001/ 2026-03-20T17:58:23.459 DEBUG:teuthology.orchestra.run.vm08:> curl http://vm08.local:8001/ 2026-03-20T17:58:23.470 INFO:teuthology.orchestra.run.vm08.stderr: % Total % Received % Xferd Average Speed Time Time Time Current 2026-03-20T17:58:23.470 INFO:teuthology.orchestra.run.vm08.stderr: Dload Upload Total Spent Left Speed 2026-03-20T17:58:23.473 INFO:teuthology.orchestra.run.vm08.stderr: 0 0 0 0 0 0 0 0 --:--:-- --:--:-- --:--:-- 0 100 187 0 187 0 0 31603 0 --:--:-- --:--:-- --:--:-- 37400 2026-03-20T17:58:23.473 INFO:teuthology.orchestra.run.vm08.stdout:anonymous 2026-03-20T17:58:23.473 INFO:teuthology.run_tasks:Running task rgw-multisite... 2026-03-20T17:58:23.534 INFO:tasks.rgw_multisite:creating realm.. 2026-03-20T17:58:23.534 INFO:tasks.util.rgw:rgwadmin: c1.client.0 : ['realm', 'create', '--rgw-realm', 'test-realm', '--default', '--cluster', 'c1', '--debug-rgw', '0', '--debug-ms', '0'] 2026-03-20T17:58:23.534 DEBUG:tasks.util.rgw:rgwadmin: cmd=['adjust-ulimits', 'ceph-coverage', '/home/ubuntu/cephtest/archive/coverage', 'radosgw-admin', '--log-to-stderr', '--format', 'json', '-n', 'client.0', '--cluster', 'c1', 'realm', 'create', '--rgw-realm', 'test-realm', '--default', '--cluster', 'c1', '--debug-rgw', '0', '--debug-ms', '0'] 2026-03-20T17:58:23.534 DEBUG:teuthology.orchestra.run.vm02:> adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage radosgw-admin --log-to-stderr --format json -n client.0 --cluster c1 realm create --rgw-realm test-realm --default --cluster c1 --debug-rgw 0 --debug-ms 0 2026-03-20T17:58:23.556 INFO:teuthology.orchestra.run.vm02.stderr:ignoring --setuser ceph since I am not root 2026-03-20T17:58:23.556 INFO:teuthology.orchestra.run.vm02.stderr:ignoring --setgroup ceph since I am not root 2026-03-20T17:58:23.587 INFO:teuthology.orchestra.run.vm02.stdout:{"id":"f5bc16fe-b476-42e1-93d5-47c86d0b7983","name":"test-realm","current_period":"6b3a6782-0294-444e-971d-8e9000c552d8","epoch":1} 2026-03-20T17:58:23.588 INFO:tasks.rgw_multisite:creating master zonegroup.. 2026-03-20T17:58:23.588 INFO:tasks.util.rgw:rgwadmin: c1.client.0 : ['zonegroup', 'set', '--rgw-zonegroup', 'a', '--rgw-realm', 'test-realm', '--default', '--cluster', 'c1', '--debug-rgw', '0', '--debug-ms', '0'] 2026-03-20T17:58:23.588 DEBUG:tasks.util.rgw:rgwadmin: cmd=['adjust-ulimits', 'ceph-coverage', '/home/ubuntu/cephtest/archive/coverage', 'radosgw-admin', '--log-to-stderr', '--format', 'json', '-n', 'client.0', '--cluster', 'c1', 'zonegroup', 'set', '--rgw-zonegroup', 'a', '--rgw-realm', 'test-realm', '--default', '--cluster', 'c1', '--debug-rgw', '0', '--debug-ms', '0'] 2026-03-20T17:58:23.588 DEBUG:teuthology.orchestra.run.vm02:> adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage radosgw-admin --log-to-stderr --format json -n client.0 --cluster c1 zonegroup set --rgw-zonegroup a --rgw-realm test-realm --default --cluster c1 --debug-rgw 0 --debug-ms 0 2026-03-20T17:58:23.650 INFO:teuthology.orchestra.run.vm02.stderr:ignoring --setuser ceph since I am not root 2026-03-20T17:58:23.651 INFO:teuthology.orchestra.run.vm02.stderr:ignoring --setgroup ceph since I am not root 2026-03-20T17:58:23.664 INFO:teuthology.orchestra.run.vm02.stderr:2026-03-20T17:58:23.661+0000 7f77510c2a80 -1 old format 2026-03-20T17:58:23.675 INFO:teuthology.orchestra.run.vm02.stdout:{"id":"a","name":"a","api_name":"a","is_master":true,"endpoints":["http://vm02.local:8000"],"hostnames":[],"hostnames_s3website":[],"master_zone":"","zones":[],"placement_targets":[{"name":"default-placement","tags":[],"storage_classes":[]}],"default_placement":"default-placement","realm_id":"f5bc16fe-b476-42e1-93d5-47c86d0b7983","sync_policy":{"groups":[]},"enabled_features":["notification_v2","resharding"]} 2026-03-20T17:58:23.675 INFO:tasks.rgw_multisite:creating master zone.. 2026-03-20T17:58:23.675 INFO:tasks.util.rgw:rgwadmin: c1.client.0 : ['zone', 'create', '--rgw-zone', 'a1', '--rgw-zonegroup', 'a', '--rgw-realm', 'test-realm', '--default', '--master', '--access-key', 'KLI823V2WXR7ZEZZ', '--secret', 'bfYGF3EVFszZr1T17wMbAaEQn3Tkb5fG', '--endpoints', 'http://vm02.local:8000', '--cluster', 'c1', '--debug-rgw', '0', '--debug-ms', '0'] 2026-03-20T17:58:23.675 DEBUG:tasks.util.rgw:rgwadmin: cmd=['adjust-ulimits', 'ceph-coverage', '/home/ubuntu/cephtest/archive/coverage', 'radosgw-admin', '--log-to-stderr', '--format', 'json', '-n', 'client.0', '--cluster', 'c1', 'zone', 'create', '--rgw-zone', 'a1', '--rgw-zonegroup', 'a', '--rgw-realm', 'test-realm', '--default', '--master', '--access-key', 'KLI823V2WXR7ZEZZ', '--secret', 'bfYGF3EVFszZr1T17wMbAaEQn3Tkb5fG', '--endpoints', 'http://vm02.local:8000', '--cluster', 'c1', '--debug-rgw', '0', '--debug-ms', '0'] 2026-03-20T17:58:23.675 DEBUG:teuthology.orchestra.run.vm02:> adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage radosgw-admin --log-to-stderr --format json -n client.0 --cluster c1 zone create --rgw-zone a1 --rgw-zonegroup a --rgw-realm test-realm --default --master --access-key KLI823V2WXR7ZEZZ --secret bfYGF3EVFszZr1T17wMbAaEQn3Tkb5fG --endpoints http://vm02.local:8000 --cluster c1 --debug-rgw 0 --debug-ms 0 2026-03-20T17:58:23.742 INFO:teuthology.orchestra.run.vm02.stderr:ignoring --setuser ceph since I am not root 2026-03-20T17:58:23.742 INFO:teuthology.orchestra.run.vm02.stderr:ignoring --setgroup ceph since I am not root 2026-03-20T17:58:23.757 INFO:teuthology.orchestra.run.vm02.stderr:NOTICE: set zone's realm_id=f5bc16fe-b476-42e1-93d5-47c86d0b7983 2026-03-20T17:58:23.777 INFO:teuthology.orchestra.run.vm02.stdout:{"id":"a5b5537f-2aa1-48e1-aabd-4c0e85b95fff","name":"a1","domain_root":"a1.rgw.meta:root","control_pool":"a1.rgw.control","dedup_pool":"a1.rgw.dedup","gc_pool":"a1.rgw.log:gc","lc_pool":"a1.rgw.log:lc","log_pool":"a1.rgw.log","intent_log_pool":"a1.rgw.log:intent","usage_log_pool":"a1.rgw.log:usage","roles_pool":"a1.rgw.meta:roles","reshard_pool":"a1.rgw.log:reshard","user_keys_pool":"a1.rgw.meta:users.keys","user_email_pool":"a1.rgw.meta:users.email","user_swift_pool":"a1.rgw.meta:users.swift","user_uid_pool":"a1.rgw.meta:users.uid","otp_pool":"a1.rgw.otp","notif_pool":"a1.rgw.log:notif","topics_pool":"a1.rgw.meta:topics","account_pool":"a1.rgw.meta:accounts","group_pool":"a1.rgw.meta:groups","system_key":{"access_key":"KLI823V2WXR7ZEZZ","secret_key":"bfYGF3EVFszZr1T17wMbAaEQn3Tkb5fG"},"placement_pools":[{"key":"default-placement","val":{"index_pool":"a1.rgw.buckets.index","storage_classes":{"STANDARD":{"data_pool":"a1.rgw.buckets.data"}},"data_extra_pool":"a1.rgw.buckets.non-ec","index_type":0,"inline_data":true}}],"realm_id":"f5bc16fe-b476-42e1-93d5-47c86d0b7983","restore_pool":"a1.rgw.log:restore"} 2026-03-20T17:58:23.777 DEBUG:teuthology.orchestra.run.vm02:> sudo ceph osd pool create a1.rgw.buckets.data 64 64 --cluster c1 2026-03-20T17:58:24.702 INFO:teuthology.orchestra.run.vm02.stderr:pool 'a1.rgw.buckets.data' created 2026-03-20T17:58:24.731 DEBUG:teuthology.orchestra.run.vm02:> sudo ceph osd pool application enable a1.rgw.buckets.data rgw --cluster c1 2026-03-20T17:58:25.707 INFO:teuthology.orchestra.run.vm02.stderr:enabled application 'rgw' on pool 'a1.rgw.buckets.data' 2026-03-20T17:58:25.743 INFO:tasks.util.rgw:rgwadmin: c1.client.0 : ['zone', 'placement', '--rgw-zone', 'a1', '--rgw-zonegroup', 'a', '--rgw-realm', 'test-realm', 'modify', '--placement-id', 'default-placement', '--compression', 'random', '--cluster', 'c1', '--debug-rgw', '0', '--debug-ms', '0'] 2026-03-20T17:58:25.743 DEBUG:tasks.util.rgw:rgwadmin: cmd=['adjust-ulimits', 'ceph-coverage', '/home/ubuntu/cephtest/archive/coverage', 'radosgw-admin', '--log-to-stderr', '--format', 'json', '-n', 'client.0', '--cluster', 'c1', 'zone', 'placement', '--rgw-zone', 'a1', '--rgw-zonegroup', 'a', '--rgw-realm', 'test-realm', 'modify', '--placement-id', 'default-placement', '--compression', 'random', '--cluster', 'c1', '--debug-rgw', '0', '--debug-ms', '0'] 2026-03-20T17:58:25.743 DEBUG:teuthology.orchestra.run.vm02:> adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage radosgw-admin --log-to-stderr --format json -n client.0 --cluster c1 zone placement --rgw-zone a1 --rgw-zonegroup a --rgw-realm test-realm modify --placement-id default-placement --compression random --cluster c1 --debug-rgw 0 --debug-ms 0 2026-03-20T17:58:25.768 INFO:teuthology.orchestra.run.vm02.stderr:ignoring --setuser ceph since I am not root 2026-03-20T17:58:25.768 INFO:teuthology.orchestra.run.vm02.stderr:ignoring --setgroup ceph since I am not root 2026-03-20T17:58:25.793 INFO:teuthology.orchestra.run.vm02.stdout:{"id":"a5b5537f-2aa1-48e1-aabd-4c0e85b95fff","name":"a1","domain_root":"a1.rgw.meta:root","control_pool":"a1.rgw.control","dedup_pool":"a1.rgw.dedup","gc_pool":"a1.rgw.log:gc","lc_pool":"a1.rgw.log:lc","log_pool":"a1.rgw.log","intent_log_pool":"a1.rgw.log:intent","usage_log_pool":"a1.rgw.log:usage","roles_pool":"a1.rgw.meta:roles","reshard_pool":"a1.rgw.log:reshard","user_keys_pool":"a1.rgw.meta:users.keys","user_email_pool":"a1.rgw.meta:users.email","user_swift_pool":"a1.rgw.meta:users.swift","user_uid_pool":"a1.rgw.meta:users.uid","otp_pool":"a1.rgw.otp","notif_pool":"a1.rgw.log:notif","topics_pool":"a1.rgw.meta:topics","account_pool":"a1.rgw.meta:accounts","group_pool":"a1.rgw.meta:groups","system_key":{"access_key":"KLI823V2WXR7ZEZZ","secret_key":"bfYGF3EVFszZr1T17wMbAaEQn3Tkb5fG"},"placement_pools":[{"key":"default-placement","val":{"index_pool":"a1.rgw.buckets.index","storage_classes":{"STANDARD":{"data_pool":"a1.rgw.buckets.data","compression_type":"random"}},"data_extra_pool":"a1.rgw.buckets.non-ec","index_type":0,"inline_data":true}}],"realm_id":"f5bc16fe-b476-42e1-93d5-47c86d0b7983","restore_pool":"a1.rgw.log:restore"} 2026-03-20T17:58:25.793 INFO:tasks.util.rgw:rgwadmin: c1.client.0 : ['period', 'update', '--rgw-zone', 'a1', '--rgw-zonegroup', 'a', '--rgw-realm', 'test-realm', '--commit', '--cluster', 'c1', '--debug-rgw', '0', '--debug-ms', '0'] 2026-03-20T17:58:25.793 DEBUG:tasks.util.rgw:rgwadmin: cmd=['adjust-ulimits', 'ceph-coverage', '/home/ubuntu/cephtest/archive/coverage', 'radosgw-admin', '--log-to-stderr', '--format', 'json', '-n', 'client.0', '--cluster', 'c1', 'period', 'update', '--rgw-zone', 'a1', '--rgw-zonegroup', 'a', '--rgw-realm', 'test-realm', '--commit', '--cluster', 'c1', '--debug-rgw', '0', '--debug-ms', '0'] 2026-03-20T17:58:25.793 DEBUG:teuthology.orchestra.run.vm02:> adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage radosgw-admin --log-to-stderr --format json -n client.0 --cluster c1 period update --rgw-zone a1 --rgw-zonegroup a --rgw-realm test-realm --commit --cluster c1 --debug-rgw 0 --debug-ms 0 2026-03-20T17:58:25.855 INFO:teuthology.orchestra.run.vm02.stderr:ignoring --setuser ceph since I am not root 2026-03-20T17:58:25.855 INFO:teuthology.orchestra.run.vm02.stderr:ignoring --setgroup ceph since I am not root 2026-03-20T17:58:25.877 INFO:teuthology.orchestra.run.vm02.stderr:2026-03-20T17:58:25.873+0000 7f9ab427ba80 0 period (6b3a6782-0294-444e-971d-8e9000c552d8 does not have zone a5b5537f-2aa1-48e1-aabd-4c0e85b95fff configured 2026-03-20T17:58:32.843 INFO:teuthology.orchestra.run.vm02.stdout:{"id":"82c662be-66b5-4c4f-b24e-ff9298fc8a6e","epoch":1,"predecessor_uuid":"6b3a6782-0294-444e-971d-8e9000c552d8","sync_status":[],"period_map":{"id":"82c662be-66b5-4c4f-b24e-ff9298fc8a6e","zonegroups":[{"id":"a","name":"a","api_name":"a","is_master":true,"endpoints":["http://vm02.local:8000"],"hostnames":[],"hostnames_s3website":[],"master_zone":"a5b5537f-2aa1-48e1-aabd-4c0e85b95fff","zones":[{"id":"a5b5537f-2aa1-48e1-aabd-4c0e85b95fff","name":"a1","endpoints":["http://vm02.local:8000"],"log_meta":false,"log_data":false,"bucket_index_max_shards":11,"read_only":false,"tier_type":"","sync_from_all":true,"sync_from":[],"redirect_zone":"","supported_features":["compress-encrypted","notification_v2","resharding"]}],"placement_targets":[{"name":"default-placement","tags":[],"storage_classes":["STANDARD"]}],"default_placement":"default-placement","realm_id":"f5bc16fe-b476-42e1-93d5-47c86d0b7983","sync_policy":{"groups":[]},"enabled_features":["notification_v2","resharding"]}],"short_zone_ids":[{"key":"a5b5537f-2aa1-48e1-aabd-4c0e85b95fff","val":3173326878}]},"master_zonegroup":"a","master_zone":"a5b5537f-2aa1-48e1-aabd-4c0e85b95fff","period_config":{"bucket_quota":{"enabled":false,"check_on_raw":false,"max_size":-1,"max_size_kb":0,"max_objects":-1},"user_quota":{"enabled":false,"check_on_raw":false,"max_size":-1,"max_size_kb":0,"max_objects":-1},"user_ratelimit":{"max_read_ops":0,"max_write_ops":0,"max_read_bytes":0,"max_write_bytes":0,"enabled":false},"bucket_ratelimit":{"max_read_ops":0,"max_write_ops":0,"max_read_bytes":0,"max_write_bytes":0,"enabled":false},"anonymous_ratelimit":{"max_read_ops":0,"max_write_ops":0,"max_read_bytes":0,"max_write_bytes":0,"enabled":false}},"realm_id":"f5bc16fe-b476-42e1-93d5-47c86d0b7983","realm_epoch":2} 2026-03-20T17:58:32.843 DEBUG:tasks.rgw.c1.client.0:waiting for process to exit 2026-03-20T17:58:32.843 INFO:teuthology.orchestra.run:waiting for 300 2026-03-20T17:58:32.887 INFO:tasks.rgw.c1.client.0.vm02.stdout:2026-03-20T17:58:32.885+0000 7fcd5242e640 -1 received signal: Terminated from /usr/bin/python3 /usr/bin/daemon-helper term radosgw --rgw-frontends beast port=8000 -n client.0 --cluster c1 -k /etc/ceph/c1.client.0.keyring --log-file /var/log/ceph/rgw.c1.client.0.log --rgw_ops_log_socket_path /home/ubuntu/cephtest/rgw.opslog.c1.client.0.sock --foreground (PID: 24133) UID: 0 2026-03-20T17:58:32.887 INFO:tasks.rgw.c1.client.0.vm02.stdout:2026-03-20T17:58:32.885+0000 7fcd55c76b40 -1 shutting down 2026-03-20T17:58:54.291 INFO:tasks.daemonwatchdog.daemon_watchdog:daemon c1.rgw.client.0 is failed for ~0s 2026-03-20T17:58:57.250 INFO:tasks.rgw.c1.client.0:Stopped 2026-03-20T17:58:57.251 INFO:tasks.rgw.c1.client.0:Restarting daemon 2026-03-20T17:58:57.251 DEBUG:teuthology.orchestra.run.vm02:> sudo adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage daemon-helper term radosgw --rgw-frontends 'beast port=8000' -n client.0 --cluster c1 -k /etc/ceph/c1.client.0.keyring --log-file /var/log/ceph/rgw.c1.client.0.log --rgw_ops_log_socket_path /home/ubuntu/cephtest/rgw.opslog.c1.client.0.sock --foreground --rgw-zone a1 --rgw-zonegroup a --rgw-realm test-realm | sudo tee /var/log/ceph/rgw.c1.client.0.stdout 2>&1 2026-03-20T17:58:57.252 INFO:tasks.rgw.c1.client.0:Started 2026-03-20T17:58:57.253 DEBUG:teuthology.orchestra.run.vm02:> curl http://vm02.local:8000 2026-03-20T17:58:57.260 INFO:teuthology.orchestra.run.vm02.stderr: % Total % Received % Xferd Average Speed Time Time Time Current 2026-03-20T17:58:57.260 INFO:teuthology.orchestra.run.vm02.stderr: Dload Upload Total Spent Left Speed 2026-03-20T17:58:57.260 INFO:teuthology.orchestra.run.vm02.stderr: 0 0 0 0 0 0 0 0 --:--:-- --:--:-- --:--:-- 0 0 0 0 0 0 0 0 0 --:--:-- --:--:-- --:--:-- 0 2026-03-20T17:58:57.261 INFO:teuthology.orchestra.run.vm02.stderr:curl: (7) Failed to connect to vm02.local port 8000 after 0 ms: Connection refused 2026-03-20T17:58:57.261 DEBUG:teuthology.orchestra.run:got remote process result: 7 2026-03-20T17:58:58.262 DEBUG:teuthology.orchestra.run.vm02:> curl http://vm02.local:8000 2026-03-20T17:58:58.274 INFO:teuthology.orchestra.run.vm02.stderr: % Total % Received % Xferd Average Speed Time Time Time Current 2026-03-20T17:58:58.274 INFO:teuthology.orchestra.run.vm02.stderr: Dload Upload Total Spent Left Speed 2026-03-20T17:58:58.276 INFO:teuthology.orchestra.run.vm02.stderr: 0 0 0 0 0 0 0 0 --:--:-- --:--:-- --:--:-- 0 100 187 0 187 0 0 33161 0 --:--:-- --:--:-- --:--:-- 37400 2026-03-20T17:58:58.277 INFO:teuthology.orchestra.run.vm02.stdout:anonymous 2026-03-20T17:58:58.277 INFO:tasks.rgw_multisite:creating admin user.. 2026-03-20T17:58:58.277 INFO:tasks.util.rgw:rgwadmin: c1.client.0 : ['user', 'create', '--uid', 'realm-admin', '--rgw-zone', 'a1', '--rgw-zonegroup', 'a', '--rgw-realm', 'test-realm', '--display-name', 'Realm Admin', '--system', '--access-key', 'KLI823V2WXR7ZEZZ', '--secret', 'bfYGF3EVFszZr1T17wMbAaEQn3Tkb5fG', '--cluster', 'c1', '--debug-rgw', '0', '--debug-ms', '0'] 2026-03-20T17:58:58.277 DEBUG:tasks.util.rgw:rgwadmin: cmd=['adjust-ulimits', 'ceph-coverage', '/home/ubuntu/cephtest/archive/coverage', 'radosgw-admin', '--log-to-stderr', '--format', 'json', '-n', 'client.0', '--cluster', 'c1', 'user', 'create', '--uid', 'realm-admin', '--rgw-zone', 'a1', '--rgw-zonegroup', 'a', '--rgw-realm', 'test-realm', '--display-name', 'Realm Admin', '--system', '--access-key', 'KLI823V2WXR7ZEZZ', '--secret', 'bfYGF3EVFszZr1T17wMbAaEQn3Tkb5fG', '--cluster', 'c1', '--debug-rgw', '0', '--debug-ms', '0'] 2026-03-20T17:58:58.277 DEBUG:teuthology.orchestra.run.vm02:> adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage radosgw-admin --log-to-stderr --format json -n client.0 --cluster c1 user create --uid realm-admin --rgw-zone a1 --rgw-zonegroup a --rgw-realm test-realm --display-name 'Realm Admin' --system --access-key KLI823V2WXR7ZEZZ --secret bfYGF3EVFszZr1T17wMbAaEQn3Tkb5fG --cluster c1 --debug-rgw 0 --debug-ms 0 2026-03-20T17:58:58.342 INFO:teuthology.orchestra.run.vm02.stderr:ignoring --setuser ceph since I am not root 2026-03-20T17:58:58.342 INFO:teuthology.orchestra.run.vm02.stderr:ignoring --setgroup ceph since I am not root 2026-03-20T17:58:58.420 INFO:teuthology.orchestra.run.vm02.stdout:{"user_id":"realm-admin","display_name":"Realm Admin","email":"","suspended":0,"max_buckets":1000,"subusers":[],"keys":[{"user":"realm-admin","access_key":"KLI823V2WXR7ZEZZ","secret_key":"bfYGF3EVFszZr1T17wMbAaEQn3Tkb5fG","active":true,"create_date":"2026-03-20T17:58:58.416271Z"}],"swift_keys":[],"caps":[],"op_mask":"read, write, delete","system":true,"default_placement":"","default_storage_class":"","placement_tags":[],"bucket_quota":{"enabled":false,"check_on_raw":false,"max_size":-1,"max_size_kb":0,"max_objects":-1},"user_quota":{"enabled":false,"check_on_raw":false,"max_size":-1,"max_size_kb":0,"max_objects":-1},"temp_url_keys":[],"type":"rgw","mfa_ids":[],"account_id":"","path":"/","create_date":"2026-03-20T17:58:58.416262Z","tags":[],"group_ids":[]} 2026-03-20T17:58:58.432 INFO:tasks.rgw_multisite:creating zone.. 2026-03-20T17:58:58.432 INFO:tasks.util.rgw:rgwadmin: c1.client.0 : ['zone', 'create', '--rgw-zone', 'a2', '--rgw-zonegroup', 'a', '--rgw-realm', 'test-realm', '--access-key', 'KLI823V2WXR7ZEZZ', '--secret', 'bfYGF3EVFszZr1T17wMbAaEQn3Tkb5fG', '--endpoints', 'http://vm05.local:8001', '--cluster', 'c1', '--debug-rgw', '0', '--debug-ms', '0'] 2026-03-20T17:58:58.432 DEBUG:tasks.util.rgw:rgwadmin: cmd=['adjust-ulimits', 'ceph-coverage', '/home/ubuntu/cephtest/archive/coverage', 'radosgw-admin', '--log-to-stderr', '--format', 'json', '-n', 'client.0', '--cluster', 'c1', 'zone', 'create', '--rgw-zone', 'a2', '--rgw-zonegroup', 'a', '--rgw-realm', 'test-realm', '--access-key', 'KLI823V2WXR7ZEZZ', '--secret', 'bfYGF3EVFszZr1T17wMbAaEQn3Tkb5fG', '--endpoints', 'http://vm05.local:8001', '--cluster', 'c1', '--debug-rgw', '0', '--debug-ms', '0'] 2026-03-20T17:58:58.432 DEBUG:teuthology.orchestra.run.vm02:> adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage radosgw-admin --log-to-stderr --format json -n client.0 --cluster c1 zone create --rgw-zone a2 --rgw-zonegroup a --rgw-realm test-realm --access-key KLI823V2WXR7ZEZZ --secret bfYGF3EVFszZr1T17wMbAaEQn3Tkb5fG --endpoints http://vm05.local:8001 --cluster c1 --debug-rgw 0 --debug-ms 0 2026-03-20T17:58:58.502 INFO:teuthology.orchestra.run.vm02.stderr:ignoring --setuser ceph since I am not root 2026-03-20T17:58:58.502 INFO:teuthology.orchestra.run.vm02.stderr:ignoring --setgroup ceph since I am not root 2026-03-20T17:58:58.524 INFO:teuthology.orchestra.run.vm02.stderr:NOTICE: set zone's realm_id=f5bc16fe-b476-42e1-93d5-47c86d0b7983 2026-03-20T17:58:58.545 INFO:teuthology.orchestra.run.vm02.stdout:{"id":"9164766d-ca5c-4a39-9306-38f195abd67d","name":"a2","domain_root":"a2.rgw.meta:root","control_pool":"a2.rgw.control","dedup_pool":"a2.rgw.dedup","gc_pool":"a2.rgw.log:gc","lc_pool":"a2.rgw.log:lc","log_pool":"a2.rgw.log","intent_log_pool":"a2.rgw.log:intent","usage_log_pool":"a2.rgw.log:usage","roles_pool":"a2.rgw.meta:roles","reshard_pool":"a2.rgw.log:reshard","user_keys_pool":"a2.rgw.meta:users.keys","user_email_pool":"a2.rgw.meta:users.email","user_swift_pool":"a2.rgw.meta:users.swift","user_uid_pool":"a2.rgw.meta:users.uid","otp_pool":"a2.rgw.otp","notif_pool":"a2.rgw.log:notif","topics_pool":"a2.rgw.meta:topics","account_pool":"a2.rgw.meta:accounts","group_pool":"a2.rgw.meta:groups","system_key":{"access_key":"KLI823V2WXR7ZEZZ","secret_key":"bfYGF3EVFszZr1T17wMbAaEQn3Tkb5fG"},"placement_pools":[{"key":"default-placement","val":{"index_pool":"a2.rgw.buckets.index","storage_classes":{"STANDARD":{"data_pool":"a2.rgw.buckets.data"}},"data_extra_pool":"a2.rgw.buckets.non-ec","index_type":0,"inline_data":true}}],"realm_id":"f5bc16fe-b476-42e1-93d5-47c86d0b7983","restore_pool":"a2.rgw.log:restore"} 2026-03-20T17:58:58.545 DEBUG:teuthology.orchestra.run.vm05:> sudo ceph osd pool create a2.rgw.buckets.data 64 64 --cluster c1 2026-03-20T17:58:58.850 INFO:teuthology.orchestra.run.vm05.stderr:pool 'a2.rgw.buckets.data' created 2026-03-20T17:58:58.876 DEBUG:teuthology.orchestra.run.vm05:> sudo ceph osd pool application enable a2.rgw.buckets.data rgw --cluster c1 2026-03-20T17:58:59.852 INFO:teuthology.orchestra.run.vm05.stderr:enabled application 'rgw' on pool 'a2.rgw.buckets.data' 2026-03-20T17:58:59.894 INFO:tasks.util.rgw:rgwadmin: c1.client.0 : ['zone', 'placement', '--rgw-zone', 'a2', '--rgw-zonegroup', 'a', '--rgw-realm', 'test-realm', 'modify', '--placement-id', 'default-placement', '--compression', 'random', '--cluster', 'c1', '--debug-rgw', '0', '--debug-ms', '0'] 2026-03-20T17:58:59.894 DEBUG:tasks.util.rgw:rgwadmin: cmd=['adjust-ulimits', 'ceph-coverage', '/home/ubuntu/cephtest/archive/coverage', 'radosgw-admin', '--log-to-stderr', '--format', 'json', '-n', 'client.0', '--cluster', 'c1', 'zone', 'placement', '--rgw-zone', 'a2', '--rgw-zonegroup', 'a', '--rgw-realm', 'test-realm', 'modify', '--placement-id', 'default-placement', '--compression', 'random', '--cluster', 'c1', '--debug-rgw', '0', '--debug-ms', '0'] 2026-03-20T17:58:59.894 DEBUG:teuthology.orchestra.run.vm02:> adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage radosgw-admin --log-to-stderr --format json -n client.0 --cluster c1 zone placement --rgw-zone a2 --rgw-zonegroup a --rgw-realm test-realm modify --placement-id default-placement --compression random --cluster c1 --debug-rgw 0 --debug-ms 0 2026-03-20T17:58:59.917 INFO:teuthology.orchestra.run.vm02.stderr:ignoring --setuser ceph since I am not root 2026-03-20T17:58:59.917 INFO:teuthology.orchestra.run.vm02.stderr:ignoring --setgroup ceph since I am not root 2026-03-20T17:58:59.939 INFO:teuthology.orchestra.run.vm02.stdout:{"id":"9164766d-ca5c-4a39-9306-38f195abd67d","name":"a2","domain_root":"a2.rgw.meta:root","control_pool":"a2.rgw.control","dedup_pool":"a2.rgw.dedup","gc_pool":"a2.rgw.log:gc","lc_pool":"a2.rgw.log:lc","log_pool":"a2.rgw.log","intent_log_pool":"a2.rgw.log:intent","usage_log_pool":"a2.rgw.log:usage","roles_pool":"a2.rgw.meta:roles","reshard_pool":"a2.rgw.log:reshard","user_keys_pool":"a2.rgw.meta:users.keys","user_email_pool":"a2.rgw.meta:users.email","user_swift_pool":"a2.rgw.meta:users.swift","user_uid_pool":"a2.rgw.meta:users.uid","otp_pool":"a2.rgw.otp","notif_pool":"a2.rgw.log:notif","topics_pool":"a2.rgw.meta:topics","account_pool":"a2.rgw.meta:accounts","group_pool":"a2.rgw.meta:groups","system_key":{"access_key":"KLI823V2WXR7ZEZZ","secret_key":"bfYGF3EVFszZr1T17wMbAaEQn3Tkb5fG"},"placement_pools":[{"key":"default-placement","val":{"index_pool":"a2.rgw.buckets.index","storage_classes":{"STANDARD":{"data_pool":"a2.rgw.buckets.data","compression_type":"random"}},"data_extra_pool":"a2.rgw.buckets.non-ec","index_type":0,"inline_data":true}}],"realm_id":"f5bc16fe-b476-42e1-93d5-47c86d0b7983","restore_pool":"a2.rgw.log:restore"} 2026-03-20T17:58:59.939 INFO:tasks.util.rgw:rgwadmin: c1.client.0 : ['period', 'update', '--rgw-zone', 'a2', '--rgw-zonegroup', 'a', '--rgw-realm', 'test-realm', '--commit', '--cluster', 'c1', '--debug-rgw', '0', '--debug-ms', '0'] 2026-03-20T17:58:59.939 DEBUG:tasks.util.rgw:rgwadmin: cmd=['adjust-ulimits', 'ceph-coverage', '/home/ubuntu/cephtest/archive/coverage', 'radosgw-admin', '--log-to-stderr', '--format', 'json', '-n', 'client.0', '--cluster', 'c1', 'period', 'update', '--rgw-zone', 'a2', '--rgw-zonegroup', 'a', '--rgw-realm', 'test-realm', '--commit', '--cluster', 'c1', '--debug-rgw', '0', '--debug-ms', '0'] 2026-03-20T17:58:59.939 DEBUG:teuthology.orchestra.run.vm02:> adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage radosgw-admin --log-to-stderr --format json -n client.0 --cluster c1 period update --rgw-zone a2 --rgw-zonegroup a --rgw-realm test-realm --commit --cluster c1 --debug-rgw 0 --debug-ms 0 2026-03-20T17:59:00.004 INFO:teuthology.orchestra.run.vm02.stderr:ignoring --setuser ceph since I am not root 2026-03-20T17:59:00.004 INFO:teuthology.orchestra.run.vm02.stderr:ignoring --setgroup ceph since I am not root 2026-03-20T17:59:00.040 INFO:teuthology.orchestra.run.vm02.stderr:2026-03-20T17:59:00.037+0000 7f6196463a80 0 period (82c662be-66b5-4c4f-b24e-ff9298fc8a6e does not have zone 9164766d-ca5c-4a39-9306-38f195abd67d configured 2026-03-20T17:59:00.296 INFO:tasks.daemonwatchdog.daemon_watchdog:daemon c1.rgw.client.0 has been restored 2026-03-20T17:59:07.046 INFO:teuthology.orchestra.run.vm02.stderr:Sending period to new master zone a5b5537f-2aa1-48e1-aabd-4c0e85b95fff 2026-03-20T17:59:07.076 INFO:teuthology.orchestra.run.vm02.stdout:{"id":"82c662be-66b5-4c4f-b24e-ff9298fc8a6e","epoch":2,"predecessor_uuid":"6b3a6782-0294-444e-971d-8e9000c552d8","sync_status":[],"period_map":{"id":"82c662be-66b5-4c4f-b24e-ff9298fc8a6e","zonegroups":[{"id":"a","name":"a","api_name":"a","is_master":true,"endpoints":["http://vm02.local:8000"],"hostnames":[],"hostnames_s3website":[],"master_zone":"a5b5537f-2aa1-48e1-aabd-4c0e85b95fff","zones":[{"id":"9164766d-ca5c-4a39-9306-38f195abd67d","name":"a2","endpoints":["http://vm05.local:8001"],"log_meta":false,"log_data":true,"bucket_index_max_shards":11,"read_only":false,"tier_type":"","sync_from_all":true,"sync_from":[],"redirect_zone":"","supported_features":["compress-encrypted","notification_v2","resharding"]},{"id":"a5b5537f-2aa1-48e1-aabd-4c0e85b95fff","name":"a1","endpoints":["http://vm02.local:8000"],"log_meta":false,"log_data":true,"bucket_index_max_shards":11,"read_only":false,"tier_type":"","sync_from_all":true,"sync_from":[],"redirect_zone":"","supported_features":["compress-encrypted","notification_v2","resharding"]}],"placement_targets":[{"name":"default-placement","tags":[],"storage_classes":["STANDARD"]}],"default_placement":"default-placement","realm_id":"f5bc16fe-b476-42e1-93d5-47c86d0b7983","sync_policy":{"groups":[]},"enabled_features":["notification_v2","resharding"]}],"short_zone_ids":[{"key":"9164766d-ca5c-4a39-9306-38f195abd67d","val":1184894607},{"key":"a5b5537f-2aa1-48e1-aabd-4c0e85b95fff","val":3173326878}]},"master_zonegroup":"a","master_zone":"a5b5537f-2aa1-48e1-aabd-4c0e85b95fff","period_config":{"bucket_quota":{"enabled":false,"check_on_raw":false,"max_size":-1,"max_size_kb":0,"max_objects":-1},"user_quota":{"enabled":false,"check_on_raw":false,"max_size":-1,"max_size_kb":0,"max_objects":-1},"user_ratelimit":{"max_read_ops":0,"max_write_ops":0,"max_read_bytes":0,"max_write_bytes":0,"enabled":false},"bucket_ratelimit":{"max_read_ops":0,"max_write_ops":0,"max_read_bytes":0,"max_write_bytes":0,"enabled":false},"anonymous_ratelimit":{"max_read_ops":0,"max_write_ops":0,"max_read_bytes":0,"max_write_bytes":0,"enabled":false}},"realm_id":"f5bc16fe-b476-42e1-93d5-47c86d0b7983","realm_epoch":2} 2026-03-20T17:59:07.077 DEBUG:tasks.rgw.c1.client.1:waiting for process to exit 2026-03-20T17:59:07.077 INFO:teuthology.orchestra.run:waiting for 300 2026-03-20T17:59:07.077 INFO:tasks.rgw.c1.client.1.vm05.stdout:2026-03-20T17:59:07.073+0000 7f171bee8640 -1 received signal: Terminated from /usr/bin/python3 /usr/bin/daemon-helper term radosgw --rgw-frontends beast port=8001 -n client.1 --cluster c1 -k /etc/ceph/c1.client.1.keyring --log-file /var/log/ceph/rgw.c1.client.1.log --rgw_ops_log_socket_path /home/ubuntu/cephtest/rgw.opslog.c1.client.1.sock --foreground (PID: 23418) UID: 0 2026-03-20T17:59:07.077 INFO:tasks.rgw.c1.client.1.vm05.stdout:2026-03-20T17:59:07.073+0000 7f171f730b40 -1 shutting down 2026-03-20T17:59:24.228 INFO:tasks.daemonwatchdog.daemon_watchdog:daemon c1.rgw.client.1 is failed for ~0s 2026-03-20T17:59:25.382 INFO:tasks.rgw.c1.client.1:Stopped 2026-03-20T17:59:25.382 INFO:tasks.rgw.c1.client.1:Restarting daemon 2026-03-20T17:59:25.382 DEBUG:teuthology.orchestra.run.vm05:> sudo adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage daemon-helper term radosgw --rgw-frontends 'beast port=8001' -n client.1 --cluster c1 -k /etc/ceph/c1.client.1.keyring --log-file /var/log/ceph/rgw.c1.client.1.log --rgw_ops_log_socket_path /home/ubuntu/cephtest/rgw.opslog.c1.client.1.sock --foreground --rgw-zone a2 --rgw-zonegroup a --rgw-realm test-realm | sudo tee /var/log/ceph/rgw.c1.client.1.stdout 2>&1 2026-03-20T17:59:25.383 INFO:tasks.rgw.c1.client.1:Started 2026-03-20T17:59:25.383 DEBUG:teuthology.orchestra.run.vm05:> curl http://vm05.local:8001 2026-03-20T17:59:25.388 INFO:teuthology.orchestra.run.vm05.stderr: % Total % Received % Xferd Average Speed Time Time Time Current 2026-03-20T17:59:25.388 INFO:teuthology.orchestra.run.vm05.stderr: Dload Upload Total Spent Left Speed 2026-03-20T17:59:25.389 INFO:teuthology.orchestra.run.vm05.stderr: 0 0 0 0 0 0 0 0 --:--:-- --:--:-- --:--:-- 0 0 0 0 0 0 0 0 0 --:--:-- --:--:-- --:--:-- 0 2026-03-20T17:59:25.389 INFO:teuthology.orchestra.run.vm05.stderr:curl: (7) Failed to connect to vm05.local port 8001 after 0 ms: Connection refused 2026-03-20T17:59:25.389 DEBUG:teuthology.orchestra.run:got remote process result: 7 2026-03-20T17:59:26.390 DEBUG:teuthology.orchestra.run.vm05:> curl http://vm05.local:8001 2026-03-20T17:59:26.396 INFO:teuthology.orchestra.run.vm05.stderr: % Total % Received % Xferd Average Speed Time Time Time Current 2026-03-20T17:59:26.397 INFO:teuthology.orchestra.run.vm05.stderr: Dload Upload Total Spent Left Speed 2026-03-20T17:59:26.398 INFO:teuthology.orchestra.run.vm05.stderr: 0 0 0 0 0 0 0 0 --:--:-- --:--:-- --:--:-- 0 100 187 0 187 0 0 102k 0 --:--:-- --:--:-- --:--:-- 182k 2026-03-20T17:59:26.398 INFO:teuthology.orchestra.run.vm05.stdout:anonymous 2026-03-20T17:59:26.398 INFO:tasks.rgw_multisite:pulling realm configuration to c2 2026-03-20T17:59:26.398 INFO:tasks.util.rgw:rgwadmin: c2.client.0 : ['realm', 'pull', '--rgw-realm', 'test-realm', '--default', '--url', 'http://vm02.local:8000', '--access-key', 'KLI823V2WXR7ZEZZ', '--secret', 'bfYGF3EVFszZr1T17wMbAaEQn3Tkb5fG', '--cluster', 'c2', '--debug-rgw', '0', '--debug-ms', '0'] 2026-03-20T17:59:26.398 DEBUG:tasks.util.rgw:rgwadmin: cmd=['adjust-ulimits', 'ceph-coverage', '/home/ubuntu/cephtest/archive/coverage', 'radosgw-admin', '--log-to-stderr', '--format', 'json', '-n', 'client.0', '--cluster', 'c2', 'realm', 'pull', '--rgw-realm', 'test-realm', '--default', '--url', 'http://vm02.local:8000', '--access-key', 'KLI823V2WXR7ZEZZ', '--secret', 'bfYGF3EVFszZr1T17wMbAaEQn3Tkb5fG', '--cluster', 'c2', '--debug-rgw', '0', '--debug-ms', '0'] 2026-03-20T17:59:26.398 DEBUG:teuthology.orchestra.run.vm06:> adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage radosgw-admin --log-to-stderr --format json -n client.0 --cluster c2 realm pull --rgw-realm test-realm --default --url http://vm02.local:8000 --access-key KLI823V2WXR7ZEZZ --secret bfYGF3EVFszZr1T17wMbAaEQn3Tkb5fG --cluster c2 --debug-rgw 0 --debug-ms 0 2026-03-20T17:59:26.420 INFO:teuthology.orchestra.run.vm06.stderr:ignoring --setuser ceph since I am not root 2026-03-20T17:59:26.420 INFO:teuthology.orchestra.run.vm06.stderr:ignoring --setgroup ceph since I am not root 2026-03-20T17:59:28.056 INFO:teuthology.orchestra.run.vm06.stdout:{"id":"f5bc16fe-b476-42e1-93d5-47c86d0b7983","name":"test-realm","current_period":"82c662be-66b5-4c4f-b24e-ff9298fc8a6e","epoch":2} 2026-03-20T17:59:28.056 INFO:tasks.rgw_multisite:creating zonegroup.. 2026-03-20T17:59:28.056 INFO:tasks.util.rgw:rgwadmin: c2.client.0 : ['zonegroup', 'set', '--rgw-zonegroup', 'b', '--rgw-realm', 'test-realm', '--default', '--cluster', 'c2', '--debug-rgw', '0', '--debug-ms', '0'] 2026-03-20T17:59:28.056 DEBUG:tasks.util.rgw:rgwadmin: cmd=['adjust-ulimits', 'ceph-coverage', '/home/ubuntu/cephtest/archive/coverage', 'radosgw-admin', '--log-to-stderr', '--format', 'json', '-n', 'client.0', '--cluster', 'c2', 'zonegroup', 'set', '--rgw-zonegroup', 'b', '--rgw-realm', 'test-realm', '--default', '--cluster', 'c2', '--debug-rgw', '0', '--debug-ms', '0'] 2026-03-20T17:59:28.056 DEBUG:teuthology.orchestra.run.vm06:> adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage radosgw-admin --log-to-stderr --format json -n client.0 --cluster c2 zonegroup set --rgw-zonegroup b --rgw-realm test-realm --default --cluster c2 --debug-rgw 0 --debug-ms 0 2026-03-20T17:59:28.122 INFO:teuthology.orchestra.run.vm06.stderr:ignoring --setuser ceph since I am not root 2026-03-20T17:59:28.122 INFO:teuthology.orchestra.run.vm06.stderr:ignoring --setgroup ceph since I am not root 2026-03-20T17:59:28.137 INFO:teuthology.orchestra.run.vm06.stderr:2026-03-20T17:59:28.131+0000 7f9eeb2c0a80 -1 old format 2026-03-20T17:59:28.148 INFO:teuthology.orchestra.run.vm06.stdout:{"id":"b","name":"b","api_name":"b","is_master":false,"endpoints":["http://vm06.local:8000"],"hostnames":[],"hostnames_s3website":[],"master_zone":"","zones":[],"placement_targets":[{"name":"default-placement","tags":[],"storage_classes":[]}],"default_placement":"default-placement","realm_id":"f5bc16fe-b476-42e1-93d5-47c86d0b7983","sync_policy":{"groups":[]},"enabled_features":["notification_v2","resharding"]} 2026-03-20T17:59:28.148 INFO:tasks.rgw_multisite:creating zone.. 2026-03-20T17:59:28.148 INFO:tasks.util.rgw:rgwadmin: c2.client.0 : ['zone', 'create', '--rgw-zone', 'b1', '--rgw-zonegroup', 'b', '--rgw-realm', 'test-realm', '--default', '--master', '--access-key', 'KLI823V2WXR7ZEZZ', '--secret', 'bfYGF3EVFszZr1T17wMbAaEQn3Tkb5fG', '--endpoints', 'http://vm06.local:8000', '--cluster', 'c2', '--debug-rgw', '0', '--debug-ms', '0'] 2026-03-20T17:59:28.148 DEBUG:tasks.util.rgw:rgwadmin: cmd=['adjust-ulimits', 'ceph-coverage', '/home/ubuntu/cephtest/archive/coverage', 'radosgw-admin', '--log-to-stderr', '--format', 'json', '-n', 'client.0', '--cluster', 'c2', 'zone', 'create', '--rgw-zone', 'b1', '--rgw-zonegroup', 'b', '--rgw-realm', 'test-realm', '--default', '--master', '--access-key', 'KLI823V2WXR7ZEZZ', '--secret', 'bfYGF3EVFszZr1T17wMbAaEQn3Tkb5fG', '--endpoints', 'http://vm06.local:8000', '--cluster', 'c2', '--debug-rgw', '0', '--debug-ms', '0'] 2026-03-20T17:59:28.148 DEBUG:teuthology.orchestra.run.vm06:> adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage radosgw-admin --log-to-stderr --format json -n client.0 --cluster c2 zone create --rgw-zone b1 --rgw-zonegroup b --rgw-realm test-realm --default --master --access-key KLI823V2WXR7ZEZZ --secret bfYGF3EVFszZr1T17wMbAaEQn3Tkb5fG --endpoints http://vm06.local:8000 --cluster c2 --debug-rgw 0 --debug-ms 0 2026-03-20T17:59:28.219 INFO:teuthology.orchestra.run.vm06.stderr:ignoring --setuser ceph since I am not root 2026-03-20T17:59:28.219 INFO:teuthology.orchestra.run.vm06.stderr:ignoring --setgroup ceph since I am not root 2026-03-20T17:59:28.235 INFO:teuthology.orchestra.run.vm06.stderr:NOTICE: set zone's realm_id=f5bc16fe-b476-42e1-93d5-47c86d0b7983 2026-03-20T17:59:28.257 INFO:teuthology.orchestra.run.vm06.stdout:{"id":"0dc28930-d0f7-4e21-9754-a0eeb8caebc5","name":"b1","domain_root":"b1.rgw.meta:root","control_pool":"b1.rgw.control","dedup_pool":"b1.rgw.dedup","gc_pool":"b1.rgw.log:gc","lc_pool":"b1.rgw.log:lc","log_pool":"b1.rgw.log","intent_log_pool":"b1.rgw.log:intent","usage_log_pool":"b1.rgw.log:usage","roles_pool":"b1.rgw.meta:roles","reshard_pool":"b1.rgw.log:reshard","user_keys_pool":"b1.rgw.meta:users.keys","user_email_pool":"b1.rgw.meta:users.email","user_swift_pool":"b1.rgw.meta:users.swift","user_uid_pool":"b1.rgw.meta:users.uid","otp_pool":"b1.rgw.otp","notif_pool":"b1.rgw.log:notif","topics_pool":"b1.rgw.meta:topics","account_pool":"b1.rgw.meta:accounts","group_pool":"b1.rgw.meta:groups","system_key":{"access_key":"KLI823V2WXR7ZEZZ","secret_key":"bfYGF3EVFszZr1T17wMbAaEQn3Tkb5fG"},"placement_pools":[{"key":"default-placement","val":{"index_pool":"b1.rgw.buckets.index","storage_classes":{"STANDARD":{"data_pool":"b1.rgw.buckets.data"}},"data_extra_pool":"b1.rgw.buckets.non-ec","index_type":0,"inline_data":true}}],"realm_id":"f5bc16fe-b476-42e1-93d5-47c86d0b7983","restore_pool":"b1.rgw.log:restore"} 2026-03-20T17:59:28.257 DEBUG:teuthology.orchestra.run.vm06:> sudo ceph osd pool create b1.rgw.buckets.data 64 64 --cluster c2 2026-03-20T17:59:29.072 INFO:teuthology.orchestra.run.vm06.stderr:pool 'b1.rgw.buckets.data' created 2026-03-20T17:59:29.099 DEBUG:teuthology.orchestra.run.vm06:> sudo ceph osd pool application enable b1.rgw.buckets.data rgw --cluster c2 2026-03-20T17:59:30.077 INFO:teuthology.orchestra.run.vm06.stderr:enabled application 'rgw' on pool 'b1.rgw.buckets.data' 2026-03-20T17:59:30.113 INFO:tasks.util.rgw:rgwadmin: c2.client.0 : ['zone', 'placement', '--rgw-zone', 'b1', '--rgw-zonegroup', 'b', '--rgw-realm', 'test-realm', 'modify', '--placement-id', 'default-placement', '--compression', 'random', '--cluster', 'c2', '--debug-rgw', '0', '--debug-ms', '0'] 2026-03-20T17:59:30.113 DEBUG:tasks.util.rgw:rgwadmin: cmd=['adjust-ulimits', 'ceph-coverage', '/home/ubuntu/cephtest/archive/coverage', 'radosgw-admin', '--log-to-stderr', '--format', 'json', '-n', 'client.0', '--cluster', 'c2', 'zone', 'placement', '--rgw-zone', 'b1', '--rgw-zonegroup', 'b', '--rgw-realm', 'test-realm', 'modify', '--placement-id', 'default-placement', '--compression', 'random', '--cluster', 'c2', '--debug-rgw', '0', '--debug-ms', '0'] 2026-03-20T17:59:30.113 DEBUG:teuthology.orchestra.run.vm06:> adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage radosgw-admin --log-to-stderr --format json -n client.0 --cluster c2 zone placement --rgw-zone b1 --rgw-zonegroup b --rgw-realm test-realm modify --placement-id default-placement --compression random --cluster c2 --debug-rgw 0 --debug-ms 0 2026-03-20T17:59:30.138 INFO:teuthology.orchestra.run.vm06.stderr:ignoring --setuser ceph since I am not root 2026-03-20T17:59:30.138 INFO:teuthology.orchestra.run.vm06.stderr:ignoring --setgroup ceph since I am not root 2026-03-20T17:59:30.166 INFO:teuthology.orchestra.run.vm06.stdout:{"id":"0dc28930-d0f7-4e21-9754-a0eeb8caebc5","name":"b1","domain_root":"b1.rgw.meta:root","control_pool":"b1.rgw.control","dedup_pool":"b1.rgw.dedup","gc_pool":"b1.rgw.log:gc","lc_pool":"b1.rgw.log:lc","log_pool":"b1.rgw.log","intent_log_pool":"b1.rgw.log:intent","usage_log_pool":"b1.rgw.log:usage","roles_pool":"b1.rgw.meta:roles","reshard_pool":"b1.rgw.log:reshard","user_keys_pool":"b1.rgw.meta:users.keys","user_email_pool":"b1.rgw.meta:users.email","user_swift_pool":"b1.rgw.meta:users.swift","user_uid_pool":"b1.rgw.meta:users.uid","otp_pool":"b1.rgw.otp","notif_pool":"b1.rgw.log:notif","topics_pool":"b1.rgw.meta:topics","account_pool":"b1.rgw.meta:accounts","group_pool":"b1.rgw.meta:groups","system_key":{"access_key":"KLI823V2WXR7ZEZZ","secret_key":"bfYGF3EVFszZr1T17wMbAaEQn3Tkb5fG"},"placement_pools":[{"key":"default-placement","val":{"index_pool":"b1.rgw.buckets.index","storage_classes":{"STANDARD":{"data_pool":"b1.rgw.buckets.data","compression_type":"random"}},"data_extra_pool":"b1.rgw.buckets.non-ec","index_type":0,"inline_data":true}}],"realm_id":"f5bc16fe-b476-42e1-93d5-47c86d0b7983","restore_pool":"b1.rgw.log:restore"} 2026-03-20T17:59:30.167 INFO:tasks.util.rgw:rgwadmin: c2.client.0 : ['period', 'update', '--rgw-zone', 'b1', '--rgw-zonegroup', 'b', '--rgw-realm', 'test-realm', '--commit', '--cluster', 'c2', '--debug-rgw', '0', '--debug-ms', '0'] 2026-03-20T17:59:30.167 DEBUG:tasks.util.rgw:rgwadmin: cmd=['adjust-ulimits', 'ceph-coverage', '/home/ubuntu/cephtest/archive/coverage', 'radosgw-admin', '--log-to-stderr', '--format', 'json', '-n', 'client.0', '--cluster', 'c2', 'period', 'update', '--rgw-zone', 'b1', '--rgw-zonegroup', 'b', '--rgw-realm', 'test-realm', '--commit', '--cluster', 'c2', '--debug-rgw', '0', '--debug-ms', '0'] 2026-03-20T17:59:30.167 DEBUG:teuthology.orchestra.run.vm06:> adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage radosgw-admin --log-to-stderr --format json -n client.0 --cluster c2 period update --rgw-zone b1 --rgw-zonegroup b --rgw-realm test-realm --commit --cluster c2 --debug-rgw 0 --debug-ms 0 2026-03-20T17:59:30.230 INFO:teuthology.orchestra.run.vm06.stderr:ignoring --setuser ceph since I am not root 2026-03-20T17:59:30.230 INFO:teuthology.orchestra.run.vm06.stderr:ignoring --setgroup ceph since I am not root 2026-03-20T17:59:30.234 INFO:tasks.daemonwatchdog.daemon_watchdog:daemon c1.rgw.client.1 has been restored 2026-03-20T17:59:30.256 INFO:teuthology.orchestra.run.vm06.stderr:2026-03-20T17:59:30.251+0000 7f2635f85a80 0 period (82c662be-66b5-4c4f-b24e-ff9298fc8a6e does not have zone 0dc28930-d0f7-4e21-9754-a0eeb8caebc5 configured 2026-03-20T17:59:37.199 INFO:teuthology.orchestra.run.vm06.stderr:Sending period to new master zone a5b5537f-2aa1-48e1-aabd-4c0e85b95fff 2026-03-20T17:59:37.242 INFO:teuthology.orchestra.run.vm06.stdout:{"id":"82c662be-66b5-4c4f-b24e-ff9298fc8a6e","epoch":3,"predecessor_uuid":"6b3a6782-0294-444e-971d-8e9000c552d8","sync_status":[],"period_map":{"id":"82c662be-66b5-4c4f-b24e-ff9298fc8a6e","zonegroups":[{"id":"a","name":"a","api_name":"a","is_master":true,"endpoints":["http://vm02.local:8000"],"hostnames":[],"hostnames_s3website":[],"master_zone":"a5b5537f-2aa1-48e1-aabd-4c0e85b95fff","zones":[{"id":"9164766d-ca5c-4a39-9306-38f195abd67d","name":"a2","endpoints":["http://vm05.local:8001"],"log_meta":false,"log_data":true,"bucket_index_max_shards":11,"read_only":false,"tier_type":"","sync_from_all":true,"sync_from":[],"redirect_zone":"","supported_features":["compress-encrypted","notification_v2","resharding"]},{"id":"a5b5537f-2aa1-48e1-aabd-4c0e85b95fff","name":"a1","endpoints":["http://vm02.local:8000"],"log_meta":false,"log_data":true,"bucket_index_max_shards":11,"read_only":false,"tier_type":"","sync_from_all":true,"sync_from":[],"redirect_zone":"","supported_features":["compress-encrypted","notification_v2","resharding"]}],"placement_targets":[{"name":"default-placement","tags":[],"storage_classes":["STANDARD"]}],"default_placement":"default-placement","realm_id":"f5bc16fe-b476-42e1-93d5-47c86d0b7983","sync_policy":{"groups":[]},"enabled_features":["notification_v2","resharding"]},{"id":"b","name":"b","api_name":"b","is_master":false,"endpoints":["http://vm06.local:8000"],"hostnames":[],"hostnames_s3website":[],"master_zone":"0dc28930-d0f7-4e21-9754-a0eeb8caebc5","zones":[{"id":"0dc28930-d0f7-4e21-9754-a0eeb8caebc5","name":"b1","endpoints":["http://vm06.local:8000"],"log_meta":false,"log_data":false,"bucket_index_max_shards":11,"read_only":false,"tier_type":"","sync_from_all":true,"sync_from":[],"redirect_zone":"","supported_features":["compress-encrypted","notification_v2","resharding"]}],"placement_targets":[{"name":"default-placement","tags":[],"storage_classes":["STANDARD"]}],"default_placement":"default-placement","realm_id":"f5bc16fe-b476-42e1-93d5-47c86d0b7983","sync_policy":{"groups":[]},"enabled_features":["notification_v2","resharding"]}],"short_zone_ids":[{"key":"0dc28930-d0f7-4e21-9754-a0eeb8caebc5","val":414735373},{"key":"9164766d-ca5c-4a39-9306-38f195abd67d","val":1184894607},{"key":"a5b5537f-2aa1-48e1-aabd-4c0e85b95fff","val":3173326878}]},"master_zonegroup":"a","master_zone":"a5b5537f-2aa1-48e1-aabd-4c0e85b95fff","period_config":{"bucket_quota":{"enabled":false,"check_on_raw":false,"max_size":-1,"max_size_kb":0,"max_objects":-1},"user_quota":{"enabled":false,"check_on_raw":false,"max_size":-1,"max_size_kb":0,"max_objects":-1},"user_ratelimit":{"max_read_ops":0,"max_write_ops":0,"max_read_bytes":0,"max_write_bytes":0,"enabled":false},"bucket_ratelimit":{"max_read_ops":0,"max_write_ops":0,"max_read_bytes":0,"max_write_bytes":0,"enabled":false},"anonymous_ratelimit":{"max_read_ops":0,"max_write_ops":0,"max_read_bytes":0,"max_write_bytes":0,"enabled":false}},"realm_id":"f5bc16fe-b476-42e1-93d5-47c86d0b7983","realm_epoch":2} 2026-03-20T17:59:37.242 DEBUG:tasks.rgw.c2.client.0:waiting for process to exit 2026-03-20T17:59:37.242 INFO:teuthology.orchestra.run:waiting for 300 2026-03-20T17:59:37.284 INFO:tasks.rgw.c2.client.0.vm06.stdout:2026-03-20T17:59:37.279+0000 7fc06b7db640 -1 received signal: Terminated from /usr/bin/python3 /usr/bin/daemon-helper term radosgw --rgw-frontends beast port=8000 -n client.0 --cluster c2 -k /etc/ceph/c2.client.0.keyring --log-file /var/log/ceph/rgw.c2.client.0.log --rgw_ops_log_socket_path /home/ubuntu/cephtest/rgw.opslog.c2.client.0.sock --foreground (PID: 24142) UID: 0 2026-03-20T17:59:37.284 INFO:tasks.rgw.c2.client.0.vm06.stdout:2026-03-20T17:59:37.279+0000 7fc06f023b40 -1 shutting down 2026-03-20T17:59:55.546 INFO:tasks.rgw.c2.client.0:Stopped 2026-03-20T17:59:55.546 INFO:tasks.rgw.c2.client.0:Restarting daemon 2026-03-20T17:59:55.546 DEBUG:teuthology.orchestra.run.vm06:> sudo adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage daemon-helper term radosgw --rgw-frontends 'beast port=8000' -n client.0 --cluster c2 -k /etc/ceph/c2.client.0.keyring --log-file /var/log/ceph/rgw.c2.client.0.log --rgw_ops_log_socket_path /home/ubuntu/cephtest/rgw.opslog.c2.client.0.sock --foreground --rgw-zone b1 --rgw-zonegroup b --rgw-realm test-realm | sudo tee /var/log/ceph/rgw.c2.client.0.stdout 2>&1 2026-03-20T17:59:55.547 INFO:tasks.rgw.c2.client.0:Started 2026-03-20T17:59:55.547 DEBUG:teuthology.orchestra.run.vm06:> curl http://vm06.local:8000 2026-03-20T17:59:55.554 INFO:teuthology.orchestra.run.vm06.stderr: % Total % Received % Xferd Average Speed Time Time Time Current 2026-03-20T17:59:55.555 INFO:teuthology.orchestra.run.vm06.stderr: Dload Upload Total Spent Left Speed 2026-03-20T17:59:55.555 INFO:teuthology.orchestra.run.vm06.stderr: 0 0 0 0 0 0 0 0 --:--:-- --:--:-- --:--:-- 0 0 0 0 0 0 0 0 0 --:--:-- --:--:-- --:--:-- 0 2026-03-20T17:59:55.555 INFO:teuthology.orchestra.run.vm06.stderr:curl: (7) Failed to connect to vm06.local port 8000 after 0 ms: Connection refused 2026-03-20T17:59:55.555 DEBUG:teuthology.orchestra.run:got remote process result: 7 2026-03-20T17:59:56.556 DEBUG:teuthology.orchestra.run.vm06:> curl http://vm06.local:8000 2026-03-20T17:59:56.572 DEBUG:teuthology.orchestra.run:got remote process result: 7 2026-03-20T17:59:56.572 INFO:teuthology.orchestra.run.vm06.stderr: % Total % Received % Xferd Average Speed Time Time Time Current 2026-03-20T17:59:56.572 INFO:teuthology.orchestra.run.vm06.stderr: Dload Upload Total Spent Left Speed 2026-03-20T17:59:56.572 INFO:teuthology.orchestra.run.vm06.stderr: 0 0 0 0 0 0 0 0 --:--:-- --:--:-- --:--:-- 0 0 0 0 0 0 0 0 0 --:--:-- --:--:-- --:--:-- 0 2026-03-20T17:59:56.572 INFO:teuthology.orchestra.run.vm06.stderr:curl: (7) Failed to connect to vm06.local port 8000 after 4 ms: Connection refused 2026-03-20T17:59:58.573 DEBUG:teuthology.orchestra.run.vm06:> curl http://vm06.local:8000 2026-03-20T17:59:58.581 INFO:teuthology.orchestra.run.vm06.stderr: % Total % Received % Xferd Average Speed Time Time Time Current 2026-03-20T17:59:58.581 INFO:teuthology.orchestra.run.vm06.stderr: Dload Upload Total Spent Left Speed 2026-03-20T17:59:58.583 INFO:teuthology.orchestra.run.vm06.stderr: 0 0 0 0 0 0 0 0 --:--:-- --:--:-- --:--:-- 0 100 187 0 187 0 0 84845 0 --:--:-- --:--:-- --:--:-- 93500 2026-03-20T17:59:58.583 INFO:teuthology.orchestra.run.vm06.stdout:anonymous 2026-03-20T17:59:58.584 INFO:tasks.rgw_multisite:pulling realm configuration to c2 2026-03-20T17:59:58.584 INFO:tasks.util.rgw:rgwadmin: c2.client.0 : ['realm', 'pull', '--rgw-realm', 'test-realm', '--default', '--url', 'http://vm02.local:8000', '--access-key', 'KLI823V2WXR7ZEZZ', '--secret', 'bfYGF3EVFszZr1T17wMbAaEQn3Tkb5fG', '--cluster', 'c2', '--debug-rgw', '0', '--debug-ms', '0'] 2026-03-20T17:59:58.584 DEBUG:tasks.util.rgw:rgwadmin: cmd=['adjust-ulimits', 'ceph-coverage', '/home/ubuntu/cephtest/archive/coverage', 'radosgw-admin', '--log-to-stderr', '--format', 'json', '-n', 'client.0', '--cluster', 'c2', 'realm', 'pull', '--rgw-realm', 'test-realm', '--default', '--url', 'http://vm02.local:8000', '--access-key', 'KLI823V2WXR7ZEZZ', '--secret', 'bfYGF3EVFszZr1T17wMbAaEQn3Tkb5fG', '--cluster', 'c2', '--debug-rgw', '0', '--debug-ms', '0'] 2026-03-20T17:59:58.584 DEBUG:teuthology.orchestra.run.vm06:> adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage radosgw-admin --log-to-stderr --format json -n client.0 --cluster c2 realm pull --rgw-realm test-realm --default --url http://vm02.local:8000 --access-key KLI823V2WXR7ZEZZ --secret bfYGF3EVFszZr1T17wMbAaEQn3Tkb5fG --cluster c2 --debug-rgw 0 --debug-ms 0 2026-03-20T17:59:58.660 INFO:teuthology.orchestra.run.vm06.stderr:ignoring --setuser ceph since I am not root 2026-03-20T17:59:58.661 INFO:teuthology.orchestra.run.vm06.stderr:ignoring --setgroup ceph since I am not root 2026-03-20T17:59:58.714 INFO:teuthology.orchestra.run.vm06.stdout:{"id":"f5bc16fe-b476-42e1-93d5-47c86d0b7983","name":"test-realm","current_period":"82c662be-66b5-4c4f-b24e-ff9298fc8a6e","epoch":2} 2026-03-20T17:59:58.715 INFO:tasks.rgw_multisite:creating zone.. 2026-03-20T17:59:58.715 INFO:tasks.util.rgw:rgwadmin: c2.client.0 : ['zone', 'create', '--rgw-zone', 'b2', '--rgw-zonegroup', 'b', '--rgw-realm', 'test-realm', '--access-key', 'KLI823V2WXR7ZEZZ', '--secret', 'bfYGF3EVFszZr1T17wMbAaEQn3Tkb5fG', '--endpoints', 'http://vm08.local:8001', '--cluster', 'c2', '--debug-rgw', '0', '--debug-ms', '0'] 2026-03-20T17:59:58.715 DEBUG:tasks.util.rgw:rgwadmin: cmd=['adjust-ulimits', 'ceph-coverage', '/home/ubuntu/cephtest/archive/coverage', 'radosgw-admin', '--log-to-stderr', '--format', 'json', '-n', 'client.0', '--cluster', 'c2', 'zone', 'create', '--rgw-zone', 'b2', '--rgw-zonegroup', 'b', '--rgw-realm', 'test-realm', '--access-key', 'KLI823V2WXR7ZEZZ', '--secret', 'bfYGF3EVFszZr1T17wMbAaEQn3Tkb5fG', '--endpoints', 'http://vm08.local:8001', '--cluster', 'c2', '--debug-rgw', '0', '--debug-ms', '0'] 2026-03-20T17:59:58.715 DEBUG:teuthology.orchestra.run.vm06:> adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage radosgw-admin --log-to-stderr --format json -n client.0 --cluster c2 zone create --rgw-zone b2 --rgw-zonegroup b --rgw-realm test-realm --access-key KLI823V2WXR7ZEZZ --secret bfYGF3EVFszZr1T17wMbAaEQn3Tkb5fG --endpoints http://vm08.local:8001 --cluster c2 --debug-rgw 0 --debug-ms 0 2026-03-20T17:59:58.784 INFO:teuthology.orchestra.run.vm06.stderr:ignoring --setuser ceph since I am not root 2026-03-20T17:59:58.784 INFO:teuthology.orchestra.run.vm06.stderr:ignoring --setgroup ceph since I am not root 2026-03-20T17:59:58.800 INFO:teuthology.orchestra.run.vm06.stderr:NOTICE: set zone's realm_id=f5bc16fe-b476-42e1-93d5-47c86d0b7983 2026-03-20T17:59:58.821 INFO:teuthology.orchestra.run.vm06.stdout:{"id":"7b70d73b-78c6-4a1d-b06e-7aa536f1b974","name":"b2","domain_root":"b2.rgw.meta:root","control_pool":"b2.rgw.control","dedup_pool":"b2.rgw.dedup","gc_pool":"b2.rgw.log:gc","lc_pool":"b2.rgw.log:lc","log_pool":"b2.rgw.log","intent_log_pool":"b2.rgw.log:intent","usage_log_pool":"b2.rgw.log:usage","roles_pool":"b2.rgw.meta:roles","reshard_pool":"b2.rgw.log:reshard","user_keys_pool":"b2.rgw.meta:users.keys","user_email_pool":"b2.rgw.meta:users.email","user_swift_pool":"b2.rgw.meta:users.swift","user_uid_pool":"b2.rgw.meta:users.uid","otp_pool":"b2.rgw.otp","notif_pool":"b2.rgw.log:notif","topics_pool":"b2.rgw.meta:topics","account_pool":"b2.rgw.meta:accounts","group_pool":"b2.rgw.meta:groups","system_key":{"access_key":"KLI823V2WXR7ZEZZ","secret_key":"bfYGF3EVFszZr1T17wMbAaEQn3Tkb5fG"},"placement_pools":[{"key":"default-placement","val":{"index_pool":"b2.rgw.buckets.index","storage_classes":{"STANDARD":{"data_pool":"b2.rgw.buckets.data"}},"data_extra_pool":"b2.rgw.buckets.non-ec","index_type":0,"inline_data":true}}],"realm_id":"f5bc16fe-b476-42e1-93d5-47c86d0b7983","restore_pool":"b2.rgw.log:restore"} 2026-03-20T17:59:58.821 DEBUG:teuthology.orchestra.run.vm08:> sudo ceph osd pool create b2.rgw.buckets.data 64 64 --cluster c2 2026-03-20T17:59:59.271 INFO:teuthology.orchestra.run.vm08.stderr:pool 'b2.rgw.buckets.data' created 2026-03-20T17:59:59.294 DEBUG:teuthology.orchestra.run.vm08:> sudo ceph osd pool application enable b2.rgw.buckets.data rgw --cluster c2 2026-03-20T18:00:00.293 INFO:teuthology.orchestra.run.vm08.stderr:enabled application 'rgw' on pool 'b2.rgw.buckets.data' 2026-03-20T18:00:00.331 INFO:tasks.util.rgw:rgwadmin: c2.client.0 : ['zone', 'placement', '--rgw-zone', 'b2', '--rgw-zonegroup', 'b', '--rgw-realm', 'test-realm', 'modify', '--placement-id', 'default-placement', '--compression', 'random', '--cluster', 'c2', '--debug-rgw', '0', '--debug-ms', '0'] 2026-03-20T18:00:00.331 DEBUG:tasks.util.rgw:rgwadmin: cmd=['adjust-ulimits', 'ceph-coverage', '/home/ubuntu/cephtest/archive/coverage', 'radosgw-admin', '--log-to-stderr', '--format', 'json', '-n', 'client.0', '--cluster', 'c2', 'zone', 'placement', '--rgw-zone', 'b2', '--rgw-zonegroup', 'b', '--rgw-realm', 'test-realm', 'modify', '--placement-id', 'default-placement', '--compression', 'random', '--cluster', 'c2', '--debug-rgw', '0', '--debug-ms', '0'] 2026-03-20T18:00:00.331 DEBUG:teuthology.orchestra.run.vm06:> adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage radosgw-admin --log-to-stderr --format json -n client.0 --cluster c2 zone placement --rgw-zone b2 --rgw-zonegroup b --rgw-realm test-realm modify --placement-id default-placement --compression random --cluster c2 --debug-rgw 0 --debug-ms 0 2026-03-20T18:00:00.354 INFO:teuthology.orchestra.run.vm06.stderr:ignoring --setuser ceph since I am not root 2026-03-20T18:00:00.354 INFO:teuthology.orchestra.run.vm06.stderr:ignoring --setgroup ceph since I am not root 2026-03-20T18:00:00.379 INFO:teuthology.orchestra.run.vm06.stdout:{"id":"7b70d73b-78c6-4a1d-b06e-7aa536f1b974","name":"b2","domain_root":"b2.rgw.meta:root","control_pool":"b2.rgw.control","dedup_pool":"b2.rgw.dedup","gc_pool":"b2.rgw.log:gc","lc_pool":"b2.rgw.log:lc","log_pool":"b2.rgw.log","intent_log_pool":"b2.rgw.log:intent","usage_log_pool":"b2.rgw.log:usage","roles_pool":"b2.rgw.meta:roles","reshard_pool":"b2.rgw.log:reshard","user_keys_pool":"b2.rgw.meta:users.keys","user_email_pool":"b2.rgw.meta:users.email","user_swift_pool":"b2.rgw.meta:users.swift","user_uid_pool":"b2.rgw.meta:users.uid","otp_pool":"b2.rgw.otp","notif_pool":"b2.rgw.log:notif","topics_pool":"b2.rgw.meta:topics","account_pool":"b2.rgw.meta:accounts","group_pool":"b2.rgw.meta:groups","system_key":{"access_key":"KLI823V2WXR7ZEZZ","secret_key":"bfYGF3EVFszZr1T17wMbAaEQn3Tkb5fG"},"placement_pools":[{"key":"default-placement","val":{"index_pool":"b2.rgw.buckets.index","storage_classes":{"STANDARD":{"data_pool":"b2.rgw.buckets.data","compression_type":"random"}},"data_extra_pool":"b2.rgw.buckets.non-ec","index_type":0,"inline_data":true}}],"realm_id":"f5bc16fe-b476-42e1-93d5-47c86d0b7983","restore_pool":"b2.rgw.log:restore"} 2026-03-20T18:00:00.380 INFO:tasks.util.rgw:rgwadmin: c2.client.0 : ['period', 'update', '--rgw-zone', 'b2', '--rgw-zonegroup', 'b', '--rgw-realm', 'test-realm', '--commit', '--cluster', 'c2', '--debug-rgw', '0', '--debug-ms', '0'] 2026-03-20T18:00:00.380 DEBUG:tasks.util.rgw:rgwadmin: cmd=['adjust-ulimits', 'ceph-coverage', '/home/ubuntu/cephtest/archive/coverage', 'radosgw-admin', '--log-to-stderr', '--format', 'json', '-n', 'client.0', '--cluster', 'c2', 'period', 'update', '--rgw-zone', 'b2', '--rgw-zonegroup', 'b', '--rgw-realm', 'test-realm', '--commit', '--cluster', 'c2', '--debug-rgw', '0', '--debug-ms', '0'] 2026-03-20T18:00:00.380 DEBUG:teuthology.orchestra.run.vm06:> adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage radosgw-admin --log-to-stderr --format json -n client.0 --cluster c2 period update --rgw-zone b2 --rgw-zonegroup b --rgw-realm test-realm --commit --cluster c2 --debug-rgw 0 --debug-ms 0 2026-03-20T18:00:00.442 INFO:teuthology.orchestra.run.vm06.stderr:ignoring --setuser ceph since I am not root 2026-03-20T18:00:00.442 INFO:teuthology.orchestra.run.vm06.stderr:ignoring --setgroup ceph since I am not root 2026-03-20T18:00:00.465 INFO:teuthology.orchestra.run.vm06.stderr:2026-03-20T18:00:00.459+0000 7f297ce60a80 0 period (82c662be-66b5-4c4f-b24e-ff9298fc8a6e does not have zone 7b70d73b-78c6-4a1d-b06e-7aa536f1b974 configured 2026-03-20T18:00:07.435 INFO:teuthology.orchestra.run.vm06.stderr:Sending period to new master zone a5b5537f-2aa1-48e1-aabd-4c0e85b95fff 2026-03-20T18:00:07.478 INFO:teuthology.orchestra.run.vm06.stdout:{"id":"82c662be-66b5-4c4f-b24e-ff9298fc8a6e","epoch":4,"predecessor_uuid":"6b3a6782-0294-444e-971d-8e9000c552d8","sync_status":[],"period_map":{"id":"82c662be-66b5-4c4f-b24e-ff9298fc8a6e","zonegroups":[{"id":"a","name":"a","api_name":"a","is_master":true,"endpoints":["http://vm02.local:8000"],"hostnames":[],"hostnames_s3website":[],"master_zone":"a5b5537f-2aa1-48e1-aabd-4c0e85b95fff","zones":[{"id":"9164766d-ca5c-4a39-9306-38f195abd67d","name":"a2","endpoints":["http://vm05.local:8001"],"log_meta":false,"log_data":true,"bucket_index_max_shards":11,"read_only":false,"tier_type":"","sync_from_all":true,"sync_from":[],"redirect_zone":"","supported_features":["compress-encrypted","notification_v2","resharding"]},{"id":"a5b5537f-2aa1-48e1-aabd-4c0e85b95fff","name":"a1","endpoints":["http://vm02.local:8000"],"log_meta":false,"log_data":true,"bucket_index_max_shards":11,"read_only":false,"tier_type":"","sync_from_all":true,"sync_from":[],"redirect_zone":"","supported_features":["compress-encrypted","notification_v2","resharding"]}],"placement_targets":[{"name":"default-placement","tags":[],"storage_classes":["STANDARD"]}],"default_placement":"default-placement","realm_id":"f5bc16fe-b476-42e1-93d5-47c86d0b7983","sync_policy":{"groups":[]},"enabled_features":["notification_v2","resharding"]},{"id":"b","name":"b","api_name":"b","is_master":false,"endpoints":["http://vm06.local:8000"],"hostnames":[],"hostnames_s3website":[],"master_zone":"0dc28930-d0f7-4e21-9754-a0eeb8caebc5","zones":[{"id":"0dc28930-d0f7-4e21-9754-a0eeb8caebc5","name":"b1","endpoints":["http://vm06.local:8000"],"log_meta":false,"log_data":true,"bucket_index_max_shards":11,"read_only":false,"tier_type":"","sync_from_all":true,"sync_from":[],"redirect_zone":"","supported_features":["compress-encrypted","notification_v2","resharding"]},{"id":"7b70d73b-78c6-4a1d-b06e-7aa536f1b974","name":"b2","endpoints":["http://vm08.local:8001"],"log_meta":false,"log_data":true,"bucket_index_max_shards":11,"read_only":false,"tier_type":"","sync_from_all":true,"sync_from":[],"redirect_zone":"","supported_features":["compress-encrypted","notification_v2","resharding"]}],"placement_targets":[{"name":"default-placement","tags":[],"storage_classes":["STANDARD"]}],"default_placement":"default-placement","realm_id":"f5bc16fe-b476-42e1-93d5-47c86d0b7983","sync_policy":{"groups":[]},"enabled_features":["notification_v2","resharding"]}],"short_zone_ids":[{"key":"0dc28930-d0f7-4e21-9754-a0eeb8caebc5","val":414735373},{"key":"7b70d73b-78c6-4a1d-b06e-7aa536f1b974","val":1041491834},{"key":"9164766d-ca5c-4a39-9306-38f195abd67d","val":1184894607},{"key":"a5b5537f-2aa1-48e1-aabd-4c0e85b95fff","val":3173326878}]},"master_zonegroup":"a","master_zone":"a5b5537f-2aa1-48e1-aabd-4c0e85b95fff","period_config":{"bucket_quota":{"enabled":false,"check_on_raw":false,"max_size":-1,"max_size_kb":0,"max_objects":-1},"user_quota":{"enabled":false,"check_on_raw":false,"max_size":-1,"max_size_kb":0,"max_objects":-1},"user_ratelimit":{"max_read_ops":0,"max_write_ops":0,"max_read_bytes":0,"max_write_bytes":0,"enabled":false},"bucket_ratelimit":{"max_read_ops":0,"max_write_ops":0,"max_read_bytes":0,"max_write_bytes":0,"enabled":false},"anonymous_ratelimit":{"max_read_ops":0,"max_write_ops":0,"max_read_bytes":0,"max_write_bytes":0,"enabled":false}},"realm_id":"f5bc16fe-b476-42e1-93d5-47c86d0b7983","realm_epoch":2} 2026-03-20T18:00:07.478 DEBUG:tasks.rgw.c2.client.1:waiting for process to exit 2026-03-20T18:00:07.478 INFO:teuthology.orchestra.run:waiting for 300 2026-03-20T18:00:07.479 INFO:tasks.rgw.c2.client.1.vm08.stdout:2026-03-20T18:00:07.477+0000 7fcdf0ed3640 -1 received signal: Terminated from /usr/bin/python3 /usr/bin/daemon-helper term radosgw --rgw-frontends beast port=8001 -n client.1 --cluster c2 -k /etc/ceph/c2.client.1.keyring --log-file /var/log/ceph/rgw.c2.client.1.log --rgw_ops_log_socket_path /home/ubuntu/cephtest/rgw.opslog.c2.client.1.sock --foreground (PID: 23478) UID: 0 2026-03-20T18:00:07.479 INFO:tasks.rgw.c2.client.1.vm08.stdout:2026-03-20T18:00:07.477+0000 7fcdf471bb40 -1 shutting down 2026-03-20T18:00:25.782 INFO:tasks.rgw.c2.client.1:Stopped 2026-03-20T18:00:25.782 INFO:tasks.rgw.c2.client.1:Restarting daemon 2026-03-20T18:00:25.783 DEBUG:teuthology.orchestra.run.vm08:> sudo adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage daemon-helper term radosgw --rgw-frontends 'beast port=8001' -n client.1 --cluster c2 -k /etc/ceph/c2.client.1.keyring --log-file /var/log/ceph/rgw.c2.client.1.log --rgw_ops_log_socket_path /home/ubuntu/cephtest/rgw.opslog.c2.client.1.sock --foreground --rgw-zone b2 --rgw-zonegroup b --rgw-realm test-realm | sudo tee /var/log/ceph/rgw.c2.client.1.stdout 2>&1 2026-03-20T18:00:25.783 INFO:tasks.rgw.c2.client.1:Started 2026-03-20T18:00:25.784 DEBUG:teuthology.orchestra.run.vm08:> curl http://vm08.local:8001 2026-03-20T18:00:25.789 INFO:teuthology.orchestra.run.vm08.stderr: % Total % Received % Xferd Average Speed Time Time Time Current 2026-03-20T18:00:25.789 INFO:teuthology.orchestra.run.vm08.stderr: Dload Upload Total Spent Left Speed 2026-03-20T18:00:25.789 INFO:teuthology.orchestra.run.vm08.stderr: 0 0 0 0 0 0 0 0 --:--:-- --:--:-- --:--:-- 0 0 0 0 0 0 0 0 0 --:--:-- --:--:-- --:--:-- 0 2026-03-20T18:00:25.789 INFO:teuthology.orchestra.run.vm08.stderr:curl: (7) Failed to connect to vm08.local port 8001 after 0 ms: Connection refused 2026-03-20T18:00:25.789 DEBUG:teuthology.orchestra.run:got remote process result: 7 2026-03-20T18:00:25.796 INFO:tasks.daemonwatchdog.daemon_watchdog:daemon c2.rgw.client.1 is failed for ~0s 2026-03-20T18:00:26.790 DEBUG:teuthology.orchestra.run.vm08:> curl http://vm08.local:8001 2026-03-20T18:00:26.803 INFO:teuthology.orchestra.run.vm08.stderr: % Total % Received % Xferd Average Speed Time Time Time Current 2026-03-20T18:00:26.803 INFO:teuthology.orchestra.run.vm08.stderr: Dload Upload Total Spent Left Speed 2026-03-20T18:00:26.803 INFO:teuthology.orchestra.run.vm08.stderr: 0 0 0 0 0 0 0 0 --:--:-- --:--:-- --:--:-- 0 100 187 0 187 0 0 39409 0 --:--:-- --:--:-- --:--:-- 46750 2026-03-20T18:00:26.803 INFO:teuthology.orchestra.run.vm08.stdout:anonymous 2026-03-20T18:00:26.803 INFO:tasks.rgw_multisite:rgw multisite configuration completed 2026-03-20T18:00:26.803 INFO:teuthology.run_tasks:Running task rgw-multisite-tests... 2026-03-20T18:00:26.806 INFO:tasks.rgw_multisite_tests:importing tests from /home/teuthos/src/github.com_kshtsk_ceph_c04fae7bd18063e6f8f9e1cf40c23f810c1f6ca4/qa/../src/test/rgw/rgw_multi 2026-03-20T18:00:26.827 INFO:tasks.rgw_multisite_tests:creating test user.. 2026-03-20T18:00:26.827 INFO:tasks.util.rgw:rgwadmin: c1.client.0 : ['account', 'create', '--account-id', 'RGW11111111111111111', '--rgw-zone', 'a1', '--rgw-zonegroup', 'a', '--rgw-realm', 'test-realm', '--cluster', 'c1', '--debug-rgw', '0', '--debug-ms', '0'] 2026-03-20T18:00:26.827 DEBUG:tasks.util.rgw:rgwadmin: cmd=['adjust-ulimits', 'ceph-coverage', '/home/ubuntu/cephtest/archive/coverage', 'radosgw-admin', '--log-to-stderr', '--format', 'json', '-n', 'client.0', '--cluster', 'c1', 'account', 'create', '--account-id', 'RGW11111111111111111', '--rgw-zone', 'a1', '--rgw-zonegroup', 'a', '--rgw-realm', 'test-realm', '--cluster', 'c1', '--debug-rgw', '0', '--debug-ms', '0'] 2026-03-20T18:00:26.827 DEBUG:teuthology.orchestra.run.vm02:> adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage radosgw-admin --log-to-stderr --format json -n client.0 --cluster c1 account create --account-id RGW11111111111111111 --rgw-zone a1 --rgw-zonegroup a --rgw-realm test-realm --cluster c1 --debug-rgw 0 --debug-ms 0 2026-03-20T18:00:26.848 INFO:teuthology.orchestra.run.vm02.stderr:ignoring --setuser ceph since I am not root 2026-03-20T18:00:26.849 INFO:teuthology.orchestra.run.vm02.stderr:ignoring --setgroup ceph since I am not root 2026-03-20T18:00:26.931 INFO:teuthology.orchestra.run.vm02.stdout:{"id":"RGW11111111111111111","tenant":"","name":"","email":"","quota":{"enabled":false,"check_on_raw":false,"max_size":-1,"max_size_kb":0,"max_objects":-1},"bucket_quota":{"enabled":false,"check_on_raw":false,"max_size":-1,"max_size_kb":0,"max_objects":-1},"max_users":1000,"max_roles":1000,"max_groups":1000,"max_buckets":1000,"max_access_keys":4} 2026-03-20T18:00:26.931 INFO:tasks.util.rgw:rgwadmin: c1.client.0 : ['user', 'create', '--uid', 'rgw-multisite-test-user', '--account-id', 'RGW11111111111111111', '--account-root', '--rgw-zone', 'a1', '--rgw-zonegroup', 'a', '--rgw-realm', 'test-realm', '--display-name', 'TestUser', '--gen-access-key', '--gen-secret', '--cluster', 'c1', '--debug-rgw', '0', '--debug-ms', '0'] 2026-03-20T18:00:26.931 DEBUG:tasks.util.rgw:rgwadmin: cmd=['adjust-ulimits', 'ceph-coverage', '/home/ubuntu/cephtest/archive/coverage', 'radosgw-admin', '--log-to-stderr', '--format', 'json', '-n', 'client.0', '--cluster', 'c1', 'user', 'create', '--uid', 'rgw-multisite-test-user', '--account-id', 'RGW11111111111111111', '--account-root', '--rgw-zone', 'a1', '--rgw-zonegroup', 'a', '--rgw-realm', 'test-realm', '--display-name', 'TestUser', '--gen-access-key', '--gen-secret', '--cluster', 'c1', '--debug-rgw', '0', '--debug-ms', '0'] 2026-03-20T18:00:26.931 DEBUG:teuthology.orchestra.run.vm02:> adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage radosgw-admin --log-to-stderr --format json -n client.0 --cluster c1 user create --uid rgw-multisite-test-user --account-id RGW11111111111111111 --account-root --rgw-zone a1 --rgw-zonegroup a --rgw-realm test-realm --display-name TestUser --gen-access-key --gen-secret --cluster c1 --debug-rgw 0 --debug-ms 0 2026-03-20T18:00:26.997 INFO:teuthology.orchestra.run.vm02.stderr:ignoring --setuser ceph since I am not root 2026-03-20T18:00:26.998 INFO:teuthology.orchestra.run.vm02.stderr:ignoring --setgroup ceph since I am not root 2026-03-20T18:00:27.076 INFO:teuthology.orchestra.run.vm02.stdout:{"user_id":"rgw-multisite-test-user","display_name":"TestUser","email":"","suspended":0,"max_buckets":1000,"subusers":[],"keys":[{"user":"rgw-multisite-test-user","access_key":"C6S7C67V5NF4C49X83GD","secret_key":"nu6i2eRB9n0yZVbnUFGNMkjXAZKcXgMadG9gwYXf","active":true,"create_date":"2026-03-20T18:00:27.069821Z"}],"swift_keys":[],"caps":[],"op_mask":"read, write, delete","default_placement":"","default_storage_class":"","placement_tags":[],"bucket_quota":{"enabled":false,"check_on_raw":false,"max_size":-1,"max_size_kb":0,"max_objects":-1},"user_quota":{"enabled":false,"check_on_raw":false,"max_size":-1,"max_size_kb":0,"max_objects":-1},"temp_url_keys":[],"type":"root","mfa_ids":[],"account_id":"RGW11111111111111111","path":"/","create_date":"2026-03-20T18:00:27.069540Z","tags":[],"group_ids":[]} 2026-03-20T18:00:27.086 INFO:tasks.rgw_multisite_tests:creating non-account user.. 2026-03-20T18:00:27.086 INFO:tasks.util.rgw:rgwadmin: c1.client.0 : ['user', 'create', '--uid', 'rgw-multisite-test-non-account-user', '--rgw-zone', 'a1', '--rgw-zonegroup', 'a', '--rgw-realm', 'test-realm', '--display-name', 'NonAccountUser', '--gen-access-key', '--gen-secret', '--cluster', 'c1', '--debug-rgw', '0', '--debug-ms', '0'] 2026-03-20T18:00:27.086 DEBUG:tasks.util.rgw:rgwadmin: cmd=['adjust-ulimits', 'ceph-coverage', '/home/ubuntu/cephtest/archive/coverage', 'radosgw-admin', '--log-to-stderr', '--format', 'json', '-n', 'client.0', '--cluster', 'c1', 'user', 'create', '--uid', 'rgw-multisite-test-non-account-user', '--rgw-zone', 'a1', '--rgw-zonegroup', 'a', '--rgw-realm', 'test-realm', '--display-name', 'NonAccountUser', '--gen-access-key', '--gen-secret', '--cluster', 'c1', '--debug-rgw', '0', '--debug-ms', '0'] 2026-03-20T18:00:27.086 DEBUG:teuthology.orchestra.run.vm02:> adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage radosgw-admin --log-to-stderr --format json -n client.0 --cluster c1 user create --uid rgw-multisite-test-non-account-user --rgw-zone a1 --rgw-zonegroup a --rgw-realm test-realm --display-name NonAccountUser --gen-access-key --gen-secret --cluster c1 --debug-rgw 0 --debug-ms 0 2026-03-20T18:00:27.148 INFO:teuthology.orchestra.run.vm02.stderr:ignoring --setuser ceph since I am not root 2026-03-20T18:00:27.148 INFO:teuthology.orchestra.run.vm02.stderr:ignoring --setgroup ceph since I am not root 2026-03-20T18:00:27.221 INFO:teuthology.orchestra.run.vm02.stdout:{"user_id":"rgw-multisite-test-non-account-user","display_name":"NonAccountUser","email":"","suspended":0,"max_buckets":1000,"subusers":[],"keys":[{"user":"rgw-multisite-test-non-account-user","access_key":"5EBIZQXEADDQN3D313PH","secret_key":"9AmP5r1AdlQiqHd2XQBQ28ppsExGjjS5gBqraBkj","active":true,"create_date":"2026-03-20T18:00:27.216961Z"}],"swift_keys":[],"caps":[],"op_mask":"read, write, delete","default_placement":"","default_storage_class":"","placement_tags":[],"bucket_quota":{"enabled":false,"check_on_raw":false,"max_size":-1,"max_size_kb":0,"max_objects":-1},"user_quota":{"enabled":false,"check_on_raw":false,"max_size":-1,"max_size_kb":0,"max_objects":-1},"temp_url_keys":[],"type":"rgw","mfa_ids":[],"account_id":"","path":"/","create_date":"2026-03-20T18:00:27.216743Z","tags":[],"group_ids":[]} 2026-03-20T18:00:27.231 INFO:tasks.rgw_multisite_tests:creating non-account alt user.. 2026-03-20T18:00:27.231 INFO:tasks.util.rgw:rgwadmin: c1.client.0 : ['user', 'create', '--uid', 'rgw-multisite-test-non-account-alt-user', '--rgw-zone', 'a1', '--rgw-zonegroup', 'a', '--rgw-realm', 'test-realm', '--display-name', 'NonAccountAltUser', '--gen-access-key', '--gen-secret', '--cluster', 'c1', '--debug-rgw', '0', '--debug-ms', '0'] 2026-03-20T18:00:27.231 DEBUG:tasks.util.rgw:rgwadmin: cmd=['adjust-ulimits', 'ceph-coverage', '/home/ubuntu/cephtest/archive/coverage', 'radosgw-admin', '--log-to-stderr', '--format', 'json', '-n', 'client.0', '--cluster', 'c1', 'user', 'create', '--uid', 'rgw-multisite-test-non-account-alt-user', '--rgw-zone', 'a1', '--rgw-zonegroup', 'a', '--rgw-realm', 'test-realm', '--display-name', 'NonAccountAltUser', '--gen-access-key', '--gen-secret', '--cluster', 'c1', '--debug-rgw', '0', '--debug-ms', '0'] 2026-03-20T18:00:27.231 DEBUG:teuthology.orchestra.run.vm02:> adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage radosgw-admin --log-to-stderr --format json -n client.0 --cluster c1 user create --uid rgw-multisite-test-non-account-alt-user --rgw-zone a1 --rgw-zonegroup a --rgw-realm test-realm --display-name NonAccountAltUser --gen-access-key --gen-secret --cluster c1 --debug-rgw 0 --debug-ms 0 2026-03-20T18:00:27.299 INFO:teuthology.orchestra.run.vm02.stderr:ignoring --setuser ceph since I am not root 2026-03-20T18:00:27.299 INFO:teuthology.orchestra.run.vm02.stderr:ignoring --setgroup ceph since I am not root 2026-03-20T18:00:27.363 INFO:teuthology.orchestra.run.vm02.stdout:{"user_id":"rgw-multisite-test-non-account-alt-user","display_name":"NonAccountAltUser","email":"","suspended":0,"max_buckets":1000,"subusers":[],"keys":[{"user":"rgw-multisite-test-non-account-alt-user","access_key":"I1VJP2P6PH3U2ULDR6QP","secret_key":"ZOOD2GLr9ikckDB76WH5bvAMTZ4JY6g5hhBOt1lL","active":true,"create_date":"2026-03-20T18:00:27.358056Z"}],"swift_keys":[],"caps":[],"op_mask":"read, write, delete","default_placement":"","default_storage_class":"","placement_tags":[],"bucket_quota":{"enabled":false,"check_on_raw":false,"max_size":-1,"max_size_kb":0,"max_objects":-1},"user_quota":{"enabled":false,"check_on_raw":false,"max_size":-1,"max_size_kb":0,"max_objects":-1},"temp_url_keys":[],"type":"rgw","mfa_ids":[],"account_id":"","path":"/","create_date":"2026-03-20T18:00:27.357876Z","tags":[],"group_ids":[]} 2026-03-20T18:00:27.371 INFO:rgw_multi.tests:meta checkpoint 2026-03-20T18:00:27.371 INFO:tasks.util.rgw:rgwadmin: c1.client.0 : ['mdlog', 'status', '--rgw-zone', 'a1', '--rgw-zonegroup', 'a', '--rgw-realm', 'test-realm', '--cluster', 'c1', '--debug-rgw', '0', '--debug-ms', '0', '--rgw-cache-enabled', 'false'] 2026-03-20T18:00:27.371 DEBUG:tasks.util.rgw:rgwadmin: cmd=['adjust-ulimits', 'ceph-coverage', '/home/ubuntu/cephtest/archive/coverage', 'radosgw-admin', '--log-to-stderr', '--format', 'json', '-n', 'client.0', '--cluster', 'c1', 'mdlog', 'status', '--rgw-zone', 'a1', '--rgw-zonegroup', 'a', '--rgw-realm', 'test-realm', '--cluster', 'c1', '--debug-rgw', '0', '--debug-ms', '0', '--rgw-cache-enabled', 'false'] 2026-03-20T18:00:27.371 DEBUG:teuthology.orchestra.run.vm02:> adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage radosgw-admin --log-to-stderr --format json -n client.0 --cluster c1 mdlog status --rgw-zone a1 --rgw-zonegroup a --rgw-realm test-realm --cluster c1 --debug-rgw 0 --debug-ms 0 --rgw-cache-enabled false 2026-03-20T18:00:27.433 INFO:teuthology.orchestra.run.vm02.stderr:ignoring --setuser ceph since I am not root 2026-03-20T18:00:27.433 INFO:teuthology.orchestra.run.vm02.stderr:ignoring --setgroup ceph since I am not root 2026-03-20T18:00:27.495 INFO:teuthology.orchestra.run.vm02.stderr:No --period given, using current period=82c662be-66b5-4c4f-b24e-ff9298fc8a6e 2026-03-20T18:00:27.504 INFO:teuthology.orchestra.run.vm02.stdout:[{"marker":"1_1774029627.072292_68.1","last_update":"2026-03-20T18:00:27.072292Z"},{"marker":"","last_update":"0.000000"},{"marker":"1_1774029627.218585_92.1","last_update":"2026-03-20T18:00:27.218585Z"},{"marker":"1_1774029627.359834_56.1","last_update":"2026-03-20T18:00:27.359834Z"}] 2026-03-20T18:00:27.504 INFO:rgw_multi.tests:starting meta checkpoint for zone=a2 2026-03-20T18:00:27.504 INFO:tasks.util.rgw:rgwadmin: c1.client.0 : ['metadata', 'sync', 'status', '--rgw-zone', 'a2', '--rgw-zonegroup', 'a', '--rgw-realm', 'test-realm', '--cluster', 'c1', '--debug-rgw', '0', '--debug-ms', '0', '--rgw-cache-enabled', 'false'] 2026-03-20T18:00:27.504 DEBUG:tasks.util.rgw:rgwadmin: cmd=['adjust-ulimits', 'ceph-coverage', '/home/ubuntu/cephtest/archive/coverage', 'radosgw-admin', '--log-to-stderr', '--format', 'json', '-n', 'client.0', '--cluster', 'c1', 'metadata', 'sync', 'status', '--rgw-zone', 'a2', '--rgw-zonegroup', 'a', '--rgw-realm', 'test-realm', '--cluster', 'c1', '--debug-rgw', '0', '--debug-ms', '0', '--rgw-cache-enabled', 'false'] 2026-03-20T18:00:27.504 DEBUG:teuthology.orchestra.run.vm02:> adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage radosgw-admin --log-to-stderr --format json -n client.0 --cluster c1 metadata sync status --rgw-zone a2 --rgw-zonegroup a --rgw-realm test-realm --cluster c1 --debug-rgw 0 --debug-ms 0 --rgw-cache-enabled false 2026-03-20T18:00:27.570 INFO:teuthology.orchestra.run.vm02.stderr:ignoring --setuser ceph since I am not root 2026-03-20T18:00:27.570 INFO:teuthology.orchestra.run.vm02.stderr:ignoring --setgroup ceph since I am not root 2026-03-20T18:00:27.629 INFO:teuthology.orchestra.run.vm02.stderr:2026-03-20T18:00:27.625+0000 7fc0d0a66a80 20 RGW-SYNC:meta: read sync status 2026-03-20T18:00:27.629 INFO:teuthology.orchestra.run.vm02.stderr:2026-03-20T18:00:27.625+0000 7fc0d0a66a80 20 RGW-SYNC:meta: read sync status 2026-03-20T18:00:27.639 INFO:teuthology.orchestra.run.vm02.stdout:{"sync_status":{"info":{"status":"sync","num_shards":4,"period":"82c662be-66b5-4c4f-b24e-ff9298fc8a6e","realm_epoch":2},"markers":[{"key":0,"val":{"state":1,"marker":"","next_step_marker":"","total_entries":1,"pos":0,"timestamp":"0.000000","realm_epoch":2}},{"key":1,"val":{"state":1,"marker":"","next_step_marker":"","total_entries":0,"pos":0,"timestamp":"0.000000","realm_epoch":2}},{"key":2,"val":{"state":1,"marker":"","next_step_marker":"","total_entries":0,"pos":0,"timestamp":"0.000000","realm_epoch":2}},{"key":3,"val":{"state":1,"marker":"","next_step_marker":"","total_entries":0,"pos":0,"timestamp":"0.000000","realm_epoch":2}}]},"full_sync":{"total":1,"complete":1},"current_time":"2026-03-20T18:00:27Z"} 2026-03-20T18:00:27.639 WARNING:rgw_multi.tests:zone a2 behind master: shard=0 master=1_1774029627.072292_68.1 target=, shard=2 master=1_1774029627.218585_92.1 target=, shard=3 master=1_1774029627.359834_56.1 target= 2026-03-20T18:00:28.931 INFO:tasks.rgw.c2.client.0.vm06.stdout:2026-03-20T18:00:28.923+0000 7f3113bd2640 -1 req 3972785660454311698 0.004000000s :post_period failed to decode period 2026-03-20T18:00:31.802 INFO:tasks.daemonwatchdog.daemon_watchdog:daemon c2.rgw.client.1 has been restored 2026-03-20T18:00:32.639 INFO:tasks.util.rgw:rgwadmin: c1.client.0 : ['metadata', 'sync', 'status', '--rgw-zone', 'a2', '--rgw-zonegroup', 'a', '--rgw-realm', 'test-realm', '--cluster', 'c1', '--debug-rgw', '0', '--debug-ms', '0', '--rgw-cache-enabled', 'false'] 2026-03-20T18:00:32.639 DEBUG:tasks.util.rgw:rgwadmin: cmd=['adjust-ulimits', 'ceph-coverage', '/home/ubuntu/cephtest/archive/coverage', 'radosgw-admin', '--log-to-stderr', '--format', 'json', '-n', 'client.0', '--cluster', 'c1', 'metadata', 'sync', 'status', '--rgw-zone', 'a2', '--rgw-zonegroup', 'a', '--rgw-realm', 'test-realm', '--cluster', 'c1', '--debug-rgw', '0', '--debug-ms', '0', '--rgw-cache-enabled', 'false'] 2026-03-20T18:00:32.639 DEBUG:teuthology.orchestra.run.vm02:> adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage radosgw-admin --log-to-stderr --format json -n client.0 --cluster c1 metadata sync status --rgw-zone a2 --rgw-zonegroup a --rgw-realm test-realm --cluster c1 --debug-rgw 0 --debug-ms 0 --rgw-cache-enabled false 2026-03-20T18:00:32.660 INFO:teuthology.orchestra.run.vm02.stderr:ignoring --setuser ceph since I am not root 2026-03-20T18:00:32.660 INFO:teuthology.orchestra.run.vm02.stderr:ignoring --setgroup ceph since I am not root 2026-03-20T18:00:32.725 INFO:teuthology.orchestra.run.vm02.stderr:2026-03-20T18:00:32.721+0000 7f8b236a1a80 20 RGW-SYNC:meta: read sync status 2026-03-20T18:00:32.726 INFO:teuthology.orchestra.run.vm02.stderr:2026-03-20T18:00:32.721+0000 7f8b236a1a80 20 RGW-SYNC:meta: read sync status 2026-03-20T18:00:32.734 INFO:teuthology.orchestra.run.vm02.stdout:{"sync_status":{"info":{"status":"sync","num_shards":4,"period":"82c662be-66b5-4c4f-b24e-ff9298fc8a6e","realm_epoch":2},"markers":[{"key":0,"val":{"state":1,"marker":"","next_step_marker":"","total_entries":1,"pos":0,"timestamp":"0.000000","realm_epoch":2}},{"key":1,"val":{"state":1,"marker":"","next_step_marker":"","total_entries":0,"pos":0,"timestamp":"0.000000","realm_epoch":2}},{"key":2,"val":{"state":1,"marker":"","next_step_marker":"","total_entries":0,"pos":0,"timestamp":"0.000000","realm_epoch":2}},{"key":3,"val":{"state":1,"marker":"","next_step_marker":"","total_entries":0,"pos":0,"timestamp":"0.000000","realm_epoch":2}}]},"full_sync":{"total":1,"complete":1},"current_time":"2026-03-20T18:00:32Z"} 2026-03-20T18:00:32.734 WARNING:rgw_multi.tests:zone a2 behind master: shard=0 master=1_1774029627.072292_68.1 target=, shard=2 master=1_1774029627.218585_92.1 target=, shard=3 master=1_1774029627.359834_56.1 target= 2026-03-20T18:00:37.735 INFO:tasks.util.rgw:rgwadmin: c1.client.0 : ['metadata', 'sync', 'status', '--rgw-zone', 'a2', '--rgw-zonegroup', 'a', '--rgw-realm', 'test-realm', '--cluster', 'c1', '--debug-rgw', '0', '--debug-ms', '0', '--rgw-cache-enabled', 'false'] 2026-03-20T18:00:37.735 DEBUG:tasks.util.rgw:rgwadmin: cmd=['adjust-ulimits', 'ceph-coverage', '/home/ubuntu/cephtest/archive/coverage', 'radosgw-admin', '--log-to-stderr', '--format', 'json', '-n', 'client.0', '--cluster', 'c1', 'metadata', 'sync', 'status', '--rgw-zone', 'a2', '--rgw-zonegroup', 'a', '--rgw-realm', 'test-realm', '--cluster', 'c1', '--debug-rgw', '0', '--debug-ms', '0', '--rgw-cache-enabled', 'false'] 2026-03-20T18:00:37.735 DEBUG:teuthology.orchestra.run.vm02:> adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage radosgw-admin --log-to-stderr --format json -n client.0 --cluster c1 metadata sync status --rgw-zone a2 --rgw-zonegroup a --rgw-realm test-realm --cluster c1 --debug-rgw 0 --debug-ms 0 --rgw-cache-enabled false 2026-03-20T18:00:37.761 INFO:teuthology.orchestra.run.vm02.stderr:ignoring --setuser ceph since I am not root 2026-03-20T18:00:37.761 INFO:teuthology.orchestra.run.vm02.stderr:ignoring --setgroup ceph since I am not root 2026-03-20T18:00:37.826 INFO:teuthology.orchestra.run.vm02.stderr:2026-03-20T18:00:37.821+0000 7fc44b705a80 20 RGW-SYNC:meta: read sync status 2026-03-20T18:00:37.827 INFO:teuthology.orchestra.run.vm02.stderr:2026-03-20T18:00:37.825+0000 7fc44b705a80 20 RGW-SYNC:meta: read sync status 2026-03-20T18:00:37.837 INFO:teuthology.orchestra.run.vm02.stdout:{"sync_status":{"info":{"status":"sync","num_shards":4,"period":"82c662be-66b5-4c4f-b24e-ff9298fc8a6e","realm_epoch":2},"markers":[{"key":0,"val":{"state":1,"marker":"","next_step_marker":"","total_entries":1,"pos":0,"timestamp":"0.000000","realm_epoch":2}},{"key":1,"val":{"state":1,"marker":"","next_step_marker":"","total_entries":0,"pos":0,"timestamp":"0.000000","realm_epoch":2}},{"key":2,"val":{"state":1,"marker":"","next_step_marker":"","total_entries":0,"pos":0,"timestamp":"0.000000","realm_epoch":2}},{"key":3,"val":{"state":1,"marker":"","next_step_marker":"","total_entries":0,"pos":0,"timestamp":"0.000000","realm_epoch":2}}]},"full_sync":{"total":1,"complete":1},"current_time":"2026-03-20T18:00:37Z"} 2026-03-20T18:00:37.837 WARNING:rgw_multi.tests:zone a2 behind master: shard=0 master=1_1774029627.072292_68.1 target=, shard=2 master=1_1774029627.218585_92.1 target=, shard=3 master=1_1774029627.359834_56.1 target= 2026-03-20T18:00:42.838 INFO:tasks.util.rgw:rgwadmin: c1.client.0 : ['metadata', 'sync', 'status', '--rgw-zone', 'a2', '--rgw-zonegroup', 'a', '--rgw-realm', 'test-realm', '--cluster', 'c1', '--debug-rgw', '0', '--debug-ms', '0', '--rgw-cache-enabled', 'false'] 2026-03-20T18:00:42.838 DEBUG:tasks.util.rgw:rgwadmin: cmd=['adjust-ulimits', 'ceph-coverage', '/home/ubuntu/cephtest/archive/coverage', 'radosgw-admin', '--log-to-stderr', '--format', 'json', '-n', 'client.0', '--cluster', 'c1', 'metadata', 'sync', 'status', '--rgw-zone', 'a2', '--rgw-zonegroup', 'a', '--rgw-realm', 'test-realm', '--cluster', 'c1', '--debug-rgw', '0', '--debug-ms', '0', '--rgw-cache-enabled', 'false'] 2026-03-20T18:00:42.838 DEBUG:teuthology.orchestra.run.vm02:> adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage radosgw-admin --log-to-stderr --format json -n client.0 --cluster c1 metadata sync status --rgw-zone a2 --rgw-zonegroup a --rgw-realm test-realm --cluster c1 --debug-rgw 0 --debug-ms 0 --rgw-cache-enabled false 2026-03-20T18:00:42.862 INFO:teuthology.orchestra.run.vm02.stderr:ignoring --setuser ceph since I am not root 2026-03-20T18:00:42.862 INFO:teuthology.orchestra.run.vm02.stderr:ignoring --setgroup ceph since I am not root 2026-03-20T18:00:42.928 INFO:teuthology.orchestra.run.vm02.stderr:2026-03-20T18:00:42.925+0000 7fe422a24a80 20 RGW-SYNC:meta: read sync status 2026-03-20T18:00:42.929 INFO:teuthology.orchestra.run.vm02.stderr:2026-03-20T18:00:42.925+0000 7fe422a24a80 20 RGW-SYNC:meta: read sync status 2026-03-20T18:00:42.937 INFO:teuthology.orchestra.run.vm02.stdout:{"sync_status":{"info":{"status":"sync","num_shards":4,"period":"82c662be-66b5-4c4f-b24e-ff9298fc8a6e","realm_epoch":2},"markers":[{"key":0,"val":{"state":1,"marker":"","next_step_marker":"","total_entries":1,"pos":0,"timestamp":"0.000000","realm_epoch":2}},{"key":1,"val":{"state":1,"marker":"","next_step_marker":"","total_entries":0,"pos":0,"timestamp":"0.000000","realm_epoch":2}},{"key":2,"val":{"state":1,"marker":"","next_step_marker":"","total_entries":0,"pos":0,"timestamp":"0.000000","realm_epoch":2}},{"key":3,"val":{"state":1,"marker":"","next_step_marker":"","total_entries":0,"pos":0,"timestamp":"0.000000","realm_epoch":2}}]},"full_sync":{"total":1,"complete":1},"current_time":"2026-03-20T18:00:42Z"} 2026-03-20T18:00:42.938 WARNING:rgw_multi.tests:zone a2 behind master: shard=0 master=1_1774029627.072292_68.1 target=, shard=2 master=1_1774029627.218585_92.1 target=, shard=3 master=1_1774029627.359834_56.1 target= 2026-03-20T18:00:47.939 INFO:tasks.util.rgw:rgwadmin: c1.client.0 : ['metadata', 'sync', 'status', '--rgw-zone', 'a2', '--rgw-zonegroup', 'a', '--rgw-realm', 'test-realm', '--cluster', 'c1', '--debug-rgw', '0', '--debug-ms', '0', '--rgw-cache-enabled', 'false'] 2026-03-20T18:00:47.939 DEBUG:tasks.util.rgw:rgwadmin: cmd=['adjust-ulimits', 'ceph-coverage', '/home/ubuntu/cephtest/archive/coverage', 'radosgw-admin', '--log-to-stderr', '--format', 'json', '-n', 'client.0', '--cluster', 'c1', 'metadata', 'sync', 'status', '--rgw-zone', 'a2', '--rgw-zonegroup', 'a', '--rgw-realm', 'test-realm', '--cluster', 'c1', '--debug-rgw', '0', '--debug-ms', '0', '--rgw-cache-enabled', 'false'] 2026-03-20T18:00:47.939 DEBUG:teuthology.orchestra.run.vm02:> adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage radosgw-admin --log-to-stderr --format json -n client.0 --cluster c1 metadata sync status --rgw-zone a2 --rgw-zonegroup a --rgw-realm test-realm --cluster c1 --debug-rgw 0 --debug-ms 0 --rgw-cache-enabled false 2026-03-20T18:00:47.960 INFO:teuthology.orchestra.run.vm02.stderr:ignoring --setuser ceph since I am not root 2026-03-20T18:00:47.960 INFO:teuthology.orchestra.run.vm02.stderr:ignoring --setgroup ceph since I am not root 2026-03-20T18:00:48.028 INFO:teuthology.orchestra.run.vm02.stderr:2026-03-20T18:00:48.025+0000 7fc516ffba80 20 RGW-SYNC:meta: read sync status 2026-03-20T18:00:48.029 INFO:teuthology.orchestra.run.vm02.stderr:2026-03-20T18:00:48.025+0000 7fc516ffba80 20 RGW-SYNC:meta: read sync status 2026-03-20T18:00:48.038 INFO:teuthology.orchestra.run.vm02.stdout:{"sync_status":{"info":{"status":"sync","num_shards":4,"period":"82c662be-66b5-4c4f-b24e-ff9298fc8a6e","realm_epoch":2},"markers":[{"key":0,"val":{"state":1,"marker":"","next_step_marker":"","total_entries":1,"pos":0,"timestamp":"0.000000","realm_epoch":2}},{"key":1,"val":{"state":1,"marker":"","next_step_marker":"","total_entries":0,"pos":0,"timestamp":"0.000000","realm_epoch":2}},{"key":2,"val":{"state":1,"marker":"","next_step_marker":"","total_entries":0,"pos":0,"timestamp":"0.000000","realm_epoch":2}},{"key":3,"val":{"state":1,"marker":"","next_step_marker":"","total_entries":0,"pos":0,"timestamp":"0.000000","realm_epoch":2}}]},"full_sync":{"total":1,"complete":1},"current_time":"2026-03-20T18:00:48Z"} 2026-03-20T18:00:48.038 WARNING:rgw_multi.tests:zone a2 behind master: shard=0 master=1_1774029627.072292_68.1 target=, shard=2 master=1_1774029627.218585_92.1 target=, shard=3 master=1_1774029627.359834_56.1 target= 2026-03-20T18:00:53.039 INFO:tasks.util.rgw:rgwadmin: c1.client.0 : ['metadata', 'sync', 'status', '--rgw-zone', 'a2', '--rgw-zonegroup', 'a', '--rgw-realm', 'test-realm', '--cluster', 'c1', '--debug-rgw', '0', '--debug-ms', '0', '--rgw-cache-enabled', 'false'] 2026-03-20T18:00:53.039 DEBUG:tasks.util.rgw:rgwadmin: cmd=['adjust-ulimits', 'ceph-coverage', '/home/ubuntu/cephtest/archive/coverage', 'radosgw-admin', '--log-to-stderr', '--format', 'json', '-n', 'client.0', '--cluster', 'c1', 'metadata', 'sync', 'status', '--rgw-zone', 'a2', '--rgw-zonegroup', 'a', '--rgw-realm', 'test-realm', '--cluster', 'c1', '--debug-rgw', '0', '--debug-ms', '0', '--rgw-cache-enabled', 'false'] 2026-03-20T18:00:53.040 DEBUG:teuthology.orchestra.run.vm02:> adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage radosgw-admin --log-to-stderr --format json -n client.0 --cluster c1 metadata sync status --rgw-zone a2 --rgw-zonegroup a --rgw-realm test-realm --cluster c1 --debug-rgw 0 --debug-ms 0 --rgw-cache-enabled false 2026-03-20T18:00:53.064 INFO:teuthology.orchestra.run.vm02.stderr:ignoring --setuser ceph since I am not root 2026-03-20T18:00:53.064 INFO:teuthology.orchestra.run.vm02.stderr:ignoring --setgroup ceph since I am not root 2026-03-20T18:00:53.139 INFO:teuthology.orchestra.run.vm02.stderr:2026-03-20T18:00:53.137+0000 7f177b3e7a80 20 RGW-SYNC:meta: read sync status 2026-03-20T18:00:53.139 INFO:teuthology.orchestra.run.vm02.stderr:2026-03-20T18:00:53.137+0000 7f177b3e7a80 20 RGW-SYNC:meta: read sync status 2026-03-20T18:00:53.150 INFO:teuthology.orchestra.run.vm02.stdout:{"sync_status":{"info":{"status":"sync","num_shards":4,"period":"82c662be-66b5-4c4f-b24e-ff9298fc8a6e","realm_epoch":2},"markers":[{"key":0,"val":{"state":1,"marker":"","next_step_marker":"","total_entries":1,"pos":0,"timestamp":"0.000000","realm_epoch":2}},{"key":1,"val":{"state":1,"marker":"","next_step_marker":"","total_entries":0,"pos":0,"timestamp":"0.000000","realm_epoch":2}},{"key":2,"val":{"state":1,"marker":"","next_step_marker":"","total_entries":0,"pos":0,"timestamp":"0.000000","realm_epoch":2}},{"key":3,"val":{"state":1,"marker":"","next_step_marker":"","total_entries":0,"pos":0,"timestamp":"0.000000","realm_epoch":2}}]},"full_sync":{"total":1,"complete":1},"current_time":"2026-03-20T18:00:53Z"} 2026-03-20T18:00:53.150 WARNING:rgw_multi.tests:zone a2 behind master: shard=0 master=1_1774029627.072292_68.1 target=, shard=2 master=1_1774029627.218585_92.1 target=, shard=3 master=1_1774029627.359834_56.1 target= 2026-03-20T18:00:58.151 INFO:tasks.util.rgw:rgwadmin: c1.client.0 : ['metadata', 'sync', 'status', '--rgw-zone', 'a2', '--rgw-zonegroup', 'a', '--rgw-realm', 'test-realm', '--cluster', 'c1', '--debug-rgw', '0', '--debug-ms', '0', '--rgw-cache-enabled', 'false'] 2026-03-20T18:00:58.151 DEBUG:tasks.util.rgw:rgwadmin: cmd=['adjust-ulimits', 'ceph-coverage', '/home/ubuntu/cephtest/archive/coverage', 'radosgw-admin', '--log-to-stderr', '--format', 'json', '-n', 'client.0', '--cluster', 'c1', 'metadata', 'sync', 'status', '--rgw-zone', 'a2', '--rgw-zonegroup', 'a', '--rgw-realm', 'test-realm', '--cluster', 'c1', '--debug-rgw', '0', '--debug-ms', '0', '--rgw-cache-enabled', 'false'] 2026-03-20T18:00:58.151 DEBUG:teuthology.orchestra.run.vm02:> adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage radosgw-admin --log-to-stderr --format json -n client.0 --cluster c1 metadata sync status --rgw-zone a2 --rgw-zonegroup a --rgw-realm test-realm --cluster c1 --debug-rgw 0 --debug-ms 0 --rgw-cache-enabled false 2026-03-20T18:00:58.173 INFO:teuthology.orchestra.run.vm02.stderr:ignoring --setuser ceph since I am not root 2026-03-20T18:00:58.173 INFO:teuthology.orchestra.run.vm02.stderr:ignoring --setgroup ceph since I am not root 2026-03-20T18:00:58.261 INFO:teuthology.orchestra.run.vm02.stderr:2026-03-20T18:00:58.257+0000 7f35cf3b3a80 20 RGW-SYNC:meta: read sync status 2026-03-20T18:00:58.262 INFO:teuthology.orchestra.run.vm02.stderr:2026-03-20T18:00:58.257+0000 7f35cf3b3a80 20 RGW-SYNC:meta: read sync status 2026-03-20T18:00:58.271 INFO:teuthology.orchestra.run.vm02.stdout:{"sync_status":{"info":{"status":"sync","num_shards":4,"period":"82c662be-66b5-4c4f-b24e-ff9298fc8a6e","realm_epoch":2},"markers":[{"key":0,"val":{"state":1,"marker":"","next_step_marker":"","total_entries":1,"pos":0,"timestamp":"0.000000","realm_epoch":2}},{"key":1,"val":{"state":1,"marker":"","next_step_marker":"","total_entries":0,"pos":0,"timestamp":"0.000000","realm_epoch":2}},{"key":2,"val":{"state":1,"marker":"","next_step_marker":"","total_entries":0,"pos":0,"timestamp":"0.000000","realm_epoch":2}},{"key":3,"val":{"state":1,"marker":"","next_step_marker":"","total_entries":0,"pos":0,"timestamp":"0.000000","realm_epoch":2}}]},"full_sync":{"total":1,"complete":1},"current_time":"2026-03-20T18:00:58Z"} 2026-03-20T18:00:58.271 WARNING:rgw_multi.tests:zone a2 behind master: shard=0 master=1_1774029627.072292_68.1 target=, shard=2 master=1_1774029627.218585_92.1 target=, shard=3 master=1_1774029627.359834_56.1 target= 2026-03-20T18:01:03.272 INFO:tasks.util.rgw:rgwadmin: c1.client.0 : ['metadata', 'sync', 'status', '--rgw-zone', 'a2', '--rgw-zonegroup', 'a', '--rgw-realm', 'test-realm', '--cluster', 'c1', '--debug-rgw', '0', '--debug-ms', '0', '--rgw-cache-enabled', 'false'] 2026-03-20T18:01:03.272 DEBUG:tasks.util.rgw:rgwadmin: cmd=['adjust-ulimits', 'ceph-coverage', '/home/ubuntu/cephtest/archive/coverage', 'radosgw-admin', '--log-to-stderr', '--format', 'json', '-n', 'client.0', '--cluster', 'c1', 'metadata', 'sync', 'status', '--rgw-zone', 'a2', '--rgw-zonegroup', 'a', '--rgw-realm', 'test-realm', '--cluster', 'c1', '--debug-rgw', '0', '--debug-ms', '0', '--rgw-cache-enabled', 'false'] 2026-03-20T18:01:03.272 DEBUG:teuthology.orchestra.run.vm02:> adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage radosgw-admin --log-to-stderr --format json -n client.0 --cluster c1 metadata sync status --rgw-zone a2 --rgw-zonegroup a --rgw-realm test-realm --cluster c1 --debug-rgw 0 --debug-ms 0 --rgw-cache-enabled false 2026-03-20T18:01:03.301 INFO:teuthology.orchestra.run.vm02.stderr:ignoring --setuser ceph since I am not root 2026-03-20T18:01:03.301 INFO:teuthology.orchestra.run.vm02.stderr:ignoring --setgroup ceph since I am not root 2026-03-20T18:01:03.392 INFO:teuthology.orchestra.run.vm02.stderr:2026-03-20T18:01:03.389+0000 7fea7a089a80 20 RGW-SYNC:meta: read sync status 2026-03-20T18:01:03.392 INFO:teuthology.orchestra.run.vm02.stderr:2026-03-20T18:01:03.389+0000 7fea7a089a80 20 RGW-SYNC:meta: read sync status 2026-03-20T18:01:03.405 INFO:teuthology.orchestra.run.vm02.stdout:{"sync_status":{"info":{"status":"sync","num_shards":4,"period":"82c662be-66b5-4c4f-b24e-ff9298fc8a6e","realm_epoch":2},"markers":[{"key":0,"val":{"state":1,"marker":"","next_step_marker":"","total_entries":1,"pos":0,"timestamp":"0.000000","realm_epoch":2}},{"key":1,"val":{"state":1,"marker":"","next_step_marker":"","total_entries":0,"pos":0,"timestamp":"0.000000","realm_epoch":2}},{"key":2,"val":{"state":1,"marker":"","next_step_marker":"","total_entries":0,"pos":0,"timestamp":"0.000000","realm_epoch":2}},{"key":3,"val":{"state":1,"marker":"","next_step_marker":"","total_entries":0,"pos":0,"timestamp":"0.000000","realm_epoch":2}}]},"full_sync":{"total":1,"complete":1},"current_time":"2026-03-20T18:01:03Z"} 2026-03-20T18:01:03.405 WARNING:rgw_multi.tests:zone a2 behind master: shard=0 master=1_1774029627.072292_68.1 target=, shard=2 master=1_1774029627.218585_92.1 target=, shard=3 master=1_1774029627.359834_56.1 target= 2026-03-20T18:01:08.406 INFO:tasks.util.rgw:rgwadmin: c1.client.0 : ['metadata', 'sync', 'status', '--rgw-zone', 'a2', '--rgw-zonegroup', 'a', '--rgw-realm', 'test-realm', '--cluster', 'c1', '--debug-rgw', '0', '--debug-ms', '0', '--rgw-cache-enabled', 'false'] 2026-03-20T18:01:08.406 DEBUG:tasks.util.rgw:rgwadmin: cmd=['adjust-ulimits', 'ceph-coverage', '/home/ubuntu/cephtest/archive/coverage', 'radosgw-admin', '--log-to-stderr', '--format', 'json', '-n', 'client.0', '--cluster', 'c1', 'metadata', 'sync', 'status', '--rgw-zone', 'a2', '--rgw-zonegroup', 'a', '--rgw-realm', 'test-realm', '--cluster', 'c1', '--debug-rgw', '0', '--debug-ms', '0', '--rgw-cache-enabled', 'false'] 2026-03-20T18:01:08.406 DEBUG:teuthology.orchestra.run.vm02:> adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage radosgw-admin --log-to-stderr --format json -n client.0 --cluster c1 metadata sync status --rgw-zone a2 --rgw-zonegroup a --rgw-realm test-realm --cluster c1 --debug-rgw 0 --debug-ms 0 --rgw-cache-enabled false 2026-03-20T18:01:08.427 INFO:teuthology.orchestra.run.vm02.stderr:ignoring --setuser ceph since I am not root 2026-03-20T18:01:08.427 INFO:teuthology.orchestra.run.vm02.stderr:ignoring --setgroup ceph since I am not root 2026-03-20T18:01:08.503 INFO:teuthology.orchestra.run.vm02.stderr:2026-03-20T18:01:08.501+0000 7f6daa332a80 20 RGW-SYNC:meta: read sync status 2026-03-20T18:01:08.504 INFO:teuthology.orchestra.run.vm02.stderr:2026-03-20T18:01:08.501+0000 7f6daa332a80 20 RGW-SYNC:meta: read sync status 2026-03-20T18:01:08.517 INFO:teuthology.orchestra.run.vm02.stdout:{"sync_status":{"info":{"status":"sync","num_shards":4,"period":"82c662be-66b5-4c4f-b24e-ff9298fc8a6e","realm_epoch":2},"markers":[{"key":0,"val":{"state":1,"marker":"","next_step_marker":"","total_entries":1,"pos":0,"timestamp":"0.000000","realm_epoch":2}},{"key":1,"val":{"state":1,"marker":"","next_step_marker":"","total_entries":0,"pos":0,"timestamp":"0.000000","realm_epoch":2}},{"key":2,"val":{"state":1,"marker":"","next_step_marker":"","total_entries":0,"pos":0,"timestamp":"0.000000","realm_epoch":2}},{"key":3,"val":{"state":1,"marker":"","next_step_marker":"","total_entries":0,"pos":0,"timestamp":"0.000000","realm_epoch":2}}]},"full_sync":{"total":1,"complete":1},"current_time":"2026-03-20T18:01:08Z"} 2026-03-20T18:01:08.517 WARNING:rgw_multi.tests:zone a2 behind master: shard=0 master=1_1774029627.072292_68.1 target=, shard=2 master=1_1774029627.218585_92.1 target=, shard=3 master=1_1774029627.359834_56.1 target= 2026-03-20T18:01:13.518 INFO:tasks.util.rgw:rgwadmin: c1.client.0 : ['metadata', 'sync', 'status', '--rgw-zone', 'a2', '--rgw-zonegroup', 'a', '--rgw-realm', 'test-realm', '--cluster', 'c1', '--debug-rgw', '0', '--debug-ms', '0', '--rgw-cache-enabled', 'false'] 2026-03-20T18:01:13.518 DEBUG:tasks.util.rgw:rgwadmin: cmd=['adjust-ulimits', 'ceph-coverage', '/home/ubuntu/cephtest/archive/coverage', 'radosgw-admin', '--log-to-stderr', '--format', 'json', '-n', 'client.0', '--cluster', 'c1', 'metadata', 'sync', 'status', '--rgw-zone', 'a2', '--rgw-zonegroup', 'a', '--rgw-realm', 'test-realm', '--cluster', 'c1', '--debug-rgw', '0', '--debug-ms', '0', '--rgw-cache-enabled', 'false'] 2026-03-20T18:01:13.518 DEBUG:teuthology.orchestra.run.vm02:> adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage radosgw-admin --log-to-stderr --format json -n client.0 --cluster c1 metadata sync status --rgw-zone a2 --rgw-zonegroup a --rgw-realm test-realm --cluster c1 --debug-rgw 0 --debug-ms 0 --rgw-cache-enabled false 2026-03-20T18:01:13.539 INFO:teuthology.orchestra.run.vm02.stderr:ignoring --setuser ceph since I am not root 2026-03-20T18:01:13.539 INFO:teuthology.orchestra.run.vm02.stderr:ignoring --setgroup ceph since I am not root 2026-03-20T18:01:13.611 INFO:teuthology.orchestra.run.vm02.stderr:2026-03-20T18:01:13.609+0000 7f009eed9a80 20 RGW-SYNC:meta: read sync status 2026-03-20T18:01:13.611 INFO:teuthology.orchestra.run.vm02.stderr:2026-03-20T18:01:13.609+0000 7f009eed9a80 20 RGW-SYNC:meta: read sync status 2026-03-20T18:01:13.624 INFO:teuthology.orchestra.run.vm02.stdout:{"sync_status":{"info":{"status":"sync","num_shards":4,"period":"82c662be-66b5-4c4f-b24e-ff9298fc8a6e","realm_epoch":2},"markers":[{"key":0,"val":{"state":1,"marker":"","next_step_marker":"","total_entries":1,"pos":0,"timestamp":"0.000000","realm_epoch":2}},{"key":1,"val":{"state":1,"marker":"","next_step_marker":"","total_entries":0,"pos":0,"timestamp":"0.000000","realm_epoch":2}},{"key":2,"val":{"state":1,"marker":"","next_step_marker":"","total_entries":0,"pos":0,"timestamp":"0.000000","realm_epoch":2}},{"key":3,"val":{"state":1,"marker":"","next_step_marker":"","total_entries":0,"pos":0,"timestamp":"0.000000","realm_epoch":2}}]},"full_sync":{"total":1,"complete":1},"current_time":"2026-03-20T18:01:13Z"} 2026-03-20T18:01:13.624 WARNING:rgw_multi.tests:zone a2 behind master: shard=0 master=1_1774029627.072292_68.1 target=, shard=2 master=1_1774029627.218585_92.1 target=, shard=3 master=1_1774029627.359834_56.1 target= 2026-03-20T18:01:18.624 INFO:tasks.util.rgw:rgwadmin: c1.client.0 : ['metadata', 'sync', 'status', '--rgw-zone', 'a2', '--rgw-zonegroup', 'a', '--rgw-realm', 'test-realm', '--cluster', 'c1', '--debug-rgw', '0', '--debug-ms', '0', '--rgw-cache-enabled', 'false'] 2026-03-20T18:01:18.625 DEBUG:tasks.util.rgw:rgwadmin: cmd=['adjust-ulimits', 'ceph-coverage', '/home/ubuntu/cephtest/archive/coverage', 'radosgw-admin', '--log-to-stderr', '--format', 'json', '-n', 'client.0', '--cluster', 'c1', 'metadata', 'sync', 'status', '--rgw-zone', 'a2', '--rgw-zonegroup', 'a', '--rgw-realm', 'test-realm', '--cluster', 'c1', '--debug-rgw', '0', '--debug-ms', '0', '--rgw-cache-enabled', 'false'] 2026-03-20T18:01:18.625 DEBUG:teuthology.orchestra.run.vm02:> adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage radosgw-admin --log-to-stderr --format json -n client.0 --cluster c1 metadata sync status --rgw-zone a2 --rgw-zonegroup a --rgw-realm test-realm --cluster c1 --debug-rgw 0 --debug-ms 0 --rgw-cache-enabled false 2026-03-20T18:01:18.646 INFO:teuthology.orchestra.run.vm02.stderr:ignoring --setuser ceph since I am not root 2026-03-20T18:01:18.646 INFO:teuthology.orchestra.run.vm02.stderr:ignoring --setgroup ceph since I am not root 2026-03-20T18:01:18.718 INFO:teuthology.orchestra.run.vm02.stderr:2026-03-20T18:01:18.713+0000 7f52a1460a80 20 RGW-SYNC:meta: read sync status 2026-03-20T18:01:18.719 INFO:teuthology.orchestra.run.vm02.stderr:2026-03-20T18:01:18.717+0000 7f52a1460a80 20 RGW-SYNC:meta: read sync status 2026-03-20T18:01:18.728 INFO:teuthology.orchestra.run.vm02.stdout:{"sync_status":{"info":{"status":"sync","num_shards":4,"period":"82c662be-66b5-4c4f-b24e-ff9298fc8a6e","realm_epoch":2},"markers":[{"key":0,"val":{"state":1,"marker":"","next_step_marker":"","total_entries":1,"pos":0,"timestamp":"0.000000","realm_epoch":2}},{"key":1,"val":{"state":1,"marker":"","next_step_marker":"","total_entries":0,"pos":0,"timestamp":"0.000000","realm_epoch":2}},{"key":2,"val":{"state":1,"marker":"","next_step_marker":"","total_entries":0,"pos":0,"timestamp":"0.000000","realm_epoch":2}},{"key":3,"val":{"state":1,"marker":"","next_step_marker":"","total_entries":0,"pos":0,"timestamp":"0.000000","realm_epoch":2}}]},"full_sync":{"total":1,"complete":1},"current_time":"2026-03-20T18:01:18Z"} 2026-03-20T18:01:18.728 WARNING:rgw_multi.tests:zone a2 behind master: shard=0 master=1_1774029627.072292_68.1 target=, shard=2 master=1_1774029627.218585_92.1 target=, shard=3 master=1_1774029627.359834_56.1 target= 2026-03-20T18:01:23.729 INFO:tasks.util.rgw:rgwadmin: c1.client.0 : ['metadata', 'sync', 'status', '--rgw-zone', 'a2', '--rgw-zonegroup', 'a', '--rgw-realm', 'test-realm', '--cluster', 'c1', '--debug-rgw', '0', '--debug-ms', '0', '--rgw-cache-enabled', 'false'] 2026-03-20T18:01:23.730 DEBUG:tasks.util.rgw:rgwadmin: cmd=['adjust-ulimits', 'ceph-coverage', '/home/ubuntu/cephtest/archive/coverage', 'radosgw-admin', '--log-to-stderr', '--format', 'json', '-n', 'client.0', '--cluster', 'c1', 'metadata', 'sync', 'status', '--rgw-zone', 'a2', '--rgw-zonegroup', 'a', '--rgw-realm', 'test-realm', '--cluster', 'c1', '--debug-rgw', '0', '--debug-ms', '0', '--rgw-cache-enabled', 'false'] 2026-03-20T18:01:23.730 DEBUG:teuthology.orchestra.run.vm02:> adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage radosgw-admin --log-to-stderr --format json -n client.0 --cluster c1 metadata sync status --rgw-zone a2 --rgw-zonegroup a --rgw-realm test-realm --cluster c1 --debug-rgw 0 --debug-ms 0 --rgw-cache-enabled false 2026-03-20T18:01:23.753 INFO:teuthology.orchestra.run.vm02.stderr:ignoring --setuser ceph since I am not root 2026-03-20T18:01:23.753 INFO:teuthology.orchestra.run.vm02.stderr:ignoring --setgroup ceph since I am not root 2026-03-20T18:01:23.821 INFO:teuthology.orchestra.run.vm02.stderr:2026-03-20T18:01:23.817+0000 7f0b291d4a80 20 RGW-SYNC:meta: read sync status 2026-03-20T18:01:23.821 INFO:teuthology.orchestra.run.vm02.stderr:2026-03-20T18:01:23.817+0000 7f0b291d4a80 20 RGW-SYNC:meta: read sync status 2026-03-20T18:01:23.831 INFO:teuthology.orchestra.run.vm02.stdout:{"sync_status":{"info":{"status":"sync","num_shards":4,"period":"82c662be-66b5-4c4f-b24e-ff9298fc8a6e","realm_epoch":2},"markers":[{"key":0,"val":{"state":1,"marker":"","next_step_marker":"","total_entries":1,"pos":0,"timestamp":"0.000000","realm_epoch":2}},{"key":1,"val":{"state":1,"marker":"","next_step_marker":"","total_entries":0,"pos":0,"timestamp":"0.000000","realm_epoch":2}},{"key":2,"val":{"state":1,"marker":"","next_step_marker":"","total_entries":0,"pos":0,"timestamp":"0.000000","realm_epoch":2}},{"key":3,"val":{"state":1,"marker":"","next_step_marker":"","total_entries":0,"pos":0,"timestamp":"0.000000","realm_epoch":2}}]},"full_sync":{"total":1,"complete":1},"current_time":"2026-03-20T18:01:23Z"} 2026-03-20T18:01:23.831 WARNING:rgw_multi.tests:zone a2 behind master: shard=0 master=1_1774029627.072292_68.1 target=, shard=2 master=1_1774029627.218585_92.1 target=, shard=3 master=1_1774029627.359834_56.1 target= 2026-03-20T18:01:28.832 INFO:tasks.util.rgw:rgwadmin: c1.client.0 : ['metadata', 'sync', 'status', '--rgw-zone', 'a2', '--rgw-zonegroup', 'a', '--rgw-realm', 'test-realm', '--cluster', 'c1', '--debug-rgw', '0', '--debug-ms', '0', '--rgw-cache-enabled', 'false'] 2026-03-20T18:01:28.833 DEBUG:tasks.util.rgw:rgwadmin: cmd=['adjust-ulimits', 'ceph-coverage', '/home/ubuntu/cephtest/archive/coverage', 'radosgw-admin', '--log-to-stderr', '--format', 'json', '-n', 'client.0', '--cluster', 'c1', 'metadata', 'sync', 'status', '--rgw-zone', 'a2', '--rgw-zonegroup', 'a', '--rgw-realm', 'test-realm', '--cluster', 'c1', '--debug-rgw', '0', '--debug-ms', '0', '--rgw-cache-enabled', 'false'] 2026-03-20T18:01:28.833 DEBUG:teuthology.orchestra.run.vm02:> adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage radosgw-admin --log-to-stderr --format json -n client.0 --cluster c1 metadata sync status --rgw-zone a2 --rgw-zonegroup a --rgw-realm test-realm --cluster c1 --debug-rgw 0 --debug-ms 0 --rgw-cache-enabled false 2026-03-20T18:01:28.854 INFO:teuthology.orchestra.run.vm02.stderr:ignoring --setuser ceph since I am not root 2026-03-20T18:01:28.854 INFO:teuthology.orchestra.run.vm02.stderr:ignoring --setgroup ceph since I am not root 2026-03-20T18:01:28.930 INFO:teuthology.orchestra.run.vm02.stderr:2026-03-20T18:01:28.925+0000 7f3c37873a80 20 RGW-SYNC:meta: read sync status 2026-03-20T18:01:28.931 INFO:teuthology.orchestra.run.vm02.stderr:2026-03-20T18:01:28.929+0000 7f3c37873a80 20 RGW-SYNC:meta: read sync status 2026-03-20T18:01:28.941 INFO:teuthology.orchestra.run.vm02.stdout:{"sync_status":{"info":{"status":"sync","num_shards":4,"period":"82c662be-66b5-4c4f-b24e-ff9298fc8a6e","realm_epoch":2},"markers":[{"key":0,"val":{"state":1,"marker":"","next_step_marker":"","total_entries":1,"pos":0,"timestamp":"0.000000","realm_epoch":2}},{"key":1,"val":{"state":1,"marker":"","next_step_marker":"","total_entries":0,"pos":0,"timestamp":"0.000000","realm_epoch":2}},{"key":2,"val":{"state":1,"marker":"","next_step_marker":"","total_entries":0,"pos":0,"timestamp":"0.000000","realm_epoch":2}},{"key":3,"val":{"state":1,"marker":"","next_step_marker":"","total_entries":0,"pos":0,"timestamp":"0.000000","realm_epoch":2}}]},"full_sync":{"total":1,"complete":1},"current_time":"2026-03-20T18:01:28Z"} 2026-03-20T18:01:28.941 WARNING:rgw_multi.tests:zone a2 behind master: shard=0 master=1_1774029627.072292_68.1 target=, shard=2 master=1_1774029627.218585_92.1 target=, shard=3 master=1_1774029627.359834_56.1 target= 2026-03-20T18:01:33.942 INFO:tasks.util.rgw:rgwadmin: c1.client.0 : ['metadata', 'sync', 'status', '--rgw-zone', 'a2', '--rgw-zonegroup', 'a', '--rgw-realm', 'test-realm', '--cluster', 'c1', '--debug-rgw', '0', '--debug-ms', '0', '--rgw-cache-enabled', 'false'] 2026-03-20T18:01:33.942 DEBUG:tasks.util.rgw:rgwadmin: cmd=['adjust-ulimits', 'ceph-coverage', '/home/ubuntu/cephtest/archive/coverage', 'radosgw-admin', '--log-to-stderr', '--format', 'json', '-n', 'client.0', '--cluster', 'c1', 'metadata', 'sync', 'status', '--rgw-zone', 'a2', '--rgw-zonegroup', 'a', '--rgw-realm', 'test-realm', '--cluster', 'c1', '--debug-rgw', '0', '--debug-ms', '0', '--rgw-cache-enabled', 'false'] 2026-03-20T18:01:33.942 DEBUG:teuthology.orchestra.run.vm02:> adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage radosgw-admin --log-to-stderr --format json -n client.0 --cluster c1 metadata sync status --rgw-zone a2 --rgw-zonegroup a --rgw-realm test-realm --cluster c1 --debug-rgw 0 --debug-ms 0 --rgw-cache-enabled false 2026-03-20T18:01:33.966 INFO:teuthology.orchestra.run.vm02.stderr:ignoring --setuser ceph since I am not root 2026-03-20T18:01:33.966 INFO:teuthology.orchestra.run.vm02.stderr:ignoring --setgroup ceph since I am not root 2026-03-20T18:01:34.038 INFO:teuthology.orchestra.run.vm02.stderr:2026-03-20T18:01:34.033+0000 7f8d3fb03a80 20 RGW-SYNC:meta: read sync status 2026-03-20T18:01:34.039 INFO:teuthology.orchestra.run.vm02.stderr:2026-03-20T18:01:34.037+0000 7f8d3fb03a80 20 RGW-SYNC:meta: read sync status 2026-03-20T18:01:34.049 INFO:teuthology.orchestra.run.vm02.stdout:{"sync_status":{"info":{"status":"sync","num_shards":4,"period":"82c662be-66b5-4c4f-b24e-ff9298fc8a6e","realm_epoch":2},"markers":[{"key":0,"val":{"state":1,"marker":"","next_step_marker":"","total_entries":1,"pos":0,"timestamp":"0.000000","realm_epoch":2}},{"key":1,"val":{"state":1,"marker":"","next_step_marker":"","total_entries":0,"pos":0,"timestamp":"0.000000","realm_epoch":2}},{"key":2,"val":{"state":1,"marker":"","next_step_marker":"","total_entries":0,"pos":0,"timestamp":"0.000000","realm_epoch":2}},{"key":3,"val":{"state":1,"marker":"","next_step_marker":"","total_entries":0,"pos":0,"timestamp":"0.000000","realm_epoch":2}}]},"full_sync":{"total":1,"complete":1},"current_time":"2026-03-20T18:01:34Z"} 2026-03-20T18:01:34.049 WARNING:rgw_multi.tests:zone a2 behind master: shard=0 master=1_1774029627.072292_68.1 target=, shard=2 master=1_1774029627.218585_92.1 target=, shard=3 master=1_1774029627.359834_56.1 target= 2026-03-20T18:01:39.050 INFO:tasks.util.rgw:rgwadmin: c1.client.0 : ['metadata', 'sync', 'status', '--rgw-zone', 'a2', '--rgw-zonegroup', 'a', '--rgw-realm', 'test-realm', '--cluster', 'c1', '--debug-rgw', '0', '--debug-ms', '0', '--rgw-cache-enabled', 'false'] 2026-03-20T18:01:39.050 DEBUG:tasks.util.rgw:rgwadmin: cmd=['adjust-ulimits', 'ceph-coverage', '/home/ubuntu/cephtest/archive/coverage', 'radosgw-admin', '--log-to-stderr', '--format', 'json', '-n', 'client.0', '--cluster', 'c1', 'metadata', 'sync', 'status', '--rgw-zone', 'a2', '--rgw-zonegroup', 'a', '--rgw-realm', 'test-realm', '--cluster', 'c1', '--debug-rgw', '0', '--debug-ms', '0', '--rgw-cache-enabled', 'false'] 2026-03-20T18:01:39.050 DEBUG:teuthology.orchestra.run.vm02:> adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage radosgw-admin --log-to-stderr --format json -n client.0 --cluster c1 metadata sync status --rgw-zone a2 --rgw-zonegroup a --rgw-realm test-realm --cluster c1 --debug-rgw 0 --debug-ms 0 --rgw-cache-enabled false 2026-03-20T18:01:39.072 INFO:teuthology.orchestra.run.vm02.stderr:ignoring --setuser ceph since I am not root 2026-03-20T18:01:39.072 INFO:teuthology.orchestra.run.vm02.stderr:ignoring --setgroup ceph since I am not root 2026-03-20T18:01:39.141 INFO:teuthology.orchestra.run.vm02.stderr:2026-03-20T18:01:39.137+0000 7f369c6caa80 20 RGW-SYNC:meta: read sync status 2026-03-20T18:01:39.141 INFO:teuthology.orchestra.run.vm02.stderr:2026-03-20T18:01:39.137+0000 7f369c6caa80 20 RGW-SYNC:meta: read sync status 2026-03-20T18:01:39.152 INFO:teuthology.orchestra.run.vm02.stdout:{"sync_status":{"info":{"status":"sync","num_shards":4,"period":"82c662be-66b5-4c4f-b24e-ff9298fc8a6e","realm_epoch":2},"markers":[{"key":0,"val":{"state":1,"marker":"","next_step_marker":"","total_entries":1,"pos":0,"timestamp":"0.000000","realm_epoch":2}},{"key":1,"val":{"state":1,"marker":"","next_step_marker":"","total_entries":0,"pos":0,"timestamp":"0.000000","realm_epoch":2}},{"key":2,"val":{"state":1,"marker":"","next_step_marker":"","total_entries":0,"pos":0,"timestamp":"0.000000","realm_epoch":2}},{"key":3,"val":{"state":1,"marker":"","next_step_marker":"","total_entries":0,"pos":0,"timestamp":"0.000000","realm_epoch":2}}]},"full_sync":{"total":1,"complete":1},"current_time":"2026-03-20T18:01:39Z"} 2026-03-20T18:01:39.152 WARNING:rgw_multi.tests:zone a2 behind master: shard=0 master=1_1774029627.072292_68.1 target=, shard=2 master=1_1774029627.218585_92.1 target=, shard=3 master=1_1774029627.359834_56.1 target= 2026-03-20T18:01:44.153 INFO:tasks.util.rgw:rgwadmin: c1.client.0 : ['metadata', 'sync', 'status', '--rgw-zone', 'a2', '--rgw-zonegroup', 'a', '--rgw-realm', 'test-realm', '--cluster', 'c1', '--debug-rgw', '0', '--debug-ms', '0', '--rgw-cache-enabled', 'false'] 2026-03-20T18:01:44.153 DEBUG:tasks.util.rgw:rgwadmin: cmd=['adjust-ulimits', 'ceph-coverage', '/home/ubuntu/cephtest/archive/coverage', 'radosgw-admin', '--log-to-stderr', '--format', 'json', '-n', 'client.0', '--cluster', 'c1', 'metadata', 'sync', 'status', '--rgw-zone', 'a2', '--rgw-zonegroup', 'a', '--rgw-realm', 'test-realm', '--cluster', 'c1', '--debug-rgw', '0', '--debug-ms', '0', '--rgw-cache-enabled', 'false'] 2026-03-20T18:01:44.153 DEBUG:teuthology.orchestra.run.vm02:> adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage radosgw-admin --log-to-stderr --format json -n client.0 --cluster c1 metadata sync status --rgw-zone a2 --rgw-zonegroup a --rgw-realm test-realm --cluster c1 --debug-rgw 0 --debug-ms 0 --rgw-cache-enabled false 2026-03-20T18:01:44.177 INFO:teuthology.orchestra.run.vm02.stderr:ignoring --setuser ceph since I am not root 2026-03-20T18:01:44.177 INFO:teuthology.orchestra.run.vm02.stderr:ignoring --setgroup ceph since I am not root 2026-03-20T18:01:44.257 INFO:teuthology.orchestra.run.vm02.stderr:2026-03-20T18:01:44.253+0000 7fc5d1067a80 20 RGW-SYNC:meta: read sync status 2026-03-20T18:01:44.258 INFO:teuthology.orchestra.run.vm02.stderr:2026-03-20T18:01:44.253+0000 7fc5d1067a80 20 RGW-SYNC:meta: read sync status 2026-03-20T18:01:44.268 INFO:teuthology.orchestra.run.vm02.stdout:{"sync_status":{"info":{"status":"sync","num_shards":4,"period":"82c662be-66b5-4c4f-b24e-ff9298fc8a6e","realm_epoch":2},"markers":[{"key":0,"val":{"state":1,"marker":"","next_step_marker":"","total_entries":1,"pos":0,"timestamp":"0.000000","realm_epoch":2}},{"key":1,"val":{"state":1,"marker":"","next_step_marker":"","total_entries":0,"pos":0,"timestamp":"0.000000","realm_epoch":2}},{"key":2,"val":{"state":1,"marker":"","next_step_marker":"","total_entries":0,"pos":0,"timestamp":"0.000000","realm_epoch":2}},{"key":3,"val":{"state":1,"marker":"","next_step_marker":"","total_entries":0,"pos":0,"timestamp":"0.000000","realm_epoch":2}}]},"full_sync":{"total":1,"complete":1},"current_time":"2026-03-20T18:01:44Z"} 2026-03-20T18:01:44.269 WARNING:rgw_multi.tests:zone a2 behind master: shard=0 master=1_1774029627.072292_68.1 target=, shard=2 master=1_1774029627.218585_92.1 target=, shard=3 master=1_1774029627.359834_56.1 target= 2026-03-20T18:01:49.269 INFO:tasks.util.rgw:rgwadmin: c1.client.0 : ['metadata', 'sync', 'status', '--rgw-zone', 'a2', '--rgw-zonegroup', 'a', '--rgw-realm', 'test-realm', '--cluster', 'c1', '--debug-rgw', '0', '--debug-ms', '0', '--rgw-cache-enabled', 'false'] 2026-03-20T18:01:49.270 DEBUG:tasks.util.rgw:rgwadmin: cmd=['adjust-ulimits', 'ceph-coverage', '/home/ubuntu/cephtest/archive/coverage', 'radosgw-admin', '--log-to-stderr', '--format', 'json', '-n', 'client.0', '--cluster', 'c1', 'metadata', 'sync', 'status', '--rgw-zone', 'a2', '--rgw-zonegroup', 'a', '--rgw-realm', 'test-realm', '--cluster', 'c1', '--debug-rgw', '0', '--debug-ms', '0', '--rgw-cache-enabled', 'false'] 2026-03-20T18:01:49.270 DEBUG:teuthology.orchestra.run.vm02:> adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage radosgw-admin --log-to-stderr --format json -n client.0 --cluster c1 metadata sync status --rgw-zone a2 --rgw-zonegroup a --rgw-realm test-realm --cluster c1 --debug-rgw 0 --debug-ms 0 --rgw-cache-enabled false 2026-03-20T18:01:49.290 INFO:teuthology.orchestra.run.vm02.stderr:ignoring --setuser ceph since I am not root 2026-03-20T18:01:49.290 INFO:teuthology.orchestra.run.vm02.stderr:ignoring --setgroup ceph since I am not root 2026-03-20T18:01:49.356 INFO:teuthology.orchestra.run.vm02.stderr:2026-03-20T18:01:49.353+0000 7fa90f31aa80 20 RGW-SYNC:meta: read sync status 2026-03-20T18:01:49.357 INFO:teuthology.orchestra.run.vm02.stderr:2026-03-20T18:01:49.353+0000 7fa90f31aa80 20 RGW-SYNC:meta: read sync status 2026-03-20T18:01:49.366 INFO:teuthology.orchestra.run.vm02.stdout:{"sync_status":{"info":{"status":"sync","num_shards":4,"period":"82c662be-66b5-4c4f-b24e-ff9298fc8a6e","realm_epoch":2},"markers":[{"key":0,"val":{"state":1,"marker":"","next_step_marker":"","total_entries":1,"pos":0,"timestamp":"0.000000","realm_epoch":2}},{"key":1,"val":{"state":1,"marker":"","next_step_marker":"","total_entries":0,"pos":0,"timestamp":"0.000000","realm_epoch":2}},{"key":2,"val":{"state":1,"marker":"","next_step_marker":"","total_entries":0,"pos":0,"timestamp":"0.000000","realm_epoch":2}},{"key":3,"val":{"state":1,"marker":"","next_step_marker":"","total_entries":0,"pos":0,"timestamp":"0.000000","realm_epoch":2}}]},"full_sync":{"total":1,"complete":1},"current_time":"2026-03-20T18:01:49Z"} 2026-03-20T18:01:49.366 WARNING:rgw_multi.tests:zone a2 behind master: shard=0 master=1_1774029627.072292_68.1 target=, shard=2 master=1_1774029627.218585_92.1 target=, shard=3 master=1_1774029627.359834_56.1 target= 2026-03-20T18:01:54.367 INFO:tasks.util.rgw:rgwadmin: c1.client.0 : ['metadata', 'sync', 'status', '--rgw-zone', 'a2', '--rgw-zonegroup', 'a', '--rgw-realm', 'test-realm', '--cluster', 'c1', '--debug-rgw', '0', '--debug-ms', '0', '--rgw-cache-enabled', 'false'] 2026-03-20T18:01:54.367 DEBUG:tasks.util.rgw:rgwadmin: cmd=['adjust-ulimits', 'ceph-coverage', '/home/ubuntu/cephtest/archive/coverage', 'radosgw-admin', '--log-to-stderr', '--format', 'json', '-n', 'client.0', '--cluster', 'c1', 'metadata', 'sync', 'status', '--rgw-zone', 'a2', '--rgw-zonegroup', 'a', '--rgw-realm', 'test-realm', '--cluster', 'c1', '--debug-rgw', '0', '--debug-ms', '0', '--rgw-cache-enabled', 'false'] 2026-03-20T18:01:54.367 DEBUG:teuthology.orchestra.run.vm02:> adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage radosgw-admin --log-to-stderr --format json -n client.0 --cluster c1 metadata sync status --rgw-zone a2 --rgw-zonegroup a --rgw-realm test-realm --cluster c1 --debug-rgw 0 --debug-ms 0 --rgw-cache-enabled false 2026-03-20T18:01:54.391 INFO:teuthology.orchestra.run.vm02.stderr:ignoring --setuser ceph since I am not root 2026-03-20T18:01:54.397 INFO:teuthology.orchestra.run.vm02.stderr:ignoring --setgroup ceph since I am not root 2026-03-20T18:01:54.466 INFO:teuthology.orchestra.run.vm02.stderr:2026-03-20T18:01:54.461+0000 7f5158a8ea80 20 RGW-SYNC:meta: read sync status 2026-03-20T18:01:54.467 INFO:teuthology.orchestra.run.vm02.stderr:2026-03-20T18:01:54.465+0000 7f5158a8ea80 20 RGW-SYNC:meta: read sync status 2026-03-20T18:01:54.478 INFO:teuthology.orchestra.run.vm02.stdout:{"sync_status":{"info":{"status":"sync","num_shards":4,"period":"82c662be-66b5-4c4f-b24e-ff9298fc8a6e","realm_epoch":2},"markers":[{"key":0,"val":{"state":1,"marker":"","next_step_marker":"","total_entries":1,"pos":0,"timestamp":"0.000000","realm_epoch":2}},{"key":1,"val":{"state":1,"marker":"","next_step_marker":"","total_entries":0,"pos":0,"timestamp":"0.000000","realm_epoch":2}},{"key":2,"val":{"state":1,"marker":"","next_step_marker":"","total_entries":0,"pos":0,"timestamp":"0.000000","realm_epoch":2}},{"key":3,"val":{"state":1,"marker":"","next_step_marker":"","total_entries":0,"pos":0,"timestamp":"0.000000","realm_epoch":2}}]},"full_sync":{"total":1,"complete":1},"current_time":"2026-03-20T18:01:54Z"} 2026-03-20T18:01:54.479 WARNING:rgw_multi.tests:zone a2 behind master: shard=0 master=1_1774029627.072292_68.1 target=, shard=2 master=1_1774029627.218585_92.1 target=, shard=3 master=1_1774029627.359834_56.1 target= 2026-03-20T18:01:59.479 INFO:tasks.util.rgw:rgwadmin: c1.client.0 : ['metadata', 'sync', 'status', '--rgw-zone', 'a2', '--rgw-zonegroup', 'a', '--rgw-realm', 'test-realm', '--cluster', 'c1', '--debug-rgw', '0', '--debug-ms', '0', '--rgw-cache-enabled', 'false'] 2026-03-20T18:01:59.479 DEBUG:tasks.util.rgw:rgwadmin: cmd=['adjust-ulimits', 'ceph-coverage', '/home/ubuntu/cephtest/archive/coverage', 'radosgw-admin', '--log-to-stderr', '--format', 'json', '-n', 'client.0', '--cluster', 'c1', 'metadata', 'sync', 'status', '--rgw-zone', 'a2', '--rgw-zonegroup', 'a', '--rgw-realm', 'test-realm', '--cluster', 'c1', '--debug-rgw', '0', '--debug-ms', '0', '--rgw-cache-enabled', 'false'] 2026-03-20T18:01:59.480 DEBUG:teuthology.orchestra.run.vm02:> adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage radosgw-admin --log-to-stderr --format json -n client.0 --cluster c1 metadata sync status --rgw-zone a2 --rgw-zonegroup a --rgw-realm test-realm --cluster c1 --debug-rgw 0 --debug-ms 0 --rgw-cache-enabled false 2026-03-20T18:01:59.503 INFO:teuthology.orchestra.run.vm02.stderr:ignoring --setuser ceph since I am not root 2026-03-20T18:01:59.503 INFO:teuthology.orchestra.run.vm02.stderr:ignoring --setgroup ceph since I am not root 2026-03-20T18:01:59.586 INFO:teuthology.orchestra.run.vm02.stderr:2026-03-20T18:01:59.581+0000 7f5e13195a80 20 RGW-SYNC:meta: read sync status 2026-03-20T18:01:59.587 INFO:teuthology.orchestra.run.vm02.stderr:2026-03-20T18:01:59.585+0000 7f5e13195a80 20 RGW-SYNC:meta: read sync status 2026-03-20T18:01:59.600 INFO:teuthology.orchestra.run.vm02.stdout:{"sync_status":{"info":{"status":"sync","num_shards":4,"period":"82c662be-66b5-4c4f-b24e-ff9298fc8a6e","realm_epoch":2},"markers":[{"key":0,"val":{"state":1,"marker":"","next_step_marker":"","total_entries":1,"pos":0,"timestamp":"0.000000","realm_epoch":2}},{"key":1,"val":{"state":1,"marker":"","next_step_marker":"","total_entries":0,"pos":0,"timestamp":"0.000000","realm_epoch":2}},{"key":2,"val":{"state":1,"marker":"","next_step_marker":"","total_entries":0,"pos":0,"timestamp":"0.000000","realm_epoch":2}},{"key":3,"val":{"state":1,"marker":"","next_step_marker":"","total_entries":0,"pos":0,"timestamp":"0.000000","realm_epoch":2}}]},"full_sync":{"total":1,"complete":1},"current_time":"2026-03-20T18:01:59Z"} 2026-03-20T18:01:59.600 WARNING:rgw_multi.tests:zone a2 behind master: shard=0 master=1_1774029627.072292_68.1 target=, shard=2 master=1_1774029627.218585_92.1 target=, shard=3 master=1_1774029627.359834_56.1 target= 2026-03-20T18:02:04.601 INFO:tasks.util.rgw:rgwadmin: c1.client.0 : ['metadata', 'sync', 'status', '--rgw-zone', 'a2', '--rgw-zonegroup', 'a', '--rgw-realm', 'test-realm', '--cluster', 'c1', '--debug-rgw', '0', '--debug-ms', '0', '--rgw-cache-enabled', 'false'] 2026-03-20T18:02:04.602 DEBUG:tasks.util.rgw:rgwadmin: cmd=['adjust-ulimits', 'ceph-coverage', '/home/ubuntu/cephtest/archive/coverage', 'radosgw-admin', '--log-to-stderr', '--format', 'json', '-n', 'client.0', '--cluster', 'c1', 'metadata', 'sync', 'status', '--rgw-zone', 'a2', '--rgw-zonegroup', 'a', '--rgw-realm', 'test-realm', '--cluster', 'c1', '--debug-rgw', '0', '--debug-ms', '0', '--rgw-cache-enabled', 'false'] 2026-03-20T18:02:04.602 DEBUG:teuthology.orchestra.run.vm02:> adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage radosgw-admin --log-to-stderr --format json -n client.0 --cluster c1 metadata sync status --rgw-zone a2 --rgw-zonegroup a --rgw-realm test-realm --cluster c1 --debug-rgw 0 --debug-ms 0 --rgw-cache-enabled false 2026-03-20T18:02:04.623 INFO:teuthology.orchestra.run.vm02.stderr:ignoring --setuser ceph since I am not root 2026-03-20T18:02:04.623 INFO:teuthology.orchestra.run.vm02.stderr:ignoring --setgroup ceph since I am not root 2026-03-20T18:02:04.696 INFO:teuthology.orchestra.run.vm02.stderr:2026-03-20T18:02:04.693+0000 7fa06dc87a80 20 RGW-SYNC:meta: read sync status 2026-03-20T18:02:04.697 INFO:teuthology.orchestra.run.vm02.stderr:2026-03-20T18:02:04.693+0000 7fa06dc87a80 20 RGW-SYNC:meta: read sync status 2026-03-20T18:02:04.708 INFO:teuthology.orchestra.run.vm02.stdout:{"sync_status":{"info":{"status":"sync","num_shards":4,"period":"82c662be-66b5-4c4f-b24e-ff9298fc8a6e","realm_epoch":2},"markers":[{"key":0,"val":{"state":1,"marker":"1_1774029627.072292_68.1","next_step_marker":"","total_entries":1,"pos":0,"timestamp":"2026-03-20T18:00:27.072292Z","realm_epoch":2}},{"key":1,"val":{"state":1,"marker":"","next_step_marker":"","total_entries":0,"pos":0,"timestamp":"0.000000","realm_epoch":2}},{"key":2,"val":{"state":1,"marker":"1_1774029627.218585_92.1","next_step_marker":"","total_entries":0,"pos":0,"timestamp":"2026-03-20T18:00:27.218585Z","realm_epoch":2}},{"key":3,"val":{"state":1,"marker":"1_1774029627.359834_56.1","next_step_marker":"","total_entries":0,"pos":0,"timestamp":"2026-03-20T18:00:27.359834Z","realm_epoch":2}}]},"full_sync":{"total":1,"complete":1},"current_time":"2026-03-20T18:02:04Z"} 2026-03-20T18:02:04.709 INFO:rgw_multi.tests:finish meta checkpoint for zone=a2 2026-03-20T18:02:04.709 INFO:rgw_multi.tests:starting meta checkpoint for zone=b1 2026-03-20T18:02:04.709 INFO:tasks.util.rgw:rgwadmin: c2.client.0 : ['metadata', 'sync', 'status', '--rgw-zone', 'b1', '--rgw-zonegroup', 'b', '--rgw-realm', 'test-realm', '--cluster', 'c2', '--debug-rgw', '0', '--debug-ms', '0', '--rgw-cache-enabled', 'false'] 2026-03-20T18:02:04.709 DEBUG:tasks.util.rgw:rgwadmin: cmd=['adjust-ulimits', 'ceph-coverage', '/home/ubuntu/cephtest/archive/coverage', 'radosgw-admin', '--log-to-stderr', '--format', 'json', '-n', 'client.0', '--cluster', 'c2', 'metadata', 'sync', 'status', '--rgw-zone', 'b1', '--rgw-zonegroup', 'b', '--rgw-realm', 'test-realm', '--cluster', 'c2', '--debug-rgw', '0', '--debug-ms', '0', '--rgw-cache-enabled', 'false'] 2026-03-20T18:02:04.709 DEBUG:teuthology.orchestra.run.vm06:> adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage radosgw-admin --log-to-stderr --format json -n client.0 --cluster c2 metadata sync status --rgw-zone b1 --rgw-zonegroup b --rgw-realm test-realm --cluster c2 --debug-rgw 0 --debug-ms 0 --rgw-cache-enabled false 2026-03-20T18:02:04.731 INFO:teuthology.orchestra.run.vm06.stderr:ignoring --setuser ceph since I am not root 2026-03-20T18:02:04.731 INFO:teuthology.orchestra.run.vm06.stderr:ignoring --setgroup ceph since I am not root 2026-03-20T18:02:04.804 INFO:teuthology.orchestra.run.vm06.stderr:2026-03-20T18:02:04.795+0000 7f3280a8ea80 20 RGW-SYNC:meta: read sync status 2026-03-20T18:02:04.805 INFO:teuthology.orchestra.run.vm06.stderr:2026-03-20T18:02:04.799+0000 7f3280a8ea80 20 RGW-SYNC:meta: read sync status 2026-03-20T18:02:04.815 INFO:teuthology.orchestra.run.vm06.stdout:{"sync_status":{"info":{"status":"sync","num_shards":4,"period":"82c662be-66b5-4c4f-b24e-ff9298fc8a6e","realm_epoch":2},"markers":[{"key":0,"val":{"state":1,"marker":"1_1774029627.072292_68.1","next_step_marker":"","total_entries":1,"pos":0,"timestamp":"2026-03-20T18:00:27.072292Z","realm_epoch":2}},{"key":1,"val":{"state":1,"marker":"","next_step_marker":"","total_entries":0,"pos":0,"timestamp":"0.000000","realm_epoch":2}},{"key":2,"val":{"state":1,"marker":"1_1774029627.218585_92.1","next_step_marker":"","total_entries":0,"pos":0,"timestamp":"2026-03-20T18:00:27.218585Z","realm_epoch":2}},{"key":3,"val":{"state":1,"marker":"1_1774029627.359834_56.1","next_step_marker":"","total_entries":0,"pos":0,"timestamp":"2026-03-20T18:00:27.359834Z","realm_epoch":2}}]},"full_sync":{"total":1,"complete":1},"current_time":"2026-03-20T18:02:04Z"} 2026-03-20T18:02:04.815 INFO:rgw_multi.tests:finish meta checkpoint for zone=b1 2026-03-20T18:02:04.815 INFO:rgw_multi.tests:starting meta checkpoint for zone=b2 2026-03-20T18:02:04.815 INFO:tasks.util.rgw:rgwadmin: c2.client.0 : ['metadata', 'sync', 'status', '--rgw-zone', 'b2', '--rgw-zonegroup', 'b', '--rgw-realm', 'test-realm', '--cluster', 'c2', '--debug-rgw', '0', '--debug-ms', '0', '--rgw-cache-enabled', 'false'] 2026-03-20T18:02:04.815 DEBUG:tasks.util.rgw:rgwadmin: cmd=['adjust-ulimits', 'ceph-coverage', '/home/ubuntu/cephtest/archive/coverage', 'radosgw-admin', '--log-to-stderr', '--format', 'json', '-n', 'client.0', '--cluster', 'c2', 'metadata', 'sync', 'status', '--rgw-zone', 'b2', '--rgw-zonegroup', 'b', '--rgw-realm', 'test-realm', '--cluster', 'c2', '--debug-rgw', '0', '--debug-ms', '0', '--rgw-cache-enabled', 'false'] 2026-03-20T18:02:04.815 DEBUG:teuthology.orchestra.run.vm06:> adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage radosgw-admin --log-to-stderr --format json -n client.0 --cluster c2 metadata sync status --rgw-zone b2 --rgw-zonegroup b --rgw-realm test-realm --cluster c2 --debug-rgw 0 --debug-ms 0 --rgw-cache-enabled false 2026-03-20T18:02:04.878 INFO:teuthology.orchestra.run.vm06.stderr:ignoring --setuser ceph since I am not root 2026-03-20T18:02:04.878 INFO:teuthology.orchestra.run.vm06.stderr:ignoring --setgroup ceph since I am not root 2026-03-20T18:02:04.948 INFO:teuthology.orchestra.run.vm06.stderr:2026-03-20T18:02:04.939+0000 7efdff5f7a80 20 RGW-SYNC:meta: read sync status 2026-03-20T18:02:04.948 INFO:teuthology.orchestra.run.vm06.stderr:2026-03-20T18:02:04.943+0000 7efdff5f7a80 20 RGW-SYNC:meta: read sync status 2026-03-20T18:02:04.958 INFO:teuthology.orchestra.run.vm06.stdout:{"sync_status":{"info":{"status":"sync","num_shards":4,"period":"82c662be-66b5-4c4f-b24e-ff9298fc8a6e","realm_epoch":2},"markers":[{"key":0,"val":{"state":1,"marker":"1_1774029627.072292_68.1","next_step_marker":"","total_entries":5,"pos":0,"timestamp":"2026-03-20T18:00:27.072292Z","realm_epoch":2}},{"key":1,"val":{"state":1,"marker":"","next_step_marker":"","total_entries":0,"pos":0,"timestamp":"0.000000","realm_epoch":2}},{"key":2,"val":{"state":1,"marker":"1_1774029627.218585_92.1","next_step_marker":"","total_entries":0,"pos":0,"timestamp":"2026-03-20T18:00:27.218585Z","realm_epoch":2}},{"key":3,"val":{"state":1,"marker":"1_1774029627.359834_56.1","next_step_marker":"","total_entries":0,"pos":0,"timestamp":"2026-03-20T18:00:27.359834Z","realm_epoch":2}}]},"full_sync":{"total":5,"complete":5},"current_time":"2026-03-20T18:02:04Z"} 2026-03-20T18:02:04.959 INFO:rgw_multi.tests:finish meta checkpoint for zone=b2 2026-03-20T18:02:04.959 INFO:rgw_multi.tests:meta checkpoint 2026-03-20T18:02:04.959 INFO:tasks.util.rgw:rgwadmin: c1.client.0 : ['mdlog', 'status', '--rgw-zone', 'a1', '--rgw-zonegroup', 'a', '--rgw-realm', 'test-realm', '--cluster', 'c1', '--debug-rgw', '0', '--debug-ms', '0', '--rgw-cache-enabled', 'false'] 2026-03-20T18:02:04.959 DEBUG:tasks.util.rgw:rgwadmin: cmd=['adjust-ulimits', 'ceph-coverage', '/home/ubuntu/cephtest/archive/coverage', 'radosgw-admin', '--log-to-stderr', '--format', 'json', '-n', 'client.0', '--cluster', 'c1', 'mdlog', 'status', '--rgw-zone', 'a1', '--rgw-zonegroup', 'a', '--rgw-realm', 'test-realm', '--cluster', 'c1', '--debug-rgw', '0', '--debug-ms', '0', '--rgw-cache-enabled', 'false'] 2026-03-20T18:02:04.959 DEBUG:teuthology.orchestra.run.vm02:> adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage radosgw-admin --log-to-stderr --format json -n client.0 --cluster c1 mdlog status --rgw-zone a1 --rgw-zonegroup a --rgw-realm test-realm --cluster c1 --debug-rgw 0 --debug-ms 0 --rgw-cache-enabled false 2026-03-20T18:02:04.980 INFO:teuthology.orchestra.run.vm02.stderr:ignoring --setuser ceph since I am not root 2026-03-20T18:02:04.980 INFO:teuthology.orchestra.run.vm02.stderr:ignoring --setgroup ceph since I am not root 2026-03-20T18:02:05.053 INFO:teuthology.orchestra.run.vm02.stderr:No --period given, using current period=82c662be-66b5-4c4f-b24e-ff9298fc8a6e 2026-03-20T18:02:05.064 INFO:teuthology.orchestra.run.vm02.stdout:[{"marker":"1_1774029627.072292_68.1","last_update":"2026-03-20T18:00:27.072292Z"},{"marker":"","last_update":"0.000000"},{"marker":"1_1774029627.218585_92.1","last_update":"2026-03-20T18:00:27.218585Z"},{"marker":"1_1774029627.359834_56.1","last_update":"2026-03-20T18:00:27.359834Z"}] 2026-03-20T18:02:05.064 INFO:rgw_multi.tests:starting meta checkpoint for zone=a2 2026-03-20T18:02:05.064 INFO:tasks.util.rgw:rgwadmin: c1.client.0 : ['metadata', 'sync', 'status', '--rgw-zone', 'a2', '--rgw-zonegroup', 'a', '--rgw-realm', 'test-realm', '--cluster', 'c1', '--debug-rgw', '0', '--debug-ms', '0', '--rgw-cache-enabled', 'false'] 2026-03-20T18:02:05.064 DEBUG:tasks.util.rgw:rgwadmin: cmd=['adjust-ulimits', 'ceph-coverage', '/home/ubuntu/cephtest/archive/coverage', 'radosgw-admin', '--log-to-stderr', '--format', 'json', '-n', 'client.0', '--cluster', 'c1', 'metadata', 'sync', 'status', '--rgw-zone', 'a2', '--rgw-zonegroup', 'a', '--rgw-realm', 'test-realm', '--cluster', 'c1', '--debug-rgw', '0', '--debug-ms', '0', '--rgw-cache-enabled', 'false'] 2026-03-20T18:02:05.064 DEBUG:teuthology.orchestra.run.vm02:> adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage radosgw-admin --log-to-stderr --format json -n client.0 --cluster c1 metadata sync status --rgw-zone a2 --rgw-zonegroup a --rgw-realm test-realm --cluster c1 --debug-rgw 0 --debug-ms 0 --rgw-cache-enabled false 2026-03-20T18:02:05.127 INFO:teuthology.orchestra.run.vm02.stderr:ignoring --setuser ceph since I am not root 2026-03-20T18:02:05.127 INFO:teuthology.orchestra.run.vm02.stderr:ignoring --setgroup ceph since I am not root 2026-03-20T18:02:05.195 INFO:teuthology.orchestra.run.vm02.stderr:2026-03-20T18:02:05.193+0000 7fc53fcbba80 20 RGW-SYNC:meta: read sync status 2026-03-20T18:02:05.196 INFO:teuthology.orchestra.run.vm02.stderr:2026-03-20T18:02:05.193+0000 7fc53fcbba80 20 RGW-SYNC:meta: read sync status 2026-03-20T18:02:05.206 INFO:teuthology.orchestra.run.vm02.stdout:{"sync_status":{"info":{"status":"sync","num_shards":4,"period":"82c662be-66b5-4c4f-b24e-ff9298fc8a6e","realm_epoch":2},"markers":[{"key":0,"val":{"state":1,"marker":"1_1774029627.072292_68.1","next_step_marker":"","total_entries":1,"pos":0,"timestamp":"2026-03-20T18:00:27.072292Z","realm_epoch":2}},{"key":1,"val":{"state":1,"marker":"","next_step_marker":"","total_entries":0,"pos":0,"timestamp":"0.000000","realm_epoch":2}},{"key":2,"val":{"state":1,"marker":"1_1774029627.218585_92.1","next_step_marker":"","total_entries":0,"pos":0,"timestamp":"2026-03-20T18:00:27.218585Z","realm_epoch":2}},{"key":3,"val":{"state":1,"marker":"1_1774029627.359834_56.1","next_step_marker":"","total_entries":0,"pos":0,"timestamp":"2026-03-20T18:00:27.359834Z","realm_epoch":2}}]},"full_sync":{"total":1,"complete":1},"current_time":"2026-03-20T18:02:05Z"} 2026-03-20T18:02:05.207 INFO:rgw_multi.tests:finish meta checkpoint for zone=a2 2026-03-20T18:02:05.207 INFO:rgw_multi.tests:starting meta checkpoint for zone=b1 2026-03-20T18:02:05.207 INFO:tasks.util.rgw:rgwadmin: c2.client.0 : ['metadata', 'sync', 'status', '--rgw-zone', 'b1', '--rgw-zonegroup', 'b', '--rgw-realm', 'test-realm', '--cluster', 'c2', '--debug-rgw', '0', '--debug-ms', '0', '--rgw-cache-enabled', 'false'] 2026-03-20T18:02:05.207 DEBUG:tasks.util.rgw:rgwadmin: cmd=['adjust-ulimits', 'ceph-coverage', '/home/ubuntu/cephtest/archive/coverage', 'radosgw-admin', '--log-to-stderr', '--format', 'json', '-n', 'client.0', '--cluster', 'c2', 'metadata', 'sync', 'status', '--rgw-zone', 'b1', '--rgw-zonegroup', 'b', '--rgw-realm', 'test-realm', '--cluster', 'c2', '--debug-rgw', '0', '--debug-ms', '0', '--rgw-cache-enabled', 'false'] 2026-03-20T18:02:05.207 DEBUG:teuthology.orchestra.run.vm06:> adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage radosgw-admin --log-to-stderr --format json -n client.0 --cluster c2 metadata sync status --rgw-zone b1 --rgw-zonegroup b --rgw-realm test-realm --cluster c2 --debug-rgw 0 --debug-ms 0 --rgw-cache-enabled false 2026-03-20T18:02:05.228 INFO:teuthology.orchestra.run.vm06.stderr:ignoring --setuser ceph since I am not root 2026-03-20T18:02:05.228 INFO:teuthology.orchestra.run.vm06.stderr:ignoring --setgroup ceph since I am not root 2026-03-20T18:02:05.297 INFO:teuthology.orchestra.run.vm06.stderr:2026-03-20T18:02:05.291+0000 7fdde783aa80 20 RGW-SYNC:meta: read sync status 2026-03-20T18:02:05.298 INFO:teuthology.orchestra.run.vm06.stderr:2026-03-20T18:02:05.291+0000 7fdde783aa80 20 RGW-SYNC:meta: read sync status 2026-03-20T18:02:05.308 INFO:teuthology.orchestra.run.vm06.stdout:{"sync_status":{"info":{"status":"sync","num_shards":4,"period":"82c662be-66b5-4c4f-b24e-ff9298fc8a6e","realm_epoch":2},"markers":[{"key":0,"val":{"state":1,"marker":"1_1774029627.072292_68.1","next_step_marker":"","total_entries":1,"pos":0,"timestamp":"2026-03-20T18:00:27.072292Z","realm_epoch":2}},{"key":1,"val":{"state":1,"marker":"","next_step_marker":"","total_entries":0,"pos":0,"timestamp":"0.000000","realm_epoch":2}},{"key":2,"val":{"state":1,"marker":"1_1774029627.218585_92.1","next_step_marker":"","total_entries":0,"pos":0,"timestamp":"2026-03-20T18:00:27.218585Z","realm_epoch":2}},{"key":3,"val":{"state":1,"marker":"1_1774029627.359834_56.1","next_step_marker":"","total_entries":0,"pos":0,"timestamp":"2026-03-20T18:00:27.359834Z","realm_epoch":2}}]},"full_sync":{"total":1,"complete":1},"current_time":"2026-03-20T18:02:05Z"} 2026-03-20T18:02:05.308 INFO:rgw_multi.tests:finish meta checkpoint for zone=b1 2026-03-20T18:02:05.308 INFO:rgw_multi.tests:starting meta checkpoint for zone=b2 2026-03-20T18:02:05.308 INFO:tasks.util.rgw:rgwadmin: c2.client.0 : ['metadata', 'sync', 'status', '--rgw-zone', 'b2', '--rgw-zonegroup', 'b', '--rgw-realm', 'test-realm', '--cluster', 'c2', '--debug-rgw', '0', '--debug-ms', '0', '--rgw-cache-enabled', 'false'] 2026-03-20T18:02:05.308 DEBUG:tasks.util.rgw:rgwadmin: cmd=['adjust-ulimits', 'ceph-coverage', '/home/ubuntu/cephtest/archive/coverage', 'radosgw-admin', '--log-to-stderr', '--format', 'json', '-n', 'client.0', '--cluster', 'c2', 'metadata', 'sync', 'status', '--rgw-zone', 'b2', '--rgw-zonegroup', 'b', '--rgw-realm', 'test-realm', '--cluster', 'c2', '--debug-rgw', '0', '--debug-ms', '0', '--rgw-cache-enabled', 'false'] 2026-03-20T18:02:05.308 DEBUG:teuthology.orchestra.run.vm06:> adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage radosgw-admin --log-to-stderr --format json -n client.0 --cluster c2 metadata sync status --rgw-zone b2 --rgw-zonegroup b --rgw-realm test-realm --cluster c2 --debug-rgw 0 --debug-ms 0 --rgw-cache-enabled false 2026-03-20T18:02:05.368 INFO:teuthology.orchestra.run.vm06.stderr:ignoring --setuser ceph since I am not root 2026-03-20T18:02:05.368 INFO:teuthology.orchestra.run.vm06.stderr:ignoring --setgroup ceph since I am not root 2026-03-20T18:02:05.434 INFO:teuthology.orchestra.run.vm06.stderr:2026-03-20T18:02:05.427+0000 7fcf9ac20a80 20 RGW-SYNC:meta: read sync status 2026-03-20T18:02:05.435 INFO:teuthology.orchestra.run.vm06.stderr:2026-03-20T18:02:05.427+0000 7fcf9ac20a80 20 RGW-SYNC:meta: read sync status 2026-03-20T18:02:05.444 INFO:teuthology.orchestra.run.vm06.stdout:{"sync_status":{"info":{"status":"sync","num_shards":4,"period":"82c662be-66b5-4c4f-b24e-ff9298fc8a6e","realm_epoch":2},"markers":[{"key":0,"val":{"state":1,"marker":"1_1774029627.072292_68.1","next_step_marker":"","total_entries":5,"pos":0,"timestamp":"2026-03-20T18:00:27.072292Z","realm_epoch":2}},{"key":1,"val":{"state":1,"marker":"","next_step_marker":"","total_entries":0,"pos":0,"timestamp":"0.000000","realm_epoch":2}},{"key":2,"val":{"state":1,"marker":"1_1774029627.218585_92.1","next_step_marker":"","total_entries":0,"pos":0,"timestamp":"2026-03-20T18:00:27.218585Z","realm_epoch":2}},{"key":3,"val":{"state":1,"marker":"1_1774029627.359834_56.1","next_step_marker":"","total_entries":0,"pos":0,"timestamp":"2026-03-20T18:00:27.359834Z","realm_epoch":2}}]},"full_sync":{"total":5,"complete":5},"current_time":"2026-03-20T18:02:05Z"} 2026-03-20T18:02:05.445 INFO:rgw_multi.tests:finish meta checkpoint for zone=b2 2026-03-20T18:02:05.445 INFO:tasks.rgw_multisite_tests:running rgw multisite tests on '/home/teuthos/src/github.com_kshtsk_ceph_c04fae7bd18063e6f8f9e1cf40c23f810c1f6ca4/qa/../src/test/rgw/rgw_multi' with args=['tests.py', '-a', '!fails_with_rgw'] 2026-03-20T18:02:05.733 INFO:tasks.util.rgw:rgwadmin: c1.client.0 : ['mdlog', 'status', '--rgw-zone', 'a1', '--rgw-zonegroup', 'a', '--rgw-realm', 'test-realm', '--cluster', 'c1', '--debug-rgw', '0', '--debug-ms', '0', '--rgw-cache-enabled', 'false'] 2026-03-20T18:02:05.733 DEBUG:tasks.util.rgw:rgwadmin: cmd=['adjust-ulimits', 'ceph-coverage', '/home/ubuntu/cephtest/archive/coverage', 'radosgw-admin', '--log-to-stderr', '--format', 'json', '-n', 'client.0', '--cluster', 'c1', 'mdlog', 'status', '--rgw-zone', 'a1', '--rgw-zonegroup', 'a', '--rgw-realm', 'test-realm', '--cluster', 'c1', '--debug-rgw', '0', '--debug-ms', '0', '--rgw-cache-enabled', 'false'] 2026-03-20T18:02:05.733 DEBUG:teuthology.orchestra.run.vm02:> adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage radosgw-admin --log-to-stderr --format json -n client.0 --cluster c1 mdlog status --rgw-zone a1 --rgw-zonegroup a --rgw-realm test-realm --cluster c1 --debug-rgw 0 --debug-ms 0 --rgw-cache-enabled false 2026-03-20T18:02:05.756 INFO:teuthology.orchestra.run.vm02.stderr:ignoring --setuser ceph since I am not root 2026-03-20T18:02:05.756 INFO:teuthology.orchestra.run.vm02.stderr:ignoring --setgroup ceph since I am not root 2026-03-20T18:02:05.822 INFO:teuthology.orchestra.run.vm02.stderr:No --period given, using current period=82c662be-66b5-4c4f-b24e-ff9298fc8a6e 2026-03-20T18:02:05.832 INFO:teuthology.orchestra.run.vm02.stdout:[{"marker":"1_1774029627.072292_68.1","last_update":"2026-03-20T18:00:27.072292Z"},{"marker":"","last_update":"0.000000"},{"marker":"1_1774029627.218585_92.1","last_update":"2026-03-20T18:00:27.218585Z"},{"marker":"1_1774029627.359834_56.1","last_update":"2026-03-20T18:00:27.359834Z"}] 2026-03-20T18:02:05.833 INFO:rgw_multi.tests:starting meta checkpoint for zone=a2 2026-03-20T18:02:05.833 INFO:tasks.util.rgw:rgwadmin: c1.client.0 : ['metadata', 'sync', 'status', '--rgw-zone', 'a2', '--rgw-zonegroup', 'a', '--rgw-realm', 'test-realm', '--cluster', 'c1', '--debug-rgw', '0', '--debug-ms', '0', '--rgw-cache-enabled', 'false'] 2026-03-20T18:02:05.833 DEBUG:tasks.util.rgw:rgwadmin: cmd=['adjust-ulimits', 'ceph-coverage', '/home/ubuntu/cephtest/archive/coverage', 'radosgw-admin', '--log-to-stderr', '--format', 'json', '-n', 'client.0', '--cluster', 'c1', 'metadata', 'sync', 'status', '--rgw-zone', 'a2', '--rgw-zonegroup', 'a', '--rgw-realm', 'test-realm', '--cluster', 'c1', '--debug-rgw', '0', '--debug-ms', '0', '--rgw-cache-enabled', 'false'] 2026-03-20T18:02:05.833 DEBUG:teuthology.orchestra.run.vm02:> adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage radosgw-admin --log-to-stderr --format json -n client.0 --cluster c1 metadata sync status --rgw-zone a2 --rgw-zonegroup a --rgw-realm test-realm --cluster c1 --debug-rgw 0 --debug-ms 0 --rgw-cache-enabled false 2026-03-20T18:02:05.895 INFO:teuthology.orchestra.run.vm02.stderr:ignoring --setuser ceph since I am not root 2026-03-20T18:02:05.895 INFO:teuthology.orchestra.run.vm02.stderr:ignoring --setgroup ceph since I am not root 2026-03-20T18:02:05.971 INFO:teuthology.orchestra.run.vm02.stderr:2026-03-20T18:02:05.965+0000 7f0e65439a80 20 RGW-SYNC:meta: read sync status 2026-03-20T18:02:05.971 INFO:teuthology.orchestra.run.vm02.stderr:2026-03-20T18:02:05.969+0000 7f0e65439a80 20 RGW-SYNC:meta: read sync status 2026-03-20T18:02:05.982 INFO:teuthology.orchestra.run.vm02.stdout:{"sync_status":{"info":{"status":"sync","num_shards":4,"period":"82c662be-66b5-4c4f-b24e-ff9298fc8a6e","realm_epoch":2},"markers":[{"key":0,"val":{"state":1,"marker":"1_1774029627.072292_68.1","next_step_marker":"","total_entries":1,"pos":0,"timestamp":"2026-03-20T18:00:27.072292Z","realm_epoch":2}},{"key":1,"val":{"state":1,"marker":"","next_step_marker":"","total_entries":0,"pos":0,"timestamp":"0.000000","realm_epoch":2}},{"key":2,"val":{"state":1,"marker":"1_1774029627.218585_92.1","next_step_marker":"","total_entries":0,"pos":0,"timestamp":"2026-03-20T18:00:27.218585Z","realm_epoch":2}},{"key":3,"val":{"state":1,"marker":"1_1774029627.359834_56.1","next_step_marker":"","total_entries":0,"pos":0,"timestamp":"2026-03-20T18:00:27.359834Z","realm_epoch":2}}]},"full_sync":{"total":1,"complete":1},"current_time":"2026-03-20T18:02:05Z"} 2026-03-20T18:02:05.982 INFO:rgw_multi.tests:finish meta checkpoint for zone=a2 2026-03-20T18:02:05.982 INFO:rgw_multi.tests:create bucket zone=a1 name=cyjkqa-1 2026-03-20T18:02:08.573 INFO:rgw_multi.tests:create bucket zone=a2 name=cyjkqa-2 2026-03-20T18:02:11.587 INFO:tasks.util.rgw:rgwadmin: c1.client.0 : ['mdlog', 'status', '--rgw-zone', 'a1', '--rgw-zonegroup', 'a', '--rgw-realm', 'test-realm', '--cluster', 'c1', '--debug-rgw', '0', '--debug-ms', '0', '--rgw-cache-enabled', 'false'] 2026-03-20T18:02:11.587 DEBUG:tasks.util.rgw:rgwadmin: cmd=['adjust-ulimits', 'ceph-coverage', '/home/ubuntu/cephtest/archive/coverage', 'radosgw-admin', '--log-to-stderr', '--format', 'json', '-n', 'client.0', '--cluster', 'c1', 'mdlog', 'status', '--rgw-zone', 'a1', '--rgw-zonegroup', 'a', '--rgw-realm', 'test-realm', '--cluster', 'c1', '--debug-rgw', '0', '--debug-ms', '0', '--rgw-cache-enabled', 'false'] 2026-03-20T18:02:11.587 DEBUG:teuthology.orchestra.run.vm02:> adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage radosgw-admin --log-to-stderr --format json -n client.0 --cluster c1 mdlog status --rgw-zone a1 --rgw-zonegroup a --rgw-realm test-realm --cluster c1 --debug-rgw 0 --debug-ms 0 --rgw-cache-enabled false 2026-03-20T18:02:11.608 INFO:teuthology.orchestra.run.vm02.stderr:ignoring --setuser ceph since I am not root 2026-03-20T18:02:11.608 INFO:teuthology.orchestra.run.vm02.stderr:ignoring --setgroup ceph since I am not root 2026-03-20T18:02:11.680 INFO:teuthology.orchestra.run.vm02.stderr:No --period given, using current period=82c662be-66b5-4c4f-b24e-ff9298fc8a6e 2026-03-20T18:02:11.695 INFO:teuthology.orchestra.run.vm02.stdout:[{"marker":"1_1774029728.581833_109.1","last_update":"2026-03-20T18:02:08.581833Z"},{"marker":"","last_update":"0.000000"},{"marker":"1_1774029728.584377_143.1","last_update":"2026-03-20T18:02:08.584377Z"},{"marker":"1_1774029627.359834_56.1","last_update":"2026-03-20T18:00:27.359834Z"}] 2026-03-20T18:02:11.695 INFO:rgw_multi.tests:starting meta checkpoint for zone=a2 2026-03-20T18:02:11.695 INFO:tasks.util.rgw:rgwadmin: c1.client.0 : ['metadata', 'sync', 'status', '--rgw-zone', 'a2', '--rgw-zonegroup', 'a', '--rgw-realm', 'test-realm', '--cluster', 'c1', '--debug-rgw', '0', '--debug-ms', '0', '--rgw-cache-enabled', 'false'] 2026-03-20T18:02:11.695 DEBUG:tasks.util.rgw:rgwadmin: cmd=['adjust-ulimits', 'ceph-coverage', '/home/ubuntu/cephtest/archive/coverage', 'radosgw-admin', '--log-to-stderr', '--format', 'json', '-n', 'client.0', '--cluster', 'c1', 'metadata', 'sync', 'status', '--rgw-zone', 'a2', '--rgw-zonegroup', 'a', '--rgw-realm', 'test-realm', '--cluster', 'c1', '--debug-rgw', '0', '--debug-ms', '0', '--rgw-cache-enabled', 'false'] 2026-03-20T18:02:11.695 DEBUG:teuthology.orchestra.run.vm02:> adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage radosgw-admin --log-to-stderr --format json -n client.0 --cluster c1 metadata sync status --rgw-zone a2 --rgw-zonegroup a --rgw-realm test-realm --cluster c1 --debug-rgw 0 --debug-ms 0 --rgw-cache-enabled false 2026-03-20T18:02:11.761 INFO:teuthology.orchestra.run.vm02.stderr:ignoring --setuser ceph since I am not root 2026-03-20T18:02:11.761 INFO:teuthology.orchestra.run.vm02.stderr:ignoring --setgroup ceph since I am not root 2026-03-20T18:02:11.832 INFO:teuthology.orchestra.run.vm02.stderr:2026-03-20T18:02:11.829+0000 7f3bad619a80 20 RGW-SYNC:meta: read sync status 2026-03-20T18:02:11.833 INFO:teuthology.orchestra.run.vm02.stderr:2026-03-20T18:02:11.829+0000 7f3bad619a80 20 RGW-SYNC:meta: read sync status 2026-03-20T18:02:11.843 INFO:teuthology.orchestra.run.vm02.stdout:{"sync_status":{"info":{"status":"sync","num_shards":4,"period":"82c662be-66b5-4c4f-b24e-ff9298fc8a6e","realm_epoch":2},"markers":[{"key":0,"val":{"state":1,"marker":"1_1774029728.581833_109.1","next_step_marker":"","total_entries":1,"pos":0,"timestamp":"2026-03-20T18:02:08.581833Z","realm_epoch":2}},{"key":1,"val":{"state":1,"marker":"","next_step_marker":"","total_entries":0,"pos":0,"timestamp":"0.000000","realm_epoch":2}},{"key":2,"val":{"state":1,"marker":"1_1774029728.584377_143.1","next_step_marker":"","total_entries":0,"pos":0,"timestamp":"2026-03-20T18:02:08.584377Z","realm_epoch":2}},{"key":3,"val":{"state":1,"marker":"1_1774029627.359834_56.1","next_step_marker":"","total_entries":0,"pos":0,"timestamp":"2026-03-20T18:00:27.359834Z","realm_epoch":2}}]},"full_sync":{"total":1,"complete":1},"current_time":"2026-03-20T18:02:11Z"} 2026-03-20T18:02:11.843 INFO:rgw_multi.tests:finish meta checkpoint for zone=a2 2026-03-20T18:02:11.855 INFO:tasks.rgw_multisite_tests:rgw_multi.tests.test_bucket_create ... ok 2026-03-20T18:02:11.966 INFO:tasks.util.rgw:rgwadmin: c1.client.0 : ['user', 'create', '--tenant', 'testx', '--uid', 'test', '--access-key', 'abcd', '--secret-key', 'efgh', '--display-name', 'tenanted-user', '--cluster', 'c1', '--debug-rgw', '0', '--debug-ms', '0'] 2026-03-20T18:02:11.966 DEBUG:tasks.util.rgw:rgwadmin: cmd=['adjust-ulimits', 'ceph-coverage', '/home/ubuntu/cephtest/archive/coverage', 'radosgw-admin', '--log-to-stderr', '--format', 'json', '-n', 'client.0', '--cluster', 'c1', 'user', 'create', '--tenant', 'testx', '--uid', 'test', '--access-key', 'abcd', '--secret-key', 'efgh', '--display-name', 'tenanted-user', '--cluster', 'c1', '--debug-rgw', '0', '--debug-ms', '0'] 2026-03-20T18:02:11.966 DEBUG:teuthology.orchestra.run.vm02:> adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage radosgw-admin --log-to-stderr --format json -n client.0 --cluster c1 user create --tenant testx --uid test --access-key abcd --secret-key efgh --display-name tenanted-user --cluster c1 --debug-rgw 0 --debug-ms 0 2026-03-20T18:02:11.991 INFO:teuthology.orchestra.run.vm02.stderr:ignoring --setuser ceph since I am not root 2026-03-20T18:02:11.991 INFO:teuthology.orchestra.run.vm02.stderr:ignoring --setgroup ceph since I am not root 2026-03-20T18:02:12.068 INFO:teuthology.orchestra.run.vm02.stdout:{"user_id":"testx$test","display_name":"tenanted-user","email":"","suspended":0,"max_buckets":1000,"subusers":[],"keys":[{"user":"testx$test","access_key":"abcd","secret_key":"efgh","active":true,"create_date":"2026-03-20T18:02:12.062858Z"}],"swift_keys":[],"caps":[],"op_mask":"read, write, delete","default_placement":"","default_storage_class":"","placement_tags":[],"bucket_quota":{"enabled":false,"check_on_raw":false,"max_size":-1,"max_size_kb":0,"max_objects":-1},"user_quota":{"enabled":false,"check_on_raw":false,"max_size":-1,"max_size_kb":0,"max_objects":-1},"temp_url_keys":[],"type":"rgw","mfa_ids":[],"account_id":"","path":"/","create_date":"2026-03-20T18:02:12.062849Z","tags":[],"group_ids":[]} 2026-03-20T18:02:12.077 INFO:tasks.util.rgw:rgwadmin: c1.client.0 : ['mdlog', 'status', '--rgw-zone', 'a1', '--rgw-zonegroup', 'a', '--rgw-realm', 'test-realm', '--cluster', 'c1', '--debug-rgw', '0', '--debug-ms', '0', '--rgw-cache-enabled', 'false'] 2026-03-20T18:02:12.077 DEBUG:tasks.util.rgw:rgwadmin: cmd=['adjust-ulimits', 'ceph-coverage', '/home/ubuntu/cephtest/archive/coverage', 'radosgw-admin', '--log-to-stderr', '--format', 'json', '-n', 'client.0', '--cluster', 'c1', 'mdlog', 'status', '--rgw-zone', 'a1', '--rgw-zonegroup', 'a', '--rgw-realm', 'test-realm', '--cluster', 'c1', '--debug-rgw', '0', '--debug-ms', '0', '--rgw-cache-enabled', 'false'] 2026-03-20T18:02:12.077 DEBUG:teuthology.orchestra.run.vm02:> adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage radosgw-admin --log-to-stderr --format json -n client.0 --cluster c1 mdlog status --rgw-zone a1 --rgw-zonegroup a --rgw-realm test-realm --cluster c1 --debug-rgw 0 --debug-ms 0 --rgw-cache-enabled false 2026-03-20T18:02:12.138 INFO:teuthology.orchestra.run.vm02.stderr:ignoring --setuser ceph since I am not root 2026-03-20T18:02:12.139 INFO:teuthology.orchestra.run.vm02.stderr:ignoring --setgroup ceph since I am not root 2026-03-20T18:02:12.213 INFO:teuthology.orchestra.run.vm02.stderr:No --period given, using current period=82c662be-66b5-4c4f-b24e-ff9298fc8a6e 2026-03-20T18:02:12.222 INFO:teuthology.orchestra.run.vm02.stdout:[{"marker":"1_1774029728.581833_109.1","last_update":"2026-03-20T18:02:08.581833Z"},{"marker":"1_1774029732.064755_60.1","last_update":"2026-03-20T18:02:12.064755Z"},{"marker":"1_1774029728.584377_143.1","last_update":"2026-03-20T18:02:08.584377Z"},{"marker":"1_1774029627.359834_56.1","last_update":"2026-03-20T18:00:27.359834Z"}] 2026-03-20T18:02:12.222 INFO:rgw_multi.tests:starting meta checkpoint for zone=a2 2026-03-20T18:02:12.222 INFO:tasks.util.rgw:rgwadmin: c1.client.0 : ['metadata', 'sync', 'status', '--rgw-zone', 'a2', '--rgw-zonegroup', 'a', '--rgw-realm', 'test-realm', '--cluster', 'c1', '--debug-rgw', '0', '--debug-ms', '0', '--rgw-cache-enabled', 'false'] 2026-03-20T18:02:12.222 DEBUG:tasks.util.rgw:rgwadmin: cmd=['adjust-ulimits', 'ceph-coverage', '/home/ubuntu/cephtest/archive/coverage', 'radosgw-admin', '--log-to-stderr', '--format', 'json', '-n', 'client.0', '--cluster', 'c1', 'metadata', 'sync', 'status', '--rgw-zone', 'a2', '--rgw-zonegroup', 'a', '--rgw-realm', 'test-realm', '--cluster', 'c1', '--debug-rgw', '0', '--debug-ms', '0', '--rgw-cache-enabled', 'false'] 2026-03-20T18:02:12.223 DEBUG:teuthology.orchestra.run.vm02:> adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage radosgw-admin --log-to-stderr --format json -n client.0 --cluster c1 metadata sync status --rgw-zone a2 --rgw-zonegroup a --rgw-realm test-realm --cluster c1 --debug-rgw 0 --debug-ms 0 --rgw-cache-enabled false 2026-03-20T18:02:12.287 INFO:teuthology.orchestra.run.vm02.stderr:ignoring --setuser ceph since I am not root 2026-03-20T18:02:12.287 INFO:teuthology.orchestra.run.vm02.stderr:ignoring --setgroup ceph since I am not root 2026-03-20T18:02:12.360 INFO:teuthology.orchestra.run.vm02.stderr:2026-03-20T18:02:12.357+0000 7f136cebfa80 20 RGW-SYNC:meta: read sync status 2026-03-20T18:02:12.360 INFO:teuthology.orchestra.run.vm02.stderr:2026-03-20T18:02:12.357+0000 7f136cebfa80 20 RGW-SYNC:meta: read sync status 2026-03-20T18:02:12.370 INFO:teuthology.orchestra.run.vm02.stdout:{"sync_status":{"info":{"status":"sync","num_shards":4,"period":"82c662be-66b5-4c4f-b24e-ff9298fc8a6e","realm_epoch":2},"markers":[{"key":0,"val":{"state":1,"marker":"1_1774029728.581833_109.1","next_step_marker":"","total_entries":1,"pos":0,"timestamp":"2026-03-20T18:02:08.581833Z","realm_epoch":2}},{"key":1,"val":{"state":1,"marker":"","next_step_marker":"","total_entries":0,"pos":0,"timestamp":"0.000000","realm_epoch":2}},{"key":2,"val":{"state":1,"marker":"1_1774029728.584377_143.1","next_step_marker":"","total_entries":0,"pos":0,"timestamp":"2026-03-20T18:02:08.584377Z","realm_epoch":2}},{"key":3,"val":{"state":1,"marker":"1_1774029627.359834_56.1","next_step_marker":"","total_entries":0,"pos":0,"timestamp":"2026-03-20T18:00:27.359834Z","realm_epoch":2}}]},"full_sync":{"total":1,"complete":1},"current_time":"2026-03-20T18:02:12Z"} 2026-03-20T18:02:12.370 WARNING:rgw_multi.tests:zone a2 behind master: shard=1 master=1_1774029732.064755_60.1 target= 2026-03-20T18:02:17.371 INFO:tasks.util.rgw:rgwadmin: c1.client.0 : ['metadata', 'sync', 'status', '--rgw-zone', 'a2', '--rgw-zonegroup', 'a', '--rgw-realm', 'test-realm', '--cluster', 'c1', '--debug-rgw', '0', '--debug-ms', '0', '--rgw-cache-enabled', 'false'] 2026-03-20T18:02:17.371 DEBUG:tasks.util.rgw:rgwadmin: cmd=['adjust-ulimits', 'ceph-coverage', '/home/ubuntu/cephtest/archive/coverage', 'radosgw-admin', '--log-to-stderr', '--format', 'json', '-n', 'client.0', '--cluster', 'c1', 'metadata', 'sync', 'status', '--rgw-zone', 'a2', '--rgw-zonegroup', 'a', '--rgw-realm', 'test-realm', '--cluster', 'c1', '--debug-rgw', '0', '--debug-ms', '0', '--rgw-cache-enabled', 'false'] 2026-03-20T18:02:17.371 DEBUG:teuthology.orchestra.run.vm02:> adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage radosgw-admin --log-to-stderr --format json -n client.0 --cluster c1 metadata sync status --rgw-zone a2 --rgw-zonegroup a --rgw-realm test-realm --cluster c1 --debug-rgw 0 --debug-ms 0 --rgw-cache-enabled false 2026-03-20T18:02:17.392 INFO:teuthology.orchestra.run.vm02.stderr:ignoring --setuser ceph since I am not root 2026-03-20T18:02:17.393 INFO:teuthology.orchestra.run.vm02.stderr:ignoring --setgroup ceph since I am not root 2026-03-20T18:02:17.464 INFO:teuthology.orchestra.run.vm02.stderr:2026-03-20T18:02:17.461+0000 7f0e399ada80 20 RGW-SYNC:meta: read sync status 2026-03-20T18:02:17.465 INFO:teuthology.orchestra.run.vm02.stderr:2026-03-20T18:02:17.461+0000 7f0e399ada80 20 RGW-SYNC:meta: read sync status 2026-03-20T18:02:17.481 INFO:teuthology.orchestra.run.vm02.stdout:{"sync_status":{"info":{"status":"sync","num_shards":4,"period":"82c662be-66b5-4c4f-b24e-ff9298fc8a6e","realm_epoch":2},"markers":[{"key":0,"val":{"state":1,"marker":"1_1774029728.581833_109.1","next_step_marker":"","total_entries":1,"pos":0,"timestamp":"2026-03-20T18:02:08.581833Z","realm_epoch":2}},{"key":1,"val":{"state":1,"marker":"1_1774029732.064755_60.1","next_step_marker":"","total_entries":0,"pos":0,"timestamp":"2026-03-20T18:02:12.064755Z","realm_epoch":2}},{"key":2,"val":{"state":1,"marker":"1_1774029728.584377_143.1","next_step_marker":"","total_entries":0,"pos":0,"timestamp":"2026-03-20T18:02:08.584377Z","realm_epoch":2}},{"key":3,"val":{"state":1,"marker":"1_1774029627.359834_56.1","next_step_marker":"","total_entries":0,"pos":0,"timestamp":"2026-03-20T18:00:27.359834Z","realm_epoch":2}}]},"full_sync":{"total":1,"complete":1},"current_time":"2026-03-20T18:02:17Z"} 2026-03-20T18:02:17.481 INFO:rgw_multi.tests:finish meta checkpoint for zone=a2 2026-03-20T18:02:17.505 INFO:tasks.util.rgw:rgwadmin: c1.client.0 : ['mdlog', 'status', '--rgw-zone', 'a1', '--rgw-zonegroup', 'a', '--rgw-realm', 'test-realm', '--cluster', 'c1', '--debug-rgw', '0', '--debug-ms', '0', '--rgw-cache-enabled', 'false'] 2026-03-20T18:02:17.505 DEBUG:tasks.util.rgw:rgwadmin: cmd=['adjust-ulimits', 'ceph-coverage', '/home/ubuntu/cephtest/archive/coverage', 'radosgw-admin', '--log-to-stderr', '--format', 'json', '-n', 'client.0', '--cluster', 'c1', 'mdlog', 'status', '--rgw-zone', 'a1', '--rgw-zonegroup', 'a', '--rgw-realm', 'test-realm', '--cluster', 'c1', '--debug-rgw', '0', '--debug-ms', '0', '--rgw-cache-enabled', 'false'] 2026-03-20T18:02:17.505 DEBUG:teuthology.orchestra.run.vm02:> adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage radosgw-admin --log-to-stderr --format json -n client.0 --cluster c1 mdlog status --rgw-zone a1 --rgw-zonegroup a --rgw-realm test-realm --cluster c1 --debug-rgw 0 --debug-ms 0 --rgw-cache-enabled false 2026-03-20T18:02:17.545 INFO:teuthology.orchestra.run.vm02.stderr:ignoring --setuser ceph since I am not root 2026-03-20T18:02:17.546 INFO:teuthology.orchestra.run.vm02.stderr:ignoring --setgroup ceph since I am not root 2026-03-20T18:02:17.626 INFO:teuthology.orchestra.run.vm02.stderr:No --period given, using current period=82c662be-66b5-4c4f-b24e-ff9298fc8a6e 2026-03-20T18:02:17.638 INFO:teuthology.orchestra.run.vm02.stdout:[{"marker":"1_1774029728.581833_109.1","last_update":"2026-03-20T18:02:08.581833Z"},{"marker":"1_1774029737.490667_61.1","last_update":"2026-03-20T18:02:17.490667Z"},{"marker":"1_1774029728.584377_143.1","last_update":"2026-03-20T18:02:08.584377Z"},{"marker":"1_1774029737.494130_80.1","last_update":"2026-03-20T18:02:17.494130Z"}] 2026-03-20T18:02:17.638 INFO:rgw_multi.tests:starting meta checkpoint for zone=a2 2026-03-20T18:02:17.638 INFO:tasks.util.rgw:rgwadmin: c1.client.0 : ['metadata', 'sync', 'status', '--rgw-zone', 'a2', '--rgw-zonegroup', 'a', '--rgw-realm', 'test-realm', '--cluster', 'c1', '--debug-rgw', '0', '--debug-ms', '0', '--rgw-cache-enabled', 'false'] 2026-03-20T18:02:17.638 DEBUG:tasks.util.rgw:rgwadmin: cmd=['adjust-ulimits', 'ceph-coverage', '/home/ubuntu/cephtest/archive/coverage', 'radosgw-admin', '--log-to-stderr', '--format', 'json', '-n', 'client.0', '--cluster', 'c1', 'metadata', 'sync', 'status', '--rgw-zone', 'a2', '--rgw-zonegroup', 'a', '--rgw-realm', 'test-realm', '--cluster', 'c1', '--debug-rgw', '0', '--debug-ms', '0', '--rgw-cache-enabled', 'false'] 2026-03-20T18:02:17.639 DEBUG:teuthology.orchestra.run.vm02:> adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage radosgw-admin --log-to-stderr --format json -n client.0 --cluster c1 metadata sync status --rgw-zone a2 --rgw-zonegroup a --rgw-realm test-realm --cluster c1 --debug-rgw 0 --debug-ms 0 --rgw-cache-enabled false 2026-03-20T18:02:17.699 INFO:teuthology.orchestra.run.vm02.stderr:ignoring --setuser ceph since I am not root 2026-03-20T18:02:17.699 INFO:teuthology.orchestra.run.vm02.stderr:ignoring --setgroup ceph since I am not root 2026-03-20T18:02:17.775 INFO:teuthology.orchestra.run.vm02.stderr:2026-03-20T18:02:17.773+0000 7f3836447a80 20 RGW-SYNC:meta: read sync status 2026-03-20T18:02:17.776 INFO:teuthology.orchestra.run.vm02.stderr:2026-03-20T18:02:17.773+0000 7f3836447a80 20 RGW-SYNC:meta: read sync status 2026-03-20T18:02:17.785 INFO:teuthology.orchestra.run.vm02.stdout:{"sync_status":{"info":{"status":"sync","num_shards":4,"period":"82c662be-66b5-4c4f-b24e-ff9298fc8a6e","realm_epoch":2},"markers":[{"key":0,"val":{"state":1,"marker":"1_1774029728.581833_109.1","next_step_marker":"","total_entries":1,"pos":0,"timestamp":"2026-03-20T18:02:08.581833Z","realm_epoch":2}},{"key":1,"val":{"state":1,"marker":"1_1774029737.490667_61.1","next_step_marker":"","total_entries":0,"pos":0,"timestamp":"2026-03-20T18:02:17.490667Z","realm_epoch":2}},{"key":2,"val":{"state":1,"marker":"1_1774029728.584377_143.1","next_step_marker":"","total_entries":0,"pos":0,"timestamp":"2026-03-20T18:02:08.584377Z","realm_epoch":2}},{"key":3,"val":{"state":1,"marker":"1_1774029737.494130_80.1","next_step_marker":"","total_entries":0,"pos":0,"timestamp":"2026-03-20T18:02:17.494130Z","realm_epoch":2}}]},"full_sync":{"total":1,"complete":1},"current_time":"2026-03-20T18:02:17Z"} 2026-03-20T18:02:17.786 INFO:rgw_multi.tests:finish meta checkpoint for zone=a2 2026-03-20T18:02:17.788 INFO:rgw_multi.tests:bucket exists in tenant namespace 2026-03-20T18:02:17.791 INFO:rgw_multi.tests:bucket does not exist in default user namespace 2026-03-20T18:02:17.791 INFO:tasks.util.rgw:rgwadmin: c1.client.0 : ['user', 'rm', '--tenant', 'testx', '--uid', 'test', '--purge-data', '--cluster', 'c1', '--debug-rgw', '0', '--debug-ms', '0'] 2026-03-20T18:02:17.791 DEBUG:tasks.util.rgw:rgwadmin: cmd=['adjust-ulimits', 'ceph-coverage', '/home/ubuntu/cephtest/archive/coverage', 'radosgw-admin', '--log-to-stderr', '--format', 'json', '-n', 'client.0', '--cluster', 'c1', 'user', 'rm', '--tenant', 'testx', '--uid', 'test', '--purge-data', '--cluster', 'c1', '--debug-rgw', '0', '--debug-ms', '0'] 2026-03-20T18:02:17.791 DEBUG:teuthology.orchestra.run.vm02:> adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage radosgw-admin --log-to-stderr --format json -n client.0 --cluster c1 user rm --tenant testx --uid test --purge-data --cluster c1 --debug-rgw 0 --debug-ms 0 2026-03-20T18:02:17.846 INFO:teuthology.orchestra.run.vm02.stderr:ignoring --setuser ceph since I am not root 2026-03-20T18:02:17.846 INFO:teuthology.orchestra.run.vm02.stderr:ignoring --setgroup ceph since I am not root 2026-03-20T18:02:18.008 INFO:tasks.rgw_multisite_tests:create a bucket from secondary zone under tenant namespace. check if it successfully syncs ... ok 2026-03-20T18:02:18.122 INFO:rgw_multi.tests:create bucket zone=a1 name=cyjkqa-3 2026-03-20T18:02:18.138 INFO:rgw_multi.tests:create bucket zone=a2 name=cyjkqa-4 2026-03-20T18:02:18.162 INFO:tasks.util.rgw:rgwadmin: c1.client.0 : ['mdlog', 'status', '--rgw-zone', 'a1', '--rgw-zonegroup', 'a', '--rgw-realm', 'test-realm', '--cluster', 'c1', '--debug-rgw', '0', '--debug-ms', '0', '--rgw-cache-enabled', 'false'] 2026-03-20T18:02:18.162 DEBUG:tasks.util.rgw:rgwadmin: cmd=['adjust-ulimits', 'ceph-coverage', '/home/ubuntu/cephtest/archive/coverage', 'radosgw-admin', '--log-to-stderr', '--format', 'json', '-n', 'client.0', '--cluster', 'c1', 'mdlog', 'status', '--rgw-zone', 'a1', '--rgw-zonegroup', 'a', '--rgw-realm', 'test-realm', '--cluster', 'c1', '--debug-rgw', '0', '--debug-ms', '0', '--rgw-cache-enabled', 'false'] 2026-03-20T18:02:18.162 DEBUG:teuthology.orchestra.run.vm02:> adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage radosgw-admin --log-to-stderr --format json -n client.0 --cluster c1 mdlog status --rgw-zone a1 --rgw-zonegroup a --rgw-realm test-realm --cluster c1 --debug-rgw 0 --debug-ms 0 --rgw-cache-enabled false 2026-03-20T18:02:18.183 INFO:teuthology.orchestra.run.vm02.stderr:ignoring --setuser ceph since I am not root 2026-03-20T18:02:18.183 INFO:teuthology.orchestra.run.vm02.stderr:ignoring --setgroup ceph since I am not root 2026-03-20T18:02:18.253 INFO:teuthology.orchestra.run.vm02.stderr:No --period given, using current period=82c662be-66b5-4c4f-b24e-ff9298fc8a6e 2026-03-20T18:02:18.263 INFO:teuthology.orchestra.run.vm02.stdout:[{"marker":"1_1774029738.147105_124.1","last_update":"2026-03-20T18:02:18.147105Z"},{"marker":"1_1774029737.992101_64.1","last_update":"2026-03-20T18:02:17.992101Z"},{"marker":"1_1774029738.149722_162.1","last_update":"2026-03-20T18:02:18.149722Z"},{"marker":"1_1774029737.945479_86.1","last_update":"2026-03-20T18:02:17.945479Z"}] 2026-03-20T18:02:18.264 INFO:rgw_multi.tests:starting meta checkpoint for zone=a2 2026-03-20T18:02:18.264 INFO:tasks.util.rgw:rgwadmin: c1.client.0 : ['metadata', 'sync', 'status', '--rgw-zone', 'a2', '--rgw-zonegroup', 'a', '--rgw-realm', 'test-realm', '--cluster', 'c1', '--debug-rgw', '0', '--debug-ms', '0', '--rgw-cache-enabled', 'false'] 2026-03-20T18:02:18.264 DEBUG:tasks.util.rgw:rgwadmin: cmd=['adjust-ulimits', 'ceph-coverage', '/home/ubuntu/cephtest/archive/coverage', 'radosgw-admin', '--log-to-stderr', '--format', 'json', '-n', 'client.0', '--cluster', 'c1', 'metadata', 'sync', 'status', '--rgw-zone', 'a2', '--rgw-zonegroup', 'a', '--rgw-realm', 'test-realm', '--cluster', 'c1', '--debug-rgw', '0', '--debug-ms', '0', '--rgw-cache-enabled', 'false'] 2026-03-20T18:02:18.264 DEBUG:teuthology.orchestra.run.vm02:> adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage radosgw-admin --log-to-stderr --format json -n client.0 --cluster c1 metadata sync status --rgw-zone a2 --rgw-zonegroup a --rgw-realm test-realm --cluster c1 --debug-rgw 0 --debug-ms 0 --rgw-cache-enabled false 2026-03-20T18:02:18.328 INFO:teuthology.orchestra.run.vm02.stderr:ignoring --setuser ceph since I am not root 2026-03-20T18:02:18.328 INFO:teuthology.orchestra.run.vm02.stderr:ignoring --setgroup ceph since I am not root 2026-03-20T18:02:18.402 INFO:teuthology.orchestra.run.vm02.stderr:2026-03-20T18:02:18.397+0000 7f25c251ca80 20 RGW-SYNC:meta: read sync status 2026-03-20T18:02:18.402 INFO:teuthology.orchestra.run.vm02.stderr:2026-03-20T18:02:18.397+0000 7f25c251ca80 20 RGW-SYNC:meta: read sync status 2026-03-20T18:02:18.412 INFO:teuthology.orchestra.run.vm02.stdout:{"sync_status":{"info":{"status":"sync","num_shards":4,"period":"82c662be-66b5-4c4f-b24e-ff9298fc8a6e","realm_epoch":2},"markers":[{"key":0,"val":{"state":1,"marker":"1_1774029738.147105_124.1","next_step_marker":"","total_entries":1,"pos":0,"timestamp":"2026-03-20T18:02:18.147105Z","realm_epoch":2}},{"key":1,"val":{"state":1,"marker":"1_1774029737.490667_61.1","next_step_marker":"","total_entries":0,"pos":0,"timestamp":"2026-03-20T18:02:17.490667Z","realm_epoch":2}},{"key":2,"val":{"state":1,"marker":"1_1774029738.149722_162.1","next_step_marker":"","total_entries":0,"pos":0,"timestamp":"2026-03-20T18:02:18.149722Z","realm_epoch":2}},{"key":3,"val":{"state":1,"marker":"1_1774029737.494130_80.1","next_step_marker":"","total_entries":0,"pos":0,"timestamp":"2026-03-20T18:02:17.494130Z","realm_epoch":2}}]},"full_sync":{"total":1,"complete":1},"current_time":"2026-03-20T18:02:18Z"} 2026-03-20T18:02:18.413 WARNING:rgw_multi.tests:zone a2 behind master: shard=1 master=1_1774029737.992101_64.1 target=1_1774029737.490667_61.1, shard=3 master=1_1774029737.945479_86.1 target=1_1774029737.494130_80.1 2026-03-20T18:02:23.413 INFO:tasks.util.rgw:rgwadmin: c1.client.0 : ['metadata', 'sync', 'status', '--rgw-zone', 'a2', '--rgw-zonegroup', 'a', '--rgw-realm', 'test-realm', '--cluster', 'c1', '--debug-rgw', '0', '--debug-ms', '0', '--rgw-cache-enabled', 'false'] 2026-03-20T18:02:23.414 DEBUG:tasks.util.rgw:rgwadmin: cmd=['adjust-ulimits', 'ceph-coverage', '/home/ubuntu/cephtest/archive/coverage', 'radosgw-admin', '--log-to-stderr', '--format', 'json', '-n', 'client.0', '--cluster', 'c1', 'metadata', 'sync', 'status', '--rgw-zone', 'a2', '--rgw-zonegroup', 'a', '--rgw-realm', 'test-realm', '--cluster', 'c1', '--debug-rgw', '0', '--debug-ms', '0', '--rgw-cache-enabled', 'false'] 2026-03-20T18:02:23.414 DEBUG:teuthology.orchestra.run.vm02:> adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage radosgw-admin --log-to-stderr --format json -n client.0 --cluster c1 metadata sync status --rgw-zone a2 --rgw-zonegroup a --rgw-realm test-realm --cluster c1 --debug-rgw 0 --debug-ms 0 --rgw-cache-enabled false 2026-03-20T18:02:23.436 INFO:teuthology.orchestra.run.vm02.stderr:ignoring --setuser ceph since I am not root 2026-03-20T18:02:23.436 INFO:teuthology.orchestra.run.vm02.stderr:ignoring --setgroup ceph since I am not root 2026-03-20T18:02:23.501 INFO:teuthology.orchestra.run.vm02.stderr:2026-03-20T18:02:23.497+0000 7f1ccb59fa80 20 RGW-SYNC:meta: read sync status 2026-03-20T18:02:23.502 INFO:teuthology.orchestra.run.vm02.stderr:2026-03-20T18:02:23.497+0000 7f1ccb59fa80 20 RGW-SYNC:meta: read sync status 2026-03-20T18:02:23.512 INFO:teuthology.orchestra.run.vm02.stdout:{"sync_status":{"info":{"status":"sync","num_shards":4,"period":"82c662be-66b5-4c4f-b24e-ff9298fc8a6e","realm_epoch":2},"markers":[{"key":0,"val":{"state":1,"marker":"1_1774029738.147105_124.1","next_step_marker":"","total_entries":1,"pos":0,"timestamp":"2026-03-20T18:02:18.147105Z","realm_epoch":2}},{"key":1,"val":{"state":1,"marker":"1_1774029737.992101_64.1","next_step_marker":"","total_entries":0,"pos":0,"timestamp":"2026-03-20T18:02:17.992101Z","realm_epoch":2}},{"key":2,"val":{"state":1,"marker":"1_1774029738.149722_162.1","next_step_marker":"","total_entries":0,"pos":0,"timestamp":"2026-03-20T18:02:18.149722Z","realm_epoch":2}},{"key":3,"val":{"state":1,"marker":"1_1774029737.945479_86.1","next_step_marker":"","total_entries":0,"pos":0,"timestamp":"2026-03-20T18:02:17.945479Z","realm_epoch":2}}]},"full_sync":{"total":1,"complete":1},"current_time":"2026-03-20T18:02:23Z"} 2026-03-20T18:02:23.512 INFO:rgw_multi.tests:finish meta checkpoint for zone=a2 2026-03-20T18:02:23.635 INFO:tasks.util.rgw:rgwadmin: c1.client.0 : ['mdlog', 'status', '--rgw-zone', 'a1', '--rgw-zonegroup', 'a', '--rgw-realm', 'test-realm', '--cluster', 'c1', '--debug-rgw', '0', '--debug-ms', '0', '--rgw-cache-enabled', 'false'] 2026-03-20T18:02:23.635 DEBUG:tasks.util.rgw:rgwadmin: cmd=['adjust-ulimits', 'ceph-coverage', '/home/ubuntu/cephtest/archive/coverage', 'radosgw-admin', '--log-to-stderr', '--format', 'json', '-n', 'client.0', '--cluster', 'c1', 'mdlog', 'status', '--rgw-zone', 'a1', '--rgw-zonegroup', 'a', '--rgw-realm', 'test-realm', '--cluster', 'c1', '--debug-rgw', '0', '--debug-ms', '0', '--rgw-cache-enabled', 'false'] 2026-03-20T18:02:23.636 DEBUG:teuthology.orchestra.run.vm02:> adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage radosgw-admin --log-to-stderr --format json -n client.0 --cluster c1 mdlog status --rgw-zone a1 --rgw-zonegroup a --rgw-realm test-realm --cluster c1 --debug-rgw 0 --debug-ms 0 --rgw-cache-enabled false 2026-03-20T18:02:23.655 INFO:teuthology.orchestra.run.vm02.stderr:ignoring --setuser ceph since I am not root 2026-03-20T18:02:23.655 INFO:teuthology.orchestra.run.vm02.stderr:ignoring --setgroup ceph since I am not root 2026-03-20T18:02:23.719 INFO:teuthology.orchestra.run.vm02.stderr:No --period given, using current period=82c662be-66b5-4c4f-b24e-ff9298fc8a6e 2026-03-20T18:02:23.729 INFO:teuthology.orchestra.run.vm02.stdout:[{"marker":"1_1774029743.576954_130.1","last_update":"2026-03-20T18:02:23.576954Z"},{"marker":"1_1774029737.992101_64.1","last_update":"2026-03-20T18:02:17.992101Z"},{"marker":"1_1774029738.149722_162.1","last_update":"2026-03-20T18:02:18.149722Z"},{"marker":"1_1774029737.945479_86.1","last_update":"2026-03-20T18:02:17.945479Z"}] 2026-03-20T18:02:23.730 INFO:rgw_multi.tests:starting meta checkpoint for zone=a2 2026-03-20T18:02:23.730 INFO:tasks.util.rgw:rgwadmin: c1.client.0 : ['metadata', 'sync', 'status', '--rgw-zone', 'a2', '--rgw-zonegroup', 'a', '--rgw-realm', 'test-realm', '--cluster', 'c1', '--debug-rgw', '0', '--debug-ms', '0', '--rgw-cache-enabled', 'false'] 2026-03-20T18:02:23.730 DEBUG:tasks.util.rgw:rgwadmin: cmd=['adjust-ulimits', 'ceph-coverage', '/home/ubuntu/cephtest/archive/coverage', 'radosgw-admin', '--log-to-stderr', '--format', 'json', '-n', 'client.0', '--cluster', 'c1', 'metadata', 'sync', 'status', '--rgw-zone', 'a2', '--rgw-zonegroup', 'a', '--rgw-realm', 'test-realm', '--cluster', 'c1', '--debug-rgw', '0', '--debug-ms', '0', '--rgw-cache-enabled', 'false'] 2026-03-20T18:02:23.730 DEBUG:teuthology.orchestra.run.vm02:> adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage radosgw-admin --log-to-stderr --format json -n client.0 --cluster c1 metadata sync status --rgw-zone a2 --rgw-zonegroup a --rgw-realm test-realm --cluster c1 --debug-rgw 0 --debug-ms 0 --rgw-cache-enabled false 2026-03-20T18:02:23.790 INFO:teuthology.orchestra.run.vm02.stderr:ignoring --setuser ceph since I am not root 2026-03-20T18:02:23.790 INFO:teuthology.orchestra.run.vm02.stderr:ignoring --setgroup ceph since I am not root 2026-03-20T18:02:23.856 INFO:teuthology.orchestra.run.vm02.stderr:2026-03-20T18:02:23.853+0000 7fac217fba80 20 RGW-SYNC:meta: read sync status 2026-03-20T18:02:23.856 INFO:teuthology.orchestra.run.vm02.stderr:2026-03-20T18:02:23.853+0000 7fac217fba80 20 RGW-SYNC:meta: read sync status 2026-03-20T18:02:23.867 INFO:teuthology.orchestra.run.vm02.stdout:{"sync_status":{"info":{"status":"sync","num_shards":4,"period":"82c662be-66b5-4c4f-b24e-ff9298fc8a6e","realm_epoch":2},"markers":[{"key":0,"val":{"state":1,"marker":"1_1774029743.576954_130.1","next_step_marker":"","total_entries":1,"pos":0,"timestamp":"2026-03-20T18:02:23.576954Z","realm_epoch":2}},{"key":1,"val":{"state":1,"marker":"1_1774029737.992101_64.1","next_step_marker":"","total_entries":0,"pos":0,"timestamp":"2026-03-20T18:02:17.992101Z","realm_epoch":2}},{"key":2,"val":{"state":1,"marker":"1_1774029738.149722_162.1","next_step_marker":"","total_entries":0,"pos":0,"timestamp":"2026-03-20T18:02:18.149722Z","realm_epoch":2}},{"key":3,"val":{"state":1,"marker":"1_1774029737.945479_86.1","next_step_marker":"","total_entries":0,"pos":0,"timestamp":"2026-03-20T18:02:17.945479Z","realm_epoch":2}}]},"full_sync":{"total":1,"complete":1},"current_time":"2026-03-20T18:02:23Z"} 2026-03-20T18:02:23.867 INFO:rgw_multi.tests:finish meta checkpoint for zone=a2 2026-03-20T18:02:23.875 INFO:tasks.rgw_multisite_tests:rgw_multi.tests.test_bucket_recreate ... ok 2026-03-20T18:02:23.922 INFO:rgw_multi.tests:create bucket zone=a1 name=cyjkqa-5 2026-03-20T18:02:23.935 INFO:rgw_multi.tests:create bucket zone=a2 name=cyjkqa-6 2026-03-20T18:02:23.956 INFO:tasks.util.rgw:rgwadmin: c1.client.0 : ['mdlog', 'status', '--rgw-zone', 'a1', '--rgw-zonegroup', 'a', '--rgw-realm', 'test-realm', '--cluster', 'c1', '--debug-rgw', '0', '--debug-ms', '0', '--rgw-cache-enabled', 'false'] 2026-03-20T18:02:23.956 DEBUG:tasks.util.rgw:rgwadmin: cmd=['adjust-ulimits', 'ceph-coverage', '/home/ubuntu/cephtest/archive/coverage', 'radosgw-admin', '--log-to-stderr', '--format', 'json', '-n', 'client.0', '--cluster', 'c1', 'mdlog', 'status', '--rgw-zone', 'a1', '--rgw-zonegroup', 'a', '--rgw-realm', 'test-realm', '--cluster', 'c1', '--debug-rgw', '0', '--debug-ms', '0', '--rgw-cache-enabled', 'false'] 2026-03-20T18:02:23.956 DEBUG:teuthology.orchestra.run.vm02:> adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage radosgw-admin --log-to-stderr --format json -n client.0 --cluster c1 mdlog status --rgw-zone a1 --rgw-zonegroup a --rgw-realm test-realm --cluster c1 --debug-rgw 0 --debug-ms 0 --rgw-cache-enabled false 2026-03-20T18:02:23.983 INFO:teuthology.orchestra.run.vm02.stderr:ignoring --setuser ceph since I am not root 2026-03-20T18:02:23.983 INFO:teuthology.orchestra.run.vm02.stderr:ignoring --setgroup ceph since I am not root 2026-03-20T18:02:24.053 INFO:teuthology.orchestra.run.vm02.stderr:No --period given, using current period=82c662be-66b5-4c4f-b24e-ff9298fc8a6e 2026-03-20T18:02:24.062 INFO:teuthology.orchestra.run.vm02.stdout:[{"marker":"1_1774029743.576954_130.1","last_update":"2026-03-20T18:02:23.576954Z"},{"marker":"1_1774029743.943239_66.1","last_update":"2026-03-20T18:02:23.943239Z"},{"marker":"1_1774029743.946226_170.1","last_update":"2026-03-20T18:02:23.946226Z"},{"marker":"1_1774029737.945479_86.1","last_update":"2026-03-20T18:02:17.945479Z"}] 2026-03-20T18:02:24.062 INFO:rgw_multi.tests:starting meta checkpoint for zone=a2 2026-03-20T18:02:24.062 INFO:tasks.util.rgw:rgwadmin: c1.client.0 : ['metadata', 'sync', 'status', '--rgw-zone', 'a2', '--rgw-zonegroup', 'a', '--rgw-realm', 'test-realm', '--cluster', 'c1', '--debug-rgw', '0', '--debug-ms', '0', '--rgw-cache-enabled', 'false'] 2026-03-20T18:02:24.062 DEBUG:tasks.util.rgw:rgwadmin: cmd=['adjust-ulimits', 'ceph-coverage', '/home/ubuntu/cephtest/archive/coverage', 'radosgw-admin', '--log-to-stderr', '--format', 'json', '-n', 'client.0', '--cluster', 'c1', 'metadata', 'sync', 'status', '--rgw-zone', 'a2', '--rgw-zonegroup', 'a', '--rgw-realm', 'test-realm', '--cluster', 'c1', '--debug-rgw', '0', '--debug-ms', '0', '--rgw-cache-enabled', 'false'] 2026-03-20T18:02:24.063 DEBUG:teuthology.orchestra.run.vm02:> adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage radosgw-admin --log-to-stderr --format json -n client.0 --cluster c1 metadata sync status --rgw-zone a2 --rgw-zonegroup a --rgw-realm test-realm --cluster c1 --debug-rgw 0 --debug-ms 0 --rgw-cache-enabled false 2026-03-20T18:02:24.126 INFO:teuthology.orchestra.run.vm02.stderr:ignoring --setuser ceph since I am not root 2026-03-20T18:02:24.126 INFO:teuthology.orchestra.run.vm02.stderr:ignoring --setgroup ceph since I am not root 2026-03-20T18:02:24.189 INFO:teuthology.orchestra.run.vm02.stderr:2026-03-20T18:02:24.185+0000 7fab850efa80 20 RGW-SYNC:meta: read sync status 2026-03-20T18:02:24.190 INFO:teuthology.orchestra.run.vm02.stderr:2026-03-20T18:02:24.185+0000 7fab850efa80 20 RGW-SYNC:meta: read sync status 2026-03-20T18:02:24.198 INFO:teuthology.orchestra.run.vm02.stdout:{"sync_status":{"info":{"status":"sync","num_shards":4,"period":"82c662be-66b5-4c4f-b24e-ff9298fc8a6e","realm_epoch":2},"markers":[{"key":0,"val":{"state":1,"marker":"1_1774029743.576954_130.1","next_step_marker":"","total_entries":1,"pos":0,"timestamp":"2026-03-20T18:02:23.576954Z","realm_epoch":2}},{"key":1,"val":{"state":1,"marker":"1_1774029743.943239_66.1","next_step_marker":"","total_entries":0,"pos":0,"timestamp":"2026-03-20T18:02:23.943239Z","realm_epoch":2}},{"key":2,"val":{"state":1,"marker":"1_1774029743.946226_170.1","next_step_marker":"","total_entries":0,"pos":0,"timestamp":"2026-03-20T18:02:23.946226Z","realm_epoch":2}},{"key":3,"val":{"state":1,"marker":"1_1774029737.945479_86.1","next_step_marker":"","total_entries":0,"pos":0,"timestamp":"2026-03-20T18:02:17.945479Z","realm_epoch":2}}]},"full_sync":{"total":1,"complete":1},"current_time":"2026-03-20T18:02:24Z"} 2026-03-20T18:02:24.198 INFO:rgw_multi.tests:finish meta checkpoint for zone=a2 2026-03-20T18:02:24.309 INFO:tasks.util.rgw:rgwadmin: c1.client.0 : ['mdlog', 'status', '--rgw-zone', 'a1', '--rgw-zonegroup', 'a', '--rgw-realm', 'test-realm', '--cluster', 'c1', '--debug-rgw', '0', '--debug-ms', '0', '--rgw-cache-enabled', 'false'] 2026-03-20T18:02:24.309 DEBUG:tasks.util.rgw:rgwadmin: cmd=['adjust-ulimits', 'ceph-coverage', '/home/ubuntu/cephtest/archive/coverage', 'radosgw-admin', '--log-to-stderr', '--format', 'json', '-n', 'client.0', '--cluster', 'c1', 'mdlog', 'status', '--rgw-zone', 'a1', '--rgw-zonegroup', 'a', '--rgw-realm', 'test-realm', '--cluster', 'c1', '--debug-rgw', '0', '--debug-ms', '0', '--rgw-cache-enabled', 'false'] 2026-03-20T18:02:24.309 DEBUG:teuthology.orchestra.run.vm02:> adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage radosgw-admin --log-to-stderr --format json -n client.0 --cluster c1 mdlog status --rgw-zone a1 --rgw-zonegroup a --rgw-realm test-realm --cluster c1 --debug-rgw 0 --debug-ms 0 --rgw-cache-enabled false 2026-03-20T18:02:24.329 INFO:teuthology.orchestra.run.vm02.stderr:ignoring --setuser ceph since I am not root 2026-03-20T18:02:24.330 INFO:teuthology.orchestra.run.vm02.stderr:ignoring --setgroup ceph since I am not root 2026-03-20T18:02:24.399 INFO:teuthology.orchestra.run.vm02.stderr:No --period given, using current period=82c662be-66b5-4c4f-b24e-ff9298fc8a6e 2026-03-20T18:02:24.411 INFO:teuthology.orchestra.run.vm02.stdout:[{"marker":"1_1774029743.576954_130.1","last_update":"2026-03-20T18:02:23.576954Z"},{"marker":"1_1774029744.265318_68.1","last_update":"2026-03-20T18:02:24.265318Z"},{"marker":"1_1774029744.262235_175.1","last_update":"2026-03-20T18:02:24.262235Z"},{"marker":"1_1774029737.945479_86.1","last_update":"2026-03-20T18:02:17.945479Z"}] 2026-03-20T18:02:24.411 INFO:rgw_multi.tests:starting meta checkpoint for zone=a2 2026-03-20T18:02:24.411 INFO:tasks.util.rgw:rgwadmin: c1.client.0 : ['metadata', 'sync', 'status', '--rgw-zone', 'a2', '--rgw-zonegroup', 'a', '--rgw-realm', 'test-realm', '--cluster', 'c1', '--debug-rgw', '0', '--debug-ms', '0', '--rgw-cache-enabled', 'false'] 2026-03-20T18:02:24.411 DEBUG:tasks.util.rgw:rgwadmin: cmd=['adjust-ulimits', 'ceph-coverage', '/home/ubuntu/cephtest/archive/coverage', 'radosgw-admin', '--log-to-stderr', '--format', 'json', '-n', 'client.0', '--cluster', 'c1', 'metadata', 'sync', 'status', '--rgw-zone', 'a2', '--rgw-zonegroup', 'a', '--rgw-realm', 'test-realm', '--cluster', 'c1', '--debug-rgw', '0', '--debug-ms', '0', '--rgw-cache-enabled', 'false'] 2026-03-20T18:02:24.411 DEBUG:teuthology.orchestra.run.vm02:> adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage radosgw-admin --log-to-stderr --format json -n client.0 --cluster c1 metadata sync status --rgw-zone a2 --rgw-zonegroup a --rgw-realm test-realm --cluster c1 --debug-rgw 0 --debug-ms 0 --rgw-cache-enabled false 2026-03-20T18:02:24.475 INFO:teuthology.orchestra.run.vm02.stderr:ignoring --setuser ceph since I am not root 2026-03-20T18:02:24.475 INFO:teuthology.orchestra.run.vm02.stderr:ignoring --setgroup ceph since I am not root 2026-03-20T18:02:24.541 INFO:teuthology.orchestra.run.vm02.stderr:2026-03-20T18:02:24.537+0000 7f210beeaa80 20 RGW-SYNC:meta: read sync status 2026-03-20T18:02:24.541 INFO:teuthology.orchestra.run.vm02.stderr:2026-03-20T18:02:24.537+0000 7f210beeaa80 20 RGW-SYNC:meta: read sync status 2026-03-20T18:02:24.550 INFO:teuthology.orchestra.run.vm02.stdout:{"sync_status":{"info":{"status":"sync","num_shards":4,"period":"82c662be-66b5-4c4f-b24e-ff9298fc8a6e","realm_epoch":2},"markers":[{"key":0,"val":{"state":1,"marker":"1_1774029743.576954_130.1","next_step_marker":"","total_entries":1,"pos":0,"timestamp":"2026-03-20T18:02:23.576954Z","realm_epoch":2}},{"key":1,"val":{"state":1,"marker":"1_1774029744.265318_68.1","next_step_marker":"","total_entries":0,"pos":0,"timestamp":"2026-03-20T18:02:24.265318Z","realm_epoch":2}},{"key":2,"val":{"state":1,"marker":"1_1774029744.262235_175.1","next_step_marker":"","total_entries":0,"pos":0,"timestamp":"2026-03-20T18:02:24.262235Z","realm_epoch":2}},{"key":3,"val":{"state":1,"marker":"1_1774029737.945479_86.1","next_step_marker":"","total_entries":0,"pos":0,"timestamp":"2026-03-20T18:02:17.945479Z","realm_epoch":2}}]},"full_sync":{"total":1,"complete":1},"current_time":"2026-03-20T18:02:24Z"} 2026-03-20T18:02:24.550 INFO:rgw_multi.tests:finish meta checkpoint for zone=a2 2026-03-20T18:02:24.550 INFO:rgw_multi.tests:Checking zone=a1 for deleted buckets 2026-03-20T18:02:24.552 CRITICAL:rgw_multi.tests:zone contains bucket cyjkqa-5 2026-03-20T18:02:24.552 ERROR:rgw_multi.tests:Zone a1 still has buckets: ['cyjkqa-5', 'cyjkqa-6'] 2026-03-20T18:02:24.553 INFO:tasks.rgw_multisite_tests:rgw_multi.tests.test_bucket_remove ... FAIL 2026-03-20T18:02:24.664 INFO:rgw_multi.tests:create bucket zone=a1 name=cyjkqa-7 2026-03-20T18:02:24.680 INFO:rgw_multi.tests:create bucket zone=a2 name=cyjkqa-8 2026-03-20T18:02:24.808 INFO:tasks.util.rgw:rgwadmin: c1.client.0 : ['mdlog', 'status', '--rgw-zone', 'a1', '--rgw-zonegroup', 'a', '--rgw-realm', 'test-realm', '--cluster', 'c1', '--debug-rgw', '0', '--debug-ms', '0', '--rgw-cache-enabled', 'false'] 2026-03-20T18:02:24.808 DEBUG:tasks.util.rgw:rgwadmin: cmd=['adjust-ulimits', 'ceph-coverage', '/home/ubuntu/cephtest/archive/coverage', 'radosgw-admin', '--log-to-stderr', '--format', 'json', '-n', 'client.0', '--cluster', 'c1', 'mdlog', 'status', '--rgw-zone', 'a1', '--rgw-zonegroup', 'a', '--rgw-realm', 'test-realm', '--cluster', 'c1', '--debug-rgw', '0', '--debug-ms', '0', '--rgw-cache-enabled', 'false'] 2026-03-20T18:02:24.808 DEBUG:teuthology.orchestra.run.vm02:> adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage radosgw-admin --log-to-stderr --format json -n client.0 --cluster c1 mdlog status --rgw-zone a1 --rgw-zonegroup a --rgw-realm test-realm --cluster c1 --debug-rgw 0 --debug-ms 0 --rgw-cache-enabled false 2026-03-20T18:02:24.828 INFO:teuthology.orchestra.run.vm02.stderr:ignoring --setuser ceph since I am not root 2026-03-20T18:02:24.828 INFO:teuthology.orchestra.run.vm02.stderr:ignoring --setgroup ceph since I am not root 2026-03-20T18:02:24.890 INFO:teuthology.orchestra.run.vm02.stderr:No --period given, using current period=82c662be-66b5-4c4f-b24e-ff9298fc8a6e 2026-03-20T18:02:24.899 INFO:teuthology.orchestra.run.vm02.stdout:[{"marker":"1_1774029743.576954_130.1","last_update":"2026-03-20T18:02:23.576954Z"},{"marker":"1_1774029744.688217_70.1","last_update":"2026-03-20T18:02:24.688217Z"},{"marker":"1_1774029744.690957_181.1","last_update":"2026-03-20T18:02:24.690957Z"},{"marker":"1_1774029737.945479_86.1","last_update":"2026-03-20T18:02:17.945479Z"}] 2026-03-20T18:02:24.900 INFO:rgw_multi.tests:starting meta checkpoint for zone=a2 2026-03-20T18:02:24.900 INFO:tasks.util.rgw:rgwadmin: c1.client.0 : ['metadata', 'sync', 'status', '--rgw-zone', 'a2', '--rgw-zonegroup', 'a', '--rgw-realm', 'test-realm', '--cluster', 'c1', '--debug-rgw', '0', '--debug-ms', '0', '--rgw-cache-enabled', 'false'] 2026-03-20T18:02:24.900 DEBUG:tasks.util.rgw:rgwadmin: cmd=['adjust-ulimits', 'ceph-coverage', '/home/ubuntu/cephtest/archive/coverage', 'radosgw-admin', '--log-to-stderr', '--format', 'json', '-n', 'client.0', '--cluster', 'c1', 'metadata', 'sync', 'status', '--rgw-zone', 'a2', '--rgw-zonegroup', 'a', '--rgw-realm', 'test-realm', '--cluster', 'c1', '--debug-rgw', '0', '--debug-ms', '0', '--rgw-cache-enabled', 'false'] 2026-03-20T18:02:24.900 DEBUG:teuthology.orchestra.run.vm02:> adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage radosgw-admin --log-to-stderr --format json -n client.0 --cluster c1 metadata sync status --rgw-zone a2 --rgw-zonegroup a --rgw-realm test-realm --cluster c1 --debug-rgw 0 --debug-ms 0 --rgw-cache-enabled false 2026-03-20T18:02:24.962 INFO:teuthology.orchestra.run.vm02.stderr:ignoring --setuser ceph since I am not root 2026-03-20T18:02:24.962 INFO:teuthology.orchestra.run.vm02.stderr:ignoring --setgroup ceph since I am not root 2026-03-20T18:02:25.037 INFO:teuthology.orchestra.run.vm02.stderr:2026-03-20T18:02:25.033+0000 7f73f261aa80 20 RGW-SYNC:meta: read sync status 2026-03-20T18:02:25.037 INFO:teuthology.orchestra.run.vm02.stderr:2026-03-20T18:02:25.033+0000 7f73f261aa80 20 RGW-SYNC:meta: read sync status 2026-03-20T18:02:25.047 INFO:teuthology.orchestra.run.vm02.stdout:{"sync_status":{"info":{"status":"sync","num_shards":4,"period":"82c662be-66b5-4c4f-b24e-ff9298fc8a6e","realm_epoch":2},"markers":[{"key":0,"val":{"state":1,"marker":"1_1774029743.576954_130.1","next_step_marker":"","total_entries":1,"pos":0,"timestamp":"2026-03-20T18:02:23.576954Z","realm_epoch":2}},{"key":1,"val":{"state":1,"marker":"1_1774029744.688217_70.1","next_step_marker":"","total_entries":0,"pos":0,"timestamp":"2026-03-20T18:02:24.688217Z","realm_epoch":2}},{"key":2,"val":{"state":1,"marker":"1_1774029744.690957_181.1","next_step_marker":"","total_entries":0,"pos":0,"timestamp":"2026-03-20T18:02:24.690957Z","realm_epoch":2}},{"key":3,"val":{"state":1,"marker":"1_1774029737.945479_86.1","next_step_marker":"","total_entries":0,"pos":0,"timestamp":"2026-03-20T18:02:17.945479Z","realm_epoch":2}}]},"full_sync":{"total":1,"complete":1},"current_time":"2026-03-20T18:02:25Z"} 2026-03-20T18:02:25.048 INFO:rgw_multi.tests:finish meta checkpoint for zone=a2 2026-03-20T18:02:25.048 INFO:tasks.util.rgw:rgwadmin: c1.client.0 : ['bucket', 'sync', 'checkpoint', '--bucket', 'cyjkqa-7', '--source-zone', 'a1', '--retry-delay-ms', '5000', '--timeout-sec', '300', '--rgw-zone', 'a2', '--rgw-zonegroup', 'a', '--rgw-realm', 'test-realm', '--cluster', 'c1', '--debug-rgw', '1', '--debug-ms', '0'] 2026-03-20T18:02:25.048 DEBUG:tasks.util.rgw:rgwadmin: cmd=['adjust-ulimits', 'ceph-coverage', '/home/ubuntu/cephtest/archive/coverage', 'radosgw-admin', '--log-to-stderr', '--format', 'json', '-n', 'client.0', '--cluster', 'c1', 'bucket', 'sync', 'checkpoint', '--bucket', 'cyjkqa-7', '--source-zone', 'a1', '--retry-delay-ms', '5000', '--timeout-sec', '300', '--rgw-zone', 'a2', '--rgw-zonegroup', 'a', '--rgw-realm', 'test-realm', '--cluster', 'c1', '--debug-rgw', '1', '--debug-ms', '0'] 2026-03-20T18:02:25.048 DEBUG:teuthology.orchestra.run.vm02:> adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage radosgw-admin --log-to-stderr --format json -n client.0 --cluster c1 bucket sync checkpoint --bucket cyjkqa-7 --source-zone a1 --retry-delay-ms 5000 --timeout-sec 300 --rgw-zone a2 --rgw-zonegroup a --rgw-realm test-realm --cluster c1 --debug-rgw 1 --debug-ms 0 2026-03-20T18:02:25.114 INFO:teuthology.orchestra.run.vm02.stderr:ignoring --setuser ceph since I am not root 2026-03-20T18:02:25.114 INFO:teuthology.orchestra.run.vm02.stderr:ignoring --setgroup ceph since I am not root 2026-03-20T18:02:25.183 INFO:teuthology.orchestra.run.vm02.stderr:2026-03-20T18:02:25.181+0000 7efe2d8e2a80 1 waiting to reach incremental sync.. 2026-03-20T18:02:30.184 INFO:teuthology.orchestra.run.vm02.stderr:2026-03-20T18:02:30.181+0000 7efe2d8e2a80 1 bucket sync caught up with source: 2026-03-20T18:02:30.185 INFO:teuthology.orchestra.run.vm02.stderr: local status: [, 00000000002.12.6, , 00000000001.14.6, , 00000000001.5.6, 00000000002.12.6, 00000000001.9.6, 00000000001.6.6, 00000000001.13.6, ] 2026-03-20T18:02:30.185 INFO:teuthology.orchestra.run.vm02.stderr: remote markers: [, 00000000002.12.6, , 00000000001.14.6, , 00000000001.5.6, 00000000002.12.6, 00000000001.9.6, 00000000001.6.6, 00000000001.13.6, ] 2026-03-20T18:02:30.185 INFO:teuthology.orchestra.run.vm02.stderr:2026-03-20T18:02:30.181+0000 7efe2d8e2a80 0 bucket checkpoint complete 2026-03-20T18:02:30.196 INFO:tasks.rgw_multi.zone_rados:comparing bucket=cyjkqa-7 zones={a2, a1} 2026-03-20T18:02:30.213 DEBUG:tasks.rgw_multi.zone_rados:bucket1 objects: 2026-03-20T18:02:30.214 DEBUG:tasks.rgw_multi.zone_rados:o=&, v=null 2026-03-20T18:02:30.214 DEBUG:tasks.rgw_multi.zone_rados:o=., v=null 2026-03-20T18:02:30.214 DEBUG:tasks.rgw_multi.zone_rados:o=.., v=null 2026-03-20T18:02:30.214 DEBUG:tasks.rgw_multi.zone_rados:o=..., v=null 2026-03-20T18:02:30.214 DEBUG:tasks.rgw_multi.zone_rados:o=.o, v=null 2026-03-20T18:02:30.214 DEBUG:tasks.rgw_multi.zone_rados:o=.o., v=null 2026-03-20T18:02:30.214 DEBUG:tasks.rgw_multi.zone_rados:o=:, v=null 2026-03-20T18:02:30.214 DEBUG:tasks.rgw_multi.zone_rados:o=_myobj, v=null 2026-03-20T18:02:30.214 DEBUG:tasks.rgw_multi.zone_rados:o=myobj, v=null 2026-03-20T18:02:30.217 DEBUG:tasks.rgw_multi.zone_rados:bucket2 objects: 2026-03-20T18:02:30.217 DEBUG:tasks.rgw_multi.zone_rados:o=&, v=null 2026-03-20T18:02:30.217 DEBUG:tasks.rgw_multi.zone_rados:o=., v=null 2026-03-20T18:02:30.217 DEBUG:tasks.rgw_multi.zone_rados:o=.., v=null 2026-03-20T18:02:30.217 DEBUG:tasks.rgw_multi.zone_rados:o=..., v=null 2026-03-20T18:02:30.217 DEBUG:tasks.rgw_multi.zone_rados:o=.o, v=null 2026-03-20T18:02:30.217 DEBUG:tasks.rgw_multi.zone_rados:o=.o., v=null 2026-03-20T18:02:30.217 DEBUG:tasks.rgw_multi.zone_rados:o=:, v=null 2026-03-20T18:02:30.217 DEBUG:tasks.rgw_multi.zone_rados:o=_myobj, v=null 2026-03-20T18:02:30.218 DEBUG:tasks.rgw_multi.zone_rados:o=myobj, v=null 2026-03-20T18:02:30.218 DEBUG:tasks.rgw_multi.zone_rados:comparing key name=& 2026-03-20T18:02:30.277 DEBUG:tasks.rgw_multi.zone_rados:comparing key name=. 2026-03-20T18:02:30.334 DEBUG:tasks.rgw_multi.zone_rados:comparing key name=.. 2026-03-20T18:02:30.391 DEBUG:tasks.rgw_multi.zone_rados:comparing key name=... 2026-03-20T18:02:30.446 DEBUG:tasks.rgw_multi.zone_rados:comparing key name=.o 2026-03-20T18:02:30.506 DEBUG:tasks.rgw_multi.zone_rados:comparing key name=.o. 2026-03-20T18:02:30.562 DEBUG:tasks.rgw_multi.zone_rados:comparing key name=: 2026-03-20T18:02:30.619 DEBUG:tasks.rgw_multi.zone_rados:comparing key name=_myobj 2026-03-20T18:02:30.674 DEBUG:tasks.rgw_multi.zone_rados:comparing key name=myobj 2026-03-20T18:02:30.736 INFO:tasks.rgw_multi.zone_rados:success, bucket identical: bucket=cyjkqa-7 zones={a2, a1} 2026-03-20T18:02:30.736 INFO:tasks.util.rgw:rgwadmin: c1.client.0 : ['bucket', 'sync', 'checkpoint', '--bucket', 'cyjkqa-8', '--source-zone', 'a2', '--retry-delay-ms', '5000', '--timeout-sec', '300', '--rgw-zone', 'a1', '--rgw-zonegroup', 'a', '--rgw-realm', 'test-realm', '--cluster', 'c1', '--debug-rgw', '1', '--debug-ms', '0'] 2026-03-20T18:02:30.736 DEBUG:tasks.util.rgw:rgwadmin: cmd=['adjust-ulimits', 'ceph-coverage', '/home/ubuntu/cephtest/archive/coverage', 'radosgw-admin', '--log-to-stderr', '--format', 'json', '-n', 'client.0', '--cluster', 'c1', 'bucket', 'sync', 'checkpoint', '--bucket', 'cyjkqa-8', '--source-zone', 'a2', '--retry-delay-ms', '5000', '--timeout-sec', '300', '--rgw-zone', 'a1', '--rgw-zonegroup', 'a', '--rgw-realm', 'test-realm', '--cluster', 'c1', '--debug-rgw', '1', '--debug-ms', '0'] 2026-03-20T18:02:30.736 DEBUG:teuthology.orchestra.run.vm02:> adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage radosgw-admin --log-to-stderr --format json -n client.0 --cluster c1 bucket sync checkpoint --bucket cyjkqa-8 --source-zone a2 --retry-delay-ms 5000 --timeout-sec 300 --rgw-zone a1 --rgw-zonegroup a --rgw-realm test-realm --cluster c1 --debug-rgw 1 --debug-ms 0 2026-03-20T18:02:30.760 INFO:teuthology.orchestra.run.vm02.stderr:ignoring --setuser ceph since I am not root 2026-03-20T18:02:30.760 INFO:teuthology.orchestra.run.vm02.stderr:ignoring --setgroup ceph since I am not root 2026-03-20T18:02:30.844 INFO:teuthology.orchestra.run.vm02.stderr:2026-03-20T18:02:30.841+0000 7fa30bf73a80 1 bucket sync caught up with source: 2026-03-20T18:02:30.845 INFO:teuthology.orchestra.run.vm02.stderr: local status: [, 00000000002.21.6, , 00000000001.17.6, , 00000000001.16.6, 00000000002.7.6, 00000000001.6.6, 00000000001.5.6, 00000000001.11.6, ] 2026-03-20T18:02:30.845 INFO:teuthology.orchestra.run.vm02.stderr: remote markers: [, 00000000002.21.6, , 00000000001.17.6, , 00000000001.16.6, 00000000002.7.6, 00000000001.6.6, 00000000001.5.6, 00000000001.11.6, ] 2026-03-20T18:02:30.845 INFO:teuthology.orchestra.run.vm02.stderr:2026-03-20T18:02:30.841+0000 7fa30bf73a80 0 bucket checkpoint complete 2026-03-20T18:02:30.854 INFO:tasks.rgw_multi.zone_rados:comparing bucket=cyjkqa-8 zones={a1, a2} 2026-03-20T18:02:30.862 DEBUG:tasks.rgw_multi.zone_rados:bucket1 objects: 2026-03-20T18:02:30.862 DEBUG:tasks.rgw_multi.zone_rados:o=&, v=null 2026-03-20T18:02:30.862 DEBUG:tasks.rgw_multi.zone_rados:o=., v=null 2026-03-20T18:02:30.862 DEBUG:tasks.rgw_multi.zone_rados:o=.., v=null 2026-03-20T18:02:30.862 DEBUG:tasks.rgw_multi.zone_rados:o=..., v=null 2026-03-20T18:02:30.862 DEBUG:tasks.rgw_multi.zone_rados:o=.o, v=null 2026-03-20T18:02:30.862 DEBUG:tasks.rgw_multi.zone_rados:o=.o., v=null 2026-03-20T18:02:30.862 DEBUG:tasks.rgw_multi.zone_rados:o=:, v=null 2026-03-20T18:02:30.862 DEBUG:tasks.rgw_multi.zone_rados:o=_myobj, v=null 2026-03-20T18:02:30.862 DEBUG:tasks.rgw_multi.zone_rados:o=myobj, v=null 2026-03-20T18:02:30.865 DEBUG:tasks.rgw_multi.zone_rados:bucket2 objects: 2026-03-20T18:02:30.865 DEBUG:tasks.rgw_multi.zone_rados:o=&, v=null 2026-03-20T18:02:30.865 DEBUG:tasks.rgw_multi.zone_rados:o=., v=null 2026-03-20T18:02:30.865 DEBUG:tasks.rgw_multi.zone_rados:o=.., v=null 2026-03-20T18:02:30.865 DEBUG:tasks.rgw_multi.zone_rados:o=..., v=null 2026-03-20T18:02:30.865 DEBUG:tasks.rgw_multi.zone_rados:o=.o, v=null 2026-03-20T18:02:30.865 DEBUG:tasks.rgw_multi.zone_rados:o=.o., v=null 2026-03-20T18:02:30.865 DEBUG:tasks.rgw_multi.zone_rados:o=:, v=null 2026-03-20T18:02:30.865 DEBUG:tasks.rgw_multi.zone_rados:o=_myobj, v=null 2026-03-20T18:02:30.865 DEBUG:tasks.rgw_multi.zone_rados:o=myobj, v=null 2026-03-20T18:02:30.865 DEBUG:tasks.rgw_multi.zone_rados:comparing key name=& 2026-03-20T18:02:30.923 DEBUG:tasks.rgw_multi.zone_rados:comparing key name=. 2026-03-20T18:02:30.978 DEBUG:tasks.rgw_multi.zone_rados:comparing key name=.. 2026-03-20T18:02:31.106 DEBUG:tasks.rgw_multi.zone_rados:comparing key name=... 2026-03-20T18:02:31.162 DEBUG:tasks.rgw_multi.zone_rados:comparing key name=.o 2026-03-20T18:02:31.218 DEBUG:tasks.rgw_multi.zone_rados:comparing key name=.o. 2026-03-20T18:02:31.269 DEBUG:tasks.rgw_multi.zone_rados:comparing key name=: 2026-03-20T18:02:31.326 DEBUG:tasks.rgw_multi.zone_rados:comparing key name=_myobj 2026-03-20T18:02:31.383 DEBUG:tasks.rgw_multi.zone_rados:comparing key name=myobj 2026-03-20T18:02:31.439 INFO:tasks.rgw_multi.zone_rados:success, bucket identical: bucket=cyjkqa-8 zones={a1, a2} 2026-03-20T18:02:31.441 INFO:tasks.rgw_multisite_tests:rgw_multi.tests.test_object_sync ... ok 2026-03-20T18:02:31.495 INFO:rgw_multi.tests:create bucket zone=a1 name=cyjkqa-9 2026-03-20T18:02:31.508 INFO:rgw_multi.tests:create bucket zone=a2 name=cyjkqa-10 2026-03-20T18:02:31.540 INFO:tasks.util.rgw:rgwadmin: c1.client.0 : ['mdlog', 'status', '--rgw-zone', 'a1', '--rgw-zonegroup', 'a', '--rgw-realm', 'test-realm', '--cluster', 'c1', '--debug-rgw', '0', '--debug-ms', '0', '--rgw-cache-enabled', 'false'] 2026-03-20T18:02:31.540 DEBUG:tasks.util.rgw:rgwadmin: cmd=['adjust-ulimits', 'ceph-coverage', '/home/ubuntu/cephtest/archive/coverage', 'radosgw-admin', '--log-to-stderr', '--format', 'json', '-n', 'client.0', '--cluster', 'c1', 'mdlog', 'status', '--rgw-zone', 'a1', '--rgw-zonegroup', 'a', '--rgw-realm', 'test-realm', '--cluster', 'c1', '--debug-rgw', '0', '--debug-ms', '0', '--rgw-cache-enabled', 'false'] 2026-03-20T18:02:31.540 DEBUG:teuthology.orchestra.run.vm02:> adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage radosgw-admin --log-to-stderr --format json -n client.0 --cluster c1 mdlog status --rgw-zone a1 --rgw-zonegroup a --rgw-realm test-realm --cluster c1 --debug-rgw 0 --debug-ms 0 --rgw-cache-enabled false 2026-03-20T18:02:31.568 INFO:teuthology.orchestra.run.vm02.stderr:ignoring --setuser ceph since I am not root 2026-03-20T18:02:31.568 INFO:teuthology.orchestra.run.vm02.stderr:ignoring --setgroup ceph since I am not root 2026-03-20T18:02:31.639 INFO:teuthology.orchestra.run.vm02.stderr:No --period given, using current period=82c662be-66b5-4c4f-b24e-ff9298fc8a6e 2026-03-20T18:02:31.651 INFO:teuthology.orchestra.run.vm02.stdout:[{"marker":"1_1774029751.515343_165.1","last_update":"2026-03-20T18:02:31.515343Z"},{"marker":"1_1774029751.502359_72.1","last_update":"2026-03-20T18:02:31.502359Z"},{"marker":"1_1774029751.504947_186.1","last_update":"2026-03-20T18:02:31.504947Z"},{"marker":"1_1774029751.517733_97.1","last_update":"2026-03-20T18:02:31.517733Z"}] 2026-03-20T18:02:31.651 INFO:rgw_multi.tests:starting meta checkpoint for zone=a2 2026-03-20T18:02:31.651 INFO:tasks.util.rgw:rgwadmin: c1.client.0 : ['metadata', 'sync', 'status', '--rgw-zone', 'a2', '--rgw-zonegroup', 'a', '--rgw-realm', 'test-realm', '--cluster', 'c1', '--debug-rgw', '0', '--debug-ms', '0', '--rgw-cache-enabled', 'false'] 2026-03-20T18:02:31.651 DEBUG:tasks.util.rgw:rgwadmin: cmd=['adjust-ulimits', 'ceph-coverage', '/home/ubuntu/cephtest/archive/coverage', 'radosgw-admin', '--log-to-stderr', '--format', 'json', '-n', 'client.0', '--cluster', 'c1', 'metadata', 'sync', 'status', '--rgw-zone', 'a2', '--rgw-zonegroup', 'a', '--rgw-realm', 'test-realm', '--cluster', 'c1', '--debug-rgw', '0', '--debug-ms', '0', '--rgw-cache-enabled', 'false'] 2026-03-20T18:02:31.651 DEBUG:teuthology.orchestra.run.vm02:> adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage radosgw-admin --log-to-stderr --format json -n client.0 --cluster c1 metadata sync status --rgw-zone a2 --rgw-zonegroup a --rgw-realm test-realm --cluster c1 --debug-rgw 0 --debug-ms 0 --rgw-cache-enabled false 2026-03-20T18:02:31.714 INFO:teuthology.orchestra.run.vm02.stderr:ignoring --setuser ceph since I am not root 2026-03-20T18:02:31.714 INFO:teuthology.orchestra.run.vm02.stderr:ignoring --setgroup ceph since I am not root 2026-03-20T18:02:31.825 INFO:teuthology.orchestra.run.vm02.stderr:2026-03-20T18:02:31.821+0000 7f2fc7da9a80 20 RGW-SYNC:meta: read sync status 2026-03-20T18:02:31.825 INFO:teuthology.orchestra.run.vm02.stderr:2026-03-20T18:02:31.821+0000 7f2fc7da9a80 20 RGW-SYNC:meta: read sync status 2026-03-20T18:02:31.836 INFO:teuthology.orchestra.run.vm02.stdout:{"sync_status":{"info":{"status":"sync","num_shards":4,"period":"82c662be-66b5-4c4f-b24e-ff9298fc8a6e","realm_epoch":2},"markers":[{"key":0,"val":{"state":1,"marker":"1_1774029751.515343_165.1","next_step_marker":"","total_entries":1,"pos":0,"timestamp":"2026-03-20T18:02:31.515343Z","realm_epoch":2}},{"key":1,"val":{"state":1,"marker":"1_1774029751.502359_72.1","next_step_marker":"","total_entries":0,"pos":0,"timestamp":"2026-03-20T18:02:31.502359Z","realm_epoch":2}},{"key":2,"val":{"state":1,"marker":"1_1774029751.504947_186.1","next_step_marker":"","total_entries":0,"pos":0,"timestamp":"2026-03-20T18:02:31.504947Z","realm_epoch":2}},{"key":3,"val":{"state":1,"marker":"1_1774029751.517733_97.1","next_step_marker":"","total_entries":0,"pos":0,"timestamp":"2026-03-20T18:02:31.517733Z","realm_epoch":2}}]},"full_sync":{"total":1,"complete":1},"current_time":"2026-03-20T18:02:31Z"} 2026-03-20T18:02:31.836 INFO:rgw_multi.tests:finish meta checkpoint for zone=a2 2026-03-20T18:02:31.837 INFO:tasks.util.rgw:rgwadmin: c1.client.0 : ['bucket', 'sync', 'checkpoint', '--bucket', 'cyjkqa-9', '--source-zone', 'a1', '--retry-delay-ms', '5000', '--timeout-sec', '300', '--rgw-zone', 'a2', '--rgw-zonegroup', 'a', '--rgw-realm', 'test-realm', '--cluster', 'c1', '--debug-rgw', '1', '--debug-ms', '0'] 2026-03-20T18:02:31.837 DEBUG:tasks.util.rgw:rgwadmin: cmd=['adjust-ulimits', 'ceph-coverage', '/home/ubuntu/cephtest/archive/coverage', 'radosgw-admin', '--log-to-stderr', '--format', 'json', '-n', 'client.0', '--cluster', 'c1', 'bucket', 'sync', 'checkpoint', '--bucket', 'cyjkqa-9', '--source-zone', 'a1', '--retry-delay-ms', '5000', '--timeout-sec', '300', '--rgw-zone', 'a2', '--rgw-zonegroup', 'a', '--rgw-realm', 'test-realm', '--cluster', 'c1', '--debug-rgw', '1', '--debug-ms', '0'] 2026-03-20T18:02:31.837 DEBUG:teuthology.orchestra.run.vm02:> adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage radosgw-admin --log-to-stderr --format json -n client.0 --cluster c1 bucket sync checkpoint --bucket cyjkqa-9 --source-zone a1 --retry-delay-ms 5000 --timeout-sec 300 --rgw-zone a2 --rgw-zonegroup a --rgw-realm test-realm --cluster c1 --debug-rgw 1 --debug-ms 0 2026-03-20T18:02:31.906 INFO:teuthology.orchestra.run.vm02.stderr:ignoring --setuser ceph since I am not root 2026-03-20T18:02:31.906 INFO:teuthology.orchestra.run.vm02.stderr:ignoring --setgroup ceph since I am not root 2026-03-20T18:02:31.981 INFO:teuthology.orchestra.run.vm02.stderr:2026-03-20T18:02:31.977+0000 7fbee4cd0a80 1 waiting to reach incremental sync.. 2026-03-20T18:02:36.982 INFO:teuthology.orchestra.run.vm02.stderr:2026-03-20T18:02:36.977+0000 7fbee4cd0a80 1 bucket sync caught up with source: 2026-03-20T18:02:36.982 INFO:teuthology.orchestra.run.vm02.stderr: local status: [, 00000000001.6.6, , , , , , , , , ] 2026-03-20T18:02:36.982 INFO:teuthology.orchestra.run.vm02.stderr: remote markers: [, 00000000001.6.6, , , , , , , , , ] 2026-03-20T18:02:36.982 INFO:teuthology.orchestra.run.vm02.stderr:2026-03-20T18:02:36.977+0000 7fbee4cd0a80 0 bucket checkpoint complete 2026-03-20T18:02:36.992 INFO:tasks.rgw_multi.zone_rados:comparing bucket=cyjkqa-9 zones={a2, a1} 2026-03-20T18:02:36.999 DEBUG:tasks.rgw_multi.zone_rados:bucket1 objects: 2026-03-20T18:02:36.999 DEBUG:tasks.rgw_multi.zone_rados:o=myobj, v=null 2026-03-20T18:02:37.002 DEBUG:tasks.rgw_multi.zone_rados:bucket2 objects: 2026-03-20T18:02:37.002 DEBUG:tasks.rgw_multi.zone_rados:o=myobj, v=null 2026-03-20T18:02:37.002 DEBUG:tasks.rgw_multi.zone_rados:comparing key name=myobj 2026-03-20T18:02:37.059 INFO:tasks.rgw_multi.zone_rados:success, bucket identical: bucket=cyjkqa-9 zones={a2, a1} 2026-03-20T18:02:37.059 INFO:tasks.util.rgw:rgwadmin: c1.client.0 : ['bucket', 'sync', 'checkpoint', '--bucket', 'cyjkqa-10', '--source-zone', 'a2', '--retry-delay-ms', '5000', '--timeout-sec', '300', '--rgw-zone', 'a1', '--rgw-zonegroup', 'a', '--rgw-realm', 'test-realm', '--cluster', 'c1', '--debug-rgw', '1', '--debug-ms', '0'] 2026-03-20T18:02:37.059 DEBUG:tasks.util.rgw:rgwadmin: cmd=['adjust-ulimits', 'ceph-coverage', '/home/ubuntu/cephtest/archive/coverage', 'radosgw-admin', '--log-to-stderr', '--format', 'json', '-n', 'client.0', '--cluster', 'c1', 'bucket', 'sync', 'checkpoint', '--bucket', 'cyjkqa-10', '--source-zone', 'a2', '--retry-delay-ms', '5000', '--timeout-sec', '300', '--rgw-zone', 'a1', '--rgw-zonegroup', 'a', '--rgw-realm', 'test-realm', '--cluster', 'c1', '--debug-rgw', '1', '--debug-ms', '0'] 2026-03-20T18:02:37.059 DEBUG:teuthology.orchestra.run.vm02:> adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage radosgw-admin --log-to-stderr --format json -n client.0 --cluster c1 bucket sync checkpoint --bucket cyjkqa-10 --source-zone a2 --retry-delay-ms 5000 --timeout-sec 300 --rgw-zone a1 --rgw-zonegroup a --rgw-realm test-realm --cluster c1 --debug-rgw 1 --debug-ms 0 2026-03-20T18:02:37.083 INFO:teuthology.orchestra.run.vm02.stderr:ignoring --setuser ceph since I am not root 2026-03-20T18:02:37.083 INFO:teuthology.orchestra.run.vm02.stderr:ignoring --setgroup ceph since I am not root 2026-03-20T18:02:37.165 INFO:teuthology.orchestra.run.vm02.stderr:2026-03-20T18:02:37.161+0000 7f128314aa80 1 bucket sync caught up with source: 2026-03-20T18:02:37.165 INFO:teuthology.orchestra.run.vm02.stderr: local status: [, 00000000001.22.6, , , , , , , , , ] 2026-03-20T18:02:37.165 INFO:teuthology.orchestra.run.vm02.stderr: remote markers: [, 00000000001.22.6, , , , , , , , , ] 2026-03-20T18:02:37.165 INFO:teuthology.orchestra.run.vm02.stderr:2026-03-20T18:02:37.161+0000 7f128314aa80 0 bucket checkpoint complete 2026-03-20T18:02:37.175 INFO:tasks.rgw_multi.zone_rados:comparing bucket=cyjkqa-10 zones={a1, a2} 2026-03-20T18:02:37.182 DEBUG:tasks.rgw_multi.zone_rados:bucket1 objects: 2026-03-20T18:02:37.182 DEBUG:tasks.rgw_multi.zone_rados:o=myobj, v=null 2026-03-20T18:02:37.184 DEBUG:tasks.rgw_multi.zone_rados:bucket2 objects: 2026-03-20T18:02:37.184 DEBUG:tasks.rgw_multi.zone_rados:o=myobj, v=null 2026-03-20T18:02:37.184 DEBUG:tasks.rgw_multi.zone_rados:comparing key name=myobj 2026-03-20T18:02:37.245 INFO:tasks.rgw_multi.zone_rados:success, bucket identical: bucket=cyjkqa-10 zones={a1, a2} 2026-03-20T18:02:37.252 INFO:tasks.util.rgw:rgwadmin: c1.client.0 : ['bucket', 'sync', 'checkpoint', '--bucket', 'cyjkqa-9', '--source-zone', 'a1', '--retry-delay-ms', '5000', '--timeout-sec', '300', '--rgw-zone', 'a2', '--rgw-zonegroup', 'a', '--rgw-realm', 'test-realm', '--cluster', 'c1', '--debug-rgw', '1', '--debug-ms', '0'] 2026-03-20T18:02:37.252 DEBUG:tasks.util.rgw:rgwadmin: cmd=['adjust-ulimits', 'ceph-coverage', '/home/ubuntu/cephtest/archive/coverage', 'radosgw-admin', '--log-to-stderr', '--format', 'json', '-n', 'client.0', '--cluster', 'c1', 'bucket', 'sync', 'checkpoint', '--bucket', 'cyjkqa-9', '--source-zone', 'a1', '--retry-delay-ms', '5000', '--timeout-sec', '300', '--rgw-zone', 'a2', '--rgw-zonegroup', 'a', '--rgw-realm', 'test-realm', '--cluster', 'c1', '--debug-rgw', '1', '--debug-ms', '0'] 2026-03-20T18:02:37.252 DEBUG:teuthology.orchestra.run.vm02:> adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage radosgw-admin --log-to-stderr --format json -n client.0 --cluster c1 bucket sync checkpoint --bucket cyjkqa-9 --source-zone a1 --retry-delay-ms 5000 --timeout-sec 300 --rgw-zone a2 --rgw-zonegroup a --rgw-realm test-realm --cluster c1 --debug-rgw 1 --debug-ms 0 2026-03-20T18:02:37.277 INFO:teuthology.orchestra.run.vm02.stderr:ignoring --setuser ceph since I am not root 2026-03-20T18:02:37.277 INFO:teuthology.orchestra.run.vm02.stderr:ignoring --setgroup ceph since I am not root 2026-03-20T18:02:37.356 INFO:teuthology.orchestra.run.vm02.stderr:2026-03-20T18:02:37.353+0000 7f8de0d0fa80 1 waiting for incremental sync to catch up: 2026-03-20T18:02:37.356 INFO:teuthology.orchestra.run.vm02.stderr: local status: [, 00000000001.6.6, , , , , , , , , ] 2026-03-20T18:02:37.356 INFO:teuthology.orchestra.run.vm02.stderr: remote markers: [, 00000000002.8.6, , , , , , , , , ] 2026-03-20T18:02:42.357 INFO:teuthology.orchestra.run.vm02.stderr:2026-03-20T18:02:42.353+0000 7f8de0d0fa80 1 bucket sync caught up with source: 2026-03-20T18:02:42.357 INFO:teuthology.orchestra.run.vm02.stderr: local status: [, 00000000002.8.6, , , , , , , , , ] 2026-03-20T18:02:42.357 INFO:teuthology.orchestra.run.vm02.stderr: remote markers: [, 00000000002.8.6, , , , , , , , , ] 2026-03-20T18:02:42.357 INFO:teuthology.orchestra.run.vm02.stderr:2026-03-20T18:02:42.353+0000 7f8de0d0fa80 0 bucket checkpoint complete 2026-03-20T18:02:42.366 INFO:tasks.rgw_multi.zone_rados:comparing bucket=cyjkqa-9 zones={a2, a1} 2026-03-20T18:02:42.372 DEBUG:tasks.rgw_multi.zone_rados:bucket1 objects: 2026-03-20T18:02:42.374 DEBUG:tasks.rgw_multi.zone_rados:bucket2 objects: 2026-03-20T18:02:42.374 INFO:tasks.rgw_multi.zone_rados:success, bucket identical: bucket=cyjkqa-9 zones={a2, a1} 2026-03-20T18:02:42.379 INFO:tasks.util.rgw:rgwadmin: c1.client.0 : ['bucket', 'sync', 'checkpoint', '--bucket', 'cyjkqa-10', '--source-zone', 'a2', '--retry-delay-ms', '5000', '--timeout-sec', '300', '--rgw-zone', 'a1', '--rgw-zonegroup', 'a', '--rgw-realm', 'test-realm', '--cluster', 'c1', '--debug-rgw', '1', '--debug-ms', '0'] 2026-03-20T18:02:42.379 DEBUG:tasks.util.rgw:rgwadmin: cmd=['adjust-ulimits', 'ceph-coverage', '/home/ubuntu/cephtest/archive/coverage', 'radosgw-admin', '--log-to-stderr', '--format', 'json', '-n', 'client.0', '--cluster', 'c1', 'bucket', 'sync', 'checkpoint', '--bucket', 'cyjkqa-10', '--source-zone', 'a2', '--retry-delay-ms', '5000', '--timeout-sec', '300', '--rgw-zone', 'a1', '--rgw-zonegroup', 'a', '--rgw-realm', 'test-realm', '--cluster', 'c1', '--debug-rgw', '1', '--debug-ms', '0'] 2026-03-20T18:02:42.379 DEBUG:teuthology.orchestra.run.vm02:> adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage radosgw-admin --log-to-stderr --format json -n client.0 --cluster c1 bucket sync checkpoint --bucket cyjkqa-10 --source-zone a2 --retry-delay-ms 5000 --timeout-sec 300 --rgw-zone a1 --rgw-zonegroup a --rgw-realm test-realm --cluster c1 --debug-rgw 1 --debug-ms 0 2026-03-20T18:02:42.430 INFO:teuthology.orchestra.run.vm02.stderr:ignoring --setuser ceph since I am not root 2026-03-20T18:02:42.430 INFO:teuthology.orchestra.run.vm02.stderr:ignoring --setgroup ceph since I am not root 2026-03-20T18:02:42.505 INFO:teuthology.orchestra.run.vm02.stderr:2026-03-20T18:02:42.501+0000 7f900b1f8a80 1 waiting for incremental sync to catch up: 2026-03-20T18:02:42.505 INFO:teuthology.orchestra.run.vm02.stderr: local status: [, 00000000001.22.6, , , , , , , , , ] 2026-03-20T18:02:42.505 INFO:teuthology.orchestra.run.vm02.stderr: remote markers: [, 00000000002.26.6, , , , , , , , , ] 2026-03-20T18:02:47.506 INFO:teuthology.orchestra.run.vm02.stderr:2026-03-20T18:02:47.501+0000 7f900b1f8a80 1 waiting for incremental sync to catch up: 2026-03-20T18:02:47.506 INFO:teuthology.orchestra.run.vm02.stderr: local status: [, 00000000001.22.6, , , , , , , , , ] 2026-03-20T18:02:47.506 INFO:teuthology.orchestra.run.vm02.stderr: remote markers: [, 00000000002.26.6, , , , , , , , , ] 2026-03-20T18:02:52.507 INFO:teuthology.orchestra.run.vm02.stderr:2026-03-20T18:02:52.505+0000 7f900b1f8a80 1 waiting for incremental sync to catch up: 2026-03-20T18:02:52.507 INFO:teuthology.orchestra.run.vm02.stderr: local status: [, 00000000001.22.6, , , , , , , , , ] 2026-03-20T18:02:52.507 INFO:teuthology.orchestra.run.vm02.stderr: remote markers: [, 00000000002.26.6, , , , , , , , , ] 2026-03-20T18:02:57.508 INFO:teuthology.orchestra.run.vm02.stderr:2026-03-20T18:02:57.505+0000 7f900b1f8a80 1 waiting for incremental sync to catch up: 2026-03-20T18:02:57.508 INFO:teuthology.orchestra.run.vm02.stderr: local status: [, 00000000001.22.6, , , , , , , , , ] 2026-03-20T18:02:57.508 INFO:teuthology.orchestra.run.vm02.stderr: remote markers: [, 00000000002.26.6, , , , , , , , , ] 2026-03-20T18:03:02.509 INFO:teuthology.orchestra.run.vm02.stderr:2026-03-20T18:03:02.505+0000 7f900b1f8a80 1 waiting for incremental sync to catch up: 2026-03-20T18:03:02.509 INFO:teuthology.orchestra.run.vm02.stderr: local status: [, 00000000001.22.6, , , , , , , , , ] 2026-03-20T18:03:02.509 INFO:teuthology.orchestra.run.vm02.stderr: remote markers: [, 00000000002.26.6, , , , , , , , , ] 2026-03-20T18:03:07.510 INFO:teuthology.orchestra.run.vm02.stderr:2026-03-20T18:03:07.505+0000 7f900b1f8a80 1 bucket sync caught up with source: 2026-03-20T18:03:07.510 INFO:teuthology.orchestra.run.vm02.stderr: local status: [, 00000000002.26.6, , , , , , , , , ] 2026-03-20T18:03:07.510 INFO:teuthology.orchestra.run.vm02.stderr: remote markers: [, 00000000002.26.6, , , , , , , , , ] 2026-03-20T18:03:07.510 INFO:teuthology.orchestra.run.vm02.stderr:2026-03-20T18:03:07.505+0000 7f900b1f8a80 0 bucket checkpoint complete 2026-03-20T18:03:07.520 INFO:tasks.rgw_multi.zone_rados:comparing bucket=cyjkqa-10 zones={a1, a2} 2026-03-20T18:03:07.526 DEBUG:tasks.rgw_multi.zone_rados:bucket1 objects: 2026-03-20T18:03:07.528 DEBUG:tasks.rgw_multi.zone_rados:bucket2 objects: 2026-03-20T18:03:07.528 INFO:tasks.rgw_multi.zone_rados:success, bucket identical: bucket=cyjkqa-10 zones={a1, a2} 2026-03-20T18:03:07.530 INFO:tasks.rgw_multisite_tests:rgw_multi.tests.test_object_delete ... ok 2026-03-20T18:03:07.651 INFO:rgw_multi.tests:create bucket zone=a1 name=cyjkqa-11 2026-03-20T18:03:07.666 INFO:rgw_multi.tests:create bucket zone=a2 name=cyjkqa-12 2026-03-20T18:03:08.198 INFO:tasks.util.rgw:rgwadmin: c1.client.0 : ['mdlog', 'status', '--rgw-zone', 'a1', '--rgw-zonegroup', 'a', '--rgw-realm', 'test-realm', '--cluster', 'c1', '--debug-rgw', '0', '--debug-ms', '0', '--rgw-cache-enabled', 'false'] 2026-03-20T18:03:08.198 DEBUG:tasks.util.rgw:rgwadmin: cmd=['adjust-ulimits', 'ceph-coverage', '/home/ubuntu/cephtest/archive/coverage', 'radosgw-admin', '--log-to-stderr', '--format', 'json', '-n', 'client.0', '--cluster', 'c1', 'mdlog', 'status', '--rgw-zone', 'a1', '--rgw-zonegroup', 'a', '--rgw-realm', 'test-realm', '--cluster', 'c1', '--debug-rgw', '0', '--debug-ms', '0', '--rgw-cache-enabled', 'false'] 2026-03-20T18:03:08.199 DEBUG:teuthology.orchestra.run.vm02:> adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage radosgw-admin --log-to-stderr --format json -n client.0 --cluster c1 mdlog status --rgw-zone a1 --rgw-zonegroup a --rgw-realm test-realm --cluster c1 --debug-rgw 0 --debug-ms 0 --rgw-cache-enabled false 2026-03-20T18:03:08.224 INFO:teuthology.orchestra.run.vm02.stderr:ignoring --setuser ceph since I am not root 2026-03-20T18:03:08.224 INFO:teuthology.orchestra.run.vm02.stderr:ignoring --setgroup ceph since I am not root 2026-03-20T18:03:08.306 INFO:teuthology.orchestra.run.vm02.stderr:No --period given, using current period=82c662be-66b5-4c4f-b24e-ff9298fc8a6e 2026-03-20T18:03:08.316 INFO:teuthology.orchestra.run.vm02.stdout:[{"marker":"1_1774029787.674910_178.1","last_update":"2026-03-20T18:03:07.674910Z"},{"marker":"1_1774029751.502359_72.1","last_update":"2026-03-20T18:02:31.502359Z"},{"marker":"1_1774029751.504947_186.1","last_update":"2026-03-20T18:02:31.504947Z"},{"marker":"1_1774029787.677348_109.1","last_update":"2026-03-20T18:03:07.677348Z"}] 2026-03-20T18:03:08.316 INFO:rgw_multi.tests:starting meta checkpoint for zone=a2 2026-03-20T18:03:08.316 INFO:tasks.util.rgw:rgwadmin: c1.client.0 : ['metadata', 'sync', 'status', '--rgw-zone', 'a2', '--rgw-zonegroup', 'a', '--rgw-realm', 'test-realm', '--cluster', 'c1', '--debug-rgw', '0', '--debug-ms', '0', '--rgw-cache-enabled', 'false'] 2026-03-20T18:03:08.316 DEBUG:tasks.util.rgw:rgwadmin: cmd=['adjust-ulimits', 'ceph-coverage', '/home/ubuntu/cephtest/archive/coverage', 'radosgw-admin', '--log-to-stderr', '--format', 'json', '-n', 'client.0', '--cluster', 'c1', 'metadata', 'sync', 'status', '--rgw-zone', 'a2', '--rgw-zonegroup', 'a', '--rgw-realm', 'test-realm', '--cluster', 'c1', '--debug-rgw', '0', '--debug-ms', '0', '--rgw-cache-enabled', 'false'] 2026-03-20T18:03:08.316 DEBUG:teuthology.orchestra.run.vm02:> adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage radosgw-admin --log-to-stderr --format json -n client.0 --cluster c1 metadata sync status --rgw-zone a2 --rgw-zonegroup a --rgw-realm test-realm --cluster c1 --debug-rgw 0 --debug-ms 0 --rgw-cache-enabled false 2026-03-20T18:03:08.383 INFO:teuthology.orchestra.run.vm02.stderr:ignoring --setuser ceph since I am not root 2026-03-20T18:03:08.383 INFO:teuthology.orchestra.run.vm02.stderr:ignoring --setgroup ceph since I am not root 2026-03-20T18:03:08.481 INFO:teuthology.orchestra.run.vm02.stderr:2026-03-20T18:03:08.477+0000 7f4bc21d8a80 20 RGW-SYNC:meta: read sync status 2026-03-20T18:03:08.481 INFO:teuthology.orchestra.run.vm02.stderr:2026-03-20T18:03:08.477+0000 7f4bc21d8a80 20 RGW-SYNC:meta: read sync status 2026-03-20T18:03:08.494 INFO:teuthology.orchestra.run.vm02.stdout:{"sync_status":{"info":{"status":"sync","num_shards":4,"period":"82c662be-66b5-4c4f-b24e-ff9298fc8a6e","realm_epoch":2},"markers":[{"key":0,"val":{"state":1,"marker":"1_1774029787.674910_178.1","next_step_marker":"","total_entries":1,"pos":0,"timestamp":"2026-03-20T18:03:07.674910Z","realm_epoch":2}},{"key":1,"val":{"state":1,"marker":"1_1774029751.502359_72.1","next_step_marker":"","total_entries":0,"pos":0,"timestamp":"2026-03-20T18:02:31.502359Z","realm_epoch":2}},{"key":2,"val":{"state":1,"marker":"1_1774029751.504947_186.1","next_step_marker":"","total_entries":0,"pos":0,"timestamp":"2026-03-20T18:02:31.504947Z","realm_epoch":2}},{"key":3,"val":{"state":1,"marker":"1_1774029787.677348_109.1","next_step_marker":"","total_entries":0,"pos":0,"timestamp":"2026-03-20T18:03:07.677348Z","realm_epoch":2}}]},"full_sync":{"total":1,"complete":1},"current_time":"2026-03-20T18:03:08Z"} 2026-03-20T18:03:08.494 INFO:rgw_multi.tests:finish meta checkpoint for zone=a2 2026-03-20T18:03:08.494 INFO:tasks.util.rgw:rgwadmin: c1.client.0 : ['bucket', 'sync', 'checkpoint', '--bucket', 'cyjkqa-11', '--source-zone', 'a1', '--retry-delay-ms', '5000', '--timeout-sec', '300', '--rgw-zone', 'a2', '--rgw-zonegroup', 'a', '--rgw-realm', 'test-realm', '--cluster', 'c1', '--debug-rgw', '1', '--debug-ms', '0'] 2026-03-20T18:03:08.494 DEBUG:tasks.util.rgw:rgwadmin: cmd=['adjust-ulimits', 'ceph-coverage', '/home/ubuntu/cephtest/archive/coverage', 'radosgw-admin', '--log-to-stderr', '--format', 'json', '-n', 'client.0', '--cluster', 'c1', 'bucket', 'sync', 'checkpoint', '--bucket', 'cyjkqa-11', '--source-zone', 'a1', '--retry-delay-ms', '5000', '--timeout-sec', '300', '--rgw-zone', 'a2', '--rgw-zonegroup', 'a', '--rgw-realm', 'test-realm', '--cluster', 'c1', '--debug-rgw', '1', '--debug-ms', '0'] 2026-03-20T18:03:08.494 DEBUG:teuthology.orchestra.run.vm02:> adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage radosgw-admin --log-to-stderr --format json -n client.0 --cluster c1 bucket sync checkpoint --bucket cyjkqa-11 --source-zone a1 --retry-delay-ms 5000 --timeout-sec 300 --rgw-zone a2 --rgw-zonegroup a --rgw-realm test-realm --cluster c1 --debug-rgw 1 --debug-ms 0 2026-03-20T18:03:08.564 INFO:teuthology.orchestra.run.vm02.stderr:ignoring --setuser ceph since I am not root 2026-03-20T18:03:08.564 INFO:teuthology.orchestra.run.vm02.stderr:ignoring --setgroup ceph since I am not root 2026-03-20T18:03:08.652 INFO:teuthology.orchestra.run.vm02.stderr:2026-03-20T18:03:08.649+0000 7fcb8a77aa80 1 waiting to reach incremental sync.. 2026-03-20T18:03:13.654 INFO:teuthology.orchestra.run.vm02.stderr:2026-03-20T18:03:13.649+0000 7fcb8a77aa80 1 bucket sync caught up with source: 2026-03-20T18:03:13.654 INFO:teuthology.orchestra.run.vm02.stderr: local status: [00000000007.32.6, 00000000004.11.6, 00000000004.18.6, 00000000004.31.6, 00000000006.19.6, 00000000005.15.6, 00000000004.33.6, 00000000005.25.6, 00000000002.34.6, 00000000004.16.6, 00000000004.17.6] 2026-03-20T18:03:13.654 INFO:teuthology.orchestra.run.vm02.stderr: remote markers: [00000000007.32.6, 00000000004.11.6, 00000000004.18.6, 00000000004.31.6, 00000000006.19.6, 00000000005.15.6, 00000000004.33.6, 00000000005.25.6, 00000000002.34.6, 00000000004.16.6, 00000000004.17.6] 2026-03-20T18:03:13.654 INFO:teuthology.orchestra.run.vm02.stderr:2026-03-20T18:03:13.649+0000 7fcb8a77aa80 0 bucket checkpoint complete 2026-03-20T18:03:13.668 INFO:tasks.rgw_multi.zone_rados:comparing bucket=cyjkqa-11 zones={a2, a1} 2026-03-20T18:03:13.679 DEBUG:tasks.rgw_multi.zone_rados:bucket1 objects: 2026-03-20T18:03:13.679 DEBUG:tasks.rgw_multi.zone_rados:o=obj1, v=null 2026-03-20T18:03:13.679 DEBUG:tasks.rgw_multi.zone_rados:o=obj10, v=null 2026-03-20T18:03:13.680 DEBUG:tasks.rgw_multi.zone_rados:o=obj11, v=null 2026-03-20T18:03:13.680 DEBUG:tasks.rgw_multi.zone_rados:o=obj12, v=null 2026-03-20T18:03:13.680 DEBUG:tasks.rgw_multi.zone_rados:o=obj13, v=null 2026-03-20T18:03:13.680 DEBUG:tasks.rgw_multi.zone_rados:o=obj14, v=null 2026-03-20T18:03:13.680 DEBUG:tasks.rgw_multi.zone_rados:o=obj15, v=null 2026-03-20T18:03:13.680 DEBUG:tasks.rgw_multi.zone_rados:o=obj16, v=null 2026-03-20T18:03:13.680 DEBUG:tasks.rgw_multi.zone_rados:o=obj17, v=null 2026-03-20T18:03:13.680 DEBUG:tasks.rgw_multi.zone_rados:o=obj18, v=null 2026-03-20T18:03:13.680 DEBUG:tasks.rgw_multi.zone_rados:o=obj19, v=null 2026-03-20T18:03:13.680 DEBUG:tasks.rgw_multi.zone_rados:o=obj2, v=null 2026-03-20T18:03:13.680 DEBUG:tasks.rgw_multi.zone_rados:o=obj20, v=null 2026-03-20T18:03:13.680 DEBUG:tasks.rgw_multi.zone_rados:o=obj21, v=null 2026-03-20T18:03:13.680 DEBUG:tasks.rgw_multi.zone_rados:o=obj22, v=null 2026-03-20T18:03:13.680 DEBUG:tasks.rgw_multi.zone_rados:o=obj23, v=null 2026-03-20T18:03:13.680 DEBUG:tasks.rgw_multi.zone_rados:o=obj24, v=null 2026-03-20T18:03:13.680 DEBUG:tasks.rgw_multi.zone_rados:o=obj25, v=null 2026-03-20T18:03:13.680 DEBUG:tasks.rgw_multi.zone_rados:o=obj26, v=null 2026-03-20T18:03:13.680 DEBUG:tasks.rgw_multi.zone_rados:o=obj27, v=null 2026-03-20T18:03:13.680 DEBUG:tasks.rgw_multi.zone_rados:o=obj28, v=null 2026-03-20T18:03:13.680 DEBUG:tasks.rgw_multi.zone_rados:o=obj29, v=null 2026-03-20T18:03:13.680 DEBUG:tasks.rgw_multi.zone_rados:o=obj3, v=null 2026-03-20T18:03:13.680 DEBUG:tasks.rgw_multi.zone_rados:o=obj30, v=null 2026-03-20T18:03:13.680 DEBUG:tasks.rgw_multi.zone_rados:o=obj31, v=null 2026-03-20T18:03:13.680 DEBUG:tasks.rgw_multi.zone_rados:o=obj32, v=null 2026-03-20T18:03:13.680 DEBUG:tasks.rgw_multi.zone_rados:o=obj33, v=null 2026-03-20T18:03:13.680 DEBUG:tasks.rgw_multi.zone_rados:o=obj34, v=null 2026-03-20T18:03:13.680 DEBUG:tasks.rgw_multi.zone_rados:o=obj35, v=null 2026-03-20T18:03:13.680 DEBUG:tasks.rgw_multi.zone_rados:o=obj36, v=null 2026-03-20T18:03:13.680 DEBUG:tasks.rgw_multi.zone_rados:o=obj37, v=null 2026-03-20T18:03:13.680 DEBUG:tasks.rgw_multi.zone_rados:o=obj38, v=null 2026-03-20T18:03:13.680 DEBUG:tasks.rgw_multi.zone_rados:o=obj39, v=null 2026-03-20T18:03:13.680 DEBUG:tasks.rgw_multi.zone_rados:o=obj4, v=null 2026-03-20T18:03:13.680 DEBUG:tasks.rgw_multi.zone_rados:o=obj40, v=null 2026-03-20T18:03:13.680 DEBUG:tasks.rgw_multi.zone_rados:o=obj41, v=null 2026-03-20T18:03:13.680 DEBUG:tasks.rgw_multi.zone_rados:o=obj42, v=null 2026-03-20T18:03:13.680 DEBUG:tasks.rgw_multi.zone_rados:o=obj43, v=null 2026-03-20T18:03:13.680 DEBUG:tasks.rgw_multi.zone_rados:o=obj44, v=null 2026-03-20T18:03:13.680 DEBUG:tasks.rgw_multi.zone_rados:o=obj45, v=null 2026-03-20T18:03:13.680 DEBUG:tasks.rgw_multi.zone_rados:o=obj46, v=null 2026-03-20T18:03:13.680 DEBUG:tasks.rgw_multi.zone_rados:o=obj47, v=null 2026-03-20T18:03:13.680 DEBUG:tasks.rgw_multi.zone_rados:o=obj48, v=null 2026-03-20T18:03:13.680 DEBUG:tasks.rgw_multi.zone_rados:o=obj49, v=null 2026-03-20T18:03:13.680 DEBUG:tasks.rgw_multi.zone_rados:o=obj5, v=null 2026-03-20T18:03:13.680 DEBUG:tasks.rgw_multi.zone_rados:o=obj6, v=null 2026-03-20T18:03:13.680 DEBUG:tasks.rgw_multi.zone_rados:o=obj7, v=null 2026-03-20T18:03:13.680 DEBUG:tasks.rgw_multi.zone_rados:o=obj8, v=null 2026-03-20T18:03:13.680 DEBUG:tasks.rgw_multi.zone_rados:o=obj9, v=null 2026-03-20T18:03:13.688 DEBUG:tasks.rgw_multi.zone_rados:bucket2 objects: 2026-03-20T18:03:13.688 DEBUG:tasks.rgw_multi.zone_rados:o=obj1, v=null 2026-03-20T18:03:13.688 DEBUG:tasks.rgw_multi.zone_rados:o=obj10, v=null 2026-03-20T18:03:13.688 DEBUG:tasks.rgw_multi.zone_rados:o=obj11, v=null 2026-03-20T18:03:13.688 DEBUG:tasks.rgw_multi.zone_rados:o=obj12, v=null 2026-03-20T18:03:13.688 DEBUG:tasks.rgw_multi.zone_rados:o=obj13, v=null 2026-03-20T18:03:13.688 DEBUG:tasks.rgw_multi.zone_rados:o=obj14, v=null 2026-03-20T18:03:13.688 DEBUG:tasks.rgw_multi.zone_rados:o=obj15, v=null 2026-03-20T18:03:13.688 DEBUG:tasks.rgw_multi.zone_rados:o=obj16, v=null 2026-03-20T18:03:13.688 DEBUG:tasks.rgw_multi.zone_rados:o=obj17, v=null 2026-03-20T18:03:13.688 DEBUG:tasks.rgw_multi.zone_rados:o=obj18, v=null 2026-03-20T18:03:13.688 DEBUG:tasks.rgw_multi.zone_rados:o=obj19, v=null 2026-03-20T18:03:13.688 DEBUG:tasks.rgw_multi.zone_rados:o=obj2, v=null 2026-03-20T18:03:13.688 DEBUG:tasks.rgw_multi.zone_rados:o=obj20, v=null 2026-03-20T18:03:13.688 DEBUG:tasks.rgw_multi.zone_rados:o=obj21, v=null 2026-03-20T18:03:13.688 DEBUG:tasks.rgw_multi.zone_rados:o=obj22, v=null 2026-03-20T18:03:13.688 DEBUG:tasks.rgw_multi.zone_rados:o=obj23, v=null 2026-03-20T18:03:13.688 DEBUG:tasks.rgw_multi.zone_rados:o=obj24, v=null 2026-03-20T18:03:13.688 DEBUG:tasks.rgw_multi.zone_rados:o=obj25, v=null 2026-03-20T18:03:13.688 DEBUG:tasks.rgw_multi.zone_rados:o=obj26, v=null 2026-03-20T18:03:13.688 DEBUG:tasks.rgw_multi.zone_rados:o=obj27, v=null 2026-03-20T18:03:13.688 DEBUG:tasks.rgw_multi.zone_rados:o=obj28, v=null 2026-03-20T18:03:13.688 DEBUG:tasks.rgw_multi.zone_rados:o=obj29, v=null 2026-03-20T18:03:13.688 DEBUG:tasks.rgw_multi.zone_rados:o=obj3, v=null 2026-03-20T18:03:13.688 DEBUG:tasks.rgw_multi.zone_rados:o=obj30, v=null 2026-03-20T18:03:13.688 DEBUG:tasks.rgw_multi.zone_rados:o=obj31, v=null 2026-03-20T18:03:13.688 DEBUG:tasks.rgw_multi.zone_rados:o=obj32, v=null 2026-03-20T18:03:13.688 DEBUG:tasks.rgw_multi.zone_rados:o=obj33, v=null 2026-03-20T18:03:13.688 DEBUG:tasks.rgw_multi.zone_rados:o=obj34, v=null 2026-03-20T18:03:13.688 DEBUG:tasks.rgw_multi.zone_rados:o=obj35, v=null 2026-03-20T18:03:13.688 DEBUG:tasks.rgw_multi.zone_rados:o=obj36, v=null 2026-03-20T18:03:13.688 DEBUG:tasks.rgw_multi.zone_rados:o=obj37, v=null 2026-03-20T18:03:13.688 DEBUG:tasks.rgw_multi.zone_rados:o=obj38, v=null 2026-03-20T18:03:13.688 DEBUG:tasks.rgw_multi.zone_rados:o=obj39, v=null 2026-03-20T18:03:13.688 DEBUG:tasks.rgw_multi.zone_rados:o=obj4, v=null 2026-03-20T18:03:13.688 DEBUG:tasks.rgw_multi.zone_rados:o=obj40, v=null 2026-03-20T18:03:13.688 DEBUG:tasks.rgw_multi.zone_rados:o=obj41, v=null 2026-03-20T18:03:13.689 DEBUG:tasks.rgw_multi.zone_rados:o=obj42, v=null 2026-03-20T18:03:13.689 DEBUG:tasks.rgw_multi.zone_rados:o=obj43, v=null 2026-03-20T18:03:13.689 DEBUG:tasks.rgw_multi.zone_rados:o=obj44, v=null 2026-03-20T18:03:13.689 DEBUG:tasks.rgw_multi.zone_rados:o=obj45, v=null 2026-03-20T18:03:13.689 DEBUG:tasks.rgw_multi.zone_rados:o=obj46, v=null 2026-03-20T18:03:13.689 DEBUG:tasks.rgw_multi.zone_rados:o=obj47, v=null 2026-03-20T18:03:13.689 DEBUG:tasks.rgw_multi.zone_rados:o=obj48, v=null 2026-03-20T18:03:13.689 DEBUG:tasks.rgw_multi.zone_rados:o=obj49, v=null 2026-03-20T18:03:13.689 DEBUG:tasks.rgw_multi.zone_rados:o=obj5, v=null 2026-03-20T18:03:13.689 DEBUG:tasks.rgw_multi.zone_rados:o=obj6, v=null 2026-03-20T18:03:13.689 DEBUG:tasks.rgw_multi.zone_rados:o=obj7, v=null 2026-03-20T18:03:13.689 DEBUG:tasks.rgw_multi.zone_rados:o=obj8, v=null 2026-03-20T18:03:13.689 DEBUG:tasks.rgw_multi.zone_rados:o=obj9, v=null 2026-03-20T18:03:13.689 DEBUG:tasks.rgw_multi.zone_rados:comparing key name=obj1 2026-03-20T18:03:13.768 DEBUG:tasks.rgw_multi.zone_rados:comparing key name=obj10 2026-03-20T18:03:13.833 DEBUG:tasks.rgw_multi.zone_rados:comparing key name=obj11 2026-03-20T18:03:13.895 DEBUG:tasks.rgw_multi.zone_rados:comparing key name=obj12 2026-03-20T18:03:13.956 DEBUG:tasks.rgw_multi.zone_rados:comparing key name=obj13 2026-03-20T18:03:14.041 DEBUG:tasks.rgw_multi.zone_rados:comparing key name=obj14 2026-03-20T18:03:14.106 DEBUG:tasks.rgw_multi.zone_rados:comparing key name=obj15 2026-03-20T18:03:14.189 DEBUG:tasks.rgw_multi.zone_rados:comparing key name=obj16 2026-03-20T18:03:14.249 DEBUG:tasks.rgw_multi.zone_rados:comparing key name=obj17 2026-03-20T18:03:14.310 DEBUG:tasks.rgw_multi.zone_rados:comparing key name=obj18 2026-03-20T18:03:14.372 DEBUG:tasks.rgw_multi.zone_rados:comparing key name=obj19 2026-03-20T18:03:14.433 DEBUG:tasks.rgw_multi.zone_rados:comparing key name=obj2 2026-03-20T18:03:14.558 DEBUG:tasks.rgw_multi.zone_rados:comparing key name=obj20 2026-03-20T18:03:14.622 DEBUG:tasks.rgw_multi.zone_rados:comparing key name=obj21 2026-03-20T18:03:14.679 DEBUG:tasks.rgw_multi.zone_rados:comparing key name=obj22 2026-03-20T18:03:14.741 DEBUG:tasks.rgw_multi.zone_rados:comparing key name=obj23 2026-03-20T18:03:14.803 DEBUG:tasks.rgw_multi.zone_rados:comparing key name=obj24 2026-03-20T18:03:14.863 DEBUG:tasks.rgw_multi.zone_rados:comparing key name=obj25 2026-03-20T18:03:14.931 DEBUG:tasks.rgw_multi.zone_rados:comparing key name=obj26 2026-03-20T18:03:14.992 DEBUG:tasks.rgw_multi.zone_rados:comparing key name=obj27 2026-03-20T18:03:15.054 DEBUG:tasks.rgw_multi.zone_rados:comparing key name=obj28 2026-03-20T18:03:15.114 DEBUG:tasks.rgw_multi.zone_rados:comparing key name=obj29 2026-03-20T18:03:15.214 DEBUG:tasks.rgw_multi.zone_rados:comparing key name=obj3 2026-03-20T18:03:15.277 DEBUG:tasks.rgw_multi.zone_rados:comparing key name=obj30 2026-03-20T18:03:15.338 DEBUG:tasks.rgw_multi.zone_rados:comparing key name=obj31 2026-03-20T18:03:15.401 DEBUG:tasks.rgw_multi.zone_rados:comparing key name=obj32 2026-03-20T18:03:15.461 DEBUG:tasks.rgw_multi.zone_rados:comparing key name=obj33 2026-03-20T18:03:15.522 DEBUG:tasks.rgw_multi.zone_rados:comparing key name=obj34 2026-03-20T18:03:15.582 DEBUG:tasks.rgw_multi.zone_rados:comparing key name=obj35 2026-03-20T18:03:15.647 DEBUG:tasks.rgw_multi.zone_rados:comparing key name=obj36 2026-03-20T18:03:15.704 DEBUG:tasks.rgw_multi.zone_rados:comparing key name=obj37 2026-03-20T18:03:15.768 DEBUG:tasks.rgw_multi.zone_rados:comparing key name=obj38 2026-03-20T18:03:15.831 DEBUG:tasks.rgw_multi.zone_rados:comparing key name=obj39 2026-03-20T18:03:15.894 DEBUG:tasks.rgw_multi.zone_rados:comparing key name=obj4 2026-03-20T18:03:15.960 DEBUG:tasks.rgw_multi.zone_rados:comparing key name=obj40 2026-03-20T18:03:16.022 DEBUG:tasks.rgw_multi.zone_rados:comparing key name=obj41 2026-03-20T18:03:16.081 DEBUG:tasks.rgw_multi.zone_rados:comparing key name=obj42 2026-03-20T18:03:16.135 DEBUG:tasks.rgw_multi.zone_rados:comparing key name=obj43 2026-03-20T18:03:16.198 DEBUG:tasks.rgw_multi.zone_rados:comparing key name=obj44 2026-03-20T18:03:16.255 DEBUG:tasks.rgw_multi.zone_rados:comparing key name=obj45 2026-03-20T18:03:16.313 DEBUG:tasks.rgw_multi.zone_rados:comparing key name=obj46 2026-03-20T18:03:16.366 DEBUG:tasks.rgw_multi.zone_rados:comparing key name=obj47 2026-03-20T18:03:16.423 DEBUG:tasks.rgw_multi.zone_rados:comparing key name=obj48 2026-03-20T18:03:16.479 DEBUG:tasks.rgw_multi.zone_rados:comparing key name=obj49 2026-03-20T18:03:16.538 DEBUG:tasks.rgw_multi.zone_rados:comparing key name=obj5 2026-03-20T18:03:16.594 DEBUG:tasks.rgw_multi.zone_rados:comparing key name=obj6 2026-03-20T18:03:16.651 DEBUG:tasks.rgw_multi.zone_rados:comparing key name=obj7 2026-03-20T18:03:16.707 DEBUG:tasks.rgw_multi.zone_rados:comparing key name=obj8 2026-03-20T18:03:16.762 DEBUG:tasks.rgw_multi.zone_rados:comparing key name=obj9 2026-03-20T18:03:16.823 INFO:tasks.rgw_multi.zone_rados:success, bucket identical: bucket=cyjkqa-11 zones={a2, a1} 2026-03-20T18:03:16.823 INFO:tasks.util.rgw:rgwadmin: c1.client.0 : ['bucket', 'sync', 'checkpoint', '--bucket', 'cyjkqa-12', '--source-zone', 'a2', '--retry-delay-ms', '5000', '--timeout-sec', '300', '--rgw-zone', 'a1', '--rgw-zonegroup', 'a', '--rgw-realm', 'test-realm', '--cluster', 'c1', '--debug-rgw', '1', '--debug-ms', '0'] 2026-03-20T18:03:16.823 DEBUG:tasks.util.rgw:rgwadmin: cmd=['adjust-ulimits', 'ceph-coverage', '/home/ubuntu/cephtest/archive/coverage', 'radosgw-admin', '--log-to-stderr', '--format', 'json', '-n', 'client.0', '--cluster', 'c1', 'bucket', 'sync', 'checkpoint', '--bucket', 'cyjkqa-12', '--source-zone', 'a2', '--retry-delay-ms', '5000', '--timeout-sec', '300', '--rgw-zone', 'a1', '--rgw-zonegroup', 'a', '--rgw-realm', 'test-realm', '--cluster', 'c1', '--debug-rgw', '1', '--debug-ms', '0'] 2026-03-20T18:03:16.824 DEBUG:teuthology.orchestra.run.vm02:> adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage radosgw-admin --log-to-stderr --format json -n client.0 --cluster c1 bucket sync checkpoint --bucket cyjkqa-12 --source-zone a2 --retry-delay-ms 5000 --timeout-sec 300 --rgw-zone a1 --rgw-zonegroup a --rgw-realm test-realm --cluster c1 --debug-rgw 1 --debug-ms 0 2026-03-20T18:03:16.849 INFO:teuthology.orchestra.run.vm02.stderr:ignoring --setuser ceph since I am not root 2026-03-20T18:03:16.849 INFO:teuthology.orchestra.run.vm02.stderr:ignoring --setgroup ceph since I am not root 2026-03-20T18:03:16.926 INFO:teuthology.orchestra.run.vm02.stderr:2026-03-20T18:03:16.921+0000 7fb7510fca80 1 bucket sync caught up with source: 2026-03-20T18:03:16.926 INFO:teuthology.orchestra.run.vm02.stderr: local status: [00000000007.28.6, 00000000004.17.6, 00000000004.20.6, 00000000004.16.6, 00000000006.27.6, 00000000005.18.6, 00000000004.16.6, 00000000005.24.6, 00000000002.36.6, 00000000004.32.6, 00000000004.20.6] 2026-03-20T18:03:16.926 INFO:teuthology.orchestra.run.vm02.stderr: remote markers: [00000000007.28.6, 00000000004.17.6, 00000000004.20.6, 00000000004.16.6, 00000000006.27.6, 00000000005.18.6, 00000000004.16.6, 00000000005.24.6, 00000000002.36.6, 00000000004.32.6, 00000000004.20.6] 2026-03-20T18:03:16.926 INFO:teuthology.orchestra.run.vm02.stderr:2026-03-20T18:03:16.921+0000 7fb7510fca80 0 bucket checkpoint complete 2026-03-20T18:03:16.935 INFO:tasks.rgw_multi.zone_rados:comparing bucket=cyjkqa-12 zones={a1, a2} 2026-03-20T18:03:16.946 DEBUG:tasks.rgw_multi.zone_rados:bucket1 objects: 2026-03-20T18:03:16.946 DEBUG:tasks.rgw_multi.zone_rados:o=obj1, v=null 2026-03-20T18:03:16.946 DEBUG:tasks.rgw_multi.zone_rados:o=obj10, v=null 2026-03-20T18:03:16.946 DEBUG:tasks.rgw_multi.zone_rados:o=obj11, v=null 2026-03-20T18:03:16.946 DEBUG:tasks.rgw_multi.zone_rados:o=obj12, v=null 2026-03-20T18:03:16.946 DEBUG:tasks.rgw_multi.zone_rados:o=obj13, v=null 2026-03-20T18:03:16.946 DEBUG:tasks.rgw_multi.zone_rados:o=obj14, v=null 2026-03-20T18:03:16.946 DEBUG:tasks.rgw_multi.zone_rados:o=obj15, v=null 2026-03-20T18:03:16.946 DEBUG:tasks.rgw_multi.zone_rados:o=obj16, v=null 2026-03-20T18:03:16.946 DEBUG:tasks.rgw_multi.zone_rados:o=obj17, v=null 2026-03-20T18:03:16.946 DEBUG:tasks.rgw_multi.zone_rados:o=obj18, v=null 2026-03-20T18:03:16.946 DEBUG:tasks.rgw_multi.zone_rados:o=obj19, v=null 2026-03-20T18:03:16.946 DEBUG:tasks.rgw_multi.zone_rados:o=obj2, v=null 2026-03-20T18:03:16.946 DEBUG:tasks.rgw_multi.zone_rados:o=obj20, v=null 2026-03-20T18:03:16.946 DEBUG:tasks.rgw_multi.zone_rados:o=obj21, v=null 2026-03-20T18:03:16.946 DEBUG:tasks.rgw_multi.zone_rados:o=obj22, v=null 2026-03-20T18:03:16.946 DEBUG:tasks.rgw_multi.zone_rados:o=obj23, v=null 2026-03-20T18:03:16.946 DEBUG:tasks.rgw_multi.zone_rados:o=obj24, v=null 2026-03-20T18:03:16.946 DEBUG:tasks.rgw_multi.zone_rados:o=obj25, v=null 2026-03-20T18:03:16.946 DEBUG:tasks.rgw_multi.zone_rados:o=obj26, v=null 2026-03-20T18:03:16.946 DEBUG:tasks.rgw_multi.zone_rados:o=obj27, v=null 2026-03-20T18:03:16.946 DEBUG:tasks.rgw_multi.zone_rados:o=obj28, v=null 2026-03-20T18:03:16.947 DEBUG:tasks.rgw_multi.zone_rados:o=obj29, v=null 2026-03-20T18:03:16.947 DEBUG:tasks.rgw_multi.zone_rados:o=obj3, v=null 2026-03-20T18:03:16.947 DEBUG:tasks.rgw_multi.zone_rados:o=obj30, v=null 2026-03-20T18:03:16.947 DEBUG:tasks.rgw_multi.zone_rados:o=obj31, v=null 2026-03-20T18:03:16.947 DEBUG:tasks.rgw_multi.zone_rados:o=obj32, v=null 2026-03-20T18:03:16.947 DEBUG:tasks.rgw_multi.zone_rados:o=obj33, v=null 2026-03-20T18:03:16.947 DEBUG:tasks.rgw_multi.zone_rados:o=obj34, v=null 2026-03-20T18:03:16.947 DEBUG:tasks.rgw_multi.zone_rados:o=obj35, v=null 2026-03-20T18:03:16.947 DEBUG:tasks.rgw_multi.zone_rados:o=obj36, v=null 2026-03-20T18:03:16.947 DEBUG:tasks.rgw_multi.zone_rados:o=obj37, v=null 2026-03-20T18:03:16.947 DEBUG:tasks.rgw_multi.zone_rados:o=obj38, v=null 2026-03-20T18:03:16.947 DEBUG:tasks.rgw_multi.zone_rados:o=obj39, v=null 2026-03-20T18:03:16.947 DEBUG:tasks.rgw_multi.zone_rados:o=obj4, v=null 2026-03-20T18:03:16.947 DEBUG:tasks.rgw_multi.zone_rados:o=obj40, v=null 2026-03-20T18:03:16.947 DEBUG:tasks.rgw_multi.zone_rados:o=obj41, v=null 2026-03-20T18:03:16.947 DEBUG:tasks.rgw_multi.zone_rados:o=obj42, v=null 2026-03-20T18:03:16.947 DEBUG:tasks.rgw_multi.zone_rados:o=obj43, v=null 2026-03-20T18:03:16.947 DEBUG:tasks.rgw_multi.zone_rados:o=obj44, v=null 2026-03-20T18:03:16.947 DEBUG:tasks.rgw_multi.zone_rados:o=obj45, v=null 2026-03-20T18:03:16.947 DEBUG:tasks.rgw_multi.zone_rados:o=obj46, v=null 2026-03-20T18:03:16.947 DEBUG:tasks.rgw_multi.zone_rados:o=obj47, v=null 2026-03-20T18:03:16.947 DEBUG:tasks.rgw_multi.zone_rados:o=obj48, v=null 2026-03-20T18:03:16.947 DEBUG:tasks.rgw_multi.zone_rados:o=obj49, v=null 2026-03-20T18:03:16.947 DEBUG:tasks.rgw_multi.zone_rados:o=obj5, v=null 2026-03-20T18:03:16.947 DEBUG:tasks.rgw_multi.zone_rados:o=obj6, v=null 2026-03-20T18:03:16.947 DEBUG:tasks.rgw_multi.zone_rados:o=obj7, v=null 2026-03-20T18:03:16.947 DEBUG:tasks.rgw_multi.zone_rados:o=obj8, v=null 2026-03-20T18:03:16.947 DEBUG:tasks.rgw_multi.zone_rados:o=obj9, v=null 2026-03-20T18:03:16.953 DEBUG:tasks.rgw_multi.zone_rados:bucket2 objects: 2026-03-20T18:03:16.953 DEBUG:tasks.rgw_multi.zone_rados:o=obj1, v=null 2026-03-20T18:03:16.953 DEBUG:tasks.rgw_multi.zone_rados:o=obj10, v=null 2026-03-20T18:03:16.953 DEBUG:tasks.rgw_multi.zone_rados:o=obj11, v=null 2026-03-20T18:03:16.953 DEBUG:tasks.rgw_multi.zone_rados:o=obj12, v=null 2026-03-20T18:03:16.953 DEBUG:tasks.rgw_multi.zone_rados:o=obj13, v=null 2026-03-20T18:03:16.953 DEBUG:tasks.rgw_multi.zone_rados:o=obj14, v=null 2026-03-20T18:03:16.953 DEBUG:tasks.rgw_multi.zone_rados:o=obj15, v=null 2026-03-20T18:03:16.953 DEBUG:tasks.rgw_multi.zone_rados:o=obj16, v=null 2026-03-20T18:03:16.953 DEBUG:tasks.rgw_multi.zone_rados:o=obj17, v=null 2026-03-20T18:03:16.953 DEBUG:tasks.rgw_multi.zone_rados:o=obj18, v=null 2026-03-20T18:03:16.953 DEBUG:tasks.rgw_multi.zone_rados:o=obj19, v=null 2026-03-20T18:03:16.953 DEBUG:tasks.rgw_multi.zone_rados:o=obj2, v=null 2026-03-20T18:03:16.953 DEBUG:tasks.rgw_multi.zone_rados:o=obj20, v=null 2026-03-20T18:03:16.953 DEBUG:tasks.rgw_multi.zone_rados:o=obj21, v=null 2026-03-20T18:03:16.953 DEBUG:tasks.rgw_multi.zone_rados:o=obj22, v=null 2026-03-20T18:03:16.953 DEBUG:tasks.rgw_multi.zone_rados:o=obj23, v=null 2026-03-20T18:03:16.953 DEBUG:tasks.rgw_multi.zone_rados:o=obj24, v=null 2026-03-20T18:03:16.953 DEBUG:tasks.rgw_multi.zone_rados:o=obj25, v=null 2026-03-20T18:03:16.953 DEBUG:tasks.rgw_multi.zone_rados:o=obj26, v=null 2026-03-20T18:03:16.953 DEBUG:tasks.rgw_multi.zone_rados:o=obj27, v=null 2026-03-20T18:03:16.953 DEBUG:tasks.rgw_multi.zone_rados:o=obj28, v=null 2026-03-20T18:03:16.953 DEBUG:tasks.rgw_multi.zone_rados:o=obj29, v=null 2026-03-20T18:03:16.953 DEBUG:tasks.rgw_multi.zone_rados:o=obj3, v=null 2026-03-20T18:03:16.953 DEBUG:tasks.rgw_multi.zone_rados:o=obj30, v=null 2026-03-20T18:03:16.953 DEBUG:tasks.rgw_multi.zone_rados:o=obj31, v=null 2026-03-20T18:03:16.953 DEBUG:tasks.rgw_multi.zone_rados:o=obj32, v=null 2026-03-20T18:03:16.953 DEBUG:tasks.rgw_multi.zone_rados:o=obj33, v=null 2026-03-20T18:03:16.953 DEBUG:tasks.rgw_multi.zone_rados:o=obj34, v=null 2026-03-20T18:03:16.953 DEBUG:tasks.rgw_multi.zone_rados:o=obj35, v=null 2026-03-20T18:03:16.953 DEBUG:tasks.rgw_multi.zone_rados:o=obj36, v=null 2026-03-20T18:03:16.953 DEBUG:tasks.rgw_multi.zone_rados:o=obj37, v=null 2026-03-20T18:03:16.953 DEBUG:tasks.rgw_multi.zone_rados:o=obj38, v=null 2026-03-20T18:03:16.953 DEBUG:tasks.rgw_multi.zone_rados:o=obj39, v=null 2026-03-20T18:03:16.953 DEBUG:tasks.rgw_multi.zone_rados:o=obj4, v=null 2026-03-20T18:03:16.953 DEBUG:tasks.rgw_multi.zone_rados:o=obj40, v=null 2026-03-20T18:03:16.953 DEBUG:tasks.rgw_multi.zone_rados:o=obj41, v=null 2026-03-20T18:03:16.953 DEBUG:tasks.rgw_multi.zone_rados:o=obj42, v=null 2026-03-20T18:03:16.953 DEBUG:tasks.rgw_multi.zone_rados:o=obj43, v=null 2026-03-20T18:03:16.953 DEBUG:tasks.rgw_multi.zone_rados:o=obj44, v=null 2026-03-20T18:03:16.953 DEBUG:tasks.rgw_multi.zone_rados:o=obj45, v=null 2026-03-20T18:03:16.953 DEBUG:tasks.rgw_multi.zone_rados:o=obj46, v=null 2026-03-20T18:03:16.953 DEBUG:tasks.rgw_multi.zone_rados:o=obj47, v=null 2026-03-20T18:03:16.953 DEBUG:tasks.rgw_multi.zone_rados:o=obj48, v=null 2026-03-20T18:03:16.953 DEBUG:tasks.rgw_multi.zone_rados:o=obj49, v=null 2026-03-20T18:03:16.953 DEBUG:tasks.rgw_multi.zone_rados:o=obj5, v=null 2026-03-20T18:03:16.953 DEBUG:tasks.rgw_multi.zone_rados:o=obj6, v=null 2026-03-20T18:03:16.953 DEBUG:tasks.rgw_multi.zone_rados:o=obj7, v=null 2026-03-20T18:03:16.953 DEBUG:tasks.rgw_multi.zone_rados:o=obj8, v=null 2026-03-20T18:03:16.953 DEBUG:tasks.rgw_multi.zone_rados:o=obj9, v=null 2026-03-20T18:03:16.953 DEBUG:tasks.rgw_multi.zone_rados:comparing key name=obj1 2026-03-20T18:03:17.011 DEBUG:tasks.rgw_multi.zone_rados:comparing key name=obj10 2026-03-20T18:03:17.067 DEBUG:tasks.rgw_multi.zone_rados:comparing key name=obj11 2026-03-20T18:03:17.123 DEBUG:tasks.rgw_multi.zone_rados:comparing key name=obj12 2026-03-20T18:03:17.184 DEBUG:tasks.rgw_multi.zone_rados:comparing key name=obj13 2026-03-20T18:03:17.243 DEBUG:tasks.rgw_multi.zone_rados:comparing key name=obj14 2026-03-20T18:03:17.302 DEBUG:tasks.rgw_multi.zone_rados:comparing key name=obj15 2026-03-20T18:03:17.362 DEBUG:tasks.rgw_multi.zone_rados:comparing key name=obj16 2026-03-20T18:03:17.414 DEBUG:tasks.rgw_multi.zone_rados:comparing key name=obj17 2026-03-20T18:03:17.470 DEBUG:tasks.rgw_multi.zone_rados:comparing key name=obj18 2026-03-20T18:03:17.526 DEBUG:tasks.rgw_multi.zone_rados:comparing key name=obj19 2026-03-20T18:03:17.583 DEBUG:tasks.rgw_multi.zone_rados:comparing key name=obj2 2026-03-20T18:03:17.640 DEBUG:tasks.rgw_multi.zone_rados:comparing key name=obj20 2026-03-20T18:03:17.699 DEBUG:tasks.rgw_multi.zone_rados:comparing key name=obj21 2026-03-20T18:03:17.755 DEBUG:tasks.rgw_multi.zone_rados:comparing key name=obj22 2026-03-20T18:03:17.812 DEBUG:tasks.rgw_multi.zone_rados:comparing key name=obj23 2026-03-20T18:03:17.868 DEBUG:tasks.rgw_multi.zone_rados:comparing key name=obj24 2026-03-20T18:03:17.924 DEBUG:tasks.rgw_multi.zone_rados:comparing key name=obj25 2026-03-20T18:03:17.986 DEBUG:tasks.rgw_multi.zone_rados:comparing key name=obj26 2026-03-20T18:03:18.053 DEBUG:tasks.rgw_multi.zone_rados:comparing key name=obj27 2026-03-20T18:03:18.120 DEBUG:tasks.rgw_multi.zone_rados:comparing key name=obj28 2026-03-20T18:03:18.185 DEBUG:tasks.rgw_multi.zone_rados:comparing key name=obj29 2026-03-20T18:03:18.248 DEBUG:tasks.rgw_multi.zone_rados:comparing key name=obj3 2026-03-20T18:03:18.311 DEBUG:tasks.rgw_multi.zone_rados:comparing key name=obj30 2026-03-20T18:03:18.372 DEBUG:tasks.rgw_multi.zone_rados:comparing key name=obj31 2026-03-20T18:03:18.435 DEBUG:tasks.rgw_multi.zone_rados:comparing key name=obj32 2026-03-20T18:03:18.499 DEBUG:tasks.rgw_multi.zone_rados:comparing key name=obj33 2026-03-20T18:03:18.564 DEBUG:tasks.rgw_multi.zone_rados:comparing key name=obj34 2026-03-20T18:03:18.620 DEBUG:tasks.rgw_multi.zone_rados:comparing key name=obj35 2026-03-20T18:03:18.681 DEBUG:tasks.rgw_multi.zone_rados:comparing key name=obj36 2026-03-20T18:03:18.742 DEBUG:tasks.rgw_multi.zone_rados:comparing key name=obj37 2026-03-20T18:03:18.806 DEBUG:tasks.rgw_multi.zone_rados:comparing key name=obj38 2026-03-20T18:03:18.869 DEBUG:tasks.rgw_multi.zone_rados:comparing key name=obj39 2026-03-20T18:03:18.934 DEBUG:tasks.rgw_multi.zone_rados:comparing key name=obj4 2026-03-20T18:03:19.001 DEBUG:tasks.rgw_multi.zone_rados:comparing key name=obj40 2026-03-20T18:03:19.064 DEBUG:tasks.rgw_multi.zone_rados:comparing key name=obj41 2026-03-20T18:03:19.128 DEBUG:tasks.rgw_multi.zone_rados:comparing key name=obj42 2026-03-20T18:03:19.189 DEBUG:tasks.rgw_multi.zone_rados:comparing key name=obj43 2026-03-20T18:03:19.299 DEBUG:tasks.rgw_multi.zone_rados:comparing key name=obj44 2026-03-20T18:03:19.355 DEBUG:tasks.rgw_multi.zone_rados:comparing key name=obj45 2026-03-20T18:03:19.414 DEBUG:tasks.rgw_multi.zone_rados:comparing key name=obj46 2026-03-20T18:03:19.476 DEBUG:tasks.rgw_multi.zone_rados:comparing key name=obj47 2026-03-20T18:03:19.532 DEBUG:tasks.rgw_multi.zone_rados:comparing key name=obj48 2026-03-20T18:03:19.587 DEBUG:tasks.rgw_multi.zone_rados:comparing key name=obj49 2026-03-20T18:03:19.643 DEBUG:tasks.rgw_multi.zone_rados:comparing key name=obj5 2026-03-20T18:03:19.704 DEBUG:tasks.rgw_multi.zone_rados:comparing key name=obj6 2026-03-20T18:03:19.766 DEBUG:tasks.rgw_multi.zone_rados:comparing key name=obj7 2026-03-20T18:03:19.823 DEBUG:tasks.rgw_multi.zone_rados:comparing key name=obj8 2026-03-20T18:03:19.880 DEBUG:tasks.rgw_multi.zone_rados:comparing key name=obj9 2026-03-20T18:03:19.943 INFO:tasks.rgw_multi.zone_rados:success, bucket identical: bucket=cyjkqa-12 zones={a1, a2} 2026-03-20T18:03:19.982 INFO:tasks.util.rgw:rgwadmin: c1.client.0 : ['bucket', 'sync', 'checkpoint', '--bucket', 'cyjkqa-11', '--source-zone', 'a1', '--retry-delay-ms', '5000', '--timeout-sec', '300', '--rgw-zone', 'a2', '--rgw-zonegroup', 'a', '--rgw-realm', 'test-realm', '--cluster', 'c1', '--debug-rgw', '1', '--debug-ms', '0'] 2026-03-20T18:03:19.982 DEBUG:tasks.util.rgw:rgwadmin: cmd=['adjust-ulimits', 'ceph-coverage', '/home/ubuntu/cephtest/archive/coverage', 'radosgw-admin', '--log-to-stderr', '--format', 'json', '-n', 'client.0', '--cluster', 'c1', 'bucket', 'sync', 'checkpoint', '--bucket', 'cyjkqa-11', '--source-zone', 'a1', '--retry-delay-ms', '5000', '--timeout-sec', '300', '--rgw-zone', 'a2', '--rgw-zonegroup', 'a', '--rgw-realm', 'test-realm', '--cluster', 'c1', '--debug-rgw', '1', '--debug-ms', '0'] 2026-03-20T18:03:19.982 DEBUG:teuthology.orchestra.run.vm02:> adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage radosgw-admin --log-to-stderr --format json -n client.0 --cluster c1 bucket sync checkpoint --bucket cyjkqa-11 --source-zone a1 --retry-delay-ms 5000 --timeout-sec 300 --rgw-zone a2 --rgw-zonegroup a --rgw-realm test-realm --cluster c1 --debug-rgw 1 --debug-ms 0 2026-03-20T18:03:20.008 INFO:teuthology.orchestra.run.vm02.stderr:ignoring --setuser ceph since I am not root 2026-03-20T18:03:20.008 INFO:teuthology.orchestra.run.vm02.stderr:ignoring --setgroup ceph since I am not root 2026-03-20T18:03:20.091 INFO:teuthology.orchestra.run.vm02.stderr:2026-03-20T18:03:20.089+0000 7fbbab9eca80 1 waiting for incremental sync to catch up: 2026-03-20T18:03:20.091 INFO:teuthology.orchestra.run.vm02.stderr: local status: [00000000007.32.6, 00000000004.11.6, 00000000004.18.6, 00000000004.31.6, 00000000006.19.6, 00000000005.15.6, 00000000004.33.6, 00000000005.25.6, 00000000002.34.6, 00000000004.16.6, 00000000004.17.6] 2026-03-20T18:03:20.092 INFO:teuthology.orchestra.run.vm02.stderr: remote markers: [00000000014.51.6, 00000000008.19.6, 00000000008.40.6, 00000000008.61.6, 00000000012.31.6, 00000000010.25.6, 00000000008.60.6, 00000000010.43.6, 00000000004.52.6, 00000000008.24.6, 00000000008.25.6] 2026-03-20T18:03:25.092 INFO:teuthology.orchestra.run.vm02.stderr:2026-03-20T18:03:25.089+0000 7fbbab9eca80 1 waiting for incremental sync to catch up: 2026-03-20T18:03:25.092 INFO:teuthology.orchestra.run.vm02.stderr: local status: [00000000007.32.6, 00000000004.11.6, 00000000004.18.6, 00000000004.31.6, 00000000006.19.6, 00000000005.15.6, 00000000004.33.6, 00000000005.25.6, 00000000002.34.6, 00000000004.16.6, 00000000004.17.6] 2026-03-20T18:03:25.092 INFO:teuthology.orchestra.run.vm02.stderr: remote markers: [00000000014.51.6, 00000000008.19.6, 00000000008.40.6, 00000000008.61.6, 00000000012.31.6, 00000000010.25.6, 00000000008.60.6, 00000000010.43.6, 00000000004.52.6, 00000000008.24.6, 00000000008.25.6] 2026-03-20T18:03:30.093 INFO:teuthology.orchestra.run.vm02.stderr:2026-03-20T18:03:30.089+0000 7fbbab9eca80 1 bucket sync caught up with source: 2026-03-20T18:03:30.093 INFO:teuthology.orchestra.run.vm02.stderr: local status: [00000000014.51.6, 00000000008.19.6, 00000000008.40.6, 00000000008.61.6, 00000000012.31.6, 00000000010.25.6, 00000000008.60.6, 00000000010.43.6, 00000000004.52.6, 00000000008.24.6, 00000000008.25.6] 2026-03-20T18:03:30.093 INFO:teuthology.orchestra.run.vm02.stderr: remote markers: [00000000014.51.6, 00000000008.19.6, 00000000008.40.6, 00000000008.61.6, 00000000012.31.6, 00000000010.25.6, 00000000008.60.6, 00000000010.43.6, 00000000004.52.6, 00000000008.24.6, 00000000008.25.6] 2026-03-20T18:03:30.093 INFO:teuthology.orchestra.run.vm02.stderr:2026-03-20T18:03:30.089+0000 7fbbab9eca80 0 bucket checkpoint complete 2026-03-20T18:03:30.143 INFO:tasks.rgw_multi.zone_rados:comparing bucket=cyjkqa-11 zones={a2, a1} 2026-03-20T18:03:30.149 DEBUG:tasks.rgw_multi.zone_rados:bucket1 objects: 2026-03-20T18:03:30.151 DEBUG:tasks.rgw_multi.zone_rados:bucket2 objects: 2026-03-20T18:03:30.151 INFO:tasks.rgw_multi.zone_rados:success, bucket identical: bucket=cyjkqa-11 zones={a2, a1} 2026-03-20T18:03:30.219 INFO:tasks.util.rgw:rgwadmin: c1.client.0 : ['bucket', 'sync', 'checkpoint', '--bucket', 'cyjkqa-12', '--source-zone', 'a2', '--retry-delay-ms', '5000', '--timeout-sec', '300', '--rgw-zone', 'a1', '--rgw-zonegroup', 'a', '--rgw-realm', 'test-realm', '--cluster', 'c1', '--debug-rgw', '1', '--debug-ms', '0'] 2026-03-20T18:03:30.219 DEBUG:tasks.util.rgw:rgwadmin: cmd=['adjust-ulimits', 'ceph-coverage', '/home/ubuntu/cephtest/archive/coverage', 'radosgw-admin', '--log-to-stderr', '--format', 'json', '-n', 'client.0', '--cluster', 'c1', 'bucket', 'sync', 'checkpoint', '--bucket', 'cyjkqa-12', '--source-zone', 'a2', '--retry-delay-ms', '5000', '--timeout-sec', '300', '--rgw-zone', 'a1', '--rgw-zonegroup', 'a', '--rgw-realm', 'test-realm', '--cluster', 'c1', '--debug-rgw', '1', '--debug-ms', '0'] 2026-03-20T18:03:30.219 DEBUG:teuthology.orchestra.run.vm02:> adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage radosgw-admin --log-to-stderr --format json -n client.0 --cluster c1 bucket sync checkpoint --bucket cyjkqa-12 --source-zone a2 --retry-delay-ms 5000 --timeout-sec 300 --rgw-zone a1 --rgw-zonegroup a --rgw-realm test-realm --cluster c1 --debug-rgw 1 --debug-ms 0 2026-03-20T18:03:30.250 INFO:teuthology.orchestra.run.vm02.stderr:ignoring --setuser ceph since I am not root 2026-03-20T18:03:30.250 INFO:teuthology.orchestra.run.vm02.stderr:ignoring --setgroup ceph since I am not root 2026-03-20T18:03:30.341 INFO:teuthology.orchestra.run.vm02.stderr:2026-03-20T18:03:30.337+0000 7fc16ff95a80 1 waiting for incremental sync to catch up: 2026-03-20T18:03:30.341 INFO:teuthology.orchestra.run.vm02.stderr: local status: [00000000007.28.6, 00000000004.17.6, 00000000004.20.6, 00000000004.16.6, 00000000006.27.6, 00000000005.18.6, 00000000004.16.6, 00000000005.24.6, 00000000002.36.6, 00000000004.32.6, 00000000004.20.6] 2026-03-20T18:03:30.341 INFO:teuthology.orchestra.run.vm02.stderr: remote markers: [00000000014.58.6, 00000000008.45.6, 00000000008.28.6, 00000000008.24.6, 00000000012.39.6, 00000000010.28.6, 00000000008.24.6, 00000000010.34.6, 00000000004.80.6, 00000000008.78.6, 00000000008.28.6] 2026-03-20T18:03:35.342 INFO:teuthology.orchestra.run.vm02.stderr:2026-03-20T18:03:35.337+0000 7fc16ff95a80 1 waiting for incremental sync to catch up: 2026-03-20T18:03:35.342 INFO:teuthology.orchestra.run.vm02.stderr: local status: [00000000007.28.6, 00000000004.17.6, 00000000004.20.6, 00000000004.16.6, 00000000006.27.6, 00000000005.18.6, 00000000004.16.6, 00000000005.24.6, 00000000002.36.6, 00000000004.32.6, 00000000004.20.6] 2026-03-20T18:03:35.342 INFO:teuthology.orchestra.run.vm02.stderr: remote markers: [00000000014.58.6, 00000000008.45.6, 00000000008.28.6, 00000000008.24.6, 00000000012.39.6, 00000000010.28.6, 00000000008.24.6, 00000000010.34.6, 00000000004.80.6, 00000000008.78.6, 00000000008.28.6] 2026-03-20T18:03:40.343 INFO:teuthology.orchestra.run.vm02.stderr:2026-03-20T18:03:40.341+0000 7fc16ff95a80 1 waiting for incremental sync to catch up: 2026-03-20T18:03:40.343 INFO:teuthology.orchestra.run.vm02.stderr: local status: [00000000007.28.6, 00000000004.17.6, 00000000004.20.6, 00000000004.16.6, 00000000006.27.6, 00000000005.18.6, 00000000004.16.6, 00000000005.24.6, 00000000002.36.6, 00000000004.32.6, 00000000004.20.6] 2026-03-20T18:03:40.343 INFO:teuthology.orchestra.run.vm02.stderr: remote markers: [00000000014.58.6, 00000000008.45.6, 00000000008.28.6, 00000000008.24.6, 00000000012.39.6, 00000000010.28.6, 00000000008.24.6, 00000000010.34.6, 00000000004.80.6, 00000000008.78.6, 00000000008.28.6] 2026-03-20T18:03:45.344 INFO:teuthology.orchestra.run.vm02.stderr:2026-03-20T18:03:45.341+0000 7fc16ff95a80 1 waiting for incremental sync to catch up: 2026-03-20T18:03:45.344 INFO:teuthology.orchestra.run.vm02.stderr: local status: [00000000007.28.6, 00000000004.17.6, 00000000004.20.6, 00000000004.16.6, 00000000006.27.6, 00000000005.18.6, 00000000004.16.6, 00000000005.24.6, 00000000002.36.6, 00000000004.32.6, 00000000004.20.6] 2026-03-20T18:03:45.344 INFO:teuthology.orchestra.run.vm02.stderr: remote markers: [00000000014.58.6, 00000000008.45.6, 00000000008.28.6, 00000000008.24.6, 00000000012.39.6, 00000000010.28.6, 00000000008.24.6, 00000000010.34.6, 00000000004.80.6, 00000000008.78.6, 00000000008.28.6] 2026-03-20T18:03:50.345 INFO:teuthology.orchestra.run.vm02.stderr:2026-03-20T18:03:50.341+0000 7fc16ff95a80 1 bucket sync caught up with source: 2026-03-20T18:03:50.346 INFO:teuthology.orchestra.run.vm02.stderr: local status: [00000000014.58.6, 00000000008.45.6, 00000000008.28.6, 00000000008.24.6, 00000000012.39.6, 00000000010.28.6, 00000000008.24.6, 00000000010.34.6, 00000000004.80.6, 00000000008.78.6, 00000000008.28.6] 2026-03-20T18:03:50.346 INFO:teuthology.orchestra.run.vm02.stderr: remote markers: [00000000014.58.6, 00000000008.45.6, 00000000008.28.6, 00000000008.24.6, 00000000012.39.6, 00000000010.28.6, 00000000008.24.6, 00000000010.34.6, 00000000004.80.6, 00000000008.78.6, 00000000008.28.6] 2026-03-20T18:03:50.346 INFO:teuthology.orchestra.run.vm02.stderr:2026-03-20T18:03:50.341+0000 7fc16ff95a80 0 bucket checkpoint complete 2026-03-20T18:03:50.355 INFO:tasks.rgw_multi.zone_rados:comparing bucket=cyjkqa-12 zones={a1, a2} 2026-03-20T18:03:50.361 DEBUG:tasks.rgw_multi.zone_rados:bucket1 objects: 2026-03-20T18:03:50.364 DEBUG:tasks.rgw_multi.zone_rados:bucket2 objects: 2026-03-20T18:03:50.364 INFO:tasks.rgw_multi.zone_rados:success, bucket identical: bucket=cyjkqa-12 zones={a1, a2} 2026-03-20T18:03:50.367 INFO:tasks.rgw_multisite_tests:rgw_multi.tests.test_multi_object_delete ... ok 2026-03-20T18:03:50.411 INFO:rgw_multi.tests:create bucket zone=a1 name=cyjkqa-13 2026-03-20T18:03:50.424 INFO:rgw_multi.tests:create bucket zone=a2 name=cyjkqa-14 2026-03-20T18:03:50.458 INFO:tasks.util.rgw:rgwadmin: c1.client.0 : ['mdlog', 'status', '--rgw-zone', 'a1', '--rgw-zonegroup', 'a', '--rgw-realm', 'test-realm', '--cluster', 'c1', '--debug-rgw', '0', '--debug-ms', '0', '--rgw-cache-enabled', 'false'] 2026-03-20T18:03:50.458 DEBUG:tasks.util.rgw:rgwadmin: cmd=['adjust-ulimits', 'ceph-coverage', '/home/ubuntu/cephtest/archive/coverage', 'radosgw-admin', '--log-to-stderr', '--format', 'json', '-n', 'client.0', '--cluster', 'c1', 'mdlog', 'status', '--rgw-zone', 'a1', '--rgw-zonegroup', 'a', '--rgw-realm', 'test-realm', '--cluster', 'c1', '--debug-rgw', '0', '--debug-ms', '0', '--rgw-cache-enabled', 'false'] 2026-03-20T18:03:50.458 DEBUG:teuthology.orchestra.run.vm02:> adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage radosgw-admin --log-to-stderr --format json -n client.0 --cluster c1 mdlog status --rgw-zone a1 --rgw-zonegroup a --rgw-realm test-realm --cluster c1 --debug-rgw 0 --debug-ms 0 --rgw-cache-enabled false 2026-03-20T18:03:50.478 INFO:teuthology.orchestra.run.vm02.stderr:ignoring --setuser ceph since I am not root 2026-03-20T18:03:50.478 INFO:teuthology.orchestra.run.vm02.stderr:ignoring --setgroup ceph since I am not root 2026-03-20T18:03:50.547 INFO:teuthology.orchestra.run.vm02.stderr:No --period given, using current period=82c662be-66b5-4c4f-b24e-ff9298fc8a6e 2026-03-20T18:03:50.559 INFO:teuthology.orchestra.run.vm02.stdout:[{"marker":"1_1774029830.455027_204.1","last_update":"2026-03-20T18:03:50.455027Z"},{"marker":"1_1774029751.502359_72.1","last_update":"2026-03-20T18:02:31.502359Z"},{"marker":"1_1774029751.504947_186.1","last_update":"2026-03-20T18:02:31.504947Z"},{"marker":"1_1774029830.435475_132.1","last_update":"2026-03-20T18:03:50.435475Z"}] 2026-03-20T18:03:50.560 INFO:rgw_multi.tests:starting meta checkpoint for zone=a2 2026-03-20T18:03:50.560 INFO:tasks.util.rgw:rgwadmin: c1.client.0 : ['metadata', 'sync', 'status', '--rgw-zone', 'a2', '--rgw-zonegroup', 'a', '--rgw-realm', 'test-realm', '--cluster', 'c1', '--debug-rgw', '0', '--debug-ms', '0', '--rgw-cache-enabled', 'false'] 2026-03-20T18:03:50.560 DEBUG:tasks.util.rgw:rgwadmin: cmd=['adjust-ulimits', 'ceph-coverage', '/home/ubuntu/cephtest/archive/coverage', 'radosgw-admin', '--log-to-stderr', '--format', 'json', '-n', 'client.0', '--cluster', 'c1', 'metadata', 'sync', 'status', '--rgw-zone', 'a2', '--rgw-zonegroup', 'a', '--rgw-realm', 'test-realm', '--cluster', 'c1', '--debug-rgw', '0', '--debug-ms', '0', '--rgw-cache-enabled', 'false'] 2026-03-20T18:03:50.560 DEBUG:teuthology.orchestra.run.vm02:> adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage radosgw-admin --log-to-stderr --format json -n client.0 --cluster c1 metadata sync status --rgw-zone a2 --rgw-zonegroup a --rgw-realm test-realm --cluster c1 --debug-rgw 0 --debug-ms 0 --rgw-cache-enabled false 2026-03-20T18:03:50.626 INFO:teuthology.orchestra.run.vm02.stderr:ignoring --setuser ceph since I am not root 2026-03-20T18:03:50.626 INFO:teuthology.orchestra.run.vm02.stderr:ignoring --setgroup ceph since I am not root 2026-03-20T18:03:50.720 INFO:teuthology.orchestra.run.vm02.stderr:2026-03-20T18:03:50.717+0000 7faaba7d8a80 20 RGW-SYNC:meta: read sync status 2026-03-20T18:03:50.721 INFO:teuthology.orchestra.run.vm02.stderr:2026-03-20T18:03:50.717+0000 7faaba7d8a80 20 RGW-SYNC:meta: read sync status 2026-03-20T18:03:50.731 INFO:teuthology.orchestra.run.vm02.stdout:{"sync_status":{"info":{"status":"sync","num_shards":4,"period":"82c662be-66b5-4c4f-b24e-ff9298fc8a6e","realm_epoch":2},"markers":[{"key":0,"val":{"state":1,"marker":"1_1774029830.455027_204.1","next_step_marker":"","total_entries":1,"pos":0,"timestamp":"2026-03-20T18:03:50.455027Z","realm_epoch":2}},{"key":1,"val":{"state":1,"marker":"1_1774029751.502359_72.1","next_step_marker":"","total_entries":0,"pos":0,"timestamp":"2026-03-20T18:02:31.502359Z","realm_epoch":2}},{"key":2,"val":{"state":1,"marker":"1_1774029751.504947_186.1","next_step_marker":"","total_entries":0,"pos":0,"timestamp":"2026-03-20T18:02:31.504947Z","realm_epoch":2}},{"key":3,"val":{"state":1,"marker":"1_1774029830.435475_132.1","next_step_marker":"","total_entries":0,"pos":0,"timestamp":"2026-03-20T18:03:50.435475Z","realm_epoch":2}}]},"full_sync":{"total":1,"complete":1},"current_time":"2026-03-20T18:03:50Z"} 2026-03-20T18:03:50.731 INFO:rgw_multi.tests:finish meta checkpoint for zone=a2 2026-03-20T18:03:50.745 INFO:tasks.util.rgw:rgwadmin: c1.client.0 : ['bucket', 'sync', 'checkpoint', '--bucket', 'cyjkqa-13', '--source-zone', 'a1', '--retry-delay-ms', '5000', '--timeout-sec', '300', '--rgw-zone', 'a2', '--rgw-zonegroup', 'a', '--rgw-realm', 'test-realm', '--cluster', 'c1', '--debug-rgw', '1', '--debug-ms', '0'] 2026-03-20T18:03:50.745 DEBUG:tasks.util.rgw:rgwadmin: cmd=['adjust-ulimits', 'ceph-coverage', '/home/ubuntu/cephtest/archive/coverage', 'radosgw-admin', '--log-to-stderr', '--format', 'json', '-n', 'client.0', '--cluster', 'c1', 'bucket', 'sync', 'checkpoint', '--bucket', 'cyjkqa-13', '--source-zone', 'a1', '--retry-delay-ms', '5000', '--timeout-sec', '300', '--rgw-zone', 'a2', '--rgw-zonegroup', 'a', '--rgw-realm', 'test-realm', '--cluster', 'c1', '--debug-rgw', '1', '--debug-ms', '0'] 2026-03-20T18:03:50.745 DEBUG:teuthology.orchestra.run.vm02:> adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage radosgw-admin --log-to-stderr --format json -n client.0 --cluster c1 bucket sync checkpoint --bucket cyjkqa-13 --source-zone a1 --retry-delay-ms 5000 --timeout-sec 300 --rgw-zone a2 --rgw-zonegroup a --rgw-realm test-realm --cluster c1 --debug-rgw 1 --debug-ms 0 2026-03-20T18:03:50.797 INFO:teuthology.orchestra.run.vm02.stderr:ignoring --setuser ceph since I am not root 2026-03-20T18:03:50.798 INFO:teuthology.orchestra.run.vm02.stderr:ignoring --setgroup ceph since I am not root 2026-03-20T18:03:50.880 INFO:teuthology.orchestra.run.vm02.stderr:2026-03-20T18:03:50.877+0000 7f267bd05a80 1 waiting to reach incremental sync.. 2026-03-20T18:03:55.883 INFO:teuthology.orchestra.run.vm02.stderr:2026-03-20T18:03:55.881+0000 7f267bd05a80 1 bucket sync caught up with source: 2026-03-20T18:03:55.883 INFO:teuthology.orchestra.run.vm02.stderr: local status: [, , , , , 00000000002.35.11, , , , , ] 2026-03-20T18:03:55.883 INFO:teuthology.orchestra.run.vm02.stderr: remote markers: [, , , , , 00000000002.35.11, , , , , ] 2026-03-20T18:03:55.883 INFO:teuthology.orchestra.run.vm02.stderr:2026-03-20T18:03:55.881+0000 7f267bd05a80 0 bucket checkpoint complete 2026-03-20T18:03:55.908 INFO:tasks.util.rgw:rgwadmin: c1.client.0 : ['bucket', 'sync', 'checkpoint', '--bucket', 'cyjkqa-14', '--source-zone', 'a2', '--retry-delay-ms', '5000', '--timeout-sec', '300', '--rgw-zone', 'a1', '--rgw-zonegroup', 'a', '--rgw-realm', 'test-realm', '--cluster', 'c1', '--debug-rgw', '1', '--debug-ms', '0'] 2026-03-20T18:03:55.908 DEBUG:tasks.util.rgw:rgwadmin: cmd=['adjust-ulimits', 'ceph-coverage', '/home/ubuntu/cephtest/archive/coverage', 'radosgw-admin', '--log-to-stderr', '--format', 'json', '-n', 'client.0', '--cluster', 'c1', 'bucket', 'sync', 'checkpoint', '--bucket', 'cyjkqa-14', '--source-zone', 'a2', '--retry-delay-ms', '5000', '--timeout-sec', '300', '--rgw-zone', 'a1', '--rgw-zonegroup', 'a', '--rgw-realm', 'test-realm', '--cluster', 'c1', '--debug-rgw', '1', '--debug-ms', '0'] 2026-03-20T18:03:55.908 DEBUG:teuthology.orchestra.run.vm02:> adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage radosgw-admin --log-to-stderr --format json -n client.0 --cluster c1 bucket sync checkpoint --bucket cyjkqa-14 --source-zone a2 --retry-delay-ms 5000 --timeout-sec 300 --rgw-zone a1 --rgw-zonegroup a --rgw-realm test-realm --cluster c1 --debug-rgw 1 --debug-ms 0 2026-03-20T18:03:55.959 INFO:teuthology.orchestra.run.vm02.stderr:ignoring --setuser ceph since I am not root 2026-03-20T18:03:55.959 INFO:teuthology.orchestra.run.vm02.stderr:ignoring --setgroup ceph since I am not root 2026-03-20T18:03:56.037 INFO:teuthology.orchestra.run.vm02.stderr:2026-03-20T18:03:56.033+0000 7fae0f401a80 1 waiting to reach incremental sync.. 2026-03-20T18:04:01.038 INFO:teuthology.orchestra.run.vm02.stderr:2026-03-20T18:04:01.033+0000 7fae0f401a80 1 bucket sync caught up with source: 2026-03-20T18:04:01.039 INFO:teuthology.orchestra.run.vm02.stderr: local status: [, , , , , 00000000002.38.11, , , , , ] 2026-03-20T18:04:01.039 INFO:teuthology.orchestra.run.vm02.stderr: remote markers: [, , , , , 00000000002.38.11, , , , , ] 2026-03-20T18:04:01.039 INFO:teuthology.orchestra.run.vm02.stderr:2026-03-20T18:04:01.033+0000 7fae0f401a80 0 bucket checkpoint complete 2026-03-20T18:04:01.320 INFO:tasks.util.rgw:rgwadmin: c1.client.0 : ['bucket', 'sync', 'checkpoint', '--bucket', 'cyjkqa-13', '--source-zone', 'a1', '--retry-delay-ms', '5000', '--timeout-sec', '300', '--rgw-zone', 'a2', '--rgw-zonegroup', 'a', '--rgw-realm', 'test-realm', '--cluster', 'c1', '--debug-rgw', '1', '--debug-ms', '0'] 2026-03-20T18:04:01.320 DEBUG:tasks.util.rgw:rgwadmin: cmd=['adjust-ulimits', 'ceph-coverage', '/home/ubuntu/cephtest/archive/coverage', 'radosgw-admin', '--log-to-stderr', '--format', 'json', '-n', 'client.0', '--cluster', 'c1', 'bucket', 'sync', 'checkpoint', '--bucket', 'cyjkqa-13', '--source-zone', 'a1', '--retry-delay-ms', '5000', '--timeout-sec', '300', '--rgw-zone', 'a2', '--rgw-zonegroup', 'a', '--rgw-realm', 'test-realm', '--cluster', 'c1', '--debug-rgw', '1', '--debug-ms', '0'] 2026-03-20T18:04:01.320 DEBUG:teuthology.orchestra.run.vm02:> adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage radosgw-admin --log-to-stderr --format json -n client.0 --cluster c1 bucket sync checkpoint --bucket cyjkqa-13 --source-zone a1 --retry-delay-ms 5000 --timeout-sec 300 --rgw-zone a2 --rgw-zonegroup a --rgw-realm test-realm --cluster c1 --debug-rgw 1 --debug-ms 0 2026-03-20T18:04:01.342 INFO:teuthology.orchestra.run.vm02.stderr:ignoring --setuser ceph since I am not root 2026-03-20T18:04:01.343 INFO:teuthology.orchestra.run.vm02.stderr:ignoring --setgroup ceph since I am not root 2026-03-20T18:04:01.426 INFO:teuthology.orchestra.run.vm02.stderr:2026-03-20T18:04:01.421+0000 7f435f503a80 1 waiting for incremental sync to catch up: 2026-03-20T18:04:01.426 INFO:teuthology.orchestra.run.vm02.stderr: local status: [, , , , , 00000000002.35.11, , , , , ] 2026-03-20T18:04:01.426 INFO:teuthology.orchestra.run.vm02.stderr: remote markers: [, , , , , 00000000003.37.10, , , , 00000000010.46.6, ] 2026-03-20T18:04:06.428 INFO:teuthology.orchestra.run.vm02.stderr:2026-03-20T18:04:06.425+0000 7f435f503a80 1 waiting for incremental sync to catch up: 2026-03-20T18:04:06.428 INFO:teuthology.orchestra.run.vm02.stderr: local status: [, , , , , 00000000002.35.11, , , 00000000002.10.11, 00000000010.46.6, ] 2026-03-20T18:04:06.428 INFO:teuthology.orchestra.run.vm02.stderr: remote markers: [, , , , , 00000000003.37.10, , , , 00000000010.46.6, ] 2026-03-20T18:04:11.429 INFO:teuthology.orchestra.run.vm02.stderr:2026-03-20T18:04:11.425+0000 7f435f503a80 1 bucket sync caught up with source: 2026-03-20T18:04:11.429 INFO:teuthology.orchestra.run.vm02.stderr: local status: [, , , , , 00000000003.37.10, , , 00000000002.10.11, 00000000010.46.6, ] 2026-03-20T18:04:11.429 INFO:teuthology.orchestra.run.vm02.stderr: remote markers: [, , , , , 00000000003.37.10, , , , 00000000010.46.6, ] 2026-03-20T18:04:11.429 INFO:teuthology.orchestra.run.vm02.stderr:2026-03-20T18:04:11.425+0000 7f435f503a80 0 bucket checkpoint complete 2026-03-20T18:04:11.439 INFO:tasks.util.rgw:rgwadmin: c1.client.0 : ['bucket', 'sync', 'checkpoint', '--bucket', 'cyjkqa-13', '--source-zone', 'a2', '--retry-delay-ms', '5000', '--timeout-sec', '300', '--rgw-zone', 'a1', '--rgw-zonegroup', 'a', '--rgw-realm', 'test-realm', '--cluster', 'c1', '--debug-rgw', '1', '--debug-ms', '0'] 2026-03-20T18:04:11.439 DEBUG:tasks.util.rgw:rgwadmin: cmd=['adjust-ulimits', 'ceph-coverage', '/home/ubuntu/cephtest/archive/coverage', 'radosgw-admin', '--log-to-stderr', '--format', 'json', '-n', 'client.0', '--cluster', 'c1', 'bucket', 'sync', 'checkpoint', '--bucket', 'cyjkqa-13', '--source-zone', 'a2', '--retry-delay-ms', '5000', '--timeout-sec', '300', '--rgw-zone', 'a1', '--rgw-zonegroup', 'a', '--rgw-realm', 'test-realm', '--cluster', 'c1', '--debug-rgw', '1', '--debug-ms', '0'] 2026-03-20T18:04:11.439 DEBUG:teuthology.orchestra.run.vm02:> adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage radosgw-admin --log-to-stderr --format json -n client.0 --cluster c1 bucket sync checkpoint --bucket cyjkqa-13 --source-zone a2 --retry-delay-ms 5000 --timeout-sec 300 --rgw-zone a1 --rgw-zonegroup a --rgw-realm test-realm --cluster c1 --debug-rgw 1 --debug-ms 0 2026-03-20T18:04:11.510 INFO:teuthology.orchestra.run.vm02.stderr:ignoring --setuser ceph since I am not root 2026-03-20T18:04:11.510 INFO:teuthology.orchestra.run.vm02.stderr:ignoring --setgroup ceph since I am not root 2026-03-20T18:04:11.586 INFO:teuthology.orchestra.run.vm02.stderr:2026-03-20T18:04:11.581+0000 7f64ff1f5a80 1 bucket sync caught up with source: 2026-03-20T18:04:11.586 INFO:teuthology.orchestra.run.vm02.stderr: local status: [, , , , , 00000000002.38.11, , , 00000000010.22.6, 00000000002.31.11, ] 2026-03-20T18:04:11.586 INFO:teuthology.orchestra.run.vm02.stderr: remote markers: [, , , , , 00000000002.38.11, , , 00000000010.22.6, 00000000002.31.11, ] 2026-03-20T18:04:11.586 INFO:teuthology.orchestra.run.vm02.stderr:2026-03-20T18:04:11.581+0000 7f64ff1f5a80 0 bucket checkpoint complete 2026-03-20T18:04:11.596 INFO:tasks.rgw_multi.zone_rados:comparing bucket=cyjkqa-13 zones={a2, a1} 2026-03-20T18:04:11.643 DEBUG:tasks.rgw_multi.zone_rados:bucket1 objects: 2026-03-20T18:04:11.643 DEBUG:tasks.rgw_multi.zone_rados:o=dummy, v=Kvuz5.4V4jOrDLNIVMav8djnl2RyrOH 2026-03-20T18:04:11.643 DEBUG:tasks.rgw_multi.zone_rados:o=obj-a1, v=ed.lk0OS7lrCg7JyG03ZVdNVXF2NsU8 2026-03-20T18:04:11.643 DEBUG:tasks.rgw_multi.zone_rados:o=obj-a2, v=nat6aXHu09AtJQaZH4K2gt9k0TwrUR0 2026-03-20T18:04:11.646 DEBUG:tasks.rgw_multi.zone_rados:bucket2 objects: 2026-03-20T18:04:11.646 DEBUG:tasks.rgw_multi.zone_rados:o=dummy, v=Kvuz5.4V4jOrDLNIVMav8djnl2RyrOH 2026-03-20T18:04:11.646 DEBUG:tasks.rgw_multi.zone_rados:o=obj-a1, v=ed.lk0OS7lrCg7JyG03ZVdNVXF2NsU8 2026-03-20T18:04:11.646 DEBUG:tasks.rgw_multi.zone_rados:o=obj-a2, v=nat6aXHu09AtJQaZH4K2gt9k0TwrUR0 2026-03-20T18:04:11.646 DEBUG:tasks.rgw_multi.zone_rados:comparing key name=dummy 2026-03-20T18:04:11.660 DEBUG:tasks.rgw_multi.zone_rados:comparing key name=obj-a1 2026-03-20T18:04:11.719 DEBUG:tasks.rgw_multi.zone_rados:comparing key name=obj-a2 2026-03-20T18:04:11.771 INFO:tasks.rgw_multi.zone_rados:success, bucket identical: bucket=cyjkqa-13 zones={a2, a1} 2026-03-20T18:04:11.771 INFO:tasks.util.rgw:rgwadmin: c1.client.0 : ['bucket', 'sync', 'checkpoint', '--bucket', 'cyjkqa-14', '--source-zone', 'a1', '--retry-delay-ms', '5000', '--timeout-sec', '300', '--rgw-zone', 'a2', '--rgw-zonegroup', 'a', '--rgw-realm', 'test-realm', '--cluster', 'c1', '--debug-rgw', '1', '--debug-ms', '0'] 2026-03-20T18:04:11.771 DEBUG:tasks.util.rgw:rgwadmin: cmd=['adjust-ulimits', 'ceph-coverage', '/home/ubuntu/cephtest/archive/coverage', 'radosgw-admin', '--log-to-stderr', '--format', 'json', '-n', 'client.0', '--cluster', 'c1', 'bucket', 'sync', 'checkpoint', '--bucket', 'cyjkqa-14', '--source-zone', 'a1', '--retry-delay-ms', '5000', '--timeout-sec', '300', '--rgw-zone', 'a2', '--rgw-zonegroup', 'a', '--rgw-realm', 'test-realm', '--cluster', 'c1', '--debug-rgw', '1', '--debug-ms', '0'] 2026-03-20T18:04:11.772 DEBUG:teuthology.orchestra.run.vm02:> adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage radosgw-admin --log-to-stderr --format json -n client.0 --cluster c1 bucket sync checkpoint --bucket cyjkqa-14 --source-zone a1 --retry-delay-ms 5000 --timeout-sec 300 --rgw-zone a2 --rgw-zonegroup a --rgw-realm test-realm --cluster c1 --debug-rgw 1 --debug-ms 0 2026-03-20T18:04:11.795 INFO:teuthology.orchestra.run.vm02.stderr:ignoring --setuser ceph since I am not root 2026-03-20T18:04:11.795 INFO:teuthology.orchestra.run.vm02.stderr:ignoring --setgroup ceph since I am not root 2026-03-20T18:04:11.873 INFO:teuthology.orchestra.run.vm02.stderr:2026-03-20T18:04:11.869+0000 7f24ff6d2a80 1 bucket sync caught up with source: 2026-03-20T18:04:11.873 INFO:teuthology.orchestra.run.vm02.stderr: local status: [, , , , , 00000000002.30.11, , , 00000000002.41.11, 00000000010.21.6, ] 2026-03-20T18:04:11.873 INFO:teuthology.orchestra.run.vm02.stderr: remote markers: [, , , , , 00000000002.30.11, , , 00000000002.41.11, 00000000010.21.6, ] 2026-03-20T18:04:11.873 INFO:teuthology.orchestra.run.vm02.stderr:2026-03-20T18:04:11.869+0000 7f24ff6d2a80 0 bucket checkpoint complete 2026-03-20T18:04:11.885 INFO:tasks.util.rgw:rgwadmin: c1.client.0 : ['bucket', 'sync', 'checkpoint', '--bucket', 'cyjkqa-14', '--source-zone', 'a2', '--retry-delay-ms', '5000', '--timeout-sec', '300', '--rgw-zone', 'a1', '--rgw-zonegroup', 'a', '--rgw-realm', 'test-realm', '--cluster', 'c1', '--debug-rgw', '1', '--debug-ms', '0'] 2026-03-20T18:04:11.885 DEBUG:tasks.util.rgw:rgwadmin: cmd=['adjust-ulimits', 'ceph-coverage', '/home/ubuntu/cephtest/archive/coverage', 'radosgw-admin', '--log-to-stderr', '--format', 'json', '-n', 'client.0', '--cluster', 'c1', 'bucket', 'sync', 'checkpoint', '--bucket', 'cyjkqa-14', '--source-zone', 'a2', '--retry-delay-ms', '5000', '--timeout-sec', '300', '--rgw-zone', 'a1', '--rgw-zonegroup', 'a', '--rgw-realm', 'test-realm', '--cluster', 'c1', '--debug-rgw', '1', '--debug-ms', '0'] 2026-03-20T18:04:11.885 DEBUG:teuthology.orchestra.run.vm02:> adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage radosgw-admin --log-to-stderr --format json -n client.0 --cluster c1 bucket sync checkpoint --bucket cyjkqa-14 --source-zone a2 --retry-delay-ms 5000 --timeout-sec 300 --rgw-zone a1 --rgw-zonegroup a --rgw-realm test-realm --cluster c1 --debug-rgw 1 --debug-ms 0 2026-03-20T18:04:11.956 INFO:teuthology.orchestra.run.vm02.stderr:ignoring --setuser ceph since I am not root 2026-03-20T18:04:11.956 INFO:teuthology.orchestra.run.vm02.stderr:ignoring --setgroup ceph since I am not root 2026-03-20T18:04:12.030 INFO:teuthology.orchestra.run.vm02.stderr:2026-03-20T18:04:12.025+0000 7efde64e0a80 1 bucket sync caught up with source: 2026-03-20T18:04:12.030 INFO:teuthology.orchestra.run.vm02.stderr: local status: [, , , , , 00000000003.40.10, , , 00000000010.58.6, 00000000002.8.11, ] 2026-03-20T18:04:12.030 INFO:teuthology.orchestra.run.vm02.stderr: remote markers: [, , , , , 00000000003.40.10, , , 00000000010.58.6, 00000000002.8.11, ] 2026-03-20T18:04:12.030 INFO:teuthology.orchestra.run.vm02.stderr:2026-03-20T18:04:12.025+0000 7efde64e0a80 0 bucket checkpoint complete 2026-03-20T18:04:12.039 INFO:tasks.rgw_multi.zone_rados:comparing bucket=cyjkqa-14 zones={a2, a1} 2026-03-20T18:04:12.045 DEBUG:tasks.rgw_multi.zone_rados:bucket1 objects: 2026-03-20T18:04:12.046 DEBUG:tasks.rgw_multi.zone_rados:o=dummy, v=WDUInOfd7d4zBU7sT3VyNNqLAOSAgE. 2026-03-20T18:04:12.046 DEBUG:tasks.rgw_multi.zone_rados:o=obj-a1, v=6NGiwjCnhz1W8-lw5BkPlg.xp4TOUMa 2026-03-20T18:04:12.046 DEBUG:tasks.rgw_multi.zone_rados:o=obj-a2, v=TwygkKhG4FKaZUEdcC15ZelIkr3wVfc 2026-03-20T18:04:12.048 DEBUG:tasks.rgw_multi.zone_rados:bucket2 objects: 2026-03-20T18:04:12.048 DEBUG:tasks.rgw_multi.zone_rados:o=dummy, v=WDUInOfd7d4zBU7sT3VyNNqLAOSAgE. 2026-03-20T18:04:12.048 DEBUG:tasks.rgw_multi.zone_rados:o=obj-a1, v=6NGiwjCnhz1W8-lw5BkPlg.xp4TOUMa 2026-03-20T18:04:12.048 DEBUG:tasks.rgw_multi.zone_rados:o=obj-a2, v=TwygkKhG4FKaZUEdcC15ZelIkr3wVfc 2026-03-20T18:04:12.048 DEBUG:tasks.rgw_multi.zone_rados:comparing key name=dummy 2026-03-20T18:04:12.063 DEBUG:tasks.rgw_multi.zone_rados:comparing key name=obj-a1 2026-03-20T18:04:12.123 DEBUG:tasks.rgw_multi.zone_rados:comparing key name=obj-a2 2026-03-20T18:04:12.180 INFO:tasks.rgw_multi.zone_rados:success, bucket identical: bucket=cyjkqa-14 zones={a2, a1} 2026-03-20T18:04:12.255 INFO:tasks.util.rgw:rgwadmin: c1.client.0 : ['bucket', 'sync', 'checkpoint', '--bucket', 'cyjkqa-13', '--source-zone', 'a1', '--retry-delay-ms', '5000', '--timeout-sec', '300', '--rgw-zone', 'a2', '--rgw-zonegroup', 'a', '--rgw-realm', 'test-realm', '--cluster', 'c1', '--debug-rgw', '1', '--debug-ms', '0'] 2026-03-20T18:04:12.255 DEBUG:tasks.util.rgw:rgwadmin: cmd=['adjust-ulimits', 'ceph-coverage', '/home/ubuntu/cephtest/archive/coverage', 'radosgw-admin', '--log-to-stderr', '--format', 'json', '-n', 'client.0', '--cluster', 'c1', 'bucket', 'sync', 'checkpoint', '--bucket', 'cyjkqa-13', '--source-zone', 'a1', '--retry-delay-ms', '5000', '--timeout-sec', '300', '--rgw-zone', 'a2', '--rgw-zonegroup', 'a', '--rgw-realm', 'test-realm', '--cluster', 'c1', '--debug-rgw', '1', '--debug-ms', '0'] 2026-03-20T18:04:12.255 DEBUG:teuthology.orchestra.run.vm02:> adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage radosgw-admin --log-to-stderr --format json -n client.0 --cluster c1 bucket sync checkpoint --bucket cyjkqa-13 --source-zone a1 --retry-delay-ms 5000 --timeout-sec 300 --rgw-zone a2 --rgw-zonegroup a --rgw-realm test-realm --cluster c1 --debug-rgw 1 --debug-ms 0 2026-03-20T18:04:12.279 INFO:teuthology.orchestra.run.vm02.stderr:ignoring --setuser ceph since I am not root 2026-03-20T18:04:12.279 INFO:teuthology.orchestra.run.vm02.stderr:ignoring --setgroup ceph since I am not root 2026-03-20T18:04:12.354 INFO:teuthology.orchestra.run.vm02.stderr:2026-03-20T18:04:12.349+0000 7f8ebd324a80 1 waiting for incremental sync to catch up: 2026-03-20T18:04:12.354 INFO:teuthology.orchestra.run.vm02.stderr: local status: [, , , , , 00000000003.37.10, , , 00000000002.10.11, 00000000010.46.6, ] 2026-03-20T18:04:12.354 INFO:teuthology.orchestra.run.vm02.stderr: remote markers: [, , , , , 00000000003.37.10, , , 00000000002.10.11, 00000000011.49.6, ] 2026-03-20T18:04:17.355 INFO:teuthology.orchestra.run.vm02.stderr:2026-03-20T18:04:17.353+0000 7f8ebd324a80 1 waiting for incremental sync to catch up: 2026-03-20T18:04:17.355 INFO:teuthology.orchestra.run.vm02.stderr: local status: [, , , , , 00000000003.37.10, , , 00000000002.10.11, 00000000010.46.6, ] 2026-03-20T18:04:17.355 INFO:teuthology.orchestra.run.vm02.stderr: remote markers: [, , , , , 00000000003.37.10, , , 00000000002.10.11, 00000000011.49.6, ] 2026-03-20T18:04:22.356 INFO:teuthology.orchestra.run.vm02.stderr:2026-03-20T18:04:22.353+0000 7f8ebd324a80 1 waiting for incremental sync to catch up: 2026-03-20T18:04:22.356 INFO:teuthology.orchestra.run.vm02.stderr: local status: [, , , , , 00000000003.37.10, , , 00000000002.10.11, 00000000010.46.6, ] 2026-03-20T18:04:22.356 INFO:teuthology.orchestra.run.vm02.stderr: remote markers: [, , , , , 00000000003.37.10, , , 00000000002.10.11, 00000000011.49.6, ] 2026-03-20T18:04:27.357 INFO:teuthology.orchestra.run.vm02.stderr:2026-03-20T18:04:27.353+0000 7f8ebd324a80 1 waiting for incremental sync to catch up: 2026-03-20T18:04:27.357 INFO:teuthology.orchestra.run.vm02.stderr: local status: [, , , , , 00000000003.37.10, , , 00000000002.10.11, 00000000010.46.6, ] 2026-03-20T18:04:27.357 INFO:teuthology.orchestra.run.vm02.stderr: remote markers: [, , , , , 00000000003.37.10, , , 00000000002.10.11, 00000000011.49.6, ] 2026-03-20T18:04:32.358 INFO:teuthology.orchestra.run.vm02.stderr:2026-03-20T18:04:32.353+0000 7f8ebd324a80 1 bucket sync caught up with source: 2026-03-20T18:04:32.358 INFO:teuthology.orchestra.run.vm02.stderr: local status: [, , , , , 00000000003.37.10, , , 00000000002.10.11, 00000000011.49.6, ] 2026-03-20T18:04:32.358 INFO:teuthology.orchestra.run.vm02.stderr: remote markers: [, , , , , 00000000003.37.10, , , 00000000002.10.11, 00000000011.49.6, ] 2026-03-20T18:04:32.358 INFO:teuthology.orchestra.run.vm02.stderr:2026-03-20T18:04:32.353+0000 7f8ebd324a80 0 bucket checkpoint complete 2026-03-20T18:04:32.369 INFO:tasks.util.rgw:rgwadmin: c1.client.0 : ['bucket', 'sync', 'checkpoint', '--bucket', 'cyjkqa-13', '--source-zone', 'a2', '--retry-delay-ms', '5000', '--timeout-sec', '300', '--rgw-zone', 'a1', '--rgw-zonegroup', 'a', '--rgw-realm', 'test-realm', '--cluster', 'c1', '--debug-rgw', '1', '--debug-ms', '0'] 2026-03-20T18:04:32.369 DEBUG:tasks.util.rgw:rgwadmin: cmd=['adjust-ulimits', 'ceph-coverage', '/home/ubuntu/cephtest/archive/coverage', 'radosgw-admin', '--log-to-stderr', '--format', 'json', '-n', 'client.0', '--cluster', 'c1', 'bucket', 'sync', 'checkpoint', '--bucket', 'cyjkqa-13', '--source-zone', 'a2', '--retry-delay-ms', '5000', '--timeout-sec', '300', '--rgw-zone', 'a1', '--rgw-zonegroup', 'a', '--rgw-realm', 'test-realm', '--cluster', 'c1', '--debug-rgw', '1', '--debug-ms', '0'] 2026-03-20T18:04:32.369 DEBUG:teuthology.orchestra.run.vm02:> adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage radosgw-admin --log-to-stderr --format json -n client.0 --cluster c1 bucket sync checkpoint --bucket cyjkqa-13 --source-zone a2 --retry-delay-ms 5000 --timeout-sec 300 --rgw-zone a1 --rgw-zonegroup a --rgw-realm test-realm --cluster c1 --debug-rgw 1 --debug-ms 0 2026-03-20T18:04:32.434 INFO:teuthology.orchestra.run.vm02.stderr:ignoring --setuser ceph since I am not root 2026-03-20T18:04:32.434 INFO:teuthology.orchestra.run.vm02.stderr:ignoring --setgroup ceph since I am not root 2026-03-20T18:04:32.511 INFO:teuthology.orchestra.run.vm02.stderr:2026-03-20T18:04:32.509+0000 7f88eb738a80 1 waiting for incremental sync to catch up: 2026-03-20T18:04:32.511 INFO:teuthology.orchestra.run.vm02.stderr: local status: [, , , , , 00000000002.38.11, , , 00000000011.25.6, 00000000002.31.11, ] 2026-03-20T18:04:32.511 INFO:teuthology.orchestra.run.vm02.stderr: remote markers: [, , , , , 00000000002.38.11, , , 00000000011.25.6, 00000000003.34.6, ] 2026-03-20T18:04:37.512 INFO:teuthology.orchestra.run.vm02.stderr:2026-03-20T18:04:37.509+0000 7f88eb738a80 1 waiting for incremental sync to catch up: 2026-03-20T18:04:37.512 INFO:teuthology.orchestra.run.vm02.stderr: local status: [, , , , , 00000000002.38.11, , , 00000000011.25.6, 00000000002.31.11, ] 2026-03-20T18:04:37.512 INFO:teuthology.orchestra.run.vm02.stderr: remote markers: [, , , , , 00000000002.38.11, , , 00000000011.25.6, 00000000003.34.6, ] 2026-03-20T18:04:42.513 INFO:teuthology.orchestra.run.vm02.stderr:2026-03-20T18:04:42.509+0000 7f88eb738a80 1 waiting for incremental sync to catch up: 2026-03-20T18:04:42.513 INFO:teuthology.orchestra.run.vm02.stderr: local status: [, , , , , 00000000002.38.11, , , 00000000011.25.6, 00000000002.31.11, ] 2026-03-20T18:04:42.513 INFO:teuthology.orchestra.run.vm02.stderr: remote markers: [, , , , , 00000000002.38.11, , , 00000000011.25.6, 00000000003.34.6, ] 2026-03-20T18:04:47.514 INFO:teuthology.orchestra.run.vm02.stderr:2026-03-20T18:04:47.509+0000 7f88eb738a80 1 waiting for incremental sync to catch up: 2026-03-20T18:04:47.514 INFO:teuthology.orchestra.run.vm02.stderr: local status: [, , , , , 00000000002.38.11, , , 00000000011.25.6, 00000000002.31.11, ] 2026-03-20T18:04:47.514 INFO:teuthology.orchestra.run.vm02.stderr: remote markers: [, , , , , 00000000002.38.11, , , 00000000011.25.6, 00000000003.34.6, ] 2026-03-20T18:04:52.516 INFO:teuthology.orchestra.run.vm02.stderr:2026-03-20T18:04:52.513+0000 7f88eb738a80 1 bucket sync caught up with source: 2026-03-20T18:04:52.516 INFO:teuthology.orchestra.run.vm02.stderr: local status: [, , , , , 00000000002.38.11, , , 00000000011.25.6, 00000000003.34.6, ] 2026-03-20T18:04:52.516 INFO:teuthology.orchestra.run.vm02.stderr: remote markers: [, , , , , 00000000002.38.11, , , 00000000011.25.6, 00000000003.34.6, ] 2026-03-20T18:04:52.516 INFO:teuthology.orchestra.run.vm02.stderr:2026-03-20T18:04:52.513+0000 7f88eb738a80 0 bucket checkpoint complete 2026-03-20T18:04:52.527 INFO:tasks.rgw_multi.zone_rados:comparing bucket=cyjkqa-13 zones={a2, a1} 2026-03-20T18:04:52.533 DEBUG:tasks.rgw_multi.zone_rados:bucket1 objects: 2026-03-20T18:04:52.533 DEBUG:tasks.rgw_multi.zone_rados:o=dummy, v=Kvuz5.4V4jOrDLNIVMav8djnl2RyrOH 2026-03-20T18:04:52.533 DEBUG:tasks.rgw_multi.zone_rados:o=obj-a1, v=ed.lk0OS7lrCg7JyG03ZVdNVXF2NsU8 2026-03-20T18:04:52.533 DEBUG:tasks.rgw_multi.zone_rados:o=obj-a2, v=nat6aXHu09AtJQaZH4K2gt9k0TwrUR0 2026-03-20T18:04:52.535 DEBUG:tasks.rgw_multi.zone_rados:bucket2 objects: 2026-03-20T18:04:52.535 DEBUG:tasks.rgw_multi.zone_rados:o=dummy, v=Kvuz5.4V4jOrDLNIVMav8djnl2RyrOH 2026-03-20T18:04:52.535 DEBUG:tasks.rgw_multi.zone_rados:o=obj-a1, v=ed.lk0OS7lrCg7JyG03ZVdNVXF2NsU8 2026-03-20T18:04:52.535 DEBUG:tasks.rgw_multi.zone_rados:o=obj-a2, v=nat6aXHu09AtJQaZH4K2gt9k0TwrUR0 2026-03-20T18:04:52.535 DEBUG:tasks.rgw_multi.zone_rados:comparing key name=dummy 2026-03-20T18:04:52.546 DEBUG:tasks.rgw_multi.zone_rados:comparing key name=obj-a1 2026-03-20T18:04:52.603 DEBUG:tasks.rgw_multi.zone_rados:comparing key name=obj-a2 2026-03-20T18:04:52.659 INFO:tasks.rgw_multi.zone_rados:success, bucket identical: bucket=cyjkqa-13 zones={a2, a1} 2026-03-20T18:04:52.659 INFO:tasks.util.rgw:rgwadmin: c1.client.0 : ['bucket', 'sync', 'checkpoint', '--bucket', 'cyjkqa-14', '--source-zone', 'a1', '--retry-delay-ms', '5000', '--timeout-sec', '300', '--rgw-zone', 'a2', '--rgw-zonegroup', 'a', '--rgw-realm', 'test-realm', '--cluster', 'c1', '--debug-rgw', '1', '--debug-ms', '0'] 2026-03-20T18:04:52.659 DEBUG:tasks.util.rgw:rgwadmin: cmd=['adjust-ulimits', 'ceph-coverage', '/home/ubuntu/cephtest/archive/coverage', 'radosgw-admin', '--log-to-stderr', '--format', 'json', '-n', 'client.0', '--cluster', 'c1', 'bucket', 'sync', 'checkpoint', '--bucket', 'cyjkqa-14', '--source-zone', 'a1', '--retry-delay-ms', '5000', '--timeout-sec', '300', '--rgw-zone', 'a2', '--rgw-zonegroup', 'a', '--rgw-realm', 'test-realm', '--cluster', 'c1', '--debug-rgw', '1', '--debug-ms', '0'] 2026-03-20T18:04:52.659 DEBUG:teuthology.orchestra.run.vm02:> adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage radosgw-admin --log-to-stderr --format json -n client.0 --cluster c1 bucket sync checkpoint --bucket cyjkqa-14 --source-zone a1 --retry-delay-ms 5000 --timeout-sec 300 --rgw-zone a2 --rgw-zonegroup a --rgw-realm test-realm --cluster c1 --debug-rgw 1 --debug-ms 0 2026-03-20T18:04:52.683 INFO:teuthology.orchestra.run.vm02.stderr:ignoring --setuser ceph since I am not root 2026-03-20T18:04:52.683 INFO:teuthology.orchestra.run.vm02.stderr:ignoring --setgroup ceph since I am not root 2026-03-20T18:04:52.756 INFO:teuthology.orchestra.run.vm02.stderr:2026-03-20T18:04:52.753+0000 7f314e008a80 1 waiting for incremental sync to catch up: 2026-03-20T18:04:52.756 INFO:teuthology.orchestra.run.vm02.stderr: local status: [, , , , , 00000000002.30.11, , , 00000000002.41.11, 00000000011.24.6, ] 2026-03-20T18:04:52.756 INFO:teuthology.orchestra.run.vm02.stderr: remote markers: [, , , , , 00000000002.30.11, , , 00000000003.44.6, 00000000011.24.6, ] 2026-03-20T18:04:57.756 INFO:teuthology.orchestra.run.vm02.stderr:2026-03-20T18:04:57.753+0000 7f314e008a80 1 bucket sync caught up with source: 2026-03-20T18:04:57.757 INFO:teuthology.orchestra.run.vm02.stderr: local status: [, , , , , 00000000002.30.11, , , 00000000003.44.6, 00000000011.24.6, ] 2026-03-20T18:04:57.757 INFO:teuthology.orchestra.run.vm02.stderr: remote markers: [, , , , , 00000000002.30.11, , , 00000000003.44.6, 00000000011.24.6, ] 2026-03-20T18:04:57.757 INFO:teuthology.orchestra.run.vm02.stderr:2026-03-20T18:04:57.753+0000 7f314e008a80 0 bucket checkpoint complete 2026-03-20T18:04:57.766 INFO:tasks.util.rgw:rgwadmin: c1.client.0 : ['bucket', 'sync', 'checkpoint', '--bucket', 'cyjkqa-14', '--source-zone', 'a2', '--retry-delay-ms', '5000', '--timeout-sec', '300', '--rgw-zone', 'a1', '--rgw-zonegroup', 'a', '--rgw-realm', 'test-realm', '--cluster', 'c1', '--debug-rgw', '1', '--debug-ms', '0'] 2026-03-20T18:04:57.767 DEBUG:tasks.util.rgw:rgwadmin: cmd=['adjust-ulimits', 'ceph-coverage', '/home/ubuntu/cephtest/archive/coverage', 'radosgw-admin', '--log-to-stderr', '--format', 'json', '-n', 'client.0', '--cluster', 'c1', 'bucket', 'sync', 'checkpoint', '--bucket', 'cyjkqa-14', '--source-zone', 'a2', '--retry-delay-ms', '5000', '--timeout-sec', '300', '--rgw-zone', 'a1', '--rgw-zonegroup', 'a', '--rgw-realm', 'test-realm', '--cluster', 'c1', '--debug-rgw', '1', '--debug-ms', '0'] 2026-03-20T18:04:57.767 DEBUG:teuthology.orchestra.run.vm02:> adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage radosgw-admin --log-to-stderr --format json -n client.0 --cluster c1 bucket sync checkpoint --bucket cyjkqa-14 --source-zone a2 --retry-delay-ms 5000 --timeout-sec 300 --rgw-zone a1 --rgw-zonegroup a --rgw-realm test-realm --cluster c1 --debug-rgw 1 --debug-ms 0 2026-03-20T18:04:57.833 INFO:teuthology.orchestra.run.vm02.stderr:ignoring --setuser ceph since I am not root 2026-03-20T18:04:57.833 INFO:teuthology.orchestra.run.vm02.stderr:ignoring --setgroup ceph since I am not root 2026-03-20T18:04:57.909 INFO:teuthology.orchestra.run.vm02.stderr:2026-03-20T18:04:57.905+0000 7fddfc0d6a80 1 bucket sync caught up with source: 2026-03-20T18:04:57.909 INFO:teuthology.orchestra.run.vm02.stderr: local status: [, , , , , 00000000003.40.10, , , 00000000011.62.6, 00000000003.11.6, ] 2026-03-20T18:04:57.909 INFO:teuthology.orchestra.run.vm02.stderr: remote markers: [, , , , , 00000000003.40.10, , , 00000000011.62.6, 00000000003.11.6, ] 2026-03-20T18:04:57.909 INFO:teuthology.orchestra.run.vm02.stderr:2026-03-20T18:04:57.905+0000 7fddfc0d6a80 0 bucket checkpoint complete 2026-03-20T18:04:57.918 INFO:tasks.rgw_multi.zone_rados:comparing bucket=cyjkqa-14 zones={a2, a1} 2026-03-20T18:04:57.924 DEBUG:tasks.rgw_multi.zone_rados:bucket1 objects: 2026-03-20T18:04:57.924 DEBUG:tasks.rgw_multi.zone_rados:o=dummy, v=WDUInOfd7d4zBU7sT3VyNNqLAOSAgE. 2026-03-20T18:04:57.924 DEBUG:tasks.rgw_multi.zone_rados:o=obj-a1, v=6NGiwjCnhz1W8-lw5BkPlg.xp4TOUMa 2026-03-20T18:04:57.924 DEBUG:tasks.rgw_multi.zone_rados:o=obj-a2, v=TwygkKhG4FKaZUEdcC15ZelIkr3wVfc 2026-03-20T18:04:57.926 DEBUG:tasks.rgw_multi.zone_rados:bucket2 objects: 2026-03-20T18:04:57.926 DEBUG:tasks.rgw_multi.zone_rados:o=dummy, v=WDUInOfd7d4zBU7sT3VyNNqLAOSAgE. 2026-03-20T18:04:57.926 DEBUG:tasks.rgw_multi.zone_rados:o=obj-a1, v=6NGiwjCnhz1W8-lw5BkPlg.xp4TOUMa 2026-03-20T18:04:57.926 DEBUG:tasks.rgw_multi.zone_rados:o=obj-a2, v=TwygkKhG4FKaZUEdcC15ZelIkr3wVfc 2026-03-20T18:04:57.926 DEBUG:tasks.rgw_multi.zone_rados:comparing key name=dummy 2026-03-20T18:04:57.939 DEBUG:tasks.rgw_multi.zone_rados:comparing key name=obj-a1 2026-03-20T18:04:58.000 DEBUG:tasks.rgw_multi.zone_rados:comparing key name=obj-a2 2026-03-20T18:04:58.061 INFO:tasks.rgw_multi.zone_rados:success, bucket identical: bucket=cyjkqa-14 zones={a2, a1} 2026-03-20T18:04:58.064 INFO:tasks.rgw_multisite_tests:rgw_multi.tests.test_versioned_object_incremental_sync ... ok 2026-03-20T18:04:58.184 INFO:tasks.util.rgw:rgwadmin: c1.client.0 : ['mdlog', 'status', '--rgw-zone', 'a1', '--rgw-zonegroup', 'a', '--rgw-realm', 'test-realm', '--cluster', 'c1', '--debug-rgw', '0', '--debug-ms', '0', '--rgw-cache-enabled', 'false'] 2026-03-20T18:04:58.184 DEBUG:tasks.util.rgw:rgwadmin: cmd=['adjust-ulimits', 'ceph-coverage', '/home/ubuntu/cephtest/archive/coverage', 'radosgw-admin', '--log-to-stderr', '--format', 'json', '-n', 'client.0', '--cluster', 'c1', 'mdlog', 'status', '--rgw-zone', 'a1', '--rgw-zonegroup', 'a', '--rgw-realm', 'test-realm', '--cluster', 'c1', '--debug-rgw', '0', '--debug-ms', '0', '--rgw-cache-enabled', 'false'] 2026-03-20T18:04:58.184 DEBUG:teuthology.orchestra.run.vm02:> adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage radosgw-admin --log-to-stderr --format json -n client.0 --cluster c1 mdlog status --rgw-zone a1 --rgw-zonegroup a --rgw-realm test-realm --cluster c1 --debug-rgw 0 --debug-ms 0 --rgw-cache-enabled false 2026-03-20T18:04:58.203 INFO:teuthology.orchestra.run.vm02.stderr:ignoring --setuser ceph since I am not root 2026-03-20T18:04:58.203 INFO:teuthology.orchestra.run.vm02.stderr:ignoring --setgroup ceph since I am not root 2026-03-20T18:04:58.276 INFO:teuthology.orchestra.run.vm02.stderr:No --period given, using current period=82c662be-66b5-4c4f-b24e-ff9298fc8a6e 2026-03-20T18:04:58.286 INFO:teuthology.orchestra.run.vm02.stdout:[{"marker":"1_1774029898.178453_222.1","last_update":"2026-03-20T18:04:58.178453Z"},{"marker":"1_1774029751.502359_72.1","last_update":"2026-03-20T18:02:31.502359Z"},{"marker":"1_1774029751.504947_186.1","last_update":"2026-03-20T18:02:31.504947Z"},{"marker":"1_1774029898.181498_142.1","last_update":"2026-03-20T18:04:58.181498Z"}] 2026-03-20T18:04:58.286 INFO:rgw_multi.tests:starting meta checkpoint for zone=a2 2026-03-20T18:04:58.286 INFO:tasks.util.rgw:rgwadmin: c1.client.0 : ['metadata', 'sync', 'status', '--rgw-zone', 'a2', '--rgw-zonegroup', 'a', '--rgw-realm', 'test-realm', '--cluster', 'c1', '--debug-rgw', '0', '--debug-ms', '0', '--rgw-cache-enabled', 'false'] 2026-03-20T18:04:58.286 DEBUG:tasks.util.rgw:rgwadmin: cmd=['adjust-ulimits', 'ceph-coverage', '/home/ubuntu/cephtest/archive/coverage', 'radosgw-admin', '--log-to-stderr', '--format', 'json', '-n', 'client.0', '--cluster', 'c1', 'metadata', 'sync', 'status', '--rgw-zone', 'a2', '--rgw-zonegroup', 'a', '--rgw-realm', 'test-realm', '--cluster', 'c1', '--debug-rgw', '0', '--debug-ms', '0', '--rgw-cache-enabled', 'false'] 2026-03-20T18:04:58.286 DEBUG:teuthology.orchestra.run.vm02:> adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage radosgw-admin --log-to-stderr --format json -n client.0 --cluster c1 metadata sync status --rgw-zone a2 --rgw-zonegroup a --rgw-realm test-realm --cluster c1 --debug-rgw 0 --debug-ms 0 --rgw-cache-enabled false 2026-03-20T18:04:58.346 INFO:teuthology.orchestra.run.vm02.stderr:ignoring --setuser ceph since I am not root 2026-03-20T18:04:58.347 INFO:teuthology.orchestra.run.vm02.stderr:ignoring --setgroup ceph since I am not root 2026-03-20T18:04:58.411 INFO:teuthology.orchestra.run.vm02.stderr:2026-03-20T18:04:58.409+0000 7f15d7937a80 20 RGW-SYNC:meta: read sync status 2026-03-20T18:04:58.412 INFO:teuthology.orchestra.run.vm02.stderr:2026-03-20T18:04:58.409+0000 7f15d7937a80 20 RGW-SYNC:meta: read sync status 2026-03-20T18:04:58.422 INFO:teuthology.orchestra.run.vm02.stdout:{"sync_status":{"info":{"status":"sync","num_shards":4,"period":"82c662be-66b5-4c4f-b24e-ff9298fc8a6e","realm_epoch":2},"markers":[{"key":0,"val":{"state":1,"marker":"1_1774029898.178453_222.1","next_step_marker":"","total_entries":1,"pos":0,"timestamp":"2026-03-20T18:04:58.178453Z","realm_epoch":2}},{"key":1,"val":{"state":1,"marker":"1_1774029751.502359_72.1","next_step_marker":"","total_entries":0,"pos":0,"timestamp":"2026-03-20T18:02:31.502359Z","realm_epoch":2}},{"key":2,"val":{"state":1,"marker":"1_1774029751.504947_186.1","next_step_marker":"","total_entries":0,"pos":0,"timestamp":"2026-03-20T18:02:31.504947Z","realm_epoch":2}},{"key":3,"val":{"state":1,"marker":"1_1774029898.181498_142.1","next_step_marker":"","total_entries":0,"pos":0,"timestamp":"2026-03-20T18:04:58.181498Z","realm_epoch":2}}]},"full_sync":{"total":1,"complete":1},"current_time":"2026-03-20T18:04:58Z"} 2026-03-20T18:04:58.422 INFO:rgw_multi.tests:finish meta checkpoint for zone=a2 2026-03-20T18:04:58.429 INFO:tasks.util.rgw:rgwadmin: c1.client.0 : ['bucket', 'sync', 'checkpoint', '--bucket', 'cyjkqa-15', '--source-zone', 'a1', '--retry-delay-ms', '5000', '--timeout-sec', '300', '--rgw-zone', 'a2', '--rgw-zonegroup', 'a', '--rgw-realm', 'test-realm', '--cluster', 'c1', '--debug-rgw', '1', '--debug-ms', '0'] 2026-03-20T18:04:58.429 DEBUG:tasks.util.rgw:rgwadmin: cmd=['adjust-ulimits', 'ceph-coverage', '/home/ubuntu/cephtest/archive/coverage', 'radosgw-admin', '--log-to-stderr', '--format', 'json', '-n', 'client.0', '--cluster', 'c1', 'bucket', 'sync', 'checkpoint', '--bucket', 'cyjkqa-15', '--source-zone', 'a1', '--retry-delay-ms', '5000', '--timeout-sec', '300', '--rgw-zone', 'a2', '--rgw-zonegroup', 'a', '--rgw-realm', 'test-realm', '--cluster', 'c1', '--debug-rgw', '1', '--debug-ms', '0'] 2026-03-20T18:04:58.429 DEBUG:teuthology.orchestra.run.vm02:> adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage radosgw-admin --log-to-stderr --format json -n client.0 --cluster c1 bucket sync checkpoint --bucket cyjkqa-15 --source-zone a1 --retry-delay-ms 5000 --timeout-sec 300 --rgw-zone a2 --rgw-zonegroup a --rgw-realm test-realm --cluster c1 --debug-rgw 1 --debug-ms 0 2026-03-20T18:04:58.486 INFO:teuthology.orchestra.run.vm02.stderr:ignoring --setuser ceph since I am not root 2026-03-20T18:04:58.487 INFO:teuthology.orchestra.run.vm02.stderr:ignoring --setgroup ceph since I am not root 2026-03-20T18:04:58.559 INFO:teuthology.orchestra.run.vm02.stderr:2026-03-20T18:04:58.557+0000 7f95e60daa80 1 waiting to reach incremental sync.. 2026-03-20T18:05:03.560 INFO:teuthology.orchestra.run.vm02.stderr:2026-03-20T18:05:03.557+0000 7f95e60daa80 1 bucket sync caught up with source: 2026-03-20T18:05:03.561 INFO:teuthology.orchestra.run.vm02.stderr: local status: [, , , , , , , , 00000000001.7.6, , ] 2026-03-20T18:05:03.561 INFO:teuthology.orchestra.run.vm02.stderr: remote markers: [, , , , , , , , 00000000001.7.6, , ] 2026-03-20T18:05:03.561 INFO:teuthology.orchestra.run.vm02.stderr:2026-03-20T18:05:03.557+0000 7f95e60daa80 0 bucket checkpoint complete 2026-03-20T18:05:03.570 INFO:tasks.util.rgw:rgwadmin: c1.client.0 : ['bucket', 'sync', 'checkpoint', '--bucket', 'cyjkqa-15', '--source-zone', 'a2', '--retry-delay-ms', '5000', '--timeout-sec', '300', '--rgw-zone', 'a1', '--rgw-zonegroup', 'a', '--rgw-realm', 'test-realm', '--cluster', 'c1', '--debug-rgw', '1', '--debug-ms', '0'] 2026-03-20T18:05:03.571 DEBUG:tasks.util.rgw:rgwadmin: cmd=['adjust-ulimits', 'ceph-coverage', '/home/ubuntu/cephtest/archive/coverage', 'radosgw-admin', '--log-to-stderr', '--format', 'json', '-n', 'client.0', '--cluster', 'c1', 'bucket', 'sync', 'checkpoint', '--bucket', 'cyjkqa-15', '--source-zone', 'a2', '--retry-delay-ms', '5000', '--timeout-sec', '300', '--rgw-zone', 'a1', '--rgw-zonegroup', 'a', '--rgw-realm', 'test-realm', '--cluster', 'c1', '--debug-rgw', '1', '--debug-ms', '0'] 2026-03-20T18:05:03.571 DEBUG:teuthology.orchestra.run.vm02:> adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage radosgw-admin --log-to-stderr --format json -n client.0 --cluster c1 bucket sync checkpoint --bucket cyjkqa-15 --source-zone a2 --retry-delay-ms 5000 --timeout-sec 300 --rgw-zone a1 --rgw-zonegroup a --rgw-realm test-realm --cluster c1 --debug-rgw 1 --debug-ms 0 2026-03-20T18:05:03.642 INFO:teuthology.orchestra.run.vm02.stderr:ignoring --setuser ceph since I am not root 2026-03-20T18:05:03.642 INFO:teuthology.orchestra.run.vm02.stderr:ignoring --setgroup ceph since I am not root 2026-03-20T18:05:03.727 INFO:teuthology.orchestra.run.vm02.stderr:2026-03-20T18:05:03.725+0000 7f7521439a80 1 waiting to reach incremental sync.. 2026-03-20T18:05:08.728 INFO:teuthology.orchestra.run.vm02.stderr:2026-03-20T18:05:08.725+0000 7f7521439a80 1 bucket sync caught up with source: 2026-03-20T18:05:08.729 INFO:teuthology.orchestra.run.vm02.stderr: local status: [, , , , , , , , 00000000001.9.6, , ] 2026-03-20T18:05:08.729 INFO:teuthology.orchestra.run.vm02.stderr: remote markers: [, , , , , , , , 00000000001.9.6, , ] 2026-03-20T18:05:08.729 INFO:teuthology.orchestra.run.vm02.stderr:2026-03-20T18:05:08.725+0000 7f7521439a80 0 bucket checkpoint complete 2026-03-20T18:05:08.739 INFO:tasks.rgw_multi.zone_rados:comparing bucket=cyjkqa-15 zones={a2, a1} 2026-03-20T18:05:08.746 DEBUG:tasks.rgw_multi.zone_rados:bucket1 objects: 2026-03-20T18:05:08.746 DEBUG:tasks.rgw_multi.zone_rados:o=obj, v=null 2026-03-20T18:05:08.749 DEBUG:tasks.rgw_multi.zone_rados:bucket2 objects: 2026-03-20T18:05:08.749 DEBUG:tasks.rgw_multi.zone_rados:o=obj, v=null 2026-03-20T18:05:08.749 DEBUG:tasks.rgw_multi.zone_rados:comparing key name=obj 2026-03-20T18:05:08.764 INFO:tasks.rgw_multi.zone_rados:success, bucket identical: bucket=cyjkqa-15 zones={a2, a1} 2026-03-20T18:05:08.768 INFO:tasks.util.rgw:rgwadmin: c1.client.0 : ['mdlog', 'status', '--rgw-zone', 'a1', '--rgw-zonegroup', 'a', '--rgw-realm', 'test-realm', '--cluster', 'c1', '--debug-rgw', '0', '--debug-ms', '0', '--rgw-cache-enabled', 'false'] 2026-03-20T18:05:08.768 DEBUG:tasks.util.rgw:rgwadmin: cmd=['adjust-ulimits', 'ceph-coverage', '/home/ubuntu/cephtest/archive/coverage', 'radosgw-admin', '--log-to-stderr', '--format', 'json', '-n', 'client.0', '--cluster', 'c1', 'mdlog', 'status', '--rgw-zone', 'a1', '--rgw-zonegroup', 'a', '--rgw-realm', 'test-realm', '--cluster', 'c1', '--debug-rgw', '0', '--debug-ms', '0', '--rgw-cache-enabled', 'false'] 2026-03-20T18:05:08.768 DEBUG:teuthology.orchestra.run.vm02:> adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage radosgw-admin --log-to-stderr --format json -n client.0 --cluster c1 mdlog status --rgw-zone a1 --rgw-zonegroup a --rgw-realm test-realm --cluster c1 --debug-rgw 0 --debug-ms 0 --rgw-cache-enabled false 2026-03-20T18:05:08.805 INFO:teuthology.orchestra.run.vm02.stderr:ignoring --setuser ceph since I am not root 2026-03-20T18:05:08.806 INFO:teuthology.orchestra.run.vm02.stderr:ignoring --setgroup ceph since I am not root 2026-03-20T18:05:08.869 INFO:teuthology.orchestra.run.vm02.stderr:No --period given, using current period=82c662be-66b5-4c4f-b24e-ff9298fc8a6e 2026-03-20T18:05:08.878 INFO:teuthology.orchestra.run.vm02.stdout:[{"marker":"1_1774029908.767156_232.1","last_update":"2026-03-20T18:05:08.767156Z"},{"marker":"1_1774029751.502359_72.1","last_update":"2026-03-20T18:02:31.502359Z"},{"marker":"1_1774029751.504947_186.1","last_update":"2026-03-20T18:02:31.504947Z"},{"marker":"1_1774029898.181498_142.1","last_update":"2026-03-20T18:04:58.181498Z"}] 2026-03-20T18:05:08.879 INFO:rgw_multi.tests:starting meta checkpoint for zone=a2 2026-03-20T18:05:08.879 INFO:tasks.util.rgw:rgwadmin: c1.client.0 : ['metadata', 'sync', 'status', '--rgw-zone', 'a2', '--rgw-zonegroup', 'a', '--rgw-realm', 'test-realm', '--cluster', 'c1', '--debug-rgw', '0', '--debug-ms', '0', '--rgw-cache-enabled', 'false'] 2026-03-20T18:05:08.879 DEBUG:tasks.util.rgw:rgwadmin: cmd=['adjust-ulimits', 'ceph-coverage', '/home/ubuntu/cephtest/archive/coverage', 'radosgw-admin', '--log-to-stderr', '--format', 'json', '-n', 'client.0', '--cluster', 'c1', 'metadata', 'sync', 'status', '--rgw-zone', 'a2', '--rgw-zonegroup', 'a', '--rgw-realm', 'test-realm', '--cluster', 'c1', '--debug-rgw', '0', '--debug-ms', '0', '--rgw-cache-enabled', 'false'] 2026-03-20T18:05:08.879 DEBUG:teuthology.orchestra.run.vm02:> adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage radosgw-admin --log-to-stderr --format json -n client.0 --cluster c1 metadata sync status --rgw-zone a2 --rgw-zonegroup a --rgw-realm test-realm --cluster c1 --debug-rgw 0 --debug-ms 0 --rgw-cache-enabled false 2026-03-20T18:05:08.941 INFO:teuthology.orchestra.run.vm02.stderr:ignoring --setuser ceph since I am not root 2026-03-20T18:05:08.942 INFO:teuthology.orchestra.run.vm02.stderr:ignoring --setgroup ceph since I am not root 2026-03-20T18:05:09.000 INFO:teuthology.orchestra.run.vm02.stderr:2026-03-20T18:05:08.997+0000 7f4b2861ea80 20 RGW-SYNC:meta: read sync status 2026-03-20T18:05:09.000 INFO:teuthology.orchestra.run.vm02.stderr:2026-03-20T18:05:08.997+0000 7f4b2861ea80 20 RGW-SYNC:meta: read sync status 2026-03-20T18:05:09.008 INFO:teuthology.orchestra.run.vm02.stdout:{"sync_status":{"info":{"status":"sync","num_shards":4,"period":"82c662be-66b5-4c4f-b24e-ff9298fc8a6e","realm_epoch":2},"markers":[{"key":0,"val":{"state":1,"marker":"1_1774029908.767156_232.1","next_step_marker":"","total_entries":1,"pos":0,"timestamp":"2026-03-20T18:05:08.767156Z","realm_epoch":2}},{"key":1,"val":{"state":1,"marker":"1_1774029751.502359_72.1","next_step_marker":"","total_entries":0,"pos":0,"timestamp":"2026-03-20T18:02:31.502359Z","realm_epoch":2}},{"key":2,"val":{"state":1,"marker":"1_1774029751.504947_186.1","next_step_marker":"","total_entries":0,"pos":0,"timestamp":"2026-03-20T18:02:31.504947Z","realm_epoch":2}},{"key":3,"val":{"state":1,"marker":"1_1774029898.181498_142.1","next_step_marker":"","total_entries":0,"pos":0,"timestamp":"2026-03-20T18:04:58.181498Z","realm_epoch":2}}]},"full_sync":{"total":1,"complete":1},"current_time":"2026-03-20T18:05:09Z"} 2026-03-20T18:05:09.009 INFO:rgw_multi.tests:finish meta checkpoint for zone=a2 2026-03-20T18:05:09.020 INFO:tasks.util.rgw:rgwadmin: c1.client.0 : ['bucket', 'sync', 'checkpoint', '--bucket', 'cyjkqa-15', '--source-zone', 'a1', '--retry-delay-ms', '5000', '--timeout-sec', '300', '--rgw-zone', 'a2', '--rgw-zonegroup', 'a', '--rgw-realm', 'test-realm', '--cluster', 'c1', '--debug-rgw', '1', '--debug-ms', '0'] 2026-03-20T18:05:09.020 DEBUG:tasks.util.rgw:rgwadmin: cmd=['adjust-ulimits', 'ceph-coverage', '/home/ubuntu/cephtest/archive/coverage', 'radosgw-admin', '--log-to-stderr', '--format', 'json', '-n', 'client.0', '--cluster', 'c1', 'bucket', 'sync', 'checkpoint', '--bucket', 'cyjkqa-15', '--source-zone', 'a1', '--retry-delay-ms', '5000', '--timeout-sec', '300', '--rgw-zone', 'a2', '--rgw-zonegroup', 'a', '--rgw-realm', 'test-realm', '--cluster', 'c1', '--debug-rgw', '1', '--debug-ms', '0'] 2026-03-20T18:05:09.020 DEBUG:teuthology.orchestra.run.vm02:> adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage radosgw-admin --log-to-stderr --format json -n client.0 --cluster c1 bucket sync checkpoint --bucket cyjkqa-15 --source-zone a1 --retry-delay-ms 5000 --timeout-sec 300 --rgw-zone a2 --rgw-zonegroup a --rgw-realm test-realm --cluster c1 --debug-rgw 1 --debug-ms 0 2026-03-20T18:05:09.072 INFO:teuthology.orchestra.run.vm02.stderr:ignoring --setuser ceph since I am not root 2026-03-20T18:05:09.072 INFO:teuthology.orchestra.run.vm02.stderr:ignoring --setgroup ceph since I am not root 2026-03-20T18:05:09.138 INFO:teuthology.orchestra.run.vm02.stderr:2026-03-20T18:05:09.133+0000 7f17dbaeca80 1 waiting for incremental sync to catch up: 2026-03-20T18:05:09.138 INFO:teuthology.orchestra.run.vm02.stderr: local status: [, , , , , , , , 00000000001.7.6, , ] 2026-03-20T18:05:09.138 INFO:teuthology.orchestra.run.vm02.stderr: remote markers: [, , , , , , , , 00000000003.10.13, , ] 2026-03-20T18:05:14.139 INFO:teuthology.orchestra.run.vm02.stderr:2026-03-20T18:05:14.137+0000 7f17dbaeca80 1 waiting for incremental sync to catch up: 2026-03-20T18:05:14.140 INFO:teuthology.orchestra.run.vm02.stderr: local status: [, , , , , , , , 00000000001.7.6, , ] 2026-03-20T18:05:14.140 INFO:teuthology.orchestra.run.vm02.stderr: remote markers: [, , , , , , , , 00000000003.10.13, , ] 2026-03-20T18:05:19.141 INFO:teuthology.orchestra.run.vm02.stderr:2026-03-20T18:05:19.137+0000 7f17dbaeca80 1 bucket sync caught up with source: 2026-03-20T18:05:19.141 INFO:teuthology.orchestra.run.vm02.stderr: local status: [, , , , , , , , 00000000003.10.13, , ] 2026-03-20T18:05:19.141 INFO:teuthology.orchestra.run.vm02.stderr: remote markers: [, , , , , , , , 00000000003.10.13, , ] 2026-03-20T18:05:19.141 INFO:teuthology.orchestra.run.vm02.stderr:2026-03-20T18:05:19.137+0000 7f17dbaeca80 0 bucket checkpoint complete 2026-03-20T18:05:19.150 INFO:tasks.util.rgw:rgwadmin: c1.client.0 : ['bucket', 'sync', 'checkpoint', '--bucket', 'cyjkqa-15', '--source-zone', 'a2', '--retry-delay-ms', '5000', '--timeout-sec', '300', '--rgw-zone', 'a1', '--rgw-zonegroup', 'a', '--rgw-realm', 'test-realm', '--cluster', 'c1', '--debug-rgw', '1', '--debug-ms', '0'] 2026-03-20T18:05:19.150 DEBUG:tasks.util.rgw:rgwadmin: cmd=['adjust-ulimits', 'ceph-coverage', '/home/ubuntu/cephtest/archive/coverage', 'radosgw-admin', '--log-to-stderr', '--format', 'json', '-n', 'client.0', '--cluster', 'c1', 'bucket', 'sync', 'checkpoint', '--bucket', 'cyjkqa-15', '--source-zone', 'a2', '--retry-delay-ms', '5000', '--timeout-sec', '300', '--rgw-zone', 'a1', '--rgw-zonegroup', 'a', '--rgw-realm', 'test-realm', '--cluster', 'c1', '--debug-rgw', '1', '--debug-ms', '0'] 2026-03-20T18:05:19.150 DEBUG:teuthology.orchestra.run.vm02:> adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage radosgw-admin --log-to-stderr --format json -n client.0 --cluster c1 bucket sync checkpoint --bucket cyjkqa-15 --source-zone a2 --retry-delay-ms 5000 --timeout-sec 300 --rgw-zone a1 --rgw-zonegroup a --rgw-realm test-realm --cluster c1 --debug-rgw 1 --debug-ms 0 2026-03-20T18:05:19.214 INFO:teuthology.orchestra.run.vm02.stderr:ignoring --setuser ceph since I am not root 2026-03-20T18:05:19.214 INFO:teuthology.orchestra.run.vm02.stderr:ignoring --setgroup ceph since I am not root 2026-03-20T18:05:19.284 INFO:teuthology.orchestra.run.vm02.stderr:2026-03-20T18:05:19.281+0000 7f336edd0a80 1 waiting for incremental sync to catch up: 2026-03-20T18:05:19.284 INFO:teuthology.orchestra.run.vm02.stderr: local status: [, , , , , , , , 00000000001.9.6, , ] 2026-03-20T18:05:19.284 INFO:teuthology.orchestra.run.vm02.stderr: remote markers: [, , , , , , , , 00000000003.13.13, , ] 2026-03-20T18:05:24.285 INFO:teuthology.orchestra.run.vm02.stderr:2026-03-20T18:05:24.281+0000 7f336edd0a80 1 waiting for incremental sync to catch up: 2026-03-20T18:05:24.285 INFO:teuthology.orchestra.run.vm02.stderr: local status: [, , , , , , , , 00000000001.9.6, , ] 2026-03-20T18:05:24.285 INFO:teuthology.orchestra.run.vm02.stderr: remote markers: [, , , , , , , , 00000000003.13.13, , ] 2026-03-20T18:05:29.286 INFO:teuthology.orchestra.run.vm02.stderr:2026-03-20T18:05:29.281+0000 7f336edd0a80 1 waiting for incremental sync to catch up: 2026-03-20T18:05:29.286 INFO:teuthology.orchestra.run.vm02.stderr: local status: [, , , , , , , , 00000000001.9.6, , ] 2026-03-20T18:05:29.286 INFO:teuthology.orchestra.run.vm02.stderr: remote markers: [, , , , , , , , 00000000003.13.13, , ] 2026-03-20T18:05:34.288 INFO:teuthology.orchestra.run.vm02.stderr:2026-03-20T18:05:34.285+0000 7f336edd0a80 1 waiting for incremental sync to catch up: 2026-03-20T18:05:34.288 INFO:teuthology.orchestra.run.vm02.stderr: local status: [, , , , , , , , 00000000001.9.6, , ] 2026-03-20T18:05:34.288 INFO:teuthology.orchestra.run.vm02.stderr: remote markers: [, , , , , , , , 00000000003.13.13, , ] 2026-03-20T18:05:39.289 INFO:teuthology.orchestra.run.vm02.stderr:2026-03-20T18:05:39.285+0000 7f336edd0a80 1 bucket sync caught up with source: 2026-03-20T18:05:39.289 INFO:teuthology.orchestra.run.vm02.stderr: local status: [, , , , , , , , 00000000003.13.13, , ] 2026-03-20T18:05:39.289 INFO:teuthology.orchestra.run.vm02.stderr: remote markers: [, , , , , , , , 00000000003.13.13, , ] 2026-03-20T18:05:39.289 INFO:teuthology.orchestra.run.vm02.stderr:2026-03-20T18:05:39.285+0000 7f336edd0a80 0 bucket checkpoint complete 2026-03-20T18:05:39.299 INFO:tasks.rgw_multi.zone_rados:comparing bucket=cyjkqa-15 zones={a2, a1} 2026-03-20T18:05:39.304 DEBUG:tasks.rgw_multi.zone_rados:bucket1 objects: 2026-03-20T18:05:39.304 DEBUG:tasks.rgw_multi.zone_rados:o=obj, v=kxg1R15xGNhHzuvVIopvrmUZuO5JhS0 2026-03-20T18:05:39.304 DEBUG:tasks.rgw_multi.zone_rados:o=obj, v=null 2026-03-20T18:05:39.307 DEBUG:tasks.rgw_multi.zone_rados:bucket2 objects: 2026-03-20T18:05:39.307 DEBUG:tasks.rgw_multi.zone_rados:o=obj, v=kxg1R15xGNhHzuvVIopvrmUZuO5JhS0 2026-03-20T18:05:39.307 DEBUG:tasks.rgw_multi.zone_rados:o=obj, v=null 2026-03-20T18:05:39.307 DEBUG:tasks.rgw_multi.zone_rados:comparing key name=obj 2026-03-20T18:05:39.320 DEBUG:tasks.rgw_multi.zone_rados:comparing key name=obj 2026-03-20T18:05:39.331 INFO:tasks.rgw_multi.zone_rados:success, bucket identical: bucket=cyjkqa-15 zones={a2, a1} 2026-03-20T18:05:39.349 INFO:tasks.util.rgw:rgwadmin: c1.client.0 : ['bucket', 'sync', 'checkpoint', '--bucket', 'cyjkqa-15', '--source-zone', 'a1', '--retry-delay-ms', '5000', '--timeout-sec', '300', '--rgw-zone', 'a2', '--rgw-zonegroup', 'a', '--rgw-realm', 'test-realm', '--cluster', 'c1', '--debug-rgw', '1', '--debug-ms', '0'] 2026-03-20T18:05:39.349 DEBUG:tasks.util.rgw:rgwadmin: cmd=['adjust-ulimits', 'ceph-coverage', '/home/ubuntu/cephtest/archive/coverage', 'radosgw-admin', '--log-to-stderr', '--format', 'json', '-n', 'client.0', '--cluster', 'c1', 'bucket', 'sync', 'checkpoint', '--bucket', 'cyjkqa-15', '--source-zone', 'a1', '--retry-delay-ms', '5000', '--timeout-sec', '300', '--rgw-zone', 'a2', '--rgw-zonegroup', 'a', '--rgw-realm', 'test-realm', '--cluster', 'c1', '--debug-rgw', '1', '--debug-ms', '0'] 2026-03-20T18:05:39.349 DEBUG:teuthology.orchestra.run.vm02:> adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage radosgw-admin --log-to-stderr --format json -n client.0 --cluster c1 bucket sync checkpoint --bucket cyjkqa-15 --source-zone a1 --retry-delay-ms 5000 --timeout-sec 300 --rgw-zone a2 --rgw-zonegroup a --rgw-realm test-realm --cluster c1 --debug-rgw 1 --debug-ms 0 2026-03-20T18:05:39.374 INFO:teuthology.orchestra.run.vm02.stderr:ignoring --setuser ceph since I am not root 2026-03-20T18:05:39.374 INFO:teuthology.orchestra.run.vm02.stderr:ignoring --setgroup ceph since I am not root 2026-03-20T18:05:39.449 INFO:teuthology.orchestra.run.vm02.stderr:2026-03-20T18:05:39.445+0000 7fd58261ba80 1 waiting for incremental sync to catch up: 2026-03-20T18:05:39.449 INFO:teuthology.orchestra.run.vm02.stderr: local status: [, , , , , , , , 00000000003.10.13, , ] 2026-03-20T18:05:39.449 INFO:teuthology.orchestra.run.vm02.stderr: remote markers: [, , , , , , , , 00000000007.18.6, , ] 2026-03-20T18:05:44.450 INFO:teuthology.orchestra.run.vm02.stderr:2026-03-20T18:05:44.445+0000 7fd58261ba80 1 waiting for incremental sync to catch up: 2026-03-20T18:05:44.450 INFO:teuthology.orchestra.run.vm02.stderr: local status: [, , , , , , , , 00000000003.10.13, , ] 2026-03-20T18:05:44.450 INFO:teuthology.orchestra.run.vm02.stderr: remote markers: [, , , , , , , , 00000000007.18.6, , ] 2026-03-20T18:05:49.451 INFO:teuthology.orchestra.run.vm02.stderr:2026-03-20T18:05:49.449+0000 7fd58261ba80 1 waiting for incremental sync to catch up: 2026-03-20T18:05:49.452 INFO:teuthology.orchestra.run.vm02.stderr: local status: [, , , , , , , , 00000000003.10.13, , ] 2026-03-20T18:05:49.452 INFO:teuthology.orchestra.run.vm02.stderr: remote markers: [, , , , , , , , 00000000007.18.6, , ] 2026-03-20T18:05:54.453 INFO:teuthology.orchestra.run.vm02.stderr:2026-03-20T18:05:54.449+0000 7fd58261ba80 1 waiting for incremental sync to catch up: 2026-03-20T18:05:54.453 INFO:teuthology.orchestra.run.vm02.stderr: local status: [, , , , , , , , 00000000003.10.13, , ] 2026-03-20T18:05:54.453 INFO:teuthology.orchestra.run.vm02.stderr: remote markers: [, , , , , , , , 00000000007.18.6, , ] 2026-03-20T18:05:59.454 INFO:teuthology.orchestra.run.vm02.stderr:2026-03-20T18:05:59.450+0000 7fd58261ba80 1 waiting for incremental sync to catch up: 2026-03-20T18:05:59.454 INFO:teuthology.orchestra.run.vm02.stderr: local status: [, , , , , , , , 00000000003.10.13, , ] 2026-03-20T18:05:59.454 INFO:teuthology.orchestra.run.vm02.stderr: remote markers: [, , , , , , , , 00000000007.18.6, , ] 2026-03-20T18:06:04.455 INFO:teuthology.orchestra.run.vm02.stderr:2026-03-20T18:06:04.454+0000 7fd58261ba80 1 bucket sync caught up with source: 2026-03-20T18:06:04.455 INFO:teuthology.orchestra.run.vm02.stderr: local status: [, , , , , , , , 00000000007.18.6, , ] 2026-03-20T18:06:04.455 INFO:teuthology.orchestra.run.vm02.stderr: remote markers: [, , , , , , , , 00000000007.18.6, , ] 2026-03-20T18:06:04.455 INFO:teuthology.orchestra.run.vm02.stderr:2026-03-20T18:06:04.454+0000 7fd58261ba80 0 bucket checkpoint complete 2026-03-20T18:06:04.465 INFO:tasks.util.rgw:rgwadmin: c1.client.0 : ['bucket', 'sync', 'checkpoint', '--bucket', 'cyjkqa-15', '--source-zone', 'a2', '--retry-delay-ms', '5000', '--timeout-sec', '300', '--rgw-zone', 'a1', '--rgw-zonegroup', 'a', '--rgw-realm', 'test-realm', '--cluster', 'c1', '--debug-rgw', '1', '--debug-ms', '0'] 2026-03-20T18:06:04.466 DEBUG:tasks.util.rgw:rgwadmin: cmd=['adjust-ulimits', 'ceph-coverage', '/home/ubuntu/cephtest/archive/coverage', 'radosgw-admin', '--log-to-stderr', '--format', 'json', '-n', 'client.0', '--cluster', 'c1', 'bucket', 'sync', 'checkpoint', '--bucket', 'cyjkqa-15', '--source-zone', 'a2', '--retry-delay-ms', '5000', '--timeout-sec', '300', '--rgw-zone', 'a1', '--rgw-zonegroup', 'a', '--rgw-realm', 'test-realm', '--cluster', 'c1', '--debug-rgw', '1', '--debug-ms', '0'] 2026-03-20T18:06:04.466 DEBUG:teuthology.orchestra.run.vm02:> adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage radosgw-admin --log-to-stderr --format json -n client.0 --cluster c1 bucket sync checkpoint --bucket cyjkqa-15 --source-zone a2 --retry-delay-ms 5000 --timeout-sec 300 --rgw-zone a1 --rgw-zonegroup a --rgw-realm test-realm --cluster c1 --debug-rgw 1 --debug-ms 0 2026-03-20T18:06:04.533 INFO:teuthology.orchestra.run.vm02.stderr:ignoring --setuser ceph since I am not root 2026-03-20T18:06:04.533 INFO:teuthology.orchestra.run.vm02.stderr:ignoring --setgroup ceph since I am not root 2026-03-20T18:06:04.606 INFO:teuthology.orchestra.run.vm02.stderr:2026-03-20T18:06:04.602+0000 7fc476950a80 1 waiting for incremental sync to catch up: 2026-03-20T18:06:04.606 INFO:teuthology.orchestra.run.vm02.stderr: local status: [, , , , , , , , 00000000003.13.13, , ] 2026-03-20T18:06:04.606 INFO:teuthology.orchestra.run.vm02.stderr: remote markers: [, , , , , , , , 00000000007.21.6, , ] 2026-03-20T18:06:09.607 INFO:teuthology.orchestra.run.vm02.stderr:2026-03-20T18:06:09.606+0000 7fc476950a80 1 waiting for incremental sync to catch up: 2026-03-20T18:06:09.607 INFO:teuthology.orchestra.run.vm02.stderr: local status: [, , , , , , , , 00000000003.13.13, , ] 2026-03-20T18:06:09.607 INFO:teuthology.orchestra.run.vm02.stderr: remote markers: [, , , , , , , , 00000000007.21.6, , ] 2026-03-20T18:06:14.609 INFO:teuthology.orchestra.run.vm02.stderr:2026-03-20T18:06:14.606+0000 7fc476950a80 1 waiting for incremental sync to catch up: 2026-03-20T18:06:14.609 INFO:teuthology.orchestra.run.vm02.stderr: local status: [, , , , , , , , 00000000003.13.13, , ] 2026-03-20T18:06:14.609 INFO:teuthology.orchestra.run.vm02.stderr: remote markers: [, , , , , , , , 00000000007.21.6, , ] 2026-03-20T18:06:19.610 INFO:teuthology.orchestra.run.vm02.stderr:2026-03-20T18:06:19.606+0000 7fc476950a80 1 waiting for incremental sync to catch up: 2026-03-20T18:06:19.610 INFO:teuthology.orchestra.run.vm02.stderr: local status: [, , , , , , , , 00000000003.13.13, , ] 2026-03-20T18:06:19.610 INFO:teuthology.orchestra.run.vm02.stderr: remote markers: [, , , , , , , , 00000000007.21.6, , ] 2026-03-20T18:06:24.611 INFO:teuthology.orchestra.run.vm02.stderr:2026-03-20T18:06:24.610+0000 7fc476950a80 1 bucket sync caught up with source: 2026-03-20T18:06:24.611 INFO:teuthology.orchestra.run.vm02.stderr: local status: [, , , , , , , , 00000000007.21.6, , ] 2026-03-20T18:06:24.611 INFO:teuthology.orchestra.run.vm02.stderr: remote markers: [, , , , , , , , 00000000007.21.6, , ] 2026-03-20T18:06:24.611 INFO:teuthology.orchestra.run.vm02.stderr:2026-03-20T18:06:24.610+0000 7fc476950a80 0 bucket checkpoint complete 2026-03-20T18:06:24.621 INFO:tasks.rgw_multi.zone_rados:comparing bucket=cyjkqa-15 zones={a2, a1} 2026-03-20T18:06:24.628 DEBUG:tasks.rgw_multi.zone_rados:bucket1 objects: 2026-03-20T18:06:24.630 DEBUG:tasks.rgw_multi.zone_rados:bucket2 objects: 2026-03-20T18:06:24.630 INFO:tasks.rgw_multi.zone_rados:success, bucket identical: bucket=cyjkqa-15 zones={a2, a1} 2026-03-20T18:06:24.633 INFO:tasks.rgw_multisite_tests:rgw_multi.tests.test_null_version_id_delete ... ok 2026-03-20T18:06:24.702 INFO:tasks.util.rgw:rgwadmin: c1.client.0 : ['mdlog', 'status', '--rgw-zone', 'a1', '--rgw-zonegroup', 'a', '--rgw-realm', 'test-realm', '--cluster', 'c1', '--debug-rgw', '0', '--debug-ms', '0', '--rgw-cache-enabled', 'false'] 2026-03-20T18:06:24.703 DEBUG:tasks.util.rgw:rgwadmin: cmd=['adjust-ulimits', 'ceph-coverage', '/home/ubuntu/cephtest/archive/coverage', 'radosgw-admin', '--log-to-stderr', '--format', 'json', '-n', 'client.0', '--cluster', 'c1', 'mdlog', 'status', '--rgw-zone', 'a1', '--rgw-zonegroup', 'a', '--rgw-realm', 'test-realm', '--cluster', 'c1', '--debug-rgw', '0', '--debug-ms', '0', '--rgw-cache-enabled', 'false'] 2026-03-20T18:06:24.703 DEBUG:teuthology.orchestra.run.vm02:> adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage radosgw-admin --log-to-stderr --format json -n client.0 --cluster c1 mdlog status --rgw-zone a1 --rgw-zonegroup a --rgw-realm test-realm --cluster c1 --debug-rgw 0 --debug-ms 0 --rgw-cache-enabled false 2026-03-20T18:06:24.724 INFO:teuthology.orchestra.run.vm02.stderr:ignoring --setuser ceph since I am not root 2026-03-20T18:06:24.724 INFO:teuthology.orchestra.run.vm02.stderr:ignoring --setgroup ceph since I am not root 2026-03-20T18:06:24.811 INFO:teuthology.orchestra.run.vm02.stderr:No --period given, using current period=82c662be-66b5-4c4f-b24e-ff9298fc8a6e 2026-03-20T18:06:24.821 INFO:teuthology.orchestra.run.vm02.stdout:[{"marker":"1_1774029984.700846_244.1","last_update":"2026-03-20T18:06:24.700846Z"},{"marker":"1_1774029751.502359_72.1","last_update":"2026-03-20T18:02:31.502359Z"},{"marker":"1_1774029751.504947_186.1","last_update":"2026-03-20T18:02:31.504947Z"},{"marker":"1_1774029984.693472_148.1","last_update":"2026-03-20T18:06:24.693472Z"}] 2026-03-20T18:06:24.821 INFO:rgw_multi.tests:starting meta checkpoint for zone=a2 2026-03-20T18:06:24.821 INFO:tasks.util.rgw:rgwadmin: c1.client.0 : ['metadata', 'sync', 'status', '--rgw-zone', 'a2', '--rgw-zonegroup', 'a', '--rgw-realm', 'test-realm', '--cluster', 'c1', '--debug-rgw', '0', '--debug-ms', '0', '--rgw-cache-enabled', 'false'] 2026-03-20T18:06:24.821 DEBUG:tasks.util.rgw:rgwadmin: cmd=['adjust-ulimits', 'ceph-coverage', '/home/ubuntu/cephtest/archive/coverage', 'radosgw-admin', '--log-to-stderr', '--format', 'json', '-n', 'client.0', '--cluster', 'c1', 'metadata', 'sync', 'status', '--rgw-zone', 'a2', '--rgw-zonegroup', 'a', '--rgw-realm', 'test-realm', '--cluster', 'c1', '--debug-rgw', '0', '--debug-ms', '0', '--rgw-cache-enabled', 'false'] 2026-03-20T18:06:24.821 DEBUG:teuthology.orchestra.run.vm02:> adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage radosgw-admin --log-to-stderr --format json -n client.0 --cluster c1 metadata sync status --rgw-zone a2 --rgw-zonegroup a --rgw-realm test-realm --cluster c1 --debug-rgw 0 --debug-ms 0 --rgw-cache-enabled false 2026-03-20T18:06:24.885 INFO:teuthology.orchestra.run.vm02.stderr:ignoring --setuser ceph since I am not root 2026-03-20T18:06:24.885 INFO:teuthology.orchestra.run.vm02.stderr:ignoring --setgroup ceph since I am not root 2026-03-20T18:06:24.961 INFO:teuthology.orchestra.run.vm02.stderr:2026-03-20T18:06:24.958+0000 7fdc6a1dea80 20 RGW-SYNC:meta: read sync status 2026-03-20T18:06:24.962 INFO:teuthology.orchestra.run.vm02.stderr:2026-03-20T18:06:24.958+0000 7fdc6a1dea80 20 RGW-SYNC:meta: read sync status 2026-03-20T18:06:24.972 INFO:teuthology.orchestra.run.vm02.stdout:{"sync_status":{"info":{"status":"sync","num_shards":4,"period":"82c662be-66b5-4c4f-b24e-ff9298fc8a6e","realm_epoch":2},"markers":[{"key":0,"val":{"state":1,"marker":"1_1774029984.700846_244.1","next_step_marker":"","total_entries":1,"pos":0,"timestamp":"2026-03-20T18:06:24.700846Z","realm_epoch":2}},{"key":1,"val":{"state":1,"marker":"1_1774029751.502359_72.1","next_step_marker":"","total_entries":0,"pos":0,"timestamp":"2026-03-20T18:02:31.502359Z","realm_epoch":2}},{"key":2,"val":{"state":1,"marker":"1_1774029751.504947_186.1","next_step_marker":"","total_entries":0,"pos":0,"timestamp":"2026-03-20T18:02:31.504947Z","realm_epoch":2}},{"key":3,"val":{"state":1,"marker":"1_1774029984.693472_148.1","next_step_marker":"","total_entries":0,"pos":0,"timestamp":"2026-03-20T18:06:24.693472Z","realm_epoch":2}}]},"full_sync":{"total":1,"complete":1},"current_time":"2026-03-20T18:06:24Z"} 2026-03-20T18:06:24.972 INFO:rgw_multi.tests:finish meta checkpoint for zone=a2 2026-03-20T18:06:24.987 INFO:tasks.util.rgw:rgwadmin: c1.client.0 : ['bucket', 'sync', 'checkpoint', '--bucket', 'cyjkqa-16', '--source-zone', 'a1', '--retry-delay-ms', '5000', '--timeout-sec', '300', '--rgw-zone', 'a2', '--rgw-zonegroup', 'a', '--rgw-realm', 'test-realm', '--cluster', 'c1', '--debug-rgw', '1', '--debug-ms', '0'] 2026-03-20T18:06:24.987 DEBUG:tasks.util.rgw:rgwadmin: cmd=['adjust-ulimits', 'ceph-coverage', '/home/ubuntu/cephtest/archive/coverage', 'radosgw-admin', '--log-to-stderr', '--format', 'json', '-n', 'client.0', '--cluster', 'c1', 'bucket', 'sync', 'checkpoint', '--bucket', 'cyjkqa-16', '--source-zone', 'a1', '--retry-delay-ms', '5000', '--timeout-sec', '300', '--rgw-zone', 'a2', '--rgw-zonegroup', 'a', '--rgw-realm', 'test-realm', '--cluster', 'c1', '--debug-rgw', '1', '--debug-ms', '0'] 2026-03-20T18:06:24.987 DEBUG:teuthology.orchestra.run.vm02:> adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage radosgw-admin --log-to-stderr --format json -n client.0 --cluster c1 bucket sync checkpoint --bucket cyjkqa-16 --source-zone a1 --retry-delay-ms 5000 --timeout-sec 300 --rgw-zone a2 --rgw-zonegroup a --rgw-realm test-realm --cluster c1 --debug-rgw 1 --debug-ms 0 2026-03-20T18:06:25.042 INFO:teuthology.orchestra.run.vm02.stderr:ignoring --setuser ceph since I am not root 2026-03-20T18:06:25.042 INFO:teuthology.orchestra.run.vm02.stderr:ignoring --setgroup ceph since I am not root 2026-03-20T18:06:25.116 INFO:teuthology.orchestra.run.vm02.stderr:2026-03-20T18:06:25.114+0000 7f3ed76e0a80 1 waiting to reach incremental sync.. 2026-03-20T18:06:30.118 INFO:teuthology.orchestra.run.vm02.stderr:2026-03-20T18:06:30.114+0000 7f3ed76e0a80 1 bucket sync caught up with source: 2026-03-20T18:06:30.118 INFO:teuthology.orchestra.run.vm02.stderr: local status: [, , , , , 00000000002.34.11, , , , , ] 2026-03-20T18:06:30.118 INFO:teuthology.orchestra.run.vm02.stderr: remote markers: [, , , , , 00000000002.34.11, , , , , ] 2026-03-20T18:06:30.118 INFO:teuthology.orchestra.run.vm02.stderr:2026-03-20T18:06:30.114+0000 7f3ed76e0a80 0 bucket checkpoint complete 2026-03-20T18:06:30.131 INFO:tasks.util.rgw:rgwadmin: c1.client.0 : ['bucket', 'sync', 'checkpoint', '--bucket', 'cyjkqa-16', '--source-zone', 'a2', '--retry-delay-ms', '5000', '--timeout-sec', '300', '--rgw-zone', 'a1', '--rgw-zonegroup', 'a', '--rgw-realm', 'test-realm', '--cluster', 'c1', '--debug-rgw', '1', '--debug-ms', '0'] 2026-03-20T18:06:30.131 DEBUG:tasks.util.rgw:rgwadmin: cmd=['adjust-ulimits', 'ceph-coverage', '/home/ubuntu/cephtest/archive/coverage', 'radosgw-admin', '--log-to-stderr', '--format', 'json', '-n', 'client.0', '--cluster', 'c1', 'bucket', 'sync', 'checkpoint', '--bucket', 'cyjkqa-16', '--source-zone', 'a2', '--retry-delay-ms', '5000', '--timeout-sec', '300', '--rgw-zone', 'a1', '--rgw-zonegroup', 'a', '--rgw-realm', 'test-realm', '--cluster', 'c1', '--debug-rgw', '1', '--debug-ms', '0'] 2026-03-20T18:06:30.131 DEBUG:teuthology.orchestra.run.vm02:> adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage radosgw-admin --log-to-stderr --format json -n client.0 --cluster c1 bucket sync checkpoint --bucket cyjkqa-16 --source-zone a2 --retry-delay-ms 5000 --timeout-sec 300 --rgw-zone a1 --rgw-zonegroup a --rgw-realm test-realm --cluster c1 --debug-rgw 1 --debug-ms 0 2026-03-20T18:06:30.200 INFO:teuthology.orchestra.run.vm02.stderr:ignoring --setuser ceph since I am not root 2026-03-20T18:06:30.200 INFO:teuthology.orchestra.run.vm02.stderr:ignoring --setgroup ceph since I am not root 2026-03-20T18:06:30.290 INFO:teuthology.orchestra.run.vm02.stderr:2026-03-20T18:06:30.286+0000 7f343b912a80 1 waiting to reach incremental sync.. 2026-03-20T18:06:35.291 INFO:teuthology.orchestra.run.vm02.stderr:2026-03-20T18:06:35.290+0000 7f343b912a80 1 bucket sync caught up with source: 2026-03-20T18:06:35.291 INFO:teuthology.orchestra.run.vm02.stderr: local status: [, , , , , 00000000002.39.11, , , , , ] 2026-03-20T18:06:35.291 INFO:teuthology.orchestra.run.vm02.stderr: remote markers: [, , , , , 00000000002.39.11, , , , , ] 2026-03-20T18:06:35.291 INFO:teuthology.orchestra.run.vm02.stderr:2026-03-20T18:06:35.290+0000 7f343b912a80 0 bucket checkpoint complete 2026-03-20T18:06:35.300 INFO:tasks.rgw_multi.zone_rados:comparing bucket=cyjkqa-16 zones={a2, a1} 2026-03-20T18:06:35.309 DEBUG:tasks.rgw_multi.zone_rados:bucket1 objects: 2026-03-20T18:06:35.309 DEBUG:tasks.rgw_multi.zone_rados:o=dummy, v=hjxJCgLzR8NX9-CihSoXGbd5Rr.anPK 2026-03-20T18:06:35.312 DEBUG:tasks.rgw_multi.zone_rados:bucket2 objects: 2026-03-20T18:06:35.312 DEBUG:tasks.rgw_multi.zone_rados:o=dummy, v=hjxJCgLzR8NX9-CihSoXGbd5Rr.anPK 2026-03-20T18:06:35.312 DEBUG:tasks.rgw_multi.zone_rados:comparing key name=dummy 2026-03-20T18:06:35.327 INFO:tasks.rgw_multi.zone_rados:success, bucket identical: bucket=cyjkqa-16 zones={a2, a1} 2026-03-20T18:06:35.518 INFO:tasks.util.rgw:rgwadmin: c1.client.0 : ['bucket', 'sync', 'checkpoint', '--bucket', 'cyjkqa-16', '--source-zone', 'a1', '--retry-delay-ms', '5000', '--timeout-sec', '300', '--rgw-zone', 'a2', '--rgw-zonegroup', 'a', '--rgw-realm', 'test-realm', '--cluster', 'c1', '--debug-rgw', '1', '--debug-ms', '0'] 2026-03-20T18:06:35.518 DEBUG:tasks.util.rgw:rgwadmin: cmd=['adjust-ulimits', 'ceph-coverage', '/home/ubuntu/cephtest/archive/coverage', 'radosgw-admin', '--log-to-stderr', '--format', 'json', '-n', 'client.0', '--cluster', 'c1', 'bucket', 'sync', 'checkpoint', '--bucket', 'cyjkqa-16', '--source-zone', 'a1', '--retry-delay-ms', '5000', '--timeout-sec', '300', '--rgw-zone', 'a2', '--rgw-zonegroup', 'a', '--rgw-realm', 'test-realm', '--cluster', 'c1', '--debug-rgw', '1', '--debug-ms', '0'] 2026-03-20T18:06:35.518 DEBUG:teuthology.orchestra.run.vm02:> adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage radosgw-admin --log-to-stderr --format json -n client.0 --cluster c1 bucket sync checkpoint --bucket cyjkqa-16 --source-zone a1 --retry-delay-ms 5000 --timeout-sec 300 --rgw-zone a2 --rgw-zonegroup a --rgw-realm test-realm --cluster c1 --debug-rgw 1 --debug-ms 0 2026-03-20T18:06:35.542 INFO:teuthology.orchestra.run.vm02.stderr:ignoring --setuser ceph since I am not root 2026-03-20T18:06:35.542 INFO:teuthology.orchestra.run.vm02.stderr:ignoring --setgroup ceph since I am not root 2026-03-20T18:06:35.642 INFO:teuthology.orchestra.run.vm02.stderr:2026-03-20T18:06:35.638+0000 7fd7604ffa80 1 waiting for incremental sync to catch up: 2026-03-20T18:06:35.642 INFO:teuthology.orchestra.run.vm02.stderr: local status: [, , , , , 00000000002.34.11, , , , , ] 2026-03-20T18:06:35.642 INFO:teuthology.orchestra.run.vm02.stderr: remote markers: [, , , , , 00000000003.36.10, , , 00000000020.71.12, , ] 2026-03-20T18:06:40.643 INFO:teuthology.orchestra.run.vm02.stderr:2026-03-20T18:06:40.642+0000 7fd7604ffa80 1 waiting for incremental sync to catch up: 2026-03-20T18:06:40.643 INFO:teuthology.orchestra.run.vm02.stderr: local status: [, , , , , 00000000002.34.11, , , 00000000031.102.12, , ] 2026-03-20T18:06:40.643 INFO:teuthology.orchestra.run.vm02.stderr: remote markers: [, , , , , 00000000003.36.10, , , 00000000020.71.12, , ] 2026-03-20T18:06:45.644 INFO:teuthology.orchestra.run.vm02.stderr:2026-03-20T18:06:45.642+0000 7fd7604ffa80 1 waiting for incremental sync to catch up: 2026-03-20T18:06:45.644 INFO:teuthology.orchestra.run.vm02.stderr: local status: [, , , , , 00000000002.34.11, , , 00000000031.102.12, , ] 2026-03-20T18:06:45.644 INFO:teuthology.orchestra.run.vm02.stderr: remote markers: [, , , , , 00000000003.36.10, , , 00000000020.71.12, , ] 2026-03-20T18:06:50.645 INFO:teuthology.orchestra.run.vm02.stderr:2026-03-20T18:06:50.642+0000 7fd7604ffa80 1 bucket sync caught up with source: 2026-03-20T18:06:50.645 INFO:teuthology.orchestra.run.vm02.stderr: local status: [, , , , , 00000000003.36.10, , , 00000000031.102.12, , ] 2026-03-20T18:06:50.645 INFO:teuthology.orchestra.run.vm02.stderr: remote markers: [, , , , , 00000000003.36.10, , , 00000000020.71.12, , ] 2026-03-20T18:06:50.645 INFO:teuthology.orchestra.run.vm02.stderr:2026-03-20T18:06:50.642+0000 7fd7604ffa80 0 bucket checkpoint complete 2026-03-20T18:06:50.655 INFO:tasks.util.rgw:rgwadmin: c1.client.0 : ['bucket', 'sync', 'checkpoint', '--bucket', 'cyjkqa-16', '--source-zone', 'a2', '--retry-delay-ms', '5000', '--timeout-sec', '300', '--rgw-zone', 'a1', '--rgw-zonegroup', 'a', '--rgw-realm', 'test-realm', '--cluster', 'c1', '--debug-rgw', '1', '--debug-ms', '0'] 2026-03-20T18:06:50.655 DEBUG:tasks.util.rgw:rgwadmin: cmd=['adjust-ulimits', 'ceph-coverage', '/home/ubuntu/cephtest/archive/coverage', 'radosgw-admin', '--log-to-stderr', '--format', 'json', '-n', 'client.0', '--cluster', 'c1', 'bucket', 'sync', 'checkpoint', '--bucket', 'cyjkqa-16', '--source-zone', 'a2', '--retry-delay-ms', '5000', '--timeout-sec', '300', '--rgw-zone', 'a1', '--rgw-zonegroup', 'a', '--rgw-realm', 'test-realm', '--cluster', 'c1', '--debug-rgw', '1', '--debug-ms', '0'] 2026-03-20T18:06:50.655 DEBUG:teuthology.orchestra.run.vm02:> adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage radosgw-admin --log-to-stderr --format json -n client.0 --cluster c1 bucket sync checkpoint --bucket cyjkqa-16 --source-zone a2 --retry-delay-ms 5000 --timeout-sec 300 --rgw-zone a1 --rgw-zonegroup a --rgw-realm test-realm --cluster c1 --debug-rgw 1 --debug-ms 0 2026-03-20T18:06:50.721 INFO:teuthology.orchestra.run.vm02.stderr:ignoring --setuser ceph since I am not root 2026-03-20T18:06:50.721 INFO:teuthology.orchestra.run.vm02.stderr:ignoring --setgroup ceph since I am not root 2026-03-20T18:06:50.791 INFO:teuthology.orchestra.run.vm02.stderr:2026-03-20T18:06:50.790+0000 7fb1d458ba80 1 bucket sync caught up with source: 2026-03-20T18:06:50.791 INFO:teuthology.orchestra.run.vm02.stderr: local status: [, , , , , 00000000002.39.11, , , 00000000031.105.9, , ] 2026-03-20T18:06:50.791 INFO:teuthology.orchestra.run.vm02.stderr: remote markers: [, , , , , 00000000002.39.11, , , 00000000031.105.9, , ] 2026-03-20T18:06:50.791 INFO:teuthology.orchestra.run.vm02.stderr:2026-03-20T18:06:50.790+0000 7fb1d458ba80 0 bucket checkpoint complete 2026-03-20T18:06:50.802 INFO:tasks.rgw_multi.zone_rados:comparing bucket=cyjkqa-16 zones={a2, a1} 2026-03-20T18:06:50.809 DEBUG:tasks.rgw_multi.zone_rados:bucket1 objects: 2026-03-20T18:06:50.809 DEBUG:tasks.rgw_multi.zone_rados:o=dummy, v=hjxJCgLzR8NX9-CihSoXGbd5Rr.anPK 2026-03-20T18:06:50.809 DEBUG:tasks.rgw_multi.zone_rados:o=obj, v=LiKLUXPhO97DVZ-ZPP1rtLXDCRY02nw 2026-03-20T18:06:50.809 DEBUG:tasks.rgw_multi.zone_rados:o=obj, v=Q4oMth0D1hkk5UlbYPszseaQ9kXi1V4 2026-03-20T18:06:50.809 DEBUG:tasks.rgw_multi.zone_rados:o=obj, v=Wlv0DCLyXwuoOIJJZSwudcvMyyiejC0 2026-03-20T18:06:50.809 DEBUG:tasks.rgw_multi.zone_rados:o=obj, v=gbuBCU47qYp3wcBzlJCFlwOkXraxR7S 2026-03-20T18:06:50.809 DEBUG:tasks.rgw_multi.zone_rados:o=obj, v=QJcoHR6le6XVBPL6WQo-OZDrDkr.eFh 2026-03-20T18:06:50.809 DEBUG:tasks.rgw_multi.zone_rados:o=obj, v=XlE.YBbWLP90seQtR6p3oRbum4Z8.EL 2026-03-20T18:06:50.809 DEBUG:tasks.rgw_multi.zone_rados:o=obj, v=Cq0dU8AYGZl9zRlgJ9mxOoqO8KJGsmR 2026-03-20T18:06:50.809 DEBUG:tasks.rgw_multi.zone_rados:o=obj, v=ll5P6RBeSd81rqT2D72IOL.7c0zDain 2026-03-20T18:06:50.809 DEBUG:tasks.rgw_multi.zone_rados:o=obj, v=He8bcZB0OqfQvpt705UC-4eGy0yzj6A 2026-03-20T18:06:50.809 DEBUG:tasks.rgw_multi.zone_rados:o=obj, v=u7-0r1n4WD3asCR2pRKYXKL9EWbNJaS 2026-03-20T18:06:50.810 DEBUG:tasks.rgw_multi.zone_rados:o=obj, v=bsNyB78Obcuh6Cco6KjltuAMj-RYvfN 2026-03-20T18:06:50.810 DEBUG:tasks.rgw_multi.zone_rados:o=obj, v=xJc994aYk8L26YLDnvksQXcbS3l6Zqq 2026-03-20T18:06:50.810 DEBUG:tasks.rgw_multi.zone_rados:o=obj, v=3VH1Vo6STEY7tUaRobiRQUp9eJO2Aig 2026-03-20T18:06:50.810 DEBUG:tasks.rgw_multi.zone_rados:o=obj, v=XfHzNiPenr5Xm9s5lXCUItg-.qGb8Hh 2026-03-20T18:06:50.810 DEBUG:tasks.rgw_multi.zone_rados:o=obj, v=BhI7jju-RmHAJnQgdJGaDdO8X3ylhZY 2026-03-20T18:06:50.810 DEBUG:tasks.rgw_multi.zone_rados:o=obj, v=UZFnI6Cw4.PeOv6mZdtyL3jn2IPcUfT 2026-03-20T18:06:50.810 DEBUG:tasks.rgw_multi.zone_rados:o=obj, v=ucPFr8bMfgUz8gu05kcPoK.TDbnu9iP 2026-03-20T18:06:50.810 DEBUG:tasks.rgw_multi.zone_rados:o=obj, v=zcVXzRZW.9RWhflPz9ddn3PPse.KM3O 2026-03-20T18:06:50.810 DEBUG:tasks.rgw_multi.zone_rados:o=obj, v=9aCJ.ZpMajCmStS5jzekYB50pyM-V3- 2026-03-20T18:06:50.810 DEBUG:tasks.rgw_multi.zone_rados:o=obj, v=NV.WyIP8YuQg9kpciRmM90wZU32iJt5 2026-03-20T18:06:50.813 DEBUG:tasks.rgw_multi.zone_rados:bucket2 objects: 2026-03-20T18:06:50.813 DEBUG:tasks.rgw_multi.zone_rados:o=dummy, v=hjxJCgLzR8NX9-CihSoXGbd5Rr.anPK 2026-03-20T18:06:50.813 DEBUG:tasks.rgw_multi.zone_rados:o=obj, v=LiKLUXPhO97DVZ-ZPP1rtLXDCRY02nw 2026-03-20T18:06:50.813 DEBUG:tasks.rgw_multi.zone_rados:o=obj, v=Q4oMth0D1hkk5UlbYPszseaQ9kXi1V4 2026-03-20T18:06:50.813 DEBUG:tasks.rgw_multi.zone_rados:o=obj, v=Wlv0DCLyXwuoOIJJZSwudcvMyyiejC0 2026-03-20T18:06:50.814 DEBUG:tasks.rgw_multi.zone_rados:o=obj, v=gbuBCU47qYp3wcBzlJCFlwOkXraxR7S 2026-03-20T18:06:50.814 DEBUG:tasks.rgw_multi.zone_rados:o=obj, v=QJcoHR6le6XVBPL6WQo-OZDrDkr.eFh 2026-03-20T18:06:50.814 DEBUG:tasks.rgw_multi.zone_rados:o=obj, v=XlE.YBbWLP90seQtR6p3oRbum4Z8.EL 2026-03-20T18:06:50.814 DEBUG:tasks.rgw_multi.zone_rados:o=obj, v=Cq0dU8AYGZl9zRlgJ9mxOoqO8KJGsmR 2026-03-20T18:06:50.814 DEBUG:tasks.rgw_multi.zone_rados:o=obj, v=ll5P6RBeSd81rqT2D72IOL.7c0zDain 2026-03-20T18:06:50.814 DEBUG:tasks.rgw_multi.zone_rados:o=obj, v=He8bcZB0OqfQvpt705UC-4eGy0yzj6A 2026-03-20T18:06:50.814 DEBUG:tasks.rgw_multi.zone_rados:o=obj, v=u7-0r1n4WD3asCR2pRKYXKL9EWbNJaS 2026-03-20T18:06:50.814 DEBUG:tasks.rgw_multi.zone_rados:o=obj, v=bsNyB78Obcuh6Cco6KjltuAMj-RYvfN 2026-03-20T18:06:50.814 DEBUG:tasks.rgw_multi.zone_rados:o=obj, v=xJc994aYk8L26YLDnvksQXcbS3l6Zqq 2026-03-20T18:06:50.814 DEBUG:tasks.rgw_multi.zone_rados:o=obj, v=3VH1Vo6STEY7tUaRobiRQUp9eJO2Aig 2026-03-20T18:06:50.814 DEBUG:tasks.rgw_multi.zone_rados:o=obj, v=XfHzNiPenr5Xm9s5lXCUItg-.qGb8Hh 2026-03-20T18:06:50.814 DEBUG:tasks.rgw_multi.zone_rados:o=obj, v=BhI7jju-RmHAJnQgdJGaDdO8X3ylhZY 2026-03-20T18:06:50.814 DEBUG:tasks.rgw_multi.zone_rados:o=obj, v=UZFnI6Cw4.PeOv6mZdtyL3jn2IPcUfT 2026-03-20T18:06:50.814 DEBUG:tasks.rgw_multi.zone_rados:o=obj, v=ucPFr8bMfgUz8gu05kcPoK.TDbnu9iP 2026-03-20T18:06:50.814 DEBUG:tasks.rgw_multi.zone_rados:o=obj, v=zcVXzRZW.9RWhflPz9ddn3PPse.KM3O 2026-03-20T18:06:50.814 DEBUG:tasks.rgw_multi.zone_rados:o=obj, v=9aCJ.ZpMajCmStS5jzekYB50pyM-V3- 2026-03-20T18:06:50.814 DEBUG:tasks.rgw_multi.zone_rados:o=obj, v=NV.WyIP8YuQg9kpciRmM90wZU32iJt5 2026-03-20T18:06:50.814 DEBUG:tasks.rgw_multi.zone_rados:comparing key name=dummy 2026-03-20T18:06:50.825 DEBUG:tasks.rgw_multi.zone_rados:comparing key name=obj 2026-03-20T18:06:50.959 DEBUG:tasks.rgw_multi.zone_rados:comparing key name=obj 2026-03-20T18:06:51.020 DEBUG:tasks.rgw_multi.zone_rados:comparing key name=obj 2026-03-20T18:06:51.076 DEBUG:tasks.rgw_multi.zone_rados:comparing key name=obj 2026-03-20T18:06:51.132 DEBUG:tasks.rgw_multi.zone_rados:comparing key name=obj 2026-03-20T18:06:51.193 DEBUG:tasks.rgw_multi.zone_rados:comparing key name=obj 2026-03-20T18:06:51.252 DEBUG:tasks.rgw_multi.zone_rados:comparing key name=obj 2026-03-20T18:06:51.315 DEBUG:tasks.rgw_multi.zone_rados:comparing key name=obj 2026-03-20T18:06:51.376 DEBUG:tasks.rgw_multi.zone_rados:comparing key name=obj 2026-03-20T18:06:51.437 DEBUG:tasks.rgw_multi.zone_rados:comparing key name=obj 2026-03-20T18:06:51.497 DEBUG:tasks.rgw_multi.zone_rados:comparing key name=obj 2026-03-20T18:06:51.552 DEBUG:tasks.rgw_multi.zone_rados:comparing key name=obj 2026-03-20T18:06:51.613 DEBUG:tasks.rgw_multi.zone_rados:comparing key name=obj 2026-03-20T18:06:51.672 DEBUG:tasks.rgw_multi.zone_rados:comparing key name=obj 2026-03-20T18:06:51.732 DEBUG:tasks.rgw_multi.zone_rados:comparing key name=obj 2026-03-20T18:06:51.792 DEBUG:tasks.rgw_multi.zone_rados:comparing key name=obj 2026-03-20T18:06:51.852 DEBUG:tasks.rgw_multi.zone_rados:comparing key name=obj 2026-03-20T18:06:51.912 DEBUG:tasks.rgw_multi.zone_rados:comparing key name=obj 2026-03-20T18:06:51.968 DEBUG:tasks.rgw_multi.zone_rados:comparing key name=obj 2026-03-20T18:06:52.024 DEBUG:tasks.rgw_multi.zone_rados:comparing key name=obj 2026-03-20T18:06:52.081 INFO:tasks.rgw_multi.zone_rados:success, bucket identical: bucket=cyjkqa-16 zones={a2, a1} 2026-03-20T18:06:52.081 INFO:tasks.util.rgw:rgwadmin: c1.client.0 : ['datalog', 'status', '--rgw-zone', 'a1', '--rgw-zonegroup', 'a', '--rgw-realm', 'test-realm', '--cluster', 'c1', '--debug-rgw', '0', '--debug-ms', '0', '--rgw-cache-enabled', 'false'] 2026-03-20T18:06:52.081 DEBUG:tasks.util.rgw:rgwadmin: cmd=['adjust-ulimits', 'ceph-coverage', '/home/ubuntu/cephtest/archive/coverage', 'radosgw-admin', '--log-to-stderr', '--format', 'json', '-n', 'client.0', '--cluster', 'c1', 'datalog', 'status', '--rgw-zone', 'a1', '--rgw-zonegroup', 'a', '--rgw-realm', 'test-realm', '--cluster', 'c1', '--debug-rgw', '0', '--debug-ms', '0', '--rgw-cache-enabled', 'false'] 2026-03-20T18:06:52.081 DEBUG:teuthology.orchestra.run.vm02:> adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage radosgw-admin --log-to-stderr --format json -n client.0 --cluster c1 datalog status --rgw-zone a1 --rgw-zonegroup a --rgw-realm test-realm --cluster c1 --debug-rgw 0 --debug-ms 0 --rgw-cache-enabled false 2026-03-20T18:06:52.103 INFO:teuthology.orchestra.run.vm02.stderr:ignoring --setuser ceph since I am not root 2026-03-20T18:06:52.103 INFO:teuthology.orchestra.run.vm02.stderr:ignoring --setgroup ceph since I am not root 2026-03-20T18:06:52.204 INFO:teuthology.orchestra.run.vm02.stdout:[{"marker":"00000000000000000000:00000000000000005046","last_update":"2026-03-20T18:03:46.500419Z"},{"marker":"00000000000000000000:00000000000000003499","last_update":"2026-03-20T18:03:46.502763Z"},{"marker":"00000000000000000000:00000000000000005614","last_update":"2026-03-20T18:06:42.537311Z"},{"marker":"00000000000000000000:00000000000000005884","last_update":"2026-03-20T18:06:42.539343Z"}] 2026-03-20T18:06:52.205 INFO:rgw_multi.tests:starting data checkpoint for target_zone=a2 source_zone=a1 2026-03-20T18:06:52.205 INFO:tasks.util.rgw:rgwadmin: c1.client.0 : ['data', 'sync', 'status', '--rgw-zone', 'a2', '--rgw-zonegroup', 'a', '--rgw-realm', 'test-realm', '--source-zone', 'a1', '--cluster', 'c1', '--debug-rgw', '0', '--debug-ms', '0', '--rgw-cache-enabled', 'false'] 2026-03-20T18:06:52.205 DEBUG:tasks.util.rgw:rgwadmin: cmd=['adjust-ulimits', 'ceph-coverage', '/home/ubuntu/cephtest/archive/coverage', 'radosgw-admin', '--log-to-stderr', '--format', 'json', '-n', 'client.0', '--cluster', 'c1', 'data', 'sync', 'status', '--rgw-zone', 'a2', '--rgw-zonegroup', 'a', '--rgw-realm', 'test-realm', '--source-zone', 'a1', '--cluster', 'c1', '--debug-rgw', '0', '--debug-ms', '0', '--rgw-cache-enabled', 'false'] 2026-03-20T18:06:52.205 DEBUG:teuthology.orchestra.run.vm02:> adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage radosgw-admin --log-to-stderr --format json -n client.0 --cluster c1 data sync status --rgw-zone a2 --rgw-zonegroup a --rgw-realm test-realm --source-zone a1 --cluster c1 --debug-rgw 0 --debug-ms 0 --rgw-cache-enabled false 2026-03-20T18:06:52.267 INFO:teuthology.orchestra.run.vm02.stderr:ignoring --setuser ceph since I am not root 2026-03-20T18:06:52.267 INFO:teuthology.orchestra.run.vm02.stderr:ignoring --setgroup ceph since I am not root 2026-03-20T18:06:52.380 INFO:teuthology.orchestra.run.vm02.stdout:{"sync_status":{"info":{"status":"sync","num_shards":4,"instance_id":11856732089314853152},"markers":[{"key":0,"val":{"status":"incremental-sync","marker":"00000000000000000000:00000000000000005046","next_step_marker":"","total_entries":0,"pos":0,"timestamp":"2026-03-20T18:03:46.500419Z"}},{"key":1,"val":{"status":"incremental-sync","marker":"00000000000000000000:00000000000000003499","next_step_marker":"","total_entries":0,"pos":0,"timestamp":"2026-03-20T18:03:46.502763Z"}},{"key":2,"val":{"status":"incremental-sync","marker":"00000000000000000000:00000000000000005614","next_step_marker":"","total_entries":0,"pos":0,"timestamp":"2026-03-20T18:06:42.537311Z"}},{"key":3,"val":{"status":"incremental-sync","marker":"00000000000000000000:00000000000000005884","next_step_marker":"","total_entries":0,"pos":0,"timestamp":"2026-03-20T18:06:42.539343Z"}}]},"full_sync":{"total":0,"complete":0},"current_time":"2026-03-20T18:06:52Z"} 2026-03-20T18:06:52.380 INFO:rgw_multi.tests:finished data checkpoint for target_zone=a2 source_zone=a1 2026-03-20T18:06:52.380 INFO:tasks.util.rgw:rgwadmin: c1.client.0 : ['datalog', 'status', '--rgw-zone', 'a2', '--rgw-zonegroup', 'a', '--rgw-realm', 'test-realm', '--cluster', 'c1', '--debug-rgw', '0', '--debug-ms', '0', '--rgw-cache-enabled', 'false'] 2026-03-20T18:06:52.380 DEBUG:tasks.util.rgw:rgwadmin: cmd=['adjust-ulimits', 'ceph-coverage', '/home/ubuntu/cephtest/archive/coverage', 'radosgw-admin', '--log-to-stderr', '--format', 'json', '-n', 'client.0', '--cluster', 'c1', 'datalog', 'status', '--rgw-zone', 'a2', '--rgw-zonegroup', 'a', '--rgw-realm', 'test-realm', '--cluster', 'c1', '--debug-rgw', '0', '--debug-ms', '0', '--rgw-cache-enabled', 'false'] 2026-03-20T18:06:52.380 DEBUG:teuthology.orchestra.run.vm02:> adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage radosgw-admin --log-to-stderr --format json -n client.0 --cluster c1 datalog status --rgw-zone a2 --rgw-zonegroup a --rgw-realm test-realm --cluster c1 --debug-rgw 0 --debug-ms 0 --rgw-cache-enabled false 2026-03-20T18:06:52.441 INFO:teuthology.orchestra.run.vm02.stderr:ignoring --setuser ceph since I am not root 2026-03-20T18:06:52.442 INFO:teuthology.orchestra.run.vm02.stderr:ignoring --setgroup ceph since I am not root 2026-03-20T18:06:52.522 INFO:teuthology.orchestra.run.vm02.stdout:[{"marker":"00000000000000000000:00000000000000005470","last_update":"2026-03-20T18:03:44.116332Z"},{"marker":"00000000000000000000:00000000000000003781","last_update":"2026-03-20T18:03:44.117950Z"},{"marker":"00000000000000000000:00000000000000006037","last_update":"2026-03-20T18:06:40.147671Z"},{"marker":"00000000000000000000:00000000000000006307","last_update":"2026-03-20T18:06:40.150286Z"}] 2026-03-20T18:06:52.522 INFO:rgw_multi.tests:starting data checkpoint for target_zone=a1 source_zone=a2 2026-03-20T18:06:52.522 INFO:tasks.util.rgw:rgwadmin: c1.client.0 : ['data', 'sync', 'status', '--rgw-zone', 'a1', '--rgw-zonegroup', 'a', '--rgw-realm', 'test-realm', '--source-zone', 'a2', '--cluster', 'c1', '--debug-rgw', '0', '--debug-ms', '0', '--rgw-cache-enabled', 'false'] 2026-03-20T18:06:52.522 DEBUG:tasks.util.rgw:rgwadmin: cmd=['adjust-ulimits', 'ceph-coverage', '/home/ubuntu/cephtest/archive/coverage', 'radosgw-admin', '--log-to-stderr', '--format', 'json', '-n', 'client.0', '--cluster', 'c1', 'data', 'sync', 'status', '--rgw-zone', 'a1', '--rgw-zonegroup', 'a', '--rgw-realm', 'test-realm', '--source-zone', 'a2', '--cluster', 'c1', '--debug-rgw', '0', '--debug-ms', '0', '--rgw-cache-enabled', 'false'] 2026-03-20T18:06:52.522 DEBUG:teuthology.orchestra.run.vm02:> adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage radosgw-admin --log-to-stderr --format json -n client.0 --cluster c1 data sync status --rgw-zone a1 --rgw-zonegroup a --rgw-realm test-realm --source-zone a2 --cluster c1 --debug-rgw 0 --debug-ms 0 --rgw-cache-enabled false 2026-03-20T18:06:52.587 INFO:teuthology.orchestra.run.vm02.stderr:ignoring --setuser ceph since I am not root 2026-03-20T18:06:52.587 INFO:teuthology.orchestra.run.vm02.stderr:ignoring --setgroup ceph since I am not root 2026-03-20T18:06:52.669 INFO:teuthology.orchestra.run.vm02.stdout:{"sync_status":{"info":{"status":"sync","num_shards":4,"instance_id":15343266579503146597},"markers":[{"key":0,"val":{"status":"incremental-sync","marker":"00000000000000000000:00000000000000005470","next_step_marker":"","total_entries":0,"pos":0,"timestamp":"2026-03-20T18:03:44.116332Z"}},{"key":1,"val":{"status":"incremental-sync","marker":"00000000000000000000:00000000000000003781","next_step_marker":"","total_entries":0,"pos":0,"timestamp":"2026-03-20T18:03:44.117950Z"}},{"key":2,"val":{"status":"incremental-sync","marker":"00000000000000000000:00000000000000006037","next_step_marker":"","total_entries":0,"pos":0,"timestamp":"2026-03-20T18:06:40.147671Z"}},{"key":3,"val":{"status":"incremental-sync","marker":"00000000000000000000:00000000000000006307","next_step_marker":"","total_entries":0,"pos":0,"timestamp":"2026-03-20T18:06:40.150286Z"}}]},"full_sync":{"total":0,"complete":0},"current_time":"2026-03-20T18:06:52Z"} 2026-03-20T18:06:52.669 INFO:rgw_multi.tests:finished data checkpoint for target_zone=a1 source_zone=a2 2026-03-20T18:06:52.672 INFO:tasks.rgw_multisite_tests:rgw_multi.tests.test_concurrent_versioned_object_incremental_sync ... ok 2026-03-20T18:06:52.792 INFO:tasks.util.rgw:rgwadmin: c1.client.0 : ['mdlog', 'status', '--rgw-zone', 'a1', '--rgw-zonegroup', 'a', '--rgw-realm', 'test-realm', '--cluster', 'c1', '--debug-rgw', '0', '--debug-ms', '0', '--rgw-cache-enabled', 'false'] 2026-03-20T18:06:52.792 DEBUG:tasks.util.rgw:rgwadmin: cmd=['adjust-ulimits', 'ceph-coverage', '/home/ubuntu/cephtest/archive/coverage', 'radosgw-admin', '--log-to-stderr', '--format', 'json', '-n', 'client.0', '--cluster', 'c1', 'mdlog', 'status', '--rgw-zone', 'a1', '--rgw-zonegroup', 'a', '--rgw-realm', 'test-realm', '--cluster', 'c1', '--debug-rgw', '0', '--debug-ms', '0', '--rgw-cache-enabled', 'false'] 2026-03-20T18:06:52.793 DEBUG:teuthology.orchestra.run.vm02:> adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage radosgw-admin --log-to-stderr --format json -n client.0 --cluster c1 mdlog status --rgw-zone a1 --rgw-zonegroup a --rgw-realm test-realm --cluster c1 --debug-rgw 0 --debug-ms 0 --rgw-cache-enabled false 2026-03-20T18:06:52.812 INFO:teuthology.orchestra.run.vm02.stderr:ignoring --setuser ceph since I am not root 2026-03-20T18:06:52.812 INFO:teuthology.orchestra.run.vm02.stderr:ignoring --setgroup ceph since I am not root 2026-03-20T18:06:52.879 INFO:teuthology.orchestra.run.vm02.stderr:No --period given, using current period=82c662be-66b5-4c4f-b24e-ff9298fc8a6e 2026-03-20T18:06:52.888 INFO:teuthology.orchestra.run.vm02.stdout:[{"marker":"1_1774030012.787471_251.1","last_update":"2026-03-20T18:06:52.787471Z"},{"marker":"1_1774029751.502359_72.1","last_update":"2026-03-20T18:02:31.502359Z"},{"marker":"1_1774029751.504947_186.1","last_update":"2026-03-20T18:02:31.504947Z"},{"marker":"1_1774030012.789970_152.1","last_update":"2026-03-20T18:06:52.789970Z"}] 2026-03-20T18:06:52.888 INFO:rgw_multi.tests:starting meta checkpoint for zone=a2 2026-03-20T18:06:52.888 INFO:tasks.util.rgw:rgwadmin: c1.client.0 : ['metadata', 'sync', 'status', '--rgw-zone', 'a2', '--rgw-zonegroup', 'a', '--rgw-realm', 'test-realm', '--cluster', 'c1', '--debug-rgw', '0', '--debug-ms', '0', '--rgw-cache-enabled', 'false'] 2026-03-20T18:06:52.888 DEBUG:tasks.util.rgw:rgwadmin: cmd=['adjust-ulimits', 'ceph-coverage', '/home/ubuntu/cephtest/archive/coverage', 'radosgw-admin', '--log-to-stderr', '--format', 'json', '-n', 'client.0', '--cluster', 'c1', 'metadata', 'sync', 'status', '--rgw-zone', 'a2', '--rgw-zonegroup', 'a', '--rgw-realm', 'test-realm', '--cluster', 'c1', '--debug-rgw', '0', '--debug-ms', '0', '--rgw-cache-enabled', 'false'] 2026-03-20T18:06:52.888 DEBUG:teuthology.orchestra.run.vm02:> adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage radosgw-admin --log-to-stderr --format json -n client.0 --cluster c1 metadata sync status --rgw-zone a2 --rgw-zonegroup a --rgw-realm test-realm --cluster c1 --debug-rgw 0 --debug-ms 0 --rgw-cache-enabled false 2026-03-20T18:06:52.951 INFO:teuthology.orchestra.run.vm02.stderr:ignoring --setuser ceph since I am not root 2026-03-20T18:06:52.951 INFO:teuthology.orchestra.run.vm02.stderr:ignoring --setgroup ceph since I am not root 2026-03-20T18:06:53.065 INFO:teuthology.orchestra.run.vm02.stderr:2026-03-20T18:06:53.062+0000 7f1ecb455a80 20 RGW-SYNC:meta: read sync status 2026-03-20T18:06:53.065 INFO:teuthology.orchestra.run.vm02.stderr:2026-03-20T18:06:53.062+0000 7f1ecb455a80 20 RGW-SYNC:meta: read sync status 2026-03-20T18:06:53.116 INFO:teuthology.orchestra.run.vm02.stdout:{"sync_status":{"info":{"status":"sync","num_shards":4,"period":"82c662be-66b5-4c4f-b24e-ff9298fc8a6e","realm_epoch":2},"markers":[{"key":0,"val":{"state":1,"marker":"1_1774030012.787471_251.1","next_step_marker":"","total_entries":1,"pos":0,"timestamp":"2026-03-20T18:06:52.787471Z","realm_epoch":2}},{"key":1,"val":{"state":1,"marker":"1_1774029751.502359_72.1","next_step_marker":"","total_entries":0,"pos":0,"timestamp":"2026-03-20T18:02:31.502359Z","realm_epoch":2}},{"key":2,"val":{"state":1,"marker":"1_1774029751.504947_186.1","next_step_marker":"","total_entries":0,"pos":0,"timestamp":"2026-03-20T18:02:31.504947Z","realm_epoch":2}},{"key":3,"val":{"state":1,"marker":"1_1774030012.789970_152.1","next_step_marker":"","total_entries":0,"pos":0,"timestamp":"2026-03-20T18:06:52.789970Z","realm_epoch":2}}]},"full_sync":{"total":1,"complete":1},"current_time":"2026-03-20T18:06:53Z"} 2026-03-20T18:06:53.116 INFO:rgw_multi.tests:finish meta checkpoint for zone=a2 2026-03-20T18:06:53.124 INFO:tasks.util.rgw:rgwadmin: c1.client.0 : ['bucket', 'sync', 'checkpoint', '--bucket', 'cyjkqa-17', '--source-zone', 'a1', '--retry-delay-ms', '5000', '--timeout-sec', '300', '--rgw-zone', 'a2', '--rgw-zonegroup', 'a', '--rgw-realm', 'test-realm', '--cluster', 'c1', '--debug-rgw', '1', '--debug-ms', '0'] 2026-03-20T18:06:53.124 DEBUG:tasks.util.rgw:rgwadmin: cmd=['adjust-ulimits', 'ceph-coverage', '/home/ubuntu/cephtest/archive/coverage', 'radosgw-admin', '--log-to-stderr', '--format', 'json', '-n', 'client.0', '--cluster', 'c1', 'bucket', 'sync', 'checkpoint', '--bucket', 'cyjkqa-17', '--source-zone', 'a1', '--retry-delay-ms', '5000', '--timeout-sec', '300', '--rgw-zone', 'a2', '--rgw-zonegroup', 'a', '--rgw-realm', 'test-realm', '--cluster', 'c1', '--debug-rgw', '1', '--debug-ms', '0'] 2026-03-20T18:06:53.124 DEBUG:teuthology.orchestra.run.vm02:> adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage radosgw-admin --log-to-stderr --format json -n client.0 --cluster c1 bucket sync checkpoint --bucket cyjkqa-17 --source-zone a1 --retry-delay-ms 5000 --timeout-sec 300 --rgw-zone a2 --rgw-zonegroup a --rgw-realm test-realm --cluster c1 --debug-rgw 1 --debug-ms 0 2026-03-20T18:06:53.181 INFO:teuthology.orchestra.run.vm02.stderr:ignoring --setuser ceph since I am not root 2026-03-20T18:06:53.181 INFO:teuthology.orchestra.run.vm02.stderr:ignoring --setgroup ceph since I am not root 2026-03-20T18:06:53.252 INFO:teuthology.orchestra.run.vm02.stderr:2026-03-20T18:06:53.250+0000 7fba63ec8a80 1 waiting to reach incremental sync.. 2026-03-20T18:06:58.253 INFO:teuthology.orchestra.run.vm02.stderr:2026-03-20T18:06:58.250+0000 7fba63ec8a80 1 bucket sync caught up with source: 2026-03-20T18:06:58.253 INFO:teuthology.orchestra.run.vm02.stderr: local status: [, , , , , , , , 00000000001.26.6, , ] 2026-03-20T18:06:58.253 INFO:teuthology.orchestra.run.vm02.stderr: remote markers: [, , , , , , , , 00000000001.26.6, , ] 2026-03-20T18:06:58.253 INFO:teuthology.orchestra.run.vm02.stderr:2026-03-20T18:06:58.250+0000 7fba63ec8a80 0 bucket checkpoint complete 2026-03-20T18:06:58.264 INFO:tasks.util.rgw:rgwadmin: c1.client.0 : ['bucket', 'sync', 'checkpoint', '--bucket', 'cyjkqa-17', '--source-zone', 'a2', '--retry-delay-ms', '5000', '--timeout-sec', '300', '--rgw-zone', 'a1', '--rgw-zonegroup', 'a', '--rgw-realm', 'test-realm', '--cluster', 'c1', '--debug-rgw', '1', '--debug-ms', '0'] 2026-03-20T18:06:58.264 DEBUG:tasks.util.rgw:rgwadmin: cmd=['adjust-ulimits', 'ceph-coverage', '/home/ubuntu/cephtest/archive/coverage', 'radosgw-admin', '--log-to-stderr', '--format', 'json', '-n', 'client.0', '--cluster', 'c1', 'bucket', 'sync', 'checkpoint', '--bucket', 'cyjkqa-17', '--source-zone', 'a2', '--retry-delay-ms', '5000', '--timeout-sec', '300', '--rgw-zone', 'a1', '--rgw-zonegroup', 'a', '--rgw-realm', 'test-realm', '--cluster', 'c1', '--debug-rgw', '1', '--debug-ms', '0'] 2026-03-20T18:06:58.264 DEBUG:teuthology.orchestra.run.vm02:> adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage radosgw-admin --log-to-stderr --format json -n client.0 --cluster c1 bucket sync checkpoint --bucket cyjkqa-17 --source-zone a2 --retry-delay-ms 5000 --timeout-sec 300 --rgw-zone a1 --rgw-zonegroup a --rgw-realm test-realm --cluster c1 --debug-rgw 1 --debug-ms 0 2026-03-20T18:06:58.330 INFO:teuthology.orchestra.run.vm02.stderr:ignoring --setuser ceph since I am not root 2026-03-20T18:06:58.330 INFO:teuthology.orchestra.run.vm02.stderr:ignoring --setgroup ceph since I am not root 2026-03-20T18:06:58.398 INFO:teuthology.orchestra.run.vm02.stderr:2026-03-20T18:06:58.394+0000 7fbbb15dca80 1 waiting to reach incremental sync.. 2026-03-20T18:07:03.400 INFO:teuthology.orchestra.run.vm02.stderr:2026-03-20T18:07:03.398+0000 7fbbb15dca80 1 bucket sync caught up with source: 2026-03-20T18:07:03.400 INFO:teuthology.orchestra.run.vm02.stderr: local status: [, , , , , , , , 00000000001.31.6, , ] 2026-03-20T18:07:03.400 INFO:teuthology.orchestra.run.vm02.stderr: remote markers: [, , , , , , , , 00000000001.31.6, , ] 2026-03-20T18:07:03.400 INFO:teuthology.orchestra.run.vm02.stderr:2026-03-20T18:07:03.398+0000 7fbbb15dca80 0 bucket checkpoint complete 2026-03-20T18:07:03.413 INFO:tasks.rgw_multi.zone_rados:comparing bucket=cyjkqa-17 zones={a2, a1} 2026-03-20T18:07:03.420 DEBUG:tasks.rgw_multi.zone_rados:bucket1 objects: 2026-03-20T18:07:03.420 DEBUG:tasks.rgw_multi.zone_rados:o=obj, v=null 2026-03-20T18:07:03.423 DEBUG:tasks.rgw_multi.zone_rados:bucket2 objects: 2026-03-20T18:07:03.423 DEBUG:tasks.rgw_multi.zone_rados:o=obj, v=null 2026-03-20T18:07:03.423 DEBUG:tasks.rgw_multi.zone_rados:comparing key name=obj 2026-03-20T18:07:03.436 INFO:tasks.rgw_multi.zone_rados:success, bucket identical: bucket=cyjkqa-17 zones={a2, a1} 2026-03-20T18:07:03.441 INFO:tasks.util.rgw:rgwadmin: c1.client.0 : ['mdlog', 'status', '--rgw-zone', 'a1', '--rgw-zonegroup', 'a', '--rgw-realm', 'test-realm', '--cluster', 'c1', '--debug-rgw', '0', '--debug-ms', '0', '--rgw-cache-enabled', 'false'] 2026-03-20T18:07:03.441 DEBUG:tasks.util.rgw:rgwadmin: cmd=['adjust-ulimits', 'ceph-coverage', '/home/ubuntu/cephtest/archive/coverage', 'radosgw-admin', '--log-to-stderr', '--format', 'json', '-n', 'client.0', '--cluster', 'c1', 'mdlog', 'status', '--rgw-zone', 'a1', '--rgw-zonegroup', 'a', '--rgw-realm', 'test-realm', '--cluster', 'c1', '--debug-rgw', '0', '--debug-ms', '0', '--rgw-cache-enabled', 'false'] 2026-03-20T18:07:03.441 DEBUG:teuthology.orchestra.run.vm02:> adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage radosgw-admin --log-to-stderr --format json -n client.0 --cluster c1 mdlog status --rgw-zone a1 --rgw-zonegroup a --rgw-realm test-realm --cluster c1 --debug-rgw 0 --debug-ms 0 --rgw-cache-enabled false 2026-03-20T18:07:03.475 INFO:teuthology.orchestra.run.vm02.stderr:ignoring --setuser ceph since I am not root 2026-03-20T18:07:03.475 INFO:teuthology.orchestra.run.vm02.stderr:ignoring --setgroup ceph since I am not root 2026-03-20T18:07:03.547 INFO:teuthology.orchestra.run.vm02.stderr:No --period given, using current period=82c662be-66b5-4c4f-b24e-ff9298fc8a6e 2026-03-20T18:07:03.561 INFO:teuthology.orchestra.run.vm02.stdout:[{"marker":"1_1774030023.439907_269.1","last_update":"2026-03-20T18:07:03.439907Z"},{"marker":"1_1774029751.502359_72.1","last_update":"2026-03-20T18:02:31.502359Z"},{"marker":"1_1774029751.504947_186.1","last_update":"2026-03-20T18:02:31.504947Z"},{"marker":"1_1774030012.789970_152.1","last_update":"2026-03-20T18:06:52.789970Z"}] 2026-03-20T18:07:03.561 INFO:rgw_multi.tests:starting meta checkpoint for zone=a2 2026-03-20T18:07:03.562 INFO:tasks.util.rgw:rgwadmin: c1.client.0 : ['metadata', 'sync', 'status', '--rgw-zone', 'a2', '--rgw-zonegroup', 'a', '--rgw-realm', 'test-realm', '--cluster', 'c1', '--debug-rgw', '0', '--debug-ms', '0', '--rgw-cache-enabled', 'false'] 2026-03-20T18:07:03.562 DEBUG:tasks.util.rgw:rgwadmin: cmd=['adjust-ulimits', 'ceph-coverage', '/home/ubuntu/cephtest/archive/coverage', 'radosgw-admin', '--log-to-stderr', '--format', 'json', '-n', 'client.0', '--cluster', 'c1', 'metadata', 'sync', 'status', '--rgw-zone', 'a2', '--rgw-zonegroup', 'a', '--rgw-realm', 'test-realm', '--cluster', 'c1', '--debug-rgw', '0', '--debug-ms', '0', '--rgw-cache-enabled', 'false'] 2026-03-20T18:07:03.562 DEBUG:teuthology.orchestra.run.vm02:> adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage radosgw-admin --log-to-stderr --format json -n client.0 --cluster c1 metadata sync status --rgw-zone a2 --rgw-zonegroup a --rgw-realm test-realm --cluster c1 --debug-rgw 0 --debug-ms 0 --rgw-cache-enabled false 2026-03-20T18:07:03.624 INFO:teuthology.orchestra.run.vm02.stderr:ignoring --setuser ceph since I am not root 2026-03-20T18:07:03.624 INFO:teuthology.orchestra.run.vm02.stderr:ignoring --setgroup ceph since I am not root 2026-03-20T18:07:03.698 INFO:teuthology.orchestra.run.vm02.stderr:2026-03-20T18:07:03.694+0000 7f4e6ec9ba80 20 RGW-SYNC:meta: read sync status 2026-03-20T18:07:03.699 INFO:teuthology.orchestra.run.vm02.stderr:2026-03-20T18:07:03.698+0000 7f4e6ec9ba80 20 RGW-SYNC:meta: read sync status 2026-03-20T18:07:03.710 INFO:teuthology.orchestra.run.vm02.stdout:{"sync_status":{"info":{"status":"sync","num_shards":4,"period":"82c662be-66b5-4c4f-b24e-ff9298fc8a6e","realm_epoch":2},"markers":[{"key":0,"val":{"state":1,"marker":"1_1774030023.439907_269.1","next_step_marker":"","total_entries":1,"pos":0,"timestamp":"2026-03-20T18:07:03.439907Z","realm_epoch":2}},{"key":1,"val":{"state":1,"marker":"1_1774029751.502359_72.1","next_step_marker":"","total_entries":0,"pos":0,"timestamp":"2026-03-20T18:02:31.502359Z","realm_epoch":2}},{"key":2,"val":{"state":1,"marker":"1_1774029751.504947_186.1","next_step_marker":"","total_entries":0,"pos":0,"timestamp":"2026-03-20T18:02:31.504947Z","realm_epoch":2}},{"key":3,"val":{"state":1,"marker":"1_1774030012.789970_152.1","next_step_marker":"","total_entries":0,"pos":0,"timestamp":"2026-03-20T18:06:52.789970Z","realm_epoch":2}}]},"full_sync":{"total":1,"complete":1},"current_time":"2026-03-20T18:07:03Z"} 2026-03-20T18:07:03.710 INFO:rgw_multi.tests:finish meta checkpoint for zone=a2 2026-03-20T18:07:03.720 INFO:tasks.util.rgw:rgwadmin: c1.client.0 : ['bucket', 'sync', 'checkpoint', '--bucket', 'cyjkqa-17', '--source-zone', 'a1', '--retry-delay-ms', '5000', '--timeout-sec', '300', '--rgw-zone', 'a2', '--rgw-zonegroup', 'a', '--rgw-realm', 'test-realm', '--cluster', 'c1', '--debug-rgw', '1', '--debug-ms', '0'] 2026-03-20T18:07:03.720 DEBUG:tasks.util.rgw:rgwadmin: cmd=['adjust-ulimits', 'ceph-coverage', '/home/ubuntu/cephtest/archive/coverage', 'radosgw-admin', '--log-to-stderr', '--format', 'json', '-n', 'client.0', '--cluster', 'c1', 'bucket', 'sync', 'checkpoint', '--bucket', 'cyjkqa-17', '--source-zone', 'a1', '--retry-delay-ms', '5000', '--timeout-sec', '300', '--rgw-zone', 'a2', '--rgw-zonegroup', 'a', '--rgw-realm', 'test-realm', '--cluster', 'c1', '--debug-rgw', '1', '--debug-ms', '0'] 2026-03-20T18:07:03.720 DEBUG:teuthology.orchestra.run.vm02:> adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage radosgw-admin --log-to-stderr --format json -n client.0 --cluster c1 bucket sync checkpoint --bucket cyjkqa-17 --source-zone a1 --retry-delay-ms 5000 --timeout-sec 300 --rgw-zone a2 --rgw-zonegroup a --rgw-realm test-realm --cluster c1 --debug-rgw 1 --debug-ms 0 2026-03-20T18:07:03.776 INFO:teuthology.orchestra.run.vm02.stderr:ignoring --setuser ceph since I am not root 2026-03-20T18:07:03.776 INFO:teuthology.orchestra.run.vm02.stderr:ignoring --setgroup ceph since I am not root 2026-03-20T18:07:03.851 INFO:teuthology.orchestra.run.vm02.stderr:2026-03-20T18:07:03.850+0000 7fccb71d6a80 1 waiting for incremental sync to catch up: 2026-03-20T18:07:03.851 INFO:teuthology.orchestra.run.vm02.stderr: local status: [, , , , , , , , 00000000001.26.6, , ] 2026-03-20T18:07:03.851 INFO:teuthology.orchestra.run.vm02.stderr: remote markers: [, , , , , , , , 00000000003.29.13, , ] 2026-03-20T18:07:08.853 INFO:teuthology.orchestra.run.vm02.stderr:2026-03-20T18:07:08.850+0000 7fccb71d6a80 1 bucket sync caught up with source: 2026-03-20T18:07:08.853 INFO:teuthology.orchestra.run.vm02.stderr: local status: [, , , , , , , , 00000000003.29.13, , ] 2026-03-20T18:07:08.853 INFO:teuthology.orchestra.run.vm02.stderr: remote markers: [, , , , , , , , 00000000003.29.13, , ] 2026-03-20T18:07:08.853 INFO:teuthology.orchestra.run.vm02.stderr:2026-03-20T18:07:08.850+0000 7fccb71d6a80 0 bucket checkpoint complete 2026-03-20T18:07:08.866 INFO:tasks.util.rgw:rgwadmin: c1.client.0 : ['bucket', 'sync', 'checkpoint', '--bucket', 'cyjkqa-17', '--source-zone', 'a2', '--retry-delay-ms', '5000', '--timeout-sec', '300', '--rgw-zone', 'a1', '--rgw-zonegroup', 'a', '--rgw-realm', 'test-realm', '--cluster', 'c1', '--debug-rgw', '1', '--debug-ms', '0'] 2026-03-20T18:07:08.867 DEBUG:tasks.util.rgw:rgwadmin: cmd=['adjust-ulimits', 'ceph-coverage', '/home/ubuntu/cephtest/archive/coverage', 'radosgw-admin', '--log-to-stderr', '--format', 'json', '-n', 'client.0', '--cluster', 'c1', 'bucket', 'sync', 'checkpoint', '--bucket', 'cyjkqa-17', '--source-zone', 'a2', '--retry-delay-ms', '5000', '--timeout-sec', '300', '--rgw-zone', 'a1', '--rgw-zonegroup', 'a', '--rgw-realm', 'test-realm', '--cluster', 'c1', '--debug-rgw', '1', '--debug-ms', '0'] 2026-03-20T18:07:08.867 DEBUG:teuthology.orchestra.run.vm02:> adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage radosgw-admin --log-to-stderr --format json -n client.0 --cluster c1 bucket sync checkpoint --bucket cyjkqa-17 --source-zone a2 --retry-delay-ms 5000 --timeout-sec 300 --rgw-zone a1 --rgw-zonegroup a --rgw-realm test-realm --cluster c1 --debug-rgw 1 --debug-ms 0 2026-03-20T18:07:08.934 INFO:teuthology.orchestra.run.vm02.stderr:ignoring --setuser ceph since I am not root 2026-03-20T18:07:08.934 INFO:teuthology.orchestra.run.vm02.stderr:ignoring --setgroup ceph since I am not root 2026-03-20T18:07:09.014 INFO:teuthology.orchestra.run.vm02.stderr:2026-03-20T18:07:09.010+0000 7f8f75d1da80 1 waiting for incremental sync to catch up: 2026-03-20T18:07:09.014 INFO:teuthology.orchestra.run.vm02.stderr: local status: [, , , , , , , , 00000000001.31.6, , ] 2026-03-20T18:07:09.014 INFO:teuthology.orchestra.run.vm02.stderr: remote markers: [, , , , , , , , 00000000003.35.13, , ] 2026-03-20T18:07:14.015 INFO:teuthology.orchestra.run.vm02.stderr:2026-03-20T18:07:14.014+0000 7f8f75d1da80 1 waiting for incremental sync to catch up: 2026-03-20T18:07:14.015 INFO:teuthology.orchestra.run.vm02.stderr: local status: [, , , , , , , , 00000000001.31.6, , ] 2026-03-20T18:07:14.015 INFO:teuthology.orchestra.run.vm02.stderr: remote markers: [, , , , , , , , 00000000003.35.13, , ] 2026-03-20T18:07:19.016 INFO:teuthology.orchestra.run.vm02.stderr:2026-03-20T18:07:19.014+0000 7f8f75d1da80 1 waiting for incremental sync to catch up: 2026-03-20T18:07:19.016 INFO:teuthology.orchestra.run.vm02.stderr: local status: [, , , , , , , , 00000000001.31.6, , ] 2026-03-20T18:07:19.016 INFO:teuthology.orchestra.run.vm02.stderr: remote markers: [, , , , , , , , 00000000003.35.13, , ] 2026-03-20T18:07:24.018 INFO:teuthology.orchestra.run.vm02.stderr:2026-03-20T18:07:24.014+0000 7f8f75d1da80 1 waiting for incremental sync to catch up: 2026-03-20T18:07:24.018 INFO:teuthology.orchestra.run.vm02.stderr: local status: [, , , , , , , , 00000000001.31.6, , ] 2026-03-20T18:07:24.018 INFO:teuthology.orchestra.run.vm02.stderr: remote markers: [, , , , , , , , 00000000003.35.13, , ] 2026-03-20T18:07:29.019 INFO:teuthology.orchestra.run.vm02.stderr:2026-03-20T18:07:29.014+0000 7f8f75d1da80 1 bucket sync caught up with source: 2026-03-20T18:07:29.019 INFO:teuthology.orchestra.run.vm02.stderr: local status: [, , , , , , , , 00000000003.35.13, , ] 2026-03-20T18:07:29.019 INFO:teuthology.orchestra.run.vm02.stderr: remote markers: [, , , , , , , , 00000000003.35.13, , ] 2026-03-20T18:07:29.019 INFO:teuthology.orchestra.run.vm02.stderr:2026-03-20T18:07:29.014+0000 7f8f75d1da80 0 bucket checkpoint complete 2026-03-20T18:07:29.029 INFO:tasks.rgw_multi.zone_rados:comparing bucket=cyjkqa-17 zones={a2, a1} 2026-03-20T18:07:29.035 DEBUG:tasks.rgw_multi.zone_rados:bucket1 objects: 2026-03-20T18:07:29.035 DEBUG:tasks.rgw_multi.zone_rados:o=obj, v=EXKYZIcMNakQQoPRmoZBlRyAv0CIIvz 2026-03-20T18:07:29.035 DEBUG:tasks.rgw_multi.zone_rados:o=obj, v=null 2026-03-20T18:07:29.038 DEBUG:tasks.rgw_multi.zone_rados:bucket2 objects: 2026-03-20T18:07:29.038 DEBUG:tasks.rgw_multi.zone_rados:o=obj, v=EXKYZIcMNakQQoPRmoZBlRyAv0CIIvz 2026-03-20T18:07:29.038 DEBUG:tasks.rgw_multi.zone_rados:o=obj, v=null 2026-03-20T18:07:29.038 DEBUG:tasks.rgw_multi.zone_rados:comparing key name=obj 2026-03-20T18:07:29.051 DEBUG:tasks.rgw_multi.zone_rados:comparing key name=obj 2026-03-20T18:07:29.064 INFO:tasks.rgw_multi.zone_rados:success, bucket identical: bucket=cyjkqa-17 zones={a2, a1} 2026-03-20T18:07:29.069 INFO:tasks.util.rgw:rgwadmin: c1.client.0 : ['mdlog', 'status', '--rgw-zone', 'a1', '--rgw-zonegroup', 'a', '--rgw-realm', 'test-realm', '--cluster', 'c1', '--debug-rgw', '0', '--debug-ms', '0', '--rgw-cache-enabled', 'false'] 2026-03-20T18:07:29.069 DEBUG:tasks.util.rgw:rgwadmin: cmd=['adjust-ulimits', 'ceph-coverage', '/home/ubuntu/cephtest/archive/coverage', 'radosgw-admin', '--log-to-stderr', '--format', 'json', '-n', 'client.0', '--cluster', 'c1', 'mdlog', 'status', '--rgw-zone', 'a1', '--rgw-zonegroup', 'a', '--rgw-realm', 'test-realm', '--cluster', 'c1', '--debug-rgw', '0', '--debug-ms', '0', '--rgw-cache-enabled', 'false'] 2026-03-20T18:07:29.069 DEBUG:teuthology.orchestra.run.vm02:> adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage radosgw-admin --log-to-stderr --format json -n client.0 --cluster c1 mdlog status --rgw-zone a1 --rgw-zonegroup a --rgw-realm test-realm --cluster c1 --debug-rgw 0 --debug-ms 0 --rgw-cache-enabled false 2026-03-20T18:07:29.097 INFO:teuthology.orchestra.run.vm02.stderr:ignoring --setuser ceph since I am not root 2026-03-20T18:07:29.097 INFO:teuthology.orchestra.run.vm02.stderr:ignoring --setgroup ceph since I am not root 2026-03-20T18:07:29.191 INFO:teuthology.orchestra.run.vm02.stderr:No --period given, using current period=82c662be-66b5-4c4f-b24e-ff9298fc8a6e 2026-03-20T18:07:29.201 INFO:teuthology.orchestra.run.vm02.stdout:[{"marker":"1_1774030049.067619_276.1","last_update":"2026-03-20T18:07:29.067619Z"},{"marker":"1_1774029751.502359_72.1","last_update":"2026-03-20T18:02:31.502359Z"},{"marker":"1_1774029751.504947_186.1","last_update":"2026-03-20T18:02:31.504947Z"},{"marker":"1_1774030012.789970_152.1","last_update":"2026-03-20T18:06:52.789970Z"}] 2026-03-20T18:07:29.202 INFO:rgw_multi.tests:starting meta checkpoint for zone=a2 2026-03-20T18:07:29.202 INFO:tasks.util.rgw:rgwadmin: c1.client.0 : ['metadata', 'sync', 'status', '--rgw-zone', 'a2', '--rgw-zonegroup', 'a', '--rgw-realm', 'test-realm', '--cluster', 'c1', '--debug-rgw', '0', '--debug-ms', '0', '--rgw-cache-enabled', 'false'] 2026-03-20T18:07:29.202 DEBUG:tasks.util.rgw:rgwadmin: cmd=['adjust-ulimits', 'ceph-coverage', '/home/ubuntu/cephtest/archive/coverage', 'radosgw-admin', '--log-to-stderr', '--format', 'json', '-n', 'client.0', '--cluster', 'c1', 'metadata', 'sync', 'status', '--rgw-zone', 'a2', '--rgw-zonegroup', 'a', '--rgw-realm', 'test-realm', '--cluster', 'c1', '--debug-rgw', '0', '--debug-ms', '0', '--rgw-cache-enabled', 'false'] 2026-03-20T18:07:29.202 DEBUG:teuthology.orchestra.run.vm02:> adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage radosgw-admin --log-to-stderr --format json -n client.0 --cluster c1 metadata sync status --rgw-zone a2 --rgw-zonegroup a --rgw-realm test-realm --cluster c1 --debug-rgw 0 --debug-ms 0 --rgw-cache-enabled false 2026-03-20T18:07:29.269 INFO:teuthology.orchestra.run.vm02.stderr:ignoring --setuser ceph since I am not root 2026-03-20T18:07:29.269 INFO:teuthology.orchestra.run.vm02.stderr:ignoring --setgroup ceph since I am not root 2026-03-20T18:07:29.348 INFO:teuthology.orchestra.run.vm02.stderr:2026-03-20T18:07:29.346+0000 7f86bcd80a80 20 RGW-SYNC:meta: read sync status 2026-03-20T18:07:29.349 INFO:teuthology.orchestra.run.vm02.stderr:2026-03-20T18:07:29.346+0000 7f86bcd80a80 20 RGW-SYNC:meta: read sync status 2026-03-20T18:07:29.360 INFO:teuthology.orchestra.run.vm02.stdout:{"sync_status":{"info":{"status":"sync","num_shards":4,"period":"82c662be-66b5-4c4f-b24e-ff9298fc8a6e","realm_epoch":2},"markers":[{"key":0,"val":{"state":1,"marker":"1_1774030049.067619_276.1","next_step_marker":"","total_entries":1,"pos":0,"timestamp":"2026-03-20T18:07:29.067619Z","realm_epoch":2}},{"key":1,"val":{"state":1,"marker":"1_1774029751.502359_72.1","next_step_marker":"","total_entries":0,"pos":0,"timestamp":"2026-03-20T18:02:31.502359Z","realm_epoch":2}},{"key":2,"val":{"state":1,"marker":"1_1774029751.504947_186.1","next_step_marker":"","total_entries":0,"pos":0,"timestamp":"2026-03-20T18:02:31.504947Z","realm_epoch":2}},{"key":3,"val":{"state":1,"marker":"1_1774030012.789970_152.1","next_step_marker":"","total_entries":0,"pos":0,"timestamp":"2026-03-20T18:06:52.789970Z","realm_epoch":2}}]},"full_sync":{"total":1,"complete":1},"current_time":"2026-03-20T18:07:29Z"} 2026-03-20T18:07:29.360 INFO:rgw_multi.tests:finish meta checkpoint for zone=a2 2026-03-20T18:07:29.373 INFO:tasks.util.rgw:rgwadmin: c1.client.0 : ['bucket', 'sync', 'checkpoint', '--bucket', 'cyjkqa-17', '--source-zone', 'a1', '--retry-delay-ms', '5000', '--timeout-sec', '300', '--rgw-zone', 'a2', '--rgw-zonegroup', 'a', '--rgw-realm', 'test-realm', '--cluster', 'c1', '--debug-rgw', '1', '--debug-ms', '0'] 2026-03-20T18:07:29.373 DEBUG:tasks.util.rgw:rgwadmin: cmd=['adjust-ulimits', 'ceph-coverage', '/home/ubuntu/cephtest/archive/coverage', 'radosgw-admin', '--log-to-stderr', '--format', 'json', '-n', 'client.0', '--cluster', 'c1', 'bucket', 'sync', 'checkpoint', '--bucket', 'cyjkqa-17', '--source-zone', 'a1', '--retry-delay-ms', '5000', '--timeout-sec', '300', '--rgw-zone', 'a2', '--rgw-zonegroup', 'a', '--rgw-realm', 'test-realm', '--cluster', 'c1', '--debug-rgw', '1', '--debug-ms', '0'] 2026-03-20T18:07:29.373 DEBUG:teuthology.orchestra.run.vm02:> adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage radosgw-admin --log-to-stderr --format json -n client.0 --cluster c1 bucket sync checkpoint --bucket cyjkqa-17 --source-zone a1 --retry-delay-ms 5000 --timeout-sec 300 --rgw-zone a2 --rgw-zonegroup a --rgw-realm test-realm --cluster c1 --debug-rgw 1 --debug-ms 0 2026-03-20T18:07:29.427 INFO:teuthology.orchestra.run.vm02.stderr:ignoring --setuser ceph since I am not root 2026-03-20T18:07:29.427 INFO:teuthology.orchestra.run.vm02.stderr:ignoring --setgroup ceph since I am not root 2026-03-20T18:07:29.519 INFO:teuthology.orchestra.run.vm02.stderr:2026-03-20T18:07:29.518+0000 7f4344506a80 1 waiting for incremental sync to catch up: 2026-03-20T18:07:29.520 INFO:teuthology.orchestra.run.vm02.stderr: local status: [, , , , , , , , 00000000003.29.13, , ] 2026-03-20T18:07:29.520 INFO:teuthology.orchestra.run.vm02.stderr: remote markers: [, , , , , , , , 00000000005.33.18, , ] 2026-03-20T18:07:34.521 INFO:teuthology.orchestra.run.vm02.stderr:2026-03-20T18:07:34.518+0000 7f4344506a80 1 waiting for incremental sync to catch up: 2026-03-20T18:07:34.521 INFO:teuthology.orchestra.run.vm02.stderr: local status: [, , , , , , , , 00000000003.29.13, , ] 2026-03-20T18:07:34.521 INFO:teuthology.orchestra.run.vm02.stderr: remote markers: [, , , , , , , , 00000000005.33.18, , ] 2026-03-20T18:07:39.521 INFO:teuthology.orchestra.run.vm02.stderr:2026-03-20T18:07:39.518+0000 7f4344506a80 1 waiting for incremental sync to catch up: 2026-03-20T18:07:39.522 INFO:teuthology.orchestra.run.vm02.stderr: local status: [, , , , , , , , 00000000003.29.13, , ] 2026-03-20T18:07:39.522 INFO:teuthology.orchestra.run.vm02.stderr: remote markers: [, , , , , , , , 00000000005.33.18, , ] 2026-03-20T18:07:44.523 INFO:teuthology.orchestra.run.vm02.stderr:2026-03-20T18:07:44.522+0000 7f4344506a80 1 waiting for incremental sync to catch up: 2026-03-20T18:07:44.523 INFO:teuthology.orchestra.run.vm02.stderr: local status: [, , , , , , , , 00000000003.29.13, , ] 2026-03-20T18:07:44.523 INFO:teuthology.orchestra.run.vm02.stderr: remote markers: [, , , , , , , , 00000000005.33.18, , ] 2026-03-20T18:07:49.523 INFO:teuthology.orchestra.run.vm02.stderr:2026-03-20T18:07:49.522+0000 7f4344506a80 1 waiting for incremental sync to catch up: 2026-03-20T18:07:49.524 INFO:teuthology.orchestra.run.vm02.stderr: local status: [, , , , , , , , 00000000003.29.13, , ] 2026-03-20T18:07:49.524 INFO:teuthology.orchestra.run.vm02.stderr: remote markers: [, , , , , , , , 00000000005.33.18, , ] 2026-03-20T18:07:54.524 INFO:teuthology.orchestra.run.vm02.stderr:2026-03-20T18:07:54.522+0000 7f4344506a80 1 bucket sync caught up with source: 2026-03-20T18:07:54.524 INFO:teuthology.orchestra.run.vm02.stderr: local status: [, , , , , , , , 00000000005.33.18, , ] 2026-03-20T18:07:54.524 INFO:teuthology.orchestra.run.vm02.stderr: remote markers: [, , , , , , , , 00000000005.33.18, , ] 2026-03-20T18:07:54.524 INFO:teuthology.orchestra.run.vm02.stderr:2026-03-20T18:07:54.522+0000 7f4344506a80 0 bucket checkpoint complete 2026-03-20T18:07:54.534 INFO:tasks.util.rgw:rgwadmin: c1.client.0 : ['bucket', 'sync', 'checkpoint', '--bucket', 'cyjkqa-17', '--source-zone', 'a2', '--retry-delay-ms', '5000', '--timeout-sec', '300', '--rgw-zone', 'a1', '--rgw-zonegroup', 'a', '--rgw-realm', 'test-realm', '--cluster', 'c1', '--debug-rgw', '1', '--debug-ms', '0'] 2026-03-20T18:07:54.534 DEBUG:tasks.util.rgw:rgwadmin: cmd=['adjust-ulimits', 'ceph-coverage', '/home/ubuntu/cephtest/archive/coverage', 'radosgw-admin', '--log-to-stderr', '--format', 'json', '-n', 'client.0', '--cluster', 'c1', 'bucket', 'sync', 'checkpoint', '--bucket', 'cyjkqa-17', '--source-zone', 'a2', '--retry-delay-ms', '5000', '--timeout-sec', '300', '--rgw-zone', 'a1', '--rgw-zonegroup', 'a', '--rgw-realm', 'test-realm', '--cluster', 'c1', '--debug-rgw', '1', '--debug-ms', '0'] 2026-03-20T18:07:54.534 DEBUG:teuthology.orchestra.run.vm02:> adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage radosgw-admin --log-to-stderr --format json -n client.0 --cluster c1 bucket sync checkpoint --bucket cyjkqa-17 --source-zone a2 --retry-delay-ms 5000 --timeout-sec 300 --rgw-zone a1 --rgw-zonegroup a --rgw-realm test-realm --cluster c1 --debug-rgw 1 --debug-ms 0 2026-03-20T18:07:54.599 INFO:teuthology.orchestra.run.vm02.stderr:ignoring --setuser ceph since I am not root 2026-03-20T18:07:54.599 INFO:teuthology.orchestra.run.vm02.stderr:ignoring --setgroup ceph since I am not root 2026-03-20T18:07:54.692 INFO:teuthology.orchestra.run.vm02.stderr:2026-03-20T18:07:54.690+0000 7f27a8944a80 1 waiting for incremental sync to catch up: 2026-03-20T18:07:54.692 INFO:teuthology.orchestra.run.vm02.stderr: local status: [, , , , , , , , 00000000003.35.13, , ] 2026-03-20T18:07:54.692 INFO:teuthology.orchestra.run.vm02.stderr: remote markers: [, , , , , , , , 00000000005.40.18, , ] 2026-03-20T18:07:59.694 INFO:teuthology.orchestra.run.vm02.stderr:2026-03-20T18:07:59.690+0000 7f27a8944a80 1 waiting for incremental sync to catch up: 2026-03-20T18:07:59.694 INFO:teuthology.orchestra.run.vm02.stderr: local status: [, , , , , , , , 00000000003.35.13, , ] 2026-03-20T18:07:59.694 INFO:teuthology.orchestra.run.vm02.stderr: remote markers: [, , , , , , , , 00000000005.40.18, , ] 2026-03-20T18:08:04.695 INFO:teuthology.orchestra.run.vm02.stderr:2026-03-20T18:08:04.694+0000 7f27a8944a80 1 waiting for incremental sync to catch up: 2026-03-20T18:08:04.695 INFO:teuthology.orchestra.run.vm02.stderr: local status: [, , , , , , , , 00000000003.35.13, , ] 2026-03-20T18:08:04.695 INFO:teuthology.orchestra.run.vm02.stderr: remote markers: [, , , , , , , , 00000000005.40.18, , ] 2026-03-20T18:08:09.696 INFO:teuthology.orchestra.run.vm02.stderr:2026-03-20T18:08:09.694+0000 7f27a8944a80 1 waiting for incremental sync to catch up: 2026-03-20T18:08:09.696 INFO:teuthology.orchestra.run.vm02.stderr: local status: [, , , , , , , , 00000000003.35.13, , ] 2026-03-20T18:08:09.696 INFO:teuthology.orchestra.run.vm02.stderr: remote markers: [, , , , , , , , 00000000005.40.18, , ] 2026-03-20T18:08:14.697 INFO:teuthology.orchestra.run.vm02.stderr:2026-03-20T18:08:14.694+0000 7f27a8944a80 1 bucket sync caught up with source: 2026-03-20T18:08:14.697 INFO:teuthology.orchestra.run.vm02.stderr: local status: [, , , , , , , , 00000000005.40.18, , ] 2026-03-20T18:08:14.697 INFO:teuthology.orchestra.run.vm02.stderr: remote markers: [, , , , , , , , 00000000005.40.18, , ] 2026-03-20T18:08:14.697 INFO:teuthology.orchestra.run.vm02.stderr:2026-03-20T18:08:14.694+0000 7f27a8944a80 0 bucket checkpoint complete 2026-03-20T18:08:14.707 INFO:tasks.rgw_multi.zone_rados:comparing bucket=cyjkqa-17 zones={a2, a1} 2026-03-20T18:08:14.714 DEBUG:tasks.rgw_multi.zone_rados:bucket1 objects: 2026-03-20T18:08:14.714 DEBUG:tasks.rgw_multi.zone_rados:o=obj, v=null 2026-03-20T18:08:14.714 DEBUG:tasks.rgw_multi.zone_rados:o=obj, v=EXKYZIcMNakQQoPRmoZBlRyAv0CIIvz 2026-03-20T18:08:14.716 DEBUG:tasks.rgw_multi.zone_rados:bucket2 objects: 2026-03-20T18:08:14.716 DEBUG:tasks.rgw_multi.zone_rados:o=obj, v=null 2026-03-20T18:08:14.716 DEBUG:tasks.rgw_multi.zone_rados:o=obj, v=EXKYZIcMNakQQoPRmoZBlRyAv0CIIvz 2026-03-20T18:08:14.716 DEBUG:tasks.rgw_multi.zone_rados:comparing key name=obj 2026-03-20T18:08:14.727 DEBUG:tasks.rgw_multi.zone_rados:comparing key name=obj 2026-03-20T18:08:14.737 INFO:tasks.rgw_multi.zone_rados:success, bucket identical: bucket=cyjkqa-17 zones={a2, a1} 2026-03-20T18:08:14.739 INFO:tasks.rgw_multisite_tests:rgw_multi.tests.test_version_suspended_incremental_sync ... ok 2026-03-20T18:08:14.781 INFO:rgw_multi.tests:create bucket zone=a1 name=cyjkqa-18 2026-03-20T18:08:14.793 INFO:rgw_multi.tests:create bucket zone=a2 name=cyjkqa-19 2026-03-20T18:08:14.892 INFO:tasks.util.rgw:rgwadmin: c1.client.0 : ['mdlog', 'status', '--rgw-zone', 'a1', '--rgw-zonegroup', 'a', '--rgw-realm', 'test-realm', '--cluster', 'c1', '--debug-rgw', '0', '--debug-ms', '0', '--rgw-cache-enabled', 'false'] 2026-03-20T18:08:14.892 DEBUG:tasks.util.rgw:rgwadmin: cmd=['adjust-ulimits', 'ceph-coverage', '/home/ubuntu/cephtest/archive/coverage', 'radosgw-admin', '--log-to-stderr', '--format', 'json', '-n', 'client.0', '--cluster', 'c1', 'mdlog', 'status', '--rgw-zone', 'a1', '--rgw-zonegroup', 'a', '--rgw-realm', 'test-realm', '--cluster', 'c1', '--debug-rgw', '0', '--debug-ms', '0', '--rgw-cache-enabled', 'false'] 2026-03-20T18:08:14.892 DEBUG:teuthology.orchestra.run.vm02:> adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage radosgw-admin --log-to-stderr --format json -n client.0 --cluster c1 mdlog status --rgw-zone a1 --rgw-zonegroup a --rgw-realm test-realm --cluster c1 --debug-rgw 0 --debug-ms 0 --rgw-cache-enabled false 2026-03-20T18:08:14.913 INFO:teuthology.orchestra.run.vm02.stderr:ignoring --setuser ceph since I am not root 2026-03-20T18:08:14.913 INFO:teuthology.orchestra.run.vm02.stderr:ignoring --setgroup ceph since I am not root 2026-03-20T18:08:14.979 INFO:teuthology.orchestra.run.vm02.stderr:No --period given, using current period=82c662be-66b5-4c4f-b24e-ff9298fc8a6e 2026-03-20T18:08:14.989 INFO:teuthology.orchestra.run.vm02.stdout:[{"marker":"1_1774030094.888047_286.1","last_update":"2026-03-20T18:08:14.888047Z"},{"marker":"1_1774029751.502359_72.1","last_update":"2026-03-20T18:02:31.502359Z"},{"marker":"1_1774029751.504947_186.1","last_update":"2026-03-20T18:02:31.504947Z"},{"marker":"1_1774030094.805591_165.1","last_update":"2026-03-20T18:08:14.805591Z"}] 2026-03-20T18:08:14.989 INFO:rgw_multi.tests:starting meta checkpoint for zone=a2 2026-03-20T18:08:14.989 INFO:tasks.util.rgw:rgwadmin: c1.client.0 : ['metadata', 'sync', 'status', '--rgw-zone', 'a2', '--rgw-zonegroup', 'a', '--rgw-realm', 'test-realm', '--cluster', 'c1', '--debug-rgw', '0', '--debug-ms', '0', '--rgw-cache-enabled', 'false'] 2026-03-20T18:08:14.989 DEBUG:tasks.util.rgw:rgwadmin: cmd=['adjust-ulimits', 'ceph-coverage', '/home/ubuntu/cephtest/archive/coverage', 'radosgw-admin', '--log-to-stderr', '--format', 'json', '-n', 'client.0', '--cluster', 'c1', 'metadata', 'sync', 'status', '--rgw-zone', 'a2', '--rgw-zonegroup', 'a', '--rgw-realm', 'test-realm', '--cluster', 'c1', '--debug-rgw', '0', '--debug-ms', '0', '--rgw-cache-enabled', 'false'] 2026-03-20T18:08:14.989 DEBUG:teuthology.orchestra.run.vm02:> adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage radosgw-admin --log-to-stderr --format json -n client.0 --cluster c1 metadata sync status --rgw-zone a2 --rgw-zonegroup a --rgw-realm test-realm --cluster c1 --debug-rgw 0 --debug-ms 0 --rgw-cache-enabled false 2026-03-20T18:08:15.050 INFO:teuthology.orchestra.run.vm02.stderr:ignoring --setuser ceph since I am not root 2026-03-20T18:08:15.050 INFO:teuthology.orchestra.run.vm02.stderr:ignoring --setgroup ceph since I am not root 2026-03-20T18:08:15.127 INFO:teuthology.orchestra.run.vm02.stderr:2026-03-20T18:08:15.126+0000 7f3ae9f54a80 20 RGW-SYNC:meta: read sync status 2026-03-20T18:08:15.128 INFO:teuthology.orchestra.run.vm02.stderr:2026-03-20T18:08:15.126+0000 7f3ae9f54a80 20 RGW-SYNC:meta: read sync status 2026-03-20T18:08:15.139 INFO:teuthology.orchestra.run.vm02.stdout:{"sync_status":{"info":{"status":"sync","num_shards":4,"period":"82c662be-66b5-4c4f-b24e-ff9298fc8a6e","realm_epoch":2},"markers":[{"key":0,"val":{"state":1,"marker":"1_1774030094.888047_286.1","next_step_marker":"","total_entries":1,"pos":0,"timestamp":"2026-03-20T18:08:14.888047Z","realm_epoch":2}},{"key":1,"val":{"state":1,"marker":"1_1774029751.502359_72.1","next_step_marker":"","total_entries":0,"pos":0,"timestamp":"2026-03-20T18:02:31.502359Z","realm_epoch":2}},{"key":2,"val":{"state":1,"marker":"1_1774029751.504947_186.1","next_step_marker":"","total_entries":0,"pos":0,"timestamp":"2026-03-20T18:02:31.504947Z","realm_epoch":2}},{"key":3,"val":{"state":1,"marker":"1_1774030094.805591_165.1","next_step_marker":"","total_entries":0,"pos":0,"timestamp":"2026-03-20T18:08:14.805591Z","realm_epoch":2}}]},"full_sync":{"total":1,"complete":1},"current_time":"2026-03-20T18:08:15Z"} 2026-03-20T18:08:15.139 INFO:rgw_multi.tests:finish meta checkpoint for zone=a2 2026-03-20T18:08:15.199 INFO:tasks.util.rgw:rgwadmin: c1.client.0 : ['bucket', 'sync', 'checkpoint', '--bucket', 'cyjkqa-18', '--source-zone', 'a1', '--retry-delay-ms', '5000', '--timeout-sec', '300', '--rgw-zone', 'a2', '--rgw-zonegroup', 'a', '--rgw-realm', 'test-realm', '--cluster', 'c1', '--debug-rgw', '1', '--debug-ms', '0'] 2026-03-20T18:08:15.199 DEBUG:tasks.util.rgw:rgwadmin: cmd=['adjust-ulimits', 'ceph-coverage', '/home/ubuntu/cephtest/archive/coverage', 'radosgw-admin', '--log-to-stderr', '--format', 'json', '-n', 'client.0', '--cluster', 'c1', 'bucket', 'sync', 'checkpoint', '--bucket', 'cyjkqa-18', '--source-zone', 'a1', '--retry-delay-ms', '5000', '--timeout-sec', '300', '--rgw-zone', 'a2', '--rgw-zonegroup', 'a', '--rgw-realm', 'test-realm', '--cluster', 'c1', '--debug-rgw', '1', '--debug-ms', '0'] 2026-03-20T18:08:15.200 DEBUG:teuthology.orchestra.run.vm02:> adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage radosgw-admin --log-to-stderr --format json -n client.0 --cluster c1 bucket sync checkpoint --bucket cyjkqa-18 --source-zone a1 --retry-delay-ms 5000 --timeout-sec 300 --rgw-zone a2 --rgw-zonegroup a --rgw-realm test-realm --cluster c1 --debug-rgw 1 --debug-ms 0 2026-03-20T18:08:15.222 INFO:teuthology.orchestra.run.vm02.stderr:ignoring --setuser ceph since I am not root 2026-03-20T18:08:15.222 INFO:teuthology.orchestra.run.vm02.stderr:ignoring --setgroup ceph since I am not root 2026-03-20T18:08:15.300 INFO:teuthology.orchestra.run.vm02.stderr:2026-03-20T18:08:15.298+0000 7f4709d43a80 1 waiting to reach incremental sync.. 2026-03-20T18:08:20.302 INFO:teuthology.orchestra.run.vm02.stderr:2026-03-20T18:08:20.298+0000 7f4709d43a80 1 bucket sync caught up with source: 2026-03-20T18:08:20.302 INFO:teuthology.orchestra.run.vm02.stderr: local status: [, , , , , , , , 00000000005.113.12, , ] 2026-03-20T18:08:20.302 INFO:teuthology.orchestra.run.vm02.stderr: remote markers: [, , , , , , , , 00000000005.113.12, , ] 2026-03-20T18:08:20.302 INFO:teuthology.orchestra.run.vm02.stderr:2026-03-20T18:08:20.298+0000 7f4709d43a80 0 bucket checkpoint complete 2026-03-20T18:08:20.314 INFO:tasks.util.rgw:rgwadmin: c1.client.0 : ['bucket', 'sync', 'checkpoint', '--bucket', 'cyjkqa-18', '--source-zone', 'a2', '--retry-delay-ms', '5000', '--timeout-sec', '300', '--rgw-zone', 'a1', '--rgw-zonegroup', 'a', '--rgw-realm', 'test-realm', '--cluster', 'c1', '--debug-rgw', '1', '--debug-ms', '0'] 2026-03-20T18:08:20.314 DEBUG:tasks.util.rgw:rgwadmin: cmd=['adjust-ulimits', 'ceph-coverage', '/home/ubuntu/cephtest/archive/coverage', 'radosgw-admin', '--log-to-stderr', '--format', 'json', '-n', 'client.0', '--cluster', 'c1', 'bucket', 'sync', 'checkpoint', '--bucket', 'cyjkqa-18', '--source-zone', 'a2', '--retry-delay-ms', '5000', '--timeout-sec', '300', '--rgw-zone', 'a1', '--rgw-zonegroup', 'a', '--rgw-realm', 'test-realm', '--cluster', 'c1', '--debug-rgw', '1', '--debug-ms', '0'] 2026-03-20T18:08:20.314 DEBUG:teuthology.orchestra.run.vm02:> adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage radosgw-admin --log-to-stderr --format json -n client.0 --cluster c1 bucket sync checkpoint --bucket cyjkqa-18 --source-zone a2 --retry-delay-ms 5000 --timeout-sec 300 --rgw-zone a1 --rgw-zonegroup a --rgw-realm test-realm --cluster c1 --debug-rgw 1 --debug-ms 0 2026-03-20T18:08:20.382 INFO:teuthology.orchestra.run.vm02.stderr:ignoring --setuser ceph since I am not root 2026-03-20T18:08:20.382 INFO:teuthology.orchestra.run.vm02.stderr:ignoring --setgroup ceph since I am not root 2026-03-20T18:08:20.452 INFO:teuthology.orchestra.run.vm02.stderr:2026-03-20T18:08:20.450+0000 7f05371c7a80 1 waiting to reach incremental sync.. 2026-03-20T18:08:25.456 INFO:teuthology.orchestra.run.vm02.stderr:2026-03-20T18:08:25.454+0000 7f05371c7a80 1 bucket sync caught up with source: 2026-03-20T18:08:25.456 INFO:teuthology.orchestra.run.vm02.stderr: local status: [, , , , , , , , 00000000003.116.6, , ] 2026-03-20T18:08:25.456 INFO:teuthology.orchestra.run.vm02.stderr: remote markers: [, , , , , , , , 00000000003.116.6, , ] 2026-03-20T18:08:25.456 INFO:teuthology.orchestra.run.vm02.stderr:2026-03-20T18:08:25.454+0000 7f05371c7a80 0 bucket checkpoint complete 2026-03-20T18:08:25.466 INFO:tasks.rgw_multi.zone_rados:comparing bucket=cyjkqa-18 zones={a2, a1} 2026-03-20T18:08:25.474 DEBUG:tasks.rgw_multi.zone_rados:bucket1 objects: 2026-03-20T18:08:25.474 DEBUG:tasks.rgw_multi.zone_rados:o=obj, v=tiP9JNK0n2QH1HqbbnEhhKO2fKoe2K6 2026-03-20T18:08:25.474 DEBUG:tasks.rgw_multi.zone_rados:o=obj, v=blt-o51iQ0dJm0J.0eHV56.cB25kKK5 2026-03-20T18:08:25.474 DEBUG:tasks.rgw_multi.zone_rados:o=obj, v=KBh47YBaVZMhHvkqhcQhdrwZk2IU8oR 2026-03-20T18:08:25.474 DEBUG:tasks.rgw_multi.zone_rados:o=obj, v=rJyXdyTm4Iarn3ajSjW5.SgwUzaCmzy 2026-03-20T18:08:25.477 DEBUG:tasks.rgw_multi.zone_rados:bucket2 objects: 2026-03-20T18:08:25.477 DEBUG:tasks.rgw_multi.zone_rados:o=obj, v=tiP9JNK0n2QH1HqbbnEhhKO2fKoe2K6 2026-03-20T18:08:25.477 DEBUG:tasks.rgw_multi.zone_rados:o=obj, v=blt-o51iQ0dJm0J.0eHV56.cB25kKK5 2026-03-20T18:08:25.477 DEBUG:tasks.rgw_multi.zone_rados:o=obj, v=KBh47YBaVZMhHvkqhcQhdrwZk2IU8oR 2026-03-20T18:08:25.477 DEBUG:tasks.rgw_multi.zone_rados:o=obj, v=rJyXdyTm4Iarn3ajSjW5.SgwUzaCmzy 2026-03-20T18:08:25.477 DEBUG:tasks.rgw_multi.zone_rados:comparing key name=obj 2026-03-20T18:08:25.489 DEBUG:tasks.rgw_multi.zone_rados:both are delete markers, skipping content comparison 2026-03-20T18:08:25.489 DEBUG:tasks.rgw_multi.zone_rados:both are delete markers, skipping content comparison 2026-03-20T18:08:25.489 DEBUG:tasks.rgw_multi.zone_rados:both are delete markers, skipping content comparison 2026-03-20T18:08:25.489 INFO:tasks.rgw_multi.zone_rados:success, bucket identical: bucket=cyjkqa-18 zones={a2, a1} 2026-03-20T18:08:25.489 INFO:tasks.util.rgw:rgwadmin: c1.client.0 : ['bucket', 'sync', 'checkpoint', '--bucket', 'cyjkqa-19', '--source-zone', 'a1', '--retry-delay-ms', '5000', '--timeout-sec', '300', '--rgw-zone', 'a2', '--rgw-zonegroup', 'a', '--rgw-realm', 'test-realm', '--cluster', 'c1', '--debug-rgw', '1', '--debug-ms', '0'] 2026-03-20T18:08:25.489 DEBUG:tasks.util.rgw:rgwadmin: cmd=['adjust-ulimits', 'ceph-coverage', '/home/ubuntu/cephtest/archive/coverage', 'radosgw-admin', '--log-to-stderr', '--format', 'json', '-n', 'client.0', '--cluster', 'c1', 'bucket', 'sync', 'checkpoint', '--bucket', 'cyjkqa-19', '--source-zone', 'a1', '--retry-delay-ms', '5000', '--timeout-sec', '300', '--rgw-zone', 'a2', '--rgw-zonegroup', 'a', '--rgw-realm', 'test-realm', '--cluster', 'c1', '--debug-rgw', '1', '--debug-ms', '0'] 2026-03-20T18:08:25.489 DEBUG:teuthology.orchestra.run.vm02:> adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage radosgw-admin --log-to-stderr --format json -n client.0 --cluster c1 bucket sync checkpoint --bucket cyjkqa-19 --source-zone a1 --retry-delay-ms 5000 --timeout-sec 300 --rgw-zone a2 --rgw-zonegroup a --rgw-realm test-realm --cluster c1 --debug-rgw 1 --debug-ms 0 2026-03-20T18:08:25.529 INFO:teuthology.orchestra.run.vm02.stderr:ignoring --setuser ceph since I am not root 2026-03-20T18:08:25.529 INFO:teuthology.orchestra.run.vm02.stderr:ignoring --setgroup ceph since I am not root 2026-03-20T18:08:25.607 INFO:teuthology.orchestra.run.vm02.stderr:2026-03-20T18:08:25.602+0000 7f12ddf3fa80 1 bucket sync caught up with source: 2026-03-20T18:08:25.607 INFO:teuthology.orchestra.run.vm02.stderr: local status: [, , , , , , , , 00000000004.24.6, , ] 2026-03-20T18:08:25.607 INFO:teuthology.orchestra.run.vm02.stderr: remote markers: [, , , , , , , , 00000000004.24.6, , ] 2026-03-20T18:08:25.607 INFO:teuthology.orchestra.run.vm02.stderr:2026-03-20T18:08:25.602+0000 7f12ddf3fa80 0 bucket checkpoint complete 2026-03-20T18:08:25.616 INFO:tasks.util.rgw:rgwadmin: c1.client.0 : ['bucket', 'sync', 'checkpoint', '--bucket', 'cyjkqa-19', '--source-zone', 'a2', '--retry-delay-ms', '5000', '--timeout-sec', '300', '--rgw-zone', 'a1', '--rgw-zonegroup', 'a', '--rgw-realm', 'test-realm', '--cluster', 'c1', '--debug-rgw', '1', '--debug-ms', '0'] 2026-03-20T18:08:25.616 DEBUG:tasks.util.rgw:rgwadmin: cmd=['adjust-ulimits', 'ceph-coverage', '/home/ubuntu/cephtest/archive/coverage', 'radosgw-admin', '--log-to-stderr', '--format', 'json', '-n', 'client.0', '--cluster', 'c1', 'bucket', 'sync', 'checkpoint', '--bucket', 'cyjkqa-19', '--source-zone', 'a2', '--retry-delay-ms', '5000', '--timeout-sec', '300', '--rgw-zone', 'a1', '--rgw-zonegroup', 'a', '--rgw-realm', 'test-realm', '--cluster', 'c1', '--debug-rgw', '1', '--debug-ms', '0'] 2026-03-20T18:08:25.616 DEBUG:teuthology.orchestra.run.vm02:> adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage radosgw-admin --log-to-stderr --format json -n client.0 --cluster c1 bucket sync checkpoint --bucket cyjkqa-19 --source-zone a2 --retry-delay-ms 5000 --timeout-sec 300 --rgw-zone a1 --rgw-zonegroup a --rgw-realm test-realm --cluster c1 --debug-rgw 1 --debug-ms 0 2026-03-20T18:08:25.682 INFO:teuthology.orchestra.run.vm02.stderr:ignoring --setuser ceph since I am not root 2026-03-20T18:08:25.682 INFO:teuthology.orchestra.run.vm02.stderr:ignoring --setgroup ceph since I am not root 2026-03-20T18:08:25.756 INFO:teuthology.orchestra.run.vm02.stderr:2026-03-20T18:08:25.754+0000 7f4146186a80 1 bucket sync caught up with source: 2026-03-20T18:08:25.756 INFO:teuthology.orchestra.run.vm02.stderr: local status: [, , , , , , , , 00000000005.28.12, , ] 2026-03-20T18:08:25.756 INFO:teuthology.orchestra.run.vm02.stderr: remote markers: [, , , , , , , , 00000000005.28.12, , ] 2026-03-20T18:08:25.756 INFO:teuthology.orchestra.run.vm02.stderr:2026-03-20T18:08:25.754+0000 7f4146186a80 0 bucket checkpoint complete 2026-03-20T18:08:25.765 INFO:tasks.rgw_multi.zone_rados:comparing bucket=cyjkqa-19 zones={a2, a1} 2026-03-20T18:08:25.772 DEBUG:tasks.rgw_multi.zone_rados:bucket1 objects: 2026-03-20T18:08:25.772 DEBUG:tasks.rgw_multi.zone_rados:o=obj, v=n5amQvZk8-EQcReyPkO4f4BGX6WWNGi 2026-03-20T18:08:25.772 DEBUG:tasks.rgw_multi.zone_rados:o=obj, v=M2AlWdo9yEy45YD2OMgIMTocYuiCLRp 2026-03-20T18:08:25.772 DEBUG:tasks.rgw_multi.zone_rados:o=obj, v=7xjqEO27yeBq5DcoaHAkaL08InII39N 2026-03-20T18:08:25.772 DEBUG:tasks.rgw_multi.zone_rados:o=obj, v=act23TZLd6arO4X.H4jy3wU-dCFu-ck 2026-03-20T18:08:25.775 DEBUG:tasks.rgw_multi.zone_rados:bucket2 objects: 2026-03-20T18:08:25.775 DEBUG:tasks.rgw_multi.zone_rados:o=obj, v=n5amQvZk8-EQcReyPkO4f4BGX6WWNGi 2026-03-20T18:08:25.775 DEBUG:tasks.rgw_multi.zone_rados:o=obj, v=M2AlWdo9yEy45YD2OMgIMTocYuiCLRp 2026-03-20T18:08:25.775 DEBUG:tasks.rgw_multi.zone_rados:o=obj, v=7xjqEO27yeBq5DcoaHAkaL08InII39N 2026-03-20T18:08:25.775 DEBUG:tasks.rgw_multi.zone_rados:o=obj, v=act23TZLd6arO4X.H4jy3wU-dCFu-ck 2026-03-20T18:08:25.775 DEBUG:tasks.rgw_multi.zone_rados:comparing key name=obj 2026-03-20T18:08:25.785 DEBUG:tasks.rgw_multi.zone_rados:both are delete markers, skipping content comparison 2026-03-20T18:08:25.785 DEBUG:tasks.rgw_multi.zone_rados:both are delete markers, skipping content comparison 2026-03-20T18:08:25.785 DEBUG:tasks.rgw_multi.zone_rados:both are delete markers, skipping content comparison 2026-03-20T18:08:25.785 INFO:tasks.rgw_multi.zone_rados:success, bucket identical: bucket=cyjkqa-19 zones={a2, a1} 2026-03-20T18:08:25.787 INFO:tasks.rgw_multisite_tests:rgw_multi.tests.test_delete_marker_full_sync ... ok 2026-03-20T18:08:25.828 INFO:rgw_multi.tests:create bucket zone=a1 name=cyjkqa-20 2026-03-20T18:08:25.843 INFO:rgw_multi.tests:create bucket zone=a2 name=cyjkqa-21 2026-03-20T18:08:25.889 INFO:tasks.util.rgw:rgwadmin: c1.client.0 : ['mdlog', 'status', '--rgw-zone', 'a1', '--rgw-zonegroup', 'a', '--rgw-realm', 'test-realm', '--cluster', 'c1', '--debug-rgw', '0', '--debug-ms', '0', '--rgw-cache-enabled', 'false'] 2026-03-20T18:08:25.889 DEBUG:tasks.util.rgw:rgwadmin: cmd=['adjust-ulimits', 'ceph-coverage', '/home/ubuntu/cephtest/archive/coverage', 'radosgw-admin', '--log-to-stderr', '--format', 'json', '-n', 'client.0', '--cluster', 'c1', 'mdlog', 'status', '--rgw-zone', 'a1', '--rgw-zonegroup', 'a', '--rgw-realm', 'test-realm', '--cluster', 'c1', '--debug-rgw', '0', '--debug-ms', '0', '--rgw-cache-enabled', 'false'] 2026-03-20T18:08:25.890 DEBUG:teuthology.orchestra.run.vm02:> adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage radosgw-admin --log-to-stderr --format json -n client.0 --cluster c1 mdlog status --rgw-zone a1 --rgw-zonegroup a --rgw-realm test-realm --cluster c1 --debug-rgw 0 --debug-ms 0 --rgw-cache-enabled false 2026-03-20T18:08:25.909 INFO:teuthology.orchestra.run.vm02.stderr:ignoring --setuser ceph since I am not root 2026-03-20T18:08:25.909 INFO:teuthology.orchestra.run.vm02.stderr:ignoring --setgroup ceph since I am not root 2026-03-20T18:08:25.993 INFO:teuthology.orchestra.run.vm02.stderr:No --period given, using current period=82c662be-66b5-4c4f-b24e-ff9298fc8a6e 2026-03-20T18:08:26.006 INFO:teuthology.orchestra.run.vm02.stdout:[{"marker":"1_1774030105.886494_300.1","last_update":"2026-03-20T18:08:25.886494Z"},{"marker":"1_1774029751.502359_72.1","last_update":"2026-03-20T18:02:31.502359Z"},{"marker":"1_1774029751.504947_186.1","last_update":"2026-03-20T18:02:31.504947Z"},{"marker":"1_1774030105.855570_171.1","last_update":"2026-03-20T18:08:25.855570Z"}] 2026-03-20T18:08:26.006 INFO:rgw_multi.tests:starting meta checkpoint for zone=a2 2026-03-20T18:08:26.006 INFO:tasks.util.rgw:rgwadmin: c1.client.0 : ['metadata', 'sync', 'status', '--rgw-zone', 'a2', '--rgw-zonegroup', 'a', '--rgw-realm', 'test-realm', '--cluster', 'c1', '--debug-rgw', '0', '--debug-ms', '0', '--rgw-cache-enabled', 'false'] 2026-03-20T18:08:26.006 DEBUG:tasks.util.rgw:rgwadmin: cmd=['adjust-ulimits', 'ceph-coverage', '/home/ubuntu/cephtest/archive/coverage', 'radosgw-admin', '--log-to-stderr', '--format', 'json', '-n', 'client.0', '--cluster', 'c1', 'metadata', 'sync', 'status', '--rgw-zone', 'a2', '--rgw-zonegroup', 'a', '--rgw-realm', 'test-realm', '--cluster', 'c1', '--debug-rgw', '0', '--debug-ms', '0', '--rgw-cache-enabled', 'false'] 2026-03-20T18:08:26.006 DEBUG:teuthology.orchestra.run.vm02:> adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage radosgw-admin --log-to-stderr --format json -n client.0 --cluster c1 metadata sync status --rgw-zone a2 --rgw-zonegroup a --rgw-realm test-realm --cluster c1 --debug-rgw 0 --debug-ms 0 --rgw-cache-enabled false 2026-03-20T18:08:26.066 INFO:teuthology.orchestra.run.vm02.stderr:ignoring --setuser ceph since I am not root 2026-03-20T18:08:26.066 INFO:teuthology.orchestra.run.vm02.stderr:ignoring --setgroup ceph since I am not root 2026-03-20T18:08:26.136 INFO:teuthology.orchestra.run.vm02.stderr:2026-03-20T18:08:26.134+0000 7f5cfc787a80 20 RGW-SYNC:meta: read sync status 2026-03-20T18:08:26.137 INFO:teuthology.orchestra.run.vm02.stderr:2026-03-20T18:08:26.134+0000 7f5cfc787a80 20 RGW-SYNC:meta: read sync status 2026-03-20T18:08:26.146 INFO:teuthology.orchestra.run.vm02.stdout:{"sync_status":{"info":{"status":"sync","num_shards":4,"period":"82c662be-66b5-4c4f-b24e-ff9298fc8a6e","realm_epoch":2},"markers":[{"key":0,"val":{"state":1,"marker":"1_1774030105.886494_300.1","next_step_marker":"","total_entries":1,"pos":0,"timestamp":"2026-03-20T18:08:25.886494Z","realm_epoch":2}},{"key":1,"val":{"state":1,"marker":"1_1774029751.502359_72.1","next_step_marker":"","total_entries":0,"pos":0,"timestamp":"2026-03-20T18:02:31.502359Z","realm_epoch":2}},{"key":2,"val":{"state":1,"marker":"1_1774029751.504947_186.1","next_step_marker":"","total_entries":0,"pos":0,"timestamp":"2026-03-20T18:02:31.504947Z","realm_epoch":2}},{"key":3,"val":{"state":1,"marker":"1_1774030105.855570_171.1","next_step_marker":"","total_entries":0,"pos":0,"timestamp":"2026-03-20T18:08:25.855570Z","realm_epoch":2}}]},"full_sync":{"total":1,"complete":1},"current_time":"2026-03-20T18:08:26Z"} 2026-03-20T18:08:26.147 INFO:rgw_multi.tests:finish meta checkpoint for zone=a2 2026-03-20T18:08:26.276 INFO:tasks.util.rgw:rgwadmin: c1.client.0 : ['bucket', 'sync', 'checkpoint', '--bucket', 'cyjkqa-20', '--source-zone', 'a1', '--retry-delay-ms', '5000', '--timeout-sec', '300', '--rgw-zone', 'a2', '--rgw-zonegroup', 'a', '--rgw-realm', 'test-realm', '--cluster', 'c1', '--debug-rgw', '1', '--debug-ms', '0'] 2026-03-20T18:08:26.276 DEBUG:tasks.util.rgw:rgwadmin: cmd=['adjust-ulimits', 'ceph-coverage', '/home/ubuntu/cephtest/archive/coverage', 'radosgw-admin', '--log-to-stderr', '--format', 'json', '-n', 'client.0', '--cluster', 'c1', 'bucket', 'sync', 'checkpoint', '--bucket', 'cyjkqa-20', '--source-zone', 'a1', '--retry-delay-ms', '5000', '--timeout-sec', '300', '--rgw-zone', 'a2', '--rgw-zonegroup', 'a', '--rgw-realm', 'test-realm', '--cluster', 'c1', '--debug-rgw', '1', '--debug-ms', '0'] 2026-03-20T18:08:26.276 DEBUG:teuthology.orchestra.run.vm02:> adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage radosgw-admin --log-to-stderr --format json -n client.0 --cluster c1 bucket sync checkpoint --bucket cyjkqa-20 --source-zone a1 --retry-delay-ms 5000 --timeout-sec 300 --rgw-zone a2 --rgw-zonegroup a --rgw-realm test-realm --cluster c1 --debug-rgw 1 --debug-ms 0 2026-03-20T18:08:26.299 INFO:teuthology.orchestra.run.vm02.stderr:ignoring --setuser ceph since I am not root 2026-03-20T18:08:26.299 INFO:teuthology.orchestra.run.vm02.stderr:ignoring --setgroup ceph since I am not root 2026-03-20T18:08:26.375 INFO:teuthology.orchestra.run.vm02.stderr:2026-03-20T18:08:26.374+0000 7f0979189a80 1 waiting to reach incremental sync.. 2026-03-20T18:08:31.376 INFO:teuthology.orchestra.run.vm02.stderr:2026-03-20T18:08:31.374+0000 7f0979189a80 1 bucket sync caught up with source: 2026-03-20T18:08:31.376 INFO:teuthology.orchestra.run.vm02.stderr: local status: [, , , , , , , , 00000000005.22.14, , ] 2026-03-20T18:08:31.376 INFO:teuthology.orchestra.run.vm02.stderr: remote markers: [, , , , , , , , 00000000005.22.14, , ] 2026-03-20T18:08:31.376 INFO:teuthology.orchestra.run.vm02.stderr:2026-03-20T18:08:31.374+0000 7f0979189a80 0 bucket checkpoint complete 2026-03-20T18:08:31.386 INFO:tasks.util.rgw:rgwadmin: c1.client.0 : ['bucket', 'sync', 'checkpoint', '--bucket', 'cyjkqa-20', '--source-zone', 'a2', '--retry-delay-ms', '5000', '--timeout-sec', '300', '--rgw-zone', 'a1', '--rgw-zonegroup', 'a', '--rgw-realm', 'test-realm', '--cluster', 'c1', '--debug-rgw', '1', '--debug-ms', '0'] 2026-03-20T18:08:31.386 DEBUG:tasks.util.rgw:rgwadmin: cmd=['adjust-ulimits', 'ceph-coverage', '/home/ubuntu/cephtest/archive/coverage', 'radosgw-admin', '--log-to-stderr', '--format', 'json', '-n', 'client.0', '--cluster', 'c1', 'bucket', 'sync', 'checkpoint', '--bucket', 'cyjkqa-20', '--source-zone', 'a2', '--retry-delay-ms', '5000', '--timeout-sec', '300', '--rgw-zone', 'a1', '--rgw-zonegroup', 'a', '--rgw-realm', 'test-realm', '--cluster', 'c1', '--debug-rgw', '1', '--debug-ms', '0'] 2026-03-20T18:08:31.386 DEBUG:teuthology.orchestra.run.vm02:> adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage radosgw-admin --log-to-stderr --format json -n client.0 --cluster c1 bucket sync checkpoint --bucket cyjkqa-20 --source-zone a2 --retry-delay-ms 5000 --timeout-sec 300 --rgw-zone a1 --rgw-zonegroup a --rgw-realm test-realm --cluster c1 --debug-rgw 1 --debug-ms 0 2026-03-20T18:08:31.449 INFO:teuthology.orchestra.run.vm02.stderr:ignoring --setuser ceph since I am not root 2026-03-20T18:08:31.449 INFO:teuthology.orchestra.run.vm02.stderr:ignoring --setgroup ceph since I am not root 2026-03-20T18:08:31.536 INFO:teuthology.orchestra.run.vm02.stderr:2026-03-20T18:08:31.534+0000 7fd5a7729a80 1 waiting to reach incremental sync.. 2026-03-20T18:08:36.538 INFO:teuthology.orchestra.run.vm02.stderr:2026-03-20T18:08:36.534+0000 7fd5a7729a80 1 bucket sync caught up with source: 2026-03-20T18:08:36.538 INFO:teuthology.orchestra.run.vm02.stderr: local status: [, , , , , , , , 00000000001.15.14, , ] 2026-03-20T18:08:36.538 INFO:teuthology.orchestra.run.vm02.stderr: remote markers: [, , , , , , , , 00000000001.15.14, , ] 2026-03-20T18:08:36.538 INFO:teuthology.orchestra.run.vm02.stderr:2026-03-20T18:08:36.534+0000 7fd5a7729a80 0 bucket checkpoint complete 2026-03-20T18:08:36.548 INFO:tasks.rgw_multi.zone_rados:comparing bucket=cyjkqa-20 zones={a2, a1} 2026-03-20T18:08:36.555 DEBUG:tasks.rgw_multi.zone_rados:bucket1 objects: 2026-03-20T18:08:36.555 DEBUG:tasks.rgw_multi.zone_rados:o=obj, v=null 2026-03-20T18:08:36.558 DEBUG:tasks.rgw_multi.zone_rados:bucket2 objects: 2026-03-20T18:08:36.558 DEBUG:tasks.rgw_multi.zone_rados:o=obj, v=null 2026-03-20T18:08:36.558 DEBUG:tasks.rgw_multi.zone_rados:both are delete markers, skipping content comparison 2026-03-20T18:08:36.558 INFO:tasks.rgw_multi.zone_rados:success, bucket identical: bucket=cyjkqa-20 zones={a2, a1} 2026-03-20T18:08:36.558 INFO:tasks.util.rgw:rgwadmin: c1.client.0 : ['bucket', 'sync', 'checkpoint', '--bucket', 'cyjkqa-21', '--source-zone', 'a1', '--retry-delay-ms', '5000', '--timeout-sec', '300', '--rgw-zone', 'a2', '--rgw-zonegroup', 'a', '--rgw-realm', 'test-realm', '--cluster', 'c1', '--debug-rgw', '1', '--debug-ms', '0'] 2026-03-20T18:08:36.558 DEBUG:tasks.util.rgw:rgwadmin: cmd=['adjust-ulimits', 'ceph-coverage', '/home/ubuntu/cephtest/archive/coverage', 'radosgw-admin', '--log-to-stderr', '--format', 'json', '-n', 'client.0', '--cluster', 'c1', 'bucket', 'sync', 'checkpoint', '--bucket', 'cyjkqa-21', '--source-zone', 'a1', '--retry-delay-ms', '5000', '--timeout-sec', '300', '--rgw-zone', 'a2', '--rgw-zonegroup', 'a', '--rgw-realm', 'test-realm', '--cluster', 'c1', '--debug-rgw', '1', '--debug-ms', '0'] 2026-03-20T18:08:36.558 DEBUG:teuthology.orchestra.run.vm02:> adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage radosgw-admin --log-to-stderr --format json -n client.0 --cluster c1 bucket sync checkpoint --bucket cyjkqa-21 --source-zone a1 --retry-delay-ms 5000 --timeout-sec 300 --rgw-zone a2 --rgw-zonegroup a --rgw-realm test-realm --cluster c1 --debug-rgw 1 --debug-ms 0 2026-03-20T18:08:36.613 INFO:teuthology.orchestra.run.vm02.stderr:ignoring --setuser ceph since I am not root 2026-03-20T18:08:36.613 INFO:teuthology.orchestra.run.vm02.stderr:ignoring --setgroup ceph since I am not root 2026-03-20T18:08:36.686 INFO:teuthology.orchestra.run.vm02.stderr:2026-03-20T18:08:36.682+0000 7fc13877da80 1 bucket sync caught up with source: 2026-03-20T18:08:36.686 INFO:teuthology.orchestra.run.vm02.stderr: local status: [, , , , , , , , 00000000001.3.14, , ] 2026-03-20T18:08:36.686 INFO:teuthology.orchestra.run.vm02.stderr: remote markers: [, , , , , , , , 00000000001.3.14, , ] 2026-03-20T18:08:36.686 INFO:teuthology.orchestra.run.vm02.stderr:2026-03-20T18:08:36.682+0000 7fc13877da80 0 bucket checkpoint complete 2026-03-20T18:08:36.695 INFO:tasks.util.rgw:rgwadmin: c1.client.0 : ['bucket', 'sync', 'checkpoint', '--bucket', 'cyjkqa-21', '--source-zone', 'a2', '--retry-delay-ms', '5000', '--timeout-sec', '300', '--rgw-zone', 'a1', '--rgw-zonegroup', 'a', '--rgw-realm', 'test-realm', '--cluster', 'c1', '--debug-rgw', '1', '--debug-ms', '0'] 2026-03-20T18:08:36.695 DEBUG:tasks.util.rgw:rgwadmin: cmd=['adjust-ulimits', 'ceph-coverage', '/home/ubuntu/cephtest/archive/coverage', 'radosgw-admin', '--log-to-stderr', '--format', 'json', '-n', 'client.0', '--cluster', 'c1', 'bucket', 'sync', 'checkpoint', '--bucket', 'cyjkqa-21', '--source-zone', 'a2', '--retry-delay-ms', '5000', '--timeout-sec', '300', '--rgw-zone', 'a1', '--rgw-zonegroup', 'a', '--rgw-realm', 'test-realm', '--cluster', 'c1', '--debug-rgw', '1', '--debug-ms', '0'] 2026-03-20T18:08:36.695 DEBUG:teuthology.orchestra.run.vm02:> adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage radosgw-admin --log-to-stderr --format json -n client.0 --cluster c1 bucket sync checkpoint --bucket cyjkqa-21 --source-zone a2 --retry-delay-ms 5000 --timeout-sec 300 --rgw-zone a1 --rgw-zonegroup a --rgw-realm test-realm --cluster c1 --debug-rgw 1 --debug-ms 0 2026-03-20T18:08:36.765 INFO:teuthology.orchestra.run.vm02.stderr:ignoring --setuser ceph since I am not root 2026-03-20T18:08:36.765 INFO:teuthology.orchestra.run.vm02.stderr:ignoring --setgroup ceph since I am not root 2026-03-20T18:08:36.839 INFO:teuthology.orchestra.run.vm02.stderr:2026-03-20T18:08:36.838+0000 7efdd8534a80 1 bucket sync caught up with source: 2026-03-20T18:08:36.839 INFO:teuthology.orchestra.run.vm02.stderr: local status: [, , , , , , , , 00000000005.18.14, , ] 2026-03-20T18:08:36.839 INFO:teuthology.orchestra.run.vm02.stderr: remote markers: [, , , , , , , , 00000000005.18.14, , ] 2026-03-20T18:08:36.839 INFO:teuthology.orchestra.run.vm02.stderr:2026-03-20T18:08:36.838+0000 7efdd8534a80 0 bucket checkpoint complete 2026-03-20T18:08:36.848 INFO:tasks.rgw_multi.zone_rados:comparing bucket=cyjkqa-21 zones={a2, a1} 2026-03-20T18:08:36.853 DEBUG:tasks.rgw_multi.zone_rados:bucket1 objects: 2026-03-20T18:08:36.853 DEBUG:tasks.rgw_multi.zone_rados:o=obj, v=null 2026-03-20T18:08:36.856 DEBUG:tasks.rgw_multi.zone_rados:bucket2 objects: 2026-03-20T18:08:36.856 DEBUG:tasks.rgw_multi.zone_rados:o=obj, v=null 2026-03-20T18:08:36.856 DEBUG:tasks.rgw_multi.zone_rados:both are delete markers, skipping content comparison 2026-03-20T18:08:36.856 INFO:tasks.rgw_multi.zone_rados:success, bucket identical: bucket=cyjkqa-21 zones={a2, a1} 2026-03-20T18:08:36.858 INFO:tasks.rgw_multisite_tests:rgw_multi.tests.test_suspended_delete_marker_full_sync ... ok 2026-03-20T18:08:36.971 INFO:tasks.util.rgw:rgwadmin: c1.client.0 : ['mdlog', 'status', '--rgw-zone', 'a1', '--rgw-zonegroup', 'a', '--rgw-realm', 'test-realm', '--cluster', 'c1', '--debug-rgw', '0', '--debug-ms', '0', '--rgw-cache-enabled', 'false'] 2026-03-20T18:08:36.971 DEBUG:tasks.util.rgw:rgwadmin: cmd=['adjust-ulimits', 'ceph-coverage', '/home/ubuntu/cephtest/archive/coverage', 'radosgw-admin', '--log-to-stderr', '--format', 'json', '-n', 'client.0', '--cluster', 'c1', 'mdlog', 'status', '--rgw-zone', 'a1', '--rgw-zonegroup', 'a', '--rgw-realm', 'test-realm', '--cluster', 'c1', '--debug-rgw', '0', '--debug-ms', '0', '--rgw-cache-enabled', 'false'] 2026-03-20T18:08:36.971 DEBUG:teuthology.orchestra.run.vm02:> adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage radosgw-admin --log-to-stderr --format json -n client.0 --cluster c1 mdlog status --rgw-zone a1 --rgw-zonegroup a --rgw-realm test-realm --cluster c1 --debug-rgw 0 --debug-ms 0 --rgw-cache-enabled false 2026-03-20T18:08:36.997 INFO:teuthology.orchestra.run.vm02.stderr:ignoring --setuser ceph since I am not root 2026-03-20T18:08:36.998 INFO:teuthology.orchestra.run.vm02.stderr:ignoring --setgroup ceph since I am not root 2026-03-20T18:08:37.063 INFO:teuthology.orchestra.run.vm02.stderr:No --period given, using current period=82c662be-66b5-4c4f-b24e-ff9298fc8a6e 2026-03-20T18:08:37.073 INFO:teuthology.orchestra.run.vm02.stdout:[{"marker":"1_1774030116.970001_320.1","last_update":"2026-03-20T18:08:36.970001Z"},{"marker":"1_1774029751.502359_72.1","last_update":"2026-03-20T18:02:31.502359Z"},{"marker":"1_1774029751.504947_186.1","last_update":"2026-03-20T18:02:31.504947Z"},{"marker":"1_1774030116.963516_172.1","last_update":"2026-03-20T18:08:36.963516Z"}] 2026-03-20T18:08:37.073 INFO:rgw_multi.tests:starting meta checkpoint for zone=a2 2026-03-20T18:08:37.073 INFO:tasks.util.rgw:rgwadmin: c1.client.0 : ['metadata', 'sync', 'status', '--rgw-zone', 'a2', '--rgw-zonegroup', 'a', '--rgw-realm', 'test-realm', '--cluster', 'c1', '--debug-rgw', '0', '--debug-ms', '0', '--rgw-cache-enabled', 'false'] 2026-03-20T18:08:37.073 DEBUG:tasks.util.rgw:rgwadmin: cmd=['adjust-ulimits', 'ceph-coverage', '/home/ubuntu/cephtest/archive/coverage', 'radosgw-admin', '--log-to-stderr', '--format', 'json', '-n', 'client.0', '--cluster', 'c1', 'metadata', 'sync', 'status', '--rgw-zone', 'a2', '--rgw-zonegroup', 'a', '--rgw-realm', 'test-realm', '--cluster', 'c1', '--debug-rgw', '0', '--debug-ms', '0', '--rgw-cache-enabled', 'false'] 2026-03-20T18:08:37.073 DEBUG:teuthology.orchestra.run.vm02:> adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage radosgw-admin --log-to-stderr --format json -n client.0 --cluster c1 metadata sync status --rgw-zone a2 --rgw-zonegroup a --rgw-realm test-realm --cluster c1 --debug-rgw 0 --debug-ms 0 --rgw-cache-enabled false 2026-03-20T18:08:37.135 INFO:teuthology.orchestra.run.vm02.stderr:ignoring --setuser ceph since I am not root 2026-03-20T18:08:37.135 INFO:teuthology.orchestra.run.vm02.stderr:ignoring --setgroup ceph since I am not root 2026-03-20T18:08:37.196 INFO:teuthology.orchestra.run.vm02.stderr:2026-03-20T18:08:37.194+0000 7fc5a267aa80 20 RGW-SYNC:meta: read sync status 2026-03-20T18:08:37.196 INFO:teuthology.orchestra.run.vm02.stderr:2026-03-20T18:08:37.194+0000 7fc5a267aa80 20 RGW-SYNC:meta: read sync status 2026-03-20T18:08:37.206 INFO:teuthology.orchestra.run.vm02.stdout:{"sync_status":{"info":{"status":"sync","num_shards":4,"period":"82c662be-66b5-4c4f-b24e-ff9298fc8a6e","realm_epoch":2},"markers":[{"key":0,"val":{"state":1,"marker":"1_1774030116.970001_320.1","next_step_marker":"","total_entries":1,"pos":0,"timestamp":"2026-03-20T18:08:36.970001Z","realm_epoch":2}},{"key":1,"val":{"state":1,"marker":"1_1774029751.502359_72.1","next_step_marker":"","total_entries":0,"pos":0,"timestamp":"2026-03-20T18:02:31.502359Z","realm_epoch":2}},{"key":2,"val":{"state":1,"marker":"1_1774029751.504947_186.1","next_step_marker":"","total_entries":0,"pos":0,"timestamp":"2026-03-20T18:02:31.504947Z","realm_epoch":2}},{"key":3,"val":{"state":1,"marker":"1_1774030116.963516_172.1","next_step_marker":"","total_entries":0,"pos":0,"timestamp":"2026-03-20T18:08:36.963516Z","realm_epoch":2}}]},"full_sync":{"total":1,"complete":1},"current_time":"2026-03-20T18:08:37Z"} 2026-03-20T18:08:37.206 INFO:rgw_multi.tests:finish meta checkpoint for zone=a2 2026-03-20T18:08:37.219 INFO:tasks.util.rgw:rgwadmin: c1.client.0 : ['bucket', 'sync', 'checkpoint', '--bucket', 'cyjkqa-22', '--source-zone', 'a1', '--retry-delay-ms', '5000', '--timeout-sec', '300', '--rgw-zone', 'a2', '--rgw-zonegroup', 'a', '--rgw-realm', 'test-realm', '--cluster', 'c1', '--debug-rgw', '1', '--debug-ms', '0'] 2026-03-20T18:08:37.219 DEBUG:tasks.util.rgw:rgwadmin: cmd=['adjust-ulimits', 'ceph-coverage', '/home/ubuntu/cephtest/archive/coverage', 'radosgw-admin', '--log-to-stderr', '--format', 'json', '-n', 'client.0', '--cluster', 'c1', 'bucket', 'sync', 'checkpoint', '--bucket', 'cyjkqa-22', '--source-zone', 'a1', '--retry-delay-ms', '5000', '--timeout-sec', '300', '--rgw-zone', 'a2', '--rgw-zonegroup', 'a', '--rgw-realm', 'test-realm', '--cluster', 'c1', '--debug-rgw', '1', '--debug-ms', '0'] 2026-03-20T18:08:37.219 DEBUG:teuthology.orchestra.run.vm02:> adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage radosgw-admin --log-to-stderr --format json -n client.0 --cluster c1 bucket sync checkpoint --bucket cyjkqa-22 --source-zone a1 --retry-delay-ms 5000 --timeout-sec 300 --rgw-zone a2 --rgw-zonegroup a --rgw-realm test-realm --cluster c1 --debug-rgw 1 --debug-ms 0 2026-03-20T18:08:37.269 INFO:teuthology.orchestra.run.vm02.stderr:ignoring --setuser ceph since I am not root 2026-03-20T18:08:37.269 INFO:teuthology.orchestra.run.vm02.stderr:ignoring --setgroup ceph since I am not root 2026-03-20T18:08:37.340 INFO:teuthology.orchestra.run.vm02.stderr:2026-03-20T18:08:37.338+0000 7fa5a2715a80 1 waiting to reach incremental sync.. 2026-03-20T18:08:42.341 INFO:teuthology.orchestra.run.vm02.stderr:2026-03-20T18:08:42.338+0000 7fa5a2715a80 1 bucket sync caught up with source: 2026-03-20T18:08:42.341 INFO:teuthology.orchestra.run.vm02.stderr: local status: [, , , , , , , , 00000000002.7.11, , ] 2026-03-20T18:08:42.341 INFO:teuthology.orchestra.run.vm02.stderr: remote markers: [, , , , , , , , 00000000002.7.11, , ] 2026-03-20T18:08:42.341 INFO:teuthology.orchestra.run.vm02.stderr:2026-03-20T18:08:42.338+0000 7fa5a2715a80 0 bucket checkpoint complete 2026-03-20T18:08:42.351 INFO:tasks.util.rgw:rgwadmin: c1.client.0 : ['bucket', 'sync', 'checkpoint', '--bucket', 'cyjkqa-22', '--source-zone', 'a2', '--retry-delay-ms', '5000', '--timeout-sec', '300', '--rgw-zone', 'a1', '--rgw-zonegroup', 'a', '--rgw-realm', 'test-realm', '--cluster', 'c1', '--debug-rgw', '1', '--debug-ms', '0'] 2026-03-20T18:08:42.351 DEBUG:tasks.util.rgw:rgwadmin: cmd=['adjust-ulimits', 'ceph-coverage', '/home/ubuntu/cephtest/archive/coverage', 'radosgw-admin', '--log-to-stderr', '--format', 'json', '-n', 'client.0', '--cluster', 'c1', 'bucket', 'sync', 'checkpoint', '--bucket', 'cyjkqa-22', '--source-zone', 'a2', '--retry-delay-ms', '5000', '--timeout-sec', '300', '--rgw-zone', 'a1', '--rgw-zonegroup', 'a', '--rgw-realm', 'test-realm', '--cluster', 'c1', '--debug-rgw', '1', '--debug-ms', '0'] 2026-03-20T18:08:42.351 DEBUG:teuthology.orchestra.run.vm02:> adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage radosgw-admin --log-to-stderr --format json -n client.0 --cluster c1 bucket sync checkpoint --bucket cyjkqa-22 --source-zone a2 --retry-delay-ms 5000 --timeout-sec 300 --rgw-zone a1 --rgw-zonegroup a --rgw-realm test-realm --cluster c1 --debug-rgw 1 --debug-ms 0 2026-03-20T18:08:42.417 INFO:teuthology.orchestra.run.vm02.stderr:ignoring --setuser ceph since I am not root 2026-03-20T18:08:42.417 INFO:teuthology.orchestra.run.vm02.stderr:ignoring --setgroup ceph since I am not root 2026-03-20T18:08:42.491 INFO:teuthology.orchestra.run.vm02.stderr:2026-03-20T18:08:42.486+0000 7ffaecc75a80 1 waiting to reach incremental sync.. 2026-03-20T18:08:43.481 DEBUG:teuthology.exit:Got signal 15; running 1 handler... 2026-03-20T18:08:43.482 DEBUG:teuthology.exit:Finished running handlers