2026-03-06T13:36:07.884 INFO:root:teuthology version: 1.2.4.dev6+g1c580df7a 2026-03-06T13:36:07.888 DEBUG:teuthology.report:Pushing job info to http://localhost:8080 2026-03-06T13:36:07.912 INFO:teuthology.run:Config: archive_path: /archive/irq0-2026-03-06_13:20:18-orch:cephadm:workunits-cobaltcore-storage-v19.2.3-fasttrack-3-none-default-vps/273 branch: cobaltcore-storage-v19.2.3-fasttrack-3 description: orch:cephadm:workunits/{0-distro/centos_9.stream agent/off mon_election/classic task/test_rgw_multisite} email: null first_in_suite: false flavor: default job_id: '273' last_in_suite: false machine_type: vps name: irq0-2026-03-06_13:20:18-orch:cephadm:workunits-cobaltcore-storage-v19.2.3-fasttrack-3-none-default-vps no_nested_subset: false os_type: centos os_version: 9.stream overrides: admin_socket: branch: cobaltcore-storage-v19.2.3-fasttrack-3 ansible.cephlab: branch: main repo: https://github.com/kshtsk/ceph-cm-ansible.git skip_tags: nagios,monitoring-scripts,hostname,pubkeys,zap,sudoers,kerberos,ntp-client,resolvconf,cpan,nfs vars: timezone: Europe/Berlin ceph: conf: global: mon election default strategy: 1 mgr: debug mgr: 20 debug ms: 1 mgr/cephadm/use_agent: false mon: debug mon: 20 debug ms: 1 debug paxos: 20 osd: debug ms: 1 debug osd: 20 osd mclock iops capacity threshold hdd: 49000 flavor: default log-ignorelist: - \(MDS_ALL_DOWN\) - \(MDS_UP_LESS_THAN_MAX\) - MON_DOWN - mons down - mon down - out of quorum - CEPHADM_STRAY_DAEMON log-only-match: - CEPHADM_ sha1: c24117fd5525679b799527bc1bd1f1dd0a2db5e2 ceph-deploy: conf: client: log file: /var/log/ceph/ceph-$name.$pid.log mon: {} cephadm: cephadm_binary_url: https://download.ceph.com/rpm-19.2.3/el9/noarch/cephadm containers: image: harbor.clyso.com/custom-ceph/ceph/ceph:cobaltcore-storage-v19.2.3-fasttrack-3 install: ceph: flavor: default sha1: c24117fd5525679b799527bc1bd1f1dd0a2db5e2 extra_system_packages: deb: - python3-xmltodict - s3cmd rpm: - bzip2 - perl-Test-Harness - python3-xmltodict - s3cmd repos: - name: ceph-source priority: 1 url: https://s3.clyso.com/ces-packages/components/ceph/rpm-19.2.3-47-gc24117fd552/el9.clyso/SRPMS - name: ceph-noarch priority: 1 url: https://s3.clyso.com/ces-packages/components/ceph/rpm-19.2.3-47-gc24117fd552/el9.clyso/noarch - name: ceph priority: 1 url: https://s3.clyso.com/ces-packages/components/ceph/rpm-19.2.3-47-gc24117fd552/el9.clyso/x86_64 selinux: allowlist: - scontext=system_u:system_r:logrotate_t:s0 - scontext=system_u:system_r:getty_t:s0 workunit: branch: tt-19.2.3-fasttrack-3-no-nvme-loop sha1: 5726a36c3452e5b72190cfceba828abc62c819b7 owner: irq0 priority: 1000 repo: https://github.com/ceph/ceph.git roles: - - host.a - mon.a - mgr.a - osd.0 - - host.b - mon.b - mgr.b - osd.1 - - host.c - mon.c - osd.2 seed: 6609 sha1: c24117fd5525679b799527bc1bd1f1dd0a2db5e2 sleep_before_teardown: 0 subset: 1/64 suite: orch:cephadm:workunits suite_branch: tt-19.2.3-fasttrack-3-no-nvme-loop suite_path: /home/teuthos/src/github.com_kshtsk_ceph_5726a36c3452e5b72190cfceba828abc62c819b7/qa suite_relpath: qa suite_repo: https://github.com/kshtsk/ceph.git suite_sha1: 5726a36c3452e5b72190cfceba828abc62c819b7 targets: vm00.local: ecdsa-sha2-nistp256 AAAAE2VjZHNhLXNoYTItbmlzdHAyNTYAAAAIbmlzdHAyNTYAAABBBAbtRXihk0uf59Zoqmi67xir7ZVI3/LMpIv6PXBUp5AjDQcl6SKXj1uDmbjwiCOdBqGF4Pc/6W6hDF4Ql8QmUxc= vm01.local: ecdsa-sha2-nistp256 AAAAE2VjZHNhLXNoYTItbmlzdHAyNTYAAAAIbmlzdHAyNTYAAABBBGznuUvPwG9TbB8LGsGA+26k2Bljz4hkp4P4qX/UXdzVcCM/2trx1Rc6OSG2UCSJvXOxO9Kik3SlvD0BBbKUzF4= vm08.local: ecdsa-sha2-nistp256 AAAAE2VjZHNhLXNoYTItbmlzdHAyNTYAAAAIbmlzdHAyNTYAAABBBMRA78P8414AciGh1BOg1I/H6Xvqpep8VkXvZ1FNU/1LD/V8UizgpOBiK6LaSeUHH1gGuKwgmaQ79rxjyztj+WE= tasks: - pexec: all: - sudo dnf remove nvme-cli -y - sudo dnf install nvmetcli nvme-cli -y - install: null - cephadm: null - cephadm.shell: host.a: - ceph mgr module enable rgw - rgw_module.apply: specs: - rgw_realm: myrealm1 rgw_zone: myzone1 rgw_zonegroup: myzonegroup1 spec: rgw_frontend_port: 5500 - cephadm.shell: host.a: - 'set -e set -x while true; do TOKEN=$(ceph rgw realm tokens | jq -r ''.[0].token''); echo $TOKEN; if [ "$TOKEN" != "master zone has no endpoint" ]; then break; fi; sleep 5; done TOKENS=$(ceph rgw realm tokens) echo $TOKENS | jq --exit-status ''.[0].realm == "myrealm1"'' echo $TOKENS | jq --exit-status ''.[0].token'' TOKEN_JSON=$(ceph rgw realm tokens | jq -r ''.[0].token'' | base64 --decode) echo $TOKEN_JSON | jq --exit-status ''.realm_name == "myrealm1"'' echo $TOKEN_JSON | jq --exit-status ''.endpoint | test("http://.+:\\d+")'' echo $TOKEN_JSON | jq --exit-status ''.realm_id | test("^[0-9a-f]{8}-[0-9a-f]{4}-[0-9a-f]{4}-[0-9a-f]{4}-[0-9a-f]{12}$")'' echo $TOKEN_JSON | jq --exit-status ''.access_key'' echo $TOKEN_JSON | jq --exit-status ''.secret'' ' teuthology: fragments_dropped: [] meta: {} postmerge: [] teuthology_branch: clyso-debian-13 teuthology_repo: https://github.com/clyso/teuthology teuthology_sha1: 1c580df7a9c7c2aadc272da296344fd99f27c444 timestamp: 2026-03-06_13:20:18 tube: vps user: irq0 verbose: false worker_log: /home/teuthos/.teuthology/dispatcher/dispatcher.vps.43333 2026-03-06T13:36:07.912 INFO:teuthology.run:suite_path is set to /home/teuthos/src/github.com_kshtsk_ceph_5726a36c3452e5b72190cfceba828abc62c819b7/qa; will attempt to use it 2026-03-06T13:36:07.913 INFO:teuthology.run:Found tasks at /home/teuthos/src/github.com_kshtsk_ceph_5726a36c3452e5b72190cfceba828abc62c819b7/qa/tasks 2026-03-06T13:36:07.913 INFO:teuthology.run_tasks:Running task internal.save_config... 2026-03-06T13:36:07.913 INFO:teuthology.task.internal:Saving configuration 2026-03-06T13:36:07.918 INFO:teuthology.run_tasks:Running task internal.check_lock... 2026-03-06T13:36:07.919 INFO:teuthology.task.internal.check_lock:Checking locks... 2026-03-06T13:36:07.925 DEBUG:teuthology.task.internal.check_lock:machine status is {'name': 'vm00.local', 'description': '/archive/irq0-2026-03-06_13:20:18-orch:cephadm:workunits-cobaltcore-storage-v19.2.3-fasttrack-3-none-default-vps/273', 'up': True, 'machine_type': 'vps', 'is_vm': True, 'vm_host': {'name': 'localhost', 'description': None, 'up': True, 'machine_type': 'libvirt', 'is_vm': False, 'vm_host': None, 'os_type': None, 'os_version': None, 'arch': None, 'locked': True, 'locked_since': None, 'locked_by': None, 'mac_address': None, 'ssh_pub_key': None}, 'os_type': 'centos', 'os_version': '9.stream', 'arch': 'x86_64', 'locked': True, 'locked_since': '2026-03-06 12:34:16.196188', 'locked_by': 'irq0', 'mac_address': '52:55:00:00:00:00', 'ssh_pub_key': 'ecdsa-sha2-nistp256 AAAAE2VjZHNhLXNoYTItbmlzdHAyNTYAAAAIbmlzdHAyNTYAAABBBAbtRXihk0uf59Zoqmi67xir7ZVI3/LMpIv6PXBUp5AjDQcl6SKXj1uDmbjwiCOdBqGF4Pc/6W6hDF4Ql8QmUxc='} 2026-03-06T13:36:07.930 DEBUG:teuthology.task.internal.check_lock:machine status is {'name': 'vm01.local', 'description': '/archive/irq0-2026-03-06_13:20:18-orch:cephadm:workunits-cobaltcore-storage-v19.2.3-fasttrack-3-none-default-vps/273', 'up': True, 'machine_type': 'vps', 'is_vm': True, 'vm_host': {'name': 'localhost', 'description': None, 'up': True, 'machine_type': 'libvirt', 'is_vm': False, 'vm_host': None, 'os_type': None, 'os_version': None, 'arch': None, 'locked': True, 'locked_since': None, 'locked_by': None, 'mac_address': None, 'ssh_pub_key': None}, 'os_type': 'centos', 'os_version': '9.stream', 'arch': 'x86_64', 'locked': True, 'locked_since': '2026-03-06 12:34:16.196460', 'locked_by': 'irq0', 'mac_address': '52:55:00:00:00:01', 'ssh_pub_key': 'ecdsa-sha2-nistp256 AAAAE2VjZHNhLXNoYTItbmlzdHAyNTYAAAAIbmlzdHAyNTYAAABBBGznuUvPwG9TbB8LGsGA+26k2Bljz4hkp4P4qX/UXdzVcCM/2trx1Rc6OSG2UCSJvXOxO9Kik3SlvD0BBbKUzF4='} 2026-03-06T13:36:07.936 DEBUG:teuthology.task.internal.check_lock:machine status is {'name': 'vm08.local', 'description': '/archive/irq0-2026-03-06_13:20:18-orch:cephadm:workunits-cobaltcore-storage-v19.2.3-fasttrack-3-none-default-vps/273', 'up': True, 'machine_type': 'vps', 'is_vm': True, 'vm_host': {'name': 'localhost', 'description': None, 'up': True, 'machine_type': 'libvirt', 'is_vm': False, 'vm_host': None, 'os_type': None, 'os_version': None, 'arch': None, 'locked': True, 'locked_since': None, 'locked_by': None, 'mac_address': None, 'ssh_pub_key': None}, 'os_type': 'centos', 'os_version': '9.stream', 'arch': 'x86_64', 'locked': True, 'locked_since': '2026-03-06 12:34:16.195721', 'locked_by': 'irq0', 'mac_address': '52:55:00:00:00:08', 'ssh_pub_key': 'ecdsa-sha2-nistp256 AAAAE2VjZHNhLXNoYTItbmlzdHAyNTYAAAAIbmlzdHAyNTYAAABBBMRA78P8414AciGh1BOg1I/H6Xvqpep8VkXvZ1FNU/1LD/V8UizgpOBiK6LaSeUHH1gGuKwgmaQ79rxjyztj+WE='} 2026-03-06T13:36:07.936 INFO:teuthology.run_tasks:Running task internal.add_remotes... 2026-03-06T13:36:07.936 INFO:teuthology.task.internal:roles: ubuntu@vm00.local - ['host.a', 'mon.a', 'mgr.a', 'osd.0'] 2026-03-06T13:36:07.937 INFO:teuthology.task.internal:roles: ubuntu@vm01.local - ['host.b', 'mon.b', 'mgr.b', 'osd.1'] 2026-03-06T13:36:07.937 INFO:teuthology.task.internal:roles: ubuntu@vm08.local - ['host.c', 'mon.c', 'osd.2'] 2026-03-06T13:36:07.937 INFO:teuthology.run_tasks:Running task console_log... 2026-03-06T13:36:07.943 DEBUG:teuthology.task.console_log:vm00 does not support IPMI; excluding 2026-03-06T13:36:07.949 DEBUG:teuthology.task.console_log:vm01 does not support IPMI; excluding 2026-03-06T13:36:07.955 DEBUG:teuthology.task.console_log:vm08 does not support IPMI; excluding 2026-03-06T13:36:07.955 DEBUG:teuthology.exit:Installing handler: Handler(exiter=, func=.kill_console_loggers at 0x7f5d5acf7d90>, signals=[15]) 2026-03-06T13:36:07.955 INFO:teuthology.run_tasks:Running task internal.connect... 2026-03-06T13:36:07.955 INFO:teuthology.task.internal:Opening connections... 2026-03-06T13:36:07.955 DEBUG:teuthology.task.internal:connecting to ubuntu@vm00.local 2026-03-06T13:36:07.956 DEBUG:teuthology.orchestra.connection:{'hostname': 'vm00.local', 'username': 'ubuntu', 'timeout': 60} 2026-03-06T13:36:08.016 DEBUG:teuthology.task.internal:connecting to ubuntu@vm01.local 2026-03-06T13:36:08.016 DEBUG:teuthology.orchestra.connection:{'hostname': 'vm01.local', 'username': 'ubuntu', 'timeout': 60} 2026-03-06T13:36:08.075 DEBUG:teuthology.task.internal:connecting to ubuntu@vm08.local 2026-03-06T13:36:08.076 DEBUG:teuthology.orchestra.connection:{'hostname': 'vm08.local', 'username': 'ubuntu', 'timeout': 60} 2026-03-06T13:36:08.134 INFO:teuthology.run_tasks:Running task internal.push_inventory... 2026-03-06T13:36:08.135 DEBUG:teuthology.orchestra.run.vm00:> uname -m 2026-03-06T13:36:08.151 INFO:teuthology.orchestra.run.vm00.stdout:x86_64 2026-03-06T13:36:08.151 DEBUG:teuthology.orchestra.run.vm00:> cat /etc/os-release 2026-03-06T13:36:08.205 INFO:teuthology.orchestra.run.vm00.stdout:NAME="CentOS Stream" 2026-03-06T13:36:08.205 INFO:teuthology.orchestra.run.vm00.stdout:VERSION="9" 2026-03-06T13:36:08.205 INFO:teuthology.orchestra.run.vm00.stdout:ID="centos" 2026-03-06T13:36:08.205 INFO:teuthology.orchestra.run.vm00.stdout:ID_LIKE="rhel fedora" 2026-03-06T13:36:08.205 INFO:teuthology.orchestra.run.vm00.stdout:VERSION_ID="9" 2026-03-06T13:36:08.205 INFO:teuthology.orchestra.run.vm00.stdout:PLATFORM_ID="platform:el9" 2026-03-06T13:36:08.205 INFO:teuthology.orchestra.run.vm00.stdout:PRETTY_NAME="CentOS Stream 9" 2026-03-06T13:36:08.205 INFO:teuthology.orchestra.run.vm00.stdout:ANSI_COLOR="0;31" 2026-03-06T13:36:08.205 INFO:teuthology.orchestra.run.vm00.stdout:LOGO="fedora-logo-icon" 2026-03-06T13:36:08.205 INFO:teuthology.orchestra.run.vm00.stdout:CPE_NAME="cpe:/o:centos:centos:9" 2026-03-06T13:36:08.205 INFO:teuthology.orchestra.run.vm00.stdout:HOME_URL="https://centos.org/" 2026-03-06T13:36:08.205 INFO:teuthology.orchestra.run.vm00.stdout:BUG_REPORT_URL="https://issues.redhat.com/" 2026-03-06T13:36:08.205 INFO:teuthology.orchestra.run.vm00.stdout:REDHAT_SUPPORT_PRODUCT="Red Hat Enterprise Linux 9" 2026-03-06T13:36:08.205 INFO:teuthology.orchestra.run.vm00.stdout:REDHAT_SUPPORT_PRODUCT_VERSION="CentOS Stream" 2026-03-06T13:36:08.205 INFO:teuthology.lock.ops:Updating vm00.local on lock server 2026-03-06T13:36:08.211 DEBUG:teuthology.orchestra.run.vm01:> uname -m 2026-03-06T13:36:08.228 INFO:teuthology.orchestra.run.vm01.stdout:x86_64 2026-03-06T13:36:08.229 DEBUG:teuthology.orchestra.run.vm01:> cat /etc/os-release 2026-03-06T13:36:08.282 INFO:teuthology.orchestra.run.vm01.stdout:NAME="CentOS Stream" 2026-03-06T13:36:08.282 INFO:teuthology.orchestra.run.vm01.stdout:VERSION="9" 2026-03-06T13:36:08.282 INFO:teuthology.orchestra.run.vm01.stdout:ID="centos" 2026-03-06T13:36:08.282 INFO:teuthology.orchestra.run.vm01.stdout:ID_LIKE="rhel fedora" 2026-03-06T13:36:08.282 INFO:teuthology.orchestra.run.vm01.stdout:VERSION_ID="9" 2026-03-06T13:36:08.282 INFO:teuthology.orchestra.run.vm01.stdout:PLATFORM_ID="platform:el9" 2026-03-06T13:36:08.282 INFO:teuthology.orchestra.run.vm01.stdout:PRETTY_NAME="CentOS Stream 9" 2026-03-06T13:36:08.282 INFO:teuthology.orchestra.run.vm01.stdout:ANSI_COLOR="0;31" 2026-03-06T13:36:08.282 INFO:teuthology.orchestra.run.vm01.stdout:LOGO="fedora-logo-icon" 2026-03-06T13:36:08.282 INFO:teuthology.orchestra.run.vm01.stdout:CPE_NAME="cpe:/o:centos:centos:9" 2026-03-06T13:36:08.282 INFO:teuthology.orchestra.run.vm01.stdout:HOME_URL="https://centos.org/" 2026-03-06T13:36:08.282 INFO:teuthology.orchestra.run.vm01.stdout:BUG_REPORT_URL="https://issues.redhat.com/" 2026-03-06T13:36:08.282 INFO:teuthology.orchestra.run.vm01.stdout:REDHAT_SUPPORT_PRODUCT="Red Hat Enterprise Linux 9" 2026-03-06T13:36:08.282 INFO:teuthology.orchestra.run.vm01.stdout:REDHAT_SUPPORT_PRODUCT_VERSION="CentOS Stream" 2026-03-06T13:36:08.282 INFO:teuthology.lock.ops:Updating vm01.local on lock server 2026-03-06T13:36:08.287 DEBUG:teuthology.orchestra.run.vm08:> uname -m 2026-03-06T13:36:08.301 INFO:teuthology.orchestra.run.vm08.stdout:x86_64 2026-03-06T13:36:08.301 DEBUG:teuthology.orchestra.run.vm08:> cat /etc/os-release 2026-03-06T13:36:08.357 INFO:teuthology.orchestra.run.vm08.stdout:NAME="CentOS Stream" 2026-03-06T13:36:08.357 INFO:teuthology.orchestra.run.vm08.stdout:VERSION="9" 2026-03-06T13:36:08.357 INFO:teuthology.orchestra.run.vm08.stdout:ID="centos" 2026-03-06T13:36:08.357 INFO:teuthology.orchestra.run.vm08.stdout:ID_LIKE="rhel fedora" 2026-03-06T13:36:08.357 INFO:teuthology.orchestra.run.vm08.stdout:VERSION_ID="9" 2026-03-06T13:36:08.357 INFO:teuthology.orchestra.run.vm08.stdout:PLATFORM_ID="platform:el9" 2026-03-06T13:36:08.357 INFO:teuthology.orchestra.run.vm08.stdout:PRETTY_NAME="CentOS Stream 9" 2026-03-06T13:36:08.357 INFO:teuthology.orchestra.run.vm08.stdout:ANSI_COLOR="0;31" 2026-03-06T13:36:08.357 INFO:teuthology.orchestra.run.vm08.stdout:LOGO="fedora-logo-icon" 2026-03-06T13:36:08.357 INFO:teuthology.orchestra.run.vm08.stdout:CPE_NAME="cpe:/o:centos:centos:9" 2026-03-06T13:36:08.357 INFO:teuthology.orchestra.run.vm08.stdout:HOME_URL="https://centos.org/" 2026-03-06T13:36:08.357 INFO:teuthology.orchestra.run.vm08.stdout:BUG_REPORT_URL="https://issues.redhat.com/" 2026-03-06T13:36:08.357 INFO:teuthology.orchestra.run.vm08.stdout:REDHAT_SUPPORT_PRODUCT="Red Hat Enterprise Linux 9" 2026-03-06T13:36:08.357 INFO:teuthology.orchestra.run.vm08.stdout:REDHAT_SUPPORT_PRODUCT_VERSION="CentOS Stream" 2026-03-06T13:36:08.357 INFO:teuthology.lock.ops:Updating vm08.local on lock server 2026-03-06T13:36:08.362 INFO:teuthology.run_tasks:Running task internal.serialize_remote_roles... 2026-03-06T13:36:08.363 INFO:teuthology.run_tasks:Running task internal.check_conflict... 2026-03-06T13:36:08.364 INFO:teuthology.task.internal:Checking for old test directory... 2026-03-06T13:36:08.364 DEBUG:teuthology.orchestra.run.vm00:> test '!' -e /home/ubuntu/cephtest 2026-03-06T13:36:08.366 DEBUG:teuthology.orchestra.run.vm01:> test '!' -e /home/ubuntu/cephtest 2026-03-06T13:36:08.367 DEBUG:teuthology.orchestra.run.vm08:> test '!' -e /home/ubuntu/cephtest 2026-03-06T13:36:08.411 INFO:teuthology.run_tasks:Running task internal.check_ceph_data... 2026-03-06T13:36:08.412 INFO:teuthology.task.internal:Checking for non-empty /var/lib/ceph... 2026-03-06T13:36:08.412 DEBUG:teuthology.orchestra.run.vm00:> test -z $(ls -A /var/lib/ceph) 2026-03-06T13:36:08.421 DEBUG:teuthology.orchestra.run.vm01:> test -z $(ls -A /var/lib/ceph) 2026-03-06T13:36:08.423 DEBUG:teuthology.orchestra.run.vm08:> test -z $(ls -A /var/lib/ceph) 2026-03-06T13:36:08.435 INFO:teuthology.orchestra.run.vm00.stderr:ls: cannot access '/var/lib/ceph': No such file or directory 2026-03-06T13:36:08.438 INFO:teuthology.orchestra.run.vm01.stderr:ls: cannot access '/var/lib/ceph': No such file or directory 2026-03-06T13:36:08.469 INFO:teuthology.orchestra.run.vm08.stderr:ls: cannot access '/var/lib/ceph': No such file or directory 2026-03-06T13:36:08.469 INFO:teuthology.run_tasks:Running task internal.vm_setup... 2026-03-06T13:36:08.477 DEBUG:teuthology.orchestra.run.vm00:> test -e /ceph-qa-ready 2026-03-06T13:36:08.493 DEBUG:teuthology.orchestra.run:got remote process result: 1 2026-03-06T13:36:08.693 DEBUG:teuthology.orchestra.run.vm01:> test -e /ceph-qa-ready 2026-03-06T13:36:08.708 DEBUG:teuthology.orchestra.run:got remote process result: 1 2026-03-06T13:36:08.897 DEBUG:teuthology.orchestra.run.vm08:> test -e /ceph-qa-ready 2026-03-06T13:36:08.911 DEBUG:teuthology.orchestra.run:got remote process result: 1 2026-03-06T13:36:09.092 INFO:teuthology.run_tasks:Running task internal.base... 2026-03-06T13:36:09.093 INFO:teuthology.task.internal:Creating test directory... 2026-03-06T13:36:09.093 DEBUG:teuthology.orchestra.run.vm00:> mkdir -p -m0755 -- /home/ubuntu/cephtest 2026-03-06T13:36:09.095 DEBUG:teuthology.orchestra.run.vm01:> mkdir -p -m0755 -- /home/ubuntu/cephtest 2026-03-06T13:36:09.097 DEBUG:teuthology.orchestra.run.vm08:> mkdir -p -m0755 -- /home/ubuntu/cephtest 2026-03-06T13:36:09.113 INFO:teuthology.run_tasks:Running task internal.archive_upload... 2026-03-06T13:36:09.115 INFO:teuthology.run_tasks:Running task internal.archive... 2026-03-06T13:36:09.116 INFO:teuthology.task.internal:Creating archive directory... 2026-03-06T13:36:09.116 DEBUG:teuthology.orchestra.run.vm00:> install -d -m0755 -- /home/ubuntu/cephtest/archive 2026-03-06T13:36:09.152 DEBUG:teuthology.orchestra.run.vm01:> install -d -m0755 -- /home/ubuntu/cephtest/archive 2026-03-06T13:36:09.154 DEBUG:teuthology.orchestra.run.vm08:> install -d -m0755 -- /home/ubuntu/cephtest/archive 2026-03-06T13:36:09.172 INFO:teuthology.run_tasks:Running task internal.coredump... 2026-03-06T13:36:09.173 INFO:teuthology.task.internal:Enabling coredump saving... 2026-03-06T13:36:09.173 DEBUG:teuthology.orchestra.run.vm00:> test -f /run/.containerenv -o -f /.dockerenv 2026-03-06T13:36:09.222 DEBUG:teuthology.orchestra.run:got remote process result: 1 2026-03-06T13:36:09.222 DEBUG:teuthology.orchestra.run.vm01:> test -f /run/.containerenv -o -f /.dockerenv 2026-03-06T13:36:09.238 DEBUG:teuthology.orchestra.run:got remote process result: 1 2026-03-06T13:36:09.239 DEBUG:teuthology.orchestra.run.vm08:> test -f /run/.containerenv -o -f /.dockerenv 2026-03-06T13:36:09.257 DEBUG:teuthology.orchestra.run:got remote process result: 1 2026-03-06T13:36:09.257 DEBUG:teuthology.orchestra.run.vm00:> install -d -m0755 -- /home/ubuntu/cephtest/archive/coredump && sudo sysctl -w kernel.core_pattern=/home/ubuntu/cephtest/archive/coredump/%t.%p.core && echo kernel.core_pattern=/home/ubuntu/cephtest/archive/coredump/%t.%p.core | sudo tee -a /etc/sysctl.conf 2026-03-06T13:36:09.264 DEBUG:teuthology.orchestra.run.vm01:> install -d -m0755 -- /home/ubuntu/cephtest/archive/coredump && sudo sysctl -w kernel.core_pattern=/home/ubuntu/cephtest/archive/coredump/%t.%p.core && echo kernel.core_pattern=/home/ubuntu/cephtest/archive/coredump/%t.%p.core | sudo tee -a /etc/sysctl.conf 2026-03-06T13:36:09.281 DEBUG:teuthology.orchestra.run.vm08:> install -d -m0755 -- /home/ubuntu/cephtest/archive/coredump && sudo sysctl -w kernel.core_pattern=/home/ubuntu/cephtest/archive/coredump/%t.%p.core && echo kernel.core_pattern=/home/ubuntu/cephtest/archive/coredump/%t.%p.core | sudo tee -a /etc/sysctl.conf 2026-03-06T13:36:09.289 INFO:teuthology.orchestra.run.vm00.stdout:kernel.core_pattern = /home/ubuntu/cephtest/archive/coredump/%t.%p.core 2026-03-06T13:36:09.302 INFO:teuthology.orchestra.run.vm00.stdout:kernel.core_pattern=/home/ubuntu/cephtest/archive/coredump/%t.%p.core 2026-03-06T13:36:09.311 INFO:teuthology.orchestra.run.vm01.stdout:kernel.core_pattern = /home/ubuntu/cephtest/archive/coredump/%t.%p.core 2026-03-06T13:36:09.323 INFO:teuthology.orchestra.run.vm01.stdout:kernel.core_pattern=/home/ubuntu/cephtest/archive/coredump/%t.%p.core 2026-03-06T13:36:09.327 INFO:teuthology.orchestra.run.vm08.stdout:kernel.core_pattern = /home/ubuntu/cephtest/archive/coredump/%t.%p.core 2026-03-06T13:36:09.337 INFO:teuthology.orchestra.run.vm08.stdout:kernel.core_pattern=/home/ubuntu/cephtest/archive/coredump/%t.%p.core 2026-03-06T13:36:09.338 INFO:teuthology.run_tasks:Running task internal.sudo... 2026-03-06T13:36:09.339 INFO:teuthology.task.internal:Configuring sudo... 2026-03-06T13:36:09.340 DEBUG:teuthology.orchestra.run.vm00:> sudo sed -i.orig.teuthology -e 's/^\([^#]*\) \(requiretty\)/\1 !\2/g' -e 's/^\([^#]*\) !\(visiblepw\)/\1 \2/g' /etc/sudoers 2026-03-06T13:36:09.347 DEBUG:teuthology.orchestra.run.vm01:> sudo sed -i.orig.teuthology -e 's/^\([^#]*\) \(requiretty\)/\1 !\2/g' -e 's/^\([^#]*\) !\(visiblepw\)/\1 \2/g' /etc/sudoers 2026-03-06T13:36:09.367 DEBUG:teuthology.orchestra.run.vm08:> sudo sed -i.orig.teuthology -e 's/^\([^#]*\) \(requiretty\)/\1 !\2/g' -e 's/^\([^#]*\) !\(visiblepw\)/\1 \2/g' /etc/sudoers 2026-03-06T13:36:09.403 INFO:teuthology.run_tasks:Running task internal.syslog... 2026-03-06T13:36:09.405 INFO:teuthology.task.internal.syslog:Starting syslog monitoring... 2026-03-06T13:36:09.406 DEBUG:teuthology.orchestra.run.vm00:> mkdir -p -m0755 -- /home/ubuntu/cephtest/archive/syslog 2026-03-06T13:36:09.413 DEBUG:teuthology.orchestra.run.vm01:> mkdir -p -m0755 -- /home/ubuntu/cephtest/archive/syslog 2026-03-06T13:36:09.433 DEBUG:teuthology.orchestra.run.vm08:> mkdir -p -m0755 -- /home/ubuntu/cephtest/archive/syslog 2026-03-06T13:36:09.459 DEBUG:teuthology.orchestra.run.vm00:> install -m 666 /dev/null /home/ubuntu/cephtest/archive/syslog/kern.log 2026-03-06T13:36:09.493 DEBUG:teuthology.orchestra.run.vm00:> install -m 666 /dev/null /home/ubuntu/cephtest/archive/syslog/misc.log 2026-03-06T13:36:09.549 DEBUG:teuthology.orchestra.run.vm00:> set -ex 2026-03-06T13:36:09.549 DEBUG:teuthology.orchestra.run.vm00:> sudo dd of=/etc/rsyslog.d/80-cephtest.conf 2026-03-06T13:36:09.608 DEBUG:teuthology.orchestra.run.vm01:> install -m 666 /dev/null /home/ubuntu/cephtest/archive/syslog/kern.log 2026-03-06T13:36:09.630 DEBUG:teuthology.orchestra.run.vm01:> install -m 666 /dev/null /home/ubuntu/cephtest/archive/syslog/misc.log 2026-03-06T13:36:09.687 DEBUG:teuthology.orchestra.run.vm01:> set -ex 2026-03-06T13:36:09.687 DEBUG:teuthology.orchestra.run.vm01:> sudo dd of=/etc/rsyslog.d/80-cephtest.conf 2026-03-06T13:36:09.744 DEBUG:teuthology.orchestra.run.vm08:> install -m 666 /dev/null /home/ubuntu/cephtest/archive/syslog/kern.log 2026-03-06T13:36:09.765 DEBUG:teuthology.orchestra.run.vm08:> install -m 666 /dev/null /home/ubuntu/cephtest/archive/syslog/misc.log 2026-03-06T13:36:09.822 DEBUG:teuthology.orchestra.run.vm08:> set -ex 2026-03-06T13:36:09.822 DEBUG:teuthology.orchestra.run.vm08:> sudo dd of=/etc/rsyslog.d/80-cephtest.conf 2026-03-06T13:36:09.880 DEBUG:teuthology.orchestra.run.vm00:> sudo service rsyslog restart 2026-03-06T13:36:09.881 DEBUG:teuthology.orchestra.run.vm01:> sudo service rsyslog restart 2026-03-06T13:36:09.883 DEBUG:teuthology.orchestra.run.vm08:> sudo service rsyslog restart 2026-03-06T13:36:09.908 INFO:teuthology.orchestra.run.vm00.stderr:Redirecting to /bin/systemctl restart rsyslog.service 2026-03-06T13:36:09.908 INFO:teuthology.orchestra.run.vm01.stderr:Redirecting to /bin/systemctl restart rsyslog.service 2026-03-06T13:36:09.948 INFO:teuthology.orchestra.run.vm08.stderr:Redirecting to /bin/systemctl restart rsyslog.service 2026-03-06T13:36:10.322 INFO:teuthology.run_tasks:Running task internal.timer... 2026-03-06T13:36:10.323 INFO:teuthology.task.internal:Starting timer... 2026-03-06T13:36:10.323 INFO:teuthology.run_tasks:Running task pcp... 2026-03-06T13:36:10.326 INFO:teuthology.run_tasks:Running task selinux... 2026-03-06T13:36:10.328 DEBUG:teuthology.task:Applying overrides for task selinux: {'allowlist': ['scontext=system_u:system_r:logrotate_t:s0', 'scontext=system_u:system_r:getty_t:s0']} 2026-03-06T13:36:10.328 INFO:teuthology.task.selinux:Excluding vm00: VMs are not yet supported 2026-03-06T13:36:10.328 INFO:teuthology.task.selinux:Excluding vm01: VMs are not yet supported 2026-03-06T13:36:10.328 INFO:teuthology.task.selinux:Excluding vm08: VMs are not yet supported 2026-03-06T13:36:10.328 DEBUG:teuthology.task.selinux:Getting current SELinux state 2026-03-06T13:36:10.328 DEBUG:teuthology.task.selinux:Existing SELinux modes: {} 2026-03-06T13:36:10.328 INFO:teuthology.task.selinux:Putting SELinux into permissive mode 2026-03-06T13:36:10.328 INFO:teuthology.run_tasks:Running task ansible.cephlab... 2026-03-06T13:36:10.329 DEBUG:teuthology.task:Applying overrides for task ansible.cephlab: {'branch': 'main', 'repo': 'https://github.com/kshtsk/ceph-cm-ansible.git', 'skip_tags': 'nagios,monitoring-scripts,hostname,pubkeys,zap,sudoers,kerberos,ntp-client,resolvconf,cpan,nfs', 'vars': {'timezone': 'Europe/Berlin'}} 2026-03-06T13:36:10.329 DEBUG:teuthology.repo_utils:Setting repo remote to https://github.com/kshtsk/ceph-cm-ansible.git 2026-03-06T13:36:10.331 INFO:teuthology.repo_utils:Fetching github.com_kshtsk_ceph-cm-ansible_main from origin 2026-03-06T13:36:10.919 DEBUG:teuthology.repo_utils:Resetting repo at /home/teuthos/src/github.com_kshtsk_ceph-cm-ansible_main to origin/main 2026-03-06T13:36:10.924 INFO:teuthology.task.ansible:Playbook: [{'import_playbook': 'ansible_managed.yml'}, {'import_playbook': 'teuthology.yml'}, {'hosts': 'testnodes', 'tasks': [{'set_fact': {'ran_from_cephlab_playbook': True}}]}, {'import_playbook': 'testnodes.yml'}, {'import_playbook': 'container-host.yml'}, {'import_playbook': 'cobbler.yml'}, {'import_playbook': 'paddles.yml'}, {'import_playbook': 'pulpito.yml'}, {'hosts': 'testnodes', 'become': True, 'tasks': [{'name': 'Touch /ceph-qa-ready', 'file': {'path': '/ceph-qa-ready', 'state': 'touch'}, 'when': 'ran_from_cephlab_playbook|bool'}]}] 2026-03-06T13:36:10.925 DEBUG:teuthology.task.ansible:Running ansible-playbook -v --extra-vars '{"ansible_ssh_user": "ubuntu", "timezone": "Europe/Berlin"}' -i /tmp/teuth_ansible_inventorypas49nnf --limit vm00.local,vm01.local,vm08.local /home/teuthos/src/github.com_kshtsk_ceph-cm-ansible_main/cephlab.yml --skip-tags nagios,monitoring-scripts,hostname,pubkeys,zap,sudoers,kerberos,ntp-client,resolvconf,cpan,nfs 2026-03-06T13:38:25.170 DEBUG:teuthology.task.ansible:Reconnecting to [Remote(name='ubuntu@vm00.local'), Remote(name='ubuntu@vm01.local'), Remote(name='ubuntu@vm08.local')] 2026-03-06T13:38:25.171 INFO:teuthology.orchestra.remote:Trying to reconnect to host 'ubuntu@vm00.local' 2026-03-06T13:38:25.171 DEBUG:teuthology.orchestra.connection:{'hostname': 'vm00.local', 'username': 'ubuntu', 'timeout': 60} 2026-03-06T13:38:25.244 DEBUG:teuthology.orchestra.run.vm00:> true 2026-03-06T13:38:25.326 INFO:teuthology.orchestra.remote:Successfully reconnected to host 'ubuntu@vm00.local' 2026-03-06T13:38:25.326 INFO:teuthology.orchestra.remote:Trying to reconnect to host 'ubuntu@vm01.local' 2026-03-06T13:38:25.326 DEBUG:teuthology.orchestra.connection:{'hostname': 'vm01.local', 'username': 'ubuntu', 'timeout': 60} 2026-03-06T13:38:25.387 DEBUG:teuthology.orchestra.run.vm01:> true 2026-03-06T13:38:25.471 INFO:teuthology.orchestra.remote:Successfully reconnected to host 'ubuntu@vm01.local' 2026-03-06T13:38:25.471 INFO:teuthology.orchestra.remote:Trying to reconnect to host 'ubuntu@vm08.local' 2026-03-06T13:38:25.472 DEBUG:teuthology.orchestra.connection:{'hostname': 'vm08.local', 'username': 'ubuntu', 'timeout': 60} 2026-03-06T13:38:25.537 DEBUG:teuthology.orchestra.run.vm08:> true 2026-03-06T13:38:25.620 INFO:teuthology.orchestra.remote:Successfully reconnected to host 'ubuntu@vm08.local' 2026-03-06T13:38:25.620 INFO:teuthology.run_tasks:Running task clock... 2026-03-06T13:38:25.622 INFO:teuthology.task.clock:Syncing clocks and checking initial clock skew... 2026-03-06T13:38:25.622 INFO:teuthology.orchestra.run:Running command with timeout 360 2026-03-06T13:38:25.622 DEBUG:teuthology.orchestra.run.vm00:> sudo systemctl stop ntp.service || sudo systemctl stop ntpd.service || sudo systemctl stop chronyd.service ; sudo ntpd -gq || sudo chronyc makestep ; sudo systemctl start ntp.service || sudo systemctl start ntpd.service || sudo systemctl start chronyd.service ; PATH=/usr/bin:/usr/sbin ntpq -p || PATH=/usr/bin:/usr/sbin chronyc sources || true 2026-03-06T13:38:25.625 INFO:teuthology.orchestra.run:Running command with timeout 360 2026-03-06T13:38:25.625 DEBUG:teuthology.orchestra.run.vm01:> sudo systemctl stop ntp.service || sudo systemctl stop ntpd.service || sudo systemctl stop chronyd.service ; sudo ntpd -gq || sudo chronyc makestep ; sudo systemctl start ntp.service || sudo systemctl start ntpd.service || sudo systemctl start chronyd.service ; PATH=/usr/bin:/usr/sbin ntpq -p || PATH=/usr/bin:/usr/sbin chronyc sources || true 2026-03-06T13:38:25.628 INFO:teuthology.orchestra.run:Running command with timeout 360 2026-03-06T13:38:25.628 DEBUG:teuthology.orchestra.run.vm08:> sudo systemctl stop ntp.service || sudo systemctl stop ntpd.service || sudo systemctl stop chronyd.service ; sudo ntpd -gq || sudo chronyc makestep ; sudo systemctl start ntp.service || sudo systemctl start ntpd.service || sudo systemctl start chronyd.service ; PATH=/usr/bin:/usr/sbin ntpq -p || PATH=/usr/bin:/usr/sbin chronyc sources || true 2026-03-06T13:38:25.655 INFO:teuthology.orchestra.run.vm00.stderr:Failed to stop ntp.service: Unit ntp.service not loaded. 2026-03-06T13:38:25.660 INFO:teuthology.orchestra.run.vm01.stderr:Failed to stop ntp.service: Unit ntp.service not loaded. 2026-03-06T13:38:25.671 INFO:teuthology.orchestra.run.vm00.stderr:Failed to stop ntpd.service: Unit ntpd.service not loaded. 2026-03-06T13:38:25.677 INFO:teuthology.orchestra.run.vm01.stderr:Failed to stop ntpd.service: Unit ntpd.service not loaded. 2026-03-06T13:38:25.698 INFO:teuthology.orchestra.run.vm00.stderr:sudo: ntpd: command not found 2026-03-06T13:38:25.700 INFO:teuthology.orchestra.run.vm08.stderr:Failed to stop ntp.service: Unit ntp.service not loaded. 2026-03-06T13:38:25.706 INFO:teuthology.orchestra.run.vm01.stderr:sudo: ntpd: command not found 2026-03-06T13:38:25.710 INFO:teuthology.orchestra.run.vm00.stdout:506 Cannot talk to daemon 2026-03-06T13:38:25.716 INFO:teuthology.orchestra.run.vm08.stderr:Failed to stop ntpd.service: Unit ntpd.service not loaded. 2026-03-06T13:38:25.721 INFO:teuthology.orchestra.run.vm01.stdout:506 Cannot talk to daemon 2026-03-06T13:38:25.725 INFO:teuthology.orchestra.run.vm00.stderr:Failed to start ntp.service: Unit ntp.service not found. 2026-03-06T13:38:25.739 INFO:teuthology.orchestra.run.vm00.stderr:Failed to start ntpd.service: Unit ntpd.service not found. 2026-03-06T13:38:25.740 INFO:teuthology.orchestra.run.vm01.stderr:Failed to start ntp.service: Unit ntp.service not found. 2026-03-06T13:38:25.743 INFO:teuthology.orchestra.run.vm08.stderr:sudo: ntpd: command not found 2026-03-06T13:38:25.756 INFO:teuthology.orchestra.run.vm08.stdout:506 Cannot talk to daemon 2026-03-06T13:38:25.758 INFO:teuthology.orchestra.run.vm01.stderr:Failed to start ntpd.service: Unit ntpd.service not found. 2026-03-06T13:38:25.771 INFO:teuthology.orchestra.run.vm08.stderr:Failed to start ntp.service: Unit ntp.service not found. 2026-03-06T13:38:25.788 INFO:teuthology.orchestra.run.vm08.stderr:Failed to start ntpd.service: Unit ntpd.service not found. 2026-03-06T13:38:25.795 INFO:teuthology.orchestra.run.vm00.stderr:bash: line 1: ntpq: command not found 2026-03-06T13:38:25.797 INFO:teuthology.orchestra.run.vm00.stdout:MS Name/IP address Stratum Poll Reach LastRx Last sample 2026-03-06T13:38:25.797 INFO:teuthology.orchestra.run.vm00.stdout:=============================================================================== 2026-03-06T13:38:25.816 INFO:teuthology.orchestra.run.vm01.stderr:bash: line 1: ntpq: command not found 2026-03-06T13:38:25.819 INFO:teuthology.orchestra.run.vm01.stdout:MS Name/IP address Stratum Poll Reach LastRx Last sample 2026-03-06T13:38:25.819 INFO:teuthology.orchestra.run.vm01.stdout:=============================================================================== 2026-03-06T13:38:25.837 INFO:teuthology.orchestra.run.vm08.stderr:bash: line 1: ntpq: command not found 2026-03-06T13:38:25.844 INFO:teuthology.orchestra.run.vm08.stdout:MS Name/IP address Stratum Poll Reach LastRx Last sample 2026-03-06T13:38:25.845 INFO:teuthology.orchestra.run.vm08.stdout:=============================================================================== 2026-03-06T13:38:25.845 INFO:teuthology.run_tasks:Running task pexec... 2026-03-06T13:38:25.847 INFO:teuthology.task.pexec:Executing custom commands... 2026-03-06T13:38:25.848 DEBUG:teuthology.orchestra.run.vm00:> TESTDIR=/home/ubuntu/cephtest bash -s 2026-03-06T13:38:25.848 DEBUG:teuthology.orchestra.run.vm01:> TESTDIR=/home/ubuntu/cephtest bash -s 2026-03-06T13:38:25.848 DEBUG:teuthology.orchestra.run.vm08:> TESTDIR=/home/ubuntu/cephtest bash -s 2026-03-06T13:38:25.850 DEBUG:teuthology.task.pexec:ubuntu@vm00.local< sudo dnf remove nvme-cli -y 2026-03-06T13:38:25.850 DEBUG:teuthology.task.pexec:ubuntu@vm00.local< sudo dnf install nvmetcli nvme-cli -y 2026-03-06T13:38:25.850 INFO:teuthology.task.pexec:Running commands on host ubuntu@vm00.local 2026-03-06T13:38:25.850 INFO:teuthology.task.pexec:sudo dnf remove nvme-cli -y 2026-03-06T13:38:25.850 INFO:teuthology.task.pexec:sudo dnf install nvmetcli nvme-cli -y 2026-03-06T13:38:25.861 DEBUG:teuthology.task.pexec:ubuntu@vm01.local< sudo dnf remove nvme-cli -y 2026-03-06T13:38:25.861 DEBUG:teuthology.task.pexec:ubuntu@vm01.local< sudo dnf install nvmetcli nvme-cli -y 2026-03-06T13:38:25.861 INFO:teuthology.task.pexec:Running commands on host ubuntu@vm01.local 2026-03-06T13:38:25.861 INFO:teuthology.task.pexec:sudo dnf remove nvme-cli -y 2026-03-06T13:38:25.861 INFO:teuthology.task.pexec:sudo dnf install nvmetcli nvme-cli -y 2026-03-06T13:38:25.889 DEBUG:teuthology.task.pexec:ubuntu@vm08.local< sudo dnf remove nvme-cli -y 2026-03-06T13:38:25.889 DEBUG:teuthology.task.pexec:ubuntu@vm08.local< sudo dnf install nvmetcli nvme-cli -y 2026-03-06T13:38:25.889 INFO:teuthology.task.pexec:Running commands on host ubuntu@vm08.local 2026-03-06T13:38:25.889 INFO:teuthology.task.pexec:sudo dnf remove nvme-cli -y 2026-03-06T13:38:25.889 INFO:teuthology.task.pexec:sudo dnf install nvmetcli nvme-cli -y 2026-03-06T13:38:26.062 INFO:teuthology.orchestra.run.vm00.stdout:No match for argument: nvme-cli 2026-03-06T13:38:26.062 INFO:teuthology.orchestra.run.vm00.stderr:No packages marked for removal. 2026-03-06T13:38:26.065 INFO:teuthology.orchestra.run.vm00.stdout:Dependencies resolved. 2026-03-06T13:38:26.066 INFO:teuthology.orchestra.run.vm01.stdout:No match for argument: nvme-cli 2026-03-06T13:38:26.066 INFO:teuthology.orchestra.run.vm01.stderr:No packages marked for removal. 2026-03-06T13:38:26.066 INFO:teuthology.orchestra.run.vm00.stdout:Nothing to do. 2026-03-06T13:38:26.066 INFO:teuthology.orchestra.run.vm00.stdout:Complete! 2026-03-06T13:38:26.068 INFO:teuthology.orchestra.run.vm08.stdout:No match for argument: nvme-cli 2026-03-06T13:38:26.068 INFO:teuthology.orchestra.run.vm08.stderr:No packages marked for removal. 2026-03-06T13:38:26.068 INFO:teuthology.orchestra.run.vm01.stdout:Dependencies resolved. 2026-03-06T13:38:26.069 INFO:teuthology.orchestra.run.vm01.stdout:Nothing to do. 2026-03-06T13:38:26.069 INFO:teuthology.orchestra.run.vm01.stdout:Complete! 2026-03-06T13:38:26.071 INFO:teuthology.orchestra.run.vm08.stdout:Dependencies resolved. 2026-03-06T13:38:26.071 INFO:teuthology.orchestra.run.vm08.stdout:Nothing to do. 2026-03-06T13:38:26.071 INFO:teuthology.orchestra.run.vm08.stdout:Complete! 2026-03-06T13:38:26.439 INFO:teuthology.orchestra.run.vm08.stdout:Last metadata expiration check: 0:01:29 ago on Fri 06 Mar 2026 01:36:57 PM CET. 2026-03-06T13:38:26.476 INFO:teuthology.orchestra.run.vm00.stdout:Last metadata expiration check: 0:01:27 ago on Fri 06 Mar 2026 01:36:59 PM CET. 2026-03-06T13:38:26.531 INFO:teuthology.orchestra.run.vm08.stdout:Dependencies resolved. 2026-03-06T13:38:26.532 INFO:teuthology.orchestra.run.vm08.stdout:================================================================================ 2026-03-06T13:38:26.532 INFO:teuthology.orchestra.run.vm08.stdout: Package Architecture Version Repository Size 2026-03-06T13:38:26.532 INFO:teuthology.orchestra.run.vm08.stdout:================================================================================ 2026-03-06T13:38:26.532 INFO:teuthology.orchestra.run.vm08.stdout:Installing: 2026-03-06T13:38:26.532 INFO:teuthology.orchestra.run.vm08.stdout: nvme-cli x86_64 2.16-1.el9 baseos 1.2 M 2026-03-06T13:38:26.532 INFO:teuthology.orchestra.run.vm08.stdout: nvmetcli noarch 0.8-3.el9 baseos 44 k 2026-03-06T13:38:26.532 INFO:teuthology.orchestra.run.vm08.stdout:Installing dependencies: 2026-03-06T13:38:26.532 INFO:teuthology.orchestra.run.vm08.stdout: python3-configshell noarch 1:1.1.30-1.el9 baseos 72 k 2026-03-06T13:38:26.532 INFO:teuthology.orchestra.run.vm08.stdout: python3-kmod x86_64 0.9-32.el9 baseos 84 k 2026-03-06T13:38:26.532 INFO:teuthology.orchestra.run.vm08.stdout: python3-pyparsing noarch 2.4.7-9.el9 baseos 150 k 2026-03-06T13:38:26.532 INFO:teuthology.orchestra.run.vm08.stdout: python3-urwid x86_64 2.1.2-4.el9 baseos 837 k 2026-03-06T13:38:26.532 INFO:teuthology.orchestra.run.vm08.stdout: 2026-03-06T13:38:26.532 INFO:teuthology.orchestra.run.vm08.stdout:Transaction Summary 2026-03-06T13:38:26.532 INFO:teuthology.orchestra.run.vm08.stdout:================================================================================ 2026-03-06T13:38:26.532 INFO:teuthology.orchestra.run.vm08.stdout:Install 6 Packages 2026-03-06T13:38:26.532 INFO:teuthology.orchestra.run.vm08.stdout: 2026-03-06T13:38:26.532 INFO:teuthology.orchestra.run.vm08.stdout:Total download size: 2.3 M 2026-03-06T13:38:26.532 INFO:teuthology.orchestra.run.vm08.stdout:Installed size: 11 M 2026-03-06T13:38:26.532 INFO:teuthology.orchestra.run.vm08.stdout:Downloading Packages: 2026-03-06T13:38:26.566 INFO:teuthology.orchestra.run.vm01.stdout:Last metadata expiration check: 0:01:00 ago on Fri 06 Mar 2026 01:37:26 PM CET. 2026-03-06T13:38:26.581 INFO:teuthology.orchestra.run.vm00.stdout:Dependencies resolved. 2026-03-06T13:38:26.582 INFO:teuthology.orchestra.run.vm00.stdout:================================================================================ 2026-03-06T13:38:26.582 INFO:teuthology.orchestra.run.vm00.stdout: Package Architecture Version Repository Size 2026-03-06T13:38:26.582 INFO:teuthology.orchestra.run.vm00.stdout:================================================================================ 2026-03-06T13:38:26.582 INFO:teuthology.orchestra.run.vm00.stdout:Installing: 2026-03-06T13:38:26.582 INFO:teuthology.orchestra.run.vm00.stdout: nvme-cli x86_64 2.16-1.el9 baseos 1.2 M 2026-03-06T13:38:26.582 INFO:teuthology.orchestra.run.vm00.stdout: nvmetcli noarch 0.8-3.el9 baseos 44 k 2026-03-06T13:38:26.582 INFO:teuthology.orchestra.run.vm00.stdout:Installing dependencies: 2026-03-06T13:38:26.582 INFO:teuthology.orchestra.run.vm00.stdout: python3-configshell noarch 1:1.1.30-1.el9 baseos 72 k 2026-03-06T13:38:26.582 INFO:teuthology.orchestra.run.vm00.stdout: python3-kmod x86_64 0.9-32.el9 baseos 84 k 2026-03-06T13:38:26.582 INFO:teuthology.orchestra.run.vm00.stdout: python3-pyparsing noarch 2.4.7-9.el9 baseos 150 k 2026-03-06T13:38:26.582 INFO:teuthology.orchestra.run.vm00.stdout: python3-urwid x86_64 2.1.2-4.el9 baseos 837 k 2026-03-06T13:38:26.582 INFO:teuthology.orchestra.run.vm00.stdout: 2026-03-06T13:38:26.582 INFO:teuthology.orchestra.run.vm00.stdout:Transaction Summary 2026-03-06T13:38:26.582 INFO:teuthology.orchestra.run.vm00.stdout:================================================================================ 2026-03-06T13:38:26.582 INFO:teuthology.orchestra.run.vm00.stdout:Install 6 Packages 2026-03-06T13:38:26.582 INFO:teuthology.orchestra.run.vm00.stdout: 2026-03-06T13:38:26.582 INFO:teuthology.orchestra.run.vm00.stdout:Total download size: 2.3 M 2026-03-06T13:38:26.582 INFO:teuthology.orchestra.run.vm00.stdout:Installed size: 11 M 2026-03-06T13:38:26.582 INFO:teuthology.orchestra.run.vm00.stdout:Downloading Packages: 2026-03-06T13:38:26.700 INFO:teuthology.orchestra.run.vm01.stdout:Dependencies resolved. 2026-03-06T13:38:26.701 INFO:teuthology.orchestra.run.vm01.stdout:================================================================================ 2026-03-06T13:38:26.701 INFO:teuthology.orchestra.run.vm01.stdout: Package Architecture Version Repository Size 2026-03-06T13:38:26.701 INFO:teuthology.orchestra.run.vm01.stdout:================================================================================ 2026-03-06T13:38:26.701 INFO:teuthology.orchestra.run.vm01.stdout:Installing: 2026-03-06T13:38:26.701 INFO:teuthology.orchestra.run.vm01.stdout: nvme-cli x86_64 2.16-1.el9 baseos 1.2 M 2026-03-06T13:38:26.701 INFO:teuthology.orchestra.run.vm01.stdout: nvmetcli noarch 0.8-3.el9 baseos 44 k 2026-03-06T13:38:26.701 INFO:teuthology.orchestra.run.vm01.stdout:Installing dependencies: 2026-03-06T13:38:26.701 INFO:teuthology.orchestra.run.vm01.stdout: python3-configshell noarch 1:1.1.30-1.el9 baseos 72 k 2026-03-06T13:38:26.701 INFO:teuthology.orchestra.run.vm01.stdout: python3-kmod x86_64 0.9-32.el9 baseos 84 k 2026-03-06T13:38:26.701 INFO:teuthology.orchestra.run.vm01.stdout: python3-pyparsing noarch 2.4.7-9.el9 baseos 150 k 2026-03-06T13:38:26.701 INFO:teuthology.orchestra.run.vm01.stdout: python3-urwid x86_64 2.1.2-4.el9 baseos 837 k 2026-03-06T13:38:26.701 INFO:teuthology.orchestra.run.vm01.stdout: 2026-03-06T13:38:26.701 INFO:teuthology.orchestra.run.vm01.stdout:Transaction Summary 2026-03-06T13:38:26.701 INFO:teuthology.orchestra.run.vm01.stdout:================================================================================ 2026-03-06T13:38:26.701 INFO:teuthology.orchestra.run.vm01.stdout:Install 6 Packages 2026-03-06T13:38:26.701 INFO:teuthology.orchestra.run.vm01.stdout: 2026-03-06T13:38:26.701 INFO:teuthology.orchestra.run.vm01.stdout:Total download size: 2.3 M 2026-03-06T13:38:26.701 INFO:teuthology.orchestra.run.vm01.stdout:Installed size: 11 M 2026-03-06T13:38:26.701 INFO:teuthology.orchestra.run.vm01.stdout:Downloading Packages: 2026-03-06T13:38:26.891 INFO:teuthology.orchestra.run.vm08.stdout:(1/6): nvmetcli-0.8-3.el9.noarch.rpm 176 kB/s | 44 kB 00:00 2026-03-06T13:38:26.920 INFO:teuthology.orchestra.run.vm01.stdout:(1/6): nvmetcli-0.8-3.el9.noarch.rpm 389 kB/s | 44 kB 00:00 2026-03-06T13:38:26.921 INFO:teuthology.orchestra.run.vm01.stdout:(2/6): python3-configshell-1.1.30-1.el9.noarch. 634 kB/s | 72 kB 00:00 2026-03-06T13:38:26.969 INFO:teuthology.orchestra.run.vm08.stdout:(2/6): python3-configshell-1.1.30-1.el9.noarch. 219 kB/s | 72 kB 00:00 2026-03-06T13:38:26.978 INFO:teuthology.orchestra.run.vm01.stdout:(3/6): python3-kmod-0.9-32.el9.x86_64.rpm 1.4 MB/s | 84 kB 00:00 2026-03-06T13:38:26.979 INFO:teuthology.orchestra.run.vm01.stdout:(4/6): python3-pyparsing-2.4.7-9.el9.noarch.rpm 2.5 MB/s | 150 kB 00:00 2026-03-06T13:38:27.036 INFO:teuthology.orchestra.run.vm08.stdout:(3/6): python3-kmod-0.9-32.el9.x86_64.rpm 583 kB/s | 84 kB 00:00 2026-03-06T13:38:27.036 INFO:teuthology.orchestra.run.vm01.stdout:(5/6): nvme-cli-2.16-1.el9.x86_64.rpm 5.0 MB/s | 1.2 MB 00:00 2026-03-06T13:38:27.071 INFO:teuthology.orchestra.run.vm01.stdout:(6/6): python3-urwid-2.1.2-4.el9.x86_64.rpm 9.0 MB/s | 837 kB 00:00 2026-03-06T13:38:27.071 INFO:teuthology.orchestra.run.vm01.stdout:-------------------------------------------------------------------------------- 2026-03-06T13:38:27.071 INFO:teuthology.orchestra.run.vm01.stdout:Total 6.3 MB/s | 2.3 MB 00:00 2026-03-06T13:38:27.078 INFO:teuthology.orchestra.run.vm08.stdout:(4/6): python3-pyparsing-2.4.7-9.el9.noarch.rpm 1.4 MB/s | 150 kB 00:00 2026-03-06T13:38:27.143 INFO:teuthology.orchestra.run.vm01.stdout:Running transaction check 2026-03-06T13:38:27.150 INFO:teuthology.orchestra.run.vm01.stdout:Transaction check succeeded. 2026-03-06T13:38:27.152 INFO:teuthology.orchestra.run.vm01.stdout:Running transaction test 2026-03-06T13:38:27.191 INFO:teuthology.orchestra.run.vm08.stdout:(5/6): nvme-cli-2.16-1.el9.x86_64.rpm 2.1 MB/s | 1.2 MB 00:00 2026-03-06T13:38:27.198 INFO:teuthology.orchestra.run.vm08.stdout:(6/6): python3-urwid-2.1.2-4.el9.x86_64.rpm 5.1 MB/s | 837 kB 00:00 2026-03-06T13:38:27.198 INFO:teuthology.orchestra.run.vm08.stdout:-------------------------------------------------------------------------------- 2026-03-06T13:38:27.198 INFO:teuthology.orchestra.run.vm08.stdout:Total 3.5 MB/s | 2.3 MB 00:00 2026-03-06T13:38:27.210 INFO:teuthology.orchestra.run.vm01.stdout:Transaction test succeeded. 2026-03-06T13:38:27.210 INFO:teuthology.orchestra.run.vm01.stdout:Running transaction 2026-03-06T13:38:27.230 INFO:teuthology.orchestra.run.vm00.stdout:(1/6): nvmetcli-0.8-3.el9.noarch.rpm 231 kB/s | 44 kB 00:00 2026-03-06T13:38:27.232 INFO:teuthology.orchestra.run.vm00.stdout:(2/6): python3-configshell-1.1.30-1.el9.noarch. 375 kB/s | 72 kB 00:00 2026-03-06T13:38:27.260 INFO:teuthology.orchestra.run.vm08.stdout:Running transaction check 2026-03-06T13:38:27.268 INFO:teuthology.orchestra.run.vm08.stdout:Transaction check succeeded. 2026-03-06T13:38:27.268 INFO:teuthology.orchestra.run.vm08.stdout:Running transaction test 2026-03-06T13:38:27.318 INFO:teuthology.orchestra.run.vm08.stdout:Transaction test succeeded. 2026-03-06T13:38:27.318 INFO:teuthology.orchestra.run.vm08.stdout:Running transaction 2026-03-06T13:38:27.326 INFO:teuthology.orchestra.run.vm00.stdout:(3/6): python3-kmod-0.9-32.el9.x86_64.rpm 876 kB/s | 84 kB 00:00 2026-03-06T13:38:27.330 INFO:teuthology.orchestra.run.vm00.stdout:(4/6): python3-pyparsing-2.4.7-9.el9.noarch.rpm 1.5 MB/s | 150 kB 00:00 2026-03-06T13:38:27.368 INFO:teuthology.orchestra.run.vm01.stdout: Preparing : 1/1 2026-03-06T13:38:27.379 INFO:teuthology.orchestra.run.vm01.stdout: Installing : python3-urwid-2.1.2-4.el9.x86_64 1/6 2026-03-06T13:38:27.390 INFO:teuthology.orchestra.run.vm01.stdout: Installing : python3-pyparsing-2.4.7-9.el9.noarch 2/6 2026-03-06T13:38:27.397 INFO:teuthology.orchestra.run.vm01.stdout: Installing : python3-configshell-1:1.1.30-1.el9.noarch 3/6 2026-03-06T13:38:27.405 INFO:teuthology.orchestra.run.vm01.stdout: Installing : python3-kmod-0.9-32.el9.x86_64 4/6 2026-03-06T13:38:27.407 INFO:teuthology.orchestra.run.vm01.stdout: Installing : nvmetcli-0.8-3.el9.noarch 5/6 2026-03-06T13:38:27.422 INFO:teuthology.orchestra.run.vm00.stdout:(5/6): nvme-cli-2.16-1.el9.x86_64.rpm 3.0 MB/s | 1.2 MB 00:00 2026-03-06T13:38:27.464 INFO:teuthology.orchestra.run.vm08.stdout: Preparing : 1/1 2026-03-06T13:38:27.476 INFO:teuthology.orchestra.run.vm08.stdout: Installing : python3-urwid-2.1.2-4.el9.x86_64 1/6 2026-03-06T13:38:27.478 INFO:teuthology.orchestra.run.vm00.stdout:(6/6): python3-urwid-2.1.2-4.el9.x86_64.rpm 5.4 MB/s | 837 kB 00:00 2026-03-06T13:38:27.478 INFO:teuthology.orchestra.run.vm00.stdout:-------------------------------------------------------------------------------- 2026-03-06T13:38:27.479 INFO:teuthology.orchestra.run.vm00.stdout:Total 2.6 MB/s | 2.3 MB 00:00 2026-03-06T13:38:27.488 INFO:teuthology.orchestra.run.vm08.stdout: Installing : python3-pyparsing-2.4.7-9.el9.noarch 2/6 2026-03-06T13:38:27.495 INFO:teuthology.orchestra.run.vm08.stdout: Installing : python3-configshell-1:1.1.30-1.el9.noarch 3/6 2026-03-06T13:38:27.503 INFO:teuthology.orchestra.run.vm08.stdout: Installing : python3-kmod-0.9-32.el9.x86_64 4/6 2026-03-06T13:38:27.504 INFO:teuthology.orchestra.run.vm08.stdout: Installing : nvmetcli-0.8-3.el9.noarch 5/6 2026-03-06T13:38:27.539 INFO:teuthology.orchestra.run.vm00.stdout:Running transaction check 2026-03-06T13:38:27.547 INFO:teuthology.orchestra.run.vm00.stdout:Transaction check succeeded. 2026-03-06T13:38:27.547 INFO:teuthology.orchestra.run.vm00.stdout:Running transaction test 2026-03-06T13:38:27.585 INFO:teuthology.orchestra.run.vm01.stdout: Running scriptlet: nvmetcli-0.8-3.el9.noarch 5/6 2026-03-06T13:38:27.590 INFO:teuthology.orchestra.run.vm01.stdout: Installing : nvme-cli-2.16-1.el9.x86_64 6/6 2026-03-06T13:38:27.596 INFO:teuthology.orchestra.run.vm00.stdout:Transaction test succeeded. 2026-03-06T13:38:27.596 INFO:teuthology.orchestra.run.vm00.stdout:Running transaction 2026-03-06T13:38:27.656 INFO:teuthology.orchestra.run.vm08.stdout: Running scriptlet: nvmetcli-0.8-3.el9.noarch 5/6 2026-03-06T13:38:27.661 INFO:teuthology.orchestra.run.vm08.stdout: Installing : nvme-cli-2.16-1.el9.x86_64 6/6 2026-03-06T13:38:27.740 INFO:teuthology.orchestra.run.vm00.stdout: Preparing : 1/1 2026-03-06T13:38:27.750 INFO:teuthology.orchestra.run.vm00.stdout: Installing : python3-urwid-2.1.2-4.el9.x86_64 1/6 2026-03-06T13:38:27.763 INFO:teuthology.orchestra.run.vm00.stdout: Installing : python3-pyparsing-2.4.7-9.el9.noarch 2/6 2026-03-06T13:38:27.770 INFO:teuthology.orchestra.run.vm00.stdout: Installing : python3-configshell-1:1.1.30-1.el9.noarch 3/6 2026-03-06T13:38:27.778 INFO:teuthology.orchestra.run.vm00.stdout: Installing : python3-kmod-0.9-32.el9.x86_64 4/6 2026-03-06T13:38:27.780 INFO:teuthology.orchestra.run.vm00.stdout: Installing : nvmetcli-0.8-3.el9.noarch 5/6 2026-03-06T13:38:27.945 INFO:teuthology.orchestra.run.vm00.stdout: Running scriptlet: nvmetcli-0.8-3.el9.noarch 5/6 2026-03-06T13:38:27.950 INFO:teuthology.orchestra.run.vm00.stdout: Installing : nvme-cli-2.16-1.el9.x86_64 6/6 2026-03-06T13:38:28.029 INFO:teuthology.orchestra.run.vm08.stdout: Running scriptlet: nvme-cli-2.16-1.el9.x86_64 6/6 2026-03-06T13:38:28.029 INFO:teuthology.orchestra.run.vm08.stdout:Created symlink /etc/systemd/system/default.target.wants/nvmefc-boot-connections.service → /usr/lib/systemd/system/nvmefc-boot-connections.service. 2026-03-06T13:38:28.030 INFO:teuthology.orchestra.run.vm08.stdout: 2026-03-06T13:38:28.036 INFO:teuthology.orchestra.run.vm01.stdout: Running scriptlet: nvme-cli-2.16-1.el9.x86_64 6/6 2026-03-06T13:38:28.036 INFO:teuthology.orchestra.run.vm01.stdout:Created symlink /etc/systemd/system/default.target.wants/nvmefc-boot-connections.service → /usr/lib/systemd/system/nvmefc-boot-connections.service. 2026-03-06T13:38:28.036 INFO:teuthology.orchestra.run.vm01.stdout: 2026-03-06T13:38:28.301 INFO:teuthology.orchestra.run.vm00.stdout: Running scriptlet: nvme-cli-2.16-1.el9.x86_64 6/6 2026-03-06T13:38:28.302 INFO:teuthology.orchestra.run.vm00.stdout:Created symlink /etc/systemd/system/default.target.wants/nvmefc-boot-connections.service → /usr/lib/systemd/system/nvmefc-boot-connections.service. 2026-03-06T13:38:28.302 INFO:teuthology.orchestra.run.vm00.stdout: 2026-03-06T13:38:28.670 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : nvme-cli-2.16-1.el9.x86_64 1/6 2026-03-06T13:38:28.670 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : nvmetcli-0.8-3.el9.noarch 2/6 2026-03-06T13:38:28.670 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-configshell-1:1.1.30-1.el9.noarch 3/6 2026-03-06T13:38:28.670 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-kmod-0.9-32.el9.x86_64 4/6 2026-03-06T13:38:28.670 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-pyparsing-2.4.7-9.el9.noarch 5/6 2026-03-06T13:38:28.673 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : nvme-cli-2.16-1.el9.x86_64 1/6 2026-03-06T13:38:28.674 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : nvmetcli-0.8-3.el9.noarch 2/6 2026-03-06T13:38:28.674 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : python3-configshell-1:1.1.30-1.el9.noarch 3/6 2026-03-06T13:38:28.675 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : python3-kmod-0.9-32.el9.x86_64 4/6 2026-03-06T13:38:28.675 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : python3-pyparsing-2.4.7-9.el9.noarch 5/6 2026-03-06T13:38:28.761 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-urwid-2.1.2-4.el9.x86_64 6/6 2026-03-06T13:38:28.761 INFO:teuthology.orchestra.run.vm01.stdout: 2026-03-06T13:38:28.761 INFO:teuthology.orchestra.run.vm01.stdout:Installed: 2026-03-06T13:38:28.761 INFO:teuthology.orchestra.run.vm01.stdout: nvme-cli-2.16-1.el9.x86_64 nvmetcli-0.8-3.el9.noarch 2026-03-06T13:38:28.761 INFO:teuthology.orchestra.run.vm01.stdout: python3-configshell-1:1.1.30-1.el9.noarch python3-kmod-0.9-32.el9.x86_64 2026-03-06T13:38:28.761 INFO:teuthology.orchestra.run.vm01.stdout: python3-pyparsing-2.4.7-9.el9.noarch python3-urwid-2.1.2-4.el9.x86_64 2026-03-06T13:38:28.761 INFO:teuthology.orchestra.run.vm01.stdout: 2026-03-06T13:38:28.761 INFO:teuthology.orchestra.run.vm01.stdout:Complete! 2026-03-06T13:38:28.780 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : python3-urwid-2.1.2-4.el9.x86_64 6/6 2026-03-06T13:38:28.780 INFO:teuthology.orchestra.run.vm08.stdout: 2026-03-06T13:38:28.780 INFO:teuthology.orchestra.run.vm08.stdout:Installed: 2026-03-06T13:38:28.780 INFO:teuthology.orchestra.run.vm08.stdout: nvme-cli-2.16-1.el9.x86_64 nvmetcli-0.8-3.el9.noarch 2026-03-06T13:38:28.780 INFO:teuthology.orchestra.run.vm08.stdout: python3-configshell-1:1.1.30-1.el9.noarch python3-kmod-0.9-32.el9.x86_64 2026-03-06T13:38:28.780 INFO:teuthology.orchestra.run.vm08.stdout: python3-pyparsing-2.4.7-9.el9.noarch python3-urwid-2.1.2-4.el9.x86_64 2026-03-06T13:38:28.780 INFO:teuthology.orchestra.run.vm08.stdout: 2026-03-06T13:38:28.780 INFO:teuthology.orchestra.run.vm08.stdout:Complete! 2026-03-06T13:38:28.838 DEBUG:teuthology.parallel:result is None 2026-03-06T13:38:28.848 DEBUG:teuthology.parallel:result is None 2026-03-06T13:38:28.879 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : nvme-cli-2.16-1.el9.x86_64 1/6 2026-03-06T13:38:28.879 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : nvmetcli-0.8-3.el9.noarch 2/6 2026-03-06T13:38:28.879 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : python3-configshell-1:1.1.30-1.el9.noarch 3/6 2026-03-06T13:38:28.879 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : python3-kmod-0.9-32.el9.x86_64 4/6 2026-03-06T13:38:28.879 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : python3-pyparsing-2.4.7-9.el9.noarch 5/6 2026-03-06T13:38:28.966 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : python3-urwid-2.1.2-4.el9.x86_64 6/6 2026-03-06T13:38:28.966 INFO:teuthology.orchestra.run.vm00.stdout: 2026-03-06T13:38:28.966 INFO:teuthology.orchestra.run.vm00.stdout:Installed: 2026-03-06T13:38:28.966 INFO:teuthology.orchestra.run.vm00.stdout: nvme-cli-2.16-1.el9.x86_64 nvmetcli-0.8-3.el9.noarch 2026-03-06T13:38:28.966 INFO:teuthology.orchestra.run.vm00.stdout: python3-configshell-1:1.1.30-1.el9.noarch python3-kmod-0.9-32.el9.x86_64 2026-03-06T13:38:28.966 INFO:teuthology.orchestra.run.vm00.stdout: python3-pyparsing-2.4.7-9.el9.noarch python3-urwid-2.1.2-4.el9.x86_64 2026-03-06T13:38:28.966 INFO:teuthology.orchestra.run.vm00.stdout: 2026-03-06T13:38:28.966 INFO:teuthology.orchestra.run.vm00.stdout:Complete! 2026-03-06T13:38:29.031 DEBUG:teuthology.parallel:result is None 2026-03-06T13:38:29.031 INFO:teuthology.run_tasks:Running task install... 2026-03-06T13:38:29.033 DEBUG:teuthology.task.install:project ceph 2026-03-06T13:38:29.033 DEBUG:teuthology.task.install:INSTALL overrides: {'ceph': {'flavor': 'default', 'sha1': 'c24117fd5525679b799527bc1bd1f1dd0a2db5e2'}, 'extra_system_packages': {'deb': ['python3-xmltodict', 's3cmd'], 'rpm': ['bzip2', 'perl-Test-Harness', 'python3-xmltodict', 's3cmd']}, 'repos': [{'name': 'ceph-source', 'priority': 1, 'url': 'https://s3.clyso.com/ces-packages/components/ceph/rpm-19.2.3-47-gc24117fd552/el9.clyso/SRPMS'}, {'name': 'ceph-noarch', 'priority': 1, 'url': 'https://s3.clyso.com/ces-packages/components/ceph/rpm-19.2.3-47-gc24117fd552/el9.clyso/noarch'}, {'name': 'ceph', 'priority': 1, 'url': 'https://s3.clyso.com/ces-packages/components/ceph/rpm-19.2.3-47-gc24117fd552/el9.clyso/x86_64'}]} 2026-03-06T13:38:29.033 DEBUG:teuthology.task.install:config {'flavor': 'default', 'sha1': 'c24117fd5525679b799527bc1bd1f1dd0a2db5e2', 'extra_system_packages': {'deb': ['python3-xmltodict', 's3cmd'], 'rpm': ['bzip2', 'perl-Test-Harness', 'python3-xmltodict', 's3cmd']}} 2026-03-06T13:38:29.033 INFO:teuthology.task.install:Using flavor: default 2026-03-06T13:38:29.036 DEBUG:teuthology.task.install:Package list is: {'deb': ['ceph', 'cephadm', 'ceph-mds', 'ceph-mgr', 'ceph-common', 'ceph-fuse', 'ceph-test', 'ceph-volume', 'radosgw', 'python3-rados', 'python3-rgw', 'python3-cephfs', 'python3-rbd', 'libcephfs2', 'libcephfs-dev', 'librados2', 'librbd1', 'rbd-fuse'], 'rpm': ['ceph-radosgw', 'ceph-test', 'ceph', 'ceph-base', 'cephadm', 'ceph-immutable-object-cache', 'ceph-mgr', 'ceph-mgr-dashboard', 'ceph-mgr-diskprediction-local', 'ceph-mgr-rook', 'ceph-mgr-cephadm', 'ceph-fuse', 'ceph-volume', 'librados-devel', 'libcephfs2', 'libcephfs-devel', 'librados2', 'librbd1', 'python3-rados', 'python3-rgw', 'python3-cephfs', 'python3-rbd', 'rbd-fuse', 'rbd-mirror', 'rbd-nbd']} 2026-03-06T13:38:29.036 INFO:teuthology.task.install:extra packages: [] 2026-03-06T13:38:29.036 DEBUG:teuthology.task.install.rpm:_update_package_list_and_install: config is {'branch': None, 'cleanup': None, 'debuginfo': None, 'downgrade_packages': [], 'exclude_packages': [], 'extra_packages': [], 'extra_system_packages': {'deb': ['python3-xmltodict', 's3cmd'], 'rpm': ['bzip2', 'perl-Test-Harness', 'python3-xmltodict', 's3cmd']}, 'extras': None, 'enable_coprs': [], 'flavor': 'default', 'install_ceph_packages': True, 'packages': {}, 'project': 'ceph', 'repos_only': False, 'sha1': 'c24117fd5525679b799527bc1bd1f1dd0a2db5e2', 'tag': None, 'wait_for_package': False, 'repos': [{'name': 'ceph-source', 'priority': 1, 'url': 'https://s3.clyso.com/ces-packages/components/ceph/rpm-19.2.3-47-gc24117fd552/el9.clyso/SRPMS'}, {'name': 'ceph-noarch', 'priority': 1, 'url': 'https://s3.clyso.com/ces-packages/components/ceph/rpm-19.2.3-47-gc24117fd552/el9.clyso/noarch'}, {'name': 'ceph', 'priority': 1, 'url': 'https://s3.clyso.com/ces-packages/components/ceph/rpm-19.2.3-47-gc24117fd552/el9.clyso/x86_64'}]} 2026-03-06T13:38:29.036 DEBUG:teuthology.task.install.rpm:Adding repos: [{'name': 'ceph-source', 'priority': 1, 'url': 'https://s3.clyso.com/ces-packages/components/ceph/rpm-19.2.3-47-gc24117fd552/el9.clyso/SRPMS'}, {'name': 'ceph-noarch', 'priority': 1, 'url': 'https://s3.clyso.com/ces-packages/components/ceph/rpm-19.2.3-47-gc24117fd552/el9.clyso/noarch'}, {'name': 'ceph', 'priority': 1, 'url': 'https://s3.clyso.com/ces-packages/components/ceph/rpm-19.2.3-47-gc24117fd552/el9.clyso/x86_64'}] 2026-03-06T13:38:29.036 DEBUG:teuthology.orchestra.run.vm00:> set -ex 2026-03-06T13:38:29.036 DEBUG:teuthology.orchestra.run.vm00:> sudo dd of=/etc/yum.repos.d/ceph-source.repo 2026-03-06T13:38:29.036 DEBUG:teuthology.task.install.rpm:_update_package_list_and_install: config is {'branch': None, 'cleanup': None, 'debuginfo': None, 'downgrade_packages': [], 'exclude_packages': [], 'extra_packages': [], 'extra_system_packages': {'deb': ['python3-xmltodict', 's3cmd'], 'rpm': ['bzip2', 'perl-Test-Harness', 'python3-xmltodict', 's3cmd']}, 'extras': None, 'enable_coprs': [], 'flavor': 'default', 'install_ceph_packages': True, 'packages': {}, 'project': 'ceph', 'repos_only': False, 'sha1': 'c24117fd5525679b799527bc1bd1f1dd0a2db5e2', 'tag': None, 'wait_for_package': False, 'repos': [{'name': 'ceph-source', 'priority': 1, 'url': 'https://s3.clyso.com/ces-packages/components/ceph/rpm-19.2.3-47-gc24117fd552/el9.clyso/SRPMS'}, {'name': 'ceph-noarch', 'priority': 1, 'url': 'https://s3.clyso.com/ces-packages/components/ceph/rpm-19.2.3-47-gc24117fd552/el9.clyso/noarch'}, {'name': 'ceph', 'priority': 1, 'url': 'https://s3.clyso.com/ces-packages/components/ceph/rpm-19.2.3-47-gc24117fd552/el9.clyso/x86_64'}]} 2026-03-06T13:38:29.036 DEBUG:teuthology.task.install.rpm:Adding repos: [{'name': 'ceph-source', 'priority': 1, 'url': 'https://s3.clyso.com/ces-packages/components/ceph/rpm-19.2.3-47-gc24117fd552/el9.clyso/SRPMS'}, {'name': 'ceph-noarch', 'priority': 1, 'url': 'https://s3.clyso.com/ces-packages/components/ceph/rpm-19.2.3-47-gc24117fd552/el9.clyso/noarch'}, {'name': 'ceph', 'priority': 1, 'url': 'https://s3.clyso.com/ces-packages/components/ceph/rpm-19.2.3-47-gc24117fd552/el9.clyso/x86_64'}] 2026-03-06T13:38:29.036 DEBUG:teuthology.orchestra.run.vm01:> set -ex 2026-03-06T13:38:29.036 DEBUG:teuthology.orchestra.run.vm01:> sudo dd of=/etc/yum.repos.d/ceph-source.repo 2026-03-06T13:38:29.036 DEBUG:teuthology.task.install.rpm:_update_package_list_and_install: config is {'branch': None, 'cleanup': None, 'debuginfo': None, 'downgrade_packages': [], 'exclude_packages': [], 'extra_packages': [], 'extra_system_packages': {'deb': ['python3-xmltodict', 's3cmd'], 'rpm': ['bzip2', 'perl-Test-Harness', 'python3-xmltodict', 's3cmd']}, 'extras': None, 'enable_coprs': [], 'flavor': 'default', 'install_ceph_packages': True, 'packages': {}, 'project': 'ceph', 'repos_only': False, 'sha1': 'c24117fd5525679b799527bc1bd1f1dd0a2db5e2', 'tag': None, 'wait_for_package': False, 'repos': [{'name': 'ceph-source', 'priority': 1, 'url': 'https://s3.clyso.com/ces-packages/components/ceph/rpm-19.2.3-47-gc24117fd552/el9.clyso/SRPMS'}, {'name': 'ceph-noarch', 'priority': 1, 'url': 'https://s3.clyso.com/ces-packages/components/ceph/rpm-19.2.3-47-gc24117fd552/el9.clyso/noarch'}, {'name': 'ceph', 'priority': 1, 'url': 'https://s3.clyso.com/ces-packages/components/ceph/rpm-19.2.3-47-gc24117fd552/el9.clyso/x86_64'}]} 2026-03-06T13:38:29.036 DEBUG:teuthology.task.install.rpm:Adding repos: [{'name': 'ceph-source', 'priority': 1, 'url': 'https://s3.clyso.com/ces-packages/components/ceph/rpm-19.2.3-47-gc24117fd552/el9.clyso/SRPMS'}, {'name': 'ceph-noarch', 'priority': 1, 'url': 'https://s3.clyso.com/ces-packages/components/ceph/rpm-19.2.3-47-gc24117fd552/el9.clyso/noarch'}, {'name': 'ceph', 'priority': 1, 'url': 'https://s3.clyso.com/ces-packages/components/ceph/rpm-19.2.3-47-gc24117fd552/el9.clyso/x86_64'}] 2026-03-06T13:38:29.036 DEBUG:teuthology.orchestra.run.vm08:> set -ex 2026-03-06T13:38:29.036 DEBUG:teuthology.orchestra.run.vm08:> sudo dd of=/etc/yum.repos.d/ceph-source.repo 2026-03-06T13:38:29.072 DEBUG:teuthology.orchestra.run.vm08:> set -ex 2026-03-06T13:38:29.072 DEBUG:teuthology.orchestra.run.vm08:> sudo dd of=/etc/yum.repos.d/ceph-noarch.repo 2026-03-06T13:38:29.079 DEBUG:teuthology.orchestra.run.vm00:> set -ex 2026-03-06T13:38:29.079 DEBUG:teuthology.orchestra.run.vm00:> sudo dd of=/etc/yum.repos.d/ceph-noarch.repo 2026-03-06T13:38:29.089 DEBUG:teuthology.orchestra.run.vm01:> set -ex 2026-03-06T13:38:29.089 DEBUG:teuthology.orchestra.run.vm01:> sudo dd of=/etc/yum.repos.d/ceph-noarch.repo 2026-03-06T13:38:29.140 DEBUG:teuthology.orchestra.run.vm08:> set -ex 2026-03-06T13:38:29.140 DEBUG:teuthology.orchestra.run.vm08:> sudo dd of=/etc/yum.repos.d/ceph.repo 2026-03-06T13:38:29.162 DEBUG:teuthology.orchestra.run.vm00:> set -ex 2026-03-06T13:38:29.163 DEBUG:teuthology.orchestra.run.vm00:> sudo dd of=/etc/yum.repos.d/ceph.repo 2026-03-06T13:38:29.164 DEBUG:teuthology.orchestra.run.vm01:> set -ex 2026-03-06T13:38:29.164 DEBUG:teuthology.orchestra.run.vm01:> sudo dd of=/etc/yum.repos.d/ceph.repo 2026-03-06T13:38:29.210 INFO:teuthology.task.install.rpm:Installing packages: ceph-radosgw, ceph-test, ceph, ceph-base, cephadm, ceph-immutable-object-cache, ceph-mgr, ceph-mgr-dashboard, ceph-mgr-diskprediction-local, ceph-mgr-rook, ceph-mgr-cephadm, ceph-fuse, ceph-volume, librados-devel, libcephfs2, libcephfs-devel, librados2, librbd1, python3-rados, python3-rgw, python3-cephfs, python3-rbd, rbd-fuse, rbd-mirror, rbd-nbd, bzip2, perl-Test-Harness, python3-xmltodict, s3cmd on remote rpm x86_64 2026-03-06T13:38:29.211 DEBUG:teuthology.orchestra.run.vm08:> sudo yum clean all 2026-03-06T13:38:29.239 INFO:teuthology.task.install.rpm:Installing packages: ceph-radosgw, ceph-test, ceph, ceph-base, cephadm, ceph-immutable-object-cache, ceph-mgr, ceph-mgr-dashboard, ceph-mgr-diskprediction-local, ceph-mgr-rook, ceph-mgr-cephadm, ceph-fuse, ceph-volume, librados-devel, libcephfs2, libcephfs-devel, librados2, librbd1, python3-rados, python3-rgw, python3-cephfs, python3-rbd, rbd-fuse, rbd-mirror, rbd-nbd, bzip2, perl-Test-Harness, python3-xmltodict, s3cmd on remote rpm x86_64 2026-03-06T13:38:29.239 DEBUG:teuthology.orchestra.run.vm01:> sudo yum clean all 2026-03-06T13:38:29.246 INFO:teuthology.task.install.rpm:Installing packages: ceph-radosgw, ceph-test, ceph, ceph-base, cephadm, ceph-immutable-object-cache, ceph-mgr, ceph-mgr-dashboard, ceph-mgr-diskprediction-local, ceph-mgr-rook, ceph-mgr-cephadm, ceph-fuse, ceph-volume, librados-devel, libcephfs2, libcephfs-devel, librados2, librbd1, python3-rados, python3-rgw, python3-cephfs, python3-rbd, rbd-fuse, rbd-mirror, rbd-nbd, bzip2, perl-Test-Harness, python3-xmltodict, s3cmd on remote rpm x86_64 2026-03-06T13:38:29.246 DEBUG:teuthology.orchestra.run.vm00:> sudo yum clean all 2026-03-06T13:38:29.402 INFO:teuthology.orchestra.run.vm08.stdout:41 files removed 2026-03-06T13:38:29.435 DEBUG:teuthology.orchestra.run.vm08:> sudo yum -y install ceph-radosgw ceph-test ceph ceph-base cephadm ceph-immutable-object-cache ceph-mgr ceph-mgr-dashboard ceph-mgr-diskprediction-local ceph-mgr-rook ceph-mgr-cephadm ceph-fuse ceph-volume librados-devel libcephfs2 libcephfs-devel librados2 librbd1 python3-rados python3-rgw python3-cephfs python3-rbd rbd-fuse rbd-mirror rbd-nbd bzip2 perl-Test-Harness python3-xmltodict s3cmd 2026-03-06T13:38:29.450 INFO:teuthology.orchestra.run.vm01.stdout:41 files removed 2026-03-06T13:38:29.461 INFO:teuthology.orchestra.run.vm00.stdout:41 files removed 2026-03-06T13:38:29.477 DEBUG:teuthology.orchestra.run.vm01:> sudo yum -y install ceph-radosgw ceph-test ceph ceph-base cephadm ceph-immutable-object-cache ceph-mgr ceph-mgr-dashboard ceph-mgr-diskprediction-local ceph-mgr-rook ceph-mgr-cephadm ceph-fuse ceph-volume librados-devel libcephfs2 libcephfs-devel librados2 librbd1 python3-rados python3-rgw python3-cephfs python3-rbd rbd-fuse rbd-mirror rbd-nbd bzip2 perl-Test-Harness python3-xmltodict s3cmd 2026-03-06T13:38:29.492 DEBUG:teuthology.orchestra.run.vm00:> sudo yum -y install ceph-radosgw ceph-test ceph ceph-base cephadm ceph-immutable-object-cache ceph-mgr ceph-mgr-dashboard ceph-mgr-diskprediction-local ceph-mgr-rook ceph-mgr-cephadm ceph-fuse ceph-volume librados-devel libcephfs2 libcephfs-devel librados2 librbd1 python3-rados python3-rgw python3-cephfs python3-rbd rbd-fuse rbd-mirror rbd-nbd bzip2 perl-Test-Harness python3-xmltodict s3cmd 2026-03-06T13:38:29.827 INFO:teuthology.orchestra.run.vm08.stdout:ceph 398 kB/s | 86 kB 00:00 2026-03-06T13:38:29.925 INFO:teuthology.orchestra.run.vm00.stdout:ceph 345 kB/s | 86 kB 00:00 2026-03-06T13:38:29.939 INFO:teuthology.orchestra.run.vm01.stdout:ceph 335 kB/s | 86 kB 00:00 2026-03-06T13:38:30.137 INFO:teuthology.orchestra.run.vm01.stdout:ceph-noarch 71 kB/s | 12 kB 00:00 2026-03-06T13:38:30.185 INFO:teuthology.orchestra.run.vm08.stdout:ceph-noarch 36 kB/s | 12 kB 00:00 2026-03-06T13:38:30.211 INFO:teuthology.orchestra.run.vm00.stdout:ceph-noarch 46 kB/s | 12 kB 00:00 2026-03-06T13:38:30.429 INFO:teuthology.orchestra.run.vm01.stdout:ceph-source 8.1 kB/s | 2.2 kB 00:00 2026-03-06T13:38:30.482 INFO:teuthology.orchestra.run.vm08.stdout:ceph-source 8.0 kB/s | 2.2 kB 00:00 2026-03-06T13:38:30.538 INFO:teuthology.orchestra.run.vm00.stdout:ceph-source 7.2 kB/s | 2.2 kB 00:00 2026-03-06T13:38:31.043 INFO:teuthology.orchestra.run.vm01.stdout:CentOS Stream 9 - BaseOS 15 MB/s | 8.9 MB 00:00 2026-03-06T13:38:31.101 INFO:teuthology.orchestra.run.vm08.stdout:CentOS Stream 9 - BaseOS 15 MB/s | 8.9 MB 00:00 2026-03-06T13:38:31.756 INFO:teuthology.orchestra.run.vm00.stdout:CentOS Stream 9 - BaseOS 7.4 MB/s | 8.9 MB 00:01 2026-03-06T13:38:33.660 INFO:teuthology.orchestra.run.vm08.stdout:CentOS Stream 9 - AppStream 14 MB/s | 27 MB 00:01 2026-03-06T13:38:35.239 INFO:teuthology.orchestra.run.vm00.stdout:CentOS Stream 9 - AppStream 9.5 MB/s | 27 MB 00:02 2026-03-06T13:38:36.544 INFO:teuthology.orchestra.run.vm01.stdout:CentOS Stream 9 - AppStream 5.8 MB/s | 27 MB 00:04 2026-03-06T13:38:37.734 INFO:teuthology.orchestra.run.vm08.stdout:CentOS Stream 9 - CRB 6.2 MB/s | 8.0 MB 00:01 2026-03-06T13:38:38.936 INFO:teuthology.orchestra.run.vm08.stdout:CentOS Stream 9 - Extras packages 55 kB/s | 20 kB 00:00 2026-03-06T13:38:39.129 INFO:teuthology.orchestra.run.vm00.stdout:CentOS Stream 9 - CRB 6.8 MB/s | 8.0 MB 00:01 2026-03-06T13:38:39.800 INFO:teuthology.orchestra.run.vm08.stdout:Extra Packages for Enterprise Linux 26 MB/s | 20 MB 00:00 2026-03-06T13:38:40.317 INFO:teuthology.orchestra.run.vm00.stdout:CentOS Stream 9 - Extras packages 56 kB/s | 20 kB 00:00 2026-03-06T13:38:40.601 INFO:teuthology.orchestra.run.vm01.stdout:CentOS Stream 9 - CRB 14 MB/s | 8.0 MB 00:00 2026-03-06T13:38:41.104 INFO:teuthology.orchestra.run.vm00.stdout:Extra Packages for Enterprise Linux 29 MB/s | 20 MB 00:00 2026-03-06T13:38:41.970 INFO:teuthology.orchestra.run.vm01.stdout:CentOS Stream 9 - Extras packages 61 kB/s | 20 kB 00:00 2026-03-06T13:38:42.918 INFO:teuthology.orchestra.run.vm01.stdout:Extra Packages for Enterprise Linux 24 MB/s | 20 MB 00:00 2026-03-06T13:38:44.487 INFO:teuthology.orchestra.run.vm08.stdout:lab-extras 65 kB/s | 50 kB 00:00 2026-03-06T13:38:45.716 INFO:teuthology.orchestra.run.vm00.stdout:lab-extras 63 kB/s | 50 kB 00:00 2026-03-06T13:38:45.924 INFO:teuthology.orchestra.run.vm08.stdout:Package librados2-2:16.2.4-5.el9.x86_64 is already installed. 2026-03-06T13:38:45.924 INFO:teuthology.orchestra.run.vm08.stdout:Package librbd1-2:16.2.4-5.el9.x86_64 is already installed. 2026-03-06T13:38:45.958 INFO:teuthology.orchestra.run.vm08.stdout:Dependencies resolved. 2026-03-06T13:38:45.963 INFO:teuthology.orchestra.run.vm08.stdout:============================================================================================== 2026-03-06T13:38:45.963 INFO:teuthology.orchestra.run.vm08.stdout: Package Arch Version Repository Size 2026-03-06T13:38:45.963 INFO:teuthology.orchestra.run.vm08.stdout:============================================================================================== 2026-03-06T13:38:45.963 INFO:teuthology.orchestra.run.vm08.stdout:Installing: 2026-03-06T13:38:45.963 INFO:teuthology.orchestra.run.vm08.stdout: bzip2 x86_64 1.0.8-11.el9 baseos 55 k 2026-03-06T13:38:45.963 INFO:teuthology.orchestra.run.vm08.stdout: ceph x86_64 2:19.2.3-47.gc24117fd552.el9.clyso ceph 6.5 k 2026-03-06T13:38:45.963 INFO:teuthology.orchestra.run.vm08.stdout: ceph-base x86_64 2:19.2.3-47.gc24117fd552.el9.clyso ceph 5.5 M 2026-03-06T13:38:45.963 INFO:teuthology.orchestra.run.vm08.stdout: ceph-fuse x86_64 2:19.2.3-47.gc24117fd552.el9.clyso ceph 1.1 M 2026-03-06T13:38:45.963 INFO:teuthology.orchestra.run.vm08.stdout: ceph-immutable-object-cache x86_64 2:19.2.3-47.gc24117fd552.el9.clyso ceph 145 k 2026-03-06T13:38:45.963 INFO:teuthology.orchestra.run.vm08.stdout: ceph-mgr x86_64 2:19.2.3-47.gc24117fd552.el9.clyso ceph 1.1 M 2026-03-06T13:38:45.963 INFO:teuthology.orchestra.run.vm08.stdout: ceph-mgr-cephadm noarch 2:19.2.3-47.gc24117fd552.el9.clyso ceph-noarch 150 k 2026-03-06T13:38:45.963 INFO:teuthology.orchestra.run.vm08.stdout: ceph-mgr-dashboard noarch 2:19.2.3-47.gc24117fd552.el9.clyso ceph-noarch 3.8 M 2026-03-06T13:38:45.963 INFO:teuthology.orchestra.run.vm08.stdout: ceph-mgr-diskprediction-local noarch 2:19.2.3-47.gc24117fd552.el9.clyso ceph-noarch 7.4 M 2026-03-06T13:38:45.963 INFO:teuthology.orchestra.run.vm08.stdout: ceph-mgr-rook noarch 2:19.2.3-47.gc24117fd552.el9.clyso ceph-noarch 49 k 2026-03-06T13:38:45.963 INFO:teuthology.orchestra.run.vm08.stdout: ceph-radosgw x86_64 2:19.2.3-47.gc24117fd552.el9.clyso ceph 11 M 2026-03-06T13:38:45.963 INFO:teuthology.orchestra.run.vm08.stdout: ceph-test x86_64 2:19.2.3-47.gc24117fd552.el9.clyso ceph 50 M 2026-03-06T13:38:45.963 INFO:teuthology.orchestra.run.vm08.stdout: ceph-volume noarch 2:19.2.3-47.gc24117fd552.el9.clyso ceph-noarch 299 k 2026-03-06T13:38:45.963 INFO:teuthology.orchestra.run.vm08.stdout: cephadm noarch 2:19.2.3-47.gc24117fd552.el9.clyso ceph-noarch 769 k 2026-03-06T13:38:45.963 INFO:teuthology.orchestra.run.vm08.stdout: libcephfs-devel x86_64 2:19.2.3-47.gc24117fd552.el9.clyso ceph 34 k 2026-03-06T13:38:45.963 INFO:teuthology.orchestra.run.vm08.stdout: libcephfs2 x86_64 2:19.2.3-47.gc24117fd552.el9.clyso ceph 998 k 2026-03-06T13:38:45.963 INFO:teuthology.orchestra.run.vm08.stdout: librados-devel x86_64 2:19.2.3-47.gc24117fd552.el9.clyso ceph 127 k 2026-03-06T13:38:45.964 INFO:teuthology.orchestra.run.vm08.stdout: perl-Test-Harness noarch 1:3.42-461.el9 appstream 295 k 2026-03-06T13:38:45.964 INFO:teuthology.orchestra.run.vm08.stdout: python3-cephfs x86_64 2:19.2.3-47.gc24117fd552.el9.clyso ceph 165 k 2026-03-06T13:38:45.964 INFO:teuthology.orchestra.run.vm08.stdout: python3-rados x86_64 2:19.2.3-47.gc24117fd552.el9.clyso ceph 322 k 2026-03-06T13:38:45.964 INFO:teuthology.orchestra.run.vm08.stdout: python3-rbd x86_64 2:19.2.3-47.gc24117fd552.el9.clyso ceph 303 k 2026-03-06T13:38:45.964 INFO:teuthology.orchestra.run.vm08.stdout: python3-rgw x86_64 2:19.2.3-47.gc24117fd552.el9.clyso ceph 100 k 2026-03-06T13:38:45.964 INFO:teuthology.orchestra.run.vm08.stdout: python3-xmltodict noarch 0.12.0-15.el9 epel 22 k 2026-03-06T13:38:45.964 INFO:teuthology.orchestra.run.vm08.stdout: rbd-fuse x86_64 2:19.2.3-47.gc24117fd552.el9.clyso ceph 85 k 2026-03-06T13:38:45.964 INFO:teuthology.orchestra.run.vm08.stdout: rbd-mirror x86_64 2:19.2.3-47.gc24117fd552.el9.clyso ceph 3.1 M 2026-03-06T13:38:45.964 INFO:teuthology.orchestra.run.vm08.stdout: rbd-nbd x86_64 2:19.2.3-47.gc24117fd552.el9.clyso ceph 171 k 2026-03-06T13:38:45.964 INFO:teuthology.orchestra.run.vm08.stdout: s3cmd noarch 2.4.0-1.el9 epel 206 k 2026-03-06T13:38:45.964 INFO:teuthology.orchestra.run.vm08.stdout:Upgrading: 2026-03-06T13:38:45.964 INFO:teuthology.orchestra.run.vm08.stdout: librados2 x86_64 2:19.2.3-47.gc24117fd552.el9.clyso ceph 3.4 M 2026-03-06T13:38:45.964 INFO:teuthology.orchestra.run.vm08.stdout: librbd1 x86_64 2:19.2.3-47.gc24117fd552.el9.clyso ceph 3.2 M 2026-03-06T13:38:45.964 INFO:teuthology.orchestra.run.vm08.stdout:Installing dependencies: 2026-03-06T13:38:45.964 INFO:teuthology.orchestra.run.vm08.stdout: abseil-cpp x86_64 20211102.0-4.el9 epel 551 k 2026-03-06T13:38:45.964 INFO:teuthology.orchestra.run.vm08.stdout: boost-program-options x86_64 1.75.0-13.el9 appstream 104 k 2026-03-06T13:38:45.964 INFO:teuthology.orchestra.run.vm08.stdout: ceph-common x86_64 2:19.2.3-47.gc24117fd552.el9.clyso ceph 22 M 2026-03-06T13:38:45.964 INFO:teuthology.orchestra.run.vm08.stdout: ceph-grafana-dashboards noarch 2:19.2.3-47.gc24117fd552.el9.clyso ceph-noarch 31 k 2026-03-06T13:38:45.964 INFO:teuthology.orchestra.run.vm08.stdout: ceph-mds x86_64 2:19.2.3-47.gc24117fd552.el9.clyso ceph 2.4 M 2026-03-06T13:38:45.964 INFO:teuthology.orchestra.run.vm08.stdout: ceph-mgr-modules-core noarch 2:19.2.3-47.gc24117fd552.el9.clyso ceph-noarch 252 k 2026-03-06T13:38:45.964 INFO:teuthology.orchestra.run.vm08.stdout: ceph-mon x86_64 2:19.2.3-47.gc24117fd552.el9.clyso ceph 4.7 M 2026-03-06T13:38:45.964 INFO:teuthology.orchestra.run.vm08.stdout: ceph-osd x86_64 2:19.2.3-47.gc24117fd552.el9.clyso ceph 17 M 2026-03-06T13:38:45.964 INFO:teuthology.orchestra.run.vm08.stdout: ceph-prometheus-alerts noarch 2:19.2.3-47.gc24117fd552.el9.clyso ceph-noarch 16 k 2026-03-06T13:38:45.964 INFO:teuthology.orchestra.run.vm08.stdout: ceph-selinux x86_64 2:19.2.3-47.gc24117fd552.el9.clyso ceph 25 k 2026-03-06T13:38:45.964 INFO:teuthology.orchestra.run.vm08.stdout: cryptsetup x86_64 2.8.1-3.el9 baseos 351 k 2026-03-06T13:38:45.964 INFO:teuthology.orchestra.run.vm08.stdout: flexiblas x86_64 3.0.4-9.el9 appstream 30 k 2026-03-06T13:38:45.964 INFO:teuthology.orchestra.run.vm08.stdout: flexiblas-netlib x86_64 3.0.4-9.el9 appstream 3.0 M 2026-03-06T13:38:45.964 INFO:teuthology.orchestra.run.vm08.stdout: flexiblas-openblas-openmp x86_64 3.0.4-9.el9 appstream 15 k 2026-03-06T13:38:45.964 INFO:teuthology.orchestra.run.vm08.stdout: fuse x86_64 2.9.9-17.el9 baseos 80 k 2026-03-06T13:38:45.964 INFO:teuthology.orchestra.run.vm08.stdout: gperftools-libs x86_64 2.9.1-3.el9 epel 308 k 2026-03-06T13:38:45.964 INFO:teuthology.orchestra.run.vm08.stdout: grpc-data noarch 1.46.7-10.el9 epel 19 k 2026-03-06T13:38:45.964 INFO:teuthology.orchestra.run.vm08.stdout: ledmon-libs x86_64 1.1.0-3.el9 baseos 40 k 2026-03-06T13:38:45.964 INFO:teuthology.orchestra.run.vm08.stdout: libarrow x86_64 9.0.0-15.el9 epel 4.4 M 2026-03-06T13:38:45.964 INFO:teuthology.orchestra.run.vm08.stdout: libarrow-doc noarch 9.0.0-15.el9 epel 25 k 2026-03-06T13:38:45.964 INFO:teuthology.orchestra.run.vm08.stdout: libcephsqlite x86_64 2:19.2.3-47.gc24117fd552.el9.clyso ceph 163 k 2026-03-06T13:38:45.964 INFO:teuthology.orchestra.run.vm08.stdout: libconfig x86_64 1.7.2-9.el9 baseos 72 k 2026-03-06T13:38:45.964 INFO:teuthology.orchestra.run.vm08.stdout: libgfortran x86_64 11.5.0-14.el9 baseos 794 k 2026-03-06T13:38:45.964 INFO:teuthology.orchestra.run.vm08.stdout: libnbd x86_64 1.20.3-4.el9 appstream 164 k 2026-03-06T13:38:45.964 INFO:teuthology.orchestra.run.vm08.stdout: liboath x86_64 2.6.12-1.el9 epel 49 k 2026-03-06T13:38:45.964 INFO:teuthology.orchestra.run.vm08.stdout: libpmemobj x86_64 1.12.1-1.el9 appstream 160 k 2026-03-06T13:38:45.964 INFO:teuthology.orchestra.run.vm08.stdout: libquadmath x86_64 11.5.0-14.el9 baseos 184 k 2026-03-06T13:38:45.964 INFO:teuthology.orchestra.run.vm08.stdout: librabbitmq x86_64 0.11.0-7.el9 appstream 45 k 2026-03-06T13:38:45.964 INFO:teuthology.orchestra.run.vm08.stdout: libradosstriper1 x86_64 2:19.2.3-47.gc24117fd552.el9.clyso ceph 503 k 2026-03-06T13:38:45.964 INFO:teuthology.orchestra.run.vm08.stdout: librdkafka x86_64 1.6.1-102.el9 appstream 662 k 2026-03-06T13:38:45.964 INFO:teuthology.orchestra.run.vm08.stdout: librgw2 x86_64 2:19.2.3-47.gc24117fd552.el9.clyso ceph 5.4 M 2026-03-06T13:38:45.964 INFO:teuthology.orchestra.run.vm08.stdout: libstoragemgmt x86_64 1.10.1-1.el9 appstream 246 k 2026-03-06T13:38:45.964 INFO:teuthology.orchestra.run.vm08.stdout: libunwind x86_64 1.6.2-1.el9 epel 67 k 2026-03-06T13:38:45.964 INFO:teuthology.orchestra.run.vm08.stdout: libxslt x86_64 1.1.34-12.el9 appstream 233 k 2026-03-06T13:38:45.964 INFO:teuthology.orchestra.run.vm08.stdout: lttng-ust x86_64 2.12.0-6.el9 appstream 292 k 2026-03-06T13:38:45.964 INFO:teuthology.orchestra.run.vm08.stdout: lua x86_64 5.4.4-4.el9 appstream 188 k 2026-03-06T13:38:45.964 INFO:teuthology.orchestra.run.vm08.stdout: lua-devel x86_64 5.4.4-4.el9 crb 22 k 2026-03-06T13:38:45.965 INFO:teuthology.orchestra.run.vm08.stdout: luarocks noarch 3.9.2-5.el9 epel 151 k 2026-03-06T13:38:45.965 INFO:teuthology.orchestra.run.vm08.stdout: mailcap noarch 2.1.49-5.el9 baseos 33 k 2026-03-06T13:38:45.965 INFO:teuthology.orchestra.run.vm08.stdout: openblas x86_64 0.3.29-1.el9 appstream 42 k 2026-03-06T13:38:45.965 INFO:teuthology.orchestra.run.vm08.stdout: openblas-openmp x86_64 0.3.29-1.el9 appstream 5.3 M 2026-03-06T13:38:45.965 INFO:teuthology.orchestra.run.vm08.stdout: parquet-libs x86_64 9.0.0-15.el9 epel 838 k 2026-03-06T13:38:45.965 INFO:teuthology.orchestra.run.vm08.stdout: pciutils x86_64 3.7.0-7.el9 baseos 93 k 2026-03-06T13:38:45.965 INFO:teuthology.orchestra.run.vm08.stdout: perl-Benchmark noarch 1.23-483.el9 appstream 26 k 2026-03-06T13:38:45.965 INFO:teuthology.orchestra.run.vm08.stdout: protobuf x86_64 3.14.0-17.el9 appstream 1.0 M 2026-03-06T13:38:45.965 INFO:teuthology.orchestra.run.vm08.stdout: protobuf-compiler x86_64 3.14.0-17.el9 crb 862 k 2026-03-06T13:38:45.965 INFO:teuthology.orchestra.run.vm08.stdout: python3-asyncssh noarch 2.13.2-5.el9 epel 548 k 2026-03-06T13:38:45.965 INFO:teuthology.orchestra.run.vm08.stdout: python3-autocommand noarch 2.2.2-8.el9 epel 29 k 2026-03-06T13:38:45.965 INFO:teuthology.orchestra.run.vm08.stdout: python3-babel noarch 2.9.1-2.el9 appstream 6.0 M 2026-03-06T13:38:45.965 INFO:teuthology.orchestra.run.vm08.stdout: python3-backports-tarfile noarch 1.2.0-1.el9 epel 60 k 2026-03-06T13:38:45.965 INFO:teuthology.orchestra.run.vm08.stdout: python3-bcrypt x86_64 3.2.2-1.el9 epel 43 k 2026-03-06T13:38:45.965 INFO:teuthology.orchestra.run.vm08.stdout: python3-cachetools noarch 4.2.4-1.el9 epel 32 k 2026-03-06T13:38:45.965 INFO:teuthology.orchestra.run.vm08.stdout: python3-ceph-argparse x86_64 2:19.2.3-47.gc24117fd552.el9.clyso ceph 45 k 2026-03-06T13:38:45.965 INFO:teuthology.orchestra.run.vm08.stdout: python3-ceph-common x86_64 2:19.2.3-47.gc24117fd552.el9.clyso ceph 142 k 2026-03-06T13:38:45.965 INFO:teuthology.orchestra.run.vm08.stdout: python3-certifi noarch 2023.05.07-4.el9 epel 14 k 2026-03-06T13:38:45.965 INFO:teuthology.orchestra.run.vm08.stdout: python3-cffi x86_64 1.14.5-5.el9 baseos 253 k 2026-03-06T13:38:45.965 INFO:teuthology.orchestra.run.vm08.stdout: python3-cheroot noarch 10.0.1-4.el9 epel 173 k 2026-03-06T13:38:45.965 INFO:teuthology.orchestra.run.vm08.stdout: python3-cherrypy noarch 18.6.1-2.el9 epel 358 k 2026-03-06T13:38:45.965 INFO:teuthology.orchestra.run.vm08.stdout: python3-cryptography x86_64 36.0.1-5.el9 baseos 1.2 M 2026-03-06T13:38:45.965 INFO:teuthology.orchestra.run.vm08.stdout: python3-devel x86_64 3.9.25-3.el9 appstream 244 k 2026-03-06T13:38:45.965 INFO:teuthology.orchestra.run.vm08.stdout: python3-google-auth noarch 1:2.45.0-1.el9 epel 254 k 2026-03-06T13:38:45.965 INFO:teuthology.orchestra.run.vm08.stdout: python3-grpcio x86_64 1.46.7-10.el9 epel 2.0 M 2026-03-06T13:38:45.965 INFO:teuthology.orchestra.run.vm08.stdout: python3-grpcio-tools x86_64 1.46.7-10.el9 epel 144 k 2026-03-06T13:38:45.965 INFO:teuthology.orchestra.run.vm08.stdout: python3-isodate noarch 0.6.1-3.el9 epel 56 k 2026-03-06T13:38:45.965 INFO:teuthology.orchestra.run.vm08.stdout: python3-jaraco noarch 8.2.1-3.el9 epel 11 k 2026-03-06T13:38:45.965 INFO:teuthology.orchestra.run.vm08.stdout: python3-jaraco-classes noarch 3.2.1-5.el9 epel 18 k 2026-03-06T13:38:45.965 INFO:teuthology.orchestra.run.vm08.stdout: python3-jaraco-collections noarch 3.0.0-8.el9 epel 23 k 2026-03-06T13:38:45.965 INFO:teuthology.orchestra.run.vm08.stdout: python3-jaraco-context noarch 6.0.1-3.el9 epel 20 k 2026-03-06T13:38:45.965 INFO:teuthology.orchestra.run.vm08.stdout: python3-jaraco-functools noarch 3.5.0-2.el9 epel 19 k 2026-03-06T13:38:45.965 INFO:teuthology.orchestra.run.vm08.stdout: python3-jaraco-text noarch 4.0.0-2.el9 epel 26 k 2026-03-06T13:38:45.965 INFO:teuthology.orchestra.run.vm08.stdout: python3-jinja2 noarch 2.11.3-8.el9 appstream 249 k 2026-03-06T13:38:45.965 INFO:teuthology.orchestra.run.vm08.stdout: python3-kubernetes noarch 1:26.1.0-3.el9 epel 1.0 M 2026-03-06T13:38:45.965 INFO:teuthology.orchestra.run.vm08.stdout: python3-libstoragemgmt x86_64 1.10.1-1.el9 appstream 177 k 2026-03-06T13:38:45.965 INFO:teuthology.orchestra.run.vm08.stdout: python3-logutils noarch 0.3.5-21.el9 epel 46 k 2026-03-06T13:38:45.965 INFO:teuthology.orchestra.run.vm08.stdout: python3-lxml x86_64 4.6.5-3.el9 appstream 1.2 M 2026-03-06T13:38:45.965 INFO:teuthology.orchestra.run.vm08.stdout: python3-mako noarch 1.1.4-6.el9 appstream 172 k 2026-03-06T13:38:45.965 INFO:teuthology.orchestra.run.vm08.stdout: python3-markupsafe x86_64 1.1.1-12.el9 appstream 35 k 2026-03-06T13:38:45.965 INFO:teuthology.orchestra.run.vm08.stdout: python3-more-itertools noarch 8.12.0-2.el9 epel 79 k 2026-03-06T13:38:45.965 INFO:teuthology.orchestra.run.vm08.stdout: python3-msgpack x86_64 1.0.3-2.el9 epel 86 k 2026-03-06T13:38:45.965 INFO:teuthology.orchestra.run.vm08.stdout: python3-natsort noarch 7.1.1-5.el9 epel 58 k 2026-03-06T13:38:45.965 INFO:teuthology.orchestra.run.vm08.stdout: python3-numpy x86_64 1:1.23.5-2.el9 appstream 6.1 M 2026-03-06T13:38:45.965 INFO:teuthology.orchestra.run.vm08.stdout: python3-numpy-f2py x86_64 1:1.23.5-2.el9 appstream 442 k 2026-03-06T13:38:45.965 INFO:teuthology.orchestra.run.vm08.stdout: python3-packaging noarch 20.9-5.el9 appstream 77 k 2026-03-06T13:38:45.965 INFO:teuthology.orchestra.run.vm08.stdout: python3-pecan noarch 1.4.2-3.el9 epel 272 k 2026-03-06T13:38:45.965 INFO:teuthology.orchestra.run.vm08.stdout: python3-ply noarch 3.11-14.el9 baseos 106 k 2026-03-06T13:38:45.965 INFO:teuthology.orchestra.run.vm08.stdout: python3-portend noarch 3.1.0-2.el9 epel 16 k 2026-03-06T13:38:45.966 INFO:teuthology.orchestra.run.vm08.stdout: python3-protobuf noarch 3.14.0-17.el9 appstream 267 k 2026-03-06T13:38:45.966 INFO:teuthology.orchestra.run.vm08.stdout: python3-pyOpenSSL noarch 21.0.0-1.el9 epel 90 k 2026-03-06T13:38:45.966 INFO:teuthology.orchestra.run.vm08.stdout: python3-pyasn1 noarch 0.4.8-7.el9 appstream 157 k 2026-03-06T13:38:45.966 INFO:teuthology.orchestra.run.vm08.stdout: python3-pyasn1-modules noarch 0.4.8-7.el9 appstream 277 k 2026-03-06T13:38:45.966 INFO:teuthology.orchestra.run.vm08.stdout: python3-pycparser noarch 2.20-6.el9 baseos 135 k 2026-03-06T13:38:45.966 INFO:teuthology.orchestra.run.vm08.stdout: python3-repoze-lru noarch 0.7-16.el9 epel 31 k 2026-03-06T13:38:45.966 INFO:teuthology.orchestra.run.vm08.stdout: python3-requests noarch 2.25.1-10.el9 baseos 126 k 2026-03-06T13:38:45.966 INFO:teuthology.orchestra.run.vm08.stdout: python3-requests-oauthlib noarch 1.3.0-12.el9 appstream 54 k 2026-03-06T13:38:45.966 INFO:teuthology.orchestra.run.vm08.stdout: python3-routes noarch 2.5.1-5.el9 epel 188 k 2026-03-06T13:38:45.966 INFO:teuthology.orchestra.run.vm08.stdout: python3-rsa noarch 4.9-2.el9 epel 59 k 2026-03-06T13:38:45.966 INFO:teuthology.orchestra.run.vm08.stdout: python3-scipy x86_64 1.9.3-2.el9 appstream 19 M 2026-03-06T13:38:45.966 INFO:teuthology.orchestra.run.vm08.stdout: python3-tempora noarch 5.0.0-2.el9 epel 36 k 2026-03-06T13:38:45.966 INFO:teuthology.orchestra.run.vm08.stdout: python3-toml noarch 0.10.2-6.el9 appstream 42 k 2026-03-06T13:38:45.966 INFO:teuthology.orchestra.run.vm08.stdout: python3-typing-extensions noarch 4.15.0-1.el9 epel 86 k 2026-03-06T13:38:45.966 INFO:teuthology.orchestra.run.vm08.stdout: python3-urllib3 noarch 1.26.5-7.el9 baseos 218 k 2026-03-06T13:38:45.966 INFO:teuthology.orchestra.run.vm08.stdout: python3-webob noarch 1.8.8-2.el9 epel 230 k 2026-03-06T13:38:45.966 INFO:teuthology.orchestra.run.vm08.stdout: python3-websocket-client noarch 1.2.3-2.el9 epel 90 k 2026-03-06T13:38:45.966 INFO:teuthology.orchestra.run.vm08.stdout: python3-werkzeug noarch 2.0.3-3.el9.1 epel 427 k 2026-03-06T13:38:45.966 INFO:teuthology.orchestra.run.vm08.stdout: python3-xmlsec x86_64 1.3.13-1.el9 epel 48 k 2026-03-06T13:38:45.966 INFO:teuthology.orchestra.run.vm08.stdout: python3-zc-lockfile noarch 2.0-10.el9 epel 20 k 2026-03-06T13:38:45.966 INFO:teuthology.orchestra.run.vm08.stdout: qatlib x86_64 25.08.0-2.el9 appstream 240 k 2026-03-06T13:38:45.966 INFO:teuthology.orchestra.run.vm08.stdout: qatzip-libs x86_64 1.3.1-1.el9 appstream 66 k 2026-03-06T13:38:45.966 INFO:teuthology.orchestra.run.vm08.stdout: re2 x86_64 1:20211101-20.el9 epel 191 k 2026-03-06T13:38:45.966 INFO:teuthology.orchestra.run.vm08.stdout: socat x86_64 1.7.4.1-8.el9 appstream 303 k 2026-03-06T13:38:45.966 INFO:teuthology.orchestra.run.vm08.stdout: thrift x86_64 0.15.0-4.el9 epel 1.6 M 2026-03-06T13:38:45.966 INFO:teuthology.orchestra.run.vm08.stdout: unzip x86_64 6.0-59.el9 baseos 182 k 2026-03-06T13:38:45.966 INFO:teuthology.orchestra.run.vm08.stdout: xmlsec1 x86_64 1.2.29-13.el9 appstream 189 k 2026-03-06T13:38:45.966 INFO:teuthology.orchestra.run.vm08.stdout: xmlsec1-openssl x86_64 1.2.29-13.el9 appstream 90 k 2026-03-06T13:38:45.966 INFO:teuthology.orchestra.run.vm08.stdout: xmlstarlet x86_64 1.6.1-20.el9 appstream 64 k 2026-03-06T13:38:45.966 INFO:teuthology.orchestra.run.vm08.stdout: zip x86_64 3.0-35.el9 baseos 266 k 2026-03-06T13:38:45.966 INFO:teuthology.orchestra.run.vm08.stdout:Installing weak dependencies: 2026-03-06T13:38:45.966 INFO:teuthology.orchestra.run.vm08.stdout: ceph-mgr-k8sevents noarch 2:19.2.3-47.gc24117fd552.el9.clyso ceph-noarch 22 k 2026-03-06T13:38:45.966 INFO:teuthology.orchestra.run.vm08.stdout: python3-influxdb noarch 5.3.1-1.el9 epel 139 k 2026-03-06T13:38:45.966 INFO:teuthology.orchestra.run.vm08.stdout: python3-saml noarch 1.16.0-1.el9 epel 125 k 2026-03-06T13:38:45.966 INFO:teuthology.orchestra.run.vm08.stdout: qatlib-service x86_64 25.08.0-2.el9 appstream 37 k 2026-03-06T13:38:45.966 INFO:teuthology.orchestra.run.vm08.stdout: smartmontools x86_64 1:7.2-10.el9 baseos 556 k 2026-03-06T13:38:45.966 INFO:teuthology.orchestra.run.vm08.stdout: 2026-03-06T13:38:45.966 INFO:teuthology.orchestra.run.vm08.stdout:Transaction Summary 2026-03-06T13:38:45.966 INFO:teuthology.orchestra.run.vm08.stdout:============================================================================================== 2026-03-06T13:38:45.966 INFO:teuthology.orchestra.run.vm08.stdout:Install 148 Packages 2026-03-06T13:38:45.966 INFO:teuthology.orchestra.run.vm08.stdout:Upgrade 2 Packages 2026-03-06T13:38:45.966 INFO:teuthology.orchestra.run.vm08.stdout: 2026-03-06T13:38:45.967 INFO:teuthology.orchestra.run.vm08.stdout:Total download size: 214 M 2026-03-06T13:38:45.967 INFO:teuthology.orchestra.run.vm08.stdout:Downloading Packages: 2026-03-06T13:38:47.114 INFO:teuthology.orchestra.run.vm00.stdout:Package librados2-2:16.2.4-5.el9.x86_64 is already installed. 2026-03-06T13:38:47.115 INFO:teuthology.orchestra.run.vm00.stdout:Package librbd1-2:16.2.4-5.el9.x86_64 is already installed. 2026-03-06T13:38:47.149 INFO:teuthology.orchestra.run.vm00.stdout:Dependencies resolved. 2026-03-06T13:38:47.154 INFO:teuthology.orchestra.run.vm00.stdout:============================================================================================== 2026-03-06T13:38:47.154 INFO:teuthology.orchestra.run.vm00.stdout: Package Arch Version Repository Size 2026-03-06T13:38:47.154 INFO:teuthology.orchestra.run.vm00.stdout:============================================================================================== 2026-03-06T13:38:47.154 INFO:teuthology.orchestra.run.vm00.stdout:Installing: 2026-03-06T13:38:47.154 INFO:teuthology.orchestra.run.vm00.stdout: bzip2 x86_64 1.0.8-11.el9 baseos 55 k 2026-03-06T13:38:47.154 INFO:teuthology.orchestra.run.vm00.stdout: ceph x86_64 2:19.2.3-47.gc24117fd552.el9.clyso ceph 6.5 k 2026-03-06T13:38:47.154 INFO:teuthology.orchestra.run.vm00.stdout: ceph-base x86_64 2:19.2.3-47.gc24117fd552.el9.clyso ceph 5.5 M 2026-03-06T13:38:47.154 INFO:teuthology.orchestra.run.vm00.stdout: ceph-fuse x86_64 2:19.2.3-47.gc24117fd552.el9.clyso ceph 1.1 M 2026-03-06T13:38:47.154 INFO:teuthology.orchestra.run.vm00.stdout: ceph-immutable-object-cache x86_64 2:19.2.3-47.gc24117fd552.el9.clyso ceph 145 k 2026-03-06T13:38:47.154 INFO:teuthology.orchestra.run.vm00.stdout: ceph-mgr x86_64 2:19.2.3-47.gc24117fd552.el9.clyso ceph 1.1 M 2026-03-06T13:38:47.154 INFO:teuthology.orchestra.run.vm00.stdout: ceph-mgr-cephadm noarch 2:19.2.3-47.gc24117fd552.el9.clyso ceph-noarch 150 k 2026-03-06T13:38:47.154 INFO:teuthology.orchestra.run.vm00.stdout: ceph-mgr-dashboard noarch 2:19.2.3-47.gc24117fd552.el9.clyso ceph-noarch 3.8 M 2026-03-06T13:38:47.154 INFO:teuthology.orchestra.run.vm00.stdout: ceph-mgr-diskprediction-local noarch 2:19.2.3-47.gc24117fd552.el9.clyso ceph-noarch 7.4 M 2026-03-06T13:38:47.154 INFO:teuthology.orchestra.run.vm00.stdout: ceph-mgr-rook noarch 2:19.2.3-47.gc24117fd552.el9.clyso ceph-noarch 49 k 2026-03-06T13:38:47.154 INFO:teuthology.orchestra.run.vm00.stdout: ceph-radosgw x86_64 2:19.2.3-47.gc24117fd552.el9.clyso ceph 11 M 2026-03-06T13:38:47.154 INFO:teuthology.orchestra.run.vm00.stdout: ceph-test x86_64 2:19.2.3-47.gc24117fd552.el9.clyso ceph 50 M 2026-03-06T13:38:47.154 INFO:teuthology.orchestra.run.vm00.stdout: ceph-volume noarch 2:19.2.3-47.gc24117fd552.el9.clyso ceph-noarch 299 k 2026-03-06T13:38:47.154 INFO:teuthology.orchestra.run.vm00.stdout: cephadm noarch 2:19.2.3-47.gc24117fd552.el9.clyso ceph-noarch 769 k 2026-03-06T13:38:47.154 INFO:teuthology.orchestra.run.vm00.stdout: libcephfs-devel x86_64 2:19.2.3-47.gc24117fd552.el9.clyso ceph 34 k 2026-03-06T13:38:47.154 INFO:teuthology.orchestra.run.vm00.stdout: libcephfs2 x86_64 2:19.2.3-47.gc24117fd552.el9.clyso ceph 998 k 2026-03-06T13:38:47.154 INFO:teuthology.orchestra.run.vm00.stdout: librados-devel x86_64 2:19.2.3-47.gc24117fd552.el9.clyso ceph 127 k 2026-03-06T13:38:47.154 INFO:teuthology.orchestra.run.vm00.stdout: perl-Test-Harness noarch 1:3.42-461.el9 appstream 295 k 2026-03-06T13:38:47.154 INFO:teuthology.orchestra.run.vm00.stdout: python3-cephfs x86_64 2:19.2.3-47.gc24117fd552.el9.clyso ceph 165 k 2026-03-06T13:38:47.154 INFO:teuthology.orchestra.run.vm00.stdout: python3-rados x86_64 2:19.2.3-47.gc24117fd552.el9.clyso ceph 322 k 2026-03-06T13:38:47.154 INFO:teuthology.orchestra.run.vm00.stdout: python3-rbd x86_64 2:19.2.3-47.gc24117fd552.el9.clyso ceph 303 k 2026-03-06T13:38:47.154 INFO:teuthology.orchestra.run.vm00.stdout: python3-rgw x86_64 2:19.2.3-47.gc24117fd552.el9.clyso ceph 100 k 2026-03-06T13:38:47.154 INFO:teuthology.orchestra.run.vm00.stdout: python3-xmltodict noarch 0.12.0-15.el9 epel 22 k 2026-03-06T13:38:47.154 INFO:teuthology.orchestra.run.vm00.stdout: rbd-fuse x86_64 2:19.2.3-47.gc24117fd552.el9.clyso ceph 85 k 2026-03-06T13:38:47.154 INFO:teuthology.orchestra.run.vm00.stdout: rbd-mirror x86_64 2:19.2.3-47.gc24117fd552.el9.clyso ceph 3.1 M 2026-03-06T13:38:47.154 INFO:teuthology.orchestra.run.vm00.stdout: rbd-nbd x86_64 2:19.2.3-47.gc24117fd552.el9.clyso ceph 171 k 2026-03-06T13:38:47.154 INFO:teuthology.orchestra.run.vm00.stdout: s3cmd noarch 2.4.0-1.el9 epel 206 k 2026-03-06T13:38:47.154 INFO:teuthology.orchestra.run.vm00.stdout:Upgrading: 2026-03-06T13:38:47.154 INFO:teuthology.orchestra.run.vm00.stdout: librados2 x86_64 2:19.2.3-47.gc24117fd552.el9.clyso ceph 3.4 M 2026-03-06T13:38:47.154 INFO:teuthology.orchestra.run.vm00.stdout: librbd1 x86_64 2:19.2.3-47.gc24117fd552.el9.clyso ceph 3.2 M 2026-03-06T13:38:47.154 INFO:teuthology.orchestra.run.vm00.stdout:Installing dependencies: 2026-03-06T13:38:47.154 INFO:teuthology.orchestra.run.vm00.stdout: abseil-cpp x86_64 20211102.0-4.el9 epel 551 k 2026-03-06T13:38:47.154 INFO:teuthology.orchestra.run.vm00.stdout: boost-program-options x86_64 1.75.0-13.el9 appstream 104 k 2026-03-06T13:38:47.154 INFO:teuthology.orchestra.run.vm00.stdout: ceph-common x86_64 2:19.2.3-47.gc24117fd552.el9.clyso ceph 22 M 2026-03-06T13:38:47.154 INFO:teuthology.orchestra.run.vm00.stdout: ceph-grafana-dashboards noarch 2:19.2.3-47.gc24117fd552.el9.clyso ceph-noarch 31 k 2026-03-06T13:38:47.154 INFO:teuthology.orchestra.run.vm00.stdout: ceph-mds x86_64 2:19.2.3-47.gc24117fd552.el9.clyso ceph 2.4 M 2026-03-06T13:38:47.154 INFO:teuthology.orchestra.run.vm00.stdout: ceph-mgr-modules-core noarch 2:19.2.3-47.gc24117fd552.el9.clyso ceph-noarch 252 k 2026-03-06T13:38:47.154 INFO:teuthology.orchestra.run.vm00.stdout: ceph-mon x86_64 2:19.2.3-47.gc24117fd552.el9.clyso ceph 4.7 M 2026-03-06T13:38:47.154 INFO:teuthology.orchestra.run.vm00.stdout: ceph-osd x86_64 2:19.2.3-47.gc24117fd552.el9.clyso ceph 17 M 2026-03-06T13:38:47.154 INFO:teuthology.orchestra.run.vm00.stdout: ceph-prometheus-alerts noarch 2:19.2.3-47.gc24117fd552.el9.clyso ceph-noarch 16 k 2026-03-06T13:38:47.154 INFO:teuthology.orchestra.run.vm00.stdout: ceph-selinux x86_64 2:19.2.3-47.gc24117fd552.el9.clyso ceph 25 k 2026-03-06T13:38:47.154 INFO:teuthology.orchestra.run.vm00.stdout: cryptsetup x86_64 2.8.1-3.el9 baseos 351 k 2026-03-06T13:38:47.154 INFO:teuthology.orchestra.run.vm00.stdout: flexiblas x86_64 3.0.4-9.el9 appstream 30 k 2026-03-06T13:38:47.154 INFO:teuthology.orchestra.run.vm00.stdout: flexiblas-netlib x86_64 3.0.4-9.el9 appstream 3.0 M 2026-03-06T13:38:47.154 INFO:teuthology.orchestra.run.vm00.stdout: flexiblas-openblas-openmp x86_64 3.0.4-9.el9 appstream 15 k 2026-03-06T13:38:47.154 INFO:teuthology.orchestra.run.vm00.stdout: fuse x86_64 2.9.9-17.el9 baseos 80 k 2026-03-06T13:38:47.154 INFO:teuthology.orchestra.run.vm00.stdout: gperftools-libs x86_64 2.9.1-3.el9 epel 308 k 2026-03-06T13:38:47.155 INFO:teuthology.orchestra.run.vm00.stdout: grpc-data noarch 1.46.7-10.el9 epel 19 k 2026-03-06T13:38:47.155 INFO:teuthology.orchestra.run.vm00.stdout: ledmon-libs x86_64 1.1.0-3.el9 baseos 40 k 2026-03-06T13:38:47.155 INFO:teuthology.orchestra.run.vm00.stdout: libarrow x86_64 9.0.0-15.el9 epel 4.4 M 2026-03-06T13:38:47.155 INFO:teuthology.orchestra.run.vm00.stdout: libarrow-doc noarch 9.0.0-15.el9 epel 25 k 2026-03-06T13:38:47.155 INFO:teuthology.orchestra.run.vm00.stdout: libcephsqlite x86_64 2:19.2.3-47.gc24117fd552.el9.clyso ceph 163 k 2026-03-06T13:38:47.155 INFO:teuthology.orchestra.run.vm00.stdout: libconfig x86_64 1.7.2-9.el9 baseos 72 k 2026-03-06T13:38:47.155 INFO:teuthology.orchestra.run.vm00.stdout: libgfortran x86_64 11.5.0-14.el9 baseos 794 k 2026-03-06T13:38:47.155 INFO:teuthology.orchestra.run.vm00.stdout: libnbd x86_64 1.20.3-4.el9 appstream 164 k 2026-03-06T13:38:47.155 INFO:teuthology.orchestra.run.vm00.stdout: liboath x86_64 2.6.12-1.el9 epel 49 k 2026-03-06T13:38:47.155 INFO:teuthology.orchestra.run.vm00.stdout: libpmemobj x86_64 1.12.1-1.el9 appstream 160 k 2026-03-06T13:38:47.155 INFO:teuthology.orchestra.run.vm00.stdout: libquadmath x86_64 11.5.0-14.el9 baseos 184 k 2026-03-06T13:38:47.155 INFO:teuthology.orchestra.run.vm00.stdout: librabbitmq x86_64 0.11.0-7.el9 appstream 45 k 2026-03-06T13:38:47.155 INFO:teuthology.orchestra.run.vm00.stdout: libradosstriper1 x86_64 2:19.2.3-47.gc24117fd552.el9.clyso ceph 503 k 2026-03-06T13:38:47.155 INFO:teuthology.orchestra.run.vm00.stdout: librdkafka x86_64 1.6.1-102.el9 appstream 662 k 2026-03-06T13:38:47.155 INFO:teuthology.orchestra.run.vm00.stdout: librgw2 x86_64 2:19.2.3-47.gc24117fd552.el9.clyso ceph 5.4 M 2026-03-06T13:38:47.155 INFO:teuthology.orchestra.run.vm00.stdout: libstoragemgmt x86_64 1.10.1-1.el9 appstream 246 k 2026-03-06T13:38:47.155 INFO:teuthology.orchestra.run.vm00.stdout: libunwind x86_64 1.6.2-1.el9 epel 67 k 2026-03-06T13:38:47.155 INFO:teuthology.orchestra.run.vm00.stdout: libxslt x86_64 1.1.34-12.el9 appstream 233 k 2026-03-06T13:38:47.155 INFO:teuthology.orchestra.run.vm00.stdout: lttng-ust x86_64 2.12.0-6.el9 appstream 292 k 2026-03-06T13:38:47.155 INFO:teuthology.orchestra.run.vm00.stdout: lua x86_64 5.4.4-4.el9 appstream 188 k 2026-03-06T13:38:47.155 INFO:teuthology.orchestra.run.vm00.stdout: lua-devel x86_64 5.4.4-4.el9 crb 22 k 2026-03-06T13:38:47.155 INFO:teuthology.orchestra.run.vm00.stdout: luarocks noarch 3.9.2-5.el9 epel 151 k 2026-03-06T13:38:47.155 INFO:teuthology.orchestra.run.vm00.stdout: mailcap noarch 2.1.49-5.el9 baseos 33 k 2026-03-06T13:38:47.155 INFO:teuthology.orchestra.run.vm00.stdout: openblas x86_64 0.3.29-1.el9 appstream 42 k 2026-03-06T13:38:47.155 INFO:teuthology.orchestra.run.vm00.stdout: openblas-openmp x86_64 0.3.29-1.el9 appstream 5.3 M 2026-03-06T13:38:47.155 INFO:teuthology.orchestra.run.vm00.stdout: parquet-libs x86_64 9.0.0-15.el9 epel 838 k 2026-03-06T13:38:47.155 INFO:teuthology.orchestra.run.vm00.stdout: pciutils x86_64 3.7.0-7.el9 baseos 93 k 2026-03-06T13:38:47.155 INFO:teuthology.orchestra.run.vm00.stdout: perl-Benchmark noarch 1.23-483.el9 appstream 26 k 2026-03-06T13:38:47.155 INFO:teuthology.orchestra.run.vm00.stdout: protobuf x86_64 3.14.0-17.el9 appstream 1.0 M 2026-03-06T13:38:47.155 INFO:teuthology.orchestra.run.vm00.stdout: protobuf-compiler x86_64 3.14.0-17.el9 crb 862 k 2026-03-06T13:38:47.155 INFO:teuthology.orchestra.run.vm00.stdout: python3-asyncssh noarch 2.13.2-5.el9 epel 548 k 2026-03-06T13:38:47.155 INFO:teuthology.orchestra.run.vm00.stdout: python3-autocommand noarch 2.2.2-8.el9 epel 29 k 2026-03-06T13:38:47.155 INFO:teuthology.orchestra.run.vm00.stdout: python3-babel noarch 2.9.1-2.el9 appstream 6.0 M 2026-03-06T13:38:47.155 INFO:teuthology.orchestra.run.vm00.stdout: python3-backports-tarfile noarch 1.2.0-1.el9 epel 60 k 2026-03-06T13:38:47.155 INFO:teuthology.orchestra.run.vm00.stdout: python3-bcrypt x86_64 3.2.2-1.el9 epel 43 k 2026-03-06T13:38:47.155 INFO:teuthology.orchestra.run.vm00.stdout: python3-cachetools noarch 4.2.4-1.el9 epel 32 k 2026-03-06T13:38:47.155 INFO:teuthology.orchestra.run.vm00.stdout: python3-ceph-argparse x86_64 2:19.2.3-47.gc24117fd552.el9.clyso ceph 45 k 2026-03-06T13:38:47.155 INFO:teuthology.orchestra.run.vm00.stdout: python3-ceph-common x86_64 2:19.2.3-47.gc24117fd552.el9.clyso ceph 142 k 2026-03-06T13:38:47.155 INFO:teuthology.orchestra.run.vm00.stdout: python3-certifi noarch 2023.05.07-4.el9 epel 14 k 2026-03-06T13:38:47.155 INFO:teuthology.orchestra.run.vm00.stdout: python3-cffi x86_64 1.14.5-5.el9 baseos 253 k 2026-03-06T13:38:47.155 INFO:teuthology.orchestra.run.vm00.stdout: python3-cheroot noarch 10.0.1-4.el9 epel 173 k 2026-03-06T13:38:47.155 INFO:teuthology.orchestra.run.vm00.stdout: python3-cherrypy noarch 18.6.1-2.el9 epel 358 k 2026-03-06T13:38:47.155 INFO:teuthology.orchestra.run.vm00.stdout: python3-cryptography x86_64 36.0.1-5.el9 baseos 1.2 M 2026-03-06T13:38:47.155 INFO:teuthology.orchestra.run.vm00.stdout: python3-devel x86_64 3.9.25-3.el9 appstream 244 k 2026-03-06T13:38:47.155 INFO:teuthology.orchestra.run.vm00.stdout: python3-google-auth noarch 1:2.45.0-1.el9 epel 254 k 2026-03-06T13:38:47.155 INFO:teuthology.orchestra.run.vm00.stdout: python3-grpcio x86_64 1.46.7-10.el9 epel 2.0 M 2026-03-06T13:38:47.155 INFO:teuthology.orchestra.run.vm00.stdout: python3-grpcio-tools x86_64 1.46.7-10.el9 epel 144 k 2026-03-06T13:38:47.155 INFO:teuthology.orchestra.run.vm00.stdout: python3-isodate noarch 0.6.1-3.el9 epel 56 k 2026-03-06T13:38:47.155 INFO:teuthology.orchestra.run.vm00.stdout: python3-jaraco noarch 8.2.1-3.el9 epel 11 k 2026-03-06T13:38:47.155 INFO:teuthology.orchestra.run.vm00.stdout: python3-jaraco-classes noarch 3.2.1-5.el9 epel 18 k 2026-03-06T13:38:47.155 INFO:teuthology.orchestra.run.vm00.stdout: python3-jaraco-collections noarch 3.0.0-8.el9 epel 23 k 2026-03-06T13:38:47.155 INFO:teuthology.orchestra.run.vm00.stdout: python3-jaraco-context noarch 6.0.1-3.el9 epel 20 k 2026-03-06T13:38:47.155 INFO:teuthology.orchestra.run.vm00.stdout: python3-jaraco-functools noarch 3.5.0-2.el9 epel 19 k 2026-03-06T13:38:47.155 INFO:teuthology.orchestra.run.vm00.stdout: python3-jaraco-text noarch 4.0.0-2.el9 epel 26 k 2026-03-06T13:38:47.155 INFO:teuthology.orchestra.run.vm00.stdout: python3-jinja2 noarch 2.11.3-8.el9 appstream 249 k 2026-03-06T13:38:47.155 INFO:teuthology.orchestra.run.vm00.stdout: python3-kubernetes noarch 1:26.1.0-3.el9 epel 1.0 M 2026-03-06T13:38:47.155 INFO:teuthology.orchestra.run.vm00.stdout: python3-libstoragemgmt x86_64 1.10.1-1.el9 appstream 177 k 2026-03-06T13:38:47.155 INFO:teuthology.orchestra.run.vm00.stdout: python3-logutils noarch 0.3.5-21.el9 epel 46 k 2026-03-06T13:38:47.155 INFO:teuthology.orchestra.run.vm00.stdout: python3-lxml x86_64 4.6.5-3.el9 appstream 1.2 M 2026-03-06T13:38:47.155 INFO:teuthology.orchestra.run.vm00.stdout: python3-mako noarch 1.1.4-6.el9 appstream 172 k 2026-03-06T13:38:47.155 INFO:teuthology.orchestra.run.vm00.stdout: python3-markupsafe x86_64 1.1.1-12.el9 appstream 35 k 2026-03-06T13:38:47.155 INFO:teuthology.orchestra.run.vm00.stdout: python3-more-itertools noarch 8.12.0-2.el9 epel 79 k 2026-03-06T13:38:47.156 INFO:teuthology.orchestra.run.vm00.stdout: python3-msgpack x86_64 1.0.3-2.el9 epel 86 k 2026-03-06T13:38:47.156 INFO:teuthology.orchestra.run.vm00.stdout: python3-natsort noarch 7.1.1-5.el9 epel 58 k 2026-03-06T13:38:47.156 INFO:teuthology.orchestra.run.vm00.stdout: python3-numpy x86_64 1:1.23.5-2.el9 appstream 6.1 M 2026-03-06T13:38:47.156 INFO:teuthology.orchestra.run.vm00.stdout: python3-numpy-f2py x86_64 1:1.23.5-2.el9 appstream 442 k 2026-03-06T13:38:47.156 INFO:teuthology.orchestra.run.vm00.stdout: python3-packaging noarch 20.9-5.el9 appstream 77 k 2026-03-06T13:38:47.156 INFO:teuthology.orchestra.run.vm00.stdout: python3-pecan noarch 1.4.2-3.el9 epel 272 k 2026-03-06T13:38:47.156 INFO:teuthology.orchestra.run.vm00.stdout: python3-ply noarch 3.11-14.el9 baseos 106 k 2026-03-06T13:38:47.156 INFO:teuthology.orchestra.run.vm00.stdout: python3-portend noarch 3.1.0-2.el9 epel 16 k 2026-03-06T13:38:47.156 INFO:teuthology.orchestra.run.vm00.stdout: python3-protobuf noarch 3.14.0-17.el9 appstream 267 k 2026-03-06T13:38:47.156 INFO:teuthology.orchestra.run.vm00.stdout: python3-pyOpenSSL noarch 21.0.0-1.el9 epel 90 k 2026-03-06T13:38:47.156 INFO:teuthology.orchestra.run.vm00.stdout: python3-pyasn1 noarch 0.4.8-7.el9 appstream 157 k 2026-03-06T13:38:47.156 INFO:teuthology.orchestra.run.vm00.stdout: python3-pyasn1-modules noarch 0.4.8-7.el9 appstream 277 k 2026-03-06T13:38:47.156 INFO:teuthology.orchestra.run.vm00.stdout: python3-pycparser noarch 2.20-6.el9 baseos 135 k 2026-03-06T13:38:47.156 INFO:teuthology.orchestra.run.vm00.stdout: python3-repoze-lru noarch 0.7-16.el9 epel 31 k 2026-03-06T13:38:47.156 INFO:teuthology.orchestra.run.vm00.stdout: python3-requests noarch 2.25.1-10.el9 baseos 126 k 2026-03-06T13:38:47.156 INFO:teuthology.orchestra.run.vm00.stdout: python3-requests-oauthlib noarch 1.3.0-12.el9 appstream 54 k 2026-03-06T13:38:47.156 INFO:teuthology.orchestra.run.vm00.stdout: python3-routes noarch 2.5.1-5.el9 epel 188 k 2026-03-06T13:38:47.156 INFO:teuthology.orchestra.run.vm00.stdout: python3-rsa noarch 4.9-2.el9 epel 59 k 2026-03-06T13:38:47.156 INFO:teuthology.orchestra.run.vm00.stdout: python3-scipy x86_64 1.9.3-2.el9 appstream 19 M 2026-03-06T13:38:47.156 INFO:teuthology.orchestra.run.vm00.stdout: python3-tempora noarch 5.0.0-2.el9 epel 36 k 2026-03-06T13:38:47.156 INFO:teuthology.orchestra.run.vm00.stdout: python3-toml noarch 0.10.2-6.el9 appstream 42 k 2026-03-06T13:38:47.156 INFO:teuthology.orchestra.run.vm00.stdout: python3-typing-extensions noarch 4.15.0-1.el9 epel 86 k 2026-03-06T13:38:47.156 INFO:teuthology.orchestra.run.vm00.stdout: python3-urllib3 noarch 1.26.5-7.el9 baseos 218 k 2026-03-06T13:38:47.156 INFO:teuthology.orchestra.run.vm00.stdout: python3-webob noarch 1.8.8-2.el9 epel 230 k 2026-03-06T13:38:47.156 INFO:teuthology.orchestra.run.vm00.stdout: python3-websocket-client noarch 1.2.3-2.el9 epel 90 k 2026-03-06T13:38:47.156 INFO:teuthology.orchestra.run.vm00.stdout: python3-werkzeug noarch 2.0.3-3.el9.1 epel 427 k 2026-03-06T13:38:47.156 INFO:teuthology.orchestra.run.vm00.stdout: python3-xmlsec x86_64 1.3.13-1.el9 epel 48 k 2026-03-06T13:38:47.156 INFO:teuthology.orchestra.run.vm00.stdout: python3-zc-lockfile noarch 2.0-10.el9 epel 20 k 2026-03-06T13:38:47.156 INFO:teuthology.orchestra.run.vm00.stdout: qatlib x86_64 25.08.0-2.el9 appstream 240 k 2026-03-06T13:38:47.156 INFO:teuthology.orchestra.run.vm00.stdout: qatzip-libs x86_64 1.3.1-1.el9 appstream 66 k 2026-03-06T13:38:47.156 INFO:teuthology.orchestra.run.vm00.stdout: re2 x86_64 1:20211101-20.el9 epel 191 k 2026-03-06T13:38:47.156 INFO:teuthology.orchestra.run.vm00.stdout: socat x86_64 1.7.4.1-8.el9 appstream 303 k 2026-03-06T13:38:47.156 INFO:teuthology.orchestra.run.vm00.stdout: thrift x86_64 0.15.0-4.el9 epel 1.6 M 2026-03-06T13:38:47.156 INFO:teuthology.orchestra.run.vm00.stdout: unzip x86_64 6.0-59.el9 baseos 182 k 2026-03-06T13:38:47.156 INFO:teuthology.orchestra.run.vm00.stdout: xmlsec1 x86_64 1.2.29-13.el9 appstream 189 k 2026-03-06T13:38:47.156 INFO:teuthology.orchestra.run.vm00.stdout: xmlsec1-openssl x86_64 1.2.29-13.el9 appstream 90 k 2026-03-06T13:38:47.156 INFO:teuthology.orchestra.run.vm00.stdout: xmlstarlet x86_64 1.6.1-20.el9 appstream 64 k 2026-03-06T13:38:47.156 INFO:teuthology.orchestra.run.vm00.stdout: zip x86_64 3.0-35.el9 baseos 266 k 2026-03-06T13:38:47.156 INFO:teuthology.orchestra.run.vm00.stdout:Installing weak dependencies: 2026-03-06T13:38:47.156 INFO:teuthology.orchestra.run.vm00.stdout: ceph-mgr-k8sevents noarch 2:19.2.3-47.gc24117fd552.el9.clyso ceph-noarch 22 k 2026-03-06T13:38:47.156 INFO:teuthology.orchestra.run.vm00.stdout: python3-influxdb noarch 5.3.1-1.el9 epel 139 k 2026-03-06T13:38:47.156 INFO:teuthology.orchestra.run.vm00.stdout: python3-saml noarch 1.16.0-1.el9 epel 125 k 2026-03-06T13:38:47.156 INFO:teuthology.orchestra.run.vm00.stdout: qatlib-service x86_64 25.08.0-2.el9 appstream 37 k 2026-03-06T13:38:47.156 INFO:teuthology.orchestra.run.vm00.stdout: smartmontools x86_64 1:7.2-10.el9 baseos 556 k 2026-03-06T13:38:47.156 INFO:teuthology.orchestra.run.vm00.stdout: 2026-03-06T13:38:47.156 INFO:teuthology.orchestra.run.vm00.stdout:Transaction Summary 2026-03-06T13:38:47.156 INFO:teuthology.orchestra.run.vm00.stdout:============================================================================================== 2026-03-06T13:38:47.156 INFO:teuthology.orchestra.run.vm00.stdout:Install 148 Packages 2026-03-06T13:38:47.156 INFO:teuthology.orchestra.run.vm00.stdout:Upgrade 2 Packages 2026-03-06T13:38:47.156 INFO:teuthology.orchestra.run.vm00.stdout: 2026-03-06T13:38:47.156 INFO:teuthology.orchestra.run.vm00.stdout:Total download size: 214 M 2026-03-06T13:38:47.156 INFO:teuthology.orchestra.run.vm00.stdout:Downloading Packages: 2026-03-06T13:38:47.266 INFO:teuthology.orchestra.run.vm08.stdout:(1/150): ceph-19.2.3-47.gc24117fd552.el9.clyso. 79 kB/s | 6.5 kB 00:00 2026-03-06T13:38:47.524 INFO:teuthology.orchestra.run.vm08.stdout:(2/150): ceph-fuse-19.2.3-47.gc24117fd552.el9.c 4.5 MB/s | 1.1 MB 00:00 2026-03-06T13:38:47.563 INFO:teuthology.orchestra.run.vm08.stdout:(3/150): ceph-immutable-object-cache-19.2.3-47. 3.7 MB/s | 145 kB 00:00 2026-03-06T13:38:47.682 INFO:teuthology.orchestra.run.vm08.stdout:(4/150): ceph-base-19.2.3-47.gc24117fd552.el9.c 11 MB/s | 5.5 MB 00:00 2026-03-06T13:38:47.725 INFO:teuthology.orchestra.run.vm08.stdout:(5/150): ceph-mds-19.2.3-47.gc24117fd552.el9.cl 15 MB/s | 2.4 MB 00:00 2026-03-06T13:38:47.781 INFO:teuthology.orchestra.run.vm08.stdout:(6/150): ceph-mgr-19.2.3-47.gc24117fd552.el9.cl 11 MB/s | 1.1 MB 00:00 2026-03-06T13:38:47.962 INFO:teuthology.orchestra.run.vm08.stdout:(7/150): ceph-mon-19.2.3-47.gc24117fd552.el9.cl 20 MB/s | 4.7 MB 00:00 2026-03-06T13:38:48.106 INFO:teuthology.orchestra.run.vm08.stdout:(8/150): ceph-common-19.2.3-47.gc24117fd552.el9 24 MB/s | 22 MB 00:00 2026-03-06T13:38:48.137 INFO:teuthology.orchestra.run.vm08.stdout:(9/150): ceph-selinux-19.2.3-47.gc24117fd552.el 812 kB/s | 25 kB 00:00 2026-03-06T13:38:48.163 INFO:teuthology.orchestra.run.vm01.stdout:lab-extras 65 kB/s | 50 kB 00:00 2026-03-06T13:38:48.426 INFO:teuthology.orchestra.run.vm08.stdout:(10/150): ceph-radosgw-19.2.3-47.gc24117fd552.e 23 MB/s | 11 MB 00:00 2026-03-06T13:38:48.486 INFO:teuthology.orchestra.run.vm08.stdout:(11/150): libcephfs-devel-19.2.3-47.gc24117fd55 563 kB/s | 34 kB 00:00 2026-03-06T13:38:48.495 INFO:teuthology.orchestra.run.vm00.stdout:(1/150): ceph-19.2.3-47.gc24117fd552.el9.clyso. 65 kB/s | 6.5 kB 00:00 2026-03-06T13:38:48.663 INFO:teuthology.orchestra.run.vm08.stdout:(12/150): ceph-osd-19.2.3-47.gc24117fd552.el9.c 19 MB/s | 17 MB 00:00 2026-03-06T13:38:48.670 INFO:teuthology.orchestra.run.vm08.stdout:(13/150): libcephfs2-19.2.3-47.gc24117fd552.el9 5.3 MB/s | 998 kB 00:00 2026-03-06T13:38:48.718 INFO:teuthology.orchestra.run.vm08.stdout:(14/150): libcephsqlite-19.2.3-47.gc24117fd552. 2.8 MB/s | 163 kB 00:00 2026-03-06T13:38:48.795 INFO:teuthology.orchestra.run.vm00.stdout:(2/150): ceph-fuse-19.2.3-47.gc24117fd552.el9.c 3.8 MB/s | 1.1 MB 00:00 2026-03-06T13:38:48.813 INFO:teuthology.orchestra.run.vm08.stdout:(15/150): libradosstriper1-19.2.3-47.gc24117fd5 5.2 MB/s | 503 kB 00:00 2026-03-06T13:38:48.868 INFO:teuthology.orchestra.run.vm00.stdout:(3/150): ceph-immutable-object-cache-19.2.3-47. 1.9 MB/s | 145 kB 00:00 2026-03-06T13:38:49.007 INFO:teuthology.orchestra.run.vm08.stdout:(16/150): librados-devel-19.2.3-47.gc24117fd552 376 kB/s | 127 kB 00:00 2026-03-06T13:38:49.046 INFO:teuthology.orchestra.run.vm08.stdout:(17/150): python3-ceph-argparse-19.2.3-47.gc241 1.1 MB/s | 45 kB 00:00 2026-03-06T13:38:49.113 INFO:teuthology.orchestra.run.vm08.stdout:(18/150): python3-ceph-common-19.2.3-47.gc24117 2.1 MB/s | 142 kB 00:00 2026-03-06T13:38:49.136 INFO:teuthology.orchestra.run.vm00.stdout:(4/150): ceph-base-19.2.3-47.gc24117fd552.el9.c 7.4 MB/s | 5.5 MB 00:00 2026-03-06T13:38:49.177 INFO:teuthology.orchestra.run.vm08.stdout:(19/150): python3-cephfs-19.2.3-47.gc24117fd552 2.5 MB/s | 165 kB 00:00 2026-03-06T13:38:49.225 INFO:teuthology.orchestra.run.vm00.stdout:(5/150): ceph-mds-19.2.3-47.gc24117fd552.el9.cl 6.8 MB/s | 2.4 MB 00:00 2026-03-06T13:38:49.246 INFO:teuthology.orchestra.run.vm08.stdout:(20/150): librgw2-19.2.3-47.gc24117fd552.el9.cl 13 MB/s | 5.4 MB 00:00 2026-03-06T13:38:49.250 INFO:teuthology.orchestra.run.vm08.stdout:(21/150): python3-rados-19.2.3-47.gc24117fd552. 4.4 MB/s | 322 kB 00:00 2026-03-06T13:38:49.265 INFO:teuthology.orchestra.run.vm00.stdout:(6/150): ceph-mgr-19.2.3-47.gc24117fd552.el9.cl 8.3 MB/s | 1.1 MB 00:00 2026-03-06T13:38:49.318 INFO:teuthology.orchestra.run.vm08.stdout:(22/150): python3-rgw-19.2.3-47.gc24117fd552.el 1.4 MB/s | 100 kB 00:00 2026-03-06T13:38:49.326 INFO:teuthology.orchestra.run.vm08.stdout:(23/150): python3-rbd-19.2.3-47.gc24117fd552.el 3.7 MB/s | 303 kB 00:00 2026-03-06T13:38:49.367 INFO:teuthology.orchestra.run.vm08.stdout:(24/150): rbd-fuse-19.2.3-47.gc24117fd552.el9.c 1.7 MB/s | 85 kB 00:00 2026-03-06T13:38:49.498 INFO:teuthology.orchestra.run.vm00.stdout:(7/150): ceph-mon-19.2.3-47.gc24117fd552.el9.cl 17 MB/s | 4.7 MB 00:00 2026-03-06T13:38:49.500 INFO:teuthology.orchestra.run.vm08.stdout:(25/150): ceph-test-19.2.3-47.gc24117fd552.el9. 36 MB/s | 50 MB 00:01 2026-03-06T13:38:49.501 INFO:teuthology.orchestra.run.vm08.stdout:(26/150): rbd-nbd-19.2.3-47.gc24117fd552.el9.cl 1.2 MB/s | 171 kB 00:00 2026-03-06T13:38:49.544 INFO:teuthology.orchestra.run.vm08.stdout:(27/150): ceph-grafana-dashboards-19.2.3-47.gc2 703 kB/s | 31 kB 00:00 2026-03-06T13:38:49.552 INFO:teuthology.orchestra.run.vm08.stdout:(28/150): ceph-mgr-cephadm-19.2.3-47.gc24117fd5 2.9 MB/s | 150 kB 00:00 2026-03-06T13:38:49.678 INFO:teuthology.orchestra.run.vm08.stdout:(29/150): rbd-mirror-19.2.3-47.gc24117fd552.el9 8.9 MB/s | 3.1 MB 00:00 2026-03-06T13:38:49.713 INFO:teuthology.orchestra.run.vm08.stdout:(30/150): ceph-mgr-k8sevents-19.2.3-47.gc24117f 634 kB/s | 22 kB 00:00 2026-03-06T13:38:49.754 INFO:teuthology.orchestra.run.vm08.stdout:(31/150): ceph-mgr-modules-core-19.2.3-47.gc241 6.0 MB/s | 252 kB 00:00 2026-03-06T13:38:49.789 INFO:teuthology.orchestra.run.vm08.stdout:(32/150): ceph-mgr-rook-19.2.3-47.gc24117fd552. 1.4 MB/s | 49 kB 00:00 2026-03-06T13:38:49.826 INFO:teuthology.orchestra.run.vm08.stdout:(33/150): ceph-prometheus-alerts-19.2.3-47.gc24 438 kB/s | 16 kB 00:00 2026-03-06T13:38:49.906 INFO:teuthology.orchestra.run.vm01.stdout:Package librados2-2:16.2.4-5.el9.x86_64 is already installed. 2026-03-06T13:38:49.906 INFO:teuthology.orchestra.run.vm01.stdout:Package librbd1-2:16.2.4-5.el9.x86_64 is already installed. 2026-03-06T13:38:49.911 INFO:teuthology.orchestra.run.vm08.stdout:(34/150): ceph-volume-19.2.3-47.gc24117fd552.el 3.5 MB/s | 299 kB 00:00 2026-03-06T13:38:49.946 INFO:teuthology.orchestra.run.vm08.stdout:(35/150): ceph-mgr-dashboard-19.2.3-47.gc24117f 9.5 MB/s | 3.8 MB 00:00 2026-03-06T13:38:49.952 INFO:teuthology.orchestra.run.vm01.stdout:Dependencies resolved. 2026-03-06T13:38:49.957 INFO:teuthology.orchestra.run.vm01.stdout:============================================================================================== 2026-03-06T13:38:49.957 INFO:teuthology.orchestra.run.vm01.stdout: Package Arch Version Repository Size 2026-03-06T13:38:49.957 INFO:teuthology.orchestra.run.vm01.stdout:============================================================================================== 2026-03-06T13:38:49.957 INFO:teuthology.orchestra.run.vm01.stdout:Installing: 2026-03-06T13:38:49.957 INFO:teuthology.orchestra.run.vm01.stdout: bzip2 x86_64 1.0.8-11.el9 baseos 55 k 2026-03-06T13:38:49.957 INFO:teuthology.orchestra.run.vm01.stdout: ceph x86_64 2:19.2.3-47.gc24117fd552.el9.clyso ceph 6.5 k 2026-03-06T13:38:49.957 INFO:teuthology.orchestra.run.vm01.stdout: ceph-base x86_64 2:19.2.3-47.gc24117fd552.el9.clyso ceph 5.5 M 2026-03-06T13:38:49.957 INFO:teuthology.orchestra.run.vm01.stdout: ceph-fuse x86_64 2:19.2.3-47.gc24117fd552.el9.clyso ceph 1.1 M 2026-03-06T13:38:49.957 INFO:teuthology.orchestra.run.vm01.stdout: ceph-immutable-object-cache x86_64 2:19.2.3-47.gc24117fd552.el9.clyso ceph 145 k 2026-03-06T13:38:49.957 INFO:teuthology.orchestra.run.vm01.stdout: ceph-mgr x86_64 2:19.2.3-47.gc24117fd552.el9.clyso ceph 1.1 M 2026-03-06T13:38:49.957 INFO:teuthology.orchestra.run.vm01.stdout: ceph-mgr-cephadm noarch 2:19.2.3-47.gc24117fd552.el9.clyso ceph-noarch 150 k 2026-03-06T13:38:49.957 INFO:teuthology.orchestra.run.vm01.stdout: ceph-mgr-dashboard noarch 2:19.2.3-47.gc24117fd552.el9.clyso ceph-noarch 3.8 M 2026-03-06T13:38:49.957 INFO:teuthology.orchestra.run.vm01.stdout: ceph-mgr-diskprediction-local noarch 2:19.2.3-47.gc24117fd552.el9.clyso ceph-noarch 7.4 M 2026-03-06T13:38:49.957 INFO:teuthology.orchestra.run.vm01.stdout: ceph-mgr-rook noarch 2:19.2.3-47.gc24117fd552.el9.clyso ceph-noarch 49 k 2026-03-06T13:38:49.957 INFO:teuthology.orchestra.run.vm01.stdout: ceph-radosgw x86_64 2:19.2.3-47.gc24117fd552.el9.clyso ceph 11 M 2026-03-06T13:38:49.957 INFO:teuthology.orchestra.run.vm01.stdout: ceph-test x86_64 2:19.2.3-47.gc24117fd552.el9.clyso ceph 50 M 2026-03-06T13:38:49.957 INFO:teuthology.orchestra.run.vm01.stdout: ceph-volume noarch 2:19.2.3-47.gc24117fd552.el9.clyso ceph-noarch 299 k 2026-03-06T13:38:49.957 INFO:teuthology.orchestra.run.vm01.stdout: cephadm noarch 2:19.2.3-47.gc24117fd552.el9.clyso ceph-noarch 769 k 2026-03-06T13:38:49.957 INFO:teuthology.orchestra.run.vm01.stdout: libcephfs-devel x86_64 2:19.2.3-47.gc24117fd552.el9.clyso ceph 34 k 2026-03-06T13:38:49.957 INFO:teuthology.orchestra.run.vm01.stdout: libcephfs2 x86_64 2:19.2.3-47.gc24117fd552.el9.clyso ceph 998 k 2026-03-06T13:38:49.957 INFO:teuthology.orchestra.run.vm01.stdout: librados-devel x86_64 2:19.2.3-47.gc24117fd552.el9.clyso ceph 127 k 2026-03-06T13:38:49.957 INFO:teuthology.orchestra.run.vm01.stdout: perl-Test-Harness noarch 1:3.42-461.el9 appstream 295 k 2026-03-06T13:38:49.957 INFO:teuthology.orchestra.run.vm01.stdout: python3-cephfs x86_64 2:19.2.3-47.gc24117fd552.el9.clyso ceph 165 k 2026-03-06T13:38:49.957 INFO:teuthology.orchestra.run.vm01.stdout: python3-rados x86_64 2:19.2.3-47.gc24117fd552.el9.clyso ceph 322 k 2026-03-06T13:38:49.957 INFO:teuthology.orchestra.run.vm01.stdout: python3-rbd x86_64 2:19.2.3-47.gc24117fd552.el9.clyso ceph 303 k 2026-03-06T13:38:49.957 INFO:teuthology.orchestra.run.vm01.stdout: python3-rgw x86_64 2:19.2.3-47.gc24117fd552.el9.clyso ceph 100 k 2026-03-06T13:38:49.957 INFO:teuthology.orchestra.run.vm01.stdout: python3-xmltodict noarch 0.12.0-15.el9 epel 22 k 2026-03-06T13:38:49.957 INFO:teuthology.orchestra.run.vm01.stdout: rbd-fuse x86_64 2:19.2.3-47.gc24117fd552.el9.clyso ceph 85 k 2026-03-06T13:38:49.957 INFO:teuthology.orchestra.run.vm01.stdout: rbd-mirror x86_64 2:19.2.3-47.gc24117fd552.el9.clyso ceph 3.1 M 2026-03-06T13:38:49.957 INFO:teuthology.orchestra.run.vm01.stdout: rbd-nbd x86_64 2:19.2.3-47.gc24117fd552.el9.clyso ceph 171 k 2026-03-06T13:38:49.957 INFO:teuthology.orchestra.run.vm01.stdout: s3cmd noarch 2.4.0-1.el9 epel 206 k 2026-03-06T13:38:49.957 INFO:teuthology.orchestra.run.vm01.stdout:Upgrading: 2026-03-06T13:38:49.957 INFO:teuthology.orchestra.run.vm01.stdout: librados2 x86_64 2:19.2.3-47.gc24117fd552.el9.clyso ceph 3.4 M 2026-03-06T13:38:49.957 INFO:teuthology.orchestra.run.vm01.stdout: librbd1 x86_64 2:19.2.3-47.gc24117fd552.el9.clyso ceph 3.2 M 2026-03-06T13:38:49.958 INFO:teuthology.orchestra.run.vm01.stdout:Installing dependencies: 2026-03-06T13:38:49.958 INFO:teuthology.orchestra.run.vm01.stdout: abseil-cpp x86_64 20211102.0-4.el9 epel 551 k 2026-03-06T13:38:49.958 INFO:teuthology.orchestra.run.vm01.stdout: boost-program-options x86_64 1.75.0-13.el9 appstream 104 k 2026-03-06T13:38:49.958 INFO:teuthology.orchestra.run.vm01.stdout: ceph-common x86_64 2:19.2.3-47.gc24117fd552.el9.clyso ceph 22 M 2026-03-06T13:38:49.958 INFO:teuthology.orchestra.run.vm01.stdout: ceph-grafana-dashboards noarch 2:19.2.3-47.gc24117fd552.el9.clyso ceph-noarch 31 k 2026-03-06T13:38:49.958 INFO:teuthology.orchestra.run.vm01.stdout: ceph-mds x86_64 2:19.2.3-47.gc24117fd552.el9.clyso ceph 2.4 M 2026-03-06T13:38:49.958 INFO:teuthology.orchestra.run.vm01.stdout: ceph-mgr-modules-core noarch 2:19.2.3-47.gc24117fd552.el9.clyso ceph-noarch 252 k 2026-03-06T13:38:49.958 INFO:teuthology.orchestra.run.vm01.stdout: ceph-mon x86_64 2:19.2.3-47.gc24117fd552.el9.clyso ceph 4.7 M 2026-03-06T13:38:49.958 INFO:teuthology.orchestra.run.vm01.stdout: ceph-osd x86_64 2:19.2.3-47.gc24117fd552.el9.clyso ceph 17 M 2026-03-06T13:38:49.958 INFO:teuthology.orchestra.run.vm01.stdout: ceph-prometheus-alerts noarch 2:19.2.3-47.gc24117fd552.el9.clyso ceph-noarch 16 k 2026-03-06T13:38:49.958 INFO:teuthology.orchestra.run.vm01.stdout: ceph-selinux x86_64 2:19.2.3-47.gc24117fd552.el9.clyso ceph 25 k 2026-03-06T13:38:49.958 INFO:teuthology.orchestra.run.vm01.stdout: cryptsetup x86_64 2.8.1-3.el9 baseos 351 k 2026-03-06T13:38:49.958 INFO:teuthology.orchestra.run.vm01.stdout: flexiblas x86_64 3.0.4-9.el9 appstream 30 k 2026-03-06T13:38:49.958 INFO:teuthology.orchestra.run.vm01.stdout: flexiblas-netlib x86_64 3.0.4-9.el9 appstream 3.0 M 2026-03-06T13:38:49.958 INFO:teuthology.orchestra.run.vm01.stdout: flexiblas-openblas-openmp x86_64 3.0.4-9.el9 appstream 15 k 2026-03-06T13:38:49.958 INFO:teuthology.orchestra.run.vm01.stdout: fuse x86_64 2.9.9-17.el9 baseos 80 k 2026-03-06T13:38:49.958 INFO:teuthology.orchestra.run.vm01.stdout: gperftools-libs x86_64 2.9.1-3.el9 epel 308 k 2026-03-06T13:38:49.958 INFO:teuthology.orchestra.run.vm01.stdout: grpc-data noarch 1.46.7-10.el9 epel 19 k 2026-03-06T13:38:49.958 INFO:teuthology.orchestra.run.vm01.stdout: ledmon-libs x86_64 1.1.0-3.el9 baseos 40 k 2026-03-06T13:38:49.958 INFO:teuthology.orchestra.run.vm01.stdout: libarrow x86_64 9.0.0-15.el9 epel 4.4 M 2026-03-06T13:38:49.958 INFO:teuthology.orchestra.run.vm01.stdout: libarrow-doc noarch 9.0.0-15.el9 epel 25 k 2026-03-06T13:38:49.958 INFO:teuthology.orchestra.run.vm01.stdout: libcephsqlite x86_64 2:19.2.3-47.gc24117fd552.el9.clyso ceph 163 k 2026-03-06T13:38:49.958 INFO:teuthology.orchestra.run.vm01.stdout: libconfig x86_64 1.7.2-9.el9 baseos 72 k 2026-03-06T13:38:49.958 INFO:teuthology.orchestra.run.vm01.stdout: libgfortran x86_64 11.5.0-14.el9 baseos 794 k 2026-03-06T13:38:49.958 INFO:teuthology.orchestra.run.vm01.stdout: libnbd x86_64 1.20.3-4.el9 appstream 164 k 2026-03-06T13:38:49.958 INFO:teuthology.orchestra.run.vm01.stdout: liboath x86_64 2.6.12-1.el9 epel 49 k 2026-03-06T13:38:49.958 INFO:teuthology.orchestra.run.vm01.stdout: libpmemobj x86_64 1.12.1-1.el9 appstream 160 k 2026-03-06T13:38:49.958 INFO:teuthology.orchestra.run.vm01.stdout: libquadmath x86_64 11.5.0-14.el9 baseos 184 k 2026-03-06T13:38:49.958 INFO:teuthology.orchestra.run.vm01.stdout: librabbitmq x86_64 0.11.0-7.el9 appstream 45 k 2026-03-06T13:38:49.958 INFO:teuthology.orchestra.run.vm01.stdout: libradosstriper1 x86_64 2:19.2.3-47.gc24117fd552.el9.clyso ceph 503 k 2026-03-06T13:38:49.958 INFO:teuthology.orchestra.run.vm01.stdout: librdkafka x86_64 1.6.1-102.el9 appstream 662 k 2026-03-06T13:38:49.958 INFO:teuthology.orchestra.run.vm01.stdout: librgw2 x86_64 2:19.2.3-47.gc24117fd552.el9.clyso ceph 5.4 M 2026-03-06T13:38:49.958 INFO:teuthology.orchestra.run.vm01.stdout: libstoragemgmt x86_64 1.10.1-1.el9 appstream 246 k 2026-03-06T13:38:49.958 INFO:teuthology.orchestra.run.vm01.stdout: libunwind x86_64 1.6.2-1.el9 epel 67 k 2026-03-06T13:38:49.958 INFO:teuthology.orchestra.run.vm01.stdout: libxslt x86_64 1.1.34-12.el9 appstream 233 k 2026-03-06T13:38:49.958 INFO:teuthology.orchestra.run.vm01.stdout: lttng-ust x86_64 2.12.0-6.el9 appstream 292 k 2026-03-06T13:38:49.958 INFO:teuthology.orchestra.run.vm01.stdout: lua x86_64 5.4.4-4.el9 appstream 188 k 2026-03-06T13:38:49.958 INFO:teuthology.orchestra.run.vm01.stdout: lua-devel x86_64 5.4.4-4.el9 crb 22 k 2026-03-06T13:38:49.958 INFO:teuthology.orchestra.run.vm01.stdout: luarocks noarch 3.9.2-5.el9 epel 151 k 2026-03-06T13:38:49.958 INFO:teuthology.orchestra.run.vm01.stdout: mailcap noarch 2.1.49-5.el9 baseos 33 k 2026-03-06T13:38:49.958 INFO:teuthology.orchestra.run.vm01.stdout: openblas x86_64 0.3.29-1.el9 appstream 42 k 2026-03-06T13:38:49.958 INFO:teuthology.orchestra.run.vm01.stdout: openblas-openmp x86_64 0.3.29-1.el9 appstream 5.3 M 2026-03-06T13:38:49.958 INFO:teuthology.orchestra.run.vm01.stdout: parquet-libs x86_64 9.0.0-15.el9 epel 838 k 2026-03-06T13:38:49.959 INFO:teuthology.orchestra.run.vm01.stdout: pciutils x86_64 3.7.0-7.el9 baseos 93 k 2026-03-06T13:38:49.959 INFO:teuthology.orchestra.run.vm01.stdout: perl-Benchmark noarch 1.23-483.el9 appstream 26 k 2026-03-06T13:38:49.959 INFO:teuthology.orchestra.run.vm01.stdout: protobuf x86_64 3.14.0-17.el9 appstream 1.0 M 2026-03-06T13:38:49.959 INFO:teuthology.orchestra.run.vm01.stdout: protobuf-compiler x86_64 3.14.0-17.el9 crb 862 k 2026-03-06T13:38:49.959 INFO:teuthology.orchestra.run.vm01.stdout: python3-asyncssh noarch 2.13.2-5.el9 epel 548 k 2026-03-06T13:38:49.959 INFO:teuthology.orchestra.run.vm01.stdout: python3-autocommand noarch 2.2.2-8.el9 epel 29 k 2026-03-06T13:38:49.959 INFO:teuthology.orchestra.run.vm01.stdout: python3-babel noarch 2.9.1-2.el9 appstream 6.0 M 2026-03-06T13:38:49.959 INFO:teuthology.orchestra.run.vm01.stdout: python3-backports-tarfile noarch 1.2.0-1.el9 epel 60 k 2026-03-06T13:38:49.959 INFO:teuthology.orchestra.run.vm01.stdout: python3-bcrypt x86_64 3.2.2-1.el9 epel 43 k 2026-03-06T13:38:49.959 INFO:teuthology.orchestra.run.vm01.stdout: python3-cachetools noarch 4.2.4-1.el9 epel 32 k 2026-03-06T13:38:49.959 INFO:teuthology.orchestra.run.vm01.stdout: python3-ceph-argparse x86_64 2:19.2.3-47.gc24117fd552.el9.clyso ceph 45 k 2026-03-06T13:38:49.959 INFO:teuthology.orchestra.run.vm01.stdout: python3-ceph-common x86_64 2:19.2.3-47.gc24117fd552.el9.clyso ceph 142 k 2026-03-06T13:38:49.959 INFO:teuthology.orchestra.run.vm01.stdout: python3-certifi noarch 2023.05.07-4.el9 epel 14 k 2026-03-06T13:38:49.959 INFO:teuthology.orchestra.run.vm01.stdout: python3-cffi x86_64 1.14.5-5.el9 baseos 253 k 2026-03-06T13:38:49.959 INFO:teuthology.orchestra.run.vm01.stdout: python3-cheroot noarch 10.0.1-4.el9 epel 173 k 2026-03-06T13:38:49.959 INFO:teuthology.orchestra.run.vm01.stdout: python3-cherrypy noarch 18.6.1-2.el9 epel 358 k 2026-03-06T13:38:49.959 INFO:teuthology.orchestra.run.vm01.stdout: python3-cryptography x86_64 36.0.1-5.el9 baseos 1.2 M 2026-03-06T13:38:49.959 INFO:teuthology.orchestra.run.vm01.stdout: python3-devel x86_64 3.9.25-3.el9 appstream 244 k 2026-03-06T13:38:49.959 INFO:teuthology.orchestra.run.vm01.stdout: python3-google-auth noarch 1:2.45.0-1.el9 epel 254 k 2026-03-06T13:38:49.959 INFO:teuthology.orchestra.run.vm01.stdout: python3-grpcio x86_64 1.46.7-10.el9 epel 2.0 M 2026-03-06T13:38:49.959 INFO:teuthology.orchestra.run.vm01.stdout: python3-grpcio-tools x86_64 1.46.7-10.el9 epel 144 k 2026-03-06T13:38:49.959 INFO:teuthology.orchestra.run.vm01.stdout: python3-isodate noarch 0.6.1-3.el9 epel 56 k 2026-03-06T13:38:49.959 INFO:teuthology.orchestra.run.vm01.stdout: python3-jaraco noarch 8.2.1-3.el9 epel 11 k 2026-03-06T13:38:49.959 INFO:teuthology.orchestra.run.vm01.stdout: python3-jaraco-classes noarch 3.2.1-5.el9 epel 18 k 2026-03-06T13:38:49.959 INFO:teuthology.orchestra.run.vm01.stdout: python3-jaraco-collections noarch 3.0.0-8.el9 epel 23 k 2026-03-06T13:38:49.959 INFO:teuthology.orchestra.run.vm01.stdout: python3-jaraco-context noarch 6.0.1-3.el9 epel 20 k 2026-03-06T13:38:49.959 INFO:teuthology.orchestra.run.vm01.stdout: python3-jaraco-functools noarch 3.5.0-2.el9 epel 19 k 2026-03-06T13:38:49.959 INFO:teuthology.orchestra.run.vm01.stdout: python3-jaraco-text noarch 4.0.0-2.el9 epel 26 k 2026-03-06T13:38:49.959 INFO:teuthology.orchestra.run.vm01.stdout: python3-jinja2 noarch 2.11.3-8.el9 appstream 249 k 2026-03-06T13:38:49.959 INFO:teuthology.orchestra.run.vm01.stdout: python3-kubernetes noarch 1:26.1.0-3.el9 epel 1.0 M 2026-03-06T13:38:49.959 INFO:teuthology.orchestra.run.vm01.stdout: python3-libstoragemgmt x86_64 1.10.1-1.el9 appstream 177 k 2026-03-06T13:38:49.959 INFO:teuthology.orchestra.run.vm01.stdout: python3-logutils noarch 0.3.5-21.el9 epel 46 k 2026-03-06T13:38:49.959 INFO:teuthology.orchestra.run.vm01.stdout: python3-lxml x86_64 4.6.5-3.el9 appstream 1.2 M 2026-03-06T13:38:49.959 INFO:teuthology.orchestra.run.vm01.stdout: python3-mako noarch 1.1.4-6.el9 appstream 172 k 2026-03-06T13:38:49.959 INFO:teuthology.orchestra.run.vm01.stdout: python3-markupsafe x86_64 1.1.1-12.el9 appstream 35 k 2026-03-06T13:38:49.959 INFO:teuthology.orchestra.run.vm01.stdout: python3-more-itertools noarch 8.12.0-2.el9 epel 79 k 2026-03-06T13:38:49.959 INFO:teuthology.orchestra.run.vm01.stdout: python3-msgpack x86_64 1.0.3-2.el9 epel 86 k 2026-03-06T13:38:49.959 INFO:teuthology.orchestra.run.vm01.stdout: python3-natsort noarch 7.1.1-5.el9 epel 58 k 2026-03-06T13:38:49.960 INFO:teuthology.orchestra.run.vm01.stdout: python3-numpy x86_64 1:1.23.5-2.el9 appstream 6.1 M 2026-03-06T13:38:49.960 INFO:teuthology.orchestra.run.vm01.stdout: python3-numpy-f2py x86_64 1:1.23.5-2.el9 appstream 442 k 2026-03-06T13:38:49.960 INFO:teuthology.orchestra.run.vm01.stdout: python3-packaging noarch 20.9-5.el9 appstream 77 k 2026-03-06T13:38:49.960 INFO:teuthology.orchestra.run.vm01.stdout: python3-pecan noarch 1.4.2-3.el9 epel 272 k 2026-03-06T13:38:49.960 INFO:teuthology.orchestra.run.vm01.stdout: python3-ply noarch 3.11-14.el9 baseos 106 k 2026-03-06T13:38:49.960 INFO:teuthology.orchestra.run.vm01.stdout: python3-portend noarch 3.1.0-2.el9 epel 16 k 2026-03-06T13:38:49.960 INFO:teuthology.orchestra.run.vm01.stdout: python3-protobuf noarch 3.14.0-17.el9 appstream 267 k 2026-03-06T13:38:49.960 INFO:teuthology.orchestra.run.vm01.stdout: python3-pyOpenSSL noarch 21.0.0-1.el9 epel 90 k 2026-03-06T13:38:49.960 INFO:teuthology.orchestra.run.vm01.stdout: python3-pyasn1 noarch 0.4.8-7.el9 appstream 157 k 2026-03-06T13:38:49.960 INFO:teuthology.orchestra.run.vm01.stdout: python3-pyasn1-modules noarch 0.4.8-7.el9 appstream 277 k 2026-03-06T13:38:49.960 INFO:teuthology.orchestra.run.vm01.stdout: python3-pycparser noarch 2.20-6.el9 baseos 135 k 2026-03-06T13:38:49.960 INFO:teuthology.orchestra.run.vm01.stdout: python3-repoze-lru noarch 0.7-16.el9 epel 31 k 2026-03-06T13:38:49.960 INFO:teuthology.orchestra.run.vm01.stdout: python3-requests noarch 2.25.1-10.el9 baseos 126 k 2026-03-06T13:38:49.960 INFO:teuthology.orchestra.run.vm01.stdout: python3-requests-oauthlib noarch 1.3.0-12.el9 appstream 54 k 2026-03-06T13:38:49.960 INFO:teuthology.orchestra.run.vm01.stdout: python3-routes noarch 2.5.1-5.el9 epel 188 k 2026-03-06T13:38:49.960 INFO:teuthology.orchestra.run.vm01.stdout: python3-rsa noarch 4.9-2.el9 epel 59 k 2026-03-06T13:38:49.960 INFO:teuthology.orchestra.run.vm01.stdout: python3-scipy x86_64 1.9.3-2.el9 appstream 19 M 2026-03-06T13:38:49.960 INFO:teuthology.orchestra.run.vm01.stdout: python3-tempora noarch 5.0.0-2.el9 epel 36 k 2026-03-06T13:38:49.960 INFO:teuthology.orchestra.run.vm01.stdout: python3-toml noarch 0.10.2-6.el9 appstream 42 k 2026-03-06T13:38:49.960 INFO:teuthology.orchestra.run.vm01.stdout: python3-typing-extensions noarch 4.15.0-1.el9 epel 86 k 2026-03-06T13:38:49.960 INFO:teuthology.orchestra.run.vm01.stdout: python3-urllib3 noarch 1.26.5-7.el9 baseos 218 k 2026-03-06T13:38:49.960 INFO:teuthology.orchestra.run.vm01.stdout: python3-webob noarch 1.8.8-2.el9 epel 230 k 2026-03-06T13:38:49.960 INFO:teuthology.orchestra.run.vm01.stdout: python3-websocket-client noarch 1.2.3-2.el9 epel 90 k 2026-03-06T13:38:49.960 INFO:teuthology.orchestra.run.vm01.stdout: python3-werkzeug noarch 2.0.3-3.el9.1 epel 427 k 2026-03-06T13:38:49.960 INFO:teuthology.orchestra.run.vm01.stdout: python3-xmlsec x86_64 1.3.13-1.el9 epel 48 k 2026-03-06T13:38:49.960 INFO:teuthology.orchestra.run.vm01.stdout: python3-zc-lockfile noarch 2.0-10.el9 epel 20 k 2026-03-06T13:38:49.960 INFO:teuthology.orchestra.run.vm01.stdout: qatlib x86_64 25.08.0-2.el9 appstream 240 k 2026-03-06T13:38:49.960 INFO:teuthology.orchestra.run.vm01.stdout: qatzip-libs x86_64 1.3.1-1.el9 appstream 66 k 2026-03-06T13:38:49.960 INFO:teuthology.orchestra.run.vm01.stdout: re2 x86_64 1:20211101-20.el9 epel 191 k 2026-03-06T13:38:49.960 INFO:teuthology.orchestra.run.vm01.stdout: socat x86_64 1.7.4.1-8.el9 appstream 303 k 2026-03-06T13:38:49.960 INFO:teuthology.orchestra.run.vm01.stdout: thrift x86_64 0.15.0-4.el9 epel 1.6 M 2026-03-06T13:38:49.960 INFO:teuthology.orchestra.run.vm01.stdout: unzip x86_64 6.0-59.el9 baseos 182 k 2026-03-06T13:38:49.960 INFO:teuthology.orchestra.run.vm01.stdout: xmlsec1 x86_64 1.2.29-13.el9 appstream 189 k 2026-03-06T13:38:49.960 INFO:teuthology.orchestra.run.vm01.stdout: xmlsec1-openssl x86_64 1.2.29-13.el9 appstream 90 k 2026-03-06T13:38:49.960 INFO:teuthology.orchestra.run.vm01.stdout: xmlstarlet x86_64 1.6.1-20.el9 appstream 64 k 2026-03-06T13:38:49.960 INFO:teuthology.orchestra.run.vm01.stdout: zip x86_64 3.0-35.el9 baseos 266 k 2026-03-06T13:38:49.960 INFO:teuthology.orchestra.run.vm01.stdout:Installing weak dependencies: 2026-03-06T13:38:49.960 INFO:teuthology.orchestra.run.vm01.stdout: ceph-mgr-k8sevents noarch 2:19.2.3-47.gc24117fd552.el9.clyso ceph-noarch 22 k 2026-03-06T13:38:49.960 INFO:teuthology.orchestra.run.vm01.stdout: python3-influxdb noarch 5.3.1-1.el9 epel 139 k 2026-03-06T13:38:49.960 INFO:teuthology.orchestra.run.vm01.stdout: python3-saml noarch 1.16.0-1.el9 epel 125 k 2026-03-06T13:38:49.961 INFO:teuthology.orchestra.run.vm01.stdout: qatlib-service x86_64 25.08.0-2.el9 appstream 37 k 2026-03-06T13:38:49.961 INFO:teuthology.orchestra.run.vm01.stdout: smartmontools x86_64 1:7.2-10.el9 baseos 556 k 2026-03-06T13:38:49.961 INFO:teuthology.orchestra.run.vm01.stdout: 2026-03-06T13:38:49.961 INFO:teuthology.orchestra.run.vm01.stdout:Transaction Summary 2026-03-06T13:38:49.961 INFO:teuthology.orchestra.run.vm01.stdout:============================================================================================== 2026-03-06T13:38:49.961 INFO:teuthology.orchestra.run.vm01.stdout:Install 148 Packages 2026-03-06T13:38:49.961 INFO:teuthology.orchestra.run.vm01.stdout:Upgrade 2 Packages 2026-03-06T13:38:49.961 INFO:teuthology.orchestra.run.vm01.stdout: 2026-03-06T13:38:49.961 INFO:teuthology.orchestra.run.vm01.stdout:Total download size: 214 M 2026-03-06T13:38:49.961 INFO:teuthology.orchestra.run.vm01.stdout:Downloading Packages: 2026-03-06T13:38:50.023 INFO:teuthology.orchestra.run.vm00.stdout:(8/150): ceph-osd-19.2.3-47.gc24117fd552.el9.cl 23 MB/s | 17 MB 00:00 2026-03-06T13:38:50.023 INFO:teuthology.orchestra.run.vm08.stdout:(36/150): ceph-mgr-diskprediction-local-19.2.3- 16 MB/s | 7.4 MB 00:00 2026-03-06T13:38:50.052 INFO:teuthology.orchestra.run.vm00.stdout:(9/150): ceph-selinux-19.2.3-47.gc24117fd552.el 876 kB/s | 25 kB 00:00 2026-03-06T13:38:50.086 INFO:teuthology.orchestra.run.vm08.stdout:(37/150): cephadm-19.2.3-47.gc24117fd552.el9.cl 4.3 MB/s | 769 kB 00:00 2026-03-06T13:38:50.224 INFO:teuthology.orchestra.run.vm08.stdout:(38/150): bzip2-1.0.8-11.el9.x86_64.rpm 197 kB/s | 55 kB 00:00 2026-03-06T13:38:50.292 INFO:teuthology.orchestra.run.vm08.stdout:(39/150): ledmon-libs-1.1.0-3.el9.x86_64.rpm 599 kB/s | 40 kB 00:00 2026-03-06T13:38:50.342 INFO:teuthology.orchestra.run.vm08.stdout:(40/150): fuse-2.9.9-17.el9.x86_64.rpm 312 kB/s | 80 kB 00:00 2026-03-06T13:38:50.361 INFO:teuthology.orchestra.run.vm08.stdout:(41/150): libconfig-1.7.2-9.el9.x86_64.rpm 1.0 MB/s | 72 kB 00:00 2026-03-06T13:38:50.413 INFO:teuthology.orchestra.run.vm08.stdout:(42/150): cryptsetup-2.8.1-3.el9.x86_64.rpm 901 kB/s | 351 kB 00:00 2026-03-06T13:38:50.478 INFO:teuthology.orchestra.run.vm08.stdout:(43/150): mailcap-2.1.49-5.el9.noarch.rpm 510 kB/s | 33 kB 00:00 2026-03-06T13:38:50.499 INFO:teuthology.orchestra.run.vm08.stdout:(44/150): libquadmath-11.5.0-14.el9.x86_64.rpm 1.3 MB/s | 184 kB 00:00 2026-03-06T13:38:50.499 INFO:teuthology.orchestra.run.vm00.stdout:(10/150): ceph-radosgw-19.2.3-47.gc24117fd552.e 11 MB/s | 11 MB 00:01 2026-03-06T13:38:50.558 INFO:teuthology.orchestra.run.vm08.stdout:(45/150): pciutils-3.7.0-7.el9.x86_64.rpm 1.1 MB/s | 93 kB 00:00 2026-03-06T13:38:50.558 INFO:teuthology.orchestra.run.vm00.stdout:(11/150): ceph-common-19.2.3-47.gc24117fd552.el 10 MB/s | 22 MB 00:02 2026-03-06T13:38:50.560 INFO:teuthology.orchestra.run.vm00.stdout:(12/150): libcephfs-devel-19.2.3-47.gc24117fd55 550 kB/s | 34 kB 00:00 2026-03-06T13:38:50.569 INFO:teuthology.orchestra.run.vm08.stdout:(46/150): python3-cffi-1.14.5-5.el9.x86_64.rpm 3.5 MB/s | 253 kB 00:00 2026-03-06T13:38:50.604 INFO:teuthology.orchestra.run.vm08.stdout:(47/150): libgfortran-11.5.0-14.el9.x86_64.rpm 3.0 MB/s | 794 kB 00:00 2026-03-06T13:38:50.636 INFO:teuthology.orchestra.run.vm00.stdout:(13/150): libcephsqlite-19.2.3-47.gc24117fd552. 2.1 MB/s | 163 kB 00:00 2026-03-06T13:38:50.644 INFO:teuthology.orchestra.run.vm08.stdout:(48/150): python3-ply-3.11-14.el9.noarch.rpm 1.4 MB/s | 106 kB 00:00 2026-03-06T13:38:50.646 INFO:teuthology.orchestra.run.vm00.stdout:(14/150): libcephfs2-19.2.3-47.gc24117fd552.el9 11 MB/s | 998 kB 00:00 2026-03-06T13:38:50.677 INFO:teuthology.orchestra.run.vm08.stdout:(49/150): python3-pycparser-2.20-6.el9.noarch.r 1.8 MB/s | 135 kB 00:00 2026-03-06T13:38:50.712 INFO:teuthology.orchestra.run.vm08.stdout:(50/150): python3-requests-2.25.1-10.el9.noarch 1.8 MB/s | 126 kB 00:00 2026-03-06T13:38:50.713 INFO:teuthology.orchestra.run.vm00.stdout:(15/150): libradosstriper1-19.2.3-47.gc24117fd5 7.4 MB/s | 503 kB 00:00 2026-03-06T13:38:50.747 INFO:teuthology.orchestra.run.vm08.stdout:(51/150): python3-urllib3-1.26.5-7.el9.noarch.r 3.0 MB/s | 218 kB 00:00 2026-03-06T13:38:50.758 INFO:teuthology.orchestra.run.vm00.stdout:(16/150): librados-devel-19.2.3-47.gc24117fd552 1.0 MB/s | 127 kB 00:00 2026-03-06T13:38:50.759 INFO:teuthology.orchestra.run.vm08.stdout:(52/150): python3-cryptography-36.0.1-5.el9.x86 6.2 MB/s | 1.2 MB 00:00 2026-03-06T13:38:50.789 INFO:teuthology.orchestra.run.vm00.stdout:(17/150): python3-ceph-argparse-19.2.3-47.gc241 1.4 MB/s | 45 kB 00:00 2026-03-06T13:38:50.818 INFO:teuthology.orchestra.run.vm08.stdout:(53/150): unzip-6.0-59.el9.x86_64.rpm 2.5 MB/s | 182 kB 00:00 2026-03-06T13:38:50.831 INFO:teuthology.orchestra.run.vm08.stdout:(54/150): zip-3.0-35.el9.x86_64.rpm 3.6 MB/s | 266 kB 00:00 2026-03-06T13:38:50.847 INFO:teuthology.orchestra.run.vm08.stdout:(55/150): smartmontools-7.2-10.el9.x86_64.rpm 4.0 MB/s | 556 kB 00:00 2026-03-06T13:38:50.858 INFO:teuthology.orchestra.run.vm00.stdout:(18/150): python3-ceph-common-19.2.3-47.gc24117 2.0 MB/s | 142 kB 00:00 2026-03-06T13:38:50.898 INFO:teuthology.orchestra.run.vm00.stdout:(19/150): python3-cephfs-19.2.3-47.gc24117fd552 4.2 MB/s | 165 kB 00:00 2026-03-06T13:38:50.948 INFO:teuthology.orchestra.run.vm00.stdout:(20/150): python3-rados-19.2.3-47.gc24117fd552. 6.3 MB/s | 322 kB 00:00 2026-03-06T13:38:50.994 INFO:teuthology.orchestra.run.vm00.stdout:(21/150): python3-rbd-19.2.3-47.gc24117fd552.el 6.5 MB/s | 303 kB 00:00 2026-03-06T13:38:51.028 INFO:teuthology.orchestra.run.vm00.stdout:(22/150): python3-rgw-19.2.3-47.gc24117fd552.el 2.9 MB/s | 100 kB 00:00 2026-03-06T13:38:51.090 INFO:teuthology.orchestra.run.vm00.stdout:(23/150): rbd-fuse-19.2.3-47.gc24117fd552.el9.c 1.4 MB/s | 85 kB 00:00 2026-03-06T13:38:51.124 INFO:teuthology.orchestra.run.vm00.stdout:(24/150): librgw2-19.2.3-47.gc24117fd552.el9.cl 13 MB/s | 5.4 MB 00:00 2026-03-06T13:38:51.162 INFO:teuthology.orchestra.run.vm00.stdout:(25/150): rbd-nbd-19.2.3-47.gc24117fd552.el9.cl 4.4 MB/s | 171 kB 00:00 2026-03-06T13:38:51.194 INFO:teuthology.orchestra.run.vm00.stdout:(26/150): ceph-grafana-dashboards-19.2.3-47.gc2 1.0 MB/s | 31 kB 00:00 2026-03-06T13:38:51.256 INFO:teuthology.orchestra.run.vm00.stdout:(27/150): ceph-mgr-cephadm-19.2.3-47.gc24117fd5 2.4 MB/s | 150 kB 00:00 2026-03-06T13:38:51.289 INFO:teuthology.orchestra.run.vm01.stdout:(1/150): ceph-19.2.3-47.gc24117fd552.el9.clyso. 63 kB/s | 6.5 kB 00:00 2026-03-06T13:38:51.307 INFO:teuthology.orchestra.run.vm00.stdout:(28/150): rbd-mirror-19.2.3-47.gc24117fd552.el9 14 MB/s | 3.1 MB 00:00 2026-03-06T13:38:51.570 INFO:teuthology.orchestra.run.vm08.stdout:(56/150): flexiblas-3.0.4-9.el9.x86_64.rpm 40 kB/s | 30 kB 00:00 2026-03-06T13:38:51.577 INFO:teuthology.orchestra.run.vm01.stdout:(2/150): ceph-fuse-19.2.3-47.gc24117fd552.el9.c 4.0 MB/s | 1.1 MB 00:00 2026-03-06T13:38:51.621 INFO:teuthology.orchestra.run.vm01.stdout:(3/150): ceph-immutable-object-cache-19.2.3-47. 3.3 MB/s | 145 kB 00:00 2026-03-06T13:38:51.621 INFO:teuthology.orchestra.run.vm00.stdout:(29/150): ceph-test-19.2.3-47.gc24117fd552.el9. 32 MB/s | 50 MB 00:01 2026-03-06T13:38:51.723 INFO:teuthology.orchestra.run.vm08.stdout:(57/150): boost-program-options-1.75.0-13.el9.x 115 kB/s | 104 kB 00:00 2026-03-06T13:38:51.805 INFO:teuthology.orchestra.run.vm00.stdout:(30/150): ceph-mgr-k8sevents-19.2.3-47.gc24117f 120 kB/s | 22 kB 00:00 2026-03-06T13:38:51.863 INFO:teuthology.orchestra.run.vm00.stdout:(31/150): ceph-mgr-diskprediction-local-19.2.3- 13 MB/s | 7.4 MB 00:00 2026-03-06T13:38:51.892 INFO:teuthology.orchestra.run.vm01.stdout:(4/150): ceph-base-19.2.3-47.gc24117fd552.el9.c 7.8 MB/s | 5.5 MB 00:00 2026-03-06T13:38:51.898 INFO:teuthology.orchestra.run.vm00.stdout:(32/150): ceph-mgr-rook-19.2.3-47.gc24117fd552. 1.3 MB/s | 49 kB 00:00 2026-03-06T13:38:51.932 INFO:teuthology.orchestra.run.vm00.stdout:(33/150): ceph-mgr-modules-core-19.2.3-47.gc241 1.9 MB/s | 252 kB 00:00 2026-03-06T13:38:51.955 INFO:teuthology.orchestra.run.vm00.stdout:(34/150): ceph-mgr-dashboard-19.2.3-47.gc24117f 5.4 MB/s | 3.8 MB 00:00 2026-03-06T13:38:51.956 INFO:teuthology.orchestra.run.vm00.stdout:(35/150): ceph-prometheus-alerts-19.2.3-47.gc24 282 kB/s | 16 kB 00:00 2026-03-06T13:38:51.974 INFO:teuthology.orchestra.run.vm01.stdout:(5/150): ceph-mds-19.2.3-47.gc24117fd552.el9.cl 6.9 MB/s | 2.4 MB 00:00 2026-03-06T13:38:51.980 INFO:teuthology.orchestra.run.vm01.stdout:(6/150): ceph-mgr-19.2.3-47.gc24117fd552.el9.cl 12 MB/s | 1.1 MB 00:00 2026-03-06T13:38:51.991 INFO:teuthology.orchestra.run.vm00.stdout:(36/150): ceph-volume-19.2.3-47.gc24117fd552.el 4.9 MB/s | 299 kB 00:00 2026-03-06T13:38:51.992 INFO:teuthology.orchestra.run.vm08.stdout:(58/150): flexiblas-openblas-openmp-3.0.4-9.el9 35 kB/s | 15 kB 00:00 2026-03-06T13:38:52.038 INFO:teuthology.orchestra.run.vm00.stdout:(37/150): cephadm-19.2.3-47.gc24117fd552.el9.cl 9.1 MB/s | 769 kB 00:00 2026-03-06T13:38:52.266 INFO:teuthology.orchestra.run.vm01.stdout:(7/150): ceph-mon-19.2.3-47.gc24117fd552.el9.cl 16 MB/s | 4.7 MB 00:00 2026-03-06T13:38:52.311 INFO:teuthology.orchestra.run.vm00.stdout:(38/150): bzip2-1.0.8-11.el9.x86_64.rpm 155 kB/s | 55 kB 00:00 2026-03-06T13:38:52.318 INFO:teuthology.orchestra.run.vm08.stdout:(59/150): flexiblas-netlib-3.0.4-9.el9.x86_64.r 2.0 MB/s | 3.0 MB 00:01 2026-03-06T13:38:52.336 INFO:teuthology.orchestra.run.vm08.stdout:(60/150): libnbd-1.20.3-4.el9.x86_64.rpm 268 kB/s | 164 kB 00:00 2026-03-06T13:38:52.435 INFO:teuthology.orchestra.run.vm00.stdout:(39/150): ledmon-libs-1.1.0-3.el9.x86_64.rpm 327 kB/s | 40 kB 00:00 2026-03-06T13:38:52.474 INFO:teuthology.orchestra.run.vm01.stdout:(8/150): ceph-common-19.2.3-47.gc24117fd552.el9 17 MB/s | 22 MB 00:01 2026-03-06T13:38:52.501 INFO:teuthology.orchestra.run.vm00.stdout:(40/150): cryptsetup-2.8.1-3.el9.x86_64.rpm 689 kB/s | 351 kB 00:00 2026-03-06T13:38:52.509 INFO:teuthology.orchestra.run.vm01.stdout:(9/150): ceph-selinux-19.2.3-47.gc24117fd552.el 733 kB/s | 25 kB 00:00 2026-03-06T13:38:52.559 INFO:teuthology.orchestra.run.vm00.stdout:(41/150): fuse-2.9.9-17.el9.x86_64.rpm 153 kB/s | 80 kB 00:00 2026-03-06T13:38:52.621 INFO:teuthology.orchestra.run.vm00.stdout:(42/150): libconfig-1.7.2-9.el9.x86_64.rpm 387 kB/s | 72 kB 00:00 2026-03-06T13:38:52.621 INFO:teuthology.orchestra.run.vm01.stdout:(10/150): ceph-osd-19.2.3-47.gc24117fd552.el9.c 27 MB/s | 17 MB 00:00 2026-03-06T13:38:52.621 INFO:teuthology.orchestra.run.vm08.stdout:(61/150): libpmemobj-1.12.1-1.el9.x86_64.rpm 255 kB/s | 160 kB 00:00 2026-03-06T13:38:52.667 INFO:teuthology.orchestra.run.vm01.stdout:(11/150): libcephfs-devel-19.2.3-47.gc24117fd55 753 kB/s | 34 kB 00:00 2026-03-06T13:38:52.703 INFO:teuthology.orchestra.run.vm00.stdout:(43/150): libgfortran-11.5.0-14.el9.x86_64.rpm 3.8 MB/s | 794 kB 00:00 2026-03-06T13:38:52.704 INFO:teuthology.orchestra.run.vm00.stdout:(44/150): mailcap-2.1.49-5.el9.noarch.rpm 399 kB/s | 33 kB 00:00 2026-03-06T13:38:52.762 INFO:teuthology.orchestra.run.vm01.stdout:(12/150): libcephfs2-19.2.3-47.gc24117fd552.el9 10 MB/s | 998 kB 00:00 2026-03-06T13:38:52.766 INFO:teuthology.orchestra.run.vm08.stdout:(62/150): librabbitmq-0.11.0-7.el9.x86_64.rpm 101 kB/s | 45 kB 00:00 2026-03-06T13:38:52.805 INFO:teuthology.orchestra.run.vm00.stdout:(45/150): pciutils-3.7.0-7.el9.x86_64.rpm 913 kB/s | 93 kB 00:00 2026-03-06T13:38:52.830 INFO:teuthology.orchestra.run.vm01.stdout:(13/150): libcephsqlite-19.2.3-47.gc24117fd552. 2.4 MB/s | 163 kB 00:00 2026-03-06T13:38:52.847 INFO:teuthology.orchestra.run.vm08.stdout:(63/150): libstoragemgmt-1.10.1-1.el9.x86_64.rp 1.1 MB/s | 246 kB 00:00 2026-03-06T13:38:52.858 INFO:teuthology.orchestra.run.vm00.stdout:(46/150): python3-cffi-1.14.5-5.el9.x86_64.rpm 1.6 MB/s | 253 kB 00:00 2026-03-06T13:38:52.864 INFO:teuthology.orchestra.run.vm01.stdout:(14/150): librados-devel-19.2.3-47.gc24117fd552 3.6 MB/s | 127 kB 00:00 2026-03-06T13:38:52.962 INFO:teuthology.orchestra.run.vm00.stdout:(47/150): python3-ply-3.11-14.el9.noarch.rpm 1.0 MB/s | 106 kB 00:00 2026-03-06T13:38:52.967 INFO:teuthology.orchestra.run.vm01.stdout:(15/150): libradosstriper1-19.2.3-47.gc24117fd5 4.8 MB/s | 503 kB 00:00 2026-03-06T13:38:52.973 INFO:teuthology.orchestra.run.vm08.stdout:(64/150): libxslt-1.1.34-12.el9.x86_64.rpm 1.1 MB/s | 233 kB 00:00 2026-03-06T13:38:53.063 INFO:teuthology.orchestra.run.vm00.stdout:(48/150): python3-cryptography-36.0.1-5.el9.x86 4.8 MB/s | 1.2 MB 00:00 2026-03-06T13:38:53.077 INFO:teuthology.orchestra.run.vm00.stdout:(49/150): python3-pycparser-2.20-6.el9.noarch.r 1.2 MB/s | 135 kB 00:00 2026-03-06T13:38:53.103 INFO:teuthology.orchestra.run.vm01.stdout:(16/150): ceph-radosgw-19.2.3-47.gc24117fd552.e 13 MB/s | 11 MB 00:00 2026-03-06T13:38:53.109 INFO:teuthology.orchestra.run.vm08.stdout:(65/150): librdkafka-1.6.1-102.el9.x86_64.rpm 858 kB/s | 662 kB 00:00 2026-03-06T13:38:53.125 INFO:teuthology.orchestra.run.vm08.stdout:(66/150): lua-5.4.4-4.el9.x86_64.rpm 1.2 MB/s | 188 kB 00:00 2026-03-06T13:38:53.159 INFO:teuthology.orchestra.run.vm00.stdout:(50/150): python3-requests-2.25.1-10.el9.noarch 1.3 MB/s | 126 kB 00:00 2026-03-06T13:38:53.164 INFO:teuthology.orchestra.run.vm00.stdout:(51/150): python3-urllib3-1.26.5-7.el9.noarch.r 2.5 MB/s | 218 kB 00:00 2026-03-06T13:38:53.179 INFO:teuthology.orchestra.run.vm01.stdout:(17/150): python3-ceph-argparse-19.2.3-47.gc241 596 kB/s | 45 kB 00:00 2026-03-06T13:38:53.230 INFO:teuthology.orchestra.run.vm01.stdout:(18/150): python3-ceph-common-19.2.3-47.gc24117 2.8 MB/s | 142 kB 00:00 2026-03-06T13:38:53.238 INFO:teuthology.orchestra.run.vm08.stdout:(67/150): openblas-0.3.29-1.el9.x86_64.rpm 327 kB/s | 42 kB 00:00 2026-03-06T13:38:53.266 INFO:teuthology.orchestra.run.vm00.stdout:(52/150): unzip-6.0-59.el9.x86_64.rpm 1.7 MB/s | 182 kB 00:00 2026-03-06T13:38:53.266 INFO:teuthology.orchestra.run.vm01.stdout:(19/150): librgw2-19.2.3-47.gc24117fd552.el9.cl 18 MB/s | 5.4 MB 00:00 2026-03-06T13:38:53.270 INFO:teuthology.orchestra.run.vm01.stdout:(20/150): python3-cephfs-19.2.3-47.gc24117fd552 4.0 MB/s | 165 kB 00:00 2026-03-06T13:38:53.289 INFO:teuthology.orchestra.run.vm00.stdout:(53/150): libquadmath-11.5.0-14.el9.x86_64.rpm 253 kB/s | 184 kB 00:00 2026-03-06T13:38:53.311 INFO:teuthology.orchestra.run.vm00.stdout:(54/150): smartmontools-7.2-10.el9.x86_64.rpm 3.6 MB/s | 556 kB 00:00 2026-03-06T13:38:53.352 INFO:teuthology.orchestra.run.vm00.stdout:(55/150): zip-3.0-35.el9.x86_64.rpm 3.0 MB/s | 266 kB 00:00 2026-03-06T13:38:53.367 INFO:teuthology.orchestra.run.vm08.stdout:(68/150): perl-Benchmark-1.23-483.el9.noarch.rp 204 kB/s | 26 kB 00:00 2026-03-06T13:38:53.376 INFO:teuthology.orchestra.run.vm01.stdout:(21/150): python3-rbd-19.2.3-47.gc24117fd552.el 2.8 MB/s | 303 kB 00:00 2026-03-06T13:38:53.382 INFO:teuthology.orchestra.run.vm01.stdout:(22/150): python3-rados-19.2.3-47.gc24117fd552. 2.7 MB/s | 322 kB 00:00 2026-03-06T13:38:53.422 INFO:teuthology.orchestra.run.vm01.stdout:(23/150): rbd-fuse-19.2.3-47.gc24117fd552.el9.c 2.1 MB/s | 85 kB 00:00 2026-03-06T13:38:53.423 INFO:teuthology.orchestra.run.vm01.stdout:(24/150): python3-rgw-19.2.3-47.gc24117fd552.el 2.1 MB/s | 100 kB 00:00 2026-03-06T13:38:53.487 INFO:teuthology.orchestra.run.vm08.stdout:(69/150): lttng-ust-2.12.0-6.el9.x86_64.rpm 458 kB/s | 292 kB 00:00 2026-03-06T13:38:53.490 INFO:teuthology.orchestra.run.vm01.stdout:(25/150): rbd-nbd-19.2.3-47.gc24117fd552.el9.cl 2.5 MB/s | 171 kB 00:00 2026-03-06T13:38:53.495 INFO:teuthology.orchestra.run.vm00.stdout:(56/150): flexiblas-3.0.4-9.el9.x86_64.rpm 161 kB/s | 30 kB 00:00 2026-03-06T13:38:53.512 INFO:teuthology.orchestra.run.vm08.stdout:(70/150): perl-Test-Harness-3.42-461.el9.noarch 2.0 MB/s | 295 kB 00:00 2026-03-06T13:38:53.515 INFO:teuthology.orchestra.run.vm00.stdout:(57/150): boost-program-options-1.75.0-13.el9.x 461 kB/s | 104 kB 00:00 2026-03-06T13:38:53.526 INFO:teuthology.orchestra.run.vm01.stdout:(26/150): ceph-grafana-dashboards-19.2.3-47.gc2 876 kB/s | 31 kB 00:00 2026-03-06T13:38:53.535 INFO:teuthology.orchestra.run.vm00.stdout:(58/150): flexiblas-openblas-openmp-3.0.4-9.el9 373 kB/s | 15 kB 00:00 2026-03-06T13:38:53.579 INFO:teuthology.orchestra.run.vm01.stdout:(27/150): ceph-mgr-cephadm-19.2.3-47.gc24117fd5 2.8 MB/s | 150 kB 00:00 2026-03-06T13:38:53.586 INFO:teuthology.orchestra.run.vm00.stdout:(59/150): libnbd-1.20.3-4.el9.x86_64.rpm 2.2 MB/s | 164 kB 00:00 2026-03-06T13:38:53.610 INFO:teuthology.orchestra.run.vm00.stdout:(60/150): libpmemobj-1.12.1-1.el9.x86_64.rpm 2.1 MB/s | 160 kB 00:00 2026-03-06T13:38:53.626 INFO:teuthology.orchestra.run.vm00.stdout:(61/150): librabbitmq-0.11.0-7.el9.x86_64.rpm 1.1 MB/s | 45 kB 00:00 2026-03-06T13:38:53.674 INFO:teuthology.orchestra.run.vm01.stdout:(28/150): rbd-mirror-19.2.3-47.gc24117fd552.el9 12 MB/s | 3.1 MB 00:00 2026-03-06T13:38:53.692 INFO:teuthology.orchestra.run.vm00.stdout:(62/150): librdkafka-1.6.1-102.el9.x86_64.rpm 7.9 MB/s | 662 kB 00:00 2026-03-06T13:38:53.707 INFO:teuthology.orchestra.run.vm00.stdout:(63/150): libstoragemgmt-1.10.1-1.el9.x86_64.rp 3.0 MB/s | 246 kB 00:00 2026-03-06T13:38:53.735 INFO:teuthology.orchestra.run.vm00.stdout:(64/150): libxslt-1.1.34-12.el9.x86_64.rpm 5.3 MB/s | 233 kB 00:00 2026-03-06T13:38:53.743 INFO:teuthology.orchestra.run.vm00.stdout:(65/150): flexiblas-netlib-3.0.4-9.el9.x86_64.r 7.6 MB/s | 3.0 MB 00:00 2026-03-06T13:38:53.754 INFO:teuthology.orchestra.run.vm00.stdout:(66/150): lttng-ust-2.12.0-6.el9.x86_64.rpm 6.1 MB/s | 292 kB 00:00 2026-03-06T13:38:53.777 INFO:teuthology.orchestra.run.vm00.stdout:(67/150): lua-5.4.4-4.el9.x86_64.rpm 4.5 MB/s | 188 kB 00:00 2026-03-06T13:38:53.780 INFO:teuthology.orchestra.run.vm00.stdout:(68/150): openblas-0.3.29-1.el9.x86_64.rpm 1.1 MB/s | 42 kB 00:00 2026-03-06T13:38:53.911 INFO:teuthology.orchestra.run.vm00.stdout:(69/150): perl-Benchmark-1.23-483.el9.noarch.rp 197 kB/s | 26 kB 00:00 2026-03-06T13:38:54.024 INFO:teuthology.orchestra.run.vm00.stdout:(70/150): openblas-openmp-0.3.29-1.el9.x86_64.r 20 MB/s | 5.3 MB 00:00 2026-03-06T13:38:54.025 INFO:teuthology.orchestra.run.vm00.stdout:(71/150): perl-Test-Harness-3.42-461.el9.noarch 1.2 MB/s | 295 kB 00:00 2026-03-06T13:38:54.030 INFO:teuthology.orchestra.run.vm00.stdout:(72/150): protobuf-3.14.0-17.el9.x86_64.rpm 8.5 MB/s | 1.0 MB 00:00 2026-03-06T13:38:54.069 INFO:teuthology.orchestra.run.vm00.stdout:(73/150): python3-devel-3.9.25-3.el9.x86_64.rpm 5.5 MB/s | 244 kB 00:00 2026-03-06T13:38:54.084 INFO:teuthology.orchestra.run.vm01.stdout:(29/150): ceph-mgr-dashboard-19.2.3-47.gc24117f 7.5 MB/s | 3.8 MB 00:00 2026-03-06T13:38:54.090 INFO:teuthology.orchestra.run.vm00.stdout:(74/150): python3-jinja2-2.11.3-8.el9.noarch.rp 4.1 MB/s | 249 kB 00:00 2026-03-06T13:38:54.107 INFO:teuthology.orchestra.run.vm01.stdout:(30/150): ceph-mgr-diskprediction-local-19.2.3- 17 MB/s | 7.4 MB 00:00 2026-03-06T13:38:54.151 INFO:teuthology.orchestra.run.vm01.stdout:(31/150): ceph-mgr-modules-core-19.2.3-47.gc241 5.6 MB/s | 252 kB 00:00 2026-03-06T13:38:54.179 INFO:teuthology.orchestra.run.vm08.stdout:(71/150): openblas-openmp-0.3.29-1.el9.x86_64.r 5.0 MB/s | 5.3 MB 00:01 2026-03-06T13:38:54.182 INFO:teuthology.orchestra.run.vm00.stdout:(75/150): python3-libstoragemgmt-1.10.1-1.el9.x 1.5 MB/s | 177 kB 00:00 2026-03-06T13:38:54.209 INFO:teuthology.orchestra.run.vm00.stdout:(76/150): python3-babel-2.9.1-2.el9.noarch.rpm 33 MB/s | 6.0 MB 00:00 2026-03-06T13:38:54.214 INFO:teuthology.orchestra.run.vm00.stdout:(77/150): python3-lxml-4.6.5-3.el9.x86_64.rpm 9.9 MB/s | 1.2 MB 00:00 2026-03-06T13:38:54.215 INFO:teuthology.orchestra.run.vm01.stdout:(32/150): ceph-mgr-rook-19.2.3-47.gc24117fd552. 778 kB/s | 49 kB 00:00 2026-03-06T13:38:54.222 INFO:teuthology.orchestra.run.vm01.stdout:(33/150): ceph-mgr-k8sevents-19.2.3-47.gc24117f 161 kB/s | 22 kB 00:00 2026-03-06T13:38:54.227 INFO:teuthology.orchestra.run.vm00.stdout:(78/150): python3-mako-1.1.4-6.el9.noarch.rpm 3.8 MB/s | 172 kB 00:00 2026-03-06T13:38:54.243 INFO:teuthology.orchestra.run.vm01.stdout:(34/150): ceph-prometheus-alerts-19.2.3-47.gc24 575 kB/s | 16 kB 00:00 2026-03-06T13:38:54.255 INFO:teuthology.orchestra.run.vm00.stdout:(79/150): python3-markupsafe-1.1.1-12.el9.x86_6 766 kB/s | 35 kB 00:00 2026-03-06T13:38:54.265 INFO:teuthology.orchestra.run.vm01.stdout:(35/150): ceph-volume-19.2.3-47.gc24117fd552.el 6.9 MB/s | 299 kB 00:00 2026-03-06T13:38:54.277 INFO:teuthology.orchestra.run.vm00.stdout:(80/150): python3-numpy-f2py-1.23.5-2.el9.x86_6 8.7 MB/s | 442 kB 00:00 2026-03-06T13:38:54.329 INFO:teuthology.orchestra.run.vm08.stdout:(72/150): protobuf-3.14.0-17.el9.x86_64.rpm 1.2 MB/s | 1.0 MB 00:00 2026-03-06T13:38:54.337 INFO:teuthology.orchestra.run.vm08.stdout:(73/150): python3-devel-3.9.25-3.el9.x86_64.rpm 1.5 MB/s | 244 kB 00:00 2026-03-06T13:38:54.346 INFO:teuthology.orchestra.run.vm00.stdout:(81/150): python3-packaging-20.9-5.el9.noarch.r 855 kB/s | 77 kB 00:00 2026-03-06T13:38:54.401 INFO:teuthology.orchestra.run.vm01.stdout:(36/150): bzip2-1.0.8-11.el9.x86_64.rpm 405 kB/s | 55 kB 00:00 2026-03-06T13:38:54.460 INFO:teuthology.orchestra.run.vm00.stdout:(82/150): python3-numpy-1.23.5-2.el9.x86_64.rpm 25 MB/s | 6.1 MB 00:00 2026-03-06T13:38:54.460 INFO:teuthology.orchestra.run.vm00.stdout:(83/150): python3-protobuf-3.14.0-17.el9.noarch 1.4 MB/s | 267 kB 00:00 2026-03-06T13:38:54.461 INFO:teuthology.orchestra.run.vm00.stdout:(84/150): python3-pyasn1-0.4.8-7.el9.noarch.rpm 1.3 MB/s | 157 kB 00:00 2026-03-06T13:38:54.468 INFO:teuthology.orchestra.run.vm08.stdout:(74/150): python3-libstoragemgmt-1.10.1-1.el9.x 1.3 MB/s | 177 kB 00:00 2026-03-06T13:38:54.493 INFO:teuthology.orchestra.run.vm01.stdout:(37/150): cryptsetup-2.8.1-3.el9.x86_64.rpm 3.7 MB/s | 351 kB 00:00 2026-03-06T13:38:54.506 INFO:teuthology.orchestra.run.vm00.stdout:(85/150): python3-requests-oauthlib-1.3.0-12.el 1.2 MB/s | 54 kB 00:00 2026-03-06T13:38:54.511 INFO:teuthology.orchestra.run.vm00.stdout:(86/150): python3-pyasn1-modules-0.4.8-7.el9.no 5.5 MB/s | 277 kB 00:00 2026-03-06T13:38:54.541 INFO:teuthology.orchestra.run.vm01.stdout:(38/150): fuse-2.9.9-17.el9.x86_64.rpm 1.6 MB/s | 80 kB 00:00 2026-03-06T13:38:54.541 INFO:teuthology.orchestra.run.vm08.stdout:(75/150): python3-babel-2.9.1-2.el9.noarch.rpm 5.8 MB/s | 6.0 MB 00:01 2026-03-06T13:38:54.554 INFO:teuthology.orchestra.run.vm00.stdout:(87/150): python3-toml-0.10.2-6.el9.noarch.rpm 860 kB/s | 42 kB 00:00 2026-03-06T13:38:54.594 INFO:teuthology.orchestra.run.vm00.stdout:(88/150): qatlib-25.08.0-2.el9.x86_64.rpm 2.8 MB/s | 240 kB 00:00 2026-03-06T13:38:54.615 INFO:teuthology.orchestra.run.vm08.stdout:(76/150): python3-lxml-4.6.5-3.el9.x86_64.rpm 8.3 MB/s | 1.2 MB 00:00 2026-03-06T13:38:54.689 INFO:teuthology.orchestra.run.vm00.stdout:(89/150): qatlib-service-25.08.0-2.el9.x86_64.r 274 kB/s | 37 kB 00:00 2026-03-06T13:38:54.767 INFO:teuthology.orchestra.run.vm01.stdout:(39/150): ledmon-libs-1.1.0-3.el9.x86_64.rpm 178 kB/s | 40 kB 00:00 2026-03-06T13:38:54.771 INFO:teuthology.orchestra.run.vm01.stdout:(40/150): cephadm-19.2.3-47.gc24117fd552.el9.cl 1.4 MB/s | 769 kB 00:00 2026-03-06T13:38:54.819 INFO:teuthology.orchestra.run.vm01.stdout:(41/150): libconfig-1.7.2-9.el9.x86_64.rpm 1.4 MB/s | 72 kB 00:00 2026-03-06T13:38:54.826 INFO:teuthology.orchestra.run.vm00.stdout:(90/150): qatzip-libs-1.3.1-1.el9.x86_64.rpm 287 kB/s | 66 kB 00:00 2026-03-06T13:38:54.858 INFO:teuthology.orchestra.run.vm01.stdout:(42/150): libquadmath-11.5.0-14.el9.x86_64.rpm 4.7 MB/s | 184 kB 00:00 2026-03-06T13:38:54.903 INFO:teuthology.orchestra.run.vm00.stdout:(91/150): python3-scipy-1.9.3-2.el9.x86_64.rpm 44 MB/s | 19 MB 00:00 2026-03-06T13:38:54.903 INFO:teuthology.orchestra.run.vm01.stdout:(43/150): mailcap-2.1.49-5.el9.noarch.rpm 736 kB/s | 33 kB 00:00 2026-03-06T13:38:54.905 INFO:teuthology.orchestra.run.vm00.stdout:(92/150): socat-1.7.4.1-8.el9.x86_64.rpm 1.4 MB/s | 303 kB 00:00 2026-03-06T13:38:54.906 INFO:teuthology.orchestra.run.vm00.stdout:(93/150): xmlsec1-1.2.29-13.el9.x86_64.rpm 2.3 MB/s | 189 kB 00:00 2026-03-06T13:38:54.942 INFO:teuthology.orchestra.run.vm08.stdout:(77/150): python3-jinja2-2.11.3-8.el9.noarch.rp 406 kB/s | 249 kB 00:00 2026-03-06T13:38:54.944 INFO:teuthology.orchestra.run.vm00.stdout:(94/150): xmlsec1-openssl-1.2.29-13.el9.x86_64. 2.2 MB/s | 90 kB 00:00 2026-03-06T13:38:54.954 INFO:teuthology.orchestra.run.vm00.stdout:(95/150): xmlstarlet-1.6.1-20.el9.x86_64.rpm 1.3 MB/s | 64 kB 00:00 2026-03-06T13:38:54.969 INFO:teuthology.orchestra.run.vm01.stdout:(44/150): pciutils-3.7.0-7.el9.x86_64.rpm 1.4 MB/s | 93 kB 00:00 2026-03-06T13:38:54.970 INFO:teuthology.orchestra.run.vm00.stdout:(96/150): abseil-cpp-20211102.0-4.el9.x86_64.rp 33 MB/s | 551 kB 00:00 2026-03-06T13:38:54.980 INFO:teuthology.orchestra.run.vm00.stdout:(97/150): gperftools-libs-2.9.1-3.el9.x86_64.rp 32 MB/s | 308 kB 00:00 2026-03-06T13:38:54.983 INFO:teuthology.orchestra.run.vm00.stdout:(98/150): grpc-data-1.46.7-10.el9.noarch.rpm 7.0 MB/s | 19 kB 00:00 2026-03-06T13:38:54.999 INFO:teuthology.orchestra.run.vm01.stdout:(45/150): libgfortran-11.5.0-14.el9.x86_64.rpm 3.4 MB/s | 794 kB 00:00 2026-03-06T13:38:55.023 INFO:teuthology.orchestra.run.vm01.stdout:(46/150): python3-cffi-1.14.5-5.el9.x86_64.rpm 4.6 MB/s | 253 kB 00:00 2026-03-06T13:38:55.034 INFO:teuthology.orchestra.run.vm08.stdout:(78/150): python3-markupsafe-1.1.1-12.el9.x86_6 83 kB/s | 35 kB 00:00 2026-03-06T13:38:55.046 INFO:teuthology.orchestra.run.vm00.stdout:(99/150): libarrow-9.0.0-15.el9.x86_64.rpm 71 MB/s | 4.4 MB 00:00 2026-03-06T13:38:55.046 INFO:teuthology.orchestra.run.vm00.stdout:(100/150): lua-devel-5.4.4-4.el9.x86_64.rpm 159 kB/s | 22 kB 00:00 2026-03-06T13:38:55.049 INFO:teuthology.orchestra.run.vm00.stdout:(101/150): libarrow-doc-9.0.0-15.el9.noarch.rpm 8.0 MB/s | 25 kB 00:00 2026-03-06T13:38:55.052 INFO:teuthology.orchestra.run.vm00.stdout:(102/150): libunwind-1.6.2-1.el9.x86_64.rpm 24 MB/s | 67 kB 00:00 2026-03-06T13:38:55.052 INFO:teuthology.orchestra.run.vm00.stdout:(103/150): liboath-2.6.12-1.el9.x86_64.rpm 7.9 MB/s | 49 kB 00:00 2026-03-06T13:38:55.055 INFO:teuthology.orchestra.run.vm00.stdout:(104/150): luarocks-3.9.2-5.el9.noarch.rpm 44 MB/s | 151 kB 00:00 2026-03-06T13:38:55.066 INFO:teuthology.orchestra.run.vm00.stdout:(105/150): python3-asyncssh-2.13.2-5.el9.noarch 53 MB/s | 548 kB 00:00 2026-03-06T13:38:55.068 INFO:teuthology.orchestra.run.vm00.stdout:(106/150): python3-autocommand-2.2.2-8.el9.noar 11 MB/s | 29 kB 00:00 2026-03-06T13:38:55.071 INFO:teuthology.orchestra.run.vm00.stdout:(107/150): python3-backports-tarfile-1.2.0-1.el 23 MB/s | 60 kB 00:00 2026-03-06T13:38:55.077 INFO:teuthology.orchestra.run.vm01.stdout:(47/150): python3-ply-3.11-14.el9.noarch.rpm 1.9 MB/s | 106 kB 00:00 2026-03-06T13:38:55.077 INFO:teuthology.orchestra.run.vm00.stdout:(108/150): python3-bcrypt-3.2.2-1.el9.x86_64.rp 7.6 MB/s | 43 kB 00:00 2026-03-06T13:38:55.084 INFO:teuthology.orchestra.run.vm00.stdout:(109/150): python3-cachetools-4.2.4-1.el9.noarc 4.6 MB/s | 32 kB 00:00 2026-03-06T13:38:55.087 INFO:teuthology.orchestra.run.vm00.stdout:(110/150): parquet-libs-9.0.0-15.el9.x86_64.rpm 24 MB/s | 838 kB 00:00 2026-03-06T13:38:55.092 INFO:teuthology.orchestra.run.vm00.stdout:(111/150): python3-certifi-2023.05.07-4.el9.noa 1.9 MB/s | 14 kB 00:00 2026-03-06T13:38:55.122 INFO:teuthology.orchestra.run.vm00.stdout:(112/150): python3-cheroot-10.0.1-4.el9.noarch. 4.9 MB/s | 173 kB 00:00 2026-03-06T13:38:55.123 INFO:teuthology.orchestra.run.vm01.stdout:(48/150): python3-pycparser-2.20-6.el9.noarch.r 2.9 MB/s | 135 kB 00:00 2026-03-06T13:38:55.134 INFO:teuthology.orchestra.run.vm00.stdout:(113/150): python3-cherrypy-18.6.1-2.el9.noarch 8.3 MB/s | 358 kB 00:00 2026-03-06T13:38:55.135 INFO:teuthology.orchestra.run.vm00.stdout:(114/150): python3-google-auth-2.45.0-1.el9.noa 19 MB/s | 254 kB 00:00 2026-03-06T13:38:55.140 INFO:teuthology.orchestra.run.vm08.stdout:(79/150): python3-mako-1.1.4-6.el9.noarch.rpm 287 kB/s | 172 kB 00:00 2026-03-06T13:38:55.147 INFO:teuthology.orchestra.run.vm00.stdout:(115/150): python3-grpcio-tools-1.46.7-10.el9.x 12 MB/s | 144 kB 00:00 2026-03-06T13:38:55.169 INFO:teuthology.orchestra.run.vm00.stdout:(116/150): python3-influxdb-5.3.1-1.el9.noarch. 6.3 MB/s | 139 kB 00:00 2026-03-06T13:38:55.175 INFO:teuthology.orchestra.run.vm00.stdout:(117/150): python3-grpcio-1.46.7-10.el9.x86_64. 51 MB/s | 2.0 MB 00:00 2026-03-06T13:38:55.177 INFO:teuthology.orchestra.run.vm00.stdout:(118/150): python3-isodate-0.6.1-3.el9.noarch.r 7.1 MB/s | 56 kB 00:00 2026-03-06T13:38:55.177 INFO:teuthology.orchestra.run.vm00.stdout:(119/150): python3-jaraco-8.2.1-3.el9.noarch.rp 3.8 MB/s | 11 kB 00:00 2026-03-06T13:38:55.179 INFO:teuthology.orchestra.run.vm00.stdout:(120/150): python3-jaraco-classes-3.2.1-5.el9.n 8.3 MB/s | 18 kB 00:00 2026-03-06T13:38:55.180 INFO:teuthology.orchestra.run.vm00.stdout:(121/150): python3-jaraco-collections-3.0.0-8.e 10 MB/s | 23 kB 00:00 2026-03-06T13:38:55.181 INFO:teuthology.orchestra.run.vm00.stdout:(122/150): python3-jaraco-context-6.0.1-3.el9.n 9.2 MB/s | 20 kB 00:00 2026-03-06T13:38:55.182 INFO:teuthology.orchestra.run.vm00.stdout:(123/150): python3-jaraco-functools-3.5.0-2.el9 8.8 MB/s | 19 kB 00:00 2026-03-06T13:38:55.184 INFO:teuthology.orchestra.run.vm00.stdout:(124/150): python3-jaraco-text-4.0.0-2.el9.noar 11 MB/s | 26 kB 00:00 2026-03-06T13:38:55.187 INFO:teuthology.orchestra.run.vm00.stdout:(125/150): python3-logutils-0.3.5-21.el9.noarch 15 MB/s | 46 kB 00:00 2026-03-06T13:38:55.191 INFO:teuthology.orchestra.run.vm00.stdout:(126/150): python3-more-itertools-8.12.0-2.el9. 23 MB/s | 79 kB 00:00 2026-03-06T13:38:55.194 INFO:teuthology.orchestra.run.vm00.stdout:(127/150): python3-msgpack-1.0.3-2.el9.x86_64.r 30 MB/s | 86 kB 00:00 2026-03-06T13:38:55.198 INFO:teuthology.orchestra.run.vm00.stdout:(128/150): python3-natsort-7.1.1-5.el9.noarch.r 16 MB/s | 58 kB 00:00 2026-03-06T13:38:55.203 INFO:teuthology.orchestra.run.vm00.stdout:(129/150): python3-pecan-1.4.2-3.el9.noarch.rpm 50 MB/s | 272 kB 00:00 2026-03-06T13:38:55.205 INFO:teuthology.orchestra.run.vm00.stdout:(130/150): python3-portend-3.1.0-2.el9.noarch.r 7.9 MB/s | 16 kB 00:00 2026-03-06T13:38:55.209 INFO:teuthology.orchestra.run.vm00.stdout:(131/150): python3-pyOpenSSL-21.0.0-1.el9.noarc 26 MB/s | 90 kB 00:00 2026-03-06T13:38:55.212 INFO:teuthology.orchestra.run.vm00.stdout:(132/150): python3-repoze-lru-0.7-16.el9.noarch 12 MB/s | 31 kB 00:00 2026-03-06T13:38:55.223 INFO:teuthology.orchestra.run.vm00.stdout:(133/150): python3-routes-2.5.1-5.el9.noarch.rp 16 MB/s | 188 kB 00:00 2026-03-06T13:38:55.224 INFO:teuthology.orchestra.run.vm08.stdout:(80/150): python3-numpy-1.23.5-2.el9.x86_64.rpm 22 MB/s | 6.1 MB 00:00 2026-03-06T13:38:55.226 INFO:teuthology.orchestra.run.vm00.stdout:(134/150): python3-rsa-4.9-2.el9.noarch.rpm 21 MB/s | 59 kB 00:00 2026-03-06T13:38:55.230 INFO:teuthology.orchestra.run.vm00.stdout:(135/150): python3-saml-1.16.0-1.el9.noarch.rpm 35 MB/s | 125 kB 00:00 2026-03-06T13:38:55.232 INFO:teuthology.orchestra.run.vm01.stdout:(49/150): python3-cryptography-36.0.1-5.el9.x86 5.3 MB/s | 1.2 MB 00:00 2026-03-06T13:38:55.233 INFO:teuthology.orchestra.run.vm00.stdout:(136/150): python3-tempora-5.0.0-2.el9.noarch.r 15 MB/s | 36 kB 00:00 2026-03-06T13:38:55.235 INFO:teuthology.orchestra.run.vm00.stdout:(137/150): python3-kubernetes-26.1.0-3.el9.noar 19 MB/s | 1.0 MB 00:00 2026-03-06T13:38:55.240 INFO:teuthology.orchestra.run.vm00.stdout:(138/150): python3-typing-extensions-4.15.0-1.e 12 MB/s | 86 kB 00:00 2026-03-06T13:38:55.244 INFO:teuthology.orchestra.run.vm00.stdout:(139/150): python3-webob-1.8.8-2.el9.noarch.rpm 28 MB/s | 230 kB 00:00 2026-03-06T13:38:55.245 INFO:teuthology.orchestra.run.vm00.stdout:(140/150): python3-websocket-client-1.2.3-2.el9 20 MB/s | 90 kB 00:00 2026-03-06T13:38:55.247 INFO:teuthology.orchestra.run.vm00.stdout:(141/150): python3-xmlsec-1.3.13-1.el9.x86_64.r 19 MB/s | 48 kB 00:00 2026-03-06T13:38:55.250 INFO:teuthology.orchestra.run.vm00.stdout:(142/150): python3-xmltodict-0.12.0-15.el9.noar 9.8 MB/s | 22 kB 00:00 2026-03-06T13:38:55.253 INFO:teuthology.orchestra.run.vm00.stdout:(143/150): python3-zc-lockfile-2.0-10.el9.noarc 8.1 MB/s | 20 kB 00:00 2026-03-06T13:38:55.256 INFO:teuthology.orchestra.run.vm00.stdout:(144/150): python3-werkzeug-2.0.3-3.el9.1.noarc 34 MB/s | 427 kB 00:00 2026-03-06T13:38:55.260 INFO:teuthology.orchestra.run.vm00.stdout:(145/150): re2-20211101-20.el9.x86_64.rpm 25 MB/s | 191 kB 00:00 2026-03-06T13:38:55.268 INFO:teuthology.orchestra.run.vm00.stdout:(146/150): s3cmd-2.4.0-1.el9.noarch.rpm 17 MB/s | 206 kB 00:00 2026-03-06T13:38:55.268 INFO:teuthology.orchestra.run.vm01.stdout:(50/150): python3-urllib3-1.26.5-7.el9.noarch.r 6.0 MB/s | 218 kB 00:00 2026-03-06T13:38:55.292 INFO:teuthology.orchestra.run.vm00.stdout:(147/150): thrift-0.15.0-4.el9.x86_64.rpm 50 MB/s | 1.6 MB 00:00 2026-03-06T13:38:55.296 INFO:teuthology.orchestra.run.vm00.stdout:(148/150): protobuf-compiler-3.14.0-17.el9.x86_ 2.4 MB/s | 862 kB 00:00 2026-03-06T13:38:55.329 INFO:teuthology.orchestra.run.vm01.stdout:(51/150): smartmontools-7.2-10.el9.x86_64.rpm 9.0 MB/s | 556 kB 00:00 2026-03-06T13:38:55.378 INFO:teuthology.orchestra.run.vm01.stdout:(52/150): python3-requests-2.25.1-10.el9.noarch 496 kB/s | 126 kB 00:00 2026-03-06T13:38:55.396 INFO:teuthology.orchestra.run.vm01.stdout:(53/150): unzip-6.0-59.el9.x86_64.rpm 2.7 MB/s | 182 kB 00:00 2026-03-06T13:38:55.464 INFO:teuthology.orchestra.run.vm01.stdout:(54/150): zip-3.0-35.el9.x86_64.rpm 3.0 MB/s | 266 kB 00:00 2026-03-06T13:38:55.558 INFO:teuthology.orchestra.run.vm01.stdout:(55/150): flexiblas-3.0.4-9.el9.x86_64.rpm 319 kB/s | 30 kB 00:00 2026-03-06T13:38:55.582 INFO:teuthology.orchestra.run.vm01.stdout:(56/150): boost-program-options-1.75.0-13.el9.x 560 kB/s | 104 kB 00:00 2026-03-06T13:38:55.614 INFO:teuthology.orchestra.run.vm01.stdout:(57/150): flexiblas-openblas-openmp-3.0.4-9.el9 469 kB/s | 15 kB 00:00 2026-03-06T13:38:55.636 INFO:teuthology.orchestra.run.vm08.stdout:(81/150): python3-packaging-20.9-5.el9.noarch.r 156 kB/s | 77 kB 00:00 2026-03-06T13:38:55.645 INFO:teuthology.orchestra.run.vm00.stdout:(149/150): librbd1-19.2.3-47.gc24117fd552.el9.c 9.0 MB/s | 3.2 MB 00:00 2026-03-06T13:38:55.708 INFO:teuthology.orchestra.run.vm01.stdout:(58/150): libnbd-1.20.3-4.el9.x86_64.rpm 1.7 MB/s | 164 kB 00:00 2026-03-06T13:38:55.743 INFO:teuthology.orchestra.run.vm08.stdout:(82/150): python3-numpy-f2py-1.23.5-2.el9.x86_6 624 kB/s | 442 kB 00:00 2026-03-06T13:38:55.754 INFO:teuthology.orchestra.run.vm00.stdout:(150/150): librados2-19.2.3-47.gc24117fd552.el9 7.1 MB/s | 3.4 MB 00:00 2026-03-06T13:38:55.760 INFO:teuthology.orchestra.run.vm00.stdout:-------------------------------------------------------------------------------- 2026-03-06T13:38:55.760 INFO:teuthology.orchestra.run.vm00.stdout:Total 25 MB/s | 214 MB 00:08 2026-03-06T13:38:55.773 INFO:teuthology.orchestra.run.vm01.stdout:(59/150): libpmemobj-1.12.1-1.el9.x86_64.rpm 2.4 MB/s | 160 kB 00:00 2026-03-06T13:38:55.818 INFO:teuthology.orchestra.run.vm01.stdout:(60/150): librabbitmq-0.11.0-7.el9.x86_64.rpm 1.0 MB/s | 45 kB 00:00 2026-03-06T13:38:55.842 INFO:teuthology.orchestra.run.vm08.stdout:(83/150): python3-protobuf-3.14.0-17.el9.noarch 434 kB/s | 267 kB 00:00 2026-03-06T13:38:56.037 INFO:teuthology.orchestra.run.vm08.stdout:(84/150): python3-pyasn1-modules-0.4.8-7.el9.no 945 kB/s | 277 kB 00:00 2026-03-06T13:38:56.038 INFO:teuthology.orchestra.run.vm01.stdout:(61/150): librdkafka-1.6.1-102.el9.x86_64.rpm 2.9 MB/s | 662 kB 00:00 2026-03-06T13:38:56.132 INFO:teuthology.orchestra.run.vm01.stdout:(62/150): libstoragemgmt-1.10.1-1.el9.x86_64.rp 2.6 MB/s | 246 kB 00:00 2026-03-06T13:38:56.192 INFO:teuthology.orchestra.run.vm08.stdout:(85/150): python3-pyasn1-0.4.8-7.el9.noarch.rpm 283 kB/s | 157 kB 00:00 2026-03-06T13:38:56.226 INFO:teuthology.orchestra.run.vm01.stdout:(63/150): libxslt-1.1.34-12.el9.x86_64.rpm 2.4 MB/s | 233 kB 00:00 2026-03-06T13:38:56.280 INFO:teuthology.orchestra.run.vm08.stdout:(86/150): python3-requests-oauthlib-1.3.0-12.el 123 kB/s | 54 kB 00:00 2026-03-06T13:38:56.321 INFO:teuthology.orchestra.run.vm01.stdout:(64/150): lttng-ust-2.12.0-6.el9.x86_64.rpm 3.0 MB/s | 292 kB 00:00 2026-03-06T13:38:56.385 INFO:teuthology.orchestra.run.vm01.stdout:(65/150): lua-5.4.4-4.el9.x86_64.rpm 2.9 MB/s | 188 kB 00:00 2026-03-06T13:38:56.417 INFO:teuthology.orchestra.run.vm01.stdout:(66/150): openblas-0.3.29-1.el9.x86_64.rpm 1.3 MB/s | 42 kB 00:00 2026-03-06T13:38:56.469 INFO:teuthology.orchestra.run.vm00.stdout:Running transaction check 2026-03-06T13:38:56.536 INFO:teuthology.orchestra.run.vm00.stdout:Transaction check succeeded. 2026-03-06T13:38:56.536 INFO:teuthology.orchestra.run.vm00.stdout:Running transaction test 2026-03-06T13:38:56.614 INFO:teuthology.orchestra.run.vm01.stdout:(67/150): flexiblas-netlib-3.0.4-9.el9.x86_64.r 2.8 MB/s | 3.0 MB 00:01 2026-03-06T13:38:56.614 INFO:teuthology.orchestra.run.vm08.stdout:(87/150): python3-toml-0.10.2-6.el9.noarch.rpm 99 kB/s | 42 kB 00:00 2026-03-06T13:38:56.646 INFO:teuthology.orchestra.run.vm01.stdout:(68/150): perl-Benchmark-1.23-483.el9.noarch.rp 830 kB/s | 26 kB 00:00 2026-03-06T13:38:56.741 INFO:teuthology.orchestra.run.vm01.stdout:(69/150): perl-Test-Harness-3.42-461.el9.noarch 3.0 MB/s | 295 kB 00:00 2026-03-06T13:38:56.908 INFO:teuthology.orchestra.run.vm08.stdout:(88/150): qatlib-25.08.0-2.el9.x86_64.rpm 382 kB/s | 240 kB 00:00 2026-03-06T13:38:57.009 INFO:teuthology.orchestra.run.vm01.stdout:(70/150): ceph-test-19.2.3-47.gc24117fd552.el9. 11 MB/s | 50 MB 00:04 2026-03-06T13:38:57.136 INFO:teuthology.orchestra.run.vm01.stdout:(71/150): protobuf-3.14.0-17.el9.x86_64.rpm 2.5 MB/s | 1.0 MB 00:00 2026-03-06T13:38:57.233 INFO:teuthology.orchestra.run.vm01.stdout:(72/150): python3-devel-3.9.25-3.el9.x86_64.rpm 2.5 MB/s | 244 kB 00:00 2026-03-06T13:38:57.309 INFO:teuthology.orchestra.run.vm08.stdout:(89/150): python3-scipy-1.9.3-2.el9.x86_64.rpm 15 MB/s | 19 MB 00:01 2026-03-06T13:38:57.329 INFO:teuthology.orchestra.run.vm01.stdout:(73/150): python3-jinja2-2.11.3-8.el9.noarch.rp 2.5 MB/s | 249 kB 00:00 2026-03-06T13:38:57.380 INFO:teuthology.orchestra.run.vm08.stdout:(90/150): qatlib-service-25.08.0-2.el9.x86_64.r 48 kB/s | 37 kB 00:00 2026-03-06T13:38:57.392 INFO:teuthology.orchestra.run.vm01.stdout:(74/150): python3-libstoragemgmt-1.10.1-1.el9.x 2.7 MB/s | 177 kB 00:00 2026-03-06T13:38:57.441 INFO:teuthology.orchestra.run.vm08.stdout:(91/150): socat-1.7.4.1-8.el9.x86_64.rpm 2.3 MB/s | 303 kB 00:00 2026-03-06T13:38:57.456 INFO:teuthology.orchestra.run.vm00.stdout:Transaction test succeeded. 2026-03-06T13:38:57.456 INFO:teuthology.orchestra.run.vm00.stdout:Running transaction 2026-03-06T13:38:57.512 INFO:teuthology.orchestra.run.vm08.stdout:(92/150): xmlsec1-1.2.29-13.el9.x86_64.rpm 1.4 MB/s | 189 kB 00:00 2026-03-06T13:38:57.570 INFO:teuthology.orchestra.run.vm08.stdout:(93/150): xmlsec1-openssl-1.2.29-13.el9.x86_64. 704 kB/s | 90 kB 00:00 2026-03-06T13:38:57.638 INFO:teuthology.orchestra.run.vm08.stdout:(94/150): lua-devel-5.4.4-4.el9.x86_64.rpm 330 kB/s | 22 kB 00:00 2026-03-06T13:38:57.675 INFO:teuthology.orchestra.run.vm01.stdout:(75/150): python3-lxml-4.6.5-3.el9.x86_64.rpm 4.3 MB/s | 1.2 MB 00:00 2026-03-06T13:38:57.738 INFO:teuthology.orchestra.run.vm01.stdout:(76/150): python3-mako-1.1.4-6.el9.noarch.rpm 2.6 MB/s | 172 kB 00:00 2026-03-06T13:38:57.759 INFO:teuthology.orchestra.run.vm08.stdout:(95/150): protobuf-compiler-3.14.0-17.el9.x86_6 7.0 MB/s | 862 kB 00:00 2026-03-06T13:38:57.770 INFO:teuthology.orchestra.run.vm01.stdout:(77/150): python3-markupsafe-1.1.1-12.el9.x86_6 1.1 MB/s | 35 kB 00:00 2026-03-06T13:38:57.773 INFO:teuthology.orchestra.run.vm08.stdout:(96/150): abseil-cpp-20211102.0-4.el9.x86_64.rp 40 MB/s | 551 kB 00:00 2026-03-06T13:38:57.779 INFO:teuthology.orchestra.run.vm08.stdout:(97/150): gperftools-libs-2.9.1-3.el9.x86_64.rp 47 MB/s | 308 kB 00:00 2026-03-06T13:38:57.782 INFO:teuthology.orchestra.run.vm08.stdout:(98/150): grpc-data-1.46.7-10.el9.noarch.rpm 9.6 MB/s | 19 kB 00:00 2026-03-06T13:38:57.850 INFO:teuthology.orchestra.run.vm08.stdout:(99/150): libarrow-9.0.0-15.el9.x86_64.rpm 65 MB/s | 4.4 MB 00:00 2026-03-06T13:38:57.851 INFO:teuthology.orchestra.run.vm08.stdout:(100/150): qatzip-libs-1.3.1-1.el9.x86_64.rpm 70 kB/s | 66 kB 00:00 2026-03-06T13:38:57.853 INFO:teuthology.orchestra.run.vm08.stdout:(101/150): libarrow-doc-9.0.0-15.el9.noarch.rpm 11 MB/s | 25 kB 00:00 2026-03-06T13:38:57.855 INFO:teuthology.orchestra.run.vm08.stdout:(102/150): libunwind-1.6.2-1.el9.x86_64.rpm 27 MB/s | 67 kB 00:00 2026-03-06T13:38:57.858 INFO:teuthology.orchestra.run.vm08.stdout:(103/150): liboath-2.6.12-1.el9.x86_64.rpm 7.9 MB/s | 49 kB 00:00 2026-03-06T13:38:57.859 INFO:teuthology.orchestra.run.vm08.stdout:(104/150): luarocks-3.9.2-5.el9.noarch.rpm 44 MB/s | 151 kB 00:00 2026-03-06T13:38:57.874 INFO:teuthology.orchestra.run.vm08.stdout:(105/150): python3-asyncssh-2.13.2-5.el9.noarch 38 MB/s | 548 kB 00:00 2026-03-06T13:38:57.879 INFO:teuthology.orchestra.run.vm08.stdout:(106/150): parquet-libs-9.0.0-15.el9.x86_64.rpm 39 MB/s | 838 kB 00:00 2026-03-06T13:38:57.879 INFO:teuthology.orchestra.run.vm08.stdout:(107/150): python3-autocommand-2.2.2-8.el9.noar 5.1 MB/s | 29 kB 00:00 2026-03-06T13:38:57.882 INFO:teuthology.orchestra.run.vm08.stdout:(108/150): python3-backports-tarfile-1.2.0-1.el 18 MB/s | 60 kB 00:00 2026-03-06T13:38:57.883 INFO:teuthology.orchestra.run.vm08.stdout:(109/150): python3-bcrypt-3.2.2-1.el9.x86_64.rp 12 MB/s | 43 kB 00:00 2026-03-06T13:38:57.885 INFO:teuthology.orchestra.run.vm08.stdout:(110/150): python3-cachetools-4.2.4-1.el9.noarc 14 MB/s | 32 kB 00:00 2026-03-06T13:38:57.885 INFO:teuthology.orchestra.run.vm08.stdout:(111/150): python3-certifi-2023.05.07-4.el9.noa 6.5 MB/s | 14 kB 00:00 2026-03-06T13:38:57.889 INFO:teuthology.orchestra.run.vm08.stdout:(112/150): python3-cheroot-10.0.1-4.el9.noarch. 47 MB/s | 173 kB 00:00 2026-03-06T13:38:57.892 INFO:teuthology.orchestra.run.vm08.stdout:(113/150): python3-cherrypy-18.6.1-2.el9.noarch 53 MB/s | 358 kB 00:00 2026-03-06T13:38:57.894 INFO:teuthology.orchestra.run.vm08.stdout:(114/150): python3-google-auth-2.45.0-1.el9.noa 47 MB/s | 254 kB 00:00 2026-03-06T13:38:57.900 INFO:teuthology.orchestra.run.vm08.stdout:(115/150): python3-grpcio-tools-1.46.7-10.el9.x 27 MB/s | 144 kB 00:00 2026-03-06T13:38:57.906 INFO:teuthology.orchestra.run.vm08.stdout:(116/150): python3-influxdb-5.3.1-1.el9.noarch. 22 MB/s | 139 kB 00:00 2026-03-06T13:38:57.912 INFO:teuthology.orchestra.run.vm08.stdout:(117/150): python3-isodate-0.6.1-3.el9.noarch.r 10 MB/s | 56 kB 00:00 2026-03-06T13:38:57.917 INFO:teuthology.orchestra.run.vm08.stdout:(118/150): python3-jaraco-8.2.1-3.el9.noarch.rp 2.3 MB/s | 11 kB 00:00 2026-03-06T13:38:57.924 INFO:teuthology.orchestra.run.vm08.stdout:(119/150): python3-grpcio-1.46.7-10.el9.x86_64. 65 MB/s | 2.0 MB 00:00 2026-03-06T13:38:57.925 INFO:teuthology.orchestra.run.vm08.stdout:(120/150): python3-jaraco-classes-3.2.1-5.el9.n 2.2 MB/s | 18 kB 00:00 2026-03-06T13:38:57.927 INFO:teuthology.orchestra.run.vm08.stdout:(121/150): python3-jaraco-collections-3.0.0-8.e 10 MB/s | 23 kB 00:00 2026-03-06T13:38:57.927 INFO:teuthology.orchestra.run.vm08.stdout:(122/150): python3-jaraco-context-6.0.1-3.el9.n 8.9 MB/s | 20 kB 00:00 2026-03-06T13:38:57.929 INFO:teuthology.orchestra.run.vm08.stdout:(123/150): python3-jaraco-functools-3.5.0-2.el9 9.3 MB/s | 19 kB 00:00 2026-03-06T13:38:57.930 INFO:teuthology.orchestra.run.vm08.stdout:(124/150): python3-jaraco-text-4.0.0-2.el9.noar 12 MB/s | 26 kB 00:00 2026-03-06T13:38:57.936 INFO:teuthology.orchestra.run.vm08.stdout:(125/150): python3-logutils-0.3.5-21.el9.noarch 8.3 MB/s | 46 kB 00:00 2026-03-06T13:38:57.942 INFO:teuthology.orchestra.run.vm08.stdout:(126/150): python3-more-itertools-8.12.0-2.el9. 13 MB/s | 79 kB 00:00 2026-03-06T13:38:57.945 INFO:teuthology.orchestra.run.vm08.stdout:(127/150): python3-kubernetes-26.1.0-3.el9.noar 64 MB/s | 1.0 MB 00:00 2026-03-06T13:38:57.946 INFO:teuthology.orchestra.run.vm08.stdout:(128/150): xmlstarlet-1.6.1-20.el9.x86_64.rpm 146 kB/s | 64 kB 00:00 2026-03-06T13:38:57.948 INFO:teuthology.orchestra.run.vm08.stdout:(129/150): python3-msgpack-1.0.3-2.el9.x86_64.r 16 MB/s | 86 kB 00:00 2026-03-06T13:38:57.949 INFO:teuthology.orchestra.run.vm08.stdout:(130/150): python3-natsort-7.1.1-5.el9.noarch.r 20 MB/s | 58 kB 00:00 2026-03-06T13:38:57.950 INFO:teuthology.orchestra.run.vm08.stdout:(131/150): python3-portend-3.1.0-2.el9.noarch.r 7.7 MB/s | 16 kB 00:00 2026-03-06T13:38:57.951 INFO:teuthology.orchestra.run.vm08.stdout:(132/150): python3-pyOpenSSL-21.0.0-1.el9.noarc 33 MB/s | 90 kB 00:00 2026-03-06T13:38:57.959 INFO:teuthology.orchestra.run.vm01.stdout:(78/150): openblas-openmp-0.3.29-1.el9.x86_64.r 3.4 MB/s | 5.3 MB 00:01 2026-03-06T13:38:57.959 INFO:teuthology.orchestra.run.vm08.stdout:(133/150): python3-repoze-lru-0.7-16.el9.noarch 3.6 MB/s | 31 kB 00:00 2026-03-06T13:38:57.960 INFO:teuthology.orchestra.run.vm08.stdout:(134/150): python3-routes-2.5.1-5.el9.noarch.rp 22 MB/s | 188 kB 00:00 2026-03-06T13:38:57.961 INFO:teuthology.orchestra.run.vm08.stdout:(135/150): python3-rsa-4.9-2.el9.noarch.rpm 23 MB/s | 59 kB 00:00 2026-03-06T13:38:57.964 INFO:teuthology.orchestra.run.vm08.stdout:(136/150): python3-saml-1.16.0-1.el9.noarch.rpm 36 MB/s | 125 kB 00:00 2026-03-06T13:38:57.965 INFO:teuthology.orchestra.run.vm08.stdout:(137/150): python3-pecan-1.4.2-3.el9.noarch.rpm 15 MB/s | 272 kB 00:00 2026-03-06T13:38:57.966 INFO:teuthology.orchestra.run.vm08.stdout:(138/150): python3-tempora-5.0.0-2.el9.noarch.r 8.2 MB/s | 36 kB 00:00 2026-03-06T13:38:57.968 INFO:teuthology.orchestra.run.vm08.stdout:(139/150): python3-typing-extensions-4.15.0-1.e 24 MB/s | 86 kB 00:00 2026-03-06T13:38:57.969 INFO:teuthology.orchestra.run.vm08.stdout:(140/150): python3-websocket-client-1.2.3-2.el9 30 MB/s | 90 kB 00:00 2026-03-06T13:38:57.975 INFO:teuthology.orchestra.run.vm08.stdout:(141/150): python3-werkzeug-2.0.3-3.el9.1.noarc 59 MB/s | 427 kB 00:00 2026-03-06T13:38:57.976 INFO:teuthology.orchestra.run.vm08.stdout:(142/150): python3-xmlsec-1.3.13-1.el9.x86_64.r 7.0 MB/s | 48 kB 00:00 2026-03-06T13:38:57.977 INFO:teuthology.orchestra.run.vm08.stdout:(143/150): python3-webob-1.8.8-2.el9.noarch.rpm 19 MB/s | 230 kB 00:00 2026-03-06T13:38:57.978 INFO:teuthology.orchestra.run.vm08.stdout:(144/150): python3-xmltodict-0.12.0-15.el9.noar 8.7 MB/s | 22 kB 00:00 2026-03-06T13:38:57.978 INFO:teuthology.orchestra.run.vm08.stdout:(145/150): python3-zc-lockfile-2.0-10.el9.noarc 8.4 MB/s | 20 kB 00:00 2026-03-06T13:38:57.983 INFO:teuthology.orchestra.run.vm08.stdout:(146/150): s3cmd-2.4.0-1.el9.noarch.rpm 43 MB/s | 206 kB 00:00 2026-03-06T13:38:57.991 INFO:teuthology.orchestra.run.vm08.stdout:(147/150): re2-20211101-20.el9.x86_64.rpm 14 MB/s | 191 kB 00:00 2026-03-06T13:38:58.004 INFO:teuthology.orchestra.run.vm08.stdout:(148/150): thrift-0.15.0-4.el9.x86_64.rpm 63 MB/s | 1.6 MB 00:00 2026-03-06T13:38:58.058 INFO:teuthology.orchestra.run.vm01.stdout:(79/150): python3-numpy-f2py-1.23.5-2.el9.x86_6 4.4 MB/s | 442 kB 00:00 2026-03-06T13:38:58.091 INFO:teuthology.orchestra.run.vm01.stdout:(80/150): python3-packaging-20.9-5.el9.noarch.r 2.3 MB/s | 77 kB 00:00 2026-03-06T13:38:58.155 INFO:teuthology.orchestra.run.vm01.stdout:(81/150): python3-protobuf-3.14.0-17.el9.noarch 4.1 MB/s | 267 kB 00:00 2026-03-06T13:38:58.189 INFO:teuthology.orchestra.run.vm01.stdout:(82/150): python3-pyasn1-0.4.8-7.el9.noarch.rpm 4.6 MB/s | 157 kB 00:00 2026-03-06T13:38:58.271 INFO:teuthology.orchestra.run.vm01.stdout:(83/150): python3-pyasn1-modules-0.4.8-7.el9.no 3.3 MB/s | 277 kB 00:00 2026-03-06T13:38:58.304 INFO:teuthology.orchestra.run.vm01.stdout:(84/150): python3-requests-oauthlib-1.3.0-12.el 1.6 MB/s | 54 kB 00:00 2026-03-06T13:38:58.420 INFO:teuthology.orchestra.run.vm08.stdout:(149/150): librbd1-19.2.3-47.gc24117fd552.el9.c 7.4 MB/s | 3.2 MB 00:00 2026-03-06T13:38:58.497 INFO:teuthology.orchestra.run.vm08.stdout:(150/150): librados2-19.2.3-47.gc24117fd552.el9 6.7 MB/s | 3.4 MB 00:00 2026-03-06T13:38:58.502 INFO:teuthology.orchestra.run.vm08.stdout:-------------------------------------------------------------------------------- 2026-03-06T13:38:58.502 INFO:teuthology.orchestra.run.vm08.stdout:Total 17 MB/s | 214 MB 00:12 2026-03-06T13:38:58.527 INFO:teuthology.orchestra.run.vm00.stdout: Preparing : 1/1 2026-03-06T13:38:58.548 INFO:teuthology.orchestra.run.vm00.stdout: Installing : python3-more-itertools-8.12.0-2.el9.noarch 1/152 2026-03-06T13:38:58.569 INFO:teuthology.orchestra.run.vm00.stdout: Installing : thrift-0.15.0-4.el9.x86_64 2/152 2026-03-06T13:38:58.766 INFO:teuthology.orchestra.run.vm00.stdout: Installing : lttng-ust-2.12.0-6.el9.x86_64 3/152 2026-03-06T13:38:58.767 INFO:teuthology.orchestra.run.vm00.stdout: Upgrading : librados2-2:19.2.3-47.gc24117fd552.el9.clyso.x86 4/152 2026-03-06T13:38:58.780 INFO:teuthology.orchestra.run.vm01.stdout:(85/150): python3-numpy-1.23.5-2.el9.x86_64.rpm 6.1 MB/s | 6.1 MB 00:01 2026-03-06T13:38:58.814 INFO:teuthology.orchestra.run.vm01.stdout:(86/150): python3-toml-0.10.2-6.el9.noarch.rpm 1.2 MB/s | 42 kB 00:00 2026-03-06T13:38:58.839 INFO:teuthology.orchestra.run.vm00.stdout: Running scriptlet: librados2-2:19.2.3-47.gc24117fd552.el9.clyso.x86 4/152 2026-03-06T13:38:58.843 INFO:teuthology.orchestra.run.vm00.stdout: Installing : libcephfs2-2:19.2.3-47.gc24117fd552.el9.clyso.x8 5/152 2026-03-06T13:38:58.849 INFO:teuthology.orchestra.run.vm01.stdout:(87/150): qatlib-25.08.0-2.el9.x86_64.rpm 6.9 MB/s | 240 kB 00:00 2026-03-06T13:38:58.881 INFO:teuthology.orchestra.run.vm01.stdout:(88/150): qatlib-service-25.08.0-2.el9.x86_64.r 1.1 MB/s | 37 kB 00:00 2026-03-06T13:38:58.883 INFO:teuthology.orchestra.run.vm00.stdout: Running scriptlet: libcephfs2-2:19.2.3-47.gc24117fd552.el9.clyso.x8 5/152 2026-03-06T13:38:58.892 INFO:teuthology.orchestra.run.vm00.stdout: Installing : python3-rados-2:19.2.3-47.gc24117fd552.el9.clyso 6/152 2026-03-06T13:38:58.904 INFO:teuthology.orchestra.run.vm00.stdout: Installing : libxslt-1.1.34-12.el9.x86_64 7/152 2026-03-06T13:38:58.908 INFO:teuthology.orchestra.run.vm00.stdout: Installing : librdkafka-1.6.1-102.el9.x86_64 8/152 2026-03-06T13:38:58.911 INFO:teuthology.orchestra.run.vm00.stdout: Installing : librabbitmq-0.11.0-7.el9.x86_64 9/152 2026-03-06T13:38:58.913 INFO:teuthology.orchestra.run.vm01.stdout:(89/150): qatzip-libs-1.3.1-1.el9.x86_64.rpm 2.0 MB/s | 66 kB 00:00 2026-03-06T13:38:58.918 INFO:teuthology.orchestra.run.vm00.stdout: Installing : python3-jaraco-8.2.1-3.el9.noarch 10/152 2026-03-06T13:38:58.971 INFO:teuthology.orchestra.run.vm00.stdout: Installing : libnbd-1.20.3-4.el9.x86_64 11/152 2026-03-06T13:38:58.977 INFO:teuthology.orchestra.run.vm01.stdout:(90/150): socat-1.7.4.1-8.el9.x86_64.rpm 4.7 MB/s | 303 kB 00:00 2026-03-06T13:38:58.981 INFO:teuthology.orchestra.run.vm00.stdout: Installing : python3-lxml-4.6.5-3.el9.x86_64 12/152 2026-03-06T13:38:58.993 INFO:teuthology.orchestra.run.vm00.stdout: Installing : xmlsec1-1.2.29-13.el9.x86_64 13/152 2026-03-06T13:38:58.994 INFO:teuthology.orchestra.run.vm00.stdout: Installing : libcephsqlite-2:19.2.3-47.gc24117fd552.el9.clyso 14/152 2026-03-06T13:38:59.011 INFO:teuthology.orchestra.run.vm01.stdout:(91/150): xmlsec1-1.2.29-13.el9.x86_64.rpm 5.4 MB/s | 189 kB 00:00 2026-03-06T13:38:59.039 INFO:teuthology.orchestra.run.vm00.stdout: Running scriptlet: libcephsqlite-2:19.2.3-47.gc24117fd552.el9.clyso 14/152 2026-03-06T13:38:59.041 INFO:teuthology.orchestra.run.vm00.stdout: Installing : libradosstriper1-2:19.2.3-47.gc24117fd552.el9.cl 15/152 2026-03-06T13:38:59.043 INFO:teuthology.orchestra.run.vm01.stdout:(92/150): xmlsec1-openssl-1.2.29-13.el9.x86_64. 2.7 MB/s | 90 kB 00:00 2026-03-06T13:38:59.060 INFO:teuthology.orchestra.run.vm00.stdout: Running scriptlet: libradosstriper1-2:19.2.3-47.gc24117fd552.el9.cl 15/152 2026-03-06T13:38:59.075 INFO:teuthology.orchestra.run.vm01.stdout:(93/150): xmlstarlet-1.6.1-20.el9.x86_64.rpm 2.0 MB/s | 64 kB 00:00 2026-03-06T13:38:59.103 INFO:teuthology.orchestra.run.vm01.stdout:(94/150): python3-babel-2.9.1-2.el9.noarch.rpm 2.8 MB/s | 6.0 MB 00:02 2026-03-06T13:38:59.104 INFO:teuthology.orchestra.run.vm00.stdout: Installing : re2-1:20211101-20.el9.x86_64 16/152 2026-03-06T13:38:59.150 INFO:teuthology.orchestra.run.vm00.stdout: Installing : libarrow-9.0.0-15.el9.x86_64 17/152 2026-03-06T13:38:59.156 INFO:teuthology.orchestra.run.vm00.stdout: Installing : python3-werkzeug-2.0.3-3.el9.1.noarch 18/152 2026-03-06T13:38:59.187 INFO:teuthology.orchestra.run.vm00.stdout: Installing : liboath-2.6.12-1.el9.x86_64 19/152 2026-03-06T13:38:59.205 INFO:teuthology.orchestra.run.vm00.stdout: Installing : python3-pyasn1-0.4.8-7.el9.noarch 20/152 2026-03-06T13:38:59.214 INFO:teuthology.orchestra.run.vm00.stdout: Installing : python3-packaging-20.9-5.el9.noarch 21/152 2026-03-06T13:38:59.226 INFO:teuthology.orchestra.run.vm00.stdout: Installing : python3-markupsafe-1.1.1-12.el9.x86_64 22/152 2026-03-06T13:38:59.228 INFO:teuthology.orchestra.run.vm01.stdout:(95/150): lua-devel-5.4.4-4.el9.x86_64.rpm 147 kB/s | 22 kB 00:00 2026-03-06T13:38:59.234 INFO:teuthology.orchestra.run.vm00.stdout: Installing : protobuf-3.14.0-17.el9.x86_64 23/152 2026-03-06T13:38:59.238 INFO:teuthology.orchestra.run.vm00.stdout: Installing : lua-5.4.4-4.el9.x86_64 24/152 2026-03-06T13:38:59.242 INFO:teuthology.orchestra.run.vm01.stdout:(96/150): abseil-cpp-20211102.0-4.el9.x86_64.rp 38 MB/s | 551 kB 00:00 2026-03-06T13:38:59.244 INFO:teuthology.orchestra.run.vm00.stdout: Installing : flexiblas-3.0.4-9.el9.x86_64 25/152 2026-03-06T13:38:59.248 INFO:teuthology.orchestra.run.vm01.stdout:(97/150): gperftools-libs-2.9.1-3.el9.x86_64.rp 53 MB/s | 308 kB 00:00 2026-03-06T13:38:59.250 INFO:teuthology.orchestra.run.vm01.stdout:(98/150): grpc-data-1.46.7-10.el9.noarch.rpm 9.6 MB/s | 19 kB 00:00 2026-03-06T13:38:59.266 INFO:teuthology.orchestra.run.vm08.stdout:Running transaction check 2026-03-06T13:38:59.279 INFO:teuthology.orchestra.run.vm00.stdout: Installing : unzip-6.0-59.el9.x86_64 26/152 2026-03-06T13:38:59.297 INFO:teuthology.orchestra.run.vm00.stdout: Installing : python3-urllib3-1.26.5-7.el9.noarch 27/152 2026-03-06T13:38:59.302 INFO:teuthology.orchestra.run.vm00.stdout: Installing : python3-requests-2.25.1-10.el9.noarch 28/152 2026-03-06T13:38:59.308 INFO:teuthology.orchestra.run.vm01.stdout:(99/150): libarrow-9.0.0-15.el9.x86_64.rpm 76 MB/s | 4.4 MB 00:00 2026-03-06T13:38:59.311 INFO:teuthology.orchestra.run.vm01.stdout:(100/150): libarrow-doc-9.0.0-15.el9.noarch.rpm 11 MB/s | 25 kB 00:00 2026-03-06T13:38:59.314 INFO:teuthology.orchestra.run.vm01.stdout:(101/150): liboath-2.6.12-1.el9.x86_64.rpm 16 MB/s | 49 kB 00:00 2026-03-06T13:38:59.318 INFO:teuthology.orchestra.run.vm00.stdout: Installing : libquadmath-11.5.0-14.el9.x86_64 29/152 2026-03-06T13:38:59.330 INFO:teuthology.orchestra.run.vm08.stdout:Transaction check succeeded. 2026-03-06T13:38:59.330 INFO:teuthology.orchestra.run.vm08.stdout:Running transaction test 2026-03-06T13:38:59.344 INFO:teuthology.orchestra.run.vm01.stdout:(102/150): libunwind-1.6.2-1.el9.x86_64.rpm 2.2 MB/s | 67 kB 00:00 2026-03-06T13:38:59.347 INFO:teuthology.orchestra.run.vm00.stdout: Installing : libgfortran-11.5.0-14.el9.x86_64 30/152 2026-03-06T13:38:59.348 INFO:teuthology.orchestra.run.vm01.stdout:(103/150): luarocks-3.9.2-5.el9.noarch.rpm 43 MB/s | 151 kB 00:00 2026-03-06T13:38:59.359 INFO:teuthology.orchestra.run.vm01.stdout:(104/150): parquet-libs-9.0.0-15.el9.x86_64.rpm 73 MB/s | 838 kB 00:00 2026-03-06T13:38:59.367 INFO:teuthology.orchestra.run.vm01.stdout:(105/150): python3-asyncssh-2.13.2-5.el9.noarch 66 MB/s | 548 kB 00:00 2026-03-06T13:38:59.369 INFO:teuthology.orchestra.run.vm01.stdout:(106/150): python3-autocommand-2.2.2-8.el9.noar 14 MB/s | 29 kB 00:00 2026-03-06T13:38:59.372 INFO:teuthology.orchestra.run.vm01.stdout:(107/150): python3-backports-tarfile-1.2.0-1.el 25 MB/s | 60 kB 00:00 2026-03-06T13:38:59.375 INFO:teuthology.orchestra.run.vm01.stdout:(108/150): python3-bcrypt-3.2.2-1.el9.x86_64.rp 18 MB/s | 43 kB 00:00 2026-03-06T13:38:59.377 INFO:teuthology.orchestra.run.vm01.stdout:(109/150): python3-cachetools-4.2.4-1.el9.noarc 15 MB/s | 32 kB 00:00 2026-03-06T13:38:59.379 INFO:teuthology.orchestra.run.vm01.stdout:(110/150): python3-certifi-2023.05.07-4.el9.noa 7.3 MB/s | 14 kB 00:00 2026-03-06T13:38:59.381 INFO:teuthology.orchestra.run.vm00.stdout: Installing : ledmon-libs-1.1.0-3.el9.x86_64 31/152 2026-03-06T13:38:59.383 INFO:teuthology.orchestra.run.vm01.stdout:(111/150): python3-cheroot-10.0.1-4.el9.noarch. 42 MB/s | 173 kB 00:00 2026-03-06T13:38:59.390 INFO:teuthology.orchestra.run.vm00.stdout: Installing : python3-ceph-common-2:19.2.3-47.gc24117fd552.el9 32/152 2026-03-06T13:38:59.392 INFO:teuthology.orchestra.run.vm01.stdout:(112/150): python3-cherrypy-18.6.1-2.el9.noarch 41 MB/s | 358 kB 00:00 2026-03-06T13:38:59.397 INFO:teuthology.orchestra.run.vm01.stdout:(113/150): python3-google-auth-2.45.0-1.el9.noa 56 MB/s | 254 kB 00:00 2026-03-06T13:38:59.403 INFO:teuthology.orchestra.run.vm00.stdout: Installing : python3-ceph-argparse-2:19.2.3-47.gc24117fd552.e 33/152 2026-03-06T13:38:59.420 INFO:teuthology.orchestra.run.vm00.stdout: Installing : python3-cephfs-2:19.2.3-47.gc24117fd552.el9.clys 34/152 2026-03-06T13:38:59.423 INFO:teuthology.orchestra.run.vm01.stdout:(114/150): python3-grpcio-1.46.7-10.el9.x86_64. 77 MB/s | 2.0 MB 00:00 2026-03-06T13:38:59.427 INFO:teuthology.orchestra.run.vm01.stdout:(115/150): python3-grpcio-tools-1.46.7-10.el9.x 40 MB/s | 144 kB 00:00 2026-03-06T13:38:59.431 INFO:teuthology.orchestra.run.vm01.stdout:(116/150): python3-influxdb-5.3.1-1.el9.noarch. 41 MB/s | 139 kB 00:00 2026-03-06T13:38:59.432 INFO:teuthology.orchestra.run.vm00.stdout: Installing : python3-requests-oauthlib-1.3.0-12.el9.noarch 35/152 2026-03-06T13:38:59.433 INFO:teuthology.orchestra.run.vm01.stdout:(117/150): python3-isodate-0.6.1-3.el9.noarch.r 21 MB/s | 56 kB 00:00 2026-03-06T13:38:59.436 INFO:teuthology.orchestra.run.vm01.stdout:(118/150): python3-jaraco-8.2.1-3.el9.noarch.rp 5.6 MB/s | 11 kB 00:00 2026-03-06T13:38:59.438 INFO:teuthology.orchestra.run.vm01.stdout:(119/150): python3-jaraco-classes-3.2.1-5.el9.n 8.5 MB/s | 18 kB 00:00 2026-03-06T13:38:59.441 INFO:teuthology.orchestra.run.vm01.stdout:(120/150): python3-jaraco-collections-3.0.0-8.e 8.4 MB/s | 23 kB 00:00 2026-03-06T13:38:59.443 INFO:teuthology.orchestra.run.vm01.stdout:(121/150): python3-jaraco-context-6.0.1-3.el9.n 10 MB/s | 20 kB 00:00 2026-03-06T13:38:59.446 INFO:teuthology.orchestra.run.vm01.stdout:(122/150): python3-jaraco-functools-3.5.0-2.el9 6.1 MB/s | 19 kB 00:00 2026-03-06T13:38:59.450 INFO:teuthology.orchestra.run.vm01.stdout:(123/150): python3-jaraco-text-4.0.0-2.el9.noar 7.5 MB/s | 26 kB 00:00 2026-03-06T13:38:59.465 INFO:teuthology.orchestra.run.vm00.stdout: Installing : zip-3.0-35.el9.x86_64 36/152 2026-03-06T13:38:59.466 INFO:teuthology.orchestra.run.vm01.stdout:(124/150): python3-kubernetes-26.1.0-3.el9.noar 67 MB/s | 1.0 MB 00:00 2026-03-06T13:38:59.468 INFO:teuthology.orchestra.run.vm01.stdout:(125/150): python3-logutils-0.3.5-21.el9.noarch 19 MB/s | 46 kB 00:00 2026-03-06T13:38:59.471 INFO:teuthology.orchestra.run.vm01.stdout:(126/150): python3-more-itertools-8.12.0-2.el9. 28 MB/s | 79 kB 00:00 2026-03-06T13:38:59.474 INFO:teuthology.orchestra.run.vm00.stdout: Installing : luarocks-3.9.2-5.el9.noarch 37/152 2026-03-06T13:38:59.477 INFO:teuthology.orchestra.run.vm01.stdout:(127/150): python3-msgpack-1.0.3-2.el9.x86_64.r 16 MB/s | 86 kB 00:00 2026-03-06T13:38:59.480 INFO:teuthology.orchestra.run.vm01.stdout:(128/150): python3-natsort-7.1.1-5.el9.noarch.r 17 MB/s | 58 kB 00:00 2026-03-06T13:38:59.484 INFO:teuthology.orchestra.run.vm00.stdout: Installing : lua-devel-5.4.4-4.el9.x86_64 38/152 2026-03-06T13:38:59.487 INFO:teuthology.orchestra.run.vm01.stdout:(129/150): python3-pecan-1.4.2-3.el9.noarch.rpm 41 MB/s | 272 kB 00:00 2026-03-06T13:38:59.490 INFO:teuthology.orchestra.run.vm01.stdout:(130/150): python3-portend-3.1.0-2.el9.noarch.r 6.9 MB/s | 16 kB 00:00 2026-03-06T13:38:59.494 INFO:teuthology.orchestra.run.vm01.stdout:(131/150): protobuf-compiler-3.14.0-17.el9.x86_ 2.2 MB/s | 862 kB 00:00 2026-03-06T13:38:59.495 INFO:teuthology.orchestra.run.vm01.stdout:(132/150): python3-pyOpenSSL-21.0.0-1.el9.noarc 16 MB/s | 90 kB 00:00 2026-03-06T13:38:59.500 INFO:teuthology.orchestra.run.vm01.stdout:(133/150): python3-routes-2.5.1-5.el9.noarch.rp 41 MB/s | 188 kB 00:00 2026-03-06T13:38:59.500 INFO:teuthology.orchestra.run.vm01.stdout:(134/150): python3-repoze-lru-0.7-16.el9.noarch 5.2 MB/s | 31 kB 00:00 2026-03-06T13:38:59.503 INFO:teuthology.orchestra.run.vm01.stdout:(135/150): python3-rsa-4.9-2.el9.noarch.rpm 19 MB/s | 59 kB 00:00 2026-03-06T13:38:59.505 INFO:teuthology.orchestra.run.vm01.stdout:(136/150): python3-saml-1.16.0-1.el9.noarch.rpm 25 MB/s | 125 kB 00:00 2026-03-06T13:38:59.506 INFO:teuthology.orchestra.run.vm01.stdout:(137/150): python3-tempora-5.0.0-2.el9.noarch.r 12 MB/s | 36 kB 00:00 2026-03-06T13:38:59.508 INFO:teuthology.orchestra.run.vm01.stdout:(138/150): python3-typing-extensions-4.15.0-1.e 31 MB/s | 86 kB 00:00 2026-03-06T13:38:59.512 INFO:teuthology.orchestra.run.vm01.stdout:(139/150): python3-webob-1.8.8-2.el9.noarch.rpm 39 MB/s | 230 kB 00:00 2026-03-06T13:38:59.513 INFO:teuthology.orchestra.run.vm01.stdout:(140/150): python3-websocket-client-1.2.3-2.el9 22 MB/s | 90 kB 00:00 2026-03-06T13:38:59.515 INFO:teuthology.orchestra.run.vm01.stdout:(141/150): python3-xmlsec-1.3.13-1.el9.x86_64.r 21 MB/s | 48 kB 00:00 2026-03-06T13:38:59.518 INFO:teuthology.orchestra.run.vm01.stdout:(142/150): python3-xmltodict-0.12.0-15.el9.noar 7.7 MB/s | 22 kB 00:00 2026-03-06T13:38:59.520 INFO:teuthology.orchestra.run.vm00.stdout: Installing : protobuf-compiler-3.14.0-17.el9.x86_64 39/152 2026-03-06T13:38:59.521 INFO:teuthology.orchestra.run.vm01.stdout:(143/150): python3-werkzeug-2.0.3-3.el9.1.noarc 50 MB/s | 427 kB 00:00 2026-03-06T13:38:59.522 INFO:teuthology.orchestra.run.vm01.stdout:(144/150): python3-zc-lockfile-2.0-10.el9.noarc 5.2 MB/s | 20 kB 00:00 2026-03-06T13:38:59.527 INFO:teuthology.orchestra.run.vm01.stdout:(145/150): re2-20211101-20.el9.x86_64.rpm 34 MB/s | 191 kB 00:00 2026-03-06T13:38:59.528 INFO:teuthology.orchestra.run.vm01.stdout:(146/150): s3cmd-2.4.0-1.el9.noarch.rpm 34 MB/s | 206 kB 00:00 2026-03-06T13:38:59.549 INFO:teuthology.orchestra.run.vm01.stdout:(147/150): thrift-0.15.0-4.el9.x86_64.rpm 74 MB/s | 1.6 MB 00:00 2026-03-06T13:38:59.595 INFO:teuthology.orchestra.run.vm00.stdout: Installing : python3-mako-1.1.4-6.el9.noarch 40/152 2026-03-06T13:38:59.613 INFO:teuthology.orchestra.run.vm00.stdout: Installing : python3-pyasn1-modules-0.4.8-7.el9.noarch 41/152 2026-03-06T13:38:59.618 INFO:teuthology.orchestra.run.vm00.stdout: Installing : python3-rsa-4.9-2.el9.noarch 42/152 2026-03-06T13:38:59.624 INFO:teuthology.orchestra.run.vm00.stdout: Installing : xmlsec1-openssl-1.2.29-13.el9.x86_64 43/152 2026-03-06T13:38:59.632 INFO:teuthology.orchestra.run.vm00.stdout: Installing : python3-xmlsec-1.3.13-1.el9.x86_64 44/152 2026-03-06T13:38:59.638 INFO:teuthology.orchestra.run.vm00.stdout: Installing : python3-jaraco-classes-3.2.1-5.el9.noarch 45/152 2026-03-06T13:38:59.649 INFO:teuthology.orchestra.run.vm00.stdout: Installing : xmlstarlet-1.6.1-20.el9.x86_64 46/152 2026-03-06T13:38:59.656 INFO:teuthology.orchestra.run.vm00.stdout: Installing : librados-devel-2:19.2.3-47.gc24117fd552.el9.clys 47/152 2026-03-06T13:38:59.661 INFO:teuthology.orchestra.run.vm00.stdout: Installing : python3-zc-lockfile-2.0-10.el9.noarch 48/152 2026-03-06T13:38:59.680 INFO:teuthology.orchestra.run.vm00.stdout: Installing : python3-xmltodict-0.12.0-15.el9.noarch 49/152 2026-03-06T13:38:59.709 INFO:teuthology.orchestra.run.vm00.stdout: Installing : python3-websocket-client-1.2.3-2.el9.noarch 50/152 2026-03-06T13:38:59.718 INFO:teuthology.orchestra.run.vm00.stdout: Installing : python3-webob-1.8.8-2.el9.noarch 51/152 2026-03-06T13:38:59.726 INFO:teuthology.orchestra.run.vm00.stdout: Installing : python3-typing-extensions-4.15.0-1.el9.noarch 52/152 2026-03-06T13:38:59.741 INFO:teuthology.orchestra.run.vm00.stdout: Installing : python3-repoze-lru-0.7-16.el9.noarch 53/152 2026-03-06T13:38:59.755 INFO:teuthology.orchestra.run.vm00.stdout: Installing : python3-routes-2.5.1-5.el9.noarch 54/152 2026-03-06T13:38:59.765 INFO:teuthology.orchestra.run.vm00.stdout: Installing : python3-natsort-7.1.1-5.el9.noarch 55/152 2026-03-06T13:38:59.791 INFO:teuthology.orchestra.run.vm00.stdout: Installing : python3-msgpack-1.0.3-2.el9.x86_64 56/152 2026-03-06T13:38:59.805 INFO:teuthology.orchestra.run.vm00.stdout: Installing : python3-influxdb-5.3.1-1.el9.noarch 57/152 2026-03-06T13:38:59.881 INFO:teuthology.orchestra.run.vm00.stdout: Installing : python3-logutils-0.3.5-21.el9.noarch 58/152 2026-03-06T13:38:59.903 INFO:teuthology.orchestra.run.vm00.stdout: Installing : python3-pecan-1.4.2-3.el9.noarch 59/152 2026-03-06T13:38:59.925 INFO:teuthology.orchestra.run.vm00.stdout: Installing : python3-isodate-0.6.1-3.el9.noarch 60/152 2026-03-06T13:38:59.933 INFO:teuthology.orchestra.run.vm00.stdout: Installing : python3-saml-1.16.0-1.el9.noarch 61/152 2026-03-06T13:38:59.944 INFO:teuthology.orchestra.run.vm00.stdout: Installing : python3-certifi-2023.05.07-4.el9.noarch 62/152 2026-03-06T13:38:59.945 INFO:teuthology.orchestra.run.vm01.stdout:(148/150): librbd1-19.2.3-47.gc24117fd552.el9.c 8.0 MB/s | 3.2 MB 00:00 2026-03-06T13:38:59.999 INFO:teuthology.orchestra.run.vm00.stdout: Installing : python3-cachetools-4.2.4-1.el9.noarch 63/152 2026-03-06T13:39:00.194 INFO:teuthology.orchestra.run.vm08.stdout:Transaction test succeeded. 2026-03-06T13:39:00.195 INFO:teuthology.orchestra.run.vm08.stdout:Running transaction 2026-03-06T13:39:00.467 INFO:teuthology.orchestra.run.vm00.stdout: Installing : python3-google-auth-1:2.45.0-1.el9.noarch 64/152 2026-03-06T13:39:00.499 INFO:teuthology.orchestra.run.vm01.stdout:(149/150): python3-scipy-1.9.3-2.el9.x86_64.rpm 8.8 MB/s | 19 MB 00:02 2026-03-06T13:39:00.509 INFO:teuthology.orchestra.run.vm00.stdout: Installing : python3-kubernetes-1:26.1.0-3.el9.noarch 65/152 2026-03-06T13:39:00.510 INFO:teuthology.orchestra.run.vm01.stdout:(150/150): librados2-19.2.3-47.gc24117fd552.el9 3.5 MB/s | 3.4 MB 00:00 2026-03-06T13:39:00.512 INFO:teuthology.orchestra.run.vm01.stdout:-------------------------------------------------------------------------------- 2026-03-06T13:39:00.512 INFO:teuthology.orchestra.run.vm01.stdout:Total 20 MB/s | 214 MB 00:10 2026-03-06T13:39:00.516 INFO:teuthology.orchestra.run.vm00.stdout: Installing : python3-backports-tarfile-1.2.0-1.el9.noarch 66/152 2026-03-06T13:39:00.525 INFO:teuthology.orchestra.run.vm00.stdout: Installing : python3-jaraco-context-6.0.1-3.el9.noarch 67/152 2026-03-06T13:39:00.531 INFO:teuthology.orchestra.run.vm00.stdout: Installing : python3-autocommand-2.2.2-8.el9.noarch 68/152 2026-03-06T13:39:00.540 INFO:teuthology.orchestra.run.vm00.stdout: Installing : libunwind-1.6.2-1.el9.x86_64 69/152 2026-03-06T13:39:00.544 INFO:teuthology.orchestra.run.vm00.stdout: Installing : gperftools-libs-2.9.1-3.el9.x86_64 70/152 2026-03-06T13:39:00.546 INFO:teuthology.orchestra.run.vm00.stdout: Installing : libarrow-doc-9.0.0-15.el9.noarch 71/152 2026-03-06T13:39:00.582 INFO:teuthology.orchestra.run.vm00.stdout: Installing : grpc-data-1.46.7-10.el9.noarch 72/152 2026-03-06T13:39:00.647 INFO:teuthology.orchestra.run.vm00.stdout: Installing : abseil-cpp-20211102.0-4.el9.x86_64 73/152 2026-03-06T13:39:00.667 INFO:teuthology.orchestra.run.vm00.stdout: Installing : python3-grpcio-1.46.7-10.el9.x86_64 74/152 2026-03-06T13:39:00.682 INFO:teuthology.orchestra.run.vm00.stdout: Installing : socat-1.7.4.1-8.el9.x86_64 75/152 2026-03-06T13:39:00.690 INFO:teuthology.orchestra.run.vm00.stdout: Installing : python3-toml-0.10.2-6.el9.noarch 76/152 2026-03-06T13:39:00.701 INFO:teuthology.orchestra.run.vm00.stdout: Installing : python3-jaraco-functools-3.5.0-2.el9.noarch 77/152 2026-03-06T13:39:00.709 INFO:teuthology.orchestra.run.vm00.stdout: Installing : python3-jaraco-text-4.0.0-2.el9.noarch 78/152 2026-03-06T13:39:00.721 INFO:teuthology.orchestra.run.vm00.stdout: Installing : python3-jaraco-collections-3.0.0-8.el9.noarch 79/152 2026-03-06T13:39:00.727 INFO:teuthology.orchestra.run.vm00.stdout: Installing : python3-tempora-5.0.0-2.el9.noarch 80/152 2026-03-06T13:39:00.769 INFO:teuthology.orchestra.run.vm00.stdout: Installing : python3-portend-3.1.0-2.el9.noarch 81/152 2026-03-06T13:39:00.784 INFO:teuthology.orchestra.run.vm00.stdout: Installing : python3-protobuf-3.14.0-17.el9.noarch 82/152 2026-03-06T13:39:00.828 INFO:teuthology.orchestra.run.vm00.stdout: Installing : python3-grpcio-tools-1.46.7-10.el9.x86_64 83/152 2026-03-06T13:39:01.138 INFO:teuthology.orchestra.run.vm00.stdout: Installing : python3-devel-3.9.25-3.el9.x86_64 84/152 2026-03-06T13:39:01.174 INFO:teuthology.orchestra.run.vm00.stdout: Installing : python3-babel-2.9.1-2.el9.noarch 85/152 2026-03-06T13:39:01.179 INFO:teuthology.orchestra.run.vm00.stdout: Installing : python3-jinja2-2.11.3-8.el9.noarch 86/152 2026-03-06T13:39:01.184 INFO:teuthology.orchestra.run.vm00.stdout: Installing : perl-Benchmark-1.23-483.el9.noarch 87/152 2026-03-06T13:39:01.200 INFO:teuthology.orchestra.run.vm08.stdout: Preparing : 1/1 2026-03-06T13:39:01.215 INFO:teuthology.orchestra.run.vm08.stdout: Installing : python3-more-itertools-8.12.0-2.el9.noarch 1/152 2026-03-06T13:39:01.229 INFO:teuthology.orchestra.run.vm08.stdout: Installing : thrift-0.15.0-4.el9.x86_64 2/152 2026-03-06T13:39:01.250 INFO:teuthology.orchestra.run.vm00.stdout: Installing : openblas-0.3.29-1.el9.x86_64 88/152 2026-03-06T13:39:01.254 INFO:teuthology.orchestra.run.vm00.stdout: Installing : openblas-openmp-0.3.29-1.el9.x86_64 89/152 2026-03-06T13:39:01.280 INFO:teuthology.orchestra.run.vm00.stdout: Installing : flexiblas-openblas-openmp-3.0.4-9.el9.x86_64 90/152 2026-03-06T13:39:01.281 INFO:teuthology.orchestra.run.vm01.stdout:Running transaction check 2026-03-06T13:39:01.345 INFO:teuthology.orchestra.run.vm01.stdout:Transaction check succeeded. 2026-03-06T13:39:01.345 INFO:teuthology.orchestra.run.vm01.stdout:Running transaction test 2026-03-06T13:39:01.405 INFO:teuthology.orchestra.run.vm08.stdout: Installing : lttng-ust-2.12.0-6.el9.x86_64 3/152 2026-03-06T13:39:01.407 INFO:teuthology.orchestra.run.vm08.stdout: Upgrading : librados2-2:19.2.3-47.gc24117fd552.el9.clyso.x86 4/152 2026-03-06T13:39:01.472 INFO:teuthology.orchestra.run.vm08.stdout: Running scriptlet: librados2-2:19.2.3-47.gc24117fd552.el9.clyso.x86 4/152 2026-03-06T13:39:01.474 INFO:teuthology.orchestra.run.vm08.stdout: Installing : libcephfs2-2:19.2.3-47.gc24117fd552.el9.clyso.x8 5/152 2026-03-06T13:39:01.505 INFO:teuthology.orchestra.run.vm08.stdout: Running scriptlet: libcephfs2-2:19.2.3-47.gc24117fd552.el9.clyso.x8 5/152 2026-03-06T13:39:01.514 INFO:teuthology.orchestra.run.vm08.stdout: Installing : python3-rados-2:19.2.3-47.gc24117fd552.el9.clyso 6/152 2026-03-06T13:39:01.526 INFO:teuthology.orchestra.run.vm08.stdout: Installing : libxslt-1.1.34-12.el9.x86_64 7/152 2026-03-06T13:39:01.530 INFO:teuthology.orchestra.run.vm08.stdout: Installing : librdkafka-1.6.1-102.el9.x86_64 8/152 2026-03-06T13:39:01.533 INFO:teuthology.orchestra.run.vm08.stdout: Installing : librabbitmq-0.11.0-7.el9.x86_64 9/152 2026-03-06T13:39:01.539 INFO:teuthology.orchestra.run.vm08.stdout: Installing : python3-jaraco-8.2.1-3.el9.noarch 10/152 2026-03-06T13:39:01.581 INFO:teuthology.orchestra.run.vm08.stdout: Installing : libnbd-1.20.3-4.el9.x86_64 11/152 2026-03-06T13:39:01.590 INFO:teuthology.orchestra.run.vm08.stdout: Installing : python3-lxml-4.6.5-3.el9.x86_64 12/152 2026-03-06T13:39:01.600 INFO:teuthology.orchestra.run.vm08.stdout: Installing : xmlsec1-1.2.29-13.el9.x86_64 13/152 2026-03-06T13:39:01.602 INFO:teuthology.orchestra.run.vm08.stdout: Installing : libcephsqlite-2:19.2.3-47.gc24117fd552.el9.clyso 14/152 2026-03-06T13:39:01.639 INFO:teuthology.orchestra.run.vm08.stdout: Running scriptlet: libcephsqlite-2:19.2.3-47.gc24117fd552.el9.clyso 14/152 2026-03-06T13:39:01.641 INFO:teuthology.orchestra.run.vm08.stdout: Installing : libradosstriper1-2:19.2.3-47.gc24117fd552.el9.cl 15/152 2026-03-06T13:39:01.658 INFO:teuthology.orchestra.run.vm08.stdout: Running scriptlet: libradosstriper1-2:19.2.3-47.gc24117fd552.el9.cl 15/152 2026-03-06T13:39:01.698 INFO:teuthology.orchestra.run.vm08.stdout: Installing : re2-1:20211101-20.el9.x86_64 16/152 2026-03-06T13:39:01.704 INFO:teuthology.orchestra.run.vm00.stdout: Installing : flexiblas-netlib-3.0.4-9.el9.x86_64 91/152 2026-03-06T13:39:01.736 INFO:teuthology.orchestra.run.vm08.stdout: Installing : libarrow-9.0.0-15.el9.x86_64 17/152 2026-03-06T13:39:01.742 INFO:teuthology.orchestra.run.vm08.stdout: Installing : python3-werkzeug-2.0.3-3.el9.1.noarch 18/152 2026-03-06T13:39:01.768 INFO:teuthology.orchestra.run.vm08.stdout: Installing : liboath-2.6.12-1.el9.x86_64 19/152 2026-03-06T13:39:01.783 INFO:teuthology.orchestra.run.vm08.stdout: Installing : python3-pyasn1-0.4.8-7.el9.noarch 20/152 2026-03-06T13:39:01.792 INFO:teuthology.orchestra.run.vm08.stdout: Installing : python3-packaging-20.9-5.el9.noarch 21/152 2026-03-06T13:39:01.803 INFO:teuthology.orchestra.run.vm08.stdout: Installing : python3-markupsafe-1.1.1-12.el9.x86_64 22/152 2026-03-06T13:39:01.803 INFO:teuthology.orchestra.run.vm00.stdout: Installing : python3-numpy-1:1.23.5-2.el9.x86_64 92/152 2026-03-06T13:39:01.810 INFO:teuthology.orchestra.run.vm08.stdout: Installing : protobuf-3.14.0-17.el9.x86_64 23/152 2026-03-06T13:39:01.814 INFO:teuthology.orchestra.run.vm08.stdout: Installing : lua-5.4.4-4.el9.x86_64 24/152 2026-03-06T13:39:01.821 INFO:teuthology.orchestra.run.vm08.stdout: Installing : flexiblas-3.0.4-9.el9.x86_64 25/152 2026-03-06T13:39:01.850 INFO:teuthology.orchestra.run.vm08.stdout: Installing : unzip-6.0-59.el9.x86_64 26/152 2026-03-06T13:39:01.868 INFO:teuthology.orchestra.run.vm08.stdout: Installing : python3-urllib3-1.26.5-7.el9.noarch 27/152 2026-03-06T13:39:01.872 INFO:teuthology.orchestra.run.vm08.stdout: Installing : python3-requests-2.25.1-10.el9.noarch 28/152 2026-03-06T13:39:01.881 INFO:teuthology.orchestra.run.vm08.stdout: Installing : libquadmath-11.5.0-14.el9.x86_64 29/152 2026-03-06T13:39:01.883 INFO:teuthology.orchestra.run.vm08.stdout: Installing : libgfortran-11.5.0-14.el9.x86_64 30/152 2026-03-06T13:39:01.914 INFO:teuthology.orchestra.run.vm08.stdout: Installing : ledmon-libs-1.1.0-3.el9.x86_64 31/152 2026-03-06T13:39:01.922 INFO:teuthology.orchestra.run.vm08.stdout: Installing : python3-ceph-common-2:19.2.3-47.gc24117fd552.el9 32/152 2026-03-06T13:39:01.933 INFO:teuthology.orchestra.run.vm08.stdout: Installing : python3-ceph-argparse-2:19.2.3-47.gc24117fd552.e 33/152 2026-03-06T13:39:01.948 INFO:teuthology.orchestra.run.vm08.stdout: Installing : python3-cephfs-2:19.2.3-47.gc24117fd552.el9.clys 34/152 2026-03-06T13:39:01.957 INFO:teuthology.orchestra.run.vm08.stdout: Installing : python3-requests-oauthlib-1.3.0-12.el9.noarch 35/152 2026-03-06T13:39:01.987 INFO:teuthology.orchestra.run.vm08.stdout: Installing : zip-3.0-35.el9.x86_64 36/152 2026-03-06T13:39:01.993 INFO:teuthology.orchestra.run.vm08.stdout: Installing : luarocks-3.9.2-5.el9.noarch 37/152 2026-03-06T13:39:02.001 INFO:teuthology.orchestra.run.vm08.stdout: Installing : lua-devel-5.4.4-4.el9.x86_64 38/152 2026-03-06T13:39:02.032 INFO:teuthology.orchestra.run.vm08.stdout: Installing : protobuf-compiler-3.14.0-17.el9.x86_64 39/152 2026-03-06T13:39:02.100 INFO:teuthology.orchestra.run.vm08.stdout: Installing : python3-mako-1.1.4-6.el9.noarch 40/152 2026-03-06T13:39:02.120 INFO:teuthology.orchestra.run.vm08.stdout: Installing : python3-pyasn1-modules-0.4.8-7.el9.noarch 41/152 2026-03-06T13:39:02.124 INFO:teuthology.orchestra.run.vm08.stdout: Installing : python3-rsa-4.9-2.el9.noarch 42/152 2026-03-06T13:39:02.131 INFO:teuthology.orchestra.run.vm08.stdout: Installing : xmlsec1-openssl-1.2.29-13.el9.x86_64 43/152 2026-03-06T13:39:02.138 INFO:teuthology.orchestra.run.vm08.stdout: Installing : python3-xmlsec-1.3.13-1.el9.x86_64 44/152 2026-03-06T13:39:02.143 INFO:teuthology.orchestra.run.vm08.stdout: Installing : python3-jaraco-classes-3.2.1-5.el9.noarch 45/152 2026-03-06T13:39:02.153 INFO:teuthology.orchestra.run.vm08.stdout: Installing : xmlstarlet-1.6.1-20.el9.x86_64 46/152 2026-03-06T13:39:02.160 INFO:teuthology.orchestra.run.vm08.stdout: Installing : librados-devel-2:19.2.3-47.gc24117fd552.el9.clys 47/152 2026-03-06T13:39:02.164 INFO:teuthology.orchestra.run.vm08.stdout: Installing : python3-zc-lockfile-2.0-10.el9.noarch 48/152 2026-03-06T13:39:02.183 INFO:teuthology.orchestra.run.vm08.stdout: Installing : python3-xmltodict-0.12.0-15.el9.noarch 49/152 2026-03-06T13:39:02.212 INFO:teuthology.orchestra.run.vm08.stdout: Installing : python3-websocket-client-1.2.3-2.el9.noarch 50/152 2026-03-06T13:39:02.219 INFO:teuthology.orchestra.run.vm08.stdout: Installing : python3-webob-1.8.8-2.el9.noarch 51/152 2026-03-06T13:39:02.227 INFO:teuthology.orchestra.run.vm08.stdout: Installing : python3-typing-extensions-4.15.0-1.el9.noarch 52/152 2026-03-06T13:39:02.241 INFO:teuthology.orchestra.run.vm01.stdout:Transaction test succeeded. 2026-03-06T13:39:02.241 INFO:teuthology.orchestra.run.vm01.stdout:Running transaction 2026-03-06T13:39:02.243 INFO:teuthology.orchestra.run.vm08.stdout: Installing : python3-repoze-lru-0.7-16.el9.noarch 53/152 2026-03-06T13:39:02.258 INFO:teuthology.orchestra.run.vm08.stdout: Installing : python3-routes-2.5.1-5.el9.noarch 54/152 2026-03-06T13:39:02.268 INFO:teuthology.orchestra.run.vm08.stdout: Installing : python3-natsort-7.1.1-5.el9.noarch 55/152 2026-03-06T13:39:02.292 INFO:teuthology.orchestra.run.vm08.stdout: Installing : python3-msgpack-1.0.3-2.el9.x86_64 56/152 2026-03-06T13:39:02.310 INFO:teuthology.orchestra.run.vm08.stdout: Installing : python3-influxdb-5.3.1-1.el9.noarch 57/152 2026-03-06T13:39:02.378 INFO:teuthology.orchestra.run.vm08.stdout: Installing : python3-logutils-0.3.5-21.el9.noarch 58/152 2026-03-06T13:39:02.395 INFO:teuthology.orchestra.run.vm08.stdout: Installing : python3-pecan-1.4.2-3.el9.noarch 59/152 2026-03-06T13:39:02.416 INFO:teuthology.orchestra.run.vm08.stdout: Installing : python3-isodate-0.6.1-3.el9.noarch 60/152 2026-03-06T13:39:02.424 INFO:teuthology.orchestra.run.vm08.stdout: Installing : python3-saml-1.16.0-1.el9.noarch 61/152 2026-03-06T13:39:02.437 INFO:teuthology.orchestra.run.vm08.stdout: Installing : python3-certifi-2023.05.07-4.el9.noarch 62/152 2026-03-06T13:39:02.498 INFO:teuthology.orchestra.run.vm08.stdout: Installing : python3-cachetools-4.2.4-1.el9.noarch 63/152 2026-03-06T13:39:02.724 INFO:teuthology.orchestra.run.vm00.stdout: Installing : python3-numpy-f2py-1:1.23.5-2.el9.x86_64 93/152 2026-03-06T13:39:02.749 INFO:teuthology.orchestra.run.vm00.stdout: Installing : python3-scipy-1.9.3-2.el9.x86_64 94/152 2026-03-06T13:39:02.924 INFO:teuthology.orchestra.run.vm08.stdout: Installing : python3-google-auth-1:2.45.0-1.el9.noarch 64/152 2026-03-06T13:39:02.929 INFO:teuthology.orchestra.run.vm00.stdout: Installing : libpmemobj-1.12.1-1.el9.x86_64 95/152 2026-03-06T13:39:02.934 INFO:teuthology.orchestra.run.vm00.stdout: Upgrading : librbd1-2:19.2.3-47.gc24117fd552.el9.clyso.x86_6 96/152 2026-03-06T13:39:02.940 INFO:teuthology.orchestra.run.vm08.stdout: Installing : python3-kubernetes-1:26.1.0-3.el9.noarch 65/152 2026-03-06T13:39:02.946 INFO:teuthology.orchestra.run.vm08.stdout: Installing : python3-backports-tarfile-1.2.0-1.el9.noarch 66/152 2026-03-06T13:39:02.955 INFO:teuthology.orchestra.run.vm08.stdout: Installing : python3-jaraco-context-6.0.1-3.el9.noarch 67/152 2026-03-06T13:39:02.959 INFO:teuthology.orchestra.run.vm08.stdout: Installing : python3-autocommand-2.2.2-8.el9.noarch 68/152 2026-03-06T13:39:02.967 INFO:teuthology.orchestra.run.vm08.stdout: Installing : libunwind-1.6.2-1.el9.x86_64 69/152 2026-03-06T13:39:02.970 INFO:teuthology.orchestra.run.vm00.stdout: Running scriptlet: librbd1-2:19.2.3-47.gc24117fd552.el9.clyso.x86_6 96/152 2026-03-06T13:39:02.971 INFO:teuthology.orchestra.run.vm08.stdout: Installing : gperftools-libs-2.9.1-3.el9.x86_64 70/152 2026-03-06T13:39:02.973 INFO:teuthology.orchestra.run.vm08.stdout: Installing : libarrow-doc-9.0.0-15.el9.noarch 71/152 2026-03-06T13:39:02.975 INFO:teuthology.orchestra.run.vm00.stdout: Installing : python3-rbd-2:19.2.3-47.gc24117fd552.el9.clyso.x 97/152 2026-03-06T13:39:02.983 INFO:teuthology.orchestra.run.vm00.stdout: Installing : boost-program-options-1.75.0-13.el9.x86_64 98/152 2026-03-06T13:39:03.004 INFO:teuthology.orchestra.run.vm08.stdout: Installing : grpc-data-1.46.7-10.el9.noarch 72/152 2026-03-06T13:39:03.057 INFO:teuthology.orchestra.run.vm08.stdout: Installing : abseil-cpp-20211102.0-4.el9.x86_64 73/152 2026-03-06T13:39:03.072 INFO:teuthology.orchestra.run.vm08.stdout: Installing : python3-grpcio-1.46.7-10.el9.x86_64 74/152 2026-03-06T13:39:03.081 INFO:teuthology.orchestra.run.vm08.stdout: Installing : socat-1.7.4.1-8.el9.x86_64 75/152 2026-03-06T13:39:03.087 INFO:teuthology.orchestra.run.vm08.stdout: Installing : python3-toml-0.10.2-6.el9.noarch 76/152 2026-03-06T13:39:03.094 INFO:teuthology.orchestra.run.vm08.stdout: Installing : python3-jaraco-functools-3.5.0-2.el9.noarch 77/152 2026-03-06T13:39:03.101 INFO:teuthology.orchestra.run.vm08.stdout: Installing : python3-jaraco-text-4.0.0-2.el9.noarch 78/152 2026-03-06T13:39:03.110 INFO:teuthology.orchestra.run.vm08.stdout: Installing : python3-jaraco-collections-3.0.0-8.el9.noarch 79/152 2026-03-06T13:39:03.117 INFO:teuthology.orchestra.run.vm08.stdout: Installing : python3-tempora-5.0.0-2.el9.noarch 80/152 2026-03-06T13:39:03.153 INFO:teuthology.orchestra.run.vm08.stdout: Installing : python3-portend-3.1.0-2.el9.noarch 81/152 2026-03-06T13:39:03.171 INFO:teuthology.orchestra.run.vm08.stdout: Installing : python3-protobuf-3.14.0-17.el9.noarch 82/152 2026-03-06T13:39:03.214 INFO:teuthology.orchestra.run.vm08.stdout: Installing : python3-grpcio-tools-1.46.7-10.el9.x86_64 83/152 2026-03-06T13:39:03.233 INFO:teuthology.orchestra.run.vm01.stdout: Preparing : 1/1 2026-03-06T13:39:03.249 INFO:teuthology.orchestra.run.vm01.stdout: Installing : python3-more-itertools-8.12.0-2.el9.noarch 1/152 2026-03-06T13:39:03.262 INFO:teuthology.orchestra.run.vm01.stdout: Installing : thrift-0.15.0-4.el9.x86_64 2/152 2026-03-06T13:39:03.275 INFO:teuthology.orchestra.run.vm00.stdout: Installing : parquet-libs-9.0.0-15.el9.x86_64 99/152 2026-03-06T13:39:03.279 INFO:teuthology.orchestra.run.vm00.stdout: Installing : librgw2-2:19.2.3-47.gc24117fd552.el9.clyso.x86_6 100/152 2026-03-06T13:39:03.302 INFO:teuthology.orchestra.run.vm00.stdout: Running scriptlet: librgw2-2:19.2.3-47.gc24117fd552.el9.clyso.x86_6 100/152 2026-03-06T13:39:03.304 INFO:teuthology.orchestra.run.vm00.stdout: Installing : python3-rgw-2:19.2.3-47.gc24117fd552.el9.clyso.x 101/152 2026-03-06T13:39:03.440 INFO:teuthology.orchestra.run.vm01.stdout: Installing : lttng-ust-2.12.0-6.el9.x86_64 3/152 2026-03-06T13:39:03.443 INFO:teuthology.orchestra.run.vm01.stdout: Upgrading : librados2-2:19.2.3-47.gc24117fd552.el9.clyso.x86 4/152 2026-03-06T13:39:03.496 INFO:teuthology.orchestra.run.vm08.stdout: Installing : python3-devel-3.9.25-3.el9.x86_64 84/152 2026-03-06T13:39:03.505 INFO:teuthology.orchestra.run.vm01.stdout: Running scriptlet: librados2-2:19.2.3-47.gc24117fd552.el9.clyso.x86 4/152 2026-03-06T13:39:03.508 INFO:teuthology.orchestra.run.vm01.stdout: Installing : libcephfs2-2:19.2.3-47.gc24117fd552.el9.clyso.x8 5/152 2026-03-06T13:39:03.529 INFO:teuthology.orchestra.run.vm08.stdout: Installing : python3-babel-2.9.1-2.el9.noarch 85/152 2026-03-06T13:39:03.534 INFO:teuthology.orchestra.run.vm08.stdout: Installing : python3-jinja2-2.11.3-8.el9.noarch 86/152 2026-03-06T13:39:03.539 INFO:teuthology.orchestra.run.vm01.stdout: Running scriptlet: libcephfs2-2:19.2.3-47.gc24117fd552.el9.clyso.x8 5/152 2026-03-06T13:39:03.540 INFO:teuthology.orchestra.run.vm08.stdout: Installing : perl-Benchmark-1.23-483.el9.noarch 87/152 2026-03-06T13:39:03.545 INFO:teuthology.orchestra.run.vm01.stdout: Installing : python3-rados-2:19.2.3-47.gc24117fd552.el9.clyso 6/152 2026-03-06T13:39:03.557 INFO:teuthology.orchestra.run.vm01.stdout: Installing : libxslt-1.1.34-12.el9.x86_64 7/152 2026-03-06T13:39:03.562 INFO:teuthology.orchestra.run.vm01.stdout: Installing : librdkafka-1.6.1-102.el9.x86_64 8/152 2026-03-06T13:39:03.565 INFO:teuthology.orchestra.run.vm01.stdout: Installing : librabbitmq-0.11.0-7.el9.x86_64 9/152 2026-03-06T13:39:03.571 INFO:teuthology.orchestra.run.vm01.stdout: Installing : python3-jaraco-8.2.1-3.el9.noarch 10/152 2026-03-06T13:39:03.605 INFO:teuthology.orchestra.run.vm08.stdout: Installing : openblas-0.3.29-1.el9.x86_64 88/152 2026-03-06T13:39:03.608 INFO:teuthology.orchestra.run.vm08.stdout: Installing : openblas-openmp-0.3.29-1.el9.x86_64 89/152 2026-03-06T13:39:03.613 INFO:teuthology.orchestra.run.vm01.stdout: Installing : libnbd-1.20.3-4.el9.x86_64 11/152 2026-03-06T13:39:03.623 INFO:teuthology.orchestra.run.vm01.stdout: Installing : python3-lxml-4.6.5-3.el9.x86_64 12/152 2026-03-06T13:39:03.634 INFO:teuthology.orchestra.run.vm01.stdout: Installing : xmlsec1-1.2.29-13.el9.x86_64 13/152 2026-03-06T13:39:03.635 INFO:teuthology.orchestra.run.vm01.stdout: Installing : libcephsqlite-2:19.2.3-47.gc24117fd552.el9.clyso 14/152 2026-03-06T13:39:03.636 INFO:teuthology.orchestra.run.vm08.stdout: Installing : flexiblas-openblas-openmp-3.0.4-9.el9.x86_64 90/152 2026-03-06T13:39:03.672 INFO:teuthology.orchestra.run.vm01.stdout: Running scriptlet: libcephsqlite-2:19.2.3-47.gc24117fd552.el9.clyso 14/152 2026-03-06T13:39:03.673 INFO:teuthology.orchestra.run.vm01.stdout: Installing : libradosstriper1-2:19.2.3-47.gc24117fd552.el9.cl 15/152 2026-03-06T13:39:03.687 INFO:teuthology.orchestra.run.vm01.stdout: Running scriptlet: libradosstriper1-2:19.2.3-47.gc24117fd552.el9.cl 15/152 2026-03-06T13:39:03.720 INFO:teuthology.orchestra.run.vm01.stdout: Installing : re2-1:20211101-20.el9.x86_64 16/152 2026-03-06T13:39:03.758 INFO:teuthology.orchestra.run.vm01.stdout: Installing : libarrow-9.0.0-15.el9.x86_64 17/152 2026-03-06T13:39:03.764 INFO:teuthology.orchestra.run.vm01.stdout: Installing : python3-werkzeug-2.0.3-3.el9.1.noarch 18/152 2026-03-06T13:39:03.791 INFO:teuthology.orchestra.run.vm01.stdout: Installing : liboath-2.6.12-1.el9.x86_64 19/152 2026-03-06T13:39:03.806 INFO:teuthology.orchestra.run.vm01.stdout: Installing : python3-pyasn1-0.4.8-7.el9.noarch 20/152 2026-03-06T13:39:03.814 INFO:teuthology.orchestra.run.vm01.stdout: Installing : python3-packaging-20.9-5.el9.noarch 21/152 2026-03-06T13:39:03.824 INFO:teuthology.orchestra.run.vm01.stdout: Installing : python3-markupsafe-1.1.1-12.el9.x86_64 22/152 2026-03-06T13:39:03.832 INFO:teuthology.orchestra.run.vm01.stdout: Installing : protobuf-3.14.0-17.el9.x86_64 23/152 2026-03-06T13:39:03.836 INFO:teuthology.orchestra.run.vm01.stdout: Installing : lua-5.4.4-4.el9.x86_64 24/152 2026-03-06T13:39:03.842 INFO:teuthology.orchestra.run.vm01.stdout: Installing : flexiblas-3.0.4-9.el9.x86_64 25/152 2026-03-06T13:39:03.874 INFO:teuthology.orchestra.run.vm01.stdout: Installing : unzip-6.0-59.el9.x86_64 26/152 2026-03-06T13:39:03.893 INFO:teuthology.orchestra.run.vm01.stdout: Installing : python3-urllib3-1.26.5-7.el9.noarch 27/152 2026-03-06T13:39:03.898 INFO:teuthology.orchestra.run.vm01.stdout: Installing : python3-requests-2.25.1-10.el9.noarch 28/152 2026-03-06T13:39:03.906 INFO:teuthology.orchestra.run.vm01.stdout: Installing : libquadmath-11.5.0-14.el9.x86_64 29/152 2026-03-06T13:39:03.909 INFO:teuthology.orchestra.run.vm01.stdout: Installing : libgfortran-11.5.0-14.el9.x86_64 30/152 2026-03-06T13:39:03.943 INFO:teuthology.orchestra.run.vm01.stdout: Installing : ledmon-libs-1.1.0-3.el9.x86_64 31/152 2026-03-06T13:39:03.950 INFO:teuthology.orchestra.run.vm01.stdout: Installing : python3-ceph-common-2:19.2.3-47.gc24117fd552.el9 32/152 2026-03-06T13:39:03.963 INFO:teuthology.orchestra.run.vm01.stdout: Installing : python3-ceph-argparse-2:19.2.3-47.gc24117fd552.e 33/152 2026-03-06T13:39:03.980 INFO:teuthology.orchestra.run.vm01.stdout: Installing : python3-cephfs-2:19.2.3-47.gc24117fd552.el9.clys 34/152 2026-03-06T13:39:03.990 INFO:teuthology.orchestra.run.vm01.stdout: Installing : python3-requests-oauthlib-1.3.0-12.el9.noarch 35/152 2026-03-06T13:39:04.025 INFO:teuthology.orchestra.run.vm01.stdout: Installing : zip-3.0-35.el9.x86_64 36/152 2026-03-06T13:39:04.030 INFO:teuthology.orchestra.run.vm01.stdout: Installing : luarocks-3.9.2-5.el9.noarch 37/152 2026-03-06T13:39:04.039 INFO:teuthology.orchestra.run.vm01.stdout: Installing : lua-devel-5.4.4-4.el9.x86_64 38/152 2026-03-06T13:39:04.051 INFO:teuthology.orchestra.run.vm08.stdout: Installing : flexiblas-netlib-3.0.4-9.el9.x86_64 91/152 2026-03-06T13:39:04.072 INFO:teuthology.orchestra.run.vm01.stdout: Installing : protobuf-compiler-3.14.0-17.el9.x86_64 39/152 2026-03-06T13:39:04.142 INFO:teuthology.orchestra.run.vm01.stdout: Installing : python3-mako-1.1.4-6.el9.noarch 40/152 2026-03-06T13:39:04.152 INFO:teuthology.orchestra.run.vm08.stdout: Installing : python3-numpy-1:1.23.5-2.el9.x86_64 92/152 2026-03-06T13:39:04.162 INFO:teuthology.orchestra.run.vm01.stdout: Installing : python3-pyasn1-modules-0.4.8-7.el9.noarch 41/152 2026-03-06T13:39:04.166 INFO:teuthology.orchestra.run.vm01.stdout: Installing : python3-rsa-4.9-2.el9.noarch 42/152 2026-03-06T13:39:04.173 INFO:teuthology.orchestra.run.vm01.stdout: Installing : xmlsec1-openssl-1.2.29-13.el9.x86_64 43/152 2026-03-06T13:39:04.182 INFO:teuthology.orchestra.run.vm01.stdout: Installing : python3-xmlsec-1.3.13-1.el9.x86_64 44/152 2026-03-06T13:39:04.189 INFO:teuthology.orchestra.run.vm01.stdout: Installing : python3-jaraco-classes-3.2.1-5.el9.noarch 45/152 2026-03-06T13:39:04.201 INFO:teuthology.orchestra.run.vm01.stdout: Installing : xmlstarlet-1.6.1-20.el9.x86_64 46/152 2026-03-06T13:39:04.209 INFO:teuthology.orchestra.run.vm01.stdout: Installing : librados-devel-2:19.2.3-47.gc24117fd552.el9.clys 47/152 2026-03-06T13:39:04.213 INFO:teuthology.orchestra.run.vm01.stdout: Installing : python3-zc-lockfile-2.0-10.el9.noarch 48/152 2026-03-06T13:39:04.235 INFO:teuthology.orchestra.run.vm01.stdout: Installing : python3-xmltodict-0.12.0-15.el9.noarch 49/152 2026-03-06T13:39:04.267 INFO:teuthology.orchestra.run.vm01.stdout: Installing : python3-websocket-client-1.2.3-2.el9.noarch 50/152 2026-03-06T13:39:04.275 INFO:teuthology.orchestra.run.vm01.stdout: Installing : python3-webob-1.8.8-2.el9.noarch 51/152 2026-03-06T13:39:04.282 INFO:teuthology.orchestra.run.vm01.stdout: Installing : python3-typing-extensions-4.15.0-1.el9.noarch 52/152 2026-03-06T13:39:04.299 INFO:teuthology.orchestra.run.vm01.stdout: Installing : python3-repoze-lru-0.7-16.el9.noarch 53/152 2026-03-06T13:39:04.312 INFO:teuthology.orchestra.run.vm01.stdout: Installing : python3-routes-2.5.1-5.el9.noarch 54/152 2026-03-06T13:39:04.322 INFO:teuthology.orchestra.run.vm01.stdout: Installing : python3-natsort-7.1.1-5.el9.noarch 55/152 2026-03-06T13:39:04.350 INFO:teuthology.orchestra.run.vm01.stdout: Installing : python3-msgpack-1.0.3-2.el9.x86_64 56/152 2026-03-06T13:39:04.364 INFO:teuthology.orchestra.run.vm01.stdout: Installing : python3-influxdb-5.3.1-1.el9.noarch 57/152 2026-03-06T13:39:04.457 INFO:teuthology.orchestra.run.vm01.stdout: Installing : python3-logutils-0.3.5-21.el9.noarch 58/152 2026-03-06T13:39:04.478 INFO:teuthology.orchestra.run.vm01.stdout: Installing : python3-pecan-1.4.2-3.el9.noarch 59/152 2026-03-06T13:39:04.501 INFO:teuthology.orchestra.run.vm01.stdout: Installing : python3-isodate-0.6.1-3.el9.noarch 60/152 2026-03-06T13:39:04.509 INFO:teuthology.orchestra.run.vm01.stdout: Installing : python3-saml-1.16.0-1.el9.noarch 61/152 2026-03-06T13:39:04.523 INFO:teuthology.orchestra.run.vm01.stdout: Installing : python3-certifi-2023.05.07-4.el9.noarch 62/152 2026-03-06T13:39:04.584 INFO:teuthology.orchestra.run.vm01.stdout: Installing : python3-cachetools-4.2.4-1.el9.noarch 63/152 2026-03-06T13:39:04.596 INFO:teuthology.orchestra.run.vm00.stdout: Running scriptlet: ceph-common-2:19.2.3-47.gc24117fd552.el9.clyso.x 102/152 2026-03-06T13:39:04.610 INFO:teuthology.orchestra.run.vm00.stdout: Installing : ceph-common-2:19.2.3-47.gc24117fd552.el9.clyso.x 102/152 2026-03-06T13:39:04.654 INFO:teuthology.orchestra.run.vm00.stdout: Running scriptlet: ceph-common-2:19.2.3-47.gc24117fd552.el9.clyso.x 102/152 2026-03-06T13:39:04.659 INFO:teuthology.orchestra.run.vm00.stdout: Installing : smartmontools-1:7.2-10.el9.x86_64 103/152 2026-03-06T13:39:04.679 INFO:teuthology.orchestra.run.vm00.stdout: Running scriptlet: smartmontools-1:7.2-10.el9.x86_64 103/152 2026-03-06T13:39:04.679 INFO:teuthology.orchestra.run.vm00.stdout:Created symlink /etc/systemd/system/multi-user.target.wants/smartd.service → /usr/lib/systemd/system/smartd.service. 2026-03-06T13:39:04.679 INFO:teuthology.orchestra.run.vm00.stdout: 2026-03-06T13:39:04.714 INFO:teuthology.orchestra.run.vm00.stdout: Installing : python3-ply-3.11-14.el9.noarch 104/152 2026-03-06T13:39:04.739 INFO:teuthology.orchestra.run.vm00.stdout: Installing : python3-pycparser-2.20-6.el9.noarch 105/152 2026-03-06T13:39:04.840 INFO:teuthology.orchestra.run.vm00.stdout: Installing : python3-cffi-1.14.5-5.el9.x86_64 106/152 2026-03-06T13:39:04.855 INFO:teuthology.orchestra.run.vm00.stdout: Installing : python3-cryptography-36.0.1-5.el9.x86_64 107/152 2026-03-06T13:39:04.888 INFO:teuthology.orchestra.run.vm00.stdout: Installing : python3-pyOpenSSL-21.0.0-1.el9.noarch 108/152 2026-03-06T13:39:04.932 INFO:teuthology.orchestra.run.vm00.stdout: Installing : python3-cheroot-10.0.1-4.el9.noarch 109/152 2026-03-06T13:39:05.000 INFO:teuthology.orchestra.run.vm08.stdout: Installing : python3-numpy-f2py-1:1.23.5-2.el9.x86_64 93/152 2026-03-06T13:39:05.001 INFO:teuthology.orchestra.run.vm00.stdout: Installing : python3-cherrypy-18.6.1-2.el9.noarch 110/152 2026-03-06T13:39:05.012 INFO:teuthology.orchestra.run.vm00.stdout: Installing : python3-asyncssh-2.13.2-5.el9.noarch 111/152 2026-03-06T13:39:05.024 INFO:teuthology.orchestra.run.vm00.stdout: Installing : python3-bcrypt-3.2.2-1.el9.x86_64 112/152 2026-03-06T13:39:05.030 INFO:teuthology.orchestra.run.vm08.stdout: Installing : python3-scipy-1.9.3-2.el9.x86_64 94/152 2026-03-06T13:39:05.031 INFO:teuthology.orchestra.run.vm00.stdout: Installing : pciutils-3.7.0-7.el9.x86_64 113/152 2026-03-06T13:39:05.036 INFO:teuthology.orchestra.run.vm00.stdout: Installing : qatlib-25.08.0-2.el9.x86_64 114/152 2026-03-06T13:39:05.039 INFO:teuthology.orchestra.run.vm00.stdout: Installing : qatlib-service-25.08.0-2.el9.x86_64 115/152 2026-03-06T13:39:05.055 INFO:teuthology.orchestra.run.vm01.stdout: Installing : python3-google-auth-1:2.45.0-1.el9.noarch 64/152 2026-03-06T13:39:05.058 INFO:teuthology.orchestra.run.vm00.stdout: Running scriptlet: qatlib-service-25.08.0-2.el9.x86_64 115/152 2026-03-06T13:39:05.075 INFO:teuthology.orchestra.run.vm01.stdout: Installing : python3-kubernetes-1:26.1.0-3.el9.noarch 65/152 2026-03-06T13:39:05.083 INFO:teuthology.orchestra.run.vm01.stdout: Installing : python3-backports-tarfile-1.2.0-1.el9.noarch 66/152 2026-03-06T13:39:05.093 INFO:teuthology.orchestra.run.vm01.stdout: Installing : python3-jaraco-context-6.0.1-3.el9.noarch 67/152 2026-03-06T13:39:05.098 INFO:teuthology.orchestra.run.vm01.stdout: Installing : python3-autocommand-2.2.2-8.el9.noarch 68/152 2026-03-06T13:39:05.106 INFO:teuthology.orchestra.run.vm01.stdout: Installing : libunwind-1.6.2-1.el9.x86_64 69/152 2026-03-06T13:39:05.110 INFO:teuthology.orchestra.run.vm01.stdout: Installing : gperftools-libs-2.9.1-3.el9.x86_64 70/152 2026-03-06T13:39:05.112 INFO:teuthology.orchestra.run.vm01.stdout: Installing : libarrow-doc-9.0.0-15.el9.noarch 71/152 2026-03-06T13:39:05.148 INFO:teuthology.orchestra.run.vm01.stdout: Installing : grpc-data-1.46.7-10.el9.noarch 72/152 2026-03-06T13:39:05.201 INFO:teuthology.orchestra.run.vm08.stdout: Installing : libpmemobj-1.12.1-1.el9.x86_64 95/152 2026-03-06T13:39:05.204 INFO:teuthology.orchestra.run.vm08.stdout: Upgrading : librbd1-2:19.2.3-47.gc24117fd552.el9.clyso.x86_6 96/152 2026-03-06T13:39:05.210 INFO:teuthology.orchestra.run.vm01.stdout: Installing : abseil-cpp-20211102.0-4.el9.x86_64 73/152 2026-03-06T13:39:05.225 INFO:teuthology.orchestra.run.vm01.stdout: Installing : python3-grpcio-1.46.7-10.el9.x86_64 74/152 2026-03-06T13:39:05.236 INFO:teuthology.orchestra.run.vm01.stdout: Installing : socat-1.7.4.1-8.el9.x86_64 75/152 2026-03-06T13:39:05.242 INFO:teuthology.orchestra.run.vm01.stdout: Installing : python3-toml-0.10.2-6.el9.noarch 76/152 2026-03-06T13:39:05.247 INFO:teuthology.orchestra.run.vm08.stdout: Running scriptlet: librbd1-2:19.2.3-47.gc24117fd552.el9.clyso.x86_6 96/152 2026-03-06T13:39:05.249 INFO:teuthology.orchestra.run.vm01.stdout: Installing : python3-jaraco-functools-3.5.0-2.el9.noarch 77/152 2026-03-06T13:39:05.251 INFO:teuthology.orchestra.run.vm08.stdout: Installing : python3-rbd-2:19.2.3-47.gc24117fd552.el9.clyso.x 97/152 2026-03-06T13:39:05.256 INFO:teuthology.orchestra.run.vm01.stdout: Installing : python3-jaraco-text-4.0.0-2.el9.noarch 78/152 2026-03-06T13:39:05.260 INFO:teuthology.orchestra.run.vm08.stdout: Installing : boost-program-options-1.75.0-13.el9.x86_64 98/152 2026-03-06T13:39:05.267 INFO:teuthology.orchestra.run.vm01.stdout: Installing : python3-jaraco-collections-3.0.0-8.el9.noarch 79/152 2026-03-06T13:39:05.273 INFO:teuthology.orchestra.run.vm01.stdout: Installing : python3-tempora-5.0.0-2.el9.noarch 80/152 2026-03-06T13:39:05.311 INFO:teuthology.orchestra.run.vm01.stdout: Installing : python3-portend-3.1.0-2.el9.noarch 81/152 2026-03-06T13:39:05.331 INFO:teuthology.orchestra.run.vm01.stdout: Installing : python3-protobuf-3.14.0-17.el9.noarch 82/152 2026-03-06T13:39:05.389 INFO:teuthology.orchestra.run.vm00.stdout: Installing : qatzip-libs-1.3.1-1.el9.x86_64 116/152 2026-03-06T13:39:05.389 INFO:teuthology.orchestra.run.vm01.stdout: Installing : python3-grpcio-tools-1.46.7-10.el9.x86_64 83/152 2026-03-06T13:39:05.396 INFO:teuthology.orchestra.run.vm00.stdout: Installing : ceph-base-2:19.2.3-47.gc24117fd552.el9.clyso.x86 117/152 2026-03-06T13:39:05.445 INFO:teuthology.orchestra.run.vm00.stdout: Running scriptlet: ceph-base-2:19.2.3-47.gc24117fd552.el9.clyso.x86 117/152 2026-03-06T13:39:05.445 INFO:teuthology.orchestra.run.vm00.stdout:Created symlink /etc/systemd/system/multi-user.target.wants/ceph.target → /usr/lib/systemd/system/ceph.target. 2026-03-06T13:39:05.445 INFO:teuthology.orchestra.run.vm00.stdout:Created symlink /etc/systemd/system/ceph.target.wants/ceph-crash.service → /usr/lib/systemd/system/ceph-crash.service. 2026-03-06T13:39:05.445 INFO:teuthology.orchestra.run.vm00.stdout: 2026-03-06T13:39:05.452 INFO:teuthology.orchestra.run.vm00.stdout: Installing : ceph-selinux-2:19.2.3-47.gc24117fd552.el9.clyso. 118/152 2026-03-06T13:39:05.543 INFO:teuthology.orchestra.run.vm08.stdout: Installing : parquet-libs-9.0.0-15.el9.x86_64 99/152 2026-03-06T13:39:05.546 INFO:teuthology.orchestra.run.vm08.stdout: Installing : librgw2-2:19.2.3-47.gc24117fd552.el9.clyso.x86_6 100/152 2026-03-06T13:39:05.572 INFO:teuthology.orchestra.run.vm08.stdout: Running scriptlet: librgw2-2:19.2.3-47.gc24117fd552.el9.clyso.x86_6 100/152 2026-03-06T13:39:05.574 INFO:teuthology.orchestra.run.vm08.stdout: Installing : python3-rgw-2:19.2.3-47.gc24117fd552.el9.clyso.x 101/152 2026-03-06T13:39:05.709 INFO:teuthology.orchestra.run.vm01.stdout: Installing : python3-devel-3.9.25-3.el9.x86_64 84/152 2026-03-06T13:39:05.742 INFO:teuthology.orchestra.run.vm01.stdout: Installing : python3-babel-2.9.1-2.el9.noarch 85/152 2026-03-06T13:39:05.746 INFO:teuthology.orchestra.run.vm01.stdout: Installing : python3-jinja2-2.11.3-8.el9.noarch 86/152 2026-03-06T13:39:05.750 INFO:teuthology.orchestra.run.vm01.stdout: Installing : perl-Benchmark-1.23-483.el9.noarch 87/152 2026-03-06T13:39:05.816 INFO:teuthology.orchestra.run.vm01.stdout: Installing : openblas-0.3.29-1.el9.x86_64 88/152 2026-03-06T13:39:05.819 INFO:teuthology.orchestra.run.vm01.stdout: Installing : openblas-openmp-0.3.29-1.el9.x86_64 89/152 2026-03-06T13:39:05.847 INFO:teuthology.orchestra.run.vm01.stdout: Installing : flexiblas-openblas-openmp-3.0.4-9.el9.x86_64 90/152 2026-03-06T13:39:06.258 INFO:teuthology.orchestra.run.vm01.stdout: Installing : flexiblas-netlib-3.0.4-9.el9.x86_64 91/152 2026-03-06T13:39:06.357 INFO:teuthology.orchestra.run.vm01.stdout: Installing : python3-numpy-1:1.23.5-2.el9.x86_64 92/152 2026-03-06T13:39:06.769 INFO:teuthology.orchestra.run.vm08.stdout: Running scriptlet: ceph-common-2:19.2.3-47.gc24117fd552.el9.clyso.x 102/152 2026-03-06T13:39:06.821 INFO:teuthology.orchestra.run.vm08.stdout: Installing : ceph-common-2:19.2.3-47.gc24117fd552.el9.clyso.x 102/152 2026-03-06T13:39:06.855 INFO:teuthology.orchestra.run.vm08.stdout: Running scriptlet: ceph-common-2:19.2.3-47.gc24117fd552.el9.clyso.x 102/152 2026-03-06T13:39:06.859 INFO:teuthology.orchestra.run.vm08.stdout: Installing : smartmontools-1:7.2-10.el9.x86_64 103/152 2026-03-06T13:39:06.876 INFO:teuthology.orchestra.run.vm08.stdout: Running scriptlet: smartmontools-1:7.2-10.el9.x86_64 103/152 2026-03-06T13:39:06.877 INFO:teuthology.orchestra.run.vm08.stdout:Created symlink /etc/systemd/system/multi-user.target.wants/smartd.service → /usr/lib/systemd/system/smartd.service. 2026-03-06T13:39:06.877 INFO:teuthology.orchestra.run.vm08.stdout: 2026-03-06T13:39:06.904 INFO:teuthology.orchestra.run.vm08.stdout: Installing : python3-ply-3.11-14.el9.noarch 104/152 2026-03-06T13:39:06.925 INFO:teuthology.orchestra.run.vm08.stdout: Installing : python3-pycparser-2.20-6.el9.noarch 105/152 2026-03-06T13:39:07.021 INFO:teuthology.orchestra.run.vm08.stdout: Installing : python3-cffi-1.14.5-5.el9.x86_64 106/152 2026-03-06T13:39:07.038 INFO:teuthology.orchestra.run.vm08.stdout: Installing : python3-cryptography-36.0.1-5.el9.x86_64 107/152 2026-03-06T13:39:07.071 INFO:teuthology.orchestra.run.vm08.stdout: Installing : python3-pyOpenSSL-21.0.0-1.el9.noarch 108/152 2026-03-06T13:39:07.113 INFO:teuthology.orchestra.run.vm08.stdout: Installing : python3-cheroot-10.0.1-4.el9.noarch 109/152 2026-03-06T13:39:07.179 INFO:teuthology.orchestra.run.vm08.stdout: Installing : python3-cherrypy-18.6.1-2.el9.noarch 110/152 2026-03-06T13:39:07.192 INFO:teuthology.orchestra.run.vm08.stdout: Installing : python3-asyncssh-2.13.2-5.el9.noarch 111/152 2026-03-06T13:39:07.200 INFO:teuthology.orchestra.run.vm08.stdout: Installing : python3-bcrypt-3.2.2-1.el9.x86_64 112/152 2026-03-06T13:39:07.208 INFO:teuthology.orchestra.run.vm08.stdout: Installing : pciutils-3.7.0-7.el9.x86_64 113/152 2026-03-06T13:39:07.213 INFO:teuthology.orchestra.run.vm08.stdout: Installing : qatlib-25.08.0-2.el9.x86_64 114/152 2026-03-06T13:39:07.215 INFO:teuthology.orchestra.run.vm08.stdout: Installing : qatlib-service-25.08.0-2.el9.x86_64 115/152 2026-03-06T13:39:07.237 INFO:teuthology.orchestra.run.vm08.stdout: Running scriptlet: qatlib-service-25.08.0-2.el9.x86_64 115/152 2026-03-06T13:39:07.291 INFO:teuthology.orchestra.run.vm01.stdout: Installing : python3-numpy-f2py-1:1.23.5-2.el9.x86_64 93/152 2026-03-06T13:39:07.318 INFO:teuthology.orchestra.run.vm01.stdout: Installing : python3-scipy-1.9.3-2.el9.x86_64 94/152 2026-03-06T13:39:07.492 INFO:teuthology.orchestra.run.vm01.stdout: Installing : libpmemobj-1.12.1-1.el9.x86_64 95/152 2026-03-06T13:39:07.537 INFO:teuthology.orchestra.run.vm01.stdout: Upgrading : librbd1-2:19.2.3-47.gc24117fd552.el9.clyso.x86_6 96/152 2026-03-06T13:39:07.565 INFO:teuthology.orchestra.run.vm08.stdout: Installing : qatzip-libs-1.3.1-1.el9.x86_64 116/152 2026-03-06T13:39:07.571 INFO:teuthology.orchestra.run.vm01.stdout: Running scriptlet: librbd1-2:19.2.3-47.gc24117fd552.el9.clyso.x86_6 96/152 2026-03-06T13:39:07.572 INFO:teuthology.orchestra.run.vm08.stdout: Installing : ceph-base-2:19.2.3-47.gc24117fd552.el9.clyso.x86 117/152 2026-03-06T13:39:07.575 INFO:teuthology.orchestra.run.vm01.stdout: Installing : python3-rbd-2:19.2.3-47.gc24117fd552.el9.clyso.x 97/152 2026-03-06T13:39:07.583 INFO:teuthology.orchestra.run.vm01.stdout: Installing : boost-program-options-1.75.0-13.el9.x86_64 98/152 2026-03-06T13:39:07.622 INFO:teuthology.orchestra.run.vm08.stdout: Running scriptlet: ceph-base-2:19.2.3-47.gc24117fd552.el9.clyso.x86 117/152 2026-03-06T13:39:07.622 INFO:teuthology.orchestra.run.vm08.stdout:Created symlink /etc/systemd/system/multi-user.target.wants/ceph.target → /usr/lib/systemd/system/ceph.target. 2026-03-06T13:39:07.622 INFO:teuthology.orchestra.run.vm08.stdout:Created symlink /etc/systemd/system/ceph.target.wants/ceph-crash.service → /usr/lib/systemd/system/ceph-crash.service. 2026-03-06T13:39:07.622 INFO:teuthology.orchestra.run.vm08.stdout: 2026-03-06T13:39:07.628 INFO:teuthology.orchestra.run.vm08.stdout: Installing : ceph-selinux-2:19.2.3-47.gc24117fd552.el9.clyso. 118/152 2026-03-06T13:39:07.862 INFO:teuthology.orchestra.run.vm01.stdout: Installing : parquet-libs-9.0.0-15.el9.x86_64 99/152 2026-03-06T13:39:07.865 INFO:teuthology.orchestra.run.vm01.stdout: Installing : librgw2-2:19.2.3-47.gc24117fd552.el9.clyso.x86_6 100/152 2026-03-06T13:39:07.887 INFO:teuthology.orchestra.run.vm01.stdout: Running scriptlet: librgw2-2:19.2.3-47.gc24117fd552.el9.clyso.x86_6 100/152 2026-03-06T13:39:07.889 INFO:teuthology.orchestra.run.vm01.stdout: Installing : python3-rgw-2:19.2.3-47.gc24117fd552.el9.clyso.x 101/152 2026-03-06T13:39:09.104 INFO:teuthology.orchestra.run.vm01.stdout: Running scriptlet: ceph-common-2:19.2.3-47.gc24117fd552.el9.clyso.x 102/152 2026-03-06T13:39:09.125 INFO:teuthology.orchestra.run.vm01.stdout: Installing : ceph-common-2:19.2.3-47.gc24117fd552.el9.clyso.x 102/152 2026-03-06T13:39:09.160 INFO:teuthology.orchestra.run.vm01.stdout: Running scriptlet: ceph-common-2:19.2.3-47.gc24117fd552.el9.clyso.x 102/152 2026-03-06T13:39:09.163 INFO:teuthology.orchestra.run.vm01.stdout: Installing : smartmontools-1:7.2-10.el9.x86_64 103/152 2026-03-06T13:39:09.182 INFO:teuthology.orchestra.run.vm01.stdout: Running scriptlet: smartmontools-1:7.2-10.el9.x86_64 103/152 2026-03-06T13:39:09.183 INFO:teuthology.orchestra.run.vm01.stdout:Created symlink /etc/systemd/system/multi-user.target.wants/smartd.service → /usr/lib/systemd/system/smartd.service. 2026-03-06T13:39:09.183 INFO:teuthology.orchestra.run.vm01.stdout: 2026-03-06T13:39:09.216 INFO:teuthology.orchestra.run.vm01.stdout: Installing : python3-ply-3.11-14.el9.noarch 104/152 2026-03-06T13:39:09.239 INFO:teuthology.orchestra.run.vm01.stdout: Installing : python3-pycparser-2.20-6.el9.noarch 105/152 2026-03-06T13:39:09.337 INFO:teuthology.orchestra.run.vm01.stdout: Installing : python3-cffi-1.14.5-5.el9.x86_64 106/152 2026-03-06T13:39:09.359 INFO:teuthology.orchestra.run.vm01.stdout: Installing : python3-cryptography-36.0.1-5.el9.x86_64 107/152 2026-03-06T13:39:09.392 INFO:teuthology.orchestra.run.vm01.stdout: Installing : python3-pyOpenSSL-21.0.0-1.el9.noarch 108/152 2026-03-06T13:39:09.433 INFO:teuthology.orchestra.run.vm01.stdout: Installing : python3-cheroot-10.0.1-4.el9.noarch 109/152 2026-03-06T13:39:09.499 INFO:teuthology.orchestra.run.vm01.stdout: Installing : python3-cherrypy-18.6.1-2.el9.noarch 110/152 2026-03-06T13:39:09.510 INFO:teuthology.orchestra.run.vm01.stdout: Installing : python3-asyncssh-2.13.2-5.el9.noarch 111/152 2026-03-06T13:39:09.516 INFO:teuthology.orchestra.run.vm01.stdout: Installing : python3-bcrypt-3.2.2-1.el9.x86_64 112/152 2026-03-06T13:39:09.522 INFO:teuthology.orchestra.run.vm01.stdout: Installing : pciutils-3.7.0-7.el9.x86_64 113/152 2026-03-06T13:39:09.526 INFO:teuthology.orchestra.run.vm01.stdout: Installing : qatlib-25.08.0-2.el9.x86_64 114/152 2026-03-06T13:39:09.528 INFO:teuthology.orchestra.run.vm01.stdout: Installing : qatlib-service-25.08.0-2.el9.x86_64 115/152 2026-03-06T13:39:09.547 INFO:teuthology.orchestra.run.vm01.stdout: Running scriptlet: qatlib-service-25.08.0-2.el9.x86_64 115/152 2026-03-06T13:39:09.870 INFO:teuthology.orchestra.run.vm01.stdout: Installing : qatzip-libs-1.3.1-1.el9.x86_64 116/152 2026-03-06T13:39:09.877 INFO:teuthology.orchestra.run.vm01.stdout: Installing : ceph-base-2:19.2.3-47.gc24117fd552.el9.clyso.x86 117/152 2026-03-06T13:39:09.914 INFO:teuthology.orchestra.run.vm01.stdout: Running scriptlet: ceph-base-2:19.2.3-47.gc24117fd552.el9.clyso.x86 117/152 2026-03-06T13:39:09.914 INFO:teuthology.orchestra.run.vm01.stdout:Created symlink /etc/systemd/system/multi-user.target.wants/ceph.target → /usr/lib/systemd/system/ceph.target. 2026-03-06T13:39:09.914 INFO:teuthology.orchestra.run.vm01.stdout:Created symlink /etc/systemd/system/ceph.target.wants/ceph-crash.service → /usr/lib/systemd/system/ceph-crash.service. 2026-03-06T13:39:09.914 INFO:teuthology.orchestra.run.vm01.stdout: 2026-03-06T13:39:09.918 INFO:teuthology.orchestra.run.vm01.stdout: Installing : ceph-selinux-2:19.2.3-47.gc24117fd552.el9.clyso. 118/152 2026-03-06T13:39:12.260 INFO:teuthology.orchestra.run.vm00.stdout: Running scriptlet: ceph-selinux-2:19.2.3-47.gc24117fd552.el9.clyso. 118/152 2026-03-06T13:39:12.260 INFO:teuthology.orchestra.run.vm00.stdout:skipping the directory /sys 2026-03-06T13:39:12.260 INFO:teuthology.orchestra.run.vm00.stdout:skipping the directory /proc 2026-03-06T13:39:12.260 INFO:teuthology.orchestra.run.vm00.stdout:skipping the directory /mnt 2026-03-06T13:39:12.260 INFO:teuthology.orchestra.run.vm00.stdout:skipping the directory /var/tmp 2026-03-06T13:39:12.260 INFO:teuthology.orchestra.run.vm00.stdout:skipping the directory /home 2026-03-06T13:39:12.260 INFO:teuthology.orchestra.run.vm00.stdout:skipping the directory /root 2026-03-06T13:39:12.260 INFO:teuthology.orchestra.run.vm00.stdout:skipping the directory /tmp 2026-03-06T13:39:12.260 INFO:teuthology.orchestra.run.vm00.stdout: 2026-03-06T13:39:12.397 INFO:teuthology.orchestra.run.vm00.stdout: Installing : ceph-mds-2:19.2.3-47.gc24117fd552.el9.clyso.x86_ 119/152 2026-03-06T13:39:12.421 INFO:teuthology.orchestra.run.vm00.stdout: Running scriptlet: ceph-mds-2:19.2.3-47.gc24117fd552.el9.clyso.x86_ 119/152 2026-03-06T13:39:12.421 INFO:teuthology.orchestra.run.vm00.stdout:Glob pattern passed to enable, but globs are not supported for this. 2026-03-06T13:39:12.421 INFO:teuthology.orchestra.run.vm00.stdout:Invalid unit name "ceph-mds@*.service" escaped as "ceph-mds@\x2a.service". 2026-03-06T13:39:12.421 INFO:teuthology.orchestra.run.vm00.stdout:Created symlink /etc/systemd/system/multi-user.target.wants/ceph-mds.target → /usr/lib/systemd/system/ceph-mds.target. 2026-03-06T13:39:12.421 INFO:teuthology.orchestra.run.vm00.stdout:Created symlink /etc/systemd/system/ceph.target.wants/ceph-mds.target → /usr/lib/systemd/system/ceph-mds.target. 2026-03-06T13:39:12.421 INFO:teuthology.orchestra.run.vm00.stdout: 2026-03-06T13:39:12.838 INFO:teuthology.orchestra.run.vm00.stdout: Installing : ceph-mon-2:19.2.3-47.gc24117fd552.el9.clyso.x86_ 120/152 2026-03-06T13:39:12.861 INFO:teuthology.orchestra.run.vm00.stdout: Running scriptlet: ceph-mon-2:19.2.3-47.gc24117fd552.el9.clyso.x86_ 120/152 2026-03-06T13:39:12.875 INFO:teuthology.orchestra.run.vm00.stdout:Glob pattern passed to enable, but globs are not supported for this. 2026-03-06T13:39:12.875 INFO:teuthology.orchestra.run.vm00.stdout:Invalid unit name "ceph-mon@*.service" escaped as "ceph-mon@\x2a.service". 2026-03-06T13:39:12.875 INFO:teuthology.orchestra.run.vm00.stdout:Created symlink /etc/systemd/system/multi-user.target.wants/ceph-mon.target → /usr/lib/systemd/system/ceph-mon.target. 2026-03-06T13:39:12.875 INFO:teuthology.orchestra.run.vm00.stdout:Created symlink /etc/systemd/system/ceph.target.wants/ceph-mon.target → /usr/lib/systemd/system/ceph-mon.target. 2026-03-06T13:39:12.875 INFO:teuthology.orchestra.run.vm00.stdout: 2026-03-06T13:39:12.968 INFO:teuthology.orchestra.run.vm00.stdout: Installing : mailcap-2.1.49-5.el9.noarch 121/152 2026-03-06T13:39:13.071 INFO:teuthology.orchestra.run.vm00.stdout: Installing : libconfig-1.7.2-9.el9.x86_64 122/152 2026-03-06T13:39:13.277 INFO:teuthology.orchestra.run.vm00.stdout: Running scriptlet: libstoragemgmt-1.10.1-1.el9.x86_64 123/152 2026-03-06T13:39:13.277 INFO:teuthology.orchestra.run.vm00.stdout:Creating group 'qat' with GID 994. 2026-03-06T13:39:13.277 INFO:teuthology.orchestra.run.vm00.stdout:Creating group 'libstoragemgmt' with GID 993. 2026-03-06T13:39:13.277 INFO:teuthology.orchestra.run.vm00.stdout:Creating user 'libstoragemgmt' (daemon account for libstoragemgmt) with UID 993 and GID 993. 2026-03-06T13:39:13.277 INFO:teuthology.orchestra.run.vm00.stdout: 2026-03-06T13:39:13.315 INFO:teuthology.orchestra.run.vm00.stdout: Installing : libstoragemgmt-1.10.1-1.el9.x86_64 123/152 2026-03-06T13:39:13.374 INFO:teuthology.orchestra.run.vm00.stdout: Running scriptlet: libstoragemgmt-1.10.1-1.el9.x86_64 123/152 2026-03-06T13:39:13.374 INFO:teuthology.orchestra.run.vm00.stdout:Created symlink /etc/systemd/system/multi-user.target.wants/libstoragemgmt.service → /usr/lib/systemd/system/libstoragemgmt.service. 2026-03-06T13:39:13.374 INFO:teuthology.orchestra.run.vm00.stdout: 2026-03-06T13:39:13.399 INFO:teuthology.orchestra.run.vm00.stdout: Installing : python3-libstoragemgmt-1.10.1-1.el9.x86_64 124/152 2026-03-06T13:39:13.428 INFO:teuthology.orchestra.run.vm00.stdout: Installing : fuse-2.9.9-17.el9.x86_64 125/152 2026-03-06T13:39:13.511 INFO:teuthology.orchestra.run.vm00.stdout: Installing : cryptsetup-2.8.1-3.el9.x86_64 126/152 2026-03-06T13:39:13.535 INFO:teuthology.orchestra.run.vm00.stdout: Installing : ceph-volume-2:19.2.3-47.gc24117fd552.el9.clyso.n 127/152 2026-03-06T13:39:13.549 INFO:teuthology.orchestra.run.vm00.stdout: Running scriptlet: ceph-volume-2:19.2.3-47.gc24117fd552.el9.clyso.n 127/152 2026-03-06T13:39:13.549 INFO:teuthology.orchestra.run.vm00.stdout:Glob pattern passed to enable, but globs are not supported for this. 2026-03-06T13:39:13.549 INFO:teuthology.orchestra.run.vm00.stdout:Invalid unit name "ceph-volume@*.service" escaped as "ceph-volume@\x2a.service". 2026-03-06T13:39:13.549 INFO:teuthology.orchestra.run.vm00.stdout: 2026-03-06T13:39:13.844 INFO:teuthology.orchestra.run.vm08.stdout: Running scriptlet: ceph-selinux-2:19.2.3-47.gc24117fd552.el9.clyso. 118/152 2026-03-06T13:39:13.844 INFO:teuthology.orchestra.run.vm08.stdout:skipping the directory /sys 2026-03-06T13:39:13.844 INFO:teuthology.orchestra.run.vm08.stdout:skipping the directory /proc 2026-03-06T13:39:13.844 INFO:teuthology.orchestra.run.vm08.stdout:skipping the directory /mnt 2026-03-06T13:39:13.844 INFO:teuthology.orchestra.run.vm08.stdout:skipping the directory /var/tmp 2026-03-06T13:39:13.844 INFO:teuthology.orchestra.run.vm08.stdout:skipping the directory /home 2026-03-06T13:39:13.844 INFO:teuthology.orchestra.run.vm08.stdout:skipping the directory /root 2026-03-06T13:39:13.844 INFO:teuthology.orchestra.run.vm08.stdout:skipping the directory /tmp 2026-03-06T13:39:13.844 INFO:teuthology.orchestra.run.vm08.stdout: 2026-03-06T13:39:13.977 INFO:teuthology.orchestra.run.vm08.stdout: Installing : ceph-mds-2:19.2.3-47.gc24117fd552.el9.clyso.x86_ 119/152 2026-03-06T13:39:14.009 INFO:teuthology.orchestra.run.vm08.stdout: Running scriptlet: ceph-mds-2:19.2.3-47.gc24117fd552.el9.clyso.x86_ 119/152 2026-03-06T13:39:14.009 INFO:teuthology.orchestra.run.vm08.stdout:Glob pattern passed to enable, but globs are not supported for this. 2026-03-06T13:39:14.009 INFO:teuthology.orchestra.run.vm08.stdout:Invalid unit name "ceph-mds@*.service" escaped as "ceph-mds@\x2a.service". 2026-03-06T13:39:14.009 INFO:teuthology.orchestra.run.vm08.stdout:Created symlink /etc/systemd/system/multi-user.target.wants/ceph-mds.target → /usr/lib/systemd/system/ceph-mds.target. 2026-03-06T13:39:14.009 INFO:teuthology.orchestra.run.vm08.stdout:Created symlink /etc/systemd/system/ceph.target.wants/ceph-mds.target → /usr/lib/systemd/system/ceph-mds.target. 2026-03-06T13:39:14.009 INFO:teuthology.orchestra.run.vm08.stdout: 2026-03-06T13:39:14.249 INFO:teuthology.orchestra.run.vm08.stdout: Installing : ceph-mon-2:19.2.3-47.gc24117fd552.el9.clyso.x86_ 120/152 2026-03-06T13:39:14.272 INFO:teuthology.orchestra.run.vm08.stdout: Running scriptlet: ceph-mon-2:19.2.3-47.gc24117fd552.el9.clyso.x86_ 120/152 2026-03-06T13:39:14.272 INFO:teuthology.orchestra.run.vm08.stdout:Glob pattern passed to enable, but globs are not supported for this. 2026-03-06T13:39:14.272 INFO:teuthology.orchestra.run.vm08.stdout:Invalid unit name "ceph-mon@*.service" escaped as "ceph-mon@\x2a.service". 2026-03-06T13:39:14.272 INFO:teuthology.orchestra.run.vm08.stdout:Created symlink /etc/systemd/system/multi-user.target.wants/ceph-mon.target → /usr/lib/systemd/system/ceph-mon.target. 2026-03-06T13:39:14.272 INFO:teuthology.orchestra.run.vm08.stdout:Created symlink /etc/systemd/system/ceph.target.wants/ceph-mon.target → /usr/lib/systemd/system/ceph-mon.target. 2026-03-06T13:39:14.272 INFO:teuthology.orchestra.run.vm08.stdout: 2026-03-06T13:39:14.281 INFO:teuthology.orchestra.run.vm08.stdout: Installing : mailcap-2.1.49-5.el9.noarch 121/152 2026-03-06T13:39:14.283 INFO:teuthology.orchestra.run.vm08.stdout: Installing : libconfig-1.7.2-9.el9.x86_64 122/152 2026-03-06T13:39:14.301 INFO:teuthology.orchestra.run.vm08.stdout: Running scriptlet: libstoragemgmt-1.10.1-1.el9.x86_64 123/152 2026-03-06T13:39:14.301 INFO:teuthology.orchestra.run.vm08.stdout:Creating group 'qat' with GID 994. 2026-03-06T13:39:14.301 INFO:teuthology.orchestra.run.vm08.stdout:Creating group 'libstoragemgmt' with GID 993. 2026-03-06T13:39:14.301 INFO:teuthology.orchestra.run.vm08.stdout:Creating user 'libstoragemgmt' (daemon account for libstoragemgmt) with UID 993 and GID 993. 2026-03-06T13:39:14.301 INFO:teuthology.orchestra.run.vm08.stdout: 2026-03-06T13:39:14.312 INFO:teuthology.orchestra.run.vm08.stdout: Installing : libstoragemgmt-1.10.1-1.el9.x86_64 123/152 2026-03-06T13:39:14.337 INFO:teuthology.orchestra.run.vm08.stdout: Running scriptlet: libstoragemgmt-1.10.1-1.el9.x86_64 123/152 2026-03-06T13:39:14.337 INFO:teuthology.orchestra.run.vm08.stdout:Created symlink /etc/systemd/system/multi-user.target.wants/libstoragemgmt.service → /usr/lib/systemd/system/libstoragemgmt.service. 2026-03-06T13:39:14.337 INFO:teuthology.orchestra.run.vm08.stdout: 2026-03-06T13:39:14.359 INFO:teuthology.orchestra.run.vm08.stdout: Installing : python3-libstoragemgmt-1.10.1-1.el9.x86_64 124/152 2026-03-06T13:39:14.388 INFO:teuthology.orchestra.run.vm08.stdout: Installing : fuse-2.9.9-17.el9.x86_64 125/152 2026-03-06T13:39:14.393 INFO:teuthology.orchestra.run.vm00.stdout: Installing : ceph-osd-2:19.2.3-47.gc24117fd552.el9.clyso.x86_ 128/152 2026-03-06T13:39:14.419 INFO:teuthology.orchestra.run.vm00.stdout: Running scriptlet: ceph-osd-2:19.2.3-47.gc24117fd552.el9.clyso.x86_ 128/152 2026-03-06T13:39:14.420 INFO:teuthology.orchestra.run.vm00.stdout:Glob pattern passed to enable, but globs are not supported for this. 2026-03-06T13:39:14.420 INFO:teuthology.orchestra.run.vm00.stdout:Invalid unit name "ceph-osd@*.service" escaped as "ceph-osd@\x2a.service". 2026-03-06T13:39:14.420 INFO:teuthology.orchestra.run.vm00.stdout:Created symlink /etc/systemd/system/multi-user.target.wants/ceph-osd.target → /usr/lib/systemd/system/ceph-osd.target. 2026-03-06T13:39:14.420 INFO:teuthology.orchestra.run.vm00.stdout:Created symlink /etc/systemd/system/ceph.target.wants/ceph-osd.target → /usr/lib/systemd/system/ceph-osd.target. 2026-03-06T13:39:14.420 INFO:teuthology.orchestra.run.vm00.stdout: 2026-03-06T13:39:14.470 INFO:teuthology.orchestra.run.vm08.stdout: Installing : cryptsetup-2.8.1-3.el9.x86_64 126/152 2026-03-06T13:39:14.475 INFO:teuthology.orchestra.run.vm08.stdout: Installing : ceph-volume-2:19.2.3-47.gc24117fd552.el9.clyso.n 127/152 2026-03-06T13:39:14.483 INFO:teuthology.orchestra.run.vm00.stdout: Running scriptlet: cephadm-2:19.2.3-47.gc24117fd552.el9.clyso.noarc 129/152 2026-03-06T13:39:14.487 INFO:teuthology.orchestra.run.vm00.stdout: Installing : cephadm-2:19.2.3-47.gc24117fd552.el9.clyso.noarc 129/152 2026-03-06T13:39:14.494 INFO:teuthology.orchestra.run.vm00.stdout: Installing : ceph-prometheus-alerts-2:19.2.3-47.gc24117fd552. 130/152 2026-03-06T13:39:14.494 INFO:teuthology.orchestra.run.vm08.stdout: Running scriptlet: ceph-volume-2:19.2.3-47.gc24117fd552.el9.clyso.n 127/152 2026-03-06T13:39:14.494 INFO:teuthology.orchestra.run.vm08.stdout:Glob pattern passed to enable, but globs are not supported for this. 2026-03-06T13:39:14.495 INFO:teuthology.orchestra.run.vm08.stdout:Invalid unit name "ceph-volume@*.service" escaped as "ceph-volume@\x2a.service". 2026-03-06T13:39:14.495 INFO:teuthology.orchestra.run.vm08.stdout: 2026-03-06T13:39:14.519 INFO:teuthology.orchestra.run.vm00.stdout: Installing : ceph-grafana-dashboards-2:19.2.3-47.gc24117fd552 131/152 2026-03-06T13:39:14.522 INFO:teuthology.orchestra.run.vm00.stdout: Installing : ceph-mgr-cephadm-2:19.2.3-47.gc24117fd552.el9.cl 132/152 2026-03-06T13:39:15.086 INFO:teuthology.orchestra.run.vm00.stdout: Running scriptlet: ceph-mgr-cephadm-2:19.2.3-47.gc24117fd552.el9.cl 132/152 2026-03-06T13:39:15.093 INFO:teuthology.orchestra.run.vm00.stdout: Installing : ceph-mgr-dashboard-2:19.2.3-47.gc24117fd552.el9. 133/152 2026-03-06T13:39:15.332 INFO:teuthology.orchestra.run.vm08.stdout: Installing : ceph-osd-2:19.2.3-47.gc24117fd552.el9.clyso.x86_ 128/152 2026-03-06T13:39:15.360 INFO:teuthology.orchestra.run.vm08.stdout: Running scriptlet: ceph-osd-2:19.2.3-47.gc24117fd552.el9.clyso.x86_ 128/152 2026-03-06T13:39:15.360 INFO:teuthology.orchestra.run.vm08.stdout:Glob pattern passed to enable, but globs are not supported for this. 2026-03-06T13:39:15.360 INFO:teuthology.orchestra.run.vm08.stdout:Invalid unit name "ceph-osd@*.service" escaped as "ceph-osd@\x2a.service". 2026-03-06T13:39:15.360 INFO:teuthology.orchestra.run.vm08.stdout:Created symlink /etc/systemd/system/multi-user.target.wants/ceph-osd.target → /usr/lib/systemd/system/ceph-osd.target. 2026-03-06T13:39:15.360 INFO:teuthology.orchestra.run.vm08.stdout:Created symlink /etc/systemd/system/ceph.target.wants/ceph-osd.target → /usr/lib/systemd/system/ceph-osd.target. 2026-03-06T13:39:15.360 INFO:teuthology.orchestra.run.vm08.stdout: 2026-03-06T13:39:15.434 INFO:teuthology.orchestra.run.vm08.stdout: Running scriptlet: cephadm-2:19.2.3-47.gc24117fd552.el9.clyso.noarc 129/152 2026-03-06T13:39:15.438 INFO:teuthology.orchestra.run.vm08.stdout: Installing : cephadm-2:19.2.3-47.gc24117fd552.el9.clyso.noarc 129/152 2026-03-06T13:39:15.447 INFO:teuthology.orchestra.run.vm08.stdout: Installing : ceph-prometheus-alerts-2:19.2.3-47.gc24117fd552. 130/152 2026-03-06T13:39:15.479 INFO:teuthology.orchestra.run.vm08.stdout: Installing : ceph-grafana-dashboards-2:19.2.3-47.gc24117fd552 131/152 2026-03-06T13:39:15.482 INFO:teuthology.orchestra.run.vm08.stdout: Installing : ceph-mgr-cephadm-2:19.2.3-47.gc24117fd552.el9.cl 132/152 2026-03-06T13:39:15.687 INFO:teuthology.orchestra.run.vm00.stdout: Running scriptlet: ceph-mgr-dashboard-2:19.2.3-47.gc24117fd552.el9. 133/152 2026-03-06T13:39:15.718 INFO:teuthology.orchestra.run.vm00.stdout: Installing : ceph-mgr-diskprediction-local-2:19.2.3-47.gc2411 134/152 2026-03-06T13:39:15.735 INFO:teuthology.orchestra.run.vm00.stdout: Running scriptlet: ceph-mgr-diskprediction-local-2:19.2.3-47.gc2411 134/152 2026-03-06T13:39:15.851 INFO:teuthology.orchestra.run.vm00.stdout: Installing : ceph-mgr-k8sevents-2:19.2.3-47.gc24117fd552.el9. 135/152 2026-03-06T13:39:15.925 INFO:teuthology.orchestra.run.vm00.stdout: Running scriptlet: ceph-mgr-k8sevents-2:19.2.3-47.gc24117fd552.el9. 135/152 2026-03-06T13:39:16.075 INFO:teuthology.orchestra.run.vm00.stdout: Installing : ceph-mgr-modules-core-2:19.2.3-47.gc24117fd552.e 136/152 2026-03-06T13:39:16.097 INFO:teuthology.orchestra.run.vm08.stdout: Running scriptlet: ceph-mgr-cephadm-2:19.2.3-47.gc24117fd552.el9.cl 132/152 2026-03-06T13:39:16.137 INFO:teuthology.orchestra.run.vm00.stdout: Installing : ceph-mgr-2:19.2.3-47.gc24117fd552.el9.clyso.x86_ 137/152 2026-03-06T13:39:16.165 INFO:teuthology.orchestra.run.vm00.stdout: Running scriptlet: ceph-mgr-2:19.2.3-47.gc24117fd552.el9.clyso.x86_ 137/152 2026-03-06T13:39:16.165 INFO:teuthology.orchestra.run.vm00.stdout:Glob pattern passed to enable, but globs are not supported for this. 2026-03-06T13:39:16.165 INFO:teuthology.orchestra.run.vm00.stdout:Invalid unit name "ceph-mgr@*.service" escaped as "ceph-mgr@\x2a.service". 2026-03-06T13:39:16.165 INFO:teuthology.orchestra.run.vm00.stdout:Created symlink /etc/systemd/system/multi-user.target.wants/ceph-mgr.target → /usr/lib/systemd/system/ceph-mgr.target. 2026-03-06T13:39:16.165 INFO:teuthology.orchestra.run.vm00.stdout:Created symlink /etc/systemd/system/ceph.target.wants/ceph-mgr.target → /usr/lib/systemd/system/ceph-mgr.target. 2026-03-06T13:39:16.165 INFO:teuthology.orchestra.run.vm00.stdout: 2026-03-06T13:39:16.186 INFO:teuthology.orchestra.run.vm08.stdout: Installing : ceph-mgr-dashboard-2:19.2.3-47.gc24117fd552.el9. 133/152 2026-03-06T13:39:16.207 INFO:teuthology.orchestra.run.vm00.stdout: Installing : ceph-mgr-rook-2:19.2.3-47.gc24117fd552.el9.clyso 138/152 2026-03-06T13:39:16.223 INFO:teuthology.orchestra.run.vm00.stdout: Running scriptlet: ceph-mgr-rook-2:19.2.3-47.gc24117fd552.el9.clyso 138/152 2026-03-06T13:39:16.302 INFO:teuthology.orchestra.run.vm00.stdout: Installing : ceph-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 139/152 2026-03-06T13:39:16.745 INFO:teuthology.orchestra.run.vm08.stdout: Running scriptlet: ceph-mgr-dashboard-2:19.2.3-47.gc24117fd552.el9. 133/152 2026-03-06T13:39:16.747 INFO:teuthology.orchestra.run.vm08.stdout: Installing : ceph-mgr-diskprediction-local-2:19.2.3-47.gc2411 134/152 2026-03-06T13:39:16.763 INFO:teuthology.orchestra.run.vm08.stdout: Running scriptlet: ceph-mgr-diskprediction-local-2:19.2.3-47.gc2411 134/152 2026-03-06T13:39:16.765 INFO:teuthology.orchestra.run.vm08.stdout: Installing : ceph-mgr-k8sevents-2:19.2.3-47.gc24117fd552.el9. 135/152 2026-03-06T13:39:16.830 INFO:teuthology.orchestra.run.vm08.stdout: Running scriptlet: ceph-mgr-k8sevents-2:19.2.3-47.gc24117fd552.el9. 135/152 2026-03-06T13:39:16.880 INFO:teuthology.orchestra.run.vm00.stdout: Installing : ceph-fuse-2:19.2.3-47.gc24117fd552.el9.clyso.x86 140/152 2026-03-06T13:39:16.884 INFO:teuthology.orchestra.run.vm00.stdout: Installing : ceph-radosgw-2:19.2.3-47.gc24117fd552.el9.clyso. 141/152 2026-03-06T13:39:16.890 INFO:teuthology.orchestra.run.vm08.stdout: Installing : ceph-mgr-modules-core-2:19.2.3-47.gc24117fd552.e 136/152 2026-03-06T13:39:16.893 INFO:teuthology.orchestra.run.vm08.stdout: Installing : ceph-mgr-2:19.2.3-47.gc24117fd552.el9.clyso.x86_ 137/152 2026-03-06T13:39:16.905 INFO:teuthology.orchestra.run.vm00.stdout: Running scriptlet: ceph-radosgw-2:19.2.3-47.gc24117fd552.el9.clyso. 141/152 2026-03-06T13:39:16.905 INFO:teuthology.orchestra.run.vm00.stdout:Glob pattern passed to enable, but globs are not supported for this. 2026-03-06T13:39:16.905 INFO:teuthology.orchestra.run.vm00.stdout:Invalid unit name "ceph-radosgw@*.service" escaped as "ceph-radosgw@\x2a.service". 2026-03-06T13:39:16.905 INFO:teuthology.orchestra.run.vm00.stdout:Created symlink /etc/systemd/system/multi-user.target.wants/ceph-radosgw.target → /usr/lib/systemd/system/ceph-radosgw.target. 2026-03-06T13:39:16.905 INFO:teuthology.orchestra.run.vm00.stdout:Created symlink /etc/systemd/system/ceph.target.wants/ceph-radosgw.target → /usr/lib/systemd/system/ceph-radosgw.target. 2026-03-06T13:39:16.905 INFO:teuthology.orchestra.run.vm00.stdout: 2026-03-06T13:39:16.918 INFO:teuthology.orchestra.run.vm00.stdout: Installing : ceph-immutable-object-cache-2:19.2.3-47.gc24117f 142/152 2026-03-06T13:39:16.919 INFO:teuthology.orchestra.run.vm08.stdout: Running scriptlet: ceph-mgr-2:19.2.3-47.gc24117fd552.el9.clyso.x86_ 137/152 2026-03-06T13:39:16.919 INFO:teuthology.orchestra.run.vm08.stdout:Glob pattern passed to enable, but globs are not supported for this. 2026-03-06T13:39:16.919 INFO:teuthology.orchestra.run.vm08.stdout:Invalid unit name "ceph-mgr@*.service" escaped as "ceph-mgr@\x2a.service". 2026-03-06T13:39:16.919 INFO:teuthology.orchestra.run.vm08.stdout:Created symlink /etc/systemd/system/multi-user.target.wants/ceph-mgr.target → /usr/lib/systemd/system/ceph-mgr.target. 2026-03-06T13:39:16.920 INFO:teuthology.orchestra.run.vm08.stdout:Created symlink /etc/systemd/system/ceph.target.wants/ceph-mgr.target → /usr/lib/systemd/system/ceph-mgr.target. 2026-03-06T13:39:16.920 INFO:teuthology.orchestra.run.vm08.stdout: 2026-03-06T13:39:16.935 INFO:teuthology.orchestra.run.vm08.stdout: Installing : ceph-mgr-rook-2:19.2.3-47.gc24117fd552.el9.clyso 138/152 2026-03-06T13:39:16.941 INFO:teuthology.orchestra.run.vm00.stdout: Running scriptlet: ceph-immutable-object-cache-2:19.2.3-47.gc24117f 142/152 2026-03-06T13:39:16.941 INFO:teuthology.orchestra.run.vm00.stdout:Glob pattern passed to enable, but globs are not supported for this. 2026-03-06T13:39:16.941 INFO:teuthology.orchestra.run.vm00.stdout:Invalid unit name "ceph-immutable-object-cache@*.service" escaped as "ceph-immutable-object-cache@\x2a.service". 2026-03-06T13:39:16.941 INFO:teuthology.orchestra.run.vm00.stdout: 2026-03-06T13:39:16.949 INFO:teuthology.orchestra.run.vm08.stdout: Running scriptlet: ceph-mgr-rook-2:19.2.3-47.gc24117fd552.el9.clyso 138/152 2026-03-06T13:39:17.008 INFO:teuthology.orchestra.run.vm08.stdout: Installing : ceph-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 139/152 2026-03-06T13:39:17.099 INFO:teuthology.orchestra.run.vm01.stdout: Running scriptlet: ceph-selinux-2:19.2.3-47.gc24117fd552.el9.clyso. 118/152 2026-03-06T13:39:17.099 INFO:teuthology.orchestra.run.vm01.stdout:skipping the directory /sys 2026-03-06T13:39:17.099 INFO:teuthology.orchestra.run.vm01.stdout:skipping the directory /proc 2026-03-06T13:39:17.099 INFO:teuthology.orchestra.run.vm01.stdout:skipping the directory /mnt 2026-03-06T13:39:17.099 INFO:teuthology.orchestra.run.vm01.stdout:skipping the directory /var/tmp 2026-03-06T13:39:17.099 INFO:teuthology.orchestra.run.vm01.stdout:skipping the directory /home 2026-03-06T13:39:17.099 INFO:teuthology.orchestra.run.vm01.stdout:skipping the directory /root 2026-03-06T13:39:17.099 INFO:teuthology.orchestra.run.vm01.stdout:skipping the directory /tmp 2026-03-06T13:39:17.099 INFO:teuthology.orchestra.run.vm01.stdout: 2026-03-06T13:39:17.104 INFO:teuthology.orchestra.run.vm00.stdout: Installing : rbd-mirror-2:19.2.3-47.gc24117fd552.el9.clyso.x8 143/152 2026-03-06T13:39:17.129 INFO:teuthology.orchestra.run.vm00.stdout: Running scriptlet: rbd-mirror-2:19.2.3-47.gc24117fd552.el9.clyso.x8 143/152 2026-03-06T13:39:17.129 INFO:teuthology.orchestra.run.vm00.stdout:Glob pattern passed to enable, but globs are not supported for this. 2026-03-06T13:39:17.129 INFO:teuthology.orchestra.run.vm00.stdout:Invalid unit name "ceph-rbd-mirror@*.service" escaped as "ceph-rbd-mirror@\x2a.service". 2026-03-06T13:39:17.129 INFO:teuthology.orchestra.run.vm00.stdout:Created symlink /etc/systemd/system/multi-user.target.wants/ceph-rbd-mirror.target → /usr/lib/systemd/system/ceph-rbd-mirror.target. 2026-03-06T13:39:17.129 INFO:teuthology.orchestra.run.vm00.stdout:Created symlink /etc/systemd/system/ceph.target.wants/ceph-rbd-mirror.target → /usr/lib/systemd/system/ceph-rbd-mirror.target. 2026-03-06T13:39:17.129 INFO:teuthology.orchestra.run.vm00.stdout: 2026-03-06T13:39:17.237 INFO:teuthology.orchestra.run.vm01.stdout: Installing : ceph-mds-2:19.2.3-47.gc24117fd552.el9.clyso.x86_ 119/152 2026-03-06T13:39:17.263 INFO:teuthology.orchestra.run.vm01.stdout: Running scriptlet: ceph-mds-2:19.2.3-47.gc24117fd552.el9.clyso.x86_ 119/152 2026-03-06T13:39:17.263 INFO:teuthology.orchestra.run.vm01.stdout:Glob pattern passed to enable, but globs are not supported for this. 2026-03-06T13:39:17.263 INFO:teuthology.orchestra.run.vm01.stdout:Invalid unit name "ceph-mds@*.service" escaped as "ceph-mds@\x2a.service". 2026-03-06T13:39:17.263 INFO:teuthology.orchestra.run.vm01.stdout:Created symlink /etc/systemd/system/multi-user.target.wants/ceph-mds.target → /usr/lib/systemd/system/ceph-mds.target. 2026-03-06T13:39:17.263 INFO:teuthology.orchestra.run.vm01.stdout:Created symlink /etc/systemd/system/ceph.target.wants/ceph-mds.target → /usr/lib/systemd/system/ceph-mds.target. 2026-03-06T13:39:17.263 INFO:teuthology.orchestra.run.vm01.stdout: 2026-03-06T13:39:17.507 INFO:teuthology.orchestra.run.vm01.stdout: Installing : ceph-mon-2:19.2.3-47.gc24117fd552.el9.clyso.x86_ 120/152 2026-03-06T13:39:17.537 INFO:teuthology.orchestra.run.vm08.stdout: Installing : ceph-fuse-2:19.2.3-47.gc24117fd552.el9.clyso.x86 140/152 2026-03-06T13:39:17.537 INFO:teuthology.orchestra.run.vm01.stdout: Running scriptlet: ceph-mon-2:19.2.3-47.gc24117fd552.el9.clyso.x86_ 120/152 2026-03-06T13:39:17.537 INFO:teuthology.orchestra.run.vm01.stdout:Glob pattern passed to enable, but globs are not supported for this. 2026-03-06T13:39:17.537 INFO:teuthology.orchestra.run.vm01.stdout:Invalid unit name "ceph-mon@*.service" escaped as "ceph-mon@\x2a.service". 2026-03-06T13:39:17.537 INFO:teuthology.orchestra.run.vm01.stdout:Created symlink /etc/systemd/system/multi-user.target.wants/ceph-mon.target → /usr/lib/systemd/system/ceph-mon.target. 2026-03-06T13:39:17.537 INFO:teuthology.orchestra.run.vm01.stdout:Created symlink /etc/systemd/system/ceph.target.wants/ceph-mon.target → /usr/lib/systemd/system/ceph-mon.target. 2026-03-06T13:39:17.538 INFO:teuthology.orchestra.run.vm01.stdout: 2026-03-06T13:39:17.541 INFO:teuthology.orchestra.run.vm08.stdout: Installing : ceph-radosgw-2:19.2.3-47.gc24117fd552.el9.clyso. 141/152 2026-03-06T13:39:17.551 INFO:teuthology.orchestra.run.vm01.stdout: Installing : mailcap-2.1.49-5.el9.noarch 121/152 2026-03-06T13:39:17.555 INFO:teuthology.orchestra.run.vm01.stdout: Installing : libconfig-1.7.2-9.el9.x86_64 122/152 2026-03-06T13:39:17.569 INFO:teuthology.orchestra.run.vm08.stdout: Running scriptlet: ceph-radosgw-2:19.2.3-47.gc24117fd552.el9.clyso. 141/152 2026-03-06T13:39:17.569 INFO:teuthology.orchestra.run.vm08.stdout:Glob pattern passed to enable, but globs are not supported for this. 2026-03-06T13:39:17.569 INFO:teuthology.orchestra.run.vm08.stdout:Invalid unit name "ceph-radosgw@*.service" escaped as "ceph-radosgw@\x2a.service". 2026-03-06T13:39:17.569 INFO:teuthology.orchestra.run.vm08.stdout:Created symlink /etc/systemd/system/multi-user.target.wants/ceph-radosgw.target → /usr/lib/systemd/system/ceph-radosgw.target. 2026-03-06T13:39:17.569 INFO:teuthology.orchestra.run.vm08.stdout:Created symlink /etc/systemd/system/ceph.target.wants/ceph-radosgw.target → /usr/lib/systemd/system/ceph-radosgw.target. 2026-03-06T13:39:17.570 INFO:teuthology.orchestra.run.vm08.stdout: 2026-03-06T13:39:17.578 INFO:teuthology.orchestra.run.vm01.stdout: Running scriptlet: libstoragemgmt-1.10.1-1.el9.x86_64 123/152 2026-03-06T13:39:17.578 INFO:teuthology.orchestra.run.vm01.stdout:Creating group 'qat' with GID 994. 2026-03-06T13:39:17.578 INFO:teuthology.orchestra.run.vm01.stdout:Creating group 'libstoragemgmt' with GID 993. 2026-03-06T13:39:17.578 INFO:teuthology.orchestra.run.vm01.stdout:Creating user 'libstoragemgmt' (daemon account for libstoragemgmt) with UID 993 and GID 993. 2026-03-06T13:39:17.578 INFO:teuthology.orchestra.run.vm01.stdout: 2026-03-06T13:39:17.584 INFO:teuthology.orchestra.run.vm08.stdout: Installing : ceph-immutable-object-cache-2:19.2.3-47.gc24117f 142/152 2026-03-06T13:39:17.592 INFO:teuthology.orchestra.run.vm01.stdout: Installing : libstoragemgmt-1.10.1-1.el9.x86_64 123/152 2026-03-06T13:39:17.615 INFO:teuthology.orchestra.run.vm08.stdout: Running scriptlet: ceph-immutable-object-cache-2:19.2.3-47.gc24117f 142/152 2026-03-06T13:39:17.615 INFO:teuthology.orchestra.run.vm08.stdout:Glob pattern passed to enable, but globs are not supported for this. 2026-03-06T13:39:17.615 INFO:teuthology.orchestra.run.vm08.stdout:Invalid unit name "ceph-immutable-object-cache@*.service" escaped as "ceph-immutable-object-cache@\x2a.service". 2026-03-06T13:39:17.615 INFO:teuthology.orchestra.run.vm08.stdout: 2026-03-06T13:39:17.631 INFO:teuthology.orchestra.run.vm01.stdout: Running scriptlet: libstoragemgmt-1.10.1-1.el9.x86_64 123/152 2026-03-06T13:39:17.631 INFO:teuthology.orchestra.run.vm01.stdout:Created symlink /etc/systemd/system/multi-user.target.wants/libstoragemgmt.service → /usr/lib/systemd/system/libstoragemgmt.service. 2026-03-06T13:39:17.631 INFO:teuthology.orchestra.run.vm01.stdout: 2026-03-06T13:39:17.656 INFO:teuthology.orchestra.run.vm01.stdout: Installing : python3-libstoragemgmt-1.10.1-1.el9.x86_64 124/152 2026-03-06T13:39:17.684 INFO:teuthology.orchestra.run.vm01.stdout: Installing : fuse-2.9.9-17.el9.x86_64 125/152 2026-03-06T13:39:17.761 INFO:teuthology.orchestra.run.vm01.stdout: Installing : cryptsetup-2.8.1-3.el9.x86_64 126/152 2026-03-06T13:39:17.766 INFO:teuthology.orchestra.run.vm01.stdout: Installing : ceph-volume-2:19.2.3-47.gc24117fd552.el9.clyso.n 127/152 2026-03-06T13:39:17.776 INFO:teuthology.orchestra.run.vm08.stdout: Installing : rbd-mirror-2:19.2.3-47.gc24117fd552.el9.clyso.x8 143/152 2026-03-06T13:39:17.781 INFO:teuthology.orchestra.run.vm01.stdout: Running scriptlet: ceph-volume-2:19.2.3-47.gc24117fd552.el9.clyso.n 127/152 2026-03-06T13:39:17.781 INFO:teuthology.orchestra.run.vm01.stdout:Glob pattern passed to enable, but globs are not supported for this. 2026-03-06T13:39:17.781 INFO:teuthology.orchestra.run.vm01.stdout:Invalid unit name "ceph-volume@*.service" escaped as "ceph-volume@\x2a.service". 2026-03-06T13:39:17.781 INFO:teuthology.orchestra.run.vm01.stdout: 2026-03-06T13:39:17.797 INFO:teuthology.orchestra.run.vm08.stdout: Running scriptlet: rbd-mirror-2:19.2.3-47.gc24117fd552.el9.clyso.x8 143/152 2026-03-06T13:39:17.797 INFO:teuthology.orchestra.run.vm08.stdout:Glob pattern passed to enable, but globs are not supported for this. 2026-03-06T13:39:17.797 INFO:teuthology.orchestra.run.vm08.stdout:Invalid unit name "ceph-rbd-mirror@*.service" escaped as "ceph-rbd-mirror@\x2a.service". 2026-03-06T13:39:17.797 INFO:teuthology.orchestra.run.vm08.stdout:Created symlink /etc/systemd/system/multi-user.target.wants/ceph-rbd-mirror.target → /usr/lib/systemd/system/ceph-rbd-mirror.target. 2026-03-06T13:39:17.797 INFO:teuthology.orchestra.run.vm08.stdout:Created symlink /etc/systemd/system/ceph.target.wants/ceph-rbd-mirror.target → /usr/lib/systemd/system/ceph-rbd-mirror.target. 2026-03-06T13:39:17.797 INFO:teuthology.orchestra.run.vm08.stdout: 2026-03-06T13:39:18.603 INFO:teuthology.orchestra.run.vm01.stdout: Installing : ceph-osd-2:19.2.3-47.gc24117fd552.el9.clyso.x86_ 128/152 2026-03-06T13:39:18.630 INFO:teuthology.orchestra.run.vm01.stdout: Running scriptlet: ceph-osd-2:19.2.3-47.gc24117fd552.el9.clyso.x86_ 128/152 2026-03-06T13:39:18.630 INFO:teuthology.orchestra.run.vm01.stdout:Glob pattern passed to enable, but globs are not supported for this. 2026-03-06T13:39:18.630 INFO:teuthology.orchestra.run.vm01.stdout:Invalid unit name "ceph-osd@*.service" escaped as "ceph-osd@\x2a.service". 2026-03-06T13:39:18.630 INFO:teuthology.orchestra.run.vm01.stdout:Created symlink /etc/systemd/system/multi-user.target.wants/ceph-osd.target → /usr/lib/systemd/system/ceph-osd.target. 2026-03-06T13:39:18.631 INFO:teuthology.orchestra.run.vm01.stdout:Created symlink /etc/systemd/system/ceph.target.wants/ceph-osd.target → /usr/lib/systemd/system/ceph-osd.target. 2026-03-06T13:39:18.631 INFO:teuthology.orchestra.run.vm01.stdout: 2026-03-06T13:39:18.704 INFO:teuthology.orchestra.run.vm01.stdout: Running scriptlet: cephadm-2:19.2.3-47.gc24117fd552.el9.clyso.noarc 129/152 2026-03-06T13:39:18.708 INFO:teuthology.orchestra.run.vm01.stdout: Installing : cephadm-2:19.2.3-47.gc24117fd552.el9.clyso.noarc 129/152 2026-03-06T13:39:18.714 INFO:teuthology.orchestra.run.vm01.stdout: Installing : ceph-prometheus-alerts-2:19.2.3-47.gc24117fd552. 130/152 2026-03-06T13:39:18.738 INFO:teuthology.orchestra.run.vm01.stdout: Installing : ceph-grafana-dashboards-2:19.2.3-47.gc24117fd552 131/152 2026-03-06T13:39:18.741 INFO:teuthology.orchestra.run.vm01.stdout: Installing : ceph-mgr-cephadm-2:19.2.3-47.gc24117fd552.el9.cl 132/152 2026-03-06T13:39:19.303 INFO:teuthology.orchestra.run.vm01.stdout: Running scriptlet: ceph-mgr-cephadm-2:19.2.3-47.gc24117fd552.el9.cl 132/152 2026-03-06T13:39:19.309 INFO:teuthology.orchestra.run.vm01.stdout: Installing : ceph-mgr-dashboard-2:19.2.3-47.gc24117fd552.el9. 133/152 2026-03-06T13:39:19.794 INFO:teuthology.orchestra.run.vm00.stdout: Installing : ceph-test-2:19.2.3-47.gc24117fd552.el9.clyso.x86 144/152 2026-03-06T13:39:19.805 INFO:teuthology.orchestra.run.vm00.stdout: Installing : rbd-fuse-2:19.2.3-47.gc24117fd552.el9.clyso.x86_ 145/152 2026-03-06T13:39:19.840 INFO:teuthology.orchestra.run.vm00.stdout: Installing : rbd-nbd-2:19.2.3-47.gc24117fd552.el9.clyso.x86_6 146/152 2026-03-06T13:39:19.848 INFO:teuthology.orchestra.run.vm00.stdout: Installing : perl-Test-Harness-1:3.42-461.el9.noarch 147/152 2026-03-06T13:39:19.849 INFO:teuthology.orchestra.run.vm01.stdout: Running scriptlet: ceph-mgr-dashboard-2:19.2.3-47.gc24117fd552.el9. 133/152 2026-03-06T13:39:19.851 INFO:teuthology.orchestra.run.vm01.stdout: Installing : ceph-mgr-diskprediction-local-2:19.2.3-47.gc2411 134/152 2026-03-06T13:39:19.865 INFO:teuthology.orchestra.run.vm01.stdout: Running scriptlet: ceph-mgr-diskprediction-local-2:19.2.3-47.gc2411 134/152 2026-03-06T13:39:19.865 INFO:teuthology.orchestra.run.vm00.stdout: Installing : libcephfs-devel-2:19.2.3-47.gc24117fd552.el9.cly 148/152 2026-03-06T13:39:19.866 INFO:teuthology.orchestra.run.vm01.stdout: Installing : ceph-mgr-k8sevents-2:19.2.3-47.gc24117fd552.el9. 135/152 2026-03-06T13:39:19.873 INFO:teuthology.orchestra.run.vm00.stdout: Installing : s3cmd-2.4.0-1.el9.noarch 149/152 2026-03-06T13:39:19.877 INFO:teuthology.orchestra.run.vm00.stdout: Installing : bzip2-1.0.8-11.el9.x86_64 150/152 2026-03-06T13:39:19.877 INFO:teuthology.orchestra.run.vm00.stdout: Cleanup : librbd1-2:16.2.4-5.el9.x86_64 151/152 2026-03-06T13:39:19.897 INFO:teuthology.orchestra.run.vm00.stdout: Running scriptlet: librbd1-2:16.2.4-5.el9.x86_64 151/152 2026-03-06T13:39:19.897 INFO:teuthology.orchestra.run.vm00.stdout: Cleanup : librados2-2:16.2.4-5.el9.x86_64 152/152 2026-03-06T13:39:19.933 INFO:teuthology.orchestra.run.vm01.stdout: Running scriptlet: ceph-mgr-k8sevents-2:19.2.3-47.gc24117fd552.el9. 135/152 2026-03-06T13:39:19.993 INFO:teuthology.orchestra.run.vm01.stdout: Installing : ceph-mgr-modules-core-2:19.2.3-47.gc24117fd552.e 136/152 2026-03-06T13:39:19.996 INFO:teuthology.orchestra.run.vm01.stdout: Installing : ceph-mgr-2:19.2.3-47.gc24117fd552.el9.clyso.x86_ 137/152 2026-03-06T13:39:20.018 INFO:teuthology.orchestra.run.vm01.stdout: Running scriptlet: ceph-mgr-2:19.2.3-47.gc24117fd552.el9.clyso.x86_ 137/152 2026-03-06T13:39:20.018 INFO:teuthology.orchestra.run.vm01.stdout:Glob pattern passed to enable, but globs are not supported for this. 2026-03-06T13:39:20.018 INFO:teuthology.orchestra.run.vm01.stdout:Invalid unit name "ceph-mgr@*.service" escaped as "ceph-mgr@\x2a.service". 2026-03-06T13:39:20.018 INFO:teuthology.orchestra.run.vm01.stdout:Created symlink /etc/systemd/system/multi-user.target.wants/ceph-mgr.target → /usr/lib/systemd/system/ceph-mgr.target. 2026-03-06T13:39:20.018 INFO:teuthology.orchestra.run.vm01.stdout:Created symlink /etc/systemd/system/ceph.target.wants/ceph-mgr.target → /usr/lib/systemd/system/ceph-mgr.target. 2026-03-06T13:39:20.018 INFO:teuthology.orchestra.run.vm01.stdout: 2026-03-06T13:39:20.033 INFO:teuthology.orchestra.run.vm01.stdout: Installing : ceph-mgr-rook-2:19.2.3-47.gc24117fd552.el9.clyso 138/152 2026-03-06T13:39:20.046 INFO:teuthology.orchestra.run.vm01.stdout: Running scriptlet: ceph-mgr-rook-2:19.2.3-47.gc24117fd552.el9.clyso 138/152 2026-03-06T13:39:20.106 INFO:teuthology.orchestra.run.vm01.stdout: Installing : ceph-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 139/152 2026-03-06T13:39:20.451 INFO:teuthology.orchestra.run.vm08.stdout: Installing : ceph-test-2:19.2.3-47.gc24117fd552.el9.clyso.x86 144/152 2026-03-06T13:39:20.463 INFO:teuthology.orchestra.run.vm08.stdout: Installing : rbd-fuse-2:19.2.3-47.gc24117fd552.el9.clyso.x86_ 145/152 2026-03-06T13:39:20.501 INFO:teuthology.orchestra.run.vm08.stdout: Installing : rbd-nbd-2:19.2.3-47.gc24117fd552.el9.clyso.x86_6 146/152 2026-03-06T13:39:20.509 INFO:teuthology.orchestra.run.vm08.stdout: Installing : perl-Test-Harness-1:3.42-461.el9.noarch 147/152 2026-03-06T13:39:20.529 INFO:teuthology.orchestra.run.vm08.stdout: Installing : libcephfs-devel-2:19.2.3-47.gc24117fd552.el9.cly 148/152 2026-03-06T13:39:20.538 INFO:teuthology.orchestra.run.vm08.stdout: Installing : s3cmd-2.4.0-1.el9.noarch 149/152 2026-03-06T13:39:20.542 INFO:teuthology.orchestra.run.vm08.stdout: Installing : bzip2-1.0.8-11.el9.x86_64 150/152 2026-03-06T13:39:20.542 INFO:teuthology.orchestra.run.vm08.stdout: Cleanup : librbd1-2:16.2.4-5.el9.x86_64 151/152 2026-03-06T13:39:20.560 INFO:teuthology.orchestra.run.vm08.stdout: Running scriptlet: librbd1-2:16.2.4-5.el9.x86_64 151/152 2026-03-06T13:39:20.560 INFO:teuthology.orchestra.run.vm08.stdout: Cleanup : librados2-2:16.2.4-5.el9.x86_64 152/152 2026-03-06T13:39:20.663 INFO:teuthology.orchestra.run.vm01.stdout: Installing : ceph-fuse-2:19.2.3-47.gc24117fd552.el9.clyso.x86 140/152 2026-03-06T13:39:20.756 INFO:teuthology.orchestra.run.vm01.stdout: Installing : ceph-radosgw-2:19.2.3-47.gc24117fd552.el9.clyso. 141/152 2026-03-06T13:39:20.778 INFO:teuthology.orchestra.run.vm01.stdout: Running scriptlet: ceph-radosgw-2:19.2.3-47.gc24117fd552.el9.clyso. 141/152 2026-03-06T13:39:20.778 INFO:teuthology.orchestra.run.vm01.stdout:Glob pattern passed to enable, but globs are not supported for this. 2026-03-06T13:39:20.778 INFO:teuthology.orchestra.run.vm01.stdout:Invalid unit name "ceph-radosgw@*.service" escaped as "ceph-radosgw@\x2a.service". 2026-03-06T13:39:20.778 INFO:teuthology.orchestra.run.vm01.stdout:Created symlink /etc/systemd/system/multi-user.target.wants/ceph-radosgw.target → /usr/lib/systemd/system/ceph-radosgw.target. 2026-03-06T13:39:20.778 INFO:teuthology.orchestra.run.vm01.stdout:Created symlink /etc/systemd/system/ceph.target.wants/ceph-radosgw.target → /usr/lib/systemd/system/ceph-radosgw.target. 2026-03-06T13:39:20.778 INFO:teuthology.orchestra.run.vm01.stdout: 2026-03-06T13:39:20.849 INFO:teuthology.orchestra.run.vm01.stdout: Installing : ceph-immutable-object-cache-2:19.2.3-47.gc24117f 142/152 2026-03-06T13:39:20.870 INFO:teuthology.orchestra.run.vm01.stdout: Running scriptlet: ceph-immutable-object-cache-2:19.2.3-47.gc24117f 142/152 2026-03-06T13:39:20.875 INFO:teuthology.orchestra.run.vm01.stdout:Glob pattern passed to enable, but globs are not supported for this. 2026-03-06T13:39:20.875 INFO:teuthology.orchestra.run.vm01.stdout:Invalid unit name "ceph-immutable-object-cache@*.service" escaped as "ceph-immutable-object-cache@\x2a.service". 2026-03-06T13:39:20.875 INFO:teuthology.orchestra.run.vm01.stdout: 2026-03-06T13:39:21.110 INFO:teuthology.orchestra.run.vm01.stdout: Installing : rbd-mirror-2:19.2.3-47.gc24117fd552.el9.clyso.x8 143/152 2026-03-06T13:39:21.134 INFO:teuthology.orchestra.run.vm01.stdout: Running scriptlet: rbd-mirror-2:19.2.3-47.gc24117fd552.el9.clyso.x8 143/152 2026-03-06T13:39:21.134 INFO:teuthology.orchestra.run.vm01.stdout:Glob pattern passed to enable, but globs are not supported for this. 2026-03-06T13:39:21.134 INFO:teuthology.orchestra.run.vm01.stdout:Invalid unit name "ceph-rbd-mirror@*.service" escaped as "ceph-rbd-mirror@\x2a.service". 2026-03-06T13:39:21.134 INFO:teuthology.orchestra.run.vm01.stdout:Created symlink /etc/systemd/system/multi-user.target.wants/ceph-rbd-mirror.target → /usr/lib/systemd/system/ceph-rbd-mirror.target. 2026-03-06T13:39:21.134 INFO:teuthology.orchestra.run.vm01.stdout:Created symlink /etc/systemd/system/ceph.target.wants/ceph-rbd-mirror.target → /usr/lib/systemd/system/ceph-rbd-mirror.target. 2026-03-06T13:39:21.134 INFO:teuthology.orchestra.run.vm01.stdout: 2026-03-06T13:39:21.264 INFO:teuthology.orchestra.run.vm00.stdout: Running scriptlet: librados2-2:16.2.4-5.el9.x86_64 152/152 2026-03-06T13:39:21.264 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : ceph-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 1/152 2026-03-06T13:39:21.264 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : ceph-base-2:19.2.3-47.gc24117fd552.el9.clyso.x86 2/152 2026-03-06T13:39:21.264 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : ceph-common-2:19.2.3-47.gc24117fd552.el9.clyso.x 3/152 2026-03-06T13:39:21.264 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : ceph-fuse-2:19.2.3-47.gc24117fd552.el9.clyso.x86 4/152 2026-03-06T13:39:21.264 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : ceph-immutable-object-cache-2:19.2.3-47.gc24117f 5/152 2026-03-06T13:39:21.264 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : ceph-mds-2:19.2.3-47.gc24117fd552.el9.clyso.x86_ 6/152 2026-03-06T13:39:21.265 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : ceph-mgr-2:19.2.3-47.gc24117fd552.el9.clyso.x86_ 7/152 2026-03-06T13:39:21.265 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : ceph-mon-2:19.2.3-47.gc24117fd552.el9.clyso.x86_ 8/152 2026-03-06T13:39:21.265 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : ceph-osd-2:19.2.3-47.gc24117fd552.el9.clyso.x86_ 9/152 2026-03-06T13:39:21.265 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : ceph-radosgw-2:19.2.3-47.gc24117fd552.el9.clyso. 10/152 2026-03-06T13:39:21.265 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : ceph-selinux-2:19.2.3-47.gc24117fd552.el9.clyso. 11/152 2026-03-06T13:39:21.265 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : ceph-test-2:19.2.3-47.gc24117fd552.el9.clyso.x86 12/152 2026-03-06T13:39:21.265 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : libcephfs-devel-2:19.2.3-47.gc24117fd552.el9.cly 13/152 2026-03-06T13:39:21.265 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : libcephfs2-2:19.2.3-47.gc24117fd552.el9.clyso.x8 14/152 2026-03-06T13:39:21.265 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : libcephsqlite-2:19.2.3-47.gc24117fd552.el9.clyso 15/152 2026-03-06T13:39:21.265 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : librados-devel-2:19.2.3-47.gc24117fd552.el9.clys 16/152 2026-03-06T13:39:21.265 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : libradosstriper1-2:19.2.3-47.gc24117fd552.el9.cl 17/152 2026-03-06T13:39:21.265 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : librgw2-2:19.2.3-47.gc24117fd552.el9.clyso.x86_6 18/152 2026-03-06T13:39:21.265 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : python3-ceph-argparse-2:19.2.3-47.gc24117fd552.e 19/152 2026-03-06T13:39:21.265 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : python3-ceph-common-2:19.2.3-47.gc24117fd552.el9 20/152 2026-03-06T13:39:21.265 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : python3-cephfs-2:19.2.3-47.gc24117fd552.el9.clys 21/152 2026-03-06T13:39:21.265 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : python3-rados-2:19.2.3-47.gc24117fd552.el9.clyso 22/152 2026-03-06T13:39:21.265 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : python3-rbd-2:19.2.3-47.gc24117fd552.el9.clyso.x 23/152 2026-03-06T13:39:21.265 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : python3-rgw-2:19.2.3-47.gc24117fd552.el9.clyso.x 24/152 2026-03-06T13:39:21.265 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : rbd-fuse-2:19.2.3-47.gc24117fd552.el9.clyso.x86_ 25/152 2026-03-06T13:39:21.265 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : rbd-mirror-2:19.2.3-47.gc24117fd552.el9.clyso.x8 26/152 2026-03-06T13:39:21.265 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : rbd-nbd-2:19.2.3-47.gc24117fd552.el9.clyso.x86_6 27/152 2026-03-06T13:39:21.265 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : ceph-grafana-dashboards-2:19.2.3-47.gc24117fd552 28/152 2026-03-06T13:39:21.265 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : ceph-mgr-cephadm-2:19.2.3-47.gc24117fd552.el9.cl 29/152 2026-03-06T13:39:21.265 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : ceph-mgr-dashboard-2:19.2.3-47.gc24117fd552.el9. 30/152 2026-03-06T13:39:21.265 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : ceph-mgr-diskprediction-local-2:19.2.3-47.gc2411 31/152 2026-03-06T13:39:21.265 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : ceph-mgr-k8sevents-2:19.2.3-47.gc24117fd552.el9. 32/152 2026-03-06T13:39:21.265 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : ceph-mgr-modules-core-2:19.2.3-47.gc24117fd552.e 33/152 2026-03-06T13:39:21.265 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : ceph-mgr-rook-2:19.2.3-47.gc24117fd552.el9.clyso 34/152 2026-03-06T13:39:21.265 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : ceph-prometheus-alerts-2:19.2.3-47.gc24117fd552. 35/152 2026-03-06T13:39:21.265 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : ceph-volume-2:19.2.3-47.gc24117fd552.el9.clyso.n 36/152 2026-03-06T13:39:21.265 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : cephadm-2:19.2.3-47.gc24117fd552.el9.clyso.noarc 37/152 2026-03-06T13:39:21.267 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : bzip2-1.0.8-11.el9.x86_64 38/152 2026-03-06T13:39:21.267 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : cryptsetup-2.8.1-3.el9.x86_64 39/152 2026-03-06T13:39:21.267 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : fuse-2.9.9-17.el9.x86_64 40/152 2026-03-06T13:39:21.267 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : ledmon-libs-1.1.0-3.el9.x86_64 41/152 2026-03-06T13:39:21.267 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : libconfig-1.7.2-9.el9.x86_64 42/152 2026-03-06T13:39:21.267 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : libgfortran-11.5.0-14.el9.x86_64 43/152 2026-03-06T13:39:21.267 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : libquadmath-11.5.0-14.el9.x86_64 44/152 2026-03-06T13:39:21.267 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : mailcap-2.1.49-5.el9.noarch 45/152 2026-03-06T13:39:21.268 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : pciutils-3.7.0-7.el9.x86_64 46/152 2026-03-06T13:39:21.268 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : python3-cffi-1.14.5-5.el9.x86_64 47/152 2026-03-06T13:39:21.268 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : python3-cryptography-36.0.1-5.el9.x86_64 48/152 2026-03-06T13:39:21.268 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : python3-ply-3.11-14.el9.noarch 49/152 2026-03-06T13:39:21.268 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : python3-pycparser-2.20-6.el9.noarch 50/152 2026-03-06T13:39:21.268 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : python3-requests-2.25.1-10.el9.noarch 51/152 2026-03-06T13:39:21.268 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : python3-urllib3-1.26.5-7.el9.noarch 52/152 2026-03-06T13:39:21.268 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : smartmontools-1:7.2-10.el9.x86_64 53/152 2026-03-06T13:39:21.268 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : unzip-6.0-59.el9.x86_64 54/152 2026-03-06T13:39:21.268 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : zip-3.0-35.el9.x86_64 55/152 2026-03-06T13:39:21.268 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : boost-program-options-1.75.0-13.el9.x86_64 56/152 2026-03-06T13:39:21.268 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : flexiblas-3.0.4-9.el9.x86_64 57/152 2026-03-06T13:39:21.268 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : flexiblas-netlib-3.0.4-9.el9.x86_64 58/152 2026-03-06T13:39:21.268 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : flexiblas-openblas-openmp-3.0.4-9.el9.x86_64 59/152 2026-03-06T13:39:21.268 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : libnbd-1.20.3-4.el9.x86_64 60/152 2026-03-06T13:39:21.268 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : libpmemobj-1.12.1-1.el9.x86_64 61/152 2026-03-06T13:39:21.268 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : librabbitmq-0.11.0-7.el9.x86_64 62/152 2026-03-06T13:39:21.268 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : librdkafka-1.6.1-102.el9.x86_64 63/152 2026-03-06T13:39:21.268 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : libstoragemgmt-1.10.1-1.el9.x86_64 64/152 2026-03-06T13:39:21.268 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : libxslt-1.1.34-12.el9.x86_64 65/152 2026-03-06T13:39:21.268 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : lttng-ust-2.12.0-6.el9.x86_64 66/152 2026-03-06T13:39:21.268 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : lua-5.4.4-4.el9.x86_64 67/152 2026-03-06T13:39:21.268 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : openblas-0.3.29-1.el9.x86_64 68/152 2026-03-06T13:39:21.268 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : openblas-openmp-0.3.29-1.el9.x86_64 69/152 2026-03-06T13:39:21.268 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : perl-Benchmark-1.23-483.el9.noarch 70/152 2026-03-06T13:39:21.268 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : perl-Test-Harness-1:3.42-461.el9.noarch 71/152 2026-03-06T13:39:21.268 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : protobuf-3.14.0-17.el9.x86_64 72/152 2026-03-06T13:39:21.268 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : python3-babel-2.9.1-2.el9.noarch 73/152 2026-03-06T13:39:21.268 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : python3-devel-3.9.25-3.el9.x86_64 74/152 2026-03-06T13:39:21.268 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : python3-jinja2-2.11.3-8.el9.noarch 75/152 2026-03-06T13:39:21.268 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : python3-libstoragemgmt-1.10.1-1.el9.x86_64 76/152 2026-03-06T13:39:21.268 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : python3-lxml-4.6.5-3.el9.x86_64 77/152 2026-03-06T13:39:21.268 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : python3-mako-1.1.4-6.el9.noarch 78/152 2026-03-06T13:39:21.268 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : python3-markupsafe-1.1.1-12.el9.x86_64 79/152 2026-03-06T13:39:21.268 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : python3-numpy-1:1.23.5-2.el9.x86_64 80/152 2026-03-06T13:39:21.268 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : python3-numpy-f2py-1:1.23.5-2.el9.x86_64 81/152 2026-03-06T13:39:21.268 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : python3-packaging-20.9-5.el9.noarch 82/152 2026-03-06T13:39:21.268 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : python3-protobuf-3.14.0-17.el9.noarch 83/152 2026-03-06T13:39:21.268 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : python3-pyasn1-0.4.8-7.el9.noarch 84/152 2026-03-06T13:39:21.268 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : python3-pyasn1-modules-0.4.8-7.el9.noarch 85/152 2026-03-06T13:39:21.268 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : python3-requests-oauthlib-1.3.0-12.el9.noarch 86/152 2026-03-06T13:39:21.268 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : python3-scipy-1.9.3-2.el9.x86_64 87/152 2026-03-06T13:39:21.268 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : python3-toml-0.10.2-6.el9.noarch 88/152 2026-03-06T13:39:21.268 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : qatlib-25.08.0-2.el9.x86_64 89/152 2026-03-06T13:39:21.268 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : qatlib-service-25.08.0-2.el9.x86_64 90/152 2026-03-06T13:39:21.268 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : qatzip-libs-1.3.1-1.el9.x86_64 91/152 2026-03-06T13:39:21.268 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : socat-1.7.4.1-8.el9.x86_64 92/152 2026-03-06T13:39:21.268 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : xmlsec1-1.2.29-13.el9.x86_64 93/152 2026-03-06T13:39:21.268 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : xmlsec1-openssl-1.2.29-13.el9.x86_64 94/152 2026-03-06T13:39:21.268 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : xmlstarlet-1.6.1-20.el9.x86_64 95/152 2026-03-06T13:39:21.268 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : lua-devel-5.4.4-4.el9.x86_64 96/152 2026-03-06T13:39:21.268 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : protobuf-compiler-3.14.0-17.el9.x86_64 97/152 2026-03-06T13:39:21.268 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : abseil-cpp-20211102.0-4.el9.x86_64 98/152 2026-03-06T13:39:21.268 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : gperftools-libs-2.9.1-3.el9.x86_64 99/152 2026-03-06T13:39:21.269 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : grpc-data-1.46.7-10.el9.noarch 100/152 2026-03-06T13:39:21.269 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : libarrow-9.0.0-15.el9.x86_64 101/152 2026-03-06T13:39:21.269 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : libarrow-doc-9.0.0-15.el9.noarch 102/152 2026-03-06T13:39:21.269 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : liboath-2.6.12-1.el9.x86_64 103/152 2026-03-06T13:39:21.269 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : libunwind-1.6.2-1.el9.x86_64 104/152 2026-03-06T13:39:21.269 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : luarocks-3.9.2-5.el9.noarch 105/152 2026-03-06T13:39:21.269 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : parquet-libs-9.0.0-15.el9.x86_64 106/152 2026-03-06T13:39:21.269 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : python3-asyncssh-2.13.2-5.el9.noarch 107/152 2026-03-06T13:39:21.269 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : python3-autocommand-2.2.2-8.el9.noarch 108/152 2026-03-06T13:39:21.269 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : python3-backports-tarfile-1.2.0-1.el9.noarch 109/152 2026-03-06T13:39:21.269 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : python3-bcrypt-3.2.2-1.el9.x86_64 110/152 2026-03-06T13:39:21.269 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : python3-cachetools-4.2.4-1.el9.noarch 111/152 2026-03-06T13:39:21.269 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : python3-certifi-2023.05.07-4.el9.noarch 112/152 2026-03-06T13:39:21.269 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : python3-cheroot-10.0.1-4.el9.noarch 113/152 2026-03-06T13:39:21.269 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : python3-cherrypy-18.6.1-2.el9.noarch 114/152 2026-03-06T13:39:21.269 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : python3-google-auth-1:2.45.0-1.el9.noarch 115/152 2026-03-06T13:39:21.269 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : python3-grpcio-1.46.7-10.el9.x86_64 116/152 2026-03-06T13:39:21.269 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : python3-grpcio-tools-1.46.7-10.el9.x86_64 117/152 2026-03-06T13:39:21.269 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : python3-influxdb-5.3.1-1.el9.noarch 118/152 2026-03-06T13:39:21.269 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : python3-isodate-0.6.1-3.el9.noarch 119/152 2026-03-06T13:39:21.269 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : python3-jaraco-8.2.1-3.el9.noarch 120/152 2026-03-06T13:39:21.269 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : python3-jaraco-classes-3.2.1-5.el9.noarch 121/152 2026-03-06T13:39:21.269 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : python3-jaraco-collections-3.0.0-8.el9.noarch 122/152 2026-03-06T13:39:21.269 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : python3-jaraco-context-6.0.1-3.el9.noarch 123/152 2026-03-06T13:39:21.269 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : python3-jaraco-functools-3.5.0-2.el9.noarch 124/152 2026-03-06T13:39:21.269 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : python3-jaraco-text-4.0.0-2.el9.noarch 125/152 2026-03-06T13:39:21.269 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : python3-kubernetes-1:26.1.0-3.el9.noarch 126/152 2026-03-06T13:39:21.269 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : python3-logutils-0.3.5-21.el9.noarch 127/152 2026-03-06T13:39:21.269 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : python3-more-itertools-8.12.0-2.el9.noarch 128/152 2026-03-06T13:39:21.269 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : python3-msgpack-1.0.3-2.el9.x86_64 129/152 2026-03-06T13:39:21.269 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : python3-natsort-7.1.1-5.el9.noarch 130/152 2026-03-06T13:39:21.269 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : python3-pecan-1.4.2-3.el9.noarch 131/152 2026-03-06T13:39:21.269 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : python3-portend-3.1.0-2.el9.noarch 132/152 2026-03-06T13:39:21.269 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : python3-pyOpenSSL-21.0.0-1.el9.noarch 133/152 2026-03-06T13:39:21.269 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : python3-repoze-lru-0.7-16.el9.noarch 134/152 2026-03-06T13:39:21.269 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : python3-routes-2.5.1-5.el9.noarch 135/152 2026-03-06T13:39:21.269 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : python3-rsa-4.9-2.el9.noarch 136/152 2026-03-06T13:39:21.269 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : python3-saml-1.16.0-1.el9.noarch 137/152 2026-03-06T13:39:21.269 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : python3-tempora-5.0.0-2.el9.noarch 138/152 2026-03-06T13:39:21.269 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : python3-typing-extensions-4.15.0-1.el9.noarch 139/152 2026-03-06T13:39:21.269 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : python3-webob-1.8.8-2.el9.noarch 140/152 2026-03-06T13:39:21.269 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : python3-websocket-client-1.2.3-2.el9.noarch 141/152 2026-03-06T13:39:21.269 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : python3-werkzeug-2.0.3-3.el9.1.noarch 142/152 2026-03-06T13:39:21.269 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : python3-xmlsec-1.3.13-1.el9.x86_64 143/152 2026-03-06T13:39:21.269 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : python3-xmltodict-0.12.0-15.el9.noarch 144/152 2026-03-06T13:39:21.269 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : python3-zc-lockfile-2.0-10.el9.noarch 145/152 2026-03-06T13:39:21.269 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : re2-1:20211101-20.el9.x86_64 146/152 2026-03-06T13:39:21.269 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : s3cmd-2.4.0-1.el9.noarch 147/152 2026-03-06T13:39:21.269 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : thrift-0.15.0-4.el9.x86_64 148/152 2026-03-06T13:39:21.269 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : librados2-2:19.2.3-47.gc24117fd552.el9.clyso.x86 149/152 2026-03-06T13:39:21.269 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : librados2-2:16.2.4-5.el9.x86_64 150/152 2026-03-06T13:39:21.269 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : librbd1-2:19.2.3-47.gc24117fd552.el9.clyso.x86_6 151/152 2026-03-06T13:39:21.628 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : librbd1-2:16.2.4-5.el9.x86_64 152/152 2026-03-06T13:39:21.628 INFO:teuthology.orchestra.run.vm00.stdout: 2026-03-06T13:39:21.628 INFO:teuthology.orchestra.run.vm00.stdout:Upgraded: 2026-03-06T13:39:21.628 INFO:teuthology.orchestra.run.vm00.stdout: librados2-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 2026-03-06T13:39:21.628 INFO:teuthology.orchestra.run.vm00.stdout: librbd1-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 2026-03-06T13:39:21.629 INFO:teuthology.orchestra.run.vm00.stdout:Installed: 2026-03-06T13:39:21.629 INFO:teuthology.orchestra.run.vm00.stdout: abseil-cpp-20211102.0-4.el9.x86_64 2026-03-06T13:39:21.629 INFO:teuthology.orchestra.run.vm00.stdout: boost-program-options-1.75.0-13.el9.x86_64 2026-03-06T13:39:21.629 INFO:teuthology.orchestra.run.vm00.stdout: bzip2-1.0.8-11.el9.x86_64 2026-03-06T13:39:21.629 INFO:teuthology.orchestra.run.vm00.stdout: ceph-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 2026-03-06T13:39:21.629 INFO:teuthology.orchestra.run.vm00.stdout: ceph-base-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 2026-03-06T13:39:21.629 INFO:teuthology.orchestra.run.vm00.stdout: ceph-common-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 2026-03-06T13:39:21.629 INFO:teuthology.orchestra.run.vm00.stdout: ceph-fuse-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 2026-03-06T13:39:21.629 INFO:teuthology.orchestra.run.vm00.stdout: ceph-grafana-dashboards-2:19.2.3-47.gc24117fd552.el9.clyso.noarch 2026-03-06T13:39:21.629 INFO:teuthology.orchestra.run.vm00.stdout: ceph-immutable-object-cache-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 2026-03-06T13:39:21.629 INFO:teuthology.orchestra.run.vm00.stdout: ceph-mds-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 2026-03-06T13:39:21.629 INFO:teuthology.orchestra.run.vm00.stdout: ceph-mgr-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 2026-03-06T13:39:21.629 INFO:teuthology.orchestra.run.vm00.stdout: ceph-mgr-cephadm-2:19.2.3-47.gc24117fd552.el9.clyso.noarch 2026-03-06T13:39:21.629 INFO:teuthology.orchestra.run.vm00.stdout: ceph-mgr-dashboard-2:19.2.3-47.gc24117fd552.el9.clyso.noarch 2026-03-06T13:39:21.629 INFO:teuthology.orchestra.run.vm00.stdout: ceph-mgr-diskprediction-local-2:19.2.3-47.gc24117fd552.el9.clyso.noarch 2026-03-06T13:39:21.629 INFO:teuthology.orchestra.run.vm00.stdout: ceph-mgr-k8sevents-2:19.2.3-47.gc24117fd552.el9.clyso.noarch 2026-03-06T13:39:21.629 INFO:teuthology.orchestra.run.vm00.stdout: ceph-mgr-modules-core-2:19.2.3-47.gc24117fd552.el9.clyso.noarch 2026-03-06T13:39:21.629 INFO:teuthology.orchestra.run.vm00.stdout: ceph-mgr-rook-2:19.2.3-47.gc24117fd552.el9.clyso.noarch 2026-03-06T13:39:21.629 INFO:teuthology.orchestra.run.vm00.stdout: ceph-mon-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 2026-03-06T13:39:21.629 INFO:teuthology.orchestra.run.vm00.stdout: ceph-osd-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 2026-03-06T13:39:21.629 INFO:teuthology.orchestra.run.vm00.stdout: ceph-prometheus-alerts-2:19.2.3-47.gc24117fd552.el9.clyso.noarch 2026-03-06T13:39:21.629 INFO:teuthology.orchestra.run.vm00.stdout: ceph-radosgw-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 2026-03-06T13:39:21.629 INFO:teuthology.orchestra.run.vm00.stdout: ceph-selinux-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 2026-03-06T13:39:21.629 INFO:teuthology.orchestra.run.vm00.stdout: ceph-test-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 2026-03-06T13:39:21.629 INFO:teuthology.orchestra.run.vm00.stdout: ceph-volume-2:19.2.3-47.gc24117fd552.el9.clyso.noarch 2026-03-06T13:39:21.629 INFO:teuthology.orchestra.run.vm00.stdout: cephadm-2:19.2.3-47.gc24117fd552.el9.clyso.noarch 2026-03-06T13:39:21.629 INFO:teuthology.orchestra.run.vm00.stdout: cryptsetup-2.8.1-3.el9.x86_64 2026-03-06T13:39:21.629 INFO:teuthology.orchestra.run.vm00.stdout: flexiblas-3.0.4-9.el9.x86_64 2026-03-06T13:39:21.629 INFO:teuthology.orchestra.run.vm00.stdout: flexiblas-netlib-3.0.4-9.el9.x86_64 2026-03-06T13:39:21.629 INFO:teuthology.orchestra.run.vm00.stdout: flexiblas-openblas-openmp-3.0.4-9.el9.x86_64 2026-03-06T13:39:21.629 INFO:teuthology.orchestra.run.vm00.stdout: fuse-2.9.9-17.el9.x86_64 2026-03-06T13:39:21.629 INFO:teuthology.orchestra.run.vm00.stdout: gperftools-libs-2.9.1-3.el9.x86_64 2026-03-06T13:39:21.629 INFO:teuthology.orchestra.run.vm00.stdout: grpc-data-1.46.7-10.el9.noarch 2026-03-06T13:39:21.629 INFO:teuthology.orchestra.run.vm00.stdout: ledmon-libs-1.1.0-3.el9.x86_64 2026-03-06T13:39:21.629 INFO:teuthology.orchestra.run.vm00.stdout: libarrow-9.0.0-15.el9.x86_64 2026-03-06T13:39:21.629 INFO:teuthology.orchestra.run.vm00.stdout: libarrow-doc-9.0.0-15.el9.noarch 2026-03-06T13:39:21.629 INFO:teuthology.orchestra.run.vm00.stdout: libcephfs-devel-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 2026-03-06T13:39:21.629 INFO:teuthology.orchestra.run.vm00.stdout: libcephfs2-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 2026-03-06T13:39:21.629 INFO:teuthology.orchestra.run.vm00.stdout: libcephsqlite-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 2026-03-06T13:39:21.630 INFO:teuthology.orchestra.run.vm00.stdout: libconfig-1.7.2-9.el9.x86_64 2026-03-06T13:39:21.630 INFO:teuthology.orchestra.run.vm00.stdout: libgfortran-11.5.0-14.el9.x86_64 2026-03-06T13:39:21.630 INFO:teuthology.orchestra.run.vm00.stdout: libnbd-1.20.3-4.el9.x86_64 2026-03-06T13:39:21.630 INFO:teuthology.orchestra.run.vm00.stdout: liboath-2.6.12-1.el9.x86_64 2026-03-06T13:39:21.630 INFO:teuthology.orchestra.run.vm00.stdout: libpmemobj-1.12.1-1.el9.x86_64 2026-03-06T13:39:21.630 INFO:teuthology.orchestra.run.vm00.stdout: libquadmath-11.5.0-14.el9.x86_64 2026-03-06T13:39:21.630 INFO:teuthology.orchestra.run.vm00.stdout: librabbitmq-0.11.0-7.el9.x86_64 2026-03-06T13:39:21.630 INFO:teuthology.orchestra.run.vm00.stdout: librados-devel-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 2026-03-06T13:39:21.630 INFO:teuthology.orchestra.run.vm00.stdout: libradosstriper1-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 2026-03-06T13:39:21.630 INFO:teuthology.orchestra.run.vm00.stdout: librdkafka-1.6.1-102.el9.x86_64 2026-03-06T13:39:21.630 INFO:teuthology.orchestra.run.vm00.stdout: librgw2-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 2026-03-06T13:39:21.630 INFO:teuthology.orchestra.run.vm00.stdout: libstoragemgmt-1.10.1-1.el9.x86_64 2026-03-06T13:39:21.630 INFO:teuthology.orchestra.run.vm00.stdout: libunwind-1.6.2-1.el9.x86_64 2026-03-06T13:39:21.630 INFO:teuthology.orchestra.run.vm00.stdout: libxslt-1.1.34-12.el9.x86_64 2026-03-06T13:39:21.630 INFO:teuthology.orchestra.run.vm00.stdout: lttng-ust-2.12.0-6.el9.x86_64 2026-03-06T13:39:21.630 INFO:teuthology.orchestra.run.vm00.stdout: lua-5.4.4-4.el9.x86_64 2026-03-06T13:39:21.630 INFO:teuthology.orchestra.run.vm00.stdout: lua-devel-5.4.4-4.el9.x86_64 2026-03-06T13:39:21.630 INFO:teuthology.orchestra.run.vm00.stdout: luarocks-3.9.2-5.el9.noarch 2026-03-06T13:39:21.630 INFO:teuthology.orchestra.run.vm00.stdout: mailcap-2.1.49-5.el9.noarch 2026-03-06T13:39:21.630 INFO:teuthology.orchestra.run.vm00.stdout: openblas-0.3.29-1.el9.x86_64 2026-03-06T13:39:21.630 INFO:teuthology.orchestra.run.vm00.stdout: openblas-openmp-0.3.29-1.el9.x86_64 2026-03-06T13:39:21.630 INFO:teuthology.orchestra.run.vm00.stdout: parquet-libs-9.0.0-15.el9.x86_64 2026-03-06T13:39:21.630 INFO:teuthology.orchestra.run.vm00.stdout: pciutils-3.7.0-7.el9.x86_64 2026-03-06T13:39:21.630 INFO:teuthology.orchestra.run.vm00.stdout: perl-Benchmark-1.23-483.el9.noarch 2026-03-06T13:39:21.630 INFO:teuthology.orchestra.run.vm00.stdout: perl-Test-Harness-1:3.42-461.el9.noarch 2026-03-06T13:39:21.630 INFO:teuthology.orchestra.run.vm00.stdout: protobuf-3.14.0-17.el9.x86_64 2026-03-06T13:39:21.630 INFO:teuthology.orchestra.run.vm00.stdout: protobuf-compiler-3.14.0-17.el9.x86_64 2026-03-06T13:39:21.630 INFO:teuthology.orchestra.run.vm00.stdout: python3-asyncssh-2.13.2-5.el9.noarch 2026-03-06T13:39:21.630 INFO:teuthology.orchestra.run.vm00.stdout: python3-autocommand-2.2.2-8.el9.noarch 2026-03-06T13:39:21.630 INFO:teuthology.orchestra.run.vm00.stdout: python3-babel-2.9.1-2.el9.noarch 2026-03-06T13:39:21.630 INFO:teuthology.orchestra.run.vm00.stdout: python3-backports-tarfile-1.2.0-1.el9.noarch 2026-03-06T13:39:21.630 INFO:teuthology.orchestra.run.vm00.stdout: python3-bcrypt-3.2.2-1.el9.x86_64 2026-03-06T13:39:21.630 INFO:teuthology.orchestra.run.vm00.stdout: python3-cachetools-4.2.4-1.el9.noarch 2026-03-06T13:39:21.630 INFO:teuthology.orchestra.run.vm00.stdout: python3-ceph-argparse-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 2026-03-06T13:39:21.630 INFO:teuthology.orchestra.run.vm00.stdout: python3-ceph-common-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 2026-03-06T13:39:21.630 INFO:teuthology.orchestra.run.vm00.stdout: python3-cephfs-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 2026-03-06T13:39:21.630 INFO:teuthology.orchestra.run.vm00.stdout: python3-certifi-2023.05.07-4.el9.noarch 2026-03-06T13:39:21.630 INFO:teuthology.orchestra.run.vm00.stdout: python3-cffi-1.14.5-5.el9.x86_64 2026-03-06T13:39:21.630 INFO:teuthology.orchestra.run.vm00.stdout: python3-cheroot-10.0.1-4.el9.noarch 2026-03-06T13:39:21.630 INFO:teuthology.orchestra.run.vm00.stdout: python3-cherrypy-18.6.1-2.el9.noarch 2026-03-06T13:39:21.630 INFO:teuthology.orchestra.run.vm00.stdout: python3-cryptography-36.0.1-5.el9.x86_64 2026-03-06T13:39:21.630 INFO:teuthology.orchestra.run.vm00.stdout: python3-devel-3.9.25-3.el9.x86_64 2026-03-06T13:39:21.630 INFO:teuthology.orchestra.run.vm00.stdout: python3-google-auth-1:2.45.0-1.el9.noarch 2026-03-06T13:39:21.630 INFO:teuthology.orchestra.run.vm00.stdout: python3-grpcio-1.46.7-10.el9.x86_64 2026-03-06T13:39:21.630 INFO:teuthology.orchestra.run.vm00.stdout: python3-grpcio-tools-1.46.7-10.el9.x86_64 2026-03-06T13:39:21.630 INFO:teuthology.orchestra.run.vm00.stdout: python3-influxdb-5.3.1-1.el9.noarch 2026-03-06T13:39:21.630 INFO:teuthology.orchestra.run.vm00.stdout: python3-isodate-0.6.1-3.el9.noarch 2026-03-06T13:39:21.630 INFO:teuthology.orchestra.run.vm00.stdout: python3-jaraco-8.2.1-3.el9.noarch 2026-03-06T13:39:21.631 INFO:teuthology.orchestra.run.vm00.stdout: python3-jaraco-classes-3.2.1-5.el9.noarch 2026-03-06T13:39:21.631 INFO:teuthology.orchestra.run.vm00.stdout: python3-jaraco-collections-3.0.0-8.el9.noarch 2026-03-06T13:39:21.631 INFO:teuthology.orchestra.run.vm00.stdout: python3-jaraco-context-6.0.1-3.el9.noarch 2026-03-06T13:39:21.631 INFO:teuthology.orchestra.run.vm00.stdout: python3-jaraco-functools-3.5.0-2.el9.noarch 2026-03-06T13:39:21.631 INFO:teuthology.orchestra.run.vm00.stdout: python3-jaraco-text-4.0.0-2.el9.noarch 2026-03-06T13:39:21.631 INFO:teuthology.orchestra.run.vm00.stdout: python3-jinja2-2.11.3-8.el9.noarch 2026-03-06T13:39:21.631 INFO:teuthology.orchestra.run.vm00.stdout: python3-kubernetes-1:26.1.0-3.el9.noarch 2026-03-06T13:39:21.631 INFO:teuthology.orchestra.run.vm00.stdout: python3-libstoragemgmt-1.10.1-1.el9.x86_64 2026-03-06T13:39:21.631 INFO:teuthology.orchestra.run.vm00.stdout: python3-logutils-0.3.5-21.el9.noarch 2026-03-06T13:39:21.631 INFO:teuthology.orchestra.run.vm00.stdout: python3-lxml-4.6.5-3.el9.x86_64 2026-03-06T13:39:21.631 INFO:teuthology.orchestra.run.vm00.stdout: python3-mako-1.1.4-6.el9.noarch 2026-03-06T13:39:21.631 INFO:teuthology.orchestra.run.vm00.stdout: python3-markupsafe-1.1.1-12.el9.x86_64 2026-03-06T13:39:21.631 INFO:teuthology.orchestra.run.vm00.stdout: python3-more-itertools-8.12.0-2.el9.noarch 2026-03-06T13:39:21.631 INFO:teuthology.orchestra.run.vm00.stdout: python3-msgpack-1.0.3-2.el9.x86_64 2026-03-06T13:39:21.631 INFO:teuthology.orchestra.run.vm00.stdout: python3-natsort-7.1.1-5.el9.noarch 2026-03-06T13:39:21.631 INFO:teuthology.orchestra.run.vm00.stdout: python3-numpy-1:1.23.5-2.el9.x86_64 2026-03-06T13:39:21.631 INFO:teuthology.orchestra.run.vm00.stdout: python3-numpy-f2py-1:1.23.5-2.el9.x86_64 2026-03-06T13:39:21.631 INFO:teuthology.orchestra.run.vm00.stdout: python3-packaging-20.9-5.el9.noarch 2026-03-06T13:39:21.631 INFO:teuthology.orchestra.run.vm00.stdout: python3-pecan-1.4.2-3.el9.noarch 2026-03-06T13:39:21.631 INFO:teuthology.orchestra.run.vm00.stdout: python3-ply-3.11-14.el9.noarch 2026-03-06T13:39:21.631 INFO:teuthology.orchestra.run.vm00.stdout: python3-portend-3.1.0-2.el9.noarch 2026-03-06T13:39:21.631 INFO:teuthology.orchestra.run.vm00.stdout: python3-protobuf-3.14.0-17.el9.noarch 2026-03-06T13:39:21.631 INFO:teuthology.orchestra.run.vm00.stdout: python3-pyOpenSSL-21.0.0-1.el9.noarch 2026-03-06T13:39:21.631 INFO:teuthology.orchestra.run.vm00.stdout: python3-pyasn1-0.4.8-7.el9.noarch 2026-03-06T13:39:21.631 INFO:teuthology.orchestra.run.vm00.stdout: python3-pyasn1-modules-0.4.8-7.el9.noarch 2026-03-06T13:39:21.631 INFO:teuthology.orchestra.run.vm00.stdout: python3-pycparser-2.20-6.el9.noarch 2026-03-06T13:39:21.631 INFO:teuthology.orchestra.run.vm00.stdout: python3-rados-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 2026-03-06T13:39:21.631 INFO:teuthology.orchestra.run.vm00.stdout: python3-rbd-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 2026-03-06T13:39:21.631 INFO:teuthology.orchestra.run.vm00.stdout: python3-repoze-lru-0.7-16.el9.noarch 2026-03-06T13:39:21.631 INFO:teuthology.orchestra.run.vm00.stdout: python3-requests-2.25.1-10.el9.noarch 2026-03-06T13:39:21.631 INFO:teuthology.orchestra.run.vm00.stdout: python3-requests-oauthlib-1.3.0-12.el9.noarch 2026-03-06T13:39:21.631 INFO:teuthology.orchestra.run.vm00.stdout: python3-rgw-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 2026-03-06T13:39:21.631 INFO:teuthology.orchestra.run.vm00.stdout: python3-routes-2.5.1-5.el9.noarch 2026-03-06T13:39:21.631 INFO:teuthology.orchestra.run.vm00.stdout: python3-rsa-4.9-2.el9.noarch 2026-03-06T13:39:21.631 INFO:teuthology.orchestra.run.vm00.stdout: python3-saml-1.16.0-1.el9.noarch 2026-03-06T13:39:21.631 INFO:teuthology.orchestra.run.vm00.stdout: python3-scipy-1.9.3-2.el9.x86_64 2026-03-06T13:39:21.631 INFO:teuthology.orchestra.run.vm00.stdout: python3-tempora-5.0.0-2.el9.noarch 2026-03-06T13:39:21.631 INFO:teuthology.orchestra.run.vm00.stdout: python3-toml-0.10.2-6.el9.noarch 2026-03-06T13:39:21.631 INFO:teuthology.orchestra.run.vm00.stdout: python3-typing-extensions-4.15.0-1.el9.noarch 2026-03-06T13:39:21.631 INFO:teuthology.orchestra.run.vm00.stdout: python3-urllib3-1.26.5-7.el9.noarch 2026-03-06T13:39:21.631 INFO:teuthology.orchestra.run.vm00.stdout: python3-webob-1.8.8-2.el9.noarch 2026-03-06T13:39:21.631 INFO:teuthology.orchestra.run.vm00.stdout: python3-websocket-client-1.2.3-2.el9.noarch 2026-03-06T13:39:21.631 INFO:teuthology.orchestra.run.vm00.stdout: python3-werkzeug-2.0.3-3.el9.1.noarch 2026-03-06T13:39:21.631 INFO:teuthology.orchestra.run.vm00.stdout: python3-xmlsec-1.3.13-1.el9.x86_64 2026-03-06T13:39:21.631 INFO:teuthology.orchestra.run.vm00.stdout: python3-xmltodict-0.12.0-15.el9.noarch 2026-03-06T13:39:21.631 INFO:teuthology.orchestra.run.vm00.stdout: python3-zc-lockfile-2.0-10.el9.noarch 2026-03-06T13:39:21.631 INFO:teuthology.orchestra.run.vm00.stdout: qatlib-25.08.0-2.el9.x86_64 2026-03-06T13:39:21.631 INFO:teuthology.orchestra.run.vm00.stdout: qatlib-service-25.08.0-2.el9.x86_64 2026-03-06T13:39:21.631 INFO:teuthology.orchestra.run.vm00.stdout: qatzip-libs-1.3.1-1.el9.x86_64 2026-03-06T13:39:21.631 INFO:teuthology.orchestra.run.vm00.stdout: rbd-fuse-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 2026-03-06T13:39:21.631 INFO:teuthology.orchestra.run.vm00.stdout: rbd-mirror-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 2026-03-06T13:39:21.631 INFO:teuthology.orchestra.run.vm00.stdout: rbd-nbd-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 2026-03-06T13:39:21.631 INFO:teuthology.orchestra.run.vm00.stdout: re2-1:20211101-20.el9.x86_64 2026-03-06T13:39:21.631 INFO:teuthology.orchestra.run.vm00.stdout: s3cmd-2.4.0-1.el9.noarch 2026-03-06T13:39:21.631 INFO:teuthology.orchestra.run.vm00.stdout: smartmontools-1:7.2-10.el9.x86_64 2026-03-06T13:39:21.631 INFO:teuthology.orchestra.run.vm00.stdout: socat-1.7.4.1-8.el9.x86_64 2026-03-06T13:39:21.631 INFO:teuthology.orchestra.run.vm00.stdout: thrift-0.15.0-4.el9.x86_64 2026-03-06T13:39:21.631 INFO:teuthology.orchestra.run.vm00.stdout: unzip-6.0-59.el9.x86_64 2026-03-06T13:39:21.631 INFO:teuthology.orchestra.run.vm00.stdout: xmlsec1-1.2.29-13.el9.x86_64 2026-03-06T13:39:21.631 INFO:teuthology.orchestra.run.vm00.stdout: xmlsec1-openssl-1.2.29-13.el9.x86_64 2026-03-06T13:39:21.631 INFO:teuthology.orchestra.run.vm00.stdout: xmlstarlet-1.6.1-20.el9.x86_64 2026-03-06T13:39:21.632 INFO:teuthology.orchestra.run.vm00.stdout: zip-3.0-35.el9.x86_64 2026-03-06T13:39:21.632 INFO:teuthology.orchestra.run.vm00.stdout: 2026-03-06T13:39:21.632 INFO:teuthology.orchestra.run.vm00.stdout:Complete! 2026-03-06T13:39:21.816 DEBUG:teuthology.parallel:result is None 2026-03-06T13:39:21.978 INFO:teuthology.orchestra.run.vm08.stdout: Running scriptlet: librados2-2:16.2.4-5.el9.x86_64 152/152 2026-03-06T13:39:21.978 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : ceph-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 1/152 2026-03-06T13:39:21.978 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : ceph-base-2:19.2.3-47.gc24117fd552.el9.clyso.x86 2/152 2026-03-06T13:39:21.978 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : ceph-common-2:19.2.3-47.gc24117fd552.el9.clyso.x 3/152 2026-03-06T13:39:21.978 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : ceph-fuse-2:19.2.3-47.gc24117fd552.el9.clyso.x86 4/152 2026-03-06T13:39:21.978 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : ceph-immutable-object-cache-2:19.2.3-47.gc24117f 5/152 2026-03-06T13:39:21.978 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : ceph-mds-2:19.2.3-47.gc24117fd552.el9.clyso.x86_ 6/152 2026-03-06T13:39:21.978 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : ceph-mgr-2:19.2.3-47.gc24117fd552.el9.clyso.x86_ 7/152 2026-03-06T13:39:21.979 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : ceph-mon-2:19.2.3-47.gc24117fd552.el9.clyso.x86_ 8/152 2026-03-06T13:39:21.979 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : ceph-osd-2:19.2.3-47.gc24117fd552.el9.clyso.x86_ 9/152 2026-03-06T13:39:21.979 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : ceph-radosgw-2:19.2.3-47.gc24117fd552.el9.clyso. 10/152 2026-03-06T13:39:21.979 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : ceph-selinux-2:19.2.3-47.gc24117fd552.el9.clyso. 11/152 2026-03-06T13:39:21.979 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : ceph-test-2:19.2.3-47.gc24117fd552.el9.clyso.x86 12/152 2026-03-06T13:39:21.979 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : libcephfs-devel-2:19.2.3-47.gc24117fd552.el9.cly 13/152 2026-03-06T13:39:21.979 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : libcephfs2-2:19.2.3-47.gc24117fd552.el9.clyso.x8 14/152 2026-03-06T13:39:21.979 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : libcephsqlite-2:19.2.3-47.gc24117fd552.el9.clyso 15/152 2026-03-06T13:39:21.979 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : librados-devel-2:19.2.3-47.gc24117fd552.el9.clys 16/152 2026-03-06T13:39:21.979 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : libradosstriper1-2:19.2.3-47.gc24117fd552.el9.cl 17/152 2026-03-06T13:39:21.979 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : librgw2-2:19.2.3-47.gc24117fd552.el9.clyso.x86_6 18/152 2026-03-06T13:39:21.979 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : python3-ceph-argparse-2:19.2.3-47.gc24117fd552.e 19/152 2026-03-06T13:39:21.979 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : python3-ceph-common-2:19.2.3-47.gc24117fd552.el9 20/152 2026-03-06T13:39:21.979 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : python3-cephfs-2:19.2.3-47.gc24117fd552.el9.clys 21/152 2026-03-06T13:39:21.979 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : python3-rados-2:19.2.3-47.gc24117fd552.el9.clyso 22/152 2026-03-06T13:39:21.979 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : python3-rbd-2:19.2.3-47.gc24117fd552.el9.clyso.x 23/152 2026-03-06T13:39:21.979 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : python3-rgw-2:19.2.3-47.gc24117fd552.el9.clyso.x 24/152 2026-03-06T13:39:21.979 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : rbd-fuse-2:19.2.3-47.gc24117fd552.el9.clyso.x86_ 25/152 2026-03-06T13:39:21.979 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : rbd-mirror-2:19.2.3-47.gc24117fd552.el9.clyso.x8 26/152 2026-03-06T13:39:21.979 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : rbd-nbd-2:19.2.3-47.gc24117fd552.el9.clyso.x86_6 27/152 2026-03-06T13:39:21.979 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : ceph-grafana-dashboards-2:19.2.3-47.gc24117fd552 28/152 2026-03-06T13:39:21.979 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : ceph-mgr-cephadm-2:19.2.3-47.gc24117fd552.el9.cl 29/152 2026-03-06T13:39:21.979 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : ceph-mgr-dashboard-2:19.2.3-47.gc24117fd552.el9. 30/152 2026-03-06T13:39:21.979 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : ceph-mgr-diskprediction-local-2:19.2.3-47.gc2411 31/152 2026-03-06T13:39:21.979 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : ceph-mgr-k8sevents-2:19.2.3-47.gc24117fd552.el9. 32/152 2026-03-06T13:39:21.979 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : ceph-mgr-modules-core-2:19.2.3-47.gc24117fd552.e 33/152 2026-03-06T13:39:21.979 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : ceph-mgr-rook-2:19.2.3-47.gc24117fd552.el9.clyso 34/152 2026-03-06T13:39:21.979 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : ceph-prometheus-alerts-2:19.2.3-47.gc24117fd552. 35/152 2026-03-06T13:39:21.979 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : ceph-volume-2:19.2.3-47.gc24117fd552.el9.clyso.n 36/152 2026-03-06T13:39:21.980 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : cephadm-2:19.2.3-47.gc24117fd552.el9.clyso.noarc 37/152 2026-03-06T13:39:21.980 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : bzip2-1.0.8-11.el9.x86_64 38/152 2026-03-06T13:39:21.980 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : cryptsetup-2.8.1-3.el9.x86_64 39/152 2026-03-06T13:39:21.980 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : fuse-2.9.9-17.el9.x86_64 40/152 2026-03-06T13:39:21.980 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : ledmon-libs-1.1.0-3.el9.x86_64 41/152 2026-03-06T13:39:21.980 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : libconfig-1.7.2-9.el9.x86_64 42/152 2026-03-06T13:39:21.980 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : libgfortran-11.5.0-14.el9.x86_64 43/152 2026-03-06T13:39:21.980 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : libquadmath-11.5.0-14.el9.x86_64 44/152 2026-03-06T13:39:21.980 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : mailcap-2.1.49-5.el9.noarch 45/152 2026-03-06T13:39:21.980 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : pciutils-3.7.0-7.el9.x86_64 46/152 2026-03-06T13:39:21.980 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : python3-cffi-1.14.5-5.el9.x86_64 47/152 2026-03-06T13:39:21.980 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : python3-cryptography-36.0.1-5.el9.x86_64 48/152 2026-03-06T13:39:21.980 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : python3-ply-3.11-14.el9.noarch 49/152 2026-03-06T13:39:21.980 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : python3-pycparser-2.20-6.el9.noarch 50/152 2026-03-06T13:39:21.980 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : python3-requests-2.25.1-10.el9.noarch 51/152 2026-03-06T13:39:21.980 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : python3-urllib3-1.26.5-7.el9.noarch 52/152 2026-03-06T13:39:21.981 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : smartmontools-1:7.2-10.el9.x86_64 53/152 2026-03-06T13:39:21.981 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : unzip-6.0-59.el9.x86_64 54/152 2026-03-06T13:39:21.981 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : zip-3.0-35.el9.x86_64 55/152 2026-03-06T13:39:21.981 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : boost-program-options-1.75.0-13.el9.x86_64 56/152 2026-03-06T13:39:21.981 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : flexiblas-3.0.4-9.el9.x86_64 57/152 2026-03-06T13:39:21.981 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : flexiblas-netlib-3.0.4-9.el9.x86_64 58/152 2026-03-06T13:39:21.981 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : flexiblas-openblas-openmp-3.0.4-9.el9.x86_64 59/152 2026-03-06T13:39:21.981 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : libnbd-1.20.3-4.el9.x86_64 60/152 2026-03-06T13:39:21.981 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : libpmemobj-1.12.1-1.el9.x86_64 61/152 2026-03-06T13:39:21.981 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : librabbitmq-0.11.0-7.el9.x86_64 62/152 2026-03-06T13:39:21.981 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : librdkafka-1.6.1-102.el9.x86_64 63/152 2026-03-06T13:39:21.981 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : libstoragemgmt-1.10.1-1.el9.x86_64 64/152 2026-03-06T13:39:21.981 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : libxslt-1.1.34-12.el9.x86_64 65/152 2026-03-06T13:39:21.981 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : lttng-ust-2.12.0-6.el9.x86_64 66/152 2026-03-06T13:39:21.981 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : lua-5.4.4-4.el9.x86_64 67/152 2026-03-06T13:39:21.981 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : openblas-0.3.29-1.el9.x86_64 68/152 2026-03-06T13:39:21.981 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : openblas-openmp-0.3.29-1.el9.x86_64 69/152 2026-03-06T13:39:21.981 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : perl-Benchmark-1.23-483.el9.noarch 70/152 2026-03-06T13:39:21.981 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : perl-Test-Harness-1:3.42-461.el9.noarch 71/152 2026-03-06T13:39:21.982 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : protobuf-3.14.0-17.el9.x86_64 72/152 2026-03-06T13:39:21.982 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : python3-babel-2.9.1-2.el9.noarch 73/152 2026-03-06T13:39:21.982 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : python3-devel-3.9.25-3.el9.x86_64 74/152 2026-03-06T13:39:21.982 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : python3-jinja2-2.11.3-8.el9.noarch 75/152 2026-03-06T13:39:21.982 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : python3-libstoragemgmt-1.10.1-1.el9.x86_64 76/152 2026-03-06T13:39:21.982 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : python3-lxml-4.6.5-3.el9.x86_64 77/152 2026-03-06T13:39:21.982 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : python3-mako-1.1.4-6.el9.noarch 78/152 2026-03-06T13:39:21.982 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : python3-markupsafe-1.1.1-12.el9.x86_64 79/152 2026-03-06T13:39:21.982 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : python3-numpy-1:1.23.5-2.el9.x86_64 80/152 2026-03-06T13:39:21.982 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : python3-numpy-f2py-1:1.23.5-2.el9.x86_64 81/152 2026-03-06T13:39:21.982 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : python3-packaging-20.9-5.el9.noarch 82/152 2026-03-06T13:39:21.982 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : python3-protobuf-3.14.0-17.el9.noarch 83/152 2026-03-06T13:39:21.982 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : python3-pyasn1-0.4.8-7.el9.noarch 84/152 2026-03-06T13:39:21.982 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : python3-pyasn1-modules-0.4.8-7.el9.noarch 85/152 2026-03-06T13:39:21.982 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : python3-requests-oauthlib-1.3.0-12.el9.noarch 86/152 2026-03-06T13:39:21.982 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : python3-scipy-1.9.3-2.el9.x86_64 87/152 2026-03-06T13:39:21.982 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : python3-toml-0.10.2-6.el9.noarch 88/152 2026-03-06T13:39:21.982 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : qatlib-25.08.0-2.el9.x86_64 89/152 2026-03-06T13:39:21.982 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : qatlib-service-25.08.0-2.el9.x86_64 90/152 2026-03-06T13:39:21.982 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : qatzip-libs-1.3.1-1.el9.x86_64 91/152 2026-03-06T13:39:21.982 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : socat-1.7.4.1-8.el9.x86_64 92/152 2026-03-06T13:39:21.982 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : xmlsec1-1.2.29-13.el9.x86_64 93/152 2026-03-06T13:39:21.982 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : xmlsec1-openssl-1.2.29-13.el9.x86_64 94/152 2026-03-06T13:39:21.983 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : xmlstarlet-1.6.1-20.el9.x86_64 95/152 2026-03-06T13:39:21.983 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : lua-devel-5.4.4-4.el9.x86_64 96/152 2026-03-06T13:39:21.983 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : protobuf-compiler-3.14.0-17.el9.x86_64 97/152 2026-03-06T13:39:21.983 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : abseil-cpp-20211102.0-4.el9.x86_64 98/152 2026-03-06T13:39:21.983 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : gperftools-libs-2.9.1-3.el9.x86_64 99/152 2026-03-06T13:39:21.983 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : grpc-data-1.46.7-10.el9.noarch 100/152 2026-03-06T13:39:21.983 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : libarrow-9.0.0-15.el9.x86_64 101/152 2026-03-06T13:39:21.983 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : libarrow-doc-9.0.0-15.el9.noarch 102/152 2026-03-06T13:39:21.983 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : liboath-2.6.12-1.el9.x86_64 103/152 2026-03-06T13:39:21.983 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : libunwind-1.6.2-1.el9.x86_64 104/152 2026-03-06T13:39:21.983 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : luarocks-3.9.2-5.el9.noarch 105/152 2026-03-06T13:39:21.983 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : parquet-libs-9.0.0-15.el9.x86_64 106/152 2026-03-06T13:39:21.983 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : python3-asyncssh-2.13.2-5.el9.noarch 107/152 2026-03-06T13:39:21.983 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : python3-autocommand-2.2.2-8.el9.noarch 108/152 2026-03-06T13:39:21.983 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : python3-backports-tarfile-1.2.0-1.el9.noarch 109/152 2026-03-06T13:39:21.983 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : python3-bcrypt-3.2.2-1.el9.x86_64 110/152 2026-03-06T13:39:21.983 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : python3-cachetools-4.2.4-1.el9.noarch 111/152 2026-03-06T13:39:21.983 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : python3-certifi-2023.05.07-4.el9.noarch 112/152 2026-03-06T13:39:21.983 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : python3-cheroot-10.0.1-4.el9.noarch 113/152 2026-03-06T13:39:21.983 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : python3-cherrypy-18.6.1-2.el9.noarch 114/152 2026-03-06T13:39:21.983 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : python3-google-auth-1:2.45.0-1.el9.noarch 115/152 2026-03-06T13:39:21.983 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : python3-grpcio-1.46.7-10.el9.x86_64 116/152 2026-03-06T13:39:21.983 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : python3-grpcio-tools-1.46.7-10.el9.x86_64 117/152 2026-03-06T13:39:21.983 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : python3-influxdb-5.3.1-1.el9.noarch 118/152 2026-03-06T13:39:21.983 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : python3-isodate-0.6.1-3.el9.noarch 119/152 2026-03-06T13:39:21.983 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : python3-jaraco-8.2.1-3.el9.noarch 120/152 2026-03-06T13:39:21.983 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : python3-jaraco-classes-3.2.1-5.el9.noarch 121/152 2026-03-06T13:39:21.983 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : python3-jaraco-collections-3.0.0-8.el9.noarch 122/152 2026-03-06T13:39:21.983 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : python3-jaraco-context-6.0.1-3.el9.noarch 123/152 2026-03-06T13:39:21.983 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : python3-jaraco-functools-3.5.0-2.el9.noarch 124/152 2026-03-06T13:39:21.984 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : python3-jaraco-text-4.0.0-2.el9.noarch 125/152 2026-03-06T13:39:21.984 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : python3-kubernetes-1:26.1.0-3.el9.noarch 126/152 2026-03-06T13:39:21.984 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : python3-logutils-0.3.5-21.el9.noarch 127/152 2026-03-06T13:39:21.984 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : python3-more-itertools-8.12.0-2.el9.noarch 128/152 2026-03-06T13:39:21.984 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : python3-msgpack-1.0.3-2.el9.x86_64 129/152 2026-03-06T13:39:21.984 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : python3-natsort-7.1.1-5.el9.noarch 130/152 2026-03-06T13:39:21.984 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : python3-pecan-1.4.2-3.el9.noarch 131/152 2026-03-06T13:39:21.984 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : python3-portend-3.1.0-2.el9.noarch 132/152 2026-03-06T13:39:21.984 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : python3-pyOpenSSL-21.0.0-1.el9.noarch 133/152 2026-03-06T13:39:21.984 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : python3-repoze-lru-0.7-16.el9.noarch 134/152 2026-03-06T13:39:21.984 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : python3-routes-2.5.1-5.el9.noarch 135/152 2026-03-06T13:39:21.984 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : python3-rsa-4.9-2.el9.noarch 136/152 2026-03-06T13:39:21.984 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : python3-saml-1.16.0-1.el9.noarch 137/152 2026-03-06T13:39:21.984 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : python3-tempora-5.0.0-2.el9.noarch 138/152 2026-03-06T13:39:21.984 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : python3-typing-extensions-4.15.0-1.el9.noarch 139/152 2026-03-06T13:39:21.984 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : python3-webob-1.8.8-2.el9.noarch 140/152 2026-03-06T13:39:21.984 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : python3-websocket-client-1.2.3-2.el9.noarch 141/152 2026-03-06T13:39:21.984 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : python3-werkzeug-2.0.3-3.el9.1.noarch 142/152 2026-03-06T13:39:21.984 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : python3-xmlsec-1.3.13-1.el9.x86_64 143/152 2026-03-06T13:39:21.984 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : python3-xmltodict-0.12.0-15.el9.noarch 144/152 2026-03-06T13:39:21.984 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : python3-zc-lockfile-2.0-10.el9.noarch 145/152 2026-03-06T13:39:21.984 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : re2-1:20211101-20.el9.x86_64 146/152 2026-03-06T13:39:21.984 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : s3cmd-2.4.0-1.el9.noarch 147/152 2026-03-06T13:39:21.984 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : thrift-0.15.0-4.el9.x86_64 148/152 2026-03-06T13:39:21.984 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : librados2-2:19.2.3-47.gc24117fd552.el9.clyso.x86 149/152 2026-03-06T13:39:21.984 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : librados2-2:16.2.4-5.el9.x86_64 150/152 2026-03-06T13:39:21.984 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : librbd1-2:19.2.3-47.gc24117fd552.el9.clyso.x86_6 151/152 2026-03-06T13:39:22.100 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : librbd1-2:16.2.4-5.el9.x86_64 152/152 2026-03-06T13:39:22.100 INFO:teuthology.orchestra.run.vm08.stdout: 2026-03-06T13:39:22.100 INFO:teuthology.orchestra.run.vm08.stdout:Upgraded: 2026-03-06T13:39:22.100 INFO:teuthology.orchestra.run.vm08.stdout: librados2-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 2026-03-06T13:39:22.100 INFO:teuthology.orchestra.run.vm08.stdout: librbd1-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 2026-03-06T13:39:22.100 INFO:teuthology.orchestra.run.vm08.stdout:Installed: 2026-03-06T13:39:22.100 INFO:teuthology.orchestra.run.vm08.stdout: abseil-cpp-20211102.0-4.el9.x86_64 2026-03-06T13:39:22.100 INFO:teuthology.orchestra.run.vm08.stdout: boost-program-options-1.75.0-13.el9.x86_64 2026-03-06T13:39:22.100 INFO:teuthology.orchestra.run.vm08.stdout: bzip2-1.0.8-11.el9.x86_64 2026-03-06T13:39:22.100 INFO:teuthology.orchestra.run.vm08.stdout: ceph-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 2026-03-06T13:39:22.100 INFO:teuthology.orchestra.run.vm08.stdout: ceph-base-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 2026-03-06T13:39:22.100 INFO:teuthology.orchestra.run.vm08.stdout: ceph-common-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 2026-03-06T13:39:22.100 INFO:teuthology.orchestra.run.vm08.stdout: ceph-fuse-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 2026-03-06T13:39:22.100 INFO:teuthology.orchestra.run.vm08.stdout: ceph-grafana-dashboards-2:19.2.3-47.gc24117fd552.el9.clyso.noarch 2026-03-06T13:39:22.100 INFO:teuthology.orchestra.run.vm08.stdout: ceph-immutable-object-cache-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 2026-03-06T13:39:22.100 INFO:teuthology.orchestra.run.vm08.stdout: ceph-mds-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 2026-03-06T13:39:22.100 INFO:teuthology.orchestra.run.vm08.stdout: ceph-mgr-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 2026-03-06T13:39:22.100 INFO:teuthology.orchestra.run.vm08.stdout: ceph-mgr-cephadm-2:19.2.3-47.gc24117fd552.el9.clyso.noarch 2026-03-06T13:39:22.100 INFO:teuthology.orchestra.run.vm08.stdout: ceph-mgr-dashboard-2:19.2.3-47.gc24117fd552.el9.clyso.noarch 2026-03-06T13:39:22.100 INFO:teuthology.orchestra.run.vm08.stdout: ceph-mgr-diskprediction-local-2:19.2.3-47.gc24117fd552.el9.clyso.noarch 2026-03-06T13:39:22.100 INFO:teuthology.orchestra.run.vm08.stdout: ceph-mgr-k8sevents-2:19.2.3-47.gc24117fd552.el9.clyso.noarch 2026-03-06T13:39:22.100 INFO:teuthology.orchestra.run.vm08.stdout: ceph-mgr-modules-core-2:19.2.3-47.gc24117fd552.el9.clyso.noarch 2026-03-06T13:39:22.100 INFO:teuthology.orchestra.run.vm08.stdout: ceph-mgr-rook-2:19.2.3-47.gc24117fd552.el9.clyso.noarch 2026-03-06T13:39:22.100 INFO:teuthology.orchestra.run.vm08.stdout: ceph-mon-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 2026-03-06T13:39:22.100 INFO:teuthology.orchestra.run.vm08.stdout: ceph-osd-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 2026-03-06T13:39:22.100 INFO:teuthology.orchestra.run.vm08.stdout: ceph-prometheus-alerts-2:19.2.3-47.gc24117fd552.el9.clyso.noarch 2026-03-06T13:39:22.100 INFO:teuthology.orchestra.run.vm08.stdout: ceph-radosgw-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 2026-03-06T13:39:22.100 INFO:teuthology.orchestra.run.vm08.stdout: ceph-selinux-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 2026-03-06T13:39:22.100 INFO:teuthology.orchestra.run.vm08.stdout: ceph-test-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 2026-03-06T13:39:22.100 INFO:teuthology.orchestra.run.vm08.stdout: ceph-volume-2:19.2.3-47.gc24117fd552.el9.clyso.noarch 2026-03-06T13:39:22.100 INFO:teuthology.orchestra.run.vm08.stdout: cephadm-2:19.2.3-47.gc24117fd552.el9.clyso.noarch 2026-03-06T13:39:22.100 INFO:teuthology.orchestra.run.vm08.stdout: cryptsetup-2.8.1-3.el9.x86_64 2026-03-06T13:39:22.100 INFO:teuthology.orchestra.run.vm08.stdout: flexiblas-3.0.4-9.el9.x86_64 2026-03-06T13:39:22.100 INFO:teuthology.orchestra.run.vm08.stdout: flexiblas-netlib-3.0.4-9.el9.x86_64 2026-03-06T13:39:22.100 INFO:teuthology.orchestra.run.vm08.stdout: flexiblas-openblas-openmp-3.0.4-9.el9.x86_64 2026-03-06T13:39:22.100 INFO:teuthology.orchestra.run.vm08.stdout: fuse-2.9.9-17.el9.x86_64 2026-03-06T13:39:22.101 INFO:teuthology.orchestra.run.vm08.stdout: gperftools-libs-2.9.1-3.el9.x86_64 2026-03-06T13:39:22.101 INFO:teuthology.orchestra.run.vm08.stdout: grpc-data-1.46.7-10.el9.noarch 2026-03-06T13:39:22.101 INFO:teuthology.orchestra.run.vm08.stdout: ledmon-libs-1.1.0-3.el9.x86_64 2026-03-06T13:39:22.101 INFO:teuthology.orchestra.run.vm08.stdout: libarrow-9.0.0-15.el9.x86_64 2026-03-06T13:39:22.101 INFO:teuthology.orchestra.run.vm08.stdout: libarrow-doc-9.0.0-15.el9.noarch 2026-03-06T13:39:22.101 INFO:teuthology.orchestra.run.vm08.stdout: libcephfs-devel-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 2026-03-06T13:39:22.101 INFO:teuthology.orchestra.run.vm08.stdout: libcephfs2-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 2026-03-06T13:39:22.101 INFO:teuthology.orchestra.run.vm08.stdout: libcephsqlite-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 2026-03-06T13:39:22.101 INFO:teuthology.orchestra.run.vm08.stdout: libconfig-1.7.2-9.el9.x86_64 2026-03-06T13:39:22.101 INFO:teuthology.orchestra.run.vm08.stdout: libgfortran-11.5.0-14.el9.x86_64 2026-03-06T13:39:22.101 INFO:teuthology.orchestra.run.vm08.stdout: libnbd-1.20.3-4.el9.x86_64 2026-03-06T13:39:22.101 INFO:teuthology.orchestra.run.vm08.stdout: liboath-2.6.12-1.el9.x86_64 2026-03-06T13:39:22.101 INFO:teuthology.orchestra.run.vm08.stdout: libpmemobj-1.12.1-1.el9.x86_64 2026-03-06T13:39:22.101 INFO:teuthology.orchestra.run.vm08.stdout: libquadmath-11.5.0-14.el9.x86_64 2026-03-06T13:39:22.101 INFO:teuthology.orchestra.run.vm08.stdout: librabbitmq-0.11.0-7.el9.x86_64 2026-03-06T13:39:22.101 INFO:teuthology.orchestra.run.vm08.stdout: librados-devel-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 2026-03-06T13:39:22.101 INFO:teuthology.orchestra.run.vm08.stdout: libradosstriper1-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 2026-03-06T13:39:22.101 INFO:teuthology.orchestra.run.vm08.stdout: librdkafka-1.6.1-102.el9.x86_64 2026-03-06T13:39:22.101 INFO:teuthology.orchestra.run.vm08.stdout: librgw2-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 2026-03-06T13:39:22.101 INFO:teuthology.orchestra.run.vm08.stdout: libstoragemgmt-1.10.1-1.el9.x86_64 2026-03-06T13:39:22.101 INFO:teuthology.orchestra.run.vm08.stdout: libunwind-1.6.2-1.el9.x86_64 2026-03-06T13:39:22.101 INFO:teuthology.orchestra.run.vm08.stdout: libxslt-1.1.34-12.el9.x86_64 2026-03-06T13:39:22.101 INFO:teuthology.orchestra.run.vm08.stdout: lttng-ust-2.12.0-6.el9.x86_64 2026-03-06T13:39:22.101 INFO:teuthology.orchestra.run.vm08.stdout: lua-5.4.4-4.el9.x86_64 2026-03-06T13:39:22.101 INFO:teuthology.orchestra.run.vm08.stdout: lua-devel-5.4.4-4.el9.x86_64 2026-03-06T13:39:22.101 INFO:teuthology.orchestra.run.vm08.stdout: luarocks-3.9.2-5.el9.noarch 2026-03-06T13:39:22.101 INFO:teuthology.orchestra.run.vm08.stdout: mailcap-2.1.49-5.el9.noarch 2026-03-06T13:39:22.101 INFO:teuthology.orchestra.run.vm08.stdout: openblas-0.3.29-1.el9.x86_64 2026-03-06T13:39:22.101 INFO:teuthology.orchestra.run.vm08.stdout: openblas-openmp-0.3.29-1.el9.x86_64 2026-03-06T13:39:22.101 INFO:teuthology.orchestra.run.vm08.stdout: parquet-libs-9.0.0-15.el9.x86_64 2026-03-06T13:39:22.101 INFO:teuthology.orchestra.run.vm08.stdout: pciutils-3.7.0-7.el9.x86_64 2026-03-06T13:39:22.101 INFO:teuthology.orchestra.run.vm08.stdout: perl-Benchmark-1.23-483.el9.noarch 2026-03-06T13:39:22.101 INFO:teuthology.orchestra.run.vm08.stdout: perl-Test-Harness-1:3.42-461.el9.noarch 2026-03-06T13:39:22.101 INFO:teuthology.orchestra.run.vm08.stdout: protobuf-3.14.0-17.el9.x86_64 2026-03-06T13:39:22.101 INFO:teuthology.orchestra.run.vm08.stdout: protobuf-compiler-3.14.0-17.el9.x86_64 2026-03-06T13:39:22.101 INFO:teuthology.orchestra.run.vm08.stdout: python3-asyncssh-2.13.2-5.el9.noarch 2026-03-06T13:39:22.101 INFO:teuthology.orchestra.run.vm08.stdout: python3-autocommand-2.2.2-8.el9.noarch 2026-03-06T13:39:22.101 INFO:teuthology.orchestra.run.vm08.stdout: python3-babel-2.9.1-2.el9.noarch 2026-03-06T13:39:22.101 INFO:teuthology.orchestra.run.vm08.stdout: python3-backports-tarfile-1.2.0-1.el9.noarch 2026-03-06T13:39:22.101 INFO:teuthology.orchestra.run.vm08.stdout: python3-bcrypt-3.2.2-1.el9.x86_64 2026-03-06T13:39:22.101 INFO:teuthology.orchestra.run.vm08.stdout: python3-cachetools-4.2.4-1.el9.noarch 2026-03-06T13:39:22.101 INFO:teuthology.orchestra.run.vm08.stdout: python3-ceph-argparse-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 2026-03-06T13:39:22.101 INFO:teuthology.orchestra.run.vm08.stdout: python3-ceph-common-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 2026-03-06T13:39:22.101 INFO:teuthology.orchestra.run.vm08.stdout: python3-cephfs-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 2026-03-06T13:39:22.101 INFO:teuthology.orchestra.run.vm08.stdout: python3-certifi-2023.05.07-4.el9.noarch 2026-03-06T13:39:22.101 INFO:teuthology.orchestra.run.vm08.stdout: python3-cffi-1.14.5-5.el9.x86_64 2026-03-06T13:39:22.101 INFO:teuthology.orchestra.run.vm08.stdout: python3-cheroot-10.0.1-4.el9.noarch 2026-03-06T13:39:22.101 INFO:teuthology.orchestra.run.vm08.stdout: python3-cherrypy-18.6.1-2.el9.noarch 2026-03-06T13:39:22.101 INFO:teuthology.orchestra.run.vm08.stdout: python3-cryptography-36.0.1-5.el9.x86_64 2026-03-06T13:39:22.101 INFO:teuthology.orchestra.run.vm08.stdout: python3-devel-3.9.25-3.el9.x86_64 2026-03-06T13:39:22.101 INFO:teuthology.orchestra.run.vm08.stdout: python3-google-auth-1:2.45.0-1.el9.noarch 2026-03-06T13:39:22.101 INFO:teuthology.orchestra.run.vm08.stdout: python3-grpcio-1.46.7-10.el9.x86_64 2026-03-06T13:39:22.101 INFO:teuthology.orchestra.run.vm08.stdout: python3-grpcio-tools-1.46.7-10.el9.x86_64 2026-03-06T13:39:22.101 INFO:teuthology.orchestra.run.vm08.stdout: python3-influxdb-5.3.1-1.el9.noarch 2026-03-06T13:39:22.101 INFO:teuthology.orchestra.run.vm08.stdout: python3-isodate-0.6.1-3.el9.noarch 2026-03-06T13:39:22.101 INFO:teuthology.orchestra.run.vm08.stdout: python3-jaraco-8.2.1-3.el9.noarch 2026-03-06T13:39:22.101 INFO:teuthology.orchestra.run.vm08.stdout: python3-jaraco-classes-3.2.1-5.el9.noarch 2026-03-06T13:39:22.101 INFO:teuthology.orchestra.run.vm08.stdout: python3-jaraco-collections-3.0.0-8.el9.noarch 2026-03-06T13:39:22.101 INFO:teuthology.orchestra.run.vm08.stdout: python3-jaraco-context-6.0.1-3.el9.noarch 2026-03-06T13:39:22.101 INFO:teuthology.orchestra.run.vm08.stdout: python3-jaraco-functools-3.5.0-2.el9.noarch 2026-03-06T13:39:22.101 INFO:teuthology.orchestra.run.vm08.stdout: python3-jaraco-text-4.0.0-2.el9.noarch 2026-03-06T13:39:22.102 INFO:teuthology.orchestra.run.vm08.stdout: python3-jinja2-2.11.3-8.el9.noarch 2026-03-06T13:39:22.102 INFO:teuthology.orchestra.run.vm08.stdout: python3-kubernetes-1:26.1.0-3.el9.noarch 2026-03-06T13:39:22.102 INFO:teuthology.orchestra.run.vm08.stdout: python3-libstoragemgmt-1.10.1-1.el9.x86_64 2026-03-06T13:39:22.102 INFO:teuthology.orchestra.run.vm08.stdout: python3-logutils-0.3.5-21.el9.noarch 2026-03-06T13:39:22.102 INFO:teuthology.orchestra.run.vm08.stdout: python3-lxml-4.6.5-3.el9.x86_64 2026-03-06T13:39:22.102 INFO:teuthology.orchestra.run.vm08.stdout: python3-mako-1.1.4-6.el9.noarch 2026-03-06T13:39:22.102 INFO:teuthology.orchestra.run.vm08.stdout: python3-markupsafe-1.1.1-12.el9.x86_64 2026-03-06T13:39:22.102 INFO:teuthology.orchestra.run.vm08.stdout: python3-more-itertools-8.12.0-2.el9.noarch 2026-03-06T13:39:22.102 INFO:teuthology.orchestra.run.vm08.stdout: python3-msgpack-1.0.3-2.el9.x86_64 2026-03-06T13:39:22.102 INFO:teuthology.orchestra.run.vm08.stdout: python3-natsort-7.1.1-5.el9.noarch 2026-03-06T13:39:22.102 INFO:teuthology.orchestra.run.vm08.stdout: python3-numpy-1:1.23.5-2.el9.x86_64 2026-03-06T13:39:22.102 INFO:teuthology.orchestra.run.vm08.stdout: python3-numpy-f2py-1:1.23.5-2.el9.x86_64 2026-03-06T13:39:22.102 INFO:teuthology.orchestra.run.vm08.stdout: python3-packaging-20.9-5.el9.noarch 2026-03-06T13:39:22.102 INFO:teuthology.orchestra.run.vm08.stdout: python3-pecan-1.4.2-3.el9.noarch 2026-03-06T13:39:22.102 INFO:teuthology.orchestra.run.vm08.stdout: python3-ply-3.11-14.el9.noarch 2026-03-06T13:39:22.102 INFO:teuthology.orchestra.run.vm08.stdout: python3-portend-3.1.0-2.el9.noarch 2026-03-06T13:39:22.102 INFO:teuthology.orchestra.run.vm08.stdout: python3-protobuf-3.14.0-17.el9.noarch 2026-03-06T13:39:22.102 INFO:teuthology.orchestra.run.vm08.stdout: python3-pyOpenSSL-21.0.0-1.el9.noarch 2026-03-06T13:39:22.102 INFO:teuthology.orchestra.run.vm08.stdout: python3-pyasn1-0.4.8-7.el9.noarch 2026-03-06T13:39:22.102 INFO:teuthology.orchestra.run.vm08.stdout: python3-pyasn1-modules-0.4.8-7.el9.noarch 2026-03-06T13:39:22.102 INFO:teuthology.orchestra.run.vm08.stdout: python3-pycparser-2.20-6.el9.noarch 2026-03-06T13:39:22.102 INFO:teuthology.orchestra.run.vm08.stdout: python3-rados-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 2026-03-06T13:39:22.102 INFO:teuthology.orchestra.run.vm08.stdout: python3-rbd-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 2026-03-06T13:39:22.102 INFO:teuthology.orchestra.run.vm08.stdout: python3-repoze-lru-0.7-16.el9.noarch 2026-03-06T13:39:22.102 INFO:teuthology.orchestra.run.vm08.stdout: python3-requests-2.25.1-10.el9.noarch 2026-03-06T13:39:22.102 INFO:teuthology.orchestra.run.vm08.stdout: python3-requests-oauthlib-1.3.0-12.el9.noarch 2026-03-06T13:39:22.102 INFO:teuthology.orchestra.run.vm08.stdout: python3-rgw-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 2026-03-06T13:39:22.102 INFO:teuthology.orchestra.run.vm08.stdout: python3-routes-2.5.1-5.el9.noarch 2026-03-06T13:39:22.102 INFO:teuthology.orchestra.run.vm08.stdout: python3-rsa-4.9-2.el9.noarch 2026-03-06T13:39:22.102 INFO:teuthology.orchestra.run.vm08.stdout: python3-saml-1.16.0-1.el9.noarch 2026-03-06T13:39:22.102 INFO:teuthology.orchestra.run.vm08.stdout: python3-scipy-1.9.3-2.el9.x86_64 2026-03-06T13:39:22.102 INFO:teuthology.orchestra.run.vm08.stdout: python3-tempora-5.0.0-2.el9.noarch 2026-03-06T13:39:22.102 INFO:teuthology.orchestra.run.vm08.stdout: python3-toml-0.10.2-6.el9.noarch 2026-03-06T13:39:22.102 INFO:teuthology.orchestra.run.vm08.stdout: python3-typing-extensions-4.15.0-1.el9.noarch 2026-03-06T13:39:22.102 INFO:teuthology.orchestra.run.vm08.stdout: python3-urllib3-1.26.5-7.el9.noarch 2026-03-06T13:39:22.102 INFO:teuthology.orchestra.run.vm08.stdout: python3-webob-1.8.8-2.el9.noarch 2026-03-06T13:39:22.102 INFO:teuthology.orchestra.run.vm08.stdout: python3-websocket-client-1.2.3-2.el9.noarch 2026-03-06T13:39:22.102 INFO:teuthology.orchestra.run.vm08.stdout: python3-werkzeug-2.0.3-3.el9.1.noarch 2026-03-06T13:39:22.102 INFO:teuthology.orchestra.run.vm08.stdout: python3-xmlsec-1.3.13-1.el9.x86_64 2026-03-06T13:39:22.102 INFO:teuthology.orchestra.run.vm08.stdout: python3-xmltodict-0.12.0-15.el9.noarch 2026-03-06T13:39:22.102 INFO:teuthology.orchestra.run.vm08.stdout: python3-zc-lockfile-2.0-10.el9.noarch 2026-03-06T13:39:22.102 INFO:teuthology.orchestra.run.vm08.stdout: qatlib-25.08.0-2.el9.x86_64 2026-03-06T13:39:22.102 INFO:teuthology.orchestra.run.vm08.stdout: qatlib-service-25.08.0-2.el9.x86_64 2026-03-06T13:39:22.102 INFO:teuthology.orchestra.run.vm08.stdout: qatzip-libs-1.3.1-1.el9.x86_64 2026-03-06T13:39:22.102 INFO:teuthology.orchestra.run.vm08.stdout: rbd-fuse-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 2026-03-06T13:39:22.102 INFO:teuthology.orchestra.run.vm08.stdout: rbd-mirror-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 2026-03-06T13:39:22.102 INFO:teuthology.orchestra.run.vm08.stdout: rbd-nbd-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 2026-03-06T13:39:22.102 INFO:teuthology.orchestra.run.vm08.stdout: re2-1:20211101-20.el9.x86_64 2026-03-06T13:39:22.102 INFO:teuthology.orchestra.run.vm08.stdout: s3cmd-2.4.0-1.el9.noarch 2026-03-06T13:39:22.102 INFO:teuthology.orchestra.run.vm08.stdout: smartmontools-1:7.2-10.el9.x86_64 2026-03-06T13:39:22.102 INFO:teuthology.orchestra.run.vm08.stdout: socat-1.7.4.1-8.el9.x86_64 2026-03-06T13:39:22.102 INFO:teuthology.orchestra.run.vm08.stdout: thrift-0.15.0-4.el9.x86_64 2026-03-06T13:39:22.102 INFO:teuthology.orchestra.run.vm08.stdout: unzip-6.0-59.el9.x86_64 2026-03-06T13:39:22.102 INFO:teuthology.orchestra.run.vm08.stdout: xmlsec1-1.2.29-13.el9.x86_64 2026-03-06T13:39:22.102 INFO:teuthology.orchestra.run.vm08.stdout: xmlsec1-openssl-1.2.29-13.el9.x86_64 2026-03-06T13:39:22.102 INFO:teuthology.orchestra.run.vm08.stdout: xmlstarlet-1.6.1-20.el9.x86_64 2026-03-06T13:39:22.103 INFO:teuthology.orchestra.run.vm08.stdout: zip-3.0-35.el9.x86_64 2026-03-06T13:39:22.103 INFO:teuthology.orchestra.run.vm08.stdout: 2026-03-06T13:39:22.103 INFO:teuthology.orchestra.run.vm08.stdout:Complete! 2026-03-06T13:39:22.201 DEBUG:teuthology.parallel:result is None 2026-03-06T13:39:23.754 INFO:teuthology.orchestra.run.vm01.stdout: Installing : ceph-test-2:19.2.3-47.gc24117fd552.el9.clyso.x86 144/152 2026-03-06T13:39:23.765 INFO:teuthology.orchestra.run.vm01.stdout: Installing : rbd-fuse-2:19.2.3-47.gc24117fd552.el9.clyso.x86_ 145/152 2026-03-06T13:39:23.799 INFO:teuthology.orchestra.run.vm01.stdout: Installing : rbd-nbd-2:19.2.3-47.gc24117fd552.el9.clyso.x86_6 146/152 2026-03-06T13:39:23.806 INFO:teuthology.orchestra.run.vm01.stdout: Installing : perl-Test-Harness-1:3.42-461.el9.noarch 147/152 2026-03-06T13:39:23.824 INFO:teuthology.orchestra.run.vm01.stdout: Installing : libcephfs-devel-2:19.2.3-47.gc24117fd552.el9.cly 148/152 2026-03-06T13:39:23.831 INFO:teuthology.orchestra.run.vm01.stdout: Installing : s3cmd-2.4.0-1.el9.noarch 149/152 2026-03-06T13:39:23.834 INFO:teuthology.orchestra.run.vm01.stdout: Installing : bzip2-1.0.8-11.el9.x86_64 150/152 2026-03-06T13:39:23.834 INFO:teuthology.orchestra.run.vm01.stdout: Cleanup : librbd1-2:16.2.4-5.el9.x86_64 151/152 2026-03-06T13:39:23.851 INFO:teuthology.orchestra.run.vm01.stdout: Running scriptlet: librbd1-2:16.2.4-5.el9.x86_64 151/152 2026-03-06T13:39:23.851 INFO:teuthology.orchestra.run.vm01.stdout: Cleanup : librados2-2:16.2.4-5.el9.x86_64 152/152 2026-03-06T13:39:25.218 INFO:teuthology.orchestra.run.vm01.stdout: Running scriptlet: librados2-2:16.2.4-5.el9.x86_64 152/152 2026-03-06T13:39:25.218 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : ceph-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 1/152 2026-03-06T13:39:25.218 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : ceph-base-2:19.2.3-47.gc24117fd552.el9.clyso.x86 2/152 2026-03-06T13:39:25.218 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : ceph-common-2:19.2.3-47.gc24117fd552.el9.clyso.x 3/152 2026-03-06T13:39:25.218 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : ceph-fuse-2:19.2.3-47.gc24117fd552.el9.clyso.x86 4/152 2026-03-06T13:39:25.218 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : ceph-immutable-object-cache-2:19.2.3-47.gc24117f 5/152 2026-03-06T13:39:25.218 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : ceph-mds-2:19.2.3-47.gc24117fd552.el9.clyso.x86_ 6/152 2026-03-06T13:39:25.218 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : ceph-mgr-2:19.2.3-47.gc24117fd552.el9.clyso.x86_ 7/152 2026-03-06T13:39:25.218 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : ceph-mon-2:19.2.3-47.gc24117fd552.el9.clyso.x86_ 8/152 2026-03-06T13:39:25.218 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : ceph-osd-2:19.2.3-47.gc24117fd552.el9.clyso.x86_ 9/152 2026-03-06T13:39:25.218 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : ceph-radosgw-2:19.2.3-47.gc24117fd552.el9.clyso. 10/152 2026-03-06T13:39:25.218 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : ceph-selinux-2:19.2.3-47.gc24117fd552.el9.clyso. 11/152 2026-03-06T13:39:25.218 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : ceph-test-2:19.2.3-47.gc24117fd552.el9.clyso.x86 12/152 2026-03-06T13:39:25.219 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : libcephfs-devel-2:19.2.3-47.gc24117fd552.el9.cly 13/152 2026-03-06T13:39:25.219 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : libcephfs2-2:19.2.3-47.gc24117fd552.el9.clyso.x8 14/152 2026-03-06T13:39:25.219 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : libcephsqlite-2:19.2.3-47.gc24117fd552.el9.clyso 15/152 2026-03-06T13:39:25.219 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : librados-devel-2:19.2.3-47.gc24117fd552.el9.clys 16/152 2026-03-06T13:39:25.219 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : libradosstriper1-2:19.2.3-47.gc24117fd552.el9.cl 17/152 2026-03-06T13:39:25.219 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : librgw2-2:19.2.3-47.gc24117fd552.el9.clyso.x86_6 18/152 2026-03-06T13:39:25.219 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-ceph-argparse-2:19.2.3-47.gc24117fd552.e 19/152 2026-03-06T13:39:25.219 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-ceph-common-2:19.2.3-47.gc24117fd552.el9 20/152 2026-03-06T13:39:25.219 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-cephfs-2:19.2.3-47.gc24117fd552.el9.clys 21/152 2026-03-06T13:39:25.219 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-rados-2:19.2.3-47.gc24117fd552.el9.clyso 22/152 2026-03-06T13:39:25.219 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-rbd-2:19.2.3-47.gc24117fd552.el9.clyso.x 23/152 2026-03-06T13:39:25.219 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-rgw-2:19.2.3-47.gc24117fd552.el9.clyso.x 24/152 2026-03-06T13:39:25.219 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : rbd-fuse-2:19.2.3-47.gc24117fd552.el9.clyso.x86_ 25/152 2026-03-06T13:39:25.219 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : rbd-mirror-2:19.2.3-47.gc24117fd552.el9.clyso.x8 26/152 2026-03-06T13:39:25.219 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : rbd-nbd-2:19.2.3-47.gc24117fd552.el9.clyso.x86_6 27/152 2026-03-06T13:39:25.219 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : ceph-grafana-dashboards-2:19.2.3-47.gc24117fd552 28/152 2026-03-06T13:39:25.219 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : ceph-mgr-cephadm-2:19.2.3-47.gc24117fd552.el9.cl 29/152 2026-03-06T13:39:25.219 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : ceph-mgr-dashboard-2:19.2.3-47.gc24117fd552.el9. 30/152 2026-03-06T13:39:25.219 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : ceph-mgr-diskprediction-local-2:19.2.3-47.gc2411 31/152 2026-03-06T13:39:25.219 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : ceph-mgr-k8sevents-2:19.2.3-47.gc24117fd552.el9. 32/152 2026-03-06T13:39:25.219 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : ceph-mgr-modules-core-2:19.2.3-47.gc24117fd552.e 33/152 2026-03-06T13:39:25.219 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : ceph-mgr-rook-2:19.2.3-47.gc24117fd552.el9.clyso 34/152 2026-03-06T13:39:25.219 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : ceph-prometheus-alerts-2:19.2.3-47.gc24117fd552. 35/152 2026-03-06T13:39:25.219 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : ceph-volume-2:19.2.3-47.gc24117fd552.el9.clyso.n 36/152 2026-03-06T13:39:25.219 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : cephadm-2:19.2.3-47.gc24117fd552.el9.clyso.noarc 37/152 2026-03-06T13:39:25.219 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : bzip2-1.0.8-11.el9.x86_64 38/152 2026-03-06T13:39:25.219 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : cryptsetup-2.8.1-3.el9.x86_64 39/152 2026-03-06T13:39:25.219 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : fuse-2.9.9-17.el9.x86_64 40/152 2026-03-06T13:39:25.220 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : ledmon-libs-1.1.0-3.el9.x86_64 41/152 2026-03-06T13:39:25.220 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : libconfig-1.7.2-9.el9.x86_64 42/152 2026-03-06T13:39:25.220 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : libgfortran-11.5.0-14.el9.x86_64 43/152 2026-03-06T13:39:25.220 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : libquadmath-11.5.0-14.el9.x86_64 44/152 2026-03-06T13:39:25.220 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : mailcap-2.1.49-5.el9.noarch 45/152 2026-03-06T13:39:25.221 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : pciutils-3.7.0-7.el9.x86_64 46/152 2026-03-06T13:39:25.221 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-cffi-1.14.5-5.el9.x86_64 47/152 2026-03-06T13:39:25.221 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-cryptography-36.0.1-5.el9.x86_64 48/152 2026-03-06T13:39:25.221 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-ply-3.11-14.el9.noarch 49/152 2026-03-06T13:39:25.221 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-pycparser-2.20-6.el9.noarch 50/152 2026-03-06T13:39:25.221 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-requests-2.25.1-10.el9.noarch 51/152 2026-03-06T13:39:25.221 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-urllib3-1.26.5-7.el9.noarch 52/152 2026-03-06T13:39:25.221 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : smartmontools-1:7.2-10.el9.x86_64 53/152 2026-03-06T13:39:25.221 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : unzip-6.0-59.el9.x86_64 54/152 2026-03-06T13:39:25.221 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : zip-3.0-35.el9.x86_64 55/152 2026-03-06T13:39:25.221 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : boost-program-options-1.75.0-13.el9.x86_64 56/152 2026-03-06T13:39:25.221 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : flexiblas-3.0.4-9.el9.x86_64 57/152 2026-03-06T13:39:25.221 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : flexiblas-netlib-3.0.4-9.el9.x86_64 58/152 2026-03-06T13:39:25.221 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : flexiblas-openblas-openmp-3.0.4-9.el9.x86_64 59/152 2026-03-06T13:39:25.221 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : libnbd-1.20.3-4.el9.x86_64 60/152 2026-03-06T13:39:25.221 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : libpmemobj-1.12.1-1.el9.x86_64 61/152 2026-03-06T13:39:25.221 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : librabbitmq-0.11.0-7.el9.x86_64 62/152 2026-03-06T13:39:25.221 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : librdkafka-1.6.1-102.el9.x86_64 63/152 2026-03-06T13:39:25.221 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : libstoragemgmt-1.10.1-1.el9.x86_64 64/152 2026-03-06T13:39:25.221 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : libxslt-1.1.34-12.el9.x86_64 65/152 2026-03-06T13:39:25.221 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : lttng-ust-2.12.0-6.el9.x86_64 66/152 2026-03-06T13:39:25.221 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : lua-5.4.4-4.el9.x86_64 67/152 2026-03-06T13:39:25.221 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : openblas-0.3.29-1.el9.x86_64 68/152 2026-03-06T13:39:25.221 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : openblas-openmp-0.3.29-1.el9.x86_64 69/152 2026-03-06T13:39:25.221 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : perl-Benchmark-1.23-483.el9.noarch 70/152 2026-03-06T13:39:25.222 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : perl-Test-Harness-1:3.42-461.el9.noarch 71/152 2026-03-06T13:39:25.222 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : protobuf-3.14.0-17.el9.x86_64 72/152 2026-03-06T13:39:25.222 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-babel-2.9.1-2.el9.noarch 73/152 2026-03-06T13:39:25.222 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-devel-3.9.25-3.el9.x86_64 74/152 2026-03-06T13:39:25.222 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-jinja2-2.11.3-8.el9.noarch 75/152 2026-03-06T13:39:25.222 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-libstoragemgmt-1.10.1-1.el9.x86_64 76/152 2026-03-06T13:39:25.222 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-lxml-4.6.5-3.el9.x86_64 77/152 2026-03-06T13:39:25.223 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-mako-1.1.4-6.el9.noarch 78/152 2026-03-06T13:39:25.223 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-markupsafe-1.1.1-12.el9.x86_64 79/152 2026-03-06T13:39:25.223 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-numpy-1:1.23.5-2.el9.x86_64 80/152 2026-03-06T13:39:25.223 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-numpy-f2py-1:1.23.5-2.el9.x86_64 81/152 2026-03-06T13:39:25.223 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-packaging-20.9-5.el9.noarch 82/152 2026-03-06T13:39:25.223 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-protobuf-3.14.0-17.el9.noarch 83/152 2026-03-06T13:39:25.223 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-pyasn1-0.4.8-7.el9.noarch 84/152 2026-03-06T13:39:25.223 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-pyasn1-modules-0.4.8-7.el9.noarch 85/152 2026-03-06T13:39:25.223 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-requests-oauthlib-1.3.0-12.el9.noarch 86/152 2026-03-06T13:39:25.223 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-scipy-1.9.3-2.el9.x86_64 87/152 2026-03-06T13:39:25.223 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-toml-0.10.2-6.el9.noarch 88/152 2026-03-06T13:39:25.223 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : qatlib-25.08.0-2.el9.x86_64 89/152 2026-03-06T13:39:25.223 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : qatlib-service-25.08.0-2.el9.x86_64 90/152 2026-03-06T13:39:25.223 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : qatzip-libs-1.3.1-1.el9.x86_64 91/152 2026-03-06T13:39:25.223 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : socat-1.7.4.1-8.el9.x86_64 92/152 2026-03-06T13:39:25.223 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : xmlsec1-1.2.29-13.el9.x86_64 93/152 2026-03-06T13:39:25.223 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : xmlsec1-openssl-1.2.29-13.el9.x86_64 94/152 2026-03-06T13:39:25.223 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : xmlstarlet-1.6.1-20.el9.x86_64 95/152 2026-03-06T13:39:25.223 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : lua-devel-5.4.4-4.el9.x86_64 96/152 2026-03-06T13:39:25.223 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : protobuf-compiler-3.14.0-17.el9.x86_64 97/152 2026-03-06T13:39:25.223 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : abseil-cpp-20211102.0-4.el9.x86_64 98/152 2026-03-06T13:39:25.223 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : gperftools-libs-2.9.1-3.el9.x86_64 99/152 2026-03-06T13:39:25.223 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : grpc-data-1.46.7-10.el9.noarch 100/152 2026-03-06T13:39:25.223 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : libarrow-9.0.0-15.el9.x86_64 101/152 2026-03-06T13:39:25.223 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : libarrow-doc-9.0.0-15.el9.noarch 102/152 2026-03-06T13:39:25.224 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : liboath-2.6.12-1.el9.x86_64 103/152 2026-03-06T13:39:25.224 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : libunwind-1.6.2-1.el9.x86_64 104/152 2026-03-06T13:39:25.224 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : luarocks-3.9.2-5.el9.noarch 105/152 2026-03-06T13:39:25.224 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : parquet-libs-9.0.0-15.el9.x86_64 106/152 2026-03-06T13:39:25.224 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-asyncssh-2.13.2-5.el9.noarch 107/152 2026-03-06T13:39:25.224 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-autocommand-2.2.2-8.el9.noarch 108/152 2026-03-06T13:39:25.224 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-backports-tarfile-1.2.0-1.el9.noarch 109/152 2026-03-06T13:39:25.224 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-bcrypt-3.2.2-1.el9.x86_64 110/152 2026-03-06T13:39:25.224 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-cachetools-4.2.4-1.el9.noarch 111/152 2026-03-06T13:39:25.224 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-certifi-2023.05.07-4.el9.noarch 112/152 2026-03-06T13:39:25.224 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-cheroot-10.0.1-4.el9.noarch 113/152 2026-03-06T13:39:25.224 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-cherrypy-18.6.1-2.el9.noarch 114/152 2026-03-06T13:39:25.224 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-google-auth-1:2.45.0-1.el9.noarch 115/152 2026-03-06T13:39:25.224 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-grpcio-1.46.7-10.el9.x86_64 116/152 2026-03-06T13:39:25.224 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-grpcio-tools-1.46.7-10.el9.x86_64 117/152 2026-03-06T13:39:25.224 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-influxdb-5.3.1-1.el9.noarch 118/152 2026-03-06T13:39:25.224 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-isodate-0.6.1-3.el9.noarch 119/152 2026-03-06T13:39:25.224 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-jaraco-8.2.1-3.el9.noarch 120/152 2026-03-06T13:39:25.224 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-jaraco-classes-3.2.1-5.el9.noarch 121/152 2026-03-06T13:39:25.224 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-jaraco-collections-3.0.0-8.el9.noarch 122/152 2026-03-06T13:39:25.224 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-jaraco-context-6.0.1-3.el9.noarch 123/152 2026-03-06T13:39:25.224 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-jaraco-functools-3.5.0-2.el9.noarch 124/152 2026-03-06T13:39:25.224 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-jaraco-text-4.0.0-2.el9.noarch 125/152 2026-03-06T13:39:25.224 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-kubernetes-1:26.1.0-3.el9.noarch 126/152 2026-03-06T13:39:25.224 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-logutils-0.3.5-21.el9.noarch 127/152 2026-03-06T13:39:25.224 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-more-itertools-8.12.0-2.el9.noarch 128/152 2026-03-06T13:39:25.224 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-msgpack-1.0.3-2.el9.x86_64 129/152 2026-03-06T13:39:25.224 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-natsort-7.1.1-5.el9.noarch 130/152 2026-03-06T13:39:25.224 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-pecan-1.4.2-3.el9.noarch 131/152 2026-03-06T13:39:25.224 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-portend-3.1.0-2.el9.noarch 132/152 2026-03-06T13:39:25.224 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-pyOpenSSL-21.0.0-1.el9.noarch 133/152 2026-03-06T13:39:25.224 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-repoze-lru-0.7-16.el9.noarch 134/152 2026-03-06T13:39:25.224 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-routes-2.5.1-5.el9.noarch 135/152 2026-03-06T13:39:25.225 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-rsa-4.9-2.el9.noarch 136/152 2026-03-06T13:39:25.225 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-saml-1.16.0-1.el9.noarch 137/152 2026-03-06T13:39:25.225 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-tempora-5.0.0-2.el9.noarch 138/152 2026-03-06T13:39:25.225 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-typing-extensions-4.15.0-1.el9.noarch 139/152 2026-03-06T13:39:25.225 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-webob-1.8.8-2.el9.noarch 140/152 2026-03-06T13:39:25.225 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-websocket-client-1.2.3-2.el9.noarch 141/152 2026-03-06T13:39:25.225 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-werkzeug-2.0.3-3.el9.1.noarch 142/152 2026-03-06T13:39:25.225 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-xmlsec-1.3.13-1.el9.x86_64 143/152 2026-03-06T13:39:25.225 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-xmltodict-0.12.0-15.el9.noarch 144/152 2026-03-06T13:39:25.225 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-zc-lockfile-2.0-10.el9.noarch 145/152 2026-03-06T13:39:25.225 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : re2-1:20211101-20.el9.x86_64 146/152 2026-03-06T13:39:25.225 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : s3cmd-2.4.0-1.el9.noarch 147/152 2026-03-06T13:39:25.225 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : thrift-0.15.0-4.el9.x86_64 148/152 2026-03-06T13:39:25.226 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : librados2-2:19.2.3-47.gc24117fd552.el9.clyso.x86 149/152 2026-03-06T13:39:25.226 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : librados2-2:16.2.4-5.el9.x86_64 150/152 2026-03-06T13:39:25.226 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : librbd1-2:19.2.3-47.gc24117fd552.el9.clyso.x86_6 151/152 2026-03-06T13:39:25.368 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : librbd1-2:16.2.4-5.el9.x86_64 152/152 2026-03-06T13:39:25.368 INFO:teuthology.orchestra.run.vm01.stdout: 2026-03-06T13:39:25.368 INFO:teuthology.orchestra.run.vm01.stdout:Upgraded: 2026-03-06T13:39:25.368 INFO:teuthology.orchestra.run.vm01.stdout: librados2-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 2026-03-06T13:39:25.368 INFO:teuthology.orchestra.run.vm01.stdout: librbd1-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 2026-03-06T13:39:25.368 INFO:teuthology.orchestra.run.vm01.stdout:Installed: 2026-03-06T13:39:25.368 INFO:teuthology.orchestra.run.vm01.stdout: abseil-cpp-20211102.0-4.el9.x86_64 2026-03-06T13:39:25.368 INFO:teuthology.orchestra.run.vm01.stdout: boost-program-options-1.75.0-13.el9.x86_64 2026-03-06T13:39:25.368 INFO:teuthology.orchestra.run.vm01.stdout: bzip2-1.0.8-11.el9.x86_64 2026-03-06T13:39:25.368 INFO:teuthology.orchestra.run.vm01.stdout: ceph-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 2026-03-06T13:39:25.368 INFO:teuthology.orchestra.run.vm01.stdout: ceph-base-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 2026-03-06T13:39:25.368 INFO:teuthology.orchestra.run.vm01.stdout: ceph-common-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 2026-03-06T13:39:25.368 INFO:teuthology.orchestra.run.vm01.stdout: ceph-fuse-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 2026-03-06T13:39:25.368 INFO:teuthology.orchestra.run.vm01.stdout: ceph-grafana-dashboards-2:19.2.3-47.gc24117fd552.el9.clyso.noarch 2026-03-06T13:39:25.368 INFO:teuthology.orchestra.run.vm01.stdout: ceph-immutable-object-cache-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 2026-03-06T13:39:25.368 INFO:teuthology.orchestra.run.vm01.stdout: ceph-mds-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 2026-03-06T13:39:25.368 INFO:teuthology.orchestra.run.vm01.stdout: ceph-mgr-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 2026-03-06T13:39:25.368 INFO:teuthology.orchestra.run.vm01.stdout: ceph-mgr-cephadm-2:19.2.3-47.gc24117fd552.el9.clyso.noarch 2026-03-06T13:39:25.368 INFO:teuthology.orchestra.run.vm01.stdout: ceph-mgr-dashboard-2:19.2.3-47.gc24117fd552.el9.clyso.noarch 2026-03-06T13:39:25.368 INFO:teuthology.orchestra.run.vm01.stdout: ceph-mgr-diskprediction-local-2:19.2.3-47.gc24117fd552.el9.clyso.noarch 2026-03-06T13:39:25.368 INFO:teuthology.orchestra.run.vm01.stdout: ceph-mgr-k8sevents-2:19.2.3-47.gc24117fd552.el9.clyso.noarch 2026-03-06T13:39:25.368 INFO:teuthology.orchestra.run.vm01.stdout: ceph-mgr-modules-core-2:19.2.3-47.gc24117fd552.el9.clyso.noarch 2026-03-06T13:39:25.368 INFO:teuthology.orchestra.run.vm01.stdout: ceph-mgr-rook-2:19.2.3-47.gc24117fd552.el9.clyso.noarch 2026-03-06T13:39:25.368 INFO:teuthology.orchestra.run.vm01.stdout: ceph-mon-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 2026-03-06T13:39:25.368 INFO:teuthology.orchestra.run.vm01.stdout: ceph-osd-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 2026-03-06T13:39:25.368 INFO:teuthology.orchestra.run.vm01.stdout: ceph-prometheus-alerts-2:19.2.3-47.gc24117fd552.el9.clyso.noarch 2026-03-06T13:39:25.368 INFO:teuthology.orchestra.run.vm01.stdout: ceph-radosgw-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 2026-03-06T13:39:25.368 INFO:teuthology.orchestra.run.vm01.stdout: ceph-selinux-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 2026-03-06T13:39:25.368 INFO:teuthology.orchestra.run.vm01.stdout: ceph-test-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 2026-03-06T13:39:25.368 INFO:teuthology.orchestra.run.vm01.stdout: ceph-volume-2:19.2.3-47.gc24117fd552.el9.clyso.noarch 2026-03-06T13:39:25.368 INFO:teuthology.orchestra.run.vm01.stdout: cephadm-2:19.2.3-47.gc24117fd552.el9.clyso.noarch 2026-03-06T13:39:25.368 INFO:teuthology.orchestra.run.vm01.stdout: cryptsetup-2.8.1-3.el9.x86_64 2026-03-06T13:39:25.368 INFO:teuthology.orchestra.run.vm01.stdout: flexiblas-3.0.4-9.el9.x86_64 2026-03-06T13:39:25.368 INFO:teuthology.orchestra.run.vm01.stdout: flexiblas-netlib-3.0.4-9.el9.x86_64 2026-03-06T13:39:25.368 INFO:teuthology.orchestra.run.vm01.stdout: flexiblas-openblas-openmp-3.0.4-9.el9.x86_64 2026-03-06T13:39:25.368 INFO:teuthology.orchestra.run.vm01.stdout: fuse-2.9.9-17.el9.x86_64 2026-03-06T13:39:25.368 INFO:teuthology.orchestra.run.vm01.stdout: gperftools-libs-2.9.1-3.el9.x86_64 2026-03-06T13:39:25.369 INFO:teuthology.orchestra.run.vm01.stdout: grpc-data-1.46.7-10.el9.noarch 2026-03-06T13:39:25.369 INFO:teuthology.orchestra.run.vm01.stdout: ledmon-libs-1.1.0-3.el9.x86_64 2026-03-06T13:39:25.369 INFO:teuthology.orchestra.run.vm01.stdout: libarrow-9.0.0-15.el9.x86_64 2026-03-06T13:39:25.369 INFO:teuthology.orchestra.run.vm01.stdout: libarrow-doc-9.0.0-15.el9.noarch 2026-03-06T13:39:25.369 INFO:teuthology.orchestra.run.vm01.stdout: libcephfs-devel-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 2026-03-06T13:39:25.369 INFO:teuthology.orchestra.run.vm01.stdout: libcephfs2-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 2026-03-06T13:39:25.369 INFO:teuthology.orchestra.run.vm01.stdout: libcephsqlite-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 2026-03-06T13:39:25.369 INFO:teuthology.orchestra.run.vm01.stdout: libconfig-1.7.2-9.el9.x86_64 2026-03-06T13:39:25.369 INFO:teuthology.orchestra.run.vm01.stdout: libgfortran-11.5.0-14.el9.x86_64 2026-03-06T13:39:25.369 INFO:teuthology.orchestra.run.vm01.stdout: libnbd-1.20.3-4.el9.x86_64 2026-03-06T13:39:25.369 INFO:teuthology.orchestra.run.vm01.stdout: liboath-2.6.12-1.el9.x86_64 2026-03-06T13:39:25.369 INFO:teuthology.orchestra.run.vm01.stdout: libpmemobj-1.12.1-1.el9.x86_64 2026-03-06T13:39:25.369 INFO:teuthology.orchestra.run.vm01.stdout: libquadmath-11.5.0-14.el9.x86_64 2026-03-06T13:39:25.369 INFO:teuthology.orchestra.run.vm01.stdout: librabbitmq-0.11.0-7.el9.x86_64 2026-03-06T13:39:25.369 INFO:teuthology.orchestra.run.vm01.stdout: librados-devel-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 2026-03-06T13:39:25.369 INFO:teuthology.orchestra.run.vm01.stdout: libradosstriper1-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 2026-03-06T13:39:25.369 INFO:teuthology.orchestra.run.vm01.stdout: librdkafka-1.6.1-102.el9.x86_64 2026-03-06T13:39:25.369 INFO:teuthology.orchestra.run.vm01.stdout: librgw2-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 2026-03-06T13:39:25.369 INFO:teuthology.orchestra.run.vm01.stdout: libstoragemgmt-1.10.1-1.el9.x86_64 2026-03-06T13:39:25.369 INFO:teuthology.orchestra.run.vm01.stdout: libunwind-1.6.2-1.el9.x86_64 2026-03-06T13:39:25.369 INFO:teuthology.orchestra.run.vm01.stdout: libxslt-1.1.34-12.el9.x86_64 2026-03-06T13:39:25.369 INFO:teuthology.orchestra.run.vm01.stdout: lttng-ust-2.12.0-6.el9.x86_64 2026-03-06T13:39:25.369 INFO:teuthology.orchestra.run.vm01.stdout: lua-5.4.4-4.el9.x86_64 2026-03-06T13:39:25.369 INFO:teuthology.orchestra.run.vm01.stdout: lua-devel-5.4.4-4.el9.x86_64 2026-03-06T13:39:25.369 INFO:teuthology.orchestra.run.vm01.stdout: luarocks-3.9.2-5.el9.noarch 2026-03-06T13:39:25.369 INFO:teuthology.orchestra.run.vm01.stdout: mailcap-2.1.49-5.el9.noarch 2026-03-06T13:39:25.369 INFO:teuthology.orchestra.run.vm01.stdout: openblas-0.3.29-1.el9.x86_64 2026-03-06T13:39:25.369 INFO:teuthology.orchestra.run.vm01.stdout: openblas-openmp-0.3.29-1.el9.x86_64 2026-03-06T13:39:25.369 INFO:teuthology.orchestra.run.vm01.stdout: parquet-libs-9.0.0-15.el9.x86_64 2026-03-06T13:39:25.369 INFO:teuthology.orchestra.run.vm01.stdout: pciutils-3.7.0-7.el9.x86_64 2026-03-06T13:39:25.369 INFO:teuthology.orchestra.run.vm01.stdout: perl-Benchmark-1.23-483.el9.noarch 2026-03-06T13:39:25.369 INFO:teuthology.orchestra.run.vm01.stdout: perl-Test-Harness-1:3.42-461.el9.noarch 2026-03-06T13:39:25.369 INFO:teuthology.orchestra.run.vm01.stdout: protobuf-3.14.0-17.el9.x86_64 2026-03-06T13:39:25.369 INFO:teuthology.orchestra.run.vm01.stdout: protobuf-compiler-3.14.0-17.el9.x86_64 2026-03-06T13:39:25.369 INFO:teuthology.orchestra.run.vm01.stdout: python3-asyncssh-2.13.2-5.el9.noarch 2026-03-06T13:39:25.369 INFO:teuthology.orchestra.run.vm01.stdout: python3-autocommand-2.2.2-8.el9.noarch 2026-03-06T13:39:25.369 INFO:teuthology.orchestra.run.vm01.stdout: python3-babel-2.9.1-2.el9.noarch 2026-03-06T13:39:25.369 INFO:teuthology.orchestra.run.vm01.stdout: python3-backports-tarfile-1.2.0-1.el9.noarch 2026-03-06T13:39:25.369 INFO:teuthology.orchestra.run.vm01.stdout: python3-bcrypt-3.2.2-1.el9.x86_64 2026-03-06T13:39:25.369 INFO:teuthology.orchestra.run.vm01.stdout: python3-cachetools-4.2.4-1.el9.noarch 2026-03-06T13:39:25.369 INFO:teuthology.orchestra.run.vm01.stdout: python3-ceph-argparse-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 2026-03-06T13:39:25.369 INFO:teuthology.orchestra.run.vm01.stdout: python3-ceph-common-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 2026-03-06T13:39:25.369 INFO:teuthology.orchestra.run.vm01.stdout: python3-cephfs-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 2026-03-06T13:39:25.369 INFO:teuthology.orchestra.run.vm01.stdout: python3-certifi-2023.05.07-4.el9.noarch 2026-03-06T13:39:25.369 INFO:teuthology.orchestra.run.vm01.stdout: python3-cffi-1.14.5-5.el9.x86_64 2026-03-06T13:39:25.369 INFO:teuthology.orchestra.run.vm01.stdout: python3-cheroot-10.0.1-4.el9.noarch 2026-03-06T13:39:25.369 INFO:teuthology.orchestra.run.vm01.stdout: python3-cherrypy-18.6.1-2.el9.noarch 2026-03-06T13:39:25.369 INFO:teuthology.orchestra.run.vm01.stdout: python3-cryptography-36.0.1-5.el9.x86_64 2026-03-06T13:39:25.369 INFO:teuthology.orchestra.run.vm01.stdout: python3-devel-3.9.25-3.el9.x86_64 2026-03-06T13:39:25.369 INFO:teuthology.orchestra.run.vm01.stdout: python3-google-auth-1:2.45.0-1.el9.noarch 2026-03-06T13:39:25.369 INFO:teuthology.orchestra.run.vm01.stdout: python3-grpcio-1.46.7-10.el9.x86_64 2026-03-06T13:39:25.369 INFO:teuthology.orchestra.run.vm01.stdout: python3-grpcio-tools-1.46.7-10.el9.x86_64 2026-03-06T13:39:25.369 INFO:teuthology.orchestra.run.vm01.stdout: python3-influxdb-5.3.1-1.el9.noarch 2026-03-06T13:39:25.369 INFO:teuthology.orchestra.run.vm01.stdout: python3-isodate-0.6.1-3.el9.noarch 2026-03-06T13:39:25.369 INFO:teuthology.orchestra.run.vm01.stdout: python3-jaraco-8.2.1-3.el9.noarch 2026-03-06T13:39:25.369 INFO:teuthology.orchestra.run.vm01.stdout: python3-jaraco-classes-3.2.1-5.el9.noarch 2026-03-06T13:39:25.369 INFO:teuthology.orchestra.run.vm01.stdout: python3-jaraco-collections-3.0.0-8.el9.noarch 2026-03-06T13:39:25.369 INFO:teuthology.orchestra.run.vm01.stdout: python3-jaraco-context-6.0.1-3.el9.noarch 2026-03-06T13:39:25.369 INFO:teuthology.orchestra.run.vm01.stdout: python3-jaraco-functools-3.5.0-2.el9.noarch 2026-03-06T13:39:25.369 INFO:teuthology.orchestra.run.vm01.stdout: python3-jaraco-text-4.0.0-2.el9.noarch 2026-03-06T13:39:25.369 INFO:teuthology.orchestra.run.vm01.stdout: python3-jinja2-2.11.3-8.el9.noarch 2026-03-06T13:39:25.369 INFO:teuthology.orchestra.run.vm01.stdout: python3-kubernetes-1:26.1.0-3.el9.noarch 2026-03-06T13:39:25.369 INFO:teuthology.orchestra.run.vm01.stdout: python3-libstoragemgmt-1.10.1-1.el9.x86_64 2026-03-06T13:39:25.370 INFO:teuthology.orchestra.run.vm01.stdout: python3-logutils-0.3.5-21.el9.noarch 2026-03-06T13:39:25.370 INFO:teuthology.orchestra.run.vm01.stdout: python3-lxml-4.6.5-3.el9.x86_64 2026-03-06T13:39:25.370 INFO:teuthology.orchestra.run.vm01.stdout: python3-mako-1.1.4-6.el9.noarch 2026-03-06T13:39:25.370 INFO:teuthology.orchestra.run.vm01.stdout: python3-markupsafe-1.1.1-12.el9.x86_64 2026-03-06T13:39:25.370 INFO:teuthology.orchestra.run.vm01.stdout: python3-more-itertools-8.12.0-2.el9.noarch 2026-03-06T13:39:25.370 INFO:teuthology.orchestra.run.vm01.stdout: python3-msgpack-1.0.3-2.el9.x86_64 2026-03-06T13:39:25.370 INFO:teuthology.orchestra.run.vm01.stdout: python3-natsort-7.1.1-5.el9.noarch 2026-03-06T13:39:25.370 INFO:teuthology.orchestra.run.vm01.stdout: python3-numpy-1:1.23.5-2.el9.x86_64 2026-03-06T13:39:25.370 INFO:teuthology.orchestra.run.vm01.stdout: python3-numpy-f2py-1:1.23.5-2.el9.x86_64 2026-03-06T13:39:25.370 INFO:teuthology.orchestra.run.vm01.stdout: python3-packaging-20.9-5.el9.noarch 2026-03-06T13:39:25.370 INFO:teuthology.orchestra.run.vm01.stdout: python3-pecan-1.4.2-3.el9.noarch 2026-03-06T13:39:25.370 INFO:teuthology.orchestra.run.vm01.stdout: python3-ply-3.11-14.el9.noarch 2026-03-06T13:39:25.370 INFO:teuthology.orchestra.run.vm01.stdout: python3-portend-3.1.0-2.el9.noarch 2026-03-06T13:39:25.370 INFO:teuthology.orchestra.run.vm01.stdout: python3-protobuf-3.14.0-17.el9.noarch 2026-03-06T13:39:25.370 INFO:teuthology.orchestra.run.vm01.stdout: python3-pyOpenSSL-21.0.0-1.el9.noarch 2026-03-06T13:39:25.370 INFO:teuthology.orchestra.run.vm01.stdout: python3-pyasn1-0.4.8-7.el9.noarch 2026-03-06T13:39:25.370 INFO:teuthology.orchestra.run.vm01.stdout: python3-pyasn1-modules-0.4.8-7.el9.noarch 2026-03-06T13:39:25.370 INFO:teuthology.orchestra.run.vm01.stdout: python3-pycparser-2.20-6.el9.noarch 2026-03-06T13:39:25.370 INFO:teuthology.orchestra.run.vm01.stdout: python3-rados-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 2026-03-06T13:39:25.370 INFO:teuthology.orchestra.run.vm01.stdout: python3-rbd-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 2026-03-06T13:39:25.370 INFO:teuthology.orchestra.run.vm01.stdout: python3-repoze-lru-0.7-16.el9.noarch 2026-03-06T13:39:25.370 INFO:teuthology.orchestra.run.vm01.stdout: python3-requests-2.25.1-10.el9.noarch 2026-03-06T13:39:25.370 INFO:teuthology.orchestra.run.vm01.stdout: python3-requests-oauthlib-1.3.0-12.el9.noarch 2026-03-06T13:39:25.370 INFO:teuthology.orchestra.run.vm01.stdout: python3-rgw-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 2026-03-06T13:39:25.370 INFO:teuthology.orchestra.run.vm01.stdout: python3-routes-2.5.1-5.el9.noarch 2026-03-06T13:39:25.370 INFO:teuthology.orchestra.run.vm01.stdout: python3-rsa-4.9-2.el9.noarch 2026-03-06T13:39:25.370 INFO:teuthology.orchestra.run.vm01.stdout: python3-saml-1.16.0-1.el9.noarch 2026-03-06T13:39:25.370 INFO:teuthology.orchestra.run.vm01.stdout: python3-scipy-1.9.3-2.el9.x86_64 2026-03-06T13:39:25.370 INFO:teuthology.orchestra.run.vm01.stdout: python3-tempora-5.0.0-2.el9.noarch 2026-03-06T13:39:25.370 INFO:teuthology.orchestra.run.vm01.stdout: python3-toml-0.10.2-6.el9.noarch 2026-03-06T13:39:25.370 INFO:teuthology.orchestra.run.vm01.stdout: python3-typing-extensions-4.15.0-1.el9.noarch 2026-03-06T13:39:25.370 INFO:teuthology.orchestra.run.vm01.stdout: python3-urllib3-1.26.5-7.el9.noarch 2026-03-06T13:39:25.370 INFO:teuthology.orchestra.run.vm01.stdout: python3-webob-1.8.8-2.el9.noarch 2026-03-06T13:39:25.370 INFO:teuthology.orchestra.run.vm01.stdout: python3-websocket-client-1.2.3-2.el9.noarch 2026-03-06T13:39:25.370 INFO:teuthology.orchestra.run.vm01.stdout: python3-werkzeug-2.0.3-3.el9.1.noarch 2026-03-06T13:39:25.370 INFO:teuthology.orchestra.run.vm01.stdout: python3-xmlsec-1.3.13-1.el9.x86_64 2026-03-06T13:39:25.370 INFO:teuthology.orchestra.run.vm01.stdout: python3-xmltodict-0.12.0-15.el9.noarch 2026-03-06T13:39:25.370 INFO:teuthology.orchestra.run.vm01.stdout: python3-zc-lockfile-2.0-10.el9.noarch 2026-03-06T13:39:25.370 INFO:teuthology.orchestra.run.vm01.stdout: qatlib-25.08.0-2.el9.x86_64 2026-03-06T13:39:25.370 INFO:teuthology.orchestra.run.vm01.stdout: qatlib-service-25.08.0-2.el9.x86_64 2026-03-06T13:39:25.370 INFO:teuthology.orchestra.run.vm01.stdout: qatzip-libs-1.3.1-1.el9.x86_64 2026-03-06T13:39:25.370 INFO:teuthology.orchestra.run.vm01.stdout: rbd-fuse-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 2026-03-06T13:39:25.370 INFO:teuthology.orchestra.run.vm01.stdout: rbd-mirror-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 2026-03-06T13:39:25.370 INFO:teuthology.orchestra.run.vm01.stdout: rbd-nbd-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 2026-03-06T13:39:25.370 INFO:teuthology.orchestra.run.vm01.stdout: re2-1:20211101-20.el9.x86_64 2026-03-06T13:39:25.370 INFO:teuthology.orchestra.run.vm01.stdout: s3cmd-2.4.0-1.el9.noarch 2026-03-06T13:39:25.370 INFO:teuthology.orchestra.run.vm01.stdout: smartmontools-1:7.2-10.el9.x86_64 2026-03-06T13:39:25.370 INFO:teuthology.orchestra.run.vm01.stdout: socat-1.7.4.1-8.el9.x86_64 2026-03-06T13:39:25.370 INFO:teuthology.orchestra.run.vm01.stdout: thrift-0.15.0-4.el9.x86_64 2026-03-06T13:39:25.370 INFO:teuthology.orchestra.run.vm01.stdout: unzip-6.0-59.el9.x86_64 2026-03-06T13:39:25.370 INFO:teuthology.orchestra.run.vm01.stdout: xmlsec1-1.2.29-13.el9.x86_64 2026-03-06T13:39:25.370 INFO:teuthology.orchestra.run.vm01.stdout: xmlsec1-openssl-1.2.29-13.el9.x86_64 2026-03-06T13:39:25.370 INFO:teuthology.orchestra.run.vm01.stdout: xmlstarlet-1.6.1-20.el9.x86_64 2026-03-06T13:39:25.370 INFO:teuthology.orchestra.run.vm01.stdout: zip-3.0-35.el9.x86_64 2026-03-06T13:39:25.370 INFO:teuthology.orchestra.run.vm01.stdout: 2026-03-06T13:39:25.370 INFO:teuthology.orchestra.run.vm01.stdout:Complete! 2026-03-06T13:39:25.500 DEBUG:teuthology.parallel:result is None 2026-03-06T13:39:25.500 INFO:teuthology.task.install:Skipping version verification because we have custom repos... 2026-03-06T13:39:25.500 INFO:teuthology.task.install:Skipping version verification because we have custom repos... 2026-03-06T13:39:25.500 INFO:teuthology.task.install:Skipping version verification because we have custom repos... 2026-03-06T13:39:25.500 INFO:teuthology.task.install.util:Shipping valgrind.supp... 2026-03-06T13:39:25.501 DEBUG:teuthology.orchestra.run.vm00:> set -ex 2026-03-06T13:39:25.501 DEBUG:teuthology.orchestra.run.vm00:> sudo dd of=/home/ubuntu/cephtest/valgrind.supp 2026-03-06T13:39:25.529 DEBUG:teuthology.orchestra.run.vm01:> set -ex 2026-03-06T13:39:25.529 DEBUG:teuthology.orchestra.run.vm01:> sudo dd of=/home/ubuntu/cephtest/valgrind.supp 2026-03-06T13:39:25.570 DEBUG:teuthology.orchestra.run.vm08:> set -ex 2026-03-06T13:39:25.570 DEBUG:teuthology.orchestra.run.vm08:> sudo dd of=/home/ubuntu/cephtest/valgrind.supp 2026-03-06T13:39:25.602 INFO:teuthology.task.install.util:Shipping 'daemon-helper'... 2026-03-06T13:39:25.602 DEBUG:teuthology.orchestra.run.vm00:> set -ex 2026-03-06T13:39:25.603 DEBUG:teuthology.orchestra.run.vm00:> sudo dd of=/usr/bin/daemon-helper 2026-03-06T13:39:25.628 DEBUG:teuthology.orchestra.run.vm00:> sudo chmod a=rx -- /usr/bin/daemon-helper 2026-03-06T13:39:25.691 DEBUG:teuthology.orchestra.run.vm01:> set -ex 2026-03-06T13:39:25.691 DEBUG:teuthology.orchestra.run.vm01:> sudo dd of=/usr/bin/daemon-helper 2026-03-06T13:39:25.722 DEBUG:teuthology.orchestra.run.vm01:> sudo chmod a=rx -- /usr/bin/daemon-helper 2026-03-06T13:39:25.795 DEBUG:teuthology.orchestra.run.vm08:> set -ex 2026-03-06T13:39:25.795 DEBUG:teuthology.orchestra.run.vm08:> sudo dd of=/usr/bin/daemon-helper 2026-03-06T13:39:25.822 DEBUG:teuthology.orchestra.run.vm08:> sudo chmod a=rx -- /usr/bin/daemon-helper 2026-03-06T13:39:25.888 INFO:teuthology.task.install.util:Shipping 'adjust-ulimits'... 2026-03-06T13:39:25.888 DEBUG:teuthology.orchestra.run.vm00:> set -ex 2026-03-06T13:39:25.888 DEBUG:teuthology.orchestra.run.vm00:> sudo dd of=/usr/bin/adjust-ulimits 2026-03-06T13:39:25.915 DEBUG:teuthology.orchestra.run.vm00:> sudo chmod a=rx -- /usr/bin/adjust-ulimits 2026-03-06T13:39:25.979 DEBUG:teuthology.orchestra.run.vm01:> set -ex 2026-03-06T13:39:25.979 DEBUG:teuthology.orchestra.run.vm01:> sudo dd of=/usr/bin/adjust-ulimits 2026-03-06T13:39:26.009 DEBUG:teuthology.orchestra.run.vm01:> sudo chmod a=rx -- /usr/bin/adjust-ulimits 2026-03-06T13:39:26.079 DEBUG:teuthology.orchestra.run.vm08:> set -ex 2026-03-06T13:39:26.079 DEBUG:teuthology.orchestra.run.vm08:> sudo dd of=/usr/bin/adjust-ulimits 2026-03-06T13:39:26.103 DEBUG:teuthology.orchestra.run.vm08:> sudo chmod a=rx -- /usr/bin/adjust-ulimits 2026-03-06T13:39:26.166 INFO:teuthology.task.install.util:Shipping 'stdin-killer'... 2026-03-06T13:39:26.166 DEBUG:teuthology.orchestra.run.vm00:> set -ex 2026-03-06T13:39:26.166 DEBUG:teuthology.orchestra.run.vm00:> sudo dd of=/usr/bin/stdin-killer 2026-03-06T13:39:26.193 DEBUG:teuthology.orchestra.run.vm00:> sudo chmod a=rx -- /usr/bin/stdin-killer 2026-03-06T13:39:26.258 DEBUG:teuthology.orchestra.run.vm01:> set -ex 2026-03-06T13:39:26.258 DEBUG:teuthology.orchestra.run.vm01:> sudo dd of=/usr/bin/stdin-killer 2026-03-06T13:39:26.300 DEBUG:teuthology.orchestra.run.vm01:> sudo chmod a=rx -- /usr/bin/stdin-killer 2026-03-06T13:39:26.377 DEBUG:teuthology.orchestra.run.vm08:> set -ex 2026-03-06T13:39:26.377 DEBUG:teuthology.orchestra.run.vm08:> sudo dd of=/usr/bin/stdin-killer 2026-03-06T13:39:26.400 DEBUG:teuthology.orchestra.run.vm08:> sudo chmod a=rx -- /usr/bin/stdin-killer 2026-03-06T13:39:26.463 INFO:teuthology.run_tasks:Running task cephadm... 2026-03-06T13:39:26.507 INFO:tasks.cephadm:Config: {'conf': {'global': {'mon election default strategy': 1}, 'mgr': {'debug mgr': 20, 'debug ms': 1, 'mgr/cephadm/use_agent': False}, 'mon': {'debug mon': 20, 'debug ms': 1, 'debug paxos': 20}, 'osd': {'debug ms': 1, 'debug osd': 20, 'osd mclock iops capacity threshold hdd': 49000}}, 'flavor': 'default', 'log-ignorelist': ['\\(MDS_ALL_DOWN\\)', '\\(MDS_UP_LESS_THAN_MAX\\)', 'MON_DOWN', 'mons down', 'mon down', 'out of quorum', 'CEPHADM_STRAY_DAEMON'], 'log-only-match': ['CEPHADM_'], 'sha1': 'c24117fd5525679b799527bc1bd1f1dd0a2db5e2', 'cephadm_binary_url': 'https://download.ceph.com/rpm-19.2.3/el9/noarch/cephadm', 'containers': {'image': 'harbor.clyso.com/custom-ceph/ceph/ceph:cobaltcore-storage-v19.2.3-fasttrack-3'}} 2026-03-06T13:39:26.507 INFO:tasks.cephadm:Provided image contains tag or digest, using it as is 2026-03-06T13:39:26.507 INFO:tasks.cephadm:Cluster image is harbor.clyso.com/custom-ceph/ceph/ceph:cobaltcore-storage-v19.2.3-fasttrack-3 2026-03-06T13:39:26.507 INFO:tasks.cephadm:Cluster fsid is 82f504c8-1959-11f1-9a16-a78d135e200b 2026-03-06T13:39:26.507 INFO:tasks.cephadm:Choosing monitor IPs and ports... 2026-03-06T13:39:26.507 INFO:tasks.cephadm:Monitor IPs: {'mon.a': '192.168.123.100', 'mon.b': '192.168.123.101', 'mon.c': '192.168.123.108'} 2026-03-06T13:39:26.507 INFO:tasks.cephadm:First mon is mon.a on vm00 2026-03-06T13:39:26.507 INFO:tasks.cephadm:First mgr is a 2026-03-06T13:39:26.507 INFO:tasks.cephadm:Normalizing hostnames... 2026-03-06T13:39:26.507 DEBUG:teuthology.orchestra.run.vm00:> sudo hostname $(hostname -s) 2026-03-06T13:39:26.532 DEBUG:teuthology.orchestra.run.vm01:> sudo hostname $(hostname -s) 2026-03-06T13:39:26.556 DEBUG:teuthology.orchestra.run.vm08:> sudo hostname $(hostname -s) 2026-03-06T13:39:26.580 INFO:tasks.cephadm:Downloading cephadm from url: https://download.ceph.com/rpm-19.2.3/el9/noarch/cephadm 2026-03-06T13:39:26.580 DEBUG:teuthology.orchestra.run.vm00:> curl --silent -L https://download.ceph.com/rpm-19.2.3/el9/noarch/cephadm > /home/ubuntu/cephtest/cephadm && ls -l /home/ubuntu/cephtest/cephadm 2026-03-06T13:39:27.776 INFO:teuthology.orchestra.run.vm00.stdout:-rw-r--r--. 1 ubuntu ubuntu 787672 Mar 6 13:39 /home/ubuntu/cephtest/cephadm 2026-03-06T13:39:27.776 DEBUG:teuthology.orchestra.run.vm01:> curl --silent -L https://download.ceph.com/rpm-19.2.3/el9/noarch/cephadm > /home/ubuntu/cephtest/cephadm && ls -l /home/ubuntu/cephtest/cephadm 2026-03-06T13:39:28.877 INFO:teuthology.orchestra.run.vm01.stdout:-rw-r--r--. 1 ubuntu ubuntu 787672 Mar 6 13:39 /home/ubuntu/cephtest/cephadm 2026-03-06T13:39:28.877 DEBUG:teuthology.orchestra.run.vm08:> curl --silent -L https://download.ceph.com/rpm-19.2.3/el9/noarch/cephadm > /home/ubuntu/cephtest/cephadm && ls -l /home/ubuntu/cephtest/cephadm 2026-03-06T13:39:29.912 INFO:teuthology.orchestra.run.vm08.stdout:-rw-r--r--. 1 ubuntu ubuntu 787672 Mar 6 13:39 /home/ubuntu/cephtest/cephadm 2026-03-06T13:39:29.913 DEBUG:teuthology.orchestra.run.vm00:> test -s /home/ubuntu/cephtest/cephadm && test $(stat -c%s /home/ubuntu/cephtest/cephadm) -gt 1000 && chmod +x /home/ubuntu/cephtest/cephadm 2026-03-06T13:39:29.928 DEBUG:teuthology.orchestra.run.vm01:> test -s /home/ubuntu/cephtest/cephadm && test $(stat -c%s /home/ubuntu/cephtest/cephadm) -gt 1000 && chmod +x /home/ubuntu/cephtest/cephadm 2026-03-06T13:39:29.949 DEBUG:teuthology.orchestra.run.vm08:> test -s /home/ubuntu/cephtest/cephadm && test $(stat -c%s /home/ubuntu/cephtest/cephadm) -gt 1000 && chmod +x /home/ubuntu/cephtest/cephadm 2026-03-06T13:39:29.969 INFO:tasks.cephadm:Pulling image harbor.clyso.com/custom-ceph/ceph/ceph:cobaltcore-storage-v19.2.3-fasttrack-3 on all hosts... 2026-03-06T13:39:29.969 DEBUG:teuthology.orchestra.run.vm00:> sudo /home/ubuntu/cephtest/cephadm --image harbor.clyso.com/custom-ceph/ceph/ceph:cobaltcore-storage-v19.2.3-fasttrack-3 pull 2026-03-06T13:39:29.971 DEBUG:teuthology.orchestra.run.vm01:> sudo /home/ubuntu/cephtest/cephadm --image harbor.clyso.com/custom-ceph/ceph/ceph:cobaltcore-storage-v19.2.3-fasttrack-3 pull 2026-03-06T13:39:29.993 DEBUG:teuthology.orchestra.run.vm08:> sudo /home/ubuntu/cephtest/cephadm --image harbor.clyso.com/custom-ceph/ceph/ceph:cobaltcore-storage-v19.2.3-fasttrack-3 pull 2026-03-06T13:39:30.284 INFO:teuthology.orchestra.run.vm00.stderr:Pulling container image harbor.clyso.com/custom-ceph/ceph/ceph:cobaltcore-storage-v19.2.3-fasttrack-3... 2026-03-06T13:39:30.340 INFO:teuthology.orchestra.run.vm08.stderr:Pulling container image harbor.clyso.com/custom-ceph/ceph/ceph:cobaltcore-storage-v19.2.3-fasttrack-3... 2026-03-06T13:39:30.373 INFO:teuthology.orchestra.run.vm01.stderr:Pulling container image harbor.clyso.com/custom-ceph/ceph/ceph:cobaltcore-storage-v19.2.3-fasttrack-3... 2026-03-06T13:39:52.488 INFO:teuthology.orchestra.run.vm08.stdout:{ 2026-03-06T13:39:52.488 INFO:teuthology.orchestra.run.vm08.stdout: "ceph_version": "ceph version 19.2.3-47-gc24117fd552 (c24117fd5525679b799527bc1bd1f1dd0a2db5e2) squid (stable)", 2026-03-06T13:39:52.488 INFO:teuthology.orchestra.run.vm08.stdout: "image_id": "306e97de47e91c2b4b24d3dc09be3b3a12039b078f343d91220102acc6628a68", 2026-03-06T13:39:52.488 INFO:teuthology.orchestra.run.vm08.stdout: "repo_digests": [ 2026-03-06T13:39:52.488 INFO:teuthology.orchestra.run.vm08.stdout: "harbor.clyso.com/custom-ceph/ceph/ceph@sha256:26363c7a4eea9ef5a0148afc7b2a22b6f486596d87a30c2a9fdcda5db3eca62b" 2026-03-06T13:39:52.488 INFO:teuthology.orchestra.run.vm08.stdout: ] 2026-03-06T13:39:52.488 INFO:teuthology.orchestra.run.vm08.stdout:} 2026-03-06T13:39:57.008 INFO:teuthology.orchestra.run.vm00.stdout:{ 2026-03-06T13:39:57.009 INFO:teuthology.orchestra.run.vm00.stdout: "ceph_version": "ceph version 19.2.3-47-gc24117fd552 (c24117fd5525679b799527bc1bd1f1dd0a2db5e2) squid (stable)", 2026-03-06T13:39:57.009 INFO:teuthology.orchestra.run.vm00.stdout: "image_id": "306e97de47e91c2b4b24d3dc09be3b3a12039b078f343d91220102acc6628a68", 2026-03-06T13:39:57.009 INFO:teuthology.orchestra.run.vm00.stdout: "repo_digests": [ 2026-03-06T13:39:57.009 INFO:teuthology.orchestra.run.vm00.stdout: "harbor.clyso.com/custom-ceph/ceph/ceph@sha256:26363c7a4eea9ef5a0148afc7b2a22b6f486596d87a30c2a9fdcda5db3eca62b" 2026-03-06T13:39:57.009 INFO:teuthology.orchestra.run.vm00.stdout: ] 2026-03-06T13:39:57.009 INFO:teuthology.orchestra.run.vm00.stdout:} 2026-03-06T13:40:00.198 INFO:teuthology.orchestra.run.vm01.stdout:{ 2026-03-06T13:40:00.198 INFO:teuthology.orchestra.run.vm01.stdout: "ceph_version": "ceph version 19.2.3-47-gc24117fd552 (c24117fd5525679b799527bc1bd1f1dd0a2db5e2) squid (stable)", 2026-03-06T13:40:00.198 INFO:teuthology.orchestra.run.vm01.stdout: "image_id": "306e97de47e91c2b4b24d3dc09be3b3a12039b078f343d91220102acc6628a68", 2026-03-06T13:40:00.198 INFO:teuthology.orchestra.run.vm01.stdout: "repo_digests": [ 2026-03-06T13:40:00.198 INFO:teuthology.orchestra.run.vm01.stdout: "harbor.clyso.com/custom-ceph/ceph/ceph@sha256:26363c7a4eea9ef5a0148afc7b2a22b6f486596d87a30c2a9fdcda5db3eca62b" 2026-03-06T13:40:00.198 INFO:teuthology.orchestra.run.vm01.stdout: ] 2026-03-06T13:40:00.198 INFO:teuthology.orchestra.run.vm01.stdout:} 2026-03-06T13:40:00.213 DEBUG:teuthology.orchestra.run.vm00:> sudo mkdir -p /etc/ceph 2026-03-06T13:40:00.239 DEBUG:teuthology.orchestra.run.vm01:> sudo mkdir -p /etc/ceph 2026-03-06T13:40:00.263 DEBUG:teuthology.orchestra.run.vm08:> sudo mkdir -p /etc/ceph 2026-03-06T13:40:00.287 DEBUG:teuthology.orchestra.run.vm00:> sudo chmod 777 /etc/ceph 2026-03-06T13:40:00.310 DEBUG:teuthology.orchestra.run.vm01:> sudo chmod 777 /etc/ceph 2026-03-06T13:40:00.332 DEBUG:teuthology.orchestra.run.vm08:> sudo chmod 777 /etc/ceph 2026-03-06T13:40:00.354 INFO:tasks.cephadm:Writing seed config... 2026-03-06T13:40:00.354 INFO:tasks.cephadm: override: [global] mon election default strategy = 1 2026-03-06T13:40:00.354 INFO:tasks.cephadm: override: [mgr] debug mgr = 20 2026-03-06T13:40:00.354 INFO:tasks.cephadm: override: [mgr] debug ms = 1 2026-03-06T13:40:00.354 INFO:tasks.cephadm: override: [mgr] mgr/cephadm/use_agent = False 2026-03-06T13:40:00.354 INFO:tasks.cephadm: override: [mon] debug mon = 20 2026-03-06T13:40:00.354 INFO:tasks.cephadm: override: [mon] debug ms = 1 2026-03-06T13:40:00.354 INFO:tasks.cephadm: override: [mon] debug paxos = 20 2026-03-06T13:40:00.354 INFO:tasks.cephadm: override: [osd] debug ms = 1 2026-03-06T13:40:00.354 INFO:tasks.cephadm: override: [osd] debug osd = 20 2026-03-06T13:40:00.354 INFO:tasks.cephadm: override: [osd] osd mclock iops capacity threshold hdd = 49000 2026-03-06T13:40:00.355 DEBUG:teuthology.orchestra.run.vm00:> set -ex 2026-03-06T13:40:00.355 DEBUG:teuthology.orchestra.run.vm00:> dd of=/home/ubuntu/cephtest/seed.ceph.conf 2026-03-06T13:40:00.369 DEBUG:tasks.cephadm:Final config: [global] # make logging friendly to teuthology log_to_file = true log_to_stderr = false log to journald = false mon cluster log to file = true mon cluster log file level = debug mon clock drift allowed = 1.000 # replicate across OSDs, not hosts osd crush chooseleaf type = 0 #osd pool default size = 2 osd pool default erasure code profile = plugin=jerasure technique=reed_sol_van k=2 m=1 crush-failure-domain=osd # enable some debugging auth debug = true ms die on old message = true ms die on bug = true debug asserts on shutdown = true # adjust warnings mon max pg per osd = 10000# >= luminous mon pg warn max object skew = 0 mon osd allow primary affinity = true mon osd allow pg remap = true mon warn on legacy crush tunables = false mon warn on crush straw calc version zero = false mon warn on no sortbitwise = false mon warn on osd down out interval zero = false mon warn on too few osds = false mon_warn_on_pool_pg_num_not_power_of_two = false # disable pg_autoscaler by default for new pools osd_pool_default_pg_autoscale_mode = off # tests delete pools mon allow pool delete = true fsid = 82f504c8-1959-11f1-9a16-a78d135e200b mon election default strategy = 1 [osd] osd scrub load threshold = 5.0 osd scrub max interval = 600 osd mclock profile = high_recovery_ops osd recover clone overlap = true osd recovery max chunk = 1048576 osd deep scrub update digest min age = 30 osd map max advance = 10 osd memory target autotune = true # debugging osd debug shutdown = true osd debug op order = true osd debug verify stray on activate = true osd debug pg log writeout = true osd debug verify cached snaps = true osd debug verify missing on start = true osd debug misdirected ops = true osd op queue = debug_random osd op queue cut off = debug_random osd shutdown pgref assert = true bdev debug aio = true osd sloppy crc = true debug ms = 1 debug osd = 20 osd mclock iops capacity threshold hdd = 49000 [mgr] mon reweight min pgs per osd = 4 mon reweight min bytes per osd = 10 mgr/telemetry/nag = false debug mgr = 20 debug ms = 1 mgr/cephadm/use_agent = False [mon] mon data avail warn = 5 mon mgr mkfs grace = 240 mon reweight min pgs per osd = 4 mon osd reporter subtree level = osd mon osd prime pg temp = true mon reweight min bytes per osd = 10 # rotate auth tickets quickly to exercise renewal paths auth mon ticket ttl = 660# 11m auth service ticket ttl = 240# 4m # don't complain about global id reclaim mon_warn_on_insecure_global_id_reclaim = false mon_warn_on_insecure_global_id_reclaim_allowed = false debug mon = 20 debug ms = 1 debug paxos = 20 [client.rgw] rgw cache enabled = true rgw enable ops log = true rgw enable usage log = true 2026-03-06T13:40:00.369 DEBUG:teuthology.orchestra.run.vm00:mon.a> sudo journalctl -f -n 0 -u ceph-82f504c8-1959-11f1-9a16-a78d135e200b@mon.a.service 2026-03-06T13:40:00.411 DEBUG:teuthology.orchestra.run.vm00:mgr.a> sudo journalctl -f -n 0 -u ceph-82f504c8-1959-11f1-9a16-a78d135e200b@mgr.a.service 2026-03-06T13:40:00.452 INFO:tasks.cephadm:Bootstrapping... 2026-03-06T13:40:00.453 DEBUG:teuthology.orchestra.run.vm00:> sudo /home/ubuntu/cephtest/cephadm --image harbor.clyso.com/custom-ceph/ceph/ceph:cobaltcore-storage-v19.2.3-fasttrack-3 -v bootstrap --fsid 82f504c8-1959-11f1-9a16-a78d135e200b --config /home/ubuntu/cephtest/seed.ceph.conf --output-config /etc/ceph/ceph.conf --output-keyring /etc/ceph/ceph.client.admin.keyring --output-pub-ssh-key /home/ubuntu/cephtest/ceph.pub --mon-id a --mgr-id a --orphan-initial-daemons --skip-monitoring-stack --mon-ip 192.168.123.100 --skip-admin-label && sudo chmod +r /etc/ceph/ceph.client.admin.keyring 2026-03-06T13:40:00.729 INFO:teuthology.orchestra.run.vm00.stdout:-------------------------------------------------------------------------------- 2026-03-06T13:40:00.729 INFO:teuthology.orchestra.run.vm00.stdout:cephadm ['--image', 'harbor.clyso.com/custom-ceph/ceph/ceph:cobaltcore-storage-v19.2.3-fasttrack-3', '-v', 'bootstrap', '--fsid', '82f504c8-1959-11f1-9a16-a78d135e200b', '--config', '/home/ubuntu/cephtest/seed.ceph.conf', '--output-config', '/etc/ceph/ceph.conf', '--output-keyring', '/etc/ceph/ceph.client.admin.keyring', '--output-pub-ssh-key', '/home/ubuntu/cephtest/ceph.pub', '--mon-id', 'a', '--mgr-id', 'a', '--orphan-initial-daemons', '--skip-monitoring-stack', '--mon-ip', '192.168.123.100', '--skip-admin-label'] 2026-03-06T13:40:00.729 INFO:teuthology.orchestra.run.vm00.stderr:Specifying an fsid for your cluster offers no advantages and may increase the likelihood of fsid conflicts. 2026-03-06T13:40:00.729 INFO:teuthology.orchestra.run.vm00.stdout:Verifying podman|docker is present... 2026-03-06T13:40:00.748 INFO:teuthology.orchestra.run.vm00.stdout:/bin/podman: stdout 5.8.0 2026-03-06T13:40:00.749 INFO:teuthology.orchestra.run.vm00.stdout:Verifying lvm2 is present... 2026-03-06T13:40:00.749 INFO:teuthology.orchestra.run.vm00.stdout:Verifying time synchronization is in place... 2026-03-06T13:40:00.755 INFO:teuthology.orchestra.run.vm00.stdout:Non-zero exit code 1 from systemctl is-enabled chrony.service 2026-03-06T13:40:00.755 INFO:teuthology.orchestra.run.vm00.stdout:systemctl: stderr Failed to get unit file state for chrony.service: No such file or directory 2026-03-06T13:40:00.760 INFO:teuthology.orchestra.run.vm00.stdout:Non-zero exit code 3 from systemctl is-active chrony.service 2026-03-06T13:40:00.760 INFO:teuthology.orchestra.run.vm00.stdout:systemctl: stdout inactive 2026-03-06T13:40:00.765 INFO:teuthology.orchestra.run.vm00.stdout:systemctl: stdout enabled 2026-03-06T13:40:00.770 INFO:teuthology.orchestra.run.vm00.stdout:systemctl: stdout active 2026-03-06T13:40:00.770 INFO:teuthology.orchestra.run.vm00.stdout:Unit chronyd.service is enabled and running 2026-03-06T13:40:00.770 INFO:teuthology.orchestra.run.vm00.stdout:Repeating the final host check... 2026-03-06T13:40:00.789 INFO:teuthology.orchestra.run.vm00.stdout:/bin/podman: stdout 5.8.0 2026-03-06T13:40:00.789 INFO:teuthology.orchestra.run.vm00.stdout:podman (/bin/podman) version 5.8.0 is present 2026-03-06T13:40:00.789 INFO:teuthology.orchestra.run.vm00.stdout:systemctl is present 2026-03-06T13:40:00.789 INFO:teuthology.orchestra.run.vm00.stdout:lvcreate is present 2026-03-06T13:40:00.795 INFO:teuthology.orchestra.run.vm00.stdout:Non-zero exit code 1 from systemctl is-enabled chrony.service 2026-03-06T13:40:00.795 INFO:teuthology.orchestra.run.vm00.stdout:systemctl: stderr Failed to get unit file state for chrony.service: No such file or directory 2026-03-06T13:40:00.799 INFO:teuthology.orchestra.run.vm00.stdout:Non-zero exit code 3 from systemctl is-active chrony.service 2026-03-06T13:40:00.799 INFO:teuthology.orchestra.run.vm00.stdout:systemctl: stdout inactive 2026-03-06T13:40:00.804 INFO:teuthology.orchestra.run.vm00.stdout:systemctl: stdout enabled 2026-03-06T13:40:00.809 INFO:teuthology.orchestra.run.vm00.stdout:systemctl: stdout active 2026-03-06T13:40:00.809 INFO:teuthology.orchestra.run.vm00.stdout:Unit chronyd.service is enabled and running 2026-03-06T13:40:00.809 INFO:teuthology.orchestra.run.vm00.stdout:Host looks OK 2026-03-06T13:40:00.809 INFO:teuthology.orchestra.run.vm00.stdout:Cluster fsid: 82f504c8-1959-11f1-9a16-a78d135e200b 2026-03-06T13:40:00.809 INFO:teuthology.orchestra.run.vm00.stdout:Acquiring lock 140573065669504 on /run/cephadm/82f504c8-1959-11f1-9a16-a78d135e200b.lock 2026-03-06T13:40:00.809 INFO:teuthology.orchestra.run.vm00.stdout:Lock 140573065669504 acquired on /run/cephadm/82f504c8-1959-11f1-9a16-a78d135e200b.lock 2026-03-06T13:40:00.809 INFO:teuthology.orchestra.run.vm00.stdout:Verifying IP 192.168.123.100 port 3300 ... 2026-03-06T13:40:00.810 INFO:teuthology.orchestra.run.vm00.stdout:Verifying IP 192.168.123.100 port 6789 ... 2026-03-06T13:40:00.810 INFO:teuthology.orchestra.run.vm00.stdout:Base mon IP(s) is [192.168.123.100:3300, 192.168.123.100:6789], mon addrv is [v2:192.168.123.100:3300,v1:192.168.123.100:6789] 2026-03-06T13:40:00.812 INFO:teuthology.orchestra.run.vm00.stdout:/sbin/ip: stdout default via 192.168.123.1 dev eth0 proto dhcp src 192.168.123.100 metric 100 2026-03-06T13:40:00.812 INFO:teuthology.orchestra.run.vm00.stdout:/sbin/ip: stdout 192.168.123.0/24 dev eth0 proto kernel scope link src 192.168.123.100 metric 100 2026-03-06T13:40:00.814 INFO:teuthology.orchestra.run.vm00.stdout:/sbin/ip: stdout ::1 dev lo proto kernel metric 256 pref medium 2026-03-06T13:40:00.814 INFO:teuthology.orchestra.run.vm00.stdout:/sbin/ip: stdout fe80::/64 dev eth0 proto kernel metric 1024 pref medium 2026-03-06T13:40:00.816 INFO:teuthology.orchestra.run.vm00.stdout:/sbin/ip: stdout 1: lo: mtu 65536 state UNKNOWN qlen 1000 2026-03-06T13:40:00.816 INFO:teuthology.orchestra.run.vm00.stdout:/sbin/ip: stdout inet6 ::1/128 scope host 2026-03-06T13:40:00.816 INFO:teuthology.orchestra.run.vm00.stdout:/sbin/ip: stdout valid_lft forever preferred_lft forever 2026-03-06T13:40:00.816 INFO:teuthology.orchestra.run.vm00.stdout:/sbin/ip: stdout 2: eth0: mtu 1500 state UP qlen 1000 2026-03-06T13:40:00.816 INFO:teuthology.orchestra.run.vm00.stdout:/sbin/ip: stdout inet6 fe80::5055:ff:fe00:0/64 scope link noprefixroute 2026-03-06T13:40:00.817 INFO:teuthology.orchestra.run.vm00.stdout:/sbin/ip: stdout valid_lft forever preferred_lft forever 2026-03-06T13:40:00.817 INFO:teuthology.orchestra.run.vm00.stdout:Mon IP `192.168.123.100` is in CIDR network `192.168.123.0/24` 2026-03-06T13:40:00.817 INFO:teuthology.orchestra.run.vm00.stdout:Mon IP `192.168.123.100` is in CIDR network `192.168.123.0/24` 2026-03-06T13:40:00.817 INFO:teuthology.orchestra.run.vm00.stdout:Inferred mon public CIDR from local network configuration ['192.168.123.0/24', '192.168.123.0/24'] 2026-03-06T13:40:00.817 INFO:teuthology.orchestra.run.vm00.stdout:Internal network (--cluster-network) has not been provided, OSD replication will default to the public_network 2026-03-06T13:40:00.817 INFO:teuthology.orchestra.run.vm00.stdout:Pulling container image harbor.clyso.com/custom-ceph/ceph/ceph:cobaltcore-storage-v19.2.3-fasttrack-3... 2026-03-06T13:40:01.465 INFO:teuthology.orchestra.run.vm00.stdout:/bin/podman: stdout 306e97de47e91c2b4b24d3dc09be3b3a12039b078f343d91220102acc6628a68 2026-03-06T13:40:01.465 INFO:teuthology.orchestra.run.vm00.stdout:/bin/podman: stderr Trying to pull harbor.clyso.com/custom-ceph/ceph/ceph:cobaltcore-storage-v19.2.3-fasttrack-3... 2026-03-06T13:40:01.465 INFO:teuthology.orchestra.run.vm00.stdout:/bin/podman: stderr Getting image source signatures 2026-03-06T13:40:01.465 INFO:teuthology.orchestra.run.vm00.stdout:/bin/podman: stderr Copying blob sha256:d21d4233fd3d4dd2f376e5ef084c47891c860682c1de15a9c0357cea5defbc91 2026-03-06T13:40:01.465 INFO:teuthology.orchestra.run.vm00.stdout:/bin/podman: stderr Copying config sha256:306e97de47e91c2b4b24d3dc09be3b3a12039b078f343d91220102acc6628a68 2026-03-06T13:40:01.465 INFO:teuthology.orchestra.run.vm00.stdout:/bin/podman: stderr Writing manifest to image destination 2026-03-06T13:40:01.684 INFO:teuthology.orchestra.run.vm00.stdout:ceph: stdout ceph version 19.2.3-47-gc24117fd552 (c24117fd5525679b799527bc1bd1f1dd0a2db5e2) squid (stable) 2026-03-06T13:40:01.684 INFO:teuthology.orchestra.run.vm00.stdout:Ceph version: ceph version 19.2.3-47-gc24117fd552 (c24117fd5525679b799527bc1bd1f1dd0a2db5e2) squid (stable) 2026-03-06T13:40:01.684 INFO:teuthology.orchestra.run.vm00.stdout:Extracting ceph user uid/gid from container image... 2026-03-06T13:40:01.795 INFO:teuthology.orchestra.run.vm00.stdout:stat: stdout 167 167 2026-03-06T13:40:01.795 INFO:teuthology.orchestra.run.vm00.stdout:Creating initial keys... 2026-03-06T13:40:01.895 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph-authtool: stdout AQAhy6ppHQBvMxAAtUXIFRLgToE6t6Mzzcyf7Q== 2026-03-06T13:40:01.975 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph-authtool: stdout AQAhy6ppyXQmORAApXL5Moym5q7A6OeRz6AsuQ== 2026-03-06T13:40:02.073 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph-authtool: stdout AQAiy6pphXA5AhAAJv0fZRVqok3M9RXb0aT8Zg== 2026-03-06T13:40:02.073 INFO:teuthology.orchestra.run.vm00.stdout:Creating initial monmap... 2026-03-06T13:40:02.171 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/monmaptool: stdout /usr/bin/monmaptool: monmap file /tmp/monmap 2026-03-06T13:40:02.171 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/monmaptool: stdout setting min_mon_release = quincy 2026-03-06T13:40:02.171 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/monmaptool: stdout /usr/bin/monmaptool: set fsid to 82f504c8-1959-11f1-9a16-a78d135e200b 2026-03-06T13:40:02.171 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/monmaptool: stdout /usr/bin/monmaptool: writing epoch 0 to /tmp/monmap (1 monitors) 2026-03-06T13:40:02.172 INFO:teuthology.orchestra.run.vm00.stdout:monmaptool for a [v2:192.168.123.100:3300,v1:192.168.123.100:6789] on /usr/bin/monmaptool: monmap file /tmp/monmap 2026-03-06T13:40:02.172 INFO:teuthology.orchestra.run.vm00.stdout:setting min_mon_release = quincy 2026-03-06T13:40:02.172 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/monmaptool: set fsid to 82f504c8-1959-11f1-9a16-a78d135e200b 2026-03-06T13:40:02.172 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/monmaptool: writing epoch 0 to /tmp/monmap (1 monitors) 2026-03-06T13:40:02.172 INFO:teuthology.orchestra.run.vm00.stdout: 2026-03-06T13:40:02.172 INFO:teuthology.orchestra.run.vm00.stdout:Creating mon... 2026-03-06T13:40:02.309 INFO:teuthology.orchestra.run.vm00.stdout:create mon.a on 2026-03-06T13:40:02.466 INFO:teuthology.orchestra.run.vm00.stdout:systemctl: stderr Removed "/etc/systemd/system/multi-user.target.wants/ceph.target". 2026-03-06T13:40:02.581 INFO:teuthology.orchestra.run.vm00.stdout:systemctl: stderr Created symlink /etc/systemd/system/multi-user.target.wants/ceph.target → /etc/systemd/system/ceph.target. 2026-03-06T13:40:02.709 INFO:teuthology.orchestra.run.vm00.stdout:systemctl: stderr Created symlink /etc/systemd/system/multi-user.target.wants/ceph-82f504c8-1959-11f1-9a16-a78d135e200b.target → /etc/systemd/system/ceph-82f504c8-1959-11f1-9a16-a78d135e200b.target. 2026-03-06T13:40:02.709 INFO:teuthology.orchestra.run.vm00.stdout:systemctl: stderr Created symlink /etc/systemd/system/ceph.target.wants/ceph-82f504c8-1959-11f1-9a16-a78d135e200b.target → /etc/systemd/system/ceph-82f504c8-1959-11f1-9a16-a78d135e200b.target. 2026-03-06T13:40:02.855 INFO:teuthology.orchestra.run.vm00.stdout:Non-zero exit code 1 from systemctl reset-failed ceph-82f504c8-1959-11f1-9a16-a78d135e200b@mon.a 2026-03-06T13:40:02.855 INFO:teuthology.orchestra.run.vm00.stdout:systemctl: stderr Failed to reset failed state of unit ceph-82f504c8-1959-11f1-9a16-a78d135e200b@mon.a.service: Unit ceph-82f504c8-1959-11f1-9a16-a78d135e200b@mon.a.service not loaded. 2026-03-06T13:40:02.991 INFO:teuthology.orchestra.run.vm00.stdout:systemctl: stderr Created symlink /etc/systemd/system/ceph-82f504c8-1959-11f1-9a16-a78d135e200b.target.wants/ceph-82f504c8-1959-11f1-9a16-a78d135e200b@mon.a.service → /etc/systemd/system/ceph-82f504c8-1959-11f1-9a16-a78d135e200b@.service. 2026-03-06T13:40:03.152 INFO:teuthology.orchestra.run.vm00.stdout:firewalld does not appear to be present 2026-03-06T13:40:03.153 INFO:teuthology.orchestra.run.vm00.stdout:Not possible to enable service . firewalld.service is not available 2026-03-06T13:40:03.153 INFO:teuthology.orchestra.run.vm00.stdout:Waiting for mon to start... 2026-03-06T13:40:03.153 INFO:teuthology.orchestra.run.vm00.stdout:Waiting for mon... 2026-03-06T13:40:03.441 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout cluster: 2026-03-06T13:40:03.441 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout id: 82f504c8-1959-11f1-9a16-a78d135e200b 2026-03-06T13:40:03.441 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout health: HEALTH_OK 2026-03-06T13:40:03.441 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout 2026-03-06T13:40:03.441 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout services: 2026-03-06T13:40:03.441 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout mon: 1 daemons, quorum a (age 0.219059s) 2026-03-06T13:40:03.441 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout mgr: no daemons active 2026-03-06T13:40:03.441 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout osd: 0 osds: 0 up, 0 in 2026-03-06T13:40:03.441 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout 2026-03-06T13:40:03.441 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout data: 2026-03-06T13:40:03.441 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout pools: 0 pools, 0 pgs 2026-03-06T13:40:03.441 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout objects: 0 objects, 0 B 2026-03-06T13:40:03.441 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout usage: 0 B used, 0 B / 0 B avail 2026-03-06T13:40:03.441 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout pgs: 2026-03-06T13:40:03.441 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout 2026-03-06T13:40:03.441 INFO:teuthology.orchestra.run.vm00.stdout:mon is available 2026-03-06T13:40:03.441 INFO:teuthology.orchestra.run.vm00.stdout:Assimilating anything we can from ceph.conf... 2026-03-06T13:40:03.716 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout 2026-03-06T13:40:03.716 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout [global] 2026-03-06T13:40:03.716 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout fsid = 82f504c8-1959-11f1-9a16-a78d135e200b 2026-03-06T13:40:03.716 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout mon_cluster_log_file_level = debug 2026-03-06T13:40:03.716 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout mon_host = [v2:192.168.123.100:3300,v1:192.168.123.100:6789] 2026-03-06T13:40:03.716 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout mon_osd_allow_pg_remap = true 2026-03-06T13:40:03.716 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout mon_osd_allow_primary_affinity = true 2026-03-06T13:40:03.716 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout mon_warn_on_no_sortbitwise = false 2026-03-06T13:40:03.716 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout osd_crush_chooseleaf_type = 0 2026-03-06T13:40:03.716 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout 2026-03-06T13:40:03.716 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout [mgr] 2026-03-06T13:40:03.716 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout mgr/cephadm/use_agent = False 2026-03-06T13:40:03.716 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout mgr/telemetry/nag = false 2026-03-06T13:40:03.716 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout 2026-03-06T13:40:03.716 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout [osd] 2026-03-06T13:40:03.716 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout osd_map_max_advance = 10 2026-03-06T13:40:03.716 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout osd_sloppy_crc = true 2026-03-06T13:40:03.716 INFO:teuthology.orchestra.run.vm00.stdout:Generating new minimal ceph.conf... 2026-03-06T13:40:03.968 INFO:teuthology.orchestra.run.vm00.stdout:Restarting the monitor... 2026-03-06T13:40:04.208 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:40:04 vm00 systemd[1]: Starting Ceph mon.a for 82f504c8-1959-11f1-9a16-a78d135e200b... 2026-03-06T13:40:04.255 INFO:teuthology.orchestra.run.vm00.stdout:Setting public_network to 192.168.123.0/24 in mon config section 2026-03-06T13:40:04.506 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:40:04 vm00 podman[51899]: 2026-03-06 13:40:04.20691487 +0100 CET m=+0.014894010 container create 0031c6cd048d438b6cd06fb48ea46d96b883426111519d2cc365e7ff3fa28546 (image=harbor.clyso.com/custom-ceph/ceph/ceph:cobaltcore-storage-v19.2.3-fasttrack-3, name=ceph-82f504c8-1959-11f1-9a16-a78d135e200b-mon-a, CEPH_REF=19.2.3-47-gc24117fd552, CEPH_SHA1=c24117fd5525679b799527bc1bd1f1dd0a2db5e2, FROM_IMAGE=rockylinux:9, GANESHA_REPO_BASEURL=https://buildlogs.centos.org/centos/$releasever-stream/storage/$basearch/nfsganesha-5/, OSD_FLAVOR=default, ceph=True, io.buildah.version=1.41.8, CEPH_GIT_REPO=https://github.com/irq0/ceph.git) 2026-03-06T13:40:04.506 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:40:04 vm00 podman[51899]: 2026-03-06 13:40:04.239184343 +0100 CET m=+0.047163483 container init 0031c6cd048d438b6cd06fb48ea46d96b883426111519d2cc365e7ff3fa28546 (image=harbor.clyso.com/custom-ceph/ceph/ceph:cobaltcore-storage-v19.2.3-fasttrack-3, name=ceph-82f504c8-1959-11f1-9a16-a78d135e200b-mon-a, ceph=True, io.buildah.version=1.41.8, CEPH_GIT_REPO=https://github.com/irq0/ceph.git, CEPH_REF=19.2.3-47-gc24117fd552, CEPH_SHA1=c24117fd5525679b799527bc1bd1f1dd0a2db5e2, FROM_IMAGE=rockylinux:9, GANESHA_REPO_BASEURL=https://buildlogs.centos.org/centos/$releasever-stream/storage/$basearch/nfsganesha-5/, OSD_FLAVOR=default) 2026-03-06T13:40:04.506 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:40:04 vm00 podman[51899]: 2026-03-06 13:40:04.243795237 +0100 CET m=+0.051774377 container start 0031c6cd048d438b6cd06fb48ea46d96b883426111519d2cc365e7ff3fa28546 (image=harbor.clyso.com/custom-ceph/ceph/ceph:cobaltcore-storage-v19.2.3-fasttrack-3, name=ceph-82f504c8-1959-11f1-9a16-a78d135e200b-mon-a, ceph=True, io.buildah.version=1.41.8, CEPH_GIT_REPO=https://github.com/irq0/ceph.git, CEPH_REF=19.2.3-47-gc24117fd552, CEPH_SHA1=c24117fd5525679b799527bc1bd1f1dd0a2db5e2, FROM_IMAGE=rockylinux:9, GANESHA_REPO_BASEURL=https://buildlogs.centos.org/centos/$releasever-stream/storage/$basearch/nfsganesha-5/, OSD_FLAVOR=default) 2026-03-06T13:40:04.506 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:40:04 vm00 bash[51899]: 0031c6cd048d438b6cd06fb48ea46d96b883426111519d2cc365e7ff3fa28546 2026-03-06T13:40:04.506 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:40:04 vm00 podman[51899]: 2026-03-06 13:40:04.200904535 +0100 CET m=+0.008883686 image pull 306e97de47e91c2b4b24d3dc09be3b3a12039b078f343d91220102acc6628a68 harbor.clyso.com/custom-ceph/ceph/ceph:cobaltcore-storage-v19.2.3-fasttrack-3 2026-03-06T13:40:04.506 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:40:04 vm00 systemd[1]: Started Ceph mon.a for 82f504c8-1959-11f1-9a16-a78d135e200b. 2026-03-06T13:40:04.506 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:40:04 vm00 ceph-mon[51914]: set uid:gid to 167:167 (ceph:ceph) 2026-03-06T13:40:04.506 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:40:04 vm00 ceph-mon[51914]: ceph version 19.2.3-47-gc24117fd552 (c24117fd5525679b799527bc1bd1f1dd0a2db5e2) squid (stable), process ceph-mon, pid 2 2026-03-06T13:40:04.506 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:40:04 vm00 ceph-mon[51914]: pidfile_write: ignore empty --pid-file 2026-03-06T13:40:04.506 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:40:04 vm00 ceph-mon[51914]: load: jerasure load: lrc 2026-03-06T13:40:04.506 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:40:04 vm00 ceph-mon[51914]: rocksdb: RocksDB version: 7.9.2 2026-03-06T13:40:04.506 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:40:04 vm00 ceph-mon[51914]: rocksdb: Git sha 0 2026-03-06T13:40:04.506 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:40:04 vm00 ceph-mon[51914]: rocksdb: Compile date 2026-03-03 21:08:28 2026-03-06T13:40:04.506 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:40:04 vm00 ceph-mon[51914]: rocksdb: DB SUMMARY 2026-03-06T13:40:04.506 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:40:04 vm00 ceph-mon[51914]: rocksdb: DB Session ID: W9FWQE4878M5Q8KNELKP 2026-03-06T13:40:04.506 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:40:04 vm00 ceph-mon[51914]: rocksdb: CURRENT file: CURRENT 2026-03-06T13:40:04.506 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:40:04 vm00 ceph-mon[51914]: rocksdb: IDENTITY file: IDENTITY 2026-03-06T13:40:04.506 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:40:04 vm00 ceph-mon[51914]: rocksdb: MANIFEST file: MANIFEST-000010 size: 179 Bytes 2026-03-06T13:40:04.506 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:40:04 vm00 ceph-mon[51914]: rocksdb: SST files in /var/lib/ceph/mon/ceph-a/store.db dir, Total Num: 1, files: 000008.sst 2026-03-06T13:40:04.506 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:40:04 vm00 ceph-mon[51914]: rocksdb: Write Ahead Log file in /var/lib/ceph/mon/ceph-a/store.db: 000009.log size: 75679 ; 2026-03-06T13:40:04.506 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:40:04 vm00 ceph-mon[51914]: rocksdb: Options.error_if_exists: 0 2026-03-06T13:40:04.506 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:40:04 vm00 ceph-mon[51914]: rocksdb: Options.create_if_missing: 0 2026-03-06T13:40:04.506 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:40:04 vm00 ceph-mon[51914]: rocksdb: Options.paranoid_checks: 1 2026-03-06T13:40:04.506 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:40:04 vm00 ceph-mon[51914]: rocksdb: Options.flush_verify_memtable_count: 1 2026-03-06T13:40:04.506 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:40:04 vm00 ceph-mon[51914]: rocksdb: Options.track_and_verify_wals_in_manifest: 0 2026-03-06T13:40:04.506 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:40:04 vm00 ceph-mon[51914]: rocksdb: Options.verify_sst_unique_id_in_manifest: 1 2026-03-06T13:40:04.506 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:40:04 vm00 ceph-mon[51914]: rocksdb: Options.env: 0x55932ee53ca0 2026-03-06T13:40:04.506 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:40:04 vm00 ceph-mon[51914]: rocksdb: Options.fs: PosixFileSystem 2026-03-06T13:40:04.506 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:40:04 vm00 ceph-mon[51914]: rocksdb: Options.info_log: 0x5593304e2320 2026-03-06T13:40:04.506 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:40:04 vm00 ceph-mon[51914]: rocksdb: Options.max_file_opening_threads: 16 2026-03-06T13:40:04.506 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:40:04 vm00 ceph-mon[51914]: rocksdb: Options.statistics: (nil) 2026-03-06T13:40:04.506 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:40:04 vm00 ceph-mon[51914]: rocksdb: Options.use_fsync: 0 2026-03-06T13:40:04.506 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:40:04 vm00 ceph-mon[51914]: rocksdb: Options.max_log_file_size: 0 2026-03-06T13:40:04.507 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:40:04 vm00 ceph-mon[51914]: rocksdb: Options.max_manifest_file_size: 1073741824 2026-03-06T13:40:04.507 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:40:04 vm00 ceph-mon[51914]: rocksdb: Options.log_file_time_to_roll: 0 2026-03-06T13:40:04.507 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:40:04 vm00 ceph-mon[51914]: rocksdb: Options.keep_log_file_num: 1000 2026-03-06T13:40:04.507 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:40:04 vm00 ceph-mon[51914]: rocksdb: Options.recycle_log_file_num: 0 2026-03-06T13:40:04.507 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:40:04 vm00 ceph-mon[51914]: rocksdb: Options.allow_fallocate: 1 2026-03-06T13:40:04.507 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:40:04 vm00 ceph-mon[51914]: rocksdb: Options.allow_mmap_reads: 0 2026-03-06T13:40:04.507 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:40:04 vm00 ceph-mon[51914]: rocksdb: Options.allow_mmap_writes: 0 2026-03-06T13:40:04.507 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:40:04 vm00 ceph-mon[51914]: rocksdb: Options.use_direct_reads: 0 2026-03-06T13:40:04.507 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:40:04 vm00 ceph-mon[51914]: rocksdb: Options.use_direct_io_for_flush_and_compaction: 0 2026-03-06T13:40:04.507 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:40:04 vm00 ceph-mon[51914]: rocksdb: Options.create_missing_column_families: 0 2026-03-06T13:40:04.507 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:40:04 vm00 ceph-mon[51914]: rocksdb: Options.db_log_dir: 2026-03-06T13:40:04.507 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:40:04 vm00 ceph-mon[51914]: rocksdb: Options.wal_dir: 2026-03-06T13:40:04.507 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:40:04 vm00 ceph-mon[51914]: rocksdb: Options.table_cache_numshardbits: 6 2026-03-06T13:40:04.507 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:40:04 vm00 ceph-mon[51914]: rocksdb: Options.WAL_ttl_seconds: 0 2026-03-06T13:40:04.507 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:40:04 vm00 ceph-mon[51914]: rocksdb: Options.WAL_size_limit_MB: 0 2026-03-06T13:40:04.507 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:40:04 vm00 ceph-mon[51914]: rocksdb: Options.max_write_batch_group_size_bytes: 1048576 2026-03-06T13:40:04.507 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:40:04 vm00 ceph-mon[51914]: rocksdb: Options.manifest_preallocation_size: 4194304 2026-03-06T13:40:04.507 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:40:04 vm00 ceph-mon[51914]: rocksdb: Options.is_fd_close_on_exec: 1 2026-03-06T13:40:04.507 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:40:04 vm00 ceph-mon[51914]: rocksdb: Options.advise_random_on_open: 1 2026-03-06T13:40:04.507 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:40:04 vm00 ceph-mon[51914]: rocksdb: Options.db_write_buffer_size: 0 2026-03-06T13:40:04.507 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:40:04 vm00 ceph-mon[51914]: rocksdb: Options.write_buffer_manager: 0x5593304e7900 2026-03-06T13:40:04.507 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:40:04 vm00 ceph-mon[51914]: rocksdb: Options.access_hint_on_compaction_start: 1 2026-03-06T13:40:04.507 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:40:04 vm00 ceph-mon[51914]: rocksdb: Options.random_access_max_buffer_size: 1048576 2026-03-06T13:40:04.507 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:40:04 vm00 ceph-mon[51914]: rocksdb: Options.use_adaptive_mutex: 0 2026-03-06T13:40:04.507 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:40:04 vm00 ceph-mon[51914]: rocksdb: Options.rate_limiter: (nil) 2026-03-06T13:40:04.507 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:40:04 vm00 ceph-mon[51914]: rocksdb: Options.sst_file_manager.rate_bytes_per_sec: 0 2026-03-06T13:40:04.507 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:40:04 vm00 ceph-mon[51914]: rocksdb: Options.wal_recovery_mode: 2 2026-03-06T13:40:04.507 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:40:04 vm00 ceph-mon[51914]: rocksdb: Options.enable_thread_tracking: 0 2026-03-06T13:40:04.507 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:40:04 vm00 ceph-mon[51914]: rocksdb: Options.enable_pipelined_write: 0 2026-03-06T13:40:04.507 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:40:04 vm00 ceph-mon[51914]: rocksdb: Options.unordered_write: 0 2026-03-06T13:40:04.507 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:40:04 vm00 ceph-mon[51914]: rocksdb: Options.allow_concurrent_memtable_write: 1 2026-03-06T13:40:04.507 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:40:04 vm00 ceph-mon[51914]: rocksdb: Options.enable_write_thread_adaptive_yield: 1 2026-03-06T13:40:04.507 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:40:04 vm00 ceph-mon[51914]: rocksdb: Options.write_thread_max_yield_usec: 100 2026-03-06T13:40:04.507 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:40:04 vm00 ceph-mon[51914]: rocksdb: Options.write_thread_slow_yield_usec: 3 2026-03-06T13:40:04.507 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:40:04 vm00 ceph-mon[51914]: rocksdb: Options.row_cache: None 2026-03-06T13:40:04.507 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:40:04 vm00 ceph-mon[51914]: rocksdb: Options.wal_filter: None 2026-03-06T13:40:04.507 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:40:04 vm00 ceph-mon[51914]: rocksdb: Options.avoid_flush_during_recovery: 0 2026-03-06T13:40:04.507 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:40:04 vm00 ceph-mon[51914]: rocksdb: Options.allow_ingest_behind: 0 2026-03-06T13:40:04.507 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:40:04 vm00 ceph-mon[51914]: rocksdb: Options.two_write_queues: 0 2026-03-06T13:40:04.507 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:40:04 vm00 ceph-mon[51914]: rocksdb: Options.manual_wal_flush: 0 2026-03-06T13:40:04.507 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:40:04 vm00 ceph-mon[51914]: rocksdb: Options.wal_compression: 0 2026-03-06T13:40:04.507 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:40:04 vm00 ceph-mon[51914]: rocksdb: Options.atomic_flush: 0 2026-03-06T13:40:04.507 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:40:04 vm00 ceph-mon[51914]: rocksdb: Options.avoid_unnecessary_blocking_io: 0 2026-03-06T13:40:04.507 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:40:04 vm00 ceph-mon[51914]: rocksdb: Options.persist_stats_to_disk: 0 2026-03-06T13:40:04.507 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:40:04 vm00 ceph-mon[51914]: rocksdb: Options.write_dbid_to_manifest: 0 2026-03-06T13:40:04.507 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:40:04 vm00 ceph-mon[51914]: rocksdb: Options.log_readahead_size: 0 2026-03-06T13:40:04.507 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:40:04 vm00 ceph-mon[51914]: rocksdb: Options.file_checksum_gen_factory: Unknown 2026-03-06T13:40:04.507 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:40:04 vm00 ceph-mon[51914]: rocksdb: Options.best_efforts_recovery: 0 2026-03-06T13:40:04.508 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:40:04 vm00 ceph-mon[51914]: rocksdb: Options.max_bgerror_resume_count: 2147483647 2026-03-06T13:40:04.508 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:40:04 vm00 ceph-mon[51914]: rocksdb: Options.bgerror_resume_retry_interval: 1000000 2026-03-06T13:40:04.508 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:40:04 vm00 ceph-mon[51914]: rocksdb: Options.allow_data_in_errors: 0 2026-03-06T13:40:04.508 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:40:04 vm00 ceph-mon[51914]: rocksdb: Options.db_host_id: __hostname__ 2026-03-06T13:40:04.508 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:40:04 vm00 ceph-mon[51914]: rocksdb: Options.enforce_single_del_contracts: true 2026-03-06T13:40:04.508 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:40:04 vm00 ceph-mon[51914]: rocksdb: Options.max_background_jobs: 2 2026-03-06T13:40:04.508 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:40:04 vm00 ceph-mon[51914]: rocksdb: Options.max_background_compactions: -1 2026-03-06T13:40:04.508 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:40:04 vm00 ceph-mon[51914]: rocksdb: Options.max_subcompactions: 1 2026-03-06T13:40:04.508 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:40:04 vm00 ceph-mon[51914]: rocksdb: Options.avoid_flush_during_shutdown: 0 2026-03-06T13:40:04.508 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:40:04 vm00 ceph-mon[51914]: rocksdb: Options.writable_file_max_buffer_size: 1048576 2026-03-06T13:40:04.508 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:40:04 vm00 ceph-mon[51914]: rocksdb: Options.delayed_write_rate : 16777216 2026-03-06T13:40:04.508 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:40:04 vm00 ceph-mon[51914]: rocksdb: Options.max_total_wal_size: 0 2026-03-06T13:40:04.508 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:40:04 vm00 ceph-mon[51914]: rocksdb: Options.delete_obsolete_files_period_micros: 21600000000 2026-03-06T13:40:04.508 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:40:04 vm00 ceph-mon[51914]: rocksdb: Options.stats_dump_period_sec: 600 2026-03-06T13:40:04.508 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:40:04 vm00 ceph-mon[51914]: rocksdb: Options.stats_persist_period_sec: 600 2026-03-06T13:40:04.508 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:40:04 vm00 ceph-mon[51914]: rocksdb: Options.stats_history_buffer_size: 1048576 2026-03-06T13:40:04.508 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:40:04 vm00 ceph-mon[51914]: rocksdb: Options.max_open_files: -1 2026-03-06T13:40:04.508 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:40:04 vm00 ceph-mon[51914]: rocksdb: Options.bytes_per_sync: 0 2026-03-06T13:40:04.508 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:40:04 vm00 ceph-mon[51914]: rocksdb: Options.wal_bytes_per_sync: 0 2026-03-06T13:40:04.508 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:40:04 vm00 ceph-mon[51914]: rocksdb: Options.strict_bytes_per_sync: 0 2026-03-06T13:40:04.508 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:40:04 vm00 ceph-mon[51914]: rocksdb: Options.compaction_readahead_size: 0 2026-03-06T13:40:04.508 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:40:04 vm00 ceph-mon[51914]: rocksdb: Options.max_background_flushes: -1 2026-03-06T13:40:04.508 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:40:04 vm00 ceph-mon[51914]: rocksdb: Compression algorithms supported: 2026-03-06T13:40:04.508 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:40:04 vm00 ceph-mon[51914]: rocksdb: kZSTD supported: 0 2026-03-06T13:40:04.508 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:40:04 vm00 ceph-mon[51914]: rocksdb: kXpressCompression supported: 0 2026-03-06T13:40:04.508 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:40:04 vm00 ceph-mon[51914]: rocksdb: kBZip2Compression supported: 0 2026-03-06T13:40:04.508 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:40:04 vm00 ceph-mon[51914]: rocksdb: kZSTDNotFinalCompression supported: 0 2026-03-06T13:40:04.508 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:40:04 vm00 ceph-mon[51914]: rocksdb: kLZ4Compression supported: 1 2026-03-06T13:40:04.508 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:40:04 vm00 ceph-mon[51914]: rocksdb: kZlibCompression supported: 1 2026-03-06T13:40:04.508 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:40:04 vm00 ceph-mon[51914]: rocksdb: kLZ4HCCompression supported: 1 2026-03-06T13:40:04.508 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:40:04 vm00 ceph-mon[51914]: rocksdb: kSnappyCompression supported: 1 2026-03-06T13:40:04.508 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:40:04 vm00 ceph-mon[51914]: rocksdb: Fast CRC32 supported: Supported on x86 2026-03-06T13:40:04.508 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:40:04 vm00 ceph-mon[51914]: rocksdb: DMutex implementation: pthread_mutex_t 2026-03-06T13:40:04.508 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:40:04 vm00 ceph-mon[51914]: rocksdb: [db/version_set.cc:5527] Recovering from manifest file: /var/lib/ceph/mon/ceph-a/store.db/MANIFEST-000010 2026-03-06T13:40:04.508 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:40:04 vm00 ceph-mon[51914]: rocksdb: [db/column_family.cc:630] --------------- Options for column family [default]: 2026-03-06T13:40:04.508 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:40:04 vm00 ceph-mon[51914]: rocksdb: Options.comparator: leveldb.BytewiseComparator 2026-03-06T13:40:04.508 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:40:04 vm00 ceph-mon[51914]: rocksdb: Options.merge_operator: 2026-03-06T13:40:04.508 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:40:04 vm00 ceph-mon[51914]: rocksdb: Options.compaction_filter: None 2026-03-06T13:40:04.508 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:40:04 vm00 ceph-mon[51914]: rocksdb: Options.compaction_filter_factory: None 2026-03-06T13:40:04.508 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:40:04 vm00 ceph-mon[51914]: rocksdb: Options.sst_partitioner_factory: None 2026-03-06T13:40:04.508 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:40:04 vm00 ceph-mon[51914]: rocksdb: Options.memtable_factory: SkipListFactory 2026-03-06T13:40:04.508 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:40:04 vm00 ceph-mon[51914]: rocksdb: Options.table_factory: BlockBasedTable 2026-03-06T13:40:04.508 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:40:04 vm00 ceph-mon[51914]: rocksdb: table_factory options: flush_block_policy_factory: FlushBlockBySizePolicyFactory (0x5593304e23e0) 2026-03-06T13:40:04.508 INFO:journalctl@ceph.mon.a.vm00.stdout: cache_index_and_filter_blocks: 1 2026-03-06T13:40:04.508 INFO:journalctl@ceph.mon.a.vm00.stdout: cache_index_and_filter_blocks_with_high_priority: 0 2026-03-06T13:40:04.508 INFO:journalctl@ceph.mon.a.vm00.stdout: pin_l0_filter_and_index_blocks_in_cache: 0 2026-03-06T13:40:04.508 INFO:journalctl@ceph.mon.a.vm00.stdout: pin_top_level_index_and_filter: 1 2026-03-06T13:40:04.508 INFO:journalctl@ceph.mon.a.vm00.stdout: index_type: 0 2026-03-06T13:40:04.509 INFO:journalctl@ceph.mon.a.vm00.stdout: data_block_index_type: 0 2026-03-06T13:40:04.509 INFO:journalctl@ceph.mon.a.vm00.stdout: index_shortening: 1 2026-03-06T13:40:04.509 INFO:journalctl@ceph.mon.a.vm00.stdout: data_block_hash_table_util_ratio: 0.750000 2026-03-06T13:40:04.509 INFO:journalctl@ceph.mon.a.vm00.stdout: checksum: 4 2026-03-06T13:40:04.509 INFO:journalctl@ceph.mon.a.vm00.stdout: no_block_cache: 0 2026-03-06T13:40:04.509 INFO:journalctl@ceph.mon.a.vm00.stdout: block_cache: 0x5593305071f0 2026-03-06T13:40:04.509 INFO:journalctl@ceph.mon.a.vm00.stdout: block_cache_name: BinnedLRUCache 2026-03-06T13:40:04.509 INFO:journalctl@ceph.mon.a.vm00.stdout: block_cache_options: 2026-03-06T13:40:04.509 INFO:journalctl@ceph.mon.a.vm00.stdout: capacity : 536870912 2026-03-06T13:40:04.509 INFO:journalctl@ceph.mon.a.vm00.stdout: num_shard_bits : 4 2026-03-06T13:40:04.509 INFO:journalctl@ceph.mon.a.vm00.stdout: strict_capacity_limit : 0 2026-03-06T13:40:04.509 INFO:journalctl@ceph.mon.a.vm00.stdout: high_pri_pool_ratio: 0.000 2026-03-06T13:40:04.509 INFO:journalctl@ceph.mon.a.vm00.stdout: block_cache_compressed: (nil) 2026-03-06T13:40:04.509 INFO:journalctl@ceph.mon.a.vm00.stdout: persistent_cache: (nil) 2026-03-06T13:40:04.509 INFO:journalctl@ceph.mon.a.vm00.stdout: block_size: 4096 2026-03-06T13:40:04.509 INFO:journalctl@ceph.mon.a.vm00.stdout: block_size_deviation: 10 2026-03-06T13:40:04.509 INFO:journalctl@ceph.mon.a.vm00.stdout: block_restart_interval: 16 2026-03-06T13:40:04.509 INFO:journalctl@ceph.mon.a.vm00.stdout: index_block_restart_interval: 1 2026-03-06T13:40:04.509 INFO:journalctl@ceph.mon.a.vm00.stdout: metadata_block_size: 4096 2026-03-06T13:40:04.509 INFO:journalctl@ceph.mon.a.vm00.stdout: partition_filters: 0 2026-03-06T13:40:04.509 INFO:journalctl@ceph.mon.a.vm00.stdout: use_delta_encoding: 1 2026-03-06T13:40:04.509 INFO:journalctl@ceph.mon.a.vm00.stdout: filter_policy: bloomfilter 2026-03-06T13:40:04.509 INFO:journalctl@ceph.mon.a.vm00.stdout: whole_key_filtering: 1 2026-03-06T13:40:04.509 INFO:journalctl@ceph.mon.a.vm00.stdout: verify_compression: 0 2026-03-06T13:40:04.509 INFO:journalctl@ceph.mon.a.vm00.stdout: read_amp_bytes_per_bit: 0 2026-03-06T13:40:04.509 INFO:journalctl@ceph.mon.a.vm00.stdout: format_version: 5 2026-03-06T13:40:04.509 INFO:journalctl@ceph.mon.a.vm00.stdout: enable_index_compression: 1 2026-03-06T13:40:04.509 INFO:journalctl@ceph.mon.a.vm00.stdout: block_align: 0 2026-03-06T13:40:04.509 INFO:journalctl@ceph.mon.a.vm00.stdout: max_auto_readahead_size: 262144 2026-03-06T13:40:04.509 INFO:journalctl@ceph.mon.a.vm00.stdout: prepopulate_block_cache: 0 2026-03-06T13:40:04.509 INFO:journalctl@ceph.mon.a.vm00.stdout: initial_auto_readahead_size: 8192 2026-03-06T13:40:04.509 INFO:journalctl@ceph.mon.a.vm00.stdout: num_file_reads_for_auto_readahead: 2 2026-03-06T13:40:04.509 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:40:04 vm00 ceph-mon[51914]: rocksdb: Options.write_buffer_size: 33554432 2026-03-06T13:40:04.509 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:40:04 vm00 ceph-mon[51914]: rocksdb: Options.max_write_buffer_number: 2 2026-03-06T13:40:04.509 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:40:04 vm00 ceph-mon[51914]: rocksdb: Options.compression: NoCompression 2026-03-06T13:40:04.509 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:40:04 vm00 ceph-mon[51914]: rocksdb: Options.bottommost_compression: Disabled 2026-03-06T13:40:04.509 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:40:04 vm00 ceph-mon[51914]: rocksdb: Options.prefix_extractor: nullptr 2026-03-06T13:40:04.509 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:40:04 vm00 ceph-mon[51914]: rocksdb: Options.memtable_insert_with_hint_prefix_extractor: nullptr 2026-03-06T13:40:04.509 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:40:04 vm00 ceph-mon[51914]: rocksdb: Options.num_levels: 7 2026-03-06T13:40:04.509 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:40:04 vm00 ceph-mon[51914]: rocksdb: Options.min_write_buffer_number_to_merge: 1 2026-03-06T13:40:04.509 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:40:04 vm00 ceph-mon[51914]: rocksdb: Options.max_write_buffer_number_to_maintain: 0 2026-03-06T13:40:04.509 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:40:04 vm00 ceph-mon[51914]: rocksdb: Options.max_write_buffer_size_to_maintain: 0 2026-03-06T13:40:04.509 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:40:04 vm00 ceph-mon[51914]: rocksdb: Options.bottommost_compression_opts.window_bits: -14 2026-03-06T13:40:04.509 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:40:04 vm00 ceph-mon[51914]: rocksdb: Options.bottommost_compression_opts.level: 32767 2026-03-06T13:40:04.509 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:40:04 vm00 ceph-mon[51914]: rocksdb: Options.bottommost_compression_opts.strategy: 0 2026-03-06T13:40:04.509 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:40:04 vm00 ceph-mon[51914]: rocksdb: Options.bottommost_compression_opts.max_dict_bytes: 0 2026-03-06T13:40:04.509 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:40:04 vm00 ceph-mon[51914]: rocksdb: Options.bottommost_compression_opts.zstd_max_train_bytes: 0 2026-03-06T13:40:04.510 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:40:04 vm00 ceph-mon[51914]: rocksdb: Options.bottommost_compression_opts.parallel_threads: 1 2026-03-06T13:40:04.510 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:40:04 vm00 ceph-mon[51914]: rocksdb: Options.bottommost_compression_opts.enabled: false 2026-03-06T13:40:04.510 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:40:04 vm00 ceph-mon[51914]: rocksdb: Options.bottommost_compression_opts.max_dict_buffer_bytes: 0 2026-03-06T13:40:04.510 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:40:04 vm00 ceph-mon[51914]: rocksdb: Options.bottommost_compression_opts.use_zstd_dict_trainer: true 2026-03-06T13:40:04.510 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:40:04 vm00 ceph-mon[51914]: rocksdb: Options.compression_opts.window_bits: -14 2026-03-06T13:40:04.510 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:40:04 vm00 ceph-mon[51914]: rocksdb: Options.compression_opts.level: 32767 2026-03-06T13:40:04.510 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:40:04 vm00 ceph-mon[51914]: rocksdb: Options.compression_opts.strategy: 0 2026-03-06T13:40:04.510 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:40:04 vm00 ceph-mon[51914]: rocksdb: Options.compression_opts.max_dict_bytes: 0 2026-03-06T13:40:04.510 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:40:04 vm00 ceph-mon[51914]: rocksdb: Options.compression_opts.zstd_max_train_bytes: 0 2026-03-06T13:40:04.510 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:40:04 vm00 ceph-mon[51914]: rocksdb: Options.compression_opts.use_zstd_dict_trainer: true 2026-03-06T13:40:04.510 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:40:04 vm00 ceph-mon[51914]: rocksdb: Options.compression_opts.parallel_threads: 1 2026-03-06T13:40:04.510 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:40:04 vm00 ceph-mon[51914]: rocksdb: Options.compression_opts.enabled: false 2026-03-06T13:40:04.510 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:40:04 vm00 ceph-mon[51914]: rocksdb: Options.compression_opts.max_dict_buffer_bytes: 0 2026-03-06T13:40:04.510 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:40:04 vm00 ceph-mon[51914]: rocksdb: Options.level0_file_num_compaction_trigger: 4 2026-03-06T13:40:04.510 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:40:04 vm00 ceph-mon[51914]: rocksdb: Options.level0_slowdown_writes_trigger: 20 2026-03-06T13:40:04.510 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:40:04 vm00 ceph-mon[51914]: rocksdb: Options.level0_stop_writes_trigger: 36 2026-03-06T13:40:04.510 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:40:04 vm00 ceph-mon[51914]: rocksdb: Options.target_file_size_base: 67108864 2026-03-06T13:40:04.510 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:40:04 vm00 ceph-mon[51914]: rocksdb: Options.target_file_size_multiplier: 1 2026-03-06T13:40:04.510 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:40:04 vm00 ceph-mon[51914]: rocksdb: Options.max_bytes_for_level_base: 268435456 2026-03-06T13:40:04.510 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:40:04 vm00 ceph-mon[51914]: rocksdb: Options.level_compaction_dynamic_level_bytes: 1 2026-03-06T13:40:04.510 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:40:04 vm00 ceph-mon[51914]: rocksdb: Options.max_bytes_for_level_multiplier: 10.000000 2026-03-06T13:40:04.510 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:40:04 vm00 ceph-mon[51914]: rocksdb: Options.max_bytes_for_level_multiplier_addtl[0]: 1 2026-03-06T13:40:04.510 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:40:04 vm00 ceph-mon[51914]: rocksdb: Options.max_bytes_for_level_multiplier_addtl[1]: 1 2026-03-06T13:40:04.510 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:40:04 vm00 ceph-mon[51914]: rocksdb: Options.max_bytes_for_level_multiplier_addtl[2]: 1 2026-03-06T13:40:04.510 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:40:04 vm00 ceph-mon[51914]: rocksdb: Options.max_bytes_for_level_multiplier_addtl[3]: 1 2026-03-06T13:40:04.510 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:40:04 vm00 ceph-mon[51914]: rocksdb: Options.max_bytes_for_level_multiplier_addtl[4]: 1 2026-03-06T13:40:04.510 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:40:04 vm00 ceph-mon[51914]: rocksdb: Options.max_bytes_for_level_multiplier_addtl[5]: 1 2026-03-06T13:40:04.510 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:40:04 vm00 ceph-mon[51914]: rocksdb: Options.max_bytes_for_level_multiplier_addtl[6]: 1 2026-03-06T13:40:04.510 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:40:04 vm00 ceph-mon[51914]: rocksdb: Options.max_sequential_skip_in_iterations: 8 2026-03-06T13:40:04.510 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:40:04 vm00 ceph-mon[51914]: rocksdb: Options.max_compaction_bytes: 1677721600 2026-03-06T13:40:04.510 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:40:04 vm00 ceph-mon[51914]: rocksdb: Options.ignore_max_compaction_bytes_for_input: true 2026-03-06T13:40:04.510 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:40:04 vm00 ceph-mon[51914]: rocksdb: Options.arena_block_size: 1048576 2026-03-06T13:40:04.510 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:40:04 vm00 ceph-mon[51914]: rocksdb: Options.soft_pending_compaction_bytes_limit: 68719476736 2026-03-06T13:40:04.510 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:40:04 vm00 ceph-mon[51914]: rocksdb: Options.hard_pending_compaction_bytes_limit: 274877906944 2026-03-06T13:40:04.510 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:40:04 vm00 ceph-mon[51914]: rocksdb: Options.disable_auto_compactions: 0 2026-03-06T13:40:04.510 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:40:04 vm00 ceph-mon[51914]: rocksdb: Options.compaction_style: kCompactionStyleLevel 2026-03-06T13:40:04.510 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:40:04 vm00 ceph-mon[51914]: rocksdb: Options.compaction_pri: kMinOverlappingRatio 2026-03-06T13:40:04.510 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:40:04 vm00 ceph-mon[51914]: rocksdb: Options.compaction_options_universal.size_ratio: 1 2026-03-06T13:40:04.510 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:40:04 vm00 ceph-mon[51914]: rocksdb: Options.compaction_options_universal.min_merge_width: 2 2026-03-06T13:40:04.510 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:40:04 vm00 ceph-mon[51914]: rocksdb: Options.compaction_options_universal.max_merge_width: 4294967295 2026-03-06T13:40:04.510 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:40:04 vm00 ceph-mon[51914]: rocksdb: Options.compaction_options_universal.max_size_amplification_percent: 200 2026-03-06T13:40:04.510 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:40:04 vm00 ceph-mon[51914]: rocksdb: Options.compaction_options_universal.compression_size_percent: -1 2026-03-06T13:40:04.510 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:40:04 vm00 ceph-mon[51914]: rocksdb: Options.compaction_options_universal.stop_style: kCompactionStopStyleTotalSize 2026-03-06T13:40:04.510 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:40:04 vm00 ceph-mon[51914]: rocksdb: Options.compaction_options_fifo.max_table_files_size: 1073741824 2026-03-06T13:40:04.510 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:40:04 vm00 ceph-mon[51914]: rocksdb: Options.compaction_options_fifo.allow_compaction: 0 2026-03-06T13:40:04.510 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:40:04 vm00 ceph-mon[51914]: rocksdb: Options.table_properties_collectors: CompactOnDeletionCollector (Sliding window size = 32768 Deletion trigger = 16384 Deletion ratio = 0); 2026-03-06T13:40:04.510 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:40:04 vm00 ceph-mon[51914]: rocksdb: Options.inplace_update_support: 0 2026-03-06T13:40:04.510 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:40:04 vm00 ceph-mon[51914]: rocksdb: Options.inplace_update_num_locks: 10000 2026-03-06T13:40:04.511 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:40:04 vm00 ceph-mon[51914]: rocksdb: Options.memtable_prefix_bloom_size_ratio: 0.000000 2026-03-06T13:40:04.511 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:40:04 vm00 ceph-mon[51914]: rocksdb: Options.memtable_whole_key_filtering: 0 2026-03-06T13:40:04.511 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:40:04 vm00 ceph-mon[51914]: rocksdb: Options.memtable_huge_page_size: 0 2026-03-06T13:40:04.511 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:40:04 vm00 ceph-mon[51914]: rocksdb: Options.bloom_locality: 0 2026-03-06T13:40:04.511 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:40:04 vm00 ceph-mon[51914]: rocksdb: Options.max_successive_merges: 0 2026-03-06T13:40:04.511 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:40:04 vm00 ceph-mon[51914]: rocksdb: Options.optimize_filters_for_hits: 0 2026-03-06T13:40:04.511 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:40:04 vm00 ceph-mon[51914]: rocksdb: Options.paranoid_file_checks: 0 2026-03-06T13:40:04.511 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:40:04 vm00 ceph-mon[51914]: rocksdb: Options.force_consistency_checks: 1 2026-03-06T13:40:04.511 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:40:04 vm00 ceph-mon[51914]: rocksdb: Options.report_bg_io_stats: 0 2026-03-06T13:40:04.511 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:40:04 vm00 ceph-mon[51914]: rocksdb: Options.ttl: 2592000 2026-03-06T13:40:04.511 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:40:04 vm00 ceph-mon[51914]: rocksdb: Options.periodic_compaction_seconds: 0 2026-03-06T13:40:04.511 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:40:04 vm00 ceph-mon[51914]: rocksdb: Options.preclude_last_level_data_seconds: 0 2026-03-06T13:40:04.511 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:40:04 vm00 ceph-mon[51914]: rocksdb: Options.preserve_internal_time_seconds: 0 2026-03-06T13:40:04.511 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:40:04 vm00 ceph-mon[51914]: rocksdb: Options.enable_blob_files: false 2026-03-06T13:40:04.511 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:40:04 vm00 ceph-mon[51914]: rocksdb: Options.min_blob_size: 0 2026-03-06T13:40:04.511 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:40:04 vm00 ceph-mon[51914]: rocksdb: Options.blob_file_size: 268435456 2026-03-06T13:40:04.511 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:40:04 vm00 ceph-mon[51914]: rocksdb: Options.blob_compression_type: NoCompression 2026-03-06T13:40:04.511 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:40:04 vm00 ceph-mon[51914]: rocksdb: Options.enable_blob_garbage_collection: false 2026-03-06T13:40:04.511 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:40:04 vm00 ceph-mon[51914]: rocksdb: Options.blob_garbage_collection_age_cutoff: 0.250000 2026-03-06T13:40:04.511 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:40:04 vm00 ceph-mon[51914]: rocksdb: Options.blob_garbage_collection_force_threshold: 1.000000 2026-03-06T13:40:04.511 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:40:04 vm00 ceph-mon[51914]: rocksdb: Options.blob_compaction_readahead_size: 0 2026-03-06T13:40:04.511 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:40:04 vm00 ceph-mon[51914]: rocksdb: Options.blob_file_starting_level: 0 2026-03-06T13:40:04.511 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:40:04 vm00 ceph-mon[51914]: rocksdb: Options.experimental_mempurge_threshold: 0.000000 2026-03-06T13:40:04.511 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:40:04 vm00 ceph-mon[51914]: rocksdb: [db/version_set.cc:5566] Recovered from manifest file:/var/lib/ceph/mon/ceph-a/store.db/MANIFEST-000010 succeeded,manifest_file_number is 10, next_file_number is 12, last_sequence is 5, log_number is 5,prev_log_number is 0,max_column_family is 0,min_log_number_to_keep is 5 2026-03-06T13:40:04.511 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:40:04 vm00 ceph-mon[51914]: rocksdb: [db/version_set.cc:5581] Column family [default] (ID 0), log number is 5 2026-03-06T13:40:04.511 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:40:04 vm00 ceph-mon[51914]: rocksdb: [db/db_impl/db_impl_open.cc:539] DB ID: 8ba37821-7457-4ff1-a3da-57f53b2f2b9f 2026-03-06T13:40:04.511 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:40:04 vm00 ceph-mon[51914]: rocksdb: EVENT_LOG_v1 {"time_micros": 1772800804265106, "job": 1, "event": "recovery_started", "wal_files": [9]} 2026-03-06T13:40:04.511 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:40:04 vm00 ceph-mon[51914]: rocksdb: [db/db_impl/db_impl_open.cc:1043] Recovering log #9 mode 2 2026-03-06T13:40:04.511 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:40:04 vm00 ceph-mon[51914]: rocksdb: EVENT_LOG_v1 {"time_micros": 1772800804266527, "cf_name": "default", "job": 1, "event": "table_file_creation", "file_number": 13, "file_size": 72760, "file_checksum": "", "file_checksum_func_name": "Unknown", "smallest_seqno": 8, "largest_seqno": 225, "table_properties": {"data_size": 71039, "index_size": 174, "index_partitions": 0, "top_level_index_size": 0, "index_key_is_user_key": 1, "index_value_is_delta_encoded": 1, "filter_size": 517, "raw_key_size": 9705, "raw_average_key_size": 49, "raw_value_size": 65518, "raw_average_value_size": 334, "num_data_blocks": 8, "num_entries": 196, "num_filter_entries": 196, "num_deletions": 3, "num_merge_operands": 0, "num_range_deletions": 0, "format_version": 0, "fixed_key_len": 0, "filter_policy": "bloomfilter", "column_family_name": "default", "column_family_id": 0, "comparator": "leveldb.BytewiseComparator", "merge_operator": "", "prefix_extractor_name": "nullptr", "property_collectors": "[CompactOnDeletionCollector]", "compression": "NoCompression", "compression_options": "window_bits=-14; level=32767; strategy=0; max_dict_bytes=0; zstd_max_train_bytes=0; enabled=0; max_dict_buffer_bytes=0; use_zstd_dict_trainer=1; ", "creation_time": 1772800804, "oldest_key_time": 0, "file_creation_time": 0, "slow_compression_estimated_data_size": 0, "fast_compression_estimated_data_size": 0, "db_id": "8ba37821-7457-4ff1-a3da-57f53b2f2b9f", "db_session_id": "W9FWQE4878M5Q8KNELKP", "orig_file_number": 13, "seqno_to_time_mapping": "N/A"}} 2026-03-06T13:40:04.511 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:40:04 vm00 ceph-mon[51914]: rocksdb: EVENT_LOG_v1 {"time_micros": 1772800804266579, "job": 1, "event": "recovery_finished"} 2026-03-06T13:40:04.511 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:40:04 vm00 ceph-mon[51914]: rocksdb: [db/version_set.cc:5047] Creating manifest 15 2026-03-06T13:40:04.511 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:40:04 vm00 ceph-mon[51914]: rocksdb: [file/delete_scheduler.cc:74] Deleted file /var/lib/ceph/mon/ceph-a/store.db/000009.log immediately, rate_bytes_per_sec 0, total_trash_size 0 max_trash_db_ratio 0.250000 2026-03-06T13:40:04.511 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:40:04 vm00 ceph-mon[51914]: rocksdb: [db/db_impl/db_impl_open.cc:1987] SstFileManager instance 0x559330508e00 2026-03-06T13:40:04.511 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:40:04 vm00 ceph-mon[51914]: rocksdb: DB pointer 0x559330620000 2026-03-06T13:40:04.511 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:40:04 vm00 ceph-mon[51914]: rocksdb: [db/db_impl/db_impl.cc:1109] ------- DUMPING STATS ------- 2026-03-06T13:40:04.511 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:40:04 vm00 ceph-mon[51914]: rocksdb: [db/db_impl/db_impl.cc:1111] 2026-03-06T13:40:04.511 INFO:journalctl@ceph.mon.a.vm00.stdout: ** DB Stats ** 2026-03-06T13:40:04.511 INFO:journalctl@ceph.mon.a.vm00.stdout: Uptime(secs): 0.0 total, 0.0 interval 2026-03-06T13:40:04.511 INFO:journalctl@ceph.mon.a.vm00.stdout: Cumulative writes: 0 writes, 0 keys, 0 commit groups, 0.0 writes per commit group, ingest: 0.00 GB, 0.00 MB/s 2026-03-06T13:40:04.511 INFO:journalctl@ceph.mon.a.vm00.stdout: Cumulative WAL: 0 writes, 0 syncs, 0.00 writes per sync, written: 0.00 GB, 0.00 MB/s 2026-03-06T13:40:04.511 INFO:journalctl@ceph.mon.a.vm00.stdout: Cumulative stall: 00:00:0.000 H:M:S, 0.0 percent 2026-03-06T13:40:04.511 INFO:journalctl@ceph.mon.a.vm00.stdout: Interval writes: 0 writes, 0 keys, 0 commit groups, 0.0 writes per commit group, ingest: 0.00 MB, 0.00 MB/s 2026-03-06T13:40:04.511 INFO:journalctl@ceph.mon.a.vm00.stdout: Interval WAL: 0 writes, 0 syncs, 0.00 writes per sync, written: 0.00 GB, 0.00 MB/s 2026-03-06T13:40:04.511 INFO:journalctl@ceph.mon.a.vm00.stdout: Interval stall: 00:00:0.000 H:M:S, 0.0 percent 2026-03-06T13:40:04.511 INFO:journalctl@ceph.mon.a.vm00.stdout: 2026-03-06T13:40:04.511 INFO:journalctl@ceph.mon.a.vm00.stdout: ** Compaction Stats [default] ** 2026-03-06T13:40:04.511 INFO:journalctl@ceph.mon.a.vm00.stdout: Level Files Size Score Read(GB) Rn(GB) Rnp1(GB) Write(GB) Wnew(GB) Moved(GB) W-Amp Rd(MB/s) Wr(MB/s) Comp(sec) CompMergeCPU(sec) Comp(cnt) Avg(sec) KeyIn KeyDrop Rblob(GB) Wblob(GB) 2026-03-06T13:40:04.511 INFO:journalctl@ceph.mon.a.vm00.stdout: ------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------ 2026-03-06T13:40:04.512 INFO:journalctl@ceph.mon.a.vm00.stdout: L0 2/0 72.91 KB 0.5 0.0 0.0 0.0 0.0 0.0 0.0 1.0 0.0 56.5 0.00 0.00 1 0.001 0 0 0.0 0.0 2026-03-06T13:40:04.512 INFO:journalctl@ceph.mon.a.vm00.stdout: Sum 2/0 72.91 KB 0.0 0.0 0.0 0.0 0.0 0.0 0.0 1.0 0.0 56.5 0.00 0.00 1 0.001 0 0 0.0 0.0 2026-03-06T13:40:04.512 INFO:journalctl@ceph.mon.a.vm00.stdout: Int 0/0 0.00 KB 0.0 0.0 0.0 0.0 0.0 0.0 0.0 1.0 0.0 56.5 0.00 0.00 1 0.001 0 0 0.0 0.0 2026-03-06T13:40:04.512 INFO:journalctl@ceph.mon.a.vm00.stdout: 2026-03-06T13:40:04.512 INFO:journalctl@ceph.mon.a.vm00.stdout: ** Compaction Stats [default] ** 2026-03-06T13:40:04.512 INFO:journalctl@ceph.mon.a.vm00.stdout: Priority Files Size Score Read(GB) Rn(GB) Rnp1(GB) Write(GB) Wnew(GB) Moved(GB) W-Amp Rd(MB/s) Wr(MB/s) Comp(sec) CompMergeCPU(sec) Comp(cnt) Avg(sec) KeyIn KeyDrop Rblob(GB) Wblob(GB) 2026-03-06T13:40:04.512 INFO:journalctl@ceph.mon.a.vm00.stdout: --------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------- 2026-03-06T13:40:04.512 INFO:journalctl@ceph.mon.a.vm00.stdout: User 0/0 0.00 KB 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0.0 56.5 0.00 0.00 1 0.001 0 0 0.0 0.0 2026-03-06T13:40:04.512 INFO:journalctl@ceph.mon.a.vm00.stdout: 2026-03-06T13:40:04.512 INFO:journalctl@ceph.mon.a.vm00.stdout: Blob file count: 0, total size: 0.0 GB, garbage size: 0.0 GB, space amp: 0.0 2026-03-06T13:40:04.512 INFO:journalctl@ceph.mon.a.vm00.stdout: 2026-03-06T13:40:04.512 INFO:journalctl@ceph.mon.a.vm00.stdout: Uptime(secs): 0.0 total, 0.0 interval 2026-03-06T13:40:04.512 INFO:journalctl@ceph.mon.a.vm00.stdout: Flush(GB): cumulative 0.000, interval 0.000 2026-03-06T13:40:04.512 INFO:journalctl@ceph.mon.a.vm00.stdout: AddFile(GB): cumulative 0.000, interval 0.000 2026-03-06T13:40:04.512 INFO:journalctl@ceph.mon.a.vm00.stdout: AddFile(Total Files): cumulative 0, interval 0 2026-03-06T13:40:04.512 INFO:journalctl@ceph.mon.a.vm00.stdout: AddFile(L0 Files): cumulative 0, interval 0 2026-03-06T13:40:04.512 INFO:journalctl@ceph.mon.a.vm00.stdout: AddFile(Keys): cumulative 0, interval 0 2026-03-06T13:40:04.512 INFO:journalctl@ceph.mon.a.vm00.stdout: Cumulative compaction: 0.00 GB write, 11.68 MB/s write, 0.00 GB read, 0.00 MB/s read, 0.0 seconds 2026-03-06T13:40:04.512 INFO:journalctl@ceph.mon.a.vm00.stdout: Interval compaction: 0.00 GB write, 11.68 MB/s write, 0.00 GB read, 0.00 MB/s read, 0.0 seconds 2026-03-06T13:40:04.512 INFO:journalctl@ceph.mon.a.vm00.stdout: Stalls(count): 0 level0_slowdown, 0 level0_slowdown_with_compaction, 0 level0_numfiles, 0 level0_numfiles_with_compaction, 0 stop for pending_compaction_bytes, 0 slowdown for pending_compaction_bytes, 0 memtable_compaction, 0 memtable_slowdown, interval 0 total count 2026-03-06T13:40:04.512 INFO:journalctl@ceph.mon.a.vm00.stdout: Block cache BinnedLRUCache@0x5593305071f0#2 capacity: 512.00 MB usage: 1.06 KB table_size: 0 occupancy: 18446744073709551615 collections: 1 last_copies: 0 last_secs: 2.4e-05 secs_since: 0 2026-03-06T13:40:04.512 INFO:journalctl@ceph.mon.a.vm00.stdout: Block cache entry stats(count,size,portion): FilterBlock(2,0.70 KB,0.00013411%) IndexBlock(2,0.36 KB,6.85453e-05%) Misc(2,0.95 KB,0.000181794%) 2026-03-06T13:40:04.512 INFO:journalctl@ceph.mon.a.vm00.stdout: 2026-03-06T13:40:04.512 INFO:journalctl@ceph.mon.a.vm00.stdout: ** File Read Latency Histogram By Level [default] ** 2026-03-06T13:40:04.512 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:40:04 vm00 ceph-mon[51914]: starting mon.a rank 0 at public addrs [v2:192.168.123.100:3300/0,v1:192.168.123.100:6789/0] at bind addrs [v2:192.168.123.100:3300/0,v1:192.168.123.100:6789/0] mon_data /var/lib/ceph/mon/ceph-a fsid 82f504c8-1959-11f1-9a16-a78d135e200b 2026-03-06T13:40:04.512 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:40:04 vm00 ceph-mon[51914]: mon.a@-1(???) e1 preinit fsid 82f504c8-1959-11f1-9a16-a78d135e200b 2026-03-06T13:40:04.512 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:40:04 vm00 ceph-mon[51914]: mon.a@-1(???).mds e1 new map 2026-03-06T13:40:04.512 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:40:04 vm00 ceph-mon[51914]: mon.a@-1(???).mds e1 print_map 2026-03-06T13:40:04.512 INFO:journalctl@ceph.mon.a.vm00.stdout: e1 2026-03-06T13:40:04.512 INFO:journalctl@ceph.mon.a.vm00.stdout: btime 2026-03-06T12:40:03:177837+0000 2026-03-06T13:40:04.512 INFO:journalctl@ceph.mon.a.vm00.stdout: enable_multiple, ever_enabled_multiple: 1,1 2026-03-06T13:40:04.512 INFO:journalctl@ceph.mon.a.vm00.stdout: default compat: compat={},rocompat={},incompat={1=base v0.20,2=client writeable ranges,3=default file layouts on dirs,4=dir inode in separate object,5=mds uses versioned encoding,6=dirfrag is stored in omap,8=no anchor table,9=file layout v2,10=snaprealm v2,11=minor log segments,12=quiesce subvolumes} 2026-03-06T13:40:04.512 INFO:journalctl@ceph.mon.a.vm00.stdout: legacy client fscid: -1 2026-03-06T13:40:04.512 INFO:journalctl@ceph.mon.a.vm00.stdout: 2026-03-06T13:40:04.512 INFO:journalctl@ceph.mon.a.vm00.stdout: No filesystems configured 2026-03-06T13:40:04.512 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:40:04 vm00 ceph-mon[51914]: mon.a@-1(???).osd e1 crush map has features 3314932999778484224, adjusting msgr requires 2026-03-06T13:40:04.512 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:40:04 vm00 ceph-mon[51914]: mon.a@-1(???).osd e1 crush map has features 288514050185494528, adjusting msgr requires 2026-03-06T13:40:04.512 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:40:04 vm00 ceph-mon[51914]: mon.a@-1(???).osd e1 crush map has features 288514050185494528, adjusting msgr requires 2026-03-06T13:40:04.512 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:40:04 vm00 ceph-mon[51914]: mon.a@-1(???).osd e1 crush map has features 288514050185494528, adjusting msgr requires 2026-03-06T13:40:04.512 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:40:04 vm00 ceph-mon[51914]: mon.a@-1(???).paxosservice(auth 1..2) refresh upgraded, format 0 -> 3 2026-03-06T13:40:04.512 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:40:04 vm00 ceph-mon[51914]: mon.a is new leader, mons a in quorum (ranks 0) 2026-03-06T13:40:04.512 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:40:04 vm00 ceph-mon[51914]: monmap epoch 1 2026-03-06T13:40:04.512 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:40:04 vm00 ceph-mon[51914]: fsid 82f504c8-1959-11f1-9a16-a78d135e200b 2026-03-06T13:40:04.512 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:40:04 vm00 ceph-mon[51914]: last_changed 2026-03-06T12:40:02.133525+0000 2026-03-06T13:40:04.512 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:40:04 vm00 ceph-mon[51914]: created 2026-03-06T12:40:02.133525+0000 2026-03-06T13:40:04.512 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:40:04 vm00 ceph-mon[51914]: min_mon_release 19 (squid) 2026-03-06T13:40:04.512 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:40:04 vm00 ceph-mon[51914]: election_strategy: 1 2026-03-06T13:40:04.513 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:40:04 vm00 ceph-mon[51914]: 0: [v2:192.168.123.100:3300/0,v1:192.168.123.100:6789/0] mon.a 2026-03-06T13:40:04.513 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:40:04 vm00 ceph-mon[51914]: fsmap 2026-03-06T13:40:04.513 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:40:04 vm00 ceph-mon[51914]: osdmap e1: 0 total, 0 up, 0 in 2026-03-06T13:40:04.513 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:40:04 vm00 ceph-mon[51914]: mgrmap e1: no daemons active 2026-03-06T13:40:04.559 INFO:teuthology.orchestra.run.vm00.stdout:Wrote config to /etc/ceph/ceph.conf 2026-03-06T13:40:04.559 INFO:teuthology.orchestra.run.vm00.stdout:Wrote keyring to /etc/ceph/ceph.client.admin.keyring 2026-03-06T13:40:04.559 INFO:teuthology.orchestra.run.vm00.stdout:Creating mgr... 2026-03-06T13:40:04.560 INFO:teuthology.orchestra.run.vm00.stdout:Verifying port 0.0.0.0:9283 ... 2026-03-06T13:40:04.560 INFO:teuthology.orchestra.run.vm00.stdout:Verifying port 0.0.0.0:8765 ... 2026-03-06T13:40:04.692 INFO:teuthology.orchestra.run.vm00.stdout:Non-zero exit code 1 from systemctl reset-failed ceph-82f504c8-1959-11f1-9a16-a78d135e200b@mgr.a 2026-03-06T13:40:04.692 INFO:teuthology.orchestra.run.vm00.stdout:systemctl: stderr Failed to reset failed state of unit ceph-82f504c8-1959-11f1-9a16-a78d135e200b@mgr.a.service: Unit ceph-82f504c8-1959-11f1-9a16-a78d135e200b@mgr.a.service not loaded. 2026-03-06T13:40:04.808 INFO:teuthology.orchestra.run.vm00.stdout:systemctl: stderr Created symlink /etc/systemd/system/ceph-82f504c8-1959-11f1-9a16-a78d135e200b.target.wants/ceph-82f504c8-1959-11f1-9a16-a78d135e200b@mgr.a.service → /etc/systemd/system/ceph-82f504c8-1959-11f1-9a16-a78d135e200b@.service. 2026-03-06T13:40:04.824 INFO:journalctl@ceph.mgr.a.vm00.stdout:Mar 06 13:40:04 vm00 systemd[1]: Starting Ceph mgr.a for 82f504c8-1959-11f1-9a16-a78d135e200b... 2026-03-06T13:40:04.959 INFO:teuthology.orchestra.run.vm00.stdout:firewalld does not appear to be present 2026-03-06T13:40:04.959 INFO:teuthology.orchestra.run.vm00.stdout:Not possible to enable service . firewalld.service is not available 2026-03-06T13:40:04.959 INFO:teuthology.orchestra.run.vm00.stdout:firewalld does not appear to be present 2026-03-06T13:40:04.959 INFO:teuthology.orchestra.run.vm00.stdout:Not possible to open ports <[9283, 8765]>. firewalld.service is not available 2026-03-06T13:40:04.959 INFO:teuthology.orchestra.run.vm00.stdout:Waiting for mgr to start... 2026-03-06T13:40:04.959 INFO:teuthology.orchestra.run.vm00.stdout:Waiting for mgr... 2026-03-06T13:40:05.082 INFO:journalctl@ceph.mgr.a.vm00.stdout:Mar 06 13:40:04 vm00 podman[52113]: 2026-03-06 13:40:04.913646347 +0100 CET m=+0.015159137 container create 28d034114368a2f278fb4633fdb2033dacf31996e90e28db94d1392dd113ebe5 (image=harbor.clyso.com/custom-ceph/ceph/ceph:cobaltcore-storage-v19.2.3-fasttrack-3, name=ceph-82f504c8-1959-11f1-9a16-a78d135e200b-mgr-a, FROM_IMAGE=rockylinux:9, GANESHA_REPO_BASEURL=https://buildlogs.centos.org/centos/$releasever-stream/storage/$basearch/nfsganesha-5/, OSD_FLAVOR=default, ceph=True, io.buildah.version=1.41.8, CEPH_GIT_REPO=https://github.com/irq0/ceph.git, CEPH_REF=19.2.3-47-gc24117fd552, CEPH_SHA1=c24117fd5525679b799527bc1bd1f1dd0a2db5e2) 2026-03-06T13:40:05.082 INFO:journalctl@ceph.mgr.a.vm00.stdout:Mar 06 13:40:04 vm00 podman[52113]: 2026-03-06 13:40:04.945587344 +0100 CET m=+0.047100145 container init 28d034114368a2f278fb4633fdb2033dacf31996e90e28db94d1392dd113ebe5 (image=harbor.clyso.com/custom-ceph/ceph/ceph:cobaltcore-storage-v19.2.3-fasttrack-3, name=ceph-82f504c8-1959-11f1-9a16-a78d135e200b-mgr-a, FROM_IMAGE=rockylinux:9, GANESHA_REPO_BASEURL=https://buildlogs.centos.org/centos/$releasever-stream/storage/$basearch/nfsganesha-5/, OSD_FLAVOR=default, ceph=True, io.buildah.version=1.41.8, CEPH_GIT_REPO=https://github.com/irq0/ceph.git, CEPH_REF=19.2.3-47-gc24117fd552, CEPH_SHA1=c24117fd5525679b799527bc1bd1f1dd0a2db5e2) 2026-03-06T13:40:05.082 INFO:journalctl@ceph.mgr.a.vm00.stdout:Mar 06 13:40:04 vm00 podman[52113]: 2026-03-06 13:40:04.948269657 +0100 CET m=+0.049782447 container start 28d034114368a2f278fb4633fdb2033dacf31996e90e28db94d1392dd113ebe5 (image=harbor.clyso.com/custom-ceph/ceph/ceph:cobaltcore-storage-v19.2.3-fasttrack-3, name=ceph-82f504c8-1959-11f1-9a16-a78d135e200b-mgr-a, CEPH_SHA1=c24117fd5525679b799527bc1bd1f1dd0a2db5e2, FROM_IMAGE=rockylinux:9, GANESHA_REPO_BASEURL=https://buildlogs.centos.org/centos/$releasever-stream/storage/$basearch/nfsganesha-5/, OSD_FLAVOR=default, ceph=True, io.buildah.version=1.41.8, CEPH_GIT_REPO=https://github.com/irq0/ceph.git, CEPH_REF=19.2.3-47-gc24117fd552) 2026-03-06T13:40:05.082 INFO:journalctl@ceph.mgr.a.vm00.stdout:Mar 06 13:40:04 vm00 bash[52113]: 28d034114368a2f278fb4633fdb2033dacf31996e90e28db94d1392dd113ebe5 2026-03-06T13:40:05.082 INFO:journalctl@ceph.mgr.a.vm00.stdout:Mar 06 13:40:04 vm00 podman[52113]: 2026-03-06 13:40:04.907611495 +0100 CET m=+0.009124285 image pull 306e97de47e91c2b4b24d3dc09be3b3a12039b078f343d91220102acc6628a68 harbor.clyso.com/custom-ceph/ceph/ceph:cobaltcore-storage-v19.2.3-fasttrack-3 2026-03-06T13:40:05.082 INFO:journalctl@ceph.mgr.a.vm00.stdout:Mar 06 13:40:04 vm00 systemd[1]: Started Ceph mgr.a for 82f504c8-1959-11f1-9a16-a78d135e200b. 2026-03-06T13:40:05.262 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout 2026-03-06T13:40:05.262 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout { 2026-03-06T13:40:05.262 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "fsid": "82f504c8-1959-11f1-9a16-a78d135e200b", 2026-03-06T13:40:05.262 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "health": { 2026-03-06T13:40:05.262 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "status": "HEALTH_OK", 2026-03-06T13:40:05.262 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "checks": {}, 2026-03-06T13:40:05.262 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "mutes": [] 2026-03-06T13:40:05.262 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout }, 2026-03-06T13:40:05.262 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "election_epoch": 5, 2026-03-06T13:40:05.262 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "quorum": [ 2026-03-06T13:40:05.262 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout 0 2026-03-06T13:40:05.262 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout ], 2026-03-06T13:40:05.262 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "quorum_names": [ 2026-03-06T13:40:05.262 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "a" 2026-03-06T13:40:05.262 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout ], 2026-03-06T13:40:05.262 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "quorum_age": 0, 2026-03-06T13:40:05.262 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "monmap": { 2026-03-06T13:40:05.262 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "epoch": 1, 2026-03-06T13:40:05.262 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "min_mon_release_name": "squid", 2026-03-06T13:40:05.262 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "num_mons": 1 2026-03-06T13:40:05.262 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout }, 2026-03-06T13:40:05.262 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "osdmap": { 2026-03-06T13:40:05.262 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "epoch": 1, 2026-03-06T13:40:05.262 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "num_osds": 0, 2026-03-06T13:40:05.262 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "num_up_osds": 0, 2026-03-06T13:40:05.262 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "osd_up_since": 0, 2026-03-06T13:40:05.262 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "num_in_osds": 0, 2026-03-06T13:40:05.262 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "osd_in_since": 0, 2026-03-06T13:40:05.262 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "num_remapped_pgs": 0 2026-03-06T13:40:05.263 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout }, 2026-03-06T13:40:05.263 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "pgmap": { 2026-03-06T13:40:05.263 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "pgs_by_state": [], 2026-03-06T13:40:05.263 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "num_pgs": 0, 2026-03-06T13:40:05.263 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "num_pools": 0, 2026-03-06T13:40:05.263 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "num_objects": 0, 2026-03-06T13:40:05.263 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "data_bytes": 0, 2026-03-06T13:40:05.263 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "bytes_used": 0, 2026-03-06T13:40:05.263 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "bytes_avail": 0, 2026-03-06T13:40:05.263 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "bytes_total": 0 2026-03-06T13:40:05.263 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout }, 2026-03-06T13:40:05.263 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "fsmap": { 2026-03-06T13:40:05.263 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "epoch": 1, 2026-03-06T13:40:05.263 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "btime": "2026-03-06T12:40:03:177837+0000", 2026-03-06T13:40:05.263 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "by_rank": [], 2026-03-06T13:40:05.263 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "up:standby": 0 2026-03-06T13:40:05.263 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout }, 2026-03-06T13:40:05.263 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "mgrmap": { 2026-03-06T13:40:05.263 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "available": false, 2026-03-06T13:40:05.263 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "num_standbys": 0, 2026-03-06T13:40:05.263 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "modules": [ 2026-03-06T13:40:05.263 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "iostat", 2026-03-06T13:40:05.263 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "nfs", 2026-03-06T13:40:05.263 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "restful" 2026-03-06T13:40:05.263 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout ], 2026-03-06T13:40:05.263 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "services": {} 2026-03-06T13:40:05.263 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout }, 2026-03-06T13:40:05.263 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "servicemap": { 2026-03-06T13:40:05.263 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "epoch": 1, 2026-03-06T13:40:05.263 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "modified": "2026-03-06T12:40:03.178379+0000", 2026-03-06T13:40:05.263 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "services": {} 2026-03-06T13:40:05.263 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout }, 2026-03-06T13:40:05.263 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "progress_events": {} 2026-03-06T13:40:05.263 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout } 2026-03-06T13:40:05.263 INFO:teuthology.orchestra.run.vm00.stdout:mgr not available, waiting (1/15)... 2026-03-06T13:40:05.332 INFO:journalctl@ceph.mgr.a.vm00.stdout:Mar 06 13:40:05 vm00 ceph-82f504c8-1959-11f1-9a16-a78d135e200b-mgr-a[52124]: 2026-03-06T12:40:05.187+0000 7f945cbd5100 -1 mgr[py] Module alerts has missing NOTIFY_TYPES member 2026-03-06T13:40:05.752 INFO:journalctl@ceph.mgr.a.vm00.stdout:Mar 06 13:40:05 vm00 ceph-82f504c8-1959-11f1-9a16-a78d135e200b-mgr-a[52124]: 2026-03-06T12:40:05.329+0000 7f945cbd5100 -1 mgr[py] Module balancer has missing NOTIFY_TYPES member 2026-03-06T13:40:05.752 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:40:05 vm00 ceph-mon[51914]: from='client.? 192.168.123.100:0/1903254732' entity='client.admin' 2026-03-06T13:40:05.752 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:40:05 vm00 ceph-mon[51914]: from='client.? 192.168.123.100:0/959624759' entity='client.admin' cmd=[{"prefix": "status", "format": "json-pretty"}]: dispatch 2026-03-06T13:40:06.752 INFO:journalctl@ceph.mgr.a.vm00.stdout:Mar 06 13:40:06 vm00 ceph-82f504c8-1959-11f1-9a16-a78d135e200b-mgr-a[52124]: 2026-03-06T12:40:06.397+0000 7f945cbd5100 -1 mgr[py] Module crash has missing NOTIFY_TYPES member 2026-03-06T13:40:07.502 INFO:journalctl@ceph.mgr.a.vm00.stdout:Mar 06 13:40:07 vm00 ceph-82f504c8-1959-11f1-9a16-a78d135e200b-mgr-a[52124]: 2026-03-06T12:40:07.210+0000 7f945cbd5100 -1 mgr[py] Module devicehealth has missing NOTIFY_TYPES member 2026-03-06T13:40:07.503 INFO:journalctl@ceph.mgr.a.vm00.stdout:Mar 06 13:40:07 vm00 ceph-82f504c8-1959-11f1-9a16-a78d135e200b-mgr-a[52124]: 2026-03-06T12:40:07.346+0000 7f945cbd5100 -1 mgr[py] Module influx has missing NOTIFY_TYPES member 2026-03-06T13:40:07.565 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout 2026-03-06T13:40:07.565 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout { 2026-03-06T13:40:07.565 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "fsid": "82f504c8-1959-11f1-9a16-a78d135e200b", 2026-03-06T13:40:07.565 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "health": { 2026-03-06T13:40:07.565 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "status": "HEALTH_OK", 2026-03-06T13:40:07.565 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "checks": {}, 2026-03-06T13:40:07.565 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "mutes": [] 2026-03-06T13:40:07.565 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout }, 2026-03-06T13:40:07.565 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "election_epoch": 5, 2026-03-06T13:40:07.565 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "quorum": [ 2026-03-06T13:40:07.565 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout 0 2026-03-06T13:40:07.565 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout ], 2026-03-06T13:40:07.565 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "quorum_names": [ 2026-03-06T13:40:07.565 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "a" 2026-03-06T13:40:07.565 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout ], 2026-03-06T13:40:07.565 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "quorum_age": 3, 2026-03-06T13:40:07.565 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "monmap": { 2026-03-06T13:40:07.565 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "epoch": 1, 2026-03-06T13:40:07.565 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "min_mon_release_name": "squid", 2026-03-06T13:40:07.565 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "num_mons": 1 2026-03-06T13:40:07.565 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout }, 2026-03-06T13:40:07.565 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "osdmap": { 2026-03-06T13:40:07.565 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "epoch": 1, 2026-03-06T13:40:07.565 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "num_osds": 0, 2026-03-06T13:40:07.565 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "num_up_osds": 0, 2026-03-06T13:40:07.565 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "osd_up_since": 0, 2026-03-06T13:40:07.565 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "num_in_osds": 0, 2026-03-06T13:40:07.565 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "osd_in_since": 0, 2026-03-06T13:40:07.565 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "num_remapped_pgs": 0 2026-03-06T13:40:07.566 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout }, 2026-03-06T13:40:07.566 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "pgmap": { 2026-03-06T13:40:07.566 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "pgs_by_state": [], 2026-03-06T13:40:07.566 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "num_pgs": 0, 2026-03-06T13:40:07.566 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "num_pools": 0, 2026-03-06T13:40:07.566 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "num_objects": 0, 2026-03-06T13:40:07.566 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "data_bytes": 0, 2026-03-06T13:40:07.566 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "bytes_used": 0, 2026-03-06T13:40:07.566 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "bytes_avail": 0, 2026-03-06T13:40:07.566 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "bytes_total": 0 2026-03-06T13:40:07.566 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout }, 2026-03-06T13:40:07.566 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "fsmap": { 2026-03-06T13:40:07.566 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "epoch": 1, 2026-03-06T13:40:07.566 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "btime": "2026-03-06T12:40:03:177837+0000", 2026-03-06T13:40:07.566 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "by_rank": [], 2026-03-06T13:40:07.566 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "up:standby": 0 2026-03-06T13:40:07.566 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout }, 2026-03-06T13:40:07.566 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "mgrmap": { 2026-03-06T13:40:07.566 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "available": false, 2026-03-06T13:40:07.566 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "num_standbys": 0, 2026-03-06T13:40:07.566 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "modules": [ 2026-03-06T13:40:07.566 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "iostat", 2026-03-06T13:40:07.566 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "nfs", 2026-03-06T13:40:07.566 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "restful" 2026-03-06T13:40:07.566 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout ], 2026-03-06T13:40:07.566 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "services": {} 2026-03-06T13:40:07.566 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout }, 2026-03-06T13:40:07.566 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "servicemap": { 2026-03-06T13:40:07.566 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "epoch": 1, 2026-03-06T13:40:07.566 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "modified": "2026-03-06T12:40:03.178379+0000", 2026-03-06T13:40:07.566 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "services": {} 2026-03-06T13:40:07.566 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout }, 2026-03-06T13:40:07.566 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "progress_events": {} 2026-03-06T13:40:07.566 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout } 2026-03-06T13:40:07.567 INFO:teuthology.orchestra.run.vm00.stdout:mgr not available, waiting (2/15)... 2026-03-06T13:40:08.002 INFO:journalctl@ceph.mgr.a.vm00.stdout:Mar 06 13:40:07 vm00 ceph-82f504c8-1959-11f1-9a16-a78d135e200b-mgr-a[52124]: 2026-03-06T12:40:07.599+0000 7f945cbd5100 -1 mgr[py] Module iostat has missing NOTIFY_TYPES member 2026-03-06T13:40:08.003 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:40:07 vm00 ceph-mon[51914]: from='client.? 192.168.123.100:0/2371423800' entity='client.admin' cmd=[{"prefix": "status", "format": "json-pretty"}]: dispatch 2026-03-06T13:40:09.611 INFO:journalctl@ceph.mgr.a.vm00.stdout:Mar 06 13:40:09 vm00 ceph-82f504c8-1959-11f1-9a16-a78d135e200b-mgr-a[52124]: 2026-03-06T12:40:09.354+0000 7f945cbd5100 -1 mgr[py] Module nfs has missing NOTIFY_TYPES member 2026-03-06T13:40:09.866 INFO:journalctl@ceph.mgr.a.vm00.stdout:Mar 06 13:40:09 vm00 ceph-82f504c8-1959-11f1-9a16-a78d135e200b-mgr-a[52124]: 2026-03-06T12:40:09.692+0000 7f945cbd5100 -1 mgr[py] Module orchestrator has missing NOTIFY_TYPES member 2026-03-06T13:40:09.866 INFO:journalctl@ceph.mgr.a.vm00.stdout:Mar 06 13:40:09 vm00 ceph-82f504c8-1959-11f1-9a16-a78d135e200b-mgr-a[52124]: 2026-03-06T12:40:09.829+0000 7f945cbd5100 -1 mgr[py] Module osd_perf_query has missing NOTIFY_TYPES member 2026-03-06T13:40:09.878 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout 2026-03-06T13:40:09.878 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout { 2026-03-06T13:40:09.878 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "fsid": "82f504c8-1959-11f1-9a16-a78d135e200b", 2026-03-06T13:40:09.878 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "health": { 2026-03-06T13:40:09.878 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "status": "HEALTH_OK", 2026-03-06T13:40:09.878 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "checks": {}, 2026-03-06T13:40:09.878 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "mutes": [] 2026-03-06T13:40:09.878 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout }, 2026-03-06T13:40:09.878 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "election_epoch": 5, 2026-03-06T13:40:09.878 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "quorum": [ 2026-03-06T13:40:09.878 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout 0 2026-03-06T13:40:09.878 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout ], 2026-03-06T13:40:09.878 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "quorum_names": [ 2026-03-06T13:40:09.878 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "a" 2026-03-06T13:40:09.878 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout ], 2026-03-06T13:40:09.878 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "quorum_age": 5, 2026-03-06T13:40:09.878 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "monmap": { 2026-03-06T13:40:09.878 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "epoch": 1, 2026-03-06T13:40:09.878 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "min_mon_release_name": "squid", 2026-03-06T13:40:09.878 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "num_mons": 1 2026-03-06T13:40:09.878 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout }, 2026-03-06T13:40:09.878 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "osdmap": { 2026-03-06T13:40:09.878 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "epoch": 1, 2026-03-06T13:40:09.878 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "num_osds": 0, 2026-03-06T13:40:09.878 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "num_up_osds": 0, 2026-03-06T13:40:09.878 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "osd_up_since": 0, 2026-03-06T13:40:09.878 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "num_in_osds": 0, 2026-03-06T13:40:09.878 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "osd_in_since": 0, 2026-03-06T13:40:09.878 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "num_remapped_pgs": 0 2026-03-06T13:40:09.878 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout }, 2026-03-06T13:40:09.879 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "pgmap": { 2026-03-06T13:40:09.879 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "pgs_by_state": [], 2026-03-06T13:40:09.879 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "num_pgs": 0, 2026-03-06T13:40:09.879 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "num_pools": 0, 2026-03-06T13:40:09.879 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "num_objects": 0, 2026-03-06T13:40:09.879 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "data_bytes": 0, 2026-03-06T13:40:09.879 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "bytes_used": 0, 2026-03-06T13:40:09.879 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "bytes_avail": 0, 2026-03-06T13:40:09.879 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "bytes_total": 0 2026-03-06T13:40:09.879 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout }, 2026-03-06T13:40:09.879 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "fsmap": { 2026-03-06T13:40:09.879 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "epoch": 1, 2026-03-06T13:40:09.879 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "btime": "2026-03-06T12:40:03:177837+0000", 2026-03-06T13:40:09.879 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "by_rank": [], 2026-03-06T13:40:09.879 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "up:standby": 0 2026-03-06T13:40:09.879 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout }, 2026-03-06T13:40:09.879 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "mgrmap": { 2026-03-06T13:40:09.879 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "available": false, 2026-03-06T13:40:09.879 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "num_standbys": 0, 2026-03-06T13:40:09.879 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "modules": [ 2026-03-06T13:40:09.879 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "iostat", 2026-03-06T13:40:09.879 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "nfs", 2026-03-06T13:40:09.879 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "restful" 2026-03-06T13:40:09.880 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout ], 2026-03-06T13:40:09.880 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "services": {} 2026-03-06T13:40:09.880 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout }, 2026-03-06T13:40:09.880 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "servicemap": { 2026-03-06T13:40:09.880 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "epoch": 1, 2026-03-06T13:40:09.880 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "modified": "2026-03-06T12:40:03.178379+0000", 2026-03-06T13:40:09.880 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "services": {} 2026-03-06T13:40:09.880 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout }, 2026-03-06T13:40:09.880 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "progress_events": {} 2026-03-06T13:40:09.880 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout } 2026-03-06T13:40:09.880 INFO:teuthology.orchestra.run.vm00.stdout:mgr not available, waiting (3/15)... 2026-03-06T13:40:10.241 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:40:09 vm00 ceph-mon[51914]: from='client.? 192.168.123.100:0/1967758096' entity='client.admin' cmd=[{"prefix": "status", "format": "json-pretty"}]: dispatch 2026-03-06T13:40:10.242 INFO:journalctl@ceph.mgr.a.vm00.stdout:Mar 06 13:40:09 vm00 ceph-82f504c8-1959-11f1-9a16-a78d135e200b-mgr-a[52124]: 2026-03-06T12:40:09.965+0000 7f945cbd5100 -1 mgr[py] Module osd_support has missing NOTIFY_TYPES member 2026-03-06T13:40:10.242 INFO:journalctl@ceph.mgr.a.vm00.stdout:Mar 06 13:40:10 vm00 ceph-82f504c8-1959-11f1-9a16-a78d135e200b-mgr-a[52124]: 2026-03-06T12:40:10.105+0000 7f945cbd5100 -1 mgr[py] Module pg_autoscaler has missing NOTIFY_TYPES member 2026-03-06T13:40:10.242 INFO:journalctl@ceph.mgr.a.vm00.stdout:Mar 06 13:40:10 vm00 ceph-82f504c8-1959-11f1-9a16-a78d135e200b-mgr-a[52124]: 2026-03-06T12:40:10.239+0000 7f945cbd5100 -1 mgr[py] Module progress has missing NOTIFY_TYPES member 2026-03-06T13:40:11.002 INFO:journalctl@ceph.mgr.a.vm00.stdout:Mar 06 13:40:10 vm00 ceph-82f504c8-1959-11f1-9a16-a78d135e200b-mgr-a[52124]: 2026-03-06T12:40:10.732+0000 7f945cbd5100 -1 mgr[py] Module prometheus has missing NOTIFY_TYPES member 2026-03-06T13:40:11.002 INFO:journalctl@ceph.mgr.a.vm00.stdout:Mar 06 13:40:10 vm00 ceph-82f504c8-1959-11f1-9a16-a78d135e200b-mgr-a[52124]: 2026-03-06T12:40:10.877+0000 7f945cbd5100 -1 mgr[py] Module rbd_support has missing NOTIFY_TYPES member 2026-03-06T13:40:11.912 INFO:journalctl@ceph.mgr.a.vm00.stdout:Mar 06 13:40:11 vm00 ceph-82f504c8-1959-11f1-9a16-a78d135e200b-mgr-a[52124]: 2026-03-06T12:40:11.577+0000 7f945cbd5100 -1 mgr[py] Module rgw has missing NOTIFY_TYPES member 2026-03-06T13:40:12.197 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout 2026-03-06T13:40:12.197 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout { 2026-03-06T13:40:12.197 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "fsid": "82f504c8-1959-11f1-9a16-a78d135e200b", 2026-03-06T13:40:12.197 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "health": { 2026-03-06T13:40:12.197 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "status": "HEALTH_OK", 2026-03-06T13:40:12.197 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "checks": {}, 2026-03-06T13:40:12.197 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "mutes": [] 2026-03-06T13:40:12.197 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout }, 2026-03-06T13:40:12.197 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "election_epoch": 5, 2026-03-06T13:40:12.197 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "quorum": [ 2026-03-06T13:40:12.197 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout 0 2026-03-06T13:40:12.197 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout ], 2026-03-06T13:40:12.197 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "quorum_names": [ 2026-03-06T13:40:12.197 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "a" 2026-03-06T13:40:12.197 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout ], 2026-03-06T13:40:12.197 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "quorum_age": 7, 2026-03-06T13:40:12.197 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "monmap": { 2026-03-06T13:40:12.197 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "epoch": 1, 2026-03-06T13:40:12.197 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "min_mon_release_name": "squid", 2026-03-06T13:40:12.197 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "num_mons": 1 2026-03-06T13:40:12.197 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout }, 2026-03-06T13:40:12.197 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "osdmap": { 2026-03-06T13:40:12.197 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "epoch": 1, 2026-03-06T13:40:12.197 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "num_osds": 0, 2026-03-06T13:40:12.198 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "num_up_osds": 0, 2026-03-06T13:40:12.198 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "osd_up_since": 0, 2026-03-06T13:40:12.198 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "num_in_osds": 0, 2026-03-06T13:40:12.198 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "osd_in_since": 0, 2026-03-06T13:40:12.198 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "num_remapped_pgs": 0 2026-03-06T13:40:12.198 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout }, 2026-03-06T13:40:12.198 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "pgmap": { 2026-03-06T13:40:12.198 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "pgs_by_state": [], 2026-03-06T13:40:12.198 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "num_pgs": 0, 2026-03-06T13:40:12.198 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "num_pools": 0, 2026-03-06T13:40:12.198 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "num_objects": 0, 2026-03-06T13:40:12.198 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "data_bytes": 0, 2026-03-06T13:40:12.198 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "bytes_used": 0, 2026-03-06T13:40:12.198 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "bytes_avail": 0, 2026-03-06T13:40:12.198 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "bytes_total": 0 2026-03-06T13:40:12.198 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout }, 2026-03-06T13:40:12.198 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "fsmap": { 2026-03-06T13:40:12.198 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "epoch": 1, 2026-03-06T13:40:12.198 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "btime": "2026-03-06T12:40:03:177837+0000", 2026-03-06T13:40:12.198 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "by_rank": [], 2026-03-06T13:40:12.198 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "up:standby": 0 2026-03-06T13:40:12.198 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout }, 2026-03-06T13:40:12.198 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "mgrmap": { 2026-03-06T13:40:12.198 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "available": false, 2026-03-06T13:40:12.198 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "num_standbys": 0, 2026-03-06T13:40:12.198 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "modules": [ 2026-03-06T13:40:12.198 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "iostat", 2026-03-06T13:40:12.198 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "nfs", 2026-03-06T13:40:12.198 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "restful" 2026-03-06T13:40:12.198 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout ], 2026-03-06T13:40:12.198 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "services": {} 2026-03-06T13:40:12.198 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout }, 2026-03-06T13:40:12.198 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "servicemap": { 2026-03-06T13:40:12.198 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "epoch": 1, 2026-03-06T13:40:12.198 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "modified": "2026-03-06T12:40:03.178379+0000", 2026-03-06T13:40:12.198 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "services": {} 2026-03-06T13:40:12.198 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout }, 2026-03-06T13:40:12.198 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "progress_events": {} 2026-03-06T13:40:12.198 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout } 2026-03-06T13:40:12.198 INFO:teuthology.orchestra.run.vm00.stdout:mgr not available, waiting (4/15)... 2026-03-06T13:40:12.502 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:40:12 vm00 ceph-mon[51914]: from='client.? 192.168.123.100:0/75567530' entity='client.admin' cmd=[{"prefix": "status", "format": "json-pretty"}]: dispatch 2026-03-06T13:40:13.002 INFO:journalctl@ceph.mgr.a.vm00.stdout:Mar 06 13:40:12 vm00 ceph-82f504c8-1959-11f1-9a16-a78d135e200b-mgr-a[52124]: 2026-03-06T12:40:12.660+0000 7f945cbd5100 -1 mgr[py] Module rook has missing NOTIFY_TYPES member 2026-03-06T13:40:13.002 INFO:journalctl@ceph.mgr.a.vm00.stdout:Mar 06 13:40:12 vm00 ceph-82f504c8-1959-11f1-9a16-a78d135e200b-mgr-a[52124]: 2026-03-06T12:40:12.782+0000 7f945cbd5100 -1 mgr[py] Module selftest has missing NOTIFY_TYPES member 2026-03-06T13:40:13.002 INFO:journalctl@ceph.mgr.a.vm00.stdout:Mar 06 13:40:12 vm00 ceph-82f504c8-1959-11f1-9a16-a78d135e200b-mgr-a[52124]: 2026-03-06T12:40:12.907+0000 7f945cbd5100 -1 mgr[py] Module snap_schedule has missing NOTIFY_TYPES member 2026-03-06T13:40:13.502 INFO:journalctl@ceph.mgr.a.vm00.stdout:Mar 06 13:40:13 vm00 ceph-82f504c8-1959-11f1-9a16-a78d135e200b-mgr-a[52124]: 2026-03-06T12:40:13.164+0000 7f945cbd5100 -1 mgr[py] Module status has missing NOTIFY_TYPES member 2026-03-06T13:40:13.502 INFO:journalctl@ceph.mgr.a.vm00.stdout:Mar 06 13:40:13 vm00 ceph-82f504c8-1959-11f1-9a16-a78d135e200b-mgr-a[52124]: 2026-03-06T12:40:13.283+0000 7f945cbd5100 -1 mgr[py] Module telegraf has missing NOTIFY_TYPES member 2026-03-06T13:40:13.906 INFO:journalctl@ceph.mgr.a.vm00.stdout:Mar 06 13:40:13 vm00 ceph-82f504c8-1959-11f1-9a16-a78d135e200b-mgr-a[52124]: 2026-03-06T12:40:13.576+0000 7f945cbd5100 -1 mgr[py] Module telemetry has missing NOTIFY_TYPES member 2026-03-06T13:40:14.222 INFO:journalctl@ceph.mgr.a.vm00.stdout:Mar 06 13:40:13 vm00 ceph-82f504c8-1959-11f1-9a16-a78d135e200b-mgr-a[52124]: 2026-03-06T12:40:13.905+0000 7f945cbd5100 -1 mgr[py] Module test_orchestrator has missing NOTIFY_TYPES member 2026-03-06T13:40:14.480 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:40:14 vm00 ceph-mon[51914]: Activating manager daemon a 2026-03-06T13:40:14.480 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:40:14 vm00 ceph-mon[51914]: mgrmap e2: a(active, starting, since 0.00415646s) 2026-03-06T13:40:14.480 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:40:14 vm00 ceph-mon[51914]: from='mgr.14100 192.168.123.100:0/2125900154' entity='mgr.a' cmd=[{"prefix": "mds metadata"}]: dispatch 2026-03-06T13:40:14.480 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:40:14 vm00 ceph-mon[51914]: from='mgr.14100 192.168.123.100:0/2125900154' entity='mgr.a' cmd=[{"prefix": "osd metadata"}]: dispatch 2026-03-06T13:40:14.480 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:40:14 vm00 ceph-mon[51914]: from='mgr.14100 192.168.123.100:0/2125900154' entity='mgr.a' cmd=[{"prefix": "mon metadata"}]: dispatch 2026-03-06T13:40:14.480 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:40:14 vm00 ceph-mon[51914]: from='mgr.14100 192.168.123.100:0/2125900154' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "a"}]: dispatch 2026-03-06T13:40:14.481 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:40:14 vm00 ceph-mon[51914]: from='mgr.14100 192.168.123.100:0/2125900154' entity='mgr.a' cmd=[{"prefix": "mgr metadata", "who": "a", "id": "a"}]: dispatch 2026-03-06T13:40:14.481 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:40:14 vm00 ceph-mon[51914]: Manager daemon a is now available 2026-03-06T13:40:14.481 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:40:14 vm00 ceph-mon[51914]: from='mgr.14100 192.168.123.100:0/2125900154' entity='mgr.a' 2026-03-06T13:40:14.481 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:40:14 vm00 ceph-mon[51914]: from='mgr.14100 192.168.123.100:0/2125900154' entity='mgr.a' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/a/mirror_snapshot_schedule"}]: dispatch 2026-03-06T13:40:14.481 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:40:14 vm00 ceph-mon[51914]: from='mgr.14100 192.168.123.100:0/2125900154' entity='mgr.a' 2026-03-06T13:40:14.481 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:40:14 vm00 ceph-mon[51914]: from='mgr.14100 192.168.123.100:0/2125900154' entity='mgr.a' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/a/trash_purge_schedule"}]: dispatch 2026-03-06T13:40:14.481 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:40:14 vm00 ceph-mon[51914]: from='mgr.14100 192.168.123.100:0/2125900154' entity='mgr.a' 2026-03-06T13:40:14.481 INFO:journalctl@ceph.mgr.a.vm00.stdout:Mar 06 13:40:14 vm00 ceph-82f504c8-1959-11f1-9a16-a78d135e200b-mgr-a[52124]: 2026-03-06T12:40:14.258+0000 7f945cbd5100 -1 mgr[py] Module volumes has missing NOTIFY_TYPES member 2026-03-06T13:40:14.481 INFO:journalctl@ceph.mgr.a.vm00.stdout:Mar 06 13:40:14 vm00 ceph-82f504c8-1959-11f1-9a16-a78d135e200b-mgr-a[52124]: 2026-03-06T12:40:14.378+0000 7f945cbd5100 -1 mgr[py] Module zabbix has missing NOTIFY_TYPES member 2026-03-06T13:40:14.490 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout 2026-03-06T13:40:14.490 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout { 2026-03-06T13:40:14.490 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "fsid": "82f504c8-1959-11f1-9a16-a78d135e200b", 2026-03-06T13:40:14.490 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "health": { 2026-03-06T13:40:14.490 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "status": "HEALTH_OK", 2026-03-06T13:40:14.490 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "checks": {}, 2026-03-06T13:40:14.490 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "mutes": [] 2026-03-06T13:40:14.490 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout }, 2026-03-06T13:40:14.491 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "election_epoch": 5, 2026-03-06T13:40:14.491 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "quorum": [ 2026-03-06T13:40:14.491 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout 0 2026-03-06T13:40:14.491 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout ], 2026-03-06T13:40:14.491 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "quorum_names": [ 2026-03-06T13:40:14.491 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "a" 2026-03-06T13:40:14.491 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout ], 2026-03-06T13:40:14.491 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "quorum_age": 10, 2026-03-06T13:40:14.491 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "monmap": { 2026-03-06T13:40:14.491 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "epoch": 1, 2026-03-06T13:40:14.491 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "min_mon_release_name": "squid", 2026-03-06T13:40:14.491 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "num_mons": 1 2026-03-06T13:40:14.491 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout }, 2026-03-06T13:40:14.491 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "osdmap": { 2026-03-06T13:40:14.491 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "epoch": 1, 2026-03-06T13:40:14.491 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "num_osds": 0, 2026-03-06T13:40:14.491 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "num_up_osds": 0, 2026-03-06T13:40:14.491 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "osd_up_since": 0, 2026-03-06T13:40:14.491 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "num_in_osds": 0, 2026-03-06T13:40:14.491 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "osd_in_since": 0, 2026-03-06T13:40:14.491 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "num_remapped_pgs": 0 2026-03-06T13:40:14.492 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout }, 2026-03-06T13:40:14.492 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "pgmap": { 2026-03-06T13:40:14.492 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "pgs_by_state": [], 2026-03-06T13:40:14.492 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "num_pgs": 0, 2026-03-06T13:40:14.492 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "num_pools": 0, 2026-03-06T13:40:14.492 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "num_objects": 0, 2026-03-06T13:40:14.492 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "data_bytes": 0, 2026-03-06T13:40:14.492 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "bytes_used": 0, 2026-03-06T13:40:14.492 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "bytes_avail": 0, 2026-03-06T13:40:14.492 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "bytes_total": 0 2026-03-06T13:40:14.492 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout }, 2026-03-06T13:40:14.492 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "fsmap": { 2026-03-06T13:40:14.492 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "epoch": 1, 2026-03-06T13:40:14.492 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "btime": "2026-03-06T12:40:03:177837+0000", 2026-03-06T13:40:14.492 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "by_rank": [], 2026-03-06T13:40:14.492 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "up:standby": 0 2026-03-06T13:40:14.492 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout }, 2026-03-06T13:40:14.492 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "mgrmap": { 2026-03-06T13:40:14.492 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "available": false, 2026-03-06T13:40:14.492 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "num_standbys": 0, 2026-03-06T13:40:14.492 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "modules": [ 2026-03-06T13:40:14.492 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "iostat", 2026-03-06T13:40:14.492 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "nfs", 2026-03-06T13:40:14.493 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "restful" 2026-03-06T13:40:14.493 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout ], 2026-03-06T13:40:14.493 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "services": {} 2026-03-06T13:40:14.493 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout }, 2026-03-06T13:40:14.493 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "servicemap": { 2026-03-06T13:40:14.493 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "epoch": 1, 2026-03-06T13:40:14.493 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "modified": "2026-03-06T12:40:03.178379+0000", 2026-03-06T13:40:14.493 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "services": {} 2026-03-06T13:40:14.493 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout }, 2026-03-06T13:40:14.493 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "progress_events": {} 2026-03-06T13:40:14.493 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout } 2026-03-06T13:40:14.493 INFO:teuthology.orchestra.run.vm00.stdout:mgr not available, waiting (5/15)... 2026-03-06T13:40:15.752 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:40:15 vm00 ceph-mon[51914]: from='client.? 192.168.123.100:0/2015628550' entity='client.admin' cmd=[{"prefix": "status", "format": "json-pretty"}]: dispatch 2026-03-06T13:40:15.752 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:40:15 vm00 ceph-mon[51914]: mgrmap e3: a(active, since 1.01043s) 2026-03-06T13:40:17.439 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:40:17 vm00 ceph-mon[51914]: mgrmap e4: a(active, since 2s) 2026-03-06T13:40:17.439 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:40:17 vm00 ceph-mon[51914]: from='client.? 192.168.123.100:0/2748542568' entity='client.admin' cmd=[{"prefix": "status", "format": "json-pretty"}]: dispatch 2026-03-06T13:40:17.450 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout 2026-03-06T13:40:17.450 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout { 2026-03-06T13:40:17.450 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "fsid": "82f504c8-1959-11f1-9a16-a78d135e200b", 2026-03-06T13:40:17.450 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "health": { 2026-03-06T13:40:17.450 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "status": "HEALTH_OK", 2026-03-06T13:40:17.451 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "checks": {}, 2026-03-06T13:40:17.451 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "mutes": [] 2026-03-06T13:40:17.451 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout }, 2026-03-06T13:40:17.451 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "election_epoch": 5, 2026-03-06T13:40:17.451 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "quorum": [ 2026-03-06T13:40:17.451 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout 0 2026-03-06T13:40:17.451 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout ], 2026-03-06T13:40:17.451 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "quorum_names": [ 2026-03-06T13:40:17.451 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "a" 2026-03-06T13:40:17.451 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout ], 2026-03-06T13:40:17.451 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "quorum_age": 12, 2026-03-06T13:40:17.451 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "monmap": { 2026-03-06T13:40:17.451 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "epoch": 1, 2026-03-06T13:40:17.451 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "min_mon_release_name": "squid", 2026-03-06T13:40:17.451 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "num_mons": 1 2026-03-06T13:40:17.451 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout }, 2026-03-06T13:40:17.451 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "osdmap": { 2026-03-06T13:40:17.451 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "epoch": 1, 2026-03-06T13:40:17.451 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "num_osds": 0, 2026-03-06T13:40:17.451 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "num_up_osds": 0, 2026-03-06T13:40:17.451 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "osd_up_since": 0, 2026-03-06T13:40:17.451 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "num_in_osds": 0, 2026-03-06T13:40:17.451 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "osd_in_since": 0, 2026-03-06T13:40:17.451 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "num_remapped_pgs": 0 2026-03-06T13:40:17.452 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout }, 2026-03-06T13:40:17.452 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "pgmap": { 2026-03-06T13:40:17.452 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "pgs_by_state": [], 2026-03-06T13:40:17.452 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "num_pgs": 0, 2026-03-06T13:40:17.452 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "num_pools": 0, 2026-03-06T13:40:17.452 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "num_objects": 0, 2026-03-06T13:40:17.452 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "data_bytes": 0, 2026-03-06T13:40:17.452 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "bytes_used": 0, 2026-03-06T13:40:17.452 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "bytes_avail": 0, 2026-03-06T13:40:17.452 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "bytes_total": 0 2026-03-06T13:40:17.452 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout }, 2026-03-06T13:40:17.452 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "fsmap": { 2026-03-06T13:40:17.452 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "epoch": 1, 2026-03-06T13:40:17.452 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "btime": "2026-03-06T12:40:03:177837+0000", 2026-03-06T13:40:17.452 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "by_rank": [], 2026-03-06T13:40:17.452 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "up:standby": 0 2026-03-06T13:40:17.452 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout }, 2026-03-06T13:40:17.452 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "mgrmap": { 2026-03-06T13:40:17.452 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "available": true, 2026-03-06T13:40:17.452 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "num_standbys": 0, 2026-03-06T13:40:17.452 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "modules": [ 2026-03-06T13:40:17.452 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "iostat", 2026-03-06T13:40:17.452 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "nfs", 2026-03-06T13:40:17.452 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "restful" 2026-03-06T13:40:17.452 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout ], 2026-03-06T13:40:17.452 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "services": {} 2026-03-06T13:40:17.452 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout }, 2026-03-06T13:40:17.452 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "servicemap": { 2026-03-06T13:40:17.452 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "epoch": 1, 2026-03-06T13:40:17.452 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "modified": "2026-03-06T12:40:03.178379+0000", 2026-03-06T13:40:17.452 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "services": {} 2026-03-06T13:40:17.452 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout }, 2026-03-06T13:40:17.452 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "progress_events": {} 2026-03-06T13:40:17.452 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout } 2026-03-06T13:40:17.452 INFO:teuthology.orchestra.run.vm00.stdout:mgr is available 2026-03-06T13:40:17.807 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout 2026-03-06T13:40:17.807 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout [global] 2026-03-06T13:40:17.807 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout fsid = 82f504c8-1959-11f1-9a16-a78d135e200b 2026-03-06T13:40:17.807 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout mon_cluster_log_file_level = debug 2026-03-06T13:40:17.807 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout mon_host = [v2:192.168.123.100:3300,v1:192.168.123.100:6789] 2026-03-06T13:40:17.807 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout mon_osd_allow_pg_remap = true 2026-03-06T13:40:17.807 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout mon_osd_allow_primary_affinity = true 2026-03-06T13:40:17.807 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout mon_warn_on_no_sortbitwise = false 2026-03-06T13:40:17.807 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout osd_crush_chooseleaf_type = 0 2026-03-06T13:40:17.807 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout 2026-03-06T13:40:17.807 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout [mgr] 2026-03-06T13:40:17.807 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout mgr/telemetry/nag = false 2026-03-06T13:40:17.807 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout 2026-03-06T13:40:17.807 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout [osd] 2026-03-06T13:40:17.807 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout osd_map_max_advance = 10 2026-03-06T13:40:17.807 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout osd_sloppy_crc = true 2026-03-06T13:40:17.807 INFO:teuthology.orchestra.run.vm00.stdout:Enabling cephadm module... 2026-03-06T13:40:18.751 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:40:18 vm00 ceph-mon[51914]: from='client.? 192.168.123.100:0/3092405290' entity='client.admin' cmd=[{"prefix": "config assimilate-conf"}]: dispatch 2026-03-06T13:40:18.752 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:40:18 vm00 ceph-mon[51914]: from='client.? 192.168.123.100:0/3092405290' entity='client.admin' cmd='[{"prefix": "config assimilate-conf"}]': finished 2026-03-06T13:40:18.752 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:40:18 vm00 ceph-mon[51914]: from='client.? 192.168.123.100:0/1516005108' entity='client.admin' cmd=[{"prefix": "mgr module enable", "module": "cephadm"}]: dispatch 2026-03-06T13:40:19.035 INFO:journalctl@ceph.mgr.a.vm00.stdout:Mar 06 13:40:18 vm00 ceph-82f504c8-1959-11f1-9a16-a78d135e200b-mgr-a[52124]: ignoring --setuser ceph since I am not root 2026-03-06T13:40:19.035 INFO:journalctl@ceph.mgr.a.vm00.stdout:Mar 06 13:40:18 vm00 ceph-82f504c8-1959-11f1-9a16-a78d135e200b-mgr-a[52124]: ignoring --setgroup ceph since I am not root 2026-03-06T13:40:19.036 INFO:journalctl@ceph.mgr.a.vm00.stdout:Mar 06 13:40:19 vm00 ceph-82f504c8-1959-11f1-9a16-a78d135e200b-mgr-a[52124]: 2026-03-06T12:40:19.033+0000 7efec1b99100 -1 mgr[py] Module alerts has missing NOTIFY_TYPES member 2026-03-06T13:40:19.163 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout { 2026-03-06T13:40:19.163 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "epoch": 5, 2026-03-06T13:40:19.163 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "available": true, 2026-03-06T13:40:19.163 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "active_name": "a", 2026-03-06T13:40:19.163 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "num_standby": 0 2026-03-06T13:40:19.163 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout } 2026-03-06T13:40:19.163 INFO:teuthology.orchestra.run.vm00.stdout:Waiting for the mgr to restart... 2026-03-06T13:40:19.163 INFO:teuthology.orchestra.run.vm00.stdout:Waiting for mgr epoch 5... 2026-03-06T13:40:19.501 INFO:journalctl@ceph.mgr.a.vm00.stdout:Mar 06 13:40:19 vm00 ceph-82f504c8-1959-11f1-9a16-a78d135e200b-mgr-a[52124]: 2026-03-06T12:40:19.165+0000 7efec1b99100 -1 mgr[py] Module balancer has missing NOTIFY_TYPES member 2026-03-06T13:40:20.252 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:40:19 vm00 ceph-mon[51914]: from='client.? 192.168.123.100:0/1516005108' entity='client.admin' cmd='[{"prefix": "mgr module enable", "module": "cephadm"}]': finished 2026-03-06T13:40:20.252 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:40:19 vm00 ceph-mon[51914]: mgrmap e5: a(active, since 4s) 2026-03-06T13:40:20.252 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:40:19 vm00 ceph-mon[51914]: from='client.? 192.168.123.100:0/943992775' entity='client.admin' cmd=[{"prefix": "mgr stat"}]: dispatch 2026-03-06T13:40:20.752 INFO:journalctl@ceph.mgr.a.vm00.stdout:Mar 06 13:40:20 vm00 ceph-82f504c8-1959-11f1-9a16-a78d135e200b-mgr-a[52124]: 2026-03-06T12:40:20.262+0000 7efec1b99100 -1 mgr[py] Module crash has missing NOTIFY_TYPES member 2026-03-06T13:40:21.436 INFO:journalctl@ceph.mgr.a.vm00.stdout:Mar 06 13:40:21 vm00 ceph-82f504c8-1959-11f1-9a16-a78d135e200b-mgr-a[52124]: 2026-03-06T12:40:21.078+0000 7efec1b99100 -1 mgr[py] Module devicehealth has missing NOTIFY_TYPES member 2026-03-06T13:40:21.436 INFO:journalctl@ceph.mgr.a.vm00.stdout:Mar 06 13:40:21 vm00 ceph-82f504c8-1959-11f1-9a16-a78d135e200b-mgr-a[52124]: 2026-03-06T12:40:21.197+0000 7efec1b99100 -1 mgr[py] Module influx has missing NOTIFY_TYPES member 2026-03-06T13:40:21.752 INFO:journalctl@ceph.mgr.a.vm00.stdout:Mar 06 13:40:21 vm00 ceph-82f504c8-1959-11f1-9a16-a78d135e200b-mgr-a[52124]: 2026-03-06T12:40:21.434+0000 7efec1b99100 -1 mgr[py] Module iostat has missing NOTIFY_TYPES member 2026-03-06T13:40:23.502 INFO:journalctl@ceph.mgr.a.vm00.stdout:Mar 06 13:40:23 vm00 ceph-82f504c8-1959-11f1-9a16-a78d135e200b-mgr-a[52124]: 2026-03-06T12:40:23.181+0000 7efec1b99100 -1 mgr[py] Module nfs has missing NOTIFY_TYPES member 2026-03-06T13:40:23.890 INFO:journalctl@ceph.mgr.a.vm00.stdout:Mar 06 13:40:23 vm00 ceph-82f504c8-1959-11f1-9a16-a78d135e200b-mgr-a[52124]: 2026-03-06T12:40:23.508+0000 7efec1b99100 -1 mgr[py] Module orchestrator has missing NOTIFY_TYPES member 2026-03-06T13:40:23.890 INFO:journalctl@ceph.mgr.a.vm00.stdout:Mar 06 13:40:23 vm00 ceph-82f504c8-1959-11f1-9a16-a78d135e200b-mgr-a[52124]: 2026-03-06T12:40:23.638+0000 7efec1b99100 -1 mgr[py] Module osd_perf_query has missing NOTIFY_TYPES member 2026-03-06T13:40:23.890 INFO:journalctl@ceph.mgr.a.vm00.stdout:Mar 06 13:40:23 vm00 ceph-82f504c8-1959-11f1-9a16-a78d135e200b-mgr-a[52124]: 2026-03-06T12:40:23.754+0000 7efec1b99100 -1 mgr[py] Module osd_support has missing NOTIFY_TYPES member 2026-03-06T13:40:24.252 INFO:journalctl@ceph.mgr.a.vm00.stdout:Mar 06 13:40:23 vm00 ceph-82f504c8-1959-11f1-9a16-a78d135e200b-mgr-a[52124]: 2026-03-06T12:40:23.888+0000 7efec1b99100 -1 mgr[py] Module pg_autoscaler has missing NOTIFY_TYPES member 2026-03-06T13:40:24.252 INFO:journalctl@ceph.mgr.a.vm00.stdout:Mar 06 13:40:24 vm00 ceph-82f504c8-1959-11f1-9a16-a78d135e200b-mgr-a[52124]: 2026-03-06T12:40:24.007+0000 7efec1b99100 -1 mgr[py] Module progress has missing NOTIFY_TYPES member 2026-03-06T13:40:25.002 INFO:journalctl@ceph.mgr.a.vm00.stdout:Mar 06 13:40:24 vm00 ceph-82f504c8-1959-11f1-9a16-a78d135e200b-mgr-a[52124]: 2026-03-06T12:40:24.503+0000 7efec1b99100 -1 mgr[py] Module prometheus has missing NOTIFY_TYPES member 2026-03-06T13:40:25.002 INFO:journalctl@ceph.mgr.a.vm00.stdout:Mar 06 13:40:24 vm00 ceph-82f504c8-1959-11f1-9a16-a78d135e200b-mgr-a[52124]: 2026-03-06T12:40:24.650+0000 7efec1b99100 -1 mgr[py] Module rbd_support has missing NOTIFY_TYPES member 2026-03-06T13:40:25.752 INFO:journalctl@ceph.mgr.a.vm00.stdout:Mar 06 13:40:25 vm00 ceph-82f504c8-1959-11f1-9a16-a78d135e200b-mgr-a[52124]: 2026-03-06T12:40:25.343+0000 7efec1b99100 -1 mgr[py] Module rgw has missing NOTIFY_TYPES member 2026-03-06T13:40:26.752 INFO:journalctl@ceph.mgr.a.vm00.stdout:Mar 06 13:40:26 vm00 ceph-82f504c8-1959-11f1-9a16-a78d135e200b-mgr-a[52124]: 2026-03-06T12:40:26.365+0000 7efec1b99100 -1 mgr[py] Module rook has missing NOTIFY_TYPES member 2026-03-06T13:40:26.752 INFO:journalctl@ceph.mgr.a.vm00.stdout:Mar 06 13:40:26 vm00 ceph-82f504c8-1959-11f1-9a16-a78d135e200b-mgr-a[52124]: 2026-03-06T12:40:26.485+0000 7efec1b99100 -1 mgr[py] Module selftest has missing NOTIFY_TYPES member 2026-03-06T13:40:26.752 INFO:journalctl@ceph.mgr.a.vm00.stdout:Mar 06 13:40:26 vm00 ceph-82f504c8-1959-11f1-9a16-a78d135e200b-mgr-a[52124]: 2026-03-06T12:40:26.608+0000 7efec1b99100 -1 mgr[py] Module snap_schedule has missing NOTIFY_TYPES member 2026-03-06T13:40:27.252 INFO:journalctl@ceph.mgr.a.vm00.stdout:Mar 06 13:40:26 vm00 ceph-82f504c8-1959-11f1-9a16-a78d135e200b-mgr-a[52124]: 2026-03-06T12:40:26.868+0000 7efec1b99100 -1 mgr[py] Module status has missing NOTIFY_TYPES member 2026-03-06T13:40:27.252 INFO:journalctl@ceph.mgr.a.vm00.stdout:Mar 06 13:40:26 vm00 ceph-82f504c8-1959-11f1-9a16-a78d135e200b-mgr-a[52124]: 2026-03-06T12:40:26.989+0000 7efec1b99100 -1 mgr[py] Module telegraf has missing NOTIFY_TYPES member 2026-03-06T13:40:27.636 INFO:journalctl@ceph.mgr.a.vm00.stdout:Mar 06 13:40:27 vm00 ceph-82f504c8-1959-11f1-9a16-a78d135e200b-mgr-a[52124]: 2026-03-06T12:40:27.297+0000 7efec1b99100 -1 mgr[py] Module telemetry has missing NOTIFY_TYPES member 2026-03-06T13:40:28.002 INFO:journalctl@ceph.mgr.a.vm00.stdout:Mar 06 13:40:27 vm00 ceph-82f504c8-1959-11f1-9a16-a78d135e200b-mgr-a[52124]: 2026-03-06T12:40:27.634+0000 7efec1b99100 -1 mgr[py] Module test_orchestrator has missing NOTIFY_TYPES member 2026-03-06T13:40:28.502 INFO:journalctl@ceph.mgr.a.vm00.stdout:Mar 06 13:40:28 vm00 ceph-82f504c8-1959-11f1-9a16-a78d135e200b-mgr-a[52124]: 2026-03-06T12:40:28.006+0000 7efec1b99100 -1 mgr[py] Module volumes has missing NOTIFY_TYPES member 2026-03-06T13:40:28.502 INFO:journalctl@ceph.mgr.a.vm00.stdout:Mar 06 13:40:28 vm00 ceph-82f504c8-1959-11f1-9a16-a78d135e200b-mgr-a[52124]: 2026-03-06T12:40:28.135+0000 7efec1b99100 -1 mgr[py] Module zabbix has missing NOTIFY_TYPES member 2026-03-06T13:40:28.502 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:40:28 vm00 ceph-mon[51914]: Active manager daemon a restarted 2026-03-06T13:40:28.502 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:40:28 vm00 ceph-mon[51914]: Activating manager daemon a 2026-03-06T13:40:28.502 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:40:28 vm00 ceph-mon[51914]: osdmap e2: 0 total, 0 up, 0 in 2026-03-06T13:40:28.502 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:40:28 vm00 ceph-mon[51914]: mgrmap e6: a(active, starting, since 0.00934805s) 2026-03-06T13:40:28.502 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:40:28 vm00 ceph-mon[51914]: from='mgr.14124 192.168.123.100:0/1010685127' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "a"}]: dispatch 2026-03-06T13:40:28.502 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:40:28 vm00 ceph-mon[51914]: from='mgr.14124 192.168.123.100:0/1010685127' entity='mgr.a' cmd=[{"prefix": "mgr metadata", "who": "a", "id": "a"}]: dispatch 2026-03-06T13:40:28.502 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:40:28 vm00 ceph-mon[51914]: from='mgr.14124 192.168.123.100:0/1010685127' entity='mgr.a' cmd=[{"prefix": "mds metadata"}]: dispatch 2026-03-06T13:40:28.502 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:40:28 vm00 ceph-mon[51914]: from='mgr.14124 192.168.123.100:0/1010685127' entity='mgr.a' cmd=[{"prefix": "osd metadata"}]: dispatch 2026-03-06T13:40:28.502 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:40:28 vm00 ceph-mon[51914]: from='mgr.14124 192.168.123.100:0/1010685127' entity='mgr.a' cmd=[{"prefix": "mon metadata"}]: dispatch 2026-03-06T13:40:28.502 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:40:28 vm00 ceph-mon[51914]: Manager daemon a is now available 2026-03-06T13:40:28.502 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:40:28 vm00 ceph-mon[51914]: from='mgr.14124 192.168.123.100:0/1010685127' entity='mgr.a' 2026-03-06T13:40:28.502 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:40:28 vm00 ceph-mon[51914]: from='mgr.14124 192.168.123.100:0/1010685127' entity='mgr.a' 2026-03-06T13:40:29.188 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout { 2026-03-06T13:40:29.188 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "mgrmap_epoch": 7, 2026-03-06T13:40:29.188 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "initialized": true 2026-03-06T13:40:29.188 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout } 2026-03-06T13:40:29.188 INFO:teuthology.orchestra.run.vm00.stdout:mgr epoch 5 is available 2026-03-06T13:40:29.188 INFO:teuthology.orchestra.run.vm00.stdout:Setting orchestrator backend to cephadm... 2026-03-06T13:40:29.502 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:40:29 vm00 ceph-mon[51914]: Found migration_current of "None". Setting to last migration. 2026-03-06T13:40:29.502 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:40:29 vm00 ceph-mon[51914]: from='mgr.14124 192.168.123.100:0/1010685127' entity='mgr.a' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/a/mirror_snapshot_schedule"}]: dispatch 2026-03-06T13:40:29.502 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:40:29 vm00 ceph-mon[51914]: from='mgr.14124 192.168.123.100:0/1010685127' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-06T13:40:29.502 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:40:29 vm00 ceph-mon[51914]: from='mgr.14124 192.168.123.100:0/1010685127' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-06T13:40:29.502 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:40:29 vm00 ceph-mon[51914]: from='mgr.14124 192.168.123.100:0/1010685127' entity='mgr.a' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/a/trash_purge_schedule"}]: dispatch 2026-03-06T13:40:29.502 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:40:29 vm00 ceph-mon[51914]: from='mgr.14124 192.168.123.100:0/1010685127' entity='mgr.a' 2026-03-06T13:40:29.502 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:40:29 vm00 ceph-mon[51914]: from='mgr.14124 192.168.123.100:0/1010685127' entity='mgr.a' 2026-03-06T13:40:29.502 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:40:29 vm00 ceph-mon[51914]: mgrmap e7: a(active, since 1.01139s) 2026-03-06T13:40:30.001 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout value unchanged 2026-03-06T13:40:30.001 INFO:teuthology.orchestra.run.vm00.stdout:Generating ssh key... 2026-03-06T13:40:30.753 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:40:30 vm00 ceph-mon[51914]: from='client.14128 -' entity='client.admin' cmd=[{"prefix": "get_command_descriptions"}]: dispatch 2026-03-06T13:40:30.753 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:40:30 vm00 ceph-mon[51914]: from='client.14128 -' entity='client.admin' cmd=[{"prefix": "mgr_status"}]: dispatch 2026-03-06T13:40:30.753 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:40:30 vm00 ceph-mon[51914]: from='client.14136 -' entity='client.admin' cmd=[{"prefix": "orch set backend", "module_name": "cephadm", "target": ["mon-mgr", ""]}]: dispatch 2026-03-06T13:40:30.753 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:40:30 vm00 ceph-mon[51914]: from='mgr.14124 192.168.123.100:0/1010685127' entity='mgr.a' 2026-03-06T13:40:30.753 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:40:30 vm00 ceph-mon[51914]: from='mgr.14124 192.168.123.100:0/1010685127' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-06T13:40:30.753 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:40:30 vm00 ceph-mon[51914]: from='mgr.14124 192.168.123.100:0/1010685127' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-06T13:40:30.753 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:40:30 vm00 ceph-mon[51914]: from='mgr.14124 192.168.123.100:0/1010685127' entity='mgr.a' 2026-03-06T13:40:30.753 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:40:30 vm00 ceph-mon[51914]: from='mgr.14124 192.168.123.100:0/1010685127' entity='mgr.a' 2026-03-06T13:40:30.753 INFO:journalctl@ceph.mgr.a.vm00.stdout:Mar 06 13:40:30 vm00 ceph-82f504c8-1959-11f1-9a16-a78d135e200b-mgr-a[52124]: Generating public/private rsa key pair. 2026-03-06T13:40:30.753 INFO:journalctl@ceph.mgr.a.vm00.stdout:Mar 06 13:40:30 vm00 ceph-82f504c8-1959-11f1-9a16-a78d135e200b-mgr-a[52124]: Your identification has been saved in /tmp/tmpyoiczwk0/key 2026-03-06T13:40:30.753 INFO:journalctl@ceph.mgr.a.vm00.stdout:Mar 06 13:40:30 vm00 ceph-82f504c8-1959-11f1-9a16-a78d135e200b-mgr-a[52124]: Your public key has been saved in /tmp/tmpyoiczwk0/key.pub 2026-03-06T13:40:30.753 INFO:journalctl@ceph.mgr.a.vm00.stdout:Mar 06 13:40:30 vm00 ceph-82f504c8-1959-11f1-9a16-a78d135e200b-mgr-a[52124]: The key fingerprint is: 2026-03-06T13:40:30.753 INFO:journalctl@ceph.mgr.a.vm00.stdout:Mar 06 13:40:30 vm00 ceph-82f504c8-1959-11f1-9a16-a78d135e200b-mgr-a[52124]: SHA256:4Uq6uRay/BJryf2lePqGWBx1schw14Diyy/q4cqYUYM ceph-82f504c8-1959-11f1-9a16-a78d135e200b 2026-03-06T13:40:30.753 INFO:journalctl@ceph.mgr.a.vm00.stdout:Mar 06 13:40:30 vm00 ceph-82f504c8-1959-11f1-9a16-a78d135e200b-mgr-a[52124]: The key's randomart image is: 2026-03-06T13:40:30.753 INFO:journalctl@ceph.mgr.a.vm00.stdout:Mar 06 13:40:30 vm00 ceph-82f504c8-1959-11f1-9a16-a78d135e200b-mgr-a[52124]: +---[RSA 3072]----+ 2026-03-06T13:40:30.753 INFO:journalctl@ceph.mgr.a.vm00.stdout:Mar 06 13:40:30 vm00 ceph-82f504c8-1959-11f1-9a16-a78d135e200b-mgr-a[52124]: | . ..++ | 2026-03-06T13:40:30.753 INFO:journalctl@ceph.mgr.a.vm00.stdout:Mar 06 13:40:30 vm00 ceph-82f504c8-1959-11f1-9a16-a78d135e200b-mgr-a[52124]: | .+oo... | 2026-03-06T13:40:30.753 INFO:journalctl@ceph.mgr.a.vm00.stdout:Mar 06 13:40:30 vm00 ceph-82f504c8-1959-11f1-9a16-a78d135e200b-mgr-a[52124]: | . oo.o | 2026-03-06T13:40:30.753 INFO:journalctl@ceph.mgr.a.vm00.stdout:Mar 06 13:40:30 vm00 ceph-82f504c8-1959-11f1-9a16-a78d135e200b-mgr-a[52124]: | . o . . | 2026-03-06T13:40:30.753 INFO:journalctl@ceph.mgr.a.vm00.stdout:Mar 06 13:40:30 vm00 ceph-82f504c8-1959-11f1-9a16-a78d135e200b-mgr-a[52124]: |E oo o. S | 2026-03-06T13:40:30.753 INFO:journalctl@ceph.mgr.a.vm00.stdout:Mar 06 13:40:30 vm00 ceph-82f504c8-1959-11f1-9a16-a78d135e200b-mgr-a[52124]: | .o.*o . | 2026-03-06T13:40:30.753 INFO:journalctl@ceph.mgr.a.vm00.stdout:Mar 06 13:40:30 vm00 ceph-82f504c8-1959-11f1-9a16-a78d135e200b-mgr-a[52124]: |.o.X.+. . | 2026-03-06T13:40:30.754 INFO:journalctl@ceph.mgr.a.vm00.stdout:Mar 06 13:40:30 vm00 ceph-82f504c8-1959-11f1-9a16-a78d135e200b-mgr-a[52124]: |o+X.=++o | 2026-03-06T13:40:30.754 INFO:journalctl@ceph.mgr.a.vm00.stdout:Mar 06 13:40:30 vm00 ceph-82f504c8-1959-11f1-9a16-a78d135e200b-mgr-a[52124]: |+=+=BO+ | 2026-03-06T13:40:30.754 INFO:journalctl@ceph.mgr.a.vm00.stdout:Mar 06 13:40:30 vm00 ceph-82f504c8-1959-11f1-9a16-a78d135e200b-mgr-a[52124]: +----[SHA256]-----+ 2026-03-06T13:40:30.892 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout ssh-rsa AAAAB3NzaC1yc2EAAAADAQABAAABgQDF/jDEuY0QPo+aznPmyq5t+YByffyF2ALnrnV7Z2nNO7hKkHZgh2kFYWSLXl7Nw3Mqq0BMAQrlG61HapL27bASaYIR6q7vQnyAtlaQCLN/ceuq8VnnnwGzsOkfwiQ2OwBbdgHql0uJqAWlsmZ25Y35gkUAUZSbH4jq29zbm+sNAPsJI7vIDcqvqkywm5/0MrVdLZoHfF/+UHYMjxE6h2y/g60XQLxAflsHXYQURVC+vXYaAnRYdS6vgs9riTP4t2svYuuSBQHEYgAMxmIVDGoRacddVkYqUN+k2B70HuYDQe+aUJO+nzoQXdj8hT4SYW251lDRGu6x7iggaV3fvDa1XwXh5SSjNJbAuwXclYX3hJDqEzWTeGf+Uu9uFyv0BeMTdk4h5+JPyANw8s5AdDLGsYYBK7CY7muFod7PNv5RnNX6Qpw0QZbdowNxUxZrc5iZ6c5Z3Vaq6cHKfTDGtajdmU809ferJkq1z7eeT5eBvt8KSnngNTYwbfHA+JCneBs= ceph-82f504c8-1959-11f1-9a16-a78d135e200b 2026-03-06T13:40:30.893 INFO:teuthology.orchestra.run.vm00.stdout:Wrote public SSH key to /home/ubuntu/cephtest/ceph.pub 2026-03-06T13:40:30.893 INFO:teuthology.orchestra.run.vm00.stdout:Adding key to root@localhost authorized_keys... 2026-03-06T13:40:30.893 INFO:teuthology.orchestra.run.vm00.stdout:Adding host vm00... 2026-03-06T13:40:31.645 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:40:31 vm00 ceph-mon[51914]: from='client.14138 -' entity='client.admin' cmd=[{"prefix": "cephadm set-user", "user": "root", "target": ["mon-mgr", ""]}]: dispatch 2026-03-06T13:40:31.645 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:40:31 vm00 ceph-mon[51914]: [06/Mar/2026:12:40:30] ENGINE Bus STARTING 2026-03-06T13:40:31.645 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:40:31 vm00 ceph-mon[51914]: [06/Mar/2026:12:40:30] ENGINE Serving on http://192.168.123.100:8765 2026-03-06T13:40:31.645 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:40:31 vm00 ceph-mon[51914]: [06/Mar/2026:12:40:30] ENGINE Serving on https://192.168.123.100:7150 2026-03-06T13:40:31.645 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:40:31 vm00 ceph-mon[51914]: [06/Mar/2026:12:40:30] ENGINE Bus STARTED 2026-03-06T13:40:31.645 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:40:31 vm00 ceph-mon[51914]: [06/Mar/2026:12:40:30] ENGINE Client ('192.168.123.100', 49792) lost — peer dropped the TLS connection suddenly, during handshake: (6, 'TLS/SSL connection has been closed (EOF) (_ssl.c:1147)') 2026-03-06T13:40:31.645 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:40:31 vm00 ceph-mon[51914]: from='client.14140 -' entity='client.admin' cmd=[{"prefix": "cephadm generate-key", "target": ["mon-mgr", ""]}]: dispatch 2026-03-06T13:40:31.645 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:40:31 vm00 ceph-mon[51914]: Generating ssh key... 2026-03-06T13:40:31.645 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:40:31 vm00 ceph-mon[51914]: mgrmap e8: a(active, since 2s) 2026-03-06T13:40:32.752 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:40:32 vm00 ceph-mon[51914]: from='client.14142 -' entity='client.admin' cmd=[{"prefix": "cephadm get-pub-key", "target": ["mon-mgr", ""]}]: dispatch 2026-03-06T13:40:32.752 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:40:32 vm00 ceph-mon[51914]: from='client.14144 -' entity='client.admin' cmd=[{"prefix": "orch host add", "hostname": "vm00", "addr": "192.168.123.100", "target": ["mon-mgr", ""]}]: dispatch 2026-03-06T13:40:33.823 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:40:33 vm00 ceph-mon[51914]: Deploying cephadm binary to vm00 2026-03-06T13:40:33.835 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout Added host 'vm00' with addr '192.168.123.100' 2026-03-06T13:40:33.835 INFO:teuthology.orchestra.run.vm00.stdout:Deploying unmanaged mon service... 2026-03-06T13:40:34.245 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout Scheduled mon update... 2026-03-06T13:40:34.245 INFO:teuthology.orchestra.run.vm00.stdout:Deploying unmanaged mgr service... 2026-03-06T13:40:34.662 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout Scheduled mgr update... 2026-03-06T13:40:35.003 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:40:34 vm00 ceph-mon[51914]: from='mgr.14124 192.168.123.100:0/1010685127' entity='mgr.a' 2026-03-06T13:40:35.003 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:40:34 vm00 ceph-mon[51914]: Added host vm00 2026-03-06T13:40:35.003 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:40:34 vm00 ceph-mon[51914]: from='mgr.14124 192.168.123.100:0/1010685127' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-06T13:40:35.003 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:40:34 vm00 ceph-mon[51914]: from='mgr.14124 192.168.123.100:0/1010685127' entity='mgr.a' 2026-03-06T13:40:35.003 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:40:34 vm00 ceph-mon[51914]: from='mgr.14124 192.168.123.100:0/1010685127' entity='mgr.a' 2026-03-06T13:40:35.810 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:40:35 vm00 ceph-mon[51914]: from='client.14146 -' entity='client.admin' cmd=[{"prefix": "orch apply", "service_type": "mon", "unmanaged": true, "target": ["mon-mgr", ""]}]: dispatch 2026-03-06T13:40:35.811 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:40:35 vm00 ceph-mon[51914]: Saving service mon spec with placement count:5 2026-03-06T13:40:35.811 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:40:35 vm00 ceph-mon[51914]: from='client.14148 -' entity='client.admin' cmd=[{"prefix": "orch apply", "service_type": "mgr", "unmanaged": true, "target": ["mon-mgr", ""]}]: dispatch 2026-03-06T13:40:35.811 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:40:35 vm00 ceph-mon[51914]: Saving service mgr spec with placement count:2 2026-03-06T13:40:35.811 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:40:35 vm00 ceph-mon[51914]: from='client.? 192.168.123.100:0/2781014848' entity='client.admin' 2026-03-06T13:40:35.937 INFO:teuthology.orchestra.run.vm00.stdout:Enabling the dashboard module... 2026-03-06T13:40:37.114 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:40:36 vm00 ceph-mon[51914]: from='client.? 192.168.123.100:0/2299623845' entity='client.admin' 2026-03-06T13:40:37.114 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:40:36 vm00 ceph-mon[51914]: from='mgr.14124 192.168.123.100:0/1010685127' entity='mgr.a' 2026-03-06T13:40:37.114 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:40:36 vm00 ceph-mon[51914]: from='client.? 192.168.123.100:0/2319074919' entity='client.admin' cmd=[{"prefix": "mgr module enable", "module": "dashboard"}]: dispatch 2026-03-06T13:40:37.114 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:40:36 vm00 ceph-mon[51914]: from='mgr.14124 192.168.123.100:0/1010685127' entity='mgr.a' 2026-03-06T13:40:37.114 INFO:journalctl@ceph.mgr.a.vm00.stdout:Mar 06 13:40:37 vm00 ceph-82f504c8-1959-11f1-9a16-a78d135e200b-mgr-a[52124]: ignoring --setuser ceph since I am not root 2026-03-06T13:40:37.114 INFO:journalctl@ceph.mgr.a.vm00.stdout:Mar 06 13:40:37 vm00 ceph-82f504c8-1959-11f1-9a16-a78d135e200b-mgr-a[52124]: ignoring --setgroup ceph since I am not root 2026-03-06T13:40:37.379 INFO:journalctl@ceph.mgr.a.vm00.stdout:Mar 06 13:40:37 vm00 ceph-82f504c8-1959-11f1-9a16-a78d135e200b-mgr-a[52124]: 2026-03-06T12:40:37.348+0000 7f8eec567100 -1 mgr[py] Module alerts has missing NOTIFY_TYPES member 2026-03-06T13:40:37.638 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout { 2026-03-06T13:40:37.638 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "epoch": 9, 2026-03-06T13:40:37.638 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "available": true, 2026-03-06T13:40:37.638 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "active_name": "a", 2026-03-06T13:40:37.639 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "num_standby": 0 2026-03-06T13:40:37.639 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout } 2026-03-06T13:40:37.639 INFO:teuthology.orchestra.run.vm00.stdout:Waiting for the mgr to restart... 2026-03-06T13:40:37.639 INFO:teuthology.orchestra.run.vm00.stdout:Waiting for mgr epoch 9... 2026-03-06T13:40:37.644 INFO:journalctl@ceph.mgr.a.vm00.stdout:Mar 06 13:40:37 vm00 ceph-82f504c8-1959-11f1-9a16-a78d135e200b-mgr-a[52124]: 2026-03-06T12:40:37.518+0000 7f8eec567100 -1 mgr[py] Module balancer has missing NOTIFY_TYPES member 2026-03-06T13:40:38.252 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:40:37 vm00 ceph-mon[51914]: from='client.? 192.168.123.100:0/2319074919' entity='client.admin' cmd='[{"prefix": "mgr module enable", "module": "dashboard"}]': finished 2026-03-06T13:40:38.252 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:40:37 vm00 ceph-mon[51914]: mgrmap e9: a(active, since 8s) 2026-03-06T13:40:38.252 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:40:37 vm00 ceph-mon[51914]: from='client.? 192.168.123.100:0/2023714721' entity='client.admin' cmd=[{"prefix": "mgr stat"}]: dispatch 2026-03-06T13:40:39.252 INFO:journalctl@ceph.mgr.a.vm00.stdout:Mar 06 13:40:38 vm00 ceph-82f504c8-1959-11f1-9a16-a78d135e200b-mgr-a[52124]: 2026-03-06T12:40:38.868+0000 7f8eec567100 -1 mgr[py] Module crash has missing NOTIFY_TYPES member 2026-03-06T13:40:40.223 INFO:journalctl@ceph.mgr.a.vm00.stdout:Mar 06 13:40:39 vm00 ceph-82f504c8-1959-11f1-9a16-a78d135e200b-mgr-a[52124]: 2026-03-06T12:40:39.818+0000 7f8eec567100 -1 mgr[py] Module devicehealth has missing NOTIFY_TYPES member 2026-03-06T13:40:40.224 INFO:journalctl@ceph.mgr.a.vm00.stdout:Mar 06 13:40:39 vm00 ceph-82f504c8-1959-11f1-9a16-a78d135e200b-mgr-a[52124]: 2026-03-06T12:40:39.952+0000 7f8eec567100 -1 mgr[py] Module influx has missing NOTIFY_TYPES member 2026-03-06T13:40:40.502 INFO:journalctl@ceph.mgr.a.vm00.stdout:Mar 06 13:40:40 vm00 ceph-82f504c8-1959-11f1-9a16-a78d135e200b-mgr-a[52124]: 2026-03-06T12:40:40.221+0000 7f8eec567100 -1 mgr[py] Module iostat has missing NOTIFY_TYPES member 2026-03-06T13:40:42.752 INFO:journalctl@ceph.mgr.a.vm00.stdout:Mar 06 13:40:42 vm00 ceph-82f504c8-1959-11f1-9a16-a78d135e200b-mgr-a[52124]: 2026-03-06T12:40:42.413+0000 7f8eec567100 -1 mgr[py] Module nfs has missing NOTIFY_TYPES member 2026-03-06T13:40:43.079 INFO:journalctl@ceph.mgr.a.vm00.stdout:Mar 06 13:40:42 vm00 ceph-82f504c8-1959-11f1-9a16-a78d135e200b-mgr-a[52124]: 2026-03-06T12:40:42.788+0000 7f8eec567100 -1 mgr[py] Module orchestrator has missing NOTIFY_TYPES member 2026-03-06T13:40:43.079 INFO:journalctl@ceph.mgr.a.vm00.stdout:Mar 06 13:40:42 vm00 ceph-82f504c8-1959-11f1-9a16-a78d135e200b-mgr-a[52124]: 2026-03-06T12:40:42.944+0000 7f8eec567100 -1 mgr[py] Module osd_perf_query has missing NOTIFY_TYPES member 2026-03-06T13:40:43.385 INFO:journalctl@ceph.mgr.a.vm00.stdout:Mar 06 13:40:43 vm00 ceph-82f504c8-1959-11f1-9a16-a78d135e200b-mgr-a[52124]: 2026-03-06T12:40:43.077+0000 7f8eec567100 -1 mgr[py] Module osd_support has missing NOTIFY_TYPES member 2026-03-06T13:40:43.385 INFO:journalctl@ceph.mgr.a.vm00.stdout:Mar 06 13:40:43 vm00 ceph-82f504c8-1959-11f1-9a16-a78d135e200b-mgr-a[52124]: 2026-03-06T12:40:43.227+0000 7f8eec567100 -1 mgr[py] Module pg_autoscaler has missing NOTIFY_TYPES member 2026-03-06T13:40:43.752 INFO:journalctl@ceph.mgr.a.vm00.stdout:Mar 06 13:40:43 vm00 ceph-82f504c8-1959-11f1-9a16-a78d135e200b-mgr-a[52124]: 2026-03-06T12:40:43.383+0000 7f8eec567100 -1 mgr[py] Module progress has missing NOTIFY_TYPES member 2026-03-06T13:40:44.252 INFO:journalctl@ceph.mgr.a.vm00.stdout:Mar 06 13:40:43 vm00 ceph-82f504c8-1959-11f1-9a16-a78d135e200b-mgr-a[52124]: 2026-03-06T12:40:43.945+0000 7f8eec567100 -1 mgr[py] Module prometheus has missing NOTIFY_TYPES member 2026-03-06T13:40:44.252 INFO:journalctl@ceph.mgr.a.vm00.stdout:Mar 06 13:40:44 vm00 ceph-82f504c8-1959-11f1-9a16-a78d135e200b-mgr-a[52124]: 2026-03-06T12:40:44.106+0000 7f8eec567100 -1 mgr[py] Module rbd_support has missing NOTIFY_TYPES member 2026-03-06T13:40:45.252 INFO:journalctl@ceph.mgr.a.vm00.stdout:Mar 06 13:40:44 vm00 ceph-82f504c8-1959-11f1-9a16-a78d135e200b-mgr-a[52124]: 2026-03-06T12:40:44.889+0000 7f8eec567100 -1 mgr[py] Module rgw has missing NOTIFY_TYPES member 2026-03-06T13:40:46.312 INFO:journalctl@ceph.mgr.a.vm00.stdout:Mar 06 13:40:46 vm00 ceph-82f504c8-1959-11f1-9a16-a78d135e200b-mgr-a[52124]: 2026-03-06T12:40:46.036+0000 7f8eec567100 -1 mgr[py] Module rook has missing NOTIFY_TYPES member 2026-03-06T13:40:46.312 INFO:journalctl@ceph.mgr.a.vm00.stdout:Mar 06 13:40:46 vm00 ceph-82f504c8-1959-11f1-9a16-a78d135e200b-mgr-a[52124]: 2026-03-06T12:40:46.175+0000 7f8eec567100 -1 mgr[py] Module selftest has missing NOTIFY_TYPES member 2026-03-06T13:40:46.592 INFO:journalctl@ceph.mgr.a.vm00.stdout:Mar 06 13:40:46 vm00 ceph-82f504c8-1959-11f1-9a16-a78d135e200b-mgr-a[52124]: 2026-03-06T12:40:46.310+0000 7f8eec567100 -1 mgr[py] Module snap_schedule has missing NOTIFY_TYPES member 2026-03-06T13:40:47.002 INFO:journalctl@ceph.mgr.a.vm00.stdout:Mar 06 13:40:46 vm00 ceph-82f504c8-1959-11f1-9a16-a78d135e200b-mgr-a[52124]: 2026-03-06T12:40:46.589+0000 7f8eec567100 -1 mgr[py] Module status has missing NOTIFY_TYPES member 2026-03-06T13:40:47.002 INFO:journalctl@ceph.mgr.a.vm00.stdout:Mar 06 13:40:46 vm00 ceph-82f504c8-1959-11f1-9a16-a78d135e200b-mgr-a[52124]: 2026-03-06T12:40:46.726+0000 7f8eec567100 -1 mgr[py] Module telegraf has missing NOTIFY_TYPES member 2026-03-06T13:40:47.431 INFO:journalctl@ceph.mgr.a.vm00.stdout:Mar 06 13:40:47 vm00 ceph-82f504c8-1959-11f1-9a16-a78d135e200b-mgr-a[52124]: 2026-03-06T12:40:47.043+0000 7f8eec567100 -1 mgr[py] Module telemetry has missing NOTIFY_TYPES member 2026-03-06T13:40:47.752 INFO:journalctl@ceph.mgr.a.vm00.stdout:Mar 06 13:40:47 vm00 ceph-82f504c8-1959-11f1-9a16-a78d135e200b-mgr-a[52124]: 2026-03-06T12:40:47.429+0000 7f8eec567100 -1 mgr[py] Module test_orchestrator has missing NOTIFY_TYPES member 2026-03-06T13:40:48.129 INFO:journalctl@ceph.mgr.a.vm00.stdout:Mar 06 13:40:47 vm00 ceph-82f504c8-1959-11f1-9a16-a78d135e200b-mgr-a[52124]: 2026-03-06T12:40:47.819+0000 7f8eec567100 -1 mgr[py] Module volumes has missing NOTIFY_TYPES member 2026-03-06T13:40:48.129 INFO:journalctl@ceph.mgr.a.vm00.stdout:Mar 06 13:40:47 vm00 ceph-82f504c8-1959-11f1-9a16-a78d135e200b-mgr-a[52124]: 2026-03-06T12:40:47.963+0000 7f8eec567100 -1 mgr[py] Module zabbix has missing NOTIFY_TYPES member 2026-03-06T13:40:48.129 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:40:48 vm00 ceph-mon[51914]: Active manager daemon a restarted 2026-03-06T13:40:48.129 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:40:48 vm00 ceph-mon[51914]: Activating manager daemon a 2026-03-06T13:40:48.129 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:40:48 vm00 ceph-mon[51914]: osdmap e3: 0 total, 0 up, 0 in 2026-03-06T13:40:48.129 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:40:48 vm00 ceph-mon[51914]: mgrmap e10: a(active, starting, since 0.00926268s) 2026-03-06T13:40:48.129 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:40:48 vm00 ceph-mon[51914]: from='mgr.14156 192.168.123.100:0/929894269' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "a"}]: dispatch 2026-03-06T13:40:48.129 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:40:48 vm00 ceph-mon[51914]: from='mgr.14156 192.168.123.100:0/929894269' entity='mgr.a' cmd=[{"prefix": "mgr metadata", "who": "a", "id": "a"}]: dispatch 2026-03-06T13:40:48.129 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:40:48 vm00 ceph-mon[51914]: from='mgr.14156 192.168.123.100:0/929894269' entity='mgr.a' cmd=[{"prefix": "mds metadata"}]: dispatch 2026-03-06T13:40:48.129 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:40:48 vm00 ceph-mon[51914]: from='mgr.14156 192.168.123.100:0/929894269' entity='mgr.a' cmd=[{"prefix": "osd metadata"}]: dispatch 2026-03-06T13:40:48.129 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:40:48 vm00 ceph-mon[51914]: from='mgr.14156 192.168.123.100:0/929894269' entity='mgr.a' cmd=[{"prefix": "mon metadata"}]: dispatch 2026-03-06T13:40:48.129 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:40:48 vm00 ceph-mon[51914]: Manager daemon a is now available 2026-03-06T13:40:48.129 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:40:48 vm00 ceph-mon[51914]: from='mgr.14156 192.168.123.100:0/929894269' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-06T13:40:48.129 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:40:48 vm00 ceph-mon[51914]: from='mgr.14156 192.168.123.100:0/929894269' entity='mgr.a' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/a/mirror_snapshot_schedule"}]: dispatch 2026-03-06T13:40:49.041 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:40:49 vm00 ceph-mon[51914]: from='mgr.14156 192.168.123.100:0/929894269' entity='mgr.a' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/a/trash_purge_schedule"}]: dispatch 2026-03-06T13:40:49.041 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:40:49 vm00 ceph-mon[51914]: from='mgr.14156 192.168.123.100:0/929894269' entity='mgr.a' 2026-03-06T13:40:49.042 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:40:49 vm00 ceph-mon[51914]: mgrmap e11: a(active, since 1.01172s) 2026-03-06T13:40:49.042 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout { 2026-03-06T13:40:49.042 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "mgrmap_epoch": 11, 2026-03-06T13:40:49.042 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "initialized": true 2026-03-06T13:40:49.042 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout } 2026-03-06T13:40:49.042 INFO:teuthology.orchestra.run.vm00.stdout:mgr epoch 9 is available 2026-03-06T13:40:49.042 INFO:teuthology.orchestra.run.vm00.stdout:Generating a dashboard self-signed certificate... 2026-03-06T13:40:49.643 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout Self-signed certificate created 2026-03-06T13:40:49.644 INFO:teuthology.orchestra.run.vm00.stdout:Creating initial admin user... 2026-03-06T13:40:50.236 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout {"username": "admin", "password": "$2b$12$VtJYC.mxbepuYfZqhO1DFOO09VoX5fKweKgNtzOkjZUGKjNAs2Vvq", "roles": ["administrator"], "name": null, "email": null, "lastUpdate": 1772800850, "enabled": true, "pwdExpirationDate": null, "pwdUpdateRequired": true} 2026-03-06T13:40:50.236 INFO:teuthology.orchestra.run.vm00.stdout:Fetching dashboard port number... 2026-03-06T13:40:50.281 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:40:50 vm00 ceph-mon[51914]: from='client.14160 -' entity='client.admin' cmd=[{"prefix": "get_command_descriptions"}]: dispatch 2026-03-06T13:40:50.282 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:40:50 vm00 ceph-mon[51914]: from='client.14160 -' entity='client.admin' cmd=[{"prefix": "mgr_status"}]: dispatch 2026-03-06T13:40:50.282 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:40:50 vm00 ceph-mon[51914]: from='mgr.14156 192.168.123.100:0/929894269' entity='mgr.a' 2026-03-06T13:40:50.282 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:40:50 vm00 ceph-mon[51914]: from='mgr.14156 192.168.123.100:0/929894269' entity='mgr.a' 2026-03-06T13:40:50.282 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:40:50 vm00 ceph-mon[51914]: from='mgr.14156 192.168.123.100:0/929894269' entity='mgr.a' 2026-03-06T13:40:50.654 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout 8443 2026-03-06T13:40:50.655 INFO:teuthology.orchestra.run.vm00.stdout:firewalld does not appear to be present 2026-03-06T13:40:50.655 INFO:teuthology.orchestra.run.vm00.stdout:Not possible to open ports <[8443]>. firewalld.service is not available 2026-03-06T13:40:50.656 INFO:teuthology.orchestra.run.vm00.stdout:Ceph Dashboard is now available at: 2026-03-06T13:40:50.656 INFO:teuthology.orchestra.run.vm00.stdout: 2026-03-06T13:40:50.656 INFO:teuthology.orchestra.run.vm00.stdout: URL: https://vm00.local:8443/ 2026-03-06T13:40:50.656 INFO:teuthology.orchestra.run.vm00.stdout: User: admin 2026-03-06T13:40:50.656 INFO:teuthology.orchestra.run.vm00.stdout: Password: se9099zkmw 2026-03-06T13:40:50.657 INFO:teuthology.orchestra.run.vm00.stdout: 2026-03-06T13:40:50.657 INFO:teuthology.orchestra.run.vm00.stdout:Saving cluster configuration to /var/lib/ceph/82f504c8-1959-11f1-9a16-a78d135e200b/config directory 2026-03-06T13:40:51.092 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stderr set mgr/dashboard/cluster/status 2026-03-06T13:40:51.093 INFO:teuthology.orchestra.run.vm00.stdout:You can access the Ceph CLI as following in case of multi-cluster or non-default config: 2026-03-06T13:40:51.093 INFO:teuthology.orchestra.run.vm00.stdout: 2026-03-06T13:40:51.093 INFO:teuthology.orchestra.run.vm00.stdout: sudo /home/ubuntu/cephtest/cephadm shell --fsid 82f504c8-1959-11f1-9a16-a78d135e200b -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring 2026-03-06T13:40:51.093 INFO:teuthology.orchestra.run.vm00.stdout: 2026-03-06T13:40:51.093 INFO:teuthology.orchestra.run.vm00.stdout:Or, if you are only running a single cluster on this host: 2026-03-06T13:40:51.093 INFO:teuthology.orchestra.run.vm00.stdout: 2026-03-06T13:40:51.093 INFO:teuthology.orchestra.run.vm00.stdout: sudo /home/ubuntu/cephtest/cephadm shell 2026-03-06T13:40:51.093 INFO:teuthology.orchestra.run.vm00.stdout: 2026-03-06T13:40:51.093 INFO:teuthology.orchestra.run.vm00.stdout:Please consider enabling telemetry to help improve Ceph: 2026-03-06T13:40:51.093 INFO:teuthology.orchestra.run.vm00.stdout: 2026-03-06T13:40:51.093 INFO:teuthology.orchestra.run.vm00.stdout: ceph telemetry on 2026-03-06T13:40:51.093 INFO:teuthology.orchestra.run.vm00.stdout: 2026-03-06T13:40:51.093 INFO:teuthology.orchestra.run.vm00.stdout:For more information see: 2026-03-06T13:40:51.093 INFO:teuthology.orchestra.run.vm00.stdout: 2026-03-06T13:40:51.093 INFO:teuthology.orchestra.run.vm00.stdout: https://docs.ceph.com/en/latest/mgr/telemetry/ 2026-03-06T13:40:51.093 INFO:teuthology.orchestra.run.vm00.stdout: 2026-03-06T13:40:51.094 INFO:teuthology.orchestra.run.vm00.stdout:Bootstrap complete. 2026-03-06T13:40:51.138 INFO:tasks.cephadm:Fetching config... 2026-03-06T13:40:51.138 DEBUG:teuthology.orchestra.run.vm00:> set -ex 2026-03-06T13:40:51.138 DEBUG:teuthology.orchestra.run.vm00:> dd if=/etc/ceph/ceph.conf of=/dev/stdout 2026-03-06T13:40:51.171 INFO:tasks.cephadm:Fetching client.admin keyring... 2026-03-06T13:40:51.171 DEBUG:teuthology.orchestra.run.vm00:> set -ex 2026-03-06T13:40:51.171 DEBUG:teuthology.orchestra.run.vm00:> dd if=/etc/ceph/ceph.client.admin.keyring of=/dev/stdout 2026-03-06T13:40:51.235 INFO:tasks.cephadm:Fetching mon keyring... 2026-03-06T13:40:51.235 DEBUG:teuthology.orchestra.run.vm00:> set -ex 2026-03-06T13:40:51.235 DEBUG:teuthology.orchestra.run.vm00:> sudo dd if=/var/lib/ceph/82f504c8-1959-11f1-9a16-a78d135e200b/mon.a/keyring of=/dev/stdout 2026-03-06T13:40:51.301 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:40:51 vm00 ceph-mon[51914]: [06/Mar/2026:12:40:49] ENGINE Bus STARTING 2026-03-06T13:40:51.301 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:40:51 vm00 ceph-mon[51914]: [06/Mar/2026:12:40:49] ENGINE Serving on https://192.168.123.100:7150 2026-03-06T13:40:51.301 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:40:51 vm00 ceph-mon[51914]: [06/Mar/2026:12:40:49] ENGINE Client ('192.168.123.100', 49358) lost — peer dropped the TLS connection suddenly, during handshake: (6, 'TLS/SSL connection has been closed (EOF) (_ssl.c:1147)') 2026-03-06T13:40:51.301 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:40:51 vm00 ceph-mon[51914]: [06/Mar/2026:12:40:49] ENGINE Serving on http://192.168.123.100:8765 2026-03-06T13:40:51.301 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:40:51 vm00 ceph-mon[51914]: [06/Mar/2026:12:40:49] ENGINE Bus STARTED 2026-03-06T13:40:51.301 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:40:51 vm00 ceph-mon[51914]: from='client.14168 -' entity='client.admin' cmd=[{"prefix": "dashboard create-self-signed-cert", "target": ["mon-mgr", ""]}]: dispatch 2026-03-06T13:40:51.301 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:40:51 vm00 ceph-mon[51914]: from='client.14170 -' entity='client.admin' cmd=[{"prefix": "dashboard ac-user-create", "username": "admin", "rolename": "administrator", "force_password": true, "pwd_update_required": true, "target": ["mon-mgr", ""]}]: dispatch 2026-03-06T13:40:51.301 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:40:51 vm00 ceph-mon[51914]: from='mgr.14156 192.168.123.100:0/929894269' entity='mgr.a' 2026-03-06T13:40:51.301 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:40:51 vm00 ceph-mon[51914]: from='client.? 192.168.123.100:0/2804616017' entity='client.admin' cmd=[{"prefix": "config get", "who": "mgr", "key": "mgr/dashboard/ssl_server_port"}]: dispatch 2026-03-06T13:40:51.301 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:40:51 vm00 ceph-mon[51914]: mgrmap e12: a(active, since 2s) 2026-03-06T13:40:51.304 INFO:tasks.cephadm:Fetching pub ssh key... 2026-03-06T13:40:51.304 DEBUG:teuthology.orchestra.run.vm00:> set -ex 2026-03-06T13:40:51.304 DEBUG:teuthology.orchestra.run.vm00:> dd if=/home/ubuntu/cephtest/ceph.pub of=/dev/stdout 2026-03-06T13:40:51.361 INFO:tasks.cephadm:Installing pub ssh key for root users... 2026-03-06T13:40:51.361 DEBUG:teuthology.orchestra.run.vm00:> sudo install -d -m 0700 /root/.ssh && echo 'ssh-rsa AAAAB3NzaC1yc2EAAAADAQABAAABgQDF/jDEuY0QPo+aznPmyq5t+YByffyF2ALnrnV7Z2nNO7hKkHZgh2kFYWSLXl7Nw3Mqq0BMAQrlG61HapL27bASaYIR6q7vQnyAtlaQCLN/ceuq8VnnnwGzsOkfwiQ2OwBbdgHql0uJqAWlsmZ25Y35gkUAUZSbH4jq29zbm+sNAPsJI7vIDcqvqkywm5/0MrVdLZoHfF/+UHYMjxE6h2y/g60XQLxAflsHXYQURVC+vXYaAnRYdS6vgs9riTP4t2svYuuSBQHEYgAMxmIVDGoRacddVkYqUN+k2B70HuYDQe+aUJO+nzoQXdj8hT4SYW251lDRGu6x7iggaV3fvDa1XwXh5SSjNJbAuwXclYX3hJDqEzWTeGf+Uu9uFyv0BeMTdk4h5+JPyANw8s5AdDLGsYYBK7CY7muFod7PNv5RnNX6Qpw0QZbdowNxUxZrc5iZ6c5Z3Vaq6cHKfTDGtajdmU809ferJkq1z7eeT5eBvt8KSnngNTYwbfHA+JCneBs= ceph-82f504c8-1959-11f1-9a16-a78d135e200b' | sudo tee -a /root/.ssh/authorized_keys && sudo chmod 0600 /root/.ssh/authorized_keys 2026-03-06T13:40:51.438 INFO:teuthology.orchestra.run.vm00.stdout:ssh-rsa AAAAB3NzaC1yc2EAAAADAQABAAABgQDF/jDEuY0QPo+aznPmyq5t+YByffyF2ALnrnV7Z2nNO7hKkHZgh2kFYWSLXl7Nw3Mqq0BMAQrlG61HapL27bASaYIR6q7vQnyAtlaQCLN/ceuq8VnnnwGzsOkfwiQ2OwBbdgHql0uJqAWlsmZ25Y35gkUAUZSbH4jq29zbm+sNAPsJI7vIDcqvqkywm5/0MrVdLZoHfF/+UHYMjxE6h2y/g60XQLxAflsHXYQURVC+vXYaAnRYdS6vgs9riTP4t2svYuuSBQHEYgAMxmIVDGoRacddVkYqUN+k2B70HuYDQe+aUJO+nzoQXdj8hT4SYW251lDRGu6x7iggaV3fvDa1XwXh5SSjNJbAuwXclYX3hJDqEzWTeGf+Uu9uFyv0BeMTdk4h5+JPyANw8s5AdDLGsYYBK7CY7muFod7PNv5RnNX6Qpw0QZbdowNxUxZrc5iZ6c5Z3Vaq6cHKfTDGtajdmU809ferJkq1z7eeT5eBvt8KSnngNTYwbfHA+JCneBs= ceph-82f504c8-1959-11f1-9a16-a78d135e200b 2026-03-06T13:40:51.448 DEBUG:teuthology.orchestra.run.vm01:> sudo install -d -m 0700 /root/.ssh && echo 'ssh-rsa AAAAB3NzaC1yc2EAAAADAQABAAABgQDF/jDEuY0QPo+aznPmyq5t+YByffyF2ALnrnV7Z2nNO7hKkHZgh2kFYWSLXl7Nw3Mqq0BMAQrlG61HapL27bASaYIR6q7vQnyAtlaQCLN/ceuq8VnnnwGzsOkfwiQ2OwBbdgHql0uJqAWlsmZ25Y35gkUAUZSbH4jq29zbm+sNAPsJI7vIDcqvqkywm5/0MrVdLZoHfF/+UHYMjxE6h2y/g60XQLxAflsHXYQURVC+vXYaAnRYdS6vgs9riTP4t2svYuuSBQHEYgAMxmIVDGoRacddVkYqUN+k2B70HuYDQe+aUJO+nzoQXdj8hT4SYW251lDRGu6x7iggaV3fvDa1XwXh5SSjNJbAuwXclYX3hJDqEzWTeGf+Uu9uFyv0BeMTdk4h5+JPyANw8s5AdDLGsYYBK7CY7muFod7PNv5RnNX6Qpw0QZbdowNxUxZrc5iZ6c5Z3Vaq6cHKfTDGtajdmU809ferJkq1z7eeT5eBvt8KSnngNTYwbfHA+JCneBs= ceph-82f504c8-1959-11f1-9a16-a78d135e200b' | sudo tee -a /root/.ssh/authorized_keys && sudo chmod 0600 /root/.ssh/authorized_keys 2026-03-06T13:40:51.480 INFO:teuthology.orchestra.run.vm01.stdout:ssh-rsa AAAAB3NzaC1yc2EAAAADAQABAAABgQDF/jDEuY0QPo+aznPmyq5t+YByffyF2ALnrnV7Z2nNO7hKkHZgh2kFYWSLXl7Nw3Mqq0BMAQrlG61HapL27bASaYIR6q7vQnyAtlaQCLN/ceuq8VnnnwGzsOkfwiQ2OwBbdgHql0uJqAWlsmZ25Y35gkUAUZSbH4jq29zbm+sNAPsJI7vIDcqvqkywm5/0MrVdLZoHfF/+UHYMjxE6h2y/g60XQLxAflsHXYQURVC+vXYaAnRYdS6vgs9riTP4t2svYuuSBQHEYgAMxmIVDGoRacddVkYqUN+k2B70HuYDQe+aUJO+nzoQXdj8hT4SYW251lDRGu6x7iggaV3fvDa1XwXh5SSjNJbAuwXclYX3hJDqEzWTeGf+Uu9uFyv0BeMTdk4h5+JPyANw8s5AdDLGsYYBK7CY7muFod7PNv5RnNX6Qpw0QZbdowNxUxZrc5iZ6c5Z3Vaq6cHKfTDGtajdmU809ferJkq1z7eeT5eBvt8KSnngNTYwbfHA+JCneBs= ceph-82f504c8-1959-11f1-9a16-a78d135e200b 2026-03-06T13:40:51.492 DEBUG:teuthology.orchestra.run.vm08:> sudo install -d -m 0700 /root/.ssh && echo 'ssh-rsa AAAAB3NzaC1yc2EAAAADAQABAAABgQDF/jDEuY0QPo+aznPmyq5t+YByffyF2ALnrnV7Z2nNO7hKkHZgh2kFYWSLXl7Nw3Mqq0BMAQrlG61HapL27bASaYIR6q7vQnyAtlaQCLN/ceuq8VnnnwGzsOkfwiQ2OwBbdgHql0uJqAWlsmZ25Y35gkUAUZSbH4jq29zbm+sNAPsJI7vIDcqvqkywm5/0MrVdLZoHfF/+UHYMjxE6h2y/g60XQLxAflsHXYQURVC+vXYaAnRYdS6vgs9riTP4t2svYuuSBQHEYgAMxmIVDGoRacddVkYqUN+k2B70HuYDQe+aUJO+nzoQXdj8hT4SYW251lDRGu6x7iggaV3fvDa1XwXh5SSjNJbAuwXclYX3hJDqEzWTeGf+Uu9uFyv0BeMTdk4h5+JPyANw8s5AdDLGsYYBK7CY7muFod7PNv5RnNX6Qpw0QZbdowNxUxZrc5iZ6c5Z3Vaq6cHKfTDGtajdmU809ferJkq1z7eeT5eBvt8KSnngNTYwbfHA+JCneBs= ceph-82f504c8-1959-11f1-9a16-a78d135e200b' | sudo tee -a /root/.ssh/authorized_keys && sudo chmod 0600 /root/.ssh/authorized_keys 2026-03-06T13:40:51.532 INFO:teuthology.orchestra.run.vm08.stdout:ssh-rsa AAAAB3NzaC1yc2EAAAADAQABAAABgQDF/jDEuY0QPo+aznPmyq5t+YByffyF2ALnrnV7Z2nNO7hKkHZgh2kFYWSLXl7Nw3Mqq0BMAQrlG61HapL27bASaYIR6q7vQnyAtlaQCLN/ceuq8VnnnwGzsOkfwiQ2OwBbdgHql0uJqAWlsmZ25Y35gkUAUZSbH4jq29zbm+sNAPsJI7vIDcqvqkywm5/0MrVdLZoHfF/+UHYMjxE6h2y/g60XQLxAflsHXYQURVC+vXYaAnRYdS6vgs9riTP4t2svYuuSBQHEYgAMxmIVDGoRacddVkYqUN+k2B70HuYDQe+aUJO+nzoQXdj8hT4SYW251lDRGu6x7iggaV3fvDa1XwXh5SSjNJbAuwXclYX3hJDqEzWTeGf+Uu9uFyv0BeMTdk4h5+JPyANw8s5AdDLGsYYBK7CY7muFod7PNv5RnNX6Qpw0QZbdowNxUxZrc5iZ6c5Z3Vaq6cHKfTDGtajdmU809ferJkq1z7eeT5eBvt8KSnngNTYwbfHA+JCneBs= ceph-82f504c8-1959-11f1-9a16-a78d135e200b 2026-03-06T13:40:51.541 DEBUG:teuthology.orchestra.run.vm00:> sudo /home/ubuntu/cephtest/cephadm --image harbor.clyso.com/custom-ceph/ceph/ceph:cobaltcore-storage-v19.2.3-fasttrack-3 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 82f504c8-1959-11f1-9a16-a78d135e200b -- ceph config set mgr mgr/cephadm/allow_ptrace true 2026-03-06T13:40:51.908 INFO:teuthology.orchestra.run.vm00.stderr:Inferring config /var/lib/ceph/82f504c8-1959-11f1-9a16-a78d135e200b/mon.a/config 2026-03-06T13:40:52.252 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:40:52 vm00 ceph-mon[51914]: from='client.? 192.168.123.100:0/4120213316' entity='client.admin' 2026-03-06T13:40:52.407 INFO:tasks.cephadm:Distributing conf and client.admin keyring to all hosts + 0755 2026-03-06T13:40:52.407 DEBUG:teuthology.orchestra.run.vm00:> sudo /home/ubuntu/cephtest/cephadm --image harbor.clyso.com/custom-ceph/ceph/ceph:cobaltcore-storage-v19.2.3-fasttrack-3 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 82f504c8-1959-11f1-9a16-a78d135e200b -- ceph orch client-keyring set client.admin '*' --mode 0755 2026-03-06T13:40:52.788 INFO:teuthology.orchestra.run.vm00.stderr:Inferring config /var/lib/ceph/82f504c8-1959-11f1-9a16-a78d135e200b/mon.a/config 2026-03-06T13:40:53.235 INFO:tasks.cephadm:Writing (initial) conf and keyring to vm01 2026-03-06T13:40:53.236 DEBUG:teuthology.orchestra.run.vm01:> set -ex 2026-03-06T13:40:53.236 DEBUG:teuthology.orchestra.run.vm01:> dd of=/etc/ceph/ceph.conf 2026-03-06T13:40:53.253 DEBUG:teuthology.orchestra.run.vm01:> set -ex 2026-03-06T13:40:53.253 DEBUG:teuthology.orchestra.run.vm01:> dd of=/etc/ceph/ceph.client.admin.keyring 2026-03-06T13:40:53.312 INFO:tasks.cephadm:Adding host vm01 to orchestrator... 2026-03-06T13:40:53.312 DEBUG:teuthology.orchestra.run.vm00:> sudo /home/ubuntu/cephtest/cephadm --image harbor.clyso.com/custom-ceph/ceph/ceph:cobaltcore-storage-v19.2.3-fasttrack-3 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 82f504c8-1959-11f1-9a16-a78d135e200b -- ceph orch host add vm01 2026-03-06T13:40:53.502 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:40:53 vm00 ceph-mon[51914]: from='client.? 192.168.123.100:0/481832294' entity='client.admin' 2026-03-06T13:40:53.502 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:40:53 vm00 ceph-mon[51914]: from='mgr.14156 192.168.123.100:0/929894269' entity='mgr.a' 2026-03-06T13:40:53.502 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:40:53 vm00 ceph-mon[51914]: from='mgr.14156 192.168.123.100:0/929894269' entity='mgr.a' 2026-03-06T13:40:53.502 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:40:53 vm00 ceph-mon[51914]: from='mgr.14156 192.168.123.100:0/929894269' entity='mgr.a' cmd=[{"prefix": "config rm", "who": "osd/host:vm00", "name": "osd_memory_target"}]: dispatch 2026-03-06T13:40:53.502 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:40:53 vm00 ceph-mon[51914]: from='mgr.14156 192.168.123.100:0/929894269' entity='mgr.a' 2026-03-06T13:40:53.502 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:40:53 vm00 ceph-mon[51914]: from='mgr.14156 192.168.123.100:0/929894269' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-06T13:40:53.502 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:40:53 vm00 ceph-mon[51914]: from='mgr.14156 192.168.123.100:0/929894269' entity='mgr.a' 2026-03-06T13:40:53.502 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:40:53 vm00 ceph-mon[51914]: from='mgr.14156 192.168.123.100:0/929894269' entity='mgr.a' 2026-03-06T13:40:53.503 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:40:53 vm00 ceph-mon[51914]: from='mgr.14156 192.168.123.100:0/929894269' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-06T13:40:53.503 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:40:53 vm00 ceph-mon[51914]: from='mgr.14156 192.168.123.100:0/929894269' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-06T13:40:53.503 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:40:53 vm00 ceph-mon[51914]: from='mgr.14156 192.168.123.100:0/929894269' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-06T13:40:53.675 INFO:teuthology.orchestra.run.vm00.stderr:Inferring config /var/lib/ceph/82f504c8-1959-11f1-9a16-a78d135e200b/mon.a/config 2026-03-06T13:40:55.252 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:40:54 vm00 ceph-mon[51914]: from='client.14178 -' entity='client.admin' cmd=[{"prefix": "orch client-keyring set", "entity": "client.admin", "placement": "*", "mode": "0755", "target": ["mon-mgr", ""]}]: dispatch 2026-03-06T13:40:55.252 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:40:54 vm00 ceph-mon[51914]: Updating vm00:/etc/ceph/ceph.conf 2026-03-06T13:40:55.252 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:40:54 vm00 ceph-mon[51914]: Updating vm00:/var/lib/ceph/82f504c8-1959-11f1-9a16-a78d135e200b/config/ceph.conf 2026-03-06T13:40:55.252 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:40:54 vm00 ceph-mon[51914]: Updating vm00:/etc/ceph/ceph.client.admin.keyring 2026-03-06T13:40:55.252 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:40:54 vm00 ceph-mon[51914]: Updating vm00:/var/lib/ceph/82f504c8-1959-11f1-9a16-a78d135e200b/config/ceph.client.admin.keyring 2026-03-06T13:40:55.252 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:40:54 vm00 ceph-mon[51914]: from='mgr.14156 192.168.123.100:0/929894269' entity='mgr.a' 2026-03-06T13:40:55.252 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:40:54 vm00 ceph-mon[51914]: from='mgr.14156 192.168.123.100:0/929894269' entity='mgr.a' 2026-03-06T13:40:55.252 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:40:54 vm00 ceph-mon[51914]: from='mgr.14156 192.168.123.100:0/929894269' entity='mgr.a' 2026-03-06T13:40:56.252 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:40:55 vm00 ceph-mon[51914]: from='client.14180 -' entity='client.admin' cmd=[{"prefix": "orch host add", "hostname": "vm01", "target": ["mon-mgr", ""]}]: dispatch 2026-03-06T13:40:56.252 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:40:55 vm00 ceph-mon[51914]: Deploying cephadm binary to vm01 2026-03-06T13:40:56.252 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:40:55 vm00 ceph-mon[51914]: mgrmap e13: a(active, since 6s) 2026-03-06T13:40:56.672 INFO:teuthology.orchestra.run.vm00.stdout:Added host 'vm01' with addr '192.168.123.101' 2026-03-06T13:40:56.757 DEBUG:teuthology.orchestra.run.vm00:> sudo /home/ubuntu/cephtest/cephadm --image harbor.clyso.com/custom-ceph/ceph/ceph:cobaltcore-storage-v19.2.3-fasttrack-3 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 82f504c8-1959-11f1-9a16-a78d135e200b -- ceph orch host ls --format=json 2026-03-06T13:40:57.129 INFO:teuthology.orchestra.run.vm00.stderr:Inferring config /var/lib/ceph/82f504c8-1959-11f1-9a16-a78d135e200b/mon.a/config 2026-03-06T13:40:57.499 INFO:teuthology.orchestra.run.vm00.stdout: 2026-03-06T13:40:57.500 INFO:teuthology.orchestra.run.vm00.stdout:[{"addr": "192.168.123.100", "hostname": "vm00", "labels": [], "status": ""}, {"addr": "192.168.123.101", "hostname": "vm01", "labels": [], "status": ""}] 2026-03-06T13:40:57.585 INFO:tasks.cephadm:Writing (initial) conf and keyring to vm08 2026-03-06T13:40:57.585 DEBUG:teuthology.orchestra.run.vm08:> set -ex 2026-03-06T13:40:57.585 DEBUG:teuthology.orchestra.run.vm08:> dd of=/etc/ceph/ceph.conf 2026-03-06T13:40:57.604 DEBUG:teuthology.orchestra.run.vm08:> set -ex 2026-03-06T13:40:57.604 DEBUG:teuthology.orchestra.run.vm08:> dd of=/etc/ceph/ceph.client.admin.keyring 2026-03-06T13:40:57.664 INFO:tasks.cephadm:Adding host vm08 to orchestrator... 2026-03-06T13:40:57.664 DEBUG:teuthology.orchestra.run.vm00:> sudo /home/ubuntu/cephtest/cephadm --image harbor.clyso.com/custom-ceph/ceph/ceph:cobaltcore-storage-v19.2.3-fasttrack-3 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 82f504c8-1959-11f1-9a16-a78d135e200b -- ceph orch host add vm08 2026-03-06T13:40:57.752 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:40:57 vm00 ceph-mon[51914]: from='mgr.14156 192.168.123.100:0/929894269' entity='mgr.a' 2026-03-06T13:40:57.752 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:40:57 vm00 ceph-mon[51914]: Added host vm01 2026-03-06T13:40:57.752 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:40:57 vm00 ceph-mon[51914]: from='mgr.14156 192.168.123.100:0/929894269' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-06T13:40:57.752 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:40:57 vm00 ceph-mon[51914]: from='mgr.14156 192.168.123.100:0/929894269' entity='mgr.a' 2026-03-06T13:40:57.752 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:40:57 vm00 ceph-mon[51914]: from='mgr.14156 192.168.123.100:0/929894269' entity='mgr.a' 2026-03-06T13:40:58.013 INFO:teuthology.orchestra.run.vm00.stderr:Inferring config /var/lib/ceph/82f504c8-1959-11f1-9a16-a78d135e200b/mon.a/config 2026-03-06T13:40:59.002 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:40:58 vm00 ceph-mon[51914]: from='client.14182 -' entity='client.admin' cmd=[{"prefix": "orch host ls", "target": ["mon-mgr", ""], "format": "json"}]: dispatch 2026-03-06T13:40:59.002 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:40:58 vm00 ceph-mon[51914]: from='mgr.14156 192.168.123.100:0/929894269' entity='mgr.a' 2026-03-06T13:41:00.002 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:40:59 vm00 ceph-mon[51914]: from='client.14184 -' entity='client.admin' cmd=[{"prefix": "orch host add", "hostname": "vm08", "target": ["mon-mgr", ""]}]: dispatch 2026-03-06T13:41:00.927 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:41:00 vm00 ceph-mon[51914]: Deploying cephadm binary to vm08 2026-03-06T13:41:00.927 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:41:00 vm00 ceph-mon[51914]: from='mgr.14156 192.168.123.100:0/929894269' entity='mgr.a' 2026-03-06T13:41:00.927 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:41:00 vm00 ceph-mon[51914]: from='mgr.14156 192.168.123.100:0/929894269' entity='mgr.a' 2026-03-06T13:41:00.927 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:41:00 vm00 ceph-mon[51914]: from='mgr.14156 192.168.123.100:0/929894269' entity='mgr.a' 2026-03-06T13:41:00.927 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:41:00 vm00 ceph-mon[51914]: from='mgr.14156 192.168.123.100:0/929894269' entity='mgr.a' 2026-03-06T13:41:00.927 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:41:00 vm00 ceph-mon[51914]: from='mgr.14156 192.168.123.100:0/929894269' entity='mgr.a' cmd=[{"prefix": "config rm", "who": "osd/host:vm01", "name": "osd_memory_target"}]: dispatch 2026-03-06T13:41:00.927 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:41:00 vm00 ceph-mon[51914]: from='mgr.14156 192.168.123.100:0/929894269' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-06T13:41:00.927 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:41:00 vm00 ceph-mon[51914]: from='mgr.14156 192.168.123.100:0/929894269' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-06T13:41:00.927 INFO:teuthology.orchestra.run.vm00.stdout:Added host 'vm08' with addr '192.168.123.108' 2026-03-06T13:41:01.000 DEBUG:teuthology.orchestra.run.vm00:> sudo /home/ubuntu/cephtest/cephadm --image harbor.clyso.com/custom-ceph/ceph/ceph:cobaltcore-storage-v19.2.3-fasttrack-3 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 82f504c8-1959-11f1-9a16-a78d135e200b -- ceph orch host ls --format=json 2026-03-06T13:41:01.357 INFO:teuthology.orchestra.run.vm00.stderr:Inferring config /var/lib/ceph/82f504c8-1959-11f1-9a16-a78d135e200b/mon.a/config 2026-03-06T13:41:01.731 INFO:teuthology.orchestra.run.vm00.stdout: 2026-03-06T13:41:01.731 INFO:teuthology.orchestra.run.vm00.stdout:[{"addr": "192.168.123.100", "hostname": "vm00", "labels": [], "status": ""}, {"addr": "192.168.123.101", "hostname": "vm01", "labels": [], "status": ""}, {"addr": "192.168.123.108", "hostname": "vm08", "labels": [], "status": ""}] 2026-03-06T13:41:01.814 INFO:tasks.cephadm:Setting crush tunables to default 2026-03-06T13:41:01.814 DEBUG:teuthology.orchestra.run.vm00:> sudo /home/ubuntu/cephtest/cephadm --image harbor.clyso.com/custom-ceph/ceph/ceph:cobaltcore-storage-v19.2.3-fasttrack-3 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 82f504c8-1959-11f1-9a16-a78d135e200b -- ceph osd crush tunables default 2026-03-06T13:41:02.002 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:41:01 vm00 ceph-mon[51914]: Updating vm01:/etc/ceph/ceph.conf 2026-03-06T13:41:02.002 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:41:01 vm00 ceph-mon[51914]: Updating vm01:/var/lib/ceph/82f504c8-1959-11f1-9a16-a78d135e200b/config/ceph.conf 2026-03-06T13:41:02.002 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:41:01 vm00 ceph-mon[51914]: Updating vm01:/etc/ceph/ceph.client.admin.keyring 2026-03-06T13:41:02.002 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:41:01 vm00 ceph-mon[51914]: Updating vm01:/var/lib/ceph/82f504c8-1959-11f1-9a16-a78d135e200b/config/ceph.client.admin.keyring 2026-03-06T13:41:02.002 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:41:01 vm00 ceph-mon[51914]: from='mgr.14156 192.168.123.100:0/929894269' entity='mgr.a' 2026-03-06T13:41:02.002 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:41:01 vm00 ceph-mon[51914]: from='mgr.14156 192.168.123.100:0/929894269' entity='mgr.a' 2026-03-06T13:41:02.002 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:41:01 vm00 ceph-mon[51914]: from='mgr.14156 192.168.123.100:0/929894269' entity='mgr.a' 2026-03-06T13:41:02.002 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:41:01 vm00 ceph-mon[51914]: from='mgr.14156 192.168.123.100:0/929894269' entity='mgr.a' 2026-03-06T13:41:02.002 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:41:01 vm00 ceph-mon[51914]: Added host vm08 2026-03-06T13:41:02.002 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:41:01 vm00 ceph-mon[51914]: from='mgr.14156 192.168.123.100:0/929894269' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-06T13:41:02.002 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:41:01 vm00 ceph-mon[51914]: from='mgr.14156 192.168.123.100:0/929894269' entity='mgr.a' 2026-03-06T13:41:02.002 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:41:01 vm00 ceph-mon[51914]: from='mgr.14156 192.168.123.100:0/929894269' entity='mgr.a' 2026-03-06T13:41:02.172 INFO:teuthology.orchestra.run.vm00.stderr:Inferring config /var/lib/ceph/82f504c8-1959-11f1-9a16-a78d135e200b/mon.a/config 2026-03-06T13:41:03.252 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:41:02 vm00 ceph-mon[51914]: from='client.14186 -' entity='client.admin' cmd=[{"prefix": "orch host ls", "target": ["mon-mgr", ""], "format": "json"}]: dispatch 2026-03-06T13:41:03.252 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:41:02 vm00 ceph-mon[51914]: from='mgr.14156 192.168.123.100:0/929894269' entity='mgr.a' 2026-03-06T13:41:03.252 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:41:02 vm00 ceph-mon[51914]: from='client.? 192.168.123.100:0/2352374010' entity='client.admin' cmd=[{"prefix": "osd crush tunables", "profile": "default"}]: dispatch 2026-03-06T13:41:03.395 INFO:teuthology.orchestra.run.vm00.stderr:adjusted tunables profile to default 2026-03-06T13:41:03.449 INFO:tasks.cephadm:Adding mon.a on vm00 2026-03-06T13:41:03.449 INFO:tasks.cephadm:Adding mon.b on vm01 2026-03-06T13:41:03.449 INFO:tasks.cephadm:Adding mon.c on vm08 2026-03-06T13:41:03.449 DEBUG:teuthology.orchestra.run.vm08:> sudo /home/ubuntu/cephtest/cephadm --image harbor.clyso.com/custom-ceph/ceph/ceph:cobaltcore-storage-v19.2.3-fasttrack-3 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 82f504c8-1959-11f1-9a16-a78d135e200b -- ceph orch apply mon '3;vm00:192.168.123.100=a;vm01:192.168.123.101=b;vm08:192.168.123.108=c' 2026-03-06T13:41:03.802 INFO:teuthology.orchestra.run.vm08.stderr:Inferring config /etc/ceph/ceph.conf 2026-03-06T13:41:03.849 INFO:teuthology.orchestra.run.vm08.stderr:Inferring config /etc/ceph/ceph.conf 2026-03-06T13:41:04.223 INFO:teuthology.orchestra.run.vm08.stdout:Scheduled mon update... 2026-03-06T13:41:04.300 DEBUG:teuthology.orchestra.run.vm01:mon.b> sudo journalctl -f -n 0 -u ceph-82f504c8-1959-11f1-9a16-a78d135e200b@mon.b.service 2026-03-06T13:41:04.303 DEBUG:teuthology.orchestra.run.vm08:mon.c> sudo journalctl -f -n 0 -u ceph-82f504c8-1959-11f1-9a16-a78d135e200b@mon.c.service 2026-03-06T13:41:04.304 INFO:tasks.cephadm:Waiting for 3 mons in monmap... 2026-03-06T13:41:04.304 DEBUG:teuthology.orchestra.run.vm08:> sudo /home/ubuntu/cephtest/cephadm --image harbor.clyso.com/custom-ceph/ceph/ceph:cobaltcore-storage-v19.2.3-fasttrack-3 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 82f504c8-1959-11f1-9a16-a78d135e200b -- ceph mon dump -f json 2026-03-06T13:41:04.740 INFO:teuthology.orchestra.run.vm08.stderr:Inferring config /var/lib/ceph/82f504c8-1959-11f1-9a16-a78d135e200b/config/ceph.conf 2026-03-06T13:41:04.752 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:41:04 vm00 ceph-mon[51914]: from='client.? 192.168.123.100:0/2352374010' entity='client.admin' cmd='[{"prefix": "osd crush tunables", "profile": "default"}]': finished 2026-03-06T13:41:04.752 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:41:04 vm00 ceph-mon[51914]: osdmap e4: 0 total, 0 up, 0 in 2026-03-06T13:41:04.752 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:41:04 vm00 ceph-mon[51914]: from='mgr.14156 192.168.123.100:0/929894269' entity='mgr.a' 2026-03-06T13:41:04.752 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:41:04 vm00 ceph-mon[51914]: from='mgr.14156 192.168.123.100:0/929894269' entity='mgr.a' 2026-03-06T13:41:05.125 INFO:teuthology.orchestra.run.vm08.stdout: 2026-03-06T13:41:05.125 INFO:teuthology.orchestra.run.vm08.stdout:{"epoch":1,"fsid":"82f504c8-1959-11f1-9a16-a78d135e200b","modified":"2026-03-06T12:40:02.133525Z","created":"2026-03-06T12:40:02.133525Z","min_mon_release":19,"min_mon_release_name":"squid","election_strategy":1,"disallowed_leaders":"","stretch_mode":false,"tiebreaker_mon":"","removed_ranks":"","features":{"persistent":["kraken","luminous","mimic","osdmap-prune","nautilus","octopus","pacific","elector-pinging","quincy","reef","squid"],"optional":[]},"mons":[{"rank":0,"name":"a","public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.100:3300","nonce":0},{"type":"v1","addr":"192.168.123.100:6789","nonce":0}]},"addr":"192.168.123.100:6789/0","public_addr":"192.168.123.100:6789/0","priority":0,"weight":0,"crush_location":"{}"}],"quorum":[0]} 2026-03-06T13:41:05.125 INFO:teuthology.orchestra.run.vm08.stderr:dumped monmap epoch 1 2026-03-06T13:41:05.752 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:41:05 vm00 ceph-mon[51914]: from='client.14190 -' entity='client.admin' cmd=[{"prefix": "orch apply", "service_type": "mon", "placement": "3;vm00:192.168.123.100=a;vm01:192.168.123.101=b;vm08:192.168.123.108=c", "target": ["mon-mgr", ""]}]: dispatch 2026-03-06T13:41:05.752 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:41:05 vm00 ceph-mon[51914]: Saving service mon spec with placement vm00:192.168.123.100=a;vm01:192.168.123.101=b;vm08:192.168.123.108=c;count:3 2026-03-06T13:41:05.752 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:41:05 vm00 ceph-mon[51914]: from='mgr.14156 192.168.123.100:0/929894269' entity='mgr.a' 2026-03-06T13:41:05.752 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:41:05 vm00 ceph-mon[51914]: from='mgr.14156 192.168.123.100:0/929894269' entity='mgr.a' 2026-03-06T13:41:05.752 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:41:05 vm00 ceph-mon[51914]: from='mgr.14156 192.168.123.100:0/929894269' entity='mgr.a' 2026-03-06T13:41:05.752 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:41:05 vm00 ceph-mon[51914]: from='mgr.14156 192.168.123.100:0/929894269' entity='mgr.a' cmd=[{"prefix": "config rm", "who": "osd/host:vm08", "name": "osd_memory_target"}]: dispatch 2026-03-06T13:41:05.752 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:41:05 vm00 ceph-mon[51914]: from='mgr.14156 192.168.123.100:0/929894269' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-06T13:41:05.752 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:41:05 vm00 ceph-mon[51914]: from='mgr.14156 192.168.123.100:0/929894269' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-06T13:41:05.752 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:41:05 vm00 ceph-mon[51914]: Updating vm08:/etc/ceph/ceph.conf 2026-03-06T13:41:05.752 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:41:05 vm00 ceph-mon[51914]: Updating vm08:/var/lib/ceph/82f504c8-1959-11f1-9a16-a78d135e200b/config/ceph.conf 2026-03-06T13:41:05.752 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:41:05 vm00 ceph-mon[51914]: Updating vm08:/etc/ceph/ceph.client.admin.keyring 2026-03-06T13:41:05.752 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:41:05 vm00 ceph-mon[51914]: Updating vm08:/var/lib/ceph/82f504c8-1959-11f1-9a16-a78d135e200b/config/ceph.client.admin.keyring 2026-03-06T13:41:05.752 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:41:05 vm00 ceph-mon[51914]: from='mgr.14156 192.168.123.100:0/929894269' entity='mgr.a' 2026-03-06T13:41:05.752 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:41:05 vm00 ceph-mon[51914]: from='mgr.14156 192.168.123.100:0/929894269' entity='mgr.a' 2026-03-06T13:41:05.752 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:41:05 vm00 ceph-mon[51914]: from='mgr.14156 192.168.123.100:0/929894269' entity='mgr.a' 2026-03-06T13:41:05.752 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:41:05 vm00 ceph-mon[51914]: from='mgr.14156 192.168.123.100:0/929894269' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "mon."}]: dispatch 2026-03-06T13:41:05.752 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:41:05 vm00 ceph-mon[51914]: from='mgr.14156 192.168.123.100:0/929894269' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-06T13:41:05.752 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:41:05 vm00 ceph-mon[51914]: Deploying daemon mon.c on vm08 2026-03-06T13:41:05.752 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:41:05 vm00 ceph-mon[51914]: from='client.? 192.168.123.108:0/3875129221' entity='client.admin' cmd=[{"prefix": "mon dump", "format": "json"}]: dispatch 2026-03-06T13:41:06.194 INFO:tasks.cephadm:Waiting for 3 mons in monmap... 2026-03-06T13:41:06.194 DEBUG:teuthology.orchestra.run.vm08:> sudo /home/ubuntu/cephtest/cephadm --image harbor.clyso.com/custom-ceph/ceph/ceph:cobaltcore-storage-v19.2.3-fasttrack-3 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 82f504c8-1959-11f1-9a16-a78d135e200b -- ceph mon dump -f json 2026-03-06T13:41:06.803 INFO:teuthology.orchestra.run.vm08.stderr:Inferring config /var/lib/ceph/82f504c8-1959-11f1-9a16-a78d135e200b/mon.c/config 2026-03-06T13:41:12.252 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:41:11 vm00 ceph-mon[51914]: from='mgr.14156 192.168.123.100:0/929894269' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "a"}]: dispatch 2026-03-06T13:41:12.252 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:41:11 vm00 ceph-mon[51914]: mon.a calling monitor election 2026-03-06T13:41:12.252 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:41:11 vm00 ceph-mon[51914]: from='mgr.14156 192.168.123.100:0/929894269' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "c"}]: dispatch 2026-03-06T13:41:12.252 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:41:11 vm00 ceph-mon[51914]: from='mgr.14156 192.168.123.100:0/929894269' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "c"}]: dispatch 2026-03-06T13:41:12.252 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:41:11 vm00 ceph-mon[51914]: pgmap v4: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-06T13:41:12.252 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:41:11 vm00 ceph-mon[51914]: from='mgr.14156 192.168.123.100:0/929894269' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "c"}]: dispatch 2026-03-06T13:41:12.252 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:41:11 vm00 ceph-mon[51914]: mon.c calling monitor election 2026-03-06T13:41:12.252 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:41:11 vm00 ceph-mon[51914]: from='mgr.14156 192.168.123.100:0/929894269' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "c"}]: dispatch 2026-03-06T13:41:12.252 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:41:11 vm00 ceph-mon[51914]: pgmap v5: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-06T13:41:12.252 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:41:11 vm00 ceph-mon[51914]: from='mgr.14156 192.168.123.100:0/929894269' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "c"}]: dispatch 2026-03-06T13:41:12.252 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:41:11 vm00 ceph-mon[51914]: from='mgr.14156 192.168.123.100:0/929894269' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "c"}]: dispatch 2026-03-06T13:41:12.252 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:41:11 vm00 ceph-mon[51914]: mon.a is new leader, mons a,c in quorum (ranks 0,1) 2026-03-06T13:41:12.252 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:41:11 vm00 ceph-mon[51914]: monmap epoch 2 2026-03-06T13:41:12.252 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:41:11 vm00 ceph-mon[51914]: fsid 82f504c8-1959-11f1-9a16-a78d135e200b 2026-03-06T13:41:12.252 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:41:11 vm00 ceph-mon[51914]: last_changed 2026-03-06T12:41:06.822758+0000 2026-03-06T13:41:12.252 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:41:11 vm00 ceph-mon[51914]: created 2026-03-06T12:40:02.133525+0000 2026-03-06T13:41:12.252 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:41:11 vm00 ceph-mon[51914]: min_mon_release 19 (squid) 2026-03-06T13:41:12.252 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:41:11 vm00 ceph-mon[51914]: election_strategy: 1 2026-03-06T13:41:12.252 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:41:11 vm00 ceph-mon[51914]: 0: [v2:192.168.123.100:3300/0,v1:192.168.123.100:6789/0] mon.a 2026-03-06T13:41:12.252 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:41:11 vm00 ceph-mon[51914]: 1: [v2:192.168.123.108:3300/0,v1:192.168.123.108:6789/0] mon.c 2026-03-06T13:41:12.252 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:41:11 vm00 ceph-mon[51914]: fsmap 2026-03-06T13:41:12.252 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:41:11 vm00 ceph-mon[51914]: osdmap e4: 0 total, 0 up, 0 in 2026-03-06T13:41:12.252 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:41:11 vm00 ceph-mon[51914]: mgrmap e13: a(active, since 23s) 2026-03-06T13:41:12.252 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:41:11 vm00 ceph-mon[51914]: overall HEALTH_OK 2026-03-06T13:41:12.252 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:41:11 vm00 ceph-mon[51914]: from='mgr.14156 192.168.123.100:0/929894269' entity='mgr.a' 2026-03-06T13:41:12.252 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:41:11 vm00 ceph-mon[51914]: from='mgr.14156 192.168.123.100:0/929894269' entity='mgr.a' 2026-03-06T13:41:12.252 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:41:11 vm00 ceph-mon[51914]: from='mgr.14156 192.168.123.100:0/929894269' entity='mgr.a' 2026-03-06T13:41:12.253 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:41:11 vm00 ceph-mon[51914]: from='mgr.14156 192.168.123.100:0/929894269' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "mon."}]: dispatch 2026-03-06T13:41:12.253 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:41:11 vm00 ceph-mon[51914]: from='mgr.14156 192.168.123.100:0/929894269' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-06T13:41:12.291 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 06 13:41:11 vm08 ceph-mon[53561]: from='mgr.14156 192.168.123.100:0/929894269' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-06T13:41:13.206 INFO:teuthology.orchestra.run.vm08.stdout: 2026-03-06T13:41:13.206 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 06 13:41:12 vm08 ceph-mon[53561]: Deploying daemon mon.b on vm01 2026-03-06T13:41:13.206 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 06 13:41:12 vm08 ceph-mon[53561]: pgmap v6: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-06T13:41:13.206 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 06 13:41:12 vm08 ceph-mon[53561]: from='mgr.14156 192.168.123.100:0/929894269' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "c"}]: dispatch 2026-03-06T13:41:13.207 INFO:teuthology.orchestra.run.vm08.stdout:{"epoch":2,"fsid":"82f504c8-1959-11f1-9a16-a78d135e200b","modified":"2026-03-06T12:41:06.822758Z","created":"2026-03-06T12:40:02.133525Z","min_mon_release":19,"min_mon_release_name":"squid","election_strategy":1,"disallowed_leaders":"","stretch_mode":false,"tiebreaker_mon":"","removed_ranks":"","features":{"persistent":["kraken","luminous","mimic","osdmap-prune","nautilus","octopus","pacific","elector-pinging","quincy","reef","squid"],"optional":[]},"mons":[{"rank":0,"name":"a","public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.100:3300","nonce":0},{"type":"v1","addr":"192.168.123.100:6789","nonce":0}]},"addr":"192.168.123.100:6789/0","public_addr":"192.168.123.100:6789/0","priority":0,"weight":0,"crush_location":"{}"},{"rank":1,"name":"c","public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.108:3300","nonce":0},{"type":"v1","addr":"192.168.123.108:6789","nonce":0}]},"addr":"192.168.123.108:6789/0","public_addr":"192.168.123.108:6789/0","priority":0,"weight":0,"crush_location":"{}"}],"quorum":[0,1]} 2026-03-06T13:41:13.207 INFO:teuthology.orchestra.run.vm08.stderr:dumped monmap epoch 2 2026-03-06T13:41:13.252 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:41:12 vm00 ceph-mon[51914]: Deploying daemon mon.b on vm01 2026-03-06T13:41:13.252 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:41:12 vm00 ceph-mon[51914]: pgmap v6: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-06T13:41:13.252 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:41:12 vm00 ceph-mon[51914]: from='mgr.14156 192.168.123.100:0/929894269' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "c"}]: dispatch 2026-03-06T13:41:13.655 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:41:13 vm01 ceph-mon[56331]: mon.b@-1(synchronizing).paxosservice(auth 1..3) refresh upgraded, format 0 -> 3 2026-03-06T13:41:14.252 INFO:journalctl@ceph.mgr.a.vm00.stdout:Mar 06 13:41:13 vm00 ceph-82f504c8-1959-11f1-9a16-a78d135e200b-mgr-a[52124]: 2026-03-06T12:41:13.820+0000 7f8ecd580640 -1 mgr.server handle_report got status from non-daemon mon.c 2026-03-06T13:41:14.268 INFO:tasks.cephadm:Waiting for 3 mons in monmap... 2026-03-06T13:41:14.268 DEBUG:teuthology.orchestra.run.vm08:> sudo /home/ubuntu/cephtest/cephadm --image harbor.clyso.com/custom-ceph/ceph/ceph:cobaltcore-storage-v19.2.3-fasttrack-3 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 82f504c8-1959-11f1-9a16-a78d135e200b -- ceph mon dump -f json 2026-03-06T13:41:14.594 INFO:teuthology.orchestra.run.vm08.stderr:Inferring config /var/lib/ceph/82f504c8-1959-11f1-9a16-a78d135e200b/mon.c/config 2026-03-06T13:41:19.002 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:41:18 vm00 ceph-mon[51914]: from='mgr.14156 192.168.123.100:0/929894269' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "a"}]: dispatch 2026-03-06T13:41:19.002 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:41:18 vm00 ceph-mon[51914]: from='mgr.14156 192.168.123.100:0/929894269' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-06T13:41:19.002 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:41:18 vm00 ceph-mon[51914]: from='mgr.14156 192.168.123.100:0/929894269' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "c"}]: dispatch 2026-03-06T13:41:19.002 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:41:18 vm00 ceph-mon[51914]: mon.a calling monitor election 2026-03-06T13:41:19.002 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:41:18 vm00 ceph-mon[51914]: mon.c calling monitor election 2026-03-06T13:41:19.002 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:41:18 vm00 ceph-mon[51914]: from='mgr.14156 192.168.123.100:0/929894269' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-06T13:41:19.002 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:41:18 vm00 ceph-mon[51914]: pgmap v7: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-06T13:41:19.003 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:41:18 vm00 ceph-mon[51914]: from='mgr.14156 192.168.123.100:0/929894269' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-06T13:41:19.003 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:41:18 vm00 ceph-mon[51914]: from='mgr.14156 192.168.123.100:0/929894269' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-06T13:41:19.003 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:41:18 vm00 ceph-mon[51914]: mon.b calling monitor election 2026-03-06T13:41:19.003 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:41:18 vm00 ceph-mon[51914]: pgmap v8: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-06T13:41:19.003 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:41:18 vm00 ceph-mon[51914]: from='mgr.14156 192.168.123.100:0/929894269' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-06T13:41:19.003 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:41:18 vm00 ceph-mon[51914]: from='mgr.14156 192.168.123.100:0/929894269' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-06T13:41:19.003 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:41:18 vm00 ceph-mon[51914]: pgmap v9: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-06T13:41:19.003 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:41:18 vm00 ceph-mon[51914]: from='mgr.14156 192.168.123.100:0/929894269' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-06T13:41:19.003 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:41:18 vm00 ceph-mon[51914]: mon.a is new leader, mons a,c,b in quorum (ranks 0,1,2) 2026-03-06T13:41:19.003 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:41:18 vm00 ceph-mon[51914]: monmap epoch 3 2026-03-06T13:41:19.003 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:41:18 vm00 ceph-mon[51914]: fsid 82f504c8-1959-11f1-9a16-a78d135e200b 2026-03-06T13:41:19.003 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:41:18 vm00 ceph-mon[51914]: last_changed 2026-03-06T12:41:13.598792+0000 2026-03-06T13:41:19.003 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:41:18 vm00 ceph-mon[51914]: created 2026-03-06T12:40:02.133525+0000 2026-03-06T13:41:19.003 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:41:18 vm00 ceph-mon[51914]: min_mon_release 19 (squid) 2026-03-06T13:41:19.003 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:41:18 vm00 ceph-mon[51914]: election_strategy: 1 2026-03-06T13:41:19.003 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:41:18 vm00 ceph-mon[51914]: 0: [v2:192.168.123.100:3300/0,v1:192.168.123.100:6789/0] mon.a 2026-03-06T13:41:19.003 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:41:18 vm00 ceph-mon[51914]: 1: [v2:192.168.123.108:3300/0,v1:192.168.123.108:6789/0] mon.c 2026-03-06T13:41:19.003 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:41:18 vm00 ceph-mon[51914]: 2: [v2:192.168.123.101:3300/0,v1:192.168.123.101:6789/0] mon.b 2026-03-06T13:41:19.003 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:41:18 vm00 ceph-mon[51914]: fsmap 2026-03-06T13:41:19.003 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:41:18 vm00 ceph-mon[51914]: osdmap e4: 0 total, 0 up, 0 in 2026-03-06T13:41:19.003 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:41:18 vm00 ceph-mon[51914]: mgrmap e13: a(active, since 30s) 2026-03-06T13:41:19.003 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:41:18 vm00 ceph-mon[51914]: overall HEALTH_OK 2026-03-06T13:41:19.003 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:41:18 vm00 ceph-mon[51914]: from='mgr.14156 192.168.123.100:0/929894269' entity='mgr.a' 2026-03-06T13:41:19.040 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 06 13:41:18 vm08 ceph-mon[53561]: from='mgr.14156 192.168.123.100:0/929894269' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "a"}]: dispatch 2026-03-06T13:41:19.040 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 06 13:41:18 vm08 ceph-mon[53561]: from='mgr.14156 192.168.123.100:0/929894269' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-06T13:41:19.040 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 06 13:41:18 vm08 ceph-mon[53561]: from='mgr.14156 192.168.123.100:0/929894269' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "c"}]: dispatch 2026-03-06T13:41:19.040 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 06 13:41:18 vm08 ceph-mon[53561]: mon.a calling monitor election 2026-03-06T13:41:19.040 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 06 13:41:18 vm08 ceph-mon[53561]: mon.c calling monitor election 2026-03-06T13:41:19.040 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 06 13:41:18 vm08 ceph-mon[53561]: from='mgr.14156 192.168.123.100:0/929894269' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-06T13:41:19.040 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 06 13:41:18 vm08 ceph-mon[53561]: pgmap v7: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-06T13:41:19.040 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 06 13:41:18 vm08 ceph-mon[53561]: from='mgr.14156 192.168.123.100:0/929894269' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-06T13:41:19.040 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 06 13:41:18 vm08 ceph-mon[53561]: from='mgr.14156 192.168.123.100:0/929894269' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-06T13:41:19.040 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 06 13:41:18 vm08 ceph-mon[53561]: mon.b calling monitor election 2026-03-06T13:41:19.041 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 06 13:41:18 vm08 ceph-mon[53561]: pgmap v8: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-06T13:41:19.041 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 06 13:41:18 vm08 ceph-mon[53561]: from='mgr.14156 192.168.123.100:0/929894269' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-06T13:41:19.041 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 06 13:41:18 vm08 ceph-mon[53561]: from='mgr.14156 192.168.123.100:0/929894269' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-06T13:41:19.041 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 06 13:41:18 vm08 ceph-mon[53561]: pgmap v9: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-06T13:41:19.041 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 06 13:41:18 vm08 ceph-mon[53561]: from='mgr.14156 192.168.123.100:0/929894269' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-06T13:41:19.041 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 06 13:41:18 vm08 ceph-mon[53561]: mon.a is new leader, mons a,c,b in quorum (ranks 0,1,2) 2026-03-06T13:41:19.041 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 06 13:41:18 vm08 ceph-mon[53561]: monmap epoch 3 2026-03-06T13:41:19.041 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 06 13:41:18 vm08 ceph-mon[53561]: fsid 82f504c8-1959-11f1-9a16-a78d135e200b 2026-03-06T13:41:19.041 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 06 13:41:18 vm08 ceph-mon[53561]: last_changed 2026-03-06T12:41:13.598792+0000 2026-03-06T13:41:19.041 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 06 13:41:18 vm08 ceph-mon[53561]: created 2026-03-06T12:40:02.133525+0000 2026-03-06T13:41:19.041 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 06 13:41:18 vm08 ceph-mon[53561]: min_mon_release 19 (squid) 2026-03-06T13:41:19.041 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 06 13:41:18 vm08 ceph-mon[53561]: election_strategy: 1 2026-03-06T13:41:19.041 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 06 13:41:18 vm08 ceph-mon[53561]: 0: [v2:192.168.123.100:3300/0,v1:192.168.123.100:6789/0] mon.a 2026-03-06T13:41:19.041 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 06 13:41:18 vm08 ceph-mon[53561]: 1: [v2:192.168.123.108:3300/0,v1:192.168.123.108:6789/0] mon.c 2026-03-06T13:41:19.041 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 06 13:41:18 vm08 ceph-mon[53561]: 2: [v2:192.168.123.101:3300/0,v1:192.168.123.101:6789/0] mon.b 2026-03-06T13:41:19.041 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 06 13:41:18 vm08 ceph-mon[53561]: fsmap 2026-03-06T13:41:19.041 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 06 13:41:18 vm08 ceph-mon[53561]: osdmap e4: 0 total, 0 up, 0 in 2026-03-06T13:41:19.041 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 06 13:41:18 vm08 ceph-mon[53561]: mgrmap e13: a(active, since 30s) 2026-03-06T13:41:19.041 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 06 13:41:18 vm08 ceph-mon[53561]: overall HEALTH_OK 2026-03-06T13:41:19.041 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 06 13:41:18 vm08 ceph-mon[53561]: from='mgr.14156 192.168.123.100:0/929894269' entity='mgr.a' 2026-03-06T13:41:19.107 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:41:18 vm01 ceph-mon[56331]: from='mgr.14156 192.168.123.100:0/929894269' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "a"}]: dispatch 2026-03-06T13:41:19.107 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:41:18 vm01 ceph-mon[56331]: from='mgr.14156 192.168.123.100:0/929894269' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-06T13:41:19.107 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:41:18 vm01 ceph-mon[56331]: from='mgr.14156 192.168.123.100:0/929894269' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "c"}]: dispatch 2026-03-06T13:41:19.107 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:41:18 vm01 ceph-mon[56331]: mon.a calling monitor election 2026-03-06T13:41:19.107 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:41:18 vm01 ceph-mon[56331]: mon.c calling monitor election 2026-03-06T13:41:19.107 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:41:18 vm01 ceph-mon[56331]: from='mgr.14156 192.168.123.100:0/929894269' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-06T13:41:19.107 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:41:18 vm01 ceph-mon[56331]: pgmap v7: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-06T13:41:19.107 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:41:18 vm01 ceph-mon[56331]: from='mgr.14156 192.168.123.100:0/929894269' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-06T13:41:19.107 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:41:18 vm01 ceph-mon[56331]: from='mgr.14156 192.168.123.100:0/929894269' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-06T13:41:19.107 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:41:18 vm01 ceph-mon[56331]: mon.b calling monitor election 2026-03-06T13:41:19.107 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:41:18 vm01 ceph-mon[56331]: pgmap v8: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-06T13:41:19.107 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:41:18 vm01 ceph-mon[56331]: from='mgr.14156 192.168.123.100:0/929894269' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-06T13:41:19.107 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:41:18 vm01 ceph-mon[56331]: from='mgr.14156 192.168.123.100:0/929894269' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-06T13:41:19.107 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:41:18 vm01 ceph-mon[56331]: pgmap v9: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-06T13:41:19.107 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:41:18 vm01 ceph-mon[56331]: from='mgr.14156 192.168.123.100:0/929894269' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-06T13:41:19.107 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:41:18 vm01 ceph-mon[56331]: mon.a is new leader, mons a,c,b in quorum (ranks 0,1,2) 2026-03-06T13:41:19.107 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:41:18 vm01 ceph-mon[56331]: monmap epoch 3 2026-03-06T13:41:19.108 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:41:18 vm01 ceph-mon[56331]: fsid 82f504c8-1959-11f1-9a16-a78d135e200b 2026-03-06T13:41:19.108 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:41:18 vm01 ceph-mon[56331]: last_changed 2026-03-06T12:41:13.598792+0000 2026-03-06T13:41:19.108 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:41:18 vm01 ceph-mon[56331]: created 2026-03-06T12:40:02.133525+0000 2026-03-06T13:41:19.108 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:41:18 vm01 ceph-mon[56331]: min_mon_release 19 (squid) 2026-03-06T13:41:19.108 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:41:18 vm01 ceph-mon[56331]: election_strategy: 1 2026-03-06T13:41:19.108 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:41:18 vm01 ceph-mon[56331]: 0: [v2:192.168.123.100:3300/0,v1:192.168.123.100:6789/0] mon.a 2026-03-06T13:41:19.108 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:41:18 vm01 ceph-mon[56331]: 1: [v2:192.168.123.108:3300/0,v1:192.168.123.108:6789/0] mon.c 2026-03-06T13:41:19.108 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:41:18 vm01 ceph-mon[56331]: 2: [v2:192.168.123.101:3300/0,v1:192.168.123.101:6789/0] mon.b 2026-03-06T13:41:19.108 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:41:18 vm01 ceph-mon[56331]: fsmap 2026-03-06T13:41:19.108 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:41:18 vm01 ceph-mon[56331]: osdmap e4: 0 total, 0 up, 0 in 2026-03-06T13:41:19.108 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:41:18 vm01 ceph-mon[56331]: mgrmap e13: a(active, since 30s) 2026-03-06T13:41:19.108 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:41:18 vm01 ceph-mon[56331]: overall HEALTH_OK 2026-03-06T13:41:19.108 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:41:18 vm01 ceph-mon[56331]: from='mgr.14156 192.168.123.100:0/929894269' entity='mgr.a' 2026-03-06T13:41:19.432 INFO:teuthology.orchestra.run.vm08.stdout: 2026-03-06T13:41:19.433 INFO:teuthology.orchestra.run.vm08.stdout:{"epoch":3,"fsid":"82f504c8-1959-11f1-9a16-a78d135e200b","modified":"2026-03-06T12:41:13.598792Z","created":"2026-03-06T12:40:02.133525Z","min_mon_release":19,"min_mon_release_name":"squid","election_strategy":1,"disallowed_leaders":"","stretch_mode":false,"tiebreaker_mon":"","removed_ranks":"","features":{"persistent":["kraken","luminous","mimic","osdmap-prune","nautilus","octopus","pacific","elector-pinging","quincy","reef","squid"],"optional":[]},"mons":[{"rank":0,"name":"a","public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.100:3300","nonce":0},{"type":"v1","addr":"192.168.123.100:6789","nonce":0}]},"addr":"192.168.123.100:6789/0","public_addr":"192.168.123.100:6789/0","priority":0,"weight":0,"crush_location":"{}"},{"rank":1,"name":"c","public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.108:3300","nonce":0},{"type":"v1","addr":"192.168.123.108:6789","nonce":0}]},"addr":"192.168.123.108:6789/0","public_addr":"192.168.123.108:6789/0","priority":0,"weight":0,"crush_location":"{}"},{"rank":2,"name":"b","public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.101:3300","nonce":0},{"type":"v1","addr":"192.168.123.101:6789","nonce":0}]},"addr":"192.168.123.101:6789/0","public_addr":"192.168.123.101:6789/0","priority":0,"weight":0,"crush_location":"{}"}],"quorum":[0,1,2]} 2026-03-06T13:41:19.433 INFO:teuthology.orchestra.run.vm08.stderr:dumped monmap epoch 3 2026-03-06T13:41:19.518 INFO:tasks.cephadm:Generating final ceph.conf file... 2026-03-06T13:41:19.519 DEBUG:teuthology.orchestra.run.vm00:> sudo /home/ubuntu/cephtest/cephadm --image harbor.clyso.com/custom-ceph/ceph/ceph:cobaltcore-storage-v19.2.3-fasttrack-3 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 82f504c8-1959-11f1-9a16-a78d135e200b -- ceph config generate-minimal-conf 2026-03-06T13:41:19.932 INFO:teuthology.orchestra.run.vm00.stderr:Inferring config /var/lib/ceph/82f504c8-1959-11f1-9a16-a78d135e200b/mon.a/config 2026-03-06T13:41:20.356 INFO:teuthology.orchestra.run.vm00.stdout:# minimal ceph.conf for 82f504c8-1959-11f1-9a16-a78d135e200b 2026-03-06T13:41:20.357 INFO:teuthology.orchestra.run.vm00.stdout:[global] 2026-03-06T13:41:20.357 INFO:teuthology.orchestra.run.vm00.stdout: fsid = 82f504c8-1959-11f1-9a16-a78d135e200b 2026-03-06T13:41:20.357 INFO:teuthology.orchestra.run.vm00.stdout: mon_host = [v2:192.168.123.100:3300/0,v1:192.168.123.100:6789/0] [v2:192.168.123.101:3300/0,v1:192.168.123.101:6789/0] [v2:192.168.123.108:3300/0,v1:192.168.123.108:6789/0] 2026-03-06T13:41:20.418 INFO:tasks.cephadm:Distributing (final) config and client.admin keyring... 2026-03-06T13:41:20.418 DEBUG:teuthology.orchestra.run.vm00:> set -ex 2026-03-06T13:41:20.418 DEBUG:teuthology.orchestra.run.vm00:> sudo dd of=/etc/ceph/ceph.conf 2026-03-06T13:41:20.449 DEBUG:teuthology.orchestra.run.vm00:> set -ex 2026-03-06T13:41:20.449 DEBUG:teuthology.orchestra.run.vm00:> sudo dd of=/etc/ceph/ceph.client.admin.keyring 2026-03-06T13:41:20.502 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:41:20 vm00 ceph-mon[51914]: from='mgr.14156 192.168.123.100:0/929894269' entity='mgr.a' 2026-03-06T13:41:20.502 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:41:20 vm00 ceph-mon[51914]: from='mgr.14156 192.168.123.100:0/929894269' entity='mgr.a' 2026-03-06T13:41:20.502 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:41:20 vm00 ceph-mon[51914]: from='mgr.14156 192.168.123.100:0/929894269' entity='mgr.a' 2026-03-06T13:41:20.502 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:41:20 vm00 ceph-mon[51914]: from='mgr.14156 192.168.123.100:0/929894269' entity='mgr.a' 2026-03-06T13:41:20.502 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:41:20 vm00 ceph-mon[51914]: from='mgr.14156 192.168.123.100:0/929894269' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-06T13:41:20.502 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:41:20 vm00 ceph-mon[51914]: from='mgr.14156 192.168.123.100:0/929894269' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-06T13:41:20.503 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:41:20 vm00 ceph-mon[51914]: Updating vm00:/etc/ceph/ceph.conf 2026-03-06T13:41:20.503 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:41:20 vm00 ceph-mon[51914]: Updating vm01:/etc/ceph/ceph.conf 2026-03-06T13:41:20.503 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:41:20 vm00 ceph-mon[51914]: Updating vm08:/etc/ceph/ceph.conf 2026-03-06T13:41:20.503 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:41:20 vm00 ceph-mon[51914]: from='client.? 192.168.123.108:0/1013204975' entity='client.admin' cmd=[{"prefix": "mon dump", "format": "json"}]: dispatch 2026-03-06T13:41:20.503 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:41:20 vm00 ceph-mon[51914]: Updating vm01:/var/lib/ceph/82f504c8-1959-11f1-9a16-a78d135e200b/config/ceph.conf 2026-03-06T13:41:20.503 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:41:20 vm00 ceph-mon[51914]: Updating vm00:/var/lib/ceph/82f504c8-1959-11f1-9a16-a78d135e200b/config/ceph.conf 2026-03-06T13:41:20.503 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:41:20 vm00 ceph-mon[51914]: Updating vm08:/var/lib/ceph/82f504c8-1959-11f1-9a16-a78d135e200b/config/ceph.conf 2026-03-06T13:41:20.503 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:41:20 vm00 ceph-mon[51914]: from='mgr.14156 192.168.123.100:0/929894269' entity='mgr.a' 2026-03-06T13:41:20.503 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:41:20 vm00 ceph-mon[51914]: from='mgr.14156 192.168.123.100:0/929894269' entity='mgr.a' 2026-03-06T13:41:20.503 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:41:20 vm00 ceph-mon[51914]: from='mgr.14156 192.168.123.100:0/929894269' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-06T13:41:20.503 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:41:20 vm00 ceph-mon[51914]: from='mgr.14156 192.168.123.100:0/929894269' entity='mgr.a' 2026-03-06T13:41:20.503 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:41:20 vm00 ceph-mon[51914]: from='mgr.14156 192.168.123.100:0/929894269' entity='mgr.a' 2026-03-06T13:41:20.503 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:41:20 vm00 ceph-mon[51914]: from='mgr.14156 192.168.123.100:0/929894269' entity='mgr.a' 2026-03-06T13:41:20.503 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:41:20 vm00 ceph-mon[51914]: from='mgr.14156 192.168.123.100:0/929894269' entity='mgr.a' 2026-03-06T13:41:20.503 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:41:20 vm00 ceph-mon[51914]: from='mgr.14156 192.168.123.100:0/929894269' entity='mgr.a' 2026-03-06T13:41:20.503 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:41:20 vm00 ceph-mon[51914]: from='mgr.14156 192.168.123.100:0/929894269' entity='mgr.a' 2026-03-06T13:41:20.503 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:41:20 vm00 ceph-mon[51914]: from='mgr.14156 192.168.123.100:0/929894269' entity='mgr.a' 2026-03-06T13:41:20.503 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:41:20 vm00 ceph-mon[51914]: from='mgr.14156 192.168.123.100:0/929894269' entity='mgr.a' 2026-03-06T13:41:20.503 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:41:20 vm00 ceph-mon[51914]: from='mgr.14156 192.168.123.100:0/929894269' entity='mgr.a' 2026-03-06T13:41:20.503 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:41:20 vm00 ceph-mon[51914]: Reconfiguring mon.a (unknown last config time)... 2026-03-06T13:41:20.503 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:41:20 vm00 ceph-mon[51914]: from='mgr.14156 192.168.123.100:0/929894269' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "mon."}]: dispatch 2026-03-06T13:41:20.503 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:41:20 vm00 ceph-mon[51914]: from='mgr.14156 192.168.123.100:0/929894269' entity='mgr.a' cmd=[{"prefix": "config get", "who": "mon", "key": "public_network"}]: dispatch 2026-03-06T13:41:20.503 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:41:20 vm00 ceph-mon[51914]: from='mgr.14156 192.168.123.100:0/929894269' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-06T13:41:20.503 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:41:20 vm00 ceph-mon[51914]: Reconfiguring daemon mon.a on vm00 2026-03-06T13:41:20.503 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:41:20 vm00 ceph-mon[51914]: pgmap v10: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-06T13:41:20.503 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:41:20 vm00 ceph-mon[51914]: from='mgr.14156 192.168.123.100:0/929894269' entity='mgr.a' 2026-03-06T13:41:20.503 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:41:20 vm00 ceph-mon[51914]: from='mgr.14156 192.168.123.100:0/929894269' entity='mgr.a' 2026-03-06T13:41:20.503 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:41:20 vm00 ceph-mon[51914]: from='mgr.14156 192.168.123.100:0/929894269' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "mon."}]: dispatch 2026-03-06T13:41:20.503 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:41:20 vm00 ceph-mon[51914]: from='mgr.14156 192.168.123.100:0/929894269' entity='mgr.a' cmd=[{"prefix": "config get", "who": "mon", "key": "public_network"}]: dispatch 2026-03-06T13:41:20.503 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:41:20 vm00 ceph-mon[51914]: from='mgr.14156 192.168.123.100:0/929894269' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-06T13:41:20.522 DEBUG:teuthology.orchestra.run.vm01:> set -ex 2026-03-06T13:41:20.547 DEBUG:teuthology.orchestra.run.vm01:> sudo dd of=/etc/ceph/ceph.conf 2026-03-06T13:41:20.575 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:41:20 vm01 ceph-mon[56331]: from='mgr.14156 192.168.123.100:0/929894269' entity='mgr.a' 2026-03-06T13:41:20.575 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:41:20 vm01 ceph-mon[56331]: from='mgr.14156 192.168.123.100:0/929894269' entity='mgr.a' 2026-03-06T13:41:20.575 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:41:20 vm01 ceph-mon[56331]: from='mgr.14156 192.168.123.100:0/929894269' entity='mgr.a' 2026-03-06T13:41:20.575 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:41:20 vm01 ceph-mon[56331]: from='mgr.14156 192.168.123.100:0/929894269' entity='mgr.a' 2026-03-06T13:41:20.575 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:41:20 vm01 ceph-mon[56331]: from='mgr.14156 192.168.123.100:0/929894269' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-06T13:41:20.575 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:41:20 vm01 ceph-mon[56331]: from='mgr.14156 192.168.123.100:0/929894269' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-06T13:41:20.575 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:41:20 vm01 ceph-mon[56331]: Updating vm00:/etc/ceph/ceph.conf 2026-03-06T13:41:20.575 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:41:20 vm01 ceph-mon[56331]: Updating vm01:/etc/ceph/ceph.conf 2026-03-06T13:41:20.575 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:41:20 vm01 ceph-mon[56331]: Updating vm08:/etc/ceph/ceph.conf 2026-03-06T13:41:20.575 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:41:20 vm01 ceph-mon[56331]: from='client.? 192.168.123.108:0/1013204975' entity='client.admin' cmd=[{"prefix": "mon dump", "format": "json"}]: dispatch 2026-03-06T13:41:20.575 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:41:20 vm01 ceph-mon[56331]: Updating vm01:/var/lib/ceph/82f504c8-1959-11f1-9a16-a78d135e200b/config/ceph.conf 2026-03-06T13:41:20.575 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:41:20 vm01 ceph-mon[56331]: Updating vm00:/var/lib/ceph/82f504c8-1959-11f1-9a16-a78d135e200b/config/ceph.conf 2026-03-06T13:41:20.575 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:41:20 vm01 ceph-mon[56331]: Updating vm08:/var/lib/ceph/82f504c8-1959-11f1-9a16-a78d135e200b/config/ceph.conf 2026-03-06T13:41:20.575 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:41:20 vm01 ceph-mon[56331]: from='mgr.14156 192.168.123.100:0/929894269' entity='mgr.a' 2026-03-06T13:41:20.575 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:41:20 vm01 ceph-mon[56331]: from='mgr.14156 192.168.123.100:0/929894269' entity='mgr.a' 2026-03-06T13:41:20.575 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:41:20 vm01 ceph-mon[56331]: from='mgr.14156 192.168.123.100:0/929894269' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-06T13:41:20.575 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:41:20 vm01 ceph-mon[56331]: from='mgr.14156 192.168.123.100:0/929894269' entity='mgr.a' 2026-03-06T13:41:20.575 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:41:20 vm01 ceph-mon[56331]: from='mgr.14156 192.168.123.100:0/929894269' entity='mgr.a' 2026-03-06T13:41:20.575 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:41:20 vm01 ceph-mon[56331]: from='mgr.14156 192.168.123.100:0/929894269' entity='mgr.a' 2026-03-06T13:41:20.575 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:41:20 vm01 ceph-mon[56331]: from='mgr.14156 192.168.123.100:0/929894269' entity='mgr.a' 2026-03-06T13:41:20.575 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:41:20 vm01 ceph-mon[56331]: from='mgr.14156 192.168.123.100:0/929894269' entity='mgr.a' 2026-03-06T13:41:20.575 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:41:20 vm01 ceph-mon[56331]: from='mgr.14156 192.168.123.100:0/929894269' entity='mgr.a' 2026-03-06T13:41:20.575 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:41:20 vm01 ceph-mon[56331]: from='mgr.14156 192.168.123.100:0/929894269' entity='mgr.a' 2026-03-06T13:41:20.575 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:41:20 vm01 ceph-mon[56331]: from='mgr.14156 192.168.123.100:0/929894269' entity='mgr.a' 2026-03-06T13:41:20.575 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:41:20 vm01 ceph-mon[56331]: from='mgr.14156 192.168.123.100:0/929894269' entity='mgr.a' 2026-03-06T13:41:20.575 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:41:20 vm01 ceph-mon[56331]: Reconfiguring mon.a (unknown last config time)... 2026-03-06T13:41:20.575 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:41:20 vm01 ceph-mon[56331]: from='mgr.14156 192.168.123.100:0/929894269' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "mon."}]: dispatch 2026-03-06T13:41:20.576 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:41:20 vm01 ceph-mon[56331]: from='mgr.14156 192.168.123.100:0/929894269' entity='mgr.a' cmd=[{"prefix": "config get", "who": "mon", "key": "public_network"}]: dispatch 2026-03-06T13:41:20.576 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:41:20 vm01 ceph-mon[56331]: from='mgr.14156 192.168.123.100:0/929894269' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-06T13:41:20.576 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:41:20 vm01 ceph-mon[56331]: Reconfiguring daemon mon.a on vm00 2026-03-06T13:41:20.576 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:41:20 vm01 ceph-mon[56331]: pgmap v10: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-06T13:41:20.576 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:41:20 vm01 ceph-mon[56331]: from='mgr.14156 192.168.123.100:0/929894269' entity='mgr.a' 2026-03-06T13:41:20.576 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:41:20 vm01 ceph-mon[56331]: from='mgr.14156 192.168.123.100:0/929894269' entity='mgr.a' 2026-03-06T13:41:20.576 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:41:20 vm01 ceph-mon[56331]: from='mgr.14156 192.168.123.100:0/929894269' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "mon."}]: dispatch 2026-03-06T13:41:20.576 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:41:20 vm01 ceph-mon[56331]: from='mgr.14156 192.168.123.100:0/929894269' entity='mgr.a' cmd=[{"prefix": "config get", "who": "mon", "key": "public_network"}]: dispatch 2026-03-06T13:41:20.576 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:41:20 vm01 ceph-mon[56331]: from='mgr.14156 192.168.123.100:0/929894269' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-06T13:41:20.579 DEBUG:teuthology.orchestra.run.vm01:> set -ex 2026-03-06T13:41:20.579 DEBUG:teuthology.orchestra.run.vm01:> sudo dd of=/etc/ceph/ceph.client.admin.keyring 2026-03-06T13:41:20.647 DEBUG:teuthology.orchestra.run.vm08:> set -ex 2026-03-06T13:41:20.647 DEBUG:teuthology.orchestra.run.vm08:> sudo dd of=/etc/ceph/ceph.conf 2026-03-06T13:41:20.671 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 06 13:41:20 vm08 ceph-mon[53561]: from='mgr.14156 192.168.123.100:0/929894269' entity='mgr.a' 2026-03-06T13:41:20.671 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 06 13:41:20 vm08 ceph-mon[53561]: from='mgr.14156 192.168.123.100:0/929894269' entity='mgr.a' 2026-03-06T13:41:20.671 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 06 13:41:20 vm08 ceph-mon[53561]: from='mgr.14156 192.168.123.100:0/929894269' entity='mgr.a' 2026-03-06T13:41:20.671 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 06 13:41:20 vm08 ceph-mon[53561]: from='mgr.14156 192.168.123.100:0/929894269' entity='mgr.a' 2026-03-06T13:41:20.671 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 06 13:41:20 vm08 ceph-mon[53561]: from='mgr.14156 192.168.123.100:0/929894269' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-06T13:41:20.671 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 06 13:41:20 vm08 ceph-mon[53561]: from='mgr.14156 192.168.123.100:0/929894269' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-06T13:41:20.671 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 06 13:41:20 vm08 ceph-mon[53561]: Updating vm00:/etc/ceph/ceph.conf 2026-03-06T13:41:20.671 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 06 13:41:20 vm08 ceph-mon[53561]: Updating vm01:/etc/ceph/ceph.conf 2026-03-06T13:41:20.671 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 06 13:41:20 vm08 ceph-mon[53561]: Updating vm08:/etc/ceph/ceph.conf 2026-03-06T13:41:20.671 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 06 13:41:20 vm08 ceph-mon[53561]: from='client.? 192.168.123.108:0/1013204975' entity='client.admin' cmd=[{"prefix": "mon dump", "format": "json"}]: dispatch 2026-03-06T13:41:20.671 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 06 13:41:20 vm08 ceph-mon[53561]: Updating vm01:/var/lib/ceph/82f504c8-1959-11f1-9a16-a78d135e200b/config/ceph.conf 2026-03-06T13:41:20.671 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 06 13:41:20 vm08 ceph-mon[53561]: Updating vm00:/var/lib/ceph/82f504c8-1959-11f1-9a16-a78d135e200b/config/ceph.conf 2026-03-06T13:41:20.671 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 06 13:41:20 vm08 ceph-mon[53561]: Updating vm08:/var/lib/ceph/82f504c8-1959-11f1-9a16-a78d135e200b/config/ceph.conf 2026-03-06T13:41:20.671 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 06 13:41:20 vm08 ceph-mon[53561]: from='mgr.14156 192.168.123.100:0/929894269' entity='mgr.a' 2026-03-06T13:41:20.671 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 06 13:41:20 vm08 ceph-mon[53561]: from='mgr.14156 192.168.123.100:0/929894269' entity='mgr.a' 2026-03-06T13:41:20.671 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 06 13:41:20 vm08 ceph-mon[53561]: from='mgr.14156 192.168.123.100:0/929894269' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-06T13:41:20.672 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 06 13:41:20 vm08 ceph-mon[53561]: from='mgr.14156 192.168.123.100:0/929894269' entity='mgr.a' 2026-03-06T13:41:20.672 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 06 13:41:20 vm08 ceph-mon[53561]: from='mgr.14156 192.168.123.100:0/929894269' entity='mgr.a' 2026-03-06T13:41:20.672 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 06 13:41:20 vm08 ceph-mon[53561]: from='mgr.14156 192.168.123.100:0/929894269' entity='mgr.a' 2026-03-06T13:41:20.672 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 06 13:41:20 vm08 ceph-mon[53561]: from='mgr.14156 192.168.123.100:0/929894269' entity='mgr.a' 2026-03-06T13:41:20.672 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 06 13:41:20 vm08 ceph-mon[53561]: from='mgr.14156 192.168.123.100:0/929894269' entity='mgr.a' 2026-03-06T13:41:20.672 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 06 13:41:20 vm08 ceph-mon[53561]: from='mgr.14156 192.168.123.100:0/929894269' entity='mgr.a' 2026-03-06T13:41:20.672 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 06 13:41:20 vm08 ceph-mon[53561]: from='mgr.14156 192.168.123.100:0/929894269' entity='mgr.a' 2026-03-06T13:41:20.672 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 06 13:41:20 vm08 ceph-mon[53561]: from='mgr.14156 192.168.123.100:0/929894269' entity='mgr.a' 2026-03-06T13:41:20.672 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 06 13:41:20 vm08 ceph-mon[53561]: from='mgr.14156 192.168.123.100:0/929894269' entity='mgr.a' 2026-03-06T13:41:20.672 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 06 13:41:20 vm08 ceph-mon[53561]: Reconfiguring mon.a (unknown last config time)... 2026-03-06T13:41:20.672 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 06 13:41:20 vm08 ceph-mon[53561]: from='mgr.14156 192.168.123.100:0/929894269' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "mon."}]: dispatch 2026-03-06T13:41:20.672 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 06 13:41:20 vm08 ceph-mon[53561]: from='mgr.14156 192.168.123.100:0/929894269' entity='mgr.a' cmd=[{"prefix": "config get", "who": "mon", "key": "public_network"}]: dispatch 2026-03-06T13:41:20.672 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 06 13:41:20 vm08 ceph-mon[53561]: from='mgr.14156 192.168.123.100:0/929894269' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-06T13:41:20.672 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 06 13:41:20 vm08 ceph-mon[53561]: Reconfiguring daemon mon.a on vm00 2026-03-06T13:41:20.672 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 06 13:41:20 vm08 ceph-mon[53561]: pgmap v10: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-06T13:41:20.672 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 06 13:41:20 vm08 ceph-mon[53561]: from='mgr.14156 192.168.123.100:0/929894269' entity='mgr.a' 2026-03-06T13:41:20.672 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 06 13:41:20 vm08 ceph-mon[53561]: from='mgr.14156 192.168.123.100:0/929894269' entity='mgr.a' 2026-03-06T13:41:20.672 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 06 13:41:20 vm08 ceph-mon[53561]: from='mgr.14156 192.168.123.100:0/929894269' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "mon."}]: dispatch 2026-03-06T13:41:20.672 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 06 13:41:20 vm08 ceph-mon[53561]: from='mgr.14156 192.168.123.100:0/929894269' entity='mgr.a' cmd=[{"prefix": "config get", "who": "mon", "key": "public_network"}]: dispatch 2026-03-06T13:41:20.672 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 06 13:41:20 vm08 ceph-mon[53561]: from='mgr.14156 192.168.123.100:0/929894269' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-06T13:41:20.676 DEBUG:teuthology.orchestra.run.vm08:> set -ex 2026-03-06T13:41:20.676 DEBUG:teuthology.orchestra.run.vm08:> sudo dd of=/etc/ceph/ceph.client.admin.keyring 2026-03-06T13:41:20.741 INFO:tasks.cephadm:Adding mgr.a on vm00 2026-03-06T13:41:20.741 INFO:tasks.cephadm:Adding mgr.b on vm01 2026-03-06T13:41:20.742 DEBUG:teuthology.orchestra.run.vm08:> sudo /home/ubuntu/cephtest/cephadm --image harbor.clyso.com/custom-ceph/ceph/ceph:cobaltcore-storage-v19.2.3-fasttrack-3 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 82f504c8-1959-11f1-9a16-a78d135e200b -- ceph orch apply mgr '2;vm00=a;vm01=b' 2026-03-06T13:41:21.002 INFO:journalctl@ceph.mgr.a.vm00.stdout:Mar 06 13:41:20 vm00 ceph-82f504c8-1959-11f1-9a16-a78d135e200b-mgr-a[52124]: 2026-03-06T12:41:20.597+0000 7f8ecd580640 -1 mgr.server handle_report got status from non-daemon mon.b 2026-03-06T13:41:21.169 INFO:teuthology.orchestra.run.vm08.stderr:Inferring config /var/lib/ceph/82f504c8-1959-11f1-9a16-a78d135e200b/mon.c/config 2026-03-06T13:41:21.527 INFO:teuthology.orchestra.run.vm08.stdout:Scheduled mgr update... 2026-03-06T13:41:21.527 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 06 13:41:21 vm08 ceph-mon[53561]: Reconfiguring mon.b (monmap changed)... 2026-03-06T13:41:21.527 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 06 13:41:21 vm08 ceph-mon[53561]: Reconfiguring daemon mon.b on vm01 2026-03-06T13:41:21.527 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 06 13:41:21 vm08 ceph-mon[53561]: from='client.? 192.168.123.100:0/2876919865' entity='client.admin' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-06T13:41:21.527 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 06 13:41:21 vm08 ceph-mon[53561]: from='mgr.14156 192.168.123.100:0/929894269' entity='mgr.a' 2026-03-06T13:41:21.527 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 06 13:41:21 vm08 ceph-mon[53561]: from='mgr.14156 192.168.123.100:0/929894269' entity='mgr.a' 2026-03-06T13:41:21.527 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 06 13:41:21 vm08 ceph-mon[53561]: Reconfiguring mon.c (monmap changed)... 2026-03-06T13:41:21.527 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 06 13:41:21 vm08 ceph-mon[53561]: from='mgr.14156 192.168.123.100:0/929894269' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "mon."}]: dispatch 2026-03-06T13:41:21.527 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 06 13:41:21 vm08 ceph-mon[53561]: from='mgr.14156 192.168.123.100:0/929894269' entity='mgr.a' cmd=[{"prefix": "config get", "who": "mon", "key": "public_network"}]: dispatch 2026-03-06T13:41:21.527 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 06 13:41:21 vm08 ceph-mon[53561]: from='mgr.14156 192.168.123.100:0/929894269' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-06T13:41:21.527 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 06 13:41:21 vm08 ceph-mon[53561]: Reconfiguring daemon mon.c on vm08 2026-03-06T13:41:21.527 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 06 13:41:21 vm08 ceph-mon[53561]: from='mgr.14156 192.168.123.100:0/929894269' entity='mgr.a' 2026-03-06T13:41:21.527 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 06 13:41:21 vm08 ceph-mon[53561]: from='mgr.14156 192.168.123.100:0/929894269' entity='mgr.a' 2026-03-06T13:41:21.527 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 06 13:41:21 vm08 ceph-mon[53561]: from='mgr.14156 192.168.123.100:0/929894269' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-06T13:41:21.527 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 06 13:41:21 vm08 ceph-mon[53561]: from='mgr.14156 192.168.123.100:0/929894269' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-06T13:41:21.527 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 06 13:41:21 vm08 ceph-mon[53561]: from='mgr.14156 192.168.123.100:0/929894269' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-06T13:41:21.527 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 06 13:41:21 vm08 ceph-mon[53561]: from='mgr.14156 192.168.123.100:0/929894269' entity='mgr.a' 2026-03-06T13:41:21.591 DEBUG:teuthology.orchestra.run.vm01:mgr.b> sudo journalctl -f -n 0 -u ceph-82f504c8-1959-11f1-9a16-a78d135e200b@mgr.b.service 2026-03-06T13:41:21.593 INFO:tasks.cephadm:Deploying OSDs... 2026-03-06T13:41:21.593 DEBUG:teuthology.orchestra.run.vm00:> set -ex 2026-03-06T13:41:21.593 DEBUG:teuthology.orchestra.run.vm00:> dd if=/scratch_devs of=/dev/stdout 2026-03-06T13:41:21.611 DEBUG:teuthology.orchestra.run:got remote process result: 1 2026-03-06T13:41:21.612 DEBUG:teuthology.orchestra.run.vm00:> ls /dev/[sv]d? 2026-03-06T13:41:21.619 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:41:21 vm01 ceph-mon[56331]: Reconfiguring mon.b (monmap changed)... 2026-03-06T13:41:21.619 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:41:21 vm01 ceph-mon[56331]: Reconfiguring daemon mon.b on vm01 2026-03-06T13:41:21.619 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:41:21 vm01 ceph-mon[56331]: from='client.? 192.168.123.100:0/2876919865' entity='client.admin' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-06T13:41:21.619 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:41:21 vm01 ceph-mon[56331]: from='mgr.14156 192.168.123.100:0/929894269' entity='mgr.a' 2026-03-06T13:41:21.619 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:41:21 vm01 ceph-mon[56331]: from='mgr.14156 192.168.123.100:0/929894269' entity='mgr.a' 2026-03-06T13:41:21.620 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:41:21 vm01 ceph-mon[56331]: Reconfiguring mon.c (monmap changed)... 2026-03-06T13:41:21.620 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:41:21 vm01 ceph-mon[56331]: from='mgr.14156 192.168.123.100:0/929894269' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "mon."}]: dispatch 2026-03-06T13:41:21.620 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:41:21 vm01 ceph-mon[56331]: from='mgr.14156 192.168.123.100:0/929894269' entity='mgr.a' cmd=[{"prefix": "config get", "who": "mon", "key": "public_network"}]: dispatch 2026-03-06T13:41:21.620 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:41:21 vm01 ceph-mon[56331]: from='mgr.14156 192.168.123.100:0/929894269' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-06T13:41:21.620 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:41:21 vm01 ceph-mon[56331]: Reconfiguring daemon mon.c on vm08 2026-03-06T13:41:21.620 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:41:21 vm01 ceph-mon[56331]: from='mgr.14156 192.168.123.100:0/929894269' entity='mgr.a' 2026-03-06T13:41:21.620 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:41:21 vm01 ceph-mon[56331]: from='mgr.14156 192.168.123.100:0/929894269' entity='mgr.a' 2026-03-06T13:41:21.620 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:41:21 vm01 ceph-mon[56331]: from='mgr.14156 192.168.123.100:0/929894269' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-06T13:41:21.620 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:41:21 vm01 ceph-mon[56331]: from='mgr.14156 192.168.123.100:0/929894269' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-06T13:41:21.620 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:41:21 vm01 ceph-mon[56331]: from='mgr.14156 192.168.123.100:0/929894269' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-06T13:41:21.620 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:41:21 vm01 ceph-mon[56331]: from='mgr.14156 192.168.123.100:0/929894269' entity='mgr.a' 2026-03-06T13:41:21.670 INFO:teuthology.orchestra.run.vm00.stdout:/dev/vda 2026-03-06T13:41:21.670 INFO:teuthology.orchestra.run.vm00.stdout:/dev/vdb 2026-03-06T13:41:21.671 INFO:teuthology.orchestra.run.vm00.stdout:/dev/vdc 2026-03-06T13:41:21.671 INFO:teuthology.orchestra.run.vm00.stdout:/dev/vdd 2026-03-06T13:41:21.671 INFO:teuthology.orchestra.run.vm00.stdout:/dev/vde 2026-03-06T13:41:21.671 WARNING:teuthology.misc:Removing root device: /dev/vda from device list 2026-03-06T13:41:21.671 DEBUG:teuthology.misc:devs=['/dev/vdb', '/dev/vdc', '/dev/vdd', '/dev/vde'] 2026-03-06T13:41:21.671 DEBUG:teuthology.orchestra.run.vm00:> stat /dev/vdb 2026-03-06T13:41:21.732 INFO:teuthology.orchestra.run.vm00.stdout: File: /dev/vdb 2026-03-06T13:41:21.732 INFO:teuthology.orchestra.run.vm00.stdout: Size: 0 Blocks: 0 IO Block: 512 block special file 2026-03-06T13:41:21.732 INFO:teuthology.orchestra.run.vm00.stdout:Device: 6h/6d Inode: 221 Links: 1 Device type: fc,10 2026-03-06T13:41:21.732 INFO:teuthology.orchestra.run.vm00.stdout:Access: (0660/brw-rw----) Uid: ( 0/ root) Gid: ( 6/ disk) 2026-03-06T13:41:21.732 INFO:teuthology.orchestra.run.vm00.stdout:Context: system_u:object_r:fixed_disk_device_t:s0 2026-03-06T13:41:21.732 INFO:teuthology.orchestra.run.vm00.stdout:Access: 2026-03-06 13:40:52.422589418 +0100 2026-03-06T13:41:21.732 INFO:teuthology.orchestra.run.vm00.stdout:Modify: 2026-03-06 13:38:28.376103554 +0100 2026-03-06T13:41:21.732 INFO:teuthology.orchestra.run.vm00.stdout:Change: 2026-03-06 13:38:28.376103554 +0100 2026-03-06T13:41:21.732 INFO:teuthology.orchestra.run.vm00.stdout: Birth: 2026-03-06 13:34:54.245000000 +0100 2026-03-06T13:41:21.732 DEBUG:teuthology.orchestra.run.vm00:> sudo dd if=/dev/vdb of=/dev/null count=1 2026-03-06T13:41:21.752 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:41:21 vm00 ceph-mon[51914]: Reconfiguring mon.b (monmap changed)... 2026-03-06T13:41:21.752 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:41:21 vm00 ceph-mon[51914]: Reconfiguring daemon mon.b on vm01 2026-03-06T13:41:21.752 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:41:21 vm00 ceph-mon[51914]: from='client.? 192.168.123.100:0/2876919865' entity='client.admin' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-06T13:41:21.752 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:41:21 vm00 ceph-mon[51914]: from='mgr.14156 192.168.123.100:0/929894269' entity='mgr.a' 2026-03-06T13:41:21.752 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:41:21 vm00 ceph-mon[51914]: from='mgr.14156 192.168.123.100:0/929894269' entity='mgr.a' 2026-03-06T13:41:21.752 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:41:21 vm00 ceph-mon[51914]: Reconfiguring mon.c (monmap changed)... 2026-03-06T13:41:21.752 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:41:21 vm00 ceph-mon[51914]: from='mgr.14156 192.168.123.100:0/929894269' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "mon."}]: dispatch 2026-03-06T13:41:21.752 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:41:21 vm00 ceph-mon[51914]: from='mgr.14156 192.168.123.100:0/929894269' entity='mgr.a' cmd=[{"prefix": "config get", "who": "mon", "key": "public_network"}]: dispatch 2026-03-06T13:41:21.752 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:41:21 vm00 ceph-mon[51914]: from='mgr.14156 192.168.123.100:0/929894269' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-06T13:41:21.752 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:41:21 vm00 ceph-mon[51914]: Reconfiguring daemon mon.c on vm08 2026-03-06T13:41:21.752 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:41:21 vm00 ceph-mon[51914]: from='mgr.14156 192.168.123.100:0/929894269' entity='mgr.a' 2026-03-06T13:41:21.752 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:41:21 vm00 ceph-mon[51914]: from='mgr.14156 192.168.123.100:0/929894269' entity='mgr.a' 2026-03-06T13:41:21.752 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:41:21 vm00 ceph-mon[51914]: from='mgr.14156 192.168.123.100:0/929894269' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-06T13:41:21.752 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:41:21 vm00 ceph-mon[51914]: from='mgr.14156 192.168.123.100:0/929894269' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-06T13:41:21.752 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:41:21 vm00 ceph-mon[51914]: from='mgr.14156 192.168.123.100:0/929894269' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-06T13:41:21.752 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:41:21 vm00 ceph-mon[51914]: from='mgr.14156 192.168.123.100:0/929894269' entity='mgr.a' 2026-03-06T13:41:21.779 INFO:teuthology.orchestra.run.vm00.stderr:1+0 records in 2026-03-06T13:41:21.779 INFO:teuthology.orchestra.run.vm00.stderr:1+0 records out 2026-03-06T13:41:21.779 INFO:teuthology.orchestra.run.vm00.stderr:512 bytes copied, 0.000199313 s, 2.6 MB/s 2026-03-06T13:41:21.780 DEBUG:teuthology.orchestra.run.vm00:> ! mount | grep -v devtmpfs | grep -q /dev/vdb 2026-03-06T13:41:21.841 DEBUG:teuthology.orchestra.run.vm00:> stat /dev/vdc 2026-03-06T13:41:21.903 INFO:teuthology.orchestra.run.vm00.stdout: File: /dev/vdc 2026-03-06T13:41:21.903 INFO:teuthology.orchestra.run.vm00.stdout: Size: 0 Blocks: 0 IO Block: 512 block special file 2026-03-06T13:41:21.903 INFO:teuthology.orchestra.run.vm00.stdout:Device: 6h/6d Inode: 224 Links: 1 Device type: fc,20 2026-03-06T13:41:21.903 INFO:teuthology.orchestra.run.vm00.stdout:Access: (0660/brw-rw----) Uid: ( 0/ root) Gid: ( 6/ disk) 2026-03-06T13:41:21.903 INFO:teuthology.orchestra.run.vm00.stdout:Context: system_u:object_r:fixed_disk_device_t:s0 2026-03-06T13:41:21.903 INFO:teuthology.orchestra.run.vm00.stdout:Access: 2026-03-06 13:40:52.461589422 +0100 2026-03-06T13:41:21.903 INFO:teuthology.orchestra.run.vm00.stdout:Modify: 2026-03-06 13:38:28.402103563 +0100 2026-03-06T13:41:21.903 INFO:teuthology.orchestra.run.vm00.stdout:Change: 2026-03-06 13:38:28.402103563 +0100 2026-03-06T13:41:21.903 INFO:teuthology.orchestra.run.vm00.stdout: Birth: 2026-03-06 13:34:54.262000000 +0100 2026-03-06T13:41:21.903 DEBUG:teuthology.orchestra.run.vm00:> sudo dd if=/dev/vdc of=/dev/null count=1 2026-03-06T13:41:21.971 INFO:teuthology.orchestra.run.vm00.stderr:1+0 records in 2026-03-06T13:41:21.971 INFO:teuthology.orchestra.run.vm00.stderr:1+0 records out 2026-03-06T13:41:21.971 INFO:teuthology.orchestra.run.vm00.stderr:512 bytes copied, 0.000117761 s, 4.3 MB/s 2026-03-06T13:41:21.972 DEBUG:teuthology.orchestra.run.vm00:> ! mount | grep -v devtmpfs | grep -q /dev/vdc 2026-03-06T13:41:22.035 DEBUG:teuthology.orchestra.run.vm00:> stat /dev/vdd 2026-03-06T13:41:22.096 INFO:teuthology.orchestra.run.vm00.stdout: File: /dev/vdd 2026-03-06T13:41:22.097 INFO:teuthology.orchestra.run.vm00.stdout: Size: 0 Blocks: 0 IO Block: 512 block special file 2026-03-06T13:41:22.097 INFO:teuthology.orchestra.run.vm00.stdout:Device: 6h/6d Inode: 256 Links: 1 Device type: fc,30 2026-03-06T13:41:22.097 INFO:teuthology.orchestra.run.vm00.stdout:Access: (0660/brw-rw----) Uid: ( 0/ root) Gid: ( 6/ disk) 2026-03-06T13:41:22.097 INFO:teuthology.orchestra.run.vm00.stdout:Context: system_u:object_r:fixed_disk_device_t:s0 2026-03-06T13:41:22.097 INFO:teuthology.orchestra.run.vm00.stdout:Access: 2026-03-06 13:40:52.497589425 +0100 2026-03-06T13:41:22.097 INFO:teuthology.orchestra.run.vm00.stdout:Modify: 2026-03-06 13:38:28.362103550 +0100 2026-03-06T13:41:22.097 INFO:teuthology.orchestra.run.vm00.stdout:Change: 2026-03-06 13:38:28.362103550 +0100 2026-03-06T13:41:22.097 INFO:teuthology.orchestra.run.vm00.stdout: Birth: 2026-03-06 13:34:54.279000000 +0100 2026-03-06T13:41:22.097 DEBUG:teuthology.orchestra.run.vm00:> sudo dd if=/dev/vdd of=/dev/null count=1 2026-03-06T13:41:22.166 INFO:teuthology.orchestra.run.vm00.stderr:1+0 records in 2026-03-06T13:41:22.166 INFO:teuthology.orchestra.run.vm00.stderr:1+0 records out 2026-03-06T13:41:22.166 INFO:teuthology.orchestra.run.vm00.stderr:512 bytes copied, 0.000223498 s, 2.3 MB/s 2026-03-06T13:41:22.168 DEBUG:teuthology.orchestra.run.vm00:> ! mount | grep -v devtmpfs | grep -q /dev/vdd 2026-03-06T13:41:22.229 DEBUG:teuthology.orchestra.run.vm00:> stat /dev/vde 2026-03-06T13:41:22.290 INFO:teuthology.orchestra.run.vm00.stdout: File: /dev/vde 2026-03-06T13:41:22.290 INFO:teuthology.orchestra.run.vm00.stdout: Size: 0 Blocks: 0 IO Block: 512 block special file 2026-03-06T13:41:22.290 INFO:teuthology.orchestra.run.vm00.stdout:Device: 6h/6d Inode: 257 Links: 1 Device type: fc,40 2026-03-06T13:41:22.290 INFO:teuthology.orchestra.run.vm00.stdout:Access: (0660/brw-rw----) Uid: ( 0/ root) Gid: ( 6/ disk) 2026-03-06T13:41:22.290 INFO:teuthology.orchestra.run.vm00.stdout:Context: system_u:object_r:fixed_disk_device_t:s0 2026-03-06T13:41:22.290 INFO:teuthology.orchestra.run.vm00.stdout:Access: 2026-03-06 13:40:52.531589429 +0100 2026-03-06T13:41:22.290 INFO:teuthology.orchestra.run.vm00.stdout:Modify: 2026-03-06 13:38:28.364103551 +0100 2026-03-06T13:41:22.290 INFO:teuthology.orchestra.run.vm00.stdout:Change: 2026-03-06 13:38:28.364103551 +0100 2026-03-06T13:41:22.290 INFO:teuthology.orchestra.run.vm00.stdout: Birth: 2026-03-06 13:34:54.286000000 +0100 2026-03-06T13:41:22.291 DEBUG:teuthology.orchestra.run.vm00:> sudo dd if=/dev/vde of=/dev/null count=1 2026-03-06T13:41:22.355 INFO:teuthology.orchestra.run.vm00.stderr:1+0 records in 2026-03-06T13:41:22.355 INFO:teuthology.orchestra.run.vm00.stderr:1+0 records out 2026-03-06T13:41:22.355 INFO:teuthology.orchestra.run.vm00.stderr:512 bytes copied, 0.000196778 s, 2.6 MB/s 2026-03-06T13:41:22.356 DEBUG:teuthology.orchestra.run.vm00:> ! mount | grep -v devtmpfs | grep -q /dev/vde 2026-03-06T13:41:22.417 DEBUG:teuthology.orchestra.run.vm01:> set -ex 2026-03-06T13:41:22.417 DEBUG:teuthology.orchestra.run.vm01:> dd if=/scratch_devs of=/dev/stdout 2026-03-06T13:41:22.436 DEBUG:teuthology.orchestra.run:got remote process result: 1 2026-03-06T13:41:22.436 DEBUG:teuthology.orchestra.run.vm01:> ls /dev/[sv]d? 2026-03-06T13:41:22.438 INFO:journalctl@ceph.mgr.b.vm01.stdout:Mar 06 13:41:22 vm01 systemd[1]: Starting Ceph mgr.b for 82f504c8-1959-11f1-9a16-a78d135e200b... 2026-03-06T13:41:22.456 INFO:teuthology.orchestra.run.vm01.stdout:/dev/vda 2026-03-06T13:41:22.456 INFO:teuthology.orchestra.run.vm01.stdout:/dev/vdb 2026-03-06T13:41:22.456 INFO:teuthology.orchestra.run.vm01.stdout:/dev/vdc 2026-03-06T13:41:22.456 INFO:teuthology.orchestra.run.vm01.stdout:/dev/vdd 2026-03-06T13:41:22.456 INFO:teuthology.orchestra.run.vm01.stdout:/dev/vde 2026-03-06T13:41:22.456 WARNING:teuthology.misc:Removing root device: /dev/vda from device list 2026-03-06T13:41:22.457 DEBUG:teuthology.misc:devs=['/dev/vdb', '/dev/vdc', '/dev/vdd', '/dev/vde'] 2026-03-06T13:41:22.457 DEBUG:teuthology.orchestra.run.vm01:> stat /dev/vdb 2026-03-06T13:41:22.518 INFO:teuthology.orchestra.run.vm01.stdout: File: /dev/vdb 2026-03-06T13:41:22.518 INFO:teuthology.orchestra.run.vm01.stdout: Size: 0 Blocks: 0 IO Block: 512 block special file 2026-03-06T13:41:22.518 INFO:teuthology.orchestra.run.vm01.stdout:Device: 6h/6d Inode: 254 Links: 1 Device type: fc,10 2026-03-06T13:41:22.518 INFO:teuthology.orchestra.run.vm01.stdout:Access: (0660/brw-rw----) Uid: ( 0/ root) Gid: ( 6/ disk) 2026-03-06T13:41:22.518 INFO:teuthology.orchestra.run.vm01.stdout:Context: system_u:object_r:fixed_disk_device_t:s0 2026-03-06T13:41:22.518 INFO:teuthology.orchestra.run.vm01.stdout:Access: 2026-03-06 13:41:00.143359101 +0100 2026-03-06T13:41:22.518 INFO:teuthology.orchestra.run.vm01.stdout:Modify: 2026-03-06 13:38:28.119112419 +0100 2026-03-06T13:41:22.518 INFO:teuthology.orchestra.run.vm01.stdout:Change: 2026-03-06 13:38:28.119112419 +0100 2026-03-06T13:41:22.518 INFO:teuthology.orchestra.run.vm01.stdout: Birth: 2026-03-06 13:35:31.268000000 +0100 2026-03-06T13:41:22.519 DEBUG:teuthology.orchestra.run.vm01:> sudo dd if=/dev/vdb of=/dev/null count=1 2026-03-06T13:41:22.593 INFO:teuthology.orchestra.run.vm01.stderr:1+0 records in 2026-03-06T13:41:22.593 INFO:teuthology.orchestra.run.vm01.stderr:1+0 records out 2026-03-06T13:41:22.593 INFO:teuthology.orchestra.run.vm01.stderr:512 bytes copied, 0.00147742 s, 347 kB/s 2026-03-06T13:41:22.594 DEBUG:teuthology.orchestra.run.vm01:> ! mount | grep -v devtmpfs | grep -q /dev/vdb 2026-03-06T13:41:22.741 DEBUG:teuthology.orchestra.run.vm01:> stat /dev/vdc 2026-03-06T13:41:22.751 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:41:22 vm01 ceph-mon[56331]: from='client.14214 -' entity='client.admin' cmd=[{"prefix": "orch apply", "service_type": "mgr", "placement": "2;vm00=a;vm01=b", "target": ["mon-mgr", ""]}]: dispatch 2026-03-06T13:41:22.751 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:41:22 vm01 ceph-mon[56331]: Saving service mgr spec with placement vm00=a;vm01=b;count:2 2026-03-06T13:41:22.751 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:41:22 vm01 ceph-mon[56331]: from='mgr.14156 192.168.123.100:0/929894269' entity='mgr.a' 2026-03-06T13:41:22.751 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:41:22 vm01 ceph-mon[56331]: from='mgr.14156 192.168.123.100:0/929894269' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-06T13:41:22.751 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:41:22 vm01 ceph-mon[56331]: from='mgr.14156 192.168.123.100:0/929894269' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-06T13:41:22.751 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:41:22 vm01 ceph-mon[56331]: from='mgr.14156 192.168.123.100:0/929894269' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-06T13:41:22.751 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:41:22 vm01 ceph-mon[56331]: from='mgr.14156 192.168.123.100:0/929894269' entity='mgr.a' 2026-03-06T13:41:22.751 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:41:22 vm01 ceph-mon[56331]: from='mgr.14156 192.168.123.100:0/929894269' entity='mgr.a' cmd=[{"prefix": "auth get-or-create", "entity": "mgr.b", "caps": ["mon", "profile mgr", "osd", "allow *", "mds", "allow *"]}]: dispatch 2026-03-06T13:41:22.751 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:41:22 vm01 ceph-mon[56331]: from='mgr.14156 192.168.123.100:0/929894269' entity='mgr.a' cmd='[{"prefix": "auth get-or-create", "entity": "mgr.b", "caps": ["mon", "profile mgr", "osd", "allow *", "mds", "allow *"]}]': finished 2026-03-06T13:41:22.751 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:41:22 vm01 ceph-mon[56331]: from='mgr.14156 192.168.123.100:0/929894269' entity='mgr.a' cmd=[{"prefix": "mgr services"}]: dispatch 2026-03-06T13:41:22.751 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:41:22 vm01 ceph-mon[56331]: from='mgr.14156 192.168.123.100:0/929894269' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-06T13:41:22.751 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:41:22 vm01 ceph-mon[56331]: Deploying daemon mgr.b on vm01 2026-03-06T13:41:22.751 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:41:22 vm01 ceph-mon[56331]: pgmap v11: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-06T13:41:22.751 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:41:22 vm01 ceph-mon[56331]: from='mgr.14156 192.168.123.100:0/929894269' entity='mgr.a' 2026-03-06T13:41:22.751 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:41:22 vm01 ceph-mon[56331]: from='mgr.14156 192.168.123.100:0/929894269' entity='mgr.a' 2026-03-06T13:41:22.751 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:41:22 vm01 ceph-mon[56331]: from='mgr.14156 192.168.123.100:0/929894269' entity='mgr.a' 2026-03-06T13:41:22.751 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:41:22 vm01 ceph-mon[56331]: from='mgr.14156 192.168.123.100:0/929894269' entity='mgr.a' 2026-03-06T13:41:22.752 INFO:journalctl@ceph.mgr.b.vm01.stdout:Mar 06 13:41:22 vm01 podman[57252]: 2026-03-06 13:41:22.436865493 +0100 CET m=+0.023073453 container create 152e255340174314a69a72952c422a109a16d57b802a9e79b2f5a060e4b501ba (image=harbor.clyso.com/custom-ceph/ceph/ceph@sha256:26363c7a4eea9ef5a0148afc7b2a22b6f486596d87a30c2a9fdcda5db3eca62b, name=ceph-82f504c8-1959-11f1-9a16-a78d135e200b-mgr-b, CEPH_GIT_REPO=https://github.com/irq0/ceph.git, CEPH_REF=19.2.3-47-gc24117fd552, CEPH_SHA1=c24117fd5525679b799527bc1bd1f1dd0a2db5e2, FROM_IMAGE=rockylinux:9, GANESHA_REPO_BASEURL=https://buildlogs.centos.org/centos/$releasever-stream/storage/$basearch/nfsganesha-5/, OSD_FLAVOR=default, ceph=True, io.buildah.version=1.41.8) 2026-03-06T13:41:22.752 INFO:journalctl@ceph.mgr.b.vm01.stdout:Mar 06 13:41:22 vm01 podman[57252]: 2026-03-06 13:41:22.470902426 +0100 CET m=+0.057110386 container init 152e255340174314a69a72952c422a109a16d57b802a9e79b2f5a060e4b501ba (image=harbor.clyso.com/custom-ceph/ceph/ceph@sha256:26363c7a4eea9ef5a0148afc7b2a22b6f486596d87a30c2a9fdcda5db3eca62b, name=ceph-82f504c8-1959-11f1-9a16-a78d135e200b-mgr-b, ceph=True, io.buildah.version=1.41.8, CEPH_GIT_REPO=https://github.com/irq0/ceph.git, CEPH_REF=19.2.3-47-gc24117fd552, CEPH_SHA1=c24117fd5525679b799527bc1bd1f1dd0a2db5e2, FROM_IMAGE=rockylinux:9, GANESHA_REPO_BASEURL=https://buildlogs.centos.org/centos/$releasever-stream/storage/$basearch/nfsganesha-5/, OSD_FLAVOR=default) 2026-03-06T13:41:22.753 INFO:journalctl@ceph.mgr.b.vm01.stdout:Mar 06 13:41:22 vm01 podman[57252]: 2026-03-06 13:41:22.473502518 +0100 CET m=+0.059710478 container start 152e255340174314a69a72952c422a109a16d57b802a9e79b2f5a060e4b501ba (image=harbor.clyso.com/custom-ceph/ceph/ceph@sha256:26363c7a4eea9ef5a0148afc7b2a22b6f486596d87a30c2a9fdcda5db3eca62b, name=ceph-82f504c8-1959-11f1-9a16-a78d135e200b-mgr-b, CEPH_GIT_REPO=https://github.com/irq0/ceph.git, CEPH_REF=19.2.3-47-gc24117fd552, CEPH_SHA1=c24117fd5525679b799527bc1bd1f1dd0a2db5e2, FROM_IMAGE=rockylinux:9, GANESHA_REPO_BASEURL=https://buildlogs.centos.org/centos/$releasever-stream/storage/$basearch/nfsganesha-5/, OSD_FLAVOR=default, ceph=True, io.buildah.version=1.41.8) 2026-03-06T13:41:22.753 INFO:journalctl@ceph.mgr.b.vm01.stdout:Mar 06 13:41:22 vm01 bash[57252]: 152e255340174314a69a72952c422a109a16d57b802a9e79b2f5a060e4b501ba 2026-03-06T13:41:22.753 INFO:journalctl@ceph.mgr.b.vm01.stdout:Mar 06 13:41:22 vm01 podman[57252]: 2026-03-06 13:41:22.428693932 +0100 CET m=+0.014901902 image pull 306e97de47e91c2b4b24d3dc09be3b3a12039b078f343d91220102acc6628a68 harbor.clyso.com/custom-ceph/ceph/ceph@sha256:26363c7a4eea9ef5a0148afc7b2a22b6f486596d87a30c2a9fdcda5db3eca62b 2026-03-06T13:41:22.753 INFO:journalctl@ceph.mgr.b.vm01.stdout:Mar 06 13:41:22 vm01 systemd[1]: Started Ceph mgr.b for 82f504c8-1959-11f1-9a16-a78d135e200b. 2026-03-06T13:41:22.753 INFO:journalctl@ceph.mgr.b.vm01.stdout:Mar 06 13:41:22 vm01 ceph-82f504c8-1959-11f1-9a16-a78d135e200b-mgr-b[57308]: 2026-03-06T12:41:22.744+0000 7f4a81416100 -1 mgr[py] Module alerts has missing NOTIFY_TYPES member 2026-03-06T13:41:22.790 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 06 13:41:22 vm08 ceph-mon[53561]: from='client.14214 -' entity='client.admin' cmd=[{"prefix": "orch apply", "service_type": "mgr", "placement": "2;vm00=a;vm01=b", "target": ["mon-mgr", ""]}]: dispatch 2026-03-06T13:41:22.790 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 06 13:41:22 vm08 ceph-mon[53561]: Saving service mgr spec with placement vm00=a;vm01=b;count:2 2026-03-06T13:41:22.790 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 06 13:41:22 vm08 ceph-mon[53561]: from='mgr.14156 192.168.123.100:0/929894269' entity='mgr.a' 2026-03-06T13:41:22.790 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 06 13:41:22 vm08 ceph-mon[53561]: from='mgr.14156 192.168.123.100:0/929894269' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-06T13:41:22.790 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 06 13:41:22 vm08 ceph-mon[53561]: from='mgr.14156 192.168.123.100:0/929894269' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-06T13:41:22.790 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 06 13:41:22 vm08 ceph-mon[53561]: from='mgr.14156 192.168.123.100:0/929894269' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-06T13:41:22.790 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 06 13:41:22 vm08 ceph-mon[53561]: from='mgr.14156 192.168.123.100:0/929894269' entity='mgr.a' 2026-03-06T13:41:22.790 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 06 13:41:22 vm08 ceph-mon[53561]: from='mgr.14156 192.168.123.100:0/929894269' entity='mgr.a' cmd=[{"prefix": "auth get-or-create", "entity": "mgr.b", "caps": ["mon", "profile mgr", "osd", "allow *", "mds", "allow *"]}]: dispatch 2026-03-06T13:41:22.790 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 06 13:41:22 vm08 ceph-mon[53561]: from='mgr.14156 192.168.123.100:0/929894269' entity='mgr.a' cmd='[{"prefix": "auth get-or-create", "entity": "mgr.b", "caps": ["mon", "profile mgr", "osd", "allow *", "mds", "allow *"]}]': finished 2026-03-06T13:41:22.790 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 06 13:41:22 vm08 ceph-mon[53561]: from='mgr.14156 192.168.123.100:0/929894269' entity='mgr.a' cmd=[{"prefix": "mgr services"}]: dispatch 2026-03-06T13:41:22.790 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 06 13:41:22 vm08 ceph-mon[53561]: from='mgr.14156 192.168.123.100:0/929894269' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-06T13:41:22.790 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 06 13:41:22 vm08 ceph-mon[53561]: Deploying daemon mgr.b on vm01 2026-03-06T13:41:22.790 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 06 13:41:22 vm08 ceph-mon[53561]: pgmap v11: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-06T13:41:22.790 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 06 13:41:22 vm08 ceph-mon[53561]: from='mgr.14156 192.168.123.100:0/929894269' entity='mgr.a' 2026-03-06T13:41:22.790 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 06 13:41:22 vm08 ceph-mon[53561]: from='mgr.14156 192.168.123.100:0/929894269' entity='mgr.a' 2026-03-06T13:41:22.790 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 06 13:41:22 vm08 ceph-mon[53561]: from='mgr.14156 192.168.123.100:0/929894269' entity='mgr.a' 2026-03-06T13:41:22.791 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 06 13:41:22 vm08 ceph-mon[53561]: from='mgr.14156 192.168.123.100:0/929894269' entity='mgr.a' 2026-03-06T13:41:22.831 INFO:teuthology.orchestra.run.vm01.stdout: File: /dev/vdc 2026-03-06T13:41:22.831 INFO:teuthology.orchestra.run.vm01.stdout: Size: 0 Blocks: 0 IO Block: 512 block special file 2026-03-06T13:41:22.831 INFO:teuthology.orchestra.run.vm01.stdout:Device: 6h/6d Inode: 255 Links: 1 Device type: fc,20 2026-03-06T13:41:22.831 INFO:teuthology.orchestra.run.vm01.stdout:Access: (0660/brw-rw----) Uid: ( 0/ root) Gid: ( 6/ disk) 2026-03-06T13:41:22.831 INFO:teuthology.orchestra.run.vm01.stdout:Context: system_u:object_r:fixed_disk_device_t:s0 2026-03-06T13:41:22.831 INFO:teuthology.orchestra.run.vm01.stdout:Access: 2026-03-06 13:41:00.171359163 +0100 2026-03-06T13:41:22.831 INFO:teuthology.orchestra.run.vm01.stdout:Modify: 2026-03-06 13:38:28.136112424 +0100 2026-03-06T13:41:22.831 INFO:teuthology.orchestra.run.vm01.stdout:Change: 2026-03-06 13:38:28.136112424 +0100 2026-03-06T13:41:22.831 INFO:teuthology.orchestra.run.vm01.stdout: Birth: 2026-03-06 13:35:31.278000000 +0100 2026-03-06T13:41:22.831 DEBUG:teuthology.orchestra.run.vm01:> sudo dd if=/dev/vdc of=/dev/null count=1 2026-03-06T13:41:23.002 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:41:22 vm00 ceph-mon[51914]: from='client.14214 -' entity='client.admin' cmd=[{"prefix": "orch apply", "service_type": "mgr", "placement": "2;vm00=a;vm01=b", "target": ["mon-mgr", ""]}]: dispatch 2026-03-06T13:41:23.002 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:41:22 vm00 ceph-mon[51914]: Saving service mgr spec with placement vm00=a;vm01=b;count:2 2026-03-06T13:41:23.002 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:41:22 vm00 ceph-mon[51914]: from='mgr.14156 192.168.123.100:0/929894269' entity='mgr.a' 2026-03-06T13:41:23.002 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:41:22 vm00 ceph-mon[51914]: from='mgr.14156 192.168.123.100:0/929894269' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-06T13:41:23.002 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:41:22 vm00 ceph-mon[51914]: from='mgr.14156 192.168.123.100:0/929894269' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-06T13:41:23.002 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:41:22 vm00 ceph-mon[51914]: from='mgr.14156 192.168.123.100:0/929894269' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-06T13:41:23.002 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:41:22 vm00 ceph-mon[51914]: from='mgr.14156 192.168.123.100:0/929894269' entity='mgr.a' 2026-03-06T13:41:23.002 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:41:22 vm00 ceph-mon[51914]: from='mgr.14156 192.168.123.100:0/929894269' entity='mgr.a' cmd=[{"prefix": "auth get-or-create", "entity": "mgr.b", "caps": ["mon", "profile mgr", "osd", "allow *", "mds", "allow *"]}]: dispatch 2026-03-06T13:41:23.002 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:41:22 vm00 ceph-mon[51914]: from='mgr.14156 192.168.123.100:0/929894269' entity='mgr.a' cmd='[{"prefix": "auth get-or-create", "entity": "mgr.b", "caps": ["mon", "profile mgr", "osd", "allow *", "mds", "allow *"]}]': finished 2026-03-06T13:41:23.002 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:41:22 vm00 ceph-mon[51914]: from='mgr.14156 192.168.123.100:0/929894269' entity='mgr.a' cmd=[{"prefix": "mgr services"}]: dispatch 2026-03-06T13:41:23.002 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:41:22 vm00 ceph-mon[51914]: from='mgr.14156 192.168.123.100:0/929894269' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-06T13:41:23.002 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:41:22 vm00 ceph-mon[51914]: Deploying daemon mgr.b on vm01 2026-03-06T13:41:23.002 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:41:22 vm00 ceph-mon[51914]: pgmap v11: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-06T13:41:23.002 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:41:22 vm00 ceph-mon[51914]: from='mgr.14156 192.168.123.100:0/929894269' entity='mgr.a' 2026-03-06T13:41:23.002 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:41:22 vm00 ceph-mon[51914]: from='mgr.14156 192.168.123.100:0/929894269' entity='mgr.a' 2026-03-06T13:41:23.002 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:41:22 vm00 ceph-mon[51914]: from='mgr.14156 192.168.123.100:0/929894269' entity='mgr.a' 2026-03-06T13:41:23.002 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:41:22 vm00 ceph-mon[51914]: from='mgr.14156 192.168.123.100:0/929894269' entity='mgr.a' 2026-03-06T13:41:23.003 INFO:journalctl@ceph.mgr.b.vm01.stdout:Mar 06 13:41:22 vm01 ceph-82f504c8-1959-11f1-9a16-a78d135e200b-mgr-b[57308]: 2026-03-06T12:41:22.904+0000 7f4a81416100 -1 mgr[py] Module balancer has missing NOTIFY_TYPES member 2026-03-06T13:41:23.014 INFO:teuthology.orchestra.run.vm01.stderr:1+0 records in 2026-03-06T13:41:23.014 INFO:teuthology.orchestra.run.vm01.stderr:1+0 records out 2026-03-06T13:41:23.014 INFO:teuthology.orchestra.run.vm01.stderr:512 bytes copied, 0.00148754 s, 344 kB/s 2026-03-06T13:41:23.015 DEBUG:teuthology.orchestra.run.vm01:> ! mount | grep -v devtmpfs | grep -q /dev/vdc 2026-03-06T13:41:23.109 DEBUG:teuthology.orchestra.run.vm01:> stat /dev/vdd 2026-03-06T13:41:23.200 INFO:teuthology.orchestra.run.vm01.stdout: File: /dev/vdd 2026-03-06T13:41:23.200 INFO:teuthology.orchestra.run.vm01.stdout: Size: 0 Blocks: 0 IO Block: 512 block special file 2026-03-06T13:41:23.200 INFO:teuthology.orchestra.run.vm01.stdout:Device: 6h/6d Inode: 256 Links: 1 Device type: fc,30 2026-03-06T13:41:23.200 INFO:teuthology.orchestra.run.vm01.stdout:Access: (0660/brw-rw----) Uid: ( 0/ root) Gid: ( 6/ disk) 2026-03-06T13:41:23.200 INFO:teuthology.orchestra.run.vm01.stdout:Context: system_u:object_r:fixed_disk_device_t:s0 2026-03-06T13:41:23.200 INFO:teuthology.orchestra.run.vm01.stdout:Access: 2026-03-06 13:41:00.198359224 +0100 2026-03-06T13:41:23.200 INFO:teuthology.orchestra.run.vm01.stdout:Modify: 2026-03-06 13:38:28.114112418 +0100 2026-03-06T13:41:23.200 INFO:teuthology.orchestra.run.vm01.stdout:Change: 2026-03-06 13:38:28.114112418 +0100 2026-03-06T13:41:23.200 INFO:teuthology.orchestra.run.vm01.stdout: Birth: 2026-03-06 13:35:31.282000000 +0100 2026-03-06T13:41:23.201 DEBUG:teuthology.orchestra.run.vm01:> sudo dd if=/dev/vdd of=/dev/null count=1 2026-03-06T13:41:23.285 INFO:teuthology.orchestra.run.vm01.stderr:1+0 records in 2026-03-06T13:41:23.287 INFO:teuthology.orchestra.run.vm01.stderr:1+0 records out 2026-03-06T13:41:23.287 INFO:teuthology.orchestra.run.vm01.stderr:512 bytes copied, 0.000360585 s, 1.4 MB/s 2026-03-06T13:41:23.287 DEBUG:teuthology.orchestra.run.vm01:> ! mount | grep -v devtmpfs | grep -q /dev/vdd 2026-03-06T13:41:23.309 DEBUG:teuthology.orchestra.run.vm01:> stat /dev/vde 2026-03-06T13:41:23.437 INFO:teuthology.orchestra.run.vm01.stdout: File: /dev/vde 2026-03-06T13:41:23.437 INFO:teuthology.orchestra.run.vm01.stdout: Size: 0 Blocks: 0 IO Block: 512 block special file 2026-03-06T13:41:23.437 INFO:teuthology.orchestra.run.vm01.stdout:Device: 6h/6d Inode: 257 Links: 1 Device type: fc,40 2026-03-06T13:41:23.437 INFO:teuthology.orchestra.run.vm01.stdout:Access: (0660/brw-rw----) Uid: ( 0/ root) Gid: ( 6/ disk) 2026-03-06T13:41:23.437 INFO:teuthology.orchestra.run.vm01.stdout:Context: system_u:object_r:fixed_disk_device_t:s0 2026-03-06T13:41:23.437 INFO:teuthology.orchestra.run.vm01.stdout:Access: 2026-03-06 13:41:00.227359289 +0100 2026-03-06T13:41:23.437 INFO:teuthology.orchestra.run.vm01.stdout:Modify: 2026-03-06 13:38:28.121112420 +0100 2026-03-06T13:41:23.437 INFO:teuthology.orchestra.run.vm01.stdout:Change: 2026-03-06 13:38:28.121112420 +0100 2026-03-06T13:41:23.437 INFO:teuthology.orchestra.run.vm01.stdout: Birth: 2026-03-06 13:35:31.286000000 +0100 2026-03-06T13:41:23.437 DEBUG:teuthology.orchestra.run.vm01:> sudo dd if=/dev/vde of=/dev/null count=1 2026-03-06T13:41:23.478 INFO:teuthology.orchestra.run.vm01.stderr:1+0 records in 2026-03-06T13:41:23.478 INFO:teuthology.orchestra.run.vm01.stderr:1+0 records out 2026-03-06T13:41:23.478 INFO:teuthology.orchestra.run.vm01.stderr:512 bytes copied, 0.000204273 s, 2.5 MB/s 2026-03-06T13:41:23.479 DEBUG:teuthology.orchestra.run.vm01:> ! mount | grep -v devtmpfs | grep -q /dev/vde 2026-03-06T13:41:23.534 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:41:23 vm01 ceph-mon[56331]: from='mgr.14156 192.168.123.100:0/929894269' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-06T13:41:23.534 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:41:23 vm01 ceph-mon[56331]: from='mgr.14156 192.168.123.100:0/929894269' entity='mgr.a' 2026-03-06T13:41:23.534 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:41:23 vm01 ceph-mon[56331]: from='mgr.14156 192.168.123.100:0/929894269' entity='mgr.a' 2026-03-06T13:41:23.534 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:41:23 vm01 ceph-mon[56331]: from='mgr.14156 192.168.123.100:0/929894269' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-06T13:41:23.534 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:41:23 vm01 ceph-mon[56331]: from='mgr.14156 192.168.123.100:0/929894269' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-06T13:41:23.534 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:41:23 vm01 ceph-mon[56331]: from='mgr.14156 192.168.123.100:0/929894269' entity='mgr.a' 2026-03-06T13:41:23.534 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:41:23 vm01 ceph-mon[56331]: from='mgr.14156 192.168.123.100:0/929894269' entity='mgr.a' cmd=[{"prefix": "auth get-or-create", "entity": "mgr.a", "caps": ["mon", "profile mgr", "osd", "allow *", "mds", "allow *"]}]: dispatch 2026-03-06T13:41:23.534 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:41:23 vm01 ceph-mon[56331]: from='mgr.14156 192.168.123.100:0/929894269' entity='mgr.a' cmd=[{"prefix": "mgr services"}]: dispatch 2026-03-06T13:41:23.534 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:41:23 vm01 ceph-mon[56331]: from='mgr.14156 192.168.123.100:0/929894269' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-06T13:41:23.552 DEBUG:teuthology.orchestra.run.vm08:> set -ex 2026-03-06T13:41:23.552 DEBUG:teuthology.orchestra.run.vm08:> dd if=/scratch_devs of=/dev/stdout 2026-03-06T13:41:23.568 DEBUG:teuthology.orchestra.run:got remote process result: 1 2026-03-06T13:41:23.568 DEBUG:teuthology.orchestra.run.vm08:> ls /dev/[sv]d? 2026-03-06T13:41:23.631 INFO:teuthology.orchestra.run.vm08.stdout:/dev/vda 2026-03-06T13:41:23.631 INFO:teuthology.orchestra.run.vm08.stdout:/dev/vdb 2026-03-06T13:41:23.631 INFO:teuthology.orchestra.run.vm08.stdout:/dev/vdc 2026-03-06T13:41:23.631 INFO:teuthology.orchestra.run.vm08.stdout:/dev/vdd 2026-03-06T13:41:23.631 INFO:teuthology.orchestra.run.vm08.stdout:/dev/vde 2026-03-06T13:41:23.632 WARNING:teuthology.misc:Removing root device: /dev/vda from device list 2026-03-06T13:41:23.632 DEBUG:teuthology.misc:devs=['/dev/vdb', '/dev/vdc', '/dev/vdd', '/dev/vde'] 2026-03-06T13:41:23.632 DEBUG:teuthology.orchestra.run.vm08:> stat /dev/vdb 2026-03-06T13:41:23.690 INFO:teuthology.orchestra.run.vm08.stdout: File: /dev/vdb 2026-03-06T13:41:23.690 INFO:teuthology.orchestra.run.vm08.stdout: Size: 0 Blocks: 0 IO Block: 512 block special file 2026-03-06T13:41:23.690 INFO:teuthology.orchestra.run.vm08.stdout:Device: 6h/6d Inode: 254 Links: 1 Device type: fc,10 2026-03-06T13:41:23.690 INFO:teuthology.orchestra.run.vm08.stdout:Access: (0660/brw-rw----) Uid: ( 0/ root) Gid: ( 6/ disk) 2026-03-06T13:41:23.690 INFO:teuthology.orchestra.run.vm08.stdout:Context: system_u:object_r:fixed_disk_device_t:s0 2026-03-06T13:41:23.690 INFO:teuthology.orchestra.run.vm08.stdout:Access: 2026-03-06 13:41:04.169630587 +0100 2026-03-06T13:41:23.690 INFO:teuthology.orchestra.run.vm08.stdout:Modify: 2026-03-06 13:38:28.103777470 +0100 2026-03-06T13:41:23.690 INFO:teuthology.orchestra.run.vm08.stdout:Change: 2026-03-06 13:38:28.103777470 +0100 2026-03-06T13:41:23.690 INFO:teuthology.orchestra.run.vm08.stdout: Birth: 2026-03-06 13:34:24.277000000 +0100 2026-03-06T13:41:23.690 DEBUG:teuthology.orchestra.run.vm08:> sudo dd if=/dev/vdb of=/dev/null count=1 2026-03-06T13:41:23.759 INFO:teuthology.orchestra.run.vm08.stderr:1+0 records in 2026-03-06T13:41:23.759 INFO:teuthology.orchestra.run.vm08.stderr:1+0 records out 2026-03-06T13:41:23.759 INFO:teuthology.orchestra.run.vm08.stderr:512 bytes copied, 0.00017037 s, 3.0 MB/s 2026-03-06T13:41:23.760 DEBUG:teuthology.orchestra.run.vm08:> ! mount | grep -v devtmpfs | grep -q /dev/vdb 2026-03-06T13:41:23.790 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 06 13:41:23 vm08 ceph-mon[53561]: from='mgr.14156 192.168.123.100:0/929894269' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-06T13:41:23.790 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 06 13:41:23 vm08 ceph-mon[53561]: from='mgr.14156 192.168.123.100:0/929894269' entity='mgr.a' 2026-03-06T13:41:23.790 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 06 13:41:23 vm08 ceph-mon[53561]: from='mgr.14156 192.168.123.100:0/929894269' entity='mgr.a' 2026-03-06T13:41:23.790 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 06 13:41:23 vm08 ceph-mon[53561]: from='mgr.14156 192.168.123.100:0/929894269' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-06T13:41:23.790 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 06 13:41:23 vm08 ceph-mon[53561]: from='mgr.14156 192.168.123.100:0/929894269' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-06T13:41:23.790 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 06 13:41:23 vm08 ceph-mon[53561]: from='mgr.14156 192.168.123.100:0/929894269' entity='mgr.a' 2026-03-06T13:41:23.790 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 06 13:41:23 vm08 ceph-mon[53561]: from='mgr.14156 192.168.123.100:0/929894269' entity='mgr.a' cmd=[{"prefix": "auth get-or-create", "entity": "mgr.a", "caps": ["mon", "profile mgr", "osd", "allow *", "mds", "allow *"]}]: dispatch 2026-03-06T13:41:23.790 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 06 13:41:23 vm08 ceph-mon[53561]: from='mgr.14156 192.168.123.100:0/929894269' entity='mgr.a' cmd=[{"prefix": "mgr services"}]: dispatch 2026-03-06T13:41:23.790 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 06 13:41:23 vm08 ceph-mon[53561]: from='mgr.14156 192.168.123.100:0/929894269' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-06T13:41:23.807 DEBUG:teuthology.orchestra.run.vm08:> stat /dev/vdc 2026-03-06T13:41:23.863 INFO:teuthology.orchestra.run.vm08.stdout: File: /dev/vdc 2026-03-06T13:41:23.863 INFO:teuthology.orchestra.run.vm08.stdout: Size: 0 Blocks: 0 IO Block: 512 block special file 2026-03-06T13:41:23.863 INFO:teuthology.orchestra.run.vm08.stdout:Device: 6h/6d Inode: 255 Links: 1 Device type: fc,20 2026-03-06T13:41:23.863 INFO:teuthology.orchestra.run.vm08.stdout:Access: (0660/brw-rw----) Uid: ( 0/ root) Gid: ( 6/ disk) 2026-03-06T13:41:23.863 INFO:teuthology.orchestra.run.vm08.stdout:Context: system_u:object_r:fixed_disk_device_t:s0 2026-03-06T13:41:23.863 INFO:teuthology.orchestra.run.vm08.stdout:Access: 2026-03-06 13:41:04.203630652 +0100 2026-03-06T13:41:23.863 INFO:teuthology.orchestra.run.vm08.stdout:Modify: 2026-03-06 13:38:28.097777468 +0100 2026-03-06T13:41:23.863 INFO:teuthology.orchestra.run.vm08.stdout:Change: 2026-03-06 13:38:28.097777468 +0100 2026-03-06T13:41:23.863 INFO:teuthology.orchestra.run.vm08.stdout: Birth: 2026-03-06 13:34:24.283000000 +0100 2026-03-06T13:41:23.863 DEBUG:teuthology.orchestra.run.vm08:> sudo dd if=/dev/vdc of=/dev/null count=1 2026-03-06T13:41:23.886 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:41:23 vm00 ceph-mon[51914]: from='mgr.14156 192.168.123.100:0/929894269' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-06T13:41:23.886 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:41:23 vm00 ceph-mon[51914]: from='mgr.14156 192.168.123.100:0/929894269' entity='mgr.a' 2026-03-06T13:41:23.886 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:41:23 vm00 ceph-mon[51914]: from='mgr.14156 192.168.123.100:0/929894269' entity='mgr.a' 2026-03-06T13:41:23.886 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:41:23 vm00 ceph-mon[51914]: from='mgr.14156 192.168.123.100:0/929894269' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-06T13:41:23.886 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:41:23 vm00 ceph-mon[51914]: from='mgr.14156 192.168.123.100:0/929894269' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-06T13:41:23.886 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:41:23 vm00 ceph-mon[51914]: from='mgr.14156 192.168.123.100:0/929894269' entity='mgr.a' 2026-03-06T13:41:23.886 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:41:23 vm00 ceph-mon[51914]: from='mgr.14156 192.168.123.100:0/929894269' entity='mgr.a' cmd=[{"prefix": "auth get-or-create", "entity": "mgr.a", "caps": ["mon", "profile mgr", "osd", "allow *", "mds", "allow *"]}]: dispatch 2026-03-06T13:41:23.886 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:41:23 vm00 ceph-mon[51914]: from='mgr.14156 192.168.123.100:0/929894269' entity='mgr.a' cmd=[{"prefix": "mgr services"}]: dispatch 2026-03-06T13:41:23.886 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:41:23 vm00 ceph-mon[51914]: from='mgr.14156 192.168.123.100:0/929894269' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-06T13:41:23.928 INFO:teuthology.orchestra.run.vm08.stderr:1+0 records in 2026-03-06T13:41:23.928 INFO:teuthology.orchestra.run.vm08.stderr:1+0 records out 2026-03-06T13:41:23.928 INFO:teuthology.orchestra.run.vm08.stderr:512 bytes copied, 0.000202069 s, 2.5 MB/s 2026-03-06T13:41:23.929 DEBUG:teuthology.orchestra.run.vm08:> ! mount | grep -v devtmpfs | grep -q /dev/vdc 2026-03-06T13:41:23.990 DEBUG:teuthology.orchestra.run.vm08:> stat /dev/vdd 2026-03-06T13:41:24.050 INFO:teuthology.orchestra.run.vm08.stdout: File: /dev/vdd 2026-03-06T13:41:24.050 INFO:teuthology.orchestra.run.vm08.stdout: Size: 0 Blocks: 0 IO Block: 512 block special file 2026-03-06T13:41:24.050 INFO:teuthology.orchestra.run.vm08.stdout:Device: 6h/6d Inode: 256 Links: 1 Device type: fc,30 2026-03-06T13:41:24.050 INFO:teuthology.orchestra.run.vm08.stdout:Access: (0660/brw-rw----) Uid: ( 0/ root) Gid: ( 6/ disk) 2026-03-06T13:41:24.050 INFO:teuthology.orchestra.run.vm08.stdout:Context: system_u:object_r:fixed_disk_device_t:s0 2026-03-06T13:41:24.050 INFO:teuthology.orchestra.run.vm08.stdout:Access: 2026-03-06 13:41:04.236630715 +0100 2026-03-06T13:41:24.050 INFO:teuthology.orchestra.run.vm08.stdout:Modify: 2026-03-06 13:38:28.092777465 +0100 2026-03-06T13:41:24.050 INFO:teuthology.orchestra.run.vm08.stdout:Change: 2026-03-06 13:38:28.092777465 +0100 2026-03-06T13:41:24.050 INFO:teuthology.orchestra.run.vm08.stdout: Birth: 2026-03-06 13:34:24.295000000 +0100 2026-03-06T13:41:24.051 DEBUG:teuthology.orchestra.run.vm08:> sudo dd if=/dev/vdd of=/dev/null count=1 2026-03-06T13:41:24.116 INFO:teuthology.orchestra.run.vm08.stderr:1+0 records in 2026-03-06T13:41:24.116 INFO:teuthology.orchestra.run.vm08.stderr:1+0 records out 2026-03-06T13:41:24.116 INFO:teuthology.orchestra.run.vm08.stderr:512 bytes copied, 0.000155241 s, 3.3 MB/s 2026-03-06T13:41:24.117 DEBUG:teuthology.orchestra.run.vm08:> ! mount | grep -v devtmpfs | grep -q /dev/vdd 2026-03-06T13:41:24.176 DEBUG:teuthology.orchestra.run.vm08:> stat /dev/vde 2026-03-06T13:41:24.236 INFO:teuthology.orchestra.run.vm08.stdout: File: /dev/vde 2026-03-06T13:41:24.236 INFO:teuthology.orchestra.run.vm08.stdout: Size: 0 Blocks: 0 IO Block: 512 block special file 2026-03-06T13:41:24.236 INFO:teuthology.orchestra.run.vm08.stdout:Device: 6h/6d Inode: 257 Links: 1 Device type: fc,40 2026-03-06T13:41:24.236 INFO:teuthology.orchestra.run.vm08.stdout:Access: (0660/brw-rw----) Uid: ( 0/ root) Gid: ( 6/ disk) 2026-03-06T13:41:24.236 INFO:teuthology.orchestra.run.vm08.stdout:Context: system_u:object_r:fixed_disk_device_t:s0 2026-03-06T13:41:24.236 INFO:teuthology.orchestra.run.vm08.stdout:Access: 2026-03-06 13:41:04.284630806 +0100 2026-03-06T13:41:24.236 INFO:teuthology.orchestra.run.vm08.stdout:Modify: 2026-03-06 13:38:28.077777458 +0100 2026-03-06T13:41:24.236 INFO:teuthology.orchestra.run.vm08.stdout:Change: 2026-03-06 13:38:28.077777458 +0100 2026-03-06T13:41:24.236 INFO:teuthology.orchestra.run.vm08.stdout: Birth: 2026-03-06 13:34:24.374000000 +0100 2026-03-06T13:41:24.236 DEBUG:teuthology.orchestra.run.vm08:> sudo dd if=/dev/vde of=/dev/null count=1 2026-03-06T13:41:24.309 INFO:teuthology.orchestra.run.vm08.stderr:1+0 records in 2026-03-06T13:41:24.309 INFO:teuthology.orchestra.run.vm08.stderr:1+0 records out 2026-03-06T13:41:24.309 INFO:teuthology.orchestra.run.vm08.stderr:512 bytes copied, 0.000164799 s, 3.1 MB/s 2026-03-06T13:41:24.310 DEBUG:teuthology.orchestra.run.vm08:> ! mount | grep -v devtmpfs | grep -q /dev/vde 2026-03-06T13:41:24.373 INFO:tasks.cephadm:Deploying osd.0 on vm00 with /dev/vde... 2026-03-06T13:41:24.373 DEBUG:teuthology.orchestra.run.vm00:> sudo /home/ubuntu/cephtest/cephadm --image harbor.clyso.com/custom-ceph/ceph/ceph:cobaltcore-storage-v19.2.3-fasttrack-3 ceph-volume -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 82f504c8-1959-11f1-9a16-a78d135e200b -- lvm zap /dev/vde 2026-03-06T13:41:24.626 INFO:journalctl@ceph.mgr.b.vm01.stdout:Mar 06 13:41:24 vm01 ceph-82f504c8-1959-11f1-9a16-a78d135e200b-mgr-b[57308]: 2026-03-06T12:41:24.276+0000 7f4a81416100 -1 mgr[py] Module crash has missing NOTIFY_TYPES member 2026-03-06T13:41:24.700 INFO:teuthology.orchestra.run.vm00.stderr:Inferring config /var/lib/ceph/82f504c8-1959-11f1-9a16-a78d135e200b/mon.a/config 2026-03-06T13:41:24.721 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:41:24 vm00 ceph-mon[51914]: Reconfiguring mgr.a (unknown last config time)... 2026-03-06T13:41:24.722 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:41:24 vm00 ceph-mon[51914]: Reconfiguring daemon mgr.a on vm00 2026-03-06T13:41:24.722 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:41:24 vm00 ceph-mon[51914]: from='mgr.14156 192.168.123.100:0/929894269' entity='mgr.a' 2026-03-06T13:41:24.722 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:41:24 vm00 ceph-mon[51914]: from='mgr.14156 192.168.123.100:0/929894269' entity='mgr.a' 2026-03-06T13:41:24.722 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:41:24 vm00 ceph-mon[51914]: from='mgr.14156 192.168.123.100:0/929894269' entity='mgr.a' 2026-03-06T13:41:24.722 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:41:24 vm00 ceph-mon[51914]: from='mgr.14156 192.168.123.100:0/929894269' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-06T13:41:24.722 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:41:24 vm00 ceph-mon[51914]: from='mgr.14156 192.168.123.100:0/929894269' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-06T13:41:24.722 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:41:24 vm00 ceph-mon[51914]: from='mgr.14156 192.168.123.100:0/929894269' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-06T13:41:24.722 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:41:24 vm00 ceph-mon[51914]: pgmap v12: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-06T13:41:24.722 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:41:24 vm00 ceph-mon[51914]: from='mgr.14156 192.168.123.100:0/929894269' entity='mgr.a' 2026-03-06T13:41:24.905 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:41:24 vm01 ceph-mon[56331]: Reconfiguring mgr.a (unknown last config time)... 2026-03-06T13:41:24.905 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:41:24 vm01 ceph-mon[56331]: Reconfiguring daemon mgr.a on vm00 2026-03-06T13:41:24.905 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:41:24 vm01 ceph-mon[56331]: from='mgr.14156 192.168.123.100:0/929894269' entity='mgr.a' 2026-03-06T13:41:24.905 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:41:24 vm01 ceph-mon[56331]: from='mgr.14156 192.168.123.100:0/929894269' entity='mgr.a' 2026-03-06T13:41:24.905 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:41:24 vm01 ceph-mon[56331]: from='mgr.14156 192.168.123.100:0/929894269' entity='mgr.a' 2026-03-06T13:41:24.905 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:41:24 vm01 ceph-mon[56331]: from='mgr.14156 192.168.123.100:0/929894269' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-06T13:41:24.905 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:41:24 vm01 ceph-mon[56331]: from='mgr.14156 192.168.123.100:0/929894269' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-06T13:41:24.905 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:41:24 vm01 ceph-mon[56331]: from='mgr.14156 192.168.123.100:0/929894269' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-06T13:41:24.905 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:41:24 vm01 ceph-mon[56331]: pgmap v12: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-06T13:41:24.905 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:41:24 vm01 ceph-mon[56331]: from='mgr.14156 192.168.123.100:0/929894269' entity='mgr.a' 2026-03-06T13:41:25.040 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 06 13:41:24 vm08 ceph-mon[53561]: Reconfiguring mgr.a (unknown last config time)... 2026-03-06T13:41:25.040 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 06 13:41:24 vm08 ceph-mon[53561]: Reconfiguring daemon mgr.a on vm00 2026-03-06T13:41:25.040 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 06 13:41:24 vm08 ceph-mon[53561]: from='mgr.14156 192.168.123.100:0/929894269' entity='mgr.a' 2026-03-06T13:41:25.040 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 06 13:41:24 vm08 ceph-mon[53561]: from='mgr.14156 192.168.123.100:0/929894269' entity='mgr.a' 2026-03-06T13:41:25.040 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 06 13:41:24 vm08 ceph-mon[53561]: from='mgr.14156 192.168.123.100:0/929894269' entity='mgr.a' 2026-03-06T13:41:25.040 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 06 13:41:24 vm08 ceph-mon[53561]: from='mgr.14156 192.168.123.100:0/929894269' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-06T13:41:25.040 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 06 13:41:24 vm08 ceph-mon[53561]: from='mgr.14156 192.168.123.100:0/929894269' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-06T13:41:25.040 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 06 13:41:24 vm08 ceph-mon[53561]: from='mgr.14156 192.168.123.100:0/929894269' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-06T13:41:25.040 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 06 13:41:24 vm08 ceph-mon[53561]: pgmap v12: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-06T13:41:25.040 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 06 13:41:24 vm08 ceph-mon[53561]: from='mgr.14156 192.168.123.100:0/929894269' entity='mgr.a' 2026-03-06T13:41:25.650 INFO:journalctl@ceph.mgr.b.vm01.stdout:Mar 06 13:41:25 vm01 ceph-82f504c8-1959-11f1-9a16-a78d135e200b-mgr-b[57308]: 2026-03-06T12:41:25.251+0000 7f4a81416100 -1 mgr[py] Module devicehealth has missing NOTIFY_TYPES member 2026-03-06T13:41:25.650 INFO:journalctl@ceph.mgr.b.vm01.stdout:Mar 06 13:41:25 vm01 ceph-82f504c8-1959-11f1-9a16-a78d135e200b-mgr-b[57308]: 2026-03-06T12:41:25.392+0000 7f4a81416100 -1 mgr[py] Module influx has missing NOTIFY_TYPES member 2026-03-06T13:41:25.650 INFO:journalctl@ceph.mgr.b.vm01.stdout:Mar 06 13:41:25 vm01 ceph-82f504c8-1959-11f1-9a16-a78d135e200b-mgr-b[57308]: 2026-03-06T12:41:25.648+0000 7f4a81416100 -1 mgr[py] Module iostat has missing NOTIFY_TYPES member 2026-03-06T13:41:25.822 INFO:teuthology.orchestra.run.vm00.stdout: 2026-03-06T13:41:25.844 DEBUG:teuthology.orchestra.run.vm00:> sudo /home/ubuntu/cephtest/cephadm --image harbor.clyso.com/custom-ceph/ceph/ceph:cobaltcore-storage-v19.2.3-fasttrack-3 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 82f504c8-1959-11f1-9a16-a78d135e200b -- ceph orch daemon add osd vm00:/dev/vde 2026-03-06T13:41:26.209 INFO:teuthology.orchestra.run.vm00.stderr:Inferring config /var/lib/ceph/82f504c8-1959-11f1-9a16-a78d135e200b/mon.a/config 2026-03-06T13:41:27.173 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:41:27 vm00 ceph-mon[51914]: pgmap v13: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-06T13:41:27.173 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:41:27 vm00 ceph-mon[51914]: from='mgr.14156 192.168.123.100:0/929894269' entity='mgr.a' cmd=[{"prefix": "osd tree", "states": ["destroyed"], "format": "json"}]: dispatch 2026-03-06T13:41:27.173 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:41:27 vm00 ceph-mon[51914]: from='mgr.14156 192.168.123.100:0/929894269' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.bootstrap-osd"}]: dispatch 2026-03-06T13:41:27.173 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:41:27 vm00 ceph-mon[51914]: from='mgr.14156 192.168.123.100:0/929894269' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-06T13:41:27.290 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 06 13:41:27 vm08 ceph-mon[53561]: pgmap v13: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-06T13:41:27.290 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 06 13:41:27 vm08 ceph-mon[53561]: from='mgr.14156 192.168.123.100:0/929894269' entity='mgr.a' cmd=[{"prefix": "osd tree", "states": ["destroyed"], "format": "json"}]: dispatch 2026-03-06T13:41:27.290 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 06 13:41:27 vm08 ceph-mon[53561]: from='mgr.14156 192.168.123.100:0/929894269' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.bootstrap-osd"}]: dispatch 2026-03-06T13:41:27.290 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 06 13:41:27 vm08 ceph-mon[53561]: from='mgr.14156 192.168.123.100:0/929894269' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-06T13:41:27.405 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:41:27 vm01 ceph-mon[56331]: pgmap v13: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-06T13:41:27.405 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:41:27 vm01 ceph-mon[56331]: from='mgr.14156 192.168.123.100:0/929894269' entity='mgr.a' cmd=[{"prefix": "osd tree", "states": ["destroyed"], "format": "json"}]: dispatch 2026-03-06T13:41:27.405 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:41:27 vm01 ceph-mon[56331]: from='mgr.14156 192.168.123.100:0/929894269' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.bootstrap-osd"}]: dispatch 2026-03-06T13:41:27.405 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:41:27 vm01 ceph-mon[56331]: from='mgr.14156 192.168.123.100:0/929894269' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-06T13:41:27.905 INFO:journalctl@ceph.mgr.b.vm01.stdout:Mar 06 13:41:27 vm01 ceph-82f504c8-1959-11f1-9a16-a78d135e200b-mgr-b[57308]: 2026-03-06T12:41:27.632+0000 7f4a81416100 -1 mgr[py] Module nfs has missing NOTIFY_TYPES member 2026-03-06T13:41:28.252 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:41:28 vm00 ceph-mon[51914]: from='client.24103 -' entity='client.admin' cmd=[{"prefix": "orch daemon add osd", "svc_arg": "vm00:/dev/vde", "target": ["mon-mgr", ""]}]: dispatch 2026-03-06T13:41:28.252 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:41:28 vm00 ceph-mon[51914]: from='client.? 192.168.123.100:0/2670807793' entity='client.bootstrap-osd' cmd=[{"prefix": "osd new", "uuid": "1d65b890-7ef0-46fe-a6c2-e544a1692d70"}]: dispatch 2026-03-06T13:41:28.252 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:41:28 vm00 ceph-mon[51914]: from='client.? ' entity='client.bootstrap-osd' cmd=[{"prefix": "osd new", "uuid": "1d65b890-7ef0-46fe-a6c2-e544a1692d70"}]: dispatch 2026-03-06T13:41:28.252 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:41:28 vm00 ceph-mon[51914]: from='client.? ' entity='client.bootstrap-osd' cmd='[{"prefix": "osd new", "uuid": "1d65b890-7ef0-46fe-a6c2-e544a1692d70"}]': finished 2026-03-06T13:41:28.252 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:41:28 vm00 ceph-mon[51914]: osdmap e5: 1 total, 0 up, 1 in 2026-03-06T13:41:28.252 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:41:28 vm00 ceph-mon[51914]: from='mgr.14156 192.168.123.100:0/929894269' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 0}]: dispatch 2026-03-06T13:41:28.310 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:41:28 vm01 ceph-mon[56331]: from='client.24103 -' entity='client.admin' cmd=[{"prefix": "orch daemon add osd", "svc_arg": "vm00:/dev/vde", "target": ["mon-mgr", ""]}]: dispatch 2026-03-06T13:41:28.310 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:41:28 vm01 ceph-mon[56331]: from='client.? 192.168.123.100:0/2670807793' entity='client.bootstrap-osd' cmd=[{"prefix": "osd new", "uuid": "1d65b890-7ef0-46fe-a6c2-e544a1692d70"}]: dispatch 2026-03-06T13:41:28.310 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:41:28 vm01 ceph-mon[56331]: from='client.? ' entity='client.bootstrap-osd' cmd=[{"prefix": "osd new", "uuid": "1d65b890-7ef0-46fe-a6c2-e544a1692d70"}]: dispatch 2026-03-06T13:41:28.310 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:41:28 vm01 ceph-mon[56331]: from='client.? ' entity='client.bootstrap-osd' cmd='[{"prefix": "osd new", "uuid": "1d65b890-7ef0-46fe-a6c2-e544a1692d70"}]': finished 2026-03-06T13:41:28.310 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:41:28 vm01 ceph-mon[56331]: osdmap e5: 1 total, 0 up, 1 in 2026-03-06T13:41:28.310 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:41:28 vm01 ceph-mon[56331]: from='mgr.14156 192.168.123.100:0/929894269' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 0}]: dispatch 2026-03-06T13:41:28.310 INFO:journalctl@ceph.mgr.b.vm01.stdout:Mar 06 13:41:28 vm01 ceph-82f504c8-1959-11f1-9a16-a78d135e200b-mgr-b[57308]: 2026-03-06T12:41:28.007+0000 7f4a81416100 -1 mgr[py] Module orchestrator has missing NOTIFY_TYPES member 2026-03-06T13:41:28.310 INFO:journalctl@ceph.mgr.b.vm01.stdout:Mar 06 13:41:28 vm01 ceph-82f504c8-1959-11f1-9a16-a78d135e200b-mgr-b[57308]: 2026-03-06T12:41:28.175+0000 7f4a81416100 -1 mgr[py] Module osd_perf_query has missing NOTIFY_TYPES member 2026-03-06T13:41:28.540 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 06 13:41:28 vm08 ceph-mon[53561]: from='client.24103 -' entity='client.admin' cmd=[{"prefix": "orch daemon add osd", "svc_arg": "vm00:/dev/vde", "target": ["mon-mgr", ""]}]: dispatch 2026-03-06T13:41:28.540 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 06 13:41:28 vm08 ceph-mon[53561]: from='client.? 192.168.123.100:0/2670807793' entity='client.bootstrap-osd' cmd=[{"prefix": "osd new", "uuid": "1d65b890-7ef0-46fe-a6c2-e544a1692d70"}]: dispatch 2026-03-06T13:41:28.540 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 06 13:41:28 vm08 ceph-mon[53561]: from='client.? ' entity='client.bootstrap-osd' cmd=[{"prefix": "osd new", "uuid": "1d65b890-7ef0-46fe-a6c2-e544a1692d70"}]: dispatch 2026-03-06T13:41:28.540 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 06 13:41:28 vm08 ceph-mon[53561]: from='client.? ' entity='client.bootstrap-osd' cmd='[{"prefix": "osd new", "uuid": "1d65b890-7ef0-46fe-a6c2-e544a1692d70"}]': finished 2026-03-06T13:41:28.540 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 06 13:41:28 vm08 ceph-mon[53561]: osdmap e5: 1 total, 0 up, 1 in 2026-03-06T13:41:28.540 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 06 13:41:28 vm08 ceph-mon[53561]: from='mgr.14156 192.168.123.100:0/929894269' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 0}]: dispatch 2026-03-06T13:41:28.595 INFO:journalctl@ceph.mgr.b.vm01.stdout:Mar 06 13:41:28 vm01 ceph-82f504c8-1959-11f1-9a16-a78d135e200b-mgr-b[57308]: 2026-03-06T12:41:28.308+0000 7f4a81416100 -1 mgr[py] Module osd_support has missing NOTIFY_TYPES member 2026-03-06T13:41:28.595 INFO:journalctl@ceph.mgr.b.vm01.stdout:Mar 06 13:41:28 vm01 ceph-82f504c8-1959-11f1-9a16-a78d135e200b-mgr-b[57308]: 2026-03-06T12:41:28.460+0000 7f4a81416100 -1 mgr[py] Module pg_autoscaler has missing NOTIFY_TYPES member 2026-03-06T13:41:28.905 INFO:journalctl@ceph.mgr.b.vm01.stdout:Mar 06 13:41:28 vm01 ceph-82f504c8-1959-11f1-9a16-a78d135e200b-mgr-b[57308]: 2026-03-06T12:41:28.593+0000 7f4a81416100 -1 mgr[py] Module progress has missing NOTIFY_TYPES member 2026-03-06T13:41:29.350 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:41:29 vm01 ceph-mon[56331]: pgmap v15: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-06T13:41:29.350 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:41:29 vm01 ceph-mon[56331]: from='client.? 192.168.123.100:0/3350344731' entity='client.bootstrap-osd' cmd=[{"prefix": "mon getmap"}]: dispatch 2026-03-06T13:41:29.350 INFO:journalctl@ceph.mgr.b.vm01.stdout:Mar 06 13:41:29 vm01 ceph-82f504c8-1959-11f1-9a16-a78d135e200b-mgr-b[57308]: 2026-03-06T12:41:29.176+0000 7f4a81416100 -1 mgr[py] Module prometheus has missing NOTIFY_TYPES member 2026-03-06T13:41:29.502 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:41:29 vm00 ceph-mon[51914]: pgmap v15: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-06T13:41:29.502 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:41:29 vm00 ceph-mon[51914]: from='client.? 192.168.123.100:0/3350344731' entity='client.bootstrap-osd' cmd=[{"prefix": "mon getmap"}]: dispatch 2026-03-06T13:41:29.540 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 06 13:41:29 vm08 ceph-mon[53561]: pgmap v15: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-06T13:41:29.540 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 06 13:41:29 vm08 ceph-mon[53561]: from='client.? 192.168.123.100:0/3350344731' entity='client.bootstrap-osd' cmd=[{"prefix": "mon getmap"}]: dispatch 2026-03-06T13:41:29.655 INFO:journalctl@ceph.mgr.b.vm01.stdout:Mar 06 13:41:29 vm01 ceph-82f504c8-1959-11f1-9a16-a78d135e200b-mgr-b[57308]: 2026-03-06T12:41:29.348+0000 7f4a81416100 -1 mgr[py] Module rbd_support has missing NOTIFY_TYPES member 2026-03-06T13:41:30.655 INFO:journalctl@ceph.mgr.b.vm01.stdout:Mar 06 13:41:30 vm01 ceph-82f504c8-1959-11f1-9a16-a78d135e200b-mgr-b[57308]: 2026-03-06T12:41:30.169+0000 7f4a81416100 -1 mgr[py] Module rgw has missing NOTIFY_TYPES member 2026-03-06T13:41:31.336 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:41:31 vm01 ceph-mon[56331]: pgmap v16: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-06T13:41:31.502 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:41:31 vm00 ceph-mon[51914]: pgmap v16: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-06T13:41:31.539 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 06 13:41:31 vm08 ceph-mon[53561]: pgmap v16: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-06T13:41:31.614 INFO:journalctl@ceph.mgr.b.vm01.stdout:Mar 06 13:41:31 vm01 ceph-82f504c8-1959-11f1-9a16-a78d135e200b-mgr-b[57308]: 2026-03-06T12:41:31.334+0000 7f4a81416100 -1 mgr[py] Module rook has missing NOTIFY_TYPES member 2026-03-06T13:41:31.614 INFO:journalctl@ceph.mgr.b.vm01.stdout:Mar 06 13:41:31 vm01 ceph-82f504c8-1959-11f1-9a16-a78d135e200b-mgr-b[57308]: 2026-03-06T12:41:31.470+0000 7f4a81416100 -1 mgr[py] Module selftest has missing NOTIFY_TYPES member 2026-03-06T13:41:31.905 INFO:journalctl@ceph.mgr.b.vm01.stdout:Mar 06 13:41:31 vm01 ceph-82f504c8-1959-11f1-9a16-a78d135e200b-mgr-b[57308]: 2026-03-06T12:41:31.612+0000 7f4a81416100 -1 mgr[py] Module snap_schedule has missing NOTIFY_TYPES member 2026-03-06T13:41:32.405 INFO:journalctl@ceph.mgr.b.vm01.stdout:Mar 06 13:41:31 vm01 ceph-82f504c8-1959-11f1-9a16-a78d135e200b-mgr-b[57308]: 2026-03-06T12:41:31.915+0000 7f4a81416100 -1 mgr[py] Module status has missing NOTIFY_TYPES member 2026-03-06T13:41:32.405 INFO:journalctl@ceph.mgr.b.vm01.stdout:Mar 06 13:41:32 vm01 ceph-82f504c8-1959-11f1-9a16-a78d135e200b-mgr-b[57308]: 2026-03-06T12:41:32.054+0000 7f4a81416100 -1 mgr[py] Module telegraf has missing NOTIFY_TYPES member 2026-03-06T13:41:32.820 INFO:journalctl@ceph.mgr.b.vm01.stdout:Mar 06 13:41:32 vm01 ceph-82f504c8-1959-11f1-9a16-a78d135e200b-mgr-b[57308]: 2026-03-06T12:41:32.407+0000 7f4a81416100 -1 mgr[py] Module telemetry has missing NOTIFY_TYPES member 2026-03-06T13:41:33.155 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:41:33 vm01 ceph-mon[56331]: pgmap v17: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-06T13:41:33.155 INFO:journalctl@ceph.mgr.b.vm01.stdout:Mar 06 13:41:32 vm01 ceph-82f504c8-1959-11f1-9a16-a78d135e200b-mgr-b[57308]: 2026-03-06T12:41:32.818+0000 7f4a81416100 -1 mgr[py] Module test_orchestrator has missing NOTIFY_TYPES member 2026-03-06T13:41:33.502 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:41:33 vm00 ceph-mon[51914]: pgmap v17: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-06T13:41:33.540 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 06 13:41:33 vm08 ceph-mon[53561]: pgmap v17: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-06T13:41:33.655 INFO:journalctl@ceph.mgr.b.vm01.stdout:Mar 06 13:41:33 vm01 ceph-82f504c8-1959-11f1-9a16-a78d135e200b-mgr-b[57308]: 2026-03-06T12:41:33.232+0000 7f4a81416100 -1 mgr[py] Module volumes has missing NOTIFY_TYPES member 2026-03-06T13:41:33.655 INFO:journalctl@ceph.mgr.b.vm01.stdout:Mar 06 13:41:33 vm01 ceph-82f504c8-1959-11f1-9a16-a78d135e200b-mgr-b[57308]: 2026-03-06T12:41:33.367+0000 7f4a81416100 -1 mgr[py] Module zabbix has missing NOTIFY_TYPES member 2026-03-06T13:41:34.388 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:41:34 vm00 ceph-mon[51914]: from='mgr.14156 192.168.123.100:0/929894269' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "osd.0"}]: dispatch 2026-03-06T13:41:34.388 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:41:34 vm00 ceph-mon[51914]: from='mgr.14156 192.168.123.100:0/929894269' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-06T13:41:34.389 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:41:34 vm00 ceph-mon[51914]: from='mgr.? 192.168.123.101:0/2213926299' entity='mgr.b' cmd=[{"prefix": "config-key get", "key": "mgr/dashboard/b/crt"}]: dispatch 2026-03-06T13:41:34.389 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:41:34 vm00 ceph-mon[51914]: Standby manager daemon b started 2026-03-06T13:41:34.389 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:41:34 vm00 ceph-mon[51914]: from='mgr.? 192.168.123.101:0/2213926299' entity='mgr.b' cmd=[{"prefix": "config-key get", "key": "mgr/dashboard/crt"}]: dispatch 2026-03-06T13:41:34.389 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:41:34 vm00 ceph-mon[51914]: from='mgr.? 192.168.123.101:0/2213926299' entity='mgr.b' cmd=[{"prefix": "config-key get", "key": "mgr/dashboard/b/key"}]: dispatch 2026-03-06T13:41:34.389 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:41:34 vm00 ceph-mon[51914]: from='mgr.? 192.168.123.101:0/2213926299' entity='mgr.b' cmd=[{"prefix": "config-key get", "key": "mgr/dashboard/key"}]: dispatch 2026-03-06T13:41:34.405 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:41:34 vm01 ceph-mon[56331]: from='mgr.14156 192.168.123.100:0/929894269' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "osd.0"}]: dispatch 2026-03-06T13:41:34.405 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:41:34 vm01 ceph-mon[56331]: from='mgr.14156 192.168.123.100:0/929894269' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-06T13:41:34.405 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:41:34 vm01 ceph-mon[56331]: from='mgr.? 192.168.123.101:0/2213926299' entity='mgr.b' cmd=[{"prefix": "config-key get", "key": "mgr/dashboard/b/crt"}]: dispatch 2026-03-06T13:41:34.405 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:41:34 vm01 ceph-mon[56331]: Standby manager daemon b started 2026-03-06T13:41:34.405 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:41:34 vm01 ceph-mon[56331]: from='mgr.? 192.168.123.101:0/2213926299' entity='mgr.b' cmd=[{"prefix": "config-key get", "key": "mgr/dashboard/crt"}]: dispatch 2026-03-06T13:41:34.405 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:41:34 vm01 ceph-mon[56331]: from='mgr.? 192.168.123.101:0/2213926299' entity='mgr.b' cmd=[{"prefix": "config-key get", "key": "mgr/dashboard/b/key"}]: dispatch 2026-03-06T13:41:34.405 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:41:34 vm01 ceph-mon[56331]: from='mgr.? 192.168.123.101:0/2213926299' entity='mgr.b' cmd=[{"prefix": "config-key get", "key": "mgr/dashboard/key"}]: dispatch 2026-03-06T13:41:34.540 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 06 13:41:34 vm08 ceph-mon[53561]: from='mgr.14156 192.168.123.100:0/929894269' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "osd.0"}]: dispatch 2026-03-06T13:41:34.540 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 06 13:41:34 vm08 ceph-mon[53561]: from='mgr.14156 192.168.123.100:0/929894269' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-06T13:41:34.540 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 06 13:41:34 vm08 ceph-mon[53561]: from='mgr.? 192.168.123.101:0/2213926299' entity='mgr.b' cmd=[{"prefix": "config-key get", "key": "mgr/dashboard/b/crt"}]: dispatch 2026-03-06T13:41:34.540 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 06 13:41:34 vm08 ceph-mon[53561]: Standby manager daemon b started 2026-03-06T13:41:34.540 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 06 13:41:34 vm08 ceph-mon[53561]: from='mgr.? 192.168.123.101:0/2213926299' entity='mgr.b' cmd=[{"prefix": "config-key get", "key": "mgr/dashboard/crt"}]: dispatch 2026-03-06T13:41:34.540 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 06 13:41:34 vm08 ceph-mon[53561]: from='mgr.? 192.168.123.101:0/2213926299' entity='mgr.b' cmd=[{"prefix": "config-key get", "key": "mgr/dashboard/b/key"}]: dispatch 2026-03-06T13:41:34.540 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 06 13:41:34 vm08 ceph-mon[53561]: from='mgr.? 192.168.123.101:0/2213926299' entity='mgr.b' cmd=[{"prefix": "config-key get", "key": "mgr/dashboard/key"}]: dispatch 2026-03-06T13:41:35.405 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:41:35 vm01 ceph-mon[56331]: Deploying daemon osd.0 on vm00 2026-03-06T13:41:35.405 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:41:35 vm01 ceph-mon[56331]: pgmap v18: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-06T13:41:35.405 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:41:35 vm01 ceph-mon[56331]: mgrmap e14: a(active, since 46s), standbys: b 2026-03-06T13:41:35.405 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:41:35 vm01 ceph-mon[56331]: from='mgr.14156 192.168.123.100:0/929894269' entity='mgr.a' cmd=[{"prefix": "mgr metadata", "who": "b", "id": "b"}]: dispatch 2026-03-06T13:41:35.502 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:41:35 vm00 ceph-mon[51914]: Deploying daemon osd.0 on vm00 2026-03-06T13:41:35.502 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:41:35 vm00 ceph-mon[51914]: pgmap v18: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-06T13:41:35.502 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:41:35 vm00 ceph-mon[51914]: mgrmap e14: a(active, since 46s), standbys: b 2026-03-06T13:41:35.502 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:41:35 vm00 ceph-mon[51914]: from='mgr.14156 192.168.123.100:0/929894269' entity='mgr.a' cmd=[{"prefix": "mgr metadata", "who": "b", "id": "b"}]: dispatch 2026-03-06T13:41:35.540 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 06 13:41:35 vm08 ceph-mon[53561]: Deploying daemon osd.0 on vm00 2026-03-06T13:41:35.540 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 06 13:41:35 vm08 ceph-mon[53561]: pgmap v18: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-06T13:41:35.540 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 06 13:41:35 vm08 ceph-mon[53561]: mgrmap e14: a(active, since 46s), standbys: b 2026-03-06T13:41:35.540 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 06 13:41:35 vm08 ceph-mon[53561]: from='mgr.14156 192.168.123.100:0/929894269' entity='mgr.a' cmd=[{"prefix": "mgr metadata", "who": "b", "id": "b"}]: dispatch 2026-03-06T13:41:36.375 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:41:36 vm00 ceph-mon[51914]: pgmap v19: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-06T13:41:36.405 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:41:36 vm01 ceph-mon[56331]: pgmap v19: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-06T13:41:36.540 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 06 13:41:36 vm08 ceph-mon[53561]: pgmap v19: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-06T13:41:37.158 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:41:37 vm00 ceph-mon[51914]: from='mgr.14156 192.168.123.100:0/929894269' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-06T13:41:37.158 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:41:37 vm00 ceph-mon[51914]: from='mgr.14156 192.168.123.100:0/929894269' entity='mgr.a' 2026-03-06T13:41:37.158 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:41:37 vm00 ceph-mon[51914]: from='mgr.14156 192.168.123.100:0/929894269' entity='mgr.a' 2026-03-06T13:41:37.405 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:41:37 vm01 ceph-mon[56331]: from='mgr.14156 192.168.123.100:0/929894269' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-06T13:41:37.405 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:41:37 vm01 ceph-mon[56331]: from='mgr.14156 192.168.123.100:0/929894269' entity='mgr.a' 2026-03-06T13:41:37.405 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:41:37 vm01 ceph-mon[56331]: from='mgr.14156 192.168.123.100:0/929894269' entity='mgr.a' 2026-03-06T13:41:37.540 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 06 13:41:37 vm08 ceph-mon[53561]: from='mgr.14156 192.168.123.100:0/929894269' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-06T13:41:37.540 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 06 13:41:37 vm08 ceph-mon[53561]: from='mgr.14156 192.168.123.100:0/929894269' entity='mgr.a' 2026-03-06T13:41:37.540 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 06 13:41:37 vm08 ceph-mon[53561]: from='mgr.14156 192.168.123.100:0/929894269' entity='mgr.a' 2026-03-06T13:41:38.310 INFO:teuthology.orchestra.run.vm00.stdout:Created osd(s) 0 on host 'vm00' 2026-03-06T13:41:38.402 DEBUG:teuthology.orchestra.run.vm00:osd.0> sudo journalctl -f -n 0 -u ceph-82f504c8-1959-11f1-9a16-a78d135e200b@osd.0.service 2026-03-06T13:41:38.403 INFO:tasks.cephadm:Deploying osd.1 on vm01 with /dev/vde... 2026-03-06T13:41:38.403 DEBUG:teuthology.orchestra.run.vm01:> sudo /home/ubuntu/cephtest/cephadm --image harbor.clyso.com/custom-ceph/ceph/ceph:cobaltcore-storage-v19.2.3-fasttrack-3 ceph-volume -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 82f504c8-1959-11f1-9a16-a78d135e200b -- lvm zap /dev/vde 2026-03-06T13:41:38.437 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:41:38 vm00 ceph-mon[51914]: from='mgr.14156 192.168.123.100:0/929894269' entity='mgr.a' 2026-03-06T13:41:38.437 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:41:38 vm00 ceph-mon[51914]: from='mgr.14156 192.168.123.100:0/929894269' entity='mgr.a' 2026-03-06T13:41:38.437 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:41:38 vm00 ceph-mon[51914]: from='mgr.14156 192.168.123.100:0/929894269' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-06T13:41:38.437 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:41:38 vm00 ceph-mon[51914]: from='mgr.14156 192.168.123.100:0/929894269' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-06T13:41:38.437 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:41:38 vm00 ceph-mon[51914]: from='mgr.14156 192.168.123.100:0/929894269' entity='mgr.a' 2026-03-06T13:41:38.437 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:41:38 vm00 ceph-mon[51914]: pgmap v20: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-06T13:41:38.540 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 06 13:41:38 vm08 ceph-mon[53561]: from='mgr.14156 192.168.123.100:0/929894269' entity='mgr.a' 2026-03-06T13:41:38.540 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 06 13:41:38 vm08 ceph-mon[53561]: from='mgr.14156 192.168.123.100:0/929894269' entity='mgr.a' 2026-03-06T13:41:38.540 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 06 13:41:38 vm08 ceph-mon[53561]: from='mgr.14156 192.168.123.100:0/929894269' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-06T13:41:38.540 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 06 13:41:38 vm08 ceph-mon[53561]: from='mgr.14156 192.168.123.100:0/929894269' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-06T13:41:38.540 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 06 13:41:38 vm08 ceph-mon[53561]: from='mgr.14156 192.168.123.100:0/929894269' entity='mgr.a' 2026-03-06T13:41:38.540 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 06 13:41:38 vm08 ceph-mon[53561]: pgmap v20: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-06T13:41:38.655 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:41:38 vm01 ceph-mon[56331]: from='mgr.14156 192.168.123.100:0/929894269' entity='mgr.a' 2026-03-06T13:41:38.655 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:41:38 vm01 ceph-mon[56331]: from='mgr.14156 192.168.123.100:0/929894269' entity='mgr.a' 2026-03-06T13:41:38.655 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:41:38 vm01 ceph-mon[56331]: from='mgr.14156 192.168.123.100:0/929894269' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-06T13:41:38.655 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:41:38 vm01 ceph-mon[56331]: from='mgr.14156 192.168.123.100:0/929894269' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-06T13:41:38.655 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:41:38 vm01 ceph-mon[56331]: from='mgr.14156 192.168.123.100:0/929894269' entity='mgr.a' 2026-03-06T13:41:38.655 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:41:38 vm01 ceph-mon[56331]: pgmap v20: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-06T13:41:38.739 INFO:journalctl@ceph.osd.0.vm00.stdout:Mar 06 13:41:38 vm00 ceph-82f504c8-1959-11f1-9a16-a78d135e200b-osd-0[61746]: 2026-03-06T12:41:38.565+0000 7fb560608740 -1 osd.0 0 log_to_monitors true 2026-03-06T13:41:38.785 INFO:teuthology.orchestra.run.vm01.stderr:Inferring config /var/lib/ceph/82f504c8-1959-11f1-9a16-a78d135e200b/mon.b/config 2026-03-06T13:41:39.540 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 06 13:41:39 vm08 ceph-mon[53561]: from='mgr.14156 192.168.123.100:0/929894269' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-06T13:41:39.540 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 06 13:41:39 vm08 ceph-mon[53561]: from='mgr.14156 192.168.123.100:0/929894269' entity='mgr.a' 2026-03-06T13:41:39.540 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 06 13:41:39 vm08 ceph-mon[53561]: from='mgr.14156 192.168.123.100:0/929894269' entity='mgr.a' 2026-03-06T13:41:39.540 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 06 13:41:39 vm08 ceph-mon[53561]: from='osd.0 [v2:192.168.123.100:6802/2576257746,v1:192.168.123.100:6803/2576257746]' entity='osd.0' cmd=[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["0"]}]: dispatch 2026-03-06T13:41:39.571 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:41:39 vm01 ceph-mon[56331]: from='mgr.14156 192.168.123.100:0/929894269' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-06T13:41:39.571 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:41:39 vm01 ceph-mon[56331]: from='mgr.14156 192.168.123.100:0/929894269' entity='mgr.a' 2026-03-06T13:41:39.571 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:41:39 vm01 ceph-mon[56331]: from='mgr.14156 192.168.123.100:0/929894269' entity='mgr.a' 2026-03-06T13:41:39.571 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:41:39 vm01 ceph-mon[56331]: from='osd.0 [v2:192.168.123.100:6802/2576257746,v1:192.168.123.100:6803/2576257746]' entity='osd.0' cmd=[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["0"]}]: dispatch 2026-03-06T13:41:39.752 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:41:39 vm00 ceph-mon[51914]: from='mgr.14156 192.168.123.100:0/929894269' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-06T13:41:39.752 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:41:39 vm00 ceph-mon[51914]: from='mgr.14156 192.168.123.100:0/929894269' entity='mgr.a' 2026-03-06T13:41:39.752 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:41:39 vm00 ceph-mon[51914]: from='mgr.14156 192.168.123.100:0/929894269' entity='mgr.a' 2026-03-06T13:41:39.752 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:41:39 vm00 ceph-mon[51914]: from='osd.0 [v2:192.168.123.100:6802/2576257746,v1:192.168.123.100:6803/2576257746]' entity='osd.0' cmd=[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["0"]}]: dispatch 2026-03-06T13:41:39.985 INFO:teuthology.orchestra.run.vm01.stdout: 2026-03-06T13:41:40.017 DEBUG:teuthology.orchestra.run.vm01:> sudo /home/ubuntu/cephtest/cephadm --image harbor.clyso.com/custom-ceph/ceph/ceph:cobaltcore-storage-v19.2.3-fasttrack-3 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 82f504c8-1959-11f1-9a16-a78d135e200b -- ceph orch daemon add osd vm01:/dev/vde 2026-03-06T13:41:40.381 INFO:teuthology.orchestra.run.vm01.stderr:Inferring config /var/lib/ceph/82f504c8-1959-11f1-9a16-a78d135e200b/mon.b/config 2026-03-06T13:41:40.655 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:41:40 vm01 ceph-mon[56331]: from='osd.0 [v2:192.168.123.100:6802/2576257746,v1:192.168.123.100:6803/2576257746]' entity='osd.0' cmd='[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["0"]}]': finished 2026-03-06T13:41:40.655 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:41:40 vm01 ceph-mon[56331]: osdmap e6: 1 total, 0 up, 1 in 2026-03-06T13:41:40.655 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:41:40 vm01 ceph-mon[56331]: from='osd.0 [v2:192.168.123.100:6802/2576257746,v1:192.168.123.100:6803/2576257746]' entity='osd.0' cmd=[{"prefix": "osd crush create-or-move", "id": 0, "weight":0.0195, "args": ["host=vm00", "root=default"]}]: dispatch 2026-03-06T13:41:40.655 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:41:40 vm01 ceph-mon[56331]: from='mgr.14156 192.168.123.100:0/929894269' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 0}]: dispatch 2026-03-06T13:41:40.655 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:41:40 vm01 ceph-mon[56331]: Detected new or changed devices on vm00 2026-03-06T13:41:40.655 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:41:40 vm01 ceph-mon[56331]: from='mgr.14156 192.168.123.100:0/929894269' entity='mgr.a' 2026-03-06T13:41:40.655 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:41:40 vm01 ceph-mon[56331]: pgmap v22: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-06T13:41:40.655 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:41:40 vm01 ceph-mon[56331]: from='mgr.14156 192.168.123.100:0/929894269' entity='mgr.a' 2026-03-06T13:41:40.655 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:41:40 vm01 ceph-mon[56331]: from='mgr.14156 192.168.123.100:0/929894269' entity='mgr.a' cmd=[{"prefix": "config rm", "who": "osd.0", "name": "osd_memory_target"}]: dispatch 2026-03-06T13:41:40.655 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:41:40 vm01 ceph-mon[56331]: Adjusting osd_memory_target on vm00 to 257.0M 2026-03-06T13:41:40.655 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:41:40 vm01 ceph-mon[56331]: Unable to set osd_memory_target on vm00 to 269530726: error parsing value: Value '269530726' is below minimum 939524096 2026-03-06T13:41:40.655 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:41:40 vm01 ceph-mon[56331]: from='mgr.14156 192.168.123.100:0/929894269' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-06T13:41:40.655 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:41:40 vm01 ceph-mon[56331]: from='mgr.14156 192.168.123.100:0/929894269' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-06T13:41:40.655 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:41:40 vm01 ceph-mon[56331]: from='mgr.14156 192.168.123.100:0/929894269' entity='mgr.a' 2026-03-06T13:41:40.752 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:41:40 vm00 ceph-mon[51914]: from='osd.0 [v2:192.168.123.100:6802/2576257746,v1:192.168.123.100:6803/2576257746]' entity='osd.0' cmd='[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["0"]}]': finished 2026-03-06T13:41:40.752 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:41:40 vm00 ceph-mon[51914]: osdmap e6: 1 total, 0 up, 1 in 2026-03-06T13:41:40.752 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:41:40 vm00 ceph-mon[51914]: from='osd.0 [v2:192.168.123.100:6802/2576257746,v1:192.168.123.100:6803/2576257746]' entity='osd.0' cmd=[{"prefix": "osd crush create-or-move", "id": 0, "weight":0.0195, "args": ["host=vm00", "root=default"]}]: dispatch 2026-03-06T13:41:40.752 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:41:40 vm00 ceph-mon[51914]: from='mgr.14156 192.168.123.100:0/929894269' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 0}]: dispatch 2026-03-06T13:41:40.752 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:41:40 vm00 ceph-mon[51914]: Detected new or changed devices on vm00 2026-03-06T13:41:40.752 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:41:40 vm00 ceph-mon[51914]: from='mgr.14156 192.168.123.100:0/929894269' entity='mgr.a' 2026-03-06T13:41:40.752 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:41:40 vm00 ceph-mon[51914]: pgmap v22: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-06T13:41:40.752 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:41:40 vm00 ceph-mon[51914]: from='mgr.14156 192.168.123.100:0/929894269' entity='mgr.a' 2026-03-06T13:41:40.752 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:41:40 vm00 ceph-mon[51914]: from='mgr.14156 192.168.123.100:0/929894269' entity='mgr.a' cmd=[{"prefix": "config rm", "who": "osd.0", "name": "osd_memory_target"}]: dispatch 2026-03-06T13:41:40.752 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:41:40 vm00 ceph-mon[51914]: Adjusting osd_memory_target on vm00 to 257.0M 2026-03-06T13:41:40.752 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:41:40 vm00 ceph-mon[51914]: Unable to set osd_memory_target on vm00 to 269530726: error parsing value: Value '269530726' is below minimum 939524096 2026-03-06T13:41:40.752 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:41:40 vm00 ceph-mon[51914]: from='mgr.14156 192.168.123.100:0/929894269' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-06T13:41:40.752 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:41:40 vm00 ceph-mon[51914]: from='mgr.14156 192.168.123.100:0/929894269' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-06T13:41:40.752 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:41:40 vm00 ceph-mon[51914]: from='mgr.14156 192.168.123.100:0/929894269' entity='mgr.a' 2026-03-06T13:41:40.790 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 06 13:41:40 vm08 ceph-mon[53561]: from='osd.0 [v2:192.168.123.100:6802/2576257746,v1:192.168.123.100:6803/2576257746]' entity='osd.0' cmd='[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["0"]}]': finished 2026-03-06T13:41:40.790 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 06 13:41:40 vm08 ceph-mon[53561]: osdmap e6: 1 total, 0 up, 1 in 2026-03-06T13:41:40.790 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 06 13:41:40 vm08 ceph-mon[53561]: from='osd.0 [v2:192.168.123.100:6802/2576257746,v1:192.168.123.100:6803/2576257746]' entity='osd.0' cmd=[{"prefix": "osd crush create-or-move", "id": 0, "weight":0.0195, "args": ["host=vm00", "root=default"]}]: dispatch 2026-03-06T13:41:40.790 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 06 13:41:40 vm08 ceph-mon[53561]: from='mgr.14156 192.168.123.100:0/929894269' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 0}]: dispatch 2026-03-06T13:41:40.790 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 06 13:41:40 vm08 ceph-mon[53561]: Detected new or changed devices on vm00 2026-03-06T13:41:40.790 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 06 13:41:40 vm08 ceph-mon[53561]: from='mgr.14156 192.168.123.100:0/929894269' entity='mgr.a' 2026-03-06T13:41:40.790 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 06 13:41:40 vm08 ceph-mon[53561]: pgmap v22: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-06T13:41:40.790 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 06 13:41:40 vm08 ceph-mon[53561]: from='mgr.14156 192.168.123.100:0/929894269' entity='mgr.a' 2026-03-06T13:41:40.790 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 06 13:41:40 vm08 ceph-mon[53561]: from='mgr.14156 192.168.123.100:0/929894269' entity='mgr.a' cmd=[{"prefix": "config rm", "who": "osd.0", "name": "osd_memory_target"}]: dispatch 2026-03-06T13:41:40.790 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 06 13:41:40 vm08 ceph-mon[53561]: Adjusting osd_memory_target on vm00 to 257.0M 2026-03-06T13:41:40.790 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 06 13:41:40 vm08 ceph-mon[53561]: Unable to set osd_memory_target on vm00 to 269530726: error parsing value: Value '269530726' is below minimum 939524096 2026-03-06T13:41:40.790 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 06 13:41:40 vm08 ceph-mon[53561]: from='mgr.14156 192.168.123.100:0/929894269' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-06T13:41:40.790 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 06 13:41:40 vm08 ceph-mon[53561]: from='mgr.14156 192.168.123.100:0/929894269' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-06T13:41:40.790 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 06 13:41:40 vm08 ceph-mon[53561]: from='mgr.14156 192.168.123.100:0/929894269' entity='mgr.a' 2026-03-06T13:41:41.406 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:41:41 vm01 ceph-mon[56331]: from='osd.0 [v2:192.168.123.100:6802/2576257746,v1:192.168.123.100:6803/2576257746]' entity='osd.0' cmd='[{"prefix": "osd crush create-or-move", "id": 0, "weight":0.0195, "args": ["host=vm00", "root=default"]}]': finished 2026-03-06T13:41:41.406 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:41:41 vm01 ceph-mon[56331]: osdmap e7: 1 total, 0 up, 1 in 2026-03-06T13:41:41.406 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:41:41 vm01 ceph-mon[56331]: from='mgr.14156 192.168.123.100:0/929894269' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 0}]: dispatch 2026-03-06T13:41:41.406 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:41:41 vm01 ceph-mon[56331]: from='mgr.14156 192.168.123.100:0/929894269' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 0}]: dispatch 2026-03-06T13:41:41.406 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:41:41 vm01 ceph-mon[56331]: from='client.24122 -' entity='client.admin' cmd=[{"prefix": "orch daemon add osd", "svc_arg": "vm01:/dev/vde", "target": ["mon-mgr", ""]}]: dispatch 2026-03-06T13:41:41.406 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:41:41 vm01 ceph-mon[56331]: from='mgr.14156 192.168.123.100:0/929894269' entity='mgr.a' cmd=[{"prefix": "osd tree", "states": ["destroyed"], "format": "json"}]: dispatch 2026-03-06T13:41:41.406 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:41:41 vm01 ceph-mon[56331]: from='mgr.14156 192.168.123.100:0/929894269' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.bootstrap-osd"}]: dispatch 2026-03-06T13:41:41.406 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:41:41 vm01 ceph-mon[56331]: from='mgr.14156 192.168.123.100:0/929894269' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-06T13:41:41.406 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:41:41 vm01 ceph-mon[56331]: from='osd.0 [v2:192.168.123.100:6802/2576257746,v1:192.168.123.100:6803/2576257746]' entity='osd.0' 2026-03-06T13:41:41.502 INFO:journalctl@ceph.osd.0.vm00.stdout:Mar 06 13:41:41 vm00 ceph-82f504c8-1959-11f1-9a16-a78d135e200b-osd-0[61746]: 2026-03-06T12:41:41.225+0000 7fb55c589640 -1 osd.0 0 waiting for initial osdmap 2026-03-06T13:41:41.502 INFO:journalctl@ceph.osd.0.vm00.stdout:Mar 06 13:41:41 vm00 ceph-82f504c8-1959-11f1-9a16-a78d135e200b-osd-0[61746]: 2026-03-06T12:41:41.231+0000 7fb557bb2640 -1 osd.0 7 set_numa_affinity unable to identify public interface '' numa node: (2) No such file or directory 2026-03-06T13:41:41.502 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:41:41 vm00 ceph-mon[51914]: from='osd.0 [v2:192.168.123.100:6802/2576257746,v1:192.168.123.100:6803/2576257746]' entity='osd.0' cmd='[{"prefix": "osd crush create-or-move", "id": 0, "weight":0.0195, "args": ["host=vm00", "root=default"]}]': finished 2026-03-06T13:41:41.502 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:41:41 vm00 ceph-mon[51914]: osdmap e7: 1 total, 0 up, 1 in 2026-03-06T13:41:41.502 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:41:41 vm00 ceph-mon[51914]: from='mgr.14156 192.168.123.100:0/929894269' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 0}]: dispatch 2026-03-06T13:41:41.502 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:41:41 vm00 ceph-mon[51914]: from='mgr.14156 192.168.123.100:0/929894269' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 0}]: dispatch 2026-03-06T13:41:41.502 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:41:41 vm00 ceph-mon[51914]: from='client.24122 -' entity='client.admin' cmd=[{"prefix": "orch daemon add osd", "svc_arg": "vm01:/dev/vde", "target": ["mon-mgr", ""]}]: dispatch 2026-03-06T13:41:41.502 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:41:41 vm00 ceph-mon[51914]: from='mgr.14156 192.168.123.100:0/929894269' entity='mgr.a' cmd=[{"prefix": "osd tree", "states": ["destroyed"], "format": "json"}]: dispatch 2026-03-06T13:41:41.502 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:41:41 vm00 ceph-mon[51914]: from='mgr.14156 192.168.123.100:0/929894269' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.bootstrap-osd"}]: dispatch 2026-03-06T13:41:41.502 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:41:41 vm00 ceph-mon[51914]: from='mgr.14156 192.168.123.100:0/929894269' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-06T13:41:41.502 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:41:41 vm00 ceph-mon[51914]: from='osd.0 [v2:192.168.123.100:6802/2576257746,v1:192.168.123.100:6803/2576257746]' entity='osd.0' 2026-03-06T13:41:41.790 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 06 13:41:41 vm08 ceph-mon[53561]: from='osd.0 [v2:192.168.123.100:6802/2576257746,v1:192.168.123.100:6803/2576257746]' entity='osd.0' cmd='[{"prefix": "osd crush create-or-move", "id": 0, "weight":0.0195, "args": ["host=vm00", "root=default"]}]': finished 2026-03-06T13:41:41.790 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 06 13:41:41 vm08 ceph-mon[53561]: osdmap e7: 1 total, 0 up, 1 in 2026-03-06T13:41:41.790 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 06 13:41:41 vm08 ceph-mon[53561]: from='mgr.14156 192.168.123.100:0/929894269' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 0}]: dispatch 2026-03-06T13:41:41.790 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 06 13:41:41 vm08 ceph-mon[53561]: from='mgr.14156 192.168.123.100:0/929894269' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 0}]: dispatch 2026-03-06T13:41:41.790 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 06 13:41:41 vm08 ceph-mon[53561]: from='client.24122 -' entity='client.admin' cmd=[{"prefix": "orch daemon add osd", "svc_arg": "vm01:/dev/vde", "target": ["mon-mgr", ""]}]: dispatch 2026-03-06T13:41:41.790 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 06 13:41:41 vm08 ceph-mon[53561]: from='mgr.14156 192.168.123.100:0/929894269' entity='mgr.a' cmd=[{"prefix": "osd tree", "states": ["destroyed"], "format": "json"}]: dispatch 2026-03-06T13:41:41.790 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 06 13:41:41 vm08 ceph-mon[53561]: from='mgr.14156 192.168.123.100:0/929894269' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.bootstrap-osd"}]: dispatch 2026-03-06T13:41:41.790 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 06 13:41:41 vm08 ceph-mon[53561]: from='mgr.14156 192.168.123.100:0/929894269' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-06T13:41:41.790 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 06 13:41:41 vm08 ceph-mon[53561]: from='osd.0 [v2:192.168.123.100:6802/2576257746,v1:192.168.123.100:6803/2576257746]' entity='osd.0' 2026-03-06T13:41:42.405 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:41:42 vm01 ceph-mon[56331]: purged_snaps scrub starts 2026-03-06T13:41:42.405 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:41:42 vm01 ceph-mon[56331]: purged_snaps scrub ok 2026-03-06T13:41:42.405 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:41:42 vm01 ceph-mon[56331]: from='mgr.14156 192.168.123.100:0/929894269' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 0}]: dispatch 2026-03-06T13:41:42.405 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:41:42 vm01 ceph-mon[56331]: pgmap v24: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-06T13:41:42.405 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:41:42 vm01 ceph-mon[56331]: from='client.? 192.168.123.101:0/3951110304' entity='client.bootstrap-osd' cmd=[{"prefix": "osd new", "uuid": "95d964b0-4ea2-45f5-a612-b4aa5749cec2"}]: dispatch 2026-03-06T13:41:42.405 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:41:42 vm01 ceph-mon[56331]: from='client.? ' entity='client.bootstrap-osd' cmd=[{"prefix": "osd new", "uuid": "95d964b0-4ea2-45f5-a612-b4aa5749cec2"}]: dispatch 2026-03-06T13:41:42.405 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:41:42 vm01 ceph-mon[56331]: from='client.? ' entity='client.bootstrap-osd' cmd='[{"prefix": "osd new", "uuid": "95d964b0-4ea2-45f5-a612-b4aa5749cec2"}]': finished 2026-03-06T13:41:42.405 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:41:42 vm01 ceph-mon[56331]: osd.0 [v2:192.168.123.100:6802/2576257746,v1:192.168.123.100:6803/2576257746] boot 2026-03-06T13:41:42.405 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:41:42 vm01 ceph-mon[56331]: osdmap e8: 2 total, 1 up, 2 in 2026-03-06T13:41:42.405 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:41:42 vm01 ceph-mon[56331]: from='mgr.14156 192.168.123.100:0/929894269' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 0}]: dispatch 2026-03-06T13:41:42.405 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:41:42 vm01 ceph-mon[56331]: from='mgr.14156 192.168.123.100:0/929894269' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 1}]: dispatch 2026-03-06T13:41:42.752 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:41:42 vm00 ceph-mon[51914]: purged_snaps scrub starts 2026-03-06T13:41:42.752 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:41:42 vm00 ceph-mon[51914]: purged_snaps scrub ok 2026-03-06T13:41:42.752 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:41:42 vm00 ceph-mon[51914]: from='mgr.14156 192.168.123.100:0/929894269' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 0}]: dispatch 2026-03-06T13:41:42.752 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:41:42 vm00 ceph-mon[51914]: pgmap v24: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-06T13:41:42.752 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:41:42 vm00 ceph-mon[51914]: from='client.? 192.168.123.101:0/3951110304' entity='client.bootstrap-osd' cmd=[{"prefix": "osd new", "uuid": "95d964b0-4ea2-45f5-a612-b4aa5749cec2"}]: dispatch 2026-03-06T13:41:42.752 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:41:42 vm00 ceph-mon[51914]: from='client.? ' entity='client.bootstrap-osd' cmd=[{"prefix": "osd new", "uuid": "95d964b0-4ea2-45f5-a612-b4aa5749cec2"}]: dispatch 2026-03-06T13:41:42.752 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:41:42 vm00 ceph-mon[51914]: from='client.? ' entity='client.bootstrap-osd' cmd='[{"prefix": "osd new", "uuid": "95d964b0-4ea2-45f5-a612-b4aa5749cec2"}]': finished 2026-03-06T13:41:42.752 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:41:42 vm00 ceph-mon[51914]: osd.0 [v2:192.168.123.100:6802/2576257746,v1:192.168.123.100:6803/2576257746] boot 2026-03-06T13:41:42.752 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:41:42 vm00 ceph-mon[51914]: osdmap e8: 2 total, 1 up, 2 in 2026-03-06T13:41:42.752 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:41:42 vm00 ceph-mon[51914]: from='mgr.14156 192.168.123.100:0/929894269' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 0}]: dispatch 2026-03-06T13:41:42.752 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:41:42 vm00 ceph-mon[51914]: from='mgr.14156 192.168.123.100:0/929894269' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 1}]: dispatch 2026-03-06T13:41:42.790 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 06 13:41:42 vm08 ceph-mon[53561]: purged_snaps scrub starts 2026-03-06T13:41:42.790 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 06 13:41:42 vm08 ceph-mon[53561]: purged_snaps scrub ok 2026-03-06T13:41:42.790 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 06 13:41:42 vm08 ceph-mon[53561]: from='mgr.14156 192.168.123.100:0/929894269' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 0}]: dispatch 2026-03-06T13:41:42.790 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 06 13:41:42 vm08 ceph-mon[53561]: pgmap v24: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-06T13:41:42.790 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 06 13:41:42 vm08 ceph-mon[53561]: from='client.? 192.168.123.101:0/3951110304' entity='client.bootstrap-osd' cmd=[{"prefix": "osd new", "uuid": "95d964b0-4ea2-45f5-a612-b4aa5749cec2"}]: dispatch 2026-03-06T13:41:42.790 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 06 13:41:42 vm08 ceph-mon[53561]: from='client.? ' entity='client.bootstrap-osd' cmd=[{"prefix": "osd new", "uuid": "95d964b0-4ea2-45f5-a612-b4aa5749cec2"}]: dispatch 2026-03-06T13:41:42.790 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 06 13:41:42 vm08 ceph-mon[53561]: from='client.? ' entity='client.bootstrap-osd' cmd='[{"prefix": "osd new", "uuid": "95d964b0-4ea2-45f5-a612-b4aa5749cec2"}]': finished 2026-03-06T13:41:42.790 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 06 13:41:42 vm08 ceph-mon[53561]: osd.0 [v2:192.168.123.100:6802/2576257746,v1:192.168.123.100:6803/2576257746] boot 2026-03-06T13:41:42.790 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 06 13:41:42 vm08 ceph-mon[53561]: osdmap e8: 2 total, 1 up, 2 in 2026-03-06T13:41:42.790 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 06 13:41:42 vm08 ceph-mon[53561]: from='mgr.14156 192.168.123.100:0/929894269' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 0}]: dispatch 2026-03-06T13:41:42.790 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 06 13:41:42 vm08 ceph-mon[53561]: from='mgr.14156 192.168.123.100:0/929894269' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 1}]: dispatch 2026-03-06T13:41:43.655 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:41:43 vm01 ceph-mon[56331]: from='client.? 192.168.123.101:0/2530642756' entity='client.bootstrap-osd' cmd=[{"prefix": "mon getmap"}]: dispatch 2026-03-06T13:41:43.752 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:41:43 vm00 ceph-mon[51914]: from='client.? 192.168.123.101:0/2530642756' entity='client.bootstrap-osd' cmd=[{"prefix": "mon getmap"}]: dispatch 2026-03-06T13:41:43.790 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 06 13:41:43 vm08 ceph-mon[53561]: from='client.? 192.168.123.101:0/2530642756' entity='client.bootstrap-osd' cmd=[{"prefix": "mon getmap"}]: dispatch 2026-03-06T13:41:44.655 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:41:44 vm01 ceph-mon[56331]: osdmap e9: 2 total, 1 up, 2 in 2026-03-06T13:41:44.655 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:41:44 vm01 ceph-mon[56331]: from='mgr.14156 192.168.123.100:0/929894269' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 1}]: dispatch 2026-03-06T13:41:44.655 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:41:44 vm01 ceph-mon[56331]: pgmap v27: 0 pgs: ; 0 B data, 426 MiB used, 20 GiB / 20 GiB avail 2026-03-06T13:41:44.752 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:41:44 vm00 ceph-mon[51914]: osdmap e9: 2 total, 1 up, 2 in 2026-03-06T13:41:44.752 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:41:44 vm00 ceph-mon[51914]: from='mgr.14156 192.168.123.100:0/929894269' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 1}]: dispatch 2026-03-06T13:41:44.752 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:41:44 vm00 ceph-mon[51914]: pgmap v27: 0 pgs: ; 0 B data, 426 MiB used, 20 GiB / 20 GiB avail 2026-03-06T13:41:44.790 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 06 13:41:44 vm08 ceph-mon[53561]: osdmap e9: 2 total, 1 up, 2 in 2026-03-06T13:41:44.790 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 06 13:41:44 vm08 ceph-mon[53561]: from='mgr.14156 192.168.123.100:0/929894269' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 1}]: dispatch 2026-03-06T13:41:44.790 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 06 13:41:44 vm08 ceph-mon[53561]: pgmap v27: 0 pgs: ; 0 B data, 426 MiB used, 20 GiB / 20 GiB avail 2026-03-06T13:41:47.051 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:41:47 vm01 ceph-mon[56331]: pgmap v28: 0 pgs: ; 0 B data, 426 MiB used, 20 GiB / 20 GiB avail 2026-03-06T13:41:47.052 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:41:47 vm01 ceph-mon[56331]: from='mgr.14156 192.168.123.100:0/929894269' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "osd.1"}]: dispatch 2026-03-06T13:41:47.052 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:41:47 vm01 ceph-mon[56331]: from='mgr.14156 192.168.123.100:0/929894269' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-06T13:41:47.502 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:41:47 vm00 ceph-mon[51914]: pgmap v28: 0 pgs: ; 0 B data, 426 MiB used, 20 GiB / 20 GiB avail 2026-03-06T13:41:47.502 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:41:47 vm00 ceph-mon[51914]: from='mgr.14156 192.168.123.100:0/929894269' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "osd.1"}]: dispatch 2026-03-06T13:41:47.502 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:41:47 vm00 ceph-mon[51914]: from='mgr.14156 192.168.123.100:0/929894269' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-06T13:41:47.540 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 06 13:41:47 vm08 ceph-mon[53561]: pgmap v28: 0 pgs: ; 0 B data, 426 MiB used, 20 GiB / 20 GiB avail 2026-03-06T13:41:47.540 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 06 13:41:47 vm08 ceph-mon[53561]: from='mgr.14156 192.168.123.100:0/929894269' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "osd.1"}]: dispatch 2026-03-06T13:41:47.540 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 06 13:41:47 vm08 ceph-mon[53561]: from='mgr.14156 192.168.123.100:0/929894269' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-06T13:41:48.054 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:41:48 vm01 ceph-mon[56331]: Deploying daemon osd.1 on vm01 2026-03-06T13:41:48.502 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:41:48 vm00 ceph-mon[51914]: Deploying daemon osd.1 on vm01 2026-03-06T13:41:48.540 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 06 13:41:48 vm08 ceph-mon[53561]: Deploying daemon osd.1 on vm01 2026-03-06T13:41:49.058 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:41:49 vm01 ceph-mon[56331]: pgmap v29: 0 pgs: ; 0 B data, 426 MiB used, 20 GiB / 20 GiB avail 2026-03-06T13:41:49.502 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:41:49 vm00 ceph-mon[51914]: pgmap v29: 0 pgs: ; 0 B data, 426 MiB used, 20 GiB / 20 GiB avail 2026-03-06T13:41:49.540 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 06 13:41:49 vm08 ceph-mon[53561]: pgmap v29: 0 pgs: ; 0 B data, 426 MiB used, 20 GiB / 20 GiB avail 2026-03-06T13:41:50.063 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:41:50 vm01 ceph-mon[56331]: from='mgr.14156 192.168.123.100:0/929894269' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-06T13:41:50.063 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:41:50 vm01 ceph-mon[56331]: from='mgr.14156 192.168.123.100:0/929894269' entity='mgr.a' 2026-03-06T13:41:50.063 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:41:50 vm01 ceph-mon[56331]: from='mgr.14156 192.168.123.100:0/929894269' entity='mgr.a' 2026-03-06T13:41:50.502 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:41:50 vm00 ceph-mon[51914]: from='mgr.14156 192.168.123.100:0/929894269' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-06T13:41:50.502 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:41:50 vm00 ceph-mon[51914]: from='mgr.14156 192.168.123.100:0/929894269' entity='mgr.a' 2026-03-06T13:41:50.502 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:41:50 vm00 ceph-mon[51914]: from='mgr.14156 192.168.123.100:0/929894269' entity='mgr.a' 2026-03-06T13:41:50.540 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 06 13:41:50 vm08 ceph-mon[53561]: from='mgr.14156 192.168.123.100:0/929894269' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-06T13:41:50.540 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 06 13:41:50 vm08 ceph-mon[53561]: from='mgr.14156 192.168.123.100:0/929894269' entity='mgr.a' 2026-03-06T13:41:50.540 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 06 13:41:50 vm08 ceph-mon[53561]: from='mgr.14156 192.168.123.100:0/929894269' entity='mgr.a' 2026-03-06T13:41:51.072 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:41:51 vm01 ceph-mon[56331]: pgmap v30: 0 pgs: ; 0 B data, 426 MiB used, 20 GiB / 20 GiB avail 2026-03-06T13:41:51.072 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:41:51 vm01 ceph-mon[56331]: from='mgr.14156 192.168.123.100:0/929894269' entity='mgr.a' 2026-03-06T13:41:51.072 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:41:51 vm01 ceph-mon[56331]: from='mgr.14156 192.168.123.100:0/929894269' entity='mgr.a' 2026-03-06T13:41:51.072 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:41:51 vm01 ceph-mon[56331]: from='mgr.14156 192.168.123.100:0/929894269' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-06T13:41:51.072 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:41:51 vm01 ceph-mon[56331]: from='mgr.14156 192.168.123.100:0/929894269' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-06T13:41:51.072 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:41:51 vm01 ceph-mon[56331]: from='mgr.14156 192.168.123.100:0/929894269' entity='mgr.a' 2026-03-06T13:41:51.502 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:41:51 vm00 ceph-mon[51914]: pgmap v30: 0 pgs: ; 0 B data, 426 MiB used, 20 GiB / 20 GiB avail 2026-03-06T13:41:51.502 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:41:51 vm00 ceph-mon[51914]: from='mgr.14156 192.168.123.100:0/929894269' entity='mgr.a' 2026-03-06T13:41:51.502 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:41:51 vm00 ceph-mon[51914]: from='mgr.14156 192.168.123.100:0/929894269' entity='mgr.a' 2026-03-06T13:41:51.502 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:41:51 vm00 ceph-mon[51914]: from='mgr.14156 192.168.123.100:0/929894269' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-06T13:41:51.502 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:41:51 vm00 ceph-mon[51914]: from='mgr.14156 192.168.123.100:0/929894269' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-06T13:41:51.502 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:41:51 vm00 ceph-mon[51914]: from='mgr.14156 192.168.123.100:0/929894269' entity='mgr.a' 2026-03-06T13:41:51.540 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 06 13:41:51 vm08 ceph-mon[53561]: pgmap v30: 0 pgs: ; 0 B data, 426 MiB used, 20 GiB / 20 GiB avail 2026-03-06T13:41:51.540 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 06 13:41:51 vm08 ceph-mon[53561]: from='mgr.14156 192.168.123.100:0/929894269' entity='mgr.a' 2026-03-06T13:41:51.540 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 06 13:41:51 vm08 ceph-mon[53561]: from='mgr.14156 192.168.123.100:0/929894269' entity='mgr.a' 2026-03-06T13:41:51.540 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 06 13:41:51 vm08 ceph-mon[53561]: from='mgr.14156 192.168.123.100:0/929894269' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-06T13:41:51.540 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 06 13:41:51 vm08 ceph-mon[53561]: from='mgr.14156 192.168.123.100:0/929894269' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-06T13:41:51.540 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 06 13:41:51 vm08 ceph-mon[53561]: from='mgr.14156 192.168.123.100:0/929894269' entity='mgr.a' 2026-03-06T13:41:51.885 INFO:teuthology.orchestra.run.vm01.stdout:Created osd(s) 1 on host 'vm01' 2026-03-06T13:41:51.985 DEBUG:teuthology.orchestra.run.vm01:osd.1> sudo journalctl -f -n 0 -u ceph-82f504c8-1959-11f1-9a16-a78d135e200b@osd.1.service 2026-03-06T13:41:51.991 INFO:tasks.cephadm:Deploying osd.2 on vm08 with /dev/vde... 2026-03-06T13:41:51.991 DEBUG:teuthology.orchestra.run.vm08:> sudo /home/ubuntu/cephtest/cephadm --image harbor.clyso.com/custom-ceph/ceph/ceph:cobaltcore-storage-v19.2.3-fasttrack-3 ceph-volume -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 82f504c8-1959-11f1-9a16-a78d135e200b -- lvm zap /dev/vde 2026-03-06T13:41:52.290 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 06 13:41:52 vm08 ceph-mon[53561]: from='mgr.14156 192.168.123.100:0/929894269' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-06T13:41:52.290 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 06 13:41:52 vm08 ceph-mon[53561]: from='mgr.14156 192.168.123.100:0/929894269' entity='mgr.a' 2026-03-06T13:41:52.290 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 06 13:41:52 vm08 ceph-mon[53561]: from='mgr.14156 192.168.123.100:0/929894269' entity='mgr.a' 2026-03-06T13:41:52.290 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 06 13:41:52 vm08 ceph-mon[53561]: from='osd.1 [v2:192.168.123.101:6800/1940387388,v1:192.168.123.101:6801/1940387388]' entity='osd.1' cmd=[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["1"]}]: dispatch 2026-03-06T13:41:52.290 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 06 13:41:52 vm08 ceph-mon[53561]: from='osd.1 ' entity='osd.1' cmd=[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["1"]}]: dispatch 2026-03-06T13:41:52.346 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:41:52 vm01 ceph-mon[56331]: from='mgr.14156 192.168.123.100:0/929894269' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-06T13:41:52.346 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:41:52 vm01 ceph-mon[56331]: from='mgr.14156 192.168.123.100:0/929894269' entity='mgr.a' 2026-03-06T13:41:52.346 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:41:52 vm01 ceph-mon[56331]: from='mgr.14156 192.168.123.100:0/929894269' entity='mgr.a' 2026-03-06T13:41:52.346 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:41:52 vm01 ceph-mon[56331]: from='osd.1 [v2:192.168.123.101:6800/1940387388,v1:192.168.123.101:6801/1940387388]' entity='osd.1' cmd=[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["1"]}]: dispatch 2026-03-06T13:41:52.346 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:41:52 vm01 ceph-mon[56331]: from='osd.1 ' entity='osd.1' cmd=[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["1"]}]: dispatch 2026-03-06T13:41:52.359 INFO:teuthology.orchestra.run.vm08.stderr:Inferring config /var/lib/ceph/82f504c8-1959-11f1-9a16-a78d135e200b/mon.c/config 2026-03-06T13:41:52.502 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:41:52 vm00 ceph-mon[51914]: from='mgr.14156 192.168.123.100:0/929894269' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-06T13:41:52.502 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:41:52 vm00 ceph-mon[51914]: from='mgr.14156 192.168.123.100:0/929894269' entity='mgr.a' 2026-03-06T13:41:52.502 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:41:52 vm00 ceph-mon[51914]: from='mgr.14156 192.168.123.100:0/929894269' entity='mgr.a' 2026-03-06T13:41:52.502 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:41:52 vm00 ceph-mon[51914]: from='osd.1 [v2:192.168.123.101:6800/1940387388,v1:192.168.123.101:6801/1940387388]' entity='osd.1' cmd=[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["1"]}]: dispatch 2026-03-06T13:41:52.502 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:41:52 vm00 ceph-mon[51914]: from='osd.1 ' entity='osd.1' cmd=[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["1"]}]: dispatch 2026-03-06T13:41:53.405 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:41:53 vm01 ceph-mon[56331]: pgmap v31: 0 pgs: ; 0 B data, 426 MiB used, 20 GiB / 20 GiB avail 2026-03-06T13:41:53.405 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:41:53 vm01 ceph-mon[56331]: from='osd.1 ' entity='osd.1' cmd='[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["1"]}]': finished 2026-03-06T13:41:53.405 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:41:53 vm01 ceph-mon[56331]: from='osd.1 [v2:192.168.123.101:6800/1940387388,v1:192.168.123.101:6801/1940387388]' entity='osd.1' cmd=[{"prefix": "osd crush create-or-move", "id": 1, "weight":0.0195, "args": ["host=vm01", "root=default"]}]: dispatch 2026-03-06T13:41:53.405 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:41:53 vm01 ceph-mon[56331]: osdmap e10: 2 total, 1 up, 2 in 2026-03-06T13:41:53.405 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:41:53 vm01 ceph-mon[56331]: from='mgr.14156 192.168.123.100:0/929894269' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 1}]: dispatch 2026-03-06T13:41:53.405 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:41:53 vm01 ceph-mon[56331]: from='osd.1 ' entity='osd.1' cmd=[{"prefix": "osd crush create-or-move", "id": 1, "weight":0.0195, "args": ["host=vm01", "root=default"]}]: dispatch 2026-03-06T13:41:53.437 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 06 13:41:53 vm08 ceph-mon[53561]: pgmap v31: 0 pgs: ; 0 B data, 426 MiB used, 20 GiB / 20 GiB avail 2026-03-06T13:41:53.437 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 06 13:41:53 vm08 ceph-mon[53561]: from='osd.1 ' entity='osd.1' cmd='[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["1"]}]': finished 2026-03-06T13:41:53.437 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 06 13:41:53 vm08 ceph-mon[53561]: from='osd.1 [v2:192.168.123.101:6800/1940387388,v1:192.168.123.101:6801/1940387388]' entity='osd.1' cmd=[{"prefix": "osd crush create-or-move", "id": 1, "weight":0.0195, "args": ["host=vm01", "root=default"]}]: dispatch 2026-03-06T13:41:53.437 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 06 13:41:53 vm08 ceph-mon[53561]: osdmap e10: 2 total, 1 up, 2 in 2026-03-06T13:41:53.437 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 06 13:41:53 vm08 ceph-mon[53561]: from='mgr.14156 192.168.123.100:0/929894269' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 1}]: dispatch 2026-03-06T13:41:53.437 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 06 13:41:53 vm08 ceph-mon[53561]: from='osd.1 ' entity='osd.1' cmd=[{"prefix": "osd crush create-or-move", "id": 1, "weight":0.0195, "args": ["host=vm01", "root=default"]}]: dispatch 2026-03-06T13:41:53.502 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:41:53 vm00 ceph-mon[51914]: pgmap v31: 0 pgs: ; 0 B data, 426 MiB used, 20 GiB / 20 GiB avail 2026-03-06T13:41:53.502 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:41:53 vm00 ceph-mon[51914]: from='osd.1 ' entity='osd.1' cmd='[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["1"]}]': finished 2026-03-06T13:41:53.502 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:41:53 vm00 ceph-mon[51914]: from='osd.1 [v2:192.168.123.101:6800/1940387388,v1:192.168.123.101:6801/1940387388]' entity='osd.1' cmd=[{"prefix": "osd crush create-or-move", "id": 1, "weight":0.0195, "args": ["host=vm01", "root=default"]}]: dispatch 2026-03-06T13:41:53.502 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:41:53 vm00 ceph-mon[51914]: osdmap e10: 2 total, 1 up, 2 in 2026-03-06T13:41:53.502 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:41:53 vm00 ceph-mon[51914]: from='mgr.14156 192.168.123.100:0/929894269' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 1}]: dispatch 2026-03-06T13:41:53.502 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:41:53 vm00 ceph-mon[51914]: from='osd.1 ' entity='osd.1' cmd=[{"prefix": "osd crush create-or-move", "id": 1, "weight":0.0195, "args": ["host=vm01", "root=default"]}]: dispatch 2026-03-06T13:41:53.575 INFO:teuthology.orchestra.run.vm08.stdout: 2026-03-06T13:41:53.595 DEBUG:teuthology.orchestra.run.vm08:> sudo /home/ubuntu/cephtest/cephadm --image harbor.clyso.com/custom-ceph/ceph/ceph:cobaltcore-storage-v19.2.3-fasttrack-3 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 82f504c8-1959-11f1-9a16-a78d135e200b -- ceph orch daemon add osd vm08:/dev/vde 2026-03-06T13:41:53.898 INFO:journalctl@ceph.osd.1.vm01.stdout:Mar 06 13:41:53 vm01 ceph-82f504c8-1959-11f1-9a16-a78d135e200b-osd-1[59988]: 2026-03-06T12:41:53.894+0000 7f2a5e07f640 -1 osd.1 0 waiting for initial osdmap 2026-03-06T13:41:53.963 INFO:teuthology.orchestra.run.vm08.stderr:Inferring config /var/lib/ceph/82f504c8-1959-11f1-9a16-a78d135e200b/mon.c/config 2026-03-06T13:41:54.155 INFO:journalctl@ceph.osd.1.vm01.stdout:Mar 06 13:41:53 vm01 ceph-82f504c8-1959-11f1-9a16-a78d135e200b-osd-1[59988]: 2026-03-06T12:41:53.904+0000 7f2a59ea9640 -1 osd.1 11 set_numa_affinity unable to identify public interface '' numa node: (2) No such file or directory 2026-03-06T13:41:54.793 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 06 13:41:54 vm08 ceph-mon[53561]: Detected new or changed devices on vm01 2026-03-06T13:41:54.793 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 06 13:41:54 vm08 ceph-mon[53561]: from='mgr.14156 192.168.123.100:0/929894269' entity='mgr.a' 2026-03-06T13:41:54.793 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 06 13:41:54 vm08 ceph-mon[53561]: from='mgr.14156 192.168.123.100:0/929894269' entity='mgr.a' 2026-03-06T13:41:54.793 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 06 13:41:54 vm08 ceph-mon[53561]: from='mgr.14156 192.168.123.100:0/929894269' entity='mgr.a' cmd=[{"prefix": "config rm", "who": "osd.1", "name": "osd_memory_target"}]: dispatch 2026-03-06T13:41:54.793 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 06 13:41:54 vm08 ceph-mon[53561]: Adjusting osd_memory_target on vm01 to 257.0M 2026-03-06T13:41:54.793 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 06 13:41:54 vm08 ceph-mon[53561]: Unable to set osd_memory_target on vm01 to 269536460: error parsing value: Value '269536460' is below minimum 939524096 2026-03-06T13:41:54.793 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 06 13:41:54 vm08 ceph-mon[53561]: from='mgr.14156 192.168.123.100:0/929894269' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-06T13:41:54.793 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 06 13:41:54 vm08 ceph-mon[53561]: from='mgr.14156 192.168.123.100:0/929894269' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-06T13:41:54.793 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 06 13:41:54 vm08 ceph-mon[53561]: from='mgr.14156 192.168.123.100:0/929894269' entity='mgr.a' 2026-03-06T13:41:54.793 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 06 13:41:54 vm08 ceph-mon[53561]: from='osd.1 ' entity='osd.1' cmd='[{"prefix": "osd crush create-or-move", "id": 1, "weight":0.0195, "args": ["host=vm01", "root=default"]}]': finished 2026-03-06T13:41:54.793 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 06 13:41:54 vm08 ceph-mon[53561]: osdmap e11: 2 total, 1 up, 2 in 2026-03-06T13:41:54.793 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 06 13:41:54 vm08 ceph-mon[53561]: from='mgr.14156 192.168.123.100:0/929894269' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 1}]: dispatch 2026-03-06T13:41:54.793 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 06 13:41:54 vm08 ceph-mon[53561]: from='mgr.14156 192.168.123.100:0/929894269' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 1}]: dispatch 2026-03-06T13:41:54.793 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 06 13:41:54 vm08 ceph-mon[53561]: pgmap v34: 0 pgs: ; 0 B data, 426 MiB used, 20 GiB / 20 GiB avail 2026-03-06T13:41:54.793 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 06 13:41:54 vm08 ceph-mon[53561]: from='mgr.14156 192.168.123.100:0/929894269' entity='mgr.a' cmd=[{"prefix": "osd tree", "states": ["destroyed"], "format": "json"}]: dispatch 2026-03-06T13:41:54.793 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 06 13:41:54 vm08 ceph-mon[53561]: from='mgr.14156 192.168.123.100:0/929894269' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.bootstrap-osd"}]: dispatch 2026-03-06T13:41:54.793 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 06 13:41:54 vm08 ceph-mon[53561]: from='mgr.14156 192.168.123.100:0/929894269' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-06T13:41:54.905 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:41:54 vm01 ceph-mon[56331]: Detected new or changed devices on vm01 2026-03-06T13:41:54.905 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:41:54 vm01 ceph-mon[56331]: from='mgr.14156 192.168.123.100:0/929894269' entity='mgr.a' 2026-03-06T13:41:54.905 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:41:54 vm01 ceph-mon[56331]: from='mgr.14156 192.168.123.100:0/929894269' entity='mgr.a' 2026-03-06T13:41:54.905 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:41:54 vm01 ceph-mon[56331]: from='mgr.14156 192.168.123.100:0/929894269' entity='mgr.a' cmd=[{"prefix": "config rm", "who": "osd.1", "name": "osd_memory_target"}]: dispatch 2026-03-06T13:41:54.905 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:41:54 vm01 ceph-mon[56331]: Adjusting osd_memory_target on vm01 to 257.0M 2026-03-06T13:41:54.905 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:41:54 vm01 ceph-mon[56331]: Unable to set osd_memory_target on vm01 to 269536460: error parsing value: Value '269536460' is below minimum 939524096 2026-03-06T13:41:54.905 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:41:54 vm01 ceph-mon[56331]: from='mgr.14156 192.168.123.100:0/929894269' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-06T13:41:54.905 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:41:54 vm01 ceph-mon[56331]: from='mgr.14156 192.168.123.100:0/929894269' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-06T13:41:54.905 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:41:54 vm01 ceph-mon[56331]: from='mgr.14156 192.168.123.100:0/929894269' entity='mgr.a' 2026-03-06T13:41:54.905 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:41:54 vm01 ceph-mon[56331]: from='osd.1 ' entity='osd.1' cmd='[{"prefix": "osd crush create-or-move", "id": 1, "weight":0.0195, "args": ["host=vm01", "root=default"]}]': finished 2026-03-06T13:41:54.905 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:41:54 vm01 ceph-mon[56331]: osdmap e11: 2 total, 1 up, 2 in 2026-03-06T13:41:54.905 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:41:54 vm01 ceph-mon[56331]: from='mgr.14156 192.168.123.100:0/929894269' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 1}]: dispatch 2026-03-06T13:41:54.905 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:41:54 vm01 ceph-mon[56331]: from='mgr.14156 192.168.123.100:0/929894269' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 1}]: dispatch 2026-03-06T13:41:54.905 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:41:54 vm01 ceph-mon[56331]: pgmap v34: 0 pgs: ; 0 B data, 426 MiB used, 20 GiB / 20 GiB avail 2026-03-06T13:41:54.905 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:41:54 vm01 ceph-mon[56331]: from='mgr.14156 192.168.123.100:0/929894269' entity='mgr.a' cmd=[{"prefix": "osd tree", "states": ["destroyed"], "format": "json"}]: dispatch 2026-03-06T13:41:54.905 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:41:54 vm01 ceph-mon[56331]: from='mgr.14156 192.168.123.100:0/929894269' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.bootstrap-osd"}]: dispatch 2026-03-06T13:41:54.905 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:41:54 vm01 ceph-mon[56331]: from='mgr.14156 192.168.123.100:0/929894269' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-06T13:41:55.002 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:41:54 vm00 ceph-mon[51914]: Detected new or changed devices on vm01 2026-03-06T13:41:55.002 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:41:54 vm00 ceph-mon[51914]: from='mgr.14156 192.168.123.100:0/929894269' entity='mgr.a' 2026-03-06T13:41:55.002 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:41:54 vm00 ceph-mon[51914]: from='mgr.14156 192.168.123.100:0/929894269' entity='mgr.a' 2026-03-06T13:41:55.002 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:41:54 vm00 ceph-mon[51914]: from='mgr.14156 192.168.123.100:0/929894269' entity='mgr.a' cmd=[{"prefix": "config rm", "who": "osd.1", "name": "osd_memory_target"}]: dispatch 2026-03-06T13:41:55.002 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:41:54 vm00 ceph-mon[51914]: Adjusting osd_memory_target on vm01 to 257.0M 2026-03-06T13:41:55.002 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:41:54 vm00 ceph-mon[51914]: Unable to set osd_memory_target on vm01 to 269536460: error parsing value: Value '269536460' is below minimum 939524096 2026-03-06T13:41:55.002 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:41:54 vm00 ceph-mon[51914]: from='mgr.14156 192.168.123.100:0/929894269' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-06T13:41:55.002 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:41:54 vm00 ceph-mon[51914]: from='mgr.14156 192.168.123.100:0/929894269' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-06T13:41:55.003 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:41:54 vm00 ceph-mon[51914]: from='mgr.14156 192.168.123.100:0/929894269' entity='mgr.a' 2026-03-06T13:41:55.003 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:41:54 vm00 ceph-mon[51914]: from='osd.1 ' entity='osd.1' cmd='[{"prefix": "osd crush create-or-move", "id": 1, "weight":0.0195, "args": ["host=vm01", "root=default"]}]': finished 2026-03-06T13:41:55.003 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:41:54 vm00 ceph-mon[51914]: osdmap e11: 2 total, 1 up, 2 in 2026-03-06T13:41:55.003 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:41:54 vm00 ceph-mon[51914]: from='mgr.14156 192.168.123.100:0/929894269' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 1}]: dispatch 2026-03-06T13:41:55.003 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:41:54 vm00 ceph-mon[51914]: from='mgr.14156 192.168.123.100:0/929894269' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 1}]: dispatch 2026-03-06T13:41:55.003 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:41:54 vm00 ceph-mon[51914]: pgmap v34: 0 pgs: ; 0 B data, 426 MiB used, 20 GiB / 20 GiB avail 2026-03-06T13:41:55.003 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:41:54 vm00 ceph-mon[51914]: from='mgr.14156 192.168.123.100:0/929894269' entity='mgr.a' cmd=[{"prefix": "osd tree", "states": ["destroyed"], "format": "json"}]: dispatch 2026-03-06T13:41:55.003 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:41:54 vm00 ceph-mon[51914]: from='mgr.14156 192.168.123.100:0/929894269' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.bootstrap-osd"}]: dispatch 2026-03-06T13:41:55.003 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:41:54 vm00 ceph-mon[51914]: from='mgr.14156 192.168.123.100:0/929894269' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-06T13:41:56.155 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:41:55 vm01 ceph-mon[56331]: purged_snaps scrub starts 2026-03-06T13:41:56.155 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:41:55 vm01 ceph-mon[56331]: purged_snaps scrub ok 2026-03-06T13:41:56.155 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:41:55 vm01 ceph-mon[56331]: from='client.14268 -' entity='client.admin' cmd=[{"prefix": "orch daemon add osd", "svc_arg": "vm08:/dev/vde", "target": ["mon-mgr", ""]}]: dispatch 2026-03-06T13:41:56.155 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:41:55 vm01 ceph-mon[56331]: from='mgr.14156 192.168.123.100:0/929894269' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 1}]: dispatch 2026-03-06T13:41:56.155 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:41:55 vm01 ceph-mon[56331]: osd.1 [v2:192.168.123.101:6800/1940387388,v1:192.168.123.101:6801/1940387388] boot 2026-03-06T13:41:56.155 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:41:55 vm01 ceph-mon[56331]: osdmap e12: 2 total, 2 up, 2 in 2026-03-06T13:41:56.155 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:41:55 vm01 ceph-mon[56331]: from='mgr.14156 192.168.123.100:0/929894269' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 1}]: dispatch 2026-03-06T13:41:56.155 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:41:55 vm01 ceph-mon[56331]: from='client.? 192.168.123.108:0/2535659898' entity='client.bootstrap-osd' cmd=[{"prefix": "osd new", "uuid": "9e99c305-cccd-4040-a383-099c1410ee92"}]: dispatch 2026-03-06T13:41:56.155 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:41:55 vm01 ceph-mon[56331]: from='client.? ' entity='client.bootstrap-osd' cmd=[{"prefix": "osd new", "uuid": "9e99c305-cccd-4040-a383-099c1410ee92"}]: dispatch 2026-03-06T13:41:56.155 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:41:55 vm01 ceph-mon[56331]: from='client.? ' entity='client.bootstrap-osd' cmd='[{"prefix": "osd new", "uuid": "9e99c305-cccd-4040-a383-099c1410ee92"}]': finished 2026-03-06T13:41:56.155 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:41:55 vm01 ceph-mon[56331]: osdmap e13: 3 total, 2 up, 3 in 2026-03-06T13:41:56.155 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:41:55 vm01 ceph-mon[56331]: from='mgr.14156 192.168.123.100:0/929894269' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 2}]: dispatch 2026-03-06T13:41:56.235 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 06 13:41:55 vm08 ceph-mon[53561]: purged_snaps scrub starts 2026-03-06T13:41:56.235 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 06 13:41:55 vm08 ceph-mon[53561]: purged_snaps scrub ok 2026-03-06T13:41:56.235 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 06 13:41:55 vm08 ceph-mon[53561]: from='client.14268 -' entity='client.admin' cmd=[{"prefix": "orch daemon add osd", "svc_arg": "vm08:/dev/vde", "target": ["mon-mgr", ""]}]: dispatch 2026-03-06T13:41:56.235 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 06 13:41:55 vm08 ceph-mon[53561]: from='mgr.14156 192.168.123.100:0/929894269' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 1}]: dispatch 2026-03-06T13:41:56.235 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 06 13:41:55 vm08 ceph-mon[53561]: osd.1 [v2:192.168.123.101:6800/1940387388,v1:192.168.123.101:6801/1940387388] boot 2026-03-06T13:41:56.235 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 06 13:41:55 vm08 ceph-mon[53561]: osdmap e12: 2 total, 2 up, 2 in 2026-03-06T13:41:56.235 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 06 13:41:55 vm08 ceph-mon[53561]: from='mgr.14156 192.168.123.100:0/929894269' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 1}]: dispatch 2026-03-06T13:41:56.235 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 06 13:41:55 vm08 ceph-mon[53561]: from='client.? 192.168.123.108:0/2535659898' entity='client.bootstrap-osd' cmd=[{"prefix": "osd new", "uuid": "9e99c305-cccd-4040-a383-099c1410ee92"}]: dispatch 2026-03-06T13:41:56.235 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 06 13:41:55 vm08 ceph-mon[53561]: from='client.? ' entity='client.bootstrap-osd' cmd=[{"prefix": "osd new", "uuid": "9e99c305-cccd-4040-a383-099c1410ee92"}]: dispatch 2026-03-06T13:41:56.235 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 06 13:41:55 vm08 ceph-mon[53561]: from='client.? ' entity='client.bootstrap-osd' cmd='[{"prefix": "osd new", "uuid": "9e99c305-cccd-4040-a383-099c1410ee92"}]': finished 2026-03-06T13:41:56.235 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 06 13:41:55 vm08 ceph-mon[53561]: osdmap e13: 3 total, 2 up, 3 in 2026-03-06T13:41:56.235 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 06 13:41:55 vm08 ceph-mon[53561]: from='mgr.14156 192.168.123.100:0/929894269' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 2}]: dispatch 2026-03-06T13:41:56.252 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:41:55 vm00 ceph-mon[51914]: purged_snaps scrub starts 2026-03-06T13:41:56.252 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:41:55 vm00 ceph-mon[51914]: purged_snaps scrub ok 2026-03-06T13:41:56.252 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:41:55 vm00 ceph-mon[51914]: from='client.14268 -' entity='client.admin' cmd=[{"prefix": "orch daemon add osd", "svc_arg": "vm08:/dev/vde", "target": ["mon-mgr", ""]}]: dispatch 2026-03-06T13:41:56.252 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:41:55 vm00 ceph-mon[51914]: from='mgr.14156 192.168.123.100:0/929894269' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 1}]: dispatch 2026-03-06T13:41:56.252 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:41:55 vm00 ceph-mon[51914]: osd.1 [v2:192.168.123.101:6800/1940387388,v1:192.168.123.101:6801/1940387388] boot 2026-03-06T13:41:56.252 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:41:55 vm00 ceph-mon[51914]: osdmap e12: 2 total, 2 up, 2 in 2026-03-06T13:41:56.252 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:41:55 vm00 ceph-mon[51914]: from='mgr.14156 192.168.123.100:0/929894269' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 1}]: dispatch 2026-03-06T13:41:56.252 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:41:55 vm00 ceph-mon[51914]: from='client.? 192.168.123.108:0/2535659898' entity='client.bootstrap-osd' cmd=[{"prefix": "osd new", "uuid": "9e99c305-cccd-4040-a383-099c1410ee92"}]: dispatch 2026-03-06T13:41:56.252 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:41:55 vm00 ceph-mon[51914]: from='client.? ' entity='client.bootstrap-osd' cmd=[{"prefix": "osd new", "uuid": "9e99c305-cccd-4040-a383-099c1410ee92"}]: dispatch 2026-03-06T13:41:56.252 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:41:55 vm00 ceph-mon[51914]: from='client.? ' entity='client.bootstrap-osd' cmd='[{"prefix": "osd new", "uuid": "9e99c305-cccd-4040-a383-099c1410ee92"}]': finished 2026-03-06T13:41:56.252 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:41:55 vm00 ceph-mon[51914]: osdmap e13: 3 total, 2 up, 3 in 2026-03-06T13:41:56.252 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:41:55 vm00 ceph-mon[51914]: from='mgr.14156 192.168.123.100:0/929894269' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 2}]: dispatch 2026-03-06T13:41:57.252 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:41:56 vm00 ceph-mon[51914]: pgmap v37: 0 pgs: ; 0 B data, 453 MiB used, 40 GiB / 40 GiB avail 2026-03-06T13:41:57.252 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:41:56 vm00 ceph-mon[51914]: from='client.? 192.168.123.108:0/706621895' entity='client.bootstrap-osd' cmd=[{"prefix": "mon getmap"}]: dispatch 2026-03-06T13:41:57.290 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 06 13:41:56 vm08 ceph-mon[53561]: pgmap v37: 0 pgs: ; 0 B data, 453 MiB used, 40 GiB / 40 GiB avail 2026-03-06T13:41:57.290 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 06 13:41:56 vm08 ceph-mon[53561]: from='client.? 192.168.123.108:0/706621895' entity='client.bootstrap-osd' cmd=[{"prefix": "mon getmap"}]: dispatch 2026-03-06T13:41:57.405 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:41:56 vm01 ceph-mon[56331]: pgmap v37: 0 pgs: ; 0 B data, 453 MiB used, 40 GiB / 40 GiB avail 2026-03-06T13:41:57.405 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:41:56 vm01 ceph-mon[56331]: from='client.? 192.168.123.108:0/706621895' entity='client.bootstrap-osd' cmd=[{"prefix": "mon getmap"}]: dispatch 2026-03-06T13:41:59.405 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:41:59 vm01 ceph-mon[56331]: pgmap v38: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-06T13:41:59.502 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:41:59 vm00 ceph-mon[51914]: pgmap v38: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-06T13:41:59.513 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 06 13:41:59 vm08 ceph-mon[53561]: pgmap v38: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-06T13:42:01.325 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 06 13:42:01 vm08 ceph-mon[53561]: pgmap v39: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-06T13:42:01.326 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 06 13:42:01 vm08 ceph-mon[53561]: from='mgr.14156 192.168.123.100:0/929894269' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "osd.2"}]: dispatch 2026-03-06T13:42:01.326 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 06 13:42:01 vm08 ceph-mon[53561]: from='mgr.14156 192.168.123.100:0/929894269' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-06T13:42:01.405 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:42:01 vm01 ceph-mon[56331]: pgmap v39: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-06T13:42:01.405 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:42:01 vm01 ceph-mon[56331]: from='mgr.14156 192.168.123.100:0/929894269' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "osd.2"}]: dispatch 2026-03-06T13:42:01.405 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:42:01 vm01 ceph-mon[56331]: from='mgr.14156 192.168.123.100:0/929894269' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-06T13:42:01.502 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:42:01 vm00 ceph-mon[51914]: pgmap v39: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-06T13:42:01.502 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:42:01 vm00 ceph-mon[51914]: from='mgr.14156 192.168.123.100:0/929894269' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "osd.2"}]: dispatch 2026-03-06T13:42:01.502 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:42:01 vm00 ceph-mon[51914]: from='mgr.14156 192.168.123.100:0/929894269' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-06T13:42:02.332 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 06 13:42:02 vm08 ceph-mon[53561]: Deploying daemon osd.2 on vm08 2026-03-06T13:42:02.405 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:42:02 vm01 ceph-mon[56331]: Deploying daemon osd.2 on vm08 2026-03-06T13:42:02.502 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:42:02 vm00 ceph-mon[51914]: Deploying daemon osd.2 on vm08 2026-03-06T13:42:03.405 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:42:03 vm01 ceph-mon[56331]: pgmap v40: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-06T13:42:03.406 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 06 13:42:03 vm08 ceph-mon[53561]: pgmap v40: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-06T13:42:03.502 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:42:03 vm00 ceph-mon[51914]: pgmap v40: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-06T13:42:05.227 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 06 13:42:05 vm08 ceph-mon[53561]: pgmap v41: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-06T13:42:05.228 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 06 13:42:05 vm08 ceph-mon[53561]: from='mgr.14156 192.168.123.100:0/929894269' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-06T13:42:05.228 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 06 13:42:05 vm08 ceph-mon[53561]: from='mgr.14156 192.168.123.100:0/929894269' entity='mgr.a' 2026-03-06T13:42:05.228 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 06 13:42:05 vm08 ceph-mon[53561]: from='mgr.14156 192.168.123.100:0/929894269' entity='mgr.a' 2026-03-06T13:42:05.228 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 06 13:42:05 vm08 ceph-mon[53561]: from='mgr.14156 192.168.123.100:0/929894269' entity='mgr.a' 2026-03-06T13:42:05.228 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 06 13:42:05 vm08 ceph-mon[53561]: from='mgr.14156 192.168.123.100:0/929894269' entity='mgr.a' 2026-03-06T13:42:05.405 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:42:05 vm01 ceph-mon[56331]: pgmap v41: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-06T13:42:05.405 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:42:05 vm01 ceph-mon[56331]: from='mgr.14156 192.168.123.100:0/929894269' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-06T13:42:05.405 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:42:05 vm01 ceph-mon[56331]: from='mgr.14156 192.168.123.100:0/929894269' entity='mgr.a' 2026-03-06T13:42:05.405 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:42:05 vm01 ceph-mon[56331]: from='mgr.14156 192.168.123.100:0/929894269' entity='mgr.a' 2026-03-06T13:42:05.405 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:42:05 vm01 ceph-mon[56331]: from='mgr.14156 192.168.123.100:0/929894269' entity='mgr.a' 2026-03-06T13:42:05.405 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:42:05 vm01 ceph-mon[56331]: from='mgr.14156 192.168.123.100:0/929894269' entity='mgr.a' 2026-03-06T13:42:05.502 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:42:05 vm00 ceph-mon[51914]: pgmap v41: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-06T13:42:05.502 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:42:05 vm00 ceph-mon[51914]: from='mgr.14156 192.168.123.100:0/929894269' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-06T13:42:05.502 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:42:05 vm00 ceph-mon[51914]: from='mgr.14156 192.168.123.100:0/929894269' entity='mgr.a' 2026-03-06T13:42:05.502 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:42:05 vm00 ceph-mon[51914]: from='mgr.14156 192.168.123.100:0/929894269' entity='mgr.a' 2026-03-06T13:42:05.502 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:42:05 vm00 ceph-mon[51914]: from='mgr.14156 192.168.123.100:0/929894269' entity='mgr.a' 2026-03-06T13:42:05.502 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:42:05 vm00 ceph-mon[51914]: from='mgr.14156 192.168.123.100:0/929894269' entity='mgr.a' 2026-03-06T13:42:06.089 INFO:teuthology.orchestra.run.vm08.stdout:Created osd(s) 2 on host 'vm08' 2026-03-06T13:42:06.173 DEBUG:teuthology.orchestra.run.vm08:osd.2> sudo journalctl -f -n 0 -u ceph-82f504c8-1959-11f1-9a16-a78d135e200b@osd.2.service 2026-03-06T13:42:06.175 INFO:tasks.cephadm:Waiting for 3 OSDs to come up... 2026-03-06T13:42:06.175 DEBUG:teuthology.orchestra.run.vm00:> sudo /home/ubuntu/cephtest/cephadm --image harbor.clyso.com/custom-ceph/ceph/ceph:cobaltcore-storage-v19.2.3-fasttrack-3 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 82f504c8-1959-11f1-9a16-a78d135e200b -- ceph osd stat -f json 2026-03-06T13:42:06.292 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 06 13:42:06 vm08 ceph-mon[53561]: from='mgr.14156 192.168.123.100:0/929894269' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-06T13:42:06.292 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 06 13:42:06 vm08 ceph-mon[53561]: from='mgr.14156 192.168.123.100:0/929894269' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-06T13:42:06.292 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 06 13:42:06 vm08 ceph-mon[53561]: from='mgr.14156 192.168.123.100:0/929894269' entity='mgr.a' 2026-03-06T13:42:06.292 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 06 13:42:06 vm08 ceph-mon[53561]: from='mgr.14156 192.168.123.100:0/929894269' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-06T13:42:06.405 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:42:06 vm01 ceph-mon[56331]: from='mgr.14156 192.168.123.100:0/929894269' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-06T13:42:06.405 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:42:06 vm01 ceph-mon[56331]: from='mgr.14156 192.168.123.100:0/929894269' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-06T13:42:06.405 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:42:06 vm01 ceph-mon[56331]: from='mgr.14156 192.168.123.100:0/929894269' entity='mgr.a' 2026-03-06T13:42:06.405 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:42:06 vm01 ceph-mon[56331]: from='mgr.14156 192.168.123.100:0/929894269' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-06T13:42:06.502 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:42:06 vm00 ceph-mon[51914]: from='mgr.14156 192.168.123.100:0/929894269' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-06T13:42:06.502 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:42:06 vm00 ceph-mon[51914]: from='mgr.14156 192.168.123.100:0/929894269' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-06T13:42:06.502 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:42:06 vm00 ceph-mon[51914]: from='mgr.14156 192.168.123.100:0/929894269' entity='mgr.a' 2026-03-06T13:42:06.502 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:42:06 vm00 ceph-mon[51914]: from='mgr.14156 192.168.123.100:0/929894269' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-06T13:42:06.547 INFO:teuthology.orchestra.run.vm00.stderr:Inferring config /var/lib/ceph/82f504c8-1959-11f1-9a16-a78d135e200b/mon.a/config 2026-03-06T13:42:06.588 INFO:journalctl@ceph.osd.2.vm08.stdout:Mar 06 13:42:06 vm08 ceph-82f504c8-1959-11f1-9a16-a78d135e200b-osd-2[57083]: 2026-03-06T12:42:06.330+0000 7fb191809740 -1 osd.2 0 log_to_monitors true 2026-03-06T13:42:06.922 INFO:teuthology.orchestra.run.vm00.stdout: 2026-03-06T13:42:07.007 INFO:teuthology.orchestra.run.vm00.stdout:{"epoch":13,"num_osds":3,"num_up_osds":2,"osd_up_since":1772800914,"num_in_osds":3,"osd_in_since":1772800915,"num_remapped_pgs":0} 2026-03-06T13:42:07.252 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:42:07 vm00 ceph-mon[51914]: pgmap v42: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-06T13:42:07.252 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:42:07 vm00 ceph-mon[51914]: from='mgr.14156 192.168.123.100:0/929894269' entity='mgr.a' 2026-03-06T13:42:07.252 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:42:07 vm00 ceph-mon[51914]: from='mgr.14156 192.168.123.100:0/929894269' entity='mgr.a' 2026-03-06T13:42:07.252 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:42:07 vm00 ceph-mon[51914]: from='osd.2 [v2:192.168.123.108:6800/4086912105,v1:192.168.123.108:6801/4086912105]' entity='osd.2' cmd=[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["2"]}]: dispatch 2026-03-06T13:42:07.252 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:42:07 vm00 ceph-mon[51914]: from='osd.2 ' entity='osd.2' cmd=[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["2"]}]: dispatch 2026-03-06T13:42:07.252 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:42:07 vm00 ceph-mon[51914]: from='client.? 192.168.123.100:0/1131489235' entity='client.admin' cmd=[{"prefix": "osd stat", "format": "json"}]: dispatch 2026-03-06T13:42:07.405 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:42:07 vm01 ceph-mon[56331]: pgmap v42: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-06T13:42:07.405 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:42:07 vm01 ceph-mon[56331]: from='mgr.14156 192.168.123.100:0/929894269' entity='mgr.a' 2026-03-06T13:42:07.405 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:42:07 vm01 ceph-mon[56331]: from='mgr.14156 192.168.123.100:0/929894269' entity='mgr.a' 2026-03-06T13:42:07.405 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:42:07 vm01 ceph-mon[56331]: from='osd.2 [v2:192.168.123.108:6800/4086912105,v1:192.168.123.108:6801/4086912105]' entity='osd.2' cmd=[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["2"]}]: dispatch 2026-03-06T13:42:07.405 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:42:07 vm01 ceph-mon[56331]: from='osd.2 ' entity='osd.2' cmd=[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["2"]}]: dispatch 2026-03-06T13:42:07.405 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:42:07 vm01 ceph-mon[56331]: from='client.? 192.168.123.100:0/1131489235' entity='client.admin' cmd=[{"prefix": "osd stat", "format": "json"}]: dispatch 2026-03-06T13:42:07.540 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 06 13:42:07 vm08 ceph-mon[53561]: pgmap v42: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-06T13:42:07.540 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 06 13:42:07 vm08 ceph-mon[53561]: from='mgr.14156 192.168.123.100:0/929894269' entity='mgr.a' 2026-03-06T13:42:07.540 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 06 13:42:07 vm08 ceph-mon[53561]: from='mgr.14156 192.168.123.100:0/929894269' entity='mgr.a' 2026-03-06T13:42:07.540 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 06 13:42:07 vm08 ceph-mon[53561]: from='osd.2 [v2:192.168.123.108:6800/4086912105,v1:192.168.123.108:6801/4086912105]' entity='osd.2' cmd=[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["2"]}]: dispatch 2026-03-06T13:42:07.540 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 06 13:42:07 vm08 ceph-mon[53561]: from='osd.2 ' entity='osd.2' cmd=[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["2"]}]: dispatch 2026-03-06T13:42:07.540 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 06 13:42:07 vm08 ceph-mon[53561]: from='client.? 192.168.123.100:0/1131489235' entity='client.admin' cmd=[{"prefix": "osd stat", "format": "json"}]: dispatch 2026-03-06T13:42:08.007 DEBUG:teuthology.orchestra.run.vm00:> sudo /home/ubuntu/cephtest/cephadm --image harbor.clyso.com/custom-ceph/ceph/ceph:cobaltcore-storage-v19.2.3-fasttrack-3 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 82f504c8-1959-11f1-9a16-a78d135e200b -- ceph osd stat -f json 2026-03-06T13:42:08.366 INFO:teuthology.orchestra.run.vm00.stderr:Inferring config /var/lib/ceph/82f504c8-1959-11f1-9a16-a78d135e200b/mon.a/config 2026-03-06T13:42:08.391 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:42:08 vm00 ceph-mon[51914]: from='osd.2 ' entity='osd.2' cmd='[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["2"]}]': finished 2026-03-06T13:42:08.391 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:42:08 vm00 ceph-mon[51914]: osdmap e14: 3 total, 2 up, 3 in 2026-03-06T13:42:08.391 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:42:08 vm00 ceph-mon[51914]: from='mgr.14156 192.168.123.100:0/929894269' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 2}]: dispatch 2026-03-06T13:42:08.391 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:42:08 vm00 ceph-mon[51914]: from='osd.2 [v2:192.168.123.108:6800/4086912105,v1:192.168.123.108:6801/4086912105]' entity='osd.2' cmd=[{"prefix": "osd crush create-or-move", "id": 2, "weight":0.0195, "args": ["host=vm08", "root=default"]}]: dispatch 2026-03-06T13:42:08.391 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:42:08 vm00 ceph-mon[51914]: from='osd.2 ' entity='osd.2' cmd=[{"prefix": "osd crush create-or-move", "id": 2, "weight":0.0195, "args": ["host=vm08", "root=default"]}]: dispatch 2026-03-06T13:42:08.391 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:42:08 vm00 ceph-mon[51914]: Detected new or changed devices on vm08 2026-03-06T13:42:08.391 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:42:08 vm00 ceph-mon[51914]: from='mgr.14156 192.168.123.100:0/929894269' entity='mgr.a' 2026-03-06T13:42:08.391 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:42:08 vm00 ceph-mon[51914]: from='mgr.14156 192.168.123.100:0/929894269' entity='mgr.a' 2026-03-06T13:42:08.391 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:42:08 vm00 ceph-mon[51914]: from='mgr.14156 192.168.123.100:0/929894269' entity='mgr.a' cmd=[{"prefix": "config rm", "who": "osd.2", "name": "osd_memory_target"}]: dispatch 2026-03-06T13:42:08.391 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:42:08 vm00 ceph-mon[51914]: Adjusting osd_memory_target on vm08 to 4353M 2026-03-06T13:42:08.391 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:42:08 vm00 ceph-mon[51914]: from='mgr.14156 192.168.123.100:0/929894269' entity='mgr.a' 2026-03-06T13:42:08.391 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:42:08 vm00 ceph-mon[51914]: from='mgr.14156 192.168.123.100:0/929894269' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-06T13:42:08.391 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:42:08 vm00 ceph-mon[51914]: from='mgr.14156 192.168.123.100:0/929894269' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-06T13:42:08.391 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:42:08 vm00 ceph-mon[51914]: from='mgr.14156 192.168.123.100:0/929894269' entity='mgr.a' 2026-03-06T13:42:08.392 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:42:08 vm00 ceph-mon[51914]: pgmap v44: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-06T13:42:08.405 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:42:08 vm01 ceph-mon[56331]: from='osd.2 ' entity='osd.2' cmd='[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["2"]}]': finished 2026-03-06T13:42:08.405 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:42:08 vm01 ceph-mon[56331]: osdmap e14: 3 total, 2 up, 3 in 2026-03-06T13:42:08.405 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:42:08 vm01 ceph-mon[56331]: from='mgr.14156 192.168.123.100:0/929894269' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 2}]: dispatch 2026-03-06T13:42:08.405 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:42:08 vm01 ceph-mon[56331]: from='osd.2 [v2:192.168.123.108:6800/4086912105,v1:192.168.123.108:6801/4086912105]' entity='osd.2' cmd=[{"prefix": "osd crush create-or-move", "id": 2, "weight":0.0195, "args": ["host=vm08", "root=default"]}]: dispatch 2026-03-06T13:42:08.405 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:42:08 vm01 ceph-mon[56331]: from='osd.2 ' entity='osd.2' cmd=[{"prefix": "osd crush create-or-move", "id": 2, "weight":0.0195, "args": ["host=vm08", "root=default"]}]: dispatch 2026-03-06T13:42:08.405 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:42:08 vm01 ceph-mon[56331]: Detected new or changed devices on vm08 2026-03-06T13:42:08.405 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:42:08 vm01 ceph-mon[56331]: from='mgr.14156 192.168.123.100:0/929894269' entity='mgr.a' 2026-03-06T13:42:08.405 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:42:08 vm01 ceph-mon[56331]: from='mgr.14156 192.168.123.100:0/929894269' entity='mgr.a' 2026-03-06T13:42:08.405 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:42:08 vm01 ceph-mon[56331]: from='mgr.14156 192.168.123.100:0/929894269' entity='mgr.a' cmd=[{"prefix": "config rm", "who": "osd.2", "name": "osd_memory_target"}]: dispatch 2026-03-06T13:42:08.405 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:42:08 vm01 ceph-mon[56331]: Adjusting osd_memory_target on vm08 to 4353M 2026-03-06T13:42:08.405 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:42:08 vm01 ceph-mon[56331]: from='mgr.14156 192.168.123.100:0/929894269' entity='mgr.a' 2026-03-06T13:42:08.405 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:42:08 vm01 ceph-mon[56331]: from='mgr.14156 192.168.123.100:0/929894269' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-06T13:42:08.405 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:42:08 vm01 ceph-mon[56331]: from='mgr.14156 192.168.123.100:0/929894269' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-06T13:42:08.405 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:42:08 vm01 ceph-mon[56331]: from='mgr.14156 192.168.123.100:0/929894269' entity='mgr.a' 2026-03-06T13:42:08.405 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:42:08 vm01 ceph-mon[56331]: pgmap v44: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-06T13:42:08.540 INFO:journalctl@ceph.osd.2.vm08.stdout:Mar 06 13:42:08 vm08 ceph-82f504c8-1959-11f1-9a16-a78d135e200b-osd-2[57083]: 2026-03-06T12:42:08.094+0000 7fb18d78a640 -1 osd.2 0 waiting for initial osdmap 2026-03-06T13:42:08.540 INFO:journalctl@ceph.osd.2.vm08.stdout:Mar 06 13:42:08 vm08 ceph-82f504c8-1959-11f1-9a16-a78d135e200b-osd-2[57083]: 2026-03-06T12:42:08.098+0000 7fb1895b4640 -1 osd.2 15 set_numa_affinity unable to identify public interface '' numa node: (2) No such file or directory 2026-03-06T13:42:08.540 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 06 13:42:08 vm08 ceph-mon[53561]: from='osd.2 ' entity='osd.2' cmd='[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["2"]}]': finished 2026-03-06T13:42:08.540 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 06 13:42:08 vm08 ceph-mon[53561]: osdmap e14: 3 total, 2 up, 3 in 2026-03-06T13:42:08.540 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 06 13:42:08 vm08 ceph-mon[53561]: from='mgr.14156 192.168.123.100:0/929894269' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 2}]: dispatch 2026-03-06T13:42:08.540 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 06 13:42:08 vm08 ceph-mon[53561]: from='osd.2 [v2:192.168.123.108:6800/4086912105,v1:192.168.123.108:6801/4086912105]' entity='osd.2' cmd=[{"prefix": "osd crush create-or-move", "id": 2, "weight":0.0195, "args": ["host=vm08", "root=default"]}]: dispatch 2026-03-06T13:42:08.540 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 06 13:42:08 vm08 ceph-mon[53561]: from='osd.2 ' entity='osd.2' cmd=[{"prefix": "osd crush create-or-move", "id": 2, "weight":0.0195, "args": ["host=vm08", "root=default"]}]: dispatch 2026-03-06T13:42:08.540 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 06 13:42:08 vm08 ceph-mon[53561]: Detected new or changed devices on vm08 2026-03-06T13:42:08.540 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 06 13:42:08 vm08 ceph-mon[53561]: from='mgr.14156 192.168.123.100:0/929894269' entity='mgr.a' 2026-03-06T13:42:08.540 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 06 13:42:08 vm08 ceph-mon[53561]: from='mgr.14156 192.168.123.100:0/929894269' entity='mgr.a' 2026-03-06T13:42:08.540 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 06 13:42:08 vm08 ceph-mon[53561]: from='mgr.14156 192.168.123.100:0/929894269' entity='mgr.a' cmd=[{"prefix": "config rm", "who": "osd.2", "name": "osd_memory_target"}]: dispatch 2026-03-06T13:42:08.540 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 06 13:42:08 vm08 ceph-mon[53561]: Adjusting osd_memory_target on vm08 to 4353M 2026-03-06T13:42:08.540 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 06 13:42:08 vm08 ceph-mon[53561]: from='mgr.14156 192.168.123.100:0/929894269' entity='mgr.a' 2026-03-06T13:42:08.540 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 06 13:42:08 vm08 ceph-mon[53561]: from='mgr.14156 192.168.123.100:0/929894269' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-06T13:42:08.540 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 06 13:42:08 vm08 ceph-mon[53561]: from='mgr.14156 192.168.123.100:0/929894269' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-06T13:42:08.540 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 06 13:42:08 vm08 ceph-mon[53561]: from='mgr.14156 192.168.123.100:0/929894269' entity='mgr.a' 2026-03-06T13:42:08.540 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 06 13:42:08 vm08 ceph-mon[53561]: pgmap v44: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-06T13:42:08.715 INFO:teuthology.orchestra.run.vm00.stdout: 2026-03-06T13:42:08.791 INFO:teuthology.orchestra.run.vm00.stdout:{"epoch":15,"num_osds":3,"num_up_osds":2,"osd_up_since":1772800914,"num_in_osds":3,"osd_in_since":1772800915,"num_remapped_pgs":0} 2026-03-06T13:42:09.405 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:42:09 vm01 ceph-mon[56331]: from='osd.2 ' entity='osd.2' cmd='[{"prefix": "osd crush create-or-move", "id": 2, "weight":0.0195, "args": ["host=vm08", "root=default"]}]': finished 2026-03-06T13:42:09.405 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:42:09 vm01 ceph-mon[56331]: osdmap e15: 3 total, 2 up, 3 in 2026-03-06T13:42:09.405 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:42:09 vm01 ceph-mon[56331]: from='mgr.14156 192.168.123.100:0/929894269' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 2}]: dispatch 2026-03-06T13:42:09.405 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:42:09 vm01 ceph-mon[56331]: from='mgr.14156 192.168.123.100:0/929894269' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 2}]: dispatch 2026-03-06T13:42:09.405 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:42:09 vm01 ceph-mon[56331]: from='client.? 192.168.123.100:0/1819223882' entity='client.admin' cmd=[{"prefix": "osd stat", "format": "json"}]: dispatch 2026-03-06T13:42:09.502 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:42:09 vm00 ceph-mon[51914]: from='osd.2 ' entity='osd.2' cmd='[{"prefix": "osd crush create-or-move", "id": 2, "weight":0.0195, "args": ["host=vm08", "root=default"]}]': finished 2026-03-06T13:42:09.502 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:42:09 vm00 ceph-mon[51914]: osdmap e15: 3 total, 2 up, 3 in 2026-03-06T13:42:09.502 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:42:09 vm00 ceph-mon[51914]: from='mgr.14156 192.168.123.100:0/929894269' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 2}]: dispatch 2026-03-06T13:42:09.502 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:42:09 vm00 ceph-mon[51914]: from='mgr.14156 192.168.123.100:0/929894269' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 2}]: dispatch 2026-03-06T13:42:09.502 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:42:09 vm00 ceph-mon[51914]: from='client.? 192.168.123.100:0/1819223882' entity='client.admin' cmd=[{"prefix": "osd stat", "format": "json"}]: dispatch 2026-03-06T13:42:09.540 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 06 13:42:09 vm08 ceph-mon[53561]: from='osd.2 ' entity='osd.2' cmd='[{"prefix": "osd crush create-or-move", "id": 2, "weight":0.0195, "args": ["host=vm08", "root=default"]}]': finished 2026-03-06T13:42:09.540 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 06 13:42:09 vm08 ceph-mon[53561]: osdmap e15: 3 total, 2 up, 3 in 2026-03-06T13:42:09.540 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 06 13:42:09 vm08 ceph-mon[53561]: from='mgr.14156 192.168.123.100:0/929894269' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 2}]: dispatch 2026-03-06T13:42:09.540 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 06 13:42:09 vm08 ceph-mon[53561]: from='mgr.14156 192.168.123.100:0/929894269' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 2}]: dispatch 2026-03-06T13:42:09.540 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 06 13:42:09 vm08 ceph-mon[53561]: from='client.? 192.168.123.100:0/1819223882' entity='client.admin' cmd=[{"prefix": "osd stat", "format": "json"}]: dispatch 2026-03-06T13:42:09.792 DEBUG:teuthology.orchestra.run.vm00:> sudo /home/ubuntu/cephtest/cephadm --image harbor.clyso.com/custom-ceph/ceph/ceph:cobaltcore-storage-v19.2.3-fasttrack-3 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 82f504c8-1959-11f1-9a16-a78d135e200b -- ceph osd stat -f json 2026-03-06T13:42:10.098 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:42:10 vm00 ceph-mon[51914]: purged_snaps scrub starts 2026-03-06T13:42:10.098 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:42:10 vm00 ceph-mon[51914]: purged_snaps scrub ok 2026-03-06T13:42:10.098 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:42:10 vm00 ceph-mon[51914]: from='mgr.14156 192.168.123.100:0/929894269' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 2}]: dispatch 2026-03-06T13:42:10.098 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:42:10 vm00 ceph-mon[51914]: osd.2 [v2:192.168.123.108:6800/4086912105,v1:192.168.123.108:6801/4086912105] boot 2026-03-06T13:42:10.098 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:42:10 vm00 ceph-mon[51914]: osdmap e16: 3 total, 3 up, 3 in 2026-03-06T13:42:10.098 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:42:10 vm00 ceph-mon[51914]: from='mgr.14156 192.168.123.100:0/929894269' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 2}]: dispatch 2026-03-06T13:42:10.098 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:42:10 vm00 ceph-mon[51914]: pgmap v47: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-06T13:42:10.098 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:42:10 vm00 ceph-mon[51914]: from='mgr.14156 192.168.123.100:0/929894269' entity='mgr.a' cmd=[{"prefix": "osd pool create", "format": "json", "pool": ".mgr", "pg_num": 1, "pg_num_min": 1, "pg_num_max": 32, "yes_i_really_mean_it": true}]: dispatch 2026-03-06T13:42:10.144 INFO:teuthology.orchestra.run.vm00.stderr:Inferring config /var/lib/ceph/82f504c8-1959-11f1-9a16-a78d135e200b/mon.a/config 2026-03-06T13:42:10.405 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:42:10 vm01 ceph-mon[56331]: purged_snaps scrub starts 2026-03-06T13:42:10.405 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:42:10 vm01 ceph-mon[56331]: purged_snaps scrub ok 2026-03-06T13:42:10.405 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:42:10 vm01 ceph-mon[56331]: from='mgr.14156 192.168.123.100:0/929894269' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 2}]: dispatch 2026-03-06T13:42:10.405 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:42:10 vm01 ceph-mon[56331]: osd.2 [v2:192.168.123.108:6800/4086912105,v1:192.168.123.108:6801/4086912105] boot 2026-03-06T13:42:10.405 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:42:10 vm01 ceph-mon[56331]: osdmap e16: 3 total, 3 up, 3 in 2026-03-06T13:42:10.405 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:42:10 vm01 ceph-mon[56331]: from='mgr.14156 192.168.123.100:0/929894269' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 2}]: dispatch 2026-03-06T13:42:10.405 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:42:10 vm01 ceph-mon[56331]: pgmap v47: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-06T13:42:10.405 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:42:10 vm01 ceph-mon[56331]: from='mgr.14156 192.168.123.100:0/929894269' entity='mgr.a' cmd=[{"prefix": "osd pool create", "format": "json", "pool": ".mgr", "pg_num": 1, "pg_num_min": 1, "pg_num_max": 32, "yes_i_really_mean_it": true}]: dispatch 2026-03-06T13:42:10.511 INFO:teuthology.orchestra.run.vm00.stdout: 2026-03-06T13:42:10.540 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 06 13:42:10 vm08 ceph-mon[53561]: purged_snaps scrub starts 2026-03-06T13:42:10.540 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 06 13:42:10 vm08 ceph-mon[53561]: purged_snaps scrub ok 2026-03-06T13:42:10.540 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 06 13:42:10 vm08 ceph-mon[53561]: from='mgr.14156 192.168.123.100:0/929894269' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 2}]: dispatch 2026-03-06T13:42:10.540 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 06 13:42:10 vm08 ceph-mon[53561]: osd.2 [v2:192.168.123.108:6800/4086912105,v1:192.168.123.108:6801/4086912105] boot 2026-03-06T13:42:10.540 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 06 13:42:10 vm08 ceph-mon[53561]: osdmap e16: 3 total, 3 up, 3 in 2026-03-06T13:42:10.540 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 06 13:42:10 vm08 ceph-mon[53561]: from='mgr.14156 192.168.123.100:0/929894269' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 2}]: dispatch 2026-03-06T13:42:10.540 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 06 13:42:10 vm08 ceph-mon[53561]: pgmap v47: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-06T13:42:10.540 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 06 13:42:10 vm08 ceph-mon[53561]: from='mgr.14156 192.168.123.100:0/929894269' entity='mgr.a' cmd=[{"prefix": "osd pool create", "format": "json", "pool": ".mgr", "pg_num": 1, "pg_num_min": 1, "pg_num_max": 32, "yes_i_really_mean_it": true}]: dispatch 2026-03-06T13:42:10.599 INFO:teuthology.orchestra.run.vm00.stdout:{"epoch":17,"num_osds":3,"num_up_osds":3,"osd_up_since":1772800929,"num_in_osds":3,"osd_in_since":1772800915,"num_remapped_pgs":0} 2026-03-06T13:42:10.599 DEBUG:teuthology.orchestra.run.vm00:> sudo /home/ubuntu/cephtest/cephadm --image harbor.clyso.com/custom-ceph/ceph/ceph:cobaltcore-storage-v19.2.3-fasttrack-3 shell --fsid 82f504c8-1959-11f1-9a16-a78d135e200b -- ceph osd dump --format=json 2026-03-06T13:42:10.942 INFO:teuthology.orchestra.run.vm00.stderr:Inferring config /var/lib/ceph/82f504c8-1959-11f1-9a16-a78d135e200b/mon.a/config 2026-03-06T13:42:11.246 INFO:journalctl@ceph.osd.0.vm00.stdout:Mar 06 13:42:11 vm00 sudo[65477]: ceph : PWD=/ ; USER=root ; COMMAND=/usr/sbin/smartctl -x --json=o /dev/vde 2026-03-06T13:42:11.246 INFO:journalctl@ceph.osd.0.vm00.stdout:Mar 06 13:42:11 vm00 sudo[65477]: pam_systemd(sudo:session): Failed to connect to system bus: No such file or directory 2026-03-06T13:42:11.246 INFO:journalctl@ceph.osd.0.vm00.stdout:Mar 06 13:42:11 vm00 sudo[65477]: pam_unix(sudo:session): session opened for user root(uid=0) by (uid=167) 2026-03-06T13:42:11.246 INFO:journalctl@ceph.osd.0.vm00.stdout:Mar 06 13:42:11 vm00 sudo[65477]: pam_unix(sudo:session): session closed for user root 2026-03-06T13:42:11.247 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:42:11 vm00 ceph-mon[51914]: from='mgr.14156 192.168.123.100:0/929894269' entity='mgr.a' cmd='[{"prefix": "osd pool create", "format": "json", "pool": ".mgr", "pg_num": 1, "pg_num_min": 1, "pg_num_max": 32, "yes_i_really_mean_it": true}]': finished 2026-03-06T13:42:11.247 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:42:11 vm00 ceph-mon[51914]: osdmap e17: 3 total, 3 up, 3 in 2026-03-06T13:42:11.247 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:42:11 vm00 ceph-mon[51914]: from='mgr.14156 192.168.123.100:0/929894269' entity='mgr.a' cmd=[{"prefix": "osd pool application enable", "format": "json", "pool": ".mgr", "app": "mgr", "yes_i_really_mean_it": true}]: dispatch 2026-03-06T13:42:11.247 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:42:11 vm00 ceph-mon[51914]: from='client.? 192.168.123.100:0/3968017203' entity='client.admin' cmd=[{"prefix": "osd stat", "format": "json"}]: dispatch 2026-03-06T13:42:11.247 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:42:11 vm00 sudo[65499]: ceph : PWD=/ ; USER=root ; COMMAND=/usr/sbin/smartctl -x --json=o /dev/vda 2026-03-06T13:42:11.247 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:42:11 vm00 sudo[65499]: pam_systemd(sudo:session): Failed to connect to system bus: No such file or directory 2026-03-06T13:42:11.247 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:42:11 vm00 sudo[65499]: pam_unix(sudo:session): session opened for user root(uid=0) by (uid=167) 2026-03-06T13:42:11.324 INFO:teuthology.orchestra.run.vm00.stdout: 2026-03-06T13:42:11.324 INFO:teuthology.orchestra.run.vm00.stdout:{"epoch":18,"fsid":"82f504c8-1959-11f1-9a16-a78d135e200b","created":"2026-03-06T12:40:03.178129+0000","modified":"2026-03-06T12:42:11.102522+0000","last_up_change":"2026-03-06T12:42:09.091745+0000","last_in_change":"2026-03-06T12:41:55.730672+0000","flags":"sortbitwise,recovery_deletes,purged_snapdirs,pglog_hardlimit","flags_num":5799936,"flags_set":["pglog_hardlimit","purged_snapdirs","recovery_deletes","sortbitwise"],"crush_version":8,"full_ratio":0.94999998807907104,"backfillfull_ratio":0.89999997615814209,"nearfull_ratio":0.85000002384185791,"cluster_snapshot":"","pool_max":1,"max_osd":3,"require_min_compat_client":"luminous","min_compat_client":"jewel","require_osd_release":"squid","allow_crimson":false,"pools":[{"pool":1,"pool_name":".mgr","create_time":"2026-03-06T12:42:10.032248+0000","flags":32769,"flags_names":"hashpspool,creating","type":1,"size":3,"min_size":2,"crush_rule":0,"peering_crush_bucket_count":0,"peering_crush_bucket_target":0,"peering_crush_bucket_barrier":0,"peering_crush_bucket_mandatory_member":2147483647,"is_stretch_pool":false,"object_hash":2,"pg_autoscale_mode":"off","pg_num":1,"pg_placement_num":1,"pg_placement_num_target":1,"pg_num_target":1,"pg_num_pending":1,"last_pg_merge_meta":{"source_pgid":"0.0","ready_epoch":0,"last_epoch_started":0,"last_epoch_clean":0,"source_version":"0'0","target_version":"0'0"},"last_change":"18","last_force_op_resend":"0","last_force_op_resend_prenautilus":"0","last_force_op_resend_preluminous":"0","auid":0,"snap_mode":"selfmanaged","snap_seq":0,"snap_epoch":0,"pool_snaps":[],"removed_snaps":"[]","quota_max_bytes":0,"quota_max_objects":0,"tiers":[],"tier_of":-1,"read_tier":-1,"write_tier":-1,"cache_mode":"none","target_max_bytes":0,"target_max_objects":0,"cache_target_dirty_ratio_micro":400000,"cache_target_dirty_high_ratio_micro":600000,"cache_target_full_ratio_micro":800000,"cache_min_flush_age":0,"cache_min_evict_age":0,"erasure_code_profile":"","hit_set_params":{"type":"none"},"hit_set_period":0,"hit_set_count":0,"use_gmt_hitset":true,"min_read_recency_for_promote":0,"min_write_recency_for_promote":0,"hit_set_grade_decay_rate":0,"hit_set_search_last_n":0,"grade_table":[],"stripe_width":0,"expected_num_objects":0,"fast_read":false,"options":{"pg_num_max":32,"pg_num_min":1},"application_metadata":{"mgr":{}},"read_balance":{"score_type":"Fair distribution","score_acting":3,"score_stable":3,"optimal_score":1,"raw_score_acting":3,"raw_score_stable":3,"primary_affinity_weighted":1,"average_primary_affinity":1,"average_primary_affinity_weighted":1}}],"osds":[{"osd":0,"uuid":"1d65b890-7ef0-46fe-a6c2-e544a1692d70","up":1,"in":1,"weight":1,"primary_affinity":1,"last_clean_begin":0,"last_clean_end":0,"up_from":8,"up_thru":0,"down_at":0,"lost_at":0,"public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.100:6802","nonce":2576257746},{"type":"v1","addr":"192.168.123.100:6803","nonce":2576257746}]},"cluster_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.100:6804","nonce":2576257746},{"type":"v1","addr":"192.168.123.100:6805","nonce":2576257746}]},"heartbeat_back_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.100:6808","nonce":2576257746},{"type":"v1","addr":"192.168.123.100:6809","nonce":2576257746}]},"heartbeat_front_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.100:6806","nonce":2576257746},{"type":"v1","addr":"192.168.123.100:6807","nonce":2576257746}]},"public_addr":"192.168.123.100:6803/2576257746","cluster_addr":"192.168.123.100:6805/2576257746","heartbeat_back_addr":"192.168.123.100:6809/2576257746","heartbeat_front_addr":"192.168.123.100:6807/2576257746","state":["exists","up"]},{"osd":1,"uuid":"95d964b0-4ea2-45f5-a612-b4aa5749cec2","up":1,"in":1,"weight":1,"primary_affinity":1,"last_clean_begin":0,"last_clean_end":0,"up_from":12,"up_thru":17,"down_at":0,"lost_at":0,"public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.101:6800","nonce":1940387388},{"type":"v1","addr":"192.168.123.101:6801","nonce":1940387388}]},"cluster_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.101:6802","nonce":1940387388},{"type":"v1","addr":"192.168.123.101:6803","nonce":1940387388}]},"heartbeat_back_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.101:6806","nonce":1940387388},{"type":"v1","addr":"192.168.123.101:6807","nonce":1940387388}]},"heartbeat_front_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.101:6804","nonce":1940387388},{"type":"v1","addr":"192.168.123.101:6805","nonce":1940387388}]},"public_addr":"192.168.123.101:6801/1940387388","cluster_addr":"192.168.123.101:6803/1940387388","heartbeat_back_addr":"192.168.123.101:6807/1940387388","heartbeat_front_addr":"192.168.123.101:6805/1940387388","state":["exists","up"]},{"osd":2,"uuid":"9e99c305-cccd-4040-a383-099c1410ee92","up":1,"in":1,"weight":1,"primary_affinity":1,"last_clean_begin":0,"last_clean_end":0,"up_from":16,"up_thru":0,"down_at":0,"lost_at":0,"public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.108:6800","nonce":4086912105},{"type":"v1","addr":"192.168.123.108:6801","nonce":4086912105}]},"cluster_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.108:6802","nonce":4086912105},{"type":"v1","addr":"192.168.123.108:6803","nonce":4086912105}]},"heartbeat_back_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.108:6806","nonce":4086912105},{"type":"v1","addr":"192.168.123.108:6807","nonce":4086912105}]},"heartbeat_front_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.108:6804","nonce":4086912105},{"type":"v1","addr":"192.168.123.108:6805","nonce":4086912105}]},"public_addr":"192.168.123.108:6801/4086912105","cluster_addr":"192.168.123.108:6803/4086912105","heartbeat_back_addr":"192.168.123.108:6807/4086912105","heartbeat_front_addr":"192.168.123.108:6805/4086912105","state":["exists","up"]}],"osd_xinfo":[{"osd":0,"down_stamp":"0.000000","laggy_probability":0,"laggy_interval":0,"features":4540701547738038271,"old_weight":0,"last_purged_snaps_scrub":"2026-03-06T12:41:39.587816+0000","dead_epoch":0},{"osd":1,"down_stamp":"0.000000","laggy_probability":0,"laggy_interval":0,"features":4540701547738038271,"old_weight":0,"last_purged_snaps_scrub":"2026-03-06T12:41:52.938720+0000","dead_epoch":0},{"osd":2,"down_stamp":"0.000000","laggy_probability":0,"laggy_interval":0,"features":4540701547738038271,"old_weight":0,"last_purged_snaps_scrub":"2026-03-06T12:42:07.364233+0000","dead_epoch":0}],"pg_upmap":[],"pg_upmap_items":[],"pg_upmap_primaries":[],"pg_temp":[],"primary_temp":[],"blocklist":{"192.168.123.100:6801/2414933134":"2026-03-07T12:40:47.966309+0000","192.168.123.100:0/310500200":"2026-03-07T12:40:47.966309+0000","192.168.123.100:0/2258966804":"2026-03-07T12:40:47.966309+0000","192.168.123.100:0/1720822170":"2026-03-07T12:40:47.966309+0000","192.168.123.100:0/3743744839":"2026-03-07T12:40:28.137920+0000","192.168.123.100:0/3253599029":"2026-03-07T12:40:28.137920+0000","192.168.123.100:6801/3456934444":"2026-03-07T12:40:28.137920+0000","192.168.123.100:6800/2414933134":"2026-03-07T12:40:47.966309+0000","192.168.123.100:6800/3456934444":"2026-03-07T12:40:28.137920+0000","192.168.123.100:0/33831712":"2026-03-07T12:40:28.137920+0000"},"range_blocklist":{},"erasure_code_profiles":{"default":{"crush-failure-domain":"osd","k":"2","m":"1","plugin":"jerasure","technique":"reed_sol_van"}},"removed_snaps_queue":[],"new_removed_snaps":[],"new_purged_snaps":[],"crush_node_flags":{},"device_class_flags":{},"stretch_mode":{"stretch_mode_enabled":false,"stretch_bucket_count":0,"degraded_stretch_mode":0,"recovering_stretch_mode":0,"stretch_mode_bucket":0}} 2026-03-06T13:42:11.390 INFO:tasks.cephadm.ceph_manager.ceph:[{'pool': 1, 'pool_name': '.mgr', 'create_time': '2026-03-06T12:42:10.032248+0000', 'flags': 32769, 'flags_names': 'hashpspool,creating', 'type': 1, 'size': 3, 'min_size': 2, 'crush_rule': 0, 'peering_crush_bucket_count': 0, 'peering_crush_bucket_target': 0, 'peering_crush_bucket_barrier': 0, 'peering_crush_bucket_mandatory_member': 2147483647, 'is_stretch_pool': False, 'object_hash': 2, 'pg_autoscale_mode': 'off', 'pg_num': 1, 'pg_placement_num': 1, 'pg_placement_num_target': 1, 'pg_num_target': 1, 'pg_num_pending': 1, 'last_pg_merge_meta': {'source_pgid': '0.0', 'ready_epoch': 0, 'last_epoch_started': 0, 'last_epoch_clean': 0, 'source_version': "0'0", 'target_version': "0'0"}, 'last_change': '18', 'last_force_op_resend': '0', 'last_force_op_resend_prenautilus': '0', 'last_force_op_resend_preluminous': '0', 'auid': 0, 'snap_mode': 'selfmanaged', 'snap_seq': 0, 'snap_epoch': 0, 'pool_snaps': [], 'removed_snaps': '[]', 'quota_max_bytes': 0, 'quota_max_objects': 0, 'tiers': [], 'tier_of': -1, 'read_tier': -1, 'write_tier': -1, 'cache_mode': 'none', 'target_max_bytes': 0, 'target_max_objects': 0, 'cache_target_dirty_ratio_micro': 400000, 'cache_target_dirty_high_ratio_micro': 600000, 'cache_target_full_ratio_micro': 800000, 'cache_min_flush_age': 0, 'cache_min_evict_age': 0, 'erasure_code_profile': '', 'hit_set_params': {'type': 'none'}, 'hit_set_period': 0, 'hit_set_count': 0, 'use_gmt_hitset': True, 'min_read_recency_for_promote': 0, 'min_write_recency_for_promote': 0, 'hit_set_grade_decay_rate': 0, 'hit_set_search_last_n': 0, 'grade_table': [], 'stripe_width': 0, 'expected_num_objects': 0, 'fast_read': False, 'options': {'pg_num_max': 32, 'pg_num_min': 1}, 'application_metadata': {'mgr': {}}, 'read_balance': {'score_type': 'Fair distribution', 'score_acting': 3, 'score_stable': 3, 'optimal_score': 1, 'raw_score_acting': 3, 'raw_score_stable': 3, 'primary_affinity_weighted': 1, 'average_primary_affinity': 1, 'average_primary_affinity_weighted': 1}}] 2026-03-06T13:42:11.391 DEBUG:teuthology.orchestra.run.vm00:> sudo /home/ubuntu/cephtest/cephadm --image harbor.clyso.com/custom-ceph/ceph/ceph:cobaltcore-storage-v19.2.3-fasttrack-3 shell --fsid 82f504c8-1959-11f1-9a16-a78d135e200b -- ceph osd pool get .mgr pg_num 2026-03-06T13:42:11.405 INFO:journalctl@ceph.osd.1.vm01.stdout:Mar 06 13:42:11 vm01 sudo[63429]: ceph : PWD=/ ; USER=root ; COMMAND=/usr/sbin/smartctl -x --json=o /dev/vde 2026-03-06T13:42:11.405 INFO:journalctl@ceph.osd.1.vm01.stdout:Mar 06 13:42:11 vm01 sudo[63429]: pam_systemd(sudo:session): Failed to connect to system bus: No such file or directory 2026-03-06T13:42:11.405 INFO:journalctl@ceph.osd.1.vm01.stdout:Mar 06 13:42:11 vm01 sudo[63429]: pam_unix(sudo:session): session opened for user root(uid=0) by (uid=167) 2026-03-06T13:42:11.405 INFO:journalctl@ceph.osd.1.vm01.stdout:Mar 06 13:42:11 vm01 sudo[63429]: pam_unix(sudo:session): session closed for user root 2026-03-06T13:42:11.405 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:42:11 vm01 ceph-mon[56331]: from='mgr.14156 192.168.123.100:0/929894269' entity='mgr.a' cmd='[{"prefix": "osd pool create", "format": "json", "pool": ".mgr", "pg_num": 1, "pg_num_min": 1, "pg_num_max": 32, "yes_i_really_mean_it": true}]': finished 2026-03-06T13:42:11.405 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:42:11 vm01 ceph-mon[56331]: osdmap e17: 3 total, 3 up, 3 in 2026-03-06T13:42:11.405 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:42:11 vm01 ceph-mon[56331]: from='mgr.14156 192.168.123.100:0/929894269' entity='mgr.a' cmd=[{"prefix": "osd pool application enable", "format": "json", "pool": ".mgr", "app": "mgr", "yes_i_really_mean_it": true}]: dispatch 2026-03-06T13:42:11.405 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:42:11 vm01 ceph-mon[56331]: from='client.? 192.168.123.100:0/3968017203' entity='client.admin' cmd=[{"prefix": "osd stat", "format": "json"}]: dispatch 2026-03-06T13:42:11.405 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:42:11 vm01 sudo[63432]: ceph : PWD=/ ; USER=root ; COMMAND=/usr/sbin/smartctl -x --json=o /dev/vda 2026-03-06T13:42:11.405 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:42:11 vm01 sudo[63432]: pam_systemd(sudo:session): Failed to connect to system bus: No such file or directory 2026-03-06T13:42:11.405 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:42:11 vm01 sudo[63432]: pam_unix(sudo:session): session opened for user root(uid=0) by (uid=167) 2026-03-06T13:42:11.405 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:42:11 vm01 sudo[63432]: pam_unix(sudo:session): session closed for user root 2026-03-06T13:42:11.502 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:42:11 vm00 sudo[65499]: pam_unix(sudo:session): session closed for user root 2026-03-06T13:42:11.539 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 06 13:42:11 vm08 ceph-mon[53561]: from='mgr.14156 192.168.123.100:0/929894269' entity='mgr.a' cmd='[{"prefix": "osd pool create", "format": "json", "pool": ".mgr", "pg_num": 1, "pg_num_min": 1, "pg_num_max": 32, "yes_i_really_mean_it": true}]': finished 2026-03-06T13:42:11.540 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 06 13:42:11 vm08 ceph-mon[53561]: osdmap e17: 3 total, 3 up, 3 in 2026-03-06T13:42:11.540 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 06 13:42:11 vm08 ceph-mon[53561]: from='mgr.14156 192.168.123.100:0/929894269' entity='mgr.a' cmd=[{"prefix": "osd pool application enable", "format": "json", "pool": ".mgr", "app": "mgr", "yes_i_really_mean_it": true}]: dispatch 2026-03-06T13:42:11.540 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 06 13:42:11 vm08 ceph-mon[53561]: from='client.? 192.168.123.100:0/3968017203' entity='client.admin' cmd=[{"prefix": "osd stat", "format": "json"}]: dispatch 2026-03-06T13:42:11.540 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 06 13:42:11 vm08 sudo[60448]: ceph : PWD=/ ; USER=root ; COMMAND=/usr/sbin/smartctl -x --json=o /dev/vda 2026-03-06T13:42:11.540 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 06 13:42:11 vm08 sudo[60448]: pam_systemd(sudo:session): Failed to connect to system bus: No such file or directory 2026-03-06T13:42:11.540 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 06 13:42:11 vm08 sudo[60448]: pam_unix(sudo:session): session opened for user root(uid=0) by (uid=167) 2026-03-06T13:42:11.540 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 06 13:42:11 vm08 sudo[60448]: pam_unix(sudo:session): session closed for user root 2026-03-06T13:42:11.540 INFO:journalctl@ceph.osd.2.vm08.stdout:Mar 06 13:42:11 vm08 sudo[60445]: ceph : PWD=/ ; USER=root ; COMMAND=/usr/sbin/smartctl -x --json=o /dev/vde 2026-03-06T13:42:11.540 INFO:journalctl@ceph.osd.2.vm08.stdout:Mar 06 13:42:11 vm08 sudo[60445]: pam_systemd(sudo:session): Failed to connect to system bus: No such file or directory 2026-03-06T13:42:11.540 INFO:journalctl@ceph.osd.2.vm08.stdout:Mar 06 13:42:11 vm08 sudo[60445]: pam_unix(sudo:session): session opened for user root(uid=0) by (uid=167) 2026-03-06T13:42:11.540 INFO:journalctl@ceph.osd.2.vm08.stdout:Mar 06 13:42:11 vm08 sudo[60445]: pam_unix(sudo:session): session closed for user root 2026-03-06T13:42:11.779 INFO:teuthology.orchestra.run.vm00.stderr:Inferring config /var/lib/ceph/82f504c8-1959-11f1-9a16-a78d135e200b/mon.a/config 2026-03-06T13:42:12.139 INFO:teuthology.orchestra.run.vm00.stdout:pg_num: 1 2026-03-06T13:42:12.198 INFO:tasks.cephadm:Setting up client nodes... 2026-03-06T13:42:12.198 INFO:tasks.ceph:Waiting until ceph daemons up and pgs clean... 2026-03-06T13:42:12.198 INFO:tasks.cephadm.ceph_manager.ceph:waiting for mgr available 2026-03-06T13:42:12.198 DEBUG:teuthology.orchestra.run.vm00:> sudo /home/ubuntu/cephtest/cephadm --image harbor.clyso.com/custom-ceph/ceph/ceph:cobaltcore-storage-v19.2.3-fasttrack-3 shell --fsid 82f504c8-1959-11f1-9a16-a78d135e200b -- ceph mgr dump --format=json 2026-03-06T13:42:12.405 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:42:12 vm01 ceph-mon[56331]: from='mgr.14156 192.168.123.100:0/929894269' entity='mgr.a' cmd='[{"prefix": "osd pool application enable", "format": "json", "pool": ".mgr", "app": "mgr", "yes_i_really_mean_it": true}]': finished 2026-03-06T13:42:12.405 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:42:12 vm01 ceph-mon[56331]: osdmap e18: 3 total, 3 up, 3 in 2026-03-06T13:42:12.405 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:42:12 vm01 ceph-mon[56331]: from='admin socket' entity='admin socket' cmd='smart' args=[json]: dispatch 2026-03-06T13:42:12.405 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:42:12 vm01 ceph-mon[56331]: from='admin socket' entity='admin socket' cmd=smart args=[json]: finished 2026-03-06T13:42:12.405 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:42:12 vm01 ceph-mon[56331]: from='mgr.14156 192.168.123.100:0/929894269' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "a"}]: dispatch 2026-03-06T13:42:12.405 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:42:12 vm01 ceph-mon[56331]: from='mgr.14156 192.168.123.100:0/929894269' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-06T13:42:12.405 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:42:12 vm01 ceph-mon[56331]: from='mgr.14156 192.168.123.100:0/929894269' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "c"}]: dispatch 2026-03-06T13:42:12.405 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:42:12 vm01 ceph-mon[56331]: from='admin socket' entity='admin socket' cmd='smart' args=[json]: dispatch 2026-03-06T13:42:12.405 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:42:12 vm01 ceph-mon[56331]: from='mgr.14156 192.168.123.100:0/929894269' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "a"}]: dispatch 2026-03-06T13:42:12.405 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:42:12 vm01 ceph-mon[56331]: from='mgr.14156 192.168.123.100:0/929894269' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-06T13:42:12.405 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:42:12 vm01 ceph-mon[56331]: from='mgr.14156 192.168.123.100:0/929894269' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "c"}]: dispatch 2026-03-06T13:42:12.405 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:42:12 vm01 ceph-mon[56331]: from='admin socket' entity='admin socket' cmd=smart args=[json]: finished 2026-03-06T13:42:12.406 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:42:12 vm01 ceph-mon[56331]: from='admin socket' entity='admin socket' cmd='smart' args=[json]: dispatch 2026-03-06T13:42:12.406 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:42:12 vm01 ceph-mon[56331]: from='mgr.14156 192.168.123.100:0/929894269' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "a"}]: dispatch 2026-03-06T13:42:12.406 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:42:12 vm01 ceph-mon[56331]: from='mgr.14156 192.168.123.100:0/929894269' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-06T13:42:12.406 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:42:12 vm01 ceph-mon[56331]: from='mgr.14156 192.168.123.100:0/929894269' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "c"}]: dispatch 2026-03-06T13:42:12.406 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:42:12 vm01 ceph-mon[56331]: from='admin socket' entity='admin socket' cmd=smart args=[json]: finished 2026-03-06T13:42:12.406 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:42:12 vm01 ceph-mon[56331]: from='client.? 192.168.123.100:0/1149976429' entity='client.admin' cmd=[{"prefix": "osd dump", "format": "json"}]: dispatch 2026-03-06T13:42:12.406 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:42:12 vm01 ceph-mon[56331]: pgmap v50: 1 pgs: 1 creating+peering; 0 B data, 79 MiB used, 60 GiB / 60 GiB avail 2026-03-06T13:42:12.503 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:42:12 vm00 ceph-mon[51914]: from='mgr.14156 192.168.123.100:0/929894269' entity='mgr.a' cmd='[{"prefix": "osd pool application enable", "format": "json", "pool": ".mgr", "app": "mgr", "yes_i_really_mean_it": true}]': finished 2026-03-06T13:42:12.503 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:42:12 vm00 ceph-mon[51914]: osdmap e18: 3 total, 3 up, 3 in 2026-03-06T13:42:12.503 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:42:12 vm00 ceph-mon[51914]: from='admin socket' entity='admin socket' cmd='smart' args=[json]: dispatch 2026-03-06T13:42:12.503 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:42:12 vm00 ceph-mon[51914]: from='admin socket' entity='admin socket' cmd=smart args=[json]: finished 2026-03-06T13:42:12.503 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:42:12 vm00 ceph-mon[51914]: from='mgr.14156 192.168.123.100:0/929894269' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "a"}]: dispatch 2026-03-06T13:42:12.503 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:42:12 vm00 ceph-mon[51914]: from='mgr.14156 192.168.123.100:0/929894269' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-06T13:42:12.503 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:42:12 vm00 ceph-mon[51914]: from='mgr.14156 192.168.123.100:0/929894269' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "c"}]: dispatch 2026-03-06T13:42:12.503 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:42:12 vm00 ceph-mon[51914]: from='admin socket' entity='admin socket' cmd='smart' args=[json]: dispatch 2026-03-06T13:42:12.503 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:42:12 vm00 ceph-mon[51914]: from='mgr.14156 192.168.123.100:0/929894269' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "a"}]: dispatch 2026-03-06T13:42:12.503 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:42:12 vm00 ceph-mon[51914]: from='mgr.14156 192.168.123.100:0/929894269' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-06T13:42:12.503 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:42:12 vm00 ceph-mon[51914]: from='mgr.14156 192.168.123.100:0/929894269' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "c"}]: dispatch 2026-03-06T13:42:12.503 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:42:12 vm00 ceph-mon[51914]: from='admin socket' entity='admin socket' cmd=smart args=[json]: finished 2026-03-06T13:42:12.503 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:42:12 vm00 ceph-mon[51914]: from='admin socket' entity='admin socket' cmd='smart' args=[json]: dispatch 2026-03-06T13:42:12.503 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:42:12 vm00 ceph-mon[51914]: from='mgr.14156 192.168.123.100:0/929894269' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "a"}]: dispatch 2026-03-06T13:42:12.503 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:42:12 vm00 ceph-mon[51914]: from='mgr.14156 192.168.123.100:0/929894269' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-06T13:42:12.503 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:42:12 vm00 ceph-mon[51914]: from='mgr.14156 192.168.123.100:0/929894269' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "c"}]: dispatch 2026-03-06T13:42:12.503 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:42:12 vm00 ceph-mon[51914]: from='admin socket' entity='admin socket' cmd=smart args=[json]: finished 2026-03-06T13:42:12.503 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:42:12 vm00 ceph-mon[51914]: from='client.? 192.168.123.100:0/1149976429' entity='client.admin' cmd=[{"prefix": "osd dump", "format": "json"}]: dispatch 2026-03-06T13:42:12.503 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:42:12 vm00 ceph-mon[51914]: pgmap v50: 1 pgs: 1 creating+peering; 0 B data, 79 MiB used, 60 GiB / 60 GiB avail 2026-03-06T13:42:12.540 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 06 13:42:12 vm08 ceph-mon[53561]: from='mgr.14156 192.168.123.100:0/929894269' entity='mgr.a' cmd='[{"prefix": "osd pool application enable", "format": "json", "pool": ".mgr", "app": "mgr", "yes_i_really_mean_it": true}]': finished 2026-03-06T13:42:12.540 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 06 13:42:12 vm08 ceph-mon[53561]: osdmap e18: 3 total, 3 up, 3 in 2026-03-06T13:42:12.540 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 06 13:42:12 vm08 ceph-mon[53561]: from='admin socket' entity='admin socket' cmd='smart' args=[json]: dispatch 2026-03-06T13:42:12.540 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 06 13:42:12 vm08 ceph-mon[53561]: from='admin socket' entity='admin socket' cmd=smart args=[json]: finished 2026-03-06T13:42:12.540 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 06 13:42:12 vm08 ceph-mon[53561]: from='mgr.14156 192.168.123.100:0/929894269' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "a"}]: dispatch 2026-03-06T13:42:12.540 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 06 13:42:12 vm08 ceph-mon[53561]: from='mgr.14156 192.168.123.100:0/929894269' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-06T13:42:12.540 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 06 13:42:12 vm08 ceph-mon[53561]: from='mgr.14156 192.168.123.100:0/929894269' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "c"}]: dispatch 2026-03-06T13:42:12.540 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 06 13:42:12 vm08 ceph-mon[53561]: from='admin socket' entity='admin socket' cmd='smart' args=[json]: dispatch 2026-03-06T13:42:12.540 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 06 13:42:12 vm08 ceph-mon[53561]: from='mgr.14156 192.168.123.100:0/929894269' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "a"}]: dispatch 2026-03-06T13:42:12.540 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 06 13:42:12 vm08 ceph-mon[53561]: from='mgr.14156 192.168.123.100:0/929894269' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-06T13:42:12.540 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 06 13:42:12 vm08 ceph-mon[53561]: from='mgr.14156 192.168.123.100:0/929894269' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "c"}]: dispatch 2026-03-06T13:42:12.540 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 06 13:42:12 vm08 ceph-mon[53561]: from='admin socket' entity='admin socket' cmd=smart args=[json]: finished 2026-03-06T13:42:12.540 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 06 13:42:12 vm08 ceph-mon[53561]: from='admin socket' entity='admin socket' cmd='smart' args=[json]: dispatch 2026-03-06T13:42:12.540 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 06 13:42:12 vm08 ceph-mon[53561]: from='mgr.14156 192.168.123.100:0/929894269' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "a"}]: dispatch 2026-03-06T13:42:12.540 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 06 13:42:12 vm08 ceph-mon[53561]: from='mgr.14156 192.168.123.100:0/929894269' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-06T13:42:12.540 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 06 13:42:12 vm08 ceph-mon[53561]: from='mgr.14156 192.168.123.100:0/929894269' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "c"}]: dispatch 2026-03-06T13:42:12.540 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 06 13:42:12 vm08 ceph-mon[53561]: from='admin socket' entity='admin socket' cmd=smart args=[json]: finished 2026-03-06T13:42:12.540 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 06 13:42:12 vm08 ceph-mon[53561]: from='client.? 192.168.123.100:0/1149976429' entity='client.admin' cmd=[{"prefix": "osd dump", "format": "json"}]: dispatch 2026-03-06T13:42:12.540 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 06 13:42:12 vm08 ceph-mon[53561]: pgmap v50: 1 pgs: 1 creating+peering; 0 B data, 79 MiB used, 60 GiB / 60 GiB avail 2026-03-06T13:42:12.557 INFO:teuthology.orchestra.run.vm00.stderr:Inferring config /var/lib/ceph/82f504c8-1959-11f1-9a16-a78d135e200b/mon.a/config 2026-03-06T13:42:12.921 INFO:teuthology.orchestra.run.vm00.stdout: 2026-03-06T13:42:12.979 INFO:teuthology.orchestra.run.vm00.stdout:{"epoch":15,"flags":0,"active_gid":14156,"active_name":"a","active_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.100:6800","nonce":3050447460},{"type":"v1","addr":"192.168.123.100:6801","nonce":3050447460}]},"active_addr":"192.168.123.100:6801/3050447460","active_change":"2026-03-06T12:40:47.966421+0000","active_mgr_features":4540701547738038271,"available":true,"standbys":[{"gid":24104,"name":"b","mgr_features":4540701547738038271,"available_modules":[{"name":"alerts","can_run":true,"error_string":"","module_options":{"interval":{"name":"interval","type":"secs","level":"advanced","flags":1,"default_value":"60","min":"","max":"","enum_allowed":[],"desc":"How frequently to reexamine health status","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"smtp_destination":{"name":"smtp_destination","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"Email address to send alerts to","long_desc":"","tags":[],"see_also":[]},"smtp_from_name":{"name":"smtp_from_name","type":"str","level":"advanced","flags":1,"default_value":"Ceph","min":"","max":"","enum_allowed":[],"desc":"Email From: name","long_desc":"","tags":[],"see_also":[]},"smtp_host":{"name":"smtp_host","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"SMTP server","long_desc":"","tags":[],"see_also":[]},"smtp_password":{"name":"smtp_password","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"Password to authenticate with","long_desc":"","tags":[],"see_also":[]},"smtp_port":{"name":"smtp_port","type":"int","level":"advanced","flags":1,"default_value":"465","min":"","max":"","enum_allowed":[],"desc":"SMTP port","long_desc":"","tags":[],"see_also":[]},"smtp_sender":{"name":"smtp_sender","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"SMTP envelope sender","long_desc":"","tags":[],"see_also":[]},"smtp_ssl":{"name":"smtp_ssl","type":"bool","level":"advanced","flags":1,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"Use SSL to connect to SMTP server","long_desc":"","tags":[],"see_also":[]},"smtp_user":{"name":"smtp_user","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"User to authenticate as","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"balancer","can_run":true,"error_string":"","module_options":{"active":{"name":"active","type":"bool","level":"advanced","flags":1,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"automatically balance PGs across cluster","long_desc":"","tags":[],"see_also":[]},"begin_time":{"name":"begin_time","type":"str","level":"advanced","flags":1,"default_value":"0000","min":"","max":"","enum_allowed":[],"desc":"beginning time of day to automatically balance","long_desc":"This is a time of day in the format HHMM.","tags":[],"see_also":[]},"begin_weekday":{"name":"begin_weekday","type":"uint","level":"advanced","flags":1,"default_value":"0","min":"0","max":"6","enum_allowed":[],"desc":"Restrict automatic balancing to this day of the week or later","long_desc":"0 = Sunday, 1 = Monday, etc.","tags":[],"see_also":[]},"crush_compat_max_iterations":{"name":"crush_compat_max_iterations","type":"uint","level":"advanced","flags":1,"default_value":"25","min":"1","max":"250","enum_allowed":[],"desc":"maximum number of iterations to attempt optimization","long_desc":"","tags":[],"see_also":[]},"crush_compat_metrics":{"name":"crush_compat_metrics","type":"str","level":"advanced","flags":1,"default_value":"pgs,objects,bytes","min":"","max":"","enum_allowed":[],"desc":"metrics with which to calculate OSD utilization","long_desc":"Value is a list of one or more of \"pgs\", \"objects\", or \"bytes\", and indicates which metrics to use to balance utilization.","tags":[],"see_also":[]},"crush_compat_step":{"name":"crush_compat_step","type":"float","level":"advanced","flags":1,"default_value":"0.5","min":"0.001","max":"0.999","enum_allowed":[],"desc":"aggressiveness of optimization","long_desc":".99 is very aggressive, .01 is less aggressive","tags":[],"see_also":[]},"end_time":{"name":"end_time","type":"str","level":"advanced","flags":1,"default_value":"2359","min":"","max":"","enum_allowed":[],"desc":"ending time of day to automatically balance","long_desc":"This is a time of day in the format HHMM.","tags":[],"see_also":[]},"end_weekday":{"name":"end_weekday","type":"uint","level":"advanced","flags":1,"default_value":"0","min":"0","max":"6","enum_allowed":[],"desc":"Restrict automatic balancing to days of the week earlier than this","long_desc":"0 = Sunday, 1 = Monday, etc.","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"min_score":{"name":"min_score","type":"float","level":"advanced","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"minimum score, below which no optimization is attempted","long_desc":"","tags":[],"see_also":[]},"mode":{"name":"mode","type":"str","level":"advanced","flags":1,"default_value":"upmap","min":"","max":"","enum_allowed":["crush-compat","none","read","upmap","upmap-read"],"desc":"Balancer mode","long_desc":"","tags":[],"see_also":[]},"pool_ids":{"name":"pool_ids","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"pools which the automatic balancing will be limited to","long_desc":"","tags":[],"see_also":[]},"sleep_interval":{"name":"sleep_interval","type":"secs","level":"advanced","flags":1,"default_value":"60","min":"","max":"","enum_allowed":[],"desc":"how frequently to wake up and attempt optimization","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"update_pg_upmap_activity":{"name":"update_pg_upmap_activity","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Updates pg_upmap activity stats to be used in `balancer status detail`","long_desc":"","tags":[],"see_also":[]},"upmap_max_deviation":{"name":"upmap_max_deviation","type":"int","level":"advanced","flags":1,"default_value":"5","min":"1","max":"","enum_allowed":[],"desc":"deviation below which no optimization is attempted","long_desc":"If the number of PGs are within this count then no optimization is attempted","tags":[],"see_also":[]},"upmap_max_optimizations":{"name":"upmap_max_optimizations","type":"uint","level":"advanced","flags":1,"default_value":"10","min":"","max":"","enum_allowed":[],"desc":"maximum upmap optimizations to make per attempt","long_desc":"","tags":[],"see_also":[]}}},{"name":"cephadm","can_run":true,"error_string":"","module_options":{"agent_down_multiplier":{"name":"agent_down_multiplier","type":"float","level":"advanced","flags":0,"default_value":"3.0","min":"","max":"","enum_allowed":[],"desc":"Multiplied by agent refresh rate to calculate how long agent must not report before being marked down","long_desc":"","tags":[],"see_also":[]},"agent_refresh_rate":{"name":"agent_refresh_rate","type":"secs","level":"advanced","flags":0,"default_value":"20","min":"","max":"","enum_allowed":[],"desc":"How often agent on each host will try to gather and send metadata","long_desc":"","tags":[],"see_also":[]},"agent_starting_port":{"name":"agent_starting_port","type":"int","level":"advanced","flags":0,"default_value":"4721","min":"","max":"","enum_allowed":[],"desc":"First port agent will try to bind to (will also try up to next 1000 subsequent ports if blocked)","long_desc":"","tags":[],"see_also":[]},"allow_ptrace":{"name":"allow_ptrace","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"allow SYS_PTRACE capability on ceph containers","long_desc":"The SYS_PTRACE capability is needed to attach to a process with gdb or strace. Enabling this options can allow debugging daemons that encounter problems at runtime.","tags":[],"see_also":[]},"autotune_interval":{"name":"autotune_interval","type":"secs","level":"advanced","flags":0,"default_value":"600","min":"","max":"","enum_allowed":[],"desc":"how frequently to autotune daemon memory","long_desc":"","tags":[],"see_also":[]},"autotune_memory_target_ratio":{"name":"autotune_memory_target_ratio","type":"float","level":"advanced","flags":0,"default_value":"0.7","min":"","max":"","enum_allowed":[],"desc":"ratio of total system memory to divide amongst autotuned daemons","long_desc":"","tags":[],"see_also":[]},"cephadm_log_destination":{"name":"cephadm_log_destination","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":["file","file,syslog","syslog"],"desc":"Destination for cephadm command's persistent logging","long_desc":"","tags":[],"see_also":[]},"cgroups_split":{"name":"cgroups_split","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"Pass --cgroups=split when cephadm creates containers (currently podman only)","long_desc":"","tags":[],"see_also":[]},"config_checks_enabled":{"name":"config_checks_enabled","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Enable or disable the cephadm configuration analysis","long_desc":"","tags":[],"see_also":[]},"config_dashboard":{"name":"config_dashboard","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"manage configs like API endpoints in Dashboard.","long_desc":"","tags":[],"see_also":[]},"container_image_alertmanager":{"name":"container_image_alertmanager","type":"str","level":"advanced","flags":0,"default_value":"quay.io/prometheus/alertmanager:v0.25.0","min":"","max":"","enum_allowed":[],"desc":"Prometheus container image","long_desc":"","tags":[],"see_also":[]},"container_image_base":{"name":"container_image_base","type":"str","level":"advanced","flags":1,"default_value":"quay.io/ceph/ceph","min":"","max":"","enum_allowed":[],"desc":"Container image name, without the tag","long_desc":"","tags":[],"see_also":[]},"container_image_elasticsearch":{"name":"container_image_elasticsearch","type":"str","level":"advanced","flags":0,"default_value":"quay.io/omrizeneva/elasticsearch:6.8.23","min":"","max":"","enum_allowed":[],"desc":"elasticsearch container image","long_desc":"","tags":[],"see_also":[]},"container_image_grafana":{"name":"container_image_grafana","type":"str","level":"advanced","flags":0,"default_value":"quay.io/ceph/grafana:10.4.0","min":"","max":"","enum_allowed":[],"desc":"Prometheus container image","long_desc":"","tags":[],"see_also":[]},"container_image_haproxy":{"name":"container_image_haproxy","type":"str","level":"advanced","flags":0,"default_value":"quay.io/ceph/haproxy:2.3","min":"","max":"","enum_allowed":[],"desc":"HAproxy container image","long_desc":"","tags":[],"see_also":[]},"container_image_jaeger_agent":{"name":"container_image_jaeger_agent","type":"str","level":"advanced","flags":0,"default_value":"quay.io/jaegertracing/jaeger-agent:1.29","min":"","max":"","enum_allowed":[],"desc":"Jaeger agent container image","long_desc":"","tags":[],"see_also":[]},"container_image_jaeger_collector":{"name":"container_image_jaeger_collector","type":"str","level":"advanced","flags":0,"default_value":"quay.io/jaegertracing/jaeger-collector:1.29","min":"","max":"","enum_allowed":[],"desc":"Jaeger collector container image","long_desc":"","tags":[],"see_also":[]},"container_image_jaeger_query":{"name":"container_image_jaeger_query","type":"str","level":"advanced","flags":0,"default_value":"quay.io/jaegertracing/jaeger-query:1.29","min":"","max":"","enum_allowed":[],"desc":"Jaeger query container image","long_desc":"","tags":[],"see_also":[]},"container_image_keepalived":{"name":"container_image_keepalived","type":"str","level":"advanced","flags":0,"default_value":"quay.io/ceph/keepalived:2.2.4","min":"","max":"","enum_allowed":[],"desc":"Keepalived container image","long_desc":"","tags":[],"see_also":[]},"container_image_loki":{"name":"container_image_loki","type":"str","level":"advanced","flags":0,"default_value":"quay.io/ceph/loki:3.0.0","min":"","max":"","enum_allowed":[],"desc":"Loki container image","long_desc":"","tags":[],"see_also":[]},"container_image_node_exporter":{"name":"container_image_node_exporter","type":"str","level":"advanced","flags":0,"default_value":"quay.io/prometheus/node-exporter:v1.7.0","min":"","max":"","enum_allowed":[],"desc":"Prometheus container image","long_desc":"","tags":[],"see_also":[]},"container_image_nvmeof":{"name":"container_image_nvmeof","type":"str","level":"advanced","flags":0,"default_value":"quay.io/ceph/nvmeof:1.2.5","min":"","max":"","enum_allowed":[],"desc":"Nvme-of container image","long_desc":"","tags":[],"see_also":[]},"container_image_prometheus":{"name":"container_image_prometheus","type":"str","level":"advanced","flags":0,"default_value":"quay.io/prometheus/prometheus:v2.51.0","min":"","max":"","enum_allowed":[],"desc":"Prometheus container image","long_desc":"","tags":[],"see_also":[]},"container_image_promtail":{"name":"container_image_promtail","type":"str","level":"advanced","flags":0,"default_value":"quay.io/ceph/promtail:3.0.0","min":"","max":"","enum_allowed":[],"desc":"Promtail container image","long_desc":"","tags":[],"see_also":[]},"container_image_samba":{"name":"container_image_samba","type":"str","level":"advanced","flags":0,"default_value":"quay.io/samba.org/samba-server:devbuilds-centos-amd64","min":"","max":"","enum_allowed":[],"desc":"Samba/SMB container image","long_desc":"","tags":[],"see_also":[]},"container_image_snmp_gateway":{"name":"container_image_snmp_gateway","type":"str","level":"advanced","flags":0,"default_value":"quay.io/ceph/snmp-notifier:v1.2.1","min":"","max":"","enum_allowed":[],"desc":"SNMP Gateway container image","long_desc":"","tags":[],"see_also":[]},"container_init":{"name":"container_init","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"Run podman/docker with `--init`","long_desc":"","tags":[],"see_also":[]},"daemon_cache_timeout":{"name":"daemon_cache_timeout","type":"secs","level":"advanced","flags":0,"default_value":"600","min":"","max":"","enum_allowed":[],"desc":"seconds to cache service (daemon) inventory","long_desc":"","tags":[],"see_also":[]},"default_cephadm_command_timeout":{"name":"default_cephadm_command_timeout","type":"int","level":"advanced","flags":0,"default_value":"900","min":"","max":"","enum_allowed":[],"desc":"Default timeout applied to cephadm commands run directly on the host (in seconds)","long_desc":"","tags":[],"see_also":[]},"default_registry":{"name":"default_registry","type":"str","level":"advanced","flags":0,"default_value":"quay.io","min":"","max":"","enum_allowed":[],"desc":"Search-registry to which we should normalize unqualified image names. This is not the default registry","long_desc":"","tags":[],"see_also":[]},"device_cache_timeout":{"name":"device_cache_timeout","type":"secs","level":"advanced","flags":0,"default_value":"1800","min":"","max":"","enum_allowed":[],"desc":"seconds to cache device inventory","long_desc":"","tags":[],"see_also":[]},"device_enhanced_scan":{"name":"device_enhanced_scan","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Use libstoragemgmt during device scans","long_desc":"","tags":[],"see_also":[]},"facts_cache_timeout":{"name":"facts_cache_timeout","type":"secs","level":"advanced","flags":0,"default_value":"60","min":"","max":"","enum_allowed":[],"desc":"seconds to cache host facts data","long_desc":"","tags":[],"see_also":[]},"grafana_dashboards_path":{"name":"grafana_dashboards_path","type":"str","level":"advanced","flags":0,"default_value":"/etc/grafana/dashboards/ceph-dashboard/","min":"","max":"","enum_allowed":[],"desc":"location of dashboards to include in grafana deployments","long_desc":"","tags":[],"see_also":[]},"host_check_interval":{"name":"host_check_interval","type":"secs","level":"advanced","flags":0,"default_value":"600","min":"","max":"","enum_allowed":[],"desc":"how frequently to perform a host check","long_desc":"","tags":[],"see_also":[]},"hw_monitoring":{"name":"hw_monitoring","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Deploy hw monitoring daemon on every host.","long_desc":"","tags":[],"see_also":[]},"inventory_list_all":{"name":"inventory_list_all","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Whether ceph-volume inventory should report more devices (mostly mappers (LVs / mpaths), partitions...)","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_refresh_metadata":{"name":"log_refresh_metadata","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Log all refresh metadata. Includes daemon, device, and host info collected regularly. Only has effect if logging at debug level","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"log to the \"cephadm\" cluster log channel\"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"manage_etc_ceph_ceph_conf":{"name":"manage_etc_ceph_ceph_conf","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Manage and own /etc/ceph/ceph.conf on the hosts.","long_desc":"","tags":[],"see_also":[]},"manage_etc_ceph_ceph_conf_hosts":{"name":"manage_etc_ceph_ceph_conf_hosts","type":"str","level":"advanced","flags":0,"default_value":"*","min":"","max":"","enum_allowed":[],"desc":"PlacementSpec describing on which hosts to manage /etc/ceph/ceph.conf","long_desc":"","tags":[],"see_also":[]},"max_count_per_host":{"name":"max_count_per_host","type":"int","level":"advanced","flags":0,"default_value":"10","min":"","max":"","enum_allowed":[],"desc":"max number of daemons per service per host","long_desc":"","tags":[],"see_also":[]},"max_osd_draining_count":{"name":"max_osd_draining_count","type":"int","level":"advanced","flags":0,"default_value":"10","min":"","max":"","enum_allowed":[],"desc":"max number of osds that will be drained simultaneously when osds are removed","long_desc":"","tags":[],"see_also":[]},"migration_current":{"name":"migration_current","type":"int","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"internal - do not modify","long_desc":"","tags":[],"see_also":[]},"mode":{"name":"mode","type":"str","level":"advanced","flags":0,"default_value":"root","min":"","max":"","enum_allowed":["cephadm-package","root"],"desc":"mode for remote execution of cephadm","long_desc":"","tags":[],"see_also":[]},"oob_default_addr":{"name":"oob_default_addr","type":"str","level":"advanced","flags":0,"default_value":"169.254.1.1","min":"","max":"","enum_allowed":[],"desc":"Default address for RedFish API (oob management).","long_desc":"","tags":[],"see_also":[]},"prometheus_alerts_path":{"name":"prometheus_alerts_path","type":"str","level":"advanced","flags":0,"default_value":"/etc/prometheus/ceph/ceph_default_alerts.yml","min":"","max":"","enum_allowed":[],"desc":"location of alerts to include in prometheus deployments","long_desc":"","tags":[],"see_also":[]},"registry_insecure":{"name":"registry_insecure","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Registry is to be considered insecure (no TLS available). Only for development purposes.","long_desc":"","tags":[],"see_also":[]},"registry_password":{"name":"registry_password","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"Custom repository password. Only used for logging into a registry.","long_desc":"","tags":[],"see_also":[]},"registry_url":{"name":"registry_url","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"Registry url for login purposes. This is not the default registry","long_desc":"","tags":[],"see_also":[]},"registry_username":{"name":"registry_username","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"Custom repository username. Only used for logging into a registry.","long_desc":"","tags":[],"see_also":[]},"secure_monitoring_stack":{"name":"secure_monitoring_stack","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Enable TLS security for all the monitoring stack daemons","long_desc":"","tags":[],"see_also":[]},"service_discovery_port":{"name":"service_discovery_port","type":"int","level":"advanced","flags":0,"default_value":"8765","min":"","max":"","enum_allowed":[],"desc":"cephadm service discovery port","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"ssh_config_file":{"name":"ssh_config_file","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"customized SSH config file to connect to managed hosts","long_desc":"","tags":[],"see_also":[]},"ssh_keepalive_count_max":{"name":"ssh_keepalive_count_max","type":"int","level":"advanced","flags":0,"default_value":"3","min":"","max":"","enum_allowed":[],"desc":"How many times ssh connections can fail liveness checks before the host is marked offline","long_desc":"","tags":[],"see_also":[]},"ssh_keepalive_interval":{"name":"ssh_keepalive_interval","type":"int","level":"advanced","flags":0,"default_value":"7","min":"","max":"","enum_allowed":[],"desc":"How often ssh connections are checked for liveness","long_desc":"","tags":[],"see_also":[]},"use_agent":{"name":"use_agent","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Use cephadm agent on each host to gather and send metadata","long_desc":"","tags":[],"see_also":[]},"use_repo_digest":{"name":"use_repo_digest","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"Automatically convert image tags to image digest. Make sure all daemons use the same image","long_desc":"","tags":[],"see_also":[]},"warn_on_failed_host_check":{"name":"warn_on_failed_host_check","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"raise a health warning if the host check fails","long_desc":"","tags":[],"see_also":[]},"warn_on_stray_daemons":{"name":"warn_on_stray_daemons","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"raise a health warning if daemons are detected that are not managed by cephadm","long_desc":"","tags":[],"see_also":[]},"warn_on_stray_hosts":{"name":"warn_on_stray_hosts","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"raise a health warning if daemons are detected on a host that is not managed by cephadm","long_desc":"","tags":[],"see_also":[]}}},{"name":"crash","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"retain_interval":{"name":"retain_interval","type":"secs","level":"advanced","flags":1,"default_value":"31536000","min":"","max":"","enum_allowed":[],"desc":"how long to retain crashes before pruning them","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"warn_recent_interval":{"name":"warn_recent_interval","type":"secs","level":"advanced","flags":1,"default_value":"1209600","min":"","max":"","enum_allowed":[],"desc":"time interval in which to warn about recent crashes","long_desc":"","tags":[],"see_also":[]}}},{"name":"dashboard","can_run":true,"error_string":"","module_options":{"ACCOUNT_LOCKOUT_ATTEMPTS":{"name":"ACCOUNT_LOCKOUT_ATTEMPTS","type":"int","level":"advanced","flags":0,"default_value":"10","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"ALERTMANAGER_API_HOST":{"name":"ALERTMANAGER_API_HOST","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"ALERTMANAGER_API_SSL_VERIFY":{"name":"ALERTMANAGER_API_SSL_VERIFY","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"AUDIT_API_ENABLED":{"name":"AUDIT_API_ENABLED","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"AUDIT_API_LOG_PAYLOAD":{"name":"AUDIT_API_LOG_PAYLOAD","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"ENABLE_BROWSABLE_API":{"name":"ENABLE_BROWSABLE_API","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"FEATURE_TOGGLE_CEPHFS":{"name":"FEATURE_TOGGLE_CEPHFS","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"FEATURE_TOGGLE_DASHBOARD":{"name":"FEATURE_TOGGLE_DASHBOARD","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"FEATURE_TOGGLE_ISCSI":{"name":"FEATURE_TOGGLE_ISCSI","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"FEATURE_TOGGLE_MIRRORING":{"name":"FEATURE_TOGGLE_MIRRORING","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"FEATURE_TOGGLE_NFS":{"name":"FEATURE_TOGGLE_NFS","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"FEATURE_TOGGLE_RBD":{"name":"FEATURE_TOGGLE_RBD","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"FEATURE_TOGGLE_RGW":{"name":"FEATURE_TOGGLE_RGW","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"GANESHA_CLUSTERS_RADOS_POOL_NAMESPACE":{"name":"GANESHA_CLUSTERS_RADOS_POOL_NAMESPACE","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"GRAFANA_API_PASSWORD":{"name":"GRAFANA_API_PASSWORD","type":"str","level":"advanced","flags":0,"default_value":"admin","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"GRAFANA_API_SSL_VERIFY":{"name":"GRAFANA_API_SSL_VERIFY","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"GRAFANA_API_URL":{"name":"GRAFANA_API_URL","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"GRAFANA_API_USERNAME":{"name":"GRAFANA_API_USERNAME","type":"str","level":"advanced","flags":0,"default_value":"admin","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"GRAFANA_FRONTEND_API_URL":{"name":"GRAFANA_FRONTEND_API_URL","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"GRAFANA_UPDATE_DASHBOARDS":{"name":"GRAFANA_UPDATE_DASHBOARDS","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"ISCSI_API_SSL_VERIFICATION":{"name":"ISCSI_API_SSL_VERIFICATION","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"ISSUE_TRACKER_API_KEY":{"name":"ISSUE_TRACKER_API_KEY","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PROMETHEUS_API_HOST":{"name":"PROMETHEUS_API_HOST","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PROMETHEUS_API_SSL_VERIFY":{"name":"PROMETHEUS_API_SSL_VERIFY","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PWD_POLICY_CHECK_COMPLEXITY_ENABLED":{"name":"PWD_POLICY_CHECK_COMPLEXITY_ENABLED","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PWD_POLICY_CHECK_EXCLUSION_LIST_ENABLED":{"name":"PWD_POLICY_CHECK_EXCLUSION_LIST_ENABLED","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PWD_POLICY_CHECK_LENGTH_ENABLED":{"name":"PWD_POLICY_CHECK_LENGTH_ENABLED","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PWD_POLICY_CHECK_OLDPWD_ENABLED":{"name":"PWD_POLICY_CHECK_OLDPWD_ENABLED","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PWD_POLICY_CHECK_REPETITIVE_CHARS_ENABLED":{"name":"PWD_POLICY_CHECK_REPETITIVE_CHARS_ENABLED","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PWD_POLICY_CHECK_SEQUENTIAL_CHARS_ENABLED":{"name":"PWD_POLICY_CHECK_SEQUENTIAL_CHARS_ENABLED","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PWD_POLICY_CHECK_USERNAME_ENABLED":{"name":"PWD_POLICY_CHECK_USERNAME_ENABLED","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PWD_POLICY_ENABLED":{"name":"PWD_POLICY_ENABLED","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PWD_POLICY_EXCLUSION_LIST":{"name":"PWD_POLICY_EXCLUSION_LIST","type":"str","level":"advanced","flags":0,"default_value":"osd,host,dashboard,pool,block,nfs,ceph,monitors,gateway,logs,crush,maps","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PWD_POLICY_MIN_COMPLEXITY":{"name":"PWD_POLICY_MIN_COMPLEXITY","type":"int","level":"advanced","flags":0,"default_value":"10","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PWD_POLICY_MIN_LENGTH":{"name":"PWD_POLICY_MIN_LENGTH","type":"int","level":"advanced","flags":0,"default_value":"8","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"REST_REQUESTS_TIMEOUT":{"name":"REST_REQUESTS_TIMEOUT","type":"int","level":"advanced","flags":0,"default_value":"45","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"RGW_API_ACCESS_KEY":{"name":"RGW_API_ACCESS_KEY","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"RGW_API_ADMIN_RESOURCE":{"name":"RGW_API_ADMIN_RESOURCE","type":"str","level":"advanced","flags":0,"default_value":"admin","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"RGW_API_SECRET_KEY":{"name":"RGW_API_SECRET_KEY","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"RGW_API_SSL_VERIFY":{"name":"RGW_API_SSL_VERIFY","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"USER_PWD_EXPIRATION_SPAN":{"name":"USER_PWD_EXPIRATION_SPAN","type":"int","level":"advanced","flags":0,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"USER_PWD_EXPIRATION_WARNING_1":{"name":"USER_PWD_EXPIRATION_WARNING_1","type":"int","level":"advanced","flags":0,"default_value":"10","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"USER_PWD_EXPIRATION_WARNING_2":{"name":"USER_PWD_EXPIRATION_WARNING_2","type":"int","level":"advanced","flags":0,"default_value":"5","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"cross_origin_url":{"name":"cross_origin_url","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"crt_file":{"name":"crt_file","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"debug":{"name":"debug","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Enable/disable debug options","long_desc":"","tags":[],"see_also":[]},"jwt_token_ttl":{"name":"jwt_token_ttl","type":"int","level":"advanced","flags":0,"default_value":"28800","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"key_file":{"name":"key_file","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"motd":{"name":"motd","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"The message of the day","long_desc":"","tags":[],"see_also":[]},"redirect_resolve_ip_addr":{"name":"redirect_resolve_ip_addr","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"server_addr":{"name":"server_addr","type":"str","level":"advanced","flags":0,"default_value":"::","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"server_port":{"name":"server_port","type":"int","level":"advanced","flags":0,"default_value":"8080","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"ssl":{"name":"ssl","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"ssl_server_port":{"name":"ssl_server_port","type":"int","level":"advanced","flags":0,"default_value":"8443","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"standby_behaviour":{"name":"standby_behaviour","type":"str","level":"advanced","flags":0,"default_value":"redirect","min":"","max":"","enum_allowed":["error","redirect"],"desc":"","long_desc":"","tags":[],"see_also":[]},"standby_error_status_code":{"name":"standby_error_status_code","type":"int","level":"advanced","flags":0,"default_value":"500","min":"400","max":"599","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"url_prefix":{"name":"url_prefix","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"devicehealth","can_run":true,"error_string":"","module_options":{"enable_monitoring":{"name":"enable_monitoring","type":"bool","level":"advanced","flags":1,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"monitor device health metrics","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"mark_out_threshold":{"name":"mark_out_threshold","type":"secs","level":"advanced","flags":1,"default_value":"2419200","min":"","max":"","enum_allowed":[],"desc":"automatically mark OSD if it may fail before this long","long_desc":"","tags":[],"see_also":[]},"pool_name":{"name":"pool_name","type":"str","level":"advanced","flags":1,"default_value":"device_health_metrics","min":"","max":"","enum_allowed":[],"desc":"name of pool in which to store device health metrics","long_desc":"","tags":[],"see_also":[]},"retention_period":{"name":"retention_period","type":"secs","level":"advanced","flags":1,"default_value":"15552000","min":"","max":"","enum_allowed":[],"desc":"how long to retain device health metrics","long_desc":"","tags":[],"see_also":[]},"scrape_frequency":{"name":"scrape_frequency","type":"secs","level":"advanced","flags":1,"default_value":"86400","min":"","max":"","enum_allowed":[],"desc":"how frequently to scrape device health metrics","long_desc":"","tags":[],"see_also":[]},"self_heal":{"name":"self_heal","type":"bool","level":"advanced","flags":1,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"preemptively heal cluster around devices that may fail","long_desc":"","tags":[],"see_also":[]},"sleep_interval":{"name":"sleep_interval","type":"secs","level":"advanced","flags":1,"default_value":"600","min":"","max":"","enum_allowed":[],"desc":"how frequently to wake up and check device health","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"warn_threshold":{"name":"warn_threshold","type":"secs","level":"advanced","flags":1,"default_value":"7257600","min":"","max":"","enum_allowed":[],"desc":"raise health warning if OSD may fail before this long","long_desc":"","tags":[],"see_also":[]}}},{"name":"influx","can_run":false,"error_string":"influxdb python module not found","module_options":{"batch_size":{"name":"batch_size","type":"int","level":"advanced","flags":0,"default_value":"5000","min":"","max":"","enum_allowed":[],"desc":"How big batches of data points should be when sending to InfluxDB.","long_desc":"","tags":[],"see_also":[]},"database":{"name":"database","type":"str","level":"advanced","flags":0,"default_value":"ceph","min":"","max":"","enum_allowed":[],"desc":"InfluxDB database name. You will need to create this database and grant write privileges to the configured username or the username must have admin privileges to create it.","long_desc":"","tags":[],"see_also":[]},"hostname":{"name":"hostname","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"InfluxDB server hostname","long_desc":"","tags":[],"see_also":[]},"interval":{"name":"interval","type":"secs","level":"advanced","flags":0,"default_value":"30","min":"5","max":"","enum_allowed":[],"desc":"Time between reports to InfluxDB. Default 30 seconds.","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"password":{"name":"password","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"password of InfluxDB server user","long_desc":"","tags":[],"see_also":[]},"port":{"name":"port","type":"int","level":"advanced","flags":0,"default_value":"8086","min":"","max":"","enum_allowed":[],"desc":"InfluxDB server port","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"ssl":{"name":"ssl","type":"str","level":"advanced","flags":0,"default_value":"false","min":"","max":"","enum_allowed":[],"desc":"Use https connection for InfluxDB server. Use \"true\" or \"false\".","long_desc":"","tags":[],"see_also":[]},"threads":{"name":"threads","type":"int","level":"advanced","flags":0,"default_value":"5","min":"1","max":"32","enum_allowed":[],"desc":"How many worker threads should be spawned for sending data to InfluxDB.","long_desc":"","tags":[],"see_also":[]},"username":{"name":"username","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"username of InfluxDB server user","long_desc":"","tags":[],"see_also":[]},"verify_ssl":{"name":"verify_ssl","type":"str","level":"advanced","flags":0,"default_value":"true","min":"","max":"","enum_allowed":[],"desc":"Verify https cert for InfluxDB server. Use \"true\" or \"false\".","long_desc":"","tags":[],"see_also":[]}}},{"name":"insights","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"iostat","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"k8sevents","can_run":true,"error_string":"","module_options":{"ceph_event_retention_days":{"name":"ceph_event_retention_days","type":"int","level":"advanced","flags":0,"default_value":"7","min":"","max":"","enum_allowed":[],"desc":"Days to hold ceph event information within local cache","long_desc":"","tags":[],"see_also":[]},"config_check_secs":{"name":"config_check_secs","type":"int","level":"advanced","flags":0,"default_value":"10","min":"10","max":"","enum_allowed":[],"desc":"interval (secs) to check for cluster configuration changes","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"localpool","can_run":true,"error_string":"","module_options":{"failure_domain":{"name":"failure_domain","type":"str","level":"advanced","flags":1,"default_value":"host","min":"","max":"","enum_allowed":[],"desc":"failure domain for any created local pool","long_desc":"what failure domain we should separate data replicas across.","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"min_size":{"name":"min_size","type":"int","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"default min_size for any created local pool","long_desc":"value to set min_size to (unchanged from Ceph's default if this option is not set)","tags":[],"see_also":[]},"num_rep":{"name":"num_rep","type":"int","level":"advanced","flags":1,"default_value":"3","min":"","max":"","enum_allowed":[],"desc":"default replica count for any created local pool","long_desc":"","tags":[],"see_also":[]},"pg_num":{"name":"pg_num","type":"int","level":"advanced","flags":1,"default_value":"128","min":"","max":"","enum_allowed":[],"desc":"default pg_num for any created local pool","long_desc":"","tags":[],"see_also":[]},"prefix":{"name":"prefix","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"name prefix for any created local pool","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"subtree":{"name":"subtree","type":"str","level":"advanced","flags":1,"default_value":"rack","min":"","max":"","enum_allowed":[],"desc":"CRUSH level for which to create a local pool","long_desc":"which CRUSH subtree type the module should create a pool for.","tags":[],"see_also":[]}}},{"name":"mds_autoscaler","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"mirroring","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"nfs","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"orchestrator","can_run":true,"error_string":"","module_options":{"fail_fs":{"name":"fail_fs","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Fail filesystem for rapid multi-rank mds upgrade","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"orchestrator":{"name":"orchestrator","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["cephadm","rook","test_orchestrator"],"desc":"Orchestrator backend","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"osd_perf_query","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"osd_support","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"pg_autoscaler","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sleep_interval":{"name":"sleep_interval","type":"secs","level":"advanced","flags":0,"default_value":"60","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"threshold":{"name":"threshold","type":"float","level":"advanced","flags":0,"default_value":"3.0","min":"1.0","max":"","enum_allowed":[],"desc":"scaling threshold","long_desc":"The factor by which the `NEW PG_NUM` must vary from the current`PG_NUM` before being accepted. Cannot be less than 1.0","tags":[],"see_also":[]}}},{"name":"progress","can_run":true,"error_string":"","module_options":{"allow_pg_recovery_event":{"name":"allow_pg_recovery_event","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"allow the module to show pg recovery progress","long_desc":"","tags":[],"see_also":[]},"enabled":{"name":"enabled","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"max_completed_events":{"name":"max_completed_events","type":"int","level":"advanced","flags":1,"default_value":"50","min":"","max":"","enum_allowed":[],"desc":"number of past completed events to remember","long_desc":"","tags":[],"see_also":[]},"sleep_interval":{"name":"sleep_interval","type":"secs","level":"advanced","flags":1,"default_value":"5","min":"","max":"","enum_allowed":[],"desc":"how long the module is going to sleep","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"prometheus","can_run":true,"error_string":"","module_options":{"cache":{"name":"cache","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"exclude_perf_counters":{"name":"exclude_perf_counters","type":"bool","level":"advanced","flags":1,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"Do not include perf-counters in the metrics output","long_desc":"Gathering perf-counters from a single Prometheus exporter can degrade ceph-mgr performance, especially in large clusters. Instead, Ceph-exporter daemons are now used by default for perf-counter gathering. This should only be disabled when no ceph-exporters are deployed.","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"rbd_stats_pools":{"name":"rbd_stats_pools","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"rbd_stats_pools_refresh_interval":{"name":"rbd_stats_pools_refresh_interval","type":"int","level":"advanced","flags":0,"default_value":"300","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"scrape_interval":{"name":"scrape_interval","type":"float","level":"advanced","flags":0,"default_value":"15.0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"server_addr":{"name":"server_addr","type":"str","level":"advanced","flags":0,"default_value":"::","min":"","max":"","enum_allowed":[],"desc":"the IPv4 or IPv6 address on which the module listens for HTTP requests","long_desc":"","tags":[],"see_also":[]},"server_port":{"name":"server_port","type":"int","level":"advanced","flags":1,"default_value":"9283","min":"","max":"","enum_allowed":[],"desc":"the port on which the module listens for HTTP requests","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"stale_cache_strategy":{"name":"stale_cache_strategy","type":"str","level":"advanced","flags":0,"default_value":"log","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"standby_behaviour":{"name":"standby_behaviour","type":"str","level":"advanced","flags":1,"default_value":"default","min":"","max":"","enum_allowed":["default","error"],"desc":"","long_desc":"","tags":[],"see_also":[]},"standby_error_status_code":{"name":"standby_error_status_code","type":"int","level":"advanced","flags":1,"default_value":"500","min":"400","max":"599","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"rbd_support","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"max_concurrent_snap_create":{"name":"max_concurrent_snap_create","type":"int","level":"advanced","flags":0,"default_value":"10","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"mirror_snapshot_schedule":{"name":"mirror_snapshot_schedule","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"trash_purge_schedule":{"name":"trash_purge_schedule","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"restful","can_run":true,"error_string":"","module_options":{"enable_auth":{"name":"enable_auth","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"key_file":{"name":"key_file","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"max_requests":{"name":"max_requests","type":"int","level":"advanced","flags":0,"default_value":"500","min":"","max":"","enum_allowed":[],"desc":"Maximum number of requests to keep in memory. When new request comes in, the oldest request will be removed if the number of requests exceeds the max request number. if un-finished request is removed, error message will be logged in the ceph-mgr log.","long_desc":"","tags":[],"see_also":[]},"server_addr":{"name":"server_addr","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"server_port":{"name":"server_port","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"rgw","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"secondary_zone_period_retry_limit":{"name":"secondary_zone_period_retry_limit","type":"int","level":"advanced","flags":0,"default_value":"5","min":"","max":"","enum_allowed":[],"desc":"RGW module period update retry limit for secondary site","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"rook","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"storage_class":{"name":"storage_class","type":"str","level":"advanced","flags":0,"default_value":"local","min":"","max":"","enum_allowed":[],"desc":"storage class name for LSO-discovered PVs","long_desc":"","tags":[],"see_also":[]}}},{"name":"selftest","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"roption1":{"name":"roption1","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"roption2":{"name":"roption2","type":"str","level":"advanced","flags":0,"default_value":"xyz","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"rwoption1":{"name":"rwoption1","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"rwoption2":{"name":"rwoption2","type":"int","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"rwoption3":{"name":"rwoption3","type":"float","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"rwoption4":{"name":"rwoption4","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"rwoption5":{"name":"rwoption5","type":"bool","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"rwoption6":{"name":"rwoption6","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"rwoption7":{"name":"rwoption7","type":"int","level":"advanced","flags":0,"default_value":"","min":"1","max":"42","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"testkey":{"name":"testkey","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"testlkey":{"name":"testlkey","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"testnewline":{"name":"testnewline","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"snap_schedule","can_run":true,"error_string":"","module_options":{"allow_m_granularity":{"name":"allow_m_granularity","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"allow minute scheduled snapshots","long_desc":"","tags":[],"see_also":[]},"dump_on_update":{"name":"dump_on_update","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"dump database to debug log on update","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"stats","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"status","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"telegraf","can_run":true,"error_string":"","module_options":{"address":{"name":"address","type":"str","level":"advanced","flags":0,"default_value":"unixgram:///tmp/telegraf.sock","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"interval":{"name":"interval","type":"secs","level":"advanced","flags":0,"default_value":"15","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"telemetry","can_run":true,"error_string":"","module_options":{"channel_basic":{"name":"channel_basic","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"Share basic cluster information (size, version)","long_desc":"","tags":[],"see_also":[]},"channel_crash":{"name":"channel_crash","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"Share metadata about Ceph daemon crashes (version, stack straces, etc)","long_desc":"","tags":[],"see_also":[]},"channel_device":{"name":"channel_device","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"Share device health metrics (e.g., SMART data, minus potentially identifying info like serial numbers)","long_desc":"","tags":[],"see_also":[]},"channel_ident":{"name":"channel_ident","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Share a user-provided description and/or contact email for the cluster","long_desc":"","tags":[],"see_also":[]},"channel_perf":{"name":"channel_perf","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Share various performance metrics of a cluster","long_desc":"","tags":[],"see_also":[]},"contact":{"name":"contact","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"description":{"name":"description","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"device_url":{"name":"device_url","type":"str","level":"advanced","flags":0,"default_value":"https://telemetry.ceph.com/device","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"enabled":{"name":"enabled","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"interval":{"name":"interval","type":"int","level":"advanced","flags":0,"default_value":"24","min":"8","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"last_opt_revision":{"name":"last_opt_revision","type":"int","level":"advanced","flags":0,"default_value":"1","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"leaderboard":{"name":"leaderboard","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"leaderboard_description":{"name":"leaderboard_description","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"organization":{"name":"organization","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"proxy":{"name":"proxy","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"url":{"name":"url","type":"str","level":"advanced","flags":0,"default_value":"https://telemetry.ceph.com/report","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"test_orchestrator","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"volumes","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"max_concurrent_clones":{"name":"max_concurrent_clones","type":"int","level":"advanced","flags":0,"default_value":"4","min":"","max":"","enum_allowed":[],"desc":"Number of asynchronous cloner threads","long_desc":"","tags":[],"see_also":[]},"periodic_async_work":{"name":"periodic_async_work","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Periodically check for async work","long_desc":"","tags":[],"see_also":[]},"snapshot_clone_delay":{"name":"snapshot_clone_delay","type":"int","level":"advanced","flags":0,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"Delay clone begin operation by snapshot_clone_delay seconds","long_desc":"","tags":[],"see_also":[]},"snapshot_clone_no_wait":{"name":"snapshot_clone_no_wait","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"Reject subvolume clone request when cloner threads are busy","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"zabbix","can_run":true,"error_string":"","module_options":{"discovery_interval":{"name":"discovery_interval","type":"uint","level":"advanced","flags":0,"default_value":"100","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"identifier":{"name":"identifier","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"interval":{"name":"interval","type":"secs","level":"advanced","flags":0,"default_value":"60","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"zabbix_host":{"name":"zabbix_host","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"zabbix_port":{"name":"zabbix_port","type":"int","level":"advanced","flags":0,"default_value":"10051","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"zabbix_sender":{"name":"zabbix_sender","type":"str","level":"advanced","flags":0,"default_value":"/usr/bin/zabbix_sender","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}}]}],"modules":["cephadm","dashboard","iostat","nfs","restful"],"available_modules":[{"name":"alerts","can_run":true,"error_string":"","module_options":{"interval":{"name":"interval","type":"secs","level":"advanced","flags":1,"default_value":"60","min":"","max":"","enum_allowed":[],"desc":"How frequently to reexamine health status","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"smtp_destination":{"name":"smtp_destination","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"Email address to send alerts to","long_desc":"","tags":[],"see_also":[]},"smtp_from_name":{"name":"smtp_from_name","type":"str","level":"advanced","flags":1,"default_value":"Ceph","min":"","max":"","enum_allowed":[],"desc":"Email From: name","long_desc":"","tags":[],"see_also":[]},"smtp_host":{"name":"smtp_host","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"SMTP server","long_desc":"","tags":[],"see_also":[]},"smtp_password":{"name":"smtp_password","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"Password to authenticate with","long_desc":"","tags":[],"see_also":[]},"smtp_port":{"name":"smtp_port","type":"int","level":"advanced","flags":1,"default_value":"465","min":"","max":"","enum_allowed":[],"desc":"SMTP port","long_desc":"","tags":[],"see_also":[]},"smtp_sender":{"name":"smtp_sender","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"SMTP envelope sender","long_desc":"","tags":[],"see_also":[]},"smtp_ssl":{"name":"smtp_ssl","type":"bool","level":"advanced","flags":1,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"Use SSL to connect to SMTP server","long_desc":"","tags":[],"see_also":[]},"smtp_user":{"name":"smtp_user","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"User to authenticate as","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"balancer","can_run":true,"error_string":"","module_options":{"active":{"name":"active","type":"bool","level":"advanced","flags":1,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"automatically balance PGs across cluster","long_desc":"","tags":[],"see_also":[]},"begin_time":{"name":"begin_time","type":"str","level":"advanced","flags":1,"default_value":"0000","min":"","max":"","enum_allowed":[],"desc":"beginning time of day to automatically balance","long_desc":"This is a time of day in the format HHMM.","tags":[],"see_also":[]},"begin_weekday":{"name":"begin_weekday","type":"uint","level":"advanced","flags":1,"default_value":"0","min":"0","max":"6","enum_allowed":[],"desc":"Restrict automatic balancing to this day of the week or later","long_desc":"0 = Sunday, 1 = Monday, etc.","tags":[],"see_also":[]},"crush_compat_max_iterations":{"name":"crush_compat_max_iterations","type":"uint","level":"advanced","flags":1,"default_value":"25","min":"1","max":"250","enum_allowed":[],"desc":"maximum number of iterations to attempt optimization","long_desc":"","tags":[],"see_also":[]},"crush_compat_metrics":{"name":"crush_compat_metrics","type":"str","level":"advanced","flags":1,"default_value":"pgs,objects,bytes","min":"","max":"","enum_allowed":[],"desc":"metrics with which to calculate OSD utilization","long_desc":"Value is a list of one or more of \"pgs\", \"objects\", or \"bytes\", and indicates which metrics to use to balance utilization.","tags":[],"see_also":[]},"crush_compat_step":{"name":"crush_compat_step","type":"float","level":"advanced","flags":1,"default_value":"0.5","min":"0.001","max":"0.999","enum_allowed":[],"desc":"aggressiveness of optimization","long_desc":".99 is very aggressive, .01 is less aggressive","tags":[],"see_also":[]},"end_time":{"name":"end_time","type":"str","level":"advanced","flags":1,"default_value":"2359","min":"","max":"","enum_allowed":[],"desc":"ending time of day to automatically balance","long_desc":"This is a time of day in the format HHMM.","tags":[],"see_also":[]},"end_weekday":{"name":"end_weekday","type":"uint","level":"advanced","flags":1,"default_value":"0","min":"0","max":"6","enum_allowed":[],"desc":"Restrict automatic balancing to days of the week earlier than this","long_desc":"0 = Sunday, 1 = Monday, etc.","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"min_score":{"name":"min_score","type":"float","level":"advanced","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"minimum score, below which no optimization is attempted","long_desc":"","tags":[],"see_also":[]},"mode":{"name":"mode","type":"str","level":"advanced","flags":1,"default_value":"upmap","min":"","max":"","enum_allowed":["crush-compat","none","read","upmap","upmap-read"],"desc":"Balancer mode","long_desc":"","tags":[],"see_also":[]},"pool_ids":{"name":"pool_ids","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"pools which the automatic balancing will be limited to","long_desc":"","tags":[],"see_also":[]},"sleep_interval":{"name":"sleep_interval","type":"secs","level":"advanced","flags":1,"default_value":"60","min":"","max":"","enum_allowed":[],"desc":"how frequently to wake up and attempt optimization","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"update_pg_upmap_activity":{"name":"update_pg_upmap_activity","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Updates pg_upmap activity stats to be used in `balancer status detail`","long_desc":"","tags":[],"see_also":[]},"upmap_max_deviation":{"name":"upmap_max_deviation","type":"int","level":"advanced","flags":1,"default_value":"5","min":"1","max":"","enum_allowed":[],"desc":"deviation below which no optimization is attempted","long_desc":"If the number of PGs are within this count then no optimization is attempted","tags":[],"see_also":[]},"upmap_max_optimizations":{"name":"upmap_max_optimizations","type":"uint","level":"advanced","flags":1,"default_value":"10","min":"","max":"","enum_allowed":[],"desc":"maximum upmap optimizations to make per attempt","long_desc":"","tags":[],"see_also":[]}}},{"name":"cephadm","can_run":true,"error_string":"","module_options":{"agent_down_multiplier":{"name":"agent_down_multiplier","type":"float","level":"advanced","flags":0,"default_value":"3.0","min":"","max":"","enum_allowed":[],"desc":"Multiplied by agent refresh rate to calculate how long agent must not report before being marked down","long_desc":"","tags":[],"see_also":[]},"agent_refresh_rate":{"name":"agent_refresh_rate","type":"secs","level":"advanced","flags":0,"default_value":"20","min":"","max":"","enum_allowed":[],"desc":"How often agent on each host will try to gather and send metadata","long_desc":"","tags":[],"see_also":[]},"agent_starting_port":{"name":"agent_starting_port","type":"int","level":"advanced","flags":0,"default_value":"4721","min":"","max":"","enum_allowed":[],"desc":"First port agent will try to bind to (will also try up to next 1000 subsequent ports if blocked)","long_desc":"","tags":[],"see_also":[]},"allow_ptrace":{"name":"allow_ptrace","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"allow SYS_PTRACE capability on ceph containers","long_desc":"The SYS_PTRACE capability is needed to attach to a process with gdb or strace. Enabling this options can allow debugging daemons that encounter problems at runtime.","tags":[],"see_also":[]},"autotune_interval":{"name":"autotune_interval","type":"secs","level":"advanced","flags":0,"default_value":"600","min":"","max":"","enum_allowed":[],"desc":"how frequently to autotune daemon memory","long_desc":"","tags":[],"see_also":[]},"autotune_memory_target_ratio":{"name":"autotune_memory_target_ratio","type":"float","level":"advanced","flags":0,"default_value":"0.7","min":"","max":"","enum_allowed":[],"desc":"ratio of total system memory to divide amongst autotuned daemons","long_desc":"","tags":[],"see_also":[]},"cephadm_log_destination":{"name":"cephadm_log_destination","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":["file","file,syslog","syslog"],"desc":"Destination for cephadm command's persistent logging","long_desc":"","tags":[],"see_also":[]},"cgroups_split":{"name":"cgroups_split","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"Pass --cgroups=split when cephadm creates containers (currently podman only)","long_desc":"","tags":[],"see_also":[]},"config_checks_enabled":{"name":"config_checks_enabled","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Enable or disable the cephadm configuration analysis","long_desc":"","tags":[],"see_also":[]},"config_dashboard":{"name":"config_dashboard","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"manage configs like API endpoints in Dashboard.","long_desc":"","tags":[],"see_also":[]},"container_image_alertmanager":{"name":"container_image_alertmanager","type":"str","level":"advanced","flags":0,"default_value":"quay.io/prometheus/alertmanager:v0.25.0","min":"","max":"","enum_allowed":[],"desc":"Prometheus container image","long_desc":"","tags":[],"see_also":[]},"container_image_base":{"name":"container_image_base","type":"str","level":"advanced","flags":1,"default_value":"quay.io/ceph/ceph","min":"","max":"","enum_allowed":[],"desc":"Container image name, without the tag","long_desc":"","tags":[],"see_also":[]},"container_image_elasticsearch":{"name":"container_image_elasticsearch","type":"str","level":"advanced","flags":0,"default_value":"quay.io/omrizeneva/elasticsearch:6.8.23","min":"","max":"","enum_allowed":[],"desc":"elasticsearch container image","long_desc":"","tags":[],"see_also":[]},"container_image_grafana":{"name":"container_image_grafana","type":"str","level":"advanced","flags":0,"default_value":"quay.io/ceph/grafana:10.4.0","min":"","max":"","enum_allowed":[],"desc":"Prometheus container image","long_desc":"","tags":[],"see_also":[]},"container_image_haproxy":{"name":"container_image_haproxy","type":"str","level":"advanced","flags":0,"default_value":"quay.io/ceph/haproxy:2.3","min":"","max":"","enum_allowed":[],"desc":"HAproxy container image","long_desc":"","tags":[],"see_also":[]},"container_image_jaeger_agent":{"name":"container_image_jaeger_agent","type":"str","level":"advanced","flags":0,"default_value":"quay.io/jaegertracing/jaeger-agent:1.29","min":"","max":"","enum_allowed":[],"desc":"Jaeger agent container image","long_desc":"","tags":[],"see_also":[]},"container_image_jaeger_collector":{"name":"container_image_jaeger_collector","type":"str","level":"advanced","flags":0,"default_value":"quay.io/jaegertracing/jaeger-collector:1.29","min":"","max":"","enum_allowed":[],"desc":"Jaeger collector container image","long_desc":"","tags":[],"see_also":[]},"container_image_jaeger_query":{"name":"container_image_jaeger_query","type":"str","level":"advanced","flags":0,"default_value":"quay.io/jaegertracing/jaeger-query:1.29","min":"","max":"","enum_allowed":[],"desc":"Jaeger query container image","long_desc":"","tags":[],"see_also":[]},"container_image_keepalived":{"name":"container_image_keepalived","type":"str","level":"advanced","flags":0,"default_value":"quay.io/ceph/keepalived:2.2.4","min":"","max":"","enum_allowed":[],"desc":"Keepalived container image","long_desc":"","tags":[],"see_also":[]},"container_image_loki":{"name":"container_image_loki","type":"str","level":"advanced","flags":0,"default_value":"quay.io/ceph/loki:3.0.0","min":"","max":"","enum_allowed":[],"desc":"Loki container image","long_desc":"","tags":[],"see_also":[]},"container_image_node_exporter":{"name":"container_image_node_exporter","type":"str","level":"advanced","flags":0,"default_value":"quay.io/prometheus/node-exporter:v1.7.0","min":"","max":"","enum_allowed":[],"desc":"Prometheus container image","long_desc":"","tags":[],"see_also":[]},"container_image_nvmeof":{"name":"container_image_nvmeof","type":"str","level":"advanced","flags":0,"default_value":"quay.io/ceph/nvmeof:1.2.5","min":"","max":"","enum_allowed":[],"desc":"Nvme-of container image","long_desc":"","tags":[],"see_also":[]},"container_image_prometheus":{"name":"container_image_prometheus","type":"str","level":"advanced","flags":0,"default_value":"quay.io/prometheus/prometheus:v2.51.0","min":"","max":"","enum_allowed":[],"desc":"Prometheus container image","long_desc":"","tags":[],"see_also":[]},"container_image_promtail":{"name":"container_image_promtail","type":"str","level":"advanced","flags":0,"default_value":"quay.io/ceph/promtail:3.0.0","min":"","max":"","enum_allowed":[],"desc":"Promtail container image","long_desc":"","tags":[],"see_also":[]},"container_image_samba":{"name":"container_image_samba","type":"str","level":"advanced","flags":0,"default_value":"quay.io/samba.org/samba-server:devbuilds-centos-amd64","min":"","max":"","enum_allowed":[],"desc":"Samba/SMB container image","long_desc":"","tags":[],"see_also":[]},"container_image_snmp_gateway":{"name":"container_image_snmp_gateway","type":"str","level":"advanced","flags":0,"default_value":"quay.io/ceph/snmp-notifier:v1.2.1","min":"","max":"","enum_allowed":[],"desc":"SNMP Gateway container image","long_desc":"","tags":[],"see_also":[]},"container_init":{"name":"container_init","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"Run podman/docker with `--init`","long_desc":"","tags":[],"see_also":[]},"daemon_cache_timeout":{"name":"daemon_cache_timeout","type":"secs","level":"advanced","flags":0,"default_value":"600","min":"","max":"","enum_allowed":[],"desc":"seconds to cache service (daemon) inventory","long_desc":"","tags":[],"see_also":[]},"default_cephadm_command_timeout":{"name":"default_cephadm_command_timeout","type":"int","level":"advanced","flags":0,"default_value":"900","min":"","max":"","enum_allowed":[],"desc":"Default timeout applied to cephadm commands run directly on the host (in seconds)","long_desc":"","tags":[],"see_also":[]},"default_registry":{"name":"default_registry","type":"str","level":"advanced","flags":0,"default_value":"quay.io","min":"","max":"","enum_allowed":[],"desc":"Search-registry to which we should normalize unqualified image names. This is not the default registry","long_desc":"","tags":[],"see_also":[]},"device_cache_timeout":{"name":"device_cache_timeout","type":"secs","level":"advanced","flags":0,"default_value":"1800","min":"","max":"","enum_allowed":[],"desc":"seconds to cache device inventory","long_desc":"","tags":[],"see_also":[]},"device_enhanced_scan":{"name":"device_enhanced_scan","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Use libstoragemgmt during device scans","long_desc":"","tags":[],"see_also":[]},"facts_cache_timeout":{"name":"facts_cache_timeout","type":"secs","level":"advanced","flags":0,"default_value":"60","min":"","max":"","enum_allowed":[],"desc":"seconds to cache host facts data","long_desc":"","tags":[],"see_also":[]},"grafana_dashboards_path":{"name":"grafana_dashboards_path","type":"str","level":"advanced","flags":0,"default_value":"/etc/grafana/dashboards/ceph-dashboard/","min":"","max":"","enum_allowed":[],"desc":"location of dashboards to include in grafana deployments","long_desc":"","tags":[],"see_also":[]},"host_check_interval":{"name":"host_check_interval","type":"secs","level":"advanced","flags":0,"default_value":"600","min":"","max":"","enum_allowed":[],"desc":"how frequently to perform a host check","long_desc":"","tags":[],"see_also":[]},"hw_monitoring":{"name":"hw_monitoring","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Deploy hw monitoring daemon on every host.","long_desc":"","tags":[],"see_also":[]},"inventory_list_all":{"name":"inventory_list_all","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Whether ceph-volume inventory should report more devices (mostly mappers (LVs / mpaths), partitions...)","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_refresh_metadata":{"name":"log_refresh_metadata","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Log all refresh metadata. Includes daemon, device, and host info collected regularly. Only has effect if logging at debug level","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"log to the \"cephadm\" cluster log channel\"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"manage_etc_ceph_ceph_conf":{"name":"manage_etc_ceph_ceph_conf","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Manage and own /etc/ceph/ceph.conf on the hosts.","long_desc":"","tags":[],"see_also":[]},"manage_etc_ceph_ceph_conf_hosts":{"name":"manage_etc_ceph_ceph_conf_hosts","type":"str","level":"advanced","flags":0,"default_value":"*","min":"","max":"","enum_allowed":[],"desc":"PlacementSpec describing on which hosts to manage /etc/ceph/ceph.conf","long_desc":"","tags":[],"see_also":[]},"max_count_per_host":{"name":"max_count_per_host","type":"int","level":"advanced","flags":0,"default_value":"10","min":"","max":"","enum_allowed":[],"desc":"max number of daemons per service per host","long_desc":"","tags":[],"see_also":[]},"max_osd_draining_count":{"name":"max_osd_draining_count","type":"int","level":"advanced","flags":0,"default_value":"10","min":"","max":"","enum_allowed":[],"desc":"max number of osds that will be drained simultaneously when osds are removed","long_desc":"","tags":[],"see_also":[]},"migration_current":{"name":"migration_current","type":"int","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"internal - do not modify","long_desc":"","tags":[],"see_also":[]},"mode":{"name":"mode","type":"str","level":"advanced","flags":0,"default_value":"root","min":"","max":"","enum_allowed":["cephadm-package","root"],"desc":"mode for remote execution of cephadm","long_desc":"","tags":[],"see_also":[]},"oob_default_addr":{"name":"oob_default_addr","type":"str","level":"advanced","flags":0,"default_value":"169.254.1.1","min":"","max":"","enum_allowed":[],"desc":"Default address for RedFish API (oob management).","long_desc":"","tags":[],"see_also":[]},"prometheus_alerts_path":{"name":"prometheus_alerts_path","type":"str","level":"advanced","flags":0,"default_value":"/etc/prometheus/ceph/ceph_default_alerts.yml","min":"","max":"","enum_allowed":[],"desc":"location of alerts to include in prometheus deployments","long_desc":"","tags":[],"see_also":[]},"registry_insecure":{"name":"registry_insecure","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Registry is to be considered insecure (no TLS available). Only for development purposes.","long_desc":"","tags":[],"see_also":[]},"registry_password":{"name":"registry_password","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"Custom repository password. Only used for logging into a registry.","long_desc":"","tags":[],"see_also":[]},"registry_url":{"name":"registry_url","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"Registry url for login purposes. This is not the default registry","long_desc":"","tags":[],"see_also":[]},"registry_username":{"name":"registry_username","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"Custom repository username. Only used for logging into a registry.","long_desc":"","tags":[],"see_also":[]},"secure_monitoring_stack":{"name":"secure_monitoring_stack","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Enable TLS security for all the monitoring stack daemons","long_desc":"","tags":[],"see_also":[]},"service_discovery_port":{"name":"service_discovery_port","type":"int","level":"advanced","flags":0,"default_value":"8765","min":"","max":"","enum_allowed":[],"desc":"cephadm service discovery port","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"ssh_config_file":{"name":"ssh_config_file","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"customized SSH config file to connect to managed hosts","long_desc":"","tags":[],"see_also":[]},"ssh_keepalive_count_max":{"name":"ssh_keepalive_count_max","type":"int","level":"advanced","flags":0,"default_value":"3","min":"","max":"","enum_allowed":[],"desc":"How many times ssh connections can fail liveness checks before the host is marked offline","long_desc":"","tags":[],"see_also":[]},"ssh_keepalive_interval":{"name":"ssh_keepalive_interval","type":"int","level":"advanced","flags":0,"default_value":"7","min":"","max":"","enum_allowed":[],"desc":"How often ssh connections are checked for liveness","long_desc":"","tags":[],"see_also":[]},"use_agent":{"name":"use_agent","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Use cephadm agent on each host to gather and send metadata","long_desc":"","tags":[],"see_also":[]},"use_repo_digest":{"name":"use_repo_digest","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"Automatically convert image tags to image digest. Make sure all daemons use the same image","long_desc":"","tags":[],"see_also":[]},"warn_on_failed_host_check":{"name":"warn_on_failed_host_check","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"raise a health warning if the host check fails","long_desc":"","tags":[],"see_also":[]},"warn_on_stray_daemons":{"name":"warn_on_stray_daemons","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"raise a health warning if daemons are detected that are not managed by cephadm","long_desc":"","tags":[],"see_also":[]},"warn_on_stray_hosts":{"name":"warn_on_stray_hosts","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"raise a health warning if daemons are detected on a host that is not managed by cephadm","long_desc":"","tags":[],"see_also":[]}}},{"name":"crash","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"retain_interval":{"name":"retain_interval","type":"secs","level":"advanced","flags":1,"default_value":"31536000","min":"","max":"","enum_allowed":[],"desc":"how long to retain crashes before pruning them","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"warn_recent_interval":{"name":"warn_recent_interval","type":"secs","level":"advanced","flags":1,"default_value":"1209600","min":"","max":"","enum_allowed":[],"desc":"time interval in which to warn about recent crashes","long_desc":"","tags":[],"see_also":[]}}},{"name":"dashboard","can_run":true,"error_string":"","module_options":{"ACCOUNT_LOCKOUT_ATTEMPTS":{"name":"ACCOUNT_LOCKOUT_ATTEMPTS","type":"int","level":"advanced","flags":0,"default_value":"10","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"ALERTMANAGER_API_HOST":{"name":"ALERTMANAGER_API_HOST","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"ALERTMANAGER_API_SSL_VERIFY":{"name":"ALERTMANAGER_API_SSL_VERIFY","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"AUDIT_API_ENABLED":{"name":"AUDIT_API_ENABLED","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"AUDIT_API_LOG_PAYLOAD":{"name":"AUDIT_API_LOG_PAYLOAD","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"ENABLE_BROWSABLE_API":{"name":"ENABLE_BROWSABLE_API","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"FEATURE_TOGGLE_CEPHFS":{"name":"FEATURE_TOGGLE_CEPHFS","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"FEATURE_TOGGLE_DASHBOARD":{"name":"FEATURE_TOGGLE_DASHBOARD","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"FEATURE_TOGGLE_ISCSI":{"name":"FEATURE_TOGGLE_ISCSI","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"FEATURE_TOGGLE_MIRRORING":{"name":"FEATURE_TOGGLE_MIRRORING","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"FEATURE_TOGGLE_NFS":{"name":"FEATURE_TOGGLE_NFS","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"FEATURE_TOGGLE_RBD":{"name":"FEATURE_TOGGLE_RBD","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"FEATURE_TOGGLE_RGW":{"name":"FEATURE_TOGGLE_RGW","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"GANESHA_CLUSTERS_RADOS_POOL_NAMESPACE":{"name":"GANESHA_CLUSTERS_RADOS_POOL_NAMESPACE","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"GRAFANA_API_PASSWORD":{"name":"GRAFANA_API_PASSWORD","type":"str","level":"advanced","flags":0,"default_value":"admin","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"GRAFANA_API_SSL_VERIFY":{"name":"GRAFANA_API_SSL_VERIFY","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"GRAFANA_API_URL":{"name":"GRAFANA_API_URL","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"GRAFANA_API_USERNAME":{"name":"GRAFANA_API_USERNAME","type":"str","level":"advanced","flags":0,"default_value":"admin","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"GRAFANA_FRONTEND_API_URL":{"name":"GRAFANA_FRONTEND_API_URL","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"GRAFANA_UPDATE_DASHBOARDS":{"name":"GRAFANA_UPDATE_DASHBOARDS","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"ISCSI_API_SSL_VERIFICATION":{"name":"ISCSI_API_SSL_VERIFICATION","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"ISSUE_TRACKER_API_KEY":{"name":"ISSUE_TRACKER_API_KEY","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PROMETHEUS_API_HOST":{"name":"PROMETHEUS_API_HOST","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PROMETHEUS_API_SSL_VERIFY":{"name":"PROMETHEUS_API_SSL_VERIFY","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PWD_POLICY_CHECK_COMPLEXITY_ENABLED":{"name":"PWD_POLICY_CHECK_COMPLEXITY_ENABLED","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PWD_POLICY_CHECK_EXCLUSION_LIST_ENABLED":{"name":"PWD_POLICY_CHECK_EXCLUSION_LIST_ENABLED","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PWD_POLICY_CHECK_LENGTH_ENABLED":{"name":"PWD_POLICY_CHECK_LENGTH_ENABLED","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PWD_POLICY_CHECK_OLDPWD_ENABLED":{"name":"PWD_POLICY_CHECK_OLDPWD_ENABLED","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PWD_POLICY_CHECK_REPETITIVE_CHARS_ENABLED":{"name":"PWD_POLICY_CHECK_REPETITIVE_CHARS_ENABLED","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PWD_POLICY_CHECK_SEQUENTIAL_CHARS_ENABLED":{"name":"PWD_POLICY_CHECK_SEQUENTIAL_CHARS_ENABLED","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PWD_POLICY_CHECK_USERNAME_ENABLED":{"name":"PWD_POLICY_CHECK_USERNAME_ENABLED","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PWD_POLICY_ENABLED":{"name":"PWD_POLICY_ENABLED","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PWD_POLICY_EXCLUSION_LIST":{"name":"PWD_POLICY_EXCLUSION_LIST","type":"str","level":"advanced","flags":0,"default_value":"osd,host,dashboard,pool,block,nfs,ceph,monitors,gateway,logs,crush,maps","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PWD_POLICY_MIN_COMPLEXITY":{"name":"PWD_POLICY_MIN_COMPLEXITY","type":"int","level":"advanced","flags":0,"default_value":"10","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PWD_POLICY_MIN_LENGTH":{"name":"PWD_POLICY_MIN_LENGTH","type":"int","level":"advanced","flags":0,"default_value":"8","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"REST_REQUESTS_TIMEOUT":{"name":"REST_REQUESTS_TIMEOUT","type":"int","level":"advanced","flags":0,"default_value":"45","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"RGW_API_ACCESS_KEY":{"name":"RGW_API_ACCESS_KEY","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"RGW_API_ADMIN_RESOURCE":{"name":"RGW_API_ADMIN_RESOURCE","type":"str","level":"advanced","flags":0,"default_value":"admin","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"RGW_API_SECRET_KEY":{"name":"RGW_API_SECRET_KEY","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"RGW_API_SSL_VERIFY":{"name":"RGW_API_SSL_VERIFY","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"USER_PWD_EXPIRATION_SPAN":{"name":"USER_PWD_EXPIRATION_SPAN","type":"int","level":"advanced","flags":0,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"USER_PWD_EXPIRATION_WARNING_1":{"name":"USER_PWD_EXPIRATION_WARNING_1","type":"int","level":"advanced","flags":0,"default_value":"10","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"USER_PWD_EXPIRATION_WARNING_2":{"name":"USER_PWD_EXPIRATION_WARNING_2","type":"int","level":"advanced","flags":0,"default_value":"5","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"cross_origin_url":{"name":"cross_origin_url","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"crt_file":{"name":"crt_file","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"debug":{"name":"debug","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Enable/disable debug options","long_desc":"","tags":[],"see_also":[]},"jwt_token_ttl":{"name":"jwt_token_ttl","type":"int","level":"advanced","flags":0,"default_value":"28800","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"key_file":{"name":"key_file","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"motd":{"name":"motd","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"The message of the day","long_desc":"","tags":[],"see_also":[]},"redirect_resolve_ip_addr":{"name":"redirect_resolve_ip_addr","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"server_addr":{"name":"server_addr","type":"str","level":"advanced","flags":0,"default_value":"::","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"server_port":{"name":"server_port","type":"int","level":"advanced","flags":0,"default_value":"8080","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"ssl":{"name":"ssl","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"ssl_server_port":{"name":"ssl_server_port","type":"int","level":"advanced","flags":0,"default_value":"8443","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"standby_behaviour":{"name":"standby_behaviour","type":"str","level":"advanced","flags":0,"default_value":"redirect","min":"","max":"","enum_allowed":["error","redirect"],"desc":"","long_desc":"","tags":[],"see_also":[]},"standby_error_status_code":{"name":"standby_error_status_code","type":"int","level":"advanced","flags":0,"default_value":"500","min":"400","max":"599","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"url_prefix":{"name":"url_prefix","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"devicehealth","can_run":true,"error_string":"","module_options":{"enable_monitoring":{"name":"enable_monitoring","type":"bool","level":"advanced","flags":1,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"monitor device health metrics","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"mark_out_threshold":{"name":"mark_out_threshold","type":"secs","level":"advanced","flags":1,"default_value":"2419200","min":"","max":"","enum_allowed":[],"desc":"automatically mark OSD if it may fail before this long","long_desc":"","tags":[],"see_also":[]},"pool_name":{"name":"pool_name","type":"str","level":"advanced","flags":1,"default_value":"device_health_metrics","min":"","max":"","enum_allowed":[],"desc":"name of pool in which to store device health metrics","long_desc":"","tags":[],"see_also":[]},"retention_period":{"name":"retention_period","type":"secs","level":"advanced","flags":1,"default_value":"15552000","min":"","max":"","enum_allowed":[],"desc":"how long to retain device health metrics","long_desc":"","tags":[],"see_also":[]},"scrape_frequency":{"name":"scrape_frequency","type":"secs","level":"advanced","flags":1,"default_value":"86400","min":"","max":"","enum_allowed":[],"desc":"how frequently to scrape device health metrics","long_desc":"","tags":[],"see_also":[]},"self_heal":{"name":"self_heal","type":"bool","level":"advanced","flags":1,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"preemptively heal cluster around devices that may fail","long_desc":"","tags":[],"see_also":[]},"sleep_interval":{"name":"sleep_interval","type":"secs","level":"advanced","flags":1,"default_value":"600","min":"","max":"","enum_allowed":[],"desc":"how frequently to wake up and check device health","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"warn_threshold":{"name":"warn_threshold","type":"secs","level":"advanced","flags":1,"default_value":"7257600","min":"","max":"","enum_allowed":[],"desc":"raise health warning if OSD may fail before this long","long_desc":"","tags":[],"see_also":[]}}},{"name":"influx","can_run":false,"error_string":"influxdb python module not found","module_options":{"batch_size":{"name":"batch_size","type":"int","level":"advanced","flags":0,"default_value":"5000","min":"","max":"","enum_allowed":[],"desc":"How big batches of data points should be when sending to InfluxDB.","long_desc":"","tags":[],"see_also":[]},"database":{"name":"database","type":"str","level":"advanced","flags":0,"default_value":"ceph","min":"","max":"","enum_allowed":[],"desc":"InfluxDB database name. You will need to create this database and grant write privileges to the configured username or the username must have admin privileges to create it.","long_desc":"","tags":[],"see_also":[]},"hostname":{"name":"hostname","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"InfluxDB server hostname","long_desc":"","tags":[],"see_also":[]},"interval":{"name":"interval","type":"secs","level":"advanced","flags":0,"default_value":"30","min":"5","max":"","enum_allowed":[],"desc":"Time between reports to InfluxDB. Default 30 seconds.","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"password":{"name":"password","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"password of InfluxDB server user","long_desc":"","tags":[],"see_also":[]},"port":{"name":"port","type":"int","level":"advanced","flags":0,"default_value":"8086","min":"","max":"","enum_allowed":[],"desc":"InfluxDB server port","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"ssl":{"name":"ssl","type":"str","level":"advanced","flags":0,"default_value":"false","min":"","max":"","enum_allowed":[],"desc":"Use https connection for InfluxDB server. Use \"true\" or \"false\".","long_desc":"","tags":[],"see_also":[]},"threads":{"name":"threads","type":"int","level":"advanced","flags":0,"default_value":"5","min":"1","max":"32","enum_allowed":[],"desc":"How many worker threads should be spawned for sending data to InfluxDB.","long_desc":"","tags":[],"see_also":[]},"username":{"name":"username","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"username of InfluxDB server user","long_desc":"","tags":[],"see_also":[]},"verify_ssl":{"name":"verify_ssl","type":"str","level":"advanced","flags":0,"default_value":"true","min":"","max":"","enum_allowed":[],"desc":"Verify https cert for InfluxDB server. Use \"true\" or \"false\".","long_desc":"","tags":[],"see_also":[]}}},{"name":"insights","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"iostat","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"k8sevents","can_run":true,"error_string":"","module_options":{"ceph_event_retention_days":{"name":"ceph_event_retention_days","type":"int","level":"advanced","flags":0,"default_value":"7","min":"","max":"","enum_allowed":[],"desc":"Days to hold ceph event information within local cache","long_desc":"","tags":[],"see_also":[]},"config_check_secs":{"name":"config_check_secs","type":"int","level":"advanced","flags":0,"default_value":"10","min":"10","max":"","enum_allowed":[],"desc":"interval (secs) to check for cluster configuration changes","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"localpool","can_run":true,"error_string":"","module_options":{"failure_domain":{"name":"failure_domain","type":"str","level":"advanced","flags":1,"default_value":"host","min":"","max":"","enum_allowed":[],"desc":"failure domain for any created local pool","long_desc":"what failure domain we should separate data replicas across.","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"min_size":{"name":"min_size","type":"int","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"default min_size for any created local pool","long_desc":"value to set min_size to (unchanged from Ceph's default if this option is not set)","tags":[],"see_also":[]},"num_rep":{"name":"num_rep","type":"int","level":"advanced","flags":1,"default_value":"3","min":"","max":"","enum_allowed":[],"desc":"default replica count for any created local pool","long_desc":"","tags":[],"see_also":[]},"pg_num":{"name":"pg_num","type":"int","level":"advanced","flags":1,"default_value":"128","min":"","max":"","enum_allowed":[],"desc":"default pg_num for any created local pool","long_desc":"","tags":[],"see_also":[]},"prefix":{"name":"prefix","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"name prefix for any created local pool","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"subtree":{"name":"subtree","type":"str","level":"advanced","flags":1,"default_value":"rack","min":"","max":"","enum_allowed":[],"desc":"CRUSH level for which to create a local pool","long_desc":"which CRUSH subtree type the module should create a pool for.","tags":[],"see_also":[]}}},{"name":"mds_autoscaler","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"mirroring","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"nfs","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"orchestrator","can_run":true,"error_string":"","module_options":{"fail_fs":{"name":"fail_fs","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Fail filesystem for rapid multi-rank mds upgrade","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"orchestrator":{"name":"orchestrator","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["cephadm","rook","test_orchestrator"],"desc":"Orchestrator backend","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"osd_perf_query","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"osd_support","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"pg_autoscaler","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sleep_interval":{"name":"sleep_interval","type":"secs","level":"advanced","flags":0,"default_value":"60","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"threshold":{"name":"threshold","type":"float","level":"advanced","flags":0,"default_value":"3.0","min":"1.0","max":"","enum_allowed":[],"desc":"scaling threshold","long_desc":"The factor by which the `NEW PG_NUM` must vary from the current`PG_NUM` before being accepted. Cannot be less than 1.0","tags":[],"see_also":[]}}},{"name":"progress","can_run":true,"error_string":"","module_options":{"allow_pg_recovery_event":{"name":"allow_pg_recovery_event","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"allow the module to show pg recovery progress","long_desc":"","tags":[],"see_also":[]},"enabled":{"name":"enabled","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"max_completed_events":{"name":"max_completed_events","type":"int","level":"advanced","flags":1,"default_value":"50","min":"","max":"","enum_allowed":[],"desc":"number of past completed events to remember","long_desc":"","tags":[],"see_also":[]},"sleep_interval":{"name":"sleep_interval","type":"secs","level":"advanced","flags":1,"default_value":"5","min":"","max":"","enum_allowed":[],"desc":"how long the module is going to sleep","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"prometheus","can_run":true,"error_string":"","module_options":{"cache":{"name":"cache","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"exclude_perf_counters":{"name":"exclude_perf_counters","type":"bool","level":"advanced","flags":1,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"Do not include perf-counters in the metrics output","long_desc":"Gathering perf-counters from a single Prometheus exporter can degrade ceph-mgr performance, especially in large clusters. Instead, Ceph-exporter daemons are now used by default for perf-counter gathering. This should only be disabled when no ceph-exporters are deployed.","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"rbd_stats_pools":{"name":"rbd_stats_pools","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"rbd_stats_pools_refresh_interval":{"name":"rbd_stats_pools_refresh_interval","type":"int","level":"advanced","flags":0,"default_value":"300","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"scrape_interval":{"name":"scrape_interval","type":"float","level":"advanced","flags":0,"default_value":"15.0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"server_addr":{"name":"server_addr","type":"str","level":"advanced","flags":0,"default_value":"::","min":"","max":"","enum_allowed":[],"desc":"the IPv4 or IPv6 address on which the module listens for HTTP requests","long_desc":"","tags":[],"see_also":[]},"server_port":{"name":"server_port","type":"int","level":"advanced","flags":1,"default_value":"9283","min":"","max":"","enum_allowed":[],"desc":"the port on which the module listens for HTTP requests","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"stale_cache_strategy":{"name":"stale_cache_strategy","type":"str","level":"advanced","flags":0,"default_value":"log","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"standby_behaviour":{"name":"standby_behaviour","type":"str","level":"advanced","flags":1,"default_value":"default","min":"","max":"","enum_allowed":["default","error"],"desc":"","long_desc":"","tags":[],"see_also":[]},"standby_error_status_code":{"name":"standby_error_status_code","type":"int","level":"advanced","flags":1,"default_value":"500","min":"400","max":"599","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"rbd_support","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"max_concurrent_snap_create":{"name":"max_concurrent_snap_create","type":"int","level":"advanced","flags":0,"default_value":"10","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"mirror_snapshot_schedule":{"name":"mirror_snapshot_schedule","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"trash_purge_schedule":{"name":"trash_purge_schedule","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"restful","can_run":true,"error_string":"","module_options":{"enable_auth":{"name":"enable_auth","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"key_file":{"name":"key_file","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"max_requests":{"name":"max_requests","type":"int","level":"advanced","flags":0,"default_value":"500","min":"","max":"","enum_allowed":[],"desc":"Maximum number of requests to keep in memory. When new request comes in, the oldest request will be removed if the number of requests exceeds the max request number. if un-finished request is removed, error message will be logged in the ceph-mgr log.","long_desc":"","tags":[],"see_also":[]},"server_addr":{"name":"server_addr","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"server_port":{"name":"server_port","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"rgw","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"secondary_zone_period_retry_limit":{"name":"secondary_zone_period_retry_limit","type":"int","level":"advanced","flags":0,"default_value":"5","min":"","max":"","enum_allowed":[],"desc":"RGW module period update retry limit for secondary site","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"rook","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"storage_class":{"name":"storage_class","type":"str","level":"advanced","flags":0,"default_value":"local","min":"","max":"","enum_allowed":[],"desc":"storage class name for LSO-discovered PVs","long_desc":"","tags":[],"see_also":[]}}},{"name":"selftest","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"roption1":{"name":"roption1","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"roption2":{"name":"roption2","type":"str","level":"advanced","flags":0,"default_value":"xyz","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"rwoption1":{"name":"rwoption1","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"rwoption2":{"name":"rwoption2","type":"int","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"rwoption3":{"name":"rwoption3","type":"float","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"rwoption4":{"name":"rwoption4","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"rwoption5":{"name":"rwoption5","type":"bool","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"rwoption6":{"name":"rwoption6","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"rwoption7":{"name":"rwoption7","type":"int","level":"advanced","flags":0,"default_value":"","min":"1","max":"42","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"testkey":{"name":"testkey","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"testlkey":{"name":"testlkey","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"testnewline":{"name":"testnewline","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"snap_schedule","can_run":true,"error_string":"","module_options":{"allow_m_granularity":{"name":"allow_m_granularity","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"allow minute scheduled snapshots","long_desc":"","tags":[],"see_also":[]},"dump_on_update":{"name":"dump_on_update","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"dump database to debug log on update","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"stats","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"status","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"telegraf","can_run":true,"error_string":"","module_options":{"address":{"name":"address","type":"str","level":"advanced","flags":0,"default_value":"unixgram:///tmp/telegraf.sock","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"interval":{"name":"interval","type":"secs","level":"advanced","flags":0,"default_value":"15","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"telemetry","can_run":true,"error_string":"","module_options":{"channel_basic":{"name":"channel_basic","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"Share basic cluster information (size, version)","long_desc":"","tags":[],"see_also":[]},"channel_crash":{"name":"channel_crash","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"Share metadata about Ceph daemon crashes (version, stack straces, etc)","long_desc":"","tags":[],"see_also":[]},"channel_device":{"name":"channel_device","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"Share device health metrics (e.g., SMART data, minus potentially identifying info like serial numbers)","long_desc":"","tags":[],"see_also":[]},"channel_ident":{"name":"channel_ident","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Share a user-provided description and/or contact email for the cluster","long_desc":"","tags":[],"see_also":[]},"channel_perf":{"name":"channel_perf","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Share various performance metrics of a cluster","long_desc":"","tags":[],"see_also":[]},"contact":{"name":"contact","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"description":{"name":"description","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"device_url":{"name":"device_url","type":"str","level":"advanced","flags":0,"default_value":"https://telemetry.ceph.com/device","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"enabled":{"name":"enabled","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"interval":{"name":"interval","type":"int","level":"advanced","flags":0,"default_value":"24","min":"8","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"last_opt_revision":{"name":"last_opt_revision","type":"int","level":"advanced","flags":0,"default_value":"1","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"leaderboard":{"name":"leaderboard","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"leaderboard_description":{"name":"leaderboard_description","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"organization":{"name":"organization","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"proxy":{"name":"proxy","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"url":{"name":"url","type":"str","level":"advanced","flags":0,"default_value":"https://telemetry.ceph.com/report","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"test_orchestrator","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"volumes","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"max_concurrent_clones":{"name":"max_concurrent_clones","type":"int","level":"advanced","flags":0,"default_value":"4","min":"","max":"","enum_allowed":[],"desc":"Number of asynchronous cloner threads","long_desc":"","tags":[],"see_also":[]},"periodic_async_work":{"name":"periodic_async_work","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Periodically check for async work","long_desc":"","tags":[],"see_also":[]},"snapshot_clone_delay":{"name":"snapshot_clone_delay","type":"int","level":"advanced","flags":0,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"Delay clone begin operation by snapshot_clone_delay seconds","long_desc":"","tags":[],"see_also":[]},"snapshot_clone_no_wait":{"name":"snapshot_clone_no_wait","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"Reject subvolume clone request when cloner threads are busy","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"zabbix","can_run":true,"error_string":"","module_options":{"discovery_interval":{"name":"discovery_interval","type":"uint","level":"advanced","flags":0,"default_value":"100","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"identifier":{"name":"identifier","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"interval":{"name":"interval","type":"secs","level":"advanced","flags":0,"default_value":"60","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"zabbix_host":{"name":"zabbix_host","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"zabbix_port":{"name":"zabbix_port","type":"int","level":"advanced","flags":0,"default_value":"10051","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"zabbix_sender":{"name":"zabbix_sender","type":"str","level":"advanced","flags":0,"default_value":"/usr/bin/zabbix_sender","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}}],"services":{"dashboard":"https://192.168.123.100:8443/"},"always_on_modules":{"octopus":["balancer","crash","devicehealth","orchestrator","pg_autoscaler","progress","rbd_support","status","telemetry","volumes"],"pacific":["balancer","crash","devicehealth","orchestrator","pg_autoscaler","progress","rbd_support","status","telemetry","volumes"],"quincy":["balancer","crash","devicehealth","orchestrator","pg_autoscaler","progress","rbd_support","status","telemetry","volumes"],"reef":["balancer","crash","devicehealth","orchestrator","pg_autoscaler","progress","rbd_support","status","telemetry","volumes"],"squid":["balancer","crash","devicehealth","orchestrator","pg_autoscaler","progress","rbd_support","status","telemetry","volumes"]},"force_disabled_modules":{},"last_failure_osd_epoch":3,"active_clients":[{"name":"devicehealth","addrvec":[{"type":"v2","addr":"192.168.123.100:0","nonce":774743829}]},{"name":"libcephsqlite","addrvec":[{"type":"v2","addr":"192.168.123.100:0","nonce":2407999995}]},{"name":"rbd_support","addrvec":[{"type":"v2","addr":"192.168.123.100:0","nonce":2789275962}]},{"name":"volumes","addrvec":[{"type":"v2","addr":"192.168.123.100:0","nonce":1183923103}]}]} 2026-03-06T13:42:12.981 INFO:tasks.cephadm.ceph_manager.ceph:mgr available! 2026-03-06T13:42:12.982 INFO:tasks.cephadm.ceph_manager.ceph:waiting for all up 2026-03-06T13:42:12.982 DEBUG:teuthology.orchestra.run.vm00:> sudo /home/ubuntu/cephtest/cephadm --image harbor.clyso.com/custom-ceph/ceph/ceph:cobaltcore-storage-v19.2.3-fasttrack-3 shell --fsid 82f504c8-1959-11f1-9a16-a78d135e200b -- ceph osd dump --format=json 2026-03-06T13:42:13.254 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:42:13 vm00 ceph-mon[51914]: osdmap e19: 3 total, 3 up, 3 in 2026-03-06T13:42:13.254 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:42:13 vm00 ceph-mon[51914]: mgrmap e15: a(active, since 84s), standbys: b 2026-03-06T13:42:13.254 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:42:13 vm00 ceph-mon[51914]: from='client.? 192.168.123.100:0/561994995' entity='client.admin' cmd=[{"prefix": "osd pool get", "pool": ".mgr", "var": "pg_num"}]: dispatch 2026-03-06T13:42:13.255 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:42:13 vm00 ceph-mon[51914]: from='client.? 192.168.123.100:0/683890953' entity='client.admin' cmd=[{"prefix": "mgr dump", "format": "json"}]: dispatch 2026-03-06T13:42:13.334 INFO:teuthology.orchestra.run.vm00.stderr:Inferring config /var/lib/ceph/82f504c8-1959-11f1-9a16-a78d135e200b/mon.a/config 2026-03-06T13:42:13.405 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:42:13 vm01 ceph-mon[56331]: osdmap e19: 3 total, 3 up, 3 in 2026-03-06T13:42:13.405 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:42:13 vm01 ceph-mon[56331]: mgrmap e15: a(active, since 84s), standbys: b 2026-03-06T13:42:13.405 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:42:13 vm01 ceph-mon[56331]: from='client.? 192.168.123.100:0/561994995' entity='client.admin' cmd=[{"prefix": "osd pool get", "pool": ".mgr", "var": "pg_num"}]: dispatch 2026-03-06T13:42:13.405 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:42:13 vm01 ceph-mon[56331]: from='client.? 192.168.123.100:0/683890953' entity='client.admin' cmd=[{"prefix": "mgr dump", "format": "json"}]: dispatch 2026-03-06T13:42:13.540 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 06 13:42:13 vm08 ceph-mon[53561]: osdmap e19: 3 total, 3 up, 3 in 2026-03-06T13:42:13.540 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 06 13:42:13 vm08 ceph-mon[53561]: mgrmap e15: a(active, since 84s), standbys: b 2026-03-06T13:42:13.540 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 06 13:42:13 vm08 ceph-mon[53561]: from='client.? 192.168.123.100:0/561994995' entity='client.admin' cmd=[{"prefix": "osd pool get", "pool": ".mgr", "var": "pg_num"}]: dispatch 2026-03-06T13:42:13.540 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 06 13:42:13 vm08 ceph-mon[53561]: from='client.? 192.168.123.100:0/683890953' entity='client.admin' cmd=[{"prefix": "mgr dump", "format": "json"}]: dispatch 2026-03-06T13:42:13.689 INFO:teuthology.orchestra.run.vm00.stdout: 2026-03-06T13:42:13.689 INFO:teuthology.orchestra.run.vm00.stdout:{"epoch":19,"fsid":"82f504c8-1959-11f1-9a16-a78d135e200b","created":"2026-03-06T12:40:03.178129+0000","modified":"2026-03-06T12:42:12.113876+0000","last_up_change":"2026-03-06T12:42:09.091745+0000","last_in_change":"2026-03-06T12:41:55.730672+0000","flags":"sortbitwise,recovery_deletes,purged_snapdirs,pglog_hardlimit","flags_num":5799936,"flags_set":["pglog_hardlimit","purged_snapdirs","recovery_deletes","sortbitwise"],"crush_version":8,"full_ratio":0.94999998807907104,"backfillfull_ratio":0.89999997615814209,"nearfull_ratio":0.85000002384185791,"cluster_snapshot":"","pool_max":1,"max_osd":3,"require_min_compat_client":"luminous","min_compat_client":"jewel","require_osd_release":"squid","allow_crimson":false,"pools":[{"pool":1,"pool_name":".mgr","create_time":"2026-03-06T12:42:10.032248+0000","flags":1,"flags_names":"hashpspool","type":1,"size":3,"min_size":2,"crush_rule":0,"peering_crush_bucket_count":0,"peering_crush_bucket_target":0,"peering_crush_bucket_barrier":0,"peering_crush_bucket_mandatory_member":2147483647,"is_stretch_pool":false,"object_hash":2,"pg_autoscale_mode":"off","pg_num":1,"pg_placement_num":1,"pg_placement_num_target":1,"pg_num_target":1,"pg_num_pending":1,"last_pg_merge_meta":{"source_pgid":"0.0","ready_epoch":0,"last_epoch_started":0,"last_epoch_clean":0,"source_version":"0'0","target_version":"0'0"},"last_change":"19","last_force_op_resend":"0","last_force_op_resend_prenautilus":"0","last_force_op_resend_preluminous":"0","auid":0,"snap_mode":"selfmanaged","snap_seq":0,"snap_epoch":0,"pool_snaps":[],"removed_snaps":"[]","quota_max_bytes":0,"quota_max_objects":0,"tiers":[],"tier_of":-1,"read_tier":-1,"write_tier":-1,"cache_mode":"none","target_max_bytes":0,"target_max_objects":0,"cache_target_dirty_ratio_micro":400000,"cache_target_dirty_high_ratio_micro":600000,"cache_target_full_ratio_micro":800000,"cache_min_flush_age":0,"cache_min_evict_age":0,"erasure_code_profile":"","hit_set_params":{"type":"none"},"hit_set_period":0,"hit_set_count":0,"use_gmt_hitset":true,"min_read_recency_for_promote":0,"min_write_recency_for_promote":0,"hit_set_grade_decay_rate":0,"hit_set_search_last_n":0,"grade_table":[],"stripe_width":0,"expected_num_objects":0,"fast_read":false,"options":{"pg_num_max":32,"pg_num_min":1},"application_metadata":{"mgr":{}},"read_balance":{"score_type":"Fair distribution","score_acting":3,"score_stable":3,"optimal_score":1,"raw_score_acting":3,"raw_score_stable":3,"primary_affinity_weighted":1,"average_primary_affinity":1,"average_primary_affinity_weighted":1}}],"osds":[{"osd":0,"uuid":"1d65b890-7ef0-46fe-a6c2-e544a1692d70","up":1,"in":1,"weight":1,"primary_affinity":1,"last_clean_begin":0,"last_clean_end":0,"up_from":8,"up_thru":0,"down_at":0,"lost_at":0,"public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.100:6802","nonce":2576257746},{"type":"v1","addr":"192.168.123.100:6803","nonce":2576257746}]},"cluster_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.100:6804","nonce":2576257746},{"type":"v1","addr":"192.168.123.100:6805","nonce":2576257746}]},"heartbeat_back_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.100:6808","nonce":2576257746},{"type":"v1","addr":"192.168.123.100:6809","nonce":2576257746}]},"heartbeat_front_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.100:6806","nonce":2576257746},{"type":"v1","addr":"192.168.123.100:6807","nonce":2576257746}]},"public_addr":"192.168.123.100:6803/2576257746","cluster_addr":"192.168.123.100:6805/2576257746","heartbeat_back_addr":"192.168.123.100:6809/2576257746","heartbeat_front_addr":"192.168.123.100:6807/2576257746","state":["exists","up"]},{"osd":1,"uuid":"95d964b0-4ea2-45f5-a612-b4aa5749cec2","up":1,"in":1,"weight":1,"primary_affinity":1,"last_clean_begin":0,"last_clean_end":0,"up_from":12,"up_thru":17,"down_at":0,"lost_at":0,"public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.101:6800","nonce":1940387388},{"type":"v1","addr":"192.168.123.101:6801","nonce":1940387388}]},"cluster_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.101:6802","nonce":1940387388},{"type":"v1","addr":"192.168.123.101:6803","nonce":1940387388}]},"heartbeat_back_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.101:6806","nonce":1940387388},{"type":"v1","addr":"192.168.123.101:6807","nonce":1940387388}]},"heartbeat_front_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.101:6804","nonce":1940387388},{"type":"v1","addr":"192.168.123.101:6805","nonce":1940387388}]},"public_addr":"192.168.123.101:6801/1940387388","cluster_addr":"192.168.123.101:6803/1940387388","heartbeat_back_addr":"192.168.123.101:6807/1940387388","heartbeat_front_addr":"192.168.123.101:6805/1940387388","state":["exists","up"]},{"osd":2,"uuid":"9e99c305-cccd-4040-a383-099c1410ee92","up":1,"in":1,"weight":1,"primary_affinity":1,"last_clean_begin":0,"last_clean_end":0,"up_from":16,"up_thru":0,"down_at":0,"lost_at":0,"public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.108:6800","nonce":4086912105},{"type":"v1","addr":"192.168.123.108:6801","nonce":4086912105}]},"cluster_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.108:6802","nonce":4086912105},{"type":"v1","addr":"192.168.123.108:6803","nonce":4086912105}]},"heartbeat_back_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.108:6806","nonce":4086912105},{"type":"v1","addr":"192.168.123.108:6807","nonce":4086912105}]},"heartbeat_front_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.108:6804","nonce":4086912105},{"type":"v1","addr":"192.168.123.108:6805","nonce":4086912105}]},"public_addr":"192.168.123.108:6801/4086912105","cluster_addr":"192.168.123.108:6803/4086912105","heartbeat_back_addr":"192.168.123.108:6807/4086912105","heartbeat_front_addr":"192.168.123.108:6805/4086912105","state":["exists","up"]}],"osd_xinfo":[{"osd":0,"down_stamp":"0.000000","laggy_probability":0,"laggy_interval":0,"features":4540701547738038271,"old_weight":0,"last_purged_snaps_scrub":"2026-03-06T12:41:39.587816+0000","dead_epoch":0},{"osd":1,"down_stamp":"0.000000","laggy_probability":0,"laggy_interval":0,"features":4540701547738038271,"old_weight":0,"last_purged_snaps_scrub":"2026-03-06T12:41:52.938720+0000","dead_epoch":0},{"osd":2,"down_stamp":"0.000000","laggy_probability":0,"laggy_interval":0,"features":4540701547738038271,"old_weight":0,"last_purged_snaps_scrub":"2026-03-06T12:42:07.364233+0000","dead_epoch":0}],"pg_upmap":[],"pg_upmap_items":[],"pg_upmap_primaries":[],"pg_temp":[],"primary_temp":[],"blocklist":{"192.168.123.100:6801/2414933134":"2026-03-07T12:40:47.966309+0000","192.168.123.100:0/310500200":"2026-03-07T12:40:47.966309+0000","192.168.123.100:0/2258966804":"2026-03-07T12:40:47.966309+0000","192.168.123.100:0/1720822170":"2026-03-07T12:40:47.966309+0000","192.168.123.100:0/3743744839":"2026-03-07T12:40:28.137920+0000","192.168.123.100:0/3253599029":"2026-03-07T12:40:28.137920+0000","192.168.123.100:6801/3456934444":"2026-03-07T12:40:28.137920+0000","192.168.123.100:6800/2414933134":"2026-03-07T12:40:47.966309+0000","192.168.123.100:6800/3456934444":"2026-03-07T12:40:28.137920+0000","192.168.123.100:0/33831712":"2026-03-07T12:40:28.137920+0000"},"range_blocklist":{},"erasure_code_profiles":{"default":{"crush-failure-domain":"osd","k":"2","m":"1","plugin":"jerasure","technique":"reed_sol_van"}},"removed_snaps_queue":[],"new_removed_snaps":[],"new_purged_snaps":[],"crush_node_flags":{},"device_class_flags":{},"stretch_mode":{"stretch_mode_enabled":false,"stretch_bucket_count":0,"degraded_stretch_mode":0,"recovering_stretch_mode":0,"stretch_mode_bucket":0}} 2026-03-06T13:42:13.772 INFO:tasks.cephadm.ceph_manager.ceph:all up! 2026-03-06T13:42:13.772 DEBUG:teuthology.orchestra.run.vm00:> sudo /home/ubuntu/cephtest/cephadm --image harbor.clyso.com/custom-ceph/ceph/ceph:cobaltcore-storage-v19.2.3-fasttrack-3 shell --fsid 82f504c8-1959-11f1-9a16-a78d135e200b -- ceph osd dump --format=json 2026-03-06T13:42:14.144 INFO:teuthology.orchestra.run.vm00.stderr:Inferring config /var/lib/ceph/82f504c8-1959-11f1-9a16-a78d135e200b/mon.a/config 2026-03-06T13:42:14.405 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:42:14 vm01 ceph-mon[56331]: from='client.? 192.168.123.100:0/565045641' entity='client.admin' cmd=[{"prefix": "osd dump", "format": "json"}]: dispatch 2026-03-06T13:42:14.405 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:42:14 vm01 ceph-mon[56331]: pgmap v52: 1 pgs: 1 creating+peering; 0 B data, 80 MiB used, 60 GiB / 60 GiB avail 2026-03-06T13:42:14.503 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:42:14 vm00 ceph-mon[51914]: from='client.? 192.168.123.100:0/565045641' entity='client.admin' cmd=[{"prefix": "osd dump", "format": "json"}]: dispatch 2026-03-06T13:42:14.503 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:42:14 vm00 ceph-mon[51914]: pgmap v52: 1 pgs: 1 creating+peering; 0 B data, 80 MiB used, 60 GiB / 60 GiB avail 2026-03-06T13:42:14.540 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 06 13:42:14 vm08 ceph-mon[53561]: from='client.? 192.168.123.100:0/565045641' entity='client.admin' cmd=[{"prefix": "osd dump", "format": "json"}]: dispatch 2026-03-06T13:42:14.540 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 06 13:42:14 vm08 ceph-mon[53561]: pgmap v52: 1 pgs: 1 creating+peering; 0 B data, 80 MiB used, 60 GiB / 60 GiB avail 2026-03-06T13:42:14.565 INFO:teuthology.orchestra.run.vm00.stdout: 2026-03-06T13:42:14.565 INFO:teuthology.orchestra.run.vm00.stdout:{"epoch":19,"fsid":"82f504c8-1959-11f1-9a16-a78d135e200b","created":"2026-03-06T12:40:03.178129+0000","modified":"2026-03-06T12:42:12.113876+0000","last_up_change":"2026-03-06T12:42:09.091745+0000","last_in_change":"2026-03-06T12:41:55.730672+0000","flags":"sortbitwise,recovery_deletes,purged_snapdirs,pglog_hardlimit","flags_num":5799936,"flags_set":["pglog_hardlimit","purged_snapdirs","recovery_deletes","sortbitwise"],"crush_version":8,"full_ratio":0.94999998807907104,"backfillfull_ratio":0.89999997615814209,"nearfull_ratio":0.85000002384185791,"cluster_snapshot":"","pool_max":1,"max_osd":3,"require_min_compat_client":"luminous","min_compat_client":"jewel","require_osd_release":"squid","allow_crimson":false,"pools":[{"pool":1,"pool_name":".mgr","create_time":"2026-03-06T12:42:10.032248+0000","flags":1,"flags_names":"hashpspool","type":1,"size":3,"min_size":2,"crush_rule":0,"peering_crush_bucket_count":0,"peering_crush_bucket_target":0,"peering_crush_bucket_barrier":0,"peering_crush_bucket_mandatory_member":2147483647,"is_stretch_pool":false,"object_hash":2,"pg_autoscale_mode":"off","pg_num":1,"pg_placement_num":1,"pg_placement_num_target":1,"pg_num_target":1,"pg_num_pending":1,"last_pg_merge_meta":{"source_pgid":"0.0","ready_epoch":0,"last_epoch_started":0,"last_epoch_clean":0,"source_version":"0'0","target_version":"0'0"},"last_change":"19","last_force_op_resend":"0","last_force_op_resend_prenautilus":"0","last_force_op_resend_preluminous":"0","auid":0,"snap_mode":"selfmanaged","snap_seq":0,"snap_epoch":0,"pool_snaps":[],"removed_snaps":"[]","quota_max_bytes":0,"quota_max_objects":0,"tiers":[],"tier_of":-1,"read_tier":-1,"write_tier":-1,"cache_mode":"none","target_max_bytes":0,"target_max_objects":0,"cache_target_dirty_ratio_micro":400000,"cache_target_dirty_high_ratio_micro":600000,"cache_target_full_ratio_micro":800000,"cache_min_flush_age":0,"cache_min_evict_age":0,"erasure_code_profile":"","hit_set_params":{"type":"none"},"hit_set_period":0,"hit_set_count":0,"use_gmt_hitset":true,"min_read_recency_for_promote":0,"min_write_recency_for_promote":0,"hit_set_grade_decay_rate":0,"hit_set_search_last_n":0,"grade_table":[],"stripe_width":0,"expected_num_objects":0,"fast_read":false,"options":{"pg_num_max":32,"pg_num_min":1},"application_metadata":{"mgr":{}},"read_balance":{"score_type":"Fair distribution","score_acting":3,"score_stable":3,"optimal_score":1,"raw_score_acting":3,"raw_score_stable":3,"primary_affinity_weighted":1,"average_primary_affinity":1,"average_primary_affinity_weighted":1}}],"osds":[{"osd":0,"uuid":"1d65b890-7ef0-46fe-a6c2-e544a1692d70","up":1,"in":1,"weight":1,"primary_affinity":1,"last_clean_begin":0,"last_clean_end":0,"up_from":8,"up_thru":0,"down_at":0,"lost_at":0,"public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.100:6802","nonce":2576257746},{"type":"v1","addr":"192.168.123.100:6803","nonce":2576257746}]},"cluster_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.100:6804","nonce":2576257746},{"type":"v1","addr":"192.168.123.100:6805","nonce":2576257746}]},"heartbeat_back_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.100:6808","nonce":2576257746},{"type":"v1","addr":"192.168.123.100:6809","nonce":2576257746}]},"heartbeat_front_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.100:6806","nonce":2576257746},{"type":"v1","addr":"192.168.123.100:6807","nonce":2576257746}]},"public_addr":"192.168.123.100:6803/2576257746","cluster_addr":"192.168.123.100:6805/2576257746","heartbeat_back_addr":"192.168.123.100:6809/2576257746","heartbeat_front_addr":"192.168.123.100:6807/2576257746","state":["exists","up"]},{"osd":1,"uuid":"95d964b0-4ea2-45f5-a612-b4aa5749cec2","up":1,"in":1,"weight":1,"primary_affinity":1,"last_clean_begin":0,"last_clean_end":0,"up_from":12,"up_thru":17,"down_at":0,"lost_at":0,"public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.101:6800","nonce":1940387388},{"type":"v1","addr":"192.168.123.101:6801","nonce":1940387388}]},"cluster_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.101:6802","nonce":1940387388},{"type":"v1","addr":"192.168.123.101:6803","nonce":1940387388}]},"heartbeat_back_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.101:6806","nonce":1940387388},{"type":"v1","addr":"192.168.123.101:6807","nonce":1940387388}]},"heartbeat_front_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.101:6804","nonce":1940387388},{"type":"v1","addr":"192.168.123.101:6805","nonce":1940387388}]},"public_addr":"192.168.123.101:6801/1940387388","cluster_addr":"192.168.123.101:6803/1940387388","heartbeat_back_addr":"192.168.123.101:6807/1940387388","heartbeat_front_addr":"192.168.123.101:6805/1940387388","state":["exists","up"]},{"osd":2,"uuid":"9e99c305-cccd-4040-a383-099c1410ee92","up":1,"in":1,"weight":1,"primary_affinity":1,"last_clean_begin":0,"last_clean_end":0,"up_from":16,"up_thru":0,"down_at":0,"lost_at":0,"public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.108:6800","nonce":4086912105},{"type":"v1","addr":"192.168.123.108:6801","nonce":4086912105}]},"cluster_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.108:6802","nonce":4086912105},{"type":"v1","addr":"192.168.123.108:6803","nonce":4086912105}]},"heartbeat_back_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.108:6806","nonce":4086912105},{"type":"v1","addr":"192.168.123.108:6807","nonce":4086912105}]},"heartbeat_front_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.108:6804","nonce":4086912105},{"type":"v1","addr":"192.168.123.108:6805","nonce":4086912105}]},"public_addr":"192.168.123.108:6801/4086912105","cluster_addr":"192.168.123.108:6803/4086912105","heartbeat_back_addr":"192.168.123.108:6807/4086912105","heartbeat_front_addr":"192.168.123.108:6805/4086912105","state":["exists","up"]}],"osd_xinfo":[{"osd":0,"down_stamp":"0.000000","laggy_probability":0,"laggy_interval":0,"features":4540701547738038271,"old_weight":0,"last_purged_snaps_scrub":"2026-03-06T12:41:39.587816+0000","dead_epoch":0},{"osd":1,"down_stamp":"0.000000","laggy_probability":0,"laggy_interval":0,"features":4540701547738038271,"old_weight":0,"last_purged_snaps_scrub":"2026-03-06T12:41:52.938720+0000","dead_epoch":0},{"osd":2,"down_stamp":"0.000000","laggy_probability":0,"laggy_interval":0,"features":4540701547738038271,"old_weight":0,"last_purged_snaps_scrub":"2026-03-06T12:42:07.364233+0000","dead_epoch":0}],"pg_upmap":[],"pg_upmap_items":[],"pg_upmap_primaries":[],"pg_temp":[],"primary_temp":[],"blocklist":{"192.168.123.100:6801/2414933134":"2026-03-07T12:40:47.966309+0000","192.168.123.100:0/310500200":"2026-03-07T12:40:47.966309+0000","192.168.123.100:0/2258966804":"2026-03-07T12:40:47.966309+0000","192.168.123.100:0/1720822170":"2026-03-07T12:40:47.966309+0000","192.168.123.100:0/3743744839":"2026-03-07T12:40:28.137920+0000","192.168.123.100:0/3253599029":"2026-03-07T12:40:28.137920+0000","192.168.123.100:6801/3456934444":"2026-03-07T12:40:28.137920+0000","192.168.123.100:6800/2414933134":"2026-03-07T12:40:47.966309+0000","192.168.123.100:6800/3456934444":"2026-03-07T12:40:28.137920+0000","192.168.123.100:0/33831712":"2026-03-07T12:40:28.137920+0000"},"range_blocklist":{},"erasure_code_profiles":{"default":{"crush-failure-domain":"osd","k":"2","m":"1","plugin":"jerasure","technique":"reed_sol_van"}},"removed_snaps_queue":[],"new_removed_snaps":[],"new_purged_snaps":[],"crush_node_flags":{},"device_class_flags":{},"stretch_mode":{"stretch_mode_enabled":false,"stretch_bucket_count":0,"degraded_stretch_mode":0,"recovering_stretch_mode":0,"stretch_mode_bucket":0}} 2026-03-06T13:42:14.644 DEBUG:teuthology.orchestra.run.vm00:> sudo /home/ubuntu/cephtest/cephadm --image harbor.clyso.com/custom-ceph/ceph/ceph:cobaltcore-storage-v19.2.3-fasttrack-3 shell --fsid 82f504c8-1959-11f1-9a16-a78d135e200b -- ceph tell osd.0 flush_pg_stats 2026-03-06T13:42:14.644 DEBUG:teuthology.orchestra.run.vm00:> sudo /home/ubuntu/cephtest/cephadm --image harbor.clyso.com/custom-ceph/ceph/ceph:cobaltcore-storage-v19.2.3-fasttrack-3 shell --fsid 82f504c8-1959-11f1-9a16-a78d135e200b -- ceph tell osd.1 flush_pg_stats 2026-03-06T13:42:14.644 DEBUG:teuthology.orchestra.run.vm00:> sudo /home/ubuntu/cephtest/cephadm --image harbor.clyso.com/custom-ceph/ceph/ceph:cobaltcore-storage-v19.2.3-fasttrack-3 shell --fsid 82f504c8-1959-11f1-9a16-a78d135e200b -- ceph tell osd.2 flush_pg_stats 2026-03-06T13:42:15.191 INFO:teuthology.orchestra.run.vm00.stderr:Inferring config /var/lib/ceph/82f504c8-1959-11f1-9a16-a78d135e200b/mon.a/config 2026-03-06T13:42:15.191 INFO:teuthology.orchestra.run.vm00.stderr:Inferring config /var/lib/ceph/82f504c8-1959-11f1-9a16-a78d135e200b/mon.a/config 2026-03-06T13:42:15.325 INFO:teuthology.orchestra.run.vm00.stderr:Inferring config /var/lib/ceph/82f504c8-1959-11f1-9a16-a78d135e200b/mon.a/config 2026-03-06T13:42:15.392 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:42:15 vm00 ceph-mon[51914]: from='client.? 192.168.123.100:0/2150538969' entity='client.admin' cmd=[{"prefix": "osd dump", "format": "json"}]: dispatch 2026-03-06T13:42:15.405 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:42:15 vm01 ceph-mon[56331]: from='client.? 192.168.123.100:0/2150538969' entity='client.admin' cmd=[{"prefix": "osd dump", "format": "json"}]: dispatch 2026-03-06T13:42:15.540 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 06 13:42:15 vm08 ceph-mon[53561]: from='client.? 192.168.123.100:0/2150538969' entity='client.admin' cmd=[{"prefix": "osd dump", "format": "json"}]: dispatch 2026-03-06T13:42:15.664 INFO:teuthology.orchestra.run.vm00.stdout:68719476739 2026-03-06T13:42:15.664 DEBUG:teuthology.orchestra.run.vm00:> sudo /home/ubuntu/cephtest/cephadm --image harbor.clyso.com/custom-ceph/ceph/ceph:cobaltcore-storage-v19.2.3-fasttrack-3 shell --fsid 82f504c8-1959-11f1-9a16-a78d135e200b -- ceph osd last-stat-seq osd.2 2026-03-06T13:42:15.783 INFO:teuthology.orchestra.run.vm00.stdout:34359738376 2026-03-06T13:42:15.783 DEBUG:teuthology.orchestra.run.vm00:> sudo /home/ubuntu/cephtest/cephadm --image harbor.clyso.com/custom-ceph/ceph/ceph:cobaltcore-storage-v19.2.3-fasttrack-3 shell --fsid 82f504c8-1959-11f1-9a16-a78d135e200b -- ceph osd last-stat-seq osd.0 2026-03-06T13:42:15.947 INFO:teuthology.orchestra.run.vm00.stdout:51539607557 2026-03-06T13:42:15.948 DEBUG:teuthology.orchestra.run.vm00:> sudo /home/ubuntu/cephtest/cephadm --image harbor.clyso.com/custom-ceph/ceph/ceph:cobaltcore-storage-v19.2.3-fasttrack-3 shell --fsid 82f504c8-1959-11f1-9a16-a78d135e200b -- ceph osd last-stat-seq osd.1 2026-03-06T13:42:16.151 INFO:teuthology.orchestra.run.vm00.stderr:Inferring config /var/lib/ceph/82f504c8-1959-11f1-9a16-a78d135e200b/mon.a/config 2026-03-06T13:42:16.238 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:42:16 vm00 ceph-mon[51914]: pgmap v53: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-06T13:42:16.405 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:42:16 vm01 ceph-mon[56331]: pgmap v53: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-06T13:42:16.500 INFO:teuthology.orchestra.run.vm00.stderr:Inferring config /var/lib/ceph/82f504c8-1959-11f1-9a16-a78d135e200b/mon.a/config 2026-03-06T13:42:16.540 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 06 13:42:16 vm08 ceph-mon[53561]: pgmap v53: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-06T13:42:16.625 INFO:teuthology.orchestra.run.vm00.stderr:Inferring config /var/lib/ceph/82f504c8-1959-11f1-9a16-a78d135e200b/mon.a/config 2026-03-06T13:42:16.805 INFO:teuthology.orchestra.run.vm00.stdout:68719476739 2026-03-06T13:42:16.873 INFO:tasks.cephadm.ceph_manager.ceph:need seq 68719476739 got 68719476739 for osd.2 2026-03-06T13:42:16.874 DEBUG:teuthology.parallel:result is None 2026-03-06T13:42:17.001 INFO:teuthology.orchestra.run.vm00.stdout:34359738376 2026-03-06T13:42:17.092 INFO:teuthology.orchestra.run.vm00.stdout:51539607558 2026-03-06T13:42:17.100 INFO:tasks.cephadm.ceph_manager.ceph:need seq 34359738376 got 34359738376 for osd.0 2026-03-06T13:42:17.100 DEBUG:teuthology.parallel:result is None 2026-03-06T13:42:17.163 INFO:tasks.cephadm.ceph_manager.ceph:need seq 51539607557 got 51539607558 for osd.1 2026-03-06T13:42:17.164 DEBUG:teuthology.parallel:result is None 2026-03-06T13:42:17.164 INFO:tasks.cephadm.ceph_manager.ceph:waiting for clean 2026-03-06T13:42:17.164 DEBUG:teuthology.orchestra.run.vm00:> sudo /home/ubuntu/cephtest/cephadm --image harbor.clyso.com/custom-ceph/ceph/ceph:cobaltcore-storage-v19.2.3-fasttrack-3 shell --fsid 82f504c8-1959-11f1-9a16-a78d135e200b -- ceph pg dump --format=json 2026-03-06T13:42:17.405 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:42:17 vm01 ceph-mon[56331]: from='client.? 192.168.123.100:0/1679205646' entity='client.admin' cmd=[{"prefix": "osd last-stat-seq", "id": 2}]: dispatch 2026-03-06T13:42:17.405 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:42:17 vm01 ceph-mon[56331]: from='client.? 192.168.123.100:0/3434377464' entity='client.admin' cmd=[{"prefix": "osd last-stat-seq", "id": 0}]: dispatch 2026-03-06T13:42:17.405 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:42:17 vm01 ceph-mon[56331]: from='client.? 192.168.123.100:0/2158854197' entity='client.admin' cmd=[{"prefix": "osd last-stat-seq", "id": 1}]: dispatch 2026-03-06T13:42:17.493 INFO:teuthology.orchestra.run.vm00.stderr:Inferring config /var/lib/ceph/82f504c8-1959-11f1-9a16-a78d135e200b/mon.a/config 2026-03-06T13:42:17.502 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:42:17 vm00 ceph-mon[51914]: from='client.? 192.168.123.100:0/1679205646' entity='client.admin' cmd=[{"prefix": "osd last-stat-seq", "id": 2}]: dispatch 2026-03-06T13:42:17.502 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:42:17 vm00 ceph-mon[51914]: from='client.? 192.168.123.100:0/3434377464' entity='client.admin' cmd=[{"prefix": "osd last-stat-seq", "id": 0}]: dispatch 2026-03-06T13:42:17.502 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:42:17 vm00 ceph-mon[51914]: from='client.? 192.168.123.100:0/2158854197' entity='client.admin' cmd=[{"prefix": "osd last-stat-seq", "id": 1}]: dispatch 2026-03-06T13:42:17.540 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 06 13:42:17 vm08 ceph-mon[53561]: from='client.? 192.168.123.100:0/1679205646' entity='client.admin' cmd=[{"prefix": "osd last-stat-seq", "id": 2}]: dispatch 2026-03-06T13:42:17.540 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 06 13:42:17 vm08 ceph-mon[53561]: from='client.? 192.168.123.100:0/3434377464' entity='client.admin' cmd=[{"prefix": "osd last-stat-seq", "id": 0}]: dispatch 2026-03-06T13:42:17.540 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 06 13:42:17 vm08 ceph-mon[53561]: from='client.? 192.168.123.100:0/2158854197' entity='client.admin' cmd=[{"prefix": "osd last-stat-seq", "id": 1}]: dispatch 2026-03-06T13:42:17.846 INFO:teuthology.orchestra.run.vm00.stdout: 2026-03-06T13:42:17.846 INFO:teuthology.orchestra.run.vm00.stderr:dumped all 2026-03-06T13:42:17.920 INFO:teuthology.orchestra.run.vm00.stdout:{"pg_ready":true,"pg_map":{"version":53,"stamp":"2026-03-06T12:42:15.988270+0000","last_osdmap_epoch":0,"last_pg_scan":0,"pg_stats_sum":{"stat_sum":{"num_bytes":459280,"num_objects":2,"num_object_clones":0,"num_object_copies":6,"num_objects_missing_on_primary":0,"num_objects_missing":0,"num_objects_degraded":0,"num_objects_misplaced":0,"num_objects_unfound":0,"num_objects_dirty":2,"num_whiteouts":0,"num_read":46,"num_read_kb":37,"num_write":57,"num_write_kb":584,"num_scrub_errors":0,"num_shallow_scrub_errors":0,"num_deep_scrub_errors":0,"num_objects_recovered":0,"num_bytes_recovered":0,"num_keys_recovered":0,"num_objects_omap":0,"num_objects_hit_set_archive":0,"num_bytes_hit_set_archive":0,"num_flush":0,"num_flush_kb":0,"num_evict":0,"num_evict_kb":0,"num_promote":0,"num_flush_mode_high":0,"num_flush_mode_low":0,"num_evict_mode_some":0,"num_evict_mode_full":0,"num_objects_pinned":0,"num_legacy_snapsets":0,"num_large_omap_objects":0,"num_objects_manifest":0,"num_omap_bytes":0,"num_omap_keys":0,"num_objects_repaired":0},"store_stats":{"total":0,"available":0,"internally_reserved":0,"allocated":0,"data_stored":0,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":0},"log_size":32,"ondisk_log_size":32,"up":3,"acting":3,"num_store_stats":0},"osd_stats_sum":{"up_from":0,"seq":0,"num_pgs":3,"num_osds":3,"num_per_pool_osds":3,"num_per_pool_omap_osds":3,"kb":62902272,"kb_used":82756,"kb_used_data":1836,"kb_used_omap":4,"kb_used_meta":80443,"kb_avail":62819516,"statfs":{"total":64411926528,"available":64327184384,"internally_reserved":0,"allocated":1880064,"data_stored":1528179,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":4770,"internal_metadata":82373982},"hb_peers":[],"snap_trim_queue_len":0,"num_snap_trimming":0,"num_shards_repaired":0,"op_queue_age_hist":{"histogram":[],"upper_bound":1},"perf_stat":{"commit_latency_ms":0,"apply_latency_ms":0,"commit_latency_ns":0,"apply_latency_ns":0},"alerts":[],"network_ping_times":[]},"pg_stats_delta":{"stat_sum":{"num_bytes":0,"num_objects":0,"num_object_clones":0,"num_object_copies":0,"num_objects_missing_on_primary":0,"num_objects_missing":0,"num_objects_degraded":0,"num_objects_misplaced":0,"num_objects_unfound":0,"num_objects_dirty":0,"num_whiteouts":0,"num_read":0,"num_read_kb":0,"num_write":0,"num_write_kb":0,"num_scrub_errors":0,"num_shallow_scrub_errors":0,"num_deep_scrub_errors":0,"num_objects_recovered":0,"num_bytes_recovered":0,"num_keys_recovered":0,"num_objects_omap":0,"num_objects_hit_set_archive":0,"num_bytes_hit_set_archive":0,"num_flush":0,"num_flush_kb":0,"num_evict":0,"num_evict_kb":0,"num_promote":0,"num_flush_mode_high":0,"num_flush_mode_low":0,"num_evict_mode_some":0,"num_evict_mode_full":0,"num_objects_pinned":0,"num_legacy_snapsets":0,"num_large_omap_objects":0,"num_objects_manifest":0,"num_omap_bytes":0,"num_omap_keys":0,"num_objects_repaired":0},"store_stats":{"total":0,"available":0,"internally_reserved":0,"allocated":0,"data_stored":0,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":0},"log_size":0,"ondisk_log_size":0,"up":0,"acting":0,"num_store_stats":0,"stamp_delta":"0.000000"},"pg_stats":[{"pgid":"1.0","version":"18'32","reported_seq":57,"reported_epoch":19,"state":"active+clean","last_fresh":"2026-03-06T12:42:12.126398+0000","last_change":"2026-03-06T12:42:11.110567+0000","last_active":"2026-03-06T12:42:12.126398+0000","last_peered":"2026-03-06T12:42:12.126398+0000","last_clean":"2026-03-06T12:42:12.126398+0000","last_became_active":"2026-03-06T12:42:11.110448+0000","last_became_peered":"2026-03-06T12:42:11.110448+0000","last_unstale":"2026-03-06T12:42:12.126398+0000","last_undegraded":"2026-03-06T12:42:12.126398+0000","last_fullsized":"2026-03-06T12:42:12.126398+0000","mapping_epoch":17,"log_start":"0'0","ondisk_log_start":"0'0","created":17,"last_epoch_clean":18,"parent":"0.0","parent_split_bits":0,"last_scrub":"0'0","last_scrub_stamp":"2026-03-06T12:42:10.100211+0000","last_deep_scrub":"0'0","last_deep_scrub_stamp":"2026-03-06T12:42:10.100211+0000","last_clean_scrub_stamp":"2026-03-06T12:42:10.100211+0000","objects_scrubbed":0,"log_size":32,"log_dups_size":0,"ondisk_log_size":32,"stats_invalid":false,"dirty_stats_invalid":false,"omap_stats_invalid":false,"hitset_stats_invalid":false,"hitset_bytes_stats_invalid":false,"pin_stats_invalid":false,"manifest_stats_invalid":false,"snaptrimq_len":0,"last_scrub_duration":0,"scrub_schedule":"periodic scrub scheduled @ 2026-03-07T16:31:21.552364+0000","scrub_duration":0,"objects_trimmed":0,"snaptrim_duration":0,"stat_sum":{"num_bytes":459280,"num_objects":2,"num_object_clones":0,"num_object_copies":6,"num_objects_missing_on_primary":0,"num_objects_missing":0,"num_objects_degraded":0,"num_objects_misplaced":0,"num_objects_unfound":0,"num_objects_dirty":2,"num_whiteouts":0,"num_read":46,"num_read_kb":37,"num_write":57,"num_write_kb":584,"num_scrub_errors":0,"num_shallow_scrub_errors":0,"num_deep_scrub_errors":0,"num_objects_recovered":0,"num_bytes_recovered":0,"num_keys_recovered":0,"num_objects_omap":0,"num_objects_hit_set_archive":0,"num_bytes_hit_set_archive":0,"num_flush":0,"num_flush_kb":0,"num_evict":0,"num_evict_kb":0,"num_promote":0,"num_flush_mode_high":0,"num_flush_mode_low":0,"num_evict_mode_some":0,"num_evict_mode_full":0,"num_objects_pinned":0,"num_legacy_snapsets":0,"num_large_omap_objects":0,"num_objects_manifest":0,"num_omap_bytes":0,"num_omap_keys":0,"num_objects_repaired":0},"up":[1,2,0],"acting":[1,2,0],"avail_no_missing":[],"object_location_counts":[],"blocked_by":[],"up_primary":1,"acting_primary":1,"purged_snaps":[]}],"pool_stats":[{"poolid":1,"num_pg":1,"stat_sum":{"num_bytes":459280,"num_objects":2,"num_object_clones":0,"num_object_copies":6,"num_objects_missing_on_primary":0,"num_objects_missing":0,"num_objects_degraded":0,"num_objects_misplaced":0,"num_objects_unfound":0,"num_objects_dirty":2,"num_whiteouts":0,"num_read":46,"num_read_kb":37,"num_write":57,"num_write_kb":584,"num_scrub_errors":0,"num_shallow_scrub_errors":0,"num_deep_scrub_errors":0,"num_objects_recovered":0,"num_bytes_recovered":0,"num_keys_recovered":0,"num_objects_omap":0,"num_objects_hit_set_archive":0,"num_bytes_hit_set_archive":0,"num_flush":0,"num_flush_kb":0,"num_evict":0,"num_evict_kb":0,"num_promote":0,"num_flush_mode_high":0,"num_flush_mode_low":0,"num_evict_mode_some":0,"num_evict_mode_full":0,"num_objects_pinned":0,"num_legacy_snapsets":0,"num_large_omap_objects":0,"num_objects_manifest":0,"num_omap_bytes":0,"num_omap_keys":0,"num_objects_repaired":0},"store_stats":{"total":0,"available":0,"internally_reserved":0,"allocated":1388544,"data_stored":1377840,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":0},"log_size":32,"ondisk_log_size":32,"up":3,"acting":3,"num_store_stats":3}],"osd_stats":[{"osd":2,"up_from":16,"seq":68719476739,"num_pgs":1,"num_osds":1,"num_per_pool_osds":1,"num_per_pool_omap_osds":1,"kb":20967424,"kb_used":27584,"kb_used_data":612,"kb_used_omap":1,"kb_used_meta":26814,"kb_avail":20939840,"statfs":{"total":21470642176,"available":21442396160,"internally_reserved":0,"allocated":626688,"data_stored":509393,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":1590,"internal_metadata":27457994},"hb_peers":[0,1],"snap_trim_queue_len":0,"num_snap_trimming":0,"num_shards_repaired":0,"op_queue_age_hist":{"histogram":[],"upper_bound":1},"perf_stat":{"commit_latency_ms":0,"apply_latency_ms":0,"commit_latency_ns":0,"apply_latency_ns":0},"alerts":[]},{"osd":1,"up_from":12,"seq":51539607558,"num_pgs":1,"num_osds":1,"num_per_pool_osds":1,"num_per_pool_omap_osds":1,"kb":20967424,"kb_used":27584,"kb_used_data":612,"kb_used_omap":1,"kb_used_meta":26814,"kb_avail":20939840,"statfs":{"total":21470642176,"available":21442396160,"internally_reserved":0,"allocated":626688,"data_stored":509393,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":1590,"internal_metadata":27457994},"hb_peers":[0,2],"snap_trim_queue_len":0,"num_snap_trimming":0,"num_shards_repaired":0,"op_queue_age_hist":{"histogram":[],"upper_bound":1},"perf_stat":{"commit_latency_ms":0,"apply_latency_ms":0,"commit_latency_ns":0,"apply_latency_ns":0},"alerts":[]},{"osd":0,"up_from":8,"seq":34359738376,"num_pgs":1,"num_osds":1,"num_per_pool_osds":1,"num_per_pool_omap_osds":1,"kb":20967424,"kb_used":27588,"kb_used_data":612,"kb_used_omap":1,"kb_used_meta":26814,"kb_avail":20939836,"statfs":{"total":21470642176,"available":21442392064,"internally_reserved":0,"allocated":626688,"data_stored":509393,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":1590,"internal_metadata":27457994},"hb_peers":[1,2],"snap_trim_queue_len":0,"num_snap_trimming":0,"num_shards_repaired":0,"op_queue_age_hist":{"histogram":[],"upper_bound":1},"perf_stat":{"commit_latency_ms":0,"apply_latency_ms":0,"commit_latency_ns":0,"apply_latency_ns":0},"alerts":[]}],"pool_statfs":[{"poolid":1,"osd":0,"total":0,"available":0,"internally_reserved":0,"allocated":462848,"data_stored":459280,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":0},{"poolid":1,"osd":1,"total":0,"available":0,"internally_reserved":0,"allocated":462848,"data_stored":459280,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":0},{"poolid":1,"osd":2,"total":0,"available":0,"internally_reserved":0,"allocated":462848,"data_stored":459280,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":0}]}} 2026-03-06T13:42:17.921 DEBUG:teuthology.orchestra.run.vm00:> sudo /home/ubuntu/cephtest/cephadm --image harbor.clyso.com/custom-ceph/ceph/ceph:cobaltcore-storage-v19.2.3-fasttrack-3 shell --fsid 82f504c8-1959-11f1-9a16-a78d135e200b -- ceph pg dump --format=json 2026-03-06T13:42:18.162 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:42:18 vm00 ceph-mon[51914]: from='client.14373 -' entity='client.admin' cmd=[{"prefix": "pg dump", "target": ["mon-mgr", ""], "format": "json"}]: dispatch 2026-03-06T13:42:18.162 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:42:18 vm00 ceph-mon[51914]: pgmap v54: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-06T13:42:18.275 INFO:teuthology.orchestra.run.vm00.stderr:Inferring config /var/lib/ceph/82f504c8-1959-11f1-9a16-a78d135e200b/mon.a/config 2026-03-06T13:42:18.540 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 06 13:42:18 vm08 ceph-mon[53561]: from='client.14373 -' entity='client.admin' cmd=[{"prefix": "pg dump", "target": ["mon-mgr", ""], "format": "json"}]: dispatch 2026-03-06T13:42:18.540 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 06 13:42:18 vm08 ceph-mon[53561]: pgmap v54: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-06T13:42:18.615 INFO:teuthology.orchestra.run.vm00.stdout: 2026-03-06T13:42:18.615 INFO:teuthology.orchestra.run.vm00.stderr:dumped all 2026-03-06T13:42:18.655 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:42:18 vm01 ceph-mon[56331]: from='client.14373 -' entity='client.admin' cmd=[{"prefix": "pg dump", "target": ["mon-mgr", ""], "format": "json"}]: dispatch 2026-03-06T13:42:18.655 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:42:18 vm01 ceph-mon[56331]: pgmap v54: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-06T13:42:18.690 INFO:teuthology.orchestra.run.vm00.stdout:{"pg_ready":true,"pg_map":{"version":54,"stamp":"2026-03-06T12:42:17.988566+0000","last_osdmap_epoch":0,"last_pg_scan":0,"pg_stats_sum":{"stat_sum":{"num_bytes":459280,"num_objects":2,"num_object_clones":0,"num_object_copies":6,"num_objects_missing_on_primary":0,"num_objects_missing":0,"num_objects_degraded":0,"num_objects_misplaced":0,"num_objects_unfound":0,"num_objects_dirty":2,"num_whiteouts":0,"num_read":46,"num_read_kb":37,"num_write":57,"num_write_kb":584,"num_scrub_errors":0,"num_shallow_scrub_errors":0,"num_deep_scrub_errors":0,"num_objects_recovered":0,"num_bytes_recovered":0,"num_keys_recovered":0,"num_objects_omap":0,"num_objects_hit_set_archive":0,"num_bytes_hit_set_archive":0,"num_flush":0,"num_flush_kb":0,"num_evict":0,"num_evict_kb":0,"num_promote":0,"num_flush_mode_high":0,"num_flush_mode_low":0,"num_evict_mode_some":0,"num_evict_mode_full":0,"num_objects_pinned":0,"num_legacy_snapsets":0,"num_large_omap_objects":0,"num_objects_manifest":0,"num_omap_bytes":0,"num_omap_keys":0,"num_objects_repaired":0},"store_stats":{"total":0,"available":0,"internally_reserved":0,"allocated":0,"data_stored":0,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":0},"log_size":32,"ondisk_log_size":32,"up":3,"acting":3,"num_store_stats":0},"osd_stats_sum":{"up_from":0,"seq":0,"num_pgs":3,"num_osds":3,"num_per_pool_osds":3,"num_per_pool_omap_osds":3,"kb":62902272,"kb_used":82756,"kb_used_data":1836,"kb_used_omap":4,"kb_used_meta":80443,"kb_avail":62819516,"statfs":{"total":64411926528,"available":64327184384,"internally_reserved":0,"allocated":1880064,"data_stored":1528179,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":4770,"internal_metadata":82373982},"hb_peers":[],"snap_trim_queue_len":0,"num_snap_trimming":0,"num_shards_repaired":0,"op_queue_age_hist":{"histogram":[],"upper_bound":1},"perf_stat":{"commit_latency_ms":0,"apply_latency_ms":0,"commit_latency_ns":0,"apply_latency_ns":0},"alerts":[],"network_ping_times":[]},"pg_stats_delta":{"stat_sum":{"num_bytes":0,"num_objects":0,"num_object_clones":0,"num_object_copies":0,"num_objects_missing_on_primary":0,"num_objects_missing":0,"num_objects_degraded":0,"num_objects_misplaced":0,"num_objects_unfound":0,"num_objects_dirty":0,"num_whiteouts":0,"num_read":0,"num_read_kb":0,"num_write":0,"num_write_kb":0,"num_scrub_errors":0,"num_shallow_scrub_errors":0,"num_deep_scrub_errors":0,"num_objects_recovered":0,"num_bytes_recovered":0,"num_keys_recovered":0,"num_objects_omap":0,"num_objects_hit_set_archive":0,"num_bytes_hit_set_archive":0,"num_flush":0,"num_flush_kb":0,"num_evict":0,"num_evict_kb":0,"num_promote":0,"num_flush_mode_high":0,"num_flush_mode_low":0,"num_evict_mode_some":0,"num_evict_mode_full":0,"num_objects_pinned":0,"num_legacy_snapsets":0,"num_large_omap_objects":0,"num_objects_manifest":0,"num_omap_bytes":0,"num_omap_keys":0,"num_objects_repaired":0},"store_stats":{"total":0,"available":0,"internally_reserved":0,"allocated":0,"data_stored":0,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":0},"log_size":0,"ondisk_log_size":0,"up":0,"acting":0,"num_store_stats":0,"stamp_delta":"2.000296"},"pg_stats":[{"pgid":"1.0","version":"18'32","reported_seq":57,"reported_epoch":19,"state":"active+clean","last_fresh":"2026-03-06T12:42:12.126398+0000","last_change":"2026-03-06T12:42:11.110567+0000","last_active":"2026-03-06T12:42:12.126398+0000","last_peered":"2026-03-06T12:42:12.126398+0000","last_clean":"2026-03-06T12:42:12.126398+0000","last_became_active":"2026-03-06T12:42:11.110448+0000","last_became_peered":"2026-03-06T12:42:11.110448+0000","last_unstale":"2026-03-06T12:42:12.126398+0000","last_undegraded":"2026-03-06T12:42:12.126398+0000","last_fullsized":"2026-03-06T12:42:12.126398+0000","mapping_epoch":17,"log_start":"0'0","ondisk_log_start":"0'0","created":17,"last_epoch_clean":18,"parent":"0.0","parent_split_bits":0,"last_scrub":"0'0","last_scrub_stamp":"2026-03-06T12:42:10.100211+0000","last_deep_scrub":"0'0","last_deep_scrub_stamp":"2026-03-06T12:42:10.100211+0000","last_clean_scrub_stamp":"2026-03-06T12:42:10.100211+0000","objects_scrubbed":0,"log_size":32,"log_dups_size":0,"ondisk_log_size":32,"stats_invalid":false,"dirty_stats_invalid":false,"omap_stats_invalid":false,"hitset_stats_invalid":false,"hitset_bytes_stats_invalid":false,"pin_stats_invalid":false,"manifest_stats_invalid":false,"snaptrimq_len":0,"last_scrub_duration":0,"scrub_schedule":"periodic scrub scheduled @ 2026-03-07T16:31:21.552364+0000","scrub_duration":0,"objects_trimmed":0,"snaptrim_duration":0,"stat_sum":{"num_bytes":459280,"num_objects":2,"num_object_clones":0,"num_object_copies":6,"num_objects_missing_on_primary":0,"num_objects_missing":0,"num_objects_degraded":0,"num_objects_misplaced":0,"num_objects_unfound":0,"num_objects_dirty":2,"num_whiteouts":0,"num_read":46,"num_read_kb":37,"num_write":57,"num_write_kb":584,"num_scrub_errors":0,"num_shallow_scrub_errors":0,"num_deep_scrub_errors":0,"num_objects_recovered":0,"num_bytes_recovered":0,"num_keys_recovered":0,"num_objects_omap":0,"num_objects_hit_set_archive":0,"num_bytes_hit_set_archive":0,"num_flush":0,"num_flush_kb":0,"num_evict":0,"num_evict_kb":0,"num_promote":0,"num_flush_mode_high":0,"num_flush_mode_low":0,"num_evict_mode_some":0,"num_evict_mode_full":0,"num_objects_pinned":0,"num_legacy_snapsets":0,"num_large_omap_objects":0,"num_objects_manifest":0,"num_omap_bytes":0,"num_omap_keys":0,"num_objects_repaired":0},"up":[1,2,0],"acting":[1,2,0],"avail_no_missing":[],"object_location_counts":[],"blocked_by":[],"up_primary":1,"acting_primary":1,"purged_snaps":[]}],"pool_stats":[{"poolid":1,"num_pg":1,"stat_sum":{"num_bytes":459280,"num_objects":2,"num_object_clones":0,"num_object_copies":6,"num_objects_missing_on_primary":0,"num_objects_missing":0,"num_objects_degraded":0,"num_objects_misplaced":0,"num_objects_unfound":0,"num_objects_dirty":2,"num_whiteouts":0,"num_read":46,"num_read_kb":37,"num_write":57,"num_write_kb":584,"num_scrub_errors":0,"num_shallow_scrub_errors":0,"num_deep_scrub_errors":0,"num_objects_recovered":0,"num_bytes_recovered":0,"num_keys_recovered":0,"num_objects_omap":0,"num_objects_hit_set_archive":0,"num_bytes_hit_set_archive":0,"num_flush":0,"num_flush_kb":0,"num_evict":0,"num_evict_kb":0,"num_promote":0,"num_flush_mode_high":0,"num_flush_mode_low":0,"num_evict_mode_some":0,"num_evict_mode_full":0,"num_objects_pinned":0,"num_legacy_snapsets":0,"num_large_omap_objects":0,"num_objects_manifest":0,"num_omap_bytes":0,"num_omap_keys":0,"num_objects_repaired":0},"store_stats":{"total":0,"available":0,"internally_reserved":0,"allocated":1388544,"data_stored":1377840,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":0},"log_size":32,"ondisk_log_size":32,"up":3,"acting":3,"num_store_stats":3}],"osd_stats":[{"osd":2,"up_from":16,"seq":68719476739,"num_pgs":1,"num_osds":1,"num_per_pool_osds":1,"num_per_pool_omap_osds":1,"kb":20967424,"kb_used":27584,"kb_used_data":612,"kb_used_omap":1,"kb_used_meta":26814,"kb_avail":20939840,"statfs":{"total":21470642176,"available":21442396160,"internally_reserved":0,"allocated":626688,"data_stored":509393,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":1590,"internal_metadata":27457994},"hb_peers":[0,1],"snap_trim_queue_len":0,"num_snap_trimming":0,"num_shards_repaired":0,"op_queue_age_hist":{"histogram":[],"upper_bound":1},"perf_stat":{"commit_latency_ms":0,"apply_latency_ms":0,"commit_latency_ns":0,"apply_latency_ns":0},"alerts":[]},{"osd":1,"up_from":12,"seq":51539607558,"num_pgs":1,"num_osds":1,"num_per_pool_osds":1,"num_per_pool_omap_osds":1,"kb":20967424,"kb_used":27584,"kb_used_data":612,"kb_used_omap":1,"kb_used_meta":26814,"kb_avail":20939840,"statfs":{"total":21470642176,"available":21442396160,"internally_reserved":0,"allocated":626688,"data_stored":509393,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":1590,"internal_metadata":27457994},"hb_peers":[0,2],"snap_trim_queue_len":0,"num_snap_trimming":0,"num_shards_repaired":0,"op_queue_age_hist":{"histogram":[],"upper_bound":1},"perf_stat":{"commit_latency_ms":0,"apply_latency_ms":0,"commit_latency_ns":0,"apply_latency_ns":0},"alerts":[]},{"osd":0,"up_from":8,"seq":34359738377,"num_pgs":1,"num_osds":1,"num_per_pool_osds":1,"num_per_pool_omap_osds":1,"kb":20967424,"kb_used":27588,"kb_used_data":612,"kb_used_omap":1,"kb_used_meta":26814,"kb_avail":20939836,"statfs":{"total":21470642176,"available":21442392064,"internally_reserved":0,"allocated":626688,"data_stored":509393,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":1590,"internal_metadata":27457994},"hb_peers":[1,2],"snap_trim_queue_len":0,"num_snap_trimming":0,"num_shards_repaired":0,"op_queue_age_hist":{"histogram":[],"upper_bound":1},"perf_stat":{"commit_latency_ms":0,"apply_latency_ms":0,"commit_latency_ns":0,"apply_latency_ns":0},"alerts":[]}],"pool_statfs":[{"poolid":1,"osd":0,"total":0,"available":0,"internally_reserved":0,"allocated":462848,"data_stored":459280,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":0},{"poolid":1,"osd":1,"total":0,"available":0,"internally_reserved":0,"allocated":462848,"data_stored":459280,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":0},{"poolid":1,"osd":2,"total":0,"available":0,"internally_reserved":0,"allocated":462848,"data_stored":459280,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":0}]}} 2026-03-06T13:42:18.690 INFO:tasks.cephadm.ceph_manager.ceph:clean! 2026-03-06T13:42:18.690 INFO:tasks.ceph:Waiting until ceph cluster ceph is healthy... 2026-03-06T13:42:18.690 INFO:tasks.cephadm.ceph_manager.ceph:wait_until_healthy 2026-03-06T13:42:18.690 DEBUG:teuthology.orchestra.run.vm00:> sudo /home/ubuntu/cephtest/cephadm --image harbor.clyso.com/custom-ceph/ceph/ceph:cobaltcore-storage-v19.2.3-fasttrack-3 shell --fsid 82f504c8-1959-11f1-9a16-a78d135e200b -- ceph health --format=json 2026-03-06T13:42:19.054 INFO:teuthology.orchestra.run.vm00.stderr:Inferring config /var/lib/ceph/82f504c8-1959-11f1-9a16-a78d135e200b/mon.a/config 2026-03-06T13:42:19.445 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:42:19 vm00 ceph-mon[51914]: from='client.14379 -' entity='client.admin' cmd=[{"prefix": "pg dump", "target": ["mon-mgr", ""], "format": "json"}]: dispatch 2026-03-06T13:42:19.445 INFO:teuthology.orchestra.run.vm00.stdout: 2026-03-06T13:42:19.445 INFO:teuthology.orchestra.run.vm00.stdout:{"status":"HEALTH_OK","checks":{},"mutes":[]} 2026-03-06T13:42:19.533 INFO:tasks.cephadm.ceph_manager.ceph:wait_until_healthy done 2026-03-06T13:42:19.533 INFO:tasks.cephadm:Setup complete, yielding 2026-03-06T13:42:19.533 INFO:teuthology.run_tasks:Running task cephadm.shell... 2026-03-06T13:42:19.535 INFO:tasks.cephadm:Running commands on role host.a host ubuntu@vm00.local 2026-03-06T13:42:19.535 DEBUG:teuthology.orchestra.run.vm00:> sudo /home/ubuntu/cephtest/cephadm --image harbor.clyso.com/custom-ceph/ceph/ceph:cobaltcore-storage-v19.2.3-fasttrack-3 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 82f504c8-1959-11f1-9a16-a78d135e200b -- bash -c 'ceph mgr module enable rgw' 2026-03-06T13:42:19.540 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 06 13:42:19 vm08 ceph-mon[53561]: from='client.14379 -' entity='client.admin' cmd=[{"prefix": "pg dump", "target": ["mon-mgr", ""], "format": "json"}]: dispatch 2026-03-06T13:42:19.655 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:42:19 vm01 ceph-mon[56331]: from='client.14379 -' entity='client.admin' cmd=[{"prefix": "pg dump", "target": ["mon-mgr", ""], "format": "json"}]: dispatch 2026-03-06T13:42:19.903 INFO:teuthology.orchestra.run.vm00.stderr:Inferring config /var/lib/ceph/82f504c8-1959-11f1-9a16-a78d135e200b/mon.a/config 2026-03-06T13:42:20.252 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:42:20 vm00 ceph-mon[51914]: from='client.? 192.168.123.100:0/2845612395' entity='client.admin' cmd=[{"prefix": "health", "format": "json"}]: dispatch 2026-03-06T13:42:20.252 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:42:20 vm00 ceph-mon[51914]: pgmap v55: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-06T13:42:20.540 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 06 13:42:20 vm08 ceph-mon[53561]: from='client.? 192.168.123.100:0/2845612395' entity='client.admin' cmd=[{"prefix": "health", "format": "json"}]: dispatch 2026-03-06T13:42:20.540 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 06 13:42:20 vm08 ceph-mon[53561]: pgmap v55: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-06T13:42:20.655 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:42:20 vm01 ceph-mon[56331]: from='client.? 192.168.123.100:0/2845612395' entity='client.admin' cmd=[{"prefix": "health", "format": "json"}]: dispatch 2026-03-06T13:42:20.655 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:42:20 vm01 ceph-mon[56331]: pgmap v55: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-06T13:42:21.273 INFO:teuthology.run_tasks:Running task rgw_module.apply... 2026-03-06T13:42:21.275 INFO:tasks.rgw_module:Applying spec(s): rgw_realm: myrealm1 rgw_zone: myzone1 rgw_zonegroup: myzonegroup1 spec: rgw_frontend_port: 5500 2026-03-06T13:42:21.276 DEBUG:teuthology.orchestra.run.vm00:> sudo /home/ubuntu/cephtest/cephadm --image harbor.clyso.com/custom-ceph/ceph/ceph:cobaltcore-storage-v19.2.3-fasttrack-3 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 82f504c8-1959-11f1-9a16-a78d135e200b -- ceph rgw realm bootstrap -i - 2026-03-06T13:42:21.491 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:42:21 vm01 ceph-mon[56331]: from='client.? 192.168.123.100:0/1505116998' entity='client.admin' cmd=[{"prefix": "mgr module enable", "module": "rgw"}]: dispatch 2026-03-06T13:42:21.491 INFO:journalctl@ceph.mgr.b.vm01.stdout:Mar 06 13:42:21 vm01 ceph-82f504c8-1959-11f1-9a16-a78d135e200b-mgr-b[57308]: ignoring --setuser ceph since I am not root 2026-03-06T13:42:21.491 INFO:journalctl@ceph.mgr.b.vm01.stdout:Mar 06 13:42:21 vm01 ceph-82f504c8-1959-11f1-9a16-a78d135e200b-mgr-b[57308]: ignoring --setgroup ceph since I am not root 2026-03-06T13:42:21.491 INFO:journalctl@ceph.mgr.b.vm01.stdout:Mar 06 13:42:21 vm01 ceph-mgr[57313]: -- 192.168.123.101:0/4061912965 <== mon.0 v2:192.168.123.100:3300/0 4 ==== auth_reply(proto 2 0 (0) Success) ==== 194+0+0 (secure 0 0 0) 0x55890af212c0 con 0x55890aeff400 2026-03-06T13:42:21.491 INFO:journalctl@ceph.mgr.b.vm01.stdout:Mar 06 13:42:21 vm01 ceph-82f504c8-1959-11f1-9a16-a78d135e200b-mgr-b[57308]: 2026-03-06T12:42:21.489+0000 7f94d75d5100 -1 mgr[py] Module alerts has missing NOTIFY_TYPES member 2026-03-06T13:42:21.502 INFO:journalctl@ceph.mgr.a.vm00.stdout:Mar 06 13:42:21 vm00 ceph-82f504c8-1959-11f1-9a16-a78d135e200b-mgr-a[52124]: ignoring --setuser ceph since I am not root 2026-03-06T13:42:21.502 INFO:journalctl@ceph.mgr.a.vm00.stdout:Mar 06 13:42:21 vm00 ceph-82f504c8-1959-11f1-9a16-a78d135e200b-mgr-a[52124]: ignoring --setgroup ceph since I am not root 2026-03-06T13:42:21.502 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:42:21 vm00 ceph-mon[51914]: from='client.? 192.168.123.100:0/1505116998' entity='client.admin' cmd=[{"prefix": "mgr module enable", "module": "rgw"}]: dispatch 2026-03-06T13:42:21.540 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 06 13:42:21 vm08 ceph-mon[53561]: from='client.? 192.168.123.100:0/1505116998' entity='client.admin' cmd=[{"prefix": "mgr module enable", "module": "rgw"}]: dispatch 2026-03-06T13:42:21.651 INFO:teuthology.orchestra.run.vm00.stdout:Inferring config /var/lib/ceph/82f504c8-1959-11f1-9a16-a78d135e200b/mon.a/config 2026-03-06T13:42:21.774 INFO:journalctl@ceph.mgr.a.vm00.stdout:Mar 06 13:42:21 vm00 ceph-82f504c8-1959-11f1-9a16-a78d135e200b-mgr-a[52124]: 2026-03-06T12:42:21.515+0000 7f64181e1100 -1 mgr[py] Module alerts has missing NOTIFY_TYPES member 2026-03-06T13:42:21.774 INFO:journalctl@ceph.mgr.a.vm00.stdout:Mar 06 13:42:21 vm00 ceph-82f504c8-1959-11f1-9a16-a78d135e200b-mgr-a[52124]: 2026-03-06T12:42:21.662+0000 7f64181e1100 -1 mgr[py] Module balancer has missing NOTIFY_TYPES member 2026-03-06T13:42:21.905 INFO:journalctl@ceph.mgr.b.vm01.stdout:Mar 06 13:42:21 vm01 ceph-82f504c8-1959-11f1-9a16-a78d135e200b-mgr-b[57308]: 2026-03-06T12:42:21.637+0000 7f94d75d5100 -1 mgr[py] Module balancer has missing NOTIFY_TYPES member 2026-03-06T13:42:22.502 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:42:22 vm00 ceph-mon[51914]: from='client.? 192.168.123.100:0/1505116998' entity='client.admin' cmd='[{"prefix": "mgr module enable", "module": "rgw"}]': finished 2026-03-06T13:42:22.502 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:42:22 vm00 ceph-mon[51914]: mgrmap e16: a(active, since 93s), standbys: b 2026-03-06T13:42:22.540 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 06 13:42:22 vm08 ceph-mon[53561]: from='client.? 192.168.123.100:0/1505116998' entity='client.admin' cmd='[{"prefix": "mgr module enable", "module": "rgw"}]': finished 2026-03-06T13:42:22.540 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 06 13:42:22 vm08 ceph-mon[53561]: mgrmap e16: a(active, since 93s), standbys: b 2026-03-06T13:42:22.655 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:42:22 vm01 ceph-mon[56331]: from='client.? 192.168.123.100:0/1505116998' entity='client.admin' cmd='[{"prefix": "mgr module enable", "module": "rgw"}]': finished 2026-03-06T13:42:22.655 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:42:22 vm01 ceph-mon[56331]: mgrmap e16: a(active, since 93s), standbys: b 2026-03-06T13:42:23.252 INFO:journalctl@ceph.mgr.a.vm00.stdout:Mar 06 13:42:22 vm00 ceph-82f504c8-1959-11f1-9a16-a78d135e200b-mgr-a[52124]: 2026-03-06T12:42:22.901+0000 7f64181e1100 -1 mgr[py] Module crash has missing NOTIFY_TYPES member 2026-03-06T13:42:23.405 INFO:journalctl@ceph.mgr.b.vm01.stdout:Mar 06 13:42:22 vm01 ceph-82f504c8-1959-11f1-9a16-a78d135e200b-mgr-b[57308]: 2026-03-06T12:42:22.910+0000 7f94d75d5100 -1 mgr[py] Module crash has missing NOTIFY_TYPES member 2026-03-06T13:42:24.155 INFO:journalctl@ceph.mgr.b.vm01.stdout:Mar 06 13:42:23 vm01 ceph-82f504c8-1959-11f1-9a16-a78d135e200b-mgr-b[57308]: 2026-03-06T12:42:23.891+0000 7f94d75d5100 -1 mgr[py] Module devicehealth has missing NOTIFY_TYPES member 2026-03-06T13:42:24.155 INFO:journalctl@ceph.mgr.b.vm01.stdout:Mar 06 13:42:24 vm01 ceph-82f504c8-1959-11f1-9a16-a78d135e200b-mgr-b[57308]: 2026-03-06T12:42:24.034+0000 7f94d75d5100 -1 mgr[py] Module influx has missing NOTIFY_TYPES member 2026-03-06T13:42:24.252 INFO:journalctl@ceph.mgr.a.vm00.stdout:Mar 06 13:42:23 vm00 ceph-82f504c8-1959-11f1-9a16-a78d135e200b-mgr-a[52124]: 2026-03-06T12:42:23.907+0000 7f64181e1100 -1 mgr[py] Module devicehealth has missing NOTIFY_TYPES member 2026-03-06T13:42:24.252 INFO:journalctl@ceph.mgr.a.vm00.stdout:Mar 06 13:42:24 vm00 ceph-82f504c8-1959-11f1-9a16-a78d135e200b-mgr-a[52124]: 2026-03-06T12:42:24.052+0000 7f64181e1100 -1 mgr[py] Module influx has missing NOTIFY_TYPES member 2026-03-06T13:42:24.655 INFO:journalctl@ceph.mgr.b.vm01.stdout:Mar 06 13:42:24 vm01 ceph-82f504c8-1959-11f1-9a16-a78d135e200b-mgr-b[57308]: 2026-03-06T12:42:24.314+0000 7f94d75d5100 -1 mgr[py] Module iostat has missing NOTIFY_TYPES member 2026-03-06T13:42:24.752 INFO:journalctl@ceph.mgr.a.vm00.stdout:Mar 06 13:42:24 vm00 ceph-82f504c8-1959-11f1-9a16-a78d135e200b-mgr-a[52124]: 2026-03-06T12:42:24.328+0000 7f64181e1100 -1 mgr[py] Module iostat has missing NOTIFY_TYPES member 2026-03-06T13:42:26.752 INFO:journalctl@ceph.mgr.a.vm00.stdout:Mar 06 13:42:26 vm00 ceph-82f504c8-1959-11f1-9a16-a78d135e200b-mgr-a[52124]: 2026-03-06T12:42:26.477+0000 7f64181e1100 -1 mgr[py] Module nfs has missing NOTIFY_TYPES member 2026-03-06T13:42:26.879 INFO:journalctl@ceph.mgr.b.vm01.stdout:Mar 06 13:42:26 vm01 ceph-82f504c8-1959-11f1-9a16-a78d135e200b-mgr-b[57308]: 2026-03-06T12:42:26.462+0000 7f94d75d5100 -1 mgr[py] Module nfs has missing NOTIFY_TYPES member 2026-03-06T13:42:27.155 INFO:journalctl@ceph.mgr.b.vm01.stdout:Mar 06 13:42:26 vm01 ceph-82f504c8-1959-11f1-9a16-a78d135e200b-mgr-b[57308]: 2026-03-06T12:42:26.877+0000 7f94d75d5100 -1 mgr[py] Module orchestrator has missing NOTIFY_TYPES member 2026-03-06T13:42:27.155 INFO:journalctl@ceph.mgr.b.vm01.stdout:Mar 06 13:42:27 vm01 ceph-82f504c8-1959-11f1-9a16-a78d135e200b-mgr-b[57308]: 2026-03-06T12:42:27.046+0000 7f94d75d5100 -1 mgr[py] Module osd_perf_query has missing NOTIFY_TYPES member 2026-03-06T13:42:27.200 INFO:journalctl@ceph.mgr.a.vm00.stdout:Mar 06 13:42:26 vm00 ceph-82f504c8-1959-11f1-9a16-a78d135e200b-mgr-a[52124]: 2026-03-06T12:42:26.892+0000 7f64181e1100 -1 mgr[py] Module orchestrator has missing NOTIFY_TYPES member 2026-03-06T13:42:27.200 INFO:journalctl@ceph.mgr.a.vm00.stdout:Mar 06 13:42:27 vm00 ceph-82f504c8-1959-11f1-9a16-a78d135e200b-mgr-a[52124]: 2026-03-06T12:42:27.057+0000 7f64181e1100 -1 mgr[py] Module osd_perf_query has missing NOTIFY_TYPES member 2026-03-06T13:42:27.502 INFO:journalctl@ceph.mgr.a.vm00.stdout:Mar 06 13:42:27 vm00 ceph-82f504c8-1959-11f1-9a16-a78d135e200b-mgr-a[52124]: 2026-03-06T12:42:27.198+0000 7f64181e1100 -1 mgr[py] Module osd_support has missing NOTIFY_TYPES member 2026-03-06T13:42:27.502 INFO:journalctl@ceph.mgr.a.vm00.stdout:Mar 06 13:42:27 vm00 ceph-82f504c8-1959-11f1-9a16-a78d135e200b-mgr-a[52124]: 2026-03-06T12:42:27.364+0000 7f64181e1100 -1 mgr[py] Module pg_autoscaler has missing NOTIFY_TYPES member 2026-03-06T13:42:27.517 INFO:journalctl@ceph.mgr.b.vm01.stdout:Mar 06 13:42:27 vm01 ceph-82f504c8-1959-11f1-9a16-a78d135e200b-mgr-b[57308]: 2026-03-06T12:42:27.190+0000 7f94d75d5100 -1 mgr[py] Module osd_support has missing NOTIFY_TYPES member 2026-03-06T13:42:27.517 INFO:journalctl@ceph.mgr.b.vm01.stdout:Mar 06 13:42:27 vm01 ceph-82f504c8-1959-11f1-9a16-a78d135e200b-mgr-b[57308]: 2026-03-06T12:42:27.369+0000 7f94d75d5100 -1 mgr[py] Module pg_autoscaler has missing NOTIFY_TYPES member 2026-03-06T13:42:27.905 INFO:journalctl@ceph.mgr.b.vm01.stdout:Mar 06 13:42:27 vm01 ceph-82f504c8-1959-11f1-9a16-a78d135e200b-mgr-b[57308]: 2026-03-06T12:42:27.515+0000 7f94d75d5100 -1 mgr[py] Module progress has missing NOTIFY_TYPES member 2026-03-06T13:42:28.002 INFO:journalctl@ceph.mgr.a.vm00.stdout:Mar 06 13:42:27 vm00 ceph-82f504c8-1959-11f1-9a16-a78d135e200b-mgr-a[52124]: 2026-03-06T12:42:27.510+0000 7f64181e1100 -1 mgr[py] Module progress has missing NOTIFY_TYPES member 2026-03-06T13:42:28.405 INFO:journalctl@ceph.mgr.b.vm01.stdout:Mar 06 13:42:28 vm01 ceph-82f504c8-1959-11f1-9a16-a78d135e200b-mgr-b[57308]: 2026-03-06T12:42:28.096+0000 7f94d75d5100 -1 mgr[py] Module prometheus has missing NOTIFY_TYPES member 2026-03-06T13:42:28.405 INFO:journalctl@ceph.mgr.b.vm01.stdout:Mar 06 13:42:28 vm01 ceph-82f504c8-1959-11f1-9a16-a78d135e200b-mgr-b[57308]: 2026-03-06T12:42:28.272+0000 7f94d75d5100 -1 mgr[py] Module rbd_support has missing NOTIFY_TYPES member 2026-03-06T13:42:28.502 INFO:journalctl@ceph.mgr.a.vm00.stdout:Mar 06 13:42:28 vm00 ceph-82f504c8-1959-11f1-9a16-a78d135e200b-mgr-a[52124]: 2026-03-06T12:42:28.107+0000 7f64181e1100 -1 mgr[py] Module prometheus has missing NOTIFY_TYPES member 2026-03-06T13:42:28.502 INFO:journalctl@ceph.mgr.a.vm00.stdout:Mar 06 13:42:28 vm00 ceph-82f504c8-1959-11f1-9a16-a78d135e200b-mgr-a[52124]: 2026-03-06T12:42:28.285+0000 7f64181e1100 -1 mgr[py] Module rbd_support has missing NOTIFY_TYPES member 2026-03-06T13:42:29.405 INFO:journalctl@ceph.mgr.b.vm01.stdout:Mar 06 13:42:29 vm01 ceph-82f504c8-1959-11f1-9a16-a78d135e200b-mgr-b[57308]: 2026-03-06T12:42:29.147+0000 7f94d75d5100 -1 mgr[py] Module rgw has missing NOTIFY_TYPES member 2026-03-06T13:42:29.502 INFO:journalctl@ceph.mgr.a.vm00.stdout:Mar 06 13:42:29 vm00 ceph-82f504c8-1959-11f1-9a16-a78d135e200b-mgr-a[52124]: 2026-03-06T12:42:29.160+0000 7f64181e1100 -1 mgr[py] Module rgw has missing NOTIFY_TYPES member 2026-03-06T13:42:30.752 INFO:journalctl@ceph.mgr.a.vm00.stdout:Mar 06 13:42:30 vm00 ceph-82f504c8-1959-11f1-9a16-a78d135e200b-mgr-a[52124]: 2026-03-06T12:42:30.476+0000 7f64181e1100 -1 mgr[py] Module rook has missing NOTIFY_TYPES member 2026-03-06T13:42:30.752 INFO:journalctl@ceph.mgr.a.vm00.stdout:Mar 06 13:42:30 vm00 ceph-82f504c8-1959-11f1-9a16-a78d135e200b-mgr-a[52124]: 2026-03-06T12:42:30.628+0000 7f64181e1100 -1 mgr[py] Module selftest has missing NOTIFY_TYPES member 2026-03-06T13:42:30.770 INFO:journalctl@ceph.mgr.b.vm01.stdout:Mar 06 13:42:30 vm01 ceph-82f504c8-1959-11f1-9a16-a78d135e200b-mgr-b[57308]: 2026-03-06T12:42:30.443+0000 7f94d75d5100 -1 mgr[py] Module rook has missing NOTIFY_TYPES member 2026-03-06T13:42:30.770 INFO:journalctl@ceph.mgr.b.vm01.stdout:Mar 06 13:42:30 vm01 ceph-82f504c8-1959-11f1-9a16-a78d135e200b-mgr-b[57308]: 2026-03-06T12:42:30.596+0000 7f94d75d5100 -1 mgr[py] Module selftest has missing NOTIFY_TYPES member 2026-03-06T13:42:31.107 INFO:journalctl@ceph.mgr.b.vm01.stdout:Mar 06 13:42:30 vm01 ceph-82f504c8-1959-11f1-9a16-a78d135e200b-mgr-b[57308]: 2026-03-06T12:42:30.768+0000 7f94d75d5100 -1 mgr[py] Module snap_schedule has missing NOTIFY_TYPES member 2026-03-06T13:42:31.115 INFO:journalctl@ceph.mgr.a.vm00.stdout:Mar 06 13:42:30 vm00 ceph-82f504c8-1959-11f1-9a16-a78d135e200b-mgr-a[52124]: 2026-03-06T12:42:30.795+0000 7f64181e1100 -1 mgr[py] Module snap_schedule has missing NOTIFY_TYPES member 2026-03-06T13:42:31.405 INFO:journalctl@ceph.mgr.b.vm01.stdout:Mar 06 13:42:31 vm01 ceph-82f504c8-1959-11f1-9a16-a78d135e200b-mgr-b[57308]: 2026-03-06T12:42:31.105+0000 7f94d75d5100 -1 mgr[py] Module status has missing NOTIFY_TYPES member 2026-03-06T13:42:31.405 INFO:journalctl@ceph.mgr.b.vm01.stdout:Mar 06 13:42:31 vm01 ceph-82f504c8-1959-11f1-9a16-a78d135e200b-mgr-b[57308]: 2026-03-06T12:42:31.255+0000 7f94d75d5100 -1 mgr[py] Module telegraf has missing NOTIFY_TYPES member 2026-03-06T13:42:31.502 INFO:journalctl@ceph.mgr.a.vm00.stdout:Mar 06 13:42:31 vm00 ceph-82f504c8-1959-11f1-9a16-a78d135e200b-mgr-a[52124]: 2026-03-06T12:42:31.113+0000 7f64181e1100 -1 mgr[py] Module status has missing NOTIFY_TYPES member 2026-03-06T13:42:31.502 INFO:journalctl@ceph.mgr.a.vm00.stdout:Mar 06 13:42:31 vm00 ceph-82f504c8-1959-11f1-9a16-a78d135e200b-mgr-a[52124]: 2026-03-06T12:42:31.260+0000 7f64181e1100 -1 mgr[py] Module telegraf has missing NOTIFY_TYPES member 2026-03-06T13:42:31.906 INFO:journalctl@ceph.mgr.b.vm01.stdout:Mar 06 13:42:31 vm01 ceph-82f504c8-1959-11f1-9a16-a78d135e200b-mgr-b[57308]: 2026-03-06T12:42:31.639+0000 7f94d75d5100 -1 mgr[py] Module telemetry has missing NOTIFY_TYPES member 2026-03-06T13:42:32.002 INFO:journalctl@ceph.mgr.a.vm00.stdout:Mar 06 13:42:31 vm00 ceph-82f504c8-1959-11f1-9a16-a78d135e200b-mgr-a[52124]: 2026-03-06T12:42:31.639+0000 7f64181e1100 -1 mgr[py] Module telemetry has missing NOTIFY_TYPES member 2026-03-06T13:42:32.405 INFO:journalctl@ceph.mgr.b.vm01.stdout:Mar 06 13:42:31 vm01 ceph-82f504c8-1959-11f1-9a16-a78d135e200b-mgr-b[57308]: 2026-03-06T12:42:31.993+0000 7f94d75d5100 -1 mgr[py] Module test_orchestrator has missing NOTIFY_TYPES member 2026-03-06T13:42:32.443 INFO:journalctl@ceph.mgr.a.vm00.stdout:Mar 06 13:42:32 vm00 ceph-82f504c8-1959-11f1-9a16-a78d135e200b-mgr-a[52124]: 2026-03-06T12:42:32.022+0000 7f64181e1100 -1 mgr[py] Module test_orchestrator has missing NOTIFY_TYPES member 2026-03-06T13:42:32.752 INFO:journalctl@ceph.mgr.a.vm00.stdout:Mar 06 13:42:32 vm00 ceph-82f504c8-1959-11f1-9a16-a78d135e200b-mgr-a[52124]: 2026-03-06T12:42:32.441+0000 7f64181e1100 -1 mgr[py] Module volumes has missing NOTIFY_TYPES member 2026-03-06T13:42:32.752 INFO:journalctl@ceph.mgr.a.vm00.stdout:Mar 06 13:42:32 vm00 ceph-82f504c8-1959-11f1-9a16-a78d135e200b-mgr-a[52124]: 2026-03-06T12:42:32.575+0000 7f64181e1100 -1 mgr[py] Module zabbix has missing NOTIFY_TYPES member 2026-03-06T13:42:32.752 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:42:32 vm00 ceph-mon[51914]: Standby manager daemon b restarted 2026-03-06T13:42:32.752 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:42:32 vm00 ceph-mon[51914]: Standby manager daemon b started 2026-03-06T13:42:32.752 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:42:32 vm00 ceph-mon[51914]: from='mgr.? 192.168.123.101:0/720156893' entity='mgr.b' cmd=[{"prefix": "config-key get", "key": "mgr/dashboard/b/crt"}]: dispatch 2026-03-06T13:42:32.752 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:42:32 vm00 ceph-mon[51914]: from='mgr.? 192.168.123.101:0/720156893' entity='mgr.b' cmd=[{"prefix": "config-key get", "key": "mgr/dashboard/crt"}]: dispatch 2026-03-06T13:42:32.752 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:42:32 vm00 ceph-mon[51914]: from='mgr.? 192.168.123.101:0/720156893' entity='mgr.b' cmd=[{"prefix": "config-key get", "key": "mgr/dashboard/b/key"}]: dispatch 2026-03-06T13:42:32.752 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:42:32 vm00 ceph-mon[51914]: from='mgr.? 192.168.123.101:0/720156893' entity='mgr.b' cmd=[{"prefix": "config-key get", "key": "mgr/dashboard/key"}]: dispatch 2026-03-06T13:42:32.753 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:42:32 vm00 ceph-mon[51914]: Active manager daemon a restarted 2026-03-06T13:42:32.753 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:42:32 vm00 ceph-mon[51914]: Activating manager daemon a 2026-03-06T13:42:32.753 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:42:32 vm00 ceph-mon[51914]: osdmap e20: 3 total, 3 up, 3 in 2026-03-06T13:42:32.753 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:42:32 vm00 ceph-mon[51914]: mgrmap e17: a(active, starting, since 0.00996817s), standbys: b 2026-03-06T13:42:32.753 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:42:32 vm00 ceph-mon[51914]: from='mgr.14400 192.168.123.100:0/3201047282' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "a"}]: dispatch 2026-03-06T13:42:32.753 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:42:32 vm00 ceph-mon[51914]: from='mgr.14400 192.168.123.100:0/3201047282' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-06T13:42:32.753 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:42:32 vm00 ceph-mon[51914]: from='mgr.14400 192.168.123.100:0/3201047282' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "c"}]: dispatch 2026-03-06T13:42:32.753 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:42:32 vm00 ceph-mon[51914]: from='mgr.14400 192.168.123.100:0/3201047282' entity='mgr.a' cmd=[{"prefix": "mgr metadata", "who": "a", "id": "a"}]: dispatch 2026-03-06T13:42:32.753 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:42:32 vm00 ceph-mon[51914]: from='mgr.14400 192.168.123.100:0/3201047282' entity='mgr.a' cmd=[{"prefix": "mgr metadata", "who": "b", "id": "b"}]: dispatch 2026-03-06T13:42:32.753 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:42:32 vm00 ceph-mon[51914]: from='mgr.14400 192.168.123.100:0/3201047282' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 0}]: dispatch 2026-03-06T13:42:32.753 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:42:32 vm00 ceph-mon[51914]: from='mgr.14400 192.168.123.100:0/3201047282' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 1}]: dispatch 2026-03-06T13:42:32.753 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:42:32 vm00 ceph-mon[51914]: from='mgr.14400 192.168.123.100:0/3201047282' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 2}]: dispatch 2026-03-06T13:42:32.753 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:42:32 vm00 ceph-mon[51914]: from='mgr.14400 192.168.123.100:0/3201047282' entity='mgr.a' cmd=[{"prefix": "mds metadata"}]: dispatch 2026-03-06T13:42:32.753 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:42:32 vm00 ceph-mon[51914]: from='mgr.14400 192.168.123.100:0/3201047282' entity='mgr.a' cmd=[{"prefix": "osd metadata"}]: dispatch 2026-03-06T13:42:32.753 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:42:32 vm00 ceph-mon[51914]: from='mgr.14400 192.168.123.100:0/3201047282' entity='mgr.a' cmd=[{"prefix": "mon metadata"}]: dispatch 2026-03-06T13:42:32.753 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:42:32 vm00 ceph-mon[51914]: Manager daemon a is now available 2026-03-06T13:42:32.903 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:42:32 vm01 ceph-mon[56331]: Standby manager daemon b restarted 2026-03-06T13:42:32.903 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:42:32 vm01 ceph-mon[56331]: Standby manager daemon b started 2026-03-06T13:42:32.903 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:42:32 vm01 ceph-mon[56331]: from='mgr.? 192.168.123.101:0/720156893' entity='mgr.b' cmd=[{"prefix": "config-key get", "key": "mgr/dashboard/b/crt"}]: dispatch 2026-03-06T13:42:32.903 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:42:32 vm01 ceph-mon[56331]: from='mgr.? 192.168.123.101:0/720156893' entity='mgr.b' cmd=[{"prefix": "config-key get", "key": "mgr/dashboard/crt"}]: dispatch 2026-03-06T13:42:32.903 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:42:32 vm01 ceph-mon[56331]: from='mgr.? 192.168.123.101:0/720156893' entity='mgr.b' cmd=[{"prefix": "config-key get", "key": "mgr/dashboard/b/key"}]: dispatch 2026-03-06T13:42:32.903 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:42:32 vm01 ceph-mon[56331]: from='mgr.? 192.168.123.101:0/720156893' entity='mgr.b' cmd=[{"prefix": "config-key get", "key": "mgr/dashboard/key"}]: dispatch 2026-03-06T13:42:32.903 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:42:32 vm01 ceph-mon[56331]: Active manager daemon a restarted 2026-03-06T13:42:32.903 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:42:32 vm01 ceph-mon[56331]: Activating manager daemon a 2026-03-06T13:42:32.903 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:42:32 vm01 ceph-mon[56331]: osdmap e20: 3 total, 3 up, 3 in 2026-03-06T13:42:32.903 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:42:32 vm01 ceph-mon[56331]: mgrmap e17: a(active, starting, since 0.00996817s), standbys: b 2026-03-06T13:42:32.903 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:42:32 vm01 ceph-mon[56331]: from='mgr.14400 192.168.123.100:0/3201047282' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "a"}]: dispatch 2026-03-06T13:42:32.903 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:42:32 vm01 ceph-mon[56331]: from='mgr.14400 192.168.123.100:0/3201047282' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-06T13:42:32.903 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:42:32 vm01 ceph-mon[56331]: from='mgr.14400 192.168.123.100:0/3201047282' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "c"}]: dispatch 2026-03-06T13:42:32.903 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:42:32 vm01 ceph-mon[56331]: from='mgr.14400 192.168.123.100:0/3201047282' entity='mgr.a' cmd=[{"prefix": "mgr metadata", "who": "a", "id": "a"}]: dispatch 2026-03-06T13:42:32.903 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:42:32 vm01 ceph-mon[56331]: from='mgr.14400 192.168.123.100:0/3201047282' entity='mgr.a' cmd=[{"prefix": "mgr metadata", "who": "b", "id": "b"}]: dispatch 2026-03-06T13:42:32.903 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:42:32 vm01 ceph-mon[56331]: from='mgr.14400 192.168.123.100:0/3201047282' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 0}]: dispatch 2026-03-06T13:42:32.903 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:42:32 vm01 ceph-mon[56331]: from='mgr.14400 192.168.123.100:0/3201047282' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 1}]: dispatch 2026-03-06T13:42:32.903 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:42:32 vm01 ceph-mon[56331]: from='mgr.14400 192.168.123.100:0/3201047282' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 2}]: dispatch 2026-03-06T13:42:32.903 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:42:32 vm01 ceph-mon[56331]: from='mgr.14400 192.168.123.100:0/3201047282' entity='mgr.a' cmd=[{"prefix": "mds metadata"}]: dispatch 2026-03-06T13:42:32.903 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:42:32 vm01 ceph-mon[56331]: from='mgr.14400 192.168.123.100:0/3201047282' entity='mgr.a' cmd=[{"prefix": "osd metadata"}]: dispatch 2026-03-06T13:42:32.903 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:42:32 vm01 ceph-mon[56331]: from='mgr.14400 192.168.123.100:0/3201047282' entity='mgr.a' cmd=[{"prefix": "mon metadata"}]: dispatch 2026-03-06T13:42:32.903 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:42:32 vm01 ceph-mon[56331]: Manager daemon a is now available 2026-03-06T13:42:32.903 INFO:journalctl@ceph.mgr.b.vm01.stdout:Mar 06 13:42:32 vm01 ceph-82f504c8-1959-11f1-9a16-a78d135e200b-mgr-b[57308]: 2026-03-06T12:42:32.418+0000 7f94d75d5100 -1 mgr[py] Module volumes has missing NOTIFY_TYPES member 2026-03-06T13:42:32.903 INFO:journalctl@ceph.mgr.b.vm01.stdout:Mar 06 13:42:32 vm01 ceph-82f504c8-1959-11f1-9a16-a78d135e200b-mgr-b[57308]: 2026-03-06T12:42:32.561+0000 7f94d75d5100 -1 mgr[py] Module zabbix has missing NOTIFY_TYPES member 2026-03-06T13:42:33.013 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 06 13:42:32 vm08 ceph-mon[53561]: Standby manager daemon b restarted 2026-03-06T13:42:33.013 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 06 13:42:32 vm08 ceph-mon[53561]: Standby manager daemon b started 2026-03-06T13:42:33.014 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 06 13:42:32 vm08 ceph-mon[53561]: from='mgr.? 192.168.123.101:0/720156893' entity='mgr.b' cmd=[{"prefix": "config-key get", "key": "mgr/dashboard/b/crt"}]: dispatch 2026-03-06T13:42:33.014 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 06 13:42:32 vm08 ceph-mon[53561]: from='mgr.? 192.168.123.101:0/720156893' entity='mgr.b' cmd=[{"prefix": "config-key get", "key": "mgr/dashboard/crt"}]: dispatch 2026-03-06T13:42:33.014 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 06 13:42:32 vm08 ceph-mon[53561]: from='mgr.? 192.168.123.101:0/720156893' entity='mgr.b' cmd=[{"prefix": "config-key get", "key": "mgr/dashboard/b/key"}]: dispatch 2026-03-06T13:42:33.014 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 06 13:42:32 vm08 ceph-mon[53561]: from='mgr.? 192.168.123.101:0/720156893' entity='mgr.b' cmd=[{"prefix": "config-key get", "key": "mgr/dashboard/key"}]: dispatch 2026-03-06T13:42:33.014 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 06 13:42:32 vm08 ceph-mon[53561]: Active manager daemon a restarted 2026-03-06T13:42:33.014 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 06 13:42:32 vm08 ceph-mon[53561]: Activating manager daemon a 2026-03-06T13:42:33.014 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 06 13:42:32 vm08 ceph-mon[53561]: osdmap e20: 3 total, 3 up, 3 in 2026-03-06T13:42:33.014 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 06 13:42:32 vm08 ceph-mon[53561]: mgrmap e17: a(active, starting, since 0.00996817s), standbys: b 2026-03-06T13:42:33.014 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 06 13:42:32 vm08 ceph-mon[53561]: from='mgr.14400 192.168.123.100:0/3201047282' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "a"}]: dispatch 2026-03-06T13:42:33.014 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 06 13:42:32 vm08 ceph-mon[53561]: from='mgr.14400 192.168.123.100:0/3201047282' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-06T13:42:33.014 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 06 13:42:32 vm08 ceph-mon[53561]: from='mgr.14400 192.168.123.100:0/3201047282' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "c"}]: dispatch 2026-03-06T13:42:33.014 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 06 13:42:32 vm08 ceph-mon[53561]: from='mgr.14400 192.168.123.100:0/3201047282' entity='mgr.a' cmd=[{"prefix": "mgr metadata", "who": "a", "id": "a"}]: dispatch 2026-03-06T13:42:33.014 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 06 13:42:32 vm08 ceph-mon[53561]: from='mgr.14400 192.168.123.100:0/3201047282' entity='mgr.a' cmd=[{"prefix": "mgr metadata", "who": "b", "id": "b"}]: dispatch 2026-03-06T13:42:33.014 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 06 13:42:32 vm08 ceph-mon[53561]: from='mgr.14400 192.168.123.100:0/3201047282' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 0}]: dispatch 2026-03-06T13:42:33.014 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 06 13:42:32 vm08 ceph-mon[53561]: from='mgr.14400 192.168.123.100:0/3201047282' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 1}]: dispatch 2026-03-06T13:42:33.014 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 06 13:42:32 vm08 ceph-mon[53561]: from='mgr.14400 192.168.123.100:0/3201047282' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 2}]: dispatch 2026-03-06T13:42:33.014 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 06 13:42:32 vm08 ceph-mon[53561]: from='mgr.14400 192.168.123.100:0/3201047282' entity='mgr.a' cmd=[{"prefix": "mds metadata"}]: dispatch 2026-03-06T13:42:33.014 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 06 13:42:32 vm08 ceph-mon[53561]: from='mgr.14400 192.168.123.100:0/3201047282' entity='mgr.a' cmd=[{"prefix": "osd metadata"}]: dispatch 2026-03-06T13:42:33.014 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 06 13:42:32 vm08 ceph-mon[53561]: from='mgr.14400 192.168.123.100:0/3201047282' entity='mgr.a' cmd=[{"prefix": "mon metadata"}]: dispatch 2026-03-06T13:42:33.014 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 06 13:42:32 vm08 ceph-mon[53561]: Manager daemon a is now available 2026-03-06T13:42:33.905 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:42:33 vm01 ceph-mon[56331]: from='mgr.14400 192.168.123.100:0/3201047282' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-06T13:42:33.906 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:42:33 vm01 ceph-mon[56331]: from='mgr.14400 192.168.123.100:0/3201047282' entity='mgr.a' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/a/mirror_snapshot_schedule"}]: dispatch 2026-03-06T13:42:33.906 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:42:33 vm01 ceph-mon[56331]: from='mgr.14400 192.168.123.100:0/3201047282' entity='mgr.a' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/a/trash_purge_schedule"}]: dispatch 2026-03-06T13:42:33.906 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:42:33 vm01 ceph-mon[56331]: mgrmap e18: a(active, since 1.02017s), standbys: b 2026-03-06T13:42:34.002 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:42:33 vm00 ceph-mon[51914]: from='mgr.14400 192.168.123.100:0/3201047282' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-06T13:42:34.003 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:42:33 vm00 ceph-mon[51914]: from='mgr.14400 192.168.123.100:0/3201047282' entity='mgr.a' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/a/mirror_snapshot_schedule"}]: dispatch 2026-03-06T13:42:34.003 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:42:33 vm00 ceph-mon[51914]: from='mgr.14400 192.168.123.100:0/3201047282' entity='mgr.a' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/a/trash_purge_schedule"}]: dispatch 2026-03-06T13:42:34.003 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:42:33 vm00 ceph-mon[51914]: mgrmap e18: a(active, since 1.02017s), standbys: b 2026-03-06T13:42:34.040 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 06 13:42:33 vm08 ceph-mon[53561]: from='mgr.14400 192.168.123.100:0/3201047282' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-06T13:42:34.040 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 06 13:42:33 vm08 ceph-mon[53561]: from='mgr.14400 192.168.123.100:0/3201047282' entity='mgr.a' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/a/mirror_snapshot_schedule"}]: dispatch 2026-03-06T13:42:34.040 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 06 13:42:33 vm08 ceph-mon[53561]: from='mgr.14400 192.168.123.100:0/3201047282' entity='mgr.a' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/a/trash_purge_schedule"}]: dispatch 2026-03-06T13:42:34.040 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 06 13:42:33 vm08 ceph-mon[53561]: mgrmap e18: a(active, since 1.02017s), standbys: b 2026-03-06T13:42:35.040 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 06 13:42:34 vm08 ceph-mon[53561]: from='client.24254 -' entity='client.admin' cmd=[{"prefix": "rgw realm bootstrap", "target": ["mon-mgr", ""]}]: dispatch 2026-03-06T13:42:35.040 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 06 13:42:34 vm08 ceph-mon[53561]: pgmap v3: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-06T13:42:35.040 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 06 13:42:34 vm08 ceph-mon[53561]: from='mgr.14400 192.168.123.100:0/3201047282' entity='mgr.a' 2026-03-06T13:42:35.040 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 06 13:42:34 vm08 ceph-mon[53561]: from='mgr.14400 192.168.123.100:0/3201047282' entity='mgr.a' 2026-03-06T13:42:35.040 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 06 13:42:34 vm08 ceph-mon[53561]: from='mgr.14400 192.168.123.100:0/3201047282' entity='mgr.a' 2026-03-06T13:42:35.040 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 06 13:42:34 vm08 ceph-mon[53561]: from='mgr.14400 192.168.123.100:0/3201047282' entity='mgr.a' 2026-03-06T13:42:35.040 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 06 13:42:34 vm08 ceph-mon[53561]: from='mgr.14400 192.168.123.100:0/3201047282' entity='mgr.a' 2026-03-06T13:42:35.040 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 06 13:42:34 vm08 ceph-mon[53561]: from='mgr.14400 192.168.123.100:0/3201047282' entity='mgr.a' 2026-03-06T13:42:35.040 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 06 13:42:34 vm08 ceph-mon[53561]: from='mgr.14400 192.168.123.100:0/3201047282' entity='mgr.a' 2026-03-06T13:42:35.040 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 06 13:42:34 vm08 ceph-mon[53561]: from='mgr.14400 192.168.123.100:0/3201047282' entity='mgr.a' 2026-03-06T13:42:35.040 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 06 13:42:34 vm08 ceph-mon[53561]: from='mgr.14400 192.168.123.100:0/3201047282' entity='mgr.a' 2026-03-06T13:42:35.040 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 06 13:42:34 vm08 ceph-mon[53561]: from='mgr.14400 192.168.123.100:0/3201047282' entity='mgr.a' cmd=[{"prefix": "config rm", "who": "osd.1", "name": "osd_memory_target"}]: dispatch 2026-03-06T13:42:35.040 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 06 13:42:34 vm08 ceph-mon[53561]: from='mgr.14400 192.168.123.100:0/3201047282' entity='mgr.a' 2026-03-06T13:42:35.040 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 06 13:42:34 vm08 ceph-mon[53561]: from='mgr.14400 192.168.123.100:0/3201047282' entity='mgr.a' cmd=[{"prefix": "config rm", "who": "osd.0", "name": "osd_memory_target"}]: dispatch 2026-03-06T13:42:35.040 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 06 13:42:34 vm08 ceph-mon[53561]: osdmap e21: 3 total, 3 up, 3 in 2026-03-06T13:42:35.040 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 06 13:42:34 vm08 ceph-mon[53561]: from='client.? 192.168.123.100:0/3693510475' entity='mgr.a' cmd=[{"prefix": "osd pool application enable","pool": ".rgw.root","app": "rgw"}]: dispatch 2026-03-06T13:42:35.040 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 06 13:42:34 vm08 ceph-mon[53561]: from='mgr.14400 192.168.123.100:0/3201047282' entity='mgr.a' 2026-03-06T13:42:35.040 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 06 13:42:34 vm08 ceph-mon[53561]: from='mgr.14400 192.168.123.100:0/3201047282' entity='mgr.a' 2026-03-06T13:42:35.040 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 06 13:42:34 vm08 ceph-mon[53561]: from='mgr.14400 192.168.123.100:0/3201047282' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-06T13:42:35.040 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 06 13:42:34 vm08 ceph-mon[53561]: from='mgr.14400 192.168.123.100:0/3201047282' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-06T13:42:35.155 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:42:34 vm01 ceph-mon[56331]: from='client.24254 -' entity='client.admin' cmd=[{"prefix": "rgw realm bootstrap", "target": ["mon-mgr", ""]}]: dispatch 2026-03-06T13:42:35.155 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:42:34 vm01 ceph-mon[56331]: pgmap v3: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-06T13:42:35.155 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:42:34 vm01 ceph-mon[56331]: from='mgr.14400 192.168.123.100:0/3201047282' entity='mgr.a' 2026-03-06T13:42:35.155 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:42:34 vm01 ceph-mon[56331]: from='mgr.14400 192.168.123.100:0/3201047282' entity='mgr.a' 2026-03-06T13:42:35.155 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:42:34 vm01 ceph-mon[56331]: from='mgr.14400 192.168.123.100:0/3201047282' entity='mgr.a' 2026-03-06T13:42:35.155 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:42:34 vm01 ceph-mon[56331]: from='mgr.14400 192.168.123.100:0/3201047282' entity='mgr.a' 2026-03-06T13:42:35.155 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:42:34 vm01 ceph-mon[56331]: from='mgr.14400 192.168.123.100:0/3201047282' entity='mgr.a' 2026-03-06T13:42:35.155 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:42:34 vm01 ceph-mon[56331]: from='mgr.14400 192.168.123.100:0/3201047282' entity='mgr.a' 2026-03-06T13:42:35.155 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:42:34 vm01 ceph-mon[56331]: from='mgr.14400 192.168.123.100:0/3201047282' entity='mgr.a' 2026-03-06T13:42:35.155 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:42:34 vm01 ceph-mon[56331]: from='mgr.14400 192.168.123.100:0/3201047282' entity='mgr.a' 2026-03-06T13:42:35.155 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:42:34 vm01 ceph-mon[56331]: from='mgr.14400 192.168.123.100:0/3201047282' entity='mgr.a' 2026-03-06T13:42:35.155 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:42:34 vm01 ceph-mon[56331]: from='mgr.14400 192.168.123.100:0/3201047282' entity='mgr.a' cmd=[{"prefix": "config rm", "who": "osd.1", "name": "osd_memory_target"}]: dispatch 2026-03-06T13:42:35.155 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:42:34 vm01 ceph-mon[56331]: from='mgr.14400 192.168.123.100:0/3201047282' entity='mgr.a' 2026-03-06T13:42:35.155 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:42:34 vm01 ceph-mon[56331]: from='mgr.14400 192.168.123.100:0/3201047282' entity='mgr.a' cmd=[{"prefix": "config rm", "who": "osd.0", "name": "osd_memory_target"}]: dispatch 2026-03-06T13:42:35.155 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:42:34 vm01 ceph-mon[56331]: osdmap e21: 3 total, 3 up, 3 in 2026-03-06T13:42:35.155 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:42:34 vm01 ceph-mon[56331]: from='client.? 192.168.123.100:0/3693510475' entity='mgr.a' cmd=[{"prefix": "osd pool application enable","pool": ".rgw.root","app": "rgw"}]: dispatch 2026-03-06T13:42:35.155 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:42:34 vm01 ceph-mon[56331]: from='mgr.14400 192.168.123.100:0/3201047282' entity='mgr.a' 2026-03-06T13:42:35.155 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:42:34 vm01 ceph-mon[56331]: from='mgr.14400 192.168.123.100:0/3201047282' entity='mgr.a' 2026-03-06T13:42:35.155 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:42:34 vm01 ceph-mon[56331]: from='mgr.14400 192.168.123.100:0/3201047282' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-06T13:42:35.155 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:42:34 vm01 ceph-mon[56331]: from='mgr.14400 192.168.123.100:0/3201047282' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-06T13:42:35.252 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:42:34 vm00 ceph-mon[51914]: from='client.24254 -' entity='client.admin' cmd=[{"prefix": "rgw realm bootstrap", "target": ["mon-mgr", ""]}]: dispatch 2026-03-06T13:42:35.252 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:42:34 vm00 ceph-mon[51914]: pgmap v3: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-06T13:42:35.252 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:42:34 vm00 ceph-mon[51914]: from='mgr.14400 192.168.123.100:0/3201047282' entity='mgr.a' 2026-03-06T13:42:35.252 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:42:34 vm00 ceph-mon[51914]: from='mgr.14400 192.168.123.100:0/3201047282' entity='mgr.a' 2026-03-06T13:42:35.252 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:42:34 vm00 ceph-mon[51914]: from='mgr.14400 192.168.123.100:0/3201047282' entity='mgr.a' 2026-03-06T13:42:35.252 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:42:34 vm00 ceph-mon[51914]: from='mgr.14400 192.168.123.100:0/3201047282' entity='mgr.a' 2026-03-06T13:42:35.252 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:42:34 vm00 ceph-mon[51914]: from='mgr.14400 192.168.123.100:0/3201047282' entity='mgr.a' 2026-03-06T13:42:35.252 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:42:34 vm00 ceph-mon[51914]: from='mgr.14400 192.168.123.100:0/3201047282' entity='mgr.a' 2026-03-06T13:42:35.252 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:42:34 vm00 ceph-mon[51914]: from='mgr.14400 192.168.123.100:0/3201047282' entity='mgr.a' 2026-03-06T13:42:35.252 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:42:34 vm00 ceph-mon[51914]: from='mgr.14400 192.168.123.100:0/3201047282' entity='mgr.a' 2026-03-06T13:42:35.252 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:42:34 vm00 ceph-mon[51914]: from='mgr.14400 192.168.123.100:0/3201047282' entity='mgr.a' 2026-03-06T13:42:35.252 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:42:34 vm00 ceph-mon[51914]: from='mgr.14400 192.168.123.100:0/3201047282' entity='mgr.a' cmd=[{"prefix": "config rm", "who": "osd.1", "name": "osd_memory_target"}]: dispatch 2026-03-06T13:42:35.252 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:42:34 vm00 ceph-mon[51914]: from='mgr.14400 192.168.123.100:0/3201047282' entity='mgr.a' 2026-03-06T13:42:35.252 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:42:34 vm00 ceph-mon[51914]: from='mgr.14400 192.168.123.100:0/3201047282' entity='mgr.a' cmd=[{"prefix": "config rm", "who": "osd.0", "name": "osd_memory_target"}]: dispatch 2026-03-06T13:42:35.252 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:42:34 vm00 ceph-mon[51914]: osdmap e21: 3 total, 3 up, 3 in 2026-03-06T13:42:35.252 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:42:34 vm00 ceph-mon[51914]: from='client.? 192.168.123.100:0/3693510475' entity='mgr.a' cmd=[{"prefix": "osd pool application enable","pool": ".rgw.root","app": "rgw"}]: dispatch 2026-03-06T13:42:35.252 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:42:34 vm00 ceph-mon[51914]: from='mgr.14400 192.168.123.100:0/3201047282' entity='mgr.a' 2026-03-06T13:42:35.252 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:42:34 vm00 ceph-mon[51914]: from='mgr.14400 192.168.123.100:0/3201047282' entity='mgr.a' 2026-03-06T13:42:35.252 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:42:34 vm00 ceph-mon[51914]: from='mgr.14400 192.168.123.100:0/3201047282' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-06T13:42:35.252 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:42:34 vm00 ceph-mon[51914]: from='mgr.14400 192.168.123.100:0/3201047282' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-06T13:42:36.040 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 06 13:42:35 vm08 ceph-mon[53561]: [06/Mar/2026:12:42:34] ENGINE Bus STARTING 2026-03-06T13:42:36.040 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 06 13:42:35 vm08 ceph-mon[53561]: pgmap v4: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-06T13:42:36.040 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 06 13:42:35 vm08 ceph-mon[53561]: Adjusting osd_memory_target on vm01 to 257.0M 2026-03-06T13:42:36.040 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 06 13:42:35 vm08 ceph-mon[53561]: Adjusting osd_memory_target on vm00 to 257.0M 2026-03-06T13:42:36.040 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 06 13:42:35 vm08 ceph-mon[53561]: Unable to set osd_memory_target on vm01 to 269536460: error parsing value: Value '269536460' is below minimum 939524096 2026-03-06T13:42:36.040 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 06 13:42:35 vm08 ceph-mon[53561]: Unable to set osd_memory_target on vm00 to 269530726: error parsing value: Value '269530726' is below minimum 939524096 2026-03-06T13:42:36.040 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 06 13:42:35 vm08 ceph-mon[53561]: [06/Mar/2026:12:42:34] ENGINE Serving on http://192.168.123.100:8765 2026-03-06T13:42:36.040 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 06 13:42:35 vm08 ceph-mon[53561]: Updating vm00:/etc/ceph/ceph.conf 2026-03-06T13:42:36.040 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 06 13:42:35 vm08 ceph-mon[53561]: Updating vm01:/etc/ceph/ceph.conf 2026-03-06T13:42:36.040 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 06 13:42:35 vm08 ceph-mon[53561]: Updating vm08:/etc/ceph/ceph.conf 2026-03-06T13:42:36.040 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 06 13:42:35 vm08 ceph-mon[53561]: [06/Mar/2026:12:42:34] ENGINE Serving on https://192.168.123.100:7150 2026-03-06T13:42:36.040 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 06 13:42:35 vm08 ceph-mon[53561]: [06/Mar/2026:12:42:34] ENGINE Bus STARTED 2026-03-06T13:42:36.040 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 06 13:42:35 vm08 ceph-mon[53561]: [06/Mar/2026:12:42:34] ENGINE Client ('192.168.123.100', 59262) lost — peer dropped the TLS connection suddenly, during handshake: (6, 'TLS/SSL connection has been closed (EOF) (_ssl.c:1147)') 2026-03-06T13:42:36.040 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 06 13:42:35 vm08 ceph-mon[53561]: Updating vm01:/var/lib/ceph/82f504c8-1959-11f1-9a16-a78d135e200b/config/ceph.conf 2026-03-06T13:42:36.040 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 06 13:42:35 vm08 ceph-mon[53561]: Updating vm08:/var/lib/ceph/82f504c8-1959-11f1-9a16-a78d135e200b/config/ceph.conf 2026-03-06T13:42:36.040 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 06 13:42:35 vm08 ceph-mon[53561]: Updating vm00:/var/lib/ceph/82f504c8-1959-11f1-9a16-a78d135e200b/config/ceph.conf 2026-03-06T13:42:36.041 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 06 13:42:35 vm08 ceph-mon[53561]: Updating vm01:/etc/ceph/ceph.client.admin.keyring 2026-03-06T13:42:36.041 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 06 13:42:35 vm08 ceph-mon[53561]: Updating vm08:/etc/ceph/ceph.client.admin.keyring 2026-03-06T13:42:36.041 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 06 13:42:35 vm08 ceph-mon[53561]: Updating vm00:/etc/ceph/ceph.client.admin.keyring 2026-03-06T13:42:36.041 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 06 13:42:35 vm08 ceph-mon[53561]: Updating vm01:/var/lib/ceph/82f504c8-1959-11f1-9a16-a78d135e200b/config/ceph.client.admin.keyring 2026-03-06T13:42:36.041 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 06 13:42:35 vm08 ceph-mon[53561]: Updating vm08:/var/lib/ceph/82f504c8-1959-11f1-9a16-a78d135e200b/config/ceph.client.admin.keyring 2026-03-06T13:42:36.041 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 06 13:42:35 vm08 ceph-mon[53561]: Updating vm00:/var/lib/ceph/82f504c8-1959-11f1-9a16-a78d135e200b/config/ceph.client.admin.keyring 2026-03-06T13:42:36.041 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 06 13:42:35 vm08 ceph-mon[53561]: from='mgr.14400 192.168.123.100:0/3201047282' entity='mgr.a' 2026-03-06T13:42:36.041 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 06 13:42:35 vm08 ceph-mon[53561]: from='mgr.14400 192.168.123.100:0/3201047282' entity='mgr.a' 2026-03-06T13:42:36.041 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 06 13:42:35 vm08 ceph-mon[53561]: from='mgr.14400 192.168.123.100:0/3201047282' entity='mgr.a' 2026-03-06T13:42:36.041 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 06 13:42:35 vm08 ceph-mon[53561]: from='mgr.14400 192.168.123.100:0/3201047282' entity='mgr.a' 2026-03-06T13:42:36.041 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 06 13:42:35 vm08 ceph-mon[53561]: from='mgr.14400 192.168.123.100:0/3201047282' entity='mgr.a' 2026-03-06T13:42:36.041 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 06 13:42:35 vm08 ceph-mon[53561]: from='mgr.14400 192.168.123.100:0/3201047282' entity='mgr.a' 2026-03-06T13:42:36.041 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 06 13:42:35 vm08 ceph-mon[53561]: from='mgr.14400 192.168.123.100:0/3201047282' entity='mgr.a' 2026-03-06T13:42:36.041 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 06 13:42:35 vm08 ceph-mon[53561]: from='mgr.14400 192.168.123.100:0/3201047282' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-06T13:42:36.041 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 06 13:42:35 vm08 ceph-mon[53561]: from='mgr.14400 192.168.123.100:0/3201047282' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-06T13:42:36.041 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 06 13:42:35 vm08 ceph-mon[53561]: from='mgr.14400 192.168.123.100:0/3201047282' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-06T13:42:36.041 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 06 13:42:35 vm08 ceph-mon[53561]: from='mgr.14400 192.168.123.100:0/3201047282' entity='mgr.a' 2026-03-06T13:42:36.041 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 06 13:42:35 vm08 ceph-mon[53561]: mgrmap e19: a(active, since 3s), standbys: b 2026-03-06T13:42:36.041 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 06 13:42:35 vm08 ceph-mon[53561]: from='client.? 192.168.123.100:0/3693510475' entity='mgr.a' cmd='[{"prefix": "osd pool application enable","pool": ".rgw.root","app": "rgw"}]': finished 2026-03-06T13:42:36.041 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 06 13:42:35 vm08 ceph-mon[53561]: osdmap e22: 3 total, 3 up, 3 in 2026-03-06T13:42:36.155 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:42:35 vm01 ceph-mon[56331]: [06/Mar/2026:12:42:34] ENGINE Bus STARTING 2026-03-06T13:42:36.155 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:42:35 vm01 ceph-mon[56331]: pgmap v4: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-06T13:42:36.155 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:42:35 vm01 ceph-mon[56331]: Adjusting osd_memory_target on vm01 to 257.0M 2026-03-06T13:42:36.155 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:42:35 vm01 ceph-mon[56331]: Adjusting osd_memory_target on vm00 to 257.0M 2026-03-06T13:42:36.155 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:42:35 vm01 ceph-mon[56331]: Unable to set osd_memory_target on vm01 to 269536460: error parsing value: Value '269536460' is below minimum 939524096 2026-03-06T13:42:36.155 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:42:35 vm01 ceph-mon[56331]: Unable to set osd_memory_target on vm00 to 269530726: error parsing value: Value '269530726' is below minimum 939524096 2026-03-06T13:42:36.155 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:42:35 vm01 ceph-mon[56331]: [06/Mar/2026:12:42:34] ENGINE Serving on http://192.168.123.100:8765 2026-03-06T13:42:36.155 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:42:35 vm01 ceph-mon[56331]: Updating vm00:/etc/ceph/ceph.conf 2026-03-06T13:42:36.155 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:42:35 vm01 ceph-mon[56331]: Updating vm01:/etc/ceph/ceph.conf 2026-03-06T13:42:36.155 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:42:35 vm01 ceph-mon[56331]: Updating vm08:/etc/ceph/ceph.conf 2026-03-06T13:42:36.155 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:42:35 vm01 ceph-mon[56331]: [06/Mar/2026:12:42:34] ENGINE Serving on https://192.168.123.100:7150 2026-03-06T13:42:36.155 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:42:35 vm01 ceph-mon[56331]: [06/Mar/2026:12:42:34] ENGINE Bus STARTED 2026-03-06T13:42:36.155 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:42:35 vm01 ceph-mon[56331]: [06/Mar/2026:12:42:34] ENGINE Client ('192.168.123.100', 59262) lost — peer dropped the TLS connection suddenly, during handshake: (6, 'TLS/SSL connection has been closed (EOF) (_ssl.c:1147)') 2026-03-06T13:42:36.155 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:42:35 vm01 ceph-mon[56331]: Updating vm01:/var/lib/ceph/82f504c8-1959-11f1-9a16-a78d135e200b/config/ceph.conf 2026-03-06T13:42:36.155 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:42:35 vm01 ceph-mon[56331]: Updating vm08:/var/lib/ceph/82f504c8-1959-11f1-9a16-a78d135e200b/config/ceph.conf 2026-03-06T13:42:36.155 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:42:35 vm01 ceph-mon[56331]: Updating vm00:/var/lib/ceph/82f504c8-1959-11f1-9a16-a78d135e200b/config/ceph.conf 2026-03-06T13:42:36.155 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:42:35 vm01 ceph-mon[56331]: Updating vm01:/etc/ceph/ceph.client.admin.keyring 2026-03-06T13:42:36.155 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:42:35 vm01 ceph-mon[56331]: Updating vm08:/etc/ceph/ceph.client.admin.keyring 2026-03-06T13:42:36.155 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:42:35 vm01 ceph-mon[56331]: Updating vm00:/etc/ceph/ceph.client.admin.keyring 2026-03-06T13:42:36.155 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:42:35 vm01 ceph-mon[56331]: Updating vm01:/var/lib/ceph/82f504c8-1959-11f1-9a16-a78d135e200b/config/ceph.client.admin.keyring 2026-03-06T13:42:36.155 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:42:35 vm01 ceph-mon[56331]: Updating vm08:/var/lib/ceph/82f504c8-1959-11f1-9a16-a78d135e200b/config/ceph.client.admin.keyring 2026-03-06T13:42:36.155 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:42:35 vm01 ceph-mon[56331]: Updating vm00:/var/lib/ceph/82f504c8-1959-11f1-9a16-a78d135e200b/config/ceph.client.admin.keyring 2026-03-06T13:42:36.155 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:42:35 vm01 ceph-mon[56331]: from='mgr.14400 192.168.123.100:0/3201047282' entity='mgr.a' 2026-03-06T13:42:36.155 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:42:35 vm01 ceph-mon[56331]: from='mgr.14400 192.168.123.100:0/3201047282' entity='mgr.a' 2026-03-06T13:42:36.155 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:42:35 vm01 ceph-mon[56331]: from='mgr.14400 192.168.123.100:0/3201047282' entity='mgr.a' 2026-03-06T13:42:36.155 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:42:35 vm01 ceph-mon[56331]: from='mgr.14400 192.168.123.100:0/3201047282' entity='mgr.a' 2026-03-06T13:42:36.155 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:42:35 vm01 ceph-mon[56331]: from='mgr.14400 192.168.123.100:0/3201047282' entity='mgr.a' 2026-03-06T13:42:36.155 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:42:35 vm01 ceph-mon[56331]: from='mgr.14400 192.168.123.100:0/3201047282' entity='mgr.a' 2026-03-06T13:42:36.155 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:42:35 vm01 ceph-mon[56331]: from='mgr.14400 192.168.123.100:0/3201047282' entity='mgr.a' 2026-03-06T13:42:36.155 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:42:35 vm01 ceph-mon[56331]: from='mgr.14400 192.168.123.100:0/3201047282' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-06T13:42:36.156 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:42:35 vm01 ceph-mon[56331]: from='mgr.14400 192.168.123.100:0/3201047282' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-06T13:42:36.156 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:42:35 vm01 ceph-mon[56331]: from='mgr.14400 192.168.123.100:0/3201047282' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-06T13:42:36.156 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:42:35 vm01 ceph-mon[56331]: from='mgr.14400 192.168.123.100:0/3201047282' entity='mgr.a' 2026-03-06T13:42:36.156 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:42:35 vm01 ceph-mon[56331]: mgrmap e19: a(active, since 3s), standbys: b 2026-03-06T13:42:36.156 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:42:35 vm01 ceph-mon[56331]: from='client.? 192.168.123.100:0/3693510475' entity='mgr.a' cmd='[{"prefix": "osd pool application enable","pool": ".rgw.root","app": "rgw"}]': finished 2026-03-06T13:42:36.156 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:42:35 vm01 ceph-mon[56331]: osdmap e22: 3 total, 3 up, 3 in 2026-03-06T13:42:36.252 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:42:35 vm00 ceph-mon[51914]: [06/Mar/2026:12:42:34] ENGINE Bus STARTING 2026-03-06T13:42:36.252 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:42:35 vm00 ceph-mon[51914]: pgmap v4: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-06T13:42:36.252 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:42:35 vm00 ceph-mon[51914]: Adjusting osd_memory_target on vm01 to 257.0M 2026-03-06T13:42:36.252 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:42:35 vm00 ceph-mon[51914]: Adjusting osd_memory_target on vm00 to 257.0M 2026-03-06T13:42:36.252 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:42:35 vm00 ceph-mon[51914]: Unable to set osd_memory_target on vm01 to 269536460: error parsing value: Value '269536460' is below minimum 939524096 2026-03-06T13:42:36.252 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:42:35 vm00 ceph-mon[51914]: Unable to set osd_memory_target on vm00 to 269530726: error parsing value: Value '269530726' is below minimum 939524096 2026-03-06T13:42:36.252 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:42:35 vm00 ceph-mon[51914]: [06/Mar/2026:12:42:34] ENGINE Serving on http://192.168.123.100:8765 2026-03-06T13:42:36.253 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:42:35 vm00 ceph-mon[51914]: Updating vm00:/etc/ceph/ceph.conf 2026-03-06T13:42:36.253 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:42:35 vm00 ceph-mon[51914]: Updating vm01:/etc/ceph/ceph.conf 2026-03-06T13:42:36.253 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:42:35 vm00 ceph-mon[51914]: Updating vm08:/etc/ceph/ceph.conf 2026-03-06T13:42:36.253 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:42:35 vm00 ceph-mon[51914]: [06/Mar/2026:12:42:34] ENGINE Serving on https://192.168.123.100:7150 2026-03-06T13:42:36.253 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:42:35 vm00 ceph-mon[51914]: [06/Mar/2026:12:42:34] ENGINE Bus STARTED 2026-03-06T13:42:36.253 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:42:35 vm00 ceph-mon[51914]: [06/Mar/2026:12:42:34] ENGINE Client ('192.168.123.100', 59262) lost — peer dropped the TLS connection suddenly, during handshake: (6, 'TLS/SSL connection has been closed (EOF) (_ssl.c:1147)') 2026-03-06T13:42:36.253 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:42:35 vm00 ceph-mon[51914]: Updating vm01:/var/lib/ceph/82f504c8-1959-11f1-9a16-a78d135e200b/config/ceph.conf 2026-03-06T13:42:36.253 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:42:35 vm00 ceph-mon[51914]: Updating vm08:/var/lib/ceph/82f504c8-1959-11f1-9a16-a78d135e200b/config/ceph.conf 2026-03-06T13:42:36.253 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:42:35 vm00 ceph-mon[51914]: Updating vm00:/var/lib/ceph/82f504c8-1959-11f1-9a16-a78d135e200b/config/ceph.conf 2026-03-06T13:42:36.253 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:42:35 vm00 ceph-mon[51914]: Updating vm01:/etc/ceph/ceph.client.admin.keyring 2026-03-06T13:42:36.253 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:42:35 vm00 ceph-mon[51914]: Updating vm08:/etc/ceph/ceph.client.admin.keyring 2026-03-06T13:42:36.253 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:42:35 vm00 ceph-mon[51914]: Updating vm00:/etc/ceph/ceph.client.admin.keyring 2026-03-06T13:42:36.253 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:42:35 vm00 ceph-mon[51914]: Updating vm01:/var/lib/ceph/82f504c8-1959-11f1-9a16-a78d135e200b/config/ceph.client.admin.keyring 2026-03-06T13:42:36.253 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:42:35 vm00 ceph-mon[51914]: Updating vm08:/var/lib/ceph/82f504c8-1959-11f1-9a16-a78d135e200b/config/ceph.client.admin.keyring 2026-03-06T13:42:36.253 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:42:35 vm00 ceph-mon[51914]: Updating vm00:/var/lib/ceph/82f504c8-1959-11f1-9a16-a78d135e200b/config/ceph.client.admin.keyring 2026-03-06T13:42:36.253 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:42:35 vm00 ceph-mon[51914]: from='mgr.14400 192.168.123.100:0/3201047282' entity='mgr.a' 2026-03-06T13:42:36.253 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:42:35 vm00 ceph-mon[51914]: from='mgr.14400 192.168.123.100:0/3201047282' entity='mgr.a' 2026-03-06T13:42:36.253 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:42:35 vm00 ceph-mon[51914]: from='mgr.14400 192.168.123.100:0/3201047282' entity='mgr.a' 2026-03-06T13:42:36.253 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:42:35 vm00 ceph-mon[51914]: from='mgr.14400 192.168.123.100:0/3201047282' entity='mgr.a' 2026-03-06T13:42:36.253 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:42:35 vm00 ceph-mon[51914]: from='mgr.14400 192.168.123.100:0/3201047282' entity='mgr.a' 2026-03-06T13:42:36.253 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:42:35 vm00 ceph-mon[51914]: from='mgr.14400 192.168.123.100:0/3201047282' entity='mgr.a' 2026-03-06T13:42:36.253 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:42:35 vm00 ceph-mon[51914]: from='mgr.14400 192.168.123.100:0/3201047282' entity='mgr.a' 2026-03-06T13:42:36.253 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:42:35 vm00 ceph-mon[51914]: from='mgr.14400 192.168.123.100:0/3201047282' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-06T13:42:36.253 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:42:35 vm00 ceph-mon[51914]: from='mgr.14400 192.168.123.100:0/3201047282' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-06T13:42:36.253 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:42:35 vm00 ceph-mon[51914]: from='mgr.14400 192.168.123.100:0/3201047282' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-06T13:42:36.253 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:42:35 vm00 ceph-mon[51914]: from='mgr.14400 192.168.123.100:0/3201047282' entity='mgr.a' 2026-03-06T13:42:36.253 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:42:35 vm00 ceph-mon[51914]: mgrmap e19: a(active, since 3s), standbys: b 2026-03-06T13:42:36.253 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:42:35 vm00 ceph-mon[51914]: from='client.? 192.168.123.100:0/3693510475' entity='mgr.a' cmd='[{"prefix": "osd pool application enable","pool": ".rgw.root","app": "rgw"}]': finished 2026-03-06T13:42:36.253 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:42:35 vm00 ceph-mon[51914]: osdmap e22: 3 total, 3 up, 3 in 2026-03-06T13:42:37.905 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:42:37 vm01 ceph-mon[56331]: pgmap v7: 33 pgs: 32 unknown, 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-06T13:42:37.905 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:42:37 vm01 ceph-mon[56331]: osdmap e23: 3 total, 3 up, 3 in 2026-03-06T13:42:37.905 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:42:37 vm01 ceph-mon[56331]: from='client.? 192.168.123.100:0/973456915' entity='mgr.a' cmd=[{"prefix": "osd pool application enable","pool": "myzone1.rgw.log","app": "rgw"}]: dispatch 2026-03-06T13:42:37.905 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:42:37 vm01 ceph-mon[56331]: from='client.? ' entity='mgr.a' cmd=[{"prefix": "osd pool application enable","pool": "myzone1.rgw.log","app": "rgw"}]: dispatch 2026-03-06T13:42:37.905 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:42:37 vm01 ceph-mon[56331]: mgrmap e20: a(active, since 4s), standbys: b 2026-03-06T13:42:38.002 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:42:37 vm00 ceph-mon[51914]: pgmap v7: 33 pgs: 32 unknown, 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-06T13:42:38.002 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:42:37 vm00 ceph-mon[51914]: osdmap e23: 3 total, 3 up, 3 in 2026-03-06T13:42:38.002 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:42:37 vm00 ceph-mon[51914]: from='client.? 192.168.123.100:0/973456915' entity='mgr.a' cmd=[{"prefix": "osd pool application enable","pool": "myzone1.rgw.log","app": "rgw"}]: dispatch 2026-03-06T13:42:38.002 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:42:37 vm00 ceph-mon[51914]: from='client.? ' entity='mgr.a' cmd=[{"prefix": "osd pool application enable","pool": "myzone1.rgw.log","app": "rgw"}]: dispatch 2026-03-06T13:42:38.002 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:42:37 vm00 ceph-mon[51914]: mgrmap e20: a(active, since 4s), standbys: b 2026-03-06T13:42:38.040 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 06 13:42:37 vm08 ceph-mon[53561]: pgmap v7: 33 pgs: 32 unknown, 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-06T13:42:38.040 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 06 13:42:37 vm08 ceph-mon[53561]: osdmap e23: 3 total, 3 up, 3 in 2026-03-06T13:42:38.040 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 06 13:42:37 vm08 ceph-mon[53561]: from='client.? 192.168.123.100:0/973456915' entity='mgr.a' cmd=[{"prefix": "osd pool application enable","pool": "myzone1.rgw.log","app": "rgw"}]: dispatch 2026-03-06T13:42:38.040 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 06 13:42:37 vm08 ceph-mon[53561]: from='client.? ' entity='mgr.a' cmd=[{"prefix": "osd pool application enable","pool": "myzone1.rgw.log","app": "rgw"}]: dispatch 2026-03-06T13:42:38.040 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 06 13:42:37 vm08 ceph-mon[53561]: mgrmap e20: a(active, since 4s), standbys: b 2026-03-06T13:42:39.002 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:42:38 vm00 ceph-mon[51914]: from='client.? ' entity='mgr.a' cmd='[{"prefix": "osd pool application enable","pool": "myzone1.rgw.log","app": "rgw"}]': finished 2026-03-06T13:42:39.002 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:42:38 vm00 ceph-mon[51914]: osdmap e24: 3 total, 3 up, 3 in 2026-03-06T13:42:39.040 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 06 13:42:38 vm08 ceph-mon[53561]: from='client.? ' entity='mgr.a' cmd='[{"prefix": "osd pool application enable","pool": "myzone1.rgw.log","app": "rgw"}]': finished 2026-03-06T13:42:39.040 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 06 13:42:38 vm08 ceph-mon[53561]: osdmap e24: 3 total, 3 up, 3 in 2026-03-06T13:42:39.155 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:42:38 vm01 ceph-mon[56331]: from='client.? ' entity='mgr.a' cmd='[{"prefix": "osd pool application enable","pool": "myzone1.rgw.log","app": "rgw"}]': finished 2026-03-06T13:42:39.155 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:42:38 vm01 ceph-mon[56331]: osdmap e24: 3 total, 3 up, 3 in 2026-03-06T13:42:40.002 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:42:39 vm00 ceph-mon[51914]: pgmap v10: 65 pgs: 64 unknown, 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-06T13:42:40.002 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:42:39 vm00 ceph-mon[51914]: osdmap e25: 3 total, 3 up, 3 in 2026-03-06T13:42:40.002 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:42:39 vm00 ceph-mon[51914]: from='client.? 192.168.123.100:0/973456915' entity='mgr.a' cmd=[{"prefix": "osd pool application enable","pool": "myzone1.rgw.control","app": "rgw"}]: dispatch 2026-03-06T13:42:40.002 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:42:39 vm00 ceph-mon[51914]: from='client.? ' entity='mgr.a' cmd=[{"prefix": "osd pool application enable","pool": "myzone1.rgw.control","app": "rgw"}]: dispatch 2026-03-06T13:42:40.040 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 06 13:42:39 vm08 ceph-mon[53561]: pgmap v10: 65 pgs: 64 unknown, 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-06T13:42:40.040 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 06 13:42:39 vm08 ceph-mon[53561]: osdmap e25: 3 total, 3 up, 3 in 2026-03-06T13:42:40.040 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 06 13:42:39 vm08 ceph-mon[53561]: from='client.? 192.168.123.100:0/973456915' entity='mgr.a' cmd=[{"prefix": "osd pool application enable","pool": "myzone1.rgw.control","app": "rgw"}]: dispatch 2026-03-06T13:42:40.040 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 06 13:42:39 vm08 ceph-mon[53561]: from='client.? ' entity='mgr.a' cmd=[{"prefix": "osd pool application enable","pool": "myzone1.rgw.control","app": "rgw"}]: dispatch 2026-03-06T13:42:40.155 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:42:39 vm01 ceph-mon[56331]: pgmap v10: 65 pgs: 64 unknown, 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-06T13:42:40.155 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:42:39 vm01 ceph-mon[56331]: osdmap e25: 3 total, 3 up, 3 in 2026-03-06T13:42:40.155 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:42:39 vm01 ceph-mon[56331]: from='client.? 192.168.123.100:0/973456915' entity='mgr.a' cmd=[{"prefix": "osd pool application enable","pool": "myzone1.rgw.control","app": "rgw"}]: dispatch 2026-03-06T13:42:40.155 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:42:39 vm01 ceph-mon[56331]: from='client.? ' entity='mgr.a' cmd=[{"prefix": "osd pool application enable","pool": "myzone1.rgw.control","app": "rgw"}]: dispatch 2026-03-06T13:42:41.002 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:42:40 vm00 ceph-mon[51914]: from='client.? ' entity='mgr.a' cmd='[{"prefix": "osd pool application enable","pool": "myzone1.rgw.control","app": "rgw"}]': finished 2026-03-06T13:42:41.002 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:42:40 vm00 ceph-mon[51914]: osdmap e26: 3 total, 3 up, 3 in 2026-03-06T13:42:41.040 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 06 13:42:40 vm08 ceph-mon[53561]: from='client.? ' entity='mgr.a' cmd='[{"prefix": "osd pool application enable","pool": "myzone1.rgw.control","app": "rgw"}]': finished 2026-03-06T13:42:41.040 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 06 13:42:40 vm08 ceph-mon[53561]: osdmap e26: 3 total, 3 up, 3 in 2026-03-06T13:42:41.155 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:42:40 vm01 ceph-mon[56331]: from='client.? ' entity='mgr.a' cmd='[{"prefix": "osd pool application enable","pool": "myzone1.rgw.control","app": "rgw"}]': finished 2026-03-06T13:42:41.155 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:42:40 vm01 ceph-mon[56331]: osdmap e26: 3 total, 3 up, 3 in 2026-03-06T13:42:42.002 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:42:41 vm00 ceph-mon[51914]: pgmap v13: 97 pgs: 32 unknown, 65 active+clean; 451 KiB data, 81 MiB used, 60 GiB / 60 GiB avail; 51 KiB/s rd, 3.5 KiB/s wr, 31 op/s 2026-03-06T13:42:42.002 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:42:41 vm00 ceph-mon[51914]: osdmap e27: 3 total, 3 up, 3 in 2026-03-06T13:42:42.002 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:42:41 vm00 ceph-mon[51914]: from='client.? 192.168.123.100:0/973456915' entity='mgr.a' cmd=[{"prefix": "osd pool application enable","pool": "myzone1.rgw.meta","app": "rgw"}]: dispatch 2026-03-06T13:42:42.002 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:42:41 vm00 ceph-mon[51914]: from='client.? ' entity='mgr.a' cmd=[{"prefix": "osd pool application enable","pool": "myzone1.rgw.meta","app": "rgw"}]: dispatch 2026-03-06T13:42:42.002 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:42:41 vm00 ceph-mon[51914]: from='client.? ' entity='mgr.a' cmd='[{"prefix": "osd pool application enable","pool": "myzone1.rgw.meta","app": "rgw"}]': finished 2026-03-06T13:42:42.002 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:42:41 vm00 ceph-mon[51914]: osdmap e28: 3 total, 3 up, 3 in 2026-03-06T13:42:42.002 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:42:41 vm00 ceph-mon[51914]: from='client.? 192.168.123.100:0/973456915' entity='mgr.a' cmd=[{"prefix": "osd pool set", "pool": "myzone1.rgw.meta", "var": "pg_autoscale_bias", "val": "4"}]: dispatch 2026-03-06T13:42:42.039 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 06 13:42:41 vm08 ceph-mon[53561]: pgmap v13: 97 pgs: 32 unknown, 65 active+clean; 451 KiB data, 81 MiB used, 60 GiB / 60 GiB avail; 51 KiB/s rd, 3.5 KiB/s wr, 31 op/s 2026-03-06T13:42:42.040 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 06 13:42:41 vm08 ceph-mon[53561]: osdmap e27: 3 total, 3 up, 3 in 2026-03-06T13:42:42.040 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 06 13:42:41 vm08 ceph-mon[53561]: from='client.? 192.168.123.100:0/973456915' entity='mgr.a' cmd=[{"prefix": "osd pool application enable","pool": "myzone1.rgw.meta","app": "rgw"}]: dispatch 2026-03-06T13:42:42.040 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 06 13:42:41 vm08 ceph-mon[53561]: from='client.? ' entity='mgr.a' cmd=[{"prefix": "osd pool application enable","pool": "myzone1.rgw.meta","app": "rgw"}]: dispatch 2026-03-06T13:42:42.040 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 06 13:42:41 vm08 ceph-mon[53561]: from='client.? ' entity='mgr.a' cmd='[{"prefix": "osd pool application enable","pool": "myzone1.rgw.meta","app": "rgw"}]': finished 2026-03-06T13:42:42.040 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 06 13:42:41 vm08 ceph-mon[53561]: osdmap e28: 3 total, 3 up, 3 in 2026-03-06T13:42:42.040 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 06 13:42:41 vm08 ceph-mon[53561]: from='client.? 192.168.123.100:0/973456915' entity='mgr.a' cmd=[{"prefix": "osd pool set", "pool": "myzone1.rgw.meta", "var": "pg_autoscale_bias", "val": "4"}]: dispatch 2026-03-06T13:42:42.155 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:42:41 vm01 ceph-mon[56331]: pgmap v13: 97 pgs: 32 unknown, 65 active+clean; 451 KiB data, 81 MiB used, 60 GiB / 60 GiB avail; 51 KiB/s rd, 3.5 KiB/s wr, 31 op/s 2026-03-06T13:42:42.155 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:42:41 vm01 ceph-mon[56331]: osdmap e27: 3 total, 3 up, 3 in 2026-03-06T13:42:42.155 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:42:41 vm01 ceph-mon[56331]: from='client.? 192.168.123.100:0/973456915' entity='mgr.a' cmd=[{"prefix": "osd pool application enable","pool": "myzone1.rgw.meta","app": "rgw"}]: dispatch 2026-03-06T13:42:42.155 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:42:41 vm01 ceph-mon[56331]: from='client.? ' entity='mgr.a' cmd=[{"prefix": "osd pool application enable","pool": "myzone1.rgw.meta","app": "rgw"}]: dispatch 2026-03-06T13:42:42.155 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:42:41 vm01 ceph-mon[56331]: from='client.? ' entity='mgr.a' cmd='[{"prefix": "osd pool application enable","pool": "myzone1.rgw.meta","app": "rgw"}]': finished 2026-03-06T13:42:42.155 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:42:41 vm01 ceph-mon[56331]: osdmap e28: 3 total, 3 up, 3 in 2026-03-06T13:42:42.155 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:42:41 vm01 ceph-mon[56331]: from='client.? 192.168.123.100:0/973456915' entity='mgr.a' cmd=[{"prefix": "osd pool set", "pool": "myzone1.rgw.meta", "var": "pg_autoscale_bias", "val": "4"}]: dispatch 2026-03-06T13:42:43.002 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:42:42 vm00 ceph-mon[51914]: from='client.? ' entity='mgr.a' cmd=[{"prefix": "osd pool set", "pool": "myzone1.rgw.meta", "var": "pg_autoscale_bias", "val": "4"}]: dispatch 2026-03-06T13:42:43.040 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 06 13:42:42 vm08 ceph-mon[53561]: from='client.? ' entity='mgr.a' cmd=[{"prefix": "osd pool set", "pool": "myzone1.rgw.meta", "var": "pg_autoscale_bias", "val": "4"}]: dispatch 2026-03-06T13:42:43.088 INFO:teuthology.orchestra.run.vm00.stdout:Realm(s) created correctly. Please, use 'ceph rgw realm tokens' to get the token. 2026-03-06T13:42:43.155 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:42:42 vm01 ceph-mon[56331]: from='client.? ' entity='mgr.a' cmd=[{"prefix": "osd pool set", "pool": "myzone1.rgw.meta", "var": "pg_autoscale_bias", "val": "4"}]: dispatch 2026-03-06T13:42:43.180 INFO:teuthology.run_tasks:Running task cephadm.shell... 2026-03-06T13:42:43.182 INFO:tasks.cephadm:Running commands on role host.a host ubuntu@vm00.local 2026-03-06T13:42:43.182 DEBUG:teuthology.orchestra.run.vm00:> sudo /home/ubuntu/cephtest/cephadm --image harbor.clyso.com/custom-ceph/ceph/ceph:cobaltcore-storage-v19.2.3-fasttrack-3 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 82f504c8-1959-11f1-9a16-a78d135e200b -- bash -c 'set -e 2026-03-06T13:42:43.182 DEBUG:teuthology.orchestra.run.vm00:> set -x 2026-03-06T13:42:43.182 DEBUG:teuthology.orchestra.run.vm00:> while true; do TOKEN=$(ceph rgw realm tokens | jq -r '"'"'.[0].token'"'"'); echo $TOKEN; if [ "$TOKEN" != "master zone has no endpoint" ]; then break; fi; sleep 5; done 2026-03-06T13:42:43.182 DEBUG:teuthology.orchestra.run.vm00:> TOKENS=$(ceph rgw realm tokens) 2026-03-06T13:42:43.182 DEBUG:teuthology.orchestra.run.vm00:> echo $TOKENS | jq --exit-status '"'"'.[0].realm == "myrealm1"'"'"' 2026-03-06T13:42:43.183 DEBUG:teuthology.orchestra.run.vm00:> echo $TOKENS | jq --exit-status '"'"'.[0].token'"'"' 2026-03-06T13:42:43.183 DEBUG:teuthology.orchestra.run.vm00:> TOKEN_JSON=$(ceph rgw realm tokens | jq -r '"'"'.[0].token'"'"' | base64 --decode) 2026-03-06T13:42:43.183 DEBUG:teuthology.orchestra.run.vm00:> echo $TOKEN_JSON | jq --exit-status '"'"'.realm_name == "myrealm1"'"'"' 2026-03-06T13:42:43.183 DEBUG:teuthology.orchestra.run.vm00:> echo $TOKEN_JSON | jq --exit-status '"'"'.endpoint | test("http://.+:\\d+")'"'"' 2026-03-06T13:42:43.183 DEBUG:teuthology.orchestra.run.vm00:> echo $TOKEN_JSON | jq --exit-status '"'"'.realm_id | test("^[0-9a-f]{8}-[0-9a-f]{4}-[0-9a-f]{4}-[0-9a-f]{4}-[0-9a-f]{12}$")'"'"' 2026-03-06T13:42:43.183 DEBUG:teuthology.orchestra.run.vm00:> echo $TOKEN_JSON | jq --exit-status '"'"'.access_key'"'"' 2026-03-06T13:42:43.183 DEBUG:teuthology.orchestra.run.vm00:> echo $TOKEN_JSON | jq --exit-status '"'"'.secret'"'"' 2026-03-06T13:42:43.183 DEBUG:teuthology.orchestra.run.vm00:> ' 2026-03-06T13:42:43.537 INFO:teuthology.orchestra.run.vm00.stderr:Inferring config /var/lib/ceph/82f504c8-1959-11f1-9a16-a78d135e200b/mon.a/config 2026-03-06T13:42:43.628 INFO:teuthology.orchestra.run.vm00.stderr:+ true 2026-03-06T13:42:43.628 INFO:teuthology.orchestra.run.vm00.stderr:++ ceph rgw realm tokens 2026-03-06T13:42:43.630 INFO:teuthology.orchestra.run.vm00.stderr:++ jq -r '.[0].token' 2026-03-06T13:42:43.800 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:42:43 vm01 ceph-mon[56331]: pgmap v16: 129 pgs: 64 unknown, 65 active+clean; 451 KiB data, 81 MiB used, 60 GiB / 60 GiB avail; 51 KiB/s rd, 3.5 KiB/s wr, 31 op/s 2026-03-06T13:42:43.800 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:42:43 vm01 ceph-mon[56331]: from='client.? ' entity='mgr.a' cmd='[{"prefix": "osd pool set", "pool": "myzone1.rgw.meta", "var": "pg_autoscale_bias", "val": "4"}]': finished 2026-03-06T13:42:43.800 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:42:43 vm01 ceph-mon[56331]: osdmap e29: 3 total, 3 up, 3 in 2026-03-06T13:42:43.800 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:42:43 vm01 ceph-mon[56331]: Saving service rgw.myrealm1.myzone1 spec with placement count:2 2026-03-06T13:42:43.800 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:42:43 vm01 ceph-mon[56331]: from='mgr.14400 192.168.123.100:0/3201047282' entity='mgr.a' 2026-03-06T13:42:43.800 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:42:43 vm01 ceph-mon[56331]: from='mgr.14400 192.168.123.100:0/3201047282' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-06T13:42:43.800 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:42:43 vm01 ceph-mon[56331]: from='mgr.14400 192.168.123.100:0/3201047282' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-06T13:42:43.800 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:42:43 vm01 ceph-mon[56331]: from='mgr.14400 192.168.123.100:0/3201047282' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-06T13:42:43.800 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:42:43 vm01 ceph-mon[56331]: from='mgr.14400 192.168.123.100:0/3201047282' entity='mgr.a' 2026-03-06T13:42:43.800 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:42:43 vm01 ceph-mon[56331]: from='mgr.14400 192.168.123.100:0/3201047282' entity='mgr.a' 2026-03-06T13:42:43.800 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:42:43 vm01 ceph-mon[56331]: from='mgr.14400 192.168.123.100:0/3201047282' entity='mgr.a' 2026-03-06T13:42:43.800 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:42:43 vm01 ceph-mon[56331]: from='mgr.14400 192.168.123.100:0/3201047282' entity='mgr.a' 2026-03-06T13:42:43.800 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:42:43 vm01 ceph-mon[56331]: from='mgr.14400 192.168.123.100:0/3201047282' entity='mgr.a' cmd=[{"prefix": "auth get-or-create", "entity": "client.rgw.myrealm1.myzone1.vm01.mmqnav", "caps": ["mon", "allow *", "mgr", "allow rw", "osd", "allow rwx tag rgw *=*"]}]: dispatch 2026-03-06T13:42:43.800 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:42:43 vm01 ceph-mon[56331]: from='mgr.14400 192.168.123.100:0/3201047282' entity='mgr.a' cmd='[{"prefix": "auth get-or-create", "entity": "client.rgw.myrealm1.myzone1.vm01.mmqnav", "caps": ["mon", "allow *", "mgr", "allow rw", "osd", "allow rwx tag rgw *=*"]}]': finished 2026-03-06T13:42:43.800 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:42:43 vm01 ceph-mon[56331]: from='mgr.14400 192.168.123.100:0/3201047282' entity='mgr.a' 2026-03-06T13:42:43.800 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:42:43 vm01 ceph-mon[56331]: from='mgr.14400 192.168.123.100:0/3201047282' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-06T13:42:43.800 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:42:43 vm01 ceph-mon[56331]: Deploying daemon rgw.myrealm1.myzone1.vm01.mmqnav on vm01 2026-03-06T13:42:44.002 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:42:43 vm00 ceph-mon[51914]: pgmap v16: 129 pgs: 64 unknown, 65 active+clean; 451 KiB data, 81 MiB used, 60 GiB / 60 GiB avail; 51 KiB/s rd, 3.5 KiB/s wr, 31 op/s 2026-03-06T13:42:44.002 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:42:43 vm00 ceph-mon[51914]: from='client.? ' entity='mgr.a' cmd='[{"prefix": "osd pool set", "pool": "myzone1.rgw.meta", "var": "pg_autoscale_bias", "val": "4"}]': finished 2026-03-06T13:42:44.002 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:42:43 vm00 ceph-mon[51914]: osdmap e29: 3 total, 3 up, 3 in 2026-03-06T13:42:44.002 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:42:43 vm00 ceph-mon[51914]: Saving service rgw.myrealm1.myzone1 spec with placement count:2 2026-03-06T13:42:44.002 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:42:43 vm00 ceph-mon[51914]: from='mgr.14400 192.168.123.100:0/3201047282' entity='mgr.a' 2026-03-06T13:42:44.002 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:42:43 vm00 ceph-mon[51914]: from='mgr.14400 192.168.123.100:0/3201047282' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-06T13:42:44.002 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:42:43 vm00 ceph-mon[51914]: from='mgr.14400 192.168.123.100:0/3201047282' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-06T13:42:44.002 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:42:43 vm00 ceph-mon[51914]: from='mgr.14400 192.168.123.100:0/3201047282' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-06T13:42:44.002 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:42:43 vm00 ceph-mon[51914]: from='mgr.14400 192.168.123.100:0/3201047282' entity='mgr.a' 2026-03-06T13:42:44.002 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:42:43 vm00 ceph-mon[51914]: from='mgr.14400 192.168.123.100:0/3201047282' entity='mgr.a' 2026-03-06T13:42:44.002 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:42:43 vm00 ceph-mon[51914]: from='mgr.14400 192.168.123.100:0/3201047282' entity='mgr.a' 2026-03-06T13:42:44.002 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:42:43 vm00 ceph-mon[51914]: from='mgr.14400 192.168.123.100:0/3201047282' entity='mgr.a' 2026-03-06T13:42:44.002 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:42:43 vm00 ceph-mon[51914]: from='mgr.14400 192.168.123.100:0/3201047282' entity='mgr.a' cmd=[{"prefix": "auth get-or-create", "entity": "client.rgw.myrealm1.myzone1.vm01.mmqnav", "caps": ["mon", "allow *", "mgr", "allow rw", "osd", "allow rwx tag rgw *=*"]}]: dispatch 2026-03-06T13:42:44.002 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:42:43 vm00 ceph-mon[51914]: from='mgr.14400 192.168.123.100:0/3201047282' entity='mgr.a' cmd='[{"prefix": "auth get-or-create", "entity": "client.rgw.myrealm1.myzone1.vm01.mmqnav", "caps": ["mon", "allow *", "mgr", "allow rw", "osd", "allow rwx tag rgw *=*"]}]': finished 2026-03-06T13:42:44.002 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:42:43 vm00 ceph-mon[51914]: from='mgr.14400 192.168.123.100:0/3201047282' entity='mgr.a' 2026-03-06T13:42:44.002 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:42:43 vm00 ceph-mon[51914]: from='mgr.14400 192.168.123.100:0/3201047282' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-06T13:42:44.002 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:42:43 vm00 ceph-mon[51914]: Deploying daemon rgw.myrealm1.myzone1.vm01.mmqnav on vm01 2026-03-06T13:42:44.040 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 06 13:42:43 vm08 ceph-mon[53561]: pgmap v16: 129 pgs: 64 unknown, 65 active+clean; 451 KiB data, 81 MiB used, 60 GiB / 60 GiB avail; 51 KiB/s rd, 3.5 KiB/s wr, 31 op/s 2026-03-06T13:42:44.040 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 06 13:42:43 vm08 ceph-mon[53561]: from='client.? ' entity='mgr.a' cmd='[{"prefix": "osd pool set", "pool": "myzone1.rgw.meta", "var": "pg_autoscale_bias", "val": "4"}]': finished 2026-03-06T13:42:44.040 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 06 13:42:43 vm08 ceph-mon[53561]: osdmap e29: 3 total, 3 up, 3 in 2026-03-06T13:42:44.040 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 06 13:42:43 vm08 ceph-mon[53561]: Saving service rgw.myrealm1.myzone1 spec with placement count:2 2026-03-06T13:42:44.040 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 06 13:42:43 vm08 ceph-mon[53561]: from='mgr.14400 192.168.123.100:0/3201047282' entity='mgr.a' 2026-03-06T13:42:44.040 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 06 13:42:43 vm08 ceph-mon[53561]: from='mgr.14400 192.168.123.100:0/3201047282' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-06T13:42:44.040 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 06 13:42:43 vm08 ceph-mon[53561]: from='mgr.14400 192.168.123.100:0/3201047282' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-06T13:42:44.040 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 06 13:42:43 vm08 ceph-mon[53561]: from='mgr.14400 192.168.123.100:0/3201047282' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-06T13:42:44.040 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 06 13:42:43 vm08 ceph-mon[53561]: from='mgr.14400 192.168.123.100:0/3201047282' entity='mgr.a' 2026-03-06T13:42:44.040 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 06 13:42:43 vm08 ceph-mon[53561]: from='mgr.14400 192.168.123.100:0/3201047282' entity='mgr.a' 2026-03-06T13:42:44.040 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 06 13:42:43 vm08 ceph-mon[53561]: from='mgr.14400 192.168.123.100:0/3201047282' entity='mgr.a' 2026-03-06T13:42:44.040 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 06 13:42:43 vm08 ceph-mon[53561]: from='mgr.14400 192.168.123.100:0/3201047282' entity='mgr.a' 2026-03-06T13:42:44.040 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 06 13:42:43 vm08 ceph-mon[53561]: from='mgr.14400 192.168.123.100:0/3201047282' entity='mgr.a' cmd=[{"prefix": "auth get-or-create", "entity": "client.rgw.myrealm1.myzone1.vm01.mmqnav", "caps": ["mon", "allow *", "mgr", "allow rw", "osd", "allow rwx tag rgw *=*"]}]: dispatch 2026-03-06T13:42:44.040 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 06 13:42:43 vm08 ceph-mon[53561]: from='mgr.14400 192.168.123.100:0/3201047282' entity='mgr.a' cmd='[{"prefix": "auth get-or-create", "entity": "client.rgw.myrealm1.myzone1.vm01.mmqnav", "caps": ["mon", "allow *", "mgr", "allow rw", "osd", "allow rwx tag rgw *=*"]}]': finished 2026-03-06T13:42:44.040 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 06 13:42:43 vm08 ceph-mon[53561]: from='mgr.14400 192.168.123.100:0/3201047282' entity='mgr.a' 2026-03-06T13:42:44.040 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 06 13:42:43 vm08 ceph-mon[53561]: from='mgr.14400 192.168.123.100:0/3201047282' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-06T13:42:44.040 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 06 13:42:43 vm08 ceph-mon[53561]: Deploying daemon rgw.myrealm1.myzone1.vm01.mmqnav on vm01 2026-03-06T13:42:44.443 INFO:teuthology.orchestra.run.vm00.stdout:master zone has no endpoint 2026-03-06T13:42:44.443 INFO:teuthology.orchestra.run.vm00.stderr:+ TOKEN='master zone has no endpoint' 2026-03-06T13:42:44.443 INFO:teuthology.orchestra.run.vm00.stderr:+ echo master zone has no endpoint 2026-03-06T13:42:44.443 INFO:teuthology.orchestra.run.vm00.stderr:+ '[' 'master zone has no endpoint' '!=' 'master zone has no endpoint' ']' 2026-03-06T13:42:44.443 INFO:teuthology.orchestra.run.vm00.stderr:+ sleep 5 2026-03-06T13:42:45.281 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 06 13:42:45 vm08 ceph-mon[53561]: from='client.14592 -' entity='client.admin' cmd=[{"prefix": "rgw realm tokens", "target": ["mon-mgr", ""]}]: dispatch 2026-03-06T13:42:45.281 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 06 13:42:45 vm08 ceph-mon[53561]: from='mgr.14400 192.168.123.100:0/3201047282' entity='mgr.a' 2026-03-06T13:42:45.281 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 06 13:42:45 vm08 ceph-mon[53561]: from='mgr.14400 192.168.123.100:0/3201047282' entity='mgr.a' 2026-03-06T13:42:45.281 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 06 13:42:45 vm08 ceph-mon[53561]: from='mgr.14400 192.168.123.100:0/3201047282' entity='mgr.a' 2026-03-06T13:42:45.281 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 06 13:42:45 vm08 ceph-mon[53561]: from='mgr.14400 192.168.123.100:0/3201047282' entity='mgr.a' cmd=[{"prefix": "auth get-or-create", "entity": "client.rgw.myrealm1.myzone1.vm08.oycfkw", "caps": ["mon", "allow *", "mgr", "allow rw", "osd", "allow rwx tag rgw *=*"]}]: dispatch 2026-03-06T13:42:45.281 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 06 13:42:45 vm08 ceph-mon[53561]: from='mgr.14400 192.168.123.100:0/3201047282' entity='mgr.a' cmd='[{"prefix": "auth get-or-create", "entity": "client.rgw.myrealm1.myzone1.vm08.oycfkw", "caps": ["mon", "allow *", "mgr", "allow rw", "osd", "allow rwx tag rgw *=*"]}]': finished 2026-03-06T13:42:45.281 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 06 13:42:45 vm08 ceph-mon[53561]: from='mgr.14400 192.168.123.100:0/3201047282' entity='mgr.a' 2026-03-06T13:42:45.281 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 06 13:42:45 vm08 ceph-mon[53561]: from='mgr.14400 192.168.123.100:0/3201047282' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-06T13:42:45.281 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 06 13:42:45 vm08 ceph-mon[53561]: Deploying daemon rgw.myrealm1.myzone1.vm08.oycfkw on vm08 2026-03-06T13:42:45.502 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:42:45 vm00 ceph-mon[51914]: from='client.14592 -' entity='client.admin' cmd=[{"prefix": "rgw realm tokens", "target": ["mon-mgr", ""]}]: dispatch 2026-03-06T13:42:45.502 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:42:45 vm00 ceph-mon[51914]: from='mgr.14400 192.168.123.100:0/3201047282' entity='mgr.a' 2026-03-06T13:42:45.502 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:42:45 vm00 ceph-mon[51914]: from='mgr.14400 192.168.123.100:0/3201047282' entity='mgr.a' 2026-03-06T13:42:45.502 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:42:45 vm00 ceph-mon[51914]: from='mgr.14400 192.168.123.100:0/3201047282' entity='mgr.a' 2026-03-06T13:42:45.502 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:42:45 vm00 ceph-mon[51914]: from='mgr.14400 192.168.123.100:0/3201047282' entity='mgr.a' cmd=[{"prefix": "auth get-or-create", "entity": "client.rgw.myrealm1.myzone1.vm08.oycfkw", "caps": ["mon", "allow *", "mgr", "allow rw", "osd", "allow rwx tag rgw *=*"]}]: dispatch 2026-03-06T13:42:45.502 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:42:45 vm00 ceph-mon[51914]: from='mgr.14400 192.168.123.100:0/3201047282' entity='mgr.a' cmd='[{"prefix": "auth get-or-create", "entity": "client.rgw.myrealm1.myzone1.vm08.oycfkw", "caps": ["mon", "allow *", "mgr", "allow rw", "osd", "allow rwx tag rgw *=*"]}]': finished 2026-03-06T13:42:45.502 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:42:45 vm00 ceph-mon[51914]: from='mgr.14400 192.168.123.100:0/3201047282' entity='mgr.a' 2026-03-06T13:42:45.502 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:42:45 vm00 ceph-mon[51914]: from='mgr.14400 192.168.123.100:0/3201047282' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-06T13:42:45.502 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:42:45 vm00 ceph-mon[51914]: Deploying daemon rgw.myrealm1.myzone1.vm08.oycfkw on vm08 2026-03-06T13:42:45.655 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:42:45 vm01 ceph-mon[56331]: from='client.14592 -' entity='client.admin' cmd=[{"prefix": "rgw realm tokens", "target": ["mon-mgr", ""]}]: dispatch 2026-03-06T13:42:45.655 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:42:45 vm01 ceph-mon[56331]: from='mgr.14400 192.168.123.100:0/3201047282' entity='mgr.a' 2026-03-06T13:42:45.655 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:42:45 vm01 ceph-mon[56331]: from='mgr.14400 192.168.123.100:0/3201047282' entity='mgr.a' 2026-03-06T13:42:45.655 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:42:45 vm01 ceph-mon[56331]: from='mgr.14400 192.168.123.100:0/3201047282' entity='mgr.a' 2026-03-06T13:42:45.655 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:42:45 vm01 ceph-mon[56331]: from='mgr.14400 192.168.123.100:0/3201047282' entity='mgr.a' cmd=[{"prefix": "auth get-or-create", "entity": "client.rgw.myrealm1.myzone1.vm08.oycfkw", "caps": ["mon", "allow *", "mgr", "allow rw", "osd", "allow rwx tag rgw *=*"]}]: dispatch 2026-03-06T13:42:45.655 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:42:45 vm01 ceph-mon[56331]: from='mgr.14400 192.168.123.100:0/3201047282' entity='mgr.a' cmd='[{"prefix": "auth get-or-create", "entity": "client.rgw.myrealm1.myzone1.vm08.oycfkw", "caps": ["mon", "allow *", "mgr", "allow rw", "osd", "allow rwx tag rgw *=*"]}]': finished 2026-03-06T13:42:45.655 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:42:45 vm01 ceph-mon[56331]: from='mgr.14400 192.168.123.100:0/3201047282' entity='mgr.a' 2026-03-06T13:42:45.655 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:42:45 vm01 ceph-mon[56331]: from='mgr.14400 192.168.123.100:0/3201047282' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-06T13:42:45.655 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:42:45 vm01 ceph-mon[56331]: Deploying daemon rgw.myrealm1.myzone1.vm08.oycfkw on vm08 2026-03-06T13:42:46.540 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 06 13:42:46 vm08 ceph-mon[53561]: pgmap v18: 129 pgs: 129 active+clean; 455 KiB data, 81 MiB used, 60 GiB / 60 GiB avail; 16 KiB/s rd, 5.1 KiB/s wr, 22 op/s 2026-03-06T13:42:46.540 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 06 13:42:46 vm08 ceph-mon[53561]: from='mgr.14400 192.168.123.100:0/3201047282' entity='mgr.a' 2026-03-06T13:42:46.540 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 06 13:42:46 vm08 ceph-mon[53561]: from='mgr.14400 192.168.123.100:0/3201047282' entity='mgr.a' 2026-03-06T13:42:46.540 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 06 13:42:46 vm08 ceph-mon[53561]: from='mgr.14400 192.168.123.100:0/3201047282' entity='mgr.a' 2026-03-06T13:42:46.540 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 06 13:42:46 vm08 ceph-mon[53561]: from='mgr.14400 192.168.123.100:0/3201047282' entity='mgr.a' cmd=[{"prefix": "rgw zone modify", "realm_name": "myrealm1", "zonegroup_name": "myzonegroup1", "zone_name": "myzone1", "realm_token": "ewogICAgInJlYWxtX25hbWUiOiAibXlyZWFsbTEiLAogICAgInJlYWxtX2lkIjogIjE1ZDk3ZTBjLTM4MzEtNDllYy1hMWUwLTdhYjE3NWExZmFiOCIsCiAgICAiZW5kcG9pbnQiOiBudWxsLAogICAgImFjY2Vzc19rZXkiOiAiNldJQ0hWUUROM0JZMURMMTEwUjMiLAogICAgInNlY3JldCI6ICI3RHl2a2VoVXhNMnVkbDlxSUhZNEE0T053Mmd5ejBLNGw4eW9xZ1B4Igp9", "zone_endpoints": ["http://192.168.123.101:5500", "http://192.168.123.108:5500"]}]: dispatch 2026-03-06T13:42:46.540 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 06 13:42:46 vm08 ceph-mon[53561]: from='mgr.14400 192.168.123.100:0/3201047282' entity='mgr.a' 2026-03-06T13:42:46.540 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 06 13:42:46 vm08 ceph-mon[53561]: from='mgr.14400 192.168.123.100:0/3201047282' entity='mgr.a' 2026-03-06T13:42:46.540 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 06 13:42:46 vm08 ceph-mon[53561]: from='mgr.14400 192.168.123.100:0/3201047282' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-06T13:42:46.655 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:42:46 vm01 ceph-mon[56331]: pgmap v18: 129 pgs: 129 active+clean; 455 KiB data, 81 MiB used, 60 GiB / 60 GiB avail; 16 KiB/s rd, 5.1 KiB/s wr, 22 op/s 2026-03-06T13:42:46.655 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:42:46 vm01 ceph-mon[56331]: from='mgr.14400 192.168.123.100:0/3201047282' entity='mgr.a' 2026-03-06T13:42:46.655 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:42:46 vm01 ceph-mon[56331]: from='mgr.14400 192.168.123.100:0/3201047282' entity='mgr.a' 2026-03-06T13:42:46.655 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:42:46 vm01 ceph-mon[56331]: from='mgr.14400 192.168.123.100:0/3201047282' entity='mgr.a' 2026-03-06T13:42:46.655 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:42:46 vm01 ceph-mon[56331]: from='mgr.14400 192.168.123.100:0/3201047282' entity='mgr.a' cmd=[{"prefix": "rgw zone modify", "realm_name": "myrealm1", "zonegroup_name": "myzonegroup1", "zone_name": "myzone1", "realm_token": "ewogICAgInJlYWxtX25hbWUiOiAibXlyZWFsbTEiLAogICAgInJlYWxtX2lkIjogIjE1ZDk3ZTBjLTM4MzEtNDllYy1hMWUwLTdhYjE3NWExZmFiOCIsCiAgICAiZW5kcG9pbnQiOiBudWxsLAogICAgImFjY2Vzc19rZXkiOiAiNldJQ0hWUUROM0JZMURMMTEwUjMiLAogICAgInNlY3JldCI6ICI3RHl2a2VoVXhNMnVkbDlxSUhZNEE0T053Mmd5ejBLNGw4eW9xZ1B4Igp9", "zone_endpoints": ["http://192.168.123.101:5500", "http://192.168.123.108:5500"]}]: dispatch 2026-03-06T13:42:46.655 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:42:46 vm01 ceph-mon[56331]: from='mgr.14400 192.168.123.100:0/3201047282' entity='mgr.a' 2026-03-06T13:42:46.655 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:42:46 vm01 ceph-mon[56331]: from='mgr.14400 192.168.123.100:0/3201047282' entity='mgr.a' 2026-03-06T13:42:46.655 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:42:46 vm01 ceph-mon[56331]: from='mgr.14400 192.168.123.100:0/3201047282' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-06T13:42:46.752 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:42:46 vm00 ceph-mon[51914]: pgmap v18: 129 pgs: 129 active+clean; 455 KiB data, 81 MiB used, 60 GiB / 60 GiB avail; 16 KiB/s rd, 5.1 KiB/s wr, 22 op/s 2026-03-06T13:42:46.752 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:42:46 vm00 ceph-mon[51914]: from='mgr.14400 192.168.123.100:0/3201047282' entity='mgr.a' 2026-03-06T13:42:46.752 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:42:46 vm00 ceph-mon[51914]: from='mgr.14400 192.168.123.100:0/3201047282' entity='mgr.a' 2026-03-06T13:42:46.752 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:42:46 vm00 ceph-mon[51914]: from='mgr.14400 192.168.123.100:0/3201047282' entity='mgr.a' 2026-03-06T13:42:46.752 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:42:46 vm00 ceph-mon[51914]: from='mgr.14400 192.168.123.100:0/3201047282' entity='mgr.a' cmd=[{"prefix": "rgw zone modify", "realm_name": "myrealm1", "zonegroup_name": "myzonegroup1", "zone_name": "myzone1", "realm_token": "ewogICAgInJlYWxtX25hbWUiOiAibXlyZWFsbTEiLAogICAgInJlYWxtX2lkIjogIjE1ZDk3ZTBjLTM4MzEtNDllYy1hMWUwLTdhYjE3NWExZmFiOCIsCiAgICAiZW5kcG9pbnQiOiBudWxsLAogICAgImFjY2Vzc19rZXkiOiAiNldJQ0hWUUROM0JZMURMMTEwUjMiLAogICAgInNlY3JldCI6ICI3RHl2a2VoVXhNMnVkbDlxSUhZNEE0T053Mmd5ejBLNGw4eW9xZ1B4Igp9", "zone_endpoints": ["http://192.168.123.101:5500", "http://192.168.123.108:5500"]}]: dispatch 2026-03-06T13:42:46.752 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:42:46 vm00 ceph-mon[51914]: from='mgr.14400 192.168.123.100:0/3201047282' entity='mgr.a' 2026-03-06T13:42:46.752 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:42:46 vm00 ceph-mon[51914]: from='mgr.14400 192.168.123.100:0/3201047282' entity='mgr.a' 2026-03-06T13:42:46.752 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:42:46 vm00 ceph-mon[51914]: from='mgr.14400 192.168.123.100:0/3201047282' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-06T13:42:47.540 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 06 13:42:47 vm08 ceph-mon[53561]: from='mon.0 -' entity='mon.' cmd=[{"prefix": "rgw zone modify", "realm_name": "myrealm1", "zonegroup_name": "myzonegroup1", "zone_name": "myzone1", "realm_token": "ewogICAgInJlYWxtX25hbWUiOiAibXlyZWFsbTEiLAogICAgInJlYWxtX2lkIjogIjE1ZDk3ZTBjLTM4MzEtNDllYy1hMWUwLTdhYjE3NWExZmFiOCIsCiAgICAiZW5kcG9pbnQiOiBudWxsLAogICAgImFjY2Vzc19rZXkiOiAiNldJQ0hWUUROM0JZMURMMTEwUjMiLAogICAgInNlY3JldCI6ICI3RHl2a2VoVXhNMnVkbDlxSUhZNEE0T053Mmd5ejBLNGw4eW9xZ1B4Igp9", "zone_endpoints": ["http://192.168.123.101:5500", "http://192.168.123.108:5500"]}]: dispatch 2026-03-06T13:42:47.540 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 06 13:42:47 vm08 ceph-mon[53561]: Saving service rgw.myrealm1.myzone1 spec with placement count:2 2026-03-06T13:42:47.540 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 06 13:42:47 vm08 ceph-mon[53561]: from='mgr.14400 192.168.123.100:0/3201047282' entity='mgr.a' 2026-03-06T13:42:47.540 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 06 13:42:47 vm08 ceph-mon[53561]: from='mgr.14400 192.168.123.100:0/3201047282' entity='mgr.a' 2026-03-06T13:42:47.540 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 06 13:42:47 vm08 ceph-mon[53561]: from='mgr.14400 192.168.123.100:0/3201047282' entity='mgr.a' 2026-03-06T13:42:47.540 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 06 13:42:47 vm08 ceph-mon[53561]: from='mgr.14400 192.168.123.100:0/3201047282' entity='mgr.a' 2026-03-06T13:42:47.540 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 06 13:42:47 vm08 ceph-mon[53561]: from='mgr.14400 192.168.123.100:0/3201047282' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-06T13:42:47.540 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 06 13:42:47 vm08 ceph-mon[53561]: from='mgr.14400 192.168.123.100:0/3201047282' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-06T13:42:47.540 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 06 13:42:47 vm08 ceph-mon[53561]: from='mgr.14400 192.168.123.100:0/3201047282' entity='mgr.a' 2026-03-06T13:42:47.540 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 06 13:42:47 vm08 ceph-mon[53561]: from='mgr.14400 192.168.123.100:0/3201047282' entity='mgr.a' 2026-03-06T13:42:47.540 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 06 13:42:47 vm08 ceph-mon[53561]: from='mgr.14400 192.168.123.100:0/3201047282' entity='mgr.a' 2026-03-06T13:42:47.540 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 06 13:42:47 vm08 ceph-mon[53561]: from='mgr.14400 192.168.123.100:0/3201047282' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-06T13:42:47.540 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 06 13:42:47 vm08 ceph-mon[53561]: from='mgr.14400 192.168.123.100:0/3201047282' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-06T13:42:47.540 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 06 13:42:47 vm08 ceph-mon[53561]: from='mgr.14400 192.168.123.100:0/3201047282' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-06T13:42:47.540 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 06 13:42:47 vm08 ceph-mon[53561]: from='mgr.14400 192.168.123.100:0/3201047282' entity='mgr.a' 2026-03-06T13:42:47.655 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:42:47 vm01 ceph-mon[56331]: from='mon.0 -' entity='mon.' cmd=[{"prefix": "rgw zone modify", "realm_name": "myrealm1", "zonegroup_name": "myzonegroup1", "zone_name": "myzone1", "realm_token": "ewogICAgInJlYWxtX25hbWUiOiAibXlyZWFsbTEiLAogICAgInJlYWxtX2lkIjogIjE1ZDk3ZTBjLTM4MzEtNDllYy1hMWUwLTdhYjE3NWExZmFiOCIsCiAgICAiZW5kcG9pbnQiOiBudWxsLAogICAgImFjY2Vzc19rZXkiOiAiNldJQ0hWUUROM0JZMURMMTEwUjMiLAogICAgInNlY3JldCI6ICI3RHl2a2VoVXhNMnVkbDlxSUhZNEE0T053Mmd5ejBLNGw4eW9xZ1B4Igp9", "zone_endpoints": ["http://192.168.123.101:5500", "http://192.168.123.108:5500"]}]: dispatch 2026-03-06T13:42:47.655 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:42:47 vm01 ceph-mon[56331]: Saving service rgw.myrealm1.myzone1 spec with placement count:2 2026-03-06T13:42:47.655 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:42:47 vm01 ceph-mon[56331]: from='mgr.14400 192.168.123.100:0/3201047282' entity='mgr.a' 2026-03-06T13:42:47.655 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:42:47 vm01 ceph-mon[56331]: from='mgr.14400 192.168.123.100:0/3201047282' entity='mgr.a' 2026-03-06T13:42:47.655 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:42:47 vm01 ceph-mon[56331]: from='mgr.14400 192.168.123.100:0/3201047282' entity='mgr.a' 2026-03-06T13:42:47.655 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:42:47 vm01 ceph-mon[56331]: from='mgr.14400 192.168.123.100:0/3201047282' entity='mgr.a' 2026-03-06T13:42:47.655 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:42:47 vm01 ceph-mon[56331]: from='mgr.14400 192.168.123.100:0/3201047282' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-06T13:42:47.655 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:42:47 vm01 ceph-mon[56331]: from='mgr.14400 192.168.123.100:0/3201047282' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-06T13:42:47.655 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:42:47 vm01 ceph-mon[56331]: from='mgr.14400 192.168.123.100:0/3201047282' entity='mgr.a' 2026-03-06T13:42:47.655 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:42:47 vm01 ceph-mon[56331]: from='mgr.14400 192.168.123.100:0/3201047282' entity='mgr.a' 2026-03-06T13:42:47.655 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:42:47 vm01 ceph-mon[56331]: from='mgr.14400 192.168.123.100:0/3201047282' entity='mgr.a' 2026-03-06T13:42:47.655 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:42:47 vm01 ceph-mon[56331]: from='mgr.14400 192.168.123.100:0/3201047282' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-06T13:42:47.655 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:42:47 vm01 ceph-mon[56331]: from='mgr.14400 192.168.123.100:0/3201047282' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-06T13:42:47.655 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:42:47 vm01 ceph-mon[56331]: from='mgr.14400 192.168.123.100:0/3201047282' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-06T13:42:47.655 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:42:47 vm01 ceph-mon[56331]: from='mgr.14400 192.168.123.100:0/3201047282' entity='mgr.a' 2026-03-06T13:42:47.752 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:42:47 vm00 ceph-mon[51914]: from='mon.0 -' entity='mon.' cmd=[{"prefix": "rgw zone modify", "realm_name": "myrealm1", "zonegroup_name": "myzonegroup1", "zone_name": "myzone1", "realm_token": "ewogICAgInJlYWxtX25hbWUiOiAibXlyZWFsbTEiLAogICAgInJlYWxtX2lkIjogIjE1ZDk3ZTBjLTM4MzEtNDllYy1hMWUwLTdhYjE3NWExZmFiOCIsCiAgICAiZW5kcG9pbnQiOiBudWxsLAogICAgImFjY2Vzc19rZXkiOiAiNldJQ0hWUUROM0JZMURMMTEwUjMiLAogICAgInNlY3JldCI6ICI3RHl2a2VoVXhNMnVkbDlxSUhZNEE0T053Mmd5ejBLNGw4eW9xZ1B4Igp9", "zone_endpoints": ["http://192.168.123.101:5500", "http://192.168.123.108:5500"]}]: dispatch 2026-03-06T13:42:47.752 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:42:47 vm00 ceph-mon[51914]: Saving service rgw.myrealm1.myzone1 spec with placement count:2 2026-03-06T13:42:47.752 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:42:47 vm00 ceph-mon[51914]: from='mgr.14400 192.168.123.100:0/3201047282' entity='mgr.a' 2026-03-06T13:42:47.752 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:42:47 vm00 ceph-mon[51914]: from='mgr.14400 192.168.123.100:0/3201047282' entity='mgr.a' 2026-03-06T13:42:47.752 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:42:47 vm00 ceph-mon[51914]: from='mgr.14400 192.168.123.100:0/3201047282' entity='mgr.a' 2026-03-06T13:42:47.752 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:42:47 vm00 ceph-mon[51914]: from='mgr.14400 192.168.123.100:0/3201047282' entity='mgr.a' 2026-03-06T13:42:47.752 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:42:47 vm00 ceph-mon[51914]: from='mgr.14400 192.168.123.100:0/3201047282' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-06T13:42:47.752 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:42:47 vm00 ceph-mon[51914]: from='mgr.14400 192.168.123.100:0/3201047282' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-06T13:42:47.752 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:42:47 vm00 ceph-mon[51914]: from='mgr.14400 192.168.123.100:0/3201047282' entity='mgr.a' 2026-03-06T13:42:47.752 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:42:47 vm00 ceph-mon[51914]: from='mgr.14400 192.168.123.100:0/3201047282' entity='mgr.a' 2026-03-06T13:42:47.752 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:42:47 vm00 ceph-mon[51914]: from='mgr.14400 192.168.123.100:0/3201047282' entity='mgr.a' 2026-03-06T13:42:47.752 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:42:47 vm00 ceph-mon[51914]: from='mgr.14400 192.168.123.100:0/3201047282' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-06T13:42:47.752 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:42:47 vm00 ceph-mon[51914]: from='mgr.14400 192.168.123.100:0/3201047282' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-06T13:42:47.752 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:42:47 vm00 ceph-mon[51914]: from='mgr.14400 192.168.123.100:0/3201047282' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-06T13:42:47.752 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:42:47 vm00 ceph-mon[51914]: from='mgr.14400 192.168.123.100:0/3201047282' entity='mgr.a' 2026-03-06T13:42:48.540 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 06 13:42:48 vm08 ceph-mon[53561]: pgmap v19: 129 pgs: 129 active+clean; 455 KiB data, 81 MiB used, 60 GiB / 60 GiB avail; 13 KiB/s rd, 4.2 KiB/s wr, 18 op/s 2026-03-06T13:42:48.540 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 06 13:42:48 vm08 ceph-mon[53561]: Checking dashboard <-> RGW credentials 2026-03-06T13:42:48.540 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 06 13:42:48 vm08 ceph-mon[53561]: Checking dashboard <-> RGW credentials 2026-03-06T13:42:48.540 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 06 13:42:48 vm08 ceph-mon[53561]: from='mgr.14400 192.168.123.100:0/3201047282' entity='mgr.a' 2026-03-06T13:42:48.655 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:42:48 vm01 ceph-mon[56331]: pgmap v19: 129 pgs: 129 active+clean; 455 KiB data, 81 MiB used, 60 GiB / 60 GiB avail; 13 KiB/s rd, 4.2 KiB/s wr, 18 op/s 2026-03-06T13:42:48.655 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:42:48 vm01 ceph-mon[56331]: Checking dashboard <-> RGW credentials 2026-03-06T13:42:48.655 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:42:48 vm01 ceph-mon[56331]: Checking dashboard <-> RGW credentials 2026-03-06T13:42:48.655 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:42:48 vm01 ceph-mon[56331]: from='mgr.14400 192.168.123.100:0/3201047282' entity='mgr.a' 2026-03-06T13:42:48.752 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:42:48 vm00 ceph-mon[51914]: pgmap v19: 129 pgs: 129 active+clean; 455 KiB data, 81 MiB used, 60 GiB / 60 GiB avail; 13 KiB/s rd, 4.2 KiB/s wr, 18 op/s 2026-03-06T13:42:48.752 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:42:48 vm00 ceph-mon[51914]: Checking dashboard <-> RGW credentials 2026-03-06T13:42:48.752 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:42:48 vm00 ceph-mon[51914]: Checking dashboard <-> RGW credentials 2026-03-06T13:42:48.752 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:42:48 vm00 ceph-mon[51914]: from='mgr.14400 192.168.123.100:0/3201047282' entity='mgr.a' 2026-03-06T13:42:49.445 INFO:teuthology.orchestra.run.vm00.stderr:+ true 2026-03-06T13:42:49.447 INFO:teuthology.orchestra.run.vm00.stderr:++ jq -r '.[0].token' 2026-03-06T13:42:49.447 INFO:teuthology.orchestra.run.vm00.stderr:++ ceph rgw realm tokens 2026-03-06T13:42:49.655 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:42:49 vm01 ceph-mon[56331]: pgmap v20: 129 pgs: 129 active+clean; 455 KiB data, 81 MiB used, 60 GiB / 60 GiB avail; 10 KiB/s rd, 3.2 KiB/s wr, 14 op/s 2026-03-06T13:42:49.752 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:42:49 vm00 ceph-mon[51914]: pgmap v20: 129 pgs: 129 active+clean; 455 KiB data, 81 MiB used, 60 GiB / 60 GiB avail; 10 KiB/s rd, 3.2 KiB/s wr, 14 op/s 2026-03-06T13:42:49.790 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 06 13:42:49 vm08 ceph-mon[53561]: pgmap v20: 129 pgs: 129 active+clean; 455 KiB data, 81 MiB used, 60 GiB / 60 GiB avail; 10 KiB/s rd, 3.2 KiB/s wr, 14 op/s 2026-03-06T13:42:49.994 INFO:teuthology.orchestra.run.vm00.stdout:ewogICAgInJlYWxtX25hbWUiOiAibXlyZWFsbTEiLAogICAgInJlYWxtX2lkIjogIjE1ZDk3ZTBjLTM4MzEtNDllYy1hMWUwLTdhYjE3NWExZmFiOCIsCiAgICAiZW5kcG9pbnQiOiAiaHR0cDovLzE5Mi4xNjguMTIzLjEwMTo1NTAwIiwKICAgICJhY2Nlc3Nfa2V5IjogIjZXSUNIVlFETjNCWTFETDExMFIzIiwKICAgICJzZWNyZXQiOiAiN0R5dmtlaFV4TTJ1ZGw5cUlIWTRBNE9OdzJneXowSzRsOHlvcWdQeCIKfQ== 2026-03-06T13:42:49.994 INFO:teuthology.orchestra.run.vm00.stderr:+ TOKEN=ewogICAgInJlYWxtX25hbWUiOiAibXlyZWFsbTEiLAogICAgInJlYWxtX2lkIjogIjE1ZDk3ZTBjLTM4MzEtNDllYy1hMWUwLTdhYjE3NWExZmFiOCIsCiAgICAiZW5kcG9pbnQiOiAiaHR0cDovLzE5Mi4xNjguMTIzLjEwMTo1NTAwIiwKICAgICJhY2Nlc3Nfa2V5IjogIjZXSUNIVlFETjNCWTFETDExMFIzIiwKICAgICJzZWNyZXQiOiAiN0R5dmtlaFV4TTJ1ZGw5cUlIWTRBNE9OdzJneXowSzRsOHlvcWdQeCIKfQ== 2026-03-06T13:42:49.994 INFO:teuthology.orchestra.run.vm00.stderr:+ echo ewogICAgInJlYWxtX25hbWUiOiAibXlyZWFsbTEiLAogICAgInJlYWxtX2lkIjogIjE1ZDk3ZTBjLTM4MzEtNDllYy1hMWUwLTdhYjE3NWExZmFiOCIsCiAgICAiZW5kcG9pbnQiOiAiaHR0cDovLzE5Mi4xNjguMTIzLjEwMTo1NTAwIiwKICAgICJhY2Nlc3Nfa2V5IjogIjZXSUNIVlFETjNCWTFETDExMFIzIiwKICAgICJzZWNyZXQiOiAiN0R5dmtlaFV4TTJ1ZGw5cUlIWTRBNE9OdzJneXowSzRsOHlvcWdQeCIKfQ== 2026-03-06T13:42:49.994 INFO:teuthology.orchestra.run.vm00.stderr:+ '[' ewogICAgInJlYWxtX25hbWUiOiAibXlyZWFsbTEiLAogICAgInJlYWxtX2lkIjogIjE1ZDk3ZTBjLTM4MzEtNDllYy1hMWUwLTdhYjE3NWExZmFiOCIsCiAgICAiZW5kcG9pbnQiOiAiaHR0cDovLzE5Mi4xNjguMTIzLjEwMTo1NTAwIiwKICAgICJhY2Nlc3Nfa2V5IjogIjZXSUNIVlFETjNCWTFETDExMFIzIiwKICAgICJzZWNyZXQiOiAiN0R5dmtlaFV4TTJ1ZGw5cUlIWTRBNE9OdzJneXowSzRsOHlvcWdQeCIKfQ== '!=' 'master zone has no endpoint' ']' 2026-03-06T13:42:49.994 INFO:teuthology.orchestra.run.vm00.stderr:+ break 2026-03-06T13:42:49.994 INFO:teuthology.orchestra.run.vm00.stderr:++ ceph rgw realm tokens 2026-03-06T13:42:50.517 INFO:teuthology.orchestra.run.vm00.stderr:+ TOKENS='[ 2026-03-06T13:42:50.517 INFO:teuthology.orchestra.run.vm00.stderr: { 2026-03-06T13:42:50.517 INFO:teuthology.orchestra.run.vm00.stderr: "realm": "myrealm1", 2026-03-06T13:42:50.517 INFO:teuthology.orchestra.run.vm00.stderr: "token": "ewogICAgInJlYWxtX25hbWUiOiAibXlyZWFsbTEiLAogICAgInJlYWxtX2lkIjogIjE1ZDk3ZTBjLTM4MzEtNDllYy1hMWUwLTdhYjE3NWExZmFiOCIsCiAgICAiZW5kcG9pbnQiOiAiaHR0cDovLzE5Mi4xNjguMTIzLjEwMTo1NTAwIiwKICAgICJhY2Nlc3Nfa2V5IjogIjZXSUNIVlFETjNCWTFETDExMFIzIiwKICAgICJzZWNyZXQiOiAiN0R5dmtlaFV4TTJ1ZGw5cUlIWTRBNE9OdzJneXowSzRsOHlvcWdQeCIKfQ==" 2026-03-06T13:42:50.517 INFO:teuthology.orchestra.run.vm00.stderr: } 2026-03-06T13:42:50.517 INFO:teuthology.orchestra.run.vm00.stderr:]' 2026-03-06T13:42:50.517 INFO:teuthology.orchestra.run.vm00.stderr:+ echo '[' '{' '"realm":' '"myrealm1",' '"token":' '"ewogICAgInJlYWxtX25hbWUiOiAibXlyZWFsbTEiLAogICAgInJlYWxtX2lkIjogIjE1ZDk3ZTBjLTM4MzEtNDllYy1hMWUwLTdhYjE3NWExZmFiOCIsCiAgICAiZW5kcG9pbnQiOiAiaHR0cDovLzE5Mi4xNjguMTIzLjEwMTo1NTAwIiwKICAgICJhY2Nlc3Nfa2V5IjogIjZXSUNIVlFETjNCWTFETDExMFIzIiwKICAgICJzZWNyZXQiOiAiN0R5dmtlaFV4TTJ1ZGw5cUlIWTRBNE9OdzJneXowSzRsOHlvcWdQeCIKfQ=="' '}' ']' 2026-03-06T13:42:50.517 INFO:teuthology.orchestra.run.vm00.stderr:+ jq --exit-status '.[0].realm == "myrealm1"' 2026-03-06T13:42:50.518 INFO:teuthology.orchestra.run.vm00.stdout:true 2026-03-06T13:42:50.519 INFO:teuthology.orchestra.run.vm00.stderr:+ echo '[' '{' '"realm":' '"myrealm1",' '"token":' '"ewogICAgInJlYWxtX25hbWUiOiAibXlyZWFsbTEiLAogICAgInJlYWxtX2lkIjogIjE1ZDk3ZTBjLTM4MzEtNDllYy1hMWUwLTdhYjE3NWExZmFiOCIsCiAgICAiZW5kcG9pbnQiOiAiaHR0cDovLzE5Mi4xNjguMTIzLjEwMTo1NTAwIiwKICAgICJhY2Nlc3Nfa2V5IjogIjZXSUNIVlFETjNCWTFETDExMFIzIiwKICAgICJzZWNyZXQiOiAiN0R5dmtlaFV4TTJ1ZGw5cUlIWTRBNE9OdzJneXowSzRsOHlvcWdQeCIKfQ=="' '}' ']' 2026-03-06T13:42:50.519 INFO:teuthology.orchestra.run.vm00.stderr:+ jq --exit-status '.[0].token' 2026-03-06T13:42:50.520 INFO:teuthology.orchestra.run.vm00.stdout:"ewogICAgInJlYWxtX25hbWUiOiAibXlyZWFsbTEiLAogICAgInJlYWxtX2lkIjogIjE1ZDk3ZTBjLTM4MzEtNDllYy1hMWUwLTdhYjE3NWExZmFiOCIsCiAgICAiZW5kcG9pbnQiOiAiaHR0cDovLzE5Mi4xNjguMTIzLjEwMTo1NTAwIiwKICAgICJhY2Nlc3Nfa2V5IjogIjZXSUNIVlFETjNCWTFETDExMFIzIiwKICAgICJzZWNyZXQiOiAiN0R5dmtlaFV4TTJ1ZGw5cUlIWTRBNE9OdzJneXowSzRsOHlvcWdQeCIKfQ==" 2026-03-06T13:42:50.521 INFO:teuthology.orchestra.run.vm00.stderr:++ ceph rgw realm tokens 2026-03-06T13:42:50.521 INFO:teuthology.orchestra.run.vm00.stderr:++ base64 --decode 2026-03-06T13:42:50.522 INFO:teuthology.orchestra.run.vm00.stderr:++ jq -r '.[0].token' 2026-03-06T13:42:50.655 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:42:50 vm01 ceph-mon[56331]: from='client.14907 -' entity='client.admin' cmd=[{"prefix": "rgw realm tokens", "target": ["mon-mgr", ""]}]: dispatch 2026-03-06T13:42:50.655 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:42:50 vm01 ceph-mon[56331]: from='client.15015 -' entity='client.admin' cmd=[{"prefix": "rgw realm tokens", "target": ["mon-mgr", ""]}]: dispatch 2026-03-06T13:42:50.752 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:42:50 vm00 ceph-mon[51914]: from='client.14907 -' entity='client.admin' cmd=[{"prefix": "rgw realm tokens", "target": ["mon-mgr", ""]}]: dispatch 2026-03-06T13:42:50.752 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:42:50 vm00 ceph-mon[51914]: from='client.15015 -' entity='client.admin' cmd=[{"prefix": "rgw realm tokens", "target": ["mon-mgr", ""]}]: dispatch 2026-03-06T13:42:50.790 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 06 13:42:50 vm08 ceph-mon[53561]: from='client.14907 -' entity='client.admin' cmd=[{"prefix": "rgw realm tokens", "target": ["mon-mgr", ""]}]: dispatch 2026-03-06T13:42:50.790 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 06 13:42:50 vm08 ceph-mon[53561]: from='client.15015 -' entity='client.admin' cmd=[{"prefix": "rgw realm tokens", "target": ["mon-mgr", ""]}]: dispatch 2026-03-06T13:42:51.110 INFO:teuthology.orchestra.run.vm00.stderr:+ TOKEN_JSON='{ 2026-03-06T13:42:51.110 INFO:teuthology.orchestra.run.vm00.stderr: "realm_name": "myrealm1", 2026-03-06T13:42:51.110 INFO:teuthology.orchestra.run.vm00.stderr: "realm_id": "15d97e0c-3831-49ec-a1e0-7ab175a1fab8", 2026-03-06T13:42:51.110 INFO:teuthology.orchestra.run.vm00.stderr: "endpoint": "http://192.168.123.101:5500", 2026-03-06T13:42:51.110 INFO:teuthology.orchestra.run.vm00.stderr: "access_key": "6WICHVQDN3BY1DL110R3", 2026-03-06T13:42:51.110 INFO:teuthology.orchestra.run.vm00.stderr: "secret": "7DyvkehUxM2udl9qIHY4A4ONw2gyz0K4l8yoqgPx" 2026-03-06T13:42:51.110 INFO:teuthology.orchestra.run.vm00.stderr:}' 2026-03-06T13:42:51.112 INFO:teuthology.orchestra.run.vm00.stderr:+ echo '{' '"realm_name":' '"myrealm1",' '"realm_id":' '"15d97e0c-3831-49ec-a1e0-7ab175a1fab8",' '"endpoint":' '"http://192.168.123.101:5500",' '"access_key":' '"6WICHVQDN3BY1DL110R3",' '"secret":' '"7DyvkehUxM2udl9qIHY4A4ONw2gyz0K4l8yoqgPx"' '}' 2026-03-06T13:42:51.112 INFO:teuthology.orchestra.run.vm00.stderr:+ jq --exit-status '.realm_name == "myrealm1"' 2026-03-06T13:42:51.114 INFO:teuthology.orchestra.run.vm00.stdout:true 2026-03-06T13:42:51.114 INFO:teuthology.orchestra.run.vm00.stderr:+ echo '{' '"realm_name":' '"myrealm1",' '"realm_id":' '"15d97e0c-3831-49ec-a1e0-7ab175a1fab8",' '"endpoint":' '"http://192.168.123.101:5500",' '"access_key":' '"6WICHVQDN3BY1DL110R3",' '"secret":' '"7DyvkehUxM2udl9qIHY4A4ONw2gyz0K4l8yoqgPx"' '}' 2026-03-06T13:42:51.114 INFO:teuthology.orchestra.run.vm00.stderr:+ jq --exit-status '.endpoint | test("http://.+:\\d+")' 2026-03-06T13:42:51.116 INFO:teuthology.orchestra.run.vm00.stdout:true 2026-03-06T13:42:51.116 INFO:teuthology.orchestra.run.vm00.stderr:+ echo '{' '"realm_name":' '"myrealm1",' '"realm_id":' '"15d97e0c-3831-49ec-a1e0-7ab175a1fab8",' '"endpoint":' '"http://192.168.123.101:5500",' '"access_key":' '"6WICHVQDN3BY1DL110R3",' '"secret":' '"7DyvkehUxM2udl9qIHY4A4ONw2gyz0K4l8yoqgPx"' '}' 2026-03-06T13:42:51.116 INFO:teuthology.orchestra.run.vm00.stderr:+ jq --exit-status '.realm_id | test("^[0-9a-f]{8}-[0-9a-f]{4}-[0-9a-f]{4}-[0-9a-f]{4}-[0-9a-f]{12}$")' 2026-03-06T13:42:51.118 INFO:teuthology.orchestra.run.vm00.stdout:true 2026-03-06T13:42:51.118 INFO:teuthology.orchestra.run.vm00.stderr:+ echo '{' '"realm_name":' '"myrealm1",' '"realm_id":' '"15d97e0c-3831-49ec-a1e0-7ab175a1fab8",' '"endpoint":' '"http://192.168.123.101:5500",' '"access_key":' '"6WICHVQDN3BY1DL110R3",' '"secret":' '"7DyvkehUxM2udl9qIHY4A4ONw2gyz0K4l8yoqgPx"' '}' 2026-03-06T13:42:51.119 INFO:teuthology.orchestra.run.vm00.stderr:+ jq --exit-status .access_key 2026-03-06T13:42:51.120 INFO:teuthology.orchestra.run.vm00.stdout:"6WICHVQDN3BY1DL110R3" 2026-03-06T13:42:51.121 INFO:teuthology.orchestra.run.vm00.stderr:+ echo '{' '"realm_name":' '"myrealm1",' '"realm_id":' '"15d97e0c-3831-49ec-a1e0-7ab175a1fab8",' '"endpoint":' '"http://192.168.123.101:5500",' '"access_key":' '"6WICHVQDN3BY1DL110R3",' '"secret":' '"7DyvkehUxM2udl9qIHY4A4ONw2gyz0K4l8yoqgPx"' '}' 2026-03-06T13:42:51.121 INFO:teuthology.orchestra.run.vm00.stderr:+ jq --exit-status .secret 2026-03-06T13:42:51.122 INFO:teuthology.orchestra.run.vm00.stdout:"7DyvkehUxM2udl9qIHY4A4ONw2gyz0K4l8yoqgPx" 2026-03-06T13:42:51.165 DEBUG:teuthology.run_tasks:Unwinding manager cephadm 2026-03-06T13:42:51.167 INFO:tasks.cephadm:Teardown begin 2026-03-06T13:42:51.167 DEBUG:teuthology.orchestra.run.vm00:> sudo rm -f /etc/ceph/ceph.conf /etc/ceph/ceph.client.admin.keyring 2026-03-06T13:42:51.192 DEBUG:teuthology.orchestra.run.vm01:> sudo rm -f /etc/ceph/ceph.conf /etc/ceph/ceph.client.admin.keyring 2026-03-06T13:42:51.223 DEBUG:teuthology.orchestra.run.vm08:> sudo rm -f /etc/ceph/ceph.conf /etc/ceph/ceph.client.admin.keyring 2026-03-06T13:42:51.250 INFO:tasks.cephadm:Cleaning up testdir ceph.* files... 2026-03-06T13:42:51.250 DEBUG:teuthology.orchestra.run.vm00:> rm -f /home/ubuntu/cephtest/seed.ceph.conf /home/ubuntu/cephtest/ceph.pub 2026-03-06T13:42:51.265 DEBUG:teuthology.orchestra.run.vm01:> rm -f /home/ubuntu/cephtest/seed.ceph.conf /home/ubuntu/cephtest/ceph.pub 2026-03-06T13:42:51.281 DEBUG:teuthology.orchestra.run.vm08:> rm -f /home/ubuntu/cephtest/seed.ceph.conf /home/ubuntu/cephtest/ceph.pub 2026-03-06T13:42:51.305 INFO:tasks.cephadm:Stopping all daemons... 2026-03-06T13:42:51.305 INFO:tasks.cephadm.mon.a:Stopping mon.a... 2026-03-06T13:42:51.305 DEBUG:teuthology.orchestra.run.vm00:> sudo systemctl stop ceph-82f504c8-1959-11f1-9a16-a78d135e200b@mon.a 2026-03-06T13:42:51.400 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:42:51 vm00 ceph-mon[51914]: pgmap v21: 129 pgs: 129 active+clean; 460 KiB data, 94 MiB used, 60 GiB / 60 GiB avail; 153 KiB/s rd, 8.3 KiB/s wr, 268 op/s 2026-03-06T13:42:51.400 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:42:51 vm00 ceph-mon[51914]: from='client.15114 -' entity='client.admin' cmd=[{"prefix": "rgw realm tokens", "target": ["mon-mgr", ""]}]: dispatch 2026-03-06T13:42:51.400 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:42:51 vm00 systemd[1]: Stopping Ceph mon.a for 82f504c8-1959-11f1-9a16-a78d135e200b... 2026-03-06T13:42:51.518 DEBUG:teuthology.orchestra.run.vm00:> sudo pkill -f 'journalctl -f -n 0 -u ceph-82f504c8-1959-11f1-9a16-a78d135e200b@mon.a.service' 2026-03-06T13:42:51.540 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 06 13:42:51 vm08 ceph-mon[53561]: pgmap v21: 129 pgs: 129 active+clean; 460 KiB data, 94 MiB used, 60 GiB / 60 GiB avail; 153 KiB/s rd, 8.3 KiB/s wr, 268 op/s 2026-03-06T13:42:51.540 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 06 13:42:51 vm08 ceph-mon[53561]: from='client.15114 -' entity='client.admin' cmd=[{"prefix": "rgw realm tokens", "target": ["mon-mgr", ""]}]: dispatch 2026-03-06T13:42:51.567 DEBUG:teuthology.orchestra.run:got remote process result: None 2026-03-06T13:42:51.567 INFO:tasks.cephadm.mon.a:Stopped mon.a 2026-03-06T13:42:51.567 INFO:tasks.cephadm.mon.c:Stopping mon.b... 2026-03-06T13:42:51.567 DEBUG:teuthology.orchestra.run.vm01:> sudo systemctl stop ceph-82f504c8-1959-11f1-9a16-a78d135e200b@mon.b 2026-03-06T13:42:51.589 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:42:51 vm01 ceph-mon[56331]: pgmap v21: 129 pgs: 129 active+clean; 460 KiB data, 94 MiB used, 60 GiB / 60 GiB avail; 153 KiB/s rd, 8.3 KiB/s wr, 268 op/s 2026-03-06T13:42:51.589 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:42:51 vm01 ceph-mon[56331]: from='client.15114 -' entity='client.admin' cmd=[{"prefix": "rgw realm tokens", "target": ["mon-mgr", ""]}]: dispatch 2026-03-06T13:42:51.817 DEBUG:teuthology.orchestra.run.vm01:> sudo pkill -f 'journalctl -f -n 0 -u ceph-82f504c8-1959-11f1-9a16-a78d135e200b@mon.b.service' 2026-03-06T13:42:51.881 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:42:51 vm01 systemd[1]: Stopping Ceph mon.b for 82f504c8-1959-11f1-9a16-a78d135e200b... 2026-03-06T13:42:51.881 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:42:51 vm01 ceph-82f504c8-1959-11f1-9a16-a78d135e200b-mon-b[56327]: 2026-03-06T12:42:51.674+0000 7f1f9a45e640 -1 received signal: Terminated from /run/podman-init -- /usr/bin/ceph-mon -n mon.b -f --setuser ceph --setgroup ceph --default-log-to-file=false --default-log-to-journald=true --default-log-to-stderr=false --default-mon-cluster-log-to-file=false --default-mon-cluster-log-to-journald=true --default-mon-cluster-log-to-stderr=false (PID: 1) UID: 0 2026-03-06T13:42:51.881 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:42:51 vm01 ceph-82f504c8-1959-11f1-9a16-a78d135e200b-mon-b[56327]: 2026-03-06T12:42:51.674+0000 7f1f9a45e640 -1 mon.b@2(peon) e3 *** Got Signal Terminated *** 2026-03-06T13:42:51.881 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:42:51 vm01 podman[65656]: 2026-03-06 13:42:51.728270806 +0100 CET m=+0.067280303 container died 4095efdfd599b6abc80027ef30a8547c874a47d0e7fe72c0d5d3fbe5ae7e9f04 (image=harbor.clyso.com/custom-ceph/ceph/ceph@sha256:26363c7a4eea9ef5a0148afc7b2a22b6f486596d87a30c2a9fdcda5db3eca62b, name=ceph-82f504c8-1959-11f1-9a16-a78d135e200b-mon-b, FROM_IMAGE=rockylinux:9, GANESHA_REPO_BASEURL=https://buildlogs.centos.org/centos/$releasever-stream/storage/$basearch/nfsganesha-5/, OSD_FLAVOR=default, ceph=True, io.buildah.version=1.41.8, CEPH_GIT_REPO=https://github.com/irq0/ceph.git, CEPH_REF=19.2.3-47-gc24117fd552, CEPH_SHA1=c24117fd5525679b799527bc1bd1f1dd0a2db5e2) 2026-03-06T13:42:51.881 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:42:51 vm01 podman[65656]: 2026-03-06 13:42:51.748960647 +0100 CET m=+0.087970144 container remove 4095efdfd599b6abc80027ef30a8547c874a47d0e7fe72c0d5d3fbe5ae7e9f04 (image=harbor.clyso.com/custom-ceph/ceph/ceph@sha256:26363c7a4eea9ef5a0148afc7b2a22b6f486596d87a30c2a9fdcda5db3eca62b, name=ceph-82f504c8-1959-11f1-9a16-a78d135e200b-mon-b, ceph=True, io.buildah.version=1.41.8, CEPH_GIT_REPO=https://github.com/irq0/ceph.git, CEPH_REF=19.2.3-47-gc24117fd552, CEPH_SHA1=c24117fd5525679b799527bc1bd1f1dd0a2db5e2, FROM_IMAGE=rockylinux:9, GANESHA_REPO_BASEURL=https://buildlogs.centos.org/centos/$releasever-stream/storage/$basearch/nfsganesha-5/, OSD_FLAVOR=default) 2026-03-06T13:42:51.881 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:42:51 vm01 bash[65656]: ceph-82f504c8-1959-11f1-9a16-a78d135e200b-mon-b 2026-03-06T13:42:51.881 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:42:51 vm01 systemd[1]: ceph-82f504c8-1959-11f1-9a16-a78d135e200b@mon.b.service: Deactivated successfully. 2026-03-06T13:42:51.881 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:42:51 vm01 systemd[1]: Stopped Ceph mon.b for 82f504c8-1959-11f1-9a16-a78d135e200b. 2026-03-06T13:42:51.881 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:42:51 vm01 systemd[1]: ceph-82f504c8-1959-11f1-9a16-a78d135e200b@mon.b.service: Consumed 1.877s CPU time. 2026-03-06T13:42:51.897 DEBUG:teuthology.orchestra.run:got remote process result: None 2026-03-06T13:42:51.897 INFO:tasks.cephadm.mon.c:Stopped mon.b 2026-03-06T13:42:51.897 INFO:tasks.cephadm.mon.c:Stopping mon.c... 2026-03-06T13:42:51.897 DEBUG:teuthology.orchestra.run.vm08:> sudo systemctl stop ceph-82f504c8-1959-11f1-9a16-a78d135e200b@mon.c 2026-03-06T13:42:52.241 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 06 13:42:51 vm08 systemd[1]: Stopping Ceph mon.c for 82f504c8-1959-11f1-9a16-a78d135e200b... 2026-03-06T13:42:52.242 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 06 13:42:52 vm08 ceph-82f504c8-1959-11f1-9a16-a78d135e200b-mon-c[53557]: 2026-03-06T12:42:52.005+0000 7fcd777f6640 -1 received signal: Terminated from /run/podman-init -- /usr/bin/ceph-mon -n mon.c -f --setuser ceph --setgroup ceph --default-log-to-file=false --default-log-to-journald=true --default-log-to-stderr=false --default-mon-cluster-log-to-file=false --default-mon-cluster-log-to-journald=true --default-mon-cluster-log-to-stderr=false (PID: 1) UID: 0 2026-03-06T13:42:52.242 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 06 13:42:52 vm08 ceph-82f504c8-1959-11f1-9a16-a78d135e200b-mon-c[53557]: 2026-03-06T12:42:52.005+0000 7fcd777f6640 -1 mon.c@1(peon) e3 *** Got Signal Terminated *** 2026-03-06T13:42:52.325 DEBUG:teuthology.orchestra.run.vm08:> sudo pkill -f 'journalctl -f -n 0 -u ceph-82f504c8-1959-11f1-9a16-a78d135e200b@mon.c.service' 2026-03-06T13:42:52.371 DEBUG:teuthology.orchestra.run:got remote process result: None 2026-03-06T13:42:52.372 INFO:tasks.cephadm.mon.c:Stopped mon.c 2026-03-06T13:42:52.372 INFO:tasks.cephadm.mgr.a:Stopping mgr.a... 2026-03-06T13:42:52.372 DEBUG:teuthology.orchestra.run.vm00:> sudo systemctl stop ceph-82f504c8-1959-11f1-9a16-a78d135e200b@mgr.a 2026-03-06T13:42:52.640 DEBUG:teuthology.orchestra.run.vm00:> sudo pkill -f 'journalctl -f -n 0 -u ceph-82f504c8-1959-11f1-9a16-a78d135e200b@mgr.a.service' 2026-03-06T13:42:52.665 INFO:journalctl@ceph.mgr.a.vm00.stdout:Mar 06 13:42:52 vm00 systemd[1]: Stopping Ceph mgr.a for 82f504c8-1959-11f1-9a16-a78d135e200b... 2026-03-06T13:42:52.665 INFO:journalctl@ceph.mgr.a.vm00.stdout:Mar 06 13:42:52 vm00 podman[72075]: 2026-03-06 13:42:52.525654032 +0100 CET m=+0.061354660 container died 28d034114368a2f278fb4633fdb2033dacf31996e90e28db94d1392dd113ebe5 (image=harbor.clyso.com/custom-ceph/ceph/ceph:cobaltcore-storage-v19.2.3-fasttrack-3, name=ceph-82f504c8-1959-11f1-9a16-a78d135e200b-mgr-a, ceph=True, io.buildah.version=1.41.8, CEPH_GIT_REPO=https://github.com/irq0/ceph.git, CEPH_REF=19.2.3-47-gc24117fd552, CEPH_SHA1=c24117fd5525679b799527bc1bd1f1dd0a2db5e2, FROM_IMAGE=rockylinux:9, GANESHA_REPO_BASEURL=https://buildlogs.centos.org/centos/$releasever-stream/storage/$basearch/nfsganesha-5/, OSD_FLAVOR=default) 2026-03-06T13:42:52.665 INFO:journalctl@ceph.mgr.a.vm00.stdout:Mar 06 13:42:52 vm00 podman[72075]: 2026-03-06 13:42:52.57482938 +0100 CET m=+0.110530008 container remove 28d034114368a2f278fb4633fdb2033dacf31996e90e28db94d1392dd113ebe5 (image=harbor.clyso.com/custom-ceph/ceph/ceph:cobaltcore-storage-v19.2.3-fasttrack-3, name=ceph-82f504c8-1959-11f1-9a16-a78d135e200b-mgr-a, CEPH_SHA1=c24117fd5525679b799527bc1bd1f1dd0a2db5e2, FROM_IMAGE=rockylinux:9, GANESHA_REPO_BASEURL=https://buildlogs.centos.org/centos/$releasever-stream/storage/$basearch/nfsganesha-5/, OSD_FLAVOR=default, ceph=True, io.buildah.version=1.41.8, CEPH_GIT_REPO=https://github.com/irq0/ceph.git, CEPH_REF=19.2.3-47-gc24117fd552) 2026-03-06T13:42:52.666 INFO:journalctl@ceph.mgr.a.vm00.stdout:Mar 06 13:42:52 vm00 bash[72075]: ceph-82f504c8-1959-11f1-9a16-a78d135e200b-mgr-a 2026-03-06T13:42:52.666 INFO:journalctl@ceph.mgr.a.vm00.stdout:Mar 06 13:42:52 vm00 systemd[1]: ceph-82f504c8-1959-11f1-9a16-a78d135e200b@mgr.a.service: Deactivated successfully. 2026-03-06T13:42:52.666 INFO:journalctl@ceph.mgr.a.vm00.stdout:Mar 06 13:42:52 vm00 systemd[1]: Stopped Ceph mgr.a for 82f504c8-1959-11f1-9a16-a78d135e200b. 2026-03-06T13:42:52.666 INFO:journalctl@ceph.mgr.a.vm00.stdout:Mar 06 13:42:52 vm00 systemd[1]: ceph-82f504c8-1959-11f1-9a16-a78d135e200b@mgr.a.service: Consumed 55.608s CPU time. 2026-03-06T13:42:52.683 DEBUG:teuthology.orchestra.run:got remote process result: None 2026-03-06T13:42:52.683 INFO:tasks.cephadm.mgr.a:Stopped mgr.a 2026-03-06T13:42:52.683 INFO:tasks.cephadm.mgr.b:Stopping mgr.b... 2026-03-06T13:42:52.683 DEBUG:teuthology.orchestra.run.vm01:> sudo systemctl stop ceph-82f504c8-1959-11f1-9a16-a78d135e200b@mgr.b 2026-03-06T13:42:52.826 INFO:journalctl@ceph.mgr.b.vm01.stdout:Mar 06 13:42:52 vm01 systemd[1]: Stopping Ceph mgr.b for 82f504c8-1959-11f1-9a16-a78d135e200b... 2026-03-06T13:42:52.928 DEBUG:teuthology.orchestra.run.vm01:> sudo pkill -f 'journalctl -f -n 0 -u ceph-82f504c8-1959-11f1-9a16-a78d135e200b@mgr.b.service' 2026-03-06T13:42:52.967 DEBUG:teuthology.orchestra.run:got remote process result: None 2026-03-06T13:42:52.967 INFO:tasks.cephadm.mgr.b:Stopped mgr.b 2026-03-06T13:42:52.967 INFO:tasks.cephadm.osd.0:Stopping osd.0... 2026-03-06T13:42:52.967 DEBUG:teuthology.orchestra.run.vm00:> sudo systemctl stop ceph-82f504c8-1959-11f1-9a16-a78d135e200b@osd.0 2026-03-06T13:42:53.252 INFO:journalctl@ceph.osd.0.vm00.stdout:Mar 06 13:42:53 vm00 systemd[1]: Stopping Ceph osd.0 for 82f504c8-1959-11f1-9a16-a78d135e200b... 2026-03-06T13:42:53.252 INFO:journalctl@ceph.osd.0.vm00.stdout:Mar 06 13:42:53 vm00 ceph-82f504c8-1959-11f1-9a16-a78d135e200b-osd-0[61746]: 2026-03-06T12:42:53.076+0000 7fb55d59d640 -1 received signal: Terminated from /run/podman-init -- /usr/bin/ceph-osd -n osd.0 -f --setuser ceph --setgroup ceph --default-log-to-file=false --default-log-to-journald=true --default-log-to-stderr=false (PID: 1) UID: 0 2026-03-06T13:42:53.252 INFO:journalctl@ceph.osd.0.vm00.stdout:Mar 06 13:42:53 vm00 ceph-82f504c8-1959-11f1-9a16-a78d135e200b-osd-0[61746]: 2026-03-06T12:42:53.076+0000 7fb55d59d640 -1 osd.0 29 *** Got signal Terminated *** 2026-03-06T13:42:53.252 INFO:journalctl@ceph.osd.0.vm00.stdout:Mar 06 13:42:53 vm00 ceph-82f504c8-1959-11f1-9a16-a78d135e200b-osd-0[61746]: 2026-03-06T12:42:53.076+0000 7fb55d59d640 -1 osd.0 29 *** Immediate shutdown (osd_fast_shutdown=true) *** 2026-03-06T13:42:58.502 INFO:journalctl@ceph.osd.0.vm00.stdout:Mar 06 13:42:58 vm00 podman[72178]: 2026-03-06 13:42:58.116817271 +0100 CET m=+5.054702018 container died 27fa8297be4f60c2ca80be6942b3377f147aac980e76f3032ee4a745cbf6f562 (image=harbor.clyso.com/custom-ceph/ceph/ceph@sha256:26363c7a4eea9ef5a0148afc7b2a22b6f486596d87a30c2a9fdcda5db3eca62b, name=ceph-82f504c8-1959-11f1-9a16-a78d135e200b-osd-0, io.buildah.version=1.41.8, CEPH_GIT_REPO=https://github.com/irq0/ceph.git, CEPH_REF=19.2.3-47-gc24117fd552, CEPH_SHA1=c24117fd5525679b799527bc1bd1f1dd0a2db5e2, FROM_IMAGE=rockylinux:9, GANESHA_REPO_BASEURL=https://buildlogs.centos.org/centos/$releasever-stream/storage/$basearch/nfsganesha-5/, OSD_FLAVOR=default, ceph=True) 2026-03-06T13:42:58.503 INFO:journalctl@ceph.osd.0.vm00.stdout:Mar 06 13:42:58 vm00 podman[72178]: 2026-03-06 13:42:58.138979079 +0100 CET m=+5.076863826 container remove 27fa8297be4f60c2ca80be6942b3377f147aac980e76f3032ee4a745cbf6f562 (image=harbor.clyso.com/custom-ceph/ceph/ceph@sha256:26363c7a4eea9ef5a0148afc7b2a22b6f486596d87a30c2a9fdcda5db3eca62b, name=ceph-82f504c8-1959-11f1-9a16-a78d135e200b-osd-0, CEPH_REF=19.2.3-47-gc24117fd552, CEPH_SHA1=c24117fd5525679b799527bc1bd1f1dd0a2db5e2, FROM_IMAGE=rockylinux:9, GANESHA_REPO_BASEURL=https://buildlogs.centos.org/centos/$releasever-stream/storage/$basearch/nfsganesha-5/, OSD_FLAVOR=default, ceph=True, io.buildah.version=1.41.8, CEPH_GIT_REPO=https://github.com/irq0/ceph.git) 2026-03-06T13:42:58.503 INFO:journalctl@ceph.osd.0.vm00.stdout:Mar 06 13:42:58 vm00 bash[72178]: ceph-82f504c8-1959-11f1-9a16-a78d135e200b-osd-0 2026-03-06T13:42:58.503 INFO:journalctl@ceph.osd.0.vm00.stdout:Mar 06 13:42:58 vm00 podman[72244]: 2026-03-06 13:42:58.260189742 +0100 CET m=+0.014596838 container create b6390b301438b8e079d889b3ae2defbcda2522ea60f497e5eb6d400a0861515b (image=harbor.clyso.com/custom-ceph/ceph/ceph@sha256:26363c7a4eea9ef5a0148afc7b2a22b6f486596d87a30c2a9fdcda5db3eca62b, name=ceph-82f504c8-1959-11f1-9a16-a78d135e200b-osd-0-deactivate, io.buildah.version=1.41.8, CEPH_GIT_REPO=https://github.com/irq0/ceph.git, CEPH_REF=19.2.3-47-gc24117fd552, CEPH_SHA1=c24117fd5525679b799527bc1bd1f1dd0a2db5e2, FROM_IMAGE=rockylinux:9, GANESHA_REPO_BASEURL=https://buildlogs.centos.org/centos/$releasever-stream/storage/$basearch/nfsganesha-5/, OSD_FLAVOR=default, ceph=True) 2026-03-06T13:42:58.503 INFO:journalctl@ceph.osd.0.vm00.stdout:Mar 06 13:42:58 vm00 podman[72244]: 2026-03-06 13:42:58.301029056 +0100 CET m=+0.055436161 container init b6390b301438b8e079d889b3ae2defbcda2522ea60f497e5eb6d400a0861515b (image=harbor.clyso.com/custom-ceph/ceph/ceph@sha256:26363c7a4eea9ef5a0148afc7b2a22b6f486596d87a30c2a9fdcda5db3eca62b, name=ceph-82f504c8-1959-11f1-9a16-a78d135e200b-osd-0-deactivate, CEPH_REF=19.2.3-47-gc24117fd552, CEPH_SHA1=c24117fd5525679b799527bc1bd1f1dd0a2db5e2, FROM_IMAGE=rockylinux:9, GANESHA_REPO_BASEURL=https://buildlogs.centos.org/centos/$releasever-stream/storage/$basearch/nfsganesha-5/, OSD_FLAVOR=default, ceph=True, io.buildah.version=1.41.8, CEPH_GIT_REPO=https://github.com/irq0/ceph.git) 2026-03-06T13:42:58.503 INFO:journalctl@ceph.osd.0.vm00.stdout:Mar 06 13:42:58 vm00 podman[72244]: 2026-03-06 13:42:58.303912414 +0100 CET m=+0.058319510 container start b6390b301438b8e079d889b3ae2defbcda2522ea60f497e5eb6d400a0861515b (image=harbor.clyso.com/custom-ceph/ceph/ceph@sha256:26363c7a4eea9ef5a0148afc7b2a22b6f486596d87a30c2a9fdcda5db3eca62b, name=ceph-82f504c8-1959-11f1-9a16-a78d135e200b-osd-0-deactivate, GANESHA_REPO_BASEURL=https://buildlogs.centos.org/centos/$releasever-stream/storage/$basearch/nfsganesha-5/, OSD_FLAVOR=default, ceph=True, io.buildah.version=1.41.8, CEPH_GIT_REPO=https://github.com/irq0/ceph.git, CEPH_REF=19.2.3-47-gc24117fd552, CEPH_SHA1=c24117fd5525679b799527bc1bd1f1dd0a2db5e2, FROM_IMAGE=rockylinux:9) 2026-03-06T13:42:58.503 INFO:journalctl@ceph.osd.0.vm00.stdout:Mar 06 13:42:58 vm00 podman[72244]: 2026-03-06 13:42:58.304588779 +0100 CET m=+0.058995875 container attach b6390b301438b8e079d889b3ae2defbcda2522ea60f497e5eb6d400a0861515b (image=harbor.clyso.com/custom-ceph/ceph/ceph@sha256:26363c7a4eea9ef5a0148afc7b2a22b6f486596d87a30c2a9fdcda5db3eca62b, name=ceph-82f504c8-1959-11f1-9a16-a78d135e200b-osd-0-deactivate, io.buildah.version=1.41.8, CEPH_GIT_REPO=https://github.com/irq0/ceph.git, CEPH_REF=19.2.3-47-gc24117fd552, CEPH_SHA1=c24117fd5525679b799527bc1bd1f1dd0a2db5e2, FROM_IMAGE=rockylinux:9, GANESHA_REPO_BASEURL=https://buildlogs.centos.org/centos/$releasever-stream/storage/$basearch/nfsganesha-5/, OSD_FLAVOR=default, ceph=True) 2026-03-06T13:42:58.503 INFO:journalctl@ceph.osd.0.vm00.stdout:Mar 06 13:42:58 vm00 podman[72244]: 2026-03-06 13:42:58.254363355 +0100 CET m=+0.008770461 image pull 306e97de47e91c2b4b24d3dc09be3b3a12039b078f343d91220102acc6628a68 harbor.clyso.com/custom-ceph/ceph/ceph@sha256:26363c7a4eea9ef5a0148afc7b2a22b6f486596d87a30c2a9fdcda5db3eca62b 2026-03-06T13:42:58.755 DEBUG:teuthology.orchestra.run.vm00:> sudo pkill -f 'journalctl -f -n 0 -u ceph-82f504c8-1959-11f1-9a16-a78d135e200b@osd.0.service' 2026-03-06T13:42:58.801 DEBUG:teuthology.orchestra.run:got remote process result: None 2026-03-06T13:42:58.801 INFO:tasks.cephadm.osd.0:Stopped osd.0 2026-03-06T13:42:58.801 INFO:tasks.cephadm.osd.1:Stopping osd.1... 2026-03-06T13:42:58.802 DEBUG:teuthology.orchestra.run.vm01:> sudo systemctl stop ceph-82f504c8-1959-11f1-9a16-a78d135e200b@osd.1 2026-03-06T13:42:59.155 INFO:journalctl@ceph.osd.1.vm01.stdout:Mar 06 13:42:58 vm01 systemd[1]: Stopping Ceph osd.1 for 82f504c8-1959-11f1-9a16-a78d135e200b... 2026-03-06T13:42:59.155 INFO:journalctl@ceph.osd.1.vm01.stdout:Mar 06 13:42:58 vm01 ceph-82f504c8-1959-11f1-9a16-a78d135e200b-osd-1[59988]: 2026-03-06T12:42:58.907+0000 7f2a5f093640 -1 received signal: Terminated from /run/podman-init -- /usr/bin/ceph-osd -n osd.1 -f --setuser ceph --setgroup ceph --default-log-to-file=false --default-log-to-journald=true --default-log-to-stderr=false (PID: 1) UID: 0 2026-03-06T13:42:59.155 INFO:journalctl@ceph.osd.1.vm01.stdout:Mar 06 13:42:58 vm01 ceph-82f504c8-1959-11f1-9a16-a78d135e200b-osd-1[59988]: 2026-03-06T12:42:58.907+0000 7f2a5f093640 -1 osd.1 29 *** Got signal Terminated *** 2026-03-06T13:42:59.155 INFO:journalctl@ceph.osd.1.vm01.stdout:Mar 06 13:42:58 vm01 ceph-82f504c8-1959-11f1-9a16-a78d135e200b-osd-1[59988]: 2026-03-06T12:42:58.907+0000 7f2a5f093640 -1 osd.1 29 *** Immediate shutdown (osd_fast_shutdown=true) *** 2026-03-06T13:43:04.203 INFO:journalctl@ceph.osd.1.vm01.stdout:Mar 06 13:43:03 vm01 podman[65864]: 2026-03-06 13:43:03.949449742 +0100 CET m=+5.056509837 container died 57ae5eb11439b347e66da9eb2a6aa2ef75ee27654b61aaba3fc3c59902cf9473 (image=harbor.clyso.com/custom-ceph/ceph/ceph@sha256:26363c7a4eea9ef5a0148afc7b2a22b6f486596d87a30c2a9fdcda5db3eca62b, name=ceph-82f504c8-1959-11f1-9a16-a78d135e200b-osd-1, CEPH_GIT_REPO=https://github.com/irq0/ceph.git, CEPH_REF=19.2.3-47-gc24117fd552, CEPH_SHA1=c24117fd5525679b799527bc1bd1f1dd0a2db5e2, FROM_IMAGE=rockylinux:9, GANESHA_REPO_BASEURL=https://buildlogs.centos.org/centos/$releasever-stream/storage/$basearch/nfsganesha-5/, OSD_FLAVOR=default, ceph=True, io.buildah.version=1.41.8) 2026-03-06T13:43:04.203 INFO:journalctl@ceph.osd.1.vm01.stdout:Mar 06 13:43:03 vm01 podman[65864]: 2026-03-06 13:43:03.967588636 +0100 CET m=+5.074648731 container remove 57ae5eb11439b347e66da9eb2a6aa2ef75ee27654b61aaba3fc3c59902cf9473 (image=harbor.clyso.com/custom-ceph/ceph/ceph@sha256:26363c7a4eea9ef5a0148afc7b2a22b6f486596d87a30c2a9fdcda5db3eca62b, name=ceph-82f504c8-1959-11f1-9a16-a78d135e200b-osd-1, io.buildah.version=1.41.8, CEPH_GIT_REPO=https://github.com/irq0/ceph.git, CEPH_REF=19.2.3-47-gc24117fd552, CEPH_SHA1=c24117fd5525679b799527bc1bd1f1dd0a2db5e2, FROM_IMAGE=rockylinux:9, GANESHA_REPO_BASEURL=https://buildlogs.centos.org/centos/$releasever-stream/storage/$basearch/nfsganesha-5/, OSD_FLAVOR=default, ceph=True) 2026-03-06T13:43:04.203 INFO:journalctl@ceph.osd.1.vm01.stdout:Mar 06 13:43:03 vm01 bash[65864]: ceph-82f504c8-1959-11f1-9a16-a78d135e200b-osd-1 2026-03-06T13:43:04.203 INFO:journalctl@ceph.osd.1.vm01.stdout:Mar 06 13:43:04 vm01 podman[65930]: 2026-03-06 13:43:04.109355493 +0100 CET m=+0.016252442 container create a907589ac7b35122c8e220b54f6495b1cd7571b8df1f0c1eb5fe631608b0f4a6 (image=harbor.clyso.com/custom-ceph/ceph/ceph@sha256:26363c7a4eea9ef5a0148afc7b2a22b6f486596d87a30c2a9fdcda5db3eca62b, name=ceph-82f504c8-1959-11f1-9a16-a78d135e200b-osd-1-deactivate, OSD_FLAVOR=default, ceph=True, io.buildah.version=1.41.8, CEPH_GIT_REPO=https://github.com/irq0/ceph.git, CEPH_REF=19.2.3-47-gc24117fd552, CEPH_SHA1=c24117fd5525679b799527bc1bd1f1dd0a2db5e2, FROM_IMAGE=rockylinux:9, GANESHA_REPO_BASEURL=https://buildlogs.centos.org/centos/$releasever-stream/storage/$basearch/nfsganesha-5/) 2026-03-06T13:43:04.203 INFO:journalctl@ceph.osd.1.vm01.stdout:Mar 06 13:43:04 vm01 podman[65930]: 2026-03-06 13:43:04.182494726 +0100 CET m=+0.089391675 container init a907589ac7b35122c8e220b54f6495b1cd7571b8df1f0c1eb5fe631608b0f4a6 (image=harbor.clyso.com/custom-ceph/ceph/ceph@sha256:26363c7a4eea9ef5a0148afc7b2a22b6f486596d87a30c2a9fdcda5db3eca62b, name=ceph-82f504c8-1959-11f1-9a16-a78d135e200b-osd-1-deactivate, io.buildah.version=1.41.8, CEPH_GIT_REPO=https://github.com/irq0/ceph.git, CEPH_REF=19.2.3-47-gc24117fd552, CEPH_SHA1=c24117fd5525679b799527bc1bd1f1dd0a2db5e2, FROM_IMAGE=rockylinux:9, GANESHA_REPO_BASEURL=https://buildlogs.centos.org/centos/$releasever-stream/storage/$basearch/nfsganesha-5/, OSD_FLAVOR=default, ceph=True) 2026-03-06T13:43:04.203 INFO:journalctl@ceph.osd.1.vm01.stdout:Mar 06 13:43:04 vm01 podman[65930]: 2026-03-06 13:43:04.187946484 +0100 CET m=+0.094843433 container start a907589ac7b35122c8e220b54f6495b1cd7571b8df1f0c1eb5fe631608b0f4a6 (image=harbor.clyso.com/custom-ceph/ceph/ceph@sha256:26363c7a4eea9ef5a0148afc7b2a22b6f486596d87a30c2a9fdcda5db3eca62b, name=ceph-82f504c8-1959-11f1-9a16-a78d135e200b-osd-1-deactivate, GANESHA_REPO_BASEURL=https://buildlogs.centos.org/centos/$releasever-stream/storage/$basearch/nfsganesha-5/, OSD_FLAVOR=default, ceph=True, io.buildah.version=1.41.8, CEPH_GIT_REPO=https://github.com/irq0/ceph.git, CEPH_REF=19.2.3-47-gc24117fd552, CEPH_SHA1=c24117fd5525679b799527bc1bd1f1dd0a2db5e2, FROM_IMAGE=rockylinux:9) 2026-03-06T13:43:04.203 INFO:journalctl@ceph.osd.1.vm01.stdout:Mar 06 13:43:04 vm01 podman[65930]: 2026-03-06 13:43:04.18970658 +0100 CET m=+0.096603529 container attach a907589ac7b35122c8e220b54f6495b1cd7571b8df1f0c1eb5fe631608b0f4a6 (image=harbor.clyso.com/custom-ceph/ceph/ceph@sha256:26363c7a4eea9ef5a0148afc7b2a22b6f486596d87a30c2a9fdcda5db3eca62b, name=ceph-82f504c8-1959-11f1-9a16-a78d135e200b-osd-1-deactivate, CEPH_GIT_REPO=https://github.com/irq0/ceph.git, CEPH_REF=19.2.3-47-gc24117fd552, CEPH_SHA1=c24117fd5525679b799527bc1bd1f1dd0a2db5e2, FROM_IMAGE=rockylinux:9, GANESHA_REPO_BASEURL=https://buildlogs.centos.org/centos/$releasever-stream/storage/$basearch/nfsganesha-5/, OSD_FLAVOR=default, ceph=True, io.buildah.version=1.41.8) 2026-03-06T13:43:04.203 INFO:journalctl@ceph.osd.1.vm01.stdout:Mar 06 13:43:04 vm01 podman[65930]: 2026-03-06 13:43:04.102795851 +0100 CET m=+0.009692810 image pull 306e97de47e91c2b4b24d3dc09be3b3a12039b078f343d91220102acc6628a68 harbor.clyso.com/custom-ceph/ceph/ceph@sha256:26363c7a4eea9ef5a0148afc7b2a22b6f486596d87a30c2a9fdcda5db3eca62b 2026-03-06T13:43:04.627 DEBUG:teuthology.orchestra.run.vm01:> sudo pkill -f 'journalctl -f -n 0 -u ceph-82f504c8-1959-11f1-9a16-a78d135e200b@osd.1.service' 2026-03-06T13:43:04.671 DEBUG:teuthology.orchestra.run:got remote process result: None 2026-03-06T13:43:04.671 INFO:tasks.cephadm.osd.1:Stopped osd.1 2026-03-06T13:43:04.671 INFO:tasks.cephadm.osd.2:Stopping osd.2... 2026-03-06T13:43:04.671 DEBUG:teuthology.orchestra.run.vm08:> sudo systemctl stop ceph-82f504c8-1959-11f1-9a16-a78d135e200b@osd.2 2026-03-06T13:43:05.040 INFO:journalctl@ceph.osd.2.vm08.stdout:Mar 06 13:43:04 vm08 systemd[1]: Stopping Ceph osd.2 for 82f504c8-1959-11f1-9a16-a78d135e200b... 2026-03-06T13:43:05.040 INFO:journalctl@ceph.osd.2.vm08.stdout:Mar 06 13:43:04 vm08 ceph-82f504c8-1959-11f1-9a16-a78d135e200b-osd-2[57083]: 2026-03-06T12:43:04.775+0000 7fb18e79e640 -1 received signal: Terminated from /run/podman-init -- /usr/bin/ceph-osd -n osd.2 -f --setuser ceph --setgroup ceph --default-log-to-file=false --default-log-to-journald=true --default-log-to-stderr=false (PID: 1) UID: 0 2026-03-06T13:43:05.040 INFO:journalctl@ceph.osd.2.vm08.stdout:Mar 06 13:43:04 vm08 ceph-82f504c8-1959-11f1-9a16-a78d135e200b-osd-2[57083]: 2026-03-06T12:43:04.775+0000 7fb18e79e640 -1 osd.2 29 *** Got signal Terminated *** 2026-03-06T13:43:05.040 INFO:journalctl@ceph.osd.2.vm08.stdout:Mar 06 13:43:04 vm08 ceph-82f504c8-1959-11f1-9a16-a78d135e200b-osd-2[57083]: 2026-03-06T12:43:04.775+0000 7fb18e79e640 -1 osd.2 29 *** Immediate shutdown (osd_fast_shutdown=true) *** 2026-03-06T13:43:10.063 INFO:journalctl@ceph.osd.2.vm08.stdout:Mar 06 13:43:09 vm08 podman[62704]: 2026-03-06 13:43:09.803577393 +0100 CET m=+5.043894781 container died 85f2716688cbe9a04a73c05de781c2539ee279bb85a747e0f95e5e2377e18921 (image=harbor.clyso.com/custom-ceph/ceph/ceph@sha256:26363c7a4eea9ef5a0148afc7b2a22b6f486596d87a30c2a9fdcda5db3eca62b, name=ceph-82f504c8-1959-11f1-9a16-a78d135e200b-osd-2, OSD_FLAVOR=default, ceph=True, io.buildah.version=1.41.8, CEPH_GIT_REPO=https://github.com/irq0/ceph.git, CEPH_REF=19.2.3-47-gc24117fd552, CEPH_SHA1=c24117fd5525679b799527bc1bd1f1dd0a2db5e2, FROM_IMAGE=rockylinux:9, GANESHA_REPO_BASEURL=https://buildlogs.centos.org/centos/$releasever-stream/storage/$basearch/nfsganesha-5/) 2026-03-06T13:43:10.063 INFO:journalctl@ceph.osd.2.vm08.stdout:Mar 06 13:43:09 vm08 podman[62704]: 2026-03-06 13:43:09.8335791 +0100 CET m=+5.073896488 container remove 85f2716688cbe9a04a73c05de781c2539ee279bb85a747e0f95e5e2377e18921 (image=harbor.clyso.com/custom-ceph/ceph/ceph@sha256:26363c7a4eea9ef5a0148afc7b2a22b6f486596d87a30c2a9fdcda5db3eca62b, name=ceph-82f504c8-1959-11f1-9a16-a78d135e200b-osd-2, CEPH_GIT_REPO=https://github.com/irq0/ceph.git, CEPH_REF=19.2.3-47-gc24117fd552, CEPH_SHA1=c24117fd5525679b799527bc1bd1f1dd0a2db5e2, FROM_IMAGE=rockylinux:9, GANESHA_REPO_BASEURL=https://buildlogs.centos.org/centos/$releasever-stream/storage/$basearch/nfsganesha-5/, OSD_FLAVOR=default, ceph=True, io.buildah.version=1.41.8) 2026-03-06T13:43:10.063 INFO:journalctl@ceph.osd.2.vm08.stdout:Mar 06 13:43:09 vm08 bash[62704]: ceph-82f504c8-1959-11f1-9a16-a78d135e200b-osd-2 2026-03-06T13:43:10.063 INFO:journalctl@ceph.osd.2.vm08.stdout:Mar 06 13:43:09 vm08 podman[62771]: 2026-03-06 13:43:09.968972081 +0100 CET m=+0.015354818 container create 939e9ebfbd48dea3105bcd983bfaf4c05747f22d1fb73eabe106e4420b8d2d8b (image=harbor.clyso.com/custom-ceph/ceph/ceph@sha256:26363c7a4eea9ef5a0148afc7b2a22b6f486596d87a30c2a9fdcda5db3eca62b, name=ceph-82f504c8-1959-11f1-9a16-a78d135e200b-osd-2-deactivate, GANESHA_REPO_BASEURL=https://buildlogs.centos.org/centos/$releasever-stream/storage/$basearch/nfsganesha-5/, OSD_FLAVOR=default, ceph=True, io.buildah.version=1.41.8, CEPH_GIT_REPO=https://github.com/irq0/ceph.git, CEPH_REF=19.2.3-47-gc24117fd552, CEPH_SHA1=c24117fd5525679b799527bc1bd1f1dd0a2db5e2, FROM_IMAGE=rockylinux:9) 2026-03-06T13:43:10.063 INFO:journalctl@ceph.osd.2.vm08.stdout:Mar 06 13:43:10 vm08 podman[62771]: 2026-03-06 13:43:10.014461381 +0100 CET m=+0.060844139 container init 939e9ebfbd48dea3105bcd983bfaf4c05747f22d1fb73eabe106e4420b8d2d8b (image=harbor.clyso.com/custom-ceph/ceph/ceph@sha256:26363c7a4eea9ef5a0148afc7b2a22b6f486596d87a30c2a9fdcda5db3eca62b, name=ceph-82f504c8-1959-11f1-9a16-a78d135e200b-osd-2-deactivate, CEPH_SHA1=c24117fd5525679b799527bc1bd1f1dd0a2db5e2, FROM_IMAGE=rockylinux:9, GANESHA_REPO_BASEURL=https://buildlogs.centos.org/centos/$releasever-stream/storage/$basearch/nfsganesha-5/, OSD_FLAVOR=default, ceph=True, io.buildah.version=1.41.8, CEPH_GIT_REPO=https://github.com/irq0/ceph.git, CEPH_REF=19.2.3-47-gc24117fd552) 2026-03-06T13:43:10.063 INFO:journalctl@ceph.osd.2.vm08.stdout:Mar 06 13:43:10 vm08 podman[62771]: 2026-03-06 13:43:10.017297844 +0100 CET m=+0.063680591 container start 939e9ebfbd48dea3105bcd983bfaf4c05747f22d1fb73eabe106e4420b8d2d8b (image=harbor.clyso.com/custom-ceph/ceph/ceph@sha256:26363c7a4eea9ef5a0148afc7b2a22b6f486596d87a30c2a9fdcda5db3eca62b, name=ceph-82f504c8-1959-11f1-9a16-a78d135e200b-osd-2-deactivate, CEPH_REF=19.2.3-47-gc24117fd552, CEPH_SHA1=c24117fd5525679b799527bc1bd1f1dd0a2db5e2, FROM_IMAGE=rockylinux:9, GANESHA_REPO_BASEURL=https://buildlogs.centos.org/centos/$releasever-stream/storage/$basearch/nfsganesha-5/, OSD_FLAVOR=default, ceph=True, io.buildah.version=1.41.8, CEPH_GIT_REPO=https://github.com/irq0/ceph.git) 2026-03-06T13:43:10.063 INFO:journalctl@ceph.osd.2.vm08.stdout:Mar 06 13:43:10 vm08 podman[62771]: 2026-03-06 13:43:10.018257281 +0100 CET m=+0.064640028 container attach 939e9ebfbd48dea3105bcd983bfaf4c05747f22d1fb73eabe106e4420b8d2d8b (image=harbor.clyso.com/custom-ceph/ceph/ceph@sha256:26363c7a4eea9ef5a0148afc7b2a22b6f486596d87a30c2a9fdcda5db3eca62b, name=ceph-82f504c8-1959-11f1-9a16-a78d135e200b-osd-2-deactivate, GANESHA_REPO_BASEURL=https://buildlogs.centos.org/centos/$releasever-stream/storage/$basearch/nfsganesha-5/, OSD_FLAVOR=default, ceph=True, io.buildah.version=1.41.8, CEPH_GIT_REPO=https://github.com/irq0/ceph.git, CEPH_REF=19.2.3-47-gc24117fd552, CEPH_SHA1=c24117fd5525679b799527bc1bd1f1dd0a2db5e2, FROM_IMAGE=rockylinux:9) 2026-03-06T13:43:10.063 INFO:journalctl@ceph.osd.2.vm08.stdout:Mar 06 13:43:10 vm08 podman[62771]: 2026-03-06 13:43:09.9626339 +0100 CET m=+0.009016658 image pull 306e97de47e91c2b4b24d3dc09be3b3a12039b078f343d91220102acc6628a68 harbor.clyso.com/custom-ceph/ceph/ceph@sha256:26363c7a4eea9ef5a0148afc7b2a22b6f486596d87a30c2a9fdcda5db3eca62b 2026-03-06T13:43:10.449 INFO:journalctl@ceph.osd.2.vm08.stdout:Mar 06 13:43:10 vm08 podman[62771]: 2026-03-06 13:43:10.44812666 +0100 CET m=+0.494509407 container died 939e9ebfbd48dea3105bcd983bfaf4c05747f22d1fb73eabe106e4420b8d2d8b (image=harbor.clyso.com/custom-ceph/ceph/ceph@sha256:26363c7a4eea9ef5a0148afc7b2a22b6f486596d87a30c2a9fdcda5db3eca62b, name=ceph-82f504c8-1959-11f1-9a16-a78d135e200b-osd-2-deactivate, CEPH_REF=19.2.3-47-gc24117fd552, CEPH_SHA1=c24117fd5525679b799527bc1bd1f1dd0a2db5e2, FROM_IMAGE=rockylinux:9, GANESHA_REPO_BASEURL=https://buildlogs.centos.org/centos/$releasever-stream/storage/$basearch/nfsganesha-5/, OSD_FLAVOR=default, ceph=True, io.buildah.version=1.41.8, CEPH_GIT_REPO=https://github.com/irq0/ceph.git) 2026-03-06T13:43:10.482 DEBUG:teuthology.orchestra.run.vm08:> sudo pkill -f 'journalctl -f -n 0 -u ceph-82f504c8-1959-11f1-9a16-a78d135e200b@osd.2.service' 2026-03-06T13:43:10.526 DEBUG:teuthology.orchestra.run:got remote process result: None 2026-03-06T13:43:10.526 INFO:tasks.cephadm.osd.2:Stopped osd.2 2026-03-06T13:43:10.526 DEBUG:teuthology.orchestra.run.vm00:> sudo /home/ubuntu/cephtest/cephadm rm-cluster --fsid 82f504c8-1959-11f1-9a16-a78d135e200b --force --keep-logs 2026-03-06T13:43:10.833 INFO:teuthology.orchestra.run.vm00.stdout:Deleting cluster with fsid: 82f504c8-1959-11f1-9a16-a78d135e200b 2026-03-06T13:43:11.868 DEBUG:teuthology.orchestra.run.vm01:> sudo /home/ubuntu/cephtest/cephadm rm-cluster --fsid 82f504c8-1959-11f1-9a16-a78d135e200b --force --keep-logs 2026-03-06T13:43:12.178 INFO:teuthology.orchestra.run.vm01.stdout:Deleting cluster with fsid: 82f504c8-1959-11f1-9a16-a78d135e200b 2026-03-06T13:43:23.665 DEBUG:teuthology.orchestra.run.vm08:> sudo /home/ubuntu/cephtest/cephadm rm-cluster --fsid 82f504c8-1959-11f1-9a16-a78d135e200b --force --keep-logs 2026-03-06T13:43:23.966 INFO:teuthology.orchestra.run.vm08.stdout:Deleting cluster with fsid: 82f504c8-1959-11f1-9a16-a78d135e200b 2026-03-06T13:43:35.259 DEBUG:teuthology.orchestra.run.vm00:> sudo rm -f /etc/ceph/ceph.conf /etc/ceph/ceph.client.admin.keyring 2026-03-06T13:43:35.287 DEBUG:teuthology.orchestra.run.vm01:> sudo rm -f /etc/ceph/ceph.conf /etc/ceph/ceph.client.admin.keyring 2026-03-06T13:43:35.314 DEBUG:teuthology.orchestra.run.vm08:> sudo rm -f /etc/ceph/ceph.conf /etc/ceph/ceph.client.admin.keyring 2026-03-06T13:43:35.342 INFO:tasks.cephadm:Archiving crash dumps... 2026-03-06T13:43:35.342 DEBUG:teuthology.misc:Transferring archived files from vm00:/var/lib/ceph/82f504c8-1959-11f1-9a16-a78d135e200b/crash to /archive/irq0-2026-03-06_13:20:18-orch:cephadm:workunits-cobaltcore-storage-v19.2.3-fasttrack-3-none-default-vps/273/remote/vm00/crash 2026-03-06T13:43:35.342 DEBUG:teuthology.orchestra.run.vm00:> sudo tar c -f - -C /var/lib/ceph/82f504c8-1959-11f1-9a16-a78d135e200b/crash -- . 2026-03-06T13:43:35.371 INFO:teuthology.orchestra.run.vm00.stderr:tar: /var/lib/ceph/82f504c8-1959-11f1-9a16-a78d135e200b/crash: Cannot open: No such file or directory 2026-03-06T13:43:35.371 INFO:teuthology.orchestra.run.vm00.stderr:tar: Error is not recoverable: exiting now 2026-03-06T13:43:35.372 DEBUG:teuthology.misc:Transferring archived files from vm01:/var/lib/ceph/82f504c8-1959-11f1-9a16-a78d135e200b/crash to /archive/irq0-2026-03-06_13:20:18-orch:cephadm:workunits-cobaltcore-storage-v19.2.3-fasttrack-3-none-default-vps/273/remote/vm01/crash 2026-03-06T13:43:35.373 DEBUG:teuthology.orchestra.run.vm01:> sudo tar c -f - -C /var/lib/ceph/82f504c8-1959-11f1-9a16-a78d135e200b/crash -- . 2026-03-06T13:43:35.402 INFO:teuthology.orchestra.run.vm01.stderr:tar: /var/lib/ceph/82f504c8-1959-11f1-9a16-a78d135e200b/crash: Cannot open: No such file or directory 2026-03-06T13:43:35.402 INFO:teuthology.orchestra.run.vm01.stderr:tar: Error is not recoverable: exiting now 2026-03-06T13:43:35.403 DEBUG:teuthology.misc:Transferring archived files from vm08:/var/lib/ceph/82f504c8-1959-11f1-9a16-a78d135e200b/crash to /archive/irq0-2026-03-06_13:20:18-orch:cephadm:workunits-cobaltcore-storage-v19.2.3-fasttrack-3-none-default-vps/273/remote/vm08/crash 2026-03-06T13:43:35.403 DEBUG:teuthology.orchestra.run.vm08:> sudo tar c -f - -C /var/lib/ceph/82f504c8-1959-11f1-9a16-a78d135e200b/crash -- . 2026-03-06T13:43:35.432 INFO:teuthology.orchestra.run.vm08.stderr:tar: /var/lib/ceph/82f504c8-1959-11f1-9a16-a78d135e200b/crash: Cannot open: No such file or directory 2026-03-06T13:43:35.432 INFO:teuthology.orchestra.run.vm08.stderr:tar: Error is not recoverable: exiting now 2026-03-06T13:43:35.433 INFO:tasks.cephadm:Checking cluster log for badness... 2026-03-06T13:43:35.433 DEBUG:teuthology.orchestra.run.vm00:> sudo egrep '\[ERR\]|\[WRN\]|\[SEC\]' /var/log/ceph/82f504c8-1959-11f1-9a16-a78d135e200b/ceph.log | egrep CEPHADM_ | egrep -v '\(MDS_ALL_DOWN\)' | egrep -v '\(MDS_UP_LESS_THAN_MAX\)' | egrep -v MON_DOWN | egrep -v 'mons down' | egrep -v 'mon down' | egrep -v 'out of quorum' | egrep -v CEPHADM_STRAY_DAEMON | head -n 1 2026-03-06T13:43:35.463 INFO:tasks.cephadm:Compressing logs... 2026-03-06T13:43:35.463 DEBUG:teuthology.orchestra.run.vm00:> time sudo find /var/log/ceph /var/log/rbd-target-api -name '*.log' -print0 | sudo xargs --max-args=1 --max-procs=0 --verbose -0 --no-run-if-empty -- gzip -5 --verbose -- 2026-03-06T13:43:35.504 DEBUG:teuthology.orchestra.run.vm01:> time sudo find /var/log/ceph /var/log/rbd-target-api -name '*.log' -print0 | sudo xargs --max-args=1 --max-procs=0 --verbose -0 --no-run-if-empty -- gzip -5 --verbose -- 2026-03-06T13:43:35.506 DEBUG:teuthology.orchestra.run.vm08:> time sudo find /var/log/ceph /var/log/rbd-target-api -name '*.log' -print0 | sudo xargs --max-args=1 --max-procs=0 --verbose -0 --no-run-if-empty -- gzip -5 --verbose -- 2026-03-06T13:43:35.526 INFO:teuthology.orchestra.run.vm01.stderr:find: ‘/var/log/rbd-target-api’: No such file or directory 2026-03-06T13:43:35.527 INFO:teuthology.orchestra.run.vm01.stderr:gzip -5 --verbose -- /var/log/ceph/cephadm.log 2026-03-06T13:43:35.527 INFO:teuthology.orchestra.run.vm01.stderr:gzip -5 --verbose -- /var/log/ceph/82f504c8-1959-11f1-9a16-a78d135e200b/ceph-volume.log 2026-03-06T13:43:35.528 INFO:teuthology.orchestra.run.vm01.stderr:/var/log/ceph/cephadm.log: 88.0% -- replaced with /var/log/ceph/cephadm.log.gz 2026-03-06T13:43:35.528 INFO:teuthology.orchestra.run.vm01.stderr:gzip -5 --verbose -- /var/log/ceph/82f504c8-1959-11f1-9a16-a78d135e200b/ceph-mon.b.log 2026-03-06T13:43:35.528 INFO:teuthology.orchestra.run.vm01.stderr:gzip -5 --verbose -- /var/log/ceph/82f504c8-1959-11f1-9a16-a78d135e200b/ceph.audit.log 2026-03-06T13:43:35.530 INFO:teuthology.orchestra.run.vm00.stderr:find: ‘/var/log/rbd-target-api’: No such file or directory 2026-03-06T13:43:35.532 INFO:teuthology.orchestra.run.vm00.stderr:gzip -5 --verbose -- /var/log/ceph/cephadm.log 2026-03-06T13:43:35.532 INFO:teuthology.orchestra.run.vm01.stderr:/var/log/ceph/82f504c8-1959-11f1-9a16-a78d135e200b/ceph-volume.log: /var/log/ceph/82f504c8-1959-11f1-9a16-a78d135e200b/ceph-mon.b.log: 93.2% -- replaced with /var/log/ceph/82f504c8-1959-11f1-9a16-a78d135e200b/ceph-volume.log.gz 2026-03-06T13:43:35.532 INFO:teuthology.orchestra.run.vm00.stderr:gzip -5 --verbose -- /var/log/ceph/82f504c8-1959-11f1-9a16-a78d135e200b/ceph-mon.a.log 2026-03-06T13:43:35.534 INFO:teuthology.orchestra.run.vm00.stderr:/var/log/ceph/cephadm.log: 90.0% -- replaced with /var/log/ceph/cephadm.log.gz 2026-03-06T13:43:35.534 INFO:teuthology.orchestra.run.vm00.stderr:gzip -5 --verbose -- /var/log/ceph/82f504c8-1959-11f1-9a16-a78d135e200b/ceph.log 2026-03-06T13:43:35.535 INFO:teuthology.orchestra.run.vm08.stderr:find: gzip -5 --verbose -- /var/log/ceph/cephadm.log 2026-03-06T13:43:35.535 INFO:teuthology.orchestra.run.vm08.stderr:‘/var/log/rbd-target-api’: No such file or directory 2026-03-06T13:43:35.536 INFO:teuthology.orchestra.run.vm01.stderr:gzip -5 --verbose -- /var/log/ceph/82f504c8-1959-11f1-9a16-a78d135e200b/ceph.log 2026-03-06T13:43:35.536 INFO:teuthology.orchestra.run.vm08.stderr:gzip -5 --verbose -- /var/log/ceph/82f504c8-1959-11f1-9a16-a78d135e200b/ceph-volume.log 2026-03-06T13:43:35.536 INFO:teuthology.orchestra.run.vm00.stderr:/var/log/ceph/82f504c8-1959-11f1-9a16-a78d135e200b/ceph-mon.a.log: gzip -5 --verbose -- /var/log/ceph/82f504c8-1959-11f1-9a16-a78d135e200b/ceph-mgr.a.log 2026-03-06T13:43:35.537 INFO:teuthology.orchestra.run.vm01.stderr:/var/log/ceph/82f504c8-1959-11f1-9a16-a78d135e200b/ceph.audit.log: 89.5% -- replaced with /var/log/ceph/82f504c8-1959-11f1-9a16-a78d135e200b/ceph.audit.log.gz 2026-03-06T13:43:35.537 INFO:teuthology.orchestra.run.vm08.stderr:/var/log/ceph/cephadm.log: 88.9% -- replaced with /var/log/ceph/cephadm.log.gz 2026-03-06T13:43:35.537 INFO:teuthology.orchestra.run.vm08.stderr:gzip -5 --verbose -- /var/log/ceph/82f504c8-1959-11f1-9a16-a78d135e200b/ceph-mon.c.log 2026-03-06T13:43:35.537 INFO:teuthology.orchestra.run.vm00.stderr:/var/log/ceph/82f504c8-1959-11f1-9a16-a78d135e200b/ceph.log: 83.6% -- replaced with /var/log/ceph/82f504c8-1959-11f1-9a16-a78d135e200b/ceph.log.gz 2026-03-06T13:43:35.537 INFO:teuthology.orchestra.run.vm00.stderr:gzip -5 --verbose -- /var/log/ceph/82f504c8-1959-11f1-9a16-a78d135e200b/ceph.audit.log 2026-03-06T13:43:35.537 INFO:teuthology.orchestra.run.vm08.stderr:gzip -5 --verbose -- /var/log/ceph/82f504c8-1959-11f1-9a16-a78d135e200b/ceph.audit.log 2026-03-06T13:43:35.538 INFO:teuthology.orchestra.run.vm01.stderr:gzip -5 --verbose -- /var/log/ceph/82f504c8-1959-11f1-9a16-a78d135e200b/ceph.cephadm.log 2026-03-06T13:43:35.539 INFO:teuthology.orchestra.run.vm01.stderr:/var/log/ceph/82f504c8-1959-11f1-9a16-a78d135e200b/ceph.log: 82.5% -- replaced with /var/log/ceph/82f504c8-1959-11f1-9a16-a78d135e200b/ceph.log.gz 2026-03-06T13:43:35.539 INFO:teuthology.orchestra.run.vm01.stderr:gzip -5 --verbose -- /var/log/ceph/82f504c8-1959-11f1-9a16-a78d135e200b/ceph-mgr.b.log 2026-03-06T13:43:35.539 INFO:teuthology.orchestra.run.vm01.stderr:/var/log/ceph/82f504c8-1959-11f1-9a16-a78d135e200b/ceph.cephadm.log: 80.4% -- replaced with /var/log/ceph/82f504c8-1959-11f1-9a16-a78d135e200b/ceph.cephadm.log.gz 2026-03-06T13:43:35.539 INFO:teuthology.orchestra.run.vm01.stderr:gzip -5 --verbose -- /var/log/ceph/82f504c8-1959-11f1-9a16-a78d135e200b/ceph-osd.1.log 2026-03-06T13:43:35.542 INFO:teuthology.orchestra.run.vm00.stderr:/var/log/ceph/82f504c8-1959-11f1-9a16-a78d135e200b/ceph-mgr.a.log: gzip -5 --verbose -- /var/log/ceph/82f504c8-1959-11f1-9a16-a78d135e200b/ceph.cephadm.log 2026-03-06T13:43:35.543 INFO:teuthology.orchestra.run.vm08.stderr:/var/log/ceph/82f504c8-1959-11f1-9a16-a78d135e200b/ceph-volume.log: /var/log/ceph/82f504c8-1959-11f1-9a16-a78d135e200b/ceph-mon.c.log: gzip -5 --verbose -- /var/log/ceph/82f504c8-1959-11f1-9a16-a78d135e200b/ceph.log 2026-03-06T13:43:35.543 INFO:teuthology.orchestra.run.vm01.stderr:/var/log/ceph/82f504c8-1959-11f1-9a16-a78d135e200b/ceph-mgr.b.log: 89.9% -- replaced with /var/log/ceph/82f504c8-1959-11f1-9a16-a78d135e200b/ceph-mgr.b.log.gz 2026-03-06T13:43:35.543 INFO:teuthology.orchestra.run.vm01.stderr:gzip -5 --verbose -- /var/log/ceph/82f504c8-1959-11f1-9a16-a78d135e200b/ceph-client.rgw.myrealm1.myzone1.vm01.mmqnav.log 2026-03-06T13:43:35.544 INFO:teuthology.orchestra.run.vm08.stderr:/var/log/ceph/82f504c8-1959-11f1-9a16-a78d135e200b/ceph.audit.log: 89.6% 93.2% -- replaced with /var/log/ceph/82f504c8-1959-11f1-9a16-a78d135e200b/ceph-volume.log.gz 2026-03-06T13:43:35.544 INFO:teuthology.orchestra.run.vm08.stderr: -- replaced with /var/log/ceph/82f504c8-1959-11f1-9a16-a78d135e200b/ceph.audit.log.gz 2026-03-06T13:43:35.544 INFO:teuthology.orchestra.run.vm08.stderr:gzip -5 --verbose -- /var/log/ceph/82f504c8-1959-11f1-9a16-a78d135e200b/ceph.cephadm.log 2026-03-06T13:43:35.544 INFO:teuthology.orchestra.run.vm08.stderr:/var/log/ceph/82f504c8-1959-11f1-9a16-a78d135e200b/ceph.log: 83.0% -- replaced with /var/log/ceph/82f504c8-1959-11f1-9a16-a78d135e200b/ceph.log.gz 2026-03-06T13:43:35.545 INFO:teuthology.orchestra.run.vm08.stderr:gzip -5 --verbose -- /var/log/ceph/82f504c8-1959-11f1-9a16-a78d135e200b/ceph-osd.2.log 2026-03-06T13:43:35.545 INFO:teuthology.orchestra.run.vm08.stderr:/var/log/ceph/82f504c8-1959-11f1-9a16-a78d135e200b/ceph.cephadm.log: 80.4% -- replaced with /var/log/ceph/82f504c8-1959-11f1-9a16-a78d135e200b/ceph.cephadm.log.gz 2026-03-06T13:43:35.545 INFO:teuthology.orchestra.run.vm00.stderr:/var/log/ceph/82f504c8-1959-11f1-9a16-a78d135e200b/ceph.audit.log: 89.4% -- replaced with /var/log/ceph/82f504c8-1959-11f1-9a16-a78d135e200b/ceph.audit.log.gz 2026-03-06T13:43:35.546 INFO:teuthology.orchestra.run.vm08.stderr:gzip -5 --verbose -- /var/log/ceph/82f504c8-1959-11f1-9a16-a78d135e200b/ceph-client.rgw.myrealm1.myzone1.vm08.oycfkw.log 2026-03-06T13:43:35.548 INFO:teuthology.orchestra.run.vm00.stderr:gzip -5 --verbose -- /var/log/ceph/82f504c8-1959-11f1-9a16-a78d135e200b/ceph-volume.log 2026-03-06T13:43:35.549 INFO:teuthology.orchestra.run.vm00.stderr:/var/log/ceph/82f504c8-1959-11f1-9a16-a78d135e200b/ceph.cephadm.log: 82.8% -- replaced with /var/log/ceph/82f504c8-1959-11f1-9a16-a78d135e200b/ceph.cephadm.log.gz 2026-03-06T13:43:35.551 INFO:teuthology.orchestra.run.vm01.stderr:/var/log/ceph/82f504c8-1959-11f1-9a16-a78d135e200b/ceph-osd.1.log: /var/log/ceph/82f504c8-1959-11f1-9a16-a78d135e200b/ceph-client.rgw.myrealm1.myzone1.vm01.mmqnav.log: 63.0% -- replaced with /var/log/ceph/82f504c8-1959-11f1-9a16-a78d135e200b/ceph-client.rgw.myrealm1.myzone1.vm01.mmqnav.log.gz 2026-03-06T13:43:35.553 INFO:teuthology.orchestra.run.vm00.stderr:gzip -5 --verbose -- /var/log/ceph/82f504c8-1959-11f1-9a16-a78d135e200b/ceph-osd.0.log 2026-03-06T13:43:35.556 INFO:teuthology.orchestra.run.vm08.stderr:/var/log/ceph/82f504c8-1959-11f1-9a16-a78d135e200b/ceph-osd.2.log: /var/log/ceph/82f504c8-1959-11f1-9a16-a78d135e200b/ceph-client.rgw.myrealm1.myzone1.vm08.oycfkw.log: 63.2% -- replaced with /var/log/ceph/82f504c8-1959-11f1-9a16-a78d135e200b/ceph-client.rgw.myrealm1.myzone1.vm08.oycfkw.log.gz 2026-03-06T13:43:35.556 INFO:teuthology.orchestra.run.vm01.stderr: 92.3% -- replaced with /var/log/ceph/82f504c8-1959-11f1-9a16-a78d135e200b/ceph-mon.b.log.gz 2026-03-06T13:43:35.561 INFO:teuthology.orchestra.run.vm00.stderr:/var/log/ceph/82f504c8-1959-11f1-9a16-a78d135e200b/ceph-volume.log: 93.2% -- replaced with /var/log/ceph/82f504c8-1959-11f1-9a16-a78d135e200b/ceph-volume.log.gz 2026-03-06T13:43:35.566 INFO:teuthology.orchestra.run.vm08.stderr: 92.4% -- replaced with /var/log/ceph/82f504c8-1959-11f1-9a16-a78d135e200b/ceph-mon.c.log.gz 2026-03-06T13:43:35.576 INFO:teuthology.orchestra.run.vm00.stderr:/var/log/ceph/82f504c8-1959-11f1-9a16-a78d135e200b/ceph-osd.0.log: 89.7% -- replaced with /var/log/ceph/82f504c8-1959-11f1-9a16-a78d135e200b/ceph-mgr.a.log.gz 2026-03-06T13:43:35.643 INFO:teuthology.orchestra.run.vm00.stderr: 91.9% -- replaced with /var/log/ceph/82f504c8-1959-11f1-9a16-a78d135e200b/ceph-mon.a.log.gz 2026-03-06T13:43:35.741 INFO:teuthology.orchestra.run.vm08.stderr: 94.5% -- replaced with /var/log/ceph/82f504c8-1959-11f1-9a16-a78d135e200b/ceph-osd.2.log.gz 2026-03-06T13:43:35.743 INFO:teuthology.orchestra.run.vm08.stderr: 2026-03-06T13:43:35.743 INFO:teuthology.orchestra.run.vm08.stderr:real 0m0.219s 2026-03-06T13:43:35.743 INFO:teuthology.orchestra.run.vm08.stderr:user 0m0.218s 2026-03-06T13:43:35.743 INFO:teuthology.orchestra.run.vm08.stderr:sys 0m0.034s 2026-03-06T13:43:35.743 INFO:teuthology.orchestra.run.vm01.stderr: 94.4% -- replaced with /var/log/ceph/82f504c8-1959-11f1-9a16-a78d135e200b/ceph-osd.1.log.gz 2026-03-06T13:43:35.745 INFO:teuthology.orchestra.run.vm01.stderr: 2026-03-06T13:43:35.745 INFO:teuthology.orchestra.run.vm01.stderr:real 0m0.227s 2026-03-06T13:43:35.745 INFO:teuthology.orchestra.run.vm01.stderr:user 0m0.229s 2026-03-06T13:43:35.745 INFO:teuthology.orchestra.run.vm01.stderr:sys 0m0.026s 2026-03-06T13:43:35.762 INFO:teuthology.orchestra.run.vm00.stderr: 94.4% -- replaced with /var/log/ceph/82f504c8-1959-11f1-9a16-a78d135e200b/ceph-osd.0.log.gz 2026-03-06T13:43:35.764 INFO:teuthology.orchestra.run.vm00.stderr: 2026-03-06T13:43:35.764 INFO:teuthology.orchestra.run.vm00.stderr:real 0m0.243s 2026-03-06T13:43:35.764 INFO:teuthology.orchestra.run.vm00.stderr:user 0m0.327s 2026-03-06T13:43:35.764 INFO:teuthology.orchestra.run.vm00.stderr:sys 0m0.029s 2026-03-06T13:43:35.765 INFO:tasks.cephadm:Archiving logs... 2026-03-06T13:43:35.765 DEBUG:teuthology.misc:Transferring archived files from vm00:/var/log/ceph to /archive/irq0-2026-03-06_13:20:18-orch:cephadm:workunits-cobaltcore-storage-v19.2.3-fasttrack-3-none-default-vps/273/remote/vm00/log 2026-03-06T13:43:35.765 DEBUG:teuthology.orchestra.run.vm00:> sudo tar c -f - -C /var/log/ceph -- . 2026-03-06T13:43:35.849 DEBUG:teuthology.misc:Transferring archived files from vm01:/var/log/ceph to /archive/irq0-2026-03-06_13:20:18-orch:cephadm:workunits-cobaltcore-storage-v19.2.3-fasttrack-3-none-default-vps/273/remote/vm01/log 2026-03-06T13:43:35.849 DEBUG:teuthology.orchestra.run.vm01:> sudo tar c -f - -C /var/log/ceph -- . 2026-03-06T13:43:35.893 DEBUG:teuthology.misc:Transferring archived files from vm08:/var/log/ceph to /archive/irq0-2026-03-06_13:20:18-orch:cephadm:workunits-cobaltcore-storage-v19.2.3-fasttrack-3-none-default-vps/273/remote/vm08/log 2026-03-06T13:43:35.894 DEBUG:teuthology.orchestra.run.vm08:> sudo tar c -f - -C /var/log/ceph -- . 2026-03-06T13:43:35.938 INFO:tasks.cephadm:Removing cluster... 2026-03-06T13:43:35.938 DEBUG:teuthology.orchestra.run.vm00:> sudo /home/ubuntu/cephtest/cephadm rm-cluster --fsid 82f504c8-1959-11f1-9a16-a78d135e200b --force 2026-03-06T13:43:36.253 INFO:teuthology.orchestra.run.vm00.stdout:Deleting cluster with fsid: 82f504c8-1959-11f1-9a16-a78d135e200b 2026-03-06T13:43:36.494 DEBUG:teuthology.orchestra.run.vm01:> sudo /home/ubuntu/cephtest/cephadm rm-cluster --fsid 82f504c8-1959-11f1-9a16-a78d135e200b --force 2026-03-06T13:43:36.792 INFO:teuthology.orchestra.run.vm01.stdout:Deleting cluster with fsid: 82f504c8-1959-11f1-9a16-a78d135e200b 2026-03-06T13:43:37.037 DEBUG:teuthology.orchestra.run.vm08:> sudo /home/ubuntu/cephtest/cephadm rm-cluster --fsid 82f504c8-1959-11f1-9a16-a78d135e200b --force 2026-03-06T13:43:37.342 INFO:teuthology.orchestra.run.vm08.stdout:Deleting cluster with fsid: 82f504c8-1959-11f1-9a16-a78d135e200b 2026-03-06T13:43:37.555 INFO:tasks.cephadm:Removing cephadm ... 2026-03-06T13:43:37.556 DEBUG:teuthology.orchestra.run.vm00:> rm -rf /home/ubuntu/cephtest/cephadm 2026-03-06T13:43:37.573 DEBUG:teuthology.orchestra.run.vm01:> rm -rf /home/ubuntu/cephtest/cephadm 2026-03-06T13:43:37.589 DEBUG:teuthology.orchestra.run.vm08:> rm -rf /home/ubuntu/cephtest/cephadm 2026-03-06T13:43:37.606 INFO:tasks.cephadm:Teardown complete 2026-03-06T13:43:37.607 DEBUG:teuthology.run_tasks:Unwinding manager install 2026-03-06T13:43:37.609 INFO:teuthology.task.install.util:Removing shipped files: /home/ubuntu/cephtest/valgrind.supp /usr/bin/daemon-helper /usr/bin/adjust-ulimits /usr/bin/stdin-killer... 2026-03-06T13:43:37.610 DEBUG:teuthology.orchestra.run.vm00:> sudo rm -f -- /home/ubuntu/cephtest/valgrind.supp /usr/bin/daemon-helper /usr/bin/adjust-ulimits /usr/bin/stdin-killer 2026-03-06T13:43:37.617 DEBUG:teuthology.orchestra.run.vm01:> sudo rm -f -- /home/ubuntu/cephtest/valgrind.supp /usr/bin/daemon-helper /usr/bin/adjust-ulimits /usr/bin/stdin-killer 2026-03-06T13:43:37.631 DEBUG:teuthology.orchestra.run.vm08:> sudo rm -f -- /home/ubuntu/cephtest/valgrind.supp /usr/bin/daemon-helper /usr/bin/adjust-ulimits /usr/bin/stdin-killer 2026-03-06T13:43:37.680 INFO:teuthology.task.install.rpm:Removing packages: ceph-radosgw, ceph-test, ceph, ceph-base, cephadm, ceph-immutable-object-cache, ceph-mgr, ceph-mgr-dashboard, ceph-mgr-diskprediction-local, ceph-mgr-rook, ceph-mgr-cephadm, ceph-fuse, ceph-volume, librados-devel, libcephfs2, libcephfs-devel, librados2, librbd1, python3-rados, python3-rgw, python3-cephfs, python3-rbd, rbd-fuse, rbd-mirror, rbd-nbd on rpm system. 2026-03-06T13:43:37.680 DEBUG:teuthology.orchestra.run.vm00:> 2026-03-06T13:43:37.680 DEBUG:teuthology.orchestra.run.vm00:> for d in ceph-radosgw ceph-test ceph ceph-base cephadm ceph-immutable-object-cache ceph-mgr ceph-mgr-dashboard ceph-mgr-diskprediction-local ceph-mgr-rook ceph-mgr-cephadm ceph-fuse ceph-volume librados-devel libcephfs2 libcephfs-devel librados2 librbd1 python3-rados python3-rgw python3-cephfs python3-rbd rbd-fuse rbd-mirror rbd-nbd ; do 2026-03-06T13:43:37.680 DEBUG:teuthology.orchestra.run.vm00:> sudo yum -y remove $d || true 2026-03-06T13:43:37.680 DEBUG:teuthology.orchestra.run.vm00:> done 2026-03-06T13:43:37.686 INFO:teuthology.task.install.rpm:Removing packages: ceph-radosgw, ceph-test, ceph, ceph-base, cephadm, ceph-immutable-object-cache, ceph-mgr, ceph-mgr-dashboard, ceph-mgr-diskprediction-local, ceph-mgr-rook, ceph-mgr-cephadm, ceph-fuse, ceph-volume, librados-devel, libcephfs2, libcephfs-devel, librados2, librbd1, python3-rados, python3-rgw, python3-cephfs, python3-rbd, rbd-fuse, rbd-mirror, rbd-nbd on rpm system. 2026-03-06T13:43:37.686 DEBUG:teuthology.orchestra.run.vm01:> 2026-03-06T13:43:37.686 DEBUG:teuthology.orchestra.run.vm01:> for d in ceph-radosgw ceph-test ceph ceph-base cephadm ceph-immutable-object-cache ceph-mgr ceph-mgr-dashboard ceph-mgr-diskprediction-local ceph-mgr-rook ceph-mgr-cephadm ceph-fuse ceph-volume librados-devel libcephfs2 libcephfs-devel librados2 librbd1 python3-rados python3-rgw python3-cephfs python3-rbd rbd-fuse rbd-mirror rbd-nbd ; do 2026-03-06T13:43:37.686 DEBUG:teuthology.orchestra.run.vm01:> sudo yum -y remove $d || true 2026-03-06T13:43:37.686 DEBUG:teuthology.orchestra.run.vm01:> done 2026-03-06T13:43:37.692 INFO:teuthology.task.install.rpm:Removing packages: ceph-radosgw, ceph-test, ceph, ceph-base, cephadm, ceph-immutable-object-cache, ceph-mgr, ceph-mgr-dashboard, ceph-mgr-diskprediction-local, ceph-mgr-rook, ceph-mgr-cephadm, ceph-fuse, ceph-volume, librados-devel, libcephfs2, libcephfs-devel, librados2, librbd1, python3-rados, python3-rgw, python3-cephfs, python3-rbd, rbd-fuse, rbd-mirror, rbd-nbd on rpm system. 2026-03-06T13:43:37.692 DEBUG:teuthology.orchestra.run.vm08:> 2026-03-06T13:43:37.692 DEBUG:teuthology.orchestra.run.vm08:> for d in ceph-radosgw ceph-test ceph ceph-base cephadm ceph-immutable-object-cache ceph-mgr ceph-mgr-dashboard ceph-mgr-diskprediction-local ceph-mgr-rook ceph-mgr-cephadm ceph-fuse ceph-volume librados-devel libcephfs2 libcephfs-devel librados2 librbd1 python3-rados python3-rgw python3-cephfs python3-rbd rbd-fuse rbd-mirror rbd-nbd ; do 2026-03-06T13:43:37.692 DEBUG:teuthology.orchestra.run.vm08:> sudo yum -y remove $d || true 2026-03-06T13:43:37.692 DEBUG:teuthology.orchestra.run.vm08:> done 2026-03-06T13:43:37.873 INFO:teuthology.orchestra.run.vm00.stdout:Dependencies resolved. 2026-03-06T13:43:37.873 INFO:teuthology.orchestra.run.vm00.stdout:================================================================================ 2026-03-06T13:43:37.873 INFO:teuthology.orchestra.run.vm00.stdout: Package Arch Version Repo Size 2026-03-06T13:43:37.873 INFO:teuthology.orchestra.run.vm00.stdout:================================================================================ 2026-03-06T13:43:37.873 INFO:teuthology.orchestra.run.vm00.stdout:Removing: 2026-03-06T13:43:37.873 INFO:teuthology.orchestra.run.vm00.stdout: ceph-radosgw x86_64 2:19.2.3-47.gc24117fd552.el9.clyso @ceph 39 M 2026-03-06T13:43:37.873 INFO:teuthology.orchestra.run.vm00.stdout:Removing unused dependencies: 2026-03-06T13:43:37.873 INFO:teuthology.orchestra.run.vm00.stdout: mailcap noarch 2.1.49-5.el9 @baseos 78 k 2026-03-06T13:43:37.873 INFO:teuthology.orchestra.run.vm00.stdout: 2026-03-06T13:43:37.873 INFO:teuthology.orchestra.run.vm00.stdout:Transaction Summary 2026-03-06T13:43:37.873 INFO:teuthology.orchestra.run.vm00.stdout:================================================================================ 2026-03-06T13:43:37.874 INFO:teuthology.orchestra.run.vm00.stdout:Remove 2 Packages 2026-03-06T13:43:37.874 INFO:teuthology.orchestra.run.vm00.stdout: 2026-03-06T13:43:37.874 INFO:teuthology.orchestra.run.vm00.stdout:Freed space: 39 M 2026-03-06T13:43:37.874 INFO:teuthology.orchestra.run.vm00.stdout:Running transaction check 2026-03-06T13:43:37.876 INFO:teuthology.orchestra.run.vm00.stdout:Transaction check succeeded. 2026-03-06T13:43:37.876 INFO:teuthology.orchestra.run.vm00.stdout:Running transaction test 2026-03-06T13:43:37.882 INFO:teuthology.orchestra.run.vm01.stdout:Dependencies resolved. 2026-03-06T13:43:37.882 INFO:teuthology.orchestra.run.vm01.stdout:================================================================================ 2026-03-06T13:43:37.882 INFO:teuthology.orchestra.run.vm01.stdout: Package Arch Version Repo Size 2026-03-06T13:43:37.883 INFO:teuthology.orchestra.run.vm01.stdout:================================================================================ 2026-03-06T13:43:37.883 INFO:teuthology.orchestra.run.vm01.stdout:Removing: 2026-03-06T13:43:37.883 INFO:teuthology.orchestra.run.vm01.stdout: ceph-radosgw x86_64 2:19.2.3-47.gc24117fd552.el9.clyso @ceph 39 M 2026-03-06T13:43:37.883 INFO:teuthology.orchestra.run.vm01.stdout:Removing unused dependencies: 2026-03-06T13:43:37.883 INFO:teuthology.orchestra.run.vm01.stdout: mailcap noarch 2.1.49-5.el9 @baseos 78 k 2026-03-06T13:43:37.883 INFO:teuthology.orchestra.run.vm01.stdout: 2026-03-06T13:43:37.883 INFO:teuthology.orchestra.run.vm01.stdout:Transaction Summary 2026-03-06T13:43:37.883 INFO:teuthology.orchestra.run.vm01.stdout:================================================================================ 2026-03-06T13:43:37.883 INFO:teuthology.orchestra.run.vm01.stdout:Remove 2 Packages 2026-03-06T13:43:37.883 INFO:teuthology.orchestra.run.vm01.stdout: 2026-03-06T13:43:37.883 INFO:teuthology.orchestra.run.vm01.stdout:Freed space: 39 M 2026-03-06T13:43:37.883 INFO:teuthology.orchestra.run.vm01.stdout:Running transaction check 2026-03-06T13:43:37.885 INFO:teuthology.orchestra.run.vm01.stdout:Transaction check succeeded. 2026-03-06T13:43:37.885 INFO:teuthology.orchestra.run.vm01.stdout:Running transaction test 2026-03-06T13:43:37.891 INFO:teuthology.orchestra.run.vm00.stdout:Transaction test succeeded. 2026-03-06T13:43:37.891 INFO:teuthology.orchestra.run.vm00.stdout:Running transaction 2026-03-06T13:43:37.899 INFO:teuthology.orchestra.run.vm01.stdout:Transaction test succeeded. 2026-03-06T13:43:37.899 INFO:teuthology.orchestra.run.vm01.stdout:Running transaction 2026-03-06T13:43:37.911 INFO:teuthology.orchestra.run.vm08.stdout:Dependencies resolved. 2026-03-06T13:43:37.911 INFO:teuthology.orchestra.run.vm08.stdout:================================================================================ 2026-03-06T13:43:37.911 INFO:teuthology.orchestra.run.vm08.stdout: Package Arch Version Repo Size 2026-03-06T13:43:37.911 INFO:teuthology.orchestra.run.vm08.stdout:================================================================================ 2026-03-06T13:43:37.911 INFO:teuthology.orchestra.run.vm08.stdout:Removing: 2026-03-06T13:43:37.911 INFO:teuthology.orchestra.run.vm08.stdout: ceph-radosgw x86_64 2:19.2.3-47.gc24117fd552.el9.clyso @ceph 39 M 2026-03-06T13:43:37.911 INFO:teuthology.orchestra.run.vm08.stdout:Removing unused dependencies: 2026-03-06T13:43:37.911 INFO:teuthology.orchestra.run.vm08.stdout: mailcap noarch 2.1.49-5.el9 @baseos 78 k 2026-03-06T13:43:37.911 INFO:teuthology.orchestra.run.vm08.stdout: 2026-03-06T13:43:37.911 INFO:teuthology.orchestra.run.vm08.stdout:Transaction Summary 2026-03-06T13:43:37.911 INFO:teuthology.orchestra.run.vm08.stdout:================================================================================ 2026-03-06T13:43:37.911 INFO:teuthology.orchestra.run.vm08.stdout:Remove 2 Packages 2026-03-06T13:43:37.911 INFO:teuthology.orchestra.run.vm08.stdout: 2026-03-06T13:43:37.911 INFO:teuthology.orchestra.run.vm08.stdout:Freed space: 39 M 2026-03-06T13:43:37.911 INFO:teuthology.orchestra.run.vm08.stdout:Running transaction check 2026-03-06T13:43:37.914 INFO:teuthology.orchestra.run.vm08.stdout:Transaction check succeeded. 2026-03-06T13:43:37.914 INFO:teuthology.orchestra.run.vm08.stdout:Running transaction test 2026-03-06T13:43:37.924 INFO:teuthology.orchestra.run.vm00.stdout: Preparing : 1/1 2026-03-06T13:43:37.928 INFO:teuthology.orchestra.run.vm08.stdout:Transaction test succeeded. 2026-03-06T13:43:37.928 INFO:teuthology.orchestra.run.vm08.stdout:Running transaction 2026-03-06T13:43:37.932 INFO:teuthology.orchestra.run.vm01.stdout: Preparing : 1/1 2026-03-06T13:43:37.946 INFO:teuthology.orchestra.run.vm00.stdout: Running scriptlet: ceph-radosgw-2:19.2.3-47.gc24117fd552.el9.clyso.x86_ 1/2 2026-03-06T13:43:37.946 INFO:teuthology.orchestra.run.vm00.stdout:Glob pattern passed to enable, but globs are not supported for this. 2026-03-06T13:43:37.946 INFO:teuthology.orchestra.run.vm00.stdout:Invalid unit name "ceph-radosgw@*.service" escaped as "ceph-radosgw@\x2a.service". 2026-03-06T13:43:37.946 INFO:teuthology.orchestra.run.vm00.stdout:Removed "/etc/systemd/system/multi-user.target.wants/ceph-radosgw.target". 2026-03-06T13:43:37.946 INFO:teuthology.orchestra.run.vm00.stdout:Removed "/etc/systemd/system/ceph.target.wants/ceph-radosgw.target". 2026-03-06T13:43:37.946 INFO:teuthology.orchestra.run.vm00.stdout: 2026-03-06T13:43:37.949 INFO:teuthology.orchestra.run.vm00.stdout: Erasing : ceph-radosgw-2:19.2.3-47.gc24117fd552.el9.clyso.x86_ 1/2 2026-03-06T13:43:37.957 INFO:teuthology.orchestra.run.vm00.stdout: Running scriptlet: ceph-radosgw-2:19.2.3-47.gc24117fd552.el9.clyso.x86_ 1/2 2026-03-06T13:43:37.957 INFO:teuthology.orchestra.run.vm01.stdout: Running scriptlet: ceph-radosgw-2:19.2.3-47.gc24117fd552.el9.clyso.x86_ 1/2 2026-03-06T13:43:37.957 INFO:teuthology.orchestra.run.vm01.stdout:Glob pattern passed to enable, but globs are not supported for this. 2026-03-06T13:43:37.957 INFO:teuthology.orchestra.run.vm01.stdout:Invalid unit name "ceph-radosgw@*.service" escaped as "ceph-radosgw@\x2a.service". 2026-03-06T13:43:37.957 INFO:teuthology.orchestra.run.vm01.stdout:Removed "/etc/systemd/system/multi-user.target.wants/ceph-radosgw.target". 2026-03-06T13:43:37.958 INFO:teuthology.orchestra.run.vm01.stdout:Removed "/etc/systemd/system/ceph.target.wants/ceph-radosgw.target". 2026-03-06T13:43:37.958 INFO:teuthology.orchestra.run.vm01.stdout: 2026-03-06T13:43:37.958 INFO:teuthology.orchestra.run.vm08.stdout: Preparing : 1/1 2026-03-06T13:43:37.961 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : ceph-radosgw-2:19.2.3-47.gc24117fd552.el9.clyso.x86_ 1/2 2026-03-06T13:43:37.971 INFO:teuthology.orchestra.run.vm01.stdout: Running scriptlet: ceph-radosgw-2:19.2.3-47.gc24117fd552.el9.clyso.x86_ 1/2 2026-03-06T13:43:37.972 INFO:teuthology.orchestra.run.vm00.stdout: Erasing : mailcap-2.1.49-5.el9.noarch 2/2 2026-03-06T13:43:37.980 INFO:teuthology.orchestra.run.vm08.stdout: Running scriptlet: ceph-radosgw-2:19.2.3-47.gc24117fd552.el9.clyso.x86_ 1/2 2026-03-06T13:43:37.980 INFO:teuthology.orchestra.run.vm08.stdout:Glob pattern passed to enable, but globs are not supported for this. 2026-03-06T13:43:37.980 INFO:teuthology.orchestra.run.vm08.stdout:Invalid unit name "ceph-radosgw@*.service" escaped as "ceph-radosgw@\x2a.service". 2026-03-06T13:43:37.980 INFO:teuthology.orchestra.run.vm08.stdout:Removed "/etc/systemd/system/multi-user.target.wants/ceph-radosgw.target". 2026-03-06T13:43:37.980 INFO:teuthology.orchestra.run.vm08.stdout:Removed "/etc/systemd/system/ceph.target.wants/ceph-radosgw.target". 2026-03-06T13:43:37.981 INFO:teuthology.orchestra.run.vm08.stdout: 2026-03-06T13:43:37.984 INFO:teuthology.orchestra.run.vm08.stdout: Erasing : ceph-radosgw-2:19.2.3-47.gc24117fd552.el9.clyso.x86_ 1/2 2026-03-06T13:43:37.987 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : mailcap-2.1.49-5.el9.noarch 2/2 2026-03-06T13:43:37.993 INFO:teuthology.orchestra.run.vm08.stdout: Running scriptlet: ceph-radosgw-2:19.2.3-47.gc24117fd552.el9.clyso.x86_ 1/2 2026-03-06T13:43:38.009 INFO:teuthology.orchestra.run.vm08.stdout: Erasing : mailcap-2.1.49-5.el9.noarch 2/2 2026-03-06T13:43:38.051 INFO:teuthology.orchestra.run.vm00.stdout: Running scriptlet: mailcap-2.1.49-5.el9.noarch 2/2 2026-03-06T13:43:38.051 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : ceph-radosgw-2:19.2.3-47.gc24117fd552.el9.clyso.x86_ 1/2 2026-03-06T13:43:38.061 INFO:teuthology.orchestra.run.vm01.stdout: Running scriptlet: mailcap-2.1.49-5.el9.noarch 2/2 2026-03-06T13:43:38.061 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : ceph-radosgw-2:19.2.3-47.gc24117fd552.el9.clyso.x86_ 1/2 2026-03-06T13:43:38.097 INFO:teuthology.orchestra.run.vm08.stdout: Running scriptlet: mailcap-2.1.49-5.el9.noarch 2/2 2026-03-06T13:43:38.098 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : ceph-radosgw-2:19.2.3-47.gc24117fd552.el9.clyso.x86_ 1/2 2026-03-06T13:43:38.110 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : mailcap-2.1.49-5.el9.noarch 2/2 2026-03-06T13:43:38.110 INFO:teuthology.orchestra.run.vm00.stdout: 2026-03-06T13:43:38.110 INFO:teuthology.orchestra.run.vm00.stdout:Removed: 2026-03-06T13:43:38.110 INFO:teuthology.orchestra.run.vm00.stdout: ceph-radosgw-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 2026-03-06T13:43:38.110 INFO:teuthology.orchestra.run.vm00.stdout: mailcap-2.1.49-5.el9.noarch 2026-03-06T13:43:38.110 INFO:teuthology.orchestra.run.vm00.stdout: 2026-03-06T13:43:38.110 INFO:teuthology.orchestra.run.vm00.stdout:Complete! 2026-03-06T13:43:38.111 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : mailcap-2.1.49-5.el9.noarch 2/2 2026-03-06T13:43:38.111 INFO:teuthology.orchestra.run.vm01.stdout: 2026-03-06T13:43:38.111 INFO:teuthology.orchestra.run.vm01.stdout:Removed: 2026-03-06T13:43:38.111 INFO:teuthology.orchestra.run.vm01.stdout: ceph-radosgw-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 2026-03-06T13:43:38.111 INFO:teuthology.orchestra.run.vm01.stdout: mailcap-2.1.49-5.el9.noarch 2026-03-06T13:43:38.111 INFO:teuthology.orchestra.run.vm01.stdout: 2026-03-06T13:43:38.111 INFO:teuthology.orchestra.run.vm01.stdout:Complete! 2026-03-06T13:43:38.160 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : mailcap-2.1.49-5.el9.noarch 2/2 2026-03-06T13:43:38.160 INFO:teuthology.orchestra.run.vm08.stdout: 2026-03-06T13:43:38.160 INFO:teuthology.orchestra.run.vm08.stdout:Removed: 2026-03-06T13:43:38.160 INFO:teuthology.orchestra.run.vm08.stdout: ceph-radosgw-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 2026-03-06T13:43:38.160 INFO:teuthology.orchestra.run.vm08.stdout: mailcap-2.1.49-5.el9.noarch 2026-03-06T13:43:38.160 INFO:teuthology.orchestra.run.vm08.stdout: 2026-03-06T13:43:38.160 INFO:teuthology.orchestra.run.vm08.stdout:Complete! 2026-03-06T13:43:38.380 INFO:teuthology.orchestra.run.vm00.stdout:Dependencies resolved. 2026-03-06T13:43:38.380 INFO:teuthology.orchestra.run.vm00.stdout:================================================================================ 2026-03-06T13:43:38.380 INFO:teuthology.orchestra.run.vm00.stdout: Package Arch Version Repository Size 2026-03-06T13:43:38.380 INFO:teuthology.orchestra.run.vm00.stdout:================================================================================ 2026-03-06T13:43:38.380 INFO:teuthology.orchestra.run.vm00.stdout:Removing: 2026-03-06T13:43:38.380 INFO:teuthology.orchestra.run.vm00.stdout: ceph-test x86_64 2:19.2.3-47.gc24117fd552.el9.clyso @ceph 210 M 2026-03-06T13:43:38.380 INFO:teuthology.orchestra.run.vm00.stdout:Removing unused dependencies: 2026-03-06T13:43:38.380 INFO:teuthology.orchestra.run.vm00.stdout: socat x86_64 1.7.4.1-8.el9 @appstream 1.1 M 2026-03-06T13:43:38.380 INFO:teuthology.orchestra.run.vm00.stdout: xmlstarlet x86_64 1.6.1-20.el9 @appstream 195 k 2026-03-06T13:43:38.380 INFO:teuthology.orchestra.run.vm00.stdout: 2026-03-06T13:43:38.380 INFO:teuthology.orchestra.run.vm00.stdout:Transaction Summary 2026-03-06T13:43:38.380 INFO:teuthology.orchestra.run.vm00.stdout:================================================================================ 2026-03-06T13:43:38.380 INFO:teuthology.orchestra.run.vm00.stdout:Remove 3 Packages 2026-03-06T13:43:38.380 INFO:teuthology.orchestra.run.vm00.stdout: 2026-03-06T13:43:38.381 INFO:teuthology.orchestra.run.vm00.stdout:Freed space: 211 M 2026-03-06T13:43:38.381 INFO:teuthology.orchestra.run.vm00.stdout:Running transaction check 2026-03-06T13:43:38.382 INFO:teuthology.orchestra.run.vm01.stdout:Dependencies resolved. 2026-03-06T13:43:38.383 INFO:teuthology.orchestra.run.vm01.stdout:================================================================================ 2026-03-06T13:43:38.383 INFO:teuthology.orchestra.run.vm01.stdout: Package Arch Version Repository Size 2026-03-06T13:43:38.383 INFO:teuthology.orchestra.run.vm01.stdout:================================================================================ 2026-03-06T13:43:38.383 INFO:teuthology.orchestra.run.vm01.stdout:Removing: 2026-03-06T13:43:38.383 INFO:teuthology.orchestra.run.vm01.stdout: ceph-test x86_64 2:19.2.3-47.gc24117fd552.el9.clyso @ceph 210 M 2026-03-06T13:43:38.383 INFO:teuthology.orchestra.run.vm01.stdout:Removing unused dependencies: 2026-03-06T13:43:38.383 INFO:teuthology.orchestra.run.vm01.stdout: socat x86_64 1.7.4.1-8.el9 @appstream 1.1 M 2026-03-06T13:43:38.383 INFO:teuthology.orchestra.run.vm01.stdout: xmlstarlet x86_64 1.6.1-20.el9 @appstream 195 k 2026-03-06T13:43:38.383 INFO:teuthology.orchestra.run.vm01.stdout: 2026-03-06T13:43:38.383 INFO:teuthology.orchestra.run.vm01.stdout:Transaction Summary 2026-03-06T13:43:38.383 INFO:teuthology.orchestra.run.vm01.stdout:================================================================================ 2026-03-06T13:43:38.383 INFO:teuthology.orchestra.run.vm01.stdout:Remove 3 Packages 2026-03-06T13:43:38.383 INFO:teuthology.orchestra.run.vm01.stdout: 2026-03-06T13:43:38.383 INFO:teuthology.orchestra.run.vm01.stdout:Freed space: 211 M 2026-03-06T13:43:38.383 INFO:teuthology.orchestra.run.vm01.stdout:Running transaction check 2026-03-06T13:43:38.384 INFO:teuthology.orchestra.run.vm00.stdout:Transaction check succeeded. 2026-03-06T13:43:38.384 INFO:teuthology.orchestra.run.vm00.stdout:Running transaction test 2026-03-06T13:43:38.386 INFO:teuthology.orchestra.run.vm01.stdout:Transaction check succeeded. 2026-03-06T13:43:38.386 INFO:teuthology.orchestra.run.vm01.stdout:Running transaction test 2026-03-06T13:43:38.400 INFO:teuthology.orchestra.run.vm08.stdout:Dependencies resolved. 2026-03-06T13:43:38.401 INFO:teuthology.orchestra.run.vm08.stdout:================================================================================ 2026-03-06T13:43:38.401 INFO:teuthology.orchestra.run.vm08.stdout: Package Arch Version Repository Size 2026-03-06T13:43:38.401 INFO:teuthology.orchestra.run.vm08.stdout:================================================================================ 2026-03-06T13:43:38.401 INFO:teuthology.orchestra.run.vm08.stdout:Removing: 2026-03-06T13:43:38.401 INFO:teuthology.orchestra.run.vm08.stdout: ceph-test x86_64 2:19.2.3-47.gc24117fd552.el9.clyso @ceph 210 M 2026-03-06T13:43:38.401 INFO:teuthology.orchestra.run.vm08.stdout:Removing unused dependencies: 2026-03-06T13:43:38.401 INFO:teuthology.orchestra.run.vm08.stdout: socat x86_64 1.7.4.1-8.el9 @appstream 1.1 M 2026-03-06T13:43:38.401 INFO:teuthology.orchestra.run.vm08.stdout: xmlstarlet x86_64 1.6.1-20.el9 @appstream 195 k 2026-03-06T13:43:38.401 INFO:teuthology.orchestra.run.vm08.stdout: 2026-03-06T13:43:38.401 INFO:teuthology.orchestra.run.vm08.stdout:Transaction Summary 2026-03-06T13:43:38.401 INFO:teuthology.orchestra.run.vm08.stdout:================================================================================ 2026-03-06T13:43:38.401 INFO:teuthology.orchestra.run.vm08.stdout:Remove 3 Packages 2026-03-06T13:43:38.401 INFO:teuthology.orchestra.run.vm08.stdout: 2026-03-06T13:43:38.401 INFO:teuthology.orchestra.run.vm08.stdout:Freed space: 211 M 2026-03-06T13:43:38.401 INFO:teuthology.orchestra.run.vm08.stdout:Running transaction check 2026-03-06T13:43:38.404 INFO:teuthology.orchestra.run.vm08.stdout:Transaction check succeeded. 2026-03-06T13:43:38.404 INFO:teuthology.orchestra.run.vm08.stdout:Running transaction test 2026-03-06T13:43:38.411 INFO:teuthology.orchestra.run.vm01.stdout:Transaction test succeeded. 2026-03-06T13:43:38.411 INFO:teuthology.orchestra.run.vm01.stdout:Running transaction 2026-03-06T13:43:38.412 INFO:teuthology.orchestra.run.vm00.stdout:Transaction test succeeded. 2026-03-06T13:43:38.412 INFO:teuthology.orchestra.run.vm00.stdout:Running transaction 2026-03-06T13:43:38.433 INFO:teuthology.orchestra.run.vm08.stdout:Transaction test succeeded. 2026-03-06T13:43:38.434 INFO:teuthology.orchestra.run.vm08.stdout:Running transaction 2026-03-06T13:43:38.482 INFO:teuthology.orchestra.run.vm01.stdout: Preparing : 1/1 2026-03-06T13:43:38.484 INFO:teuthology.orchestra.run.vm00.stdout: Preparing : 1/1 2026-03-06T13:43:38.489 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : ceph-test-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 1/3 2026-03-06T13:43:38.491 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : socat-1.7.4.1-8.el9.x86_64 2/3 2026-03-06T13:43:38.491 INFO:teuthology.orchestra.run.vm00.stdout: Erasing : ceph-test-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 1/3 2026-03-06T13:43:38.494 INFO:teuthology.orchestra.run.vm00.stdout: Erasing : socat-1.7.4.1-8.el9.x86_64 2/3 2026-03-06T13:43:38.506 INFO:teuthology.orchestra.run.vm08.stdout: Preparing : 1/1 2026-03-06T13:43:38.509 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : xmlstarlet-1.6.1-20.el9.x86_64 3/3 2026-03-06T13:43:38.510 INFO:teuthology.orchestra.run.vm00.stdout: Erasing : xmlstarlet-1.6.1-20.el9.x86_64 3/3 2026-03-06T13:43:38.513 INFO:teuthology.orchestra.run.vm08.stdout: Erasing : ceph-test-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 1/3 2026-03-06T13:43:38.516 INFO:teuthology.orchestra.run.vm08.stdout: Erasing : socat-1.7.4.1-8.el9.x86_64 2/3 2026-03-06T13:43:38.535 INFO:teuthology.orchestra.run.vm08.stdout: Erasing : xmlstarlet-1.6.1-20.el9.x86_64 3/3 2026-03-06T13:43:38.594 INFO:teuthology.orchestra.run.vm00.stdout: Running scriptlet: xmlstarlet-1.6.1-20.el9.x86_64 3/3 2026-03-06T13:43:38.594 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : ceph-test-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 1/3 2026-03-06T13:43:38.594 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : socat-1.7.4.1-8.el9.x86_64 2/3 2026-03-06T13:43:38.603 INFO:teuthology.orchestra.run.vm01.stdout: Running scriptlet: xmlstarlet-1.6.1-20.el9.x86_64 3/3 2026-03-06T13:43:38.603 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : ceph-test-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 1/3 2026-03-06T13:43:38.603 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : socat-1.7.4.1-8.el9.x86_64 2/3 2026-03-06T13:43:38.611 INFO:teuthology.orchestra.run.vm08.stdout: Running scriptlet: xmlstarlet-1.6.1-20.el9.x86_64 3/3 2026-03-06T13:43:38.611 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : ceph-test-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 1/3 2026-03-06T13:43:38.611 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : socat-1.7.4.1-8.el9.x86_64 2/3 2026-03-06T13:43:38.654 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : xmlstarlet-1.6.1-20.el9.x86_64 3/3 2026-03-06T13:43:38.654 INFO:teuthology.orchestra.run.vm00.stdout: 2026-03-06T13:43:38.654 INFO:teuthology.orchestra.run.vm00.stdout:Removed: 2026-03-06T13:43:38.654 INFO:teuthology.orchestra.run.vm00.stdout: ceph-test-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 2026-03-06T13:43:38.654 INFO:teuthology.orchestra.run.vm00.stdout: socat-1.7.4.1-8.el9.x86_64 2026-03-06T13:43:38.654 INFO:teuthology.orchestra.run.vm00.stdout: xmlstarlet-1.6.1-20.el9.x86_64 2026-03-06T13:43:38.654 INFO:teuthology.orchestra.run.vm00.stdout: 2026-03-06T13:43:38.654 INFO:teuthology.orchestra.run.vm00.stdout:Complete! 2026-03-06T13:43:38.657 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : xmlstarlet-1.6.1-20.el9.x86_64 3/3 2026-03-06T13:43:38.657 INFO:teuthology.orchestra.run.vm01.stdout: 2026-03-06T13:43:38.657 INFO:teuthology.orchestra.run.vm01.stdout:Removed: 2026-03-06T13:43:38.657 INFO:teuthology.orchestra.run.vm01.stdout: ceph-test-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 2026-03-06T13:43:38.657 INFO:teuthology.orchestra.run.vm01.stdout: socat-1.7.4.1-8.el9.x86_64 2026-03-06T13:43:38.657 INFO:teuthology.orchestra.run.vm01.stdout: xmlstarlet-1.6.1-20.el9.x86_64 2026-03-06T13:43:38.657 INFO:teuthology.orchestra.run.vm01.stdout: 2026-03-06T13:43:38.657 INFO:teuthology.orchestra.run.vm01.stdout:Complete! 2026-03-06T13:43:38.672 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : xmlstarlet-1.6.1-20.el9.x86_64 3/3 2026-03-06T13:43:38.672 INFO:teuthology.orchestra.run.vm08.stdout: 2026-03-06T13:43:38.672 INFO:teuthology.orchestra.run.vm08.stdout:Removed: 2026-03-06T13:43:38.672 INFO:teuthology.orchestra.run.vm08.stdout: ceph-test-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 2026-03-06T13:43:38.672 INFO:teuthology.orchestra.run.vm08.stdout: socat-1.7.4.1-8.el9.x86_64 2026-03-06T13:43:38.672 INFO:teuthology.orchestra.run.vm08.stdout: xmlstarlet-1.6.1-20.el9.x86_64 2026-03-06T13:43:38.672 INFO:teuthology.orchestra.run.vm08.stdout: 2026-03-06T13:43:38.672 INFO:teuthology.orchestra.run.vm08.stdout:Complete! 2026-03-06T13:43:38.863 INFO:teuthology.orchestra.run.vm00.stdout:Dependencies resolved. 2026-03-06T13:43:38.863 INFO:teuthology.orchestra.run.vm00.stdout:================================================================================ 2026-03-06T13:43:38.864 INFO:teuthology.orchestra.run.vm00.stdout: Package Arch Version Repository Size 2026-03-06T13:43:38.864 INFO:teuthology.orchestra.run.vm00.stdout:================================================================================ 2026-03-06T13:43:38.864 INFO:teuthology.orchestra.run.vm00.stdout:Removing: 2026-03-06T13:43:38.864 INFO:teuthology.orchestra.run.vm00.stdout: ceph x86_64 2:19.2.3-47.gc24117fd552.el9.clyso @ceph 0 2026-03-06T13:43:38.864 INFO:teuthology.orchestra.run.vm00.stdout:Removing unused dependencies: 2026-03-06T13:43:38.864 INFO:teuthology.orchestra.run.vm00.stdout: ceph-mds x86_64 2:19.2.3-47.gc24117fd552.el9.clyso @ceph 7.4 M 2026-03-06T13:43:38.864 INFO:teuthology.orchestra.run.vm00.stdout: ceph-mon x86_64 2:19.2.3-47.gc24117fd552.el9.clyso @ceph 18 M 2026-03-06T13:43:38.864 INFO:teuthology.orchestra.run.vm00.stdout: lua x86_64 5.4.4-4.el9 @appstream 593 k 2026-03-06T13:43:38.864 INFO:teuthology.orchestra.run.vm00.stdout: lua-devel x86_64 5.4.4-4.el9 @crb 49 k 2026-03-06T13:43:38.864 INFO:teuthology.orchestra.run.vm00.stdout: luarocks noarch 3.9.2-5.el9 @epel 692 k 2026-03-06T13:43:38.864 INFO:teuthology.orchestra.run.vm00.stdout: unzip x86_64 6.0-59.el9 @baseos 389 k 2026-03-06T13:43:38.864 INFO:teuthology.orchestra.run.vm00.stdout: zip x86_64 3.0-35.el9 @baseos 724 k 2026-03-06T13:43:38.864 INFO:teuthology.orchestra.run.vm00.stdout: 2026-03-06T13:43:38.864 INFO:teuthology.orchestra.run.vm00.stdout:Transaction Summary 2026-03-06T13:43:38.864 INFO:teuthology.orchestra.run.vm00.stdout:================================================================================ 2026-03-06T13:43:38.864 INFO:teuthology.orchestra.run.vm00.stdout:Remove 8 Packages 2026-03-06T13:43:38.864 INFO:teuthology.orchestra.run.vm00.stdout: 2026-03-06T13:43:38.864 INFO:teuthology.orchestra.run.vm00.stdout:Freed space: 28 M 2026-03-06T13:43:38.864 INFO:teuthology.orchestra.run.vm00.stdout:Running transaction check 2026-03-06T13:43:38.867 INFO:teuthology.orchestra.run.vm00.stdout:Transaction check succeeded. 2026-03-06T13:43:38.867 INFO:teuthology.orchestra.run.vm00.stdout:Running transaction test 2026-03-06T13:43:38.882 INFO:teuthology.orchestra.run.vm01.stdout:Dependencies resolved. 2026-03-06T13:43:38.883 INFO:teuthology.orchestra.run.vm01.stdout:================================================================================ 2026-03-06T13:43:38.883 INFO:teuthology.orchestra.run.vm01.stdout: Package Arch Version Repository Size 2026-03-06T13:43:38.883 INFO:teuthology.orchestra.run.vm01.stdout:================================================================================ 2026-03-06T13:43:38.883 INFO:teuthology.orchestra.run.vm01.stdout:Removing: 2026-03-06T13:43:38.883 INFO:teuthology.orchestra.run.vm01.stdout: ceph x86_64 2:19.2.3-47.gc24117fd552.el9.clyso @ceph 0 2026-03-06T13:43:38.883 INFO:teuthology.orchestra.run.vm01.stdout:Removing unused dependencies: 2026-03-06T13:43:38.883 INFO:teuthology.orchestra.run.vm01.stdout: ceph-mds x86_64 2:19.2.3-47.gc24117fd552.el9.clyso @ceph 7.4 M 2026-03-06T13:43:38.883 INFO:teuthology.orchestra.run.vm01.stdout: ceph-mon x86_64 2:19.2.3-47.gc24117fd552.el9.clyso @ceph 18 M 2026-03-06T13:43:38.883 INFO:teuthology.orchestra.run.vm01.stdout: lua x86_64 5.4.4-4.el9 @appstream 593 k 2026-03-06T13:43:38.883 INFO:teuthology.orchestra.run.vm01.stdout: lua-devel x86_64 5.4.4-4.el9 @crb 49 k 2026-03-06T13:43:38.883 INFO:teuthology.orchestra.run.vm01.stdout: luarocks noarch 3.9.2-5.el9 @epel 692 k 2026-03-06T13:43:38.883 INFO:teuthology.orchestra.run.vm01.stdout: unzip x86_64 6.0-59.el9 @baseos 389 k 2026-03-06T13:43:38.883 INFO:teuthology.orchestra.run.vm01.stdout: zip x86_64 3.0-35.el9 @baseos 724 k 2026-03-06T13:43:38.883 INFO:teuthology.orchestra.run.vm01.stdout: 2026-03-06T13:43:38.883 INFO:teuthology.orchestra.run.vm01.stdout:Transaction Summary 2026-03-06T13:43:38.883 INFO:teuthology.orchestra.run.vm01.stdout:================================================================================ 2026-03-06T13:43:38.883 INFO:teuthology.orchestra.run.vm01.stdout:Remove 8 Packages 2026-03-06T13:43:38.883 INFO:teuthology.orchestra.run.vm01.stdout: 2026-03-06T13:43:38.883 INFO:teuthology.orchestra.run.vm01.stdout:Freed space: 28 M 2026-03-06T13:43:38.883 INFO:teuthology.orchestra.run.vm01.stdout:Running transaction check 2026-03-06T13:43:38.886 INFO:teuthology.orchestra.run.vm01.stdout:Transaction check succeeded. 2026-03-06T13:43:38.886 INFO:teuthology.orchestra.run.vm01.stdout:Running transaction test 2026-03-06T13:43:38.890 INFO:teuthology.orchestra.run.vm00.stdout:Transaction test succeeded. 2026-03-06T13:43:38.890 INFO:teuthology.orchestra.run.vm00.stdout:Running transaction 2026-03-06T13:43:38.901 INFO:teuthology.orchestra.run.vm08.stdout:Dependencies resolved. 2026-03-06T13:43:38.902 INFO:teuthology.orchestra.run.vm08.stdout:================================================================================ 2026-03-06T13:43:38.902 INFO:teuthology.orchestra.run.vm08.stdout: Package Arch Version Repository Size 2026-03-06T13:43:38.902 INFO:teuthology.orchestra.run.vm08.stdout:================================================================================ 2026-03-06T13:43:38.902 INFO:teuthology.orchestra.run.vm08.stdout:Removing: 2026-03-06T13:43:38.902 INFO:teuthology.orchestra.run.vm08.stdout: ceph x86_64 2:19.2.3-47.gc24117fd552.el9.clyso @ceph 0 2026-03-06T13:43:38.902 INFO:teuthology.orchestra.run.vm08.stdout:Removing unused dependencies: 2026-03-06T13:43:38.902 INFO:teuthology.orchestra.run.vm08.stdout: ceph-mds x86_64 2:19.2.3-47.gc24117fd552.el9.clyso @ceph 7.4 M 2026-03-06T13:43:38.902 INFO:teuthology.orchestra.run.vm08.stdout: ceph-mon x86_64 2:19.2.3-47.gc24117fd552.el9.clyso @ceph 18 M 2026-03-06T13:43:38.902 INFO:teuthology.orchestra.run.vm08.stdout: lua x86_64 5.4.4-4.el9 @appstream 593 k 2026-03-06T13:43:38.902 INFO:teuthology.orchestra.run.vm08.stdout: lua-devel x86_64 5.4.4-4.el9 @crb 49 k 2026-03-06T13:43:38.902 INFO:teuthology.orchestra.run.vm08.stdout: luarocks noarch 3.9.2-5.el9 @epel 692 k 2026-03-06T13:43:38.902 INFO:teuthology.orchestra.run.vm08.stdout: unzip x86_64 6.0-59.el9 @baseos 389 k 2026-03-06T13:43:38.902 INFO:teuthology.orchestra.run.vm08.stdout: zip x86_64 3.0-35.el9 @baseos 724 k 2026-03-06T13:43:38.902 INFO:teuthology.orchestra.run.vm08.stdout: 2026-03-06T13:43:38.902 INFO:teuthology.orchestra.run.vm08.stdout:Transaction Summary 2026-03-06T13:43:38.902 INFO:teuthology.orchestra.run.vm08.stdout:================================================================================ 2026-03-06T13:43:38.902 INFO:teuthology.orchestra.run.vm08.stdout:Remove 8 Packages 2026-03-06T13:43:38.902 INFO:teuthology.orchestra.run.vm08.stdout: 2026-03-06T13:43:38.903 INFO:teuthology.orchestra.run.vm08.stdout:Freed space: 28 M 2026-03-06T13:43:38.903 INFO:teuthology.orchestra.run.vm08.stdout:Running transaction check 2026-03-06T13:43:38.905 INFO:teuthology.orchestra.run.vm08.stdout:Transaction check succeeded. 2026-03-06T13:43:38.905 INFO:teuthology.orchestra.run.vm08.stdout:Running transaction test 2026-03-06T13:43:38.911 INFO:teuthology.orchestra.run.vm01.stdout:Transaction test succeeded. 2026-03-06T13:43:38.911 INFO:teuthology.orchestra.run.vm01.stdout:Running transaction 2026-03-06T13:43:38.928 INFO:teuthology.orchestra.run.vm08.stdout:Transaction test succeeded. 2026-03-06T13:43:38.929 INFO:teuthology.orchestra.run.vm08.stdout:Running transaction 2026-03-06T13:43:38.933 INFO:teuthology.orchestra.run.vm00.stdout: Preparing : 1/1 2026-03-06T13:43:38.938 INFO:teuthology.orchestra.run.vm00.stdout: Erasing : ceph-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 1/8 2026-03-06T13:43:38.941 INFO:teuthology.orchestra.run.vm00.stdout: Erasing : luarocks-3.9.2-5.el9.noarch 2/8 2026-03-06T13:43:38.943 INFO:teuthology.orchestra.run.vm00.stdout: Erasing : lua-devel-5.4.4-4.el9.x86_64 3/8 2026-03-06T13:43:38.945 INFO:teuthology.orchestra.run.vm00.stdout: Erasing : zip-3.0-35.el9.x86_64 4/8 2026-03-06T13:43:38.948 INFO:teuthology.orchestra.run.vm00.stdout: Erasing : unzip-6.0-59.el9.x86_64 5/8 2026-03-06T13:43:38.950 INFO:teuthology.orchestra.run.vm00.stdout: Erasing : lua-5.4.4-4.el9.x86_64 6/8 2026-03-06T13:43:38.955 INFO:teuthology.orchestra.run.vm01.stdout: Preparing : 1/1 2026-03-06T13:43:38.961 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : ceph-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 1/8 2026-03-06T13:43:38.967 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : luarocks-3.9.2-5.el9.noarch 2/8 2026-03-06T13:43:38.968 INFO:teuthology.orchestra.run.vm00.stdout: Running scriptlet: ceph-mds-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 7/8 2026-03-06T13:43:38.968 INFO:teuthology.orchestra.run.vm00.stdout:Glob pattern passed to enable, but globs are not supported for this. 2026-03-06T13:43:38.968 INFO:teuthology.orchestra.run.vm00.stdout:Invalid unit name "ceph-mds@*.service" escaped as "ceph-mds@\x2a.service". 2026-03-06T13:43:38.968 INFO:teuthology.orchestra.run.vm00.stdout:Removed "/etc/systemd/system/multi-user.target.wants/ceph-mds.target". 2026-03-06T13:43:38.968 INFO:teuthology.orchestra.run.vm00.stdout:Removed "/etc/systemd/system/ceph.target.wants/ceph-mds.target". 2026-03-06T13:43:38.968 INFO:teuthology.orchestra.run.vm00.stdout: 2026-03-06T13:43:38.968 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : lua-devel-5.4.4-4.el9.x86_64 3/8 2026-03-06T13:43:38.969 INFO:teuthology.orchestra.run.vm00.stdout: Erasing : ceph-mds-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 7/8 2026-03-06T13:43:38.972 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : zip-3.0-35.el9.x86_64 4/8 2026-03-06T13:43:38.972 INFO:teuthology.orchestra.run.vm08.stdout: Preparing : 1/1 2026-03-06T13:43:38.974 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : unzip-6.0-59.el9.x86_64 5/8 2026-03-06T13:43:38.975 INFO:teuthology.orchestra.run.vm00.stdout: Running scriptlet: ceph-mds-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 7/8 2026-03-06T13:43:38.976 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : lua-5.4.4-4.el9.x86_64 6/8 2026-03-06T13:43:38.978 INFO:teuthology.orchestra.run.vm08.stdout: Erasing : ceph-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 1/8 2026-03-06T13:43:38.982 INFO:teuthology.orchestra.run.vm08.stdout: Erasing : luarocks-3.9.2-5.el9.noarch 2/8 2026-03-06T13:43:38.984 INFO:teuthology.orchestra.run.vm08.stdout: Erasing : lua-devel-5.4.4-4.el9.x86_64 3/8 2026-03-06T13:43:38.988 INFO:teuthology.orchestra.run.vm08.stdout: Erasing : zip-3.0-35.el9.x86_64 4/8 2026-03-06T13:43:38.991 INFO:teuthology.orchestra.run.vm08.stdout: Erasing : unzip-6.0-59.el9.x86_64 5/8 2026-03-06T13:43:38.992 INFO:teuthology.orchestra.run.vm00.stdout: Running scriptlet: ceph-mon-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 8/8 2026-03-06T13:43:38.992 INFO:teuthology.orchestra.run.vm00.stdout:Glob pattern passed to enable, but globs are not supported for this. 2026-03-06T13:43:38.992 INFO:teuthology.orchestra.run.vm00.stdout:Invalid unit name "ceph-mon@*.service" escaped as "ceph-mon@\x2a.service". 2026-03-06T13:43:38.992 INFO:teuthology.orchestra.run.vm00.stdout:Removed "/etc/systemd/system/multi-user.target.wants/ceph-mon.target". 2026-03-06T13:43:38.992 INFO:teuthology.orchestra.run.vm00.stdout:Removed "/etc/systemd/system/ceph.target.wants/ceph-mon.target". 2026-03-06T13:43:38.992 INFO:teuthology.orchestra.run.vm00.stdout: 2026-03-06T13:43:38.993 INFO:teuthology.orchestra.run.vm08.stdout: Erasing : lua-5.4.4-4.el9.x86_64 6/8 2026-03-06T13:43:38.993 INFO:teuthology.orchestra.run.vm00.stdout: Erasing : ceph-mon-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 8/8 2026-03-06T13:43:39.000 INFO:teuthology.orchestra.run.vm01.stdout: Running scriptlet: ceph-mds-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 7/8 2026-03-06T13:43:39.000 INFO:teuthology.orchestra.run.vm01.stdout:Glob pattern passed to enable, but globs are not supported for this. 2026-03-06T13:43:39.000 INFO:teuthology.orchestra.run.vm01.stdout:Invalid unit name "ceph-mds@*.service" escaped as "ceph-mds@\x2a.service". 2026-03-06T13:43:39.000 INFO:teuthology.orchestra.run.vm01.stdout:Removed "/etc/systemd/system/multi-user.target.wants/ceph-mds.target". 2026-03-06T13:43:39.000 INFO:teuthology.orchestra.run.vm01.stdout:Removed "/etc/systemd/system/ceph.target.wants/ceph-mds.target". 2026-03-06T13:43:39.000 INFO:teuthology.orchestra.run.vm01.stdout: 2026-03-06T13:43:39.001 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : ceph-mds-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 7/8 2026-03-06T13:43:39.008 INFO:teuthology.orchestra.run.vm01.stdout: Running scriptlet: ceph-mds-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 7/8 2026-03-06T13:43:39.017 INFO:teuthology.orchestra.run.vm08.stdout: Running scriptlet: ceph-mds-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 7/8 2026-03-06T13:43:39.017 INFO:teuthology.orchestra.run.vm08.stdout:Glob pattern passed to enable, but globs are not supported for this. 2026-03-06T13:43:39.017 INFO:teuthology.orchestra.run.vm08.stdout:Invalid unit name "ceph-mds@*.service" escaped as "ceph-mds@\x2a.service". 2026-03-06T13:43:39.017 INFO:teuthology.orchestra.run.vm08.stdout:Removed "/etc/systemd/system/multi-user.target.wants/ceph-mds.target". 2026-03-06T13:43:39.017 INFO:teuthology.orchestra.run.vm08.stdout:Removed "/etc/systemd/system/ceph.target.wants/ceph-mds.target". 2026-03-06T13:43:39.017 INFO:teuthology.orchestra.run.vm08.stdout: 2026-03-06T13:43:39.019 INFO:teuthology.orchestra.run.vm08.stdout: Erasing : ceph-mds-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 7/8 2026-03-06T13:43:39.026 INFO:teuthology.orchestra.run.vm08.stdout: Running scriptlet: ceph-mds-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 7/8 2026-03-06T13:43:39.030 INFO:teuthology.orchestra.run.vm01.stdout: Running scriptlet: ceph-mon-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 8/8 2026-03-06T13:43:39.030 INFO:teuthology.orchestra.run.vm01.stdout:Glob pattern passed to enable, but globs are not supported for this. 2026-03-06T13:43:39.030 INFO:teuthology.orchestra.run.vm01.stdout:Invalid unit name "ceph-mon@*.service" escaped as "ceph-mon@\x2a.service". 2026-03-06T13:43:39.030 INFO:teuthology.orchestra.run.vm01.stdout:Removed "/etc/systemd/system/multi-user.target.wants/ceph-mon.target". 2026-03-06T13:43:39.031 INFO:teuthology.orchestra.run.vm01.stdout:Removed "/etc/systemd/system/ceph.target.wants/ceph-mon.target". 2026-03-06T13:43:39.031 INFO:teuthology.orchestra.run.vm01.stdout: 2026-03-06T13:43:39.033 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : ceph-mon-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 8/8 2026-03-06T13:43:39.044 INFO:teuthology.orchestra.run.vm08.stdout: Running scriptlet: ceph-mon-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 8/8 2026-03-06T13:43:39.044 INFO:teuthology.orchestra.run.vm08.stdout:Glob pattern passed to enable, but globs are not supported for this. 2026-03-06T13:43:39.044 INFO:teuthology.orchestra.run.vm08.stdout:Invalid unit name "ceph-mon@*.service" escaped as "ceph-mon@\x2a.service". 2026-03-06T13:43:39.044 INFO:teuthology.orchestra.run.vm08.stdout:Removed "/etc/systemd/system/multi-user.target.wants/ceph-mon.target". 2026-03-06T13:43:39.044 INFO:teuthology.orchestra.run.vm08.stdout:Removed "/etc/systemd/system/ceph.target.wants/ceph-mon.target". 2026-03-06T13:43:39.044 INFO:teuthology.orchestra.run.vm08.stdout: 2026-03-06T13:43:39.046 INFO:teuthology.orchestra.run.vm08.stdout: Erasing : ceph-mon-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 8/8 2026-03-06T13:43:39.071 INFO:teuthology.orchestra.run.vm00.stdout: Running scriptlet: ceph-mon-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 8/8 2026-03-06T13:43:39.072 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : ceph-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 1/8 2026-03-06T13:43:39.072 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : ceph-mds-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 2/8 2026-03-06T13:43:39.072 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : ceph-mon-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 3/8 2026-03-06T13:43:39.072 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : lua-5.4.4-4.el9.x86_64 4/8 2026-03-06T13:43:39.072 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : lua-devel-5.4.4-4.el9.x86_64 5/8 2026-03-06T13:43:39.072 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : luarocks-3.9.2-5.el9.noarch 6/8 2026-03-06T13:43:39.072 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : unzip-6.0-59.el9.x86_64 7/8 2026-03-06T13:43:39.122 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : zip-3.0-35.el9.x86_64 8/8 2026-03-06T13:43:39.122 INFO:teuthology.orchestra.run.vm00.stdout: 2026-03-06T13:43:39.122 INFO:teuthology.orchestra.run.vm00.stdout:Removed: 2026-03-06T13:43:39.122 INFO:teuthology.orchestra.run.vm00.stdout: ceph-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 2026-03-06T13:43:39.123 INFO:teuthology.orchestra.run.vm00.stdout: ceph-mds-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 2026-03-06T13:43:39.123 INFO:teuthology.orchestra.run.vm00.stdout: ceph-mon-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 2026-03-06T13:43:39.123 INFO:teuthology.orchestra.run.vm00.stdout: lua-5.4.4-4.el9.x86_64 2026-03-06T13:43:39.123 INFO:teuthology.orchestra.run.vm00.stdout: lua-devel-5.4.4-4.el9.x86_64 2026-03-06T13:43:39.123 INFO:teuthology.orchestra.run.vm00.stdout: luarocks-3.9.2-5.el9.noarch 2026-03-06T13:43:39.123 INFO:teuthology.orchestra.run.vm00.stdout: unzip-6.0-59.el9.x86_64 2026-03-06T13:43:39.123 INFO:teuthology.orchestra.run.vm00.stdout: zip-3.0-35.el9.x86_64 2026-03-06T13:43:39.123 INFO:teuthology.orchestra.run.vm00.stdout: 2026-03-06T13:43:39.123 INFO:teuthology.orchestra.run.vm00.stdout:Complete! 2026-03-06T13:43:39.123 INFO:teuthology.orchestra.run.vm01.stdout: Running scriptlet: ceph-mon-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 8/8 2026-03-06T13:43:39.124 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : ceph-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 1/8 2026-03-06T13:43:39.124 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : ceph-mds-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 2/8 2026-03-06T13:43:39.124 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : ceph-mon-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 3/8 2026-03-06T13:43:39.124 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : lua-5.4.4-4.el9.x86_64 4/8 2026-03-06T13:43:39.124 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : lua-devel-5.4.4-4.el9.x86_64 5/8 2026-03-06T13:43:39.124 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : luarocks-3.9.2-5.el9.noarch 6/8 2026-03-06T13:43:39.124 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : unzip-6.0-59.el9.x86_64 7/8 2026-03-06T13:43:39.147 INFO:teuthology.orchestra.run.vm08.stdout: Running scriptlet: ceph-mon-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 8/8 2026-03-06T13:43:39.147 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : ceph-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 1/8 2026-03-06T13:43:39.147 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : ceph-mds-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 2/8 2026-03-06T13:43:39.147 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : ceph-mon-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 3/8 2026-03-06T13:43:39.147 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : lua-5.4.4-4.el9.x86_64 4/8 2026-03-06T13:43:39.147 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : lua-devel-5.4.4-4.el9.x86_64 5/8 2026-03-06T13:43:39.147 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : luarocks-3.9.2-5.el9.noarch 6/8 2026-03-06T13:43:39.147 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : unzip-6.0-59.el9.x86_64 7/8 2026-03-06T13:43:39.183 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : zip-3.0-35.el9.x86_64 8/8 2026-03-06T13:43:39.183 INFO:teuthology.orchestra.run.vm01.stdout: 2026-03-06T13:43:39.183 INFO:teuthology.orchestra.run.vm01.stdout:Removed: 2026-03-06T13:43:39.183 INFO:teuthology.orchestra.run.vm01.stdout: ceph-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 2026-03-06T13:43:39.183 INFO:teuthology.orchestra.run.vm01.stdout: ceph-mds-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 2026-03-06T13:43:39.183 INFO:teuthology.orchestra.run.vm01.stdout: ceph-mon-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 2026-03-06T13:43:39.183 INFO:teuthology.orchestra.run.vm01.stdout: lua-5.4.4-4.el9.x86_64 2026-03-06T13:43:39.183 INFO:teuthology.orchestra.run.vm01.stdout: lua-devel-5.4.4-4.el9.x86_64 2026-03-06T13:43:39.183 INFO:teuthology.orchestra.run.vm01.stdout: luarocks-3.9.2-5.el9.noarch 2026-03-06T13:43:39.183 INFO:teuthology.orchestra.run.vm01.stdout: unzip-6.0-59.el9.x86_64 2026-03-06T13:43:39.183 INFO:teuthology.orchestra.run.vm01.stdout: zip-3.0-35.el9.x86_64 2026-03-06T13:43:39.183 INFO:teuthology.orchestra.run.vm01.stdout: 2026-03-06T13:43:39.183 INFO:teuthology.orchestra.run.vm01.stdout:Complete! 2026-03-06T13:43:39.209 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : zip-3.0-35.el9.x86_64 8/8 2026-03-06T13:43:39.209 INFO:teuthology.orchestra.run.vm08.stdout: 2026-03-06T13:43:39.209 INFO:teuthology.orchestra.run.vm08.stdout:Removed: 2026-03-06T13:43:39.209 INFO:teuthology.orchestra.run.vm08.stdout: ceph-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 2026-03-06T13:43:39.209 INFO:teuthology.orchestra.run.vm08.stdout: ceph-mds-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 2026-03-06T13:43:39.209 INFO:teuthology.orchestra.run.vm08.stdout: ceph-mon-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 2026-03-06T13:43:39.209 INFO:teuthology.orchestra.run.vm08.stdout: lua-5.4.4-4.el9.x86_64 2026-03-06T13:43:39.209 INFO:teuthology.orchestra.run.vm08.stdout: lua-devel-5.4.4-4.el9.x86_64 2026-03-06T13:43:39.209 INFO:teuthology.orchestra.run.vm08.stdout: luarocks-3.9.2-5.el9.noarch 2026-03-06T13:43:39.209 INFO:teuthology.orchestra.run.vm08.stdout: unzip-6.0-59.el9.x86_64 2026-03-06T13:43:39.209 INFO:teuthology.orchestra.run.vm08.stdout: zip-3.0-35.el9.x86_64 2026-03-06T13:43:39.209 INFO:teuthology.orchestra.run.vm08.stdout: 2026-03-06T13:43:39.209 INFO:teuthology.orchestra.run.vm08.stdout:Complete! 2026-03-06T13:43:39.338 INFO:teuthology.orchestra.run.vm00.stdout:Dependencies resolved. 2026-03-06T13:43:39.345 INFO:teuthology.orchestra.run.vm00.stdout:=================================================================================================== 2026-03-06T13:43:39.345 INFO:teuthology.orchestra.run.vm00.stdout: Package Arch Version Repository Size 2026-03-06T13:43:39.345 INFO:teuthology.orchestra.run.vm00.stdout:=================================================================================================== 2026-03-06T13:43:39.345 INFO:teuthology.orchestra.run.vm00.stdout:Removing: 2026-03-06T13:43:39.345 INFO:teuthology.orchestra.run.vm00.stdout: ceph-base x86_64 2:19.2.3-47.gc24117fd552.el9.clyso @ceph 23 M 2026-03-06T13:43:39.345 INFO:teuthology.orchestra.run.vm00.stdout:Removing dependent packages: 2026-03-06T13:43:39.345 INFO:teuthology.orchestra.run.vm00.stdout: ceph-immutable-object-cache x86_64 2:19.2.3-47.gc24117fd552.el9.clyso @ceph 431 k 2026-03-06T13:43:39.345 INFO:teuthology.orchestra.run.vm00.stdout: ceph-mgr x86_64 2:19.2.3-47.gc24117fd552.el9.clyso @ceph 3.4 M 2026-03-06T13:43:39.345 INFO:teuthology.orchestra.run.vm00.stdout: ceph-mgr-cephadm noarch 2:19.2.3-47.gc24117fd552.el9.clyso @ceph-noarch 803 k 2026-03-06T13:43:39.345 INFO:teuthology.orchestra.run.vm00.stdout: ceph-mgr-dashboard noarch 2:19.2.3-47.gc24117fd552.el9.clyso @ceph-noarch 88 M 2026-03-06T13:43:39.345 INFO:teuthology.orchestra.run.vm00.stdout: ceph-mgr-diskprediction-local noarch 2:19.2.3-47.gc24117fd552.el9.clyso @ceph-noarch 66 M 2026-03-06T13:43:39.345 INFO:teuthology.orchestra.run.vm00.stdout: ceph-mgr-rook noarch 2:19.2.3-47.gc24117fd552.el9.clyso @ceph-noarch 563 k 2026-03-06T13:43:39.345 INFO:teuthology.orchestra.run.vm00.stdout: ceph-osd x86_64 2:19.2.3-47.gc24117fd552.el9.clyso @ceph 59 M 2026-03-06T13:43:39.345 INFO:teuthology.orchestra.run.vm00.stdout: ceph-volume noarch 2:19.2.3-47.gc24117fd552.el9.clyso @ceph-noarch 1.4 M 2026-03-06T13:43:39.345 INFO:teuthology.orchestra.run.vm00.stdout: rbd-mirror x86_64 2:19.2.3-47.gc24117fd552.el9.clyso @ceph 13 M 2026-03-06T13:43:39.345 INFO:teuthology.orchestra.run.vm00.stdout:Removing unused dependencies: 2026-03-06T13:43:39.346 INFO:teuthology.orchestra.run.vm00.stdout: abseil-cpp x86_64 20211102.0-4.el9 @epel 1.9 M 2026-03-06T13:43:39.346 INFO:teuthology.orchestra.run.vm00.stdout: ceph-common x86_64 2:19.2.3-47.gc24117fd552.el9.clyso @ceph 85 M 2026-03-06T13:43:39.346 INFO:teuthology.orchestra.run.vm00.stdout: ceph-grafana-dashboards noarch 2:19.2.3-47.gc24117fd552.el9.clyso @ceph-noarch 626 k 2026-03-06T13:43:39.346 INFO:teuthology.orchestra.run.vm00.stdout: ceph-mgr-k8sevents noarch 2:19.2.3-47.gc24117fd552.el9.clyso @ceph-noarch 60 k 2026-03-06T13:43:39.346 INFO:teuthology.orchestra.run.vm00.stdout: ceph-mgr-modules-core noarch 2:19.2.3-47.gc24117fd552.el9.clyso @ceph-noarch 1.5 M 2026-03-06T13:43:39.346 INFO:teuthology.orchestra.run.vm00.stdout: ceph-prometheus-alerts noarch 2:19.2.3-47.gc24117fd552.el9.clyso @ceph-noarch 51 k 2026-03-06T13:43:39.346 INFO:teuthology.orchestra.run.vm00.stdout: ceph-selinux x86_64 2:19.2.3-47.gc24117fd552.el9.clyso @ceph 138 k 2026-03-06T13:43:39.346 INFO:teuthology.orchestra.run.vm00.stdout: cryptsetup x86_64 2.8.1-3.el9 @baseos 770 k 2026-03-06T13:43:39.346 INFO:teuthology.orchestra.run.vm00.stdout: flexiblas x86_64 3.0.4-9.el9 @appstream 68 k 2026-03-06T13:43:39.346 INFO:teuthology.orchestra.run.vm00.stdout: flexiblas-netlib x86_64 3.0.4-9.el9 @appstream 11 M 2026-03-06T13:43:39.346 INFO:teuthology.orchestra.run.vm00.stdout: flexiblas-openblas-openmp x86_64 3.0.4-9.el9 @appstream 39 k 2026-03-06T13:43:39.346 INFO:teuthology.orchestra.run.vm00.stdout: gperftools-libs x86_64 2.9.1-3.el9 @epel 1.4 M 2026-03-06T13:43:39.346 INFO:teuthology.orchestra.run.vm00.stdout: grpc-data noarch 1.46.7-10.el9 @epel 13 k 2026-03-06T13:43:39.346 INFO:teuthology.orchestra.run.vm00.stdout: ledmon-libs x86_64 1.1.0-3.el9 @baseos 80 k 2026-03-06T13:43:39.346 INFO:teuthology.orchestra.run.vm00.stdout: libcephsqlite x86_64 2:19.2.3-47.gc24117fd552.el9.clyso @ceph 425 k 2026-03-06T13:43:39.346 INFO:teuthology.orchestra.run.vm00.stdout: libconfig x86_64 1.7.2-9.el9 @baseos 220 k 2026-03-06T13:43:39.346 INFO:teuthology.orchestra.run.vm00.stdout: libgfortran x86_64 11.5.0-14.el9 @baseos 2.8 M 2026-03-06T13:43:39.346 INFO:teuthology.orchestra.run.vm00.stdout: liboath x86_64 2.6.12-1.el9 @epel 94 k 2026-03-06T13:43:39.346 INFO:teuthology.orchestra.run.vm00.stdout: libquadmath x86_64 11.5.0-14.el9 @baseos 330 k 2026-03-06T13:43:39.346 INFO:teuthology.orchestra.run.vm00.stdout: libradosstriper1 x86_64 2:19.2.3-47.gc24117fd552.el9.clyso @ceph 1.6 M 2026-03-06T13:43:39.346 INFO:teuthology.orchestra.run.vm00.stdout: libstoragemgmt x86_64 1.10.1-1.el9 @appstream 685 k 2026-03-06T13:43:39.346 INFO:teuthology.orchestra.run.vm00.stdout: libunwind x86_64 1.6.2-1.el9 @epel 170 k 2026-03-06T13:43:39.346 INFO:teuthology.orchestra.run.vm00.stdout: libxslt x86_64 1.1.34-12.el9 @appstream 743 k 2026-03-06T13:43:39.346 INFO:teuthology.orchestra.run.vm00.stdout: openblas x86_64 0.3.29-1.el9 @appstream 112 k 2026-03-06T13:43:39.346 INFO:teuthology.orchestra.run.vm00.stdout: openblas-openmp x86_64 0.3.29-1.el9 @appstream 46 M 2026-03-06T13:43:39.346 INFO:teuthology.orchestra.run.vm00.stdout: pciutils x86_64 3.7.0-7.el9 @baseos 216 k 2026-03-06T13:43:39.346 INFO:teuthology.orchestra.run.vm00.stdout: protobuf x86_64 3.14.0-17.el9 @appstream 3.5 M 2026-03-06T13:43:39.346 INFO:teuthology.orchestra.run.vm00.stdout: protobuf-compiler x86_64 3.14.0-17.el9 @crb 2.9 M 2026-03-06T13:43:39.346 INFO:teuthology.orchestra.run.vm00.stdout: python3-asyncssh noarch 2.13.2-5.el9 @epel 3.9 M 2026-03-06T13:43:39.346 INFO:teuthology.orchestra.run.vm00.stdout: python3-autocommand noarch 2.2.2-8.el9 @epel 82 k 2026-03-06T13:43:39.346 INFO:teuthology.orchestra.run.vm00.stdout: python3-babel noarch 2.9.1-2.el9 @appstream 27 M 2026-03-06T13:43:39.346 INFO:teuthology.orchestra.run.vm00.stdout: python3-backports-tarfile noarch 1.2.0-1.el9 @epel 254 k 2026-03-06T13:43:39.346 INFO:teuthology.orchestra.run.vm00.stdout: python3-bcrypt x86_64 3.2.2-1.el9 @epel 87 k 2026-03-06T13:43:39.346 INFO:teuthology.orchestra.run.vm00.stdout: python3-cachetools noarch 4.2.4-1.el9 @epel 93 k 2026-03-06T13:43:39.346 INFO:teuthology.orchestra.run.vm00.stdout: python3-ceph-common x86_64 2:19.2.3-47.gc24117fd552.el9.clyso @ceph 702 k 2026-03-06T13:43:39.346 INFO:teuthology.orchestra.run.vm00.stdout: python3-certifi noarch 2023.05.07-4.el9 @epel 6.3 k 2026-03-06T13:43:39.346 INFO:teuthology.orchestra.run.vm00.stdout: python3-cffi x86_64 1.14.5-5.el9 @baseos 1.0 M 2026-03-06T13:43:39.346 INFO:teuthology.orchestra.run.vm00.stdout: python3-chardet noarch 4.0.0-5.el9 @anaconda 1.4 M 2026-03-06T13:43:39.346 INFO:teuthology.orchestra.run.vm00.stdout: python3-cheroot noarch 10.0.1-4.el9 @epel 682 k 2026-03-06T13:43:39.346 INFO:teuthology.orchestra.run.vm00.stdout: python3-cherrypy noarch 18.6.1-2.el9 @epel 1.1 M 2026-03-06T13:43:39.346 INFO:teuthology.orchestra.run.vm00.stdout: python3-cryptography x86_64 36.0.1-5.el9 @baseos 4.5 M 2026-03-06T13:43:39.346 INFO:teuthology.orchestra.run.vm00.stdout: python3-devel x86_64 3.9.25-3.el9 @appstream 765 k 2026-03-06T13:43:39.346 INFO:teuthology.orchestra.run.vm00.stdout: python3-google-auth noarch 1:2.45.0-1.el9 @epel 1.4 M 2026-03-06T13:43:39.346 INFO:teuthology.orchestra.run.vm00.stdout: python3-grpcio x86_64 1.46.7-10.el9 @epel 6.7 M 2026-03-06T13:43:39.346 INFO:teuthology.orchestra.run.vm00.stdout: python3-grpcio-tools x86_64 1.46.7-10.el9 @epel 418 k 2026-03-06T13:43:39.346 INFO:teuthology.orchestra.run.vm00.stdout: python3-idna noarch 2.10-7.el9.1 @anaconda 513 k 2026-03-06T13:43:39.346 INFO:teuthology.orchestra.run.vm00.stdout: python3-influxdb noarch 5.3.1-1.el9 @epel 747 k 2026-03-06T13:43:39.346 INFO:teuthology.orchestra.run.vm00.stdout: python3-isodate noarch 0.6.1-3.el9 @epel 203 k 2026-03-06T13:43:39.346 INFO:teuthology.orchestra.run.vm00.stdout: python3-jaraco noarch 8.2.1-3.el9 @epel 3.7 k 2026-03-06T13:43:39.346 INFO:teuthology.orchestra.run.vm00.stdout: python3-jaraco-classes noarch 3.2.1-5.el9 @epel 24 k 2026-03-06T13:43:39.346 INFO:teuthology.orchestra.run.vm00.stdout: python3-jaraco-collections noarch 3.0.0-8.el9 @epel 55 k 2026-03-06T13:43:39.346 INFO:teuthology.orchestra.run.vm00.stdout: python3-jaraco-context noarch 6.0.1-3.el9 @epel 31 k 2026-03-06T13:43:39.346 INFO:teuthology.orchestra.run.vm00.stdout: python3-jaraco-functools noarch 3.5.0-2.el9 @epel 33 k 2026-03-06T13:43:39.346 INFO:teuthology.orchestra.run.vm00.stdout: python3-jaraco-text noarch 4.0.0-2.el9 @epel 51 k 2026-03-06T13:43:39.347 INFO:teuthology.orchestra.run.vm00.stdout: python3-jinja2 noarch 2.11.3-8.el9 @appstream 1.1 M 2026-03-06T13:43:39.347 INFO:teuthology.orchestra.run.vm00.stdout: python3-jsonpatch noarch 1.21-16.el9 @koji-override-0 55 k 2026-03-06T13:43:39.347 INFO:teuthology.orchestra.run.vm00.stdout: python3-jsonpointer noarch 2.0-4.el9 @koji-override-0 34 k 2026-03-06T13:43:39.347 INFO:teuthology.orchestra.run.vm00.stdout: python3-kubernetes noarch 1:26.1.0-3.el9 @epel 21 M 2026-03-06T13:43:39.347 INFO:teuthology.orchestra.run.vm00.stdout: python3-libstoragemgmt x86_64 1.10.1-1.el9 @appstream 832 k 2026-03-06T13:43:39.347 INFO:teuthology.orchestra.run.vm00.stdout: python3-logutils noarch 0.3.5-21.el9 @epel 126 k 2026-03-06T13:43:39.347 INFO:teuthology.orchestra.run.vm00.stdout: python3-lxml x86_64 4.6.5-3.el9 @appstream 4.2 M 2026-03-06T13:43:39.347 INFO:teuthology.orchestra.run.vm00.stdout: python3-mako noarch 1.1.4-6.el9 @appstream 534 k 2026-03-06T13:43:39.347 INFO:teuthology.orchestra.run.vm00.stdout: python3-markupsafe x86_64 1.1.1-12.el9 @appstream 60 k 2026-03-06T13:43:39.347 INFO:teuthology.orchestra.run.vm00.stdout: python3-more-itertools noarch 8.12.0-2.el9 @epel 378 k 2026-03-06T13:43:39.347 INFO:teuthology.orchestra.run.vm00.stdout: python3-msgpack x86_64 1.0.3-2.el9 @epel 264 k 2026-03-06T13:43:39.347 INFO:teuthology.orchestra.run.vm00.stdout: python3-natsort noarch 7.1.1-5.el9 @epel 215 k 2026-03-06T13:43:39.347 INFO:teuthology.orchestra.run.vm00.stdout: python3-numpy x86_64 1:1.23.5-2.el9 @appstream 30 M 2026-03-06T13:43:39.347 INFO:teuthology.orchestra.run.vm00.stdout: python3-numpy-f2py x86_64 1:1.23.5-2.el9 @appstream 1.7 M 2026-03-06T13:43:39.347 INFO:teuthology.orchestra.run.vm00.stdout: python3-oauthlib noarch 3.1.1-5.el9 @koji-override-0 888 k 2026-03-06T13:43:39.347 INFO:teuthology.orchestra.run.vm00.stdout: python3-packaging noarch 20.9-5.el9 @appstream 248 k 2026-03-06T13:43:39.347 INFO:teuthology.orchestra.run.vm00.stdout: python3-pecan noarch 1.4.2-3.el9 @epel 1.3 M 2026-03-06T13:43:39.347 INFO:teuthology.orchestra.run.vm00.stdout: python3-ply noarch 3.11-14.el9 @baseos 430 k 2026-03-06T13:43:39.347 INFO:teuthology.orchestra.run.vm00.stdout: python3-portend noarch 3.1.0-2.el9 @epel 20 k 2026-03-06T13:43:39.347 INFO:teuthology.orchestra.run.vm00.stdout: python3-prettytable noarch 0.7.2-27.el9 @koji-override-0 166 k 2026-03-06T13:43:39.347 INFO:teuthology.orchestra.run.vm00.stdout: python3-protobuf noarch 3.14.0-17.el9 @appstream 1.4 M 2026-03-06T13:43:39.347 INFO:teuthology.orchestra.run.vm00.stdout: python3-pyOpenSSL noarch 21.0.0-1.el9 @epel 389 k 2026-03-06T13:43:39.347 INFO:teuthology.orchestra.run.vm00.stdout: python3-pyasn1 noarch 0.4.8-7.el9 @appstream 622 k 2026-03-06T13:43:39.347 INFO:teuthology.orchestra.run.vm00.stdout: python3-pyasn1-modules noarch 0.4.8-7.el9 @appstream 1.0 M 2026-03-06T13:43:39.347 INFO:teuthology.orchestra.run.vm00.stdout: python3-pycparser noarch 2.20-6.el9 @baseos 745 k 2026-03-06T13:43:39.347 INFO:teuthology.orchestra.run.vm00.stdout: python3-pysocks noarch 1.7.1-12.el9 @anaconda 88 k 2026-03-06T13:43:39.347 INFO:teuthology.orchestra.run.vm00.stdout: python3-pytz noarch 2021.1-5.el9 @koji-override-0 176 k 2026-03-06T13:43:39.347 INFO:teuthology.orchestra.run.vm00.stdout: python3-repoze-lru noarch 0.7-16.el9 @epel 83 k 2026-03-06T13:43:39.347 INFO:teuthology.orchestra.run.vm00.stdout: python3-requests noarch 2.25.1-10.el9 @baseos 405 k 2026-03-06T13:43:39.347 INFO:teuthology.orchestra.run.vm00.stdout: python3-requests-oauthlib noarch 1.3.0-12.el9 @appstream 119 k 2026-03-06T13:43:39.347 INFO:teuthology.orchestra.run.vm00.stdout: python3-routes noarch 2.5.1-5.el9 @epel 459 k 2026-03-06T13:43:39.347 INFO:teuthology.orchestra.run.vm00.stdout: python3-rsa noarch 4.9-2.el9 @epel 202 k 2026-03-06T13:43:39.347 INFO:teuthology.orchestra.run.vm00.stdout: python3-saml noarch 1.16.0-1.el9 @epel 730 k 2026-03-06T13:43:39.347 INFO:teuthology.orchestra.run.vm00.stdout: python3-scipy x86_64 1.9.3-2.el9 @appstream 76 M 2026-03-06T13:43:39.347 INFO:teuthology.orchestra.run.vm00.stdout: python3-tempora noarch 5.0.0-2.el9 @epel 96 k 2026-03-06T13:43:39.347 INFO:teuthology.orchestra.run.vm00.stdout: python3-toml noarch 0.10.2-6.el9 @appstream 99 k 2026-03-06T13:43:39.347 INFO:teuthology.orchestra.run.vm00.stdout: python3-typing-extensions noarch 4.15.0-1.el9 @epel 447 k 2026-03-06T13:43:39.347 INFO:teuthology.orchestra.run.vm00.stdout: python3-urllib3 noarch 1.26.5-7.el9 @baseos 746 k 2026-03-06T13:43:39.347 INFO:teuthology.orchestra.run.vm00.stdout: python3-webob noarch 1.8.8-2.el9 @epel 1.2 M 2026-03-06T13:43:39.347 INFO:teuthology.orchestra.run.vm00.stdout: python3-websocket-client noarch 1.2.3-2.el9 @epel 319 k 2026-03-06T13:43:39.347 INFO:teuthology.orchestra.run.vm00.stdout: python3-werkzeug noarch 2.0.3-3.el9.1 @epel 1.9 M 2026-03-06T13:43:39.347 INFO:teuthology.orchestra.run.vm00.stdout: python3-xmlsec x86_64 1.3.13-1.el9 @epel 158 k 2026-03-06T13:43:39.347 INFO:teuthology.orchestra.run.vm00.stdout: python3-zc-lockfile noarch 2.0-10.el9 @epel 35 k 2026-03-06T13:43:39.347 INFO:teuthology.orchestra.run.vm00.stdout: qatlib x86_64 25.08.0-2.el9 @appstream 639 k 2026-03-06T13:43:39.347 INFO:teuthology.orchestra.run.vm00.stdout: qatlib-service x86_64 25.08.0-2.el9 @appstream 69 k 2026-03-06T13:43:39.347 INFO:teuthology.orchestra.run.vm00.stdout: qatzip-libs x86_64 1.3.1-1.el9 @appstream 148 k 2026-03-06T13:43:39.347 INFO:teuthology.orchestra.run.vm00.stdout: smartmontools x86_64 1:7.2-10.el9 @baseos 1.9 M 2026-03-06T13:43:39.347 INFO:teuthology.orchestra.run.vm00.stdout: xmlsec1 x86_64 1.2.29-13.el9 @appstream 596 k 2026-03-06T13:43:39.347 INFO:teuthology.orchestra.run.vm00.stdout: xmlsec1-openssl x86_64 1.2.29-13.el9 @appstream 281 k 2026-03-06T13:43:39.347 INFO:teuthology.orchestra.run.vm00.stdout: 2026-03-06T13:43:39.347 INFO:teuthology.orchestra.run.vm00.stdout:Transaction Summary 2026-03-06T13:43:39.347 INFO:teuthology.orchestra.run.vm00.stdout:=================================================================================================== 2026-03-06T13:43:39.347 INFO:teuthology.orchestra.run.vm00.stdout:Remove 113 Packages 2026-03-06T13:43:39.347 INFO:teuthology.orchestra.run.vm00.stdout: 2026-03-06T13:43:39.348 INFO:teuthology.orchestra.run.vm00.stdout:Freed space: 623 M 2026-03-06T13:43:39.348 INFO:teuthology.orchestra.run.vm00.stdout:Running transaction check 2026-03-06T13:43:39.378 INFO:teuthology.orchestra.run.vm00.stdout:Transaction check succeeded. 2026-03-06T13:43:39.378 INFO:teuthology.orchestra.run.vm00.stdout:Running transaction test 2026-03-06T13:43:39.420 INFO:teuthology.orchestra.run.vm01.stdout:Dependencies resolved. 2026-03-06T13:43:39.426 INFO:teuthology.orchestra.run.vm01.stdout:=================================================================================================== 2026-03-06T13:43:39.427 INFO:teuthology.orchestra.run.vm01.stdout: Package Arch Version Repository Size 2026-03-06T13:43:39.427 INFO:teuthology.orchestra.run.vm01.stdout:=================================================================================================== 2026-03-06T13:43:39.427 INFO:teuthology.orchestra.run.vm01.stdout:Removing: 2026-03-06T13:43:39.427 INFO:teuthology.orchestra.run.vm01.stdout: ceph-base x86_64 2:19.2.3-47.gc24117fd552.el9.clyso @ceph 23 M 2026-03-06T13:43:39.427 INFO:teuthology.orchestra.run.vm01.stdout:Removing dependent packages: 2026-03-06T13:43:39.427 INFO:teuthology.orchestra.run.vm01.stdout: ceph-immutable-object-cache x86_64 2:19.2.3-47.gc24117fd552.el9.clyso @ceph 431 k 2026-03-06T13:43:39.427 INFO:teuthology.orchestra.run.vm01.stdout: ceph-mgr x86_64 2:19.2.3-47.gc24117fd552.el9.clyso @ceph 3.4 M 2026-03-06T13:43:39.427 INFO:teuthology.orchestra.run.vm01.stdout: ceph-mgr-cephadm noarch 2:19.2.3-47.gc24117fd552.el9.clyso @ceph-noarch 803 k 2026-03-06T13:43:39.427 INFO:teuthology.orchestra.run.vm01.stdout: ceph-mgr-dashboard noarch 2:19.2.3-47.gc24117fd552.el9.clyso @ceph-noarch 88 M 2026-03-06T13:43:39.427 INFO:teuthology.orchestra.run.vm01.stdout: ceph-mgr-diskprediction-local noarch 2:19.2.3-47.gc24117fd552.el9.clyso @ceph-noarch 66 M 2026-03-06T13:43:39.427 INFO:teuthology.orchestra.run.vm01.stdout: ceph-mgr-rook noarch 2:19.2.3-47.gc24117fd552.el9.clyso @ceph-noarch 563 k 2026-03-06T13:43:39.427 INFO:teuthology.orchestra.run.vm01.stdout: ceph-osd x86_64 2:19.2.3-47.gc24117fd552.el9.clyso @ceph 59 M 2026-03-06T13:43:39.427 INFO:teuthology.orchestra.run.vm01.stdout: ceph-volume noarch 2:19.2.3-47.gc24117fd552.el9.clyso @ceph-noarch 1.4 M 2026-03-06T13:43:39.427 INFO:teuthology.orchestra.run.vm01.stdout: rbd-mirror x86_64 2:19.2.3-47.gc24117fd552.el9.clyso @ceph 13 M 2026-03-06T13:43:39.427 INFO:teuthology.orchestra.run.vm01.stdout:Removing unused dependencies: 2026-03-06T13:43:39.427 INFO:teuthology.orchestra.run.vm01.stdout: abseil-cpp x86_64 20211102.0-4.el9 @epel 1.9 M 2026-03-06T13:43:39.427 INFO:teuthology.orchestra.run.vm01.stdout: ceph-common x86_64 2:19.2.3-47.gc24117fd552.el9.clyso @ceph 85 M 2026-03-06T13:43:39.427 INFO:teuthology.orchestra.run.vm01.stdout: ceph-grafana-dashboards noarch 2:19.2.3-47.gc24117fd552.el9.clyso @ceph-noarch 626 k 2026-03-06T13:43:39.427 INFO:teuthology.orchestra.run.vm01.stdout: ceph-mgr-k8sevents noarch 2:19.2.3-47.gc24117fd552.el9.clyso @ceph-noarch 60 k 2026-03-06T13:43:39.427 INFO:teuthology.orchestra.run.vm01.stdout: ceph-mgr-modules-core noarch 2:19.2.3-47.gc24117fd552.el9.clyso @ceph-noarch 1.5 M 2026-03-06T13:43:39.427 INFO:teuthology.orchestra.run.vm01.stdout: ceph-prometheus-alerts noarch 2:19.2.3-47.gc24117fd552.el9.clyso @ceph-noarch 51 k 2026-03-06T13:43:39.427 INFO:teuthology.orchestra.run.vm01.stdout: ceph-selinux x86_64 2:19.2.3-47.gc24117fd552.el9.clyso @ceph 138 k 2026-03-06T13:43:39.427 INFO:teuthology.orchestra.run.vm01.stdout: cryptsetup x86_64 2.8.1-3.el9 @baseos 770 k 2026-03-06T13:43:39.427 INFO:teuthology.orchestra.run.vm01.stdout: flexiblas x86_64 3.0.4-9.el9 @appstream 68 k 2026-03-06T13:43:39.427 INFO:teuthology.orchestra.run.vm01.stdout: flexiblas-netlib x86_64 3.0.4-9.el9 @appstream 11 M 2026-03-06T13:43:39.427 INFO:teuthology.orchestra.run.vm01.stdout: flexiblas-openblas-openmp x86_64 3.0.4-9.el9 @appstream 39 k 2026-03-06T13:43:39.427 INFO:teuthology.orchestra.run.vm01.stdout: gperftools-libs x86_64 2.9.1-3.el9 @epel 1.4 M 2026-03-06T13:43:39.427 INFO:teuthology.orchestra.run.vm01.stdout: grpc-data noarch 1.46.7-10.el9 @epel 13 k 2026-03-06T13:43:39.427 INFO:teuthology.orchestra.run.vm01.stdout: ledmon-libs x86_64 1.1.0-3.el9 @baseos 80 k 2026-03-06T13:43:39.427 INFO:teuthology.orchestra.run.vm01.stdout: libcephsqlite x86_64 2:19.2.3-47.gc24117fd552.el9.clyso @ceph 425 k 2026-03-06T13:43:39.427 INFO:teuthology.orchestra.run.vm01.stdout: libconfig x86_64 1.7.2-9.el9 @baseos 220 k 2026-03-06T13:43:39.427 INFO:teuthology.orchestra.run.vm01.stdout: libgfortran x86_64 11.5.0-14.el9 @baseos 2.8 M 2026-03-06T13:43:39.427 INFO:teuthology.orchestra.run.vm01.stdout: liboath x86_64 2.6.12-1.el9 @epel 94 k 2026-03-06T13:43:39.427 INFO:teuthology.orchestra.run.vm01.stdout: libquadmath x86_64 11.5.0-14.el9 @baseos 330 k 2026-03-06T13:43:39.427 INFO:teuthology.orchestra.run.vm01.stdout: libradosstriper1 x86_64 2:19.2.3-47.gc24117fd552.el9.clyso @ceph 1.6 M 2026-03-06T13:43:39.428 INFO:teuthology.orchestra.run.vm01.stdout: libstoragemgmt x86_64 1.10.1-1.el9 @appstream 685 k 2026-03-06T13:43:39.428 INFO:teuthology.orchestra.run.vm01.stdout: libunwind x86_64 1.6.2-1.el9 @epel 170 k 2026-03-06T13:43:39.428 INFO:teuthology.orchestra.run.vm01.stdout: libxslt x86_64 1.1.34-12.el9 @appstream 743 k 2026-03-06T13:43:39.428 INFO:teuthology.orchestra.run.vm01.stdout: openblas x86_64 0.3.29-1.el9 @appstream 112 k 2026-03-06T13:43:39.428 INFO:teuthology.orchestra.run.vm01.stdout: openblas-openmp x86_64 0.3.29-1.el9 @appstream 46 M 2026-03-06T13:43:39.428 INFO:teuthology.orchestra.run.vm01.stdout: pciutils x86_64 3.7.0-7.el9 @baseos 216 k 2026-03-06T13:43:39.428 INFO:teuthology.orchestra.run.vm01.stdout: protobuf x86_64 3.14.0-17.el9 @appstream 3.5 M 2026-03-06T13:43:39.428 INFO:teuthology.orchestra.run.vm01.stdout: protobuf-compiler x86_64 3.14.0-17.el9 @crb 2.9 M 2026-03-06T13:43:39.428 INFO:teuthology.orchestra.run.vm01.stdout: python3-asyncssh noarch 2.13.2-5.el9 @epel 3.9 M 2026-03-06T13:43:39.428 INFO:teuthology.orchestra.run.vm01.stdout: python3-autocommand noarch 2.2.2-8.el9 @epel 82 k 2026-03-06T13:43:39.428 INFO:teuthology.orchestra.run.vm01.stdout: python3-babel noarch 2.9.1-2.el9 @appstream 27 M 2026-03-06T13:43:39.428 INFO:teuthology.orchestra.run.vm01.stdout: python3-backports-tarfile noarch 1.2.0-1.el9 @epel 254 k 2026-03-06T13:43:39.428 INFO:teuthology.orchestra.run.vm01.stdout: python3-bcrypt x86_64 3.2.2-1.el9 @epel 87 k 2026-03-06T13:43:39.428 INFO:teuthology.orchestra.run.vm01.stdout: python3-cachetools noarch 4.2.4-1.el9 @epel 93 k 2026-03-06T13:43:39.428 INFO:teuthology.orchestra.run.vm01.stdout: python3-ceph-common x86_64 2:19.2.3-47.gc24117fd552.el9.clyso @ceph 702 k 2026-03-06T13:43:39.428 INFO:teuthology.orchestra.run.vm01.stdout: python3-certifi noarch 2023.05.07-4.el9 @epel 6.3 k 2026-03-06T13:43:39.428 INFO:teuthology.orchestra.run.vm01.stdout: python3-cffi x86_64 1.14.5-5.el9 @baseos 1.0 M 2026-03-06T13:43:39.428 INFO:teuthology.orchestra.run.vm01.stdout: python3-chardet noarch 4.0.0-5.el9 @anaconda 1.4 M 2026-03-06T13:43:39.428 INFO:teuthology.orchestra.run.vm01.stdout: python3-cheroot noarch 10.0.1-4.el9 @epel 682 k 2026-03-06T13:43:39.428 INFO:teuthology.orchestra.run.vm01.stdout: python3-cherrypy noarch 18.6.1-2.el9 @epel 1.1 M 2026-03-06T13:43:39.428 INFO:teuthology.orchestra.run.vm01.stdout: python3-cryptography x86_64 36.0.1-5.el9 @baseos 4.5 M 2026-03-06T13:43:39.428 INFO:teuthology.orchestra.run.vm01.stdout: python3-devel x86_64 3.9.25-3.el9 @appstream 765 k 2026-03-06T13:43:39.428 INFO:teuthology.orchestra.run.vm01.stdout: python3-google-auth noarch 1:2.45.0-1.el9 @epel 1.4 M 2026-03-06T13:43:39.428 INFO:teuthology.orchestra.run.vm01.stdout: python3-grpcio x86_64 1.46.7-10.el9 @epel 6.7 M 2026-03-06T13:43:39.428 INFO:teuthology.orchestra.run.vm01.stdout: python3-grpcio-tools x86_64 1.46.7-10.el9 @epel 418 k 2026-03-06T13:43:39.428 INFO:teuthology.orchestra.run.vm01.stdout: python3-idna noarch 2.10-7.el9.1 @anaconda 513 k 2026-03-06T13:43:39.428 INFO:teuthology.orchestra.run.vm01.stdout: python3-influxdb noarch 5.3.1-1.el9 @epel 747 k 2026-03-06T13:43:39.428 INFO:teuthology.orchestra.run.vm01.stdout: python3-isodate noarch 0.6.1-3.el9 @epel 203 k 2026-03-06T13:43:39.428 INFO:teuthology.orchestra.run.vm01.stdout: python3-jaraco noarch 8.2.1-3.el9 @epel 3.7 k 2026-03-06T13:43:39.428 INFO:teuthology.orchestra.run.vm01.stdout: python3-jaraco-classes noarch 3.2.1-5.el9 @epel 24 k 2026-03-06T13:43:39.428 INFO:teuthology.orchestra.run.vm01.stdout: python3-jaraco-collections noarch 3.0.0-8.el9 @epel 55 k 2026-03-06T13:43:39.428 INFO:teuthology.orchestra.run.vm01.stdout: python3-jaraco-context noarch 6.0.1-3.el9 @epel 31 k 2026-03-06T13:43:39.428 INFO:teuthology.orchestra.run.vm01.stdout: python3-jaraco-functools noarch 3.5.0-2.el9 @epel 33 k 2026-03-06T13:43:39.428 INFO:teuthology.orchestra.run.vm01.stdout: python3-jaraco-text noarch 4.0.0-2.el9 @epel 51 k 2026-03-06T13:43:39.428 INFO:teuthology.orchestra.run.vm01.stdout: python3-jinja2 noarch 2.11.3-8.el9 @appstream 1.1 M 2026-03-06T13:43:39.428 INFO:teuthology.orchestra.run.vm01.stdout: python3-jsonpatch noarch 1.21-16.el9 @koji-override-0 55 k 2026-03-06T13:43:39.428 INFO:teuthology.orchestra.run.vm01.stdout: python3-jsonpointer noarch 2.0-4.el9 @koji-override-0 34 k 2026-03-06T13:43:39.428 INFO:teuthology.orchestra.run.vm01.stdout: python3-kubernetes noarch 1:26.1.0-3.el9 @epel 21 M 2026-03-06T13:43:39.428 INFO:teuthology.orchestra.run.vm01.stdout: python3-libstoragemgmt x86_64 1.10.1-1.el9 @appstream 832 k 2026-03-06T13:43:39.428 INFO:teuthology.orchestra.run.vm01.stdout: python3-logutils noarch 0.3.5-21.el9 @epel 126 k 2026-03-06T13:43:39.428 INFO:teuthology.orchestra.run.vm01.stdout: python3-lxml x86_64 4.6.5-3.el9 @appstream 4.2 M 2026-03-06T13:43:39.428 INFO:teuthology.orchestra.run.vm01.stdout: python3-mako noarch 1.1.4-6.el9 @appstream 534 k 2026-03-06T13:43:39.428 INFO:teuthology.orchestra.run.vm01.stdout: python3-markupsafe x86_64 1.1.1-12.el9 @appstream 60 k 2026-03-06T13:43:39.428 INFO:teuthology.orchestra.run.vm01.stdout: python3-more-itertools noarch 8.12.0-2.el9 @epel 378 k 2026-03-06T13:43:39.429 INFO:teuthology.orchestra.run.vm01.stdout: python3-msgpack x86_64 1.0.3-2.el9 @epel 264 k 2026-03-06T13:43:39.429 INFO:teuthology.orchestra.run.vm01.stdout: python3-natsort noarch 7.1.1-5.el9 @epel 215 k 2026-03-06T13:43:39.429 INFO:teuthology.orchestra.run.vm01.stdout: python3-numpy x86_64 1:1.23.5-2.el9 @appstream 30 M 2026-03-06T13:43:39.429 INFO:teuthology.orchestra.run.vm01.stdout: python3-numpy-f2py x86_64 1:1.23.5-2.el9 @appstream 1.7 M 2026-03-06T13:43:39.429 INFO:teuthology.orchestra.run.vm01.stdout: python3-oauthlib noarch 3.1.1-5.el9 @koji-override-0 888 k 2026-03-06T13:43:39.429 INFO:teuthology.orchestra.run.vm01.stdout: python3-packaging noarch 20.9-5.el9 @appstream 248 k 2026-03-06T13:43:39.429 INFO:teuthology.orchestra.run.vm01.stdout: python3-pecan noarch 1.4.2-3.el9 @epel 1.3 M 2026-03-06T13:43:39.429 INFO:teuthology.orchestra.run.vm01.stdout: python3-ply noarch 3.11-14.el9 @baseos 430 k 2026-03-06T13:43:39.429 INFO:teuthology.orchestra.run.vm01.stdout: python3-portend noarch 3.1.0-2.el9 @epel 20 k 2026-03-06T13:43:39.429 INFO:teuthology.orchestra.run.vm01.stdout: python3-prettytable noarch 0.7.2-27.el9 @koji-override-0 166 k 2026-03-06T13:43:39.429 INFO:teuthology.orchestra.run.vm01.stdout: python3-protobuf noarch 3.14.0-17.el9 @appstream 1.4 M 2026-03-06T13:43:39.429 INFO:teuthology.orchestra.run.vm01.stdout: python3-pyOpenSSL noarch 21.0.0-1.el9 @epel 389 k 2026-03-06T13:43:39.429 INFO:teuthology.orchestra.run.vm01.stdout: python3-pyasn1 noarch 0.4.8-7.el9 @appstream 622 k 2026-03-06T13:43:39.429 INFO:teuthology.orchestra.run.vm01.stdout: python3-pyasn1-modules noarch 0.4.8-7.el9 @appstream 1.0 M 2026-03-06T13:43:39.429 INFO:teuthology.orchestra.run.vm01.stdout: python3-pycparser noarch 2.20-6.el9 @baseos 745 k 2026-03-06T13:43:39.429 INFO:teuthology.orchestra.run.vm01.stdout: python3-pysocks noarch 1.7.1-12.el9 @anaconda 88 k 2026-03-06T13:43:39.429 INFO:teuthology.orchestra.run.vm01.stdout: python3-pytz noarch 2021.1-5.el9 @koji-override-0 176 k 2026-03-06T13:43:39.429 INFO:teuthology.orchestra.run.vm01.stdout: python3-repoze-lru noarch 0.7-16.el9 @epel 83 k 2026-03-06T13:43:39.429 INFO:teuthology.orchestra.run.vm01.stdout: python3-requests noarch 2.25.1-10.el9 @baseos 405 k 2026-03-06T13:43:39.429 INFO:teuthology.orchestra.run.vm01.stdout: python3-requests-oauthlib noarch 1.3.0-12.el9 @appstream 119 k 2026-03-06T13:43:39.429 INFO:teuthology.orchestra.run.vm01.stdout: python3-routes noarch 2.5.1-5.el9 @epel 459 k 2026-03-06T13:43:39.429 INFO:teuthology.orchestra.run.vm01.stdout: python3-rsa noarch 4.9-2.el9 @epel 202 k 2026-03-06T13:43:39.429 INFO:teuthology.orchestra.run.vm01.stdout: python3-saml noarch 1.16.0-1.el9 @epel 730 k 2026-03-06T13:43:39.429 INFO:teuthology.orchestra.run.vm01.stdout: python3-scipy x86_64 1.9.3-2.el9 @appstream 76 M 2026-03-06T13:43:39.429 INFO:teuthology.orchestra.run.vm01.stdout: python3-tempora noarch 5.0.0-2.el9 @epel 96 k 2026-03-06T13:43:39.429 INFO:teuthology.orchestra.run.vm01.stdout: python3-toml noarch 0.10.2-6.el9 @appstream 99 k 2026-03-06T13:43:39.429 INFO:teuthology.orchestra.run.vm01.stdout: python3-typing-extensions noarch 4.15.0-1.el9 @epel 447 k 2026-03-06T13:43:39.429 INFO:teuthology.orchestra.run.vm01.stdout: python3-urllib3 noarch 1.26.5-7.el9 @baseos 746 k 2026-03-06T13:43:39.429 INFO:teuthology.orchestra.run.vm01.stdout: python3-webob noarch 1.8.8-2.el9 @epel 1.2 M 2026-03-06T13:43:39.429 INFO:teuthology.orchestra.run.vm01.stdout: python3-websocket-client noarch 1.2.3-2.el9 @epel 319 k 2026-03-06T13:43:39.429 INFO:teuthology.orchestra.run.vm01.stdout: python3-werkzeug noarch 2.0.3-3.el9.1 @epel 1.9 M 2026-03-06T13:43:39.429 INFO:teuthology.orchestra.run.vm01.stdout: python3-xmlsec x86_64 1.3.13-1.el9 @epel 158 k 2026-03-06T13:43:39.429 INFO:teuthology.orchestra.run.vm01.stdout: python3-zc-lockfile noarch 2.0-10.el9 @epel 35 k 2026-03-06T13:43:39.429 INFO:teuthology.orchestra.run.vm01.stdout: qatlib x86_64 25.08.0-2.el9 @appstream 639 k 2026-03-06T13:43:39.429 INFO:teuthology.orchestra.run.vm01.stdout: qatlib-service x86_64 25.08.0-2.el9 @appstream 69 k 2026-03-06T13:43:39.429 INFO:teuthology.orchestra.run.vm01.stdout: qatzip-libs x86_64 1.3.1-1.el9 @appstream 148 k 2026-03-06T13:43:39.429 INFO:teuthology.orchestra.run.vm01.stdout: smartmontools x86_64 1:7.2-10.el9 @baseos 1.9 M 2026-03-06T13:43:39.429 INFO:teuthology.orchestra.run.vm01.stdout: xmlsec1 x86_64 1.2.29-13.el9 @appstream 596 k 2026-03-06T13:43:39.429 INFO:teuthology.orchestra.run.vm01.stdout: xmlsec1-openssl x86_64 1.2.29-13.el9 @appstream 281 k 2026-03-06T13:43:39.429 INFO:teuthology.orchestra.run.vm01.stdout: 2026-03-06T13:43:39.429 INFO:teuthology.orchestra.run.vm01.stdout:Transaction Summary 2026-03-06T13:43:39.429 INFO:teuthology.orchestra.run.vm01.stdout:=================================================================================================== 2026-03-06T13:43:39.430 INFO:teuthology.orchestra.run.vm01.stdout:Remove 113 Packages 2026-03-06T13:43:39.430 INFO:teuthology.orchestra.run.vm01.stdout: 2026-03-06T13:43:39.430 INFO:teuthology.orchestra.run.vm01.stdout:Freed space: 623 M 2026-03-06T13:43:39.430 INFO:teuthology.orchestra.run.vm01.stdout:Running transaction check 2026-03-06T13:43:39.453 INFO:teuthology.orchestra.run.vm08.stdout:Dependencies resolved. 2026-03-06T13:43:39.459 INFO:teuthology.orchestra.run.vm08.stdout:=================================================================================================== 2026-03-06T13:43:39.459 INFO:teuthology.orchestra.run.vm08.stdout: Package Arch Version Repository Size 2026-03-06T13:43:39.460 INFO:teuthology.orchestra.run.vm08.stdout:=================================================================================================== 2026-03-06T13:43:39.460 INFO:teuthology.orchestra.run.vm08.stdout:Removing: 2026-03-06T13:43:39.460 INFO:teuthology.orchestra.run.vm08.stdout: ceph-base x86_64 2:19.2.3-47.gc24117fd552.el9.clyso @ceph 23 M 2026-03-06T13:43:39.460 INFO:teuthology.orchestra.run.vm08.stdout:Removing dependent packages: 2026-03-06T13:43:39.460 INFO:teuthology.orchestra.run.vm08.stdout: ceph-immutable-object-cache x86_64 2:19.2.3-47.gc24117fd552.el9.clyso @ceph 431 k 2026-03-06T13:43:39.460 INFO:teuthology.orchestra.run.vm08.stdout: ceph-mgr x86_64 2:19.2.3-47.gc24117fd552.el9.clyso @ceph 3.4 M 2026-03-06T13:43:39.460 INFO:teuthology.orchestra.run.vm08.stdout: ceph-mgr-cephadm noarch 2:19.2.3-47.gc24117fd552.el9.clyso @ceph-noarch 803 k 2026-03-06T13:43:39.460 INFO:teuthology.orchestra.run.vm08.stdout: ceph-mgr-dashboard noarch 2:19.2.3-47.gc24117fd552.el9.clyso @ceph-noarch 88 M 2026-03-06T13:43:39.460 INFO:teuthology.orchestra.run.vm08.stdout: ceph-mgr-diskprediction-local noarch 2:19.2.3-47.gc24117fd552.el9.clyso @ceph-noarch 66 M 2026-03-06T13:43:39.460 INFO:teuthology.orchestra.run.vm08.stdout: ceph-mgr-rook noarch 2:19.2.3-47.gc24117fd552.el9.clyso @ceph-noarch 563 k 2026-03-06T13:43:39.460 INFO:teuthology.orchestra.run.vm08.stdout: ceph-osd x86_64 2:19.2.3-47.gc24117fd552.el9.clyso @ceph 59 M 2026-03-06T13:43:39.460 INFO:teuthology.orchestra.run.vm08.stdout: ceph-volume noarch 2:19.2.3-47.gc24117fd552.el9.clyso @ceph-noarch 1.4 M 2026-03-06T13:43:39.460 INFO:teuthology.orchestra.run.vm08.stdout: rbd-mirror x86_64 2:19.2.3-47.gc24117fd552.el9.clyso @ceph 13 M 2026-03-06T13:43:39.460 INFO:teuthology.orchestra.run.vm08.stdout:Removing unused dependencies: 2026-03-06T13:43:39.460 INFO:teuthology.orchestra.run.vm08.stdout: abseil-cpp x86_64 20211102.0-4.el9 @epel 1.9 M 2026-03-06T13:43:39.460 INFO:teuthology.orchestra.run.vm08.stdout: ceph-common x86_64 2:19.2.3-47.gc24117fd552.el9.clyso @ceph 85 M 2026-03-06T13:43:39.460 INFO:teuthology.orchestra.run.vm08.stdout: ceph-grafana-dashboards noarch 2:19.2.3-47.gc24117fd552.el9.clyso @ceph-noarch 626 k 2026-03-06T13:43:39.460 INFO:teuthology.orchestra.run.vm08.stdout: ceph-mgr-k8sevents noarch 2:19.2.3-47.gc24117fd552.el9.clyso @ceph-noarch 60 k 2026-03-06T13:43:39.460 INFO:teuthology.orchestra.run.vm08.stdout: ceph-mgr-modules-core noarch 2:19.2.3-47.gc24117fd552.el9.clyso @ceph-noarch 1.5 M 2026-03-06T13:43:39.460 INFO:teuthology.orchestra.run.vm08.stdout: ceph-prometheus-alerts noarch 2:19.2.3-47.gc24117fd552.el9.clyso @ceph-noarch 51 k 2026-03-06T13:43:39.460 INFO:teuthology.orchestra.run.vm08.stdout: ceph-selinux x86_64 2:19.2.3-47.gc24117fd552.el9.clyso @ceph 138 k 2026-03-06T13:43:39.460 INFO:teuthology.orchestra.run.vm08.stdout: cryptsetup x86_64 2.8.1-3.el9 @baseos 770 k 2026-03-06T13:43:39.460 INFO:teuthology.orchestra.run.vm08.stdout: flexiblas x86_64 3.0.4-9.el9 @appstream 68 k 2026-03-06T13:43:39.460 INFO:teuthology.orchestra.run.vm08.stdout: flexiblas-netlib x86_64 3.0.4-9.el9 @appstream 11 M 2026-03-06T13:43:39.460 INFO:teuthology.orchestra.run.vm08.stdout: flexiblas-openblas-openmp x86_64 3.0.4-9.el9 @appstream 39 k 2026-03-06T13:43:39.460 INFO:teuthology.orchestra.run.vm08.stdout: gperftools-libs x86_64 2.9.1-3.el9 @epel 1.4 M 2026-03-06T13:43:39.460 INFO:teuthology.orchestra.run.vm08.stdout: grpc-data noarch 1.46.7-10.el9 @epel 13 k 2026-03-06T13:43:39.460 INFO:teuthology.orchestra.run.vm08.stdout: ledmon-libs x86_64 1.1.0-3.el9 @baseos 80 k 2026-03-06T13:43:39.460 INFO:teuthology.orchestra.run.vm08.stdout: libcephsqlite x86_64 2:19.2.3-47.gc24117fd552.el9.clyso @ceph 425 k 2026-03-06T13:43:39.460 INFO:teuthology.orchestra.run.vm08.stdout: libconfig x86_64 1.7.2-9.el9 @baseos 220 k 2026-03-06T13:43:39.460 INFO:teuthology.orchestra.run.vm08.stdout: libgfortran x86_64 11.5.0-14.el9 @baseos 2.8 M 2026-03-06T13:43:39.460 INFO:teuthology.orchestra.run.vm08.stdout: liboath x86_64 2.6.12-1.el9 @epel 94 k 2026-03-06T13:43:39.460 INFO:teuthology.orchestra.run.vm08.stdout: libquadmath x86_64 11.5.0-14.el9 @baseos 330 k 2026-03-06T13:43:39.460 INFO:teuthology.orchestra.run.vm08.stdout: libradosstriper1 x86_64 2:19.2.3-47.gc24117fd552.el9.clyso @ceph 1.6 M 2026-03-06T13:43:39.460 INFO:teuthology.orchestra.run.vm08.stdout: libstoragemgmt x86_64 1.10.1-1.el9 @appstream 685 k 2026-03-06T13:43:39.460 INFO:teuthology.orchestra.run.vm08.stdout: libunwind x86_64 1.6.2-1.el9 @epel 170 k 2026-03-06T13:43:39.460 INFO:teuthology.orchestra.run.vm08.stdout: libxslt x86_64 1.1.34-12.el9 @appstream 743 k 2026-03-06T13:43:39.460 INFO:teuthology.orchestra.run.vm08.stdout: openblas x86_64 0.3.29-1.el9 @appstream 112 k 2026-03-06T13:43:39.460 INFO:teuthology.orchestra.run.vm08.stdout: openblas-openmp x86_64 0.3.29-1.el9 @appstream 46 M 2026-03-06T13:43:39.460 INFO:teuthology.orchestra.run.vm08.stdout: pciutils x86_64 3.7.0-7.el9 @baseos 216 k 2026-03-06T13:43:39.460 INFO:teuthology.orchestra.run.vm08.stdout: protobuf x86_64 3.14.0-17.el9 @appstream 3.5 M 2026-03-06T13:43:39.460 INFO:teuthology.orchestra.run.vm08.stdout: protobuf-compiler x86_64 3.14.0-17.el9 @crb 2.9 M 2026-03-06T13:43:39.460 INFO:teuthology.orchestra.run.vm08.stdout: python3-asyncssh noarch 2.13.2-5.el9 @epel 3.9 M 2026-03-06T13:43:39.460 INFO:teuthology.orchestra.run.vm08.stdout: python3-autocommand noarch 2.2.2-8.el9 @epel 82 k 2026-03-06T13:43:39.460 INFO:teuthology.orchestra.run.vm08.stdout: python3-babel noarch 2.9.1-2.el9 @appstream 27 M 2026-03-06T13:43:39.460 INFO:teuthology.orchestra.run.vm08.stdout: python3-backports-tarfile noarch 1.2.0-1.el9 @epel 254 k 2026-03-06T13:43:39.460 INFO:teuthology.orchestra.run.vm08.stdout: python3-bcrypt x86_64 3.2.2-1.el9 @epel 87 k 2026-03-06T13:43:39.460 INFO:teuthology.orchestra.run.vm08.stdout: python3-cachetools noarch 4.2.4-1.el9 @epel 93 k 2026-03-06T13:43:39.460 INFO:teuthology.orchestra.run.vm08.stdout: python3-ceph-common x86_64 2:19.2.3-47.gc24117fd552.el9.clyso @ceph 702 k 2026-03-06T13:43:39.460 INFO:teuthology.orchestra.run.vm08.stdout: python3-certifi noarch 2023.05.07-4.el9 @epel 6.3 k 2026-03-06T13:43:39.460 INFO:teuthology.orchestra.run.vm08.stdout: python3-cffi x86_64 1.14.5-5.el9 @baseos 1.0 M 2026-03-06T13:43:39.460 INFO:teuthology.orchestra.run.vm08.stdout: python3-chardet noarch 4.0.0-5.el9 @anaconda 1.4 M 2026-03-06T13:43:39.460 INFO:teuthology.orchestra.run.vm08.stdout: python3-cheroot noarch 10.0.1-4.el9 @epel 682 k 2026-03-06T13:43:39.460 INFO:teuthology.orchestra.run.vm08.stdout: python3-cherrypy noarch 18.6.1-2.el9 @epel 1.1 M 2026-03-06T13:43:39.460 INFO:teuthology.orchestra.run.vm08.stdout: python3-cryptography x86_64 36.0.1-5.el9 @baseos 4.5 M 2026-03-06T13:43:39.460 INFO:teuthology.orchestra.run.vm08.stdout: python3-devel x86_64 3.9.25-3.el9 @appstream 765 k 2026-03-06T13:43:39.460 INFO:teuthology.orchestra.run.vm08.stdout: python3-google-auth noarch 1:2.45.0-1.el9 @epel 1.4 M 2026-03-06T13:43:39.460 INFO:teuthology.orchestra.run.vm08.stdout: python3-grpcio x86_64 1.46.7-10.el9 @epel 6.7 M 2026-03-06T13:43:39.460 INFO:teuthology.orchestra.run.vm08.stdout: python3-grpcio-tools x86_64 1.46.7-10.el9 @epel 418 k 2026-03-06T13:43:39.461 INFO:teuthology.orchestra.run.vm08.stdout: python3-idna noarch 2.10-7.el9.1 @anaconda 513 k 2026-03-06T13:43:39.461 INFO:teuthology.orchestra.run.vm08.stdout: python3-influxdb noarch 5.3.1-1.el9 @epel 747 k 2026-03-06T13:43:39.461 INFO:teuthology.orchestra.run.vm08.stdout: python3-isodate noarch 0.6.1-3.el9 @epel 203 k 2026-03-06T13:43:39.461 INFO:teuthology.orchestra.run.vm08.stdout: python3-jaraco noarch 8.2.1-3.el9 @epel 3.7 k 2026-03-06T13:43:39.461 INFO:teuthology.orchestra.run.vm08.stdout: python3-jaraco-classes noarch 3.2.1-5.el9 @epel 24 k 2026-03-06T13:43:39.461 INFO:teuthology.orchestra.run.vm08.stdout: python3-jaraco-collections noarch 3.0.0-8.el9 @epel 55 k 2026-03-06T13:43:39.461 INFO:teuthology.orchestra.run.vm08.stdout: python3-jaraco-context noarch 6.0.1-3.el9 @epel 31 k 2026-03-06T13:43:39.461 INFO:teuthology.orchestra.run.vm08.stdout: python3-jaraco-functools noarch 3.5.0-2.el9 @epel 33 k 2026-03-06T13:43:39.461 INFO:teuthology.orchestra.run.vm08.stdout: python3-jaraco-text noarch 4.0.0-2.el9 @epel 51 k 2026-03-06T13:43:39.461 INFO:teuthology.orchestra.run.vm08.stdout: python3-jinja2 noarch 2.11.3-8.el9 @appstream 1.1 M 2026-03-06T13:43:39.461 INFO:teuthology.orchestra.run.vm08.stdout: python3-jsonpatch noarch 1.21-16.el9 @koji-override-0 55 k 2026-03-06T13:43:39.461 INFO:teuthology.orchestra.run.vm08.stdout: python3-jsonpointer noarch 2.0-4.el9 @koji-override-0 34 k 2026-03-06T13:43:39.461 INFO:teuthology.orchestra.run.vm08.stdout: python3-kubernetes noarch 1:26.1.0-3.el9 @epel 21 M 2026-03-06T13:43:39.461 INFO:teuthology.orchestra.run.vm08.stdout: python3-libstoragemgmt x86_64 1.10.1-1.el9 @appstream 832 k 2026-03-06T13:43:39.461 INFO:teuthology.orchestra.run.vm08.stdout: python3-logutils noarch 0.3.5-21.el9 @epel 126 k 2026-03-06T13:43:39.461 INFO:teuthology.orchestra.run.vm08.stdout: python3-lxml x86_64 4.6.5-3.el9 @appstream 4.2 M 2026-03-06T13:43:39.461 INFO:teuthology.orchestra.run.vm08.stdout: python3-mako noarch 1.1.4-6.el9 @appstream 534 k 2026-03-06T13:43:39.461 INFO:teuthology.orchestra.run.vm08.stdout: python3-markupsafe x86_64 1.1.1-12.el9 @appstream 60 k 2026-03-06T13:43:39.461 INFO:teuthology.orchestra.run.vm08.stdout: python3-more-itertools noarch 8.12.0-2.el9 @epel 378 k 2026-03-06T13:43:39.461 INFO:teuthology.orchestra.run.vm08.stdout: python3-msgpack x86_64 1.0.3-2.el9 @epel 264 k 2026-03-06T13:43:39.461 INFO:teuthology.orchestra.run.vm08.stdout: python3-natsort noarch 7.1.1-5.el9 @epel 215 k 2026-03-06T13:43:39.461 INFO:teuthology.orchestra.run.vm08.stdout: python3-numpy x86_64 1:1.23.5-2.el9 @appstream 30 M 2026-03-06T13:43:39.461 INFO:teuthology.orchestra.run.vm08.stdout: python3-numpy-f2py x86_64 1:1.23.5-2.el9 @appstream 1.7 M 2026-03-06T13:43:39.461 INFO:teuthology.orchestra.run.vm08.stdout: python3-oauthlib noarch 3.1.1-5.el9 @koji-override-0 888 k 2026-03-06T13:43:39.461 INFO:teuthology.orchestra.run.vm08.stdout: python3-packaging noarch 20.9-5.el9 @appstream 248 k 2026-03-06T13:43:39.461 INFO:teuthology.orchestra.run.vm08.stdout: python3-pecan noarch 1.4.2-3.el9 @epel 1.3 M 2026-03-06T13:43:39.461 INFO:teuthology.orchestra.run.vm08.stdout: python3-ply noarch 3.11-14.el9 @baseos 430 k 2026-03-06T13:43:39.461 INFO:teuthology.orchestra.run.vm08.stdout: python3-portend noarch 3.1.0-2.el9 @epel 20 k 2026-03-06T13:43:39.461 INFO:teuthology.orchestra.run.vm08.stdout: python3-prettytable noarch 0.7.2-27.el9 @koji-override-0 166 k 2026-03-06T13:43:39.461 INFO:teuthology.orchestra.run.vm08.stdout: python3-protobuf noarch 3.14.0-17.el9 @appstream 1.4 M 2026-03-06T13:43:39.461 INFO:teuthology.orchestra.run.vm08.stdout: python3-pyOpenSSL noarch 21.0.0-1.el9 @epel 389 k 2026-03-06T13:43:39.461 INFO:teuthology.orchestra.run.vm08.stdout: python3-pyasn1 noarch 0.4.8-7.el9 @appstream 622 k 2026-03-06T13:43:39.461 INFO:teuthology.orchestra.run.vm08.stdout: python3-pyasn1-modules noarch 0.4.8-7.el9 @appstream 1.0 M 2026-03-06T13:43:39.461 INFO:teuthology.orchestra.run.vm08.stdout: python3-pycparser noarch 2.20-6.el9 @baseos 745 k 2026-03-06T13:43:39.461 INFO:teuthology.orchestra.run.vm08.stdout: python3-pysocks noarch 1.7.1-12.el9 @anaconda 88 k 2026-03-06T13:43:39.461 INFO:teuthology.orchestra.run.vm08.stdout: python3-pytz noarch 2021.1-5.el9 @koji-override-0 176 k 2026-03-06T13:43:39.461 INFO:teuthology.orchestra.run.vm08.stdout: python3-repoze-lru noarch 0.7-16.el9 @epel 83 k 2026-03-06T13:43:39.461 INFO:teuthology.orchestra.run.vm08.stdout: python3-requests noarch 2.25.1-10.el9 @baseos 405 k 2026-03-06T13:43:39.461 INFO:teuthology.orchestra.run.vm08.stdout: python3-requests-oauthlib noarch 1.3.0-12.el9 @appstream 119 k 2026-03-06T13:43:39.461 INFO:teuthology.orchestra.run.vm08.stdout: python3-routes noarch 2.5.1-5.el9 @epel 459 k 2026-03-06T13:43:39.461 INFO:teuthology.orchestra.run.vm08.stdout: python3-rsa noarch 4.9-2.el9 @epel 202 k 2026-03-06T13:43:39.461 INFO:teuthology.orchestra.run.vm08.stdout: python3-saml noarch 1.16.0-1.el9 @epel 730 k 2026-03-06T13:43:39.461 INFO:teuthology.orchestra.run.vm08.stdout: python3-scipy x86_64 1.9.3-2.el9 @appstream 76 M 2026-03-06T13:43:39.461 INFO:teuthology.orchestra.run.vm08.stdout: python3-tempora noarch 5.0.0-2.el9 @epel 96 k 2026-03-06T13:43:39.461 INFO:teuthology.orchestra.run.vm08.stdout: python3-toml noarch 0.10.2-6.el9 @appstream 99 k 2026-03-06T13:43:39.461 INFO:teuthology.orchestra.run.vm08.stdout: python3-typing-extensions noarch 4.15.0-1.el9 @epel 447 k 2026-03-06T13:43:39.461 INFO:teuthology.orchestra.run.vm08.stdout: python3-urllib3 noarch 1.26.5-7.el9 @baseos 746 k 2026-03-06T13:43:39.461 INFO:teuthology.orchestra.run.vm08.stdout: python3-webob noarch 1.8.8-2.el9 @epel 1.2 M 2026-03-06T13:43:39.461 INFO:teuthology.orchestra.run.vm08.stdout: python3-websocket-client noarch 1.2.3-2.el9 @epel 319 k 2026-03-06T13:43:39.461 INFO:teuthology.orchestra.run.vm08.stdout: python3-werkzeug noarch 2.0.3-3.el9.1 @epel 1.9 M 2026-03-06T13:43:39.461 INFO:teuthology.orchestra.run.vm08.stdout: python3-xmlsec x86_64 1.3.13-1.el9 @epel 158 k 2026-03-06T13:43:39.461 INFO:teuthology.orchestra.run.vm08.stdout: python3-zc-lockfile noarch 2.0-10.el9 @epel 35 k 2026-03-06T13:43:39.461 INFO:teuthology.orchestra.run.vm08.stdout: qatlib x86_64 25.08.0-2.el9 @appstream 639 k 2026-03-06T13:43:39.461 INFO:teuthology.orchestra.run.vm08.stdout: qatlib-service x86_64 25.08.0-2.el9 @appstream 69 k 2026-03-06T13:43:39.461 INFO:teuthology.orchestra.run.vm08.stdout: qatzip-libs x86_64 1.3.1-1.el9 @appstream 148 k 2026-03-06T13:43:39.462 INFO:teuthology.orchestra.run.vm08.stdout: smartmontools x86_64 1:7.2-10.el9 @baseos 1.9 M 2026-03-06T13:43:39.462 INFO:teuthology.orchestra.run.vm08.stdout: xmlsec1 x86_64 1.2.29-13.el9 @appstream 596 k 2026-03-06T13:43:39.462 INFO:teuthology.orchestra.run.vm08.stdout: xmlsec1-openssl x86_64 1.2.29-13.el9 @appstream 281 k 2026-03-06T13:43:39.462 INFO:teuthology.orchestra.run.vm08.stdout: 2026-03-06T13:43:39.462 INFO:teuthology.orchestra.run.vm08.stdout:Transaction Summary 2026-03-06T13:43:39.462 INFO:teuthology.orchestra.run.vm08.stdout:=================================================================================================== 2026-03-06T13:43:39.462 INFO:teuthology.orchestra.run.vm08.stdout:Remove 113 Packages 2026-03-06T13:43:39.462 INFO:teuthology.orchestra.run.vm08.stdout: 2026-03-06T13:43:39.462 INFO:teuthology.orchestra.run.vm01.stdout:Transaction check succeeded. 2026-03-06T13:43:39.462 INFO:teuthology.orchestra.run.vm01.stdout:Running transaction test 2026-03-06T13:43:39.462 INFO:teuthology.orchestra.run.vm08.stdout:Freed space: 623 M 2026-03-06T13:43:39.462 INFO:teuthology.orchestra.run.vm08.stdout:Running transaction check 2026-03-06T13:43:39.491 INFO:teuthology.orchestra.run.vm08.stdout:Transaction check succeeded. 2026-03-06T13:43:39.491 INFO:teuthology.orchestra.run.vm08.stdout:Running transaction test 2026-03-06T13:43:39.497 INFO:teuthology.orchestra.run.vm00.stdout:Transaction test succeeded. 2026-03-06T13:43:39.497 INFO:teuthology.orchestra.run.vm00.stdout:Running transaction 2026-03-06T13:43:39.581 INFO:teuthology.orchestra.run.vm01.stdout:Transaction test succeeded. 2026-03-06T13:43:39.581 INFO:teuthology.orchestra.run.vm01.stdout:Running transaction 2026-03-06T13:43:39.611 INFO:teuthology.orchestra.run.vm08.stdout:Transaction test succeeded. 2026-03-06T13:43:39.611 INFO:teuthology.orchestra.run.vm08.stdout:Running transaction 2026-03-06T13:43:39.667 INFO:teuthology.orchestra.run.vm00.stdout: Preparing : 1/1 2026-03-06T13:43:39.667 INFO:teuthology.orchestra.run.vm00.stdout: Erasing : ceph-mgr-rook-2:19.2.3-47.gc24117fd552.el9.clyso 1/113 2026-03-06T13:43:39.676 INFO:teuthology.orchestra.run.vm00.stdout: Running scriptlet: ceph-mgr-rook-2:19.2.3-47.gc24117fd552.el9.clyso 1/113 2026-03-06T13:43:39.698 INFO:teuthology.orchestra.run.vm00.stdout: Running scriptlet: ceph-mgr-2:19.2.3-47.gc24117fd552.el9.clyso.x86_ 2/113 2026-03-06T13:43:39.698 INFO:teuthology.orchestra.run.vm00.stdout:Glob pattern passed to enable, but globs are not supported for this. 2026-03-06T13:43:39.698 INFO:teuthology.orchestra.run.vm00.stdout:Invalid unit name "ceph-mgr@*.service" escaped as "ceph-mgr@\x2a.service". 2026-03-06T13:43:39.698 INFO:teuthology.orchestra.run.vm00.stdout:Removed "/etc/systemd/system/multi-user.target.wants/ceph-mgr.target". 2026-03-06T13:43:39.698 INFO:teuthology.orchestra.run.vm00.stdout:Removed "/etc/systemd/system/ceph.target.wants/ceph-mgr.target". 2026-03-06T13:43:39.698 INFO:teuthology.orchestra.run.vm00.stdout: 2026-03-06T13:43:39.699 INFO:teuthology.orchestra.run.vm00.stdout: Erasing : ceph-mgr-2:19.2.3-47.gc24117fd552.el9.clyso.x86_ 2/113 2026-03-06T13:43:39.714 INFO:teuthology.orchestra.run.vm00.stdout: Running scriptlet: ceph-mgr-2:19.2.3-47.gc24117fd552.el9.clyso.x86_ 2/113 2026-03-06T13:43:39.740 INFO:teuthology.orchestra.run.vm00.stdout: Erasing : ceph-mgr-modules-core-2:19.2.3-47.gc24117fd552.e 3/113 2026-03-06T13:43:39.741 INFO:teuthology.orchestra.run.vm00.stdout: Erasing : ceph-mgr-dashboard-2:19.2.3-47.gc24117fd552.el9. 4/113 2026-03-06T13:43:39.749 INFO:teuthology.orchestra.run.vm01.stdout: Preparing : 1/1 2026-03-06T13:43:39.749 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : ceph-mgr-rook-2:19.2.3-47.gc24117fd552.el9.clyso 1/113 2026-03-06T13:43:39.755 INFO:teuthology.orchestra.run.vm00.stdout: Running scriptlet: ceph-mgr-dashboard-2:19.2.3-47.gc24117fd552.el9. 4/113 2026-03-06T13:43:39.758 INFO:teuthology.orchestra.run.vm01.stdout: Running scriptlet: ceph-mgr-rook-2:19.2.3-47.gc24117fd552.el9.clyso 1/113 2026-03-06T13:43:39.759 INFO:teuthology.orchestra.run.vm00.stdout: Erasing : python3-influxdb-5.3.1-1.el9.noarch 5/113 2026-03-06T13:43:39.760 INFO:teuthology.orchestra.run.vm00.stdout: Erasing : ceph-mgr-cephadm-2:19.2.3-47.gc24117fd552.el9.cl 6/113 2026-03-06T13:43:39.774 INFO:teuthology.orchestra.run.vm00.stdout: Running scriptlet: ceph-mgr-cephadm-2:19.2.3-47.gc24117fd552.el9.cl 6/113 2026-03-06T13:43:39.778 INFO:teuthology.orchestra.run.vm01.stdout: Running scriptlet: ceph-mgr-2:19.2.3-47.gc24117fd552.el9.clyso.x86_ 2/113 2026-03-06T13:43:39.778 INFO:teuthology.orchestra.run.vm01.stdout:Glob pattern passed to enable, but globs are not supported for this. 2026-03-06T13:43:39.778 INFO:teuthology.orchestra.run.vm01.stdout:Invalid unit name "ceph-mgr@*.service" escaped as "ceph-mgr@\x2a.service". 2026-03-06T13:43:39.778 INFO:teuthology.orchestra.run.vm01.stdout:Removed "/etc/systemd/system/multi-user.target.wants/ceph-mgr.target". 2026-03-06T13:43:39.778 INFO:teuthology.orchestra.run.vm01.stdout:Removed "/etc/systemd/system/ceph.target.wants/ceph-mgr.target". 2026-03-06T13:43:39.778 INFO:teuthology.orchestra.run.vm01.stdout: 2026-03-06T13:43:39.778 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : ceph-mgr-2:19.2.3-47.gc24117fd552.el9.clyso.x86_ 2/113 2026-03-06T13:43:39.779 INFO:teuthology.orchestra.run.vm08.stdout: Preparing : 1/1 2026-03-06T13:43:39.779 INFO:teuthology.orchestra.run.vm08.stdout: Erasing : ceph-mgr-rook-2:19.2.3-47.gc24117fd552.el9.clyso 1/113 2026-03-06T13:43:39.781 INFO:teuthology.orchestra.run.vm00.stdout: Erasing : python3-cherrypy-18.6.1-2.el9.noarch 7/113 2026-03-06T13:43:39.786 INFO:teuthology.orchestra.run.vm00.stdout: Erasing : python3-cheroot-10.0.1-4.el9.noarch 8/113 2026-03-06T13:43:39.786 INFO:teuthology.orchestra.run.vm08.stdout: Running scriptlet: ceph-mgr-rook-2:19.2.3-47.gc24117fd552.el9.clyso 1/113 2026-03-06T13:43:39.792 INFO:teuthology.orchestra.run.vm01.stdout: Running scriptlet: ceph-mgr-2:19.2.3-47.gc24117fd552.el9.clyso.x86_ 2/113 2026-03-06T13:43:39.794 INFO:teuthology.orchestra.run.vm00.stdout: Erasing : python3-grpcio-tools-1.46.7-10.el9.x86_64 9/113 2026-03-06T13:43:39.798 INFO:teuthology.orchestra.run.vm00.stdout: Erasing : python3-grpcio-1.46.7-10.el9.x86_64 10/113 2026-03-06T13:43:39.805 INFO:teuthology.orchestra.run.vm08.stdout: Running scriptlet: ceph-mgr-2:19.2.3-47.gc24117fd552.el9.clyso.x86_ 2/113 2026-03-06T13:43:39.805 INFO:teuthology.orchestra.run.vm08.stdout:Glob pattern passed to enable, but globs are not supported for this. 2026-03-06T13:43:39.805 INFO:teuthology.orchestra.run.vm08.stdout:Invalid unit name "ceph-mgr@*.service" escaped as "ceph-mgr@\x2a.service". 2026-03-06T13:43:39.805 INFO:teuthology.orchestra.run.vm08.stdout:Removed "/etc/systemd/system/multi-user.target.wants/ceph-mgr.target". 2026-03-06T13:43:39.805 INFO:teuthology.orchestra.run.vm08.stdout:Removed "/etc/systemd/system/ceph.target.wants/ceph-mgr.target". 2026-03-06T13:43:39.805 INFO:teuthology.orchestra.run.vm08.stdout: 2026-03-06T13:43:39.806 INFO:teuthology.orchestra.run.vm08.stdout: Erasing : ceph-mgr-2:19.2.3-47.gc24117fd552.el9.clyso.x86_ 2/113 2026-03-06T13:43:39.818 INFO:teuthology.orchestra.run.vm08.stdout: Running scriptlet: ceph-mgr-2:19.2.3-47.gc24117fd552.el9.clyso.x86_ 2/113 2026-03-06T13:43:39.820 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : ceph-mgr-modules-core-2:19.2.3-47.gc24117fd552.e 3/113 2026-03-06T13:43:39.820 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : ceph-mgr-dashboard-2:19.2.3-47.gc24117fd552.el9. 4/113 2026-03-06T13:43:39.823 INFO:teuthology.orchestra.run.vm00.stdout: Running scriptlet: ceph-osd-2:19.2.3-47.gc24117fd552.el9.clyso.x86_ 11/113 2026-03-06T13:43:39.823 INFO:teuthology.orchestra.run.vm00.stdout:Glob pattern passed to enable, but globs are not supported for this. 2026-03-06T13:43:39.823 INFO:teuthology.orchestra.run.vm00.stdout:Invalid unit name "ceph-osd@*.service" escaped as "ceph-osd@\x2a.service". 2026-03-06T13:43:39.823 INFO:teuthology.orchestra.run.vm00.stdout:Removed "/etc/systemd/system/multi-user.target.wants/ceph-osd.target". 2026-03-06T13:43:39.823 INFO:teuthology.orchestra.run.vm00.stdout:Removed "/etc/systemd/system/ceph.target.wants/ceph-osd.target". 2026-03-06T13:43:39.823 INFO:teuthology.orchestra.run.vm00.stdout: 2026-03-06T13:43:39.829 INFO:teuthology.orchestra.run.vm00.stdout: Erasing : ceph-osd-2:19.2.3-47.gc24117fd552.el9.clyso.x86_ 11/113 2026-03-06T13:43:39.833 INFO:teuthology.orchestra.run.vm01.stdout: Running scriptlet: ceph-mgr-dashboard-2:19.2.3-47.gc24117fd552.el9. 4/113 2026-03-06T13:43:39.838 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : python3-influxdb-5.3.1-1.el9.noarch 5/113 2026-03-06T13:43:39.838 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : ceph-mgr-cephadm-2:19.2.3-47.gc24117fd552.el9.cl 6/113 2026-03-06T13:43:39.839 INFO:teuthology.orchestra.run.vm00.stdout: Running scriptlet: ceph-osd-2:19.2.3-47.gc24117fd552.el9.clyso.x86_ 11/113 2026-03-06T13:43:39.842 INFO:teuthology.orchestra.run.vm08.stdout: Erasing : ceph-mgr-modules-core-2:19.2.3-47.gc24117fd552.e 3/113 2026-03-06T13:43:39.842 INFO:teuthology.orchestra.run.vm08.stdout: Erasing : ceph-mgr-dashboard-2:19.2.3-47.gc24117fd552.el9. 4/113 2026-03-06T13:43:39.850 INFO:teuthology.orchestra.run.vm01.stdout: Running scriptlet: ceph-mgr-cephadm-2:19.2.3-47.gc24117fd552.el9.cl 6/113 2026-03-06T13:43:39.856 INFO:teuthology.orchestra.run.vm08.stdout: Running scriptlet: ceph-mgr-dashboard-2:19.2.3-47.gc24117fd552.el9. 4/113 2026-03-06T13:43:39.858 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : python3-cherrypy-18.6.1-2.el9.noarch 7/113 2026-03-06T13:43:39.861 INFO:teuthology.orchestra.run.vm08.stdout: Erasing : python3-influxdb-5.3.1-1.el9.noarch 5/113 2026-03-06T13:43:39.861 INFO:teuthology.orchestra.run.vm08.stdout: Erasing : ceph-mgr-cephadm-2:19.2.3-47.gc24117fd552.el9.cl 6/113 2026-03-06T13:43:39.861 INFO:teuthology.orchestra.run.vm00.stdout: Running scriptlet: ceph-volume-2:19.2.3-47.gc24117fd552.el9.clyso.n 12/113 2026-03-06T13:43:39.861 INFO:teuthology.orchestra.run.vm00.stdout:Glob pattern passed to enable, but globs are not supported for this. 2026-03-06T13:43:39.861 INFO:teuthology.orchestra.run.vm00.stdout:Invalid unit name "ceph-volume@*.service" escaped as "ceph-volume@\x2a.service". 2026-03-06T13:43:39.862 INFO:teuthology.orchestra.run.vm00.stdout: 2026-03-06T13:43:39.863 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : python3-cheroot-10.0.1-4.el9.noarch 8/113 2026-03-06T13:43:39.871 INFO:teuthology.orchestra.run.vm00.stdout: Erasing : ceph-volume-2:19.2.3-47.gc24117fd552.el9.clyso.n 12/113 2026-03-06T13:43:39.872 INFO:teuthology.orchestra.run.vm08.stdout: Running scriptlet: ceph-mgr-cephadm-2:19.2.3-47.gc24117fd552.el9.cl 6/113 2026-03-06T13:43:39.872 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : python3-grpcio-tools-1.46.7-10.el9.x86_64 9/113 2026-03-06T13:43:39.876 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : python3-grpcio-1.46.7-10.el9.x86_64 10/113 2026-03-06T13:43:39.878 INFO:teuthology.orchestra.run.vm08.stdout: Erasing : python3-cherrypy-18.6.1-2.el9.noarch 7/113 2026-03-06T13:43:39.882 INFO:teuthology.orchestra.run.vm08.stdout: Erasing : python3-cheroot-10.0.1-4.el9.noarch 8/113 2026-03-06T13:43:39.882 INFO:teuthology.orchestra.run.vm00.stdout: Running scriptlet: ceph-volume-2:19.2.3-47.gc24117fd552.el9.clyso.n 12/113 2026-03-06T13:43:39.885 INFO:teuthology.orchestra.run.vm00.stdout: Erasing : python3-jaraco-collections-3.0.0-8.el9.noarch 13/113 2026-03-06T13:43:39.890 INFO:teuthology.orchestra.run.vm00.stdout: Erasing : python3-jaraco-text-4.0.0-2.el9.noarch 14/113 2026-03-06T13:43:39.891 INFO:teuthology.orchestra.run.vm08.stdout: Erasing : python3-grpcio-tools-1.46.7-10.el9.x86_64 9/113 2026-03-06T13:43:39.899 INFO:teuthology.orchestra.run.vm01.stdout: Running scriptlet: ceph-osd-2:19.2.3-47.gc24117fd552.el9.clyso.x86_ 11/113 2026-03-06T13:43:39.899 INFO:teuthology.orchestra.run.vm01.stdout:Glob pattern passed to enable, but globs are not supported for this. 2026-03-06T13:43:39.899 INFO:teuthology.orchestra.run.vm01.stdout:Invalid unit name "ceph-osd@*.service" escaped as "ceph-osd@\x2a.service". 2026-03-06T13:43:39.899 INFO:teuthology.orchestra.run.vm01.stdout:Removed "/etc/systemd/system/multi-user.target.wants/ceph-osd.target". 2026-03-06T13:43:39.899 INFO:teuthology.orchestra.run.vm01.stdout:Removed "/etc/systemd/system/ceph.target.wants/ceph-osd.target". 2026-03-06T13:43:39.899 INFO:teuthology.orchestra.run.vm01.stdout: 2026-03-06T13:43:39.902 INFO:teuthology.orchestra.run.vm08.stdout: Erasing : python3-grpcio-1.46.7-10.el9.x86_64 10/113 2026-03-06T13:43:39.904 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : ceph-osd-2:19.2.3-47.gc24117fd552.el9.clyso.x86_ 11/113 2026-03-06T13:43:39.905 INFO:teuthology.orchestra.run.vm00.stdout: Erasing : python3-jinja2-2.11.3-8.el9.noarch 15/113 2026-03-06T13:43:39.915 INFO:teuthology.orchestra.run.vm01.stdout: Running scriptlet: ceph-osd-2:19.2.3-47.gc24117fd552.el9.clyso.x86_ 11/113 2026-03-06T13:43:39.917 INFO:teuthology.orchestra.run.vm00.stdout: Erasing : python3-saml-1.16.0-1.el9.noarch 16/113 2026-03-06T13:43:39.925 INFO:teuthology.orchestra.run.vm08.stdout: Running scriptlet: ceph-osd-2:19.2.3-47.gc24117fd552.el9.clyso.x86_ 11/113 2026-03-06T13:43:39.925 INFO:teuthology.orchestra.run.vm08.stdout:Glob pattern passed to enable, but globs are not supported for this. 2026-03-06T13:43:39.925 INFO:teuthology.orchestra.run.vm08.stdout:Invalid unit name "ceph-osd@*.service" escaped as "ceph-osd@\x2a.service". 2026-03-06T13:43:39.925 INFO:teuthology.orchestra.run.vm08.stdout:Removed "/etc/systemd/system/multi-user.target.wants/ceph-osd.target". 2026-03-06T13:43:39.925 INFO:teuthology.orchestra.run.vm08.stdout:Removed "/etc/systemd/system/ceph.target.wants/ceph-osd.target". 2026-03-06T13:43:39.925 INFO:teuthology.orchestra.run.vm08.stdout: 2026-03-06T13:43:39.931 INFO:teuthology.orchestra.run.vm08.stdout: Erasing : ceph-osd-2:19.2.3-47.gc24117fd552.el9.clyso.x86_ 11/113 2026-03-06T13:43:39.934 INFO:teuthology.orchestra.run.vm01.stdout: Running scriptlet: ceph-volume-2:19.2.3-47.gc24117fd552.el9.clyso.n 12/113 2026-03-06T13:43:39.934 INFO:teuthology.orchestra.run.vm01.stdout:Glob pattern passed to enable, but globs are not supported for this. 2026-03-06T13:43:39.934 INFO:teuthology.orchestra.run.vm01.stdout:Invalid unit name "ceph-volume@*.service" escaped as "ceph-volume@\x2a.service". 2026-03-06T13:43:39.934 INFO:teuthology.orchestra.run.vm01.stdout: 2026-03-06T13:43:39.943 INFO:teuthology.orchestra.run.vm08.stdout: Running scriptlet: ceph-osd-2:19.2.3-47.gc24117fd552.el9.clyso.x86_ 11/113 2026-03-06T13:43:39.944 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : ceph-volume-2:19.2.3-47.gc24117fd552.el9.clyso.n 12/113 2026-03-06T13:43:39.953 INFO:teuthology.orchestra.run.vm01.stdout: Running scriptlet: ceph-volume-2:19.2.3-47.gc24117fd552.el9.clyso.n 12/113 2026-03-06T13:43:39.955 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : python3-jaraco-collections-3.0.0-8.el9.noarch 13/113 2026-03-06T13:43:39.959 INFO:teuthology.orchestra.run.vm00.stdout: Erasing : python3-pecan-1.4.2-3.el9.noarch 17/113 2026-03-06T13:43:39.960 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : python3-jaraco-text-4.0.0-2.el9.noarch 14/113 2026-03-06T13:43:39.963 INFO:teuthology.orchestra.run.vm08.stdout: Running scriptlet: ceph-volume-2:19.2.3-47.gc24117fd552.el9.clyso.n 12/113 2026-03-06T13:43:39.963 INFO:teuthology.orchestra.run.vm08.stdout:Glob pattern passed to enable, but globs are not supported for this. 2026-03-06T13:43:39.963 INFO:teuthology.orchestra.run.vm08.stdout:Invalid unit name "ceph-volume@*.service" escaped as "ceph-volume@\x2a.service". 2026-03-06T13:43:39.963 INFO:teuthology.orchestra.run.vm08.stdout: 2026-03-06T13:43:39.966 INFO:teuthology.orchestra.run.vm00.stdout: Erasing : python3-babel-2.9.1-2.el9.noarch 18/113 2026-03-06T13:43:39.966 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : python3-jinja2-2.11.3-8.el9.noarch 15/113 2026-03-06T13:43:39.969 INFO:teuthology.orchestra.run.vm00.stdout: Erasing : python3-jaraco-classes-3.2.1-5.el9.noarch 19/113 2026-03-06T13:43:39.971 INFO:teuthology.orchestra.run.vm08.stdout: Erasing : ceph-volume-2:19.2.3-47.gc24117fd552.el9.clyso.n 12/113 2026-03-06T13:43:39.977 INFO:teuthology.orchestra.run.vm00.stdout: Erasing : python3-pyOpenSSL-21.0.0-1.el9.noarch 20/113 2026-03-06T13:43:39.977 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : python3-saml-1.16.0-1.el9.noarch 16/113 2026-03-06T13:43:39.981 INFO:teuthology.orchestra.run.vm08.stdout: Running scriptlet: ceph-volume-2:19.2.3-47.gc24117fd552.el9.clyso.n 12/113 2026-03-06T13:43:39.984 INFO:teuthology.orchestra.run.vm08.stdout: Erasing : python3-jaraco-collections-3.0.0-8.el9.noarch 13/113 2026-03-06T13:43:39.989 INFO:teuthology.orchestra.run.vm00.stdout: Erasing : python3-asyncssh-2.13.2-5.el9.noarch 21/113 2026-03-06T13:43:39.989 INFO:teuthology.orchestra.run.vm08.stdout: Erasing : python3-jaraco-text-4.0.0-2.el9.noarch 14/113 2026-03-06T13:43:39.989 INFO:teuthology.orchestra.run.vm00.stdout: Erasing : ceph-mgr-diskprediction-local-2:19.2.3-47.gc2411 22/113 2026-03-06T13:43:39.993 INFO:teuthology.orchestra.run.vm08.stdout: Erasing : python3-jinja2-2.11.3-8.el9.noarch 15/113 2026-03-06T13:43:39.998 INFO:teuthology.orchestra.run.vm00.stdout: Running scriptlet: ceph-mgr-diskprediction-local-2:19.2.3-47.gc2411 22/113 2026-03-06T13:43:40.005 INFO:teuthology.orchestra.run.vm08.stdout: Erasing : python3-saml-1.16.0-1.el9.noarch 16/113 2026-03-06T13:43:40.009 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : python3-pecan-1.4.2-3.el9.noarch 17/113 2026-03-06T13:43:40.016 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : python3-babel-2.9.1-2.el9.noarch 18/113 2026-03-06T13:43:40.019 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : python3-jaraco-classes-3.2.1-5.el9.noarch 19/113 2026-03-06T13:43:40.027 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : python3-pyOpenSSL-21.0.0-1.el9.noarch 20/113 2026-03-06T13:43:40.037 INFO:teuthology.orchestra.run.vm08.stdout: Erasing : python3-pecan-1.4.2-3.el9.noarch 17/113 2026-03-06T13:43:40.039 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : python3-asyncssh-2.13.2-5.el9.noarch 21/113 2026-03-06T13:43:40.039 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : ceph-mgr-diskprediction-local-2:19.2.3-47.gc2411 22/113 2026-03-06T13:43:40.044 INFO:teuthology.orchestra.run.vm08.stdout: Erasing : python3-babel-2.9.1-2.el9.noarch 18/113 2026-03-06T13:43:40.047 INFO:teuthology.orchestra.run.vm08.stdout: Erasing : python3-jaraco-classes-3.2.1-5.el9.noarch 19/113 2026-03-06T13:43:40.048 INFO:teuthology.orchestra.run.vm01.stdout: Running scriptlet: ceph-mgr-diskprediction-local-2:19.2.3-47.gc2411 22/113 2026-03-06T13:43:40.057 INFO:teuthology.orchestra.run.vm08.stdout: Erasing : python3-pyOpenSSL-21.0.0-1.el9.noarch 20/113 2026-03-06T13:43:40.070 INFO:teuthology.orchestra.run.vm08.stdout: Erasing : python3-asyncssh-2.13.2-5.el9.noarch 21/113 2026-03-06T13:43:40.070 INFO:teuthology.orchestra.run.vm08.stdout: Erasing : ceph-mgr-diskprediction-local-2:19.2.3-47.gc2411 22/113 2026-03-06T13:43:40.078 INFO:teuthology.orchestra.run.vm08.stdout: Running scriptlet: ceph-mgr-diskprediction-local-2:19.2.3-47.gc2411 22/113 2026-03-06T13:43:40.098 INFO:teuthology.orchestra.run.vm00.stdout: Erasing : python3-jsonpatch-1.21-16.el9.noarch 23/113 2026-03-06T13:43:40.115 INFO:teuthology.orchestra.run.vm00.stdout: Erasing : python3-scipy-1.9.3-2.el9.x86_64 24/113 2026-03-06T13:43:40.122 INFO:teuthology.orchestra.run.vm00.stdout: Erasing : python3-xmlsec-1.3.13-1.el9.x86_64 25/113 2026-03-06T13:43:40.125 INFO:teuthology.orchestra.run.vm00.stdout: Erasing : python3-lxml-4.6.5-3.el9.x86_64 26/113 2026-03-06T13:43:40.141 INFO:teuthology.orchestra.run.vm00.stdout: Running scriptlet: libstoragemgmt-1.10.1-1.el9.x86_64 27/113 2026-03-06T13:43:40.141 INFO:teuthology.orchestra.run.vm00.stdout:Removed "/etc/systemd/system/multi-user.target.wants/libstoragemgmt.service". 2026-03-06T13:43:40.141 INFO:teuthology.orchestra.run.vm00.stdout: 2026-03-06T13:43:40.142 INFO:teuthology.orchestra.run.vm00.stdout: Erasing : libstoragemgmt-1.10.1-1.el9.x86_64 27/113 2026-03-06T13:43:40.147 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : python3-jsonpatch-1.21-16.el9.noarch 23/113 2026-03-06T13:43:40.162 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : python3-scipy-1.9.3-2.el9.x86_64 24/113 2026-03-06T13:43:40.169 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : python3-xmlsec-1.3.13-1.el9.x86_64 25/113 2026-03-06T13:43:40.170 INFO:teuthology.orchestra.run.vm00.stdout: Running scriptlet: libstoragemgmt-1.10.1-1.el9.x86_64 27/113 2026-03-06T13:43:40.173 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : python3-lxml-4.6.5-3.el9.x86_64 26/113 2026-03-06T13:43:40.175 INFO:teuthology.orchestra.run.vm00.stdout: Erasing : python3-numpy-f2py-1:1.23.5-2.el9.x86_64 28/113 2026-03-06T13:43:40.175 INFO:teuthology.orchestra.run.vm08.stdout: Erasing : python3-jsonpatch-1.21-16.el9.noarch 23/113 2026-03-06T13:43:40.177 INFO:teuthology.orchestra.run.vm00.stdout: Erasing : xmlsec1-openssl-1.2.29-13.el9.x86_64 29/113 2026-03-06T13:43:40.186 INFO:teuthology.orchestra.run.vm01.stdout: Running scriptlet: libstoragemgmt-1.10.1-1.el9.x86_64 27/113 2026-03-06T13:43:40.186 INFO:teuthology.orchestra.run.vm01.stdout:Removed "/etc/systemd/system/multi-user.target.wants/libstoragemgmt.service". 2026-03-06T13:43:40.186 INFO:teuthology.orchestra.run.vm01.stdout: 2026-03-06T13:43:40.187 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : libstoragemgmt-1.10.1-1.el9.x86_64 27/113 2026-03-06T13:43:40.192 INFO:teuthology.orchestra.run.vm08.stdout: Erasing : python3-scipy-1.9.3-2.el9.x86_64 24/113 2026-03-06T13:43:40.192 INFO:teuthology.orchestra.run.vm00.stdout: Erasing : xmlsec1-1.2.29-13.el9.x86_64 30/113 2026-03-06T13:43:40.197 INFO:teuthology.orchestra.run.vm00.stdout: Erasing : python3-cryptography-36.0.1-5.el9.x86_64 31/113 2026-03-06T13:43:40.200 INFO:teuthology.orchestra.run.vm00.stdout: Erasing : protobuf-compiler-3.14.0-17.el9.x86_64 32/113 2026-03-06T13:43:40.201 INFO:teuthology.orchestra.run.vm08.stdout: Erasing : python3-xmlsec-1.3.13-1.el9.x86_64 25/113 2026-03-06T13:43:40.202 INFO:teuthology.orchestra.run.vm00.stdout: Erasing : python3-bcrypt-3.2.2-1.el9.x86_64 33/113 2026-03-06T13:43:40.205 INFO:teuthology.orchestra.run.vm08.stdout: Erasing : python3-lxml-4.6.5-3.el9.x86_64 26/113 2026-03-06T13:43:40.212 INFO:teuthology.orchestra.run.vm01.stdout: Running scriptlet: libstoragemgmt-1.10.1-1.el9.x86_64 27/113 2026-03-06T13:43:40.216 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : python3-numpy-f2py-1:1.23.5-2.el9.x86_64 28/113 2026-03-06T13:43:40.218 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : xmlsec1-openssl-1.2.29-13.el9.x86_64 29/113 2026-03-06T13:43:40.220 INFO:teuthology.orchestra.run.vm08.stdout: Running scriptlet: libstoragemgmt-1.10.1-1.el9.x86_64 27/113 2026-03-06T13:43:40.220 INFO:teuthology.orchestra.run.vm08.stdout:Removed "/etc/systemd/system/multi-user.target.wants/libstoragemgmt.service". 2026-03-06T13:43:40.220 INFO:teuthology.orchestra.run.vm08.stdout: 2026-03-06T13:43:40.221 INFO:teuthology.orchestra.run.vm08.stdout: Erasing : libstoragemgmt-1.10.1-1.el9.x86_64 27/113 2026-03-06T13:43:40.227 INFO:teuthology.orchestra.run.vm00.stdout: Running scriptlet: rbd-mirror-2:19.2.3-47.gc24117fd552.el9.clyso.x8 34/113 2026-03-06T13:43:40.227 INFO:teuthology.orchestra.run.vm00.stdout:Glob pattern passed to enable, but globs are not supported for this. 2026-03-06T13:43:40.227 INFO:teuthology.orchestra.run.vm00.stdout:Invalid unit name "ceph-rbd-mirror@*.service" escaped as "ceph-rbd-mirror@\x2a.service". 2026-03-06T13:43:40.227 INFO:teuthology.orchestra.run.vm00.stdout:Removed "/etc/systemd/system/multi-user.target.wants/ceph-rbd-mirror.target". 2026-03-06T13:43:40.227 INFO:teuthology.orchestra.run.vm00.stdout:Removed "/etc/systemd/system/ceph.target.wants/ceph-rbd-mirror.target". 2026-03-06T13:43:40.227 INFO:teuthology.orchestra.run.vm00.stdout: 2026-03-06T13:43:40.229 INFO:teuthology.orchestra.run.vm00.stdout: Erasing : rbd-mirror-2:19.2.3-47.gc24117fd552.el9.clyso.x8 34/113 2026-03-06T13:43:40.231 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : xmlsec1-1.2.29-13.el9.x86_64 30/113 2026-03-06T13:43:40.237 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : python3-cryptography-36.0.1-5.el9.x86_64 31/113 2026-03-06T13:43:40.240 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : protobuf-compiler-3.14.0-17.el9.x86_64 32/113 2026-03-06T13:43:40.241 INFO:teuthology.orchestra.run.vm00.stdout: Running scriptlet: rbd-mirror-2:19.2.3-47.gc24117fd552.el9.clyso.x8 34/113 2026-03-06T13:43:40.242 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : python3-bcrypt-3.2.2-1.el9.x86_64 33/113 2026-03-06T13:43:40.245 INFO:teuthology.orchestra.run.vm00.stdout: Erasing : python3-mako-1.1.4-6.el9.noarch 35/113 2026-03-06T13:43:40.247 INFO:teuthology.orchestra.run.vm00.stdout: Erasing : python3-jaraco-context-6.0.1-3.el9.noarch 36/113 2026-03-06T13:43:40.248 INFO:teuthology.orchestra.run.vm08.stdout: Running scriptlet: libstoragemgmt-1.10.1-1.el9.x86_64 27/113 2026-03-06T13:43:40.250 INFO:teuthology.orchestra.run.vm00.stdout: Erasing : python3-portend-3.1.0-2.el9.noarch 37/113 2026-03-06T13:43:40.253 INFO:teuthology.orchestra.run.vm08.stdout: Erasing : python3-numpy-f2py-1:1.23.5-2.el9.x86_64 28/113 2026-03-06T13:43:40.253 INFO:teuthology.orchestra.run.vm00.stdout: Erasing : python3-tempora-5.0.0-2.el9.noarch 38/113 2026-03-06T13:43:40.255 INFO:teuthology.orchestra.run.vm08.stdout: Erasing : xmlsec1-openssl-1.2.29-13.el9.x86_64 29/113 2026-03-06T13:43:40.256 INFO:teuthology.orchestra.run.vm00.stdout: Erasing : python3-jaraco-functools-3.5.0-2.el9.noarch 39/113 2026-03-06T13:43:40.259 INFO:teuthology.orchestra.run.vm00.stdout: Erasing : python3-routes-2.5.1-5.el9.noarch 40/113 2026-03-06T13:43:40.259 INFO:teuthology.orchestra.run.vm00.stdout: Erasing : ceph-mgr-k8sevents-2:19.2.3-47.gc24117fd552.el9. 41/113 2026-03-06T13:43:40.266 INFO:teuthology.orchestra.run.vm01.stdout: Running scriptlet: rbd-mirror-2:19.2.3-47.gc24117fd552.el9.clyso.x8 34/113 2026-03-06T13:43:40.266 INFO:teuthology.orchestra.run.vm01.stdout:Glob pattern passed to enable, but globs are not supported for this. 2026-03-06T13:43:40.266 INFO:teuthology.orchestra.run.vm01.stdout:Invalid unit name "ceph-rbd-mirror@*.service" escaped as "ceph-rbd-mirror@\x2a.service". 2026-03-06T13:43:40.266 INFO:teuthology.orchestra.run.vm01.stdout:Removed "/etc/systemd/system/multi-user.target.wants/ceph-rbd-mirror.target". 2026-03-06T13:43:40.266 INFO:teuthology.orchestra.run.vm01.stdout:Removed "/etc/systemd/system/ceph.target.wants/ceph-rbd-mirror.target". 2026-03-06T13:43:40.266 INFO:teuthology.orchestra.run.vm01.stdout: 2026-03-06T13:43:40.267 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : rbd-mirror-2:19.2.3-47.gc24117fd552.el9.clyso.x8 34/113 2026-03-06T13:43:40.268 INFO:teuthology.orchestra.run.vm08.stdout: Erasing : xmlsec1-1.2.29-13.el9.x86_64 30/113 2026-03-06T13:43:40.273 INFO:teuthology.orchestra.run.vm08.stdout: Erasing : python3-cryptography-36.0.1-5.el9.x86_64 31/113 2026-03-06T13:43:40.275 INFO:teuthology.orchestra.run.vm08.stdout: Erasing : protobuf-compiler-3.14.0-17.el9.x86_64 32/113 2026-03-06T13:43:40.277 INFO:teuthology.orchestra.run.vm08.stdout: Erasing : python3-bcrypt-3.2.2-1.el9.x86_64 33/113 2026-03-06T13:43:40.281 INFO:teuthology.orchestra.run.vm01.stdout: Running scriptlet: rbd-mirror-2:19.2.3-47.gc24117fd552.el9.clyso.x8 34/113 2026-03-06T13:43:40.285 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : python3-mako-1.1.4-6.el9.noarch 35/113 2026-03-06T13:43:40.287 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : python3-jaraco-context-6.0.1-3.el9.noarch 36/113 2026-03-06T13:43:40.290 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : python3-portend-3.1.0-2.el9.noarch 37/113 2026-03-06T13:43:40.292 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : python3-tempora-5.0.0-2.el9.noarch 38/113 2026-03-06T13:43:40.295 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : python3-jaraco-functools-3.5.0-2.el9.noarch 39/113 2026-03-06T13:43:40.298 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : python3-routes-2.5.1-5.el9.noarch 40/113 2026-03-06T13:43:40.298 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : ceph-mgr-k8sevents-2:19.2.3-47.gc24117fd552.el9. 41/113 2026-03-06T13:43:40.301 INFO:teuthology.orchestra.run.vm08.stdout: Running scriptlet: rbd-mirror-2:19.2.3-47.gc24117fd552.el9.clyso.x8 34/113 2026-03-06T13:43:40.301 INFO:teuthology.orchestra.run.vm08.stdout:Glob pattern passed to enable, but globs are not supported for this. 2026-03-06T13:43:40.301 INFO:teuthology.orchestra.run.vm08.stdout:Invalid unit name "ceph-rbd-mirror@*.service" escaped as "ceph-rbd-mirror@\x2a.service". 2026-03-06T13:43:40.301 INFO:teuthology.orchestra.run.vm08.stdout:Removed "/etc/systemd/system/multi-user.target.wants/ceph-rbd-mirror.target". 2026-03-06T13:43:40.301 INFO:teuthology.orchestra.run.vm08.stdout:Removed "/etc/systemd/system/ceph.target.wants/ceph-rbd-mirror.target". 2026-03-06T13:43:40.301 INFO:teuthology.orchestra.run.vm08.stdout: 2026-03-06T13:43:40.302 INFO:teuthology.orchestra.run.vm08.stdout: Erasing : rbd-mirror-2:19.2.3-47.gc24117fd552.el9.clyso.x8 34/113 2026-03-06T13:43:40.315 INFO:teuthology.orchestra.run.vm08.stdout: Running scriptlet: rbd-mirror-2:19.2.3-47.gc24117fd552.el9.clyso.x8 34/113 2026-03-06T13:43:40.315 INFO:teuthology.orchestra.run.vm00.stdout: Running scriptlet: ceph-mgr-k8sevents-2:19.2.3-47.gc24117fd552.el9. 41/113 2026-03-06T13:43:40.319 INFO:teuthology.orchestra.run.vm08.stdout: Erasing : python3-mako-1.1.4-6.el9.noarch 35/113 2026-03-06T13:43:40.322 INFO:teuthology.orchestra.run.vm08.stdout: Erasing : python3-jaraco-context-6.0.1-3.el9.noarch 36/113 2026-03-06T13:43:40.326 INFO:teuthology.orchestra.run.vm00.stdout: Erasing : python3-kubernetes-1:26.1.0-3.el9.noarch 42/113 2026-03-06T13:43:40.330 INFO:teuthology.orchestra.run.vm08.stdout: Erasing : python3-portend-3.1.0-2.el9.noarch 37/113 2026-03-06T13:43:40.333 INFO:teuthology.orchestra.run.vm08.stdout: Erasing : python3-tempora-5.0.0-2.el9.noarch 38/113 2026-03-06T13:43:40.333 INFO:teuthology.orchestra.run.vm00.stdout: Erasing : python3-requests-oauthlib-1.3.0-12.el9.noarch 43/113 2026-03-06T13:43:40.337 INFO:teuthology.orchestra.run.vm08.stdout: Erasing : python3-jaraco-functools-3.5.0-2.el9.noarch 39/113 2026-03-06T13:43:40.340 INFO:teuthology.orchestra.run.vm08.stdout: Erasing : python3-routes-2.5.1-5.el9.noarch 40/113 2026-03-06T13:43:40.340 INFO:teuthology.orchestra.run.vm08.stdout: Erasing : ceph-mgr-k8sevents-2:19.2.3-47.gc24117fd552.el9. 41/113 2026-03-06T13:43:40.343 INFO:teuthology.orchestra.run.vm00.stdout: Erasing : python3-requests-2.25.1-10.el9.noarch 44/113 2026-03-06T13:43:40.347 INFO:teuthology.orchestra.run.vm00.stdout: Erasing : python3-google-auth-1:2.45.0-1.el9.noarch 45/113 2026-03-06T13:43:40.355 INFO:teuthology.orchestra.run.vm01.stdout: Running scriptlet: ceph-mgr-k8sevents-2:19.2.3-47.gc24117fd552.el9. 41/113 2026-03-06T13:43:40.359 INFO:teuthology.orchestra.run.vm00.stdout: Erasing : python3-rsa-4.9-2.el9.noarch 46/113 2026-03-06T13:43:40.366 INFO:teuthology.orchestra.run.vm00.stdout: Erasing : python3-pyasn1-modules-0.4.8-7.el9.noarch 47/113 2026-03-06T13:43:40.366 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : python3-kubernetes-1:26.1.0-3.el9.noarch 42/113 2026-03-06T13:43:40.370 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : python3-requests-oauthlib-1.3.0-12.el9.noarch 43/113 2026-03-06T13:43:40.371 INFO:teuthology.orchestra.run.vm00.stdout: Erasing : python3-urllib3-1.26.5-7.el9.noarch 48/113 2026-03-06T13:43:40.375 INFO:teuthology.orchestra.run.vm00.stdout: Erasing : python3-cffi-1.14.5-5.el9.x86_64 49/113 2026-03-06T13:43:40.379 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : python3-requests-2.25.1-10.el9.noarch 44/113 2026-03-06T13:43:40.384 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : python3-google-auth-1:2.45.0-1.el9.noarch 45/113 2026-03-06T13:43:40.395 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : python3-rsa-4.9-2.el9.noarch 46/113 2026-03-06T13:43:40.399 INFO:teuthology.orchestra.run.vm08.stdout: Running scriptlet: ceph-mgr-k8sevents-2:19.2.3-47.gc24117fd552.el9. 41/113 2026-03-06T13:43:40.401 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : python3-pyasn1-modules-0.4.8-7.el9.noarch 47/113 2026-03-06T13:43:40.405 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : python3-urllib3-1.26.5-7.el9.noarch 48/113 2026-03-06T13:43:40.410 INFO:teuthology.orchestra.run.vm08.stdout: Erasing : python3-kubernetes-1:26.1.0-3.el9.noarch 42/113 2026-03-06T13:43:40.410 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : python3-cffi-1.14.5-5.el9.x86_64 49/113 2026-03-06T13:43:40.414 INFO:teuthology.orchestra.run.vm08.stdout: Erasing : python3-requests-oauthlib-1.3.0-12.el9.noarch 43/113 2026-03-06T13:43:40.425 INFO:teuthology.orchestra.run.vm00.stdout: Erasing : python3-pycparser-2.20-6.el9.noarch 50/113 2026-03-06T13:43:40.425 INFO:teuthology.orchestra.run.vm08.stdout: Erasing : python3-requests-2.25.1-10.el9.noarch 44/113 2026-03-06T13:43:40.430 INFO:teuthology.orchestra.run.vm08.stdout: Erasing : python3-google-auth-1:2.45.0-1.el9.noarch 45/113 2026-03-06T13:43:40.437 INFO:teuthology.orchestra.run.vm00.stdout: Erasing : python3-numpy-1:1.23.5-2.el9.x86_64 51/113 2026-03-06T13:43:40.440 INFO:teuthology.orchestra.run.vm00.stdout: Erasing : flexiblas-netlib-3.0.4-9.el9.x86_64 52/113 2026-03-06T13:43:40.441 INFO:teuthology.orchestra.run.vm08.stdout: Erasing : python3-rsa-4.9-2.el9.noarch 46/113 2026-03-06T13:43:40.445 INFO:teuthology.orchestra.run.vm00.stdout: Erasing : flexiblas-openblas-openmp-3.0.4-9.el9.x86_64 53/113 2026-03-06T13:43:40.447 INFO:teuthology.orchestra.run.vm00.stdout: Erasing : openblas-openmp-0.3.29-1.el9.x86_64 54/113 2026-03-06T13:43:40.447 INFO:teuthology.orchestra.run.vm08.stdout: Erasing : python3-pyasn1-modules-0.4.8-7.el9.noarch 47/113 2026-03-06T13:43:40.450 INFO:teuthology.orchestra.run.vm00.stdout: Erasing : libgfortran-11.5.0-14.el9.x86_64 55/113 2026-03-06T13:43:40.452 INFO:teuthology.orchestra.run.vm08.stdout: Erasing : python3-urllib3-1.26.5-7.el9.noarch 48/113 2026-03-06T13:43:40.453 INFO:teuthology.orchestra.run.vm00.stdout: Erasing : python3-libstoragemgmt-1.10.1-1.el9.x86_64 56/113 2026-03-06T13:43:40.457 INFO:teuthology.orchestra.run.vm08.stdout: Erasing : python3-cffi-1.14.5-5.el9.x86_64 49/113 2026-03-06T13:43:40.461 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : python3-pycparser-2.20-6.el9.noarch 50/113 2026-03-06T13:43:40.473 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : python3-numpy-1:1.23.5-2.el9.x86_64 51/113 2026-03-06T13:43:40.476 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : flexiblas-netlib-3.0.4-9.el9.x86_64 52/113 2026-03-06T13:43:40.478 INFO:teuthology.orchestra.run.vm00.stdout: Running scriptlet: ceph-immutable-object-cache-2:19.2.3-47.gc24117f 57/113 2026-03-06T13:43:40.478 INFO:teuthology.orchestra.run.vm00.stdout:Glob pattern passed to enable, but globs are not supported for this. 2026-03-06T13:43:40.478 INFO:teuthology.orchestra.run.vm00.stdout:Invalid unit name "ceph-immutable-object-cache@*.service" escaped as "ceph-immutable-object-cache@\x2a.service". 2026-03-06T13:43:40.478 INFO:teuthology.orchestra.run.vm00.stdout: 2026-03-06T13:43:40.478 INFO:teuthology.orchestra.run.vm00.stdout: Erasing : ceph-immutable-object-cache-2:19.2.3-47.gc24117f 57/113 2026-03-06T13:43:40.481 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : flexiblas-openblas-openmp-3.0.4-9.el9.x86_64 53/113 2026-03-06T13:43:40.484 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : openblas-openmp-0.3.29-1.el9.x86_64 54/113 2026-03-06T13:43:40.487 INFO:teuthology.orchestra.run.vm00.stdout: Running scriptlet: ceph-immutable-object-cache-2:19.2.3-47.gc24117f 57/113 2026-03-06T13:43:40.488 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : libgfortran-11.5.0-14.el9.x86_64 55/113 2026-03-06T13:43:40.489 INFO:teuthology.orchestra.run.vm00.stdout: Erasing : openblas-0.3.29-1.el9.x86_64 58/113 2026-03-06T13:43:40.491 INFO:teuthology.orchestra.run.vm00.stdout: Erasing : flexiblas-3.0.4-9.el9.x86_64 59/113 2026-03-06T13:43:40.492 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : python3-libstoragemgmt-1.10.1-1.el9.x86_64 56/113 2026-03-06T13:43:40.494 INFO:teuthology.orchestra.run.vm00.stdout: Erasing : python3-ply-3.11-14.el9.noarch 60/113 2026-03-06T13:43:40.496 INFO:teuthology.orchestra.run.vm00.stdout: Erasing : python3-idna-2.10-7.el9.1.noarch 61/113 2026-03-06T13:43:40.502 INFO:teuthology.orchestra.run.vm00.stdout: Erasing : python3-pysocks-1.7.1-12.el9.noarch 62/113 2026-03-06T13:43:40.506 INFO:teuthology.orchestra.run.vm00.stdout: Erasing : python3-pyasn1-0.4.8-7.el9.noarch 63/113 2026-03-06T13:43:40.508 INFO:teuthology.orchestra.run.vm08.stdout: Erasing : python3-pycparser-2.20-6.el9.noarch 50/113 2026-03-06T13:43:40.512 INFO:teuthology.orchestra.run.vm00.stdout: Erasing : python3-cachetools-4.2.4-1.el9.noarch 64/113 2026-03-06T13:43:40.516 INFO:teuthology.orchestra.run.vm01.stdout: Running scriptlet: ceph-immutable-object-cache-2:19.2.3-47.gc24117f 57/113 2026-03-06T13:43:40.516 INFO:teuthology.orchestra.run.vm01.stdout:Glob pattern passed to enable, but globs are not supported for this. 2026-03-06T13:43:40.516 INFO:teuthology.orchestra.run.vm01.stdout:Invalid unit name "ceph-immutable-object-cache@*.service" escaped as "ceph-immutable-object-cache@\x2a.service". 2026-03-06T13:43:40.516 INFO:teuthology.orchestra.run.vm01.stdout: 2026-03-06T13:43:40.516 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : ceph-immutable-object-cache-2:19.2.3-47.gc24117f 57/113 2026-03-06T13:43:40.519 INFO:teuthology.orchestra.run.vm08.stdout: Erasing : python3-numpy-1:1.23.5-2.el9.x86_64 51/113 2026-03-06T13:43:40.521 INFO:teuthology.orchestra.run.vm00.stdout: Erasing : python3-chardet-4.0.0-5.el9.noarch 65/113 2026-03-06T13:43:40.523 INFO:teuthology.orchestra.run.vm08.stdout: Erasing : flexiblas-netlib-3.0.4-9.el9.x86_64 52/113 2026-03-06T13:43:40.524 INFO:teuthology.orchestra.run.vm01.stdout: Running scriptlet: ceph-immutable-object-cache-2:19.2.3-47.gc24117f 57/113 2026-03-06T13:43:40.526 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : openblas-0.3.29-1.el9.x86_64 58/113 2026-03-06T13:43:40.527 INFO:teuthology.orchestra.run.vm00.stdout: Erasing : python3-oauthlib-3.1.1-5.el9.noarch 66/113 2026-03-06T13:43:40.528 INFO:teuthology.orchestra.run.vm08.stdout: Erasing : flexiblas-openblas-openmp-3.0.4-9.el9.x86_64 53/113 2026-03-06T13:43:40.528 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : flexiblas-3.0.4-9.el9.x86_64 59/113 2026-03-06T13:43:40.530 INFO:teuthology.orchestra.run.vm00.stdout: Erasing : python3-websocket-client-1.2.3-2.el9.noarch 67/113 2026-03-06T13:43:40.530 INFO:teuthology.orchestra.run.vm08.stdout: Erasing : openblas-openmp-0.3.29-1.el9.x86_64 54/113 2026-03-06T13:43:40.531 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : python3-ply-3.11-14.el9.noarch 60/113 2026-03-06T13:43:40.533 INFO:teuthology.orchestra.run.vm00.stdout: Erasing : python3-certifi-2023.05.07-4.el9.noarch 68/113 2026-03-06T13:43:40.534 INFO:teuthology.orchestra.run.vm08.stdout: Erasing : libgfortran-11.5.0-14.el9.x86_64 55/113 2026-03-06T13:43:40.534 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : python3-idna-2.10-7.el9.1.noarch 61/113 2026-03-06T13:43:40.536 INFO:teuthology.orchestra.run.vm00.stdout: Erasing : python3-repoze-lru-0.7-16.el9.noarch 69/113 2026-03-06T13:43:40.537 INFO:teuthology.orchestra.run.vm08.stdout: Erasing : python3-libstoragemgmt-1.10.1-1.el9.x86_64 56/113 2026-03-06T13:43:40.538 INFO:teuthology.orchestra.run.vm00.stdout: Erasing : python3-jaraco-8.2.1-3.el9.noarch 70/113 2026-03-06T13:43:40.540 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : python3-pysocks-1.7.1-12.el9.noarch 62/113 2026-03-06T13:43:40.541 INFO:teuthology.orchestra.run.vm00.stdout: Erasing : python3-more-itertools-8.12.0-2.el9.noarch 71/113 2026-03-06T13:43:40.543 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : python3-pyasn1-0.4.8-7.el9.noarch 63/113 2026-03-06T13:43:40.544 INFO:teuthology.orchestra.run.vm00.stdout: Erasing : python3-toml-0.10.2-6.el9.noarch 72/113 2026-03-06T13:43:40.547 INFO:teuthology.orchestra.run.vm00.stdout: Erasing : python3-pytz-2021.1-5.el9.noarch 73/113 2026-03-06T13:43:40.549 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : python3-cachetools-4.2.4-1.el9.noarch 64/113 2026-03-06T13:43:40.555 INFO:teuthology.orchestra.run.vm00.stdout: Erasing : python3-backports-tarfile-1.2.0-1.el9.noarch 74/113 2026-03-06T13:43:40.558 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : python3-chardet-4.0.0-5.el9.noarch 65/113 2026-03-06T13:43:40.560 INFO:teuthology.orchestra.run.vm00.stdout: Erasing : python3-devel-3.9.25-3.el9.x86_64 75/113 2026-03-06T13:43:40.563 INFO:teuthology.orchestra.run.vm08.stdout: Running scriptlet: ceph-immutable-object-cache-2:19.2.3-47.gc24117f 57/113 2026-03-06T13:43:40.563 INFO:teuthology.orchestra.run.vm08.stdout:Glob pattern passed to enable, but globs are not supported for this. 2026-03-06T13:43:40.563 INFO:teuthology.orchestra.run.vm08.stdout:Invalid unit name "ceph-immutable-object-cache@*.service" escaped as "ceph-immutable-object-cache@\x2a.service". 2026-03-06T13:43:40.563 INFO:teuthology.orchestra.run.vm08.stdout: 2026-03-06T13:43:40.564 INFO:teuthology.orchestra.run.vm08.stdout: Erasing : ceph-immutable-object-cache-2:19.2.3-47.gc24117f 57/113 2026-03-06T13:43:40.564 INFO:teuthology.orchestra.run.vm00.stdout: Erasing : python3-jsonpointer-2.0-4.el9.noarch 76/113 2026-03-06T13:43:40.566 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : python3-oauthlib-3.1.1-5.el9.noarch 66/113 2026-03-06T13:43:40.567 INFO:teuthology.orchestra.run.vm00.stdout: Erasing : python3-typing-extensions-4.15.0-1.el9.noarch 77/113 2026-03-06T13:43:40.571 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : python3-websocket-client-1.2.3-2.el9.noarch 67/113 2026-03-06T13:43:40.572 INFO:teuthology.orchestra.run.vm08.stdout: Running scriptlet: ceph-immutable-object-cache-2:19.2.3-47.gc24117f 57/113 2026-03-06T13:43:40.573 INFO:teuthology.orchestra.run.vm00.stdout: Erasing : python3-logutils-0.3.5-21.el9.noarch 78/113 2026-03-06T13:43:40.573 INFO:teuthology.orchestra.run.vm08.stdout: Erasing : openblas-0.3.29-1.el9.x86_64 58/113 2026-03-06T13:43:40.574 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : python3-certifi-2023.05.07-4.el9.noarch 68/113 2026-03-06T13:43:40.576 INFO:teuthology.orchestra.run.vm08.stdout: Erasing : flexiblas-3.0.4-9.el9.x86_64 59/113 2026-03-06T13:43:40.577 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : python3-repoze-lru-0.7-16.el9.noarch 69/113 2026-03-06T13:43:40.579 INFO:teuthology.orchestra.run.vm00.stdout: Erasing : python3-webob-1.8.8-2.el9.noarch 79/113 2026-03-06T13:43:40.579 INFO:teuthology.orchestra.run.vm08.stdout: Erasing : python3-ply-3.11-14.el9.noarch 60/113 2026-03-06T13:43:40.580 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : python3-jaraco-8.2.1-3.el9.noarch 70/113 2026-03-06T13:43:40.582 INFO:teuthology.orchestra.run.vm00.stdout: Erasing : python3-isodate-0.6.1-3.el9.noarch 80/113 2026-03-06T13:43:40.582 INFO:teuthology.orchestra.run.vm08.stdout: Erasing : python3-idna-2.10-7.el9.1.noarch 61/113 2026-03-06T13:43:40.584 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : python3-more-itertools-8.12.0-2.el9.noarch 71/113 2026-03-06T13:43:40.587 INFO:teuthology.orchestra.run.vm00.stdout: Erasing : python3-autocommand-2.2.2-8.el9.noarch 81/113 2026-03-06T13:43:40.588 INFO:teuthology.orchestra.run.vm08.stdout: Erasing : python3-pysocks-1.7.1-12.el9.noarch 62/113 2026-03-06T13:43:40.588 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : python3-toml-0.10.2-6.el9.noarch 72/113 2026-03-06T13:43:40.590 INFO:teuthology.orchestra.run.vm00.stdout: Erasing : python3-packaging-20.9-5.el9.noarch 82/113 2026-03-06T13:43:40.592 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : python3-pytz-2021.1-5.el9.noarch 73/113 2026-03-06T13:43:40.592 INFO:teuthology.orchestra.run.vm08.stdout: Erasing : python3-pyasn1-0.4.8-7.el9.noarch 63/113 2026-03-06T13:43:40.597 INFO:teuthology.orchestra.run.vm00.stdout: Erasing : grpc-data-1.46.7-10.el9.noarch 83/113 2026-03-06T13:43:40.598 INFO:teuthology.orchestra.run.vm08.stdout: Erasing : python3-cachetools-4.2.4-1.el9.noarch 64/113 2026-03-06T13:43:40.602 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : python3-backports-tarfile-1.2.0-1.el9.noarch 74/113 2026-03-06T13:43:40.602 INFO:teuthology.orchestra.run.vm00.stdout: Erasing : python3-protobuf-3.14.0-17.el9.noarch 84/113 2026-03-06T13:43:40.607 INFO:teuthology.orchestra.run.vm00.stdout: Erasing : python3-zc-lockfile-2.0-10.el9.noarch 85/113 2026-03-06T13:43:40.607 INFO:teuthology.orchestra.run.vm08.stdout: Erasing : python3-chardet-4.0.0-5.el9.noarch 65/113 2026-03-06T13:43:40.608 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : python3-devel-3.9.25-3.el9.x86_64 75/113 2026-03-06T13:43:40.610 INFO:teuthology.orchestra.run.vm00.stdout: Erasing : python3-natsort-7.1.1-5.el9.noarch 86/113 2026-03-06T13:43:40.611 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : python3-jsonpointer-2.0-4.el9.noarch 76/113 2026-03-06T13:43:40.612 INFO:teuthology.orchestra.run.vm00.stdout: Erasing : ceph-grafana-dashboards-2:19.2.3-47.gc24117fd552 87/113 2026-03-06T13:43:40.614 INFO:teuthology.orchestra.run.vm08.stdout: Erasing : python3-oauthlib-3.1.1-5.el9.noarch 66/113 2026-03-06T13:43:40.615 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : python3-typing-extensions-4.15.0-1.el9.noarch 77/113 2026-03-06T13:43:40.618 INFO:teuthology.orchestra.run.vm00.stdout: Erasing : ceph-prometheus-alerts-2:19.2.3-47.gc24117fd552. 88/113 2026-03-06T13:43:40.618 INFO:teuthology.orchestra.run.vm08.stdout: Erasing : python3-websocket-client-1.2.3-2.el9.noarch 67/113 2026-03-06T13:43:40.621 INFO:teuthology.orchestra.run.vm08.stdout: Erasing : python3-certifi-2023.05.07-4.el9.noarch 68/113 2026-03-06T13:43:40.621 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : python3-logutils-0.3.5-21.el9.noarch 78/113 2026-03-06T13:43:40.622 INFO:teuthology.orchestra.run.vm00.stdout: Erasing : python3-werkzeug-2.0.3-3.el9.1.noarch 89/113 2026-03-06T13:43:40.624 INFO:teuthology.orchestra.run.vm08.stdout: Erasing : python3-repoze-lru-0.7-16.el9.noarch 69/113 2026-03-06T13:43:40.626 INFO:teuthology.orchestra.run.vm08.stdout: Erasing : python3-jaraco-8.2.1-3.el9.noarch 70/113 2026-03-06T13:43:40.627 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : python3-webob-1.8.8-2.el9.noarch 79/113 2026-03-06T13:43:40.629 INFO:teuthology.orchestra.run.vm08.stdout: Erasing : python3-more-itertools-8.12.0-2.el9.noarch 71/113 2026-03-06T13:43:40.631 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : python3-isodate-0.6.1-3.el9.noarch 80/113 2026-03-06T13:43:40.632 INFO:teuthology.orchestra.run.vm08.stdout: Erasing : python3-toml-0.10.2-6.el9.noarch 72/113 2026-03-06T13:43:40.635 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : python3-autocommand-2.2.2-8.el9.noarch 81/113 2026-03-06T13:43:40.636 INFO:teuthology.orchestra.run.vm08.stdout: Erasing : python3-pytz-2021.1-5.el9.noarch 73/113 2026-03-06T13:43:40.638 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : python3-packaging-20.9-5.el9.noarch 82/113 2026-03-06T13:43:40.644 INFO:teuthology.orchestra.run.vm08.stdout: Erasing : python3-backports-tarfile-1.2.0-1.el9.noarch 74/113 2026-03-06T13:43:40.644 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : grpc-data-1.46.7-10.el9.noarch 83/113 2026-03-06T13:43:40.648 INFO:teuthology.orchestra.run.vm00.stdout: Running scriptlet: ceph-base-2:19.2.3-47.gc24117fd552.el9.clyso.x86 90/113 2026-03-06T13:43:40.648 INFO:teuthology.orchestra.run.vm00.stdout:Removed "/etc/systemd/system/ceph.target.wants/ceph-crash.service". 2026-03-06T13:43:40.648 INFO:teuthology.orchestra.run.vm00.stdout: 2026-03-06T13:43:40.650 INFO:teuthology.orchestra.run.vm08.stdout: Erasing : python3-devel-3.9.25-3.el9.x86_64 75/113 2026-03-06T13:43:40.650 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : python3-protobuf-3.14.0-17.el9.noarch 84/113 2026-03-06T13:43:40.653 INFO:teuthology.orchestra.run.vm08.stdout: Erasing : python3-jsonpointer-2.0-4.el9.noarch 76/113 2026-03-06T13:43:40.654 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : python3-zc-lockfile-2.0-10.el9.noarch 85/113 2026-03-06T13:43:40.657 INFO:teuthology.orchestra.run.vm08.stdout: Erasing : python3-typing-extensions-4.15.0-1.el9.noarch 77/113 2026-03-06T13:43:40.658 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : python3-natsort-7.1.1-5.el9.noarch 86/113 2026-03-06T13:43:40.659 INFO:teuthology.orchestra.run.vm00.stdout: Erasing : ceph-base-2:19.2.3-47.gc24117fd552.el9.clyso.x86 90/113 2026-03-06T13:43:40.660 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : ceph-grafana-dashboards-2:19.2.3-47.gc24117fd552 87/113 2026-03-06T13:43:40.662 INFO:teuthology.orchestra.run.vm08.stdout: Erasing : python3-logutils-0.3.5-21.el9.noarch 78/113 2026-03-06T13:43:40.667 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : ceph-prometheus-alerts-2:19.2.3-47.gc24117fd552. 88/113 2026-03-06T13:43:40.668 INFO:teuthology.orchestra.run.vm08.stdout: Erasing : python3-webob-1.8.8-2.el9.noarch 79/113 2026-03-06T13:43:40.672 INFO:teuthology.orchestra.run.vm08.stdout: Erasing : python3-isodate-0.6.1-3.el9.noarch 80/113 2026-03-06T13:43:40.673 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : python3-werkzeug-2.0.3-3.el9.1.noarch 89/113 2026-03-06T13:43:40.676 INFO:teuthology.orchestra.run.vm08.stdout: Erasing : python3-autocommand-2.2.2-8.el9.noarch 81/113 2026-03-06T13:43:40.679 INFO:teuthology.orchestra.run.vm08.stdout: Erasing : python3-packaging-20.9-5.el9.noarch 82/113 2026-03-06T13:43:40.685 INFO:teuthology.orchestra.run.vm08.stdout: Erasing : grpc-data-1.46.7-10.el9.noarch 83/113 2026-03-06T13:43:40.689 INFO:teuthology.orchestra.run.vm08.stdout: Erasing : python3-protobuf-3.14.0-17.el9.noarch 84/113 2026-03-06T13:43:40.692 INFO:teuthology.orchestra.run.vm08.stdout: Erasing : python3-zc-lockfile-2.0-10.el9.noarch 85/113 2026-03-06T13:43:40.694 INFO:teuthology.orchestra.run.vm00.stdout: Running scriptlet: ceph-base-2:19.2.3-47.gc24117fd552.el9.clyso.x86 90/113 2026-03-06T13:43:40.694 INFO:teuthology.orchestra.run.vm00.stdout: Erasing : ceph-common-2:19.2.3-47.gc24117fd552.el9.clyso.x 91/113 2026-03-06T13:43:40.695 INFO:teuthology.orchestra.run.vm01.stdout: Running scriptlet: ceph-base-2:19.2.3-47.gc24117fd552.el9.clyso.x86 90/113 2026-03-06T13:43:40.695 INFO:teuthology.orchestra.run.vm01.stdout:Removed "/etc/systemd/system/ceph.target.wants/ceph-crash.service". 2026-03-06T13:43:40.695 INFO:teuthology.orchestra.run.vm01.stdout: 2026-03-06T13:43:40.695 INFO:teuthology.orchestra.run.vm08.stdout: Erasing : python3-natsort-7.1.1-5.el9.noarch 86/113 2026-03-06T13:43:40.697 INFO:teuthology.orchestra.run.vm08.stdout: Erasing : ceph-grafana-dashboards-2:19.2.3-47.gc24117fd552 87/113 2026-03-06T13:43:40.704 INFO:teuthology.orchestra.run.vm08.stdout: Erasing : ceph-prometheus-alerts-2:19.2.3-47.gc24117fd552. 88/113 2026-03-06T13:43:40.704 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : ceph-base-2:19.2.3-47.gc24117fd552.el9.clyso.x86 90/113 2026-03-06T13:43:40.708 INFO:teuthology.orchestra.run.vm08.stdout: Erasing : python3-werkzeug-2.0.3-3.el9.1.noarch 89/113 2026-03-06T13:43:40.709 INFO:teuthology.orchestra.run.vm00.stdout: Running scriptlet: ceph-common-2:19.2.3-47.gc24117fd552.el9.clyso.x 91/113 2026-03-06T13:43:40.714 INFO:teuthology.orchestra.run.vm00.stdout: Erasing : qatzip-libs-1.3.1-1.el9.x86_64 92/113 2026-03-06T13:43:40.718 INFO:teuthology.orchestra.run.vm00.stdout: Erasing : python3-ceph-common-2:19.2.3-47.gc24117fd552.el9 93/113 2026-03-06T13:43:40.720 INFO:teuthology.orchestra.run.vm00.stdout: Erasing : python3-prettytable-0.7.2-27.el9.noarch 94/113 2026-03-06T13:43:40.720 INFO:teuthology.orchestra.run.vm00.stdout: Erasing : ceph-selinux-2:19.2.3-47.gc24117fd552.el9.clyso. 95/113 2026-03-06T13:43:40.729 INFO:teuthology.orchestra.run.vm08.stdout: Running scriptlet: ceph-base-2:19.2.3-47.gc24117fd552.el9.clyso.x86 90/113 2026-03-06T13:43:40.729 INFO:teuthology.orchestra.run.vm08.stdout:Removed "/etc/systemd/system/ceph.target.wants/ceph-crash.service". 2026-03-06T13:43:40.729 INFO:teuthology.orchestra.run.vm08.stdout: 2026-03-06T13:43:40.732 INFO:teuthology.orchestra.run.vm01.stdout: Running scriptlet: ceph-base-2:19.2.3-47.gc24117fd552.el9.clyso.x86 90/113 2026-03-06T13:43:40.732 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : ceph-common-2:19.2.3-47.gc24117fd552.el9.clyso.x 91/113 2026-03-06T13:43:40.737 INFO:teuthology.orchestra.run.vm08.stdout: Erasing : ceph-base-2:19.2.3-47.gc24117fd552.el9.clyso.x86 90/113 2026-03-06T13:43:40.767 INFO:teuthology.orchestra.run.vm01.stdout: Running scriptlet: ceph-common-2:19.2.3-47.gc24117fd552.el9.clyso.x 91/113 2026-03-06T13:43:40.773 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : qatzip-libs-1.3.1-1.el9.x86_64 92/113 2026-03-06T13:43:40.776 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : python3-ceph-common-2:19.2.3-47.gc24117fd552.el9 93/113 2026-03-06T13:43:40.778 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : python3-prettytable-0.7.2-27.el9.noarch 94/113 2026-03-06T13:43:40.778 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : ceph-selinux-2:19.2.3-47.gc24117fd552.el9.clyso. 95/113 2026-03-06T13:43:40.782 INFO:teuthology.orchestra.run.vm08.stdout: Running scriptlet: ceph-base-2:19.2.3-47.gc24117fd552.el9.clyso.x86 90/113 2026-03-06T13:43:40.783 INFO:teuthology.orchestra.run.vm08.stdout: Erasing : ceph-common-2:19.2.3-47.gc24117fd552.el9.clyso.x 91/113 2026-03-06T13:43:40.796 INFO:teuthology.orchestra.run.vm08.stdout: Running scriptlet: ceph-common-2:19.2.3-47.gc24117fd552.el9.clyso.x 91/113 2026-03-06T13:43:40.801 INFO:teuthology.orchestra.run.vm08.stdout: Erasing : qatzip-libs-1.3.1-1.el9.x86_64 92/113 2026-03-06T13:43:40.804 INFO:teuthology.orchestra.run.vm08.stdout: Erasing : python3-ceph-common-2:19.2.3-47.gc24117fd552.el9 93/113 2026-03-06T13:43:40.806 INFO:teuthology.orchestra.run.vm08.stdout: Erasing : python3-prettytable-0.7.2-27.el9.noarch 94/113 2026-03-06T13:43:40.806 INFO:teuthology.orchestra.run.vm08.stdout: Erasing : ceph-selinux-2:19.2.3-47.gc24117fd552.el9.clyso. 95/113 2026-03-06T13:43:46.499 INFO:teuthology.orchestra.run.vm00.stdout: Running scriptlet: ceph-selinux-2:19.2.3-47.gc24117fd552.el9.clyso. 95/113 2026-03-06T13:43:46.499 INFO:teuthology.orchestra.run.vm00.stdout:skipping the directory /sys 2026-03-06T13:43:46.499 INFO:teuthology.orchestra.run.vm00.stdout:skipping the directory /proc 2026-03-06T13:43:46.499 INFO:teuthology.orchestra.run.vm00.stdout:skipping the directory /mnt 2026-03-06T13:43:46.499 INFO:teuthology.orchestra.run.vm00.stdout:skipping the directory /var/tmp 2026-03-06T13:43:46.499 INFO:teuthology.orchestra.run.vm00.stdout:skipping the directory /home 2026-03-06T13:43:46.499 INFO:teuthology.orchestra.run.vm00.stdout:skipping the directory /root 2026-03-06T13:43:46.499 INFO:teuthology.orchestra.run.vm00.stdout:skipping the directory /tmp 2026-03-06T13:43:46.499 INFO:teuthology.orchestra.run.vm00.stdout: 2026-03-06T13:43:46.509 INFO:teuthology.orchestra.run.vm00.stdout: Erasing : qatlib-25.08.0-2.el9.x86_64 96/113 2026-03-06T13:43:46.527 INFO:teuthology.orchestra.run.vm00.stdout: Running scriptlet: qatlib-service-25.08.0-2.el9.x86_64 97/113 2026-03-06T13:43:46.527 INFO:teuthology.orchestra.run.vm00.stdout: Erasing : qatlib-service-25.08.0-2.el9.x86_64 97/113 2026-03-06T13:43:46.535 INFO:teuthology.orchestra.run.vm00.stdout: Running scriptlet: qatlib-service-25.08.0-2.el9.x86_64 97/113 2026-03-06T13:43:46.538 INFO:teuthology.orchestra.run.vm00.stdout: Erasing : gperftools-libs-2.9.1-3.el9.x86_64 98/113 2026-03-06T13:43:46.541 INFO:teuthology.orchestra.run.vm00.stdout: Erasing : libunwind-1.6.2-1.el9.x86_64 99/113 2026-03-06T13:43:46.544 INFO:teuthology.orchestra.run.vm00.stdout: Erasing : pciutils-3.7.0-7.el9.x86_64 100/113 2026-03-06T13:43:46.546 INFO:teuthology.orchestra.run.vm00.stdout: Erasing : liboath-2.6.12-1.el9.x86_64 101/113 2026-03-06T13:43:46.546 INFO:teuthology.orchestra.run.vm00.stdout: Erasing : libradosstriper1-2:19.2.3-47.gc24117fd552.el9.cl 102/113 2026-03-06T13:43:46.560 INFO:teuthology.orchestra.run.vm00.stdout: Running scriptlet: libradosstriper1-2:19.2.3-47.gc24117fd552.el9.cl 102/113 2026-03-06T13:43:46.576 INFO:teuthology.orchestra.run.vm00.stdout: Running scriptlet: smartmontools-1:7.2-10.el9.x86_64 103/113 2026-03-06T13:43:46.577 INFO:teuthology.orchestra.run.vm00.stdout:Removed "/etc/systemd/system/multi-user.target.wants/smartd.service". 2026-03-06T13:43:46.577 INFO:teuthology.orchestra.run.vm00.stdout: 2026-03-06T13:43:46.578 INFO:teuthology.orchestra.run.vm00.stdout: Erasing : smartmontools-1:7.2-10.el9.x86_64 103/113 2026-03-06T13:43:46.599 INFO:teuthology.orchestra.run.vm00.stdout: Running scriptlet: smartmontools-1:7.2-10.el9.x86_64 103/113 2026-03-06T13:43:46.601 INFO:teuthology.orchestra.run.vm00.stdout: Erasing : ledmon-libs-1.1.0-3.el9.x86_64 104/113 2026-03-06T13:43:46.608 INFO:teuthology.orchestra.run.vm00.stdout: Erasing : libquadmath-11.5.0-14.el9.x86_64 105/113 2026-03-06T13:43:46.610 INFO:teuthology.orchestra.run.vm00.stdout: Erasing : python3-markupsafe-1.1.1-12.el9.x86_64 106/113 2026-03-06T13:43:46.613 INFO:teuthology.orchestra.run.vm00.stdout: Erasing : protobuf-3.14.0-17.el9.x86_64 107/113 2026-03-06T13:43:46.616 INFO:teuthology.orchestra.run.vm00.stdout: Erasing : libxslt-1.1.34-12.el9.x86_64 108/113 2026-03-06T13:43:46.622 INFO:teuthology.orchestra.run.vm00.stdout: Erasing : libconfig-1.7.2-9.el9.x86_64 109/113 2026-03-06T13:43:46.631 INFO:teuthology.orchestra.run.vm00.stdout: Erasing : cryptsetup-2.8.1-3.el9.x86_64 110/113 2026-03-06T13:43:46.636 INFO:teuthology.orchestra.run.vm00.stdout: Erasing : abseil-cpp-20211102.0-4.el9.x86_64 111/113 2026-03-06T13:43:46.639 INFO:teuthology.orchestra.run.vm00.stdout: Erasing : python3-msgpack-1.0.3-2.el9.x86_64 112/113 2026-03-06T13:43:46.639 INFO:teuthology.orchestra.run.vm00.stdout: Erasing : libcephsqlite-2:19.2.3-47.gc24117fd552.el9.clyso 113/113 2026-03-06T13:43:46.779 INFO:teuthology.orchestra.run.vm00.stdout: Running scriptlet: libcephsqlite-2:19.2.3-47.gc24117fd552.el9.clyso 113/113 2026-03-06T13:43:46.779 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : abseil-cpp-20211102.0-4.el9.x86_64 1/113 2026-03-06T13:43:46.779 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : ceph-base-2:19.2.3-47.gc24117fd552.el9.clyso.x86 2/113 2026-03-06T13:43:46.780 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : ceph-common-2:19.2.3-47.gc24117fd552.el9.clyso.x 3/113 2026-03-06T13:43:46.780 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : ceph-grafana-dashboards-2:19.2.3-47.gc24117fd552 4/113 2026-03-06T13:43:46.780 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : ceph-immutable-object-cache-2:19.2.3-47.gc24117f 5/113 2026-03-06T13:43:46.780 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : ceph-mgr-2:19.2.3-47.gc24117fd552.el9.clyso.x86_ 6/113 2026-03-06T13:43:46.780 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : ceph-mgr-cephadm-2:19.2.3-47.gc24117fd552.el9.cl 7/113 2026-03-06T13:43:46.780 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : ceph-mgr-dashboard-2:19.2.3-47.gc24117fd552.el9. 8/113 2026-03-06T13:43:46.780 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : ceph-mgr-diskprediction-local-2:19.2.3-47.gc2411 9/113 2026-03-06T13:43:46.782 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : ceph-mgr-k8sevents-2:19.2.3-47.gc24117fd552.el9. 10/113 2026-03-06T13:43:46.782 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : ceph-mgr-modules-core-2:19.2.3-47.gc24117fd552.e 11/113 2026-03-06T13:43:46.782 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : ceph-mgr-rook-2:19.2.3-47.gc24117fd552.el9.clyso 12/113 2026-03-06T13:43:46.782 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : ceph-osd-2:19.2.3-47.gc24117fd552.el9.clyso.x86_ 13/113 2026-03-06T13:43:46.782 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : ceph-prometheus-alerts-2:19.2.3-47.gc24117fd552. 14/113 2026-03-06T13:43:46.782 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : ceph-selinux-2:19.2.3-47.gc24117fd552.el9.clyso. 15/113 2026-03-06T13:43:46.782 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : ceph-volume-2:19.2.3-47.gc24117fd552.el9.clyso.n 16/113 2026-03-06T13:43:46.782 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : cryptsetup-2.8.1-3.el9.x86_64 17/113 2026-03-06T13:43:46.782 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : flexiblas-3.0.4-9.el9.x86_64 18/113 2026-03-06T13:43:46.782 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : flexiblas-netlib-3.0.4-9.el9.x86_64 19/113 2026-03-06T13:43:46.782 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : flexiblas-openblas-openmp-3.0.4-9.el9.x86_64 20/113 2026-03-06T13:43:46.782 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : gperftools-libs-2.9.1-3.el9.x86_64 21/113 2026-03-06T13:43:46.782 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : grpc-data-1.46.7-10.el9.noarch 22/113 2026-03-06T13:43:46.782 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : ledmon-libs-1.1.0-3.el9.x86_64 23/113 2026-03-06T13:43:46.782 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : libcephsqlite-2:19.2.3-47.gc24117fd552.el9.clyso 24/113 2026-03-06T13:43:46.782 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : libconfig-1.7.2-9.el9.x86_64 25/113 2026-03-06T13:43:46.782 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : libgfortran-11.5.0-14.el9.x86_64 26/113 2026-03-06T13:43:46.782 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : liboath-2.6.12-1.el9.x86_64 27/113 2026-03-06T13:43:46.782 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : libquadmath-11.5.0-14.el9.x86_64 28/113 2026-03-06T13:43:46.782 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : libradosstriper1-2:19.2.3-47.gc24117fd552.el9.cl 29/113 2026-03-06T13:43:46.782 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : libstoragemgmt-1.10.1-1.el9.x86_64 30/113 2026-03-06T13:43:46.782 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : libunwind-1.6.2-1.el9.x86_64 31/113 2026-03-06T13:43:46.782 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : libxslt-1.1.34-12.el9.x86_64 32/113 2026-03-06T13:43:46.782 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : openblas-0.3.29-1.el9.x86_64 33/113 2026-03-06T13:43:46.782 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : openblas-openmp-0.3.29-1.el9.x86_64 34/113 2026-03-06T13:43:46.782 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : pciutils-3.7.0-7.el9.x86_64 35/113 2026-03-06T13:43:46.782 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : protobuf-3.14.0-17.el9.x86_64 36/113 2026-03-06T13:43:46.782 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : protobuf-compiler-3.14.0-17.el9.x86_64 37/113 2026-03-06T13:43:46.782 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : python3-asyncssh-2.13.2-5.el9.noarch 38/113 2026-03-06T13:43:46.782 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : python3-autocommand-2.2.2-8.el9.noarch 39/113 2026-03-06T13:43:46.782 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : python3-babel-2.9.1-2.el9.noarch 40/113 2026-03-06T13:43:46.782 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : python3-backports-tarfile-1.2.0-1.el9.noarch 41/113 2026-03-06T13:43:46.782 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : python3-bcrypt-3.2.2-1.el9.x86_64 42/113 2026-03-06T13:43:46.782 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : python3-cachetools-4.2.4-1.el9.noarch 43/113 2026-03-06T13:43:46.782 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : python3-ceph-common-2:19.2.3-47.gc24117fd552.el9 44/113 2026-03-06T13:43:46.782 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : python3-certifi-2023.05.07-4.el9.noarch 45/113 2026-03-06T13:43:46.782 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : python3-cffi-1.14.5-5.el9.x86_64 46/113 2026-03-06T13:43:46.782 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : python3-chardet-4.0.0-5.el9.noarch 47/113 2026-03-06T13:43:46.782 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : python3-cheroot-10.0.1-4.el9.noarch 48/113 2026-03-06T13:43:46.782 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : python3-cherrypy-18.6.1-2.el9.noarch 49/113 2026-03-06T13:43:46.782 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : python3-cryptography-36.0.1-5.el9.x86_64 50/113 2026-03-06T13:43:46.782 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : python3-devel-3.9.25-3.el9.x86_64 51/113 2026-03-06T13:43:46.782 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : python3-google-auth-1:2.45.0-1.el9.noarch 52/113 2026-03-06T13:43:46.782 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : python3-grpcio-1.46.7-10.el9.x86_64 53/113 2026-03-06T13:43:46.782 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : python3-grpcio-tools-1.46.7-10.el9.x86_64 54/113 2026-03-06T13:43:46.783 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : python3-idna-2.10-7.el9.1.noarch 55/113 2026-03-06T13:43:46.783 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : python3-influxdb-5.3.1-1.el9.noarch 56/113 2026-03-06T13:43:46.783 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : python3-isodate-0.6.1-3.el9.noarch 57/113 2026-03-06T13:43:46.783 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : python3-jaraco-8.2.1-3.el9.noarch 58/113 2026-03-06T13:43:46.783 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : python3-jaraco-classes-3.2.1-5.el9.noarch 59/113 2026-03-06T13:43:46.783 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : python3-jaraco-collections-3.0.0-8.el9.noarch 60/113 2026-03-06T13:43:46.783 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : python3-jaraco-context-6.0.1-3.el9.noarch 61/113 2026-03-06T13:43:46.783 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : python3-jaraco-functools-3.5.0-2.el9.noarch 62/113 2026-03-06T13:43:46.783 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : python3-jaraco-text-4.0.0-2.el9.noarch 63/113 2026-03-06T13:43:46.783 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : python3-jinja2-2.11.3-8.el9.noarch 64/113 2026-03-06T13:43:46.783 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : python3-jsonpatch-1.21-16.el9.noarch 65/113 2026-03-06T13:43:46.783 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : python3-jsonpointer-2.0-4.el9.noarch 66/113 2026-03-06T13:43:46.783 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : python3-kubernetes-1:26.1.0-3.el9.noarch 67/113 2026-03-06T13:43:46.783 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : python3-libstoragemgmt-1.10.1-1.el9.x86_64 68/113 2026-03-06T13:43:46.783 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : python3-logutils-0.3.5-21.el9.noarch 69/113 2026-03-06T13:43:46.783 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : python3-lxml-4.6.5-3.el9.x86_64 70/113 2026-03-06T13:43:46.783 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : python3-mako-1.1.4-6.el9.noarch 71/113 2026-03-06T13:43:46.783 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : python3-markupsafe-1.1.1-12.el9.x86_64 72/113 2026-03-06T13:43:46.783 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : python3-more-itertools-8.12.0-2.el9.noarch 73/113 2026-03-06T13:43:46.783 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : python3-msgpack-1.0.3-2.el9.x86_64 74/113 2026-03-06T13:43:46.783 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : python3-natsort-7.1.1-5.el9.noarch 75/113 2026-03-06T13:43:46.783 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : python3-numpy-1:1.23.5-2.el9.x86_64 76/113 2026-03-06T13:43:46.783 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : python3-numpy-f2py-1:1.23.5-2.el9.x86_64 77/113 2026-03-06T13:43:46.784 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : python3-oauthlib-3.1.1-5.el9.noarch 78/113 2026-03-06T13:43:46.784 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : python3-packaging-20.9-5.el9.noarch 79/113 2026-03-06T13:43:46.784 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : python3-pecan-1.4.2-3.el9.noarch 80/113 2026-03-06T13:43:46.784 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : python3-ply-3.11-14.el9.noarch 81/113 2026-03-06T13:43:46.784 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : python3-portend-3.1.0-2.el9.noarch 82/113 2026-03-06T13:43:46.784 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : python3-prettytable-0.7.2-27.el9.noarch 83/113 2026-03-06T13:43:46.784 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : python3-protobuf-3.14.0-17.el9.noarch 84/113 2026-03-06T13:43:46.784 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : python3-pyOpenSSL-21.0.0-1.el9.noarch 85/113 2026-03-06T13:43:46.784 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : python3-pyasn1-0.4.8-7.el9.noarch 86/113 2026-03-06T13:43:46.784 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : python3-pyasn1-modules-0.4.8-7.el9.noarch 87/113 2026-03-06T13:43:46.784 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : python3-pycparser-2.20-6.el9.noarch 88/113 2026-03-06T13:43:46.784 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : python3-pysocks-1.7.1-12.el9.noarch 89/113 2026-03-06T13:43:46.784 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : python3-pytz-2021.1-5.el9.noarch 90/113 2026-03-06T13:43:46.784 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : python3-repoze-lru-0.7-16.el9.noarch 91/113 2026-03-06T13:43:46.784 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : python3-requests-2.25.1-10.el9.noarch 92/113 2026-03-06T13:43:46.784 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : python3-requests-oauthlib-1.3.0-12.el9.noarch 93/113 2026-03-06T13:43:46.784 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : python3-routes-2.5.1-5.el9.noarch 94/113 2026-03-06T13:43:46.784 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : python3-rsa-4.9-2.el9.noarch 95/113 2026-03-06T13:43:46.784 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : python3-saml-1.16.0-1.el9.noarch 96/113 2026-03-06T13:43:46.784 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : python3-scipy-1.9.3-2.el9.x86_64 97/113 2026-03-06T13:43:46.784 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : python3-tempora-5.0.0-2.el9.noarch 98/113 2026-03-06T13:43:46.784 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : python3-toml-0.10.2-6.el9.noarch 99/113 2026-03-06T13:43:46.784 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : python3-typing-extensions-4.15.0-1.el9.noarch 100/113 2026-03-06T13:43:46.784 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : python3-urllib3-1.26.5-7.el9.noarch 101/113 2026-03-06T13:43:46.784 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : python3-webob-1.8.8-2.el9.noarch 102/113 2026-03-06T13:43:46.784 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : python3-websocket-client-1.2.3-2.el9.noarch 103/113 2026-03-06T13:43:46.784 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : python3-werkzeug-2.0.3-3.el9.1.noarch 104/113 2026-03-06T13:43:46.784 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : python3-xmlsec-1.3.13-1.el9.x86_64 105/113 2026-03-06T13:43:46.784 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : python3-zc-lockfile-2.0-10.el9.noarch 106/113 2026-03-06T13:43:46.784 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : qatlib-25.08.0-2.el9.x86_64 107/113 2026-03-06T13:43:46.784 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : qatlib-service-25.08.0-2.el9.x86_64 108/113 2026-03-06T13:43:46.784 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : qatzip-libs-1.3.1-1.el9.x86_64 109/113 2026-03-06T13:43:46.784 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : rbd-mirror-2:19.2.3-47.gc24117fd552.el9.clyso.x8 110/113 2026-03-06T13:43:46.784 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : smartmontools-1:7.2-10.el9.x86_64 111/113 2026-03-06T13:43:46.784 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : xmlsec1-1.2.29-13.el9.x86_64 112/113 2026-03-06T13:43:46.887 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : xmlsec1-openssl-1.2.29-13.el9.x86_64 113/113 2026-03-06T13:43:46.887 INFO:teuthology.orchestra.run.vm00.stdout: 2026-03-06T13:43:46.887 INFO:teuthology.orchestra.run.vm00.stdout:Removed: 2026-03-06T13:43:46.887 INFO:teuthology.orchestra.run.vm00.stdout: abseil-cpp-20211102.0-4.el9.x86_64 2026-03-06T13:43:46.887 INFO:teuthology.orchestra.run.vm00.stdout: ceph-base-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 2026-03-06T13:43:46.887 INFO:teuthology.orchestra.run.vm00.stdout: ceph-common-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 2026-03-06T13:43:46.887 INFO:teuthology.orchestra.run.vm00.stdout: ceph-grafana-dashboards-2:19.2.3-47.gc24117fd552.el9.clyso.noarch 2026-03-06T13:43:46.887 INFO:teuthology.orchestra.run.vm00.stdout: ceph-immutable-object-cache-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 2026-03-06T13:43:46.887 INFO:teuthology.orchestra.run.vm00.stdout: ceph-mgr-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 2026-03-06T13:43:46.887 INFO:teuthology.orchestra.run.vm00.stdout: ceph-mgr-cephadm-2:19.2.3-47.gc24117fd552.el9.clyso.noarch 2026-03-06T13:43:46.887 INFO:teuthology.orchestra.run.vm00.stdout: ceph-mgr-dashboard-2:19.2.3-47.gc24117fd552.el9.clyso.noarch 2026-03-06T13:43:46.887 INFO:teuthology.orchestra.run.vm00.stdout: ceph-mgr-diskprediction-local-2:19.2.3-47.gc24117fd552.el9.clyso.noarch 2026-03-06T13:43:46.887 INFO:teuthology.orchestra.run.vm00.stdout: ceph-mgr-k8sevents-2:19.2.3-47.gc24117fd552.el9.clyso.noarch 2026-03-06T13:43:46.887 INFO:teuthology.orchestra.run.vm00.stdout: ceph-mgr-modules-core-2:19.2.3-47.gc24117fd552.el9.clyso.noarch 2026-03-06T13:43:46.887 INFO:teuthology.orchestra.run.vm00.stdout: ceph-mgr-rook-2:19.2.3-47.gc24117fd552.el9.clyso.noarch 2026-03-06T13:43:46.887 INFO:teuthology.orchestra.run.vm00.stdout: ceph-osd-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 2026-03-06T13:43:46.887 INFO:teuthology.orchestra.run.vm00.stdout: ceph-prometheus-alerts-2:19.2.3-47.gc24117fd552.el9.clyso.noarch 2026-03-06T13:43:46.887 INFO:teuthology.orchestra.run.vm00.stdout: ceph-selinux-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 2026-03-06T13:43:46.887 INFO:teuthology.orchestra.run.vm00.stdout: ceph-volume-2:19.2.3-47.gc24117fd552.el9.clyso.noarch 2026-03-06T13:43:46.887 INFO:teuthology.orchestra.run.vm00.stdout: cryptsetup-2.8.1-3.el9.x86_64 2026-03-06T13:43:46.887 INFO:teuthology.orchestra.run.vm00.stdout: flexiblas-3.0.4-9.el9.x86_64 2026-03-06T13:43:46.887 INFO:teuthology.orchestra.run.vm00.stdout: flexiblas-netlib-3.0.4-9.el9.x86_64 2026-03-06T13:43:46.887 INFO:teuthology.orchestra.run.vm00.stdout: flexiblas-openblas-openmp-3.0.4-9.el9.x86_64 2026-03-06T13:43:46.887 INFO:teuthology.orchestra.run.vm00.stdout: gperftools-libs-2.9.1-3.el9.x86_64 2026-03-06T13:43:46.887 INFO:teuthology.orchestra.run.vm00.stdout: grpc-data-1.46.7-10.el9.noarch 2026-03-06T13:43:46.887 INFO:teuthology.orchestra.run.vm00.stdout: ledmon-libs-1.1.0-3.el9.x86_64 2026-03-06T13:43:46.888 INFO:teuthology.orchestra.run.vm00.stdout: libcephsqlite-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 2026-03-06T13:43:46.888 INFO:teuthology.orchestra.run.vm00.stdout: libconfig-1.7.2-9.el9.x86_64 2026-03-06T13:43:46.888 INFO:teuthology.orchestra.run.vm00.stdout: libgfortran-11.5.0-14.el9.x86_64 2026-03-06T13:43:46.888 INFO:teuthology.orchestra.run.vm00.stdout: liboath-2.6.12-1.el9.x86_64 2026-03-06T13:43:46.888 INFO:teuthology.orchestra.run.vm00.stdout: libquadmath-11.5.0-14.el9.x86_64 2026-03-06T13:43:46.888 INFO:teuthology.orchestra.run.vm00.stdout: libradosstriper1-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 2026-03-06T13:43:46.888 INFO:teuthology.orchestra.run.vm00.stdout: libstoragemgmt-1.10.1-1.el9.x86_64 2026-03-06T13:43:46.888 INFO:teuthology.orchestra.run.vm00.stdout: libunwind-1.6.2-1.el9.x86_64 2026-03-06T13:43:46.888 INFO:teuthology.orchestra.run.vm00.stdout: libxslt-1.1.34-12.el9.x86_64 2026-03-06T13:43:46.888 INFO:teuthology.orchestra.run.vm00.stdout: openblas-0.3.29-1.el9.x86_64 2026-03-06T13:43:46.888 INFO:teuthology.orchestra.run.vm00.stdout: openblas-openmp-0.3.29-1.el9.x86_64 2026-03-06T13:43:46.888 INFO:teuthology.orchestra.run.vm00.stdout: pciutils-3.7.0-7.el9.x86_64 2026-03-06T13:43:46.888 INFO:teuthology.orchestra.run.vm00.stdout: protobuf-3.14.0-17.el9.x86_64 2026-03-06T13:43:46.888 INFO:teuthology.orchestra.run.vm00.stdout: protobuf-compiler-3.14.0-17.el9.x86_64 2026-03-06T13:43:46.888 INFO:teuthology.orchestra.run.vm00.stdout: python3-asyncssh-2.13.2-5.el9.noarch 2026-03-06T13:43:46.888 INFO:teuthology.orchestra.run.vm00.stdout: python3-autocommand-2.2.2-8.el9.noarch 2026-03-06T13:43:46.888 INFO:teuthology.orchestra.run.vm00.stdout: python3-babel-2.9.1-2.el9.noarch 2026-03-06T13:43:46.888 INFO:teuthology.orchestra.run.vm00.stdout: python3-backports-tarfile-1.2.0-1.el9.noarch 2026-03-06T13:43:46.888 INFO:teuthology.orchestra.run.vm00.stdout: python3-bcrypt-3.2.2-1.el9.x86_64 2026-03-06T13:43:46.888 INFO:teuthology.orchestra.run.vm00.stdout: python3-cachetools-4.2.4-1.el9.noarch 2026-03-06T13:43:46.888 INFO:teuthology.orchestra.run.vm00.stdout: python3-ceph-common-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 2026-03-06T13:43:46.888 INFO:teuthology.orchestra.run.vm00.stdout: python3-certifi-2023.05.07-4.el9.noarch 2026-03-06T13:43:46.888 INFO:teuthology.orchestra.run.vm00.stdout: python3-cffi-1.14.5-5.el9.x86_64 2026-03-06T13:43:46.888 INFO:teuthology.orchestra.run.vm00.stdout: python3-chardet-4.0.0-5.el9.noarch 2026-03-06T13:43:46.888 INFO:teuthology.orchestra.run.vm00.stdout: python3-cheroot-10.0.1-4.el9.noarch 2026-03-06T13:43:46.888 INFO:teuthology.orchestra.run.vm00.stdout: python3-cherrypy-18.6.1-2.el9.noarch 2026-03-06T13:43:46.888 INFO:teuthology.orchestra.run.vm00.stdout: python3-cryptography-36.0.1-5.el9.x86_64 2026-03-06T13:43:46.888 INFO:teuthology.orchestra.run.vm00.stdout: python3-devel-3.9.25-3.el9.x86_64 2026-03-06T13:43:46.888 INFO:teuthology.orchestra.run.vm00.stdout: python3-google-auth-1:2.45.0-1.el9.noarch 2026-03-06T13:43:46.888 INFO:teuthology.orchestra.run.vm00.stdout: python3-grpcio-1.46.7-10.el9.x86_64 2026-03-06T13:43:46.888 INFO:teuthology.orchestra.run.vm00.stdout: python3-grpcio-tools-1.46.7-10.el9.x86_64 2026-03-06T13:43:46.888 INFO:teuthology.orchestra.run.vm00.stdout: python3-idna-2.10-7.el9.1.noarch 2026-03-06T13:43:46.888 INFO:teuthology.orchestra.run.vm00.stdout: python3-influxdb-5.3.1-1.el9.noarch 2026-03-06T13:43:46.888 INFO:teuthology.orchestra.run.vm00.stdout: python3-isodate-0.6.1-3.el9.noarch 2026-03-06T13:43:46.888 INFO:teuthology.orchestra.run.vm00.stdout: python3-jaraco-8.2.1-3.el9.noarch 2026-03-06T13:43:46.888 INFO:teuthology.orchestra.run.vm00.stdout: python3-jaraco-classes-3.2.1-5.el9.noarch 2026-03-06T13:43:46.888 INFO:teuthology.orchestra.run.vm00.stdout: python3-jaraco-collections-3.0.0-8.el9.noarch 2026-03-06T13:43:46.888 INFO:teuthology.orchestra.run.vm00.stdout: python3-jaraco-context-6.0.1-3.el9.noarch 2026-03-06T13:43:46.888 INFO:teuthology.orchestra.run.vm00.stdout: python3-jaraco-functools-3.5.0-2.el9.noarch 2026-03-06T13:43:46.888 INFO:teuthology.orchestra.run.vm00.stdout: python3-jaraco-text-4.0.0-2.el9.noarch 2026-03-06T13:43:46.888 INFO:teuthology.orchestra.run.vm00.stdout: python3-jinja2-2.11.3-8.el9.noarch 2026-03-06T13:43:46.888 INFO:teuthology.orchestra.run.vm00.stdout: python3-jsonpatch-1.21-16.el9.noarch 2026-03-06T13:43:46.888 INFO:teuthology.orchestra.run.vm00.stdout: python3-jsonpointer-2.0-4.el9.noarch 2026-03-06T13:43:46.888 INFO:teuthology.orchestra.run.vm00.stdout: python3-kubernetes-1:26.1.0-3.el9.noarch 2026-03-06T13:43:46.888 INFO:teuthology.orchestra.run.vm00.stdout: python3-libstoragemgmt-1.10.1-1.el9.x86_64 2026-03-06T13:43:46.888 INFO:teuthology.orchestra.run.vm00.stdout: python3-logutils-0.3.5-21.el9.noarch 2026-03-06T13:43:46.888 INFO:teuthology.orchestra.run.vm00.stdout: python3-lxml-4.6.5-3.el9.x86_64 2026-03-06T13:43:46.888 INFO:teuthology.orchestra.run.vm00.stdout: python3-mako-1.1.4-6.el9.noarch 2026-03-06T13:43:46.888 INFO:teuthology.orchestra.run.vm00.stdout: python3-markupsafe-1.1.1-12.el9.x86_64 2026-03-06T13:43:46.888 INFO:teuthology.orchestra.run.vm00.stdout: python3-more-itertools-8.12.0-2.el9.noarch 2026-03-06T13:43:46.888 INFO:teuthology.orchestra.run.vm00.stdout: python3-msgpack-1.0.3-2.el9.x86_64 2026-03-06T13:43:46.888 INFO:teuthology.orchestra.run.vm00.stdout: python3-natsort-7.1.1-5.el9.noarch 2026-03-06T13:43:46.888 INFO:teuthology.orchestra.run.vm00.stdout: python3-numpy-1:1.23.5-2.el9.x86_64 2026-03-06T13:43:46.888 INFO:teuthology.orchestra.run.vm00.stdout: python3-numpy-f2py-1:1.23.5-2.el9.x86_64 2026-03-06T13:43:46.888 INFO:teuthology.orchestra.run.vm00.stdout: python3-oauthlib-3.1.1-5.el9.noarch 2026-03-06T13:43:46.888 INFO:teuthology.orchestra.run.vm00.stdout: python3-packaging-20.9-5.el9.noarch 2026-03-06T13:43:46.888 INFO:teuthology.orchestra.run.vm00.stdout: python3-pecan-1.4.2-3.el9.noarch 2026-03-06T13:43:46.888 INFO:teuthology.orchestra.run.vm00.stdout: python3-ply-3.11-14.el9.noarch 2026-03-06T13:43:46.889 INFO:teuthology.orchestra.run.vm00.stdout: python3-portend-3.1.0-2.el9.noarch 2026-03-06T13:43:46.889 INFO:teuthology.orchestra.run.vm00.stdout: python3-prettytable-0.7.2-27.el9.noarch 2026-03-06T13:43:46.889 INFO:teuthology.orchestra.run.vm00.stdout: python3-protobuf-3.14.0-17.el9.noarch 2026-03-06T13:43:46.889 INFO:teuthology.orchestra.run.vm00.stdout: python3-pyOpenSSL-21.0.0-1.el9.noarch 2026-03-06T13:43:46.889 INFO:teuthology.orchestra.run.vm00.stdout: python3-pyasn1-0.4.8-7.el9.noarch 2026-03-06T13:43:46.889 INFO:teuthology.orchestra.run.vm00.stdout: python3-pyasn1-modules-0.4.8-7.el9.noarch 2026-03-06T13:43:46.889 INFO:teuthology.orchestra.run.vm00.stdout: python3-pycparser-2.20-6.el9.noarch 2026-03-06T13:43:46.889 INFO:teuthology.orchestra.run.vm00.stdout: python3-pysocks-1.7.1-12.el9.noarch 2026-03-06T13:43:46.889 INFO:teuthology.orchestra.run.vm00.stdout: python3-pytz-2021.1-5.el9.noarch 2026-03-06T13:43:46.889 INFO:teuthology.orchestra.run.vm00.stdout: python3-repoze-lru-0.7-16.el9.noarch 2026-03-06T13:43:46.889 INFO:teuthology.orchestra.run.vm00.stdout: python3-requests-2.25.1-10.el9.noarch 2026-03-06T13:43:46.889 INFO:teuthology.orchestra.run.vm00.stdout: python3-requests-oauthlib-1.3.0-12.el9.noarch 2026-03-06T13:43:46.889 INFO:teuthology.orchestra.run.vm00.stdout: python3-routes-2.5.1-5.el9.noarch 2026-03-06T13:43:46.889 INFO:teuthology.orchestra.run.vm00.stdout: python3-rsa-4.9-2.el9.noarch 2026-03-06T13:43:46.889 INFO:teuthology.orchestra.run.vm00.stdout: python3-saml-1.16.0-1.el9.noarch 2026-03-06T13:43:46.889 INFO:teuthology.orchestra.run.vm00.stdout: python3-scipy-1.9.3-2.el9.x86_64 2026-03-06T13:43:46.889 INFO:teuthology.orchestra.run.vm00.stdout: python3-tempora-5.0.0-2.el9.noarch 2026-03-06T13:43:46.889 INFO:teuthology.orchestra.run.vm00.stdout: python3-toml-0.10.2-6.el9.noarch 2026-03-06T13:43:46.889 INFO:teuthology.orchestra.run.vm00.stdout: python3-typing-extensions-4.15.0-1.el9.noarch 2026-03-06T13:43:46.889 INFO:teuthology.orchestra.run.vm00.stdout: python3-urllib3-1.26.5-7.el9.noarch 2026-03-06T13:43:46.889 INFO:teuthology.orchestra.run.vm00.stdout: python3-webob-1.8.8-2.el9.noarch 2026-03-06T13:43:46.889 INFO:teuthology.orchestra.run.vm00.stdout: python3-websocket-client-1.2.3-2.el9.noarch 2026-03-06T13:43:46.889 INFO:teuthology.orchestra.run.vm00.stdout: python3-werkzeug-2.0.3-3.el9.1.noarch 2026-03-06T13:43:46.889 INFO:teuthology.orchestra.run.vm00.stdout: python3-xmlsec-1.3.13-1.el9.x86_64 2026-03-06T13:43:46.889 INFO:teuthology.orchestra.run.vm00.stdout: python3-zc-lockfile-2.0-10.el9.noarch 2026-03-06T13:43:46.889 INFO:teuthology.orchestra.run.vm00.stdout: qatlib-25.08.0-2.el9.x86_64 2026-03-06T13:43:46.889 INFO:teuthology.orchestra.run.vm00.stdout: qatlib-service-25.08.0-2.el9.x86_64 2026-03-06T13:43:46.889 INFO:teuthology.orchestra.run.vm00.stdout: qatzip-libs-1.3.1-1.el9.x86_64 2026-03-06T13:43:46.889 INFO:teuthology.orchestra.run.vm00.stdout: rbd-mirror-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 2026-03-06T13:43:46.889 INFO:teuthology.orchestra.run.vm00.stdout: smartmontools-1:7.2-10.el9.x86_64 2026-03-06T13:43:46.889 INFO:teuthology.orchestra.run.vm00.stdout: xmlsec1-1.2.29-13.el9.x86_64 2026-03-06T13:43:46.889 INFO:teuthology.orchestra.run.vm00.stdout: xmlsec1-openssl-1.2.29-13.el9.x86_64 2026-03-06T13:43:46.889 INFO:teuthology.orchestra.run.vm00.stdout: 2026-03-06T13:43:46.889 INFO:teuthology.orchestra.run.vm00.stdout:Complete! 2026-03-06T13:43:46.895 INFO:teuthology.orchestra.run.vm08.stdout: Running scriptlet: ceph-selinux-2:19.2.3-47.gc24117fd552.el9.clyso. 95/113 2026-03-06T13:43:46.895 INFO:teuthology.orchestra.run.vm08.stdout:skipping the directory /sys 2026-03-06T13:43:46.895 INFO:teuthology.orchestra.run.vm08.stdout:skipping the directory /proc 2026-03-06T13:43:46.895 INFO:teuthology.orchestra.run.vm08.stdout:skipping the directory /mnt 2026-03-06T13:43:46.895 INFO:teuthology.orchestra.run.vm08.stdout:skipping the directory /var/tmp 2026-03-06T13:43:46.895 INFO:teuthology.orchestra.run.vm08.stdout:skipping the directory /home 2026-03-06T13:43:46.895 INFO:teuthology.orchestra.run.vm08.stdout:skipping the directory /root 2026-03-06T13:43:46.895 INFO:teuthology.orchestra.run.vm08.stdout:skipping the directory /tmp 2026-03-06T13:43:46.895 INFO:teuthology.orchestra.run.vm08.stdout: 2026-03-06T13:43:46.908 INFO:teuthology.orchestra.run.vm08.stdout: Erasing : qatlib-25.08.0-2.el9.x86_64 96/113 2026-03-06T13:43:46.931 INFO:teuthology.orchestra.run.vm01.stdout: Running scriptlet: ceph-selinux-2:19.2.3-47.gc24117fd552.el9.clyso. 95/113 2026-03-06T13:43:46.931 INFO:teuthology.orchestra.run.vm01.stdout:skipping the directory /sys 2026-03-06T13:43:46.931 INFO:teuthology.orchestra.run.vm01.stdout:skipping the directory /proc 2026-03-06T13:43:46.931 INFO:teuthology.orchestra.run.vm01.stdout:skipping the directory /mnt 2026-03-06T13:43:46.931 INFO:teuthology.orchestra.run.vm01.stdout:skipping the directory /var/tmp 2026-03-06T13:43:46.931 INFO:teuthology.orchestra.run.vm01.stdout:skipping the directory /home 2026-03-06T13:43:46.931 INFO:teuthology.orchestra.run.vm01.stdout:skipping the directory /root 2026-03-06T13:43:46.931 INFO:teuthology.orchestra.run.vm01.stdout:skipping the directory /tmp 2026-03-06T13:43:46.931 INFO:teuthology.orchestra.run.vm01.stdout: 2026-03-06T13:43:46.932 INFO:teuthology.orchestra.run.vm08.stdout: Running scriptlet: qatlib-service-25.08.0-2.el9.x86_64 97/113 2026-03-06T13:43:46.932 INFO:teuthology.orchestra.run.vm08.stdout: Erasing : qatlib-service-25.08.0-2.el9.x86_64 97/113 2026-03-06T13:43:46.944 INFO:teuthology.orchestra.run.vm08.stdout: Running scriptlet: qatlib-service-25.08.0-2.el9.x86_64 97/113 2026-03-06T13:43:46.944 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : qatlib-25.08.0-2.el9.x86_64 96/113 2026-03-06T13:43:46.947 INFO:teuthology.orchestra.run.vm08.stdout: Erasing : gperftools-libs-2.9.1-3.el9.x86_64 98/113 2026-03-06T13:43:46.951 INFO:teuthology.orchestra.run.vm08.stdout: Erasing : libunwind-1.6.2-1.el9.x86_64 99/113 2026-03-06T13:43:46.954 INFO:teuthology.orchestra.run.vm08.stdout: Erasing : pciutils-3.7.0-7.el9.x86_64 100/113 2026-03-06T13:43:46.957 INFO:teuthology.orchestra.run.vm08.stdout: Erasing : liboath-2.6.12-1.el9.x86_64 101/113 2026-03-06T13:43:46.957 INFO:teuthology.orchestra.run.vm08.stdout: Erasing : libradosstriper1-2:19.2.3-47.gc24117fd552.el9.cl 102/113 2026-03-06T13:43:46.963 INFO:teuthology.orchestra.run.vm01.stdout: Running scriptlet: qatlib-service-25.08.0-2.el9.x86_64 97/113 2026-03-06T13:43:46.963 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : qatlib-service-25.08.0-2.el9.x86_64 97/113 2026-03-06T13:43:46.972 INFO:teuthology.orchestra.run.vm01.stdout: Running scriptlet: qatlib-service-25.08.0-2.el9.x86_64 97/113 2026-03-06T13:43:46.975 INFO:teuthology.orchestra.run.vm08.stdout: Running scriptlet: libradosstriper1-2:19.2.3-47.gc24117fd552.el9.cl 102/113 2026-03-06T13:43:46.976 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : gperftools-libs-2.9.1-3.el9.x86_64 98/113 2026-03-06T13:43:46.979 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : libunwind-1.6.2-1.el9.x86_64 99/113 2026-03-06T13:43:46.982 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : pciutils-3.7.0-7.el9.x86_64 100/113 2026-03-06T13:43:46.985 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : liboath-2.6.12-1.el9.x86_64 101/113 2026-03-06T13:43:46.985 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : libradosstriper1-2:19.2.3-47.gc24117fd552.el9.cl 102/113 2026-03-06T13:43:46.990 INFO:teuthology.orchestra.run.vm08.stdout: Running scriptlet: smartmontools-1:7.2-10.el9.x86_64 103/113 2026-03-06T13:43:46.990 INFO:teuthology.orchestra.run.vm08.stdout:Removed "/etc/systemd/system/multi-user.target.wants/smartd.service". 2026-03-06T13:43:46.990 INFO:teuthology.orchestra.run.vm08.stdout: 2026-03-06T13:43:46.992 INFO:teuthology.orchestra.run.vm08.stdout: Erasing : smartmontools-1:7.2-10.el9.x86_64 103/113 2026-03-06T13:43:46.998 INFO:teuthology.orchestra.run.vm01.stdout: Running scriptlet: libradosstriper1-2:19.2.3-47.gc24117fd552.el9.cl 102/113 2026-03-06T13:43:47.003 INFO:teuthology.orchestra.run.vm08.stdout: Running scriptlet: smartmontools-1:7.2-10.el9.x86_64 103/113 2026-03-06T13:43:47.006 INFO:teuthology.orchestra.run.vm08.stdout: Erasing : ledmon-libs-1.1.0-3.el9.x86_64 104/113 2026-03-06T13:43:47.009 INFO:teuthology.orchestra.run.vm08.stdout: Erasing : libquadmath-11.5.0-14.el9.x86_64 105/113 2026-03-06T13:43:47.010 INFO:teuthology.orchestra.run.vm01.stdout: Running scriptlet: smartmontools-1:7.2-10.el9.x86_64 103/113 2026-03-06T13:43:47.010 INFO:teuthology.orchestra.run.vm01.stdout:Removed "/etc/systemd/system/multi-user.target.wants/smartd.service". 2026-03-06T13:43:47.010 INFO:teuthology.orchestra.run.vm01.stdout: 2026-03-06T13:43:47.012 INFO:teuthology.orchestra.run.vm08.stdout: Erasing : python3-markupsafe-1.1.1-12.el9.x86_64 106/113 2026-03-06T13:43:47.012 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : smartmontools-1:7.2-10.el9.x86_64 103/113 2026-03-06T13:43:47.016 INFO:teuthology.orchestra.run.vm08.stdout: Erasing : protobuf-3.14.0-17.el9.x86_64 107/113 2026-03-06T13:43:47.022 INFO:teuthology.orchestra.run.vm08.stdout: Erasing : libxslt-1.1.34-12.el9.x86_64 108/113 2026-03-06T13:43:47.022 INFO:teuthology.orchestra.run.vm01.stdout: Running scriptlet: smartmontools-1:7.2-10.el9.x86_64 103/113 2026-03-06T13:43:47.025 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : ledmon-libs-1.1.0-3.el9.x86_64 104/113 2026-03-06T13:43:47.029 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : libquadmath-11.5.0-14.el9.x86_64 105/113 2026-03-06T13:43:47.029 INFO:teuthology.orchestra.run.vm08.stdout: Erasing : libconfig-1.7.2-9.el9.x86_64 109/113 2026-03-06T13:43:47.033 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : python3-markupsafe-1.1.1-12.el9.x86_64 106/113 2026-03-06T13:43:47.037 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : protobuf-3.14.0-17.el9.x86_64 107/113 2026-03-06T13:43:47.040 INFO:teuthology.orchestra.run.vm08.stdout: Erasing : cryptsetup-2.8.1-3.el9.x86_64 110/113 2026-03-06T13:43:47.041 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : libxslt-1.1.34-12.el9.x86_64 108/113 2026-03-06T13:43:47.046 INFO:teuthology.orchestra.run.vm08.stdout: Erasing : abseil-cpp-20211102.0-4.el9.x86_64 111/113 2026-03-06T13:43:47.048 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : libconfig-1.7.2-9.el9.x86_64 109/113 2026-03-06T13:43:47.049 INFO:teuthology.orchestra.run.vm08.stdout: Erasing : python3-msgpack-1.0.3-2.el9.x86_64 112/113 2026-03-06T13:43:47.049 INFO:teuthology.orchestra.run.vm08.stdout: Erasing : libcephsqlite-2:19.2.3-47.gc24117fd552.el9.clyso 113/113 2026-03-06T13:43:47.058 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : cryptsetup-2.8.1-3.el9.x86_64 110/113 2026-03-06T13:43:47.064 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : abseil-cpp-20211102.0-4.el9.x86_64 111/113 2026-03-06T13:43:47.067 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : python3-msgpack-1.0.3-2.el9.x86_64 112/113 2026-03-06T13:43:47.068 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : libcephsqlite-2:19.2.3-47.gc24117fd552.el9.clyso 113/113 2026-03-06T13:43:47.113 INFO:teuthology.orchestra.run.vm00.stdout:Dependencies resolved. 2026-03-06T13:43:47.113 INFO:teuthology.orchestra.run.vm00.stdout:================================================================================ 2026-03-06T13:43:47.113 INFO:teuthology.orchestra.run.vm00.stdout: Package Arch Version Repository Size 2026-03-06T13:43:47.113 INFO:teuthology.orchestra.run.vm00.stdout:================================================================================ 2026-03-06T13:43:47.114 INFO:teuthology.orchestra.run.vm00.stdout:Removing: 2026-03-06T13:43:47.114 INFO:teuthology.orchestra.run.vm00.stdout: cephadm noarch 2:19.2.3-47.gc24117fd552.el9.clyso @ceph-noarch 775 k 2026-03-06T13:43:47.114 INFO:teuthology.orchestra.run.vm00.stdout: 2026-03-06T13:43:47.114 INFO:teuthology.orchestra.run.vm00.stdout:Transaction Summary 2026-03-06T13:43:47.114 INFO:teuthology.orchestra.run.vm00.stdout:================================================================================ 2026-03-06T13:43:47.114 INFO:teuthology.orchestra.run.vm00.stdout:Remove 1 Package 2026-03-06T13:43:47.114 INFO:teuthology.orchestra.run.vm00.stdout: 2026-03-06T13:43:47.114 INFO:teuthology.orchestra.run.vm00.stdout:Freed space: 775 k 2026-03-06T13:43:47.114 INFO:teuthology.orchestra.run.vm00.stdout:Running transaction check 2026-03-06T13:43:47.116 INFO:teuthology.orchestra.run.vm00.stdout:Transaction check succeeded. 2026-03-06T13:43:47.116 INFO:teuthology.orchestra.run.vm00.stdout:Running transaction test 2026-03-06T13:43:47.117 INFO:teuthology.orchestra.run.vm00.stdout:Transaction test succeeded. 2026-03-06T13:43:47.117 INFO:teuthology.orchestra.run.vm00.stdout:Running transaction 2026-03-06T13:43:47.135 INFO:teuthology.orchestra.run.vm00.stdout: Preparing : 1/1 2026-03-06T13:43:47.135 INFO:teuthology.orchestra.run.vm00.stdout: Erasing : cephadm-2:19.2.3-47.gc24117fd552.el9.clyso.noarch 1/1 2026-03-06T13:43:47.154 INFO:teuthology.orchestra.run.vm08.stdout: Running scriptlet: libcephsqlite-2:19.2.3-47.gc24117fd552.el9.clyso 113/113 2026-03-06T13:43:47.154 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : abseil-cpp-20211102.0-4.el9.x86_64 1/113 2026-03-06T13:43:47.154 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : ceph-base-2:19.2.3-47.gc24117fd552.el9.clyso.x86 2/113 2026-03-06T13:43:47.154 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : ceph-common-2:19.2.3-47.gc24117fd552.el9.clyso.x 3/113 2026-03-06T13:43:47.154 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : ceph-grafana-dashboards-2:19.2.3-47.gc24117fd552 4/113 2026-03-06T13:43:47.154 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : ceph-immutable-object-cache-2:19.2.3-47.gc24117f 5/113 2026-03-06T13:43:47.154 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : ceph-mgr-2:19.2.3-47.gc24117fd552.el9.clyso.x86_ 6/113 2026-03-06T13:43:47.154 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : ceph-mgr-cephadm-2:19.2.3-47.gc24117fd552.el9.cl 7/113 2026-03-06T13:43:47.154 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : ceph-mgr-dashboard-2:19.2.3-47.gc24117fd552.el9. 8/113 2026-03-06T13:43:47.154 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : ceph-mgr-diskprediction-local-2:19.2.3-47.gc2411 9/113 2026-03-06T13:43:47.154 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : ceph-mgr-k8sevents-2:19.2.3-47.gc24117fd552.el9. 10/113 2026-03-06T13:43:47.155 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : ceph-mgr-modules-core-2:19.2.3-47.gc24117fd552.e 11/113 2026-03-06T13:43:47.155 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : ceph-mgr-rook-2:19.2.3-47.gc24117fd552.el9.clyso 12/113 2026-03-06T13:43:47.155 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : ceph-osd-2:19.2.3-47.gc24117fd552.el9.clyso.x86_ 13/113 2026-03-06T13:43:47.155 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : ceph-prometheus-alerts-2:19.2.3-47.gc24117fd552. 14/113 2026-03-06T13:43:47.155 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : ceph-selinux-2:19.2.3-47.gc24117fd552.el9.clyso. 15/113 2026-03-06T13:43:47.155 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : ceph-volume-2:19.2.3-47.gc24117fd552.el9.clyso.n 16/113 2026-03-06T13:43:47.155 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : cryptsetup-2.8.1-3.el9.x86_64 17/113 2026-03-06T13:43:47.155 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : flexiblas-3.0.4-9.el9.x86_64 18/113 2026-03-06T13:43:47.155 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : flexiblas-netlib-3.0.4-9.el9.x86_64 19/113 2026-03-06T13:43:47.155 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : flexiblas-openblas-openmp-3.0.4-9.el9.x86_64 20/113 2026-03-06T13:43:47.155 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : gperftools-libs-2.9.1-3.el9.x86_64 21/113 2026-03-06T13:43:47.155 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : grpc-data-1.46.7-10.el9.noarch 22/113 2026-03-06T13:43:47.155 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : ledmon-libs-1.1.0-3.el9.x86_64 23/113 2026-03-06T13:43:47.155 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : libcephsqlite-2:19.2.3-47.gc24117fd552.el9.clyso 24/113 2026-03-06T13:43:47.155 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : libconfig-1.7.2-9.el9.x86_64 25/113 2026-03-06T13:43:47.155 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : libgfortran-11.5.0-14.el9.x86_64 26/113 2026-03-06T13:43:47.155 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : liboath-2.6.12-1.el9.x86_64 27/113 2026-03-06T13:43:47.155 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : libquadmath-11.5.0-14.el9.x86_64 28/113 2026-03-06T13:43:47.155 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : libradosstriper1-2:19.2.3-47.gc24117fd552.el9.cl 29/113 2026-03-06T13:43:47.155 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : libstoragemgmt-1.10.1-1.el9.x86_64 30/113 2026-03-06T13:43:47.155 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : libunwind-1.6.2-1.el9.x86_64 31/113 2026-03-06T13:43:47.155 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : libxslt-1.1.34-12.el9.x86_64 32/113 2026-03-06T13:43:47.155 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : openblas-0.3.29-1.el9.x86_64 33/113 2026-03-06T13:43:47.155 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : openblas-openmp-0.3.29-1.el9.x86_64 34/113 2026-03-06T13:43:47.155 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : pciutils-3.7.0-7.el9.x86_64 35/113 2026-03-06T13:43:47.155 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : protobuf-3.14.0-17.el9.x86_64 36/113 2026-03-06T13:43:47.155 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : protobuf-compiler-3.14.0-17.el9.x86_64 37/113 2026-03-06T13:43:47.155 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : python3-asyncssh-2.13.2-5.el9.noarch 38/113 2026-03-06T13:43:47.155 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : python3-autocommand-2.2.2-8.el9.noarch 39/113 2026-03-06T13:43:47.155 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : python3-babel-2.9.1-2.el9.noarch 40/113 2026-03-06T13:43:47.155 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : python3-backports-tarfile-1.2.0-1.el9.noarch 41/113 2026-03-06T13:43:47.155 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : python3-bcrypt-3.2.2-1.el9.x86_64 42/113 2026-03-06T13:43:47.155 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : python3-cachetools-4.2.4-1.el9.noarch 43/113 2026-03-06T13:43:47.155 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : python3-ceph-common-2:19.2.3-47.gc24117fd552.el9 44/113 2026-03-06T13:43:47.155 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : python3-certifi-2023.05.07-4.el9.noarch 45/113 2026-03-06T13:43:47.155 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : python3-cffi-1.14.5-5.el9.x86_64 46/113 2026-03-06T13:43:47.155 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : python3-chardet-4.0.0-5.el9.noarch 47/113 2026-03-06T13:43:47.155 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : python3-cheroot-10.0.1-4.el9.noarch 48/113 2026-03-06T13:43:47.155 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : python3-cherrypy-18.6.1-2.el9.noarch 49/113 2026-03-06T13:43:47.155 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : python3-cryptography-36.0.1-5.el9.x86_64 50/113 2026-03-06T13:43:47.155 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : python3-devel-3.9.25-3.el9.x86_64 51/113 2026-03-06T13:43:47.155 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : python3-google-auth-1:2.45.0-1.el9.noarch 52/113 2026-03-06T13:43:47.155 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : python3-grpcio-1.46.7-10.el9.x86_64 53/113 2026-03-06T13:43:47.155 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : python3-grpcio-tools-1.46.7-10.el9.x86_64 54/113 2026-03-06T13:43:47.155 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : python3-idna-2.10-7.el9.1.noarch 55/113 2026-03-06T13:43:47.155 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : python3-influxdb-5.3.1-1.el9.noarch 56/113 2026-03-06T13:43:47.155 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : python3-isodate-0.6.1-3.el9.noarch 57/113 2026-03-06T13:43:47.155 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : python3-jaraco-8.2.1-3.el9.noarch 58/113 2026-03-06T13:43:47.155 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : python3-jaraco-classes-3.2.1-5.el9.noarch 59/113 2026-03-06T13:43:47.156 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : python3-jaraco-collections-3.0.0-8.el9.noarch 60/113 2026-03-06T13:43:47.156 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : python3-jaraco-context-6.0.1-3.el9.noarch 61/113 2026-03-06T13:43:47.156 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : python3-jaraco-functools-3.5.0-2.el9.noarch 62/113 2026-03-06T13:43:47.156 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : python3-jaraco-text-4.0.0-2.el9.noarch 63/113 2026-03-06T13:43:47.156 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : python3-jinja2-2.11.3-8.el9.noarch 64/113 2026-03-06T13:43:47.156 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : python3-jsonpatch-1.21-16.el9.noarch 65/113 2026-03-06T13:43:47.156 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : python3-jsonpointer-2.0-4.el9.noarch 66/113 2026-03-06T13:43:47.156 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : python3-kubernetes-1:26.1.0-3.el9.noarch 67/113 2026-03-06T13:43:47.156 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : python3-libstoragemgmt-1.10.1-1.el9.x86_64 68/113 2026-03-06T13:43:47.156 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : python3-logutils-0.3.5-21.el9.noarch 69/113 2026-03-06T13:43:47.156 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : python3-lxml-4.6.5-3.el9.x86_64 70/113 2026-03-06T13:43:47.156 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : python3-mako-1.1.4-6.el9.noarch 71/113 2026-03-06T13:43:47.156 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : python3-markupsafe-1.1.1-12.el9.x86_64 72/113 2026-03-06T13:43:47.156 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : python3-more-itertools-8.12.0-2.el9.noarch 73/113 2026-03-06T13:43:47.156 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : python3-msgpack-1.0.3-2.el9.x86_64 74/113 2026-03-06T13:43:47.156 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : python3-natsort-7.1.1-5.el9.noarch 75/113 2026-03-06T13:43:47.156 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : python3-numpy-1:1.23.5-2.el9.x86_64 76/113 2026-03-06T13:43:47.156 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : python3-numpy-f2py-1:1.23.5-2.el9.x86_64 77/113 2026-03-06T13:43:47.156 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : python3-oauthlib-3.1.1-5.el9.noarch 78/113 2026-03-06T13:43:47.156 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : python3-packaging-20.9-5.el9.noarch 79/113 2026-03-06T13:43:47.156 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : python3-pecan-1.4.2-3.el9.noarch 80/113 2026-03-06T13:43:47.156 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : python3-ply-3.11-14.el9.noarch 81/113 2026-03-06T13:43:47.156 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : python3-portend-3.1.0-2.el9.noarch 82/113 2026-03-06T13:43:47.156 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : python3-prettytable-0.7.2-27.el9.noarch 83/113 2026-03-06T13:43:47.156 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : python3-protobuf-3.14.0-17.el9.noarch 84/113 2026-03-06T13:43:47.156 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : python3-pyOpenSSL-21.0.0-1.el9.noarch 85/113 2026-03-06T13:43:47.156 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : python3-pyasn1-0.4.8-7.el9.noarch 86/113 2026-03-06T13:43:47.156 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : python3-pyasn1-modules-0.4.8-7.el9.noarch 87/113 2026-03-06T13:43:47.156 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : python3-pycparser-2.20-6.el9.noarch 88/113 2026-03-06T13:43:47.156 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : python3-pysocks-1.7.1-12.el9.noarch 89/113 2026-03-06T13:43:47.156 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : python3-pytz-2021.1-5.el9.noarch 90/113 2026-03-06T13:43:47.156 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : python3-repoze-lru-0.7-16.el9.noarch 91/113 2026-03-06T13:43:47.156 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : python3-requests-2.25.1-10.el9.noarch 92/113 2026-03-06T13:43:47.156 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : python3-requests-oauthlib-1.3.0-12.el9.noarch 93/113 2026-03-06T13:43:47.156 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : python3-routes-2.5.1-5.el9.noarch 94/113 2026-03-06T13:43:47.156 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : python3-rsa-4.9-2.el9.noarch 95/113 2026-03-06T13:43:47.156 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : python3-saml-1.16.0-1.el9.noarch 96/113 2026-03-06T13:43:47.156 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : python3-scipy-1.9.3-2.el9.x86_64 97/113 2026-03-06T13:43:47.156 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : python3-tempora-5.0.0-2.el9.noarch 98/113 2026-03-06T13:43:47.156 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : python3-toml-0.10.2-6.el9.noarch 99/113 2026-03-06T13:43:47.156 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : python3-typing-extensions-4.15.0-1.el9.noarch 100/113 2026-03-06T13:43:47.156 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : python3-urllib3-1.26.5-7.el9.noarch 101/113 2026-03-06T13:43:47.156 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : python3-webob-1.8.8-2.el9.noarch 102/113 2026-03-06T13:43:47.156 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : python3-websocket-client-1.2.3-2.el9.noarch 103/113 2026-03-06T13:43:47.156 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : python3-werkzeug-2.0.3-3.el9.1.noarch 104/113 2026-03-06T13:43:47.156 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : python3-xmlsec-1.3.13-1.el9.x86_64 105/113 2026-03-06T13:43:47.156 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : python3-zc-lockfile-2.0-10.el9.noarch 106/113 2026-03-06T13:43:47.156 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : qatlib-25.08.0-2.el9.x86_64 107/113 2026-03-06T13:43:47.156 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : qatlib-service-25.08.0-2.el9.x86_64 108/113 2026-03-06T13:43:47.157 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : qatzip-libs-1.3.1-1.el9.x86_64 109/113 2026-03-06T13:43:47.157 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : rbd-mirror-2:19.2.3-47.gc24117fd552.el9.clyso.x8 110/113 2026-03-06T13:43:47.157 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : smartmontools-1:7.2-10.el9.x86_64 111/113 2026-03-06T13:43:47.157 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : xmlsec1-1.2.29-13.el9.x86_64 112/113 2026-03-06T13:43:47.170 INFO:teuthology.orchestra.run.vm01.stdout: Running scriptlet: libcephsqlite-2:19.2.3-47.gc24117fd552.el9.clyso 113/113 2026-03-06T13:43:47.170 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : abseil-cpp-20211102.0-4.el9.x86_64 1/113 2026-03-06T13:43:47.170 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : ceph-base-2:19.2.3-47.gc24117fd552.el9.clyso.x86 2/113 2026-03-06T13:43:47.170 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : ceph-common-2:19.2.3-47.gc24117fd552.el9.clyso.x 3/113 2026-03-06T13:43:47.170 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : ceph-grafana-dashboards-2:19.2.3-47.gc24117fd552 4/113 2026-03-06T13:43:47.170 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : ceph-immutable-object-cache-2:19.2.3-47.gc24117f 5/113 2026-03-06T13:43:47.170 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : ceph-mgr-2:19.2.3-47.gc24117fd552.el9.clyso.x86_ 6/113 2026-03-06T13:43:47.170 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : ceph-mgr-cephadm-2:19.2.3-47.gc24117fd552.el9.cl 7/113 2026-03-06T13:43:47.170 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : ceph-mgr-dashboard-2:19.2.3-47.gc24117fd552.el9. 8/113 2026-03-06T13:43:47.170 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : ceph-mgr-diskprediction-local-2:19.2.3-47.gc2411 9/113 2026-03-06T13:43:47.170 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : ceph-mgr-k8sevents-2:19.2.3-47.gc24117fd552.el9. 10/113 2026-03-06T13:43:47.170 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : ceph-mgr-modules-core-2:19.2.3-47.gc24117fd552.e 11/113 2026-03-06T13:43:47.170 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : ceph-mgr-rook-2:19.2.3-47.gc24117fd552.el9.clyso 12/113 2026-03-06T13:43:47.170 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : ceph-osd-2:19.2.3-47.gc24117fd552.el9.clyso.x86_ 13/113 2026-03-06T13:43:47.171 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : ceph-prometheus-alerts-2:19.2.3-47.gc24117fd552. 14/113 2026-03-06T13:43:47.171 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : ceph-selinux-2:19.2.3-47.gc24117fd552.el9.clyso. 15/113 2026-03-06T13:43:47.171 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : ceph-volume-2:19.2.3-47.gc24117fd552.el9.clyso.n 16/113 2026-03-06T13:43:47.171 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : cryptsetup-2.8.1-3.el9.x86_64 17/113 2026-03-06T13:43:47.171 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : flexiblas-3.0.4-9.el9.x86_64 18/113 2026-03-06T13:43:47.171 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : flexiblas-netlib-3.0.4-9.el9.x86_64 19/113 2026-03-06T13:43:47.171 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : flexiblas-openblas-openmp-3.0.4-9.el9.x86_64 20/113 2026-03-06T13:43:47.171 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : gperftools-libs-2.9.1-3.el9.x86_64 21/113 2026-03-06T13:43:47.171 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : grpc-data-1.46.7-10.el9.noarch 22/113 2026-03-06T13:43:47.171 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : ledmon-libs-1.1.0-3.el9.x86_64 23/113 2026-03-06T13:43:47.171 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : libcephsqlite-2:19.2.3-47.gc24117fd552.el9.clyso 24/113 2026-03-06T13:43:47.171 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : libconfig-1.7.2-9.el9.x86_64 25/113 2026-03-06T13:43:47.171 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : libgfortran-11.5.0-14.el9.x86_64 26/113 2026-03-06T13:43:47.171 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : liboath-2.6.12-1.el9.x86_64 27/113 2026-03-06T13:43:47.171 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : libquadmath-11.5.0-14.el9.x86_64 28/113 2026-03-06T13:43:47.171 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : libradosstriper1-2:19.2.3-47.gc24117fd552.el9.cl 29/113 2026-03-06T13:43:47.171 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : libstoragemgmt-1.10.1-1.el9.x86_64 30/113 2026-03-06T13:43:47.171 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : libunwind-1.6.2-1.el9.x86_64 31/113 2026-03-06T13:43:47.171 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : libxslt-1.1.34-12.el9.x86_64 32/113 2026-03-06T13:43:47.171 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : openblas-0.3.29-1.el9.x86_64 33/113 2026-03-06T13:43:47.171 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : openblas-openmp-0.3.29-1.el9.x86_64 34/113 2026-03-06T13:43:47.171 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : pciutils-3.7.0-7.el9.x86_64 35/113 2026-03-06T13:43:47.171 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : protobuf-3.14.0-17.el9.x86_64 36/113 2026-03-06T13:43:47.172 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : protobuf-compiler-3.14.0-17.el9.x86_64 37/113 2026-03-06T13:43:47.172 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-asyncssh-2.13.2-5.el9.noarch 38/113 2026-03-06T13:43:47.172 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-autocommand-2.2.2-8.el9.noarch 39/113 2026-03-06T13:43:47.172 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-babel-2.9.1-2.el9.noarch 40/113 2026-03-06T13:43:47.172 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-backports-tarfile-1.2.0-1.el9.noarch 41/113 2026-03-06T13:43:47.172 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-bcrypt-3.2.2-1.el9.x86_64 42/113 2026-03-06T13:43:47.172 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-cachetools-4.2.4-1.el9.noarch 43/113 2026-03-06T13:43:47.172 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-ceph-common-2:19.2.3-47.gc24117fd552.el9 44/113 2026-03-06T13:43:47.172 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-certifi-2023.05.07-4.el9.noarch 45/113 2026-03-06T13:43:47.172 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-cffi-1.14.5-5.el9.x86_64 46/113 2026-03-06T13:43:47.173 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-chardet-4.0.0-5.el9.noarch 47/113 2026-03-06T13:43:47.173 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-cheroot-10.0.1-4.el9.noarch 48/113 2026-03-06T13:43:47.173 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-cherrypy-18.6.1-2.el9.noarch 49/113 2026-03-06T13:43:47.173 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-cryptography-36.0.1-5.el9.x86_64 50/113 2026-03-06T13:43:47.173 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-devel-3.9.25-3.el9.x86_64 51/113 2026-03-06T13:43:47.173 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-google-auth-1:2.45.0-1.el9.noarch 52/113 2026-03-06T13:43:47.173 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-grpcio-1.46.7-10.el9.x86_64 53/113 2026-03-06T13:43:47.173 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-grpcio-tools-1.46.7-10.el9.x86_64 54/113 2026-03-06T13:43:47.173 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-idna-2.10-7.el9.1.noarch 55/113 2026-03-06T13:43:47.173 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-influxdb-5.3.1-1.el9.noarch 56/113 2026-03-06T13:43:47.173 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-isodate-0.6.1-3.el9.noarch 57/113 2026-03-06T13:43:47.173 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-jaraco-8.2.1-3.el9.noarch 58/113 2026-03-06T13:43:47.173 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-jaraco-classes-3.2.1-5.el9.noarch 59/113 2026-03-06T13:43:47.173 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-jaraco-collections-3.0.0-8.el9.noarch 60/113 2026-03-06T13:43:47.173 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-jaraco-context-6.0.1-3.el9.noarch 61/113 2026-03-06T13:43:47.173 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-jaraco-functools-3.5.0-2.el9.noarch 62/113 2026-03-06T13:43:47.173 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-jaraco-text-4.0.0-2.el9.noarch 63/113 2026-03-06T13:43:47.173 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-jinja2-2.11.3-8.el9.noarch 64/113 2026-03-06T13:43:47.173 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-jsonpatch-1.21-16.el9.noarch 65/113 2026-03-06T13:43:47.173 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-jsonpointer-2.0-4.el9.noarch 66/113 2026-03-06T13:43:47.173 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-kubernetes-1:26.1.0-3.el9.noarch 67/113 2026-03-06T13:43:47.173 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-libstoragemgmt-1.10.1-1.el9.x86_64 68/113 2026-03-06T13:43:47.173 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-logutils-0.3.5-21.el9.noarch 69/113 2026-03-06T13:43:47.174 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-lxml-4.6.5-3.el9.x86_64 70/113 2026-03-06T13:43:47.174 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-mako-1.1.4-6.el9.noarch 71/113 2026-03-06T13:43:47.174 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-markupsafe-1.1.1-12.el9.x86_64 72/113 2026-03-06T13:43:47.174 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-more-itertools-8.12.0-2.el9.noarch 73/113 2026-03-06T13:43:47.174 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-msgpack-1.0.3-2.el9.x86_64 74/113 2026-03-06T13:43:47.174 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-natsort-7.1.1-5.el9.noarch 75/113 2026-03-06T13:43:47.174 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-numpy-1:1.23.5-2.el9.x86_64 76/113 2026-03-06T13:43:47.174 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-numpy-f2py-1:1.23.5-2.el9.x86_64 77/113 2026-03-06T13:43:47.174 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-oauthlib-3.1.1-5.el9.noarch 78/113 2026-03-06T13:43:47.174 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-packaging-20.9-5.el9.noarch 79/113 2026-03-06T13:43:47.174 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-pecan-1.4.2-3.el9.noarch 80/113 2026-03-06T13:43:47.174 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-ply-3.11-14.el9.noarch 81/113 2026-03-06T13:43:47.174 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-portend-3.1.0-2.el9.noarch 82/113 2026-03-06T13:43:47.174 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-prettytable-0.7.2-27.el9.noarch 83/113 2026-03-06T13:43:47.174 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-protobuf-3.14.0-17.el9.noarch 84/113 2026-03-06T13:43:47.174 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-pyOpenSSL-21.0.0-1.el9.noarch 85/113 2026-03-06T13:43:47.174 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-pyasn1-0.4.8-7.el9.noarch 86/113 2026-03-06T13:43:47.174 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-pyasn1-modules-0.4.8-7.el9.noarch 87/113 2026-03-06T13:43:47.174 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-pycparser-2.20-6.el9.noarch 88/113 2026-03-06T13:43:47.174 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-pysocks-1.7.1-12.el9.noarch 89/113 2026-03-06T13:43:47.174 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-pytz-2021.1-5.el9.noarch 90/113 2026-03-06T13:43:47.174 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-repoze-lru-0.7-16.el9.noarch 91/113 2026-03-06T13:43:47.174 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-requests-2.25.1-10.el9.noarch 92/113 2026-03-06T13:43:47.174 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-requests-oauthlib-1.3.0-12.el9.noarch 93/113 2026-03-06T13:43:47.174 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-routes-2.5.1-5.el9.noarch 94/113 2026-03-06T13:43:47.174 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-rsa-4.9-2.el9.noarch 95/113 2026-03-06T13:43:47.174 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-saml-1.16.0-1.el9.noarch 96/113 2026-03-06T13:43:47.174 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-scipy-1.9.3-2.el9.x86_64 97/113 2026-03-06T13:43:47.174 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-tempora-5.0.0-2.el9.noarch 98/113 2026-03-06T13:43:47.174 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-toml-0.10.2-6.el9.noarch 99/113 2026-03-06T13:43:47.174 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-typing-extensions-4.15.0-1.el9.noarch 100/113 2026-03-06T13:43:47.174 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-urllib3-1.26.5-7.el9.noarch 101/113 2026-03-06T13:43:47.174 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-webob-1.8.8-2.el9.noarch 102/113 2026-03-06T13:43:47.175 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-websocket-client-1.2.3-2.el9.noarch 103/113 2026-03-06T13:43:47.175 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-werkzeug-2.0.3-3.el9.1.noarch 104/113 2026-03-06T13:43:47.175 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-xmlsec-1.3.13-1.el9.x86_64 105/113 2026-03-06T13:43:47.175 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-zc-lockfile-2.0-10.el9.noarch 106/113 2026-03-06T13:43:47.175 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : qatlib-25.08.0-2.el9.x86_64 107/113 2026-03-06T13:43:47.175 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : qatlib-service-25.08.0-2.el9.x86_64 108/113 2026-03-06T13:43:47.175 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : qatzip-libs-1.3.1-1.el9.x86_64 109/113 2026-03-06T13:43:47.175 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : rbd-mirror-2:19.2.3-47.gc24117fd552.el9.clyso.x8 110/113 2026-03-06T13:43:47.175 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : smartmontools-1:7.2-10.el9.x86_64 111/113 2026-03-06T13:43:47.175 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : xmlsec1-1.2.29-13.el9.x86_64 112/113 2026-03-06T13:43:47.245 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : xmlsec1-openssl-1.2.29-13.el9.x86_64 113/113 2026-03-06T13:43:47.245 INFO:teuthology.orchestra.run.vm08.stdout: 2026-03-06T13:43:47.245 INFO:teuthology.orchestra.run.vm08.stdout:Removed: 2026-03-06T13:43:47.245 INFO:teuthology.orchestra.run.vm08.stdout: abseil-cpp-20211102.0-4.el9.x86_64 2026-03-06T13:43:47.245 INFO:teuthology.orchestra.run.vm08.stdout: ceph-base-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 2026-03-06T13:43:47.245 INFO:teuthology.orchestra.run.vm08.stdout: ceph-common-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 2026-03-06T13:43:47.245 INFO:teuthology.orchestra.run.vm08.stdout: ceph-grafana-dashboards-2:19.2.3-47.gc24117fd552.el9.clyso.noarch 2026-03-06T13:43:47.245 INFO:teuthology.orchestra.run.vm08.stdout: ceph-immutable-object-cache-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 2026-03-06T13:43:47.245 INFO:teuthology.orchestra.run.vm08.stdout: ceph-mgr-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 2026-03-06T13:43:47.245 INFO:teuthology.orchestra.run.vm08.stdout: ceph-mgr-cephadm-2:19.2.3-47.gc24117fd552.el9.clyso.noarch 2026-03-06T13:43:47.245 INFO:teuthology.orchestra.run.vm08.stdout: ceph-mgr-dashboard-2:19.2.3-47.gc24117fd552.el9.clyso.noarch 2026-03-06T13:43:47.245 INFO:teuthology.orchestra.run.vm08.stdout: ceph-mgr-diskprediction-local-2:19.2.3-47.gc24117fd552.el9.clyso.noarch 2026-03-06T13:43:47.245 INFO:teuthology.orchestra.run.vm08.stdout: ceph-mgr-k8sevents-2:19.2.3-47.gc24117fd552.el9.clyso.noarch 2026-03-06T13:43:47.245 INFO:teuthology.orchestra.run.vm08.stdout: ceph-mgr-modules-core-2:19.2.3-47.gc24117fd552.el9.clyso.noarch 2026-03-06T13:43:47.245 INFO:teuthology.orchestra.run.vm08.stdout: ceph-mgr-rook-2:19.2.3-47.gc24117fd552.el9.clyso.noarch 2026-03-06T13:43:47.245 INFO:teuthology.orchestra.run.vm08.stdout: ceph-osd-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 2026-03-06T13:43:47.245 INFO:teuthology.orchestra.run.vm08.stdout: ceph-prometheus-alerts-2:19.2.3-47.gc24117fd552.el9.clyso.noarch 2026-03-06T13:43:47.245 INFO:teuthology.orchestra.run.vm08.stdout: ceph-selinux-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 2026-03-06T13:43:47.245 INFO:teuthology.orchestra.run.vm08.stdout: ceph-volume-2:19.2.3-47.gc24117fd552.el9.clyso.noarch 2026-03-06T13:43:47.245 INFO:teuthology.orchestra.run.vm08.stdout: cryptsetup-2.8.1-3.el9.x86_64 2026-03-06T13:43:47.245 INFO:teuthology.orchestra.run.vm08.stdout: flexiblas-3.0.4-9.el9.x86_64 2026-03-06T13:43:47.245 INFO:teuthology.orchestra.run.vm08.stdout: flexiblas-netlib-3.0.4-9.el9.x86_64 2026-03-06T13:43:47.245 INFO:teuthology.orchestra.run.vm08.stdout: flexiblas-openblas-openmp-3.0.4-9.el9.x86_64 2026-03-06T13:43:47.245 INFO:teuthology.orchestra.run.vm08.stdout: gperftools-libs-2.9.1-3.el9.x86_64 2026-03-06T13:43:47.245 INFO:teuthology.orchestra.run.vm08.stdout: grpc-data-1.46.7-10.el9.noarch 2026-03-06T13:43:47.245 INFO:teuthology.orchestra.run.vm08.stdout: ledmon-libs-1.1.0-3.el9.x86_64 2026-03-06T13:43:47.245 INFO:teuthology.orchestra.run.vm08.stdout: libcephsqlite-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 2026-03-06T13:43:47.245 INFO:teuthology.orchestra.run.vm08.stdout: libconfig-1.7.2-9.el9.x86_64 2026-03-06T13:43:47.245 INFO:teuthology.orchestra.run.vm08.stdout: libgfortran-11.5.0-14.el9.x86_64 2026-03-06T13:43:47.245 INFO:teuthology.orchestra.run.vm08.stdout: liboath-2.6.12-1.el9.x86_64 2026-03-06T13:43:47.245 INFO:teuthology.orchestra.run.vm08.stdout: libquadmath-11.5.0-14.el9.x86_64 2026-03-06T13:43:47.245 INFO:teuthology.orchestra.run.vm08.stdout: libradosstriper1-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 2026-03-06T13:43:47.245 INFO:teuthology.orchestra.run.vm08.stdout: libstoragemgmt-1.10.1-1.el9.x86_64 2026-03-06T13:43:47.245 INFO:teuthology.orchestra.run.vm08.stdout: libunwind-1.6.2-1.el9.x86_64 2026-03-06T13:43:47.245 INFO:teuthology.orchestra.run.vm08.stdout: libxslt-1.1.34-12.el9.x86_64 2026-03-06T13:43:47.245 INFO:teuthology.orchestra.run.vm08.stdout: openblas-0.3.29-1.el9.x86_64 2026-03-06T13:43:47.245 INFO:teuthology.orchestra.run.vm08.stdout: openblas-openmp-0.3.29-1.el9.x86_64 2026-03-06T13:43:47.246 INFO:teuthology.orchestra.run.vm08.stdout: pciutils-3.7.0-7.el9.x86_64 2026-03-06T13:43:47.246 INFO:teuthology.orchestra.run.vm08.stdout: protobuf-3.14.0-17.el9.x86_64 2026-03-06T13:43:47.246 INFO:teuthology.orchestra.run.vm08.stdout: protobuf-compiler-3.14.0-17.el9.x86_64 2026-03-06T13:43:47.246 INFO:teuthology.orchestra.run.vm08.stdout: python3-asyncssh-2.13.2-5.el9.noarch 2026-03-06T13:43:47.246 INFO:teuthology.orchestra.run.vm08.stdout: python3-autocommand-2.2.2-8.el9.noarch 2026-03-06T13:43:47.246 INFO:teuthology.orchestra.run.vm08.stdout: python3-babel-2.9.1-2.el9.noarch 2026-03-06T13:43:47.246 INFO:teuthology.orchestra.run.vm08.stdout: python3-backports-tarfile-1.2.0-1.el9.noarch 2026-03-06T13:43:47.246 INFO:teuthology.orchestra.run.vm08.stdout: python3-bcrypt-3.2.2-1.el9.x86_64 2026-03-06T13:43:47.246 INFO:teuthology.orchestra.run.vm08.stdout: python3-cachetools-4.2.4-1.el9.noarch 2026-03-06T13:43:47.246 INFO:teuthology.orchestra.run.vm08.stdout: python3-ceph-common-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 2026-03-06T13:43:47.246 INFO:teuthology.orchestra.run.vm08.stdout: python3-certifi-2023.05.07-4.el9.noarch 2026-03-06T13:43:47.246 INFO:teuthology.orchestra.run.vm08.stdout: python3-cffi-1.14.5-5.el9.x86_64 2026-03-06T13:43:47.246 INFO:teuthology.orchestra.run.vm08.stdout: python3-chardet-4.0.0-5.el9.noarch 2026-03-06T13:43:47.246 INFO:teuthology.orchestra.run.vm08.stdout: python3-cheroot-10.0.1-4.el9.noarch 2026-03-06T13:43:47.246 INFO:teuthology.orchestra.run.vm08.stdout: python3-cherrypy-18.6.1-2.el9.noarch 2026-03-06T13:43:47.246 INFO:teuthology.orchestra.run.vm08.stdout: python3-cryptography-36.0.1-5.el9.x86_64 2026-03-06T13:43:47.246 INFO:teuthology.orchestra.run.vm08.stdout: python3-devel-3.9.25-3.el9.x86_64 2026-03-06T13:43:47.246 INFO:teuthology.orchestra.run.vm08.stdout: python3-google-auth-1:2.45.0-1.el9.noarch 2026-03-06T13:43:47.246 INFO:teuthology.orchestra.run.vm08.stdout: python3-grpcio-1.46.7-10.el9.x86_64 2026-03-06T13:43:47.246 INFO:teuthology.orchestra.run.vm08.stdout: python3-grpcio-tools-1.46.7-10.el9.x86_64 2026-03-06T13:43:47.246 INFO:teuthology.orchestra.run.vm08.stdout: python3-idna-2.10-7.el9.1.noarch 2026-03-06T13:43:47.246 INFO:teuthology.orchestra.run.vm08.stdout: python3-influxdb-5.3.1-1.el9.noarch 2026-03-06T13:43:47.246 INFO:teuthology.orchestra.run.vm08.stdout: python3-isodate-0.6.1-3.el9.noarch 2026-03-06T13:43:47.246 INFO:teuthology.orchestra.run.vm08.stdout: python3-jaraco-8.2.1-3.el9.noarch 2026-03-06T13:43:47.246 INFO:teuthology.orchestra.run.vm08.stdout: python3-jaraco-classes-3.2.1-5.el9.noarch 2026-03-06T13:43:47.246 INFO:teuthology.orchestra.run.vm08.stdout: python3-jaraco-collections-3.0.0-8.el9.noarch 2026-03-06T13:43:47.246 INFO:teuthology.orchestra.run.vm08.stdout: python3-jaraco-context-6.0.1-3.el9.noarch 2026-03-06T13:43:47.246 INFO:teuthology.orchestra.run.vm08.stdout: python3-jaraco-functools-3.5.0-2.el9.noarch 2026-03-06T13:43:47.246 INFO:teuthology.orchestra.run.vm08.stdout: python3-jaraco-text-4.0.0-2.el9.noarch 2026-03-06T13:43:47.246 INFO:teuthology.orchestra.run.vm08.stdout: python3-jinja2-2.11.3-8.el9.noarch 2026-03-06T13:43:47.246 INFO:teuthology.orchestra.run.vm08.stdout: python3-jsonpatch-1.21-16.el9.noarch 2026-03-06T13:43:47.246 INFO:teuthology.orchestra.run.vm08.stdout: python3-jsonpointer-2.0-4.el9.noarch 2026-03-06T13:43:47.246 INFO:teuthology.orchestra.run.vm08.stdout: python3-kubernetes-1:26.1.0-3.el9.noarch 2026-03-06T13:43:47.246 INFO:teuthology.orchestra.run.vm08.stdout: python3-libstoragemgmt-1.10.1-1.el9.x86_64 2026-03-06T13:43:47.246 INFO:teuthology.orchestra.run.vm08.stdout: python3-logutils-0.3.5-21.el9.noarch 2026-03-06T13:43:47.246 INFO:teuthology.orchestra.run.vm08.stdout: python3-lxml-4.6.5-3.el9.x86_64 2026-03-06T13:43:47.246 INFO:teuthology.orchestra.run.vm08.stdout: python3-mako-1.1.4-6.el9.noarch 2026-03-06T13:43:47.246 INFO:teuthology.orchestra.run.vm08.stdout: python3-markupsafe-1.1.1-12.el9.x86_64 2026-03-06T13:43:47.246 INFO:teuthology.orchestra.run.vm08.stdout: python3-more-itertools-8.12.0-2.el9.noarch 2026-03-06T13:43:47.246 INFO:teuthology.orchestra.run.vm08.stdout: python3-msgpack-1.0.3-2.el9.x86_64 2026-03-06T13:43:47.246 INFO:teuthology.orchestra.run.vm08.stdout: python3-natsort-7.1.1-5.el9.noarch 2026-03-06T13:43:47.246 INFO:teuthology.orchestra.run.vm08.stdout: python3-numpy-1:1.23.5-2.el9.x86_64 2026-03-06T13:43:47.246 INFO:teuthology.orchestra.run.vm08.stdout: python3-numpy-f2py-1:1.23.5-2.el9.x86_64 2026-03-06T13:43:47.246 INFO:teuthology.orchestra.run.vm08.stdout: python3-oauthlib-3.1.1-5.el9.noarch 2026-03-06T13:43:47.246 INFO:teuthology.orchestra.run.vm08.stdout: python3-packaging-20.9-5.el9.noarch 2026-03-06T13:43:47.246 INFO:teuthology.orchestra.run.vm08.stdout: python3-pecan-1.4.2-3.el9.noarch 2026-03-06T13:43:47.246 INFO:teuthology.orchestra.run.vm08.stdout: python3-ply-3.11-14.el9.noarch 2026-03-06T13:43:47.246 INFO:teuthology.orchestra.run.vm08.stdout: python3-portend-3.1.0-2.el9.noarch 2026-03-06T13:43:47.246 INFO:teuthology.orchestra.run.vm08.stdout: python3-prettytable-0.7.2-27.el9.noarch 2026-03-06T13:43:47.246 INFO:teuthology.orchestra.run.vm08.stdout: python3-protobuf-3.14.0-17.el9.noarch 2026-03-06T13:43:47.246 INFO:teuthology.orchestra.run.vm08.stdout: python3-pyOpenSSL-21.0.0-1.el9.noarch 2026-03-06T13:43:47.246 INFO:teuthology.orchestra.run.vm08.stdout: python3-pyasn1-0.4.8-7.el9.noarch 2026-03-06T13:43:47.246 INFO:teuthology.orchestra.run.vm08.stdout: python3-pyasn1-modules-0.4.8-7.el9.noarch 2026-03-06T13:43:47.246 INFO:teuthology.orchestra.run.vm08.stdout: python3-pycparser-2.20-6.el9.noarch 2026-03-06T13:43:47.246 INFO:teuthology.orchestra.run.vm08.stdout: python3-pysocks-1.7.1-12.el9.noarch 2026-03-06T13:43:47.246 INFO:teuthology.orchestra.run.vm08.stdout: python3-pytz-2021.1-5.el9.noarch 2026-03-06T13:43:47.246 INFO:teuthology.orchestra.run.vm08.stdout: python3-repoze-lru-0.7-16.el9.noarch 2026-03-06T13:43:47.246 INFO:teuthology.orchestra.run.vm08.stdout: python3-requests-2.25.1-10.el9.noarch 2026-03-06T13:43:47.246 INFO:teuthology.orchestra.run.vm08.stdout: python3-requests-oauthlib-1.3.0-12.el9.noarch 2026-03-06T13:43:47.246 INFO:teuthology.orchestra.run.vm08.stdout: python3-routes-2.5.1-5.el9.noarch 2026-03-06T13:43:47.246 INFO:teuthology.orchestra.run.vm08.stdout: python3-rsa-4.9-2.el9.noarch 2026-03-06T13:43:47.246 INFO:teuthology.orchestra.run.vm08.stdout: python3-saml-1.16.0-1.el9.noarch 2026-03-06T13:43:47.246 INFO:teuthology.orchestra.run.vm08.stdout: python3-scipy-1.9.3-2.el9.x86_64 2026-03-06T13:43:47.246 INFO:teuthology.orchestra.run.vm08.stdout: python3-tempora-5.0.0-2.el9.noarch 2026-03-06T13:43:47.247 INFO:teuthology.orchestra.run.vm08.stdout: python3-toml-0.10.2-6.el9.noarch 2026-03-06T13:43:47.247 INFO:teuthology.orchestra.run.vm08.stdout: python3-typing-extensions-4.15.0-1.el9.noarch 2026-03-06T13:43:47.247 INFO:teuthology.orchestra.run.vm08.stdout: python3-urllib3-1.26.5-7.el9.noarch 2026-03-06T13:43:47.247 INFO:teuthology.orchestra.run.vm08.stdout: python3-webob-1.8.8-2.el9.noarch 2026-03-06T13:43:47.247 INFO:teuthology.orchestra.run.vm08.stdout: python3-websocket-client-1.2.3-2.el9.noarch 2026-03-06T13:43:47.247 INFO:teuthology.orchestra.run.vm08.stdout: python3-werkzeug-2.0.3-3.el9.1.noarch 2026-03-06T13:43:47.247 INFO:teuthology.orchestra.run.vm08.stdout: python3-xmlsec-1.3.13-1.el9.x86_64 2026-03-06T13:43:47.247 INFO:teuthology.orchestra.run.vm08.stdout: python3-zc-lockfile-2.0-10.el9.noarch 2026-03-06T13:43:47.247 INFO:teuthology.orchestra.run.vm08.stdout: qatlib-25.08.0-2.el9.x86_64 2026-03-06T13:43:47.247 INFO:teuthology.orchestra.run.vm08.stdout: qatlib-service-25.08.0-2.el9.x86_64 2026-03-06T13:43:47.247 INFO:teuthology.orchestra.run.vm08.stdout: qatzip-libs-1.3.1-1.el9.x86_64 2026-03-06T13:43:47.247 INFO:teuthology.orchestra.run.vm08.stdout: rbd-mirror-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 2026-03-06T13:43:47.247 INFO:teuthology.orchestra.run.vm08.stdout: smartmontools-1:7.2-10.el9.x86_64 2026-03-06T13:43:47.247 INFO:teuthology.orchestra.run.vm08.stdout: xmlsec1-1.2.29-13.el9.x86_64 2026-03-06T13:43:47.247 INFO:teuthology.orchestra.run.vm08.stdout: xmlsec1-openssl-1.2.29-13.el9.x86_64 2026-03-06T13:43:47.247 INFO:teuthology.orchestra.run.vm08.stdout: 2026-03-06T13:43:47.247 INFO:teuthology.orchestra.run.vm08.stdout:Complete! 2026-03-06T13:43:47.267 INFO:teuthology.orchestra.run.vm00.stdout: Running scriptlet: cephadm-2:19.2.3-47.gc24117fd552.el9.clyso.noarch 1/1 2026-03-06T13:43:47.267 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : xmlsec1-openssl-1.2.29-13.el9.x86_64 113/113 2026-03-06T13:43:47.267 INFO:teuthology.orchestra.run.vm01.stdout: 2026-03-06T13:43:47.267 INFO:teuthology.orchestra.run.vm01.stdout:Removed: 2026-03-06T13:43:47.267 INFO:teuthology.orchestra.run.vm01.stdout: abseil-cpp-20211102.0-4.el9.x86_64 2026-03-06T13:43:47.267 INFO:teuthology.orchestra.run.vm01.stdout: ceph-base-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 2026-03-06T13:43:47.267 INFO:teuthology.orchestra.run.vm01.stdout: ceph-common-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 2026-03-06T13:43:47.267 INFO:teuthology.orchestra.run.vm01.stdout: ceph-grafana-dashboards-2:19.2.3-47.gc24117fd552.el9.clyso.noarch 2026-03-06T13:43:47.267 INFO:teuthology.orchestra.run.vm01.stdout: ceph-immutable-object-cache-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 2026-03-06T13:43:47.267 INFO:teuthology.orchestra.run.vm01.stdout: ceph-mgr-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 2026-03-06T13:43:47.267 INFO:teuthology.orchestra.run.vm01.stdout: ceph-mgr-cephadm-2:19.2.3-47.gc24117fd552.el9.clyso.noarch 2026-03-06T13:43:47.267 INFO:teuthology.orchestra.run.vm01.stdout: ceph-mgr-dashboard-2:19.2.3-47.gc24117fd552.el9.clyso.noarch 2026-03-06T13:43:47.267 INFO:teuthology.orchestra.run.vm01.stdout: ceph-mgr-diskprediction-local-2:19.2.3-47.gc24117fd552.el9.clyso.noarch 2026-03-06T13:43:47.267 INFO:teuthology.orchestra.run.vm01.stdout: ceph-mgr-k8sevents-2:19.2.3-47.gc24117fd552.el9.clyso.noarch 2026-03-06T13:43:47.267 INFO:teuthology.orchestra.run.vm01.stdout: ceph-mgr-modules-core-2:19.2.3-47.gc24117fd552.el9.clyso.noarch 2026-03-06T13:43:47.267 INFO:teuthology.orchestra.run.vm01.stdout: ceph-mgr-rook-2:19.2.3-47.gc24117fd552.el9.clyso.noarch 2026-03-06T13:43:47.267 INFO:teuthology.orchestra.run.vm01.stdout: ceph-osd-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 2026-03-06T13:43:47.267 INFO:teuthology.orchestra.run.vm01.stdout: ceph-prometheus-alerts-2:19.2.3-47.gc24117fd552.el9.clyso.noarch 2026-03-06T13:43:47.267 INFO:teuthology.orchestra.run.vm01.stdout: ceph-selinux-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 2026-03-06T13:43:47.268 INFO:teuthology.orchestra.run.vm01.stdout: ceph-volume-2:19.2.3-47.gc24117fd552.el9.clyso.noarch 2026-03-06T13:43:47.268 INFO:teuthology.orchestra.run.vm01.stdout: cryptsetup-2.8.1-3.el9.x86_64 2026-03-06T13:43:47.268 INFO:teuthology.orchestra.run.vm01.stdout: flexiblas-3.0.4-9.el9.x86_64 2026-03-06T13:43:47.268 INFO:teuthology.orchestra.run.vm01.stdout: flexiblas-netlib-3.0.4-9.el9.x86_64 2026-03-06T13:43:47.268 INFO:teuthology.orchestra.run.vm01.stdout: flexiblas-openblas-openmp-3.0.4-9.el9.x86_64 2026-03-06T13:43:47.268 INFO:teuthology.orchestra.run.vm01.stdout: gperftools-libs-2.9.1-3.el9.x86_64 2026-03-06T13:43:47.268 INFO:teuthology.orchestra.run.vm01.stdout: grpc-data-1.46.7-10.el9.noarch 2026-03-06T13:43:47.268 INFO:teuthology.orchestra.run.vm01.stdout: ledmon-libs-1.1.0-3.el9.x86_64 2026-03-06T13:43:47.268 INFO:teuthology.orchestra.run.vm01.stdout: libcephsqlite-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 2026-03-06T13:43:47.268 INFO:teuthology.orchestra.run.vm01.stdout: libconfig-1.7.2-9.el9.x86_64 2026-03-06T13:43:47.268 INFO:teuthology.orchestra.run.vm01.stdout: libgfortran-11.5.0-14.el9.x86_64 2026-03-06T13:43:47.268 INFO:teuthology.orchestra.run.vm01.stdout: liboath-2.6.12-1.el9.x86_64 2026-03-06T13:43:47.268 INFO:teuthology.orchestra.run.vm01.stdout: libquadmath-11.5.0-14.el9.x86_64 2026-03-06T13:43:47.268 INFO:teuthology.orchestra.run.vm01.stdout: libradosstriper1-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 2026-03-06T13:43:47.268 INFO:teuthology.orchestra.run.vm01.stdout: libstoragemgmt-1.10.1-1.el9.x86_64 2026-03-06T13:43:47.268 INFO:teuthology.orchestra.run.vm01.stdout: libunwind-1.6.2-1.el9.x86_64 2026-03-06T13:43:47.268 INFO:teuthology.orchestra.run.vm01.stdout: libxslt-1.1.34-12.el9.x86_64 2026-03-06T13:43:47.268 INFO:teuthology.orchestra.run.vm01.stdout: openblas-0.3.29-1.el9.x86_64 2026-03-06T13:43:47.268 INFO:teuthology.orchestra.run.vm01.stdout: openblas-openmp-0.3.29-1.el9.x86_64 2026-03-06T13:43:47.268 INFO:teuthology.orchestra.run.vm01.stdout: pciutils-3.7.0-7.el9.x86_64 2026-03-06T13:43:47.268 INFO:teuthology.orchestra.run.vm01.stdout: protobuf-3.14.0-17.el9.x86_64 2026-03-06T13:43:47.268 INFO:teuthology.orchestra.run.vm01.stdout: protobuf-compiler-3.14.0-17.el9.x86_64 2026-03-06T13:43:47.268 INFO:teuthology.orchestra.run.vm01.stdout: python3-asyncssh-2.13.2-5.el9.noarch 2026-03-06T13:43:47.268 INFO:teuthology.orchestra.run.vm01.stdout: python3-autocommand-2.2.2-8.el9.noarch 2026-03-06T13:43:47.268 INFO:teuthology.orchestra.run.vm01.stdout: python3-babel-2.9.1-2.el9.noarch 2026-03-06T13:43:47.268 INFO:teuthology.orchestra.run.vm01.stdout: python3-backports-tarfile-1.2.0-1.el9.noarch 2026-03-06T13:43:47.268 INFO:teuthology.orchestra.run.vm01.stdout: python3-bcrypt-3.2.2-1.el9.x86_64 2026-03-06T13:43:47.268 INFO:teuthology.orchestra.run.vm01.stdout: python3-cachetools-4.2.4-1.el9.noarch 2026-03-06T13:43:47.268 INFO:teuthology.orchestra.run.vm01.stdout: python3-ceph-common-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 2026-03-06T13:43:47.268 INFO:teuthology.orchestra.run.vm01.stdout: python3-certifi-2023.05.07-4.el9.noarch 2026-03-06T13:43:47.268 INFO:teuthology.orchestra.run.vm01.stdout: python3-cffi-1.14.5-5.el9.x86_64 2026-03-06T13:43:47.268 INFO:teuthology.orchestra.run.vm01.stdout: python3-chardet-4.0.0-5.el9.noarch 2026-03-06T13:43:47.268 INFO:teuthology.orchestra.run.vm01.stdout: python3-cheroot-10.0.1-4.el9.noarch 2026-03-06T13:43:47.268 INFO:teuthology.orchestra.run.vm01.stdout: python3-cherrypy-18.6.1-2.el9.noarch 2026-03-06T13:43:47.268 INFO:teuthology.orchestra.run.vm01.stdout: python3-cryptography-36.0.1-5.el9.x86_64 2026-03-06T13:43:47.268 INFO:teuthology.orchestra.run.vm01.stdout: python3-devel-3.9.25-3.el9.x86_64 2026-03-06T13:43:47.268 INFO:teuthology.orchestra.run.vm01.stdout: python3-google-auth-1:2.45.0-1.el9.noarch 2026-03-06T13:43:47.268 INFO:teuthology.orchestra.run.vm01.stdout: python3-grpcio-1.46.7-10.el9.x86_64 2026-03-06T13:43:47.268 INFO:teuthology.orchestra.run.vm01.stdout: python3-grpcio-tools-1.46.7-10.el9.x86_64 2026-03-06T13:43:47.268 INFO:teuthology.orchestra.run.vm01.stdout: python3-idna-2.10-7.el9.1.noarch 2026-03-06T13:43:47.268 INFO:teuthology.orchestra.run.vm01.stdout: python3-influxdb-5.3.1-1.el9.noarch 2026-03-06T13:43:47.268 INFO:teuthology.orchestra.run.vm01.stdout: python3-isodate-0.6.1-3.el9.noarch 2026-03-06T13:43:47.268 INFO:teuthology.orchestra.run.vm01.stdout: python3-jaraco-8.2.1-3.el9.noarch 2026-03-06T13:43:47.268 INFO:teuthology.orchestra.run.vm01.stdout: python3-jaraco-classes-3.2.1-5.el9.noarch 2026-03-06T13:43:47.268 INFO:teuthology.orchestra.run.vm01.stdout: python3-jaraco-collections-3.0.0-8.el9.noarch 2026-03-06T13:43:47.268 INFO:teuthology.orchestra.run.vm01.stdout: python3-jaraco-context-6.0.1-3.el9.noarch 2026-03-06T13:43:47.268 INFO:teuthology.orchestra.run.vm01.stdout: python3-jaraco-functools-3.5.0-2.el9.noarch 2026-03-06T13:43:47.268 INFO:teuthology.orchestra.run.vm01.stdout: python3-jaraco-text-4.0.0-2.el9.noarch 2026-03-06T13:43:47.268 INFO:teuthology.orchestra.run.vm01.stdout: python3-jinja2-2.11.3-8.el9.noarch 2026-03-06T13:43:47.268 INFO:teuthology.orchestra.run.vm01.stdout: python3-jsonpatch-1.21-16.el9.noarch 2026-03-06T13:43:47.268 INFO:teuthology.orchestra.run.vm01.stdout: python3-jsonpointer-2.0-4.el9.noarch 2026-03-06T13:43:47.268 INFO:teuthology.orchestra.run.vm01.stdout: python3-kubernetes-1:26.1.0-3.el9.noarch 2026-03-06T13:43:47.268 INFO:teuthology.orchestra.run.vm01.stdout: python3-libstoragemgmt-1.10.1-1.el9.x86_64 2026-03-06T13:43:47.268 INFO:teuthology.orchestra.run.vm01.stdout: python3-logutils-0.3.5-21.el9.noarch 2026-03-06T13:43:47.268 INFO:teuthology.orchestra.run.vm01.stdout: python3-lxml-4.6.5-3.el9.x86_64 2026-03-06T13:43:47.268 INFO:teuthology.orchestra.run.vm01.stdout: python3-mako-1.1.4-6.el9.noarch 2026-03-06T13:43:47.268 INFO:teuthology.orchestra.run.vm01.stdout: python3-markupsafe-1.1.1-12.el9.x86_64 2026-03-06T13:43:47.268 INFO:teuthology.orchestra.run.vm01.stdout: python3-more-itertools-8.12.0-2.el9.noarch 2026-03-06T13:43:47.268 INFO:teuthology.orchestra.run.vm01.stdout: python3-msgpack-1.0.3-2.el9.x86_64 2026-03-06T13:43:47.268 INFO:teuthology.orchestra.run.vm01.stdout: python3-natsort-7.1.1-5.el9.noarch 2026-03-06T13:43:47.268 INFO:teuthology.orchestra.run.vm01.stdout: python3-numpy-1:1.23.5-2.el9.x86_64 2026-03-06T13:43:47.268 INFO:teuthology.orchestra.run.vm01.stdout: python3-numpy-f2py-1:1.23.5-2.el9.x86_64 2026-03-06T13:43:47.268 INFO:teuthology.orchestra.run.vm01.stdout: python3-oauthlib-3.1.1-5.el9.noarch 2026-03-06T13:43:47.269 INFO:teuthology.orchestra.run.vm01.stdout: python3-packaging-20.9-5.el9.noarch 2026-03-06T13:43:47.269 INFO:teuthology.orchestra.run.vm01.stdout: python3-pecan-1.4.2-3.el9.noarch 2026-03-06T13:43:47.269 INFO:teuthology.orchestra.run.vm01.stdout: python3-ply-3.11-14.el9.noarch 2026-03-06T13:43:47.269 INFO:teuthology.orchestra.run.vm01.stdout: python3-portend-3.1.0-2.el9.noarch 2026-03-06T13:43:47.269 INFO:teuthology.orchestra.run.vm01.stdout: python3-prettytable-0.7.2-27.el9.noarch 2026-03-06T13:43:47.269 INFO:teuthology.orchestra.run.vm01.stdout: python3-protobuf-3.14.0-17.el9.noarch 2026-03-06T13:43:47.269 INFO:teuthology.orchestra.run.vm01.stdout: python3-pyOpenSSL-21.0.0-1.el9.noarch 2026-03-06T13:43:47.269 INFO:teuthology.orchestra.run.vm01.stdout: python3-pyasn1-0.4.8-7.el9.noarch 2026-03-06T13:43:47.269 INFO:teuthology.orchestra.run.vm01.stdout: python3-pyasn1-modules-0.4.8-7.el9.noarch 2026-03-06T13:43:47.269 INFO:teuthology.orchestra.run.vm01.stdout: python3-pycparser-2.20-6.el9.noarch 2026-03-06T13:43:47.269 INFO:teuthology.orchestra.run.vm01.stdout: python3-pysocks-1.7.1-12.el9.noarch 2026-03-06T13:43:47.269 INFO:teuthology.orchestra.run.vm01.stdout: python3-pytz-2021.1-5.el9.noarch 2026-03-06T13:43:47.269 INFO:teuthology.orchestra.run.vm01.stdout: python3-repoze-lru-0.7-16.el9.noarch 2026-03-06T13:43:47.269 INFO:teuthology.orchestra.run.vm01.stdout: python3-requests-2.25.1-10.el9.noarch 2026-03-06T13:43:47.269 INFO:teuthology.orchestra.run.vm01.stdout: python3-requests-oauthlib-1.3.0-12.el9.noarch 2026-03-06T13:43:47.269 INFO:teuthology.orchestra.run.vm01.stdout: python3-routes-2.5.1-5.el9.noarch 2026-03-06T13:43:47.269 INFO:teuthology.orchestra.run.vm01.stdout: python3-rsa-4.9-2.el9.noarch 2026-03-06T13:43:47.269 INFO:teuthology.orchestra.run.vm01.stdout: python3-saml-1.16.0-1.el9.noarch 2026-03-06T13:43:47.269 INFO:teuthology.orchestra.run.vm01.stdout: python3-scipy-1.9.3-2.el9.x86_64 2026-03-06T13:43:47.269 INFO:teuthology.orchestra.run.vm01.stdout: python3-tempora-5.0.0-2.el9.noarch 2026-03-06T13:43:47.269 INFO:teuthology.orchestra.run.vm01.stdout: python3-toml-0.10.2-6.el9.noarch 2026-03-06T13:43:47.269 INFO:teuthology.orchestra.run.vm01.stdout: python3-typing-extensions-4.15.0-1.el9.noarch 2026-03-06T13:43:47.269 INFO:teuthology.orchestra.run.vm01.stdout: python3-urllib3-1.26.5-7.el9.noarch 2026-03-06T13:43:47.269 INFO:teuthology.orchestra.run.vm01.stdout: python3-webob-1.8.8-2.el9.noarch 2026-03-06T13:43:47.269 INFO:teuthology.orchestra.run.vm01.stdout: python3-websocket-client-1.2.3-2.el9.noarch 2026-03-06T13:43:47.269 INFO:teuthology.orchestra.run.vm01.stdout: python3-werkzeug-2.0.3-3.el9.1.noarch 2026-03-06T13:43:47.269 INFO:teuthology.orchestra.run.vm01.stdout: python3-xmlsec-1.3.13-1.el9.x86_64 2026-03-06T13:43:47.269 INFO:teuthology.orchestra.run.vm01.stdout: python3-zc-lockfile-2.0-10.el9.noarch 2026-03-06T13:43:47.269 INFO:teuthology.orchestra.run.vm01.stdout: qatlib-25.08.0-2.el9.x86_64 2026-03-06T13:43:47.269 INFO:teuthology.orchestra.run.vm01.stdout: qatlib-service-25.08.0-2.el9.x86_64 2026-03-06T13:43:47.269 INFO:teuthology.orchestra.run.vm01.stdout: qatzip-libs-1.3.1-1.el9.x86_64 2026-03-06T13:43:47.269 INFO:teuthology.orchestra.run.vm01.stdout: rbd-mirror-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 2026-03-06T13:43:47.269 INFO:teuthology.orchestra.run.vm01.stdout: smartmontools-1:7.2-10.el9.x86_64 2026-03-06T13:43:47.269 INFO:teuthology.orchestra.run.vm01.stdout: xmlsec1-1.2.29-13.el9.x86_64 2026-03-06T13:43:47.269 INFO:teuthology.orchestra.run.vm01.stdout: xmlsec1-openssl-1.2.29-13.el9.x86_64 2026-03-06T13:43:47.269 INFO:teuthology.orchestra.run.vm01.stdout: 2026-03-06T13:43:47.269 INFO:teuthology.orchestra.run.vm01.stdout:Complete! 2026-03-06T13:43:47.316 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : cephadm-2:19.2.3-47.gc24117fd552.el9.clyso.noarch 1/1 2026-03-06T13:43:47.316 INFO:teuthology.orchestra.run.vm00.stdout: 2026-03-06T13:43:47.316 INFO:teuthology.orchestra.run.vm00.stdout:Removed: 2026-03-06T13:43:47.316 INFO:teuthology.orchestra.run.vm00.stdout: cephadm-2:19.2.3-47.gc24117fd552.el9.clyso.noarch 2026-03-06T13:43:47.316 INFO:teuthology.orchestra.run.vm00.stdout: 2026-03-06T13:43:47.316 INFO:teuthology.orchestra.run.vm00.stdout:Complete! 2026-03-06T13:43:47.449 INFO:teuthology.orchestra.run.vm08.stdout:Dependencies resolved. 2026-03-06T13:43:47.449 INFO:teuthology.orchestra.run.vm08.stdout:================================================================================ 2026-03-06T13:43:47.449 INFO:teuthology.orchestra.run.vm08.stdout: Package Arch Version Repository Size 2026-03-06T13:43:47.449 INFO:teuthology.orchestra.run.vm08.stdout:================================================================================ 2026-03-06T13:43:47.449 INFO:teuthology.orchestra.run.vm08.stdout:Removing: 2026-03-06T13:43:47.449 INFO:teuthology.orchestra.run.vm08.stdout: cephadm noarch 2:19.2.3-47.gc24117fd552.el9.clyso @ceph-noarch 775 k 2026-03-06T13:43:47.449 INFO:teuthology.orchestra.run.vm08.stdout: 2026-03-06T13:43:47.449 INFO:teuthology.orchestra.run.vm08.stdout:Transaction Summary 2026-03-06T13:43:47.449 INFO:teuthology.orchestra.run.vm08.stdout:================================================================================ 2026-03-06T13:43:47.449 INFO:teuthology.orchestra.run.vm08.stdout:Remove 1 Package 2026-03-06T13:43:47.449 INFO:teuthology.orchestra.run.vm08.stdout: 2026-03-06T13:43:47.449 INFO:teuthology.orchestra.run.vm08.stdout:Freed space: 775 k 2026-03-06T13:43:47.450 INFO:teuthology.orchestra.run.vm08.stdout:Running transaction check 2026-03-06T13:43:47.451 INFO:teuthology.orchestra.run.vm08.stdout:Transaction check succeeded. 2026-03-06T13:43:47.451 INFO:teuthology.orchestra.run.vm08.stdout:Running transaction test 2026-03-06T13:43:47.452 INFO:teuthology.orchestra.run.vm08.stdout:Transaction test succeeded. 2026-03-06T13:43:47.453 INFO:teuthology.orchestra.run.vm08.stdout:Running transaction 2026-03-06T13:43:47.469 INFO:teuthology.orchestra.run.vm08.stdout: Preparing : 1/1 2026-03-06T13:43:47.469 INFO:teuthology.orchestra.run.vm08.stdout: Erasing : cephadm-2:19.2.3-47.gc24117fd552.el9.clyso.noarch 1/1 2026-03-06T13:43:47.496 INFO:teuthology.orchestra.run.vm01.stdout:Dependencies resolved. 2026-03-06T13:43:47.497 INFO:teuthology.orchestra.run.vm01.stdout:================================================================================ 2026-03-06T13:43:47.497 INFO:teuthology.orchestra.run.vm01.stdout: Package Arch Version Repository Size 2026-03-06T13:43:47.497 INFO:teuthology.orchestra.run.vm01.stdout:================================================================================ 2026-03-06T13:43:47.497 INFO:teuthology.orchestra.run.vm01.stdout:Removing: 2026-03-06T13:43:47.497 INFO:teuthology.orchestra.run.vm01.stdout: cephadm noarch 2:19.2.3-47.gc24117fd552.el9.clyso @ceph-noarch 775 k 2026-03-06T13:43:47.497 INFO:teuthology.orchestra.run.vm01.stdout: 2026-03-06T13:43:47.497 INFO:teuthology.orchestra.run.vm01.stdout:Transaction Summary 2026-03-06T13:43:47.497 INFO:teuthology.orchestra.run.vm01.stdout:================================================================================ 2026-03-06T13:43:47.497 INFO:teuthology.orchestra.run.vm01.stdout:Remove 1 Package 2026-03-06T13:43:47.497 INFO:teuthology.orchestra.run.vm01.stdout: 2026-03-06T13:43:47.497 INFO:teuthology.orchestra.run.vm01.stdout:Freed space: 775 k 2026-03-06T13:43:47.497 INFO:teuthology.orchestra.run.vm01.stdout:Running transaction check 2026-03-06T13:43:47.499 INFO:teuthology.orchestra.run.vm01.stdout:Transaction check succeeded. 2026-03-06T13:43:47.499 INFO:teuthology.orchestra.run.vm01.stdout:Running transaction test 2026-03-06T13:43:47.501 INFO:teuthology.orchestra.run.vm01.stdout:Transaction test succeeded. 2026-03-06T13:43:47.501 INFO:teuthology.orchestra.run.vm01.stdout:Running transaction 2026-03-06T13:43:47.510 INFO:teuthology.orchestra.run.vm00.stdout:No match for argument: ceph-immutable-object-cache 2026-03-06T13:43:47.510 INFO:teuthology.orchestra.run.vm00.stderr:No packages marked for removal. 2026-03-06T13:43:47.513 INFO:teuthology.orchestra.run.vm00.stdout:Dependencies resolved. 2026-03-06T13:43:47.514 INFO:teuthology.orchestra.run.vm00.stdout:Nothing to do. 2026-03-06T13:43:47.514 INFO:teuthology.orchestra.run.vm00.stdout:Complete! 2026-03-06T13:43:47.518 INFO:teuthology.orchestra.run.vm01.stdout: Preparing : 1/1 2026-03-06T13:43:47.518 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : cephadm-2:19.2.3-47.gc24117fd552.el9.clyso.noarch 1/1 2026-03-06T13:43:47.589 INFO:teuthology.orchestra.run.vm08.stdout: Running scriptlet: cephadm-2:19.2.3-47.gc24117fd552.el9.clyso.noarch 1/1 2026-03-06T13:43:47.628 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : cephadm-2:19.2.3-47.gc24117fd552.el9.clyso.noarch 1/1 2026-03-06T13:43:47.628 INFO:teuthology.orchestra.run.vm08.stdout: 2026-03-06T13:43:47.628 INFO:teuthology.orchestra.run.vm08.stdout:Removed: 2026-03-06T13:43:47.628 INFO:teuthology.orchestra.run.vm08.stdout: cephadm-2:19.2.3-47.gc24117fd552.el9.clyso.noarch 2026-03-06T13:43:47.628 INFO:teuthology.orchestra.run.vm08.stdout: 2026-03-06T13:43:47.628 INFO:teuthology.orchestra.run.vm08.stdout:Complete! 2026-03-06T13:43:47.656 INFO:teuthology.orchestra.run.vm01.stdout: Running scriptlet: cephadm-2:19.2.3-47.gc24117fd552.el9.clyso.noarch 1/1 2026-03-06T13:43:47.686 INFO:teuthology.orchestra.run.vm00.stdout:No match for argument: ceph-mgr 2026-03-06T13:43:47.686 INFO:teuthology.orchestra.run.vm00.stderr:No packages marked for removal. 2026-03-06T13:43:47.688 INFO:teuthology.orchestra.run.vm00.stdout:Dependencies resolved. 2026-03-06T13:43:47.688 INFO:teuthology.orchestra.run.vm00.stdout:Nothing to do. 2026-03-06T13:43:47.688 INFO:teuthology.orchestra.run.vm00.stdout:Complete! 2026-03-06T13:43:47.698 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : cephadm-2:19.2.3-47.gc24117fd552.el9.clyso.noarch 1/1 2026-03-06T13:43:47.698 INFO:teuthology.orchestra.run.vm01.stdout: 2026-03-06T13:43:47.698 INFO:teuthology.orchestra.run.vm01.stdout:Removed: 2026-03-06T13:43:47.698 INFO:teuthology.orchestra.run.vm01.stdout: cephadm-2:19.2.3-47.gc24117fd552.el9.clyso.noarch 2026-03-06T13:43:47.698 INFO:teuthology.orchestra.run.vm01.stdout: 2026-03-06T13:43:47.698 INFO:teuthology.orchestra.run.vm01.stdout:Complete! 2026-03-06T13:43:47.804 INFO:teuthology.orchestra.run.vm08.stdout:No match for argument: ceph-immutable-object-cache 2026-03-06T13:43:47.804 INFO:teuthology.orchestra.run.vm08.stderr:No packages marked for removal. 2026-03-06T13:43:47.806 INFO:teuthology.orchestra.run.vm08.stdout:Dependencies resolved. 2026-03-06T13:43:47.806 INFO:teuthology.orchestra.run.vm08.stdout:Nothing to do. 2026-03-06T13:43:47.806 INFO:teuthology.orchestra.run.vm08.stdout:Complete! 2026-03-06T13:43:47.859 INFO:teuthology.orchestra.run.vm00.stdout:No match for argument: ceph-mgr-dashboard 2026-03-06T13:43:47.860 INFO:teuthology.orchestra.run.vm00.stderr:No packages marked for removal. 2026-03-06T13:43:47.862 INFO:teuthology.orchestra.run.vm00.stdout:Dependencies resolved. 2026-03-06T13:43:47.862 INFO:teuthology.orchestra.run.vm00.stdout:Nothing to do. 2026-03-06T13:43:47.862 INFO:teuthology.orchestra.run.vm00.stdout:Complete! 2026-03-06T13:43:47.891 INFO:teuthology.orchestra.run.vm01.stdout:No match for argument: ceph-immutable-object-cache 2026-03-06T13:43:47.891 INFO:teuthology.orchestra.run.vm01.stderr:No packages marked for removal. 2026-03-06T13:43:47.893 INFO:teuthology.orchestra.run.vm01.stdout:Dependencies resolved. 2026-03-06T13:43:47.893 INFO:teuthology.orchestra.run.vm01.stdout:Nothing to do. 2026-03-06T13:43:47.893 INFO:teuthology.orchestra.run.vm01.stdout:Complete! 2026-03-06T13:43:47.975 INFO:teuthology.orchestra.run.vm08.stdout:No match for argument: ceph-mgr 2026-03-06T13:43:47.976 INFO:teuthology.orchestra.run.vm08.stderr:No packages marked for removal. 2026-03-06T13:43:47.977 INFO:teuthology.orchestra.run.vm08.stdout:Dependencies resolved. 2026-03-06T13:43:47.978 INFO:teuthology.orchestra.run.vm08.stdout:Nothing to do. 2026-03-06T13:43:47.978 INFO:teuthology.orchestra.run.vm08.stdout:Complete! 2026-03-06T13:43:48.042 INFO:teuthology.orchestra.run.vm00.stdout:No match for argument: ceph-mgr-diskprediction-local 2026-03-06T13:43:48.042 INFO:teuthology.orchestra.run.vm00.stderr:No packages marked for removal. 2026-03-06T13:43:48.044 INFO:teuthology.orchestra.run.vm00.stdout:Dependencies resolved. 2026-03-06T13:43:48.045 INFO:teuthology.orchestra.run.vm00.stdout:Nothing to do. 2026-03-06T13:43:48.045 INFO:teuthology.orchestra.run.vm00.stdout:Complete! 2026-03-06T13:43:48.075 INFO:teuthology.orchestra.run.vm01.stdout:No match for argument: ceph-mgr 2026-03-06T13:43:48.075 INFO:teuthology.orchestra.run.vm01.stderr:No packages marked for removal. 2026-03-06T13:43:48.077 INFO:teuthology.orchestra.run.vm01.stdout:Dependencies resolved. 2026-03-06T13:43:48.077 INFO:teuthology.orchestra.run.vm01.stdout:Nothing to do. 2026-03-06T13:43:48.077 INFO:teuthology.orchestra.run.vm01.stdout:Complete! 2026-03-06T13:43:48.140 INFO:teuthology.orchestra.run.vm08.stdout:No match for argument: ceph-mgr-dashboard 2026-03-06T13:43:48.140 INFO:teuthology.orchestra.run.vm08.stderr:No packages marked for removal. 2026-03-06T13:43:48.142 INFO:teuthology.orchestra.run.vm08.stdout:Dependencies resolved. 2026-03-06T13:43:48.143 INFO:teuthology.orchestra.run.vm08.stdout:Nothing to do. 2026-03-06T13:43:48.143 INFO:teuthology.orchestra.run.vm08.stdout:Complete! 2026-03-06T13:43:48.220 INFO:teuthology.orchestra.run.vm00.stdout:No match for argument: ceph-mgr-rook 2026-03-06T13:43:48.220 INFO:teuthology.orchestra.run.vm00.stderr:No packages marked for removal. 2026-03-06T13:43:48.222 INFO:teuthology.orchestra.run.vm00.stdout:Dependencies resolved. 2026-03-06T13:43:48.223 INFO:teuthology.orchestra.run.vm00.stdout:Nothing to do. 2026-03-06T13:43:48.223 INFO:teuthology.orchestra.run.vm00.stdout:Complete! 2026-03-06T13:43:48.240 INFO:teuthology.orchestra.run.vm01.stdout:No match for argument: ceph-mgr-dashboard 2026-03-06T13:43:48.240 INFO:teuthology.orchestra.run.vm01.stderr:No packages marked for removal. 2026-03-06T13:43:48.242 INFO:teuthology.orchestra.run.vm01.stdout:Dependencies resolved. 2026-03-06T13:43:48.242 INFO:teuthology.orchestra.run.vm01.stdout:Nothing to do. 2026-03-06T13:43:48.242 INFO:teuthology.orchestra.run.vm01.stdout:Complete! 2026-03-06T13:43:48.306 INFO:teuthology.orchestra.run.vm08.stdout:No match for argument: ceph-mgr-diskprediction-local 2026-03-06T13:43:48.306 INFO:teuthology.orchestra.run.vm08.stderr:No packages marked for removal. 2026-03-06T13:43:48.308 INFO:teuthology.orchestra.run.vm08.stdout:Dependencies resolved. 2026-03-06T13:43:48.309 INFO:teuthology.orchestra.run.vm08.stdout:Nothing to do. 2026-03-06T13:43:48.309 INFO:teuthology.orchestra.run.vm08.stdout:Complete! 2026-03-06T13:43:48.396 INFO:teuthology.orchestra.run.vm00.stdout:No match for argument: ceph-mgr-cephadm 2026-03-06T13:43:48.396 INFO:teuthology.orchestra.run.vm00.stderr:No packages marked for removal. 2026-03-06T13:43:48.398 INFO:teuthology.orchestra.run.vm00.stdout:Dependencies resolved. 2026-03-06T13:43:48.399 INFO:teuthology.orchestra.run.vm00.stdout:Nothing to do. 2026-03-06T13:43:48.399 INFO:teuthology.orchestra.run.vm00.stdout:Complete! 2026-03-06T13:43:48.421 INFO:teuthology.orchestra.run.vm01.stdout:No match for argument: ceph-mgr-diskprediction-local 2026-03-06T13:43:48.421 INFO:teuthology.orchestra.run.vm01.stderr:No packages marked for removal. 2026-03-06T13:43:48.423 INFO:teuthology.orchestra.run.vm01.stdout:Dependencies resolved. 2026-03-06T13:43:48.424 INFO:teuthology.orchestra.run.vm01.stdout:Nothing to do. 2026-03-06T13:43:48.424 INFO:teuthology.orchestra.run.vm01.stdout:Complete! 2026-03-06T13:43:48.468 INFO:teuthology.orchestra.run.vm08.stdout:No match for argument: ceph-mgr-rook 2026-03-06T13:43:48.468 INFO:teuthology.orchestra.run.vm08.stderr:No packages marked for removal. 2026-03-06T13:43:48.470 INFO:teuthology.orchestra.run.vm08.stdout:Dependencies resolved. 2026-03-06T13:43:48.471 INFO:teuthology.orchestra.run.vm08.stdout:Nothing to do. 2026-03-06T13:43:48.471 INFO:teuthology.orchestra.run.vm08.stdout:Complete! 2026-03-06T13:43:48.579 INFO:teuthology.orchestra.run.vm00.stdout:Dependencies resolved. 2026-03-06T13:43:48.579 INFO:teuthology.orchestra.run.vm00.stdout:================================================================================ 2026-03-06T13:43:48.579 INFO:teuthology.orchestra.run.vm00.stdout: Package Arch Version Repository Size 2026-03-06T13:43:48.580 INFO:teuthology.orchestra.run.vm00.stdout:================================================================================ 2026-03-06T13:43:48.580 INFO:teuthology.orchestra.run.vm00.stdout:Removing: 2026-03-06T13:43:48.580 INFO:teuthology.orchestra.run.vm00.stdout: ceph-fuse x86_64 2:19.2.3-47.gc24117fd552.el9.clyso @ceph 3.6 M 2026-03-06T13:43:48.580 INFO:teuthology.orchestra.run.vm00.stdout:Removing unused dependencies: 2026-03-06T13:43:48.580 INFO:teuthology.orchestra.run.vm00.stdout: fuse x86_64 2.9.9-17.el9 @baseos 214 k 2026-03-06T13:43:48.580 INFO:teuthology.orchestra.run.vm00.stdout: 2026-03-06T13:43:48.580 INFO:teuthology.orchestra.run.vm00.stdout:Transaction Summary 2026-03-06T13:43:48.580 INFO:teuthology.orchestra.run.vm00.stdout:================================================================================ 2026-03-06T13:43:48.580 INFO:teuthology.orchestra.run.vm00.stdout:Remove 2 Packages 2026-03-06T13:43:48.580 INFO:teuthology.orchestra.run.vm00.stdout: 2026-03-06T13:43:48.580 INFO:teuthology.orchestra.run.vm00.stdout:Freed space: 3.8 M 2026-03-06T13:43:48.580 INFO:teuthology.orchestra.run.vm00.stdout:Running transaction check 2026-03-06T13:43:48.582 INFO:teuthology.orchestra.run.vm00.stdout:Transaction check succeeded. 2026-03-06T13:43:48.582 INFO:teuthology.orchestra.run.vm00.stdout:Running transaction test 2026-03-06T13:43:48.595 INFO:teuthology.orchestra.run.vm01.stdout:No match for argument: ceph-mgr-rook 2026-03-06T13:43:48.595 INFO:teuthology.orchestra.run.vm01.stderr:No packages marked for removal. 2026-03-06T13:43:48.595 INFO:teuthology.orchestra.run.vm00.stdout:Transaction test succeeded. 2026-03-06T13:43:48.596 INFO:teuthology.orchestra.run.vm00.stdout:Running transaction 2026-03-06T13:43:48.597 INFO:teuthology.orchestra.run.vm01.stdout:Dependencies resolved. 2026-03-06T13:43:48.598 INFO:teuthology.orchestra.run.vm01.stdout:Nothing to do. 2026-03-06T13:43:48.598 INFO:teuthology.orchestra.run.vm01.stdout:Complete! 2026-03-06T13:43:48.625 INFO:teuthology.orchestra.run.vm00.stdout: Preparing : 1/1 2026-03-06T13:43:48.629 INFO:teuthology.orchestra.run.vm00.stdout: Erasing : ceph-fuse-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 1/2 2026-03-06T13:43:48.638 INFO:teuthology.orchestra.run.vm08.stdout:No match for argument: ceph-mgr-cephadm 2026-03-06T13:43:48.638 INFO:teuthology.orchestra.run.vm08.stderr:No packages marked for removal. 2026-03-06T13:43:48.640 INFO:teuthology.orchestra.run.vm08.stdout:Dependencies resolved. 2026-03-06T13:43:48.641 INFO:teuthology.orchestra.run.vm08.stdout:Nothing to do. 2026-03-06T13:43:48.641 INFO:teuthology.orchestra.run.vm08.stdout:Complete! 2026-03-06T13:43:48.644 INFO:teuthology.orchestra.run.vm00.stdout: Erasing : fuse-2.9.9-17.el9.x86_64 2/2 2026-03-06T13:43:48.716 INFO:teuthology.orchestra.run.vm00.stdout: Running scriptlet: fuse-2.9.9-17.el9.x86_64 2/2 2026-03-06T13:43:48.716 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : ceph-fuse-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 1/2 2026-03-06T13:43:48.758 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : fuse-2.9.9-17.el9.x86_64 2/2 2026-03-06T13:43:48.758 INFO:teuthology.orchestra.run.vm00.stdout: 2026-03-06T13:43:48.758 INFO:teuthology.orchestra.run.vm00.stdout:Removed: 2026-03-06T13:43:48.758 INFO:teuthology.orchestra.run.vm00.stdout: ceph-fuse-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 fuse-2.9.9-17.el9.x86_64 2026-03-06T13:43:48.758 INFO:teuthology.orchestra.run.vm00.stdout: 2026-03-06T13:43:48.758 INFO:teuthology.orchestra.run.vm00.stdout:Complete! 2026-03-06T13:43:48.772 INFO:teuthology.orchestra.run.vm01.stdout:No match for argument: ceph-mgr-cephadm 2026-03-06T13:43:48.773 INFO:teuthology.orchestra.run.vm01.stderr:No packages marked for removal. 2026-03-06T13:43:48.774 INFO:teuthology.orchestra.run.vm01.stdout:Dependencies resolved. 2026-03-06T13:43:48.775 INFO:teuthology.orchestra.run.vm01.stdout:Nothing to do. 2026-03-06T13:43:48.775 INFO:teuthology.orchestra.run.vm01.stdout:Complete! 2026-03-06T13:43:48.827 INFO:teuthology.orchestra.run.vm08.stdout:Dependencies resolved. 2026-03-06T13:43:48.828 INFO:teuthology.orchestra.run.vm08.stdout:================================================================================ 2026-03-06T13:43:48.828 INFO:teuthology.orchestra.run.vm08.stdout: Package Arch Version Repository Size 2026-03-06T13:43:48.828 INFO:teuthology.orchestra.run.vm08.stdout:================================================================================ 2026-03-06T13:43:48.828 INFO:teuthology.orchestra.run.vm08.stdout:Removing: 2026-03-06T13:43:48.828 INFO:teuthology.orchestra.run.vm08.stdout: ceph-fuse x86_64 2:19.2.3-47.gc24117fd552.el9.clyso @ceph 3.6 M 2026-03-06T13:43:48.828 INFO:teuthology.orchestra.run.vm08.stdout:Removing unused dependencies: 2026-03-06T13:43:48.828 INFO:teuthology.orchestra.run.vm08.stdout: fuse x86_64 2.9.9-17.el9 @baseos 214 k 2026-03-06T13:43:48.828 INFO:teuthology.orchestra.run.vm08.stdout: 2026-03-06T13:43:48.828 INFO:teuthology.orchestra.run.vm08.stdout:Transaction Summary 2026-03-06T13:43:48.828 INFO:teuthology.orchestra.run.vm08.stdout:================================================================================ 2026-03-06T13:43:48.828 INFO:teuthology.orchestra.run.vm08.stdout:Remove 2 Packages 2026-03-06T13:43:48.828 INFO:teuthology.orchestra.run.vm08.stdout: 2026-03-06T13:43:48.828 INFO:teuthology.orchestra.run.vm08.stdout:Freed space: 3.8 M 2026-03-06T13:43:48.828 INFO:teuthology.orchestra.run.vm08.stdout:Running transaction check 2026-03-06T13:43:48.832 INFO:teuthology.orchestra.run.vm08.stdout:Transaction check succeeded. 2026-03-06T13:43:48.832 INFO:teuthology.orchestra.run.vm08.stdout:Running transaction test 2026-03-06T13:43:48.845 INFO:teuthology.orchestra.run.vm08.stdout:Transaction test succeeded. 2026-03-06T13:43:48.846 INFO:teuthology.orchestra.run.vm08.stdout:Running transaction 2026-03-06T13:43:48.875 INFO:teuthology.orchestra.run.vm08.stdout: Preparing : 1/1 2026-03-06T13:43:48.878 INFO:teuthology.orchestra.run.vm08.stdout: Erasing : ceph-fuse-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 1/2 2026-03-06T13:43:48.892 INFO:teuthology.orchestra.run.vm08.stdout: Erasing : fuse-2.9.9-17.el9.x86_64 2/2 2026-03-06T13:43:48.954 INFO:teuthology.orchestra.run.vm08.stdout: Running scriptlet: fuse-2.9.9-17.el9.x86_64 2/2 2026-03-06T13:43:48.954 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : ceph-fuse-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 1/2 2026-03-06T13:43:48.955 INFO:teuthology.orchestra.run.vm00.stdout:No match for argument: ceph-volume 2026-03-06T13:43:48.955 INFO:teuthology.orchestra.run.vm00.stderr:No packages marked for removal. 2026-03-06T13:43:48.957 INFO:teuthology.orchestra.run.vm00.stdout:Dependencies resolved. 2026-03-06T13:43:48.958 INFO:teuthology.orchestra.run.vm00.stdout:Nothing to do. 2026-03-06T13:43:48.958 INFO:teuthology.orchestra.run.vm00.stdout:Complete! 2026-03-06T13:43:48.959 INFO:teuthology.orchestra.run.vm01.stdout:Dependencies resolved. 2026-03-06T13:43:48.960 INFO:teuthology.orchestra.run.vm01.stdout:================================================================================ 2026-03-06T13:43:48.960 INFO:teuthology.orchestra.run.vm01.stdout: Package Arch Version Repository Size 2026-03-06T13:43:48.960 INFO:teuthology.orchestra.run.vm01.stdout:================================================================================ 2026-03-06T13:43:48.960 INFO:teuthology.orchestra.run.vm01.stdout:Removing: 2026-03-06T13:43:48.960 INFO:teuthology.orchestra.run.vm01.stdout: ceph-fuse x86_64 2:19.2.3-47.gc24117fd552.el9.clyso @ceph 3.6 M 2026-03-06T13:43:48.960 INFO:teuthology.orchestra.run.vm01.stdout:Removing unused dependencies: 2026-03-06T13:43:48.960 INFO:teuthology.orchestra.run.vm01.stdout: fuse x86_64 2.9.9-17.el9 @baseos 214 k 2026-03-06T13:43:48.960 INFO:teuthology.orchestra.run.vm01.stdout: 2026-03-06T13:43:48.960 INFO:teuthology.orchestra.run.vm01.stdout:Transaction Summary 2026-03-06T13:43:48.960 INFO:teuthology.orchestra.run.vm01.stdout:================================================================================ 2026-03-06T13:43:48.960 INFO:teuthology.orchestra.run.vm01.stdout:Remove 2 Packages 2026-03-06T13:43:48.960 INFO:teuthology.orchestra.run.vm01.stdout: 2026-03-06T13:43:48.960 INFO:teuthology.orchestra.run.vm01.stdout:Freed space: 3.8 M 2026-03-06T13:43:48.960 INFO:teuthology.orchestra.run.vm01.stdout:Running transaction check 2026-03-06T13:43:48.963 INFO:teuthology.orchestra.run.vm01.stdout:Transaction check succeeded. 2026-03-06T13:43:48.963 INFO:teuthology.orchestra.run.vm01.stdout:Running transaction test 2026-03-06T13:43:48.977 INFO:teuthology.orchestra.run.vm01.stdout:Transaction test succeeded. 2026-03-06T13:43:48.977 INFO:teuthology.orchestra.run.vm01.stdout:Running transaction 2026-03-06T13:43:48.995 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : fuse-2.9.9-17.el9.x86_64 2/2 2026-03-06T13:43:48.995 INFO:teuthology.orchestra.run.vm08.stdout: 2026-03-06T13:43:48.995 INFO:teuthology.orchestra.run.vm08.stdout:Removed: 2026-03-06T13:43:48.995 INFO:teuthology.orchestra.run.vm08.stdout: ceph-fuse-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 fuse-2.9.9-17.el9.x86_64 2026-03-06T13:43:48.995 INFO:teuthology.orchestra.run.vm08.stdout: 2026-03-06T13:43:48.995 INFO:teuthology.orchestra.run.vm08.stdout:Complete! 2026-03-06T13:43:49.007 INFO:teuthology.orchestra.run.vm01.stdout: Preparing : 1/1 2026-03-06T13:43:49.011 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : ceph-fuse-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 1/2 2026-03-06T13:43:49.024 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : fuse-2.9.9-17.el9.x86_64 2/2 2026-03-06T13:43:49.093 INFO:teuthology.orchestra.run.vm01.stdout: Running scriptlet: fuse-2.9.9-17.el9.x86_64 2/2 2026-03-06T13:43:49.093 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : ceph-fuse-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 1/2 2026-03-06T13:43:49.137 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : fuse-2.9.9-17.el9.x86_64 2/2 2026-03-06T13:43:49.137 INFO:teuthology.orchestra.run.vm01.stdout: 2026-03-06T13:43:49.137 INFO:teuthology.orchestra.run.vm01.stdout:Removed: 2026-03-06T13:43:49.137 INFO:teuthology.orchestra.run.vm01.stdout: ceph-fuse-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 fuse-2.9.9-17.el9.x86_64 2026-03-06T13:43:49.137 INFO:teuthology.orchestra.run.vm01.stdout: 2026-03-06T13:43:49.137 INFO:teuthology.orchestra.run.vm01.stdout:Complete! 2026-03-06T13:43:49.152 INFO:teuthology.orchestra.run.vm00.stdout:Dependencies resolved. 2026-03-06T13:43:49.153 INFO:teuthology.orchestra.run.vm00.stdout:================================================================================ 2026-03-06T13:43:49.153 INFO:teuthology.orchestra.run.vm00.stdout: Package Arch Version Repo Size 2026-03-06T13:43:49.153 INFO:teuthology.orchestra.run.vm00.stdout:================================================================================ 2026-03-06T13:43:49.153 INFO:teuthology.orchestra.run.vm00.stdout:Removing: 2026-03-06T13:43:49.153 INFO:teuthology.orchestra.run.vm00.stdout: librados-devel x86_64 2:19.2.3-47.gc24117fd552.el9.clyso @ceph 456 k 2026-03-06T13:43:49.153 INFO:teuthology.orchestra.run.vm00.stdout:Removing dependent packages: 2026-03-06T13:43:49.153 INFO:teuthology.orchestra.run.vm00.stdout: libcephfs-devel x86_64 2:19.2.3-47.gc24117fd552.el9.clyso @ceph 153 k 2026-03-06T13:43:49.153 INFO:teuthology.orchestra.run.vm00.stdout: 2026-03-06T13:43:49.153 INFO:teuthology.orchestra.run.vm00.stdout:Transaction Summary 2026-03-06T13:43:49.153 INFO:teuthology.orchestra.run.vm00.stdout:================================================================================ 2026-03-06T13:43:49.153 INFO:teuthology.orchestra.run.vm00.stdout:Remove 2 Packages 2026-03-06T13:43:49.153 INFO:teuthology.orchestra.run.vm00.stdout: 2026-03-06T13:43:49.153 INFO:teuthology.orchestra.run.vm00.stdout:Freed space: 610 k 2026-03-06T13:43:49.153 INFO:teuthology.orchestra.run.vm00.stdout:Running transaction check 2026-03-06T13:43:49.155 INFO:teuthology.orchestra.run.vm00.stdout:Transaction check succeeded. 2026-03-06T13:43:49.155 INFO:teuthology.orchestra.run.vm00.stdout:Running transaction test 2026-03-06T13:43:49.165 INFO:teuthology.orchestra.run.vm00.stdout:Transaction test succeeded. 2026-03-06T13:43:49.165 INFO:teuthology.orchestra.run.vm00.stdout:Running transaction 2026-03-06T13:43:49.183 INFO:teuthology.orchestra.run.vm08.stdout:No match for argument: ceph-volume 2026-03-06T13:43:49.183 INFO:teuthology.orchestra.run.vm08.stderr:No packages marked for removal. 2026-03-06T13:43:49.185 INFO:teuthology.orchestra.run.vm08.stdout:Dependencies resolved. 2026-03-06T13:43:49.186 INFO:teuthology.orchestra.run.vm08.stdout:Nothing to do. 2026-03-06T13:43:49.186 INFO:teuthology.orchestra.run.vm08.stdout:Complete! 2026-03-06T13:43:49.194 INFO:teuthology.orchestra.run.vm00.stdout: Preparing : 1/1 2026-03-06T13:43:49.196 INFO:teuthology.orchestra.run.vm00.stdout: Erasing : libcephfs-devel-2:19.2.3-47.gc24117fd552.el9.clyso.x 1/2 2026-03-06T13:43:49.210 INFO:teuthology.orchestra.run.vm00.stdout: Erasing : librados-devel-2:19.2.3-47.gc24117fd552.el9.clyso.x8 2/2 2026-03-06T13:43:49.273 INFO:teuthology.orchestra.run.vm00.stdout: Running scriptlet: librados-devel-2:19.2.3-47.gc24117fd552.el9.clyso.x8 2/2 2026-03-06T13:43:49.273 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : libcephfs-devel-2:19.2.3-47.gc24117fd552.el9.clyso.x 1/2 2026-03-06T13:43:49.319 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : librados-devel-2:19.2.3-47.gc24117fd552.el9.clyso.x8 2/2 2026-03-06T13:43:49.319 INFO:teuthology.orchestra.run.vm00.stdout: 2026-03-06T13:43:49.319 INFO:teuthology.orchestra.run.vm00.stdout:Removed: 2026-03-06T13:43:49.319 INFO:teuthology.orchestra.run.vm00.stdout: libcephfs-devel-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 2026-03-06T13:43:49.319 INFO:teuthology.orchestra.run.vm00.stdout: librados-devel-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 2026-03-06T13:43:49.319 INFO:teuthology.orchestra.run.vm00.stdout: 2026-03-06T13:43:49.319 INFO:teuthology.orchestra.run.vm00.stdout:Complete! 2026-03-06T13:43:49.344 INFO:teuthology.orchestra.run.vm01.stdout:No match for argument: ceph-volume 2026-03-06T13:43:49.344 INFO:teuthology.orchestra.run.vm01.stderr:No packages marked for removal. 2026-03-06T13:43:49.346 INFO:teuthology.orchestra.run.vm01.stdout:Dependencies resolved. 2026-03-06T13:43:49.347 INFO:teuthology.orchestra.run.vm01.stdout:Nothing to do. 2026-03-06T13:43:49.347 INFO:teuthology.orchestra.run.vm01.stdout:Complete! 2026-03-06T13:43:49.378 INFO:teuthology.orchestra.run.vm08.stdout:Dependencies resolved. 2026-03-06T13:43:49.378 INFO:teuthology.orchestra.run.vm08.stdout:================================================================================ 2026-03-06T13:43:49.378 INFO:teuthology.orchestra.run.vm08.stdout: Package Arch Version Repo Size 2026-03-06T13:43:49.378 INFO:teuthology.orchestra.run.vm08.stdout:================================================================================ 2026-03-06T13:43:49.378 INFO:teuthology.orchestra.run.vm08.stdout:Removing: 2026-03-06T13:43:49.378 INFO:teuthology.orchestra.run.vm08.stdout: librados-devel x86_64 2:19.2.3-47.gc24117fd552.el9.clyso @ceph 456 k 2026-03-06T13:43:49.378 INFO:teuthology.orchestra.run.vm08.stdout:Removing dependent packages: 2026-03-06T13:43:49.378 INFO:teuthology.orchestra.run.vm08.stdout: libcephfs-devel x86_64 2:19.2.3-47.gc24117fd552.el9.clyso @ceph 153 k 2026-03-06T13:43:49.378 INFO:teuthology.orchestra.run.vm08.stdout: 2026-03-06T13:43:49.378 INFO:teuthology.orchestra.run.vm08.stdout:Transaction Summary 2026-03-06T13:43:49.378 INFO:teuthology.orchestra.run.vm08.stdout:================================================================================ 2026-03-06T13:43:49.378 INFO:teuthology.orchestra.run.vm08.stdout:Remove 2 Packages 2026-03-06T13:43:49.378 INFO:teuthology.orchestra.run.vm08.stdout: 2026-03-06T13:43:49.378 INFO:teuthology.orchestra.run.vm08.stdout:Freed space: 610 k 2026-03-06T13:43:49.379 INFO:teuthology.orchestra.run.vm08.stdout:Running transaction check 2026-03-06T13:43:49.380 INFO:teuthology.orchestra.run.vm08.stdout:Transaction check succeeded. 2026-03-06T13:43:49.380 INFO:teuthology.orchestra.run.vm08.stdout:Running transaction test 2026-03-06T13:43:49.390 INFO:teuthology.orchestra.run.vm08.stdout:Transaction test succeeded. 2026-03-06T13:43:49.391 INFO:teuthology.orchestra.run.vm08.stdout:Running transaction 2026-03-06T13:43:49.416 INFO:teuthology.orchestra.run.vm08.stdout: Preparing : 1/1 2026-03-06T13:43:49.418 INFO:teuthology.orchestra.run.vm08.stdout: Erasing : libcephfs-devel-2:19.2.3-47.gc24117fd552.el9.clyso.x 1/2 2026-03-06T13:43:49.432 INFO:teuthology.orchestra.run.vm08.stdout: Erasing : librados-devel-2:19.2.3-47.gc24117fd552.el9.clyso.x8 2/2 2026-03-06T13:43:49.492 INFO:teuthology.orchestra.run.vm08.stdout: Running scriptlet: librados-devel-2:19.2.3-47.gc24117fd552.el9.clyso.x8 2/2 2026-03-06T13:43:49.492 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : libcephfs-devel-2:19.2.3-47.gc24117fd552.el9.clyso.x 1/2 2026-03-06T13:43:49.513 INFO:teuthology.orchestra.run.vm00.stdout:Dependencies resolved. 2026-03-06T13:43:49.514 INFO:teuthology.orchestra.run.vm00.stdout:================================================================================ 2026-03-06T13:43:49.514 INFO:teuthology.orchestra.run.vm00.stdout: Package Arch Version Repo Size 2026-03-06T13:43:49.514 INFO:teuthology.orchestra.run.vm00.stdout:================================================================================ 2026-03-06T13:43:49.514 INFO:teuthology.orchestra.run.vm00.stdout:Removing: 2026-03-06T13:43:49.514 INFO:teuthology.orchestra.run.vm00.stdout: libcephfs2 x86_64 2:19.2.3-47.gc24117fd552.el9.clyso @ceph 3.0 M 2026-03-06T13:43:49.514 INFO:teuthology.orchestra.run.vm00.stdout:Removing dependent packages: 2026-03-06T13:43:49.514 INFO:teuthology.orchestra.run.vm00.stdout: python3-cephfs x86_64 2:19.2.3-47.gc24117fd552.el9.clyso @ceph 514 k 2026-03-06T13:43:49.514 INFO:teuthology.orchestra.run.vm00.stdout:Removing unused dependencies: 2026-03-06T13:43:49.514 INFO:teuthology.orchestra.run.vm00.stdout: python3-ceph-argparse x86_64 2:19.2.3-47.gc24117fd552.el9.clyso @ceph 187 k 2026-03-06T13:43:49.514 INFO:teuthology.orchestra.run.vm00.stdout: 2026-03-06T13:43:49.514 INFO:teuthology.orchestra.run.vm00.stdout:Transaction Summary 2026-03-06T13:43:49.514 INFO:teuthology.orchestra.run.vm00.stdout:================================================================================ 2026-03-06T13:43:49.514 INFO:teuthology.orchestra.run.vm00.stdout:Remove 3 Packages 2026-03-06T13:43:49.514 INFO:teuthology.orchestra.run.vm00.stdout: 2026-03-06T13:43:49.514 INFO:teuthology.orchestra.run.vm00.stdout:Freed space: 3.7 M 2026-03-06T13:43:49.514 INFO:teuthology.orchestra.run.vm00.stdout:Running transaction check 2026-03-06T13:43:49.516 INFO:teuthology.orchestra.run.vm00.stdout:Transaction check succeeded. 2026-03-06T13:43:49.516 INFO:teuthology.orchestra.run.vm00.stdout:Running transaction test 2026-03-06T13:43:49.532 INFO:teuthology.orchestra.run.vm00.stdout:Transaction test succeeded. 2026-03-06T13:43:49.532 INFO:teuthology.orchestra.run.vm00.stdout:Running transaction 2026-03-06T13:43:49.533 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : librados-devel-2:19.2.3-47.gc24117fd552.el9.clyso.x8 2/2 2026-03-06T13:43:49.533 INFO:teuthology.orchestra.run.vm08.stdout: 2026-03-06T13:43:49.533 INFO:teuthology.orchestra.run.vm08.stdout:Removed: 2026-03-06T13:43:49.533 INFO:teuthology.orchestra.run.vm08.stdout: libcephfs-devel-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 2026-03-06T13:43:49.533 INFO:teuthology.orchestra.run.vm08.stdout: librados-devel-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 2026-03-06T13:43:49.533 INFO:teuthology.orchestra.run.vm08.stdout: 2026-03-06T13:43:49.533 INFO:teuthology.orchestra.run.vm08.stdout:Complete! 2026-03-06T13:43:49.546 INFO:teuthology.orchestra.run.vm01.stdout:Dependencies resolved. 2026-03-06T13:43:49.547 INFO:teuthology.orchestra.run.vm01.stdout:================================================================================ 2026-03-06T13:43:49.547 INFO:teuthology.orchestra.run.vm01.stdout: Package Arch Version Repo Size 2026-03-06T13:43:49.547 INFO:teuthology.orchestra.run.vm01.stdout:================================================================================ 2026-03-06T13:43:49.547 INFO:teuthology.orchestra.run.vm01.stdout:Removing: 2026-03-06T13:43:49.547 INFO:teuthology.orchestra.run.vm01.stdout: librados-devel x86_64 2:19.2.3-47.gc24117fd552.el9.clyso @ceph 456 k 2026-03-06T13:43:49.547 INFO:teuthology.orchestra.run.vm01.stdout:Removing dependent packages: 2026-03-06T13:43:49.547 INFO:teuthology.orchestra.run.vm01.stdout: libcephfs-devel x86_64 2:19.2.3-47.gc24117fd552.el9.clyso @ceph 153 k 2026-03-06T13:43:49.547 INFO:teuthology.orchestra.run.vm01.stdout: 2026-03-06T13:43:49.547 INFO:teuthology.orchestra.run.vm01.stdout:Transaction Summary 2026-03-06T13:43:49.547 INFO:teuthology.orchestra.run.vm01.stdout:================================================================================ 2026-03-06T13:43:49.547 INFO:teuthology.orchestra.run.vm01.stdout:Remove 2 Packages 2026-03-06T13:43:49.547 INFO:teuthology.orchestra.run.vm01.stdout: 2026-03-06T13:43:49.547 INFO:teuthology.orchestra.run.vm01.stdout:Freed space: 610 k 2026-03-06T13:43:49.547 INFO:teuthology.orchestra.run.vm01.stdout:Running transaction check 2026-03-06T13:43:49.549 INFO:teuthology.orchestra.run.vm01.stdout:Transaction check succeeded. 2026-03-06T13:43:49.549 INFO:teuthology.orchestra.run.vm01.stdout:Running transaction test 2026-03-06T13:43:49.560 INFO:teuthology.orchestra.run.vm01.stdout:Transaction test succeeded. 2026-03-06T13:43:49.560 INFO:teuthology.orchestra.run.vm01.stdout:Running transaction 2026-03-06T13:43:49.563 INFO:teuthology.orchestra.run.vm00.stdout: Preparing : 1/1 2026-03-06T13:43:49.566 INFO:teuthology.orchestra.run.vm00.stdout: Erasing : python3-cephfs-2:19.2.3-47.gc24117fd552.el9.clyso.x8 1/3 2026-03-06T13:43:49.567 INFO:teuthology.orchestra.run.vm00.stdout: Erasing : python3-ceph-argparse-2:19.2.3-47.gc24117fd552.el9.c 2/3 2026-03-06T13:43:49.567 INFO:teuthology.orchestra.run.vm00.stdout: Erasing : libcephfs2-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 3/3 2026-03-06T13:43:49.587 INFO:teuthology.orchestra.run.vm01.stdout: Preparing : 1/1 2026-03-06T13:43:49.589 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : libcephfs-devel-2:19.2.3-47.gc24117fd552.el9.clyso.x 1/2 2026-03-06T13:43:49.602 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : librados-devel-2:19.2.3-47.gc24117fd552.el9.clyso.x8 2/2 2026-03-06T13:43:49.648 INFO:teuthology.orchestra.run.vm00.stdout: Running scriptlet: libcephfs2-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 3/3 2026-03-06T13:43:49.648 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : libcephfs2-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 1/3 2026-03-06T13:43:49.648 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : python3-ceph-argparse-2:19.2.3-47.gc24117fd552.el9.c 2/3 2026-03-06T13:43:49.665 INFO:teuthology.orchestra.run.vm01.stdout: Running scriptlet: librados-devel-2:19.2.3-47.gc24117fd552.el9.clyso.x8 2/2 2026-03-06T13:43:49.665 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : libcephfs-devel-2:19.2.3-47.gc24117fd552.el9.clyso.x 1/2 2026-03-06T13:43:49.691 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : python3-cephfs-2:19.2.3-47.gc24117fd552.el9.clyso.x8 3/3 2026-03-06T13:43:49.691 INFO:teuthology.orchestra.run.vm00.stdout: 2026-03-06T13:43:49.691 INFO:teuthology.orchestra.run.vm00.stdout:Removed: 2026-03-06T13:43:49.691 INFO:teuthology.orchestra.run.vm00.stdout: libcephfs2-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 2026-03-06T13:43:49.691 INFO:teuthology.orchestra.run.vm00.stdout: python3-ceph-argparse-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 2026-03-06T13:43:49.691 INFO:teuthology.orchestra.run.vm00.stdout: python3-cephfs-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 2026-03-06T13:43:49.691 INFO:teuthology.orchestra.run.vm00.stdout: 2026-03-06T13:43:49.691 INFO:teuthology.orchestra.run.vm00.stdout:Complete! 2026-03-06T13:43:49.712 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : librados-devel-2:19.2.3-47.gc24117fd552.el9.clyso.x8 2/2 2026-03-06T13:43:49.712 INFO:teuthology.orchestra.run.vm01.stdout: 2026-03-06T13:43:49.712 INFO:teuthology.orchestra.run.vm01.stdout:Removed: 2026-03-06T13:43:49.712 INFO:teuthology.orchestra.run.vm01.stdout: libcephfs-devel-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 2026-03-06T13:43:49.712 INFO:teuthology.orchestra.run.vm01.stdout: librados-devel-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 2026-03-06T13:43:49.712 INFO:teuthology.orchestra.run.vm01.stdout: 2026-03-06T13:43:49.712 INFO:teuthology.orchestra.run.vm01.stdout:Complete! 2026-03-06T13:43:49.739 INFO:teuthology.orchestra.run.vm08.stdout:Dependencies resolved. 2026-03-06T13:43:49.740 INFO:teuthology.orchestra.run.vm08.stdout:================================================================================ 2026-03-06T13:43:49.740 INFO:teuthology.orchestra.run.vm08.stdout: Package Arch Version Repo Size 2026-03-06T13:43:49.740 INFO:teuthology.orchestra.run.vm08.stdout:================================================================================ 2026-03-06T13:43:49.740 INFO:teuthology.orchestra.run.vm08.stdout:Removing: 2026-03-06T13:43:49.740 INFO:teuthology.orchestra.run.vm08.stdout: libcephfs2 x86_64 2:19.2.3-47.gc24117fd552.el9.clyso @ceph 3.0 M 2026-03-06T13:43:49.740 INFO:teuthology.orchestra.run.vm08.stdout:Removing dependent packages: 2026-03-06T13:43:49.740 INFO:teuthology.orchestra.run.vm08.stdout: python3-cephfs x86_64 2:19.2.3-47.gc24117fd552.el9.clyso @ceph 514 k 2026-03-06T13:43:49.740 INFO:teuthology.orchestra.run.vm08.stdout:Removing unused dependencies: 2026-03-06T13:43:49.740 INFO:teuthology.orchestra.run.vm08.stdout: python3-ceph-argparse x86_64 2:19.2.3-47.gc24117fd552.el9.clyso @ceph 187 k 2026-03-06T13:43:49.740 INFO:teuthology.orchestra.run.vm08.stdout: 2026-03-06T13:43:49.740 INFO:teuthology.orchestra.run.vm08.stdout:Transaction Summary 2026-03-06T13:43:49.740 INFO:teuthology.orchestra.run.vm08.stdout:================================================================================ 2026-03-06T13:43:49.740 INFO:teuthology.orchestra.run.vm08.stdout:Remove 3 Packages 2026-03-06T13:43:49.740 INFO:teuthology.orchestra.run.vm08.stdout: 2026-03-06T13:43:49.740 INFO:teuthology.orchestra.run.vm08.stdout:Freed space: 3.7 M 2026-03-06T13:43:49.740 INFO:teuthology.orchestra.run.vm08.stdout:Running transaction check 2026-03-06T13:43:49.742 INFO:teuthology.orchestra.run.vm08.stdout:Transaction check succeeded. 2026-03-06T13:43:49.742 INFO:teuthology.orchestra.run.vm08.stdout:Running transaction test 2026-03-06T13:43:49.760 INFO:teuthology.orchestra.run.vm08.stdout:Transaction test succeeded. 2026-03-06T13:43:49.761 INFO:teuthology.orchestra.run.vm08.stdout:Running transaction 2026-03-06T13:43:49.796 INFO:teuthology.orchestra.run.vm08.stdout: Preparing : 1/1 2026-03-06T13:43:49.798 INFO:teuthology.orchestra.run.vm08.stdout: Erasing : python3-cephfs-2:19.2.3-47.gc24117fd552.el9.clyso.x8 1/3 2026-03-06T13:43:49.800 INFO:teuthology.orchestra.run.vm08.stdout: Erasing : python3-ceph-argparse-2:19.2.3-47.gc24117fd552.el9.c 2/3 2026-03-06T13:43:49.800 INFO:teuthology.orchestra.run.vm08.stdout: Erasing : libcephfs2-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 3/3 2026-03-06T13:43:49.858 INFO:teuthology.orchestra.run.vm08.stdout: Running scriptlet: libcephfs2-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 3/3 2026-03-06T13:43:49.858 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : libcephfs2-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 1/3 2026-03-06T13:43:49.858 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : python3-ceph-argparse-2:19.2.3-47.gc24117fd552.el9.c 2/3 2026-03-06T13:43:49.877 INFO:teuthology.orchestra.run.vm00.stdout:No match for argument: libcephfs-devel 2026-03-06T13:43:49.877 INFO:teuthology.orchestra.run.vm00.stderr:No packages marked for removal. 2026-03-06T13:43:49.880 INFO:teuthology.orchestra.run.vm00.stdout:Dependencies resolved. 2026-03-06T13:43:49.880 INFO:teuthology.orchestra.run.vm00.stdout:Nothing to do. 2026-03-06T13:43:49.880 INFO:teuthology.orchestra.run.vm00.stdout:Complete! 2026-03-06T13:43:49.897 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : python3-cephfs-2:19.2.3-47.gc24117fd552.el9.clyso.x8 3/3 2026-03-06T13:43:49.897 INFO:teuthology.orchestra.run.vm08.stdout: 2026-03-06T13:43:49.897 INFO:teuthology.orchestra.run.vm08.stdout:Removed: 2026-03-06T13:43:49.897 INFO:teuthology.orchestra.run.vm08.stdout: libcephfs2-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 2026-03-06T13:43:49.897 INFO:teuthology.orchestra.run.vm08.stdout: python3-ceph-argparse-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 2026-03-06T13:43:49.897 INFO:teuthology.orchestra.run.vm08.stdout: python3-cephfs-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 2026-03-06T13:43:49.897 INFO:teuthology.orchestra.run.vm08.stdout: 2026-03-06T13:43:49.897 INFO:teuthology.orchestra.run.vm08.stdout:Complete! 2026-03-06T13:43:49.912 INFO:teuthology.orchestra.run.vm01.stdout:Dependencies resolved. 2026-03-06T13:43:49.912 INFO:teuthology.orchestra.run.vm01.stdout:================================================================================ 2026-03-06T13:43:49.913 INFO:teuthology.orchestra.run.vm01.stdout: Package Arch Version Repo Size 2026-03-06T13:43:49.913 INFO:teuthology.orchestra.run.vm01.stdout:================================================================================ 2026-03-06T13:43:49.913 INFO:teuthology.orchestra.run.vm01.stdout:Removing: 2026-03-06T13:43:49.913 INFO:teuthology.orchestra.run.vm01.stdout: libcephfs2 x86_64 2:19.2.3-47.gc24117fd552.el9.clyso @ceph 3.0 M 2026-03-06T13:43:49.913 INFO:teuthology.orchestra.run.vm01.stdout:Removing dependent packages: 2026-03-06T13:43:49.913 INFO:teuthology.orchestra.run.vm01.stdout: python3-cephfs x86_64 2:19.2.3-47.gc24117fd552.el9.clyso @ceph 514 k 2026-03-06T13:43:49.913 INFO:teuthology.orchestra.run.vm01.stdout:Removing unused dependencies: 2026-03-06T13:43:49.913 INFO:teuthology.orchestra.run.vm01.stdout: python3-ceph-argparse x86_64 2:19.2.3-47.gc24117fd552.el9.clyso @ceph 187 k 2026-03-06T13:43:49.913 INFO:teuthology.orchestra.run.vm01.stdout: 2026-03-06T13:43:49.913 INFO:teuthology.orchestra.run.vm01.stdout:Transaction Summary 2026-03-06T13:43:49.913 INFO:teuthology.orchestra.run.vm01.stdout:================================================================================ 2026-03-06T13:43:49.913 INFO:teuthology.orchestra.run.vm01.stdout:Remove 3 Packages 2026-03-06T13:43:49.913 INFO:teuthology.orchestra.run.vm01.stdout: 2026-03-06T13:43:49.913 INFO:teuthology.orchestra.run.vm01.stdout:Freed space: 3.7 M 2026-03-06T13:43:49.913 INFO:teuthology.orchestra.run.vm01.stdout:Running transaction check 2026-03-06T13:43:49.915 INFO:teuthology.orchestra.run.vm01.stdout:Transaction check succeeded. 2026-03-06T13:43:49.915 INFO:teuthology.orchestra.run.vm01.stdout:Running transaction test 2026-03-06T13:43:49.932 INFO:teuthology.orchestra.run.vm01.stdout:Transaction test succeeded. 2026-03-06T13:43:49.932 INFO:teuthology.orchestra.run.vm01.stdout:Running transaction 2026-03-06T13:43:49.968 INFO:teuthology.orchestra.run.vm01.stdout: Preparing : 1/1 2026-03-06T13:43:49.970 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : python3-cephfs-2:19.2.3-47.gc24117fd552.el9.clyso.x8 1/3 2026-03-06T13:43:49.971 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : python3-ceph-argparse-2:19.2.3-47.gc24117fd552.el9.c 2/3 2026-03-06T13:43:49.972 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : libcephfs2-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 3/3 2026-03-06T13:43:50.037 INFO:teuthology.orchestra.run.vm01.stdout: Running scriptlet: libcephfs2-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 3/3 2026-03-06T13:43:50.037 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : libcephfs2-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 1/3 2026-03-06T13:43:50.037 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-ceph-argparse-2:19.2.3-47.gc24117fd552.el9.c 2/3 2026-03-06T13:43:50.062 INFO:teuthology.orchestra.run.vm08.stdout:No match for argument: libcephfs-devel 2026-03-06T13:43:50.062 INFO:teuthology.orchestra.run.vm08.stderr:No packages marked for removal. 2026-03-06T13:43:50.064 INFO:teuthology.orchestra.run.vm08.stdout:Dependencies resolved. 2026-03-06T13:43:50.064 INFO:teuthology.orchestra.run.vm08.stdout:Nothing to do. 2026-03-06T13:43:50.064 INFO:teuthology.orchestra.run.vm08.stdout:Complete! 2026-03-06T13:43:50.080 INFO:teuthology.orchestra.run.vm00.stdout:Dependencies resolved. 2026-03-06T13:43:50.081 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-cephfs-2:19.2.3-47.gc24117fd552.el9.clyso.x8 3/3 2026-03-06T13:43:50.081 INFO:teuthology.orchestra.run.vm01.stdout: 2026-03-06T13:43:50.081 INFO:teuthology.orchestra.run.vm01.stdout:Removed: 2026-03-06T13:43:50.081 INFO:teuthology.orchestra.run.vm01.stdout: libcephfs2-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 2026-03-06T13:43:50.081 INFO:teuthology.orchestra.run.vm01.stdout: python3-ceph-argparse-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 2026-03-06T13:43:50.082 INFO:teuthology.orchestra.run.vm01.stdout: python3-cephfs-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 2026-03-06T13:43:50.082 INFO:teuthology.orchestra.run.vm01.stdout: 2026-03-06T13:43:50.082 INFO:teuthology.orchestra.run.vm01.stdout:Complete! 2026-03-06T13:43:50.082 INFO:teuthology.orchestra.run.vm00.stdout:================================================================================ 2026-03-06T13:43:50.082 INFO:teuthology.orchestra.run.vm00.stdout: Package Arch Version Repository Size 2026-03-06T13:43:50.082 INFO:teuthology.orchestra.run.vm00.stdout:================================================================================ 2026-03-06T13:43:50.082 INFO:teuthology.orchestra.run.vm00.stdout:Removing: 2026-03-06T13:43:50.082 INFO:teuthology.orchestra.run.vm00.stdout: librados2 x86_64 2:19.2.3-47.gc24117fd552.el9.clyso @ceph 13 M 2026-03-06T13:43:50.082 INFO:teuthology.orchestra.run.vm00.stdout:Removing dependent packages: 2026-03-06T13:43:50.082 INFO:teuthology.orchestra.run.vm00.stdout: python3-rados x86_64 2:19.2.3-47.gc24117fd552.el9.clyso @ceph 1.1 M 2026-03-06T13:43:50.082 INFO:teuthology.orchestra.run.vm00.stdout: python3-rbd x86_64 2:19.2.3-47.gc24117fd552.el9.clyso @ceph 1.1 M 2026-03-06T13:43:50.082 INFO:teuthology.orchestra.run.vm00.stdout: python3-rgw x86_64 2:19.2.3-47.gc24117fd552.el9.clyso @ceph 265 k 2026-03-06T13:43:50.082 INFO:teuthology.orchestra.run.vm00.stdout: qemu-kvm-block-rbd x86_64 17:10.1.0-15.el9 @appstream 37 k 2026-03-06T13:43:50.082 INFO:teuthology.orchestra.run.vm00.stdout: rbd-fuse x86_64 2:19.2.3-47.gc24117fd552.el9.clyso @ceph 227 k 2026-03-06T13:43:50.082 INFO:teuthology.orchestra.run.vm00.stdout: rbd-nbd x86_64 2:19.2.3-47.gc24117fd552.el9.clyso @ceph 494 k 2026-03-06T13:43:50.082 INFO:teuthology.orchestra.run.vm00.stdout:Removing unused dependencies: 2026-03-06T13:43:50.082 INFO:teuthology.orchestra.run.vm00.stdout: boost-program-options 2026-03-06T13:43:50.082 INFO:teuthology.orchestra.run.vm00.stdout: x86_64 1.75.0-13.el9 @appstream 276 k 2026-03-06T13:43:50.082 INFO:teuthology.orchestra.run.vm00.stdout: libarrow x86_64 9.0.0-15.el9 @epel 18 M 2026-03-06T13:43:50.082 INFO:teuthology.orchestra.run.vm00.stdout: libarrow-doc noarch 9.0.0-15.el9 @epel 122 k 2026-03-06T13:43:50.082 INFO:teuthology.orchestra.run.vm00.stdout: libnbd x86_64 1.20.3-4.el9 @appstream 453 k 2026-03-06T13:43:50.082 INFO:teuthology.orchestra.run.vm00.stdout: libpmemobj x86_64 1.12.1-1.el9 @appstream 383 k 2026-03-06T13:43:50.082 INFO:teuthology.orchestra.run.vm00.stdout: librabbitmq x86_64 0.11.0-7.el9 @appstream 102 k 2026-03-06T13:43:50.082 INFO:teuthology.orchestra.run.vm00.stdout: librbd1 x86_64 2:19.2.3-47.gc24117fd552.el9.clyso @ceph 13 M 2026-03-06T13:43:50.082 INFO:teuthology.orchestra.run.vm00.stdout: librdkafka x86_64 1.6.1-102.el9 @appstream 2.0 M 2026-03-06T13:43:50.082 INFO:teuthology.orchestra.run.vm00.stdout: librgw2 x86_64 2:19.2.3-47.gc24117fd552.el9.clyso @ceph 19 M 2026-03-06T13:43:50.082 INFO:teuthology.orchestra.run.vm00.stdout: lttng-ust x86_64 2.12.0-6.el9 @appstream 1.0 M 2026-03-06T13:43:50.082 INFO:teuthology.orchestra.run.vm00.stdout: parquet-libs x86_64 9.0.0-15.el9 @epel 2.8 M 2026-03-06T13:43:50.082 INFO:teuthology.orchestra.run.vm00.stdout: re2 x86_64 1:20211101-20.el9 @epel 472 k 2026-03-06T13:43:50.082 INFO:teuthology.orchestra.run.vm00.stdout: thrift x86_64 0.15.0-4.el9 @epel 4.8 M 2026-03-06T13:43:50.082 INFO:teuthology.orchestra.run.vm00.stdout: 2026-03-06T13:43:50.082 INFO:teuthology.orchestra.run.vm00.stdout:Transaction Summary 2026-03-06T13:43:50.082 INFO:teuthology.orchestra.run.vm00.stdout:================================================================================ 2026-03-06T13:43:50.082 INFO:teuthology.orchestra.run.vm00.stdout:Remove 20 Packages 2026-03-06T13:43:50.082 INFO:teuthology.orchestra.run.vm00.stdout: 2026-03-06T13:43:50.083 INFO:teuthology.orchestra.run.vm00.stdout:Freed space: 79 M 2026-03-06T13:43:50.083 INFO:teuthology.orchestra.run.vm00.stdout:Running transaction check 2026-03-06T13:43:50.086 INFO:teuthology.orchestra.run.vm00.stdout:Transaction check succeeded. 2026-03-06T13:43:50.086 INFO:teuthology.orchestra.run.vm00.stdout:Running transaction test 2026-03-06T13:43:50.110 INFO:teuthology.orchestra.run.vm00.stdout:Transaction test succeeded. 2026-03-06T13:43:50.110 INFO:teuthology.orchestra.run.vm00.stdout:Running transaction 2026-03-06T13:43:50.151 INFO:teuthology.orchestra.run.vm00.stdout: Preparing : 1/1 2026-03-06T13:43:50.153 INFO:teuthology.orchestra.run.vm00.stdout: Erasing : rbd-nbd-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 1/20 2026-03-06T13:43:50.156 INFO:teuthology.orchestra.run.vm00.stdout: Erasing : rbd-fuse-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 2/20 2026-03-06T13:43:50.159 INFO:teuthology.orchestra.run.vm00.stdout: Erasing : python3-rgw-2:19.2.3-47.gc24117fd552.el9.clyso.x86 3/20 2026-03-06T13:43:50.159 INFO:teuthology.orchestra.run.vm00.stdout: Erasing : librgw2-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 4/20 2026-03-06T13:43:50.172 INFO:teuthology.orchestra.run.vm00.stdout: Running scriptlet: librgw2-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 4/20 2026-03-06T13:43:50.174 INFO:teuthology.orchestra.run.vm00.stdout: Erasing : parquet-libs-9.0.0-15.el9.x86_64 5/20 2026-03-06T13:43:50.178 INFO:teuthology.orchestra.run.vm00.stdout: Erasing : python3-rbd-2:19.2.3-47.gc24117fd552.el9.clyso.x86 6/20 2026-03-06T13:43:50.179 INFO:teuthology.orchestra.run.vm00.stdout: Erasing : python3-rados-2:19.2.3-47.gc24117fd552.el9.clyso.x 7/20 2026-03-06T13:43:50.181 INFO:teuthology.orchestra.run.vm00.stdout: Erasing : qemu-kvm-block-rbd-17:10.1.0-15.el9.x86_64 8/20 2026-03-06T13:43:50.183 INFO:teuthology.orchestra.run.vm00.stdout: Erasing : libarrow-doc-9.0.0-15.el9.noarch 9/20 2026-03-06T13:43:50.183 INFO:teuthology.orchestra.run.vm00.stdout: Erasing : librbd1-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 10/20 2026-03-06T13:43:50.197 INFO:teuthology.orchestra.run.vm00.stdout: Running scriptlet: librbd1-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 10/20 2026-03-06T13:43:50.197 INFO:teuthology.orchestra.run.vm00.stdout: Erasing : librados2-2:19.2.3-47.gc24117fd552.el9.clyso.x86_6 11/20 2026-03-06T13:43:50.197 INFO:teuthology.orchestra.run.vm00.stdout:warning: file /etc/ceph: remove failed: No such file or directory 2026-03-06T13:43:50.197 INFO:teuthology.orchestra.run.vm00.stdout: 2026-03-06T13:43:50.210 INFO:teuthology.orchestra.run.vm00.stdout: Running scriptlet: librados2-2:19.2.3-47.gc24117fd552.el9.clyso.x86_6 11/20 2026-03-06T13:43:50.212 INFO:teuthology.orchestra.run.vm00.stdout: Erasing : libarrow-9.0.0-15.el9.x86_64 12/20 2026-03-06T13:43:50.216 INFO:teuthology.orchestra.run.vm00.stdout: Erasing : re2-1:20211101-20.el9.x86_64 13/20 2026-03-06T13:43:50.219 INFO:teuthology.orchestra.run.vm00.stdout: Erasing : lttng-ust-2.12.0-6.el9.x86_64 14/20 2026-03-06T13:43:50.222 INFO:teuthology.orchestra.run.vm00.stdout: Erasing : thrift-0.15.0-4.el9.x86_64 15/20 2026-03-06T13:43:50.224 INFO:teuthology.orchestra.run.vm00.stdout: Erasing : libnbd-1.20.3-4.el9.x86_64 16/20 2026-03-06T13:43:50.227 INFO:teuthology.orchestra.run.vm00.stdout: Erasing : libpmemobj-1.12.1-1.el9.x86_64 17/20 2026-03-06T13:43:50.229 INFO:teuthology.orchestra.run.vm00.stdout: Erasing : boost-program-options-1.75.0-13.el9.x86_64 18/20 2026-03-06T13:43:50.231 INFO:teuthology.orchestra.run.vm00.stdout: Erasing : librabbitmq-0.11.0-7.el9.x86_64 19/20 2026-03-06T13:43:50.237 INFO:teuthology.orchestra.run.vm08.stdout:Dependencies resolved. 2026-03-06T13:43:50.238 INFO:teuthology.orchestra.run.vm08.stdout:================================================================================ 2026-03-06T13:43:50.239 INFO:teuthology.orchestra.run.vm08.stdout: Package Arch Version Repository Size 2026-03-06T13:43:50.239 INFO:teuthology.orchestra.run.vm08.stdout:================================================================================ 2026-03-06T13:43:50.239 INFO:teuthology.orchestra.run.vm08.stdout:Removing: 2026-03-06T13:43:50.239 INFO:teuthology.orchestra.run.vm08.stdout: librados2 x86_64 2:19.2.3-47.gc24117fd552.el9.clyso @ceph 13 M 2026-03-06T13:43:50.239 INFO:teuthology.orchestra.run.vm08.stdout:Removing dependent packages: 2026-03-06T13:43:50.239 INFO:teuthology.orchestra.run.vm08.stdout: python3-rados x86_64 2:19.2.3-47.gc24117fd552.el9.clyso @ceph 1.1 M 2026-03-06T13:43:50.239 INFO:teuthology.orchestra.run.vm08.stdout: python3-rbd x86_64 2:19.2.3-47.gc24117fd552.el9.clyso @ceph 1.1 M 2026-03-06T13:43:50.239 INFO:teuthology.orchestra.run.vm08.stdout: python3-rgw x86_64 2:19.2.3-47.gc24117fd552.el9.clyso @ceph 265 k 2026-03-06T13:43:50.239 INFO:teuthology.orchestra.run.vm08.stdout: qemu-kvm-block-rbd x86_64 17:10.1.0-15.el9 @appstream 37 k 2026-03-06T13:43:50.239 INFO:teuthology.orchestra.run.vm08.stdout: rbd-fuse x86_64 2:19.2.3-47.gc24117fd552.el9.clyso @ceph 227 k 2026-03-06T13:43:50.239 INFO:teuthology.orchestra.run.vm08.stdout: rbd-nbd x86_64 2:19.2.3-47.gc24117fd552.el9.clyso @ceph 494 k 2026-03-06T13:43:50.239 INFO:teuthology.orchestra.run.vm08.stdout:Removing unused dependencies: 2026-03-06T13:43:50.239 INFO:teuthology.orchestra.run.vm08.stdout: boost-program-options 2026-03-06T13:43:50.239 INFO:teuthology.orchestra.run.vm08.stdout: x86_64 1.75.0-13.el9 @appstream 276 k 2026-03-06T13:43:50.239 INFO:teuthology.orchestra.run.vm08.stdout: libarrow x86_64 9.0.0-15.el9 @epel 18 M 2026-03-06T13:43:50.239 INFO:teuthology.orchestra.run.vm08.stdout: libarrow-doc noarch 9.0.0-15.el9 @epel 122 k 2026-03-06T13:43:50.239 INFO:teuthology.orchestra.run.vm08.stdout: libnbd x86_64 1.20.3-4.el9 @appstream 453 k 2026-03-06T13:43:50.239 INFO:teuthology.orchestra.run.vm08.stdout: libpmemobj x86_64 1.12.1-1.el9 @appstream 383 k 2026-03-06T13:43:50.239 INFO:teuthology.orchestra.run.vm08.stdout: librabbitmq x86_64 0.11.0-7.el9 @appstream 102 k 2026-03-06T13:43:50.239 INFO:teuthology.orchestra.run.vm08.stdout: librbd1 x86_64 2:19.2.3-47.gc24117fd552.el9.clyso @ceph 13 M 2026-03-06T13:43:50.239 INFO:teuthology.orchestra.run.vm08.stdout: librdkafka x86_64 1.6.1-102.el9 @appstream 2.0 M 2026-03-06T13:43:50.239 INFO:teuthology.orchestra.run.vm08.stdout: librgw2 x86_64 2:19.2.3-47.gc24117fd552.el9.clyso @ceph 19 M 2026-03-06T13:43:50.239 INFO:teuthology.orchestra.run.vm08.stdout: lttng-ust x86_64 2.12.0-6.el9 @appstream 1.0 M 2026-03-06T13:43:50.239 INFO:teuthology.orchestra.run.vm08.stdout: parquet-libs x86_64 9.0.0-15.el9 @epel 2.8 M 2026-03-06T13:43:50.239 INFO:teuthology.orchestra.run.vm08.stdout: re2 x86_64 1:20211101-20.el9 @epel 472 k 2026-03-06T13:43:50.239 INFO:teuthology.orchestra.run.vm08.stdout: thrift x86_64 0.15.0-4.el9 @epel 4.8 M 2026-03-06T13:43:50.239 INFO:teuthology.orchestra.run.vm08.stdout: 2026-03-06T13:43:50.239 INFO:teuthology.orchestra.run.vm08.stdout:Transaction Summary 2026-03-06T13:43:50.239 INFO:teuthology.orchestra.run.vm08.stdout:================================================================================ 2026-03-06T13:43:50.239 INFO:teuthology.orchestra.run.vm08.stdout:Remove 20 Packages 2026-03-06T13:43:50.239 INFO:teuthology.orchestra.run.vm08.stdout: 2026-03-06T13:43:50.239 INFO:teuthology.orchestra.run.vm08.stdout:Freed space: 79 M 2026-03-06T13:43:50.239 INFO:teuthology.orchestra.run.vm08.stdout:Running transaction check 2026-03-06T13:43:50.243 INFO:teuthology.orchestra.run.vm08.stdout:Transaction check succeeded. 2026-03-06T13:43:50.243 INFO:teuthology.orchestra.run.vm08.stdout:Running transaction test 2026-03-06T13:43:50.245 INFO:teuthology.orchestra.run.vm00.stdout: Erasing : librdkafka-1.6.1-102.el9.x86_64 20/20 2026-03-06T13:43:50.263 INFO:teuthology.orchestra.run.vm01.stdout:No match for argument: libcephfs-devel 2026-03-06T13:43:50.263 INFO:teuthology.orchestra.run.vm01.stderr:No packages marked for removal. 2026-03-06T13:43:50.265 INFO:teuthology.orchestra.run.vm08.stdout:Transaction test succeeded. 2026-03-06T13:43:50.265 INFO:teuthology.orchestra.run.vm08.stdout:Running transaction 2026-03-06T13:43:50.265 INFO:teuthology.orchestra.run.vm01.stdout:Dependencies resolved. 2026-03-06T13:43:50.266 INFO:teuthology.orchestra.run.vm01.stdout:Nothing to do. 2026-03-06T13:43:50.266 INFO:teuthology.orchestra.run.vm01.stdout:Complete! 2026-03-06T13:43:50.304 INFO:teuthology.orchestra.run.vm00.stdout: Running scriptlet: librdkafka-1.6.1-102.el9.x86_64 20/20 2026-03-06T13:43:50.304 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : boost-program-options-1.75.0-13.el9.x86_64 1/20 2026-03-06T13:43:50.304 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : libarrow-9.0.0-15.el9.x86_64 2/20 2026-03-06T13:43:50.304 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : libarrow-doc-9.0.0-15.el9.noarch 3/20 2026-03-06T13:43:50.304 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : libnbd-1.20.3-4.el9.x86_64 4/20 2026-03-06T13:43:50.304 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : libpmemobj-1.12.1-1.el9.x86_64 5/20 2026-03-06T13:43:50.304 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : librabbitmq-0.11.0-7.el9.x86_64 6/20 2026-03-06T13:43:50.304 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : librados2-2:19.2.3-47.gc24117fd552.el9.clyso.x86_6 7/20 2026-03-06T13:43:50.304 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : librbd1-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 8/20 2026-03-06T13:43:50.304 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : librdkafka-1.6.1-102.el9.x86_64 9/20 2026-03-06T13:43:50.304 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : librgw2-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 10/20 2026-03-06T13:43:50.304 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : lttng-ust-2.12.0-6.el9.x86_64 11/20 2026-03-06T13:43:50.304 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : parquet-libs-9.0.0-15.el9.x86_64 12/20 2026-03-06T13:43:50.304 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : python3-rados-2:19.2.3-47.gc24117fd552.el9.clyso.x 13/20 2026-03-06T13:43:50.304 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : python3-rbd-2:19.2.3-47.gc24117fd552.el9.clyso.x86 14/20 2026-03-06T13:43:50.304 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : python3-rgw-2:19.2.3-47.gc24117fd552.el9.clyso.x86 15/20 2026-03-06T13:43:50.304 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : qemu-kvm-block-rbd-17:10.1.0-15.el9.x86_64 16/20 2026-03-06T13:43:50.304 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : rbd-fuse-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 17/20 2026-03-06T13:43:50.304 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : rbd-nbd-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 18/20 2026-03-06T13:43:50.305 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : re2-1:20211101-20.el9.x86_64 19/20 2026-03-06T13:43:50.305 INFO:teuthology.orchestra.run.vm08.stdout: Preparing : 1/1 2026-03-06T13:43:50.308 INFO:teuthology.orchestra.run.vm08.stdout: Erasing : rbd-nbd-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 1/20 2026-03-06T13:43:50.310 INFO:teuthology.orchestra.run.vm08.stdout: Erasing : rbd-fuse-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 2/20 2026-03-06T13:43:50.313 INFO:teuthology.orchestra.run.vm08.stdout: Erasing : python3-rgw-2:19.2.3-47.gc24117fd552.el9.clyso.x86 3/20 2026-03-06T13:43:50.313 INFO:teuthology.orchestra.run.vm08.stdout: Erasing : librgw2-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 4/20 2026-03-06T13:43:50.328 INFO:teuthology.orchestra.run.vm08.stdout: Running scriptlet: librgw2-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 4/20 2026-03-06T13:43:50.330 INFO:teuthology.orchestra.run.vm08.stdout: Erasing : parquet-libs-9.0.0-15.el9.x86_64 5/20 2026-03-06T13:43:50.332 INFO:teuthology.orchestra.run.vm08.stdout: Erasing : python3-rbd-2:19.2.3-47.gc24117fd552.el9.clyso.x86 6/20 2026-03-06T13:43:50.334 INFO:teuthology.orchestra.run.vm08.stdout: Erasing : python3-rados-2:19.2.3-47.gc24117fd552.el9.clyso.x 7/20 2026-03-06T13:43:50.336 INFO:teuthology.orchestra.run.vm08.stdout: Erasing : qemu-kvm-block-rbd-17:10.1.0-15.el9.x86_64 8/20 2026-03-06T13:43:50.339 INFO:teuthology.orchestra.run.vm08.stdout: Erasing : libarrow-doc-9.0.0-15.el9.noarch 9/20 2026-03-06T13:43:50.339 INFO:teuthology.orchestra.run.vm08.stdout: Erasing : librbd1-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 10/20 2026-03-06T13:43:50.346 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : thrift-0.15.0-4.el9.x86_64 20/20 2026-03-06T13:43:50.346 INFO:teuthology.orchestra.run.vm00.stdout: 2026-03-06T13:43:50.346 INFO:teuthology.orchestra.run.vm00.stdout:Removed: 2026-03-06T13:43:50.346 INFO:teuthology.orchestra.run.vm00.stdout: boost-program-options-1.75.0-13.el9.x86_64 2026-03-06T13:43:50.346 INFO:teuthology.orchestra.run.vm00.stdout: libarrow-9.0.0-15.el9.x86_64 2026-03-06T13:43:50.346 INFO:teuthology.orchestra.run.vm00.stdout: libarrow-doc-9.0.0-15.el9.noarch 2026-03-06T13:43:50.346 INFO:teuthology.orchestra.run.vm00.stdout: libnbd-1.20.3-4.el9.x86_64 2026-03-06T13:43:50.346 INFO:teuthology.orchestra.run.vm00.stdout: libpmemobj-1.12.1-1.el9.x86_64 2026-03-06T13:43:50.346 INFO:teuthology.orchestra.run.vm00.stdout: librabbitmq-0.11.0-7.el9.x86_64 2026-03-06T13:43:50.346 INFO:teuthology.orchestra.run.vm00.stdout: librados2-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 2026-03-06T13:43:50.346 INFO:teuthology.orchestra.run.vm00.stdout: librbd1-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 2026-03-06T13:43:50.346 INFO:teuthology.orchestra.run.vm00.stdout: librdkafka-1.6.1-102.el9.x86_64 2026-03-06T13:43:50.346 INFO:teuthology.orchestra.run.vm00.stdout: librgw2-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 2026-03-06T13:43:50.346 INFO:teuthology.orchestra.run.vm00.stdout: lttng-ust-2.12.0-6.el9.x86_64 2026-03-06T13:43:50.346 INFO:teuthology.orchestra.run.vm00.stdout: parquet-libs-9.0.0-15.el9.x86_64 2026-03-06T13:43:50.346 INFO:teuthology.orchestra.run.vm00.stdout: python3-rados-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 2026-03-06T13:43:50.346 INFO:teuthology.orchestra.run.vm00.stdout: python3-rbd-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 2026-03-06T13:43:50.346 INFO:teuthology.orchestra.run.vm00.stdout: python3-rgw-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 2026-03-06T13:43:50.346 INFO:teuthology.orchestra.run.vm00.stdout: qemu-kvm-block-rbd-17:10.1.0-15.el9.x86_64 2026-03-06T13:43:50.346 INFO:teuthology.orchestra.run.vm00.stdout: rbd-fuse-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 2026-03-06T13:43:50.346 INFO:teuthology.orchestra.run.vm00.stdout: rbd-nbd-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 2026-03-06T13:43:50.346 INFO:teuthology.orchestra.run.vm00.stdout: re2-1:20211101-20.el9.x86_64 2026-03-06T13:43:50.346 INFO:teuthology.orchestra.run.vm00.stdout: thrift-0.15.0-4.el9.x86_64 2026-03-06T13:43:50.347 INFO:teuthology.orchestra.run.vm00.stdout: 2026-03-06T13:43:50.347 INFO:teuthology.orchestra.run.vm00.stdout:Complete! 2026-03-06T13:43:50.354 INFO:teuthology.orchestra.run.vm08.stdout: Running scriptlet: librbd1-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 10/20 2026-03-06T13:43:50.355 INFO:teuthology.orchestra.run.vm08.stdout: Erasing : librados2-2:19.2.3-47.gc24117fd552.el9.clyso.x86_6 11/20 2026-03-06T13:43:50.355 INFO:teuthology.orchestra.run.vm08.stdout:warning: file /etc/ceph: remove failed: No such file or directory 2026-03-06T13:43:50.355 INFO:teuthology.orchestra.run.vm08.stdout: 2026-03-06T13:43:50.367 INFO:teuthology.orchestra.run.vm08.stdout: Running scriptlet: librados2-2:19.2.3-47.gc24117fd552.el9.clyso.x86_6 11/20 2026-03-06T13:43:50.369 INFO:teuthology.orchestra.run.vm08.stdout: Erasing : libarrow-9.0.0-15.el9.x86_64 12/20 2026-03-06T13:43:50.372 INFO:teuthology.orchestra.run.vm08.stdout: Erasing : re2-1:20211101-20.el9.x86_64 13/20 2026-03-06T13:43:50.375 INFO:teuthology.orchestra.run.vm08.stdout: Erasing : lttng-ust-2.12.0-6.el9.x86_64 14/20 2026-03-06T13:43:50.378 INFO:teuthology.orchestra.run.vm08.stdout: Erasing : thrift-0.15.0-4.el9.x86_64 15/20 2026-03-06T13:43:50.380 INFO:teuthology.orchestra.run.vm08.stdout: Erasing : libnbd-1.20.3-4.el9.x86_64 16/20 2026-03-06T13:43:50.382 INFO:teuthology.orchestra.run.vm08.stdout: Erasing : libpmemobj-1.12.1-1.el9.x86_64 17/20 2026-03-06T13:43:50.384 INFO:teuthology.orchestra.run.vm08.stdout: Erasing : boost-program-options-1.75.0-13.el9.x86_64 18/20 2026-03-06T13:43:50.386 INFO:teuthology.orchestra.run.vm08.stdout: Erasing : librabbitmq-0.11.0-7.el9.x86_64 19/20 2026-03-06T13:43:50.400 INFO:teuthology.orchestra.run.vm08.stdout: Erasing : librdkafka-1.6.1-102.el9.x86_64 20/20 2026-03-06T13:43:50.458 INFO:teuthology.orchestra.run.vm08.stdout: Running scriptlet: librdkafka-1.6.1-102.el9.x86_64 20/20 2026-03-06T13:43:50.458 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : boost-program-options-1.75.0-13.el9.x86_64 1/20 2026-03-06T13:43:50.458 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : libarrow-9.0.0-15.el9.x86_64 2/20 2026-03-06T13:43:50.458 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : libarrow-doc-9.0.0-15.el9.noarch 3/20 2026-03-06T13:43:50.458 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : libnbd-1.20.3-4.el9.x86_64 4/20 2026-03-06T13:43:50.458 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : libpmemobj-1.12.1-1.el9.x86_64 5/20 2026-03-06T13:43:50.458 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : librabbitmq-0.11.0-7.el9.x86_64 6/20 2026-03-06T13:43:50.458 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : librados2-2:19.2.3-47.gc24117fd552.el9.clyso.x86_6 7/20 2026-03-06T13:43:50.458 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : librbd1-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 8/20 2026-03-06T13:43:50.458 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : librdkafka-1.6.1-102.el9.x86_64 9/20 2026-03-06T13:43:50.458 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : librgw2-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 10/20 2026-03-06T13:43:50.458 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : lttng-ust-2.12.0-6.el9.x86_64 11/20 2026-03-06T13:43:50.458 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : parquet-libs-9.0.0-15.el9.x86_64 12/20 2026-03-06T13:43:50.458 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : python3-rados-2:19.2.3-47.gc24117fd552.el9.clyso.x 13/20 2026-03-06T13:43:50.458 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : python3-rbd-2:19.2.3-47.gc24117fd552.el9.clyso.x86 14/20 2026-03-06T13:43:50.458 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : python3-rgw-2:19.2.3-47.gc24117fd552.el9.clyso.x86 15/20 2026-03-06T13:43:50.458 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : qemu-kvm-block-rbd-17:10.1.0-15.el9.x86_64 16/20 2026-03-06T13:43:50.458 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : rbd-fuse-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 17/20 2026-03-06T13:43:50.458 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : rbd-nbd-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 18/20 2026-03-06T13:43:50.458 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : re2-1:20211101-20.el9.x86_64 19/20 2026-03-06T13:43:50.460 INFO:teuthology.orchestra.run.vm01.stdout:Dependencies resolved. 2026-03-06T13:43:50.462 INFO:teuthology.orchestra.run.vm01.stdout:================================================================================ 2026-03-06T13:43:50.462 INFO:teuthology.orchestra.run.vm01.stdout: Package Arch Version Repository Size 2026-03-06T13:43:50.462 INFO:teuthology.orchestra.run.vm01.stdout:================================================================================ 2026-03-06T13:43:50.462 INFO:teuthology.orchestra.run.vm01.stdout:Removing: 2026-03-06T13:43:50.462 INFO:teuthology.orchestra.run.vm01.stdout: librados2 x86_64 2:19.2.3-47.gc24117fd552.el9.clyso @ceph 13 M 2026-03-06T13:43:50.462 INFO:teuthology.orchestra.run.vm01.stdout:Removing dependent packages: 2026-03-06T13:43:50.462 INFO:teuthology.orchestra.run.vm01.stdout: python3-rados x86_64 2:19.2.3-47.gc24117fd552.el9.clyso @ceph 1.1 M 2026-03-06T13:43:50.462 INFO:teuthology.orchestra.run.vm01.stdout: python3-rbd x86_64 2:19.2.3-47.gc24117fd552.el9.clyso @ceph 1.1 M 2026-03-06T13:43:50.462 INFO:teuthology.orchestra.run.vm01.stdout: python3-rgw x86_64 2:19.2.3-47.gc24117fd552.el9.clyso @ceph 265 k 2026-03-06T13:43:50.462 INFO:teuthology.orchestra.run.vm01.stdout: qemu-kvm-block-rbd x86_64 17:10.1.0-15.el9 @appstream 37 k 2026-03-06T13:43:50.462 INFO:teuthology.orchestra.run.vm01.stdout: rbd-fuse x86_64 2:19.2.3-47.gc24117fd552.el9.clyso @ceph 227 k 2026-03-06T13:43:50.462 INFO:teuthology.orchestra.run.vm01.stdout: rbd-nbd x86_64 2:19.2.3-47.gc24117fd552.el9.clyso @ceph 494 k 2026-03-06T13:43:50.462 INFO:teuthology.orchestra.run.vm01.stdout:Removing unused dependencies: 2026-03-06T13:43:50.462 INFO:teuthology.orchestra.run.vm01.stdout: boost-program-options 2026-03-06T13:43:50.462 INFO:teuthology.orchestra.run.vm01.stdout: x86_64 1.75.0-13.el9 @appstream 276 k 2026-03-06T13:43:50.462 INFO:teuthology.orchestra.run.vm01.stdout: libarrow x86_64 9.0.0-15.el9 @epel 18 M 2026-03-06T13:43:50.462 INFO:teuthology.orchestra.run.vm01.stdout: libarrow-doc noarch 9.0.0-15.el9 @epel 122 k 2026-03-06T13:43:50.462 INFO:teuthology.orchestra.run.vm01.stdout: libnbd x86_64 1.20.3-4.el9 @appstream 453 k 2026-03-06T13:43:50.462 INFO:teuthology.orchestra.run.vm01.stdout: libpmemobj x86_64 1.12.1-1.el9 @appstream 383 k 2026-03-06T13:43:50.463 INFO:teuthology.orchestra.run.vm01.stdout: librabbitmq x86_64 0.11.0-7.el9 @appstream 102 k 2026-03-06T13:43:50.463 INFO:teuthology.orchestra.run.vm01.stdout: librbd1 x86_64 2:19.2.3-47.gc24117fd552.el9.clyso @ceph 13 M 2026-03-06T13:43:50.463 INFO:teuthology.orchestra.run.vm01.stdout: librdkafka x86_64 1.6.1-102.el9 @appstream 2.0 M 2026-03-06T13:43:50.463 INFO:teuthology.orchestra.run.vm01.stdout: librgw2 x86_64 2:19.2.3-47.gc24117fd552.el9.clyso @ceph 19 M 2026-03-06T13:43:50.463 INFO:teuthology.orchestra.run.vm01.stdout: lttng-ust x86_64 2.12.0-6.el9 @appstream 1.0 M 2026-03-06T13:43:50.463 INFO:teuthology.orchestra.run.vm01.stdout: parquet-libs x86_64 9.0.0-15.el9 @epel 2.8 M 2026-03-06T13:43:50.463 INFO:teuthology.orchestra.run.vm01.stdout: re2 x86_64 1:20211101-20.el9 @epel 472 k 2026-03-06T13:43:50.463 INFO:teuthology.orchestra.run.vm01.stdout: thrift x86_64 0.15.0-4.el9 @epel 4.8 M 2026-03-06T13:43:50.463 INFO:teuthology.orchestra.run.vm01.stdout: 2026-03-06T13:43:50.463 INFO:teuthology.orchestra.run.vm01.stdout:Transaction Summary 2026-03-06T13:43:50.463 INFO:teuthology.orchestra.run.vm01.stdout:================================================================================ 2026-03-06T13:43:50.463 INFO:teuthology.orchestra.run.vm01.stdout:Remove 20 Packages 2026-03-06T13:43:50.463 INFO:teuthology.orchestra.run.vm01.stdout: 2026-03-06T13:43:50.463 INFO:teuthology.orchestra.run.vm01.stdout:Freed space: 79 M 2026-03-06T13:43:50.463 INFO:teuthology.orchestra.run.vm01.stdout:Running transaction check 2026-03-06T13:43:50.467 INFO:teuthology.orchestra.run.vm01.stdout:Transaction check succeeded. 2026-03-06T13:43:50.467 INFO:teuthology.orchestra.run.vm01.stdout:Running transaction test 2026-03-06T13:43:50.490 INFO:teuthology.orchestra.run.vm01.stdout:Transaction test succeeded. 2026-03-06T13:43:50.490 INFO:teuthology.orchestra.run.vm01.stdout:Running transaction 2026-03-06T13:43:50.497 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : thrift-0.15.0-4.el9.x86_64 20/20 2026-03-06T13:43:50.497 INFO:teuthology.orchestra.run.vm08.stdout: 2026-03-06T13:43:50.497 INFO:teuthology.orchestra.run.vm08.stdout:Removed: 2026-03-06T13:43:50.497 INFO:teuthology.orchestra.run.vm08.stdout: boost-program-options-1.75.0-13.el9.x86_64 2026-03-06T13:43:50.497 INFO:teuthology.orchestra.run.vm08.stdout: libarrow-9.0.0-15.el9.x86_64 2026-03-06T13:43:50.497 INFO:teuthology.orchestra.run.vm08.stdout: libarrow-doc-9.0.0-15.el9.noarch 2026-03-06T13:43:50.497 INFO:teuthology.orchestra.run.vm08.stdout: libnbd-1.20.3-4.el9.x86_64 2026-03-06T13:43:50.497 INFO:teuthology.orchestra.run.vm08.stdout: libpmemobj-1.12.1-1.el9.x86_64 2026-03-06T13:43:50.497 INFO:teuthology.orchestra.run.vm08.stdout: librabbitmq-0.11.0-7.el9.x86_64 2026-03-06T13:43:50.497 INFO:teuthology.orchestra.run.vm08.stdout: librados2-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 2026-03-06T13:43:50.497 INFO:teuthology.orchestra.run.vm08.stdout: librbd1-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 2026-03-06T13:43:50.497 INFO:teuthology.orchestra.run.vm08.stdout: librdkafka-1.6.1-102.el9.x86_64 2026-03-06T13:43:50.497 INFO:teuthology.orchestra.run.vm08.stdout: librgw2-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 2026-03-06T13:43:50.497 INFO:teuthology.orchestra.run.vm08.stdout: lttng-ust-2.12.0-6.el9.x86_64 2026-03-06T13:43:50.497 INFO:teuthology.orchestra.run.vm08.stdout: parquet-libs-9.0.0-15.el9.x86_64 2026-03-06T13:43:50.497 INFO:teuthology.orchestra.run.vm08.stdout: python3-rados-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 2026-03-06T13:43:50.497 INFO:teuthology.orchestra.run.vm08.stdout: python3-rbd-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 2026-03-06T13:43:50.497 INFO:teuthology.orchestra.run.vm08.stdout: python3-rgw-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 2026-03-06T13:43:50.497 INFO:teuthology.orchestra.run.vm08.stdout: qemu-kvm-block-rbd-17:10.1.0-15.el9.x86_64 2026-03-06T13:43:50.497 INFO:teuthology.orchestra.run.vm08.stdout: rbd-fuse-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 2026-03-06T13:43:50.497 INFO:teuthology.orchestra.run.vm08.stdout: rbd-nbd-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 2026-03-06T13:43:50.497 INFO:teuthology.orchestra.run.vm08.stdout: re2-1:20211101-20.el9.x86_64 2026-03-06T13:43:50.498 INFO:teuthology.orchestra.run.vm08.stdout: thrift-0.15.0-4.el9.x86_64 2026-03-06T13:43:50.498 INFO:teuthology.orchestra.run.vm08.stdout: 2026-03-06T13:43:50.498 INFO:teuthology.orchestra.run.vm08.stdout:Complete! 2026-03-06T13:43:50.533 INFO:teuthology.orchestra.run.vm01.stdout: Preparing : 1/1 2026-03-06T13:43:50.536 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : rbd-nbd-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 1/20 2026-03-06T13:43:50.538 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : rbd-fuse-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 2/20 2026-03-06T13:43:50.541 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : python3-rgw-2:19.2.3-47.gc24117fd552.el9.clyso.x86 3/20 2026-03-06T13:43:50.541 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : librgw2-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 4/20 2026-03-06T13:43:50.556 INFO:teuthology.orchestra.run.vm01.stdout: Running scriptlet: librgw2-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 4/20 2026-03-06T13:43:50.558 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : parquet-libs-9.0.0-15.el9.x86_64 5/20 2026-03-06T13:43:50.560 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : python3-rbd-2:19.2.3-47.gc24117fd552.el9.clyso.x86 6/20 2026-03-06T13:43:50.560 INFO:teuthology.orchestra.run.vm00.stdout:No match for argument: librbd1 2026-03-06T13:43:50.561 INFO:teuthology.orchestra.run.vm00.stderr:No packages marked for removal. 2026-03-06T13:43:50.562 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : python3-rados-2:19.2.3-47.gc24117fd552.el9.clyso.x 7/20 2026-03-06T13:43:50.563 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : qemu-kvm-block-rbd-17:10.1.0-15.el9.x86_64 8/20 2026-03-06T13:43:50.564 INFO:teuthology.orchestra.run.vm00.stdout:Dependencies resolved. 2026-03-06T13:43:50.565 INFO:teuthology.orchestra.run.vm00.stdout:Nothing to do. 2026-03-06T13:43:50.565 INFO:teuthology.orchestra.run.vm00.stdout:Complete! 2026-03-06T13:43:50.565 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : libarrow-doc-9.0.0-15.el9.noarch 9/20 2026-03-06T13:43:50.566 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : librbd1-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 10/20 2026-03-06T13:43:50.583 INFO:teuthology.orchestra.run.vm01.stdout: Running scriptlet: librbd1-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 10/20 2026-03-06T13:43:50.584 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : librados2-2:19.2.3-47.gc24117fd552.el9.clyso.x86_6 11/20 2026-03-06T13:43:50.584 INFO:teuthology.orchestra.run.vm01.stdout:warning: file /etc/ceph: remove failed: No such file or directory 2026-03-06T13:43:50.584 INFO:teuthology.orchestra.run.vm01.stdout: 2026-03-06T13:43:50.598 INFO:teuthology.orchestra.run.vm01.stdout: Running scriptlet: librados2-2:19.2.3-47.gc24117fd552.el9.clyso.x86_6 11/20 2026-03-06T13:43:50.601 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : libarrow-9.0.0-15.el9.x86_64 12/20 2026-03-06T13:43:50.643 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : re2-1:20211101-20.el9.x86_64 13/20 2026-03-06T13:43:50.647 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : lttng-ust-2.12.0-6.el9.x86_64 14/20 2026-03-06T13:43:50.649 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : thrift-0.15.0-4.el9.x86_64 15/20 2026-03-06T13:43:50.652 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : libnbd-1.20.3-4.el9.x86_64 16/20 2026-03-06T13:43:50.654 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : libpmemobj-1.12.1-1.el9.x86_64 17/20 2026-03-06T13:43:50.656 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : boost-program-options-1.75.0-13.el9.x86_64 18/20 2026-03-06T13:43:50.659 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : librabbitmq-0.11.0-7.el9.x86_64 19/20 2026-03-06T13:43:50.673 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : librdkafka-1.6.1-102.el9.x86_64 20/20 2026-03-06T13:43:50.698 INFO:teuthology.orchestra.run.vm08.stdout:No match for argument: librbd1 2026-03-06T13:43:50.698 INFO:teuthology.orchestra.run.vm08.stderr:No packages marked for removal. 2026-03-06T13:43:50.700 INFO:teuthology.orchestra.run.vm08.stdout:Dependencies resolved. 2026-03-06T13:43:50.701 INFO:teuthology.orchestra.run.vm08.stdout:Nothing to do. 2026-03-06T13:43:50.702 INFO:teuthology.orchestra.run.vm08.stdout:Complete! 2026-03-06T13:43:50.734 INFO:teuthology.orchestra.run.vm01.stdout: Running scriptlet: librdkafka-1.6.1-102.el9.x86_64 20/20 2026-03-06T13:43:50.734 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : boost-program-options-1.75.0-13.el9.x86_64 1/20 2026-03-06T13:43:50.734 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : libarrow-9.0.0-15.el9.x86_64 2/20 2026-03-06T13:43:50.734 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : libarrow-doc-9.0.0-15.el9.noarch 3/20 2026-03-06T13:43:50.734 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : libnbd-1.20.3-4.el9.x86_64 4/20 2026-03-06T13:43:50.734 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : libpmemobj-1.12.1-1.el9.x86_64 5/20 2026-03-06T13:43:50.734 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : librabbitmq-0.11.0-7.el9.x86_64 6/20 2026-03-06T13:43:50.734 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : librados2-2:19.2.3-47.gc24117fd552.el9.clyso.x86_6 7/20 2026-03-06T13:43:50.734 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : librbd1-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 8/20 2026-03-06T13:43:50.734 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : librdkafka-1.6.1-102.el9.x86_64 9/20 2026-03-06T13:43:50.734 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : librgw2-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 10/20 2026-03-06T13:43:50.734 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : lttng-ust-2.12.0-6.el9.x86_64 11/20 2026-03-06T13:43:50.734 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : parquet-libs-9.0.0-15.el9.x86_64 12/20 2026-03-06T13:43:50.734 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-rados-2:19.2.3-47.gc24117fd552.el9.clyso.x 13/20 2026-03-06T13:43:50.734 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-rbd-2:19.2.3-47.gc24117fd552.el9.clyso.x86 14/20 2026-03-06T13:43:50.734 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-rgw-2:19.2.3-47.gc24117fd552.el9.clyso.x86 15/20 2026-03-06T13:43:50.734 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : qemu-kvm-block-rbd-17:10.1.0-15.el9.x86_64 16/20 2026-03-06T13:43:50.734 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : rbd-fuse-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 17/20 2026-03-06T13:43:50.734 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : rbd-nbd-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 18/20 2026-03-06T13:43:50.734 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : re2-1:20211101-20.el9.x86_64 19/20 2026-03-06T13:43:50.756 INFO:teuthology.orchestra.run.vm00.stdout:No match for argument: python3-rados 2026-03-06T13:43:50.756 INFO:teuthology.orchestra.run.vm00.stderr:No packages marked for removal. 2026-03-06T13:43:50.759 INFO:teuthology.orchestra.run.vm00.stdout:Dependencies resolved. 2026-03-06T13:43:50.759 INFO:teuthology.orchestra.run.vm00.stdout:Nothing to do. 2026-03-06T13:43:50.759 INFO:teuthology.orchestra.run.vm00.stdout:Complete! 2026-03-06T13:43:50.782 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : thrift-0.15.0-4.el9.x86_64 20/20 2026-03-06T13:43:50.782 INFO:teuthology.orchestra.run.vm01.stdout: 2026-03-06T13:43:50.782 INFO:teuthology.orchestra.run.vm01.stdout:Removed: 2026-03-06T13:43:50.783 INFO:teuthology.orchestra.run.vm01.stdout: boost-program-options-1.75.0-13.el9.x86_64 2026-03-06T13:43:50.783 INFO:teuthology.orchestra.run.vm01.stdout: libarrow-9.0.0-15.el9.x86_64 2026-03-06T13:43:50.783 INFO:teuthology.orchestra.run.vm01.stdout: libarrow-doc-9.0.0-15.el9.noarch 2026-03-06T13:43:50.783 INFO:teuthology.orchestra.run.vm01.stdout: libnbd-1.20.3-4.el9.x86_64 2026-03-06T13:43:50.783 INFO:teuthology.orchestra.run.vm01.stdout: libpmemobj-1.12.1-1.el9.x86_64 2026-03-06T13:43:50.783 INFO:teuthology.orchestra.run.vm01.stdout: librabbitmq-0.11.0-7.el9.x86_64 2026-03-06T13:43:50.783 INFO:teuthology.orchestra.run.vm01.stdout: librados2-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 2026-03-06T13:43:50.783 INFO:teuthology.orchestra.run.vm01.stdout: librbd1-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 2026-03-06T13:43:50.783 INFO:teuthology.orchestra.run.vm01.stdout: librdkafka-1.6.1-102.el9.x86_64 2026-03-06T13:43:50.783 INFO:teuthology.orchestra.run.vm01.stdout: librgw2-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 2026-03-06T13:43:50.783 INFO:teuthology.orchestra.run.vm01.stdout: lttng-ust-2.12.0-6.el9.x86_64 2026-03-06T13:43:50.783 INFO:teuthology.orchestra.run.vm01.stdout: parquet-libs-9.0.0-15.el9.x86_64 2026-03-06T13:43:50.783 INFO:teuthology.orchestra.run.vm01.stdout: python3-rados-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 2026-03-06T13:43:50.783 INFO:teuthology.orchestra.run.vm01.stdout: python3-rbd-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 2026-03-06T13:43:50.783 INFO:teuthology.orchestra.run.vm01.stdout: python3-rgw-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 2026-03-06T13:43:50.783 INFO:teuthology.orchestra.run.vm01.stdout: qemu-kvm-block-rbd-17:10.1.0-15.el9.x86_64 2026-03-06T13:43:50.783 INFO:teuthology.orchestra.run.vm01.stdout: rbd-fuse-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 2026-03-06T13:43:50.783 INFO:teuthology.orchestra.run.vm01.stdout: rbd-nbd-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 2026-03-06T13:43:50.783 INFO:teuthology.orchestra.run.vm01.stdout: re2-1:20211101-20.el9.x86_64 2026-03-06T13:43:50.783 INFO:teuthology.orchestra.run.vm01.stdout: thrift-0.15.0-4.el9.x86_64 2026-03-06T13:43:50.783 INFO:teuthology.orchestra.run.vm01.stdout: 2026-03-06T13:43:50.783 INFO:teuthology.orchestra.run.vm01.stdout:Complete! 2026-03-06T13:43:50.893 INFO:teuthology.orchestra.run.vm08.stdout:No match for argument: python3-rados 2026-03-06T13:43:50.893 INFO:teuthology.orchestra.run.vm08.stderr:No packages marked for removal. 2026-03-06T13:43:50.895 INFO:teuthology.orchestra.run.vm08.stdout:Dependencies resolved. 2026-03-06T13:43:50.895 INFO:teuthology.orchestra.run.vm08.stdout:Nothing to do. 2026-03-06T13:43:50.895 INFO:teuthology.orchestra.run.vm08.stdout:Complete! 2026-03-06T13:43:50.933 INFO:teuthology.orchestra.run.vm00.stdout:No match for argument: python3-rgw 2026-03-06T13:43:50.933 INFO:teuthology.orchestra.run.vm00.stderr:No packages marked for removal. 2026-03-06T13:43:50.935 INFO:teuthology.orchestra.run.vm00.stdout:Dependencies resolved. 2026-03-06T13:43:50.936 INFO:teuthology.orchestra.run.vm00.stdout:Nothing to do. 2026-03-06T13:43:50.936 INFO:teuthology.orchestra.run.vm00.stdout:Complete! 2026-03-06T13:43:51.006 INFO:teuthology.orchestra.run.vm01.stdout:No match for argument: librbd1 2026-03-06T13:43:51.006 INFO:teuthology.orchestra.run.vm01.stderr:No packages marked for removal. 2026-03-06T13:43:51.008 INFO:teuthology.orchestra.run.vm01.stdout:Dependencies resolved. 2026-03-06T13:43:51.008 INFO:teuthology.orchestra.run.vm01.stdout:Nothing to do. 2026-03-06T13:43:51.009 INFO:teuthology.orchestra.run.vm01.stdout:Complete! 2026-03-06T13:43:51.063 INFO:teuthology.orchestra.run.vm08.stdout:No match for argument: python3-rgw 2026-03-06T13:43:51.063 INFO:teuthology.orchestra.run.vm08.stderr:No packages marked for removal. 2026-03-06T13:43:51.065 INFO:teuthology.orchestra.run.vm08.stdout:Dependencies resolved. 2026-03-06T13:43:51.066 INFO:teuthology.orchestra.run.vm08.stdout:Nothing to do. 2026-03-06T13:43:51.066 INFO:teuthology.orchestra.run.vm08.stdout:Complete! 2026-03-06T13:43:51.108 INFO:teuthology.orchestra.run.vm00.stdout:No match for argument: python3-cephfs 2026-03-06T13:43:51.109 INFO:teuthology.orchestra.run.vm00.stderr:No packages marked for removal. 2026-03-06T13:43:51.111 INFO:teuthology.orchestra.run.vm00.stdout:Dependencies resolved. 2026-03-06T13:43:51.112 INFO:teuthology.orchestra.run.vm00.stdout:Nothing to do. 2026-03-06T13:43:51.112 INFO:teuthology.orchestra.run.vm00.stdout:Complete! 2026-03-06T13:43:51.200 INFO:teuthology.orchestra.run.vm01.stdout:No match for argument: python3-rados 2026-03-06T13:43:51.201 INFO:teuthology.orchestra.run.vm01.stderr:No packages marked for removal. 2026-03-06T13:43:51.203 INFO:teuthology.orchestra.run.vm01.stdout:Dependencies resolved. 2026-03-06T13:43:51.204 INFO:teuthology.orchestra.run.vm01.stdout:Nothing to do. 2026-03-06T13:43:51.204 INFO:teuthology.orchestra.run.vm01.stdout:Complete! 2026-03-06T13:43:51.243 INFO:teuthology.orchestra.run.vm08.stdout:No match for argument: python3-cephfs 2026-03-06T13:43:51.243 INFO:teuthology.orchestra.run.vm08.stderr:No packages marked for removal. 2026-03-06T13:43:51.245 INFO:teuthology.orchestra.run.vm08.stdout:Dependencies resolved. 2026-03-06T13:43:51.246 INFO:teuthology.orchestra.run.vm08.stdout:Nothing to do. 2026-03-06T13:43:51.246 INFO:teuthology.orchestra.run.vm08.stdout:Complete! 2026-03-06T13:43:51.283 INFO:teuthology.orchestra.run.vm00.stdout:No match for argument: python3-rbd 2026-03-06T13:43:51.283 INFO:teuthology.orchestra.run.vm00.stderr:No packages marked for removal. 2026-03-06T13:43:51.285 INFO:teuthology.orchestra.run.vm00.stdout:Dependencies resolved. 2026-03-06T13:43:51.286 INFO:teuthology.orchestra.run.vm00.stdout:Nothing to do. 2026-03-06T13:43:51.286 INFO:teuthology.orchestra.run.vm00.stdout:Complete! 2026-03-06T13:43:51.386 INFO:teuthology.orchestra.run.vm01.stdout:No match for argument: python3-rgw 2026-03-06T13:43:51.386 INFO:teuthology.orchestra.run.vm01.stderr:No packages marked for removal. 2026-03-06T13:43:51.389 INFO:teuthology.orchestra.run.vm01.stdout:Dependencies resolved. 2026-03-06T13:43:51.389 INFO:teuthology.orchestra.run.vm01.stdout:Nothing to do. 2026-03-06T13:43:51.389 INFO:teuthology.orchestra.run.vm01.stdout:Complete! 2026-03-06T13:43:51.421 INFO:teuthology.orchestra.run.vm08.stdout:No match for argument: python3-rbd 2026-03-06T13:43:51.421 INFO:teuthology.orchestra.run.vm08.stderr:No packages marked for removal. 2026-03-06T13:43:51.424 INFO:teuthology.orchestra.run.vm08.stdout:Dependencies resolved. 2026-03-06T13:43:51.425 INFO:teuthology.orchestra.run.vm08.stdout:Nothing to do. 2026-03-06T13:43:51.425 INFO:teuthology.orchestra.run.vm08.stdout:Complete! 2026-03-06T13:43:51.469 INFO:teuthology.orchestra.run.vm00.stdout:No match for argument: rbd-fuse 2026-03-06T13:43:51.469 INFO:teuthology.orchestra.run.vm00.stderr:No packages marked for removal. 2026-03-06T13:43:51.472 INFO:teuthology.orchestra.run.vm00.stdout:Dependencies resolved. 2026-03-06T13:43:51.473 INFO:teuthology.orchestra.run.vm00.stdout:Nothing to do. 2026-03-06T13:43:51.473 INFO:teuthology.orchestra.run.vm00.stdout:Complete! 2026-03-06T13:43:51.576 INFO:teuthology.orchestra.run.vm01.stdout:No match for argument: python3-cephfs 2026-03-06T13:43:51.576 INFO:teuthology.orchestra.run.vm01.stderr:No packages marked for removal. 2026-03-06T13:43:51.579 INFO:teuthology.orchestra.run.vm01.stdout:Dependencies resolved. 2026-03-06T13:43:51.580 INFO:teuthology.orchestra.run.vm01.stdout:Nothing to do. 2026-03-06T13:43:51.580 INFO:teuthology.orchestra.run.vm01.stdout:Complete! 2026-03-06T13:43:51.616 INFO:teuthology.orchestra.run.vm08.stdout:No match for argument: rbd-fuse 2026-03-06T13:43:51.616 INFO:teuthology.orchestra.run.vm08.stderr:No packages marked for removal. 2026-03-06T13:43:51.618 INFO:teuthology.orchestra.run.vm08.stdout:Dependencies resolved. 2026-03-06T13:43:51.618 INFO:teuthology.orchestra.run.vm08.stdout:Nothing to do. 2026-03-06T13:43:51.619 INFO:teuthology.orchestra.run.vm08.stdout:Complete! 2026-03-06T13:43:51.656 INFO:teuthology.orchestra.run.vm00.stdout:No match for argument: rbd-mirror 2026-03-06T13:43:51.656 INFO:teuthology.orchestra.run.vm00.stderr:No packages marked for removal. 2026-03-06T13:43:51.659 INFO:teuthology.orchestra.run.vm00.stdout:Dependencies resolved. 2026-03-06T13:43:51.660 INFO:teuthology.orchestra.run.vm00.stdout:Nothing to do. 2026-03-06T13:43:51.660 INFO:teuthology.orchestra.run.vm00.stdout:Complete! 2026-03-06T13:43:51.769 INFO:teuthology.orchestra.run.vm01.stdout:No match for argument: python3-rbd 2026-03-06T13:43:51.769 INFO:teuthology.orchestra.run.vm01.stderr:No packages marked for removal. 2026-03-06T13:43:51.771 INFO:teuthology.orchestra.run.vm01.stdout:Dependencies resolved. 2026-03-06T13:43:51.771 INFO:teuthology.orchestra.run.vm01.stdout:Nothing to do. 2026-03-06T13:43:51.771 INFO:teuthology.orchestra.run.vm01.stdout:Complete! 2026-03-06T13:43:51.790 INFO:teuthology.orchestra.run.vm08.stdout:No match for argument: rbd-mirror 2026-03-06T13:43:51.791 INFO:teuthology.orchestra.run.vm08.stderr:No packages marked for removal. 2026-03-06T13:43:51.793 INFO:teuthology.orchestra.run.vm08.stdout:Dependencies resolved. 2026-03-06T13:43:51.794 INFO:teuthology.orchestra.run.vm08.stdout:Nothing to do. 2026-03-06T13:43:51.794 INFO:teuthology.orchestra.run.vm08.stdout:Complete! 2026-03-06T13:43:51.842 INFO:teuthology.orchestra.run.vm00.stdout:No match for argument: rbd-nbd 2026-03-06T13:43:51.843 INFO:teuthology.orchestra.run.vm00.stderr:No packages marked for removal. 2026-03-06T13:43:51.845 INFO:teuthology.orchestra.run.vm00.stdout:Dependencies resolved. 2026-03-06T13:43:51.845 INFO:teuthology.orchestra.run.vm00.stdout:Nothing to do. 2026-03-06T13:43:51.845 INFO:teuthology.orchestra.run.vm00.stdout:Complete! 2026-03-06T13:43:51.869 DEBUG:teuthology.orchestra.run.vm00:> sudo yum clean all 2026-03-06T13:43:51.948 INFO:teuthology.orchestra.run.vm01.stdout:No match for argument: rbd-fuse 2026-03-06T13:43:51.948 INFO:teuthology.orchestra.run.vm01.stderr:No packages marked for removal. 2026-03-06T13:43:51.950 INFO:teuthology.orchestra.run.vm01.stdout:Dependencies resolved. 2026-03-06T13:43:51.950 INFO:teuthology.orchestra.run.vm01.stdout:Nothing to do. 2026-03-06T13:43:51.951 INFO:teuthology.orchestra.run.vm01.stdout:Complete! 2026-03-06T13:43:51.968 INFO:teuthology.orchestra.run.vm08.stdout:No match for argument: rbd-nbd 2026-03-06T13:43:51.968 INFO:teuthology.orchestra.run.vm08.stderr:No packages marked for removal. 2026-03-06T13:43:51.970 INFO:teuthology.orchestra.run.vm08.stdout:Dependencies resolved. 2026-03-06T13:43:51.971 INFO:teuthology.orchestra.run.vm08.stdout:Nothing to do. 2026-03-06T13:43:51.971 INFO:teuthology.orchestra.run.vm08.stdout:Complete! 2026-03-06T13:43:51.990 DEBUG:teuthology.orchestra.run.vm08:> sudo yum clean all 2026-03-06T13:43:52.000 INFO:teuthology.orchestra.run.vm00.stdout:56 files removed 2026-03-06T13:43:52.019 DEBUG:teuthology.orchestra.run.vm00:> sudo rm /etc/yum.repos.d/ceph-source.repo 2026-03-06T13:43:52.050 DEBUG:teuthology.orchestra.run.vm00:> sudo rm /etc/yum.repos.d/ceph-noarch.repo 2026-03-06T13:43:52.123 INFO:teuthology.orchestra.run.vm08.stdout:56 files removed 2026-03-06T13:43:52.124 DEBUG:teuthology.orchestra.run.vm00:> sudo rm /etc/yum.repos.d/ceph.repo 2026-03-06T13:43:52.130 INFO:teuthology.orchestra.run.vm01.stdout:No match for argument: rbd-mirror 2026-03-06T13:43:52.130 INFO:teuthology.orchestra.run.vm01.stderr:No packages marked for removal. 2026-03-06T13:43:52.132 INFO:teuthology.orchestra.run.vm01.stdout:Dependencies resolved. 2026-03-06T13:43:52.133 INFO:teuthology.orchestra.run.vm01.stdout:Nothing to do. 2026-03-06T13:43:52.133 INFO:teuthology.orchestra.run.vm01.stdout:Complete! 2026-03-06T13:43:52.139 DEBUG:teuthology.orchestra.run.vm08:> sudo rm /etc/yum.repos.d/ceph-source.repo 2026-03-06T13:43:52.167 DEBUG:teuthology.orchestra.run.vm08:> sudo rm /etc/yum.repos.d/ceph-noarch.repo 2026-03-06T13:43:52.194 DEBUG:teuthology.orchestra.run.vm00:> sudo yum clean expire-cache 2026-03-06T13:43:52.234 DEBUG:teuthology.orchestra.run.vm08:> sudo rm /etc/yum.repos.d/ceph.repo 2026-03-06T13:43:52.294 INFO:teuthology.orchestra.run.vm01.stdout:No match for argument: rbd-nbd 2026-03-06T13:43:52.295 INFO:teuthology.orchestra.run.vm01.stderr:No packages marked for removal. 2026-03-06T13:43:52.297 DEBUG:teuthology.orchestra.run.vm08:> sudo yum clean expire-cache 2026-03-06T13:43:52.297 INFO:teuthology.orchestra.run.vm01.stdout:Dependencies resolved. 2026-03-06T13:43:52.297 INFO:teuthology.orchestra.run.vm01.stdout:Nothing to do. 2026-03-06T13:43:52.298 INFO:teuthology.orchestra.run.vm01.stdout:Complete! 2026-03-06T13:43:52.322 DEBUG:teuthology.orchestra.run.vm01:> sudo yum clean all 2026-03-06T13:43:52.359 INFO:teuthology.orchestra.run.vm00.stdout:Cache was expired 2026-03-06T13:43:52.360 INFO:teuthology.orchestra.run.vm00.stdout:0 files removed 2026-03-06T13:43:52.383 DEBUG:teuthology.parallel:result is None 2026-03-06T13:43:52.444 INFO:teuthology.orchestra.run.vm08.stdout:Cache was expired 2026-03-06T13:43:52.444 INFO:teuthology.orchestra.run.vm08.stdout:0 files removed 2026-03-06T13:43:52.455 INFO:teuthology.orchestra.run.vm01.stdout:56 files removed 2026-03-06T13:43:52.461 DEBUG:teuthology.parallel:result is None 2026-03-06T13:43:52.483 DEBUG:teuthology.orchestra.run.vm01:> sudo rm /etc/yum.repos.d/ceph-source.repo 2026-03-06T13:43:52.507 DEBUG:teuthology.orchestra.run.vm01:> sudo rm /etc/yum.repos.d/ceph-noarch.repo 2026-03-06T13:43:52.578 DEBUG:teuthology.orchestra.run.vm01:> sudo rm /etc/yum.repos.d/ceph.repo 2026-03-06T13:43:52.644 DEBUG:teuthology.orchestra.run.vm01:> sudo yum clean expire-cache 2026-03-06T13:43:52.802 INFO:teuthology.orchestra.run.vm01.stdout:Cache was expired 2026-03-06T13:43:52.802 INFO:teuthology.orchestra.run.vm01.stdout:0 files removed 2026-03-06T13:43:52.819 DEBUG:teuthology.parallel:result is None 2026-03-06T13:43:52.820 INFO:teuthology.task.install:Removing ceph sources lists on ubuntu@vm00.local 2026-03-06T13:43:52.820 INFO:teuthology.task.install:Removing ceph sources lists on ubuntu@vm01.local 2026-03-06T13:43:52.820 INFO:teuthology.task.install:Removing ceph sources lists on ubuntu@vm08.local 2026-03-06T13:43:52.820 DEBUG:teuthology.orchestra.run.vm00:> sudo rm -f /etc/yum.repos.d/ceph.repo 2026-03-06T13:43:52.820 DEBUG:teuthology.orchestra.run.vm01:> sudo rm -f /etc/yum.repos.d/ceph.repo 2026-03-06T13:43:52.820 DEBUG:teuthology.orchestra.run.vm08:> sudo rm -f /etc/yum.repos.d/ceph.repo 2026-03-06T13:43:52.846 DEBUG:teuthology.orchestra.run.vm08:> sudo mv -f /etc/yum/pluginconf.d/priorities.conf.orig /etc/yum/pluginconf.d/priorities.conf 2026-03-06T13:43:52.848 DEBUG:teuthology.orchestra.run.vm00:> sudo mv -f /etc/yum/pluginconf.d/priorities.conf.orig /etc/yum/pluginconf.d/priorities.conf 2026-03-06T13:43:52.850 DEBUG:teuthology.orchestra.run.vm01:> sudo mv -f /etc/yum/pluginconf.d/priorities.conf.orig /etc/yum/pluginconf.d/priorities.conf 2026-03-06T13:43:52.911 INFO:teuthology.orchestra.run.vm08.stderr:mv: cannot stat '/etc/yum/pluginconf.d/priorities.conf.orig': No such file or directory 2026-03-06T13:43:52.912 DEBUG:teuthology.orchestra.run:got remote process result: 1 2026-03-06T13:43:52.912 DEBUG:teuthology.parallel:result is None 2026-03-06T13:43:52.916 INFO:teuthology.orchestra.run.vm00.stderr:mv: cannot stat '/etc/yum/pluginconf.d/priorities.conf.orig': No such file or directory 2026-03-06T13:43:52.917 DEBUG:teuthology.orchestra.run:got remote process result: 1 2026-03-06T13:43:52.918 DEBUG:teuthology.parallel:result is None 2026-03-06T13:43:52.920 INFO:teuthology.orchestra.run.vm01.stderr:mv: cannot stat '/etc/yum/pluginconf.d/priorities.conf.orig': No such file or directory 2026-03-06T13:43:52.921 DEBUG:teuthology.orchestra.run:got remote process result: 1 2026-03-06T13:43:52.921 DEBUG:teuthology.parallel:result is None 2026-03-06T13:43:52.921 DEBUG:teuthology.run_tasks:Unwinding manager clock 2026-03-06T13:43:52.924 INFO:teuthology.task.clock:Checking final clock skew... 2026-03-06T13:43:52.924 DEBUG:teuthology.orchestra.run.vm00:> PATH=/usr/bin:/usr/sbin ntpq -p || PATH=/usr/bin:/usr/sbin chronyc sources || true 2026-03-06T13:43:52.960 DEBUG:teuthology.orchestra.run.vm01:> PATH=/usr/bin:/usr/sbin ntpq -p || PATH=/usr/bin:/usr/sbin chronyc sources || true 2026-03-06T13:43:52.964 DEBUG:teuthology.orchestra.run.vm08:> PATH=/usr/bin:/usr/sbin ntpq -p || PATH=/usr/bin:/usr/sbin chronyc sources || true 2026-03-06T13:43:52.978 INFO:teuthology.orchestra.run.vm00.stderr:bash: line 1: ntpq: command not found 2026-03-06T13:43:52.979 INFO:teuthology.orchestra.run.vm08.stderr:bash: line 1: ntpq: command not found 2026-03-06T13:43:52.981 INFO:teuthology.orchestra.run.vm01.stderr:bash: line 1: ntpq: command not found 2026-03-06T13:43:52.982 INFO:teuthology.orchestra.run.vm00.stdout:MS Name/IP address Stratum Poll Reach LastRx Last sample 2026-03-06T13:43:52.983 INFO:teuthology.orchestra.run.vm00.stdout:=============================================================================== 2026-03-06T13:43:52.983 INFO:teuthology.orchestra.run.vm00.stdout:^+ ntp2.wup-de.hosts.301-mo> 2 6 377 62 -1205us[-1381us] +/- 21ms 2026-03-06T13:43:52.983 INFO:teuthology.orchestra.run.vm00.stdout:^? ovh.saclay.org 2 7 40 326 +1016us[+2499us] +/- 23ms 2026-03-06T13:43:52.983 INFO:teuthology.orchestra.run.vm00.stdout:^+ tor.nocabal.de 2 6 377 61 +127us[ -49us] +/- 45ms 2026-03-06T13:43:52.983 INFO:teuthology.orchestra.run.vm00.stdout:^* ntp5.kernfusion.at 2 6 377 60 +68us[ -108us] +/- 17ms 2026-03-06T13:43:52.983 INFO:teuthology.orchestra.run.vm08.stdout:MS Name/IP address Stratum Poll Reach LastRx Last sample 2026-03-06T13:43:52.983 INFO:teuthology.orchestra.run.vm08.stdout:=============================================================================== 2026-03-06T13:43:52.983 INFO:teuthology.orchestra.run.vm08.stdout:^+ tor.nocabal.de 2 6 377 59 +388us[ +388us] +/- 45ms 2026-03-06T13:43:52.983 INFO:teuthology.orchestra.run.vm08.stdout:^* ntp5.kernfusion.at 2 6 377 62 +258us[ +139us] +/- 17ms 2026-03-06T13:43:52.983 INFO:teuthology.orchestra.run.vm08.stdout:^- 46.21.2.169.static.isp.r> 1 6 37 59 +1340us[+1340us] +/- 15ms 2026-03-06T13:43:52.983 INFO:teuthology.orchestra.run.vm08.stdout:^+ ntp2.wup-de.hosts.301-mo> 2 6 377 61 -1030us[-1030us] +/- 21ms 2026-03-06T13:43:52.984 INFO:teuthology.orchestra.run.vm01.stdout:MS Name/IP address Stratum Poll Reach LastRx Last sample 2026-03-06T13:43:52.984 INFO:teuthology.orchestra.run.vm01.stdout:=============================================================================== 2026-03-06T13:43:52.984 INFO:teuthology.orchestra.run.vm01.stdout:^? ovh.saclay.org 2 7 3 53 +1427us[+1427us] +/- 23ms 2026-03-06T13:43:52.984 INFO:teuthology.orchestra.run.vm01.stdout:^+ tor.nocabal.de 2 6 377 61 +587us[ +613us] +/- 45ms 2026-03-06T13:43:52.984 INFO:teuthology.orchestra.run.vm01.stdout:^+ ntp5.kernfusion.at 2 6 377 62 +546us[ +572us] +/- 17ms 2026-03-06T13:43:52.984 INFO:teuthology.orchestra.run.vm01.stdout:^* ntp2.wup-de.hosts.301-mo> 2 6 377 61 -764us[ -738us] +/- 21ms 2026-03-06T13:43:52.985 DEBUG:teuthology.run_tasks:Unwinding manager ansible.cephlab 2026-03-06T13:43:52.987 INFO:teuthology.task.ansible:Skipping ansible cleanup... 2026-03-06T13:43:52.987 DEBUG:teuthology.run_tasks:Unwinding manager selinux 2026-03-06T13:43:52.989 DEBUG:teuthology.run_tasks:Unwinding manager pcp 2026-03-06T13:43:52.991 DEBUG:teuthology.run_tasks:Unwinding manager internal.timer 2026-03-06T13:43:52.993 INFO:teuthology.task.internal:Duration was 462.669780 seconds 2026-03-06T13:43:52.993 DEBUG:teuthology.run_tasks:Unwinding manager internal.syslog 2026-03-06T13:43:52.995 INFO:teuthology.task.internal.syslog:Shutting down syslog monitoring... 2026-03-06T13:43:52.995 DEBUG:teuthology.orchestra.run.vm00:> sudo rm -f -- /etc/rsyslog.d/80-cephtest.conf && sudo service rsyslog restart 2026-03-06T13:43:53.026 DEBUG:teuthology.orchestra.run.vm01:> sudo rm -f -- /etc/rsyslog.d/80-cephtest.conf && sudo service rsyslog restart 2026-03-06T13:43:53.028 DEBUG:teuthology.orchestra.run.vm08:> sudo rm -f -- /etc/rsyslog.d/80-cephtest.conf && sudo service rsyslog restart 2026-03-06T13:43:53.061 INFO:teuthology.orchestra.run.vm00.stderr:Redirecting to /bin/systemctl restart rsyslog.service 2026-03-06T13:43:53.068 INFO:teuthology.orchestra.run.vm08.stderr:Redirecting to /bin/systemctl restart rsyslog.service 2026-03-06T13:43:53.069 INFO:teuthology.orchestra.run.vm01.stderr:Redirecting to /bin/systemctl restart rsyslog.service 2026-03-06T13:43:53.317 INFO:teuthology.task.internal.syslog:Checking logs for errors... 2026-03-06T13:43:53.318 DEBUG:teuthology.task.internal.syslog:Checking ubuntu@vm00.local 2026-03-06T13:43:53.318 DEBUG:teuthology.orchestra.run.vm00:> grep -E --binary-files=text '\bBUG\b|\bINFO\b|\bDEADLOCK\b' /home/ubuntu/cephtest/archive/syslog/kern.log | grep -v 'task .* blocked for more than .* seconds' | grep -v 'lockdep is turned off' | grep -v 'trying to register non-static key' | grep -v 'DEBUG: fsize' | grep -v CRON | grep -v 'BUG: bad unlock balance detected' | grep -v 'inconsistent lock state' | grep -v '*** DEADLOCK ***' | grep -v 'INFO: possible irq lock inversion dependency detected' | grep -v 'INFO: NMI handler (perf_event_nmi_handler) took too long to run' | grep -v 'INFO: recovery required on readonly' | grep -v 'ceph-create-keys: INFO' | grep -v INFO:ceph-create-keys | grep -v 'Loaded datasource DataSourceOpenStack' | grep -v 'container-storage-setup: INFO: Volume group backing root filesystem could not be determined' | grep -E -v '\bsalt-master\b|\bsalt-minion\b|\bsalt-api\b' | grep -v ceph-crash | grep -E -v '\btcmu-runner\b.*\bINFO\b' | head -n 1 2026-03-06T13:43:53.348 DEBUG:teuthology.task.internal.syslog:Checking ubuntu@vm01.local 2026-03-06T13:43:53.348 DEBUG:teuthology.orchestra.run.vm01:> grep -E --binary-files=text '\bBUG\b|\bINFO\b|\bDEADLOCK\b' /home/ubuntu/cephtest/archive/syslog/kern.log | grep -v 'task .* blocked for more than .* seconds' | grep -v 'lockdep is turned off' | grep -v 'trying to register non-static key' | grep -v 'DEBUG: fsize' | grep -v CRON | grep -v 'BUG: bad unlock balance detected' | grep -v 'inconsistent lock state' | grep -v '*** DEADLOCK ***' | grep -v 'INFO: possible irq lock inversion dependency detected' | grep -v 'INFO: NMI handler (perf_event_nmi_handler) took too long to run' | grep -v 'INFO: recovery required on readonly' | grep -v 'ceph-create-keys: INFO' | grep -v INFO:ceph-create-keys | grep -v 'Loaded datasource DataSourceOpenStack' | grep -v 'container-storage-setup: INFO: Volume group backing root filesystem could not be determined' | grep -E -v '\bsalt-master\b|\bsalt-minion\b|\bsalt-api\b' | grep -v ceph-crash | grep -E -v '\btcmu-runner\b.*\bINFO\b' | head -n 1 2026-03-06T13:43:53.376 DEBUG:teuthology.task.internal.syslog:Checking ubuntu@vm08.local 2026-03-06T13:43:53.376 DEBUG:teuthology.orchestra.run.vm08:> grep -E --binary-files=text '\bBUG\b|\bINFO\b|\bDEADLOCK\b' /home/ubuntu/cephtest/archive/syslog/kern.log | grep -v 'task .* blocked for more than .* seconds' | grep -v 'lockdep is turned off' | grep -v 'trying to register non-static key' | grep -v 'DEBUG: fsize' | grep -v CRON | grep -v 'BUG: bad unlock balance detected' | grep -v 'inconsistent lock state' | grep -v '*** DEADLOCK ***' | grep -v 'INFO: possible irq lock inversion dependency detected' | grep -v 'INFO: NMI handler (perf_event_nmi_handler) took too long to run' | grep -v 'INFO: recovery required on readonly' | grep -v 'ceph-create-keys: INFO' | grep -v INFO:ceph-create-keys | grep -v 'Loaded datasource DataSourceOpenStack' | grep -v 'container-storage-setup: INFO: Volume group backing root filesystem could not be determined' | grep -E -v '\bsalt-master\b|\bsalt-minion\b|\bsalt-api\b' | grep -v ceph-crash | grep -E -v '\btcmu-runner\b.*\bINFO\b' | head -n 1 2026-03-06T13:43:53.404 INFO:teuthology.task.internal.syslog:Gathering journactl... 2026-03-06T13:43:53.404 DEBUG:teuthology.orchestra.run.vm00:> sudo journalctl > /home/ubuntu/cephtest/archive/syslog/journalctl.log 2026-03-06T13:43:53.406 DEBUG:teuthology.orchestra.run.vm01:> sudo journalctl > /home/ubuntu/cephtest/archive/syslog/journalctl.log 2026-03-06T13:43:53.419 DEBUG:teuthology.orchestra.run.vm08:> sudo journalctl > /home/ubuntu/cephtest/archive/syslog/journalctl.log 2026-03-06T13:43:53.860 INFO:teuthology.task.internal.syslog:Compressing syslogs... 2026-03-06T13:43:53.861 DEBUG:teuthology.orchestra.run.vm00:> find /home/ubuntu/cephtest/archive/syslog -name '*.log' -print0 | sudo xargs -0 --max-args=1 --max-procs=0 --verbose --no-run-if-empty -- gzip -5 --verbose -- 2026-03-06T13:43:53.862 DEBUG:teuthology.orchestra.run.vm01:> find /home/ubuntu/cephtest/archive/syslog -name '*.log' -print0 | sudo xargs -0 --max-args=1 --max-procs=0 --verbose --no-run-if-empty -- gzip -5 --verbose -- 2026-03-06T13:43:53.864 DEBUG:teuthology.orchestra.run.vm08:> find /home/ubuntu/cephtest/archive/syslog -name '*.log' -print0 | sudo xargs -0 --max-args=1 --max-procs=0 --verbose --no-run-if-empty -- gzip -5 --verbose -- 2026-03-06T13:43:53.883 INFO:teuthology.orchestra.run.vm00.stderr:gzip -5 --verbose -- /home/ubuntu/cephtest/archive/syslog/kern.log 2026-03-06T13:43:53.884 INFO:teuthology.orchestra.run.vm00.stderr:gzip -5 --verbose -- /home/ubuntu/cephtest/archive/syslog/misc.log 2026-03-06T13:43:53.884 INFO:teuthology.orchestra.run.vm00.stderr:/home/ubuntu/cephtest/archive/syslog/kern.log: gzip 0.0% -- replaced with /home/ubuntu/cephtest/archive/syslog/kern.log.gz 2026-03-06T13:43:53.884 INFO:teuthology.orchestra.run.vm00.stderr: -5 --verbose -- /home/ubuntu/cephtest/archive/syslog/journalctl.log 2026-03-06T13:43:53.884 INFO:teuthology.orchestra.run.vm00.stderr:/home/ubuntu/cephtest/archive/syslog/misc.log: 0.0% -- replaced with /home/ubuntu/cephtest/archive/syslog/misc.log.gz 2026-03-06T13:43:53.888 INFO:teuthology.orchestra.run.vm08.stderr:gzip -5 --verbose -- /home/ubuntu/cephtest/archive/syslog/kern.log 2026-03-06T13:43:53.888 INFO:teuthology.orchestra.run.vm08.stderr:gzip -5 --verbose -- /home/ubuntu/cephtest/archive/syslog/misc.log 2026-03-06T13:43:53.888 INFO:teuthology.orchestra.run.vm08.stderr:/home/ubuntu/cephtest/archive/syslog/kern.log: 0.0% -- replaced with /home/ubuntu/cephtest/archive/syslog/kern.log.gz 2026-03-06T13:43:53.888 INFO:teuthology.orchestra.run.vm08.stderr:gzip -5 --verbose -- /home/ubuntu/cephtest/archive/syslog/journalctl.log 2026-03-06T13:43:53.889 INFO:teuthology.orchestra.run.vm08.stderr:/home/ubuntu/cephtest/archive/syslog/misc.log: /home/ubuntu/cephtest/archive/syslog/journalctl.log: 0.0% -- replaced with /home/ubuntu/cephtest/archive/syslog/misc.log.gz 2026-03-06T13:43:53.891 INFO:teuthology.orchestra.run.vm01.stderr:gzip -5 --verbose -- /home/ubuntu/cephtest/archive/syslog/kern.log 2026-03-06T13:43:53.892 INFO:teuthology.orchestra.run.vm01.stderr:gzip -5 --verbose -- /home/ubuntu/cephtest/archive/syslog/misc.log 2026-03-06T13:43:53.892 INFO:teuthology.orchestra.run.vm01.stderr:gzip -5 --verbose -- /home/ubuntu/cephtest/archive/syslog/journalctl.log 2026-03-06T13:43:53.892 INFO:teuthology.orchestra.run.vm01.stderr:/home/ubuntu/cephtest/archive/syslog/kern.log: 0.0% -- replaced with /home/ubuntu/cephtest/archive/syslog/kern.log.gz 2026-03-06T13:43:53.892 INFO:teuthology.orchestra.run.vm01.stderr:/home/ubuntu/cephtest/archive/syslog/misc.log: 0.0% -- replaced with /home/ubuntu/cephtest/archive/syslog/misc.log.gz 2026-03-06T13:43:54.009 INFO:teuthology.orchestra.run.vm08.stderr: 98.3% -- replaced with /home/ubuntu/cephtest/archive/syslog/journalctl.log.gz 2026-03-06T13:43:54.025 INFO:teuthology.orchestra.run.vm00.stderr:/home/ubuntu/cephtest/archive/syslog/journalctl.log: 98.2% -- replaced with /home/ubuntu/cephtest/archive/syslog/journalctl.log.gz 2026-03-06T13:43:54.047 INFO:teuthology.orchestra.run.vm01.stderr:/home/ubuntu/cephtest/archive/syslog/journalctl.log: 98.4% -- replaced with /home/ubuntu/cephtest/archive/syslog/journalctl.log.gz 2026-03-06T13:43:54.050 DEBUG:teuthology.run_tasks:Unwinding manager internal.sudo 2026-03-06T13:43:54.052 INFO:teuthology.task.internal:Restoring /etc/sudoers... 2026-03-06T13:43:54.052 DEBUG:teuthology.orchestra.run.vm00:> sudo mv -f /etc/sudoers.orig.teuthology /etc/sudoers 2026-03-06T13:43:54.095 DEBUG:teuthology.orchestra.run.vm01:> sudo mv -f /etc/sudoers.orig.teuthology /etc/sudoers 2026-03-06T13:43:54.125 DEBUG:teuthology.orchestra.run.vm08:> sudo mv -f /etc/sudoers.orig.teuthology /etc/sudoers 2026-03-06T13:43:54.154 DEBUG:teuthology.run_tasks:Unwinding manager internal.coredump 2026-03-06T13:43:54.157 DEBUG:teuthology.orchestra.run.vm00:> sudo sysctl -w kernel.core_pattern=core && sudo bash -c 'for f in `find /home/ubuntu/cephtest/archive/coredump -type f`; do file $f | grep -q systemd-sysusers && rm $f || true ; done' && rmdir --ignore-fail-on-non-empty -- /home/ubuntu/cephtest/archive/coredump 2026-03-06T13:43:54.159 DEBUG:teuthology.orchestra.run.vm01:> sudo sysctl -w kernel.core_pattern=core && sudo bash -c 'for f in `find /home/ubuntu/cephtest/archive/coredump -type f`; do file $f | grep -q systemd-sysusers && rm $f || true ; done' && rmdir --ignore-fail-on-non-empty -- /home/ubuntu/cephtest/archive/coredump 2026-03-06T13:43:54.167 DEBUG:teuthology.orchestra.run.vm08:> sudo sysctl -w kernel.core_pattern=core && sudo bash -c 'for f in `find /home/ubuntu/cephtest/archive/coredump -type f`; do file $f | grep -q systemd-sysusers && rm $f || true ; done' && rmdir --ignore-fail-on-non-empty -- /home/ubuntu/cephtest/archive/coredump 2026-03-06T13:43:54.184 INFO:teuthology.orchestra.run.vm00.stdout:kernel.core_pattern = core 2026-03-06T13:43:54.188 INFO:teuthology.orchestra.run.vm01.stdout:kernel.core_pattern = core 2026-03-06T13:43:54.226 INFO:teuthology.orchestra.run.vm08.stdout:kernel.core_pattern = core 2026-03-06T13:43:54.239 DEBUG:teuthology.orchestra.run.vm00:> test -e /home/ubuntu/cephtest/archive/coredump 2026-03-06T13:43:54.254 DEBUG:teuthology.orchestra.run:got remote process result: 1 2026-03-06T13:43:54.254 DEBUG:teuthology.orchestra.run.vm01:> test -e /home/ubuntu/cephtest/archive/coredump 2026-03-06T13:43:54.270 DEBUG:teuthology.orchestra.run:got remote process result: 1 2026-03-06T13:43:54.270 DEBUG:teuthology.orchestra.run.vm08:> test -e /home/ubuntu/cephtest/archive/coredump 2026-03-06T13:43:54.296 DEBUG:teuthology.orchestra.run:got remote process result: 1 2026-03-06T13:43:54.296 DEBUG:teuthology.run_tasks:Unwinding manager internal.archive 2026-03-06T13:43:54.298 INFO:teuthology.task.internal:Transferring archived files... 2026-03-06T13:43:54.299 DEBUG:teuthology.misc:Transferring archived files from vm00:/home/ubuntu/cephtest/archive to /archive/irq0-2026-03-06_13:20:18-orch:cephadm:workunits-cobaltcore-storage-v19.2.3-fasttrack-3-none-default-vps/273/remote/vm00 2026-03-06T13:43:54.299 DEBUG:teuthology.orchestra.run.vm00:> sudo tar c -f - -C /home/ubuntu/cephtest/archive -- . 2026-03-06T13:43:54.326 DEBUG:teuthology.misc:Transferring archived files from vm01:/home/ubuntu/cephtest/archive to /archive/irq0-2026-03-06_13:20:18-orch:cephadm:workunits-cobaltcore-storage-v19.2.3-fasttrack-3-none-default-vps/273/remote/vm01 2026-03-06T13:43:54.327 DEBUG:teuthology.orchestra.run.vm01:> sudo tar c -f - -C /home/ubuntu/cephtest/archive -- . 2026-03-06T13:43:54.357 DEBUG:teuthology.misc:Transferring archived files from vm08:/home/ubuntu/cephtest/archive to /archive/irq0-2026-03-06_13:20:18-orch:cephadm:workunits-cobaltcore-storage-v19.2.3-fasttrack-3-none-default-vps/273/remote/vm08 2026-03-06T13:43:54.357 DEBUG:teuthology.orchestra.run.vm08:> sudo tar c -f - -C /home/ubuntu/cephtest/archive -- . 2026-03-06T13:43:54.393 INFO:teuthology.task.internal:Removing archive directory... 2026-03-06T13:43:54.393 DEBUG:teuthology.orchestra.run.vm00:> rm -rf -- /home/ubuntu/cephtest/archive 2026-03-06T13:43:54.395 DEBUG:teuthology.orchestra.run.vm01:> rm -rf -- /home/ubuntu/cephtest/archive 2026-03-06T13:43:54.399 DEBUG:teuthology.orchestra.run.vm08:> rm -rf -- /home/ubuntu/cephtest/archive 2026-03-06T13:43:54.452 DEBUG:teuthology.run_tasks:Unwinding manager internal.archive_upload 2026-03-06T13:43:54.456 INFO:teuthology.task.internal:Not uploading archives. 2026-03-06T13:43:54.456 DEBUG:teuthology.run_tasks:Unwinding manager internal.base 2026-03-06T13:43:54.459 INFO:teuthology.task.internal:Tidying up after the test... 2026-03-06T13:43:54.459 DEBUG:teuthology.orchestra.run.vm00:> find /home/ubuntu/cephtest -ls ; rmdir -- /home/ubuntu/cephtest 2026-03-06T13:43:54.461 DEBUG:teuthology.orchestra.run.vm01:> find /home/ubuntu/cephtest -ls ; rmdir -- /home/ubuntu/cephtest 2026-03-06T13:43:54.463 DEBUG:teuthology.orchestra.run.vm08:> find /home/ubuntu/cephtest -ls ; rmdir -- /home/ubuntu/cephtest 2026-03-06T13:43:54.478 INFO:teuthology.orchestra.run.vm00.stdout: 8532139 0 drwxr-xr-x 2 ubuntu ubuntu 6 Mar 6 13:43 /home/ubuntu/cephtest 2026-03-06T13:43:54.480 INFO:teuthology.orchestra.run.vm01.stdout: 8532150 0 drwxr-xr-x 2 ubuntu ubuntu 6 Mar 6 13:43 /home/ubuntu/cephtest 2026-03-06T13:43:54.509 INFO:teuthology.orchestra.run.vm08.stdout: 8532144 0 drwxr-xr-x 2 ubuntu ubuntu 6 Mar 6 13:43 /home/ubuntu/cephtest 2026-03-06T13:43:54.510 DEBUG:teuthology.run_tasks:Unwinding manager console_log 2026-03-06T13:43:54.516 INFO:teuthology.run:Summary data: description: orch:cephadm:workunits/{0-distro/centos_9.stream agent/off mon_election/classic task/test_rgw_multisite} duration: 462.66978001594543 flavor: default owner: irq0 success: true 2026-03-06T13:43:54.516 DEBUG:teuthology.report:Pushing job info to http://localhost:8080 2026-03-06T13:43:54.539 INFO:teuthology.run:pass