2026-03-06T13:40:26.754 INFO:root:teuthology version: 1.2.4.dev6+g1c580df7a 2026-03-06T13:40:26.759 DEBUG:teuthology.report:Pushing job info to http://localhost:8080 2026-03-06T13:40:26.778 INFO:teuthology.run:Config: archive_path: /archive/irq0-2026-03-06_13:20:18-orch:cephadm:workunits-cobaltcore-storage-v19.2.3-fasttrack-3-none-default-vps/274 branch: cobaltcore-storage-v19.2.3-fasttrack-3 description: orch:cephadm:workunits/{0-distro/centos_9.stream_runc agent/on mon_election/connectivity task/test_set_mon_crush_locations} email: null first_in_suite: false flavor: default job_id: '274' last_in_suite: false machine_type: vps name: irq0-2026-03-06_13:20:18-orch:cephadm:workunits-cobaltcore-storage-v19.2.3-fasttrack-3-none-default-vps no_nested_subset: false os_type: centos os_version: 9.stream overrides: admin_socket: branch: cobaltcore-storage-v19.2.3-fasttrack-3 ansible.cephlab: branch: main repo: https://github.com/kshtsk/ceph-cm-ansible.git skip_tags: nagios,monitoring-scripts,hostname,pubkeys,zap,sudoers,kerberos,ntp-client,resolvconf,cpan,nfs vars: timezone: Europe/Berlin ceph: conf: global: mon election default strategy: 3 mgr: debug mgr: 20 debug ms: 1 mgr/cephadm/use_agent: true mon: debug mon: 20 debug ms: 1 debug paxos: 20 osd: debug ms: 1 debug osd: 20 osd mclock iops capacity threshold hdd: 49000 flavor: default log-ignorelist: - \(MDS_ALL_DOWN\) - \(MDS_UP_LESS_THAN_MAX\) - MON_DOWN - POOL_APP_NOT_ENABLED - mon down - mons down - out of quorum - CEPHADM_FAILED_DAEMON log-only-match: - CEPHADM_ sha1: c24117fd5525679b799527bc1bd1f1dd0a2db5e2 ceph-deploy: conf: client: log file: /var/log/ceph/ceph-$name.$pid.log mon: {} cephadm: cephadm_binary_url: https://download.ceph.com/rpm-19.2.3/el9/noarch/cephadm containers: image: harbor.clyso.com/custom-ceph/ceph/ceph:cobaltcore-storage-v19.2.3-fasttrack-3 install: ceph: flavor: default sha1: c24117fd5525679b799527bc1bd1f1dd0a2db5e2 extra_system_packages: deb: - python3-xmltodict - s3cmd rpm: - bzip2 - perl-Test-Harness - python3-xmltodict - s3cmd repos: - name: ceph-source priority: 1 url: https://s3.clyso.com/ces-packages/components/ceph/rpm-19.2.3-47-gc24117fd552/el9.clyso/SRPMS - name: ceph-noarch priority: 1 url: https://s3.clyso.com/ces-packages/components/ceph/rpm-19.2.3-47-gc24117fd552/el9.clyso/noarch - name: ceph priority: 1 url: https://s3.clyso.com/ces-packages/components/ceph/rpm-19.2.3-47-gc24117fd552/el9.clyso/x86_64 selinux: allowlist: - scontext=system_u:system_r:logrotate_t:s0 workunit: branch: tt-19.2.3-fasttrack-3-no-nvme-loop sha1: 5726a36c3452e5b72190cfceba828abc62c819b7 owner: irq0 priority: 1000 repo: https://github.com/ceph/ceph.git roles: - - host.a - osd.0 - mon.a - mgr.a - - host.b - osd.1 - mon.b - mgr.b - - host.c - osd.2 - mon.c seed: 6609 sha1: c24117fd5525679b799527bc1bd1f1dd0a2db5e2 sleep_before_teardown: 0 subset: 1/64 suite: orch:cephadm:workunits suite_branch: tt-19.2.3-fasttrack-3-no-nvme-loop suite_path: /home/teuthos/src/github.com_kshtsk_ceph_5726a36c3452e5b72190cfceba828abc62c819b7/qa suite_relpath: qa suite_repo: https://github.com/kshtsk/ceph.git suite_sha1: 5726a36c3452e5b72190cfceba828abc62c819b7 targets: vm02.local: ecdsa-sha2-nistp256 AAAAE2VjZHNhLXNoYTItbmlzdHAyNTYAAAAIbmlzdHAyNTYAAABBBLsOHiFlGFjAAfhRhA6Bz/Z5pWZ2J4sLC6KnWMM+09lOXylxa0+JGwargIdI7AF1/lF802u10EQR8zedr3bqNLA= vm04.local: ecdsa-sha2-nistp256 AAAAE2VjZHNhLXNoYTItbmlzdHAyNTYAAAAIbmlzdHAyNTYAAABBBJ+Hxit9X+qnF4d1oBi7hr5O3CtelUX2YUsvNZfVQeRW2CFA/ZVmxbJpGlaOKcEePh4iGoqDu5I6h9wQCNVSHes= vm06.local: ecdsa-sha2-nistp256 AAAAE2VjZHNhLXNoYTItbmlzdHAyNTYAAAAIbmlzdHAyNTYAAABBBG6dqd5XlqISzyVyR/vlR0sZ8bhIxgkzgmJ6QgnftHP9qECkRTzYUbnq1+LNVPULsU+f4gByp/dzBlkdq5d8Vyg= tasks: - pexec: all: - sudo dnf remove nvme-cli -y - sudo dnf install runc nvmetcli nvme-cli -y - sudo sed -i 's/^#runtime = "crun"/runtime = "runc"/g' /usr/share/containers/containers.conf - sudo sed -i 's/runtime = "crun"/#runtime = "crun"/g' /usr/share/containers/containers.conf - install: null - cephadm: null - cephadm.apply: specs: - placement: count: 3 service_id: foo service_type: mon spec: crush_locations: host.a: - datacenter=a host.b: - datacenter=b - rack=2 host.c: - datacenter=a - rack=3 - cephadm.shell: host.a: - "set -ex\n# since we don't know the real hostnames before the test, the next\n\ # bit is in order to replace the fake hostnames \"host.a/b/c\" with\n# the actual\ \ names cephadm knows the host by within the mon spec\nceph orch host ls --format\ \ json | jq -r '.[] | .hostname' > realnames\necho $'host.a\\nhost.b\\nhost.c'\ \ > fakenames\necho $'a\\nb\\nc' > mon_ids\necho $'{datacenter=a}\\n{datacenter=b,rack=2}\\\ n{datacenter=a,rack=3}' > crush_locs\nceph orch ls --service-name mon --export\ \ > mon.yaml\nMONSPEC=`cat mon.yaml`\necho \"$MONSPEC\"\nwhile read realname\ \ <&3 && read fakename <&4; do\n MONSPEC=\"${MONSPEC//$fakename/$realname}\"\ \ndone 3 mon.yaml\ncat mon.yaml\n\ # now the spec should have the real hostnames, so let's re-apply\nceph orch\ \ apply -i mon.yaml\nsleep 90\nceph orch ps --refresh\nceph orch ls --service-name\ \ mon --export > mon.yaml; ceph orch apply -i mon.yaml\nsleep 90\nceph mon dump\n\ ceph mon dump --format json\n# verify all the crush locations got set from \"\ ceph mon dump\" output\nwhile read monid <&3 && read crushloc <&4; do\n ceph\ \ mon dump --format json | jq --arg monid \"$monid\" --arg crushloc \"$crushloc\"\ \ -e '.mons | .[] | select(.name == $monid) | .crush_location == $crushloc'\n\ done 3, func=.kill_console_loggers at 0x7f8ebd0a3be0>, signals=[15]) 2026-03-06T13:40:26.815 INFO:teuthology.run_tasks:Running task internal.connect... 2026-03-06T13:40:26.815 INFO:teuthology.task.internal:Opening connections... 2026-03-06T13:40:26.815 DEBUG:teuthology.task.internal:connecting to ubuntu@vm02.local 2026-03-06T13:40:26.816 DEBUG:teuthology.orchestra.connection:{'hostname': 'vm02.local', 'username': 'ubuntu', 'timeout': 60} 2026-03-06T13:40:26.875 DEBUG:teuthology.task.internal:connecting to ubuntu@vm04.local 2026-03-06T13:40:26.876 DEBUG:teuthology.orchestra.connection:{'hostname': 'vm04.local', 'username': 'ubuntu', 'timeout': 60} 2026-03-06T13:40:26.935 DEBUG:teuthology.task.internal:connecting to ubuntu@vm06.local 2026-03-06T13:40:26.935 DEBUG:teuthology.orchestra.connection:{'hostname': 'vm06.local', 'username': 'ubuntu', 'timeout': 60} 2026-03-06T13:40:26.995 INFO:teuthology.run_tasks:Running task internal.push_inventory... 2026-03-06T13:40:26.996 DEBUG:teuthology.orchestra.run.vm02:> uname -m 2026-03-06T13:40:27.011 INFO:teuthology.orchestra.run.vm02.stdout:x86_64 2026-03-06T13:40:27.011 DEBUG:teuthology.orchestra.run.vm02:> cat /etc/os-release 2026-03-06T13:40:27.067 INFO:teuthology.orchestra.run.vm02.stdout:NAME="CentOS Stream" 2026-03-06T13:40:27.067 INFO:teuthology.orchestra.run.vm02.stdout:VERSION="9" 2026-03-06T13:40:27.067 INFO:teuthology.orchestra.run.vm02.stdout:ID="centos" 2026-03-06T13:40:27.067 INFO:teuthology.orchestra.run.vm02.stdout:ID_LIKE="rhel fedora" 2026-03-06T13:40:27.067 INFO:teuthology.orchestra.run.vm02.stdout:VERSION_ID="9" 2026-03-06T13:40:27.067 INFO:teuthology.orchestra.run.vm02.stdout:PLATFORM_ID="platform:el9" 2026-03-06T13:40:27.067 INFO:teuthology.orchestra.run.vm02.stdout:PRETTY_NAME="CentOS Stream 9" 2026-03-06T13:40:27.067 INFO:teuthology.orchestra.run.vm02.stdout:ANSI_COLOR="0;31" 2026-03-06T13:40:27.067 INFO:teuthology.orchestra.run.vm02.stdout:LOGO="fedora-logo-icon" 2026-03-06T13:40:27.067 INFO:teuthology.orchestra.run.vm02.stdout:CPE_NAME="cpe:/o:centos:centos:9" 2026-03-06T13:40:27.067 INFO:teuthology.orchestra.run.vm02.stdout:HOME_URL="https://centos.org/" 2026-03-06T13:40:27.067 INFO:teuthology.orchestra.run.vm02.stdout:BUG_REPORT_URL="https://issues.redhat.com/" 2026-03-06T13:40:27.067 INFO:teuthology.orchestra.run.vm02.stdout:REDHAT_SUPPORT_PRODUCT="Red Hat Enterprise Linux 9" 2026-03-06T13:40:27.067 INFO:teuthology.orchestra.run.vm02.stdout:REDHAT_SUPPORT_PRODUCT_VERSION="CentOS Stream" 2026-03-06T13:40:27.068 INFO:teuthology.lock.ops:Updating vm02.local on lock server 2026-03-06T13:40:27.072 DEBUG:teuthology.orchestra.run.vm04:> uname -m 2026-03-06T13:40:27.086 INFO:teuthology.orchestra.run.vm04.stdout:x86_64 2026-03-06T13:40:27.087 DEBUG:teuthology.orchestra.run.vm04:> cat /etc/os-release 2026-03-06T13:40:27.141 INFO:teuthology.orchestra.run.vm04.stdout:NAME="CentOS Stream" 2026-03-06T13:40:27.141 INFO:teuthology.orchestra.run.vm04.stdout:VERSION="9" 2026-03-06T13:40:27.141 INFO:teuthology.orchestra.run.vm04.stdout:ID="centos" 2026-03-06T13:40:27.141 INFO:teuthology.orchestra.run.vm04.stdout:ID_LIKE="rhel fedora" 2026-03-06T13:40:27.141 INFO:teuthology.orchestra.run.vm04.stdout:VERSION_ID="9" 2026-03-06T13:40:27.141 INFO:teuthology.orchestra.run.vm04.stdout:PLATFORM_ID="platform:el9" 2026-03-06T13:40:27.141 INFO:teuthology.orchestra.run.vm04.stdout:PRETTY_NAME="CentOS Stream 9" 2026-03-06T13:40:27.141 INFO:teuthology.orchestra.run.vm04.stdout:ANSI_COLOR="0;31" 2026-03-06T13:40:27.141 INFO:teuthology.orchestra.run.vm04.stdout:LOGO="fedora-logo-icon" 2026-03-06T13:40:27.141 INFO:teuthology.orchestra.run.vm04.stdout:CPE_NAME="cpe:/o:centos:centos:9" 2026-03-06T13:40:27.141 INFO:teuthology.orchestra.run.vm04.stdout:HOME_URL="https://centos.org/" 2026-03-06T13:40:27.141 INFO:teuthology.orchestra.run.vm04.stdout:BUG_REPORT_URL="https://issues.redhat.com/" 2026-03-06T13:40:27.141 INFO:teuthology.orchestra.run.vm04.stdout:REDHAT_SUPPORT_PRODUCT="Red Hat Enterprise Linux 9" 2026-03-06T13:40:27.141 INFO:teuthology.orchestra.run.vm04.stdout:REDHAT_SUPPORT_PRODUCT_VERSION="CentOS Stream" 2026-03-06T13:40:27.141 INFO:teuthology.lock.ops:Updating vm04.local on lock server 2026-03-06T13:40:27.146 DEBUG:teuthology.orchestra.run.vm06:> uname -m 2026-03-06T13:40:27.163 INFO:teuthology.orchestra.run.vm06.stdout:x86_64 2026-03-06T13:40:27.163 DEBUG:teuthology.orchestra.run.vm06:> cat /etc/os-release 2026-03-06T13:40:27.219 INFO:teuthology.orchestra.run.vm06.stdout:NAME="CentOS Stream" 2026-03-06T13:40:27.219 INFO:teuthology.orchestra.run.vm06.stdout:VERSION="9" 2026-03-06T13:40:27.219 INFO:teuthology.orchestra.run.vm06.stdout:ID="centos" 2026-03-06T13:40:27.219 INFO:teuthology.orchestra.run.vm06.stdout:ID_LIKE="rhel fedora" 2026-03-06T13:40:27.219 INFO:teuthology.orchestra.run.vm06.stdout:VERSION_ID="9" 2026-03-06T13:40:27.219 INFO:teuthology.orchestra.run.vm06.stdout:PLATFORM_ID="platform:el9" 2026-03-06T13:40:27.219 INFO:teuthology.orchestra.run.vm06.stdout:PRETTY_NAME="CentOS Stream 9" 2026-03-06T13:40:27.219 INFO:teuthology.orchestra.run.vm06.stdout:ANSI_COLOR="0;31" 2026-03-06T13:40:27.219 INFO:teuthology.orchestra.run.vm06.stdout:LOGO="fedora-logo-icon" 2026-03-06T13:40:27.219 INFO:teuthology.orchestra.run.vm06.stdout:CPE_NAME="cpe:/o:centos:centos:9" 2026-03-06T13:40:27.219 INFO:teuthology.orchestra.run.vm06.stdout:HOME_URL="https://centos.org/" 2026-03-06T13:40:27.219 INFO:teuthology.orchestra.run.vm06.stdout:BUG_REPORT_URL="https://issues.redhat.com/" 2026-03-06T13:40:27.219 INFO:teuthology.orchestra.run.vm06.stdout:REDHAT_SUPPORT_PRODUCT="Red Hat Enterprise Linux 9" 2026-03-06T13:40:27.219 INFO:teuthology.orchestra.run.vm06.stdout:REDHAT_SUPPORT_PRODUCT_VERSION="CentOS Stream" 2026-03-06T13:40:27.219 INFO:teuthology.lock.ops:Updating vm06.local on lock server 2026-03-06T13:40:27.224 INFO:teuthology.run_tasks:Running task internal.serialize_remote_roles... 2026-03-06T13:40:27.225 INFO:teuthology.run_tasks:Running task internal.check_conflict... 2026-03-06T13:40:27.226 INFO:teuthology.task.internal:Checking for old test directory... 2026-03-06T13:40:27.226 DEBUG:teuthology.orchestra.run.vm02:> test '!' -e /home/ubuntu/cephtest 2026-03-06T13:40:27.228 DEBUG:teuthology.orchestra.run.vm04:> test '!' -e /home/ubuntu/cephtest 2026-03-06T13:40:27.230 DEBUG:teuthology.orchestra.run.vm06:> test '!' -e /home/ubuntu/cephtest 2026-03-06T13:40:27.274 INFO:teuthology.run_tasks:Running task internal.check_ceph_data... 2026-03-06T13:40:27.275 INFO:teuthology.task.internal:Checking for non-empty /var/lib/ceph... 2026-03-06T13:40:27.275 DEBUG:teuthology.orchestra.run.vm02:> test -z $(ls -A /var/lib/ceph) 2026-03-06T13:40:27.283 DEBUG:teuthology.orchestra.run.vm04:> test -z $(ls -A /var/lib/ceph) 2026-03-06T13:40:27.285 DEBUG:teuthology.orchestra.run.vm06:> test -z $(ls -A /var/lib/ceph) 2026-03-06T13:40:27.296 INFO:teuthology.orchestra.run.vm02.stderr:ls: cannot access '/var/lib/ceph': No such file or directory 2026-03-06T13:40:27.298 INFO:teuthology.orchestra.run.vm04.stderr:ls: cannot access '/var/lib/ceph': No such file or directory 2026-03-06T13:40:27.329 INFO:teuthology.orchestra.run.vm06.stderr:ls: cannot access '/var/lib/ceph': No such file or directory 2026-03-06T13:40:27.330 INFO:teuthology.run_tasks:Running task internal.vm_setup... 2026-03-06T13:40:27.337 DEBUG:teuthology.orchestra.run.vm02:> test -e /ceph-qa-ready 2026-03-06T13:40:27.350 DEBUG:teuthology.orchestra.run:got remote process result: 1 2026-03-06T13:40:27.536 DEBUG:teuthology.orchestra.run.vm04:> test -e /ceph-qa-ready 2026-03-06T13:40:27.550 DEBUG:teuthology.orchestra.run:got remote process result: 1 2026-03-06T13:40:27.737 DEBUG:teuthology.orchestra.run.vm06:> test -e /ceph-qa-ready 2026-03-06T13:40:27.751 DEBUG:teuthology.orchestra.run:got remote process result: 1 2026-03-06T13:40:27.956 INFO:teuthology.run_tasks:Running task internal.base... 2026-03-06T13:40:27.957 INFO:teuthology.task.internal:Creating test directory... 2026-03-06T13:40:27.958 DEBUG:teuthology.orchestra.run.vm02:> mkdir -p -m0755 -- /home/ubuntu/cephtest 2026-03-06T13:40:27.959 DEBUG:teuthology.orchestra.run.vm04:> mkdir -p -m0755 -- /home/ubuntu/cephtest 2026-03-06T13:40:27.961 DEBUG:teuthology.orchestra.run.vm06:> mkdir -p -m0755 -- /home/ubuntu/cephtest 2026-03-06T13:40:27.979 INFO:teuthology.run_tasks:Running task internal.archive_upload... 2026-03-06T13:40:27.980 INFO:teuthology.run_tasks:Running task internal.archive... 2026-03-06T13:40:27.981 INFO:teuthology.task.internal:Creating archive directory... 2026-03-06T13:40:27.981 DEBUG:teuthology.orchestra.run.vm02:> install -d -m0755 -- /home/ubuntu/cephtest/archive 2026-03-06T13:40:28.016 DEBUG:teuthology.orchestra.run.vm04:> install -d -m0755 -- /home/ubuntu/cephtest/archive 2026-03-06T13:40:28.021 DEBUG:teuthology.orchestra.run.vm06:> install -d -m0755 -- /home/ubuntu/cephtest/archive 2026-03-06T13:40:28.041 INFO:teuthology.run_tasks:Running task internal.coredump... 2026-03-06T13:40:28.042 INFO:teuthology.task.internal:Enabling coredump saving... 2026-03-06T13:40:28.042 DEBUG:teuthology.orchestra.run.vm02:> test -f /run/.containerenv -o -f /.dockerenv 2026-03-06T13:40:28.087 DEBUG:teuthology.orchestra.run:got remote process result: 1 2026-03-06T13:40:28.087 DEBUG:teuthology.orchestra.run.vm04:> test -f /run/.containerenv -o -f /.dockerenv 2026-03-06T13:40:28.102 DEBUG:teuthology.orchestra.run:got remote process result: 1 2026-03-06T13:40:28.102 DEBUG:teuthology.orchestra.run.vm06:> test -f /run/.containerenv -o -f /.dockerenv 2026-03-06T13:40:28.116 DEBUG:teuthology.orchestra.run:got remote process result: 1 2026-03-06T13:40:28.116 DEBUG:teuthology.orchestra.run.vm02:> install -d -m0755 -- /home/ubuntu/cephtest/archive/coredump && sudo sysctl -w kernel.core_pattern=/home/ubuntu/cephtest/archive/coredump/%t.%p.core && echo kernel.core_pattern=/home/ubuntu/cephtest/archive/coredump/%t.%p.core | sudo tee -a /etc/sysctl.conf 2026-03-06T13:40:28.129 DEBUG:teuthology.orchestra.run.vm04:> install -d -m0755 -- /home/ubuntu/cephtest/archive/coredump && sudo sysctl -w kernel.core_pattern=/home/ubuntu/cephtest/archive/coredump/%t.%p.core && echo kernel.core_pattern=/home/ubuntu/cephtest/archive/coredump/%t.%p.core | sudo tee -a /etc/sysctl.conf 2026-03-06T13:40:28.144 DEBUG:teuthology.orchestra.run.vm06:> install -d -m0755 -- /home/ubuntu/cephtest/archive/coredump && sudo sysctl -w kernel.core_pattern=/home/ubuntu/cephtest/archive/coredump/%t.%p.core && echo kernel.core_pattern=/home/ubuntu/cephtest/archive/coredump/%t.%p.core | sudo tee -a /etc/sysctl.conf 2026-03-06T13:40:28.150 INFO:teuthology.orchestra.run.vm02.stdout:kernel.core_pattern = /home/ubuntu/cephtest/archive/coredump/%t.%p.core 2026-03-06T13:40:28.159 INFO:teuthology.orchestra.run.vm02.stdout:kernel.core_pattern=/home/ubuntu/cephtest/archive/coredump/%t.%p.core 2026-03-06T13:40:28.167 INFO:teuthology.orchestra.run.vm04.stdout:kernel.core_pattern = /home/ubuntu/cephtest/archive/coredump/%t.%p.core 2026-03-06T13:40:28.176 INFO:teuthology.orchestra.run.vm04.stdout:kernel.core_pattern=/home/ubuntu/cephtest/archive/coredump/%t.%p.core 2026-03-06T13:40:28.185 INFO:teuthology.orchestra.run.vm06.stdout:kernel.core_pattern = /home/ubuntu/cephtest/archive/coredump/%t.%p.core 2026-03-06T13:40:28.195 INFO:teuthology.orchestra.run.vm06.stdout:kernel.core_pattern=/home/ubuntu/cephtest/archive/coredump/%t.%p.core 2026-03-06T13:40:28.196 INFO:teuthology.run_tasks:Running task internal.sudo... 2026-03-06T13:40:28.198 INFO:teuthology.task.internal:Configuring sudo... 2026-03-06T13:40:28.198 DEBUG:teuthology.orchestra.run.vm02:> sudo sed -i.orig.teuthology -e 's/^\([^#]*\) \(requiretty\)/\1 !\2/g' -e 's/^\([^#]*\) !\(visiblepw\)/\1 \2/g' /etc/sudoers 2026-03-06T13:40:28.204 DEBUG:teuthology.orchestra.run.vm04:> sudo sed -i.orig.teuthology -e 's/^\([^#]*\) \(requiretty\)/\1 !\2/g' -e 's/^\([^#]*\) !\(visiblepw\)/\1 \2/g' /etc/sudoers 2026-03-06T13:40:28.220 DEBUG:teuthology.orchestra.run.vm06:> sudo sed -i.orig.teuthology -e 's/^\([^#]*\) \(requiretty\)/\1 !\2/g' -e 's/^\([^#]*\) !\(visiblepw\)/\1 \2/g' /etc/sudoers 2026-03-06T13:40:28.262 INFO:teuthology.run_tasks:Running task internal.syslog... 2026-03-06T13:40:28.264 INFO:teuthology.task.internal.syslog:Starting syslog monitoring... 2026-03-06T13:40:28.264 DEBUG:teuthology.orchestra.run.vm02:> mkdir -p -m0755 -- /home/ubuntu/cephtest/archive/syslog 2026-03-06T13:40:28.271 DEBUG:teuthology.orchestra.run.vm04:> mkdir -p -m0755 -- /home/ubuntu/cephtest/archive/syslog 2026-03-06T13:40:28.284 DEBUG:teuthology.orchestra.run.vm06:> mkdir -p -m0755 -- /home/ubuntu/cephtest/archive/syslog 2026-03-06T13:40:28.319 DEBUG:teuthology.orchestra.run.vm02:> install -m 666 /dev/null /home/ubuntu/cephtest/archive/syslog/kern.log 2026-03-06T13:40:28.348 DEBUG:teuthology.orchestra.run.vm02:> install -m 666 /dev/null /home/ubuntu/cephtest/archive/syslog/misc.log 2026-03-06T13:40:28.404 DEBUG:teuthology.orchestra.run.vm02:> set -ex 2026-03-06T13:40:28.404 DEBUG:teuthology.orchestra.run.vm02:> sudo dd of=/etc/rsyslog.d/80-cephtest.conf 2026-03-06T13:40:28.461 DEBUG:teuthology.orchestra.run.vm04:> install -m 666 /dev/null /home/ubuntu/cephtest/archive/syslog/kern.log 2026-03-06T13:40:28.482 DEBUG:teuthology.orchestra.run.vm04:> install -m 666 /dev/null /home/ubuntu/cephtest/archive/syslog/misc.log 2026-03-06T13:40:28.539 DEBUG:teuthology.orchestra.run.vm04:> set -ex 2026-03-06T13:40:28.539 DEBUG:teuthology.orchestra.run.vm04:> sudo dd of=/etc/rsyslog.d/80-cephtest.conf 2026-03-06T13:40:28.597 DEBUG:teuthology.orchestra.run.vm06:> install -m 666 /dev/null /home/ubuntu/cephtest/archive/syslog/kern.log 2026-03-06T13:40:28.620 DEBUG:teuthology.orchestra.run.vm06:> install -m 666 /dev/null /home/ubuntu/cephtest/archive/syslog/misc.log 2026-03-06T13:40:28.679 DEBUG:teuthology.orchestra.run.vm06:> set -ex 2026-03-06T13:40:28.679 DEBUG:teuthology.orchestra.run.vm06:> sudo dd of=/etc/rsyslog.d/80-cephtest.conf 2026-03-06T13:40:28.738 DEBUG:teuthology.orchestra.run.vm02:> sudo service rsyslog restart 2026-03-06T13:40:28.739 DEBUG:teuthology.orchestra.run.vm04:> sudo service rsyslog restart 2026-03-06T13:40:28.741 DEBUG:teuthology.orchestra.run.vm06:> sudo service rsyslog restart 2026-03-06T13:40:28.767 INFO:teuthology.orchestra.run.vm04.stderr:Redirecting to /bin/systemctl restart rsyslog.service 2026-03-06T13:40:28.768 INFO:teuthology.orchestra.run.vm02.stderr:Redirecting to /bin/systemctl restart rsyslog.service 2026-03-06T13:40:28.805 INFO:teuthology.orchestra.run.vm06.stderr:Redirecting to /bin/systemctl restart rsyslog.service 2026-03-06T13:40:29.134 INFO:teuthology.run_tasks:Running task internal.timer... 2026-03-06T13:40:29.136 INFO:teuthology.task.internal:Starting timer... 2026-03-06T13:40:29.136 INFO:teuthology.run_tasks:Running task pcp... 2026-03-06T13:40:29.138 INFO:teuthology.run_tasks:Running task selinux... 2026-03-06T13:40:29.141 DEBUG:teuthology.task:Applying overrides for task selinux: {'allowlist': ['scontext=system_u:system_r:logrotate_t:s0']} 2026-03-06T13:40:29.141 INFO:teuthology.task.selinux:Excluding vm02: VMs are not yet supported 2026-03-06T13:40:29.141 INFO:teuthology.task.selinux:Excluding vm04: VMs are not yet supported 2026-03-06T13:40:29.141 INFO:teuthology.task.selinux:Excluding vm06: VMs are not yet supported 2026-03-06T13:40:29.141 DEBUG:teuthology.task.selinux:Getting current SELinux state 2026-03-06T13:40:29.141 DEBUG:teuthology.task.selinux:Existing SELinux modes: {} 2026-03-06T13:40:29.141 INFO:teuthology.task.selinux:Putting SELinux into permissive mode 2026-03-06T13:40:29.141 INFO:teuthology.run_tasks:Running task ansible.cephlab... 2026-03-06T13:40:29.143 DEBUG:teuthology.task:Applying overrides for task ansible.cephlab: {'branch': 'main', 'repo': 'https://github.com/kshtsk/ceph-cm-ansible.git', 'skip_tags': 'nagios,monitoring-scripts,hostname,pubkeys,zap,sudoers,kerberos,ntp-client,resolvconf,cpan,nfs', 'vars': {'timezone': 'Europe/Berlin'}} 2026-03-06T13:40:29.143 DEBUG:teuthology.repo_utils:Setting repo remote to https://github.com/kshtsk/ceph-cm-ansible.git 2026-03-06T13:40:29.144 INFO:teuthology.repo_utils:Fetching github.com_kshtsk_ceph-cm-ansible_main from origin 2026-03-06T13:40:29.610 DEBUG:teuthology.repo_utils:Resetting repo at /home/teuthos/src/github.com_kshtsk_ceph-cm-ansible_main to origin/main 2026-03-06T13:40:29.618 INFO:teuthology.task.ansible:Playbook: [{'import_playbook': 'ansible_managed.yml'}, {'import_playbook': 'teuthology.yml'}, {'hosts': 'testnodes', 'tasks': [{'set_fact': {'ran_from_cephlab_playbook': True}}]}, {'import_playbook': 'testnodes.yml'}, {'import_playbook': 'container-host.yml'}, {'import_playbook': 'cobbler.yml'}, {'import_playbook': 'paddles.yml'}, {'import_playbook': 'pulpito.yml'}, {'hosts': 'testnodes', 'become': True, 'tasks': [{'name': 'Touch /ceph-qa-ready', 'file': {'path': '/ceph-qa-ready', 'state': 'touch'}, 'when': 'ran_from_cephlab_playbook|bool'}]}] 2026-03-06T13:40:29.618 DEBUG:teuthology.task.ansible:Running ansible-playbook -v --extra-vars '{"ansible_ssh_user": "ubuntu", "timezone": "Europe/Berlin"}' -i /tmp/teuth_ansible_inventory11mu4i6r --limit vm02.local,vm04.local,vm06.local /home/teuthos/src/github.com_kshtsk_ceph-cm-ansible_main/cephlab.yml --skip-tags nagios,monitoring-scripts,hostname,pubkeys,zap,sudoers,kerberos,ntp-client,resolvconf,cpan,nfs 2026-03-06T13:43:06.424 DEBUG:teuthology.task.ansible:Reconnecting to [Remote(name='ubuntu@vm02.local'), Remote(name='ubuntu@vm04.local'), Remote(name='ubuntu@vm06.local')] 2026-03-06T13:43:06.424 INFO:teuthology.orchestra.remote:Trying to reconnect to host 'ubuntu@vm02.local' 2026-03-06T13:43:06.425 DEBUG:teuthology.orchestra.connection:{'hostname': 'vm02.local', 'username': 'ubuntu', 'timeout': 60} 2026-03-06T13:43:06.485 DEBUG:teuthology.orchestra.run.vm02:> true 2026-03-06T13:43:06.558 INFO:teuthology.orchestra.remote:Successfully reconnected to host 'ubuntu@vm02.local' 2026-03-06T13:43:06.558 INFO:teuthology.orchestra.remote:Trying to reconnect to host 'ubuntu@vm04.local' 2026-03-06T13:43:06.558 DEBUG:teuthology.orchestra.connection:{'hostname': 'vm04.local', 'username': 'ubuntu', 'timeout': 60} 2026-03-06T13:43:06.622 DEBUG:teuthology.orchestra.run.vm04:> true 2026-03-06T13:43:06.700 INFO:teuthology.orchestra.remote:Successfully reconnected to host 'ubuntu@vm04.local' 2026-03-06T13:43:06.700 INFO:teuthology.orchestra.remote:Trying to reconnect to host 'ubuntu@vm06.local' 2026-03-06T13:43:06.700 DEBUG:teuthology.orchestra.connection:{'hostname': 'vm06.local', 'username': 'ubuntu', 'timeout': 60} 2026-03-06T13:43:06.764 DEBUG:teuthology.orchestra.run.vm06:> true 2026-03-06T13:43:06.839 INFO:teuthology.orchestra.remote:Successfully reconnected to host 'ubuntu@vm06.local' 2026-03-06T13:43:06.839 INFO:teuthology.run_tasks:Running task clock... 2026-03-06T13:43:06.850 INFO:teuthology.task.clock:Syncing clocks and checking initial clock skew... 2026-03-06T13:43:06.850 INFO:teuthology.orchestra.run:Running command with timeout 360 2026-03-06T13:43:06.850 DEBUG:teuthology.orchestra.run.vm02:> sudo systemctl stop ntp.service || sudo systemctl stop ntpd.service || sudo systemctl stop chronyd.service ; sudo ntpd -gq || sudo chronyc makestep ; sudo systemctl start ntp.service || sudo systemctl start ntpd.service || sudo systemctl start chronyd.service ; PATH=/usr/bin:/usr/sbin ntpq -p || PATH=/usr/bin:/usr/sbin chronyc sources || true 2026-03-06T13:43:06.852 INFO:teuthology.orchestra.run:Running command with timeout 360 2026-03-06T13:43:06.852 DEBUG:teuthology.orchestra.run.vm04:> sudo systemctl stop ntp.service || sudo systemctl stop ntpd.service || sudo systemctl stop chronyd.service ; sudo ntpd -gq || sudo chronyc makestep ; sudo systemctl start ntp.service || sudo systemctl start ntpd.service || sudo systemctl start chronyd.service ; PATH=/usr/bin:/usr/sbin ntpq -p || PATH=/usr/bin:/usr/sbin chronyc sources || true 2026-03-06T13:43:06.854 INFO:teuthology.orchestra.run:Running command with timeout 360 2026-03-06T13:43:06.854 DEBUG:teuthology.orchestra.run.vm06:> sudo systemctl stop ntp.service || sudo systemctl stop ntpd.service || sudo systemctl stop chronyd.service ; sudo ntpd -gq || sudo chronyc makestep ; sudo systemctl start ntp.service || sudo systemctl start ntpd.service || sudo systemctl start chronyd.service ; PATH=/usr/bin:/usr/sbin ntpq -p || PATH=/usr/bin:/usr/sbin chronyc sources || true 2026-03-06T13:43:06.882 INFO:teuthology.orchestra.run.vm02.stderr:Failed to stop ntp.service: Unit ntp.service not loaded. 2026-03-06T13:43:06.897 INFO:teuthology.orchestra.run.vm04.stderr:Failed to stop ntp.service: Unit ntp.service not loaded. 2026-03-06T13:43:06.898 INFO:teuthology.orchestra.run.vm02.stderr:Failed to stop ntpd.service: Unit ntpd.service not loaded. 2026-03-06T13:43:06.925 INFO:teuthology.orchestra.run.vm04.stderr:Failed to stop ntpd.service: Unit ntpd.service not loaded. 2026-03-06T13:43:06.927 INFO:teuthology.orchestra.run.vm02.stderr:sudo: ntpd: command not found 2026-03-06T13:43:06.932 INFO:teuthology.orchestra.run.vm06.stderr:Failed to stop ntp.service: Unit ntp.service not loaded. 2026-03-06T13:43:06.939 INFO:teuthology.orchestra.run.vm02.stdout:506 Cannot talk to daemon 2026-03-06T13:43:06.944 INFO:teuthology.orchestra.run.vm04.stderr:sudo: ntpd: command not found 2026-03-06T13:43:06.950 INFO:teuthology.orchestra.run.vm02.stderr:Failed to start ntp.service: Unit ntp.service not found. 2026-03-06T13:43:06.951 INFO:teuthology.orchestra.run.vm06.stderr:Failed to stop ntpd.service: Unit ntpd.service not loaded. 2026-03-06T13:43:06.957 INFO:teuthology.orchestra.run.vm04.stdout:506 Cannot talk to daemon 2026-03-06T13:43:06.961 INFO:teuthology.orchestra.run.vm02.stderr:Failed to start ntpd.service: Unit ntpd.service not found. 2026-03-06T13:43:06.973 INFO:teuthology.orchestra.run.vm04.stderr:Failed to start ntp.service: Unit ntp.service not found. 2026-03-06T13:43:06.981 INFO:teuthology.orchestra.run.vm06.stderr:sudo: ntpd: command not found 2026-03-06T13:43:06.994 INFO:teuthology.orchestra.run.vm04.stderr:Failed to start ntpd.service: Unit ntpd.service not found. 2026-03-06T13:43:06.997 INFO:teuthology.orchestra.run.vm06.stdout:506 Cannot talk to daemon 2026-03-06T13:43:07.002 INFO:teuthology.orchestra.run.vm02.stderr:bash: line 1: ntpq: command not found 2026-03-06T13:43:07.004 INFO:teuthology.orchestra.run.vm02.stdout:MS Name/IP address Stratum Poll Reach LastRx Last sample 2026-03-06T13:43:07.004 INFO:teuthology.orchestra.run.vm02.stdout:=============================================================================== 2026-03-06T13:43:07.012 INFO:teuthology.orchestra.run.vm06.stderr:Failed to start ntp.service: Unit ntp.service not found. 2026-03-06T13:43:07.027 INFO:teuthology.orchestra.run.vm06.stderr:Failed to start ntpd.service: Unit ntpd.service not found. 2026-03-06T13:43:07.046 INFO:teuthology.orchestra.run.vm04.stderr:bash: line 1: ntpq: command not found 2026-03-06T13:43:07.049 INFO:teuthology.orchestra.run.vm04.stdout:MS Name/IP address Stratum Poll Reach LastRx Last sample 2026-03-06T13:43:07.050 INFO:teuthology.orchestra.run.vm04.stdout:=============================================================================== 2026-03-06T13:43:07.083 INFO:teuthology.orchestra.run.vm06.stderr:bash: line 1: ntpq: command not found 2026-03-06T13:43:07.086 INFO:teuthology.orchestra.run.vm06.stdout:MS Name/IP address Stratum Poll Reach LastRx Last sample 2026-03-06T13:43:07.086 INFO:teuthology.orchestra.run.vm06.stdout:=============================================================================== 2026-03-06T13:43:07.088 INFO:teuthology.run_tasks:Running task pexec... 2026-03-06T13:43:07.091 INFO:teuthology.task.pexec:Executing custom commands... 2026-03-06T13:43:07.091 DEBUG:teuthology.orchestra.run.vm02:> TESTDIR=/home/ubuntu/cephtest bash -s 2026-03-06T13:43:07.091 DEBUG:teuthology.orchestra.run.vm04:> TESTDIR=/home/ubuntu/cephtest bash -s 2026-03-06T13:43:07.091 DEBUG:teuthology.orchestra.run.vm06:> TESTDIR=/home/ubuntu/cephtest bash -s 2026-03-06T13:43:07.092 DEBUG:teuthology.task.pexec:ubuntu@vm02.local< sudo dnf remove nvme-cli -y 2026-03-06T13:43:07.092 DEBUG:teuthology.task.pexec:ubuntu@vm02.local< sudo dnf install runc nvmetcli nvme-cli -y 2026-03-06T13:43:07.092 DEBUG:teuthology.task.pexec:ubuntu@vm02.local< sudo sed -i 's/^#runtime = "crun"/runtime = "runc"/g' /usr/share/containers/containers.conf 2026-03-06T13:43:07.093 DEBUG:teuthology.task.pexec:ubuntu@vm02.local< sudo sed -i 's/runtime = "crun"/#runtime = "crun"/g' /usr/share/containers/containers.conf 2026-03-06T13:43:07.093 INFO:teuthology.task.pexec:Running commands on host ubuntu@vm02.local 2026-03-06T13:43:07.093 INFO:teuthology.task.pexec:sudo dnf remove nvme-cli -y 2026-03-06T13:43:07.093 INFO:teuthology.task.pexec:sudo dnf install runc nvmetcli nvme-cli -y 2026-03-06T13:43:07.093 INFO:teuthology.task.pexec:sudo sed -i 's/^#runtime = "crun"/runtime = "runc"/g' /usr/share/containers/containers.conf 2026-03-06T13:43:07.093 INFO:teuthology.task.pexec:sudo sed -i 's/runtime = "crun"/#runtime = "crun"/g' /usr/share/containers/containers.conf 2026-03-06T13:43:07.093 DEBUG:teuthology.task.pexec:ubuntu@vm04.local< sudo dnf remove nvme-cli -y 2026-03-06T13:43:07.093 DEBUG:teuthology.task.pexec:ubuntu@vm04.local< sudo dnf install runc nvmetcli nvme-cli -y 2026-03-06T13:43:07.093 DEBUG:teuthology.task.pexec:ubuntu@vm04.local< sudo sed -i 's/^#runtime = "crun"/runtime = "runc"/g' /usr/share/containers/containers.conf 2026-03-06T13:43:07.093 DEBUG:teuthology.task.pexec:ubuntu@vm04.local< sudo sed -i 's/runtime = "crun"/#runtime = "crun"/g' /usr/share/containers/containers.conf 2026-03-06T13:43:07.093 INFO:teuthology.task.pexec:Running commands on host ubuntu@vm04.local 2026-03-06T13:43:07.093 INFO:teuthology.task.pexec:sudo dnf remove nvme-cli -y 2026-03-06T13:43:07.093 INFO:teuthology.task.pexec:sudo dnf install runc nvmetcli nvme-cli -y 2026-03-06T13:43:07.093 INFO:teuthology.task.pexec:sudo sed -i 's/^#runtime = "crun"/runtime = "runc"/g' /usr/share/containers/containers.conf 2026-03-06T13:43:07.093 INFO:teuthology.task.pexec:sudo sed -i 's/runtime = "crun"/#runtime = "crun"/g' /usr/share/containers/containers.conf 2026-03-06T13:43:07.095 DEBUG:teuthology.task.pexec:ubuntu@vm06.local< sudo dnf remove nvme-cli -y 2026-03-06T13:43:07.095 DEBUG:teuthology.task.pexec:ubuntu@vm06.local< sudo dnf install runc nvmetcli nvme-cli -y 2026-03-06T13:43:07.095 DEBUG:teuthology.task.pexec:ubuntu@vm06.local< sudo sed -i 's/^#runtime = "crun"/runtime = "runc"/g' /usr/share/containers/containers.conf 2026-03-06T13:43:07.096 DEBUG:teuthology.task.pexec:ubuntu@vm06.local< sudo sed -i 's/runtime = "crun"/#runtime = "crun"/g' /usr/share/containers/containers.conf 2026-03-06T13:43:07.096 INFO:teuthology.task.pexec:Running commands on host ubuntu@vm06.local 2026-03-06T13:43:07.096 INFO:teuthology.task.pexec:sudo dnf remove nvme-cli -y 2026-03-06T13:43:07.096 INFO:teuthology.task.pexec:sudo dnf install runc nvmetcli nvme-cli -y 2026-03-06T13:43:07.096 INFO:teuthology.task.pexec:sudo sed -i 's/^#runtime = "crun"/runtime = "runc"/g' /usr/share/containers/containers.conf 2026-03-06T13:43:07.096 INFO:teuthology.task.pexec:sudo sed -i 's/runtime = "crun"/#runtime = "crun"/g' /usr/share/containers/containers.conf 2026-03-06T13:43:07.288 INFO:teuthology.orchestra.run.vm02.stdout:No match for argument: nvme-cli 2026-03-06T13:43:07.288 INFO:teuthology.orchestra.run.vm02.stderr:No packages marked for removal. 2026-03-06T13:43:07.291 INFO:teuthology.orchestra.run.vm02.stdout:Dependencies resolved. 2026-03-06T13:43:07.291 INFO:teuthology.orchestra.run.vm02.stdout:Nothing to do. 2026-03-06T13:43:07.291 INFO:teuthology.orchestra.run.vm02.stdout:Complete! 2026-03-06T13:43:07.300 INFO:teuthology.orchestra.run.vm04.stdout:No match for argument: nvme-cli 2026-03-06T13:43:07.300 INFO:teuthology.orchestra.run.vm04.stderr:No packages marked for removal. 2026-03-06T13:43:07.303 INFO:teuthology.orchestra.run.vm04.stdout:Dependencies resolved. 2026-03-06T13:43:07.303 INFO:teuthology.orchestra.run.vm04.stdout:Nothing to do. 2026-03-06T13:43:07.304 INFO:teuthology.orchestra.run.vm04.stdout:Complete! 2026-03-06T13:43:07.367 INFO:teuthology.orchestra.run.vm06.stdout:No match for argument: nvme-cli 2026-03-06T13:43:07.368 INFO:teuthology.orchestra.run.vm06.stderr:No packages marked for removal. 2026-03-06T13:43:07.371 INFO:teuthology.orchestra.run.vm06.stdout:Dependencies resolved. 2026-03-06T13:43:07.376 INFO:teuthology.orchestra.run.vm06.stdout:Nothing to do. 2026-03-06T13:43:07.376 INFO:teuthology.orchestra.run.vm06.stdout:Complete! 2026-03-06T13:43:07.691 INFO:teuthology.orchestra.run.vm02.stdout:Last metadata expiration check: 0:02:05 ago on Fri 06 Mar 2026 01:41:02 PM CET. 2026-03-06T13:43:07.717 INFO:teuthology.orchestra.run.vm04.stdout:Last metadata expiration check: 0:01:50 ago on Fri 06 Mar 2026 01:41:17 PM CET. 2026-03-06T13:43:07.789 INFO:teuthology.orchestra.run.vm02.stdout:Dependencies resolved. 2026-03-06T13:43:07.789 INFO:teuthology.orchestra.run.vm02.stdout:================================================================================ 2026-03-06T13:43:07.789 INFO:teuthology.orchestra.run.vm02.stdout: Package Arch Version Repository Size 2026-03-06T13:43:07.789 INFO:teuthology.orchestra.run.vm02.stdout:================================================================================ 2026-03-06T13:43:07.789 INFO:teuthology.orchestra.run.vm02.stdout:Installing: 2026-03-06T13:43:07.789 INFO:teuthology.orchestra.run.vm02.stdout: nvme-cli x86_64 2.16-1.el9 baseos 1.2 M 2026-03-06T13:43:07.789 INFO:teuthology.orchestra.run.vm02.stdout: nvmetcli noarch 0.8-3.el9 baseos 44 k 2026-03-06T13:43:07.789 INFO:teuthology.orchestra.run.vm02.stdout: runc x86_64 4:1.4.0-2.el9 appstream 4.0 M 2026-03-06T13:43:07.789 INFO:teuthology.orchestra.run.vm02.stdout:Installing dependencies: 2026-03-06T13:43:07.789 INFO:teuthology.orchestra.run.vm02.stdout: python3-configshell noarch 1:1.1.30-1.el9 baseos 72 k 2026-03-06T13:43:07.789 INFO:teuthology.orchestra.run.vm02.stdout: python3-kmod x86_64 0.9-32.el9 baseos 84 k 2026-03-06T13:43:07.789 INFO:teuthology.orchestra.run.vm02.stdout: python3-pyparsing noarch 2.4.7-9.el9 baseos 150 k 2026-03-06T13:43:07.789 INFO:teuthology.orchestra.run.vm02.stdout: python3-urwid x86_64 2.1.2-4.el9 baseos 837 k 2026-03-06T13:43:07.789 INFO:teuthology.orchestra.run.vm02.stdout: 2026-03-06T13:43:07.789 INFO:teuthology.orchestra.run.vm02.stdout:Transaction Summary 2026-03-06T13:43:07.789 INFO:teuthology.orchestra.run.vm02.stdout:================================================================================ 2026-03-06T13:43:07.789 INFO:teuthology.orchestra.run.vm02.stdout:Install 7 Packages 2026-03-06T13:43:07.789 INFO:teuthology.orchestra.run.vm02.stdout: 2026-03-06T13:43:07.790 INFO:teuthology.orchestra.run.vm02.stdout:Total download size: 6.3 M 2026-03-06T13:43:07.790 INFO:teuthology.orchestra.run.vm02.stdout:Installed size: 24 M 2026-03-06T13:43:07.790 INFO:teuthology.orchestra.run.vm02.stdout:Downloading Packages: 2026-03-06T13:43:07.823 INFO:teuthology.orchestra.run.vm04.stdout:Dependencies resolved. 2026-03-06T13:43:07.824 INFO:teuthology.orchestra.run.vm04.stdout:================================================================================ 2026-03-06T13:43:07.824 INFO:teuthology.orchestra.run.vm04.stdout: Package Arch Version Repository Size 2026-03-06T13:43:07.824 INFO:teuthology.orchestra.run.vm04.stdout:================================================================================ 2026-03-06T13:43:07.824 INFO:teuthology.orchestra.run.vm04.stdout:Installing: 2026-03-06T13:43:07.824 INFO:teuthology.orchestra.run.vm04.stdout: nvme-cli x86_64 2.16-1.el9 baseos 1.2 M 2026-03-06T13:43:07.824 INFO:teuthology.orchestra.run.vm04.stdout: nvmetcli noarch 0.8-3.el9 baseos 44 k 2026-03-06T13:43:07.824 INFO:teuthology.orchestra.run.vm04.stdout: runc x86_64 4:1.4.0-2.el9 appstream 4.0 M 2026-03-06T13:43:07.824 INFO:teuthology.orchestra.run.vm04.stdout:Installing dependencies: 2026-03-06T13:43:07.824 INFO:teuthology.orchestra.run.vm04.stdout: python3-configshell noarch 1:1.1.30-1.el9 baseos 72 k 2026-03-06T13:43:07.824 INFO:teuthology.orchestra.run.vm04.stdout: python3-kmod x86_64 0.9-32.el9 baseos 84 k 2026-03-06T13:43:07.824 INFO:teuthology.orchestra.run.vm04.stdout: python3-pyparsing noarch 2.4.7-9.el9 baseos 150 k 2026-03-06T13:43:07.824 INFO:teuthology.orchestra.run.vm04.stdout: python3-urwid x86_64 2.1.2-4.el9 baseos 837 k 2026-03-06T13:43:07.824 INFO:teuthology.orchestra.run.vm04.stdout: 2026-03-06T13:43:07.824 INFO:teuthology.orchestra.run.vm04.stdout:Transaction Summary 2026-03-06T13:43:07.824 INFO:teuthology.orchestra.run.vm04.stdout:================================================================================ 2026-03-06T13:43:07.824 INFO:teuthology.orchestra.run.vm04.stdout:Install 7 Packages 2026-03-06T13:43:07.824 INFO:teuthology.orchestra.run.vm04.stdout: 2026-03-06T13:43:07.824 INFO:teuthology.orchestra.run.vm04.stdout:Total download size: 6.3 M 2026-03-06T13:43:07.824 INFO:teuthology.orchestra.run.vm04.stdout:Installed size: 24 M 2026-03-06T13:43:07.824 INFO:teuthology.orchestra.run.vm04.stdout:Downloading Packages: 2026-03-06T13:43:07.903 INFO:teuthology.orchestra.run.vm06.stdout:Last metadata expiration check: 0:01:51 ago on Fri 06 Mar 2026 01:41:16 PM CET. 2026-03-06T13:43:08.024 INFO:teuthology.orchestra.run.vm06.stdout:Dependencies resolved. 2026-03-06T13:43:08.024 INFO:teuthology.orchestra.run.vm06.stdout:================================================================================ 2026-03-06T13:43:08.024 INFO:teuthology.orchestra.run.vm06.stdout: Package Arch Version Repository Size 2026-03-06T13:43:08.024 INFO:teuthology.orchestra.run.vm06.stdout:================================================================================ 2026-03-06T13:43:08.024 INFO:teuthology.orchestra.run.vm06.stdout:Installing: 2026-03-06T13:43:08.024 INFO:teuthology.orchestra.run.vm06.stdout: nvme-cli x86_64 2.16-1.el9 baseos 1.2 M 2026-03-06T13:43:08.024 INFO:teuthology.orchestra.run.vm06.stdout: nvmetcli noarch 0.8-3.el9 baseos 44 k 2026-03-06T13:43:08.025 INFO:teuthology.orchestra.run.vm06.stdout: runc x86_64 4:1.4.0-2.el9 appstream 4.0 M 2026-03-06T13:43:08.025 INFO:teuthology.orchestra.run.vm06.stdout:Installing dependencies: 2026-03-06T13:43:08.025 INFO:teuthology.orchestra.run.vm06.stdout: python3-configshell noarch 1:1.1.30-1.el9 baseos 72 k 2026-03-06T13:43:08.025 INFO:teuthology.orchestra.run.vm06.stdout: python3-kmod x86_64 0.9-32.el9 baseos 84 k 2026-03-06T13:43:08.025 INFO:teuthology.orchestra.run.vm06.stdout: python3-pyparsing noarch 2.4.7-9.el9 baseos 150 k 2026-03-06T13:43:08.025 INFO:teuthology.orchestra.run.vm06.stdout: python3-urwid x86_64 2.1.2-4.el9 baseos 837 k 2026-03-06T13:43:08.025 INFO:teuthology.orchestra.run.vm06.stdout: 2026-03-06T13:43:08.025 INFO:teuthology.orchestra.run.vm06.stdout:Transaction Summary 2026-03-06T13:43:08.025 INFO:teuthology.orchestra.run.vm06.stdout:================================================================================ 2026-03-06T13:43:08.025 INFO:teuthology.orchestra.run.vm06.stdout:Install 7 Packages 2026-03-06T13:43:08.025 INFO:teuthology.orchestra.run.vm06.stdout: 2026-03-06T13:43:08.025 INFO:teuthology.orchestra.run.vm06.stdout:Total download size: 6.3 M 2026-03-06T13:43:08.025 INFO:teuthology.orchestra.run.vm06.stdout:Installed size: 24 M 2026-03-06T13:43:08.025 INFO:teuthology.orchestra.run.vm06.stdout:Downloading Packages: 2026-03-06T13:43:08.693 INFO:teuthology.orchestra.run.vm04.stdout:(1/7): nvmetcli-0.8-3.el9.noarch.rpm 190 kB/s | 44 kB 00:00 2026-03-06T13:43:08.693 INFO:teuthology.orchestra.run.vm04.stdout:(2/7): python3-configshell-1.1.30-1.el9.noarch. 311 kB/s | 72 kB 00:00 2026-03-06T13:43:08.717 INFO:teuthology.orchestra.run.vm06.stdout:(1/7): nvmetcli-0.8-3.el9.noarch.rpm 350 kB/s | 44 kB 00:00 2026-03-06T13:43:08.747 INFO:teuthology.orchestra.run.vm06.stdout:(2/7): python3-configshell-1.1.30-1.el9.noarch. 462 kB/s | 72 kB 00:00 2026-03-06T13:43:08.787 INFO:teuthology.orchestra.run.vm04.stdout:(3/7): python3-pyparsing-2.4.7-9.el9.noarch.rpm 1.6 MB/s | 150 kB 00:00 2026-03-06T13:43:08.803 INFO:teuthology.orchestra.run.vm04.stdout:(4/7): python3-kmod-0.9-32.el9.x86_64.rpm 767 kB/s | 84 kB 00:00 2026-03-06T13:43:08.811 INFO:teuthology.orchestra.run.vm06.stdout:(3/7): python3-kmod-0.9-32.el9.x86_64.rpm 901 kB/s | 84 kB 00:00 2026-03-06T13:43:08.816 INFO:teuthology.orchestra.run.vm02.stdout:(1/7): nvmetcli-0.8-3.el9.noarch.rpm 1.7 MB/s | 44 kB 00:00 2026-03-06T13:43:08.835 INFO:teuthology.orchestra.run.vm02.stdout:(2/7): python3-configshell-1.1.30-1.el9.noarch. 1.6 MB/s | 72 kB 00:00 2026-03-06T13:43:08.844 INFO:teuthology.orchestra.run.vm02.stdout:(3/7): nvme-cli-2.16-1.el9.x86_64.rpm 22 MB/s | 1.2 MB 00:00 2026-03-06T13:43:08.847 INFO:teuthology.orchestra.run.vm02.stdout:(4/7): python3-kmod-0.9-32.el9.x86_64.rpm 2.7 MB/s | 84 kB 00:00 2026-03-06T13:43:08.863 INFO:teuthology.orchestra.run.vm02.stdout:(5/7): python3-pyparsing-2.4.7-9.el9.noarch.rpm 5.3 MB/s | 150 kB 00:00 2026-03-06T13:43:08.878 INFO:teuthology.orchestra.run.vm02.stdout:(6/7): python3-urwid-2.1.2-4.el9.x86_64.rpm 24 MB/s | 837 kB 00:00 2026-03-06T13:43:08.878 INFO:teuthology.orchestra.run.vm06.stdout:(4/7): python3-pyparsing-2.4.7-9.el9.noarch.rpm 1.1 MB/s | 150 kB 00:00 2026-03-06T13:43:09.113 INFO:teuthology.orchestra.run.vm04.stdout:(5/7): nvme-cli-2.16-1.el9.x86_64.rpm 1.8 MB/s | 1.2 MB 00:00 2026-03-06T13:43:09.119 INFO:teuthology.orchestra.run.vm04.stdout:(6/7): python3-urwid-2.1.2-4.el9.x86_64.rpm 2.5 MB/s | 837 kB 00:00 2026-03-06T13:43:09.156 INFO:teuthology.orchestra.run.vm06.stdout:(5/7): python3-urwid-2.1.2-4.el9.x86_64.rpm 2.4 MB/s | 837 kB 00:00 2026-03-06T13:43:09.187 INFO:teuthology.orchestra.run.vm06.stdout:(6/7): nvme-cli-2.16-1.el9.x86_64.rpm 1.9 MB/s | 1.2 MB 00:00 2026-03-06T13:43:09.431 INFO:teuthology.orchestra.run.vm06.stdout:(7/7): runc-1.4.0-2.el9.x86_64.rpm 7.2 MB/s | 4.0 MB 00:00 2026-03-06T13:43:09.431 INFO:teuthology.orchestra.run.vm06.stdout:-------------------------------------------------------------------------------- 2026-03-06T13:43:09.431 INFO:teuthology.orchestra.run.vm06.stdout:Total 4.5 MB/s | 6.3 MB 00:01 2026-03-06T13:43:09.459 INFO:teuthology.orchestra.run.vm04.stdout:(7/7): runc-1.4.0-2.el9.x86_64.rpm 6.0 MB/s | 4.0 MB 00:00 2026-03-06T13:43:09.459 INFO:teuthology.orchestra.run.vm04.stdout:-------------------------------------------------------------------------------- 2026-03-06T13:43:09.459 INFO:teuthology.orchestra.run.vm04.stdout:Total 3.8 MB/s | 6.3 MB 00:01 2026-03-06T13:43:09.521 INFO:teuthology.orchestra.run.vm06.stdout:Running transaction check 2026-03-06T13:43:09.532 INFO:teuthology.orchestra.run.vm06.stdout:Transaction check succeeded. 2026-03-06T13:43:09.532 INFO:teuthology.orchestra.run.vm06.stdout:Running transaction test 2026-03-06T13:43:09.533 INFO:teuthology.orchestra.run.vm04.stdout:Running transaction check 2026-03-06T13:43:09.542 INFO:teuthology.orchestra.run.vm04.stdout:Transaction check succeeded. 2026-03-06T13:43:09.542 INFO:teuthology.orchestra.run.vm04.stdout:Running transaction test 2026-03-06T13:43:09.571 INFO:teuthology.orchestra.run.vm02.stdout:(7/7): runc-1.4.0-2.el9.x86_64.rpm 5.5 MB/s | 4.0 MB 00:00 2026-03-06T13:43:09.572 INFO:teuthology.orchestra.run.vm02.stdout:-------------------------------------------------------------------------------- 2026-03-06T13:43:09.573 INFO:teuthology.orchestra.run.vm02.stdout:Total 3.5 MB/s | 6.3 MB 00:01 2026-03-06T13:43:09.610 INFO:teuthology.orchestra.run.vm04.stdout:Transaction test succeeded. 2026-03-06T13:43:09.611 INFO:teuthology.orchestra.run.vm04.stdout:Running transaction 2026-03-06T13:43:09.621 INFO:teuthology.orchestra.run.vm06.stdout:Transaction test succeeded. 2026-03-06T13:43:09.621 INFO:teuthology.orchestra.run.vm06.stdout:Running transaction 2026-03-06T13:43:09.662 INFO:teuthology.orchestra.run.vm02.stdout:Running transaction check 2026-03-06T13:43:09.670 INFO:teuthology.orchestra.run.vm02.stdout:Transaction check succeeded. 2026-03-06T13:43:09.670 INFO:teuthology.orchestra.run.vm02.stdout:Running transaction test 2026-03-06T13:43:09.734 INFO:teuthology.orchestra.run.vm02.stdout:Transaction test succeeded. 2026-03-06T13:43:09.735 INFO:teuthology.orchestra.run.vm02.stdout:Running transaction 2026-03-06T13:43:09.791 INFO:teuthology.orchestra.run.vm04.stdout: Preparing : 1/1 2026-03-06T13:43:09.803 INFO:teuthology.orchestra.run.vm04.stdout: Installing : python3-urwid-2.1.2-4.el9.x86_64 1/7 2026-03-06T13:43:09.815 INFO:teuthology.orchestra.run.vm04.stdout: Installing : python3-pyparsing-2.4.7-9.el9.noarch 2/7 2026-03-06T13:43:09.824 INFO:teuthology.orchestra.run.vm04.stdout: Installing : python3-configshell-1:1.1.30-1.el9.noarch 3/7 2026-03-06T13:43:09.834 INFO:teuthology.orchestra.run.vm04.stdout: Installing : python3-kmod-0.9-32.el9.x86_64 4/7 2026-03-06T13:43:09.836 INFO:teuthology.orchestra.run.vm04.stdout: Installing : nvmetcli-0.8-3.el9.noarch 5/7 2026-03-06T13:43:09.840 INFO:teuthology.orchestra.run.vm06.stdout: Preparing : 1/1 2026-03-06T13:43:09.855 INFO:teuthology.orchestra.run.vm06.stdout: Installing : python3-urwid-2.1.2-4.el9.x86_64 1/7 2026-03-06T13:43:09.871 INFO:teuthology.orchestra.run.vm06.stdout: Installing : python3-pyparsing-2.4.7-9.el9.noarch 2/7 2026-03-06T13:43:09.880 INFO:teuthology.orchestra.run.vm06.stdout: Installing : python3-configshell-1:1.1.30-1.el9.noarch 3/7 2026-03-06T13:43:09.890 INFO:teuthology.orchestra.run.vm06.stdout: Installing : python3-kmod-0.9-32.el9.x86_64 4/7 2026-03-06T13:43:09.893 INFO:teuthology.orchestra.run.vm06.stdout: Installing : nvmetcli-0.8-3.el9.noarch 5/7 2026-03-06T13:43:09.896 INFO:teuthology.orchestra.run.vm04.stdout: Running scriptlet: nvmetcli-0.8-3.el9.noarch 5/7 2026-03-06T13:43:09.897 INFO:teuthology.orchestra.run.vm02.stdout: Preparing : 1/1 2026-03-06T13:43:09.909 INFO:teuthology.orchestra.run.vm02.stdout: Installing : python3-urwid-2.1.2-4.el9.x86_64 1/7 2026-03-06T13:43:09.920 INFO:teuthology.orchestra.run.vm02.stdout: Installing : python3-pyparsing-2.4.7-9.el9.noarch 2/7 2026-03-06T13:43:09.927 INFO:teuthology.orchestra.run.vm02.stdout: Installing : python3-configshell-1:1.1.30-1.el9.noarch 3/7 2026-03-06T13:43:09.935 INFO:teuthology.orchestra.run.vm02.stdout: Installing : python3-kmod-0.9-32.el9.x86_64 4/7 2026-03-06T13:43:09.937 INFO:teuthology.orchestra.run.vm02.stdout: Installing : nvmetcli-0.8-3.el9.noarch 5/7 2026-03-06T13:43:09.955 INFO:teuthology.orchestra.run.vm06.stdout: Running scriptlet: nvmetcli-0.8-3.el9.noarch 5/7 2026-03-06T13:43:09.990 INFO:teuthology.orchestra.run.vm02.stdout: Running scriptlet: nvmetcli-0.8-3.el9.noarch 5/7 2026-03-06T13:43:10.041 INFO:teuthology.orchestra.run.vm04.stdout: Installing : runc-4:1.4.0-2.el9.x86_64 6/7 2026-03-06T13:43:10.047 INFO:teuthology.orchestra.run.vm04.stdout: Installing : nvme-cli-2.16-1.el9.x86_64 7/7 2026-03-06T13:43:10.127 INFO:teuthology.orchestra.run.vm02.stdout: Installing : runc-4:1.4.0-2.el9.x86_64 6/7 2026-03-06T13:43:10.142 INFO:teuthology.orchestra.run.vm06.stdout: Installing : runc-4:1.4.0-2.el9.x86_64 6/7 2026-03-06T13:43:10.337 INFO:teuthology.orchestra.run.vm02.stdout: Installing : nvme-cli-2.16-1.el9.x86_64 7/7 2026-03-06T13:43:10.340 INFO:teuthology.orchestra.run.vm06.stdout: Installing : nvme-cli-2.16-1.el9.x86_64 7/7 2026-03-06T13:43:10.425 INFO:teuthology.orchestra.run.vm04.stdout: Running scriptlet: nvme-cli-2.16-1.el9.x86_64 7/7 2026-03-06T13:43:10.425 INFO:teuthology.orchestra.run.vm04.stdout:Created symlink /etc/systemd/system/default.target.wants/nvmefc-boot-connections.service → /usr/lib/systemd/system/nvmefc-boot-connections.service. 2026-03-06T13:43:10.425 INFO:teuthology.orchestra.run.vm04.stdout: 2026-03-06T13:43:10.655 INFO:teuthology.orchestra.run.vm02.stdout: Running scriptlet: nvme-cli-2.16-1.el9.x86_64 7/7 2026-03-06T13:43:10.655 INFO:teuthology.orchestra.run.vm02.stdout:Created symlink /etc/systemd/system/default.target.wants/nvmefc-boot-connections.service → /usr/lib/systemd/system/nvmefc-boot-connections.service. 2026-03-06T13:43:10.655 INFO:teuthology.orchestra.run.vm02.stdout: 2026-03-06T13:43:10.753 INFO:teuthology.orchestra.run.vm06.stdout: Running scriptlet: nvme-cli-2.16-1.el9.x86_64 7/7 2026-03-06T13:43:10.753 INFO:teuthology.orchestra.run.vm06.stdout:Created symlink /etc/systemd/system/default.target.wants/nvmefc-boot-connections.service → /usr/lib/systemd/system/nvmefc-boot-connections.service. 2026-03-06T13:43:10.753 INFO:teuthology.orchestra.run.vm06.stdout: 2026-03-06T13:43:11.117 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : nvme-cli-2.16-1.el9.x86_64 1/7 2026-03-06T13:43:11.118 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : nvmetcli-0.8-3.el9.noarch 2/7 2026-03-06T13:43:11.118 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : python3-configshell-1:1.1.30-1.el9.noarch 3/7 2026-03-06T13:43:11.118 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : python3-kmod-0.9-32.el9.x86_64 4/7 2026-03-06T13:43:11.118 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : python3-pyparsing-2.4.7-9.el9.noarch 5/7 2026-03-06T13:43:11.118 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : python3-urwid-2.1.2-4.el9.x86_64 6/7 2026-03-06T13:43:11.128 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : nvme-cli-2.16-1.el9.x86_64 1/7 2026-03-06T13:43:11.129 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : nvmetcli-0.8-3.el9.noarch 2/7 2026-03-06T13:43:11.129 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-configshell-1:1.1.30-1.el9.noarch 3/7 2026-03-06T13:43:11.129 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-kmod-0.9-32.el9.x86_64 4/7 2026-03-06T13:43:11.129 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-pyparsing-2.4.7-9.el9.noarch 5/7 2026-03-06T13:43:11.129 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-urwid-2.1.2-4.el9.x86_64 6/7 2026-03-06T13:43:11.185 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : runc-4:1.4.0-2.el9.x86_64 7/7 2026-03-06T13:43:11.185 INFO:teuthology.orchestra.run.vm02.stdout: 2026-03-06T13:43:11.185 INFO:teuthology.orchestra.run.vm02.stdout:Installed: 2026-03-06T13:43:11.185 INFO:teuthology.orchestra.run.vm02.stdout: nvme-cli-2.16-1.el9.x86_64 nvmetcli-0.8-3.el9.noarch 2026-03-06T13:43:11.185 INFO:teuthology.orchestra.run.vm02.stdout: python3-configshell-1:1.1.30-1.el9.noarch python3-kmod-0.9-32.el9.x86_64 2026-03-06T13:43:11.185 INFO:teuthology.orchestra.run.vm02.stdout: python3-pyparsing-2.4.7-9.el9.noarch python3-urwid-2.1.2-4.el9.x86_64 2026-03-06T13:43:11.185 INFO:teuthology.orchestra.run.vm02.stdout: runc-4:1.4.0-2.el9.x86_64 2026-03-06T13:43:11.185 INFO:teuthology.orchestra.run.vm02.stdout: 2026-03-06T13:43:11.185 INFO:teuthology.orchestra.run.vm02.stdout:Complete! 2026-03-06T13:43:11.219 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : runc-4:1.4.0-2.el9.x86_64 7/7 2026-03-06T13:43:11.219 INFO:teuthology.orchestra.run.vm04.stdout: 2026-03-06T13:43:11.219 INFO:teuthology.orchestra.run.vm04.stdout:Installed: 2026-03-06T13:43:11.219 INFO:teuthology.orchestra.run.vm04.stdout: nvme-cli-2.16-1.el9.x86_64 nvmetcli-0.8-3.el9.noarch 2026-03-06T13:43:11.219 INFO:teuthology.orchestra.run.vm04.stdout: python3-configshell-1:1.1.30-1.el9.noarch python3-kmod-0.9-32.el9.x86_64 2026-03-06T13:43:11.219 INFO:teuthology.orchestra.run.vm04.stdout: python3-pyparsing-2.4.7-9.el9.noarch python3-urwid-2.1.2-4.el9.x86_64 2026-03-06T13:43:11.219 INFO:teuthology.orchestra.run.vm04.stdout: runc-4:1.4.0-2.el9.x86_64 2026-03-06T13:43:11.219 INFO:teuthology.orchestra.run.vm04.stdout: 2026-03-06T13:43:11.219 INFO:teuthology.orchestra.run.vm04.stdout:Complete! 2026-03-06T13:43:11.267 DEBUG:teuthology.parallel:result is None 2026-03-06T13:43:11.323 DEBUG:teuthology.parallel:result is None 2026-03-06T13:43:11.378 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : nvme-cli-2.16-1.el9.x86_64 1/7 2026-03-06T13:43:11.379 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : nvmetcli-0.8-3.el9.noarch 2/7 2026-03-06T13:43:11.379 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-configshell-1:1.1.30-1.el9.noarch 3/7 2026-03-06T13:43:11.379 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-kmod-0.9-32.el9.x86_64 4/7 2026-03-06T13:43:11.379 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-pyparsing-2.4.7-9.el9.noarch 5/7 2026-03-06T13:43:11.379 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-urwid-2.1.2-4.el9.x86_64 6/7 2026-03-06T13:43:11.490 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : runc-4:1.4.0-2.el9.x86_64 7/7 2026-03-06T13:43:11.490 INFO:teuthology.orchestra.run.vm06.stdout: 2026-03-06T13:43:11.490 INFO:teuthology.orchestra.run.vm06.stdout:Installed: 2026-03-06T13:43:11.490 INFO:teuthology.orchestra.run.vm06.stdout: nvme-cli-2.16-1.el9.x86_64 nvmetcli-0.8-3.el9.noarch 2026-03-06T13:43:11.490 INFO:teuthology.orchestra.run.vm06.stdout: python3-configshell-1:1.1.30-1.el9.noarch python3-kmod-0.9-32.el9.x86_64 2026-03-06T13:43:11.490 INFO:teuthology.orchestra.run.vm06.stdout: python3-pyparsing-2.4.7-9.el9.noarch python3-urwid-2.1.2-4.el9.x86_64 2026-03-06T13:43:11.490 INFO:teuthology.orchestra.run.vm06.stdout: runc-4:1.4.0-2.el9.x86_64 2026-03-06T13:43:11.490 INFO:teuthology.orchestra.run.vm06.stdout: 2026-03-06T13:43:11.490 INFO:teuthology.orchestra.run.vm06.stdout:Complete! 2026-03-06T13:43:11.628 DEBUG:teuthology.parallel:result is None 2026-03-06T13:43:11.629 INFO:teuthology.run_tasks:Running task install... 2026-03-06T13:43:11.631 DEBUG:teuthology.task.install:project ceph 2026-03-06T13:43:11.631 DEBUG:teuthology.task.install:INSTALL overrides: {'ceph': {'flavor': 'default', 'sha1': 'c24117fd5525679b799527bc1bd1f1dd0a2db5e2'}, 'extra_system_packages': {'deb': ['python3-xmltodict', 's3cmd'], 'rpm': ['bzip2', 'perl-Test-Harness', 'python3-xmltodict', 's3cmd']}, 'repos': [{'name': 'ceph-source', 'priority': 1, 'url': 'https://s3.clyso.com/ces-packages/components/ceph/rpm-19.2.3-47-gc24117fd552/el9.clyso/SRPMS'}, {'name': 'ceph-noarch', 'priority': 1, 'url': 'https://s3.clyso.com/ces-packages/components/ceph/rpm-19.2.3-47-gc24117fd552/el9.clyso/noarch'}, {'name': 'ceph', 'priority': 1, 'url': 'https://s3.clyso.com/ces-packages/components/ceph/rpm-19.2.3-47-gc24117fd552/el9.clyso/x86_64'}]} 2026-03-06T13:43:11.631 DEBUG:teuthology.task.install:config {'flavor': 'default', 'sha1': 'c24117fd5525679b799527bc1bd1f1dd0a2db5e2', 'extra_system_packages': {'deb': ['python3-xmltodict', 's3cmd'], 'rpm': ['bzip2', 'perl-Test-Harness', 'python3-xmltodict', 's3cmd']}} 2026-03-06T13:43:11.631 INFO:teuthology.task.install:Using flavor: default 2026-03-06T13:43:11.634 DEBUG:teuthology.task.install:Package list is: {'deb': ['ceph', 'cephadm', 'ceph-mds', 'ceph-mgr', 'ceph-common', 'ceph-fuse', 'ceph-test', 'ceph-volume', 'radosgw', 'python3-rados', 'python3-rgw', 'python3-cephfs', 'python3-rbd', 'libcephfs2', 'libcephfs-dev', 'librados2', 'librbd1', 'rbd-fuse'], 'rpm': ['ceph-radosgw', 'ceph-test', 'ceph', 'ceph-base', 'cephadm', 'ceph-immutable-object-cache', 'ceph-mgr', 'ceph-mgr-dashboard', 'ceph-mgr-diskprediction-local', 'ceph-mgr-rook', 'ceph-mgr-cephadm', 'ceph-fuse', 'ceph-volume', 'librados-devel', 'libcephfs2', 'libcephfs-devel', 'librados2', 'librbd1', 'python3-rados', 'python3-rgw', 'python3-cephfs', 'python3-rbd', 'rbd-fuse', 'rbd-mirror', 'rbd-nbd']} 2026-03-06T13:43:11.634 INFO:teuthology.task.install:extra packages: [] 2026-03-06T13:43:11.635 DEBUG:teuthology.task.install.rpm:_update_package_list_and_install: config is {'branch': None, 'cleanup': None, 'debuginfo': None, 'downgrade_packages': [], 'exclude_packages': [], 'extra_packages': [], 'extra_system_packages': {'deb': ['python3-xmltodict', 's3cmd'], 'rpm': ['bzip2', 'perl-Test-Harness', 'python3-xmltodict', 's3cmd']}, 'extras': None, 'enable_coprs': [], 'flavor': 'default', 'install_ceph_packages': True, 'packages': {}, 'project': 'ceph', 'repos_only': False, 'sha1': 'c24117fd5525679b799527bc1bd1f1dd0a2db5e2', 'tag': None, 'wait_for_package': False, 'repos': [{'name': 'ceph-source', 'priority': 1, 'url': 'https://s3.clyso.com/ces-packages/components/ceph/rpm-19.2.3-47-gc24117fd552/el9.clyso/SRPMS'}, {'name': 'ceph-noarch', 'priority': 1, 'url': 'https://s3.clyso.com/ces-packages/components/ceph/rpm-19.2.3-47-gc24117fd552/el9.clyso/noarch'}, {'name': 'ceph', 'priority': 1, 'url': 'https://s3.clyso.com/ces-packages/components/ceph/rpm-19.2.3-47-gc24117fd552/el9.clyso/x86_64'}]} 2026-03-06T13:43:11.635 DEBUG:teuthology.task.install.rpm:Adding repos: [{'name': 'ceph-source', 'priority': 1, 'url': 'https://s3.clyso.com/ces-packages/components/ceph/rpm-19.2.3-47-gc24117fd552/el9.clyso/SRPMS'}, {'name': 'ceph-noarch', 'priority': 1, 'url': 'https://s3.clyso.com/ces-packages/components/ceph/rpm-19.2.3-47-gc24117fd552/el9.clyso/noarch'}, {'name': 'ceph', 'priority': 1, 'url': 'https://s3.clyso.com/ces-packages/components/ceph/rpm-19.2.3-47-gc24117fd552/el9.clyso/x86_64'}] 2026-03-06T13:43:11.635 DEBUG:teuthology.orchestra.run.vm02:> set -ex 2026-03-06T13:43:11.635 DEBUG:teuthology.orchestra.run.vm02:> sudo dd of=/etc/yum.repos.d/ceph-source.repo 2026-03-06T13:43:11.635 DEBUG:teuthology.task.install.rpm:_update_package_list_and_install: config is {'branch': None, 'cleanup': None, 'debuginfo': None, 'downgrade_packages': [], 'exclude_packages': [], 'extra_packages': [], 'extra_system_packages': {'deb': ['python3-xmltodict', 's3cmd'], 'rpm': ['bzip2', 'perl-Test-Harness', 'python3-xmltodict', 's3cmd']}, 'extras': None, 'enable_coprs': [], 'flavor': 'default', 'install_ceph_packages': True, 'packages': {}, 'project': 'ceph', 'repos_only': False, 'sha1': 'c24117fd5525679b799527bc1bd1f1dd0a2db5e2', 'tag': None, 'wait_for_package': False, 'repos': [{'name': 'ceph-source', 'priority': 1, 'url': 'https://s3.clyso.com/ces-packages/components/ceph/rpm-19.2.3-47-gc24117fd552/el9.clyso/SRPMS'}, {'name': 'ceph-noarch', 'priority': 1, 'url': 'https://s3.clyso.com/ces-packages/components/ceph/rpm-19.2.3-47-gc24117fd552/el9.clyso/noarch'}, {'name': 'ceph', 'priority': 1, 'url': 'https://s3.clyso.com/ces-packages/components/ceph/rpm-19.2.3-47-gc24117fd552/el9.clyso/x86_64'}]} 2026-03-06T13:43:11.635 DEBUG:teuthology.task.install.rpm:Adding repos: [{'name': 'ceph-source', 'priority': 1, 'url': 'https://s3.clyso.com/ces-packages/components/ceph/rpm-19.2.3-47-gc24117fd552/el9.clyso/SRPMS'}, {'name': 'ceph-noarch', 'priority': 1, 'url': 'https://s3.clyso.com/ces-packages/components/ceph/rpm-19.2.3-47-gc24117fd552/el9.clyso/noarch'}, {'name': 'ceph', 'priority': 1, 'url': 'https://s3.clyso.com/ces-packages/components/ceph/rpm-19.2.3-47-gc24117fd552/el9.clyso/x86_64'}] 2026-03-06T13:43:11.635 DEBUG:teuthology.orchestra.run.vm04:> set -ex 2026-03-06T13:43:11.635 DEBUG:teuthology.orchestra.run.vm04:> sudo dd of=/etc/yum.repos.d/ceph-source.repo 2026-03-06T13:43:11.635 DEBUG:teuthology.task.install.rpm:_update_package_list_and_install: config is {'branch': None, 'cleanup': None, 'debuginfo': None, 'downgrade_packages': [], 'exclude_packages': [], 'extra_packages': [], 'extra_system_packages': {'deb': ['python3-xmltodict', 's3cmd'], 'rpm': ['bzip2', 'perl-Test-Harness', 'python3-xmltodict', 's3cmd']}, 'extras': None, 'enable_coprs': [], 'flavor': 'default', 'install_ceph_packages': True, 'packages': {}, 'project': 'ceph', 'repos_only': False, 'sha1': 'c24117fd5525679b799527bc1bd1f1dd0a2db5e2', 'tag': None, 'wait_for_package': False, 'repos': [{'name': 'ceph-source', 'priority': 1, 'url': 'https://s3.clyso.com/ces-packages/components/ceph/rpm-19.2.3-47-gc24117fd552/el9.clyso/SRPMS'}, {'name': 'ceph-noarch', 'priority': 1, 'url': 'https://s3.clyso.com/ces-packages/components/ceph/rpm-19.2.3-47-gc24117fd552/el9.clyso/noarch'}, {'name': 'ceph', 'priority': 1, 'url': 'https://s3.clyso.com/ces-packages/components/ceph/rpm-19.2.3-47-gc24117fd552/el9.clyso/x86_64'}]} 2026-03-06T13:43:11.636 DEBUG:teuthology.task.install.rpm:Adding repos: [{'name': 'ceph-source', 'priority': 1, 'url': 'https://s3.clyso.com/ces-packages/components/ceph/rpm-19.2.3-47-gc24117fd552/el9.clyso/SRPMS'}, {'name': 'ceph-noarch', 'priority': 1, 'url': 'https://s3.clyso.com/ces-packages/components/ceph/rpm-19.2.3-47-gc24117fd552/el9.clyso/noarch'}, {'name': 'ceph', 'priority': 1, 'url': 'https://s3.clyso.com/ces-packages/components/ceph/rpm-19.2.3-47-gc24117fd552/el9.clyso/x86_64'}] 2026-03-06T13:43:11.636 DEBUG:teuthology.orchestra.run.vm06:> set -ex 2026-03-06T13:43:11.636 DEBUG:teuthology.orchestra.run.vm06:> sudo dd of=/etc/yum.repos.d/ceph-source.repo 2026-03-06T13:43:11.660 DEBUG:teuthology.orchestra.run.vm02:> set -ex 2026-03-06T13:43:11.660 DEBUG:teuthology.orchestra.run.vm02:> sudo dd of=/etc/yum.repos.d/ceph-noarch.repo 2026-03-06T13:43:11.670 DEBUG:teuthology.orchestra.run.vm04:> set -ex 2026-03-06T13:43:11.670 DEBUG:teuthology.orchestra.run.vm04:> sudo dd of=/etc/yum.repos.d/ceph-noarch.repo 2026-03-06T13:43:11.674 DEBUG:teuthology.orchestra.run.vm06:> set -ex 2026-03-06T13:43:11.674 DEBUG:teuthology.orchestra.run.vm06:> sudo dd of=/etc/yum.repos.d/ceph-noarch.repo 2026-03-06T13:43:11.725 DEBUG:teuthology.orchestra.run.vm02:> set -ex 2026-03-06T13:43:11.725 DEBUG:teuthology.orchestra.run.vm02:> sudo dd of=/etc/yum.repos.d/ceph.repo 2026-03-06T13:43:11.750 DEBUG:teuthology.orchestra.run.vm04:> set -ex 2026-03-06T13:43:11.750 DEBUG:teuthology.orchestra.run.vm04:> sudo dd of=/etc/yum.repos.d/ceph.repo 2026-03-06T13:43:11.757 DEBUG:teuthology.orchestra.run.vm06:> set -ex 2026-03-06T13:43:11.757 DEBUG:teuthology.orchestra.run.vm06:> sudo dd of=/etc/yum.repos.d/ceph.repo 2026-03-06T13:43:11.791 INFO:teuthology.task.install.rpm:Installing packages: ceph-radosgw, ceph-test, ceph, ceph-base, cephadm, ceph-immutable-object-cache, ceph-mgr, ceph-mgr-dashboard, ceph-mgr-diskprediction-local, ceph-mgr-rook, ceph-mgr-cephadm, ceph-fuse, ceph-volume, librados-devel, libcephfs2, libcephfs-devel, librados2, librbd1, python3-rados, python3-rgw, python3-cephfs, python3-rbd, rbd-fuse, rbd-mirror, rbd-nbd, bzip2, perl-Test-Harness, python3-xmltodict, s3cmd on remote rpm x86_64 2026-03-06T13:43:11.791 DEBUG:teuthology.orchestra.run.vm02:> sudo yum clean all 2026-03-06T13:43:11.824 INFO:teuthology.task.install.rpm:Installing packages: ceph-radosgw, ceph-test, ceph, ceph-base, cephadm, ceph-immutable-object-cache, ceph-mgr, ceph-mgr-dashboard, ceph-mgr-diskprediction-local, ceph-mgr-rook, ceph-mgr-cephadm, ceph-fuse, ceph-volume, librados-devel, libcephfs2, libcephfs-devel, librados2, librbd1, python3-rados, python3-rgw, python3-cephfs, python3-rbd, rbd-fuse, rbd-mirror, rbd-nbd, bzip2, perl-Test-Harness, python3-xmltodict, s3cmd on remote rpm x86_64 2026-03-06T13:43:11.824 DEBUG:teuthology.orchestra.run.vm04:> sudo yum clean all 2026-03-06T13:43:11.833 INFO:teuthology.task.install.rpm:Installing packages: ceph-radosgw, ceph-test, ceph, ceph-base, cephadm, ceph-immutable-object-cache, ceph-mgr, ceph-mgr-dashboard, ceph-mgr-diskprediction-local, ceph-mgr-rook, ceph-mgr-cephadm, ceph-fuse, ceph-volume, librados-devel, libcephfs2, libcephfs-devel, librados2, librbd1, python3-rados, python3-rgw, python3-cephfs, python3-rbd, rbd-fuse, rbd-mirror, rbd-nbd, bzip2, perl-Test-Harness, python3-xmltodict, s3cmd on remote rpm x86_64 2026-03-06T13:43:11.833 DEBUG:teuthology.orchestra.run.vm06:> sudo yum clean all 2026-03-06T13:43:11.961 INFO:teuthology.orchestra.run.vm02.stdout:41 files removed 2026-03-06T13:43:11.979 DEBUG:teuthology.orchestra.run.vm02:> sudo yum -y install ceph-radosgw ceph-test ceph ceph-base cephadm ceph-immutable-object-cache ceph-mgr ceph-mgr-dashboard ceph-mgr-diskprediction-local ceph-mgr-rook ceph-mgr-cephadm ceph-fuse ceph-volume librados-devel libcephfs2 libcephfs-devel librados2 librbd1 python3-rados python3-rgw python3-cephfs python3-rbd rbd-fuse rbd-mirror rbd-nbd bzip2 perl-Test-Harness python3-xmltodict s3cmd 2026-03-06T13:43:12.016 INFO:teuthology.orchestra.run.vm04.stdout:41 files removed 2026-03-06T13:43:12.051 DEBUG:teuthology.orchestra.run.vm04:> sudo yum -y install ceph-radosgw ceph-test ceph ceph-base cephadm ceph-immutable-object-cache ceph-mgr ceph-mgr-dashboard ceph-mgr-diskprediction-local ceph-mgr-rook ceph-mgr-cephadm ceph-fuse ceph-volume librados-devel libcephfs2 libcephfs-devel librados2 librbd1 python3-rados python3-rgw python3-cephfs python3-rbd rbd-fuse rbd-mirror rbd-nbd bzip2 perl-Test-Harness python3-xmltodict s3cmd 2026-03-06T13:43:12.069 INFO:teuthology.orchestra.run.vm06.stdout:41 files removed 2026-03-06T13:43:12.103 DEBUG:teuthology.orchestra.run.vm06:> sudo yum -y install ceph-radosgw ceph-test ceph ceph-base cephadm ceph-immutable-object-cache ceph-mgr ceph-mgr-dashboard ceph-mgr-diskprediction-local ceph-mgr-rook ceph-mgr-cephadm ceph-fuse ceph-volume librados-devel libcephfs2 libcephfs-devel librados2 librbd1 python3-rados python3-rgw python3-cephfs python3-rbd rbd-fuse rbd-mirror rbd-nbd bzip2 perl-Test-Harness python3-xmltodict s3cmd 2026-03-06T13:43:12.380 INFO:teuthology.orchestra.run.vm02.stdout:ceph 382 kB/s | 86 kB 00:00 2026-03-06T13:43:12.473 INFO:teuthology.orchestra.run.vm04.stdout:ceph 358 kB/s | 86 kB 00:00 2026-03-06T13:43:12.540 INFO:teuthology.orchestra.run.vm06.stdout:ceph 378 kB/s | 86 kB 00:00 2026-03-06T13:43:12.576 INFO:teuthology.orchestra.run.vm02.stdout:ceph-noarch 70 kB/s | 12 kB 00:00 2026-03-06T13:43:12.675 INFO:teuthology.orchestra.run.vm04.stdout:ceph-noarch 69 kB/s | 12 kB 00:00 2026-03-06T13:43:12.751 INFO:teuthology.orchestra.run.vm06.stdout:ceph-noarch 69 kB/s | 12 kB 00:00 2026-03-06T13:43:12.757 INFO:teuthology.orchestra.run.vm02.stdout:ceph-source 14 kB/s | 2.2 kB 00:00 2026-03-06T13:43:12.868 INFO:teuthology.orchestra.run.vm04.stdout:ceph-source 13 kB/s | 2.2 kB 00:00 2026-03-06T13:43:12.949 INFO:teuthology.orchestra.run.vm06.stdout:ceph-source 13 kB/s | 2.2 kB 00:00 2026-03-06T13:43:13.433 INFO:teuthology.orchestra.run.vm04.stdout:CentOS Stream 9 - BaseOS 16 MB/s | 8.9 MB 00:00 2026-03-06T13:43:14.039 INFO:teuthology.orchestra.run.vm06.stdout:CentOS Stream 9 - BaseOS 8.3 MB/s | 8.9 MB 00:01 2026-03-06T13:43:14.605 INFO:teuthology.orchestra.run.vm02.stdout:CentOS Stream 9 - BaseOS 4.9 MB/s | 8.9 MB 00:01 2026-03-06T13:43:16.841 INFO:teuthology.orchestra.run.vm06.stdout:CentOS Stream 9 - AppStream 14 MB/s | 27 MB 00:01 2026-03-06T13:43:17.119 INFO:teuthology.orchestra.run.vm02.stdout:CentOS Stream 9 - AppStream 15 MB/s | 27 MB 00:01 2026-03-06T13:43:17.597 INFO:teuthology.orchestra.run.vm04.stdout:CentOS Stream 9 - AppStream 7.8 MB/s | 27 MB 00:03 2026-03-06T13:43:20.921 INFO:teuthology.orchestra.run.vm02.stdout:CentOS Stream 9 - CRB 7.6 MB/s | 8.0 MB 00:01 2026-03-06T13:43:21.058 INFO:teuthology.orchestra.run.vm04.stdout:CentOS Stream 9 - CRB 12 MB/s | 8.0 MB 00:00 2026-03-06T13:43:21.162 INFO:teuthology.orchestra.run.vm06.stdout:CentOS Stream 9 - CRB 9.7 MB/s | 8.0 MB 00:00 2026-03-06T13:43:22.406 INFO:teuthology.orchestra.run.vm06.stdout:CentOS Stream 9 - Extras packages 51 kB/s | 20 kB 00:00 2026-03-06T13:43:22.569 INFO:teuthology.orchestra.run.vm02.stdout:CentOS Stream 9 - Extras packages 25 kB/s | 20 kB 00:00 2026-03-06T13:43:22.645 INFO:teuthology.orchestra.run.vm04.stdout:CentOS Stream 9 - Extras packages 30 kB/s | 20 kB 00:00 2026-03-06T13:43:23.234 INFO:teuthology.orchestra.run.vm04.stdout:Extra Packages for Enterprise Linux 40 MB/s | 20 MB 00:00 2026-03-06T13:43:23.323 INFO:teuthology.orchestra.run.vm06.stdout:Extra Packages for Enterprise Linux 25 MB/s | 20 MB 00:00 2026-03-06T13:43:23.572 INFO:teuthology.orchestra.run.vm02.stdout:Extra Packages for Enterprise Linux 22 MB/s | 20 MB 00:00 2026-03-06T13:43:28.139 INFO:teuthology.orchestra.run.vm04.stdout:lab-extras 64 kB/s | 50 kB 00:00 2026-03-06T13:43:28.141 INFO:teuthology.orchestra.run.vm06.stdout:lab-extras 65 kB/s | 50 kB 00:00 2026-03-06T13:43:28.323 INFO:teuthology.orchestra.run.vm02.stdout:lab-extras 65 kB/s | 50 kB 00:00 2026-03-06T13:43:29.584 INFO:teuthology.orchestra.run.vm04.stdout:Package librados2-2:16.2.4-5.el9.x86_64 is already installed. 2026-03-06T13:43:29.585 INFO:teuthology.orchestra.run.vm04.stdout:Package librbd1-2:16.2.4-5.el9.x86_64 is already installed. 2026-03-06T13:43:29.620 INFO:teuthology.orchestra.run.vm04.stdout:Dependencies resolved. 2026-03-06T13:43:29.624 INFO:teuthology.orchestra.run.vm04.stdout:============================================================================================== 2026-03-06T13:43:29.624 INFO:teuthology.orchestra.run.vm04.stdout: Package Arch Version Repository Size 2026-03-06T13:43:29.624 INFO:teuthology.orchestra.run.vm04.stdout:============================================================================================== 2026-03-06T13:43:29.624 INFO:teuthology.orchestra.run.vm04.stdout:Installing: 2026-03-06T13:43:29.624 INFO:teuthology.orchestra.run.vm04.stdout: bzip2 x86_64 1.0.8-11.el9 baseos 55 k 2026-03-06T13:43:29.624 INFO:teuthology.orchestra.run.vm04.stdout: ceph x86_64 2:19.2.3-47.gc24117fd552.el9.clyso ceph 6.5 k 2026-03-06T13:43:29.624 INFO:teuthology.orchestra.run.vm04.stdout: ceph-base x86_64 2:19.2.3-47.gc24117fd552.el9.clyso ceph 5.5 M 2026-03-06T13:43:29.624 INFO:teuthology.orchestra.run.vm04.stdout: ceph-fuse x86_64 2:19.2.3-47.gc24117fd552.el9.clyso ceph 1.1 M 2026-03-06T13:43:29.624 INFO:teuthology.orchestra.run.vm04.stdout: ceph-immutable-object-cache x86_64 2:19.2.3-47.gc24117fd552.el9.clyso ceph 145 k 2026-03-06T13:43:29.624 INFO:teuthology.orchestra.run.vm04.stdout: ceph-mgr x86_64 2:19.2.3-47.gc24117fd552.el9.clyso ceph 1.1 M 2026-03-06T13:43:29.624 INFO:teuthology.orchestra.run.vm04.stdout: ceph-mgr-cephadm noarch 2:19.2.3-47.gc24117fd552.el9.clyso ceph-noarch 150 k 2026-03-06T13:43:29.624 INFO:teuthology.orchestra.run.vm04.stdout: ceph-mgr-dashboard noarch 2:19.2.3-47.gc24117fd552.el9.clyso ceph-noarch 3.8 M 2026-03-06T13:43:29.624 INFO:teuthology.orchestra.run.vm04.stdout: ceph-mgr-diskprediction-local noarch 2:19.2.3-47.gc24117fd552.el9.clyso ceph-noarch 7.4 M 2026-03-06T13:43:29.624 INFO:teuthology.orchestra.run.vm04.stdout: ceph-mgr-rook noarch 2:19.2.3-47.gc24117fd552.el9.clyso ceph-noarch 49 k 2026-03-06T13:43:29.624 INFO:teuthology.orchestra.run.vm04.stdout: ceph-radosgw x86_64 2:19.2.3-47.gc24117fd552.el9.clyso ceph 11 M 2026-03-06T13:43:29.624 INFO:teuthology.orchestra.run.vm04.stdout: ceph-test x86_64 2:19.2.3-47.gc24117fd552.el9.clyso ceph 50 M 2026-03-06T13:43:29.625 INFO:teuthology.orchestra.run.vm04.stdout: ceph-volume noarch 2:19.2.3-47.gc24117fd552.el9.clyso ceph-noarch 299 k 2026-03-06T13:43:29.625 INFO:teuthology.orchestra.run.vm04.stdout: cephadm noarch 2:19.2.3-47.gc24117fd552.el9.clyso ceph-noarch 769 k 2026-03-06T13:43:29.625 INFO:teuthology.orchestra.run.vm04.stdout: libcephfs-devel x86_64 2:19.2.3-47.gc24117fd552.el9.clyso ceph 34 k 2026-03-06T13:43:29.625 INFO:teuthology.orchestra.run.vm04.stdout: libcephfs2 x86_64 2:19.2.3-47.gc24117fd552.el9.clyso ceph 998 k 2026-03-06T13:43:29.625 INFO:teuthology.orchestra.run.vm04.stdout: librados-devel x86_64 2:19.2.3-47.gc24117fd552.el9.clyso ceph 127 k 2026-03-06T13:43:29.625 INFO:teuthology.orchestra.run.vm04.stdout: perl-Test-Harness noarch 1:3.42-461.el9 appstream 295 k 2026-03-06T13:43:29.625 INFO:teuthology.orchestra.run.vm04.stdout: python3-cephfs x86_64 2:19.2.3-47.gc24117fd552.el9.clyso ceph 165 k 2026-03-06T13:43:29.625 INFO:teuthology.orchestra.run.vm04.stdout: python3-rados x86_64 2:19.2.3-47.gc24117fd552.el9.clyso ceph 322 k 2026-03-06T13:43:29.625 INFO:teuthology.orchestra.run.vm04.stdout: python3-rbd x86_64 2:19.2.3-47.gc24117fd552.el9.clyso ceph 303 k 2026-03-06T13:43:29.625 INFO:teuthology.orchestra.run.vm04.stdout: python3-rgw x86_64 2:19.2.3-47.gc24117fd552.el9.clyso ceph 100 k 2026-03-06T13:43:29.625 INFO:teuthology.orchestra.run.vm04.stdout: python3-xmltodict noarch 0.12.0-15.el9 epel 22 k 2026-03-06T13:43:29.625 INFO:teuthology.orchestra.run.vm04.stdout: rbd-fuse x86_64 2:19.2.3-47.gc24117fd552.el9.clyso ceph 85 k 2026-03-06T13:43:29.625 INFO:teuthology.orchestra.run.vm04.stdout: rbd-mirror x86_64 2:19.2.3-47.gc24117fd552.el9.clyso ceph 3.1 M 2026-03-06T13:43:29.625 INFO:teuthology.orchestra.run.vm04.stdout: rbd-nbd x86_64 2:19.2.3-47.gc24117fd552.el9.clyso ceph 171 k 2026-03-06T13:43:29.625 INFO:teuthology.orchestra.run.vm04.stdout: s3cmd noarch 2.4.0-1.el9 epel 206 k 2026-03-06T13:43:29.625 INFO:teuthology.orchestra.run.vm04.stdout:Upgrading: 2026-03-06T13:43:29.625 INFO:teuthology.orchestra.run.vm04.stdout: librados2 x86_64 2:19.2.3-47.gc24117fd552.el9.clyso ceph 3.4 M 2026-03-06T13:43:29.625 INFO:teuthology.orchestra.run.vm04.stdout: librbd1 x86_64 2:19.2.3-47.gc24117fd552.el9.clyso ceph 3.2 M 2026-03-06T13:43:29.625 INFO:teuthology.orchestra.run.vm04.stdout:Installing dependencies: 2026-03-06T13:43:29.625 INFO:teuthology.orchestra.run.vm04.stdout: abseil-cpp x86_64 20211102.0-4.el9 epel 551 k 2026-03-06T13:43:29.625 INFO:teuthology.orchestra.run.vm04.stdout: boost-program-options x86_64 1.75.0-13.el9 appstream 104 k 2026-03-06T13:43:29.625 INFO:teuthology.orchestra.run.vm04.stdout: ceph-common x86_64 2:19.2.3-47.gc24117fd552.el9.clyso ceph 22 M 2026-03-06T13:43:29.625 INFO:teuthology.orchestra.run.vm04.stdout: ceph-grafana-dashboards noarch 2:19.2.3-47.gc24117fd552.el9.clyso ceph-noarch 31 k 2026-03-06T13:43:29.625 INFO:teuthology.orchestra.run.vm04.stdout: ceph-mds x86_64 2:19.2.3-47.gc24117fd552.el9.clyso ceph 2.4 M 2026-03-06T13:43:29.625 INFO:teuthology.orchestra.run.vm04.stdout: ceph-mgr-modules-core noarch 2:19.2.3-47.gc24117fd552.el9.clyso ceph-noarch 252 k 2026-03-06T13:43:29.625 INFO:teuthology.orchestra.run.vm04.stdout: ceph-mon x86_64 2:19.2.3-47.gc24117fd552.el9.clyso ceph 4.7 M 2026-03-06T13:43:29.625 INFO:teuthology.orchestra.run.vm04.stdout: ceph-osd x86_64 2:19.2.3-47.gc24117fd552.el9.clyso ceph 17 M 2026-03-06T13:43:29.625 INFO:teuthology.orchestra.run.vm04.stdout: ceph-prometheus-alerts noarch 2:19.2.3-47.gc24117fd552.el9.clyso ceph-noarch 16 k 2026-03-06T13:43:29.625 INFO:teuthology.orchestra.run.vm04.stdout: ceph-selinux x86_64 2:19.2.3-47.gc24117fd552.el9.clyso ceph 25 k 2026-03-06T13:43:29.625 INFO:teuthology.orchestra.run.vm04.stdout: cryptsetup x86_64 2.8.1-3.el9 baseos 351 k 2026-03-06T13:43:29.625 INFO:teuthology.orchestra.run.vm04.stdout: flexiblas x86_64 3.0.4-9.el9 appstream 30 k 2026-03-06T13:43:29.625 INFO:teuthology.orchestra.run.vm04.stdout: flexiblas-netlib x86_64 3.0.4-9.el9 appstream 3.0 M 2026-03-06T13:43:29.625 INFO:teuthology.orchestra.run.vm04.stdout: flexiblas-openblas-openmp x86_64 3.0.4-9.el9 appstream 15 k 2026-03-06T13:43:29.625 INFO:teuthology.orchestra.run.vm04.stdout: fuse x86_64 2.9.9-17.el9 baseos 80 k 2026-03-06T13:43:29.625 INFO:teuthology.orchestra.run.vm04.stdout: gperftools-libs x86_64 2.9.1-3.el9 epel 308 k 2026-03-06T13:43:29.625 INFO:teuthology.orchestra.run.vm04.stdout: grpc-data noarch 1.46.7-10.el9 epel 19 k 2026-03-06T13:43:29.625 INFO:teuthology.orchestra.run.vm04.stdout: ledmon-libs x86_64 1.1.0-3.el9 baseos 40 k 2026-03-06T13:43:29.625 INFO:teuthology.orchestra.run.vm04.stdout: libarrow x86_64 9.0.0-15.el9 epel 4.4 M 2026-03-06T13:43:29.625 INFO:teuthology.orchestra.run.vm04.stdout: libarrow-doc noarch 9.0.0-15.el9 epel 25 k 2026-03-06T13:43:29.625 INFO:teuthology.orchestra.run.vm04.stdout: libcephsqlite x86_64 2:19.2.3-47.gc24117fd552.el9.clyso ceph 163 k 2026-03-06T13:43:29.625 INFO:teuthology.orchestra.run.vm04.stdout: libconfig x86_64 1.7.2-9.el9 baseos 72 k 2026-03-06T13:43:29.625 INFO:teuthology.orchestra.run.vm04.stdout: libgfortran x86_64 11.5.0-14.el9 baseos 794 k 2026-03-06T13:43:29.625 INFO:teuthology.orchestra.run.vm04.stdout: libnbd x86_64 1.20.3-4.el9 appstream 164 k 2026-03-06T13:43:29.625 INFO:teuthology.orchestra.run.vm04.stdout: liboath x86_64 2.6.12-1.el9 epel 49 k 2026-03-06T13:43:29.625 INFO:teuthology.orchestra.run.vm04.stdout: libpmemobj x86_64 1.12.1-1.el9 appstream 160 k 2026-03-06T13:43:29.625 INFO:teuthology.orchestra.run.vm04.stdout: libquadmath x86_64 11.5.0-14.el9 baseos 184 k 2026-03-06T13:43:29.625 INFO:teuthology.orchestra.run.vm04.stdout: librabbitmq x86_64 0.11.0-7.el9 appstream 45 k 2026-03-06T13:43:29.625 INFO:teuthology.orchestra.run.vm04.stdout: libradosstriper1 x86_64 2:19.2.3-47.gc24117fd552.el9.clyso ceph 503 k 2026-03-06T13:43:29.625 INFO:teuthology.orchestra.run.vm04.stdout: librdkafka x86_64 1.6.1-102.el9 appstream 662 k 2026-03-06T13:43:29.625 INFO:teuthology.orchestra.run.vm04.stdout: librgw2 x86_64 2:19.2.3-47.gc24117fd552.el9.clyso ceph 5.4 M 2026-03-06T13:43:29.625 INFO:teuthology.orchestra.run.vm04.stdout: libstoragemgmt x86_64 1.10.1-1.el9 appstream 246 k 2026-03-06T13:43:29.625 INFO:teuthology.orchestra.run.vm04.stdout: libunwind x86_64 1.6.2-1.el9 epel 67 k 2026-03-06T13:43:29.625 INFO:teuthology.orchestra.run.vm04.stdout: libxslt x86_64 1.1.34-12.el9 appstream 233 k 2026-03-06T13:43:29.626 INFO:teuthology.orchestra.run.vm04.stdout: lttng-ust x86_64 2.12.0-6.el9 appstream 292 k 2026-03-06T13:43:29.626 INFO:teuthology.orchestra.run.vm04.stdout: lua x86_64 5.4.4-4.el9 appstream 188 k 2026-03-06T13:43:29.626 INFO:teuthology.orchestra.run.vm04.stdout: lua-devel x86_64 5.4.4-4.el9 crb 22 k 2026-03-06T13:43:29.626 INFO:teuthology.orchestra.run.vm04.stdout: luarocks noarch 3.9.2-5.el9 epel 151 k 2026-03-06T13:43:29.626 INFO:teuthology.orchestra.run.vm04.stdout: mailcap noarch 2.1.49-5.el9 baseos 33 k 2026-03-06T13:43:29.626 INFO:teuthology.orchestra.run.vm04.stdout: openblas x86_64 0.3.29-1.el9 appstream 42 k 2026-03-06T13:43:29.626 INFO:teuthology.orchestra.run.vm04.stdout: openblas-openmp x86_64 0.3.29-1.el9 appstream 5.3 M 2026-03-06T13:43:29.626 INFO:teuthology.orchestra.run.vm04.stdout: parquet-libs x86_64 9.0.0-15.el9 epel 838 k 2026-03-06T13:43:29.626 INFO:teuthology.orchestra.run.vm04.stdout: pciutils x86_64 3.7.0-7.el9 baseos 93 k 2026-03-06T13:43:29.626 INFO:teuthology.orchestra.run.vm04.stdout: perl-Benchmark noarch 1.23-483.el9 appstream 26 k 2026-03-06T13:43:29.626 INFO:teuthology.orchestra.run.vm04.stdout: protobuf x86_64 3.14.0-17.el9 appstream 1.0 M 2026-03-06T13:43:29.626 INFO:teuthology.orchestra.run.vm04.stdout: protobuf-compiler x86_64 3.14.0-17.el9 crb 862 k 2026-03-06T13:43:29.626 INFO:teuthology.orchestra.run.vm04.stdout: python3-asyncssh noarch 2.13.2-5.el9 epel 548 k 2026-03-06T13:43:29.626 INFO:teuthology.orchestra.run.vm04.stdout: python3-autocommand noarch 2.2.2-8.el9 epel 29 k 2026-03-06T13:43:29.626 INFO:teuthology.orchestra.run.vm04.stdout: python3-babel noarch 2.9.1-2.el9 appstream 6.0 M 2026-03-06T13:43:29.626 INFO:teuthology.orchestra.run.vm04.stdout: python3-backports-tarfile noarch 1.2.0-1.el9 epel 60 k 2026-03-06T13:43:29.626 INFO:teuthology.orchestra.run.vm04.stdout: python3-bcrypt x86_64 3.2.2-1.el9 epel 43 k 2026-03-06T13:43:29.626 INFO:teuthology.orchestra.run.vm04.stdout: python3-cachetools noarch 4.2.4-1.el9 epel 32 k 2026-03-06T13:43:29.626 INFO:teuthology.orchestra.run.vm04.stdout: python3-ceph-argparse x86_64 2:19.2.3-47.gc24117fd552.el9.clyso ceph 45 k 2026-03-06T13:43:29.626 INFO:teuthology.orchestra.run.vm04.stdout: python3-ceph-common x86_64 2:19.2.3-47.gc24117fd552.el9.clyso ceph 142 k 2026-03-06T13:43:29.626 INFO:teuthology.orchestra.run.vm04.stdout: python3-certifi noarch 2023.05.07-4.el9 epel 14 k 2026-03-06T13:43:29.626 INFO:teuthology.orchestra.run.vm04.stdout: python3-cffi x86_64 1.14.5-5.el9 baseos 253 k 2026-03-06T13:43:29.626 INFO:teuthology.orchestra.run.vm04.stdout: python3-cheroot noarch 10.0.1-4.el9 epel 173 k 2026-03-06T13:43:29.626 INFO:teuthology.orchestra.run.vm04.stdout: python3-cherrypy noarch 18.6.1-2.el9 epel 358 k 2026-03-06T13:43:29.626 INFO:teuthology.orchestra.run.vm04.stdout: python3-cryptography x86_64 36.0.1-5.el9 baseos 1.2 M 2026-03-06T13:43:29.626 INFO:teuthology.orchestra.run.vm04.stdout: python3-devel x86_64 3.9.25-3.el9 appstream 244 k 2026-03-06T13:43:29.626 INFO:teuthology.orchestra.run.vm04.stdout: python3-google-auth noarch 1:2.45.0-1.el9 epel 254 k 2026-03-06T13:43:29.626 INFO:teuthology.orchestra.run.vm04.stdout: python3-grpcio x86_64 1.46.7-10.el9 epel 2.0 M 2026-03-06T13:43:29.626 INFO:teuthology.orchestra.run.vm04.stdout: python3-grpcio-tools x86_64 1.46.7-10.el9 epel 144 k 2026-03-06T13:43:29.626 INFO:teuthology.orchestra.run.vm04.stdout: python3-isodate noarch 0.6.1-3.el9 epel 56 k 2026-03-06T13:43:29.626 INFO:teuthology.orchestra.run.vm04.stdout: python3-jaraco noarch 8.2.1-3.el9 epel 11 k 2026-03-06T13:43:29.626 INFO:teuthology.orchestra.run.vm04.stdout: python3-jaraco-classes noarch 3.2.1-5.el9 epel 18 k 2026-03-06T13:43:29.626 INFO:teuthology.orchestra.run.vm04.stdout: python3-jaraco-collections noarch 3.0.0-8.el9 epel 23 k 2026-03-06T13:43:29.626 INFO:teuthology.orchestra.run.vm04.stdout: python3-jaraco-context noarch 6.0.1-3.el9 epel 20 k 2026-03-06T13:43:29.626 INFO:teuthology.orchestra.run.vm04.stdout: python3-jaraco-functools noarch 3.5.0-2.el9 epel 19 k 2026-03-06T13:43:29.626 INFO:teuthology.orchestra.run.vm04.stdout: python3-jaraco-text noarch 4.0.0-2.el9 epel 26 k 2026-03-06T13:43:29.626 INFO:teuthology.orchestra.run.vm04.stdout: python3-jinja2 noarch 2.11.3-8.el9 appstream 249 k 2026-03-06T13:43:29.626 INFO:teuthology.orchestra.run.vm04.stdout: python3-kubernetes noarch 1:26.1.0-3.el9 epel 1.0 M 2026-03-06T13:43:29.626 INFO:teuthology.orchestra.run.vm04.stdout: python3-libstoragemgmt x86_64 1.10.1-1.el9 appstream 177 k 2026-03-06T13:43:29.626 INFO:teuthology.orchestra.run.vm04.stdout: python3-logutils noarch 0.3.5-21.el9 epel 46 k 2026-03-06T13:43:29.626 INFO:teuthology.orchestra.run.vm04.stdout: python3-lxml x86_64 4.6.5-3.el9 appstream 1.2 M 2026-03-06T13:43:29.626 INFO:teuthology.orchestra.run.vm04.stdout: python3-mako noarch 1.1.4-6.el9 appstream 172 k 2026-03-06T13:43:29.626 INFO:teuthology.orchestra.run.vm04.stdout: python3-markupsafe x86_64 1.1.1-12.el9 appstream 35 k 2026-03-06T13:43:29.626 INFO:teuthology.orchestra.run.vm04.stdout: python3-more-itertools noarch 8.12.0-2.el9 epel 79 k 2026-03-06T13:43:29.626 INFO:teuthology.orchestra.run.vm04.stdout: python3-msgpack x86_64 1.0.3-2.el9 epel 86 k 2026-03-06T13:43:29.626 INFO:teuthology.orchestra.run.vm04.stdout: python3-natsort noarch 7.1.1-5.el9 epel 58 k 2026-03-06T13:43:29.626 INFO:teuthology.orchestra.run.vm04.stdout: python3-numpy x86_64 1:1.23.5-2.el9 appstream 6.1 M 2026-03-06T13:43:29.626 INFO:teuthology.orchestra.run.vm04.stdout: python3-numpy-f2py x86_64 1:1.23.5-2.el9 appstream 442 k 2026-03-06T13:43:29.626 INFO:teuthology.orchestra.run.vm04.stdout: python3-packaging noarch 20.9-5.el9 appstream 77 k 2026-03-06T13:43:29.626 INFO:teuthology.orchestra.run.vm04.stdout: python3-pecan noarch 1.4.2-3.el9 epel 272 k 2026-03-06T13:43:29.627 INFO:teuthology.orchestra.run.vm04.stdout: python3-ply noarch 3.11-14.el9 baseos 106 k 2026-03-06T13:43:29.627 INFO:teuthology.orchestra.run.vm04.stdout: python3-portend noarch 3.1.0-2.el9 epel 16 k 2026-03-06T13:43:29.627 INFO:teuthology.orchestra.run.vm04.stdout: python3-protobuf noarch 3.14.0-17.el9 appstream 267 k 2026-03-06T13:43:29.627 INFO:teuthology.orchestra.run.vm04.stdout: python3-pyOpenSSL noarch 21.0.0-1.el9 epel 90 k 2026-03-06T13:43:29.627 INFO:teuthology.orchestra.run.vm04.stdout: python3-pyasn1 noarch 0.4.8-7.el9 appstream 157 k 2026-03-06T13:43:29.627 INFO:teuthology.orchestra.run.vm04.stdout: python3-pyasn1-modules noarch 0.4.8-7.el9 appstream 277 k 2026-03-06T13:43:29.627 INFO:teuthology.orchestra.run.vm04.stdout: python3-pycparser noarch 2.20-6.el9 baseos 135 k 2026-03-06T13:43:29.627 INFO:teuthology.orchestra.run.vm04.stdout: python3-repoze-lru noarch 0.7-16.el9 epel 31 k 2026-03-06T13:43:29.627 INFO:teuthology.orchestra.run.vm04.stdout: python3-requests noarch 2.25.1-10.el9 baseos 126 k 2026-03-06T13:43:29.627 INFO:teuthology.orchestra.run.vm04.stdout: python3-requests-oauthlib noarch 1.3.0-12.el9 appstream 54 k 2026-03-06T13:43:29.627 INFO:teuthology.orchestra.run.vm04.stdout: python3-routes noarch 2.5.1-5.el9 epel 188 k 2026-03-06T13:43:29.627 INFO:teuthology.orchestra.run.vm04.stdout: python3-rsa noarch 4.9-2.el9 epel 59 k 2026-03-06T13:43:29.627 INFO:teuthology.orchestra.run.vm04.stdout: python3-scipy x86_64 1.9.3-2.el9 appstream 19 M 2026-03-06T13:43:29.627 INFO:teuthology.orchestra.run.vm04.stdout: python3-tempora noarch 5.0.0-2.el9 epel 36 k 2026-03-06T13:43:29.627 INFO:teuthology.orchestra.run.vm04.stdout: python3-toml noarch 0.10.2-6.el9 appstream 42 k 2026-03-06T13:43:29.627 INFO:teuthology.orchestra.run.vm04.stdout: python3-typing-extensions noarch 4.15.0-1.el9 epel 86 k 2026-03-06T13:43:29.627 INFO:teuthology.orchestra.run.vm04.stdout: python3-urllib3 noarch 1.26.5-7.el9 baseos 218 k 2026-03-06T13:43:29.627 INFO:teuthology.orchestra.run.vm04.stdout: python3-webob noarch 1.8.8-2.el9 epel 230 k 2026-03-06T13:43:29.627 INFO:teuthology.orchestra.run.vm04.stdout: python3-websocket-client noarch 1.2.3-2.el9 epel 90 k 2026-03-06T13:43:29.627 INFO:teuthology.orchestra.run.vm04.stdout: python3-werkzeug noarch 2.0.3-3.el9.1 epel 427 k 2026-03-06T13:43:29.627 INFO:teuthology.orchestra.run.vm04.stdout: python3-xmlsec x86_64 1.3.13-1.el9 epel 48 k 2026-03-06T13:43:29.627 INFO:teuthology.orchestra.run.vm04.stdout: python3-zc-lockfile noarch 2.0-10.el9 epel 20 k 2026-03-06T13:43:29.627 INFO:teuthology.orchestra.run.vm04.stdout: qatlib x86_64 25.08.0-2.el9 appstream 240 k 2026-03-06T13:43:29.627 INFO:teuthology.orchestra.run.vm04.stdout: qatzip-libs x86_64 1.3.1-1.el9 appstream 66 k 2026-03-06T13:43:29.627 INFO:teuthology.orchestra.run.vm04.stdout: re2 x86_64 1:20211101-20.el9 epel 191 k 2026-03-06T13:43:29.627 INFO:teuthology.orchestra.run.vm04.stdout: socat x86_64 1.7.4.1-8.el9 appstream 303 k 2026-03-06T13:43:29.627 INFO:teuthology.orchestra.run.vm04.stdout: thrift x86_64 0.15.0-4.el9 epel 1.6 M 2026-03-06T13:43:29.627 INFO:teuthology.orchestra.run.vm04.stdout: unzip x86_64 6.0-59.el9 baseos 182 k 2026-03-06T13:43:29.627 INFO:teuthology.orchestra.run.vm04.stdout: xmlsec1 x86_64 1.2.29-13.el9 appstream 189 k 2026-03-06T13:43:29.627 INFO:teuthology.orchestra.run.vm04.stdout: xmlsec1-openssl x86_64 1.2.29-13.el9 appstream 90 k 2026-03-06T13:43:29.627 INFO:teuthology.orchestra.run.vm04.stdout: xmlstarlet x86_64 1.6.1-20.el9 appstream 64 k 2026-03-06T13:43:29.627 INFO:teuthology.orchestra.run.vm04.stdout: zip x86_64 3.0-35.el9 baseos 266 k 2026-03-06T13:43:29.627 INFO:teuthology.orchestra.run.vm04.stdout:Installing weak dependencies: 2026-03-06T13:43:29.627 INFO:teuthology.orchestra.run.vm04.stdout: ceph-mgr-k8sevents noarch 2:19.2.3-47.gc24117fd552.el9.clyso ceph-noarch 22 k 2026-03-06T13:43:29.627 INFO:teuthology.orchestra.run.vm04.stdout: python3-influxdb noarch 5.3.1-1.el9 epel 139 k 2026-03-06T13:43:29.627 INFO:teuthology.orchestra.run.vm04.stdout: python3-saml noarch 1.16.0-1.el9 epel 125 k 2026-03-06T13:43:29.627 INFO:teuthology.orchestra.run.vm04.stdout: qatlib-service x86_64 25.08.0-2.el9 appstream 37 k 2026-03-06T13:43:29.627 INFO:teuthology.orchestra.run.vm04.stdout: smartmontools x86_64 1:7.2-10.el9 baseos 556 k 2026-03-06T13:43:29.627 INFO:teuthology.orchestra.run.vm04.stdout: 2026-03-06T13:43:29.627 INFO:teuthology.orchestra.run.vm04.stdout:Transaction Summary 2026-03-06T13:43:29.627 INFO:teuthology.orchestra.run.vm04.stdout:============================================================================================== 2026-03-06T13:43:29.627 INFO:teuthology.orchestra.run.vm04.stdout:Install 148 Packages 2026-03-06T13:43:29.627 INFO:teuthology.orchestra.run.vm04.stdout:Upgrade 2 Packages 2026-03-06T13:43:29.627 INFO:teuthology.orchestra.run.vm04.stdout: 2026-03-06T13:43:29.628 INFO:teuthology.orchestra.run.vm04.stdout:Total download size: 214 M 2026-03-06T13:43:29.628 INFO:teuthology.orchestra.run.vm04.stdout:Downloading Packages: 2026-03-06T13:43:29.639 INFO:teuthology.orchestra.run.vm06.stdout:Package librados2-2:16.2.4-5.el9.x86_64 is already installed. 2026-03-06T13:43:29.640 INFO:teuthology.orchestra.run.vm06.stdout:Package librbd1-2:16.2.4-5.el9.x86_64 is already installed. 2026-03-06T13:43:29.673 INFO:teuthology.orchestra.run.vm06.stdout:Dependencies resolved. 2026-03-06T13:43:29.677 INFO:teuthology.orchestra.run.vm06.stdout:============================================================================================== 2026-03-06T13:43:29.677 INFO:teuthology.orchestra.run.vm06.stdout: Package Arch Version Repository Size 2026-03-06T13:43:29.677 INFO:teuthology.orchestra.run.vm06.stdout:============================================================================================== 2026-03-06T13:43:29.678 INFO:teuthology.orchestra.run.vm06.stdout:Installing: 2026-03-06T13:43:29.678 INFO:teuthology.orchestra.run.vm06.stdout: bzip2 x86_64 1.0.8-11.el9 baseos 55 k 2026-03-06T13:43:29.678 INFO:teuthology.orchestra.run.vm06.stdout: ceph x86_64 2:19.2.3-47.gc24117fd552.el9.clyso ceph 6.5 k 2026-03-06T13:43:29.678 INFO:teuthology.orchestra.run.vm06.stdout: ceph-base x86_64 2:19.2.3-47.gc24117fd552.el9.clyso ceph 5.5 M 2026-03-06T13:43:29.678 INFO:teuthology.orchestra.run.vm06.stdout: ceph-fuse x86_64 2:19.2.3-47.gc24117fd552.el9.clyso ceph 1.1 M 2026-03-06T13:43:29.678 INFO:teuthology.orchestra.run.vm06.stdout: ceph-immutable-object-cache x86_64 2:19.2.3-47.gc24117fd552.el9.clyso ceph 145 k 2026-03-06T13:43:29.678 INFO:teuthology.orchestra.run.vm06.stdout: ceph-mgr x86_64 2:19.2.3-47.gc24117fd552.el9.clyso ceph 1.1 M 2026-03-06T13:43:29.678 INFO:teuthology.orchestra.run.vm06.stdout: ceph-mgr-cephadm noarch 2:19.2.3-47.gc24117fd552.el9.clyso ceph-noarch 150 k 2026-03-06T13:43:29.678 INFO:teuthology.orchestra.run.vm06.stdout: ceph-mgr-dashboard noarch 2:19.2.3-47.gc24117fd552.el9.clyso ceph-noarch 3.8 M 2026-03-06T13:43:29.678 INFO:teuthology.orchestra.run.vm06.stdout: ceph-mgr-diskprediction-local noarch 2:19.2.3-47.gc24117fd552.el9.clyso ceph-noarch 7.4 M 2026-03-06T13:43:29.678 INFO:teuthology.orchestra.run.vm06.stdout: ceph-mgr-rook noarch 2:19.2.3-47.gc24117fd552.el9.clyso ceph-noarch 49 k 2026-03-06T13:43:29.678 INFO:teuthology.orchestra.run.vm06.stdout: ceph-radosgw x86_64 2:19.2.3-47.gc24117fd552.el9.clyso ceph 11 M 2026-03-06T13:43:29.678 INFO:teuthology.orchestra.run.vm06.stdout: ceph-test x86_64 2:19.2.3-47.gc24117fd552.el9.clyso ceph 50 M 2026-03-06T13:43:29.678 INFO:teuthology.orchestra.run.vm06.stdout: ceph-volume noarch 2:19.2.3-47.gc24117fd552.el9.clyso ceph-noarch 299 k 2026-03-06T13:43:29.678 INFO:teuthology.orchestra.run.vm06.stdout: cephadm noarch 2:19.2.3-47.gc24117fd552.el9.clyso ceph-noarch 769 k 2026-03-06T13:43:29.678 INFO:teuthology.orchestra.run.vm06.stdout: libcephfs-devel x86_64 2:19.2.3-47.gc24117fd552.el9.clyso ceph 34 k 2026-03-06T13:43:29.678 INFO:teuthology.orchestra.run.vm06.stdout: libcephfs2 x86_64 2:19.2.3-47.gc24117fd552.el9.clyso ceph 998 k 2026-03-06T13:43:29.678 INFO:teuthology.orchestra.run.vm06.stdout: librados-devel x86_64 2:19.2.3-47.gc24117fd552.el9.clyso ceph 127 k 2026-03-06T13:43:29.678 INFO:teuthology.orchestra.run.vm06.stdout: perl-Test-Harness noarch 1:3.42-461.el9 appstream 295 k 2026-03-06T13:43:29.678 INFO:teuthology.orchestra.run.vm06.stdout: python3-cephfs x86_64 2:19.2.3-47.gc24117fd552.el9.clyso ceph 165 k 2026-03-06T13:43:29.678 INFO:teuthology.orchestra.run.vm06.stdout: python3-rados x86_64 2:19.2.3-47.gc24117fd552.el9.clyso ceph 322 k 2026-03-06T13:43:29.678 INFO:teuthology.orchestra.run.vm06.stdout: python3-rbd x86_64 2:19.2.3-47.gc24117fd552.el9.clyso ceph 303 k 2026-03-06T13:43:29.678 INFO:teuthology.orchestra.run.vm06.stdout: python3-rgw x86_64 2:19.2.3-47.gc24117fd552.el9.clyso ceph 100 k 2026-03-06T13:43:29.678 INFO:teuthology.orchestra.run.vm06.stdout: python3-xmltodict noarch 0.12.0-15.el9 epel 22 k 2026-03-06T13:43:29.678 INFO:teuthology.orchestra.run.vm06.stdout: rbd-fuse x86_64 2:19.2.3-47.gc24117fd552.el9.clyso ceph 85 k 2026-03-06T13:43:29.678 INFO:teuthology.orchestra.run.vm06.stdout: rbd-mirror x86_64 2:19.2.3-47.gc24117fd552.el9.clyso ceph 3.1 M 2026-03-06T13:43:29.678 INFO:teuthology.orchestra.run.vm06.stdout: rbd-nbd x86_64 2:19.2.3-47.gc24117fd552.el9.clyso ceph 171 k 2026-03-06T13:43:29.678 INFO:teuthology.orchestra.run.vm06.stdout: s3cmd noarch 2.4.0-1.el9 epel 206 k 2026-03-06T13:43:29.678 INFO:teuthology.orchestra.run.vm06.stdout:Upgrading: 2026-03-06T13:43:29.678 INFO:teuthology.orchestra.run.vm06.stdout: librados2 x86_64 2:19.2.3-47.gc24117fd552.el9.clyso ceph 3.4 M 2026-03-06T13:43:29.678 INFO:teuthology.orchestra.run.vm06.stdout: librbd1 x86_64 2:19.2.3-47.gc24117fd552.el9.clyso ceph 3.2 M 2026-03-06T13:43:29.678 INFO:teuthology.orchestra.run.vm06.stdout:Installing dependencies: 2026-03-06T13:43:29.678 INFO:teuthology.orchestra.run.vm06.stdout: abseil-cpp x86_64 20211102.0-4.el9 epel 551 k 2026-03-06T13:43:29.678 INFO:teuthology.orchestra.run.vm06.stdout: boost-program-options x86_64 1.75.0-13.el9 appstream 104 k 2026-03-06T13:43:29.678 INFO:teuthology.orchestra.run.vm06.stdout: ceph-common x86_64 2:19.2.3-47.gc24117fd552.el9.clyso ceph 22 M 2026-03-06T13:43:29.678 INFO:teuthology.orchestra.run.vm06.stdout: ceph-grafana-dashboards noarch 2:19.2.3-47.gc24117fd552.el9.clyso ceph-noarch 31 k 2026-03-06T13:43:29.678 INFO:teuthology.orchestra.run.vm06.stdout: ceph-mds x86_64 2:19.2.3-47.gc24117fd552.el9.clyso ceph 2.4 M 2026-03-06T13:43:29.678 INFO:teuthology.orchestra.run.vm06.stdout: ceph-mgr-modules-core noarch 2:19.2.3-47.gc24117fd552.el9.clyso ceph-noarch 252 k 2026-03-06T13:43:29.678 INFO:teuthology.orchestra.run.vm06.stdout: ceph-mon x86_64 2:19.2.3-47.gc24117fd552.el9.clyso ceph 4.7 M 2026-03-06T13:43:29.678 INFO:teuthology.orchestra.run.vm06.stdout: ceph-osd x86_64 2:19.2.3-47.gc24117fd552.el9.clyso ceph 17 M 2026-03-06T13:43:29.678 INFO:teuthology.orchestra.run.vm06.stdout: ceph-prometheus-alerts noarch 2:19.2.3-47.gc24117fd552.el9.clyso ceph-noarch 16 k 2026-03-06T13:43:29.678 INFO:teuthology.orchestra.run.vm06.stdout: ceph-selinux x86_64 2:19.2.3-47.gc24117fd552.el9.clyso ceph 25 k 2026-03-06T13:43:29.678 INFO:teuthology.orchestra.run.vm06.stdout: cryptsetup x86_64 2.8.1-3.el9 baseos 351 k 2026-03-06T13:43:29.678 INFO:teuthology.orchestra.run.vm06.stdout: flexiblas x86_64 3.0.4-9.el9 appstream 30 k 2026-03-06T13:43:29.678 INFO:teuthology.orchestra.run.vm06.stdout: flexiblas-netlib x86_64 3.0.4-9.el9 appstream 3.0 M 2026-03-06T13:43:29.678 INFO:teuthology.orchestra.run.vm06.stdout: flexiblas-openblas-openmp x86_64 3.0.4-9.el9 appstream 15 k 2026-03-06T13:43:29.678 INFO:teuthology.orchestra.run.vm06.stdout: fuse x86_64 2.9.9-17.el9 baseos 80 k 2026-03-06T13:43:29.678 INFO:teuthology.orchestra.run.vm06.stdout: gperftools-libs x86_64 2.9.1-3.el9 epel 308 k 2026-03-06T13:43:29.678 INFO:teuthology.orchestra.run.vm06.stdout: grpc-data noarch 1.46.7-10.el9 epel 19 k 2026-03-06T13:43:29.679 INFO:teuthology.orchestra.run.vm06.stdout: ledmon-libs x86_64 1.1.0-3.el9 baseos 40 k 2026-03-06T13:43:29.679 INFO:teuthology.orchestra.run.vm06.stdout: libarrow x86_64 9.0.0-15.el9 epel 4.4 M 2026-03-06T13:43:29.679 INFO:teuthology.orchestra.run.vm06.stdout: libarrow-doc noarch 9.0.0-15.el9 epel 25 k 2026-03-06T13:43:29.679 INFO:teuthology.orchestra.run.vm06.stdout: libcephsqlite x86_64 2:19.2.3-47.gc24117fd552.el9.clyso ceph 163 k 2026-03-06T13:43:29.679 INFO:teuthology.orchestra.run.vm06.stdout: libconfig x86_64 1.7.2-9.el9 baseos 72 k 2026-03-06T13:43:29.679 INFO:teuthology.orchestra.run.vm06.stdout: libgfortran x86_64 11.5.0-14.el9 baseos 794 k 2026-03-06T13:43:29.679 INFO:teuthology.orchestra.run.vm06.stdout: libnbd x86_64 1.20.3-4.el9 appstream 164 k 2026-03-06T13:43:29.679 INFO:teuthology.orchestra.run.vm06.stdout: liboath x86_64 2.6.12-1.el9 epel 49 k 2026-03-06T13:43:29.679 INFO:teuthology.orchestra.run.vm06.stdout: libpmemobj x86_64 1.12.1-1.el9 appstream 160 k 2026-03-06T13:43:29.679 INFO:teuthology.orchestra.run.vm06.stdout: libquadmath x86_64 11.5.0-14.el9 baseos 184 k 2026-03-06T13:43:29.679 INFO:teuthology.orchestra.run.vm06.stdout: librabbitmq x86_64 0.11.0-7.el9 appstream 45 k 2026-03-06T13:43:29.679 INFO:teuthology.orchestra.run.vm06.stdout: libradosstriper1 x86_64 2:19.2.3-47.gc24117fd552.el9.clyso ceph 503 k 2026-03-06T13:43:29.679 INFO:teuthology.orchestra.run.vm06.stdout: librdkafka x86_64 1.6.1-102.el9 appstream 662 k 2026-03-06T13:43:29.679 INFO:teuthology.orchestra.run.vm06.stdout: librgw2 x86_64 2:19.2.3-47.gc24117fd552.el9.clyso ceph 5.4 M 2026-03-06T13:43:29.679 INFO:teuthology.orchestra.run.vm06.stdout: libstoragemgmt x86_64 1.10.1-1.el9 appstream 246 k 2026-03-06T13:43:29.679 INFO:teuthology.orchestra.run.vm06.stdout: libunwind x86_64 1.6.2-1.el9 epel 67 k 2026-03-06T13:43:29.679 INFO:teuthology.orchestra.run.vm06.stdout: libxslt x86_64 1.1.34-12.el9 appstream 233 k 2026-03-06T13:43:29.679 INFO:teuthology.orchestra.run.vm06.stdout: lttng-ust x86_64 2.12.0-6.el9 appstream 292 k 2026-03-06T13:43:29.679 INFO:teuthology.orchestra.run.vm06.stdout: lua x86_64 5.4.4-4.el9 appstream 188 k 2026-03-06T13:43:29.679 INFO:teuthology.orchestra.run.vm06.stdout: lua-devel x86_64 5.4.4-4.el9 crb 22 k 2026-03-06T13:43:29.679 INFO:teuthology.orchestra.run.vm06.stdout: luarocks noarch 3.9.2-5.el9 epel 151 k 2026-03-06T13:43:29.679 INFO:teuthology.orchestra.run.vm06.stdout: mailcap noarch 2.1.49-5.el9 baseos 33 k 2026-03-06T13:43:29.679 INFO:teuthology.orchestra.run.vm06.stdout: openblas x86_64 0.3.29-1.el9 appstream 42 k 2026-03-06T13:43:29.679 INFO:teuthology.orchestra.run.vm06.stdout: openblas-openmp x86_64 0.3.29-1.el9 appstream 5.3 M 2026-03-06T13:43:29.679 INFO:teuthology.orchestra.run.vm06.stdout: parquet-libs x86_64 9.0.0-15.el9 epel 838 k 2026-03-06T13:43:29.679 INFO:teuthology.orchestra.run.vm06.stdout: pciutils x86_64 3.7.0-7.el9 baseos 93 k 2026-03-06T13:43:29.679 INFO:teuthology.orchestra.run.vm06.stdout: perl-Benchmark noarch 1.23-483.el9 appstream 26 k 2026-03-06T13:43:29.679 INFO:teuthology.orchestra.run.vm06.stdout: protobuf x86_64 3.14.0-17.el9 appstream 1.0 M 2026-03-06T13:43:29.679 INFO:teuthology.orchestra.run.vm06.stdout: protobuf-compiler x86_64 3.14.0-17.el9 crb 862 k 2026-03-06T13:43:29.679 INFO:teuthology.orchestra.run.vm06.stdout: python3-asyncssh noarch 2.13.2-5.el9 epel 548 k 2026-03-06T13:43:29.679 INFO:teuthology.orchestra.run.vm06.stdout: python3-autocommand noarch 2.2.2-8.el9 epel 29 k 2026-03-06T13:43:29.679 INFO:teuthology.orchestra.run.vm06.stdout: python3-babel noarch 2.9.1-2.el9 appstream 6.0 M 2026-03-06T13:43:29.679 INFO:teuthology.orchestra.run.vm06.stdout: python3-backports-tarfile noarch 1.2.0-1.el9 epel 60 k 2026-03-06T13:43:29.679 INFO:teuthology.orchestra.run.vm06.stdout: python3-bcrypt x86_64 3.2.2-1.el9 epel 43 k 2026-03-06T13:43:29.679 INFO:teuthology.orchestra.run.vm06.stdout: python3-cachetools noarch 4.2.4-1.el9 epel 32 k 2026-03-06T13:43:29.679 INFO:teuthology.orchestra.run.vm06.stdout: python3-ceph-argparse x86_64 2:19.2.3-47.gc24117fd552.el9.clyso ceph 45 k 2026-03-06T13:43:29.679 INFO:teuthology.orchestra.run.vm06.stdout: python3-ceph-common x86_64 2:19.2.3-47.gc24117fd552.el9.clyso ceph 142 k 2026-03-06T13:43:29.679 INFO:teuthology.orchestra.run.vm06.stdout: python3-certifi noarch 2023.05.07-4.el9 epel 14 k 2026-03-06T13:43:29.679 INFO:teuthology.orchestra.run.vm06.stdout: python3-cffi x86_64 1.14.5-5.el9 baseos 253 k 2026-03-06T13:43:29.679 INFO:teuthology.orchestra.run.vm06.stdout: python3-cheroot noarch 10.0.1-4.el9 epel 173 k 2026-03-06T13:43:29.679 INFO:teuthology.orchestra.run.vm06.stdout: python3-cherrypy noarch 18.6.1-2.el9 epel 358 k 2026-03-06T13:43:29.679 INFO:teuthology.orchestra.run.vm06.stdout: python3-cryptography x86_64 36.0.1-5.el9 baseos 1.2 M 2026-03-06T13:43:29.679 INFO:teuthology.orchestra.run.vm06.stdout: python3-devel x86_64 3.9.25-3.el9 appstream 244 k 2026-03-06T13:43:29.679 INFO:teuthology.orchestra.run.vm06.stdout: python3-google-auth noarch 1:2.45.0-1.el9 epel 254 k 2026-03-06T13:43:29.679 INFO:teuthology.orchestra.run.vm06.stdout: python3-grpcio x86_64 1.46.7-10.el9 epel 2.0 M 2026-03-06T13:43:29.679 INFO:teuthology.orchestra.run.vm06.stdout: python3-grpcio-tools x86_64 1.46.7-10.el9 epel 144 k 2026-03-06T13:43:29.679 INFO:teuthology.orchestra.run.vm06.stdout: python3-isodate noarch 0.6.1-3.el9 epel 56 k 2026-03-06T13:43:29.679 INFO:teuthology.orchestra.run.vm06.stdout: python3-jaraco noarch 8.2.1-3.el9 epel 11 k 2026-03-06T13:43:29.679 INFO:teuthology.orchestra.run.vm06.stdout: python3-jaraco-classes noarch 3.2.1-5.el9 epel 18 k 2026-03-06T13:43:29.679 INFO:teuthology.orchestra.run.vm06.stdout: python3-jaraco-collections noarch 3.0.0-8.el9 epel 23 k 2026-03-06T13:43:29.679 INFO:teuthology.orchestra.run.vm06.stdout: python3-jaraco-context noarch 6.0.1-3.el9 epel 20 k 2026-03-06T13:43:29.679 INFO:teuthology.orchestra.run.vm06.stdout: python3-jaraco-functools noarch 3.5.0-2.el9 epel 19 k 2026-03-06T13:43:29.679 INFO:teuthology.orchestra.run.vm06.stdout: python3-jaraco-text noarch 4.0.0-2.el9 epel 26 k 2026-03-06T13:43:29.679 INFO:teuthology.orchestra.run.vm06.stdout: python3-jinja2 noarch 2.11.3-8.el9 appstream 249 k 2026-03-06T13:43:29.679 INFO:teuthology.orchestra.run.vm06.stdout: python3-kubernetes noarch 1:26.1.0-3.el9 epel 1.0 M 2026-03-06T13:43:29.679 INFO:teuthology.orchestra.run.vm06.stdout: python3-libstoragemgmt x86_64 1.10.1-1.el9 appstream 177 k 2026-03-06T13:43:29.679 INFO:teuthology.orchestra.run.vm06.stdout: python3-logutils noarch 0.3.5-21.el9 epel 46 k 2026-03-06T13:43:29.680 INFO:teuthology.orchestra.run.vm06.stdout: python3-lxml x86_64 4.6.5-3.el9 appstream 1.2 M 2026-03-06T13:43:29.680 INFO:teuthology.orchestra.run.vm06.stdout: python3-mako noarch 1.1.4-6.el9 appstream 172 k 2026-03-06T13:43:29.680 INFO:teuthology.orchestra.run.vm06.stdout: python3-markupsafe x86_64 1.1.1-12.el9 appstream 35 k 2026-03-06T13:43:29.680 INFO:teuthology.orchestra.run.vm06.stdout: python3-more-itertools noarch 8.12.0-2.el9 epel 79 k 2026-03-06T13:43:29.680 INFO:teuthology.orchestra.run.vm06.stdout: python3-msgpack x86_64 1.0.3-2.el9 epel 86 k 2026-03-06T13:43:29.680 INFO:teuthology.orchestra.run.vm06.stdout: python3-natsort noarch 7.1.1-5.el9 epel 58 k 2026-03-06T13:43:29.680 INFO:teuthology.orchestra.run.vm06.stdout: python3-numpy x86_64 1:1.23.5-2.el9 appstream 6.1 M 2026-03-06T13:43:29.680 INFO:teuthology.orchestra.run.vm06.stdout: python3-numpy-f2py x86_64 1:1.23.5-2.el9 appstream 442 k 2026-03-06T13:43:29.680 INFO:teuthology.orchestra.run.vm06.stdout: python3-packaging noarch 20.9-5.el9 appstream 77 k 2026-03-06T13:43:29.680 INFO:teuthology.orchestra.run.vm06.stdout: python3-pecan noarch 1.4.2-3.el9 epel 272 k 2026-03-06T13:43:29.680 INFO:teuthology.orchestra.run.vm06.stdout: python3-ply noarch 3.11-14.el9 baseos 106 k 2026-03-06T13:43:29.680 INFO:teuthology.orchestra.run.vm06.stdout: python3-portend noarch 3.1.0-2.el9 epel 16 k 2026-03-06T13:43:29.680 INFO:teuthology.orchestra.run.vm06.stdout: python3-protobuf noarch 3.14.0-17.el9 appstream 267 k 2026-03-06T13:43:29.680 INFO:teuthology.orchestra.run.vm06.stdout: python3-pyOpenSSL noarch 21.0.0-1.el9 epel 90 k 2026-03-06T13:43:29.680 INFO:teuthology.orchestra.run.vm06.stdout: python3-pyasn1 noarch 0.4.8-7.el9 appstream 157 k 2026-03-06T13:43:29.680 INFO:teuthology.orchestra.run.vm06.stdout: python3-pyasn1-modules noarch 0.4.8-7.el9 appstream 277 k 2026-03-06T13:43:29.680 INFO:teuthology.orchestra.run.vm06.stdout: python3-pycparser noarch 2.20-6.el9 baseos 135 k 2026-03-06T13:43:29.680 INFO:teuthology.orchestra.run.vm06.stdout: python3-repoze-lru noarch 0.7-16.el9 epel 31 k 2026-03-06T13:43:29.680 INFO:teuthology.orchestra.run.vm06.stdout: python3-requests noarch 2.25.1-10.el9 baseos 126 k 2026-03-06T13:43:29.680 INFO:teuthology.orchestra.run.vm06.stdout: python3-requests-oauthlib noarch 1.3.0-12.el9 appstream 54 k 2026-03-06T13:43:29.680 INFO:teuthology.orchestra.run.vm06.stdout: python3-routes noarch 2.5.1-5.el9 epel 188 k 2026-03-06T13:43:29.680 INFO:teuthology.orchestra.run.vm06.stdout: python3-rsa noarch 4.9-2.el9 epel 59 k 2026-03-06T13:43:29.680 INFO:teuthology.orchestra.run.vm06.stdout: python3-scipy x86_64 1.9.3-2.el9 appstream 19 M 2026-03-06T13:43:29.680 INFO:teuthology.orchestra.run.vm06.stdout: python3-tempora noarch 5.0.0-2.el9 epel 36 k 2026-03-06T13:43:29.680 INFO:teuthology.orchestra.run.vm06.stdout: python3-toml noarch 0.10.2-6.el9 appstream 42 k 2026-03-06T13:43:29.680 INFO:teuthology.orchestra.run.vm06.stdout: python3-typing-extensions noarch 4.15.0-1.el9 epel 86 k 2026-03-06T13:43:29.680 INFO:teuthology.orchestra.run.vm06.stdout: python3-urllib3 noarch 1.26.5-7.el9 baseos 218 k 2026-03-06T13:43:29.680 INFO:teuthology.orchestra.run.vm06.stdout: python3-webob noarch 1.8.8-2.el9 epel 230 k 2026-03-06T13:43:29.680 INFO:teuthology.orchestra.run.vm06.stdout: python3-websocket-client noarch 1.2.3-2.el9 epel 90 k 2026-03-06T13:43:29.680 INFO:teuthology.orchestra.run.vm06.stdout: python3-werkzeug noarch 2.0.3-3.el9.1 epel 427 k 2026-03-06T13:43:29.680 INFO:teuthology.orchestra.run.vm06.stdout: python3-xmlsec x86_64 1.3.13-1.el9 epel 48 k 2026-03-06T13:43:29.680 INFO:teuthology.orchestra.run.vm06.stdout: python3-zc-lockfile noarch 2.0-10.el9 epel 20 k 2026-03-06T13:43:29.680 INFO:teuthology.orchestra.run.vm06.stdout: qatlib x86_64 25.08.0-2.el9 appstream 240 k 2026-03-06T13:43:29.680 INFO:teuthology.orchestra.run.vm06.stdout: qatzip-libs x86_64 1.3.1-1.el9 appstream 66 k 2026-03-06T13:43:29.680 INFO:teuthology.orchestra.run.vm06.stdout: re2 x86_64 1:20211101-20.el9 epel 191 k 2026-03-06T13:43:29.680 INFO:teuthology.orchestra.run.vm06.stdout: socat x86_64 1.7.4.1-8.el9 appstream 303 k 2026-03-06T13:43:29.680 INFO:teuthology.orchestra.run.vm06.stdout: thrift x86_64 0.15.0-4.el9 epel 1.6 M 2026-03-06T13:43:29.680 INFO:teuthology.orchestra.run.vm06.stdout: unzip x86_64 6.0-59.el9 baseos 182 k 2026-03-06T13:43:29.680 INFO:teuthology.orchestra.run.vm06.stdout: xmlsec1 x86_64 1.2.29-13.el9 appstream 189 k 2026-03-06T13:43:29.680 INFO:teuthology.orchestra.run.vm06.stdout: xmlsec1-openssl x86_64 1.2.29-13.el9 appstream 90 k 2026-03-06T13:43:29.680 INFO:teuthology.orchestra.run.vm06.stdout: xmlstarlet x86_64 1.6.1-20.el9 appstream 64 k 2026-03-06T13:43:29.680 INFO:teuthology.orchestra.run.vm06.stdout: zip x86_64 3.0-35.el9 baseos 266 k 2026-03-06T13:43:29.680 INFO:teuthology.orchestra.run.vm06.stdout:Installing weak dependencies: 2026-03-06T13:43:29.680 INFO:teuthology.orchestra.run.vm06.stdout: ceph-mgr-k8sevents noarch 2:19.2.3-47.gc24117fd552.el9.clyso ceph-noarch 22 k 2026-03-06T13:43:29.680 INFO:teuthology.orchestra.run.vm06.stdout: python3-influxdb noarch 5.3.1-1.el9 epel 139 k 2026-03-06T13:43:29.680 INFO:teuthology.orchestra.run.vm06.stdout: python3-saml noarch 1.16.0-1.el9 epel 125 k 2026-03-06T13:43:29.680 INFO:teuthology.orchestra.run.vm06.stdout: qatlib-service x86_64 25.08.0-2.el9 appstream 37 k 2026-03-06T13:43:29.680 INFO:teuthology.orchestra.run.vm06.stdout: smartmontools x86_64 1:7.2-10.el9 baseos 556 k 2026-03-06T13:43:29.680 INFO:teuthology.orchestra.run.vm06.stdout: 2026-03-06T13:43:29.680 INFO:teuthology.orchestra.run.vm06.stdout:Transaction Summary 2026-03-06T13:43:29.680 INFO:teuthology.orchestra.run.vm06.stdout:============================================================================================== 2026-03-06T13:43:29.680 INFO:teuthology.orchestra.run.vm06.stdout:Install 148 Packages 2026-03-06T13:43:29.680 INFO:teuthology.orchestra.run.vm06.stdout:Upgrade 2 Packages 2026-03-06T13:43:29.680 INFO:teuthology.orchestra.run.vm06.stdout: 2026-03-06T13:43:29.681 INFO:teuthology.orchestra.run.vm06.stdout:Total download size: 214 M 2026-03-06T13:43:29.681 INFO:teuthology.orchestra.run.vm06.stdout:Downloading Packages: 2026-03-06T13:43:29.758 INFO:teuthology.orchestra.run.vm02.stdout:Package librados2-2:16.2.4-5.el9.x86_64 is already installed. 2026-03-06T13:43:29.758 INFO:teuthology.orchestra.run.vm02.stdout:Package librbd1-2:16.2.4-5.el9.x86_64 is already installed. 2026-03-06T13:43:29.791 INFO:teuthology.orchestra.run.vm02.stdout:Dependencies resolved. 2026-03-06T13:43:29.795 INFO:teuthology.orchestra.run.vm02.stdout:============================================================================================== 2026-03-06T13:43:29.795 INFO:teuthology.orchestra.run.vm02.stdout: Package Arch Version Repository Size 2026-03-06T13:43:29.796 INFO:teuthology.orchestra.run.vm02.stdout:============================================================================================== 2026-03-06T13:43:29.796 INFO:teuthology.orchestra.run.vm02.stdout:Installing: 2026-03-06T13:43:29.796 INFO:teuthology.orchestra.run.vm02.stdout: bzip2 x86_64 1.0.8-11.el9 baseos 55 k 2026-03-06T13:43:29.796 INFO:teuthology.orchestra.run.vm02.stdout: ceph x86_64 2:19.2.3-47.gc24117fd552.el9.clyso ceph 6.5 k 2026-03-06T13:43:29.796 INFO:teuthology.orchestra.run.vm02.stdout: ceph-base x86_64 2:19.2.3-47.gc24117fd552.el9.clyso ceph 5.5 M 2026-03-06T13:43:29.796 INFO:teuthology.orchestra.run.vm02.stdout: ceph-fuse x86_64 2:19.2.3-47.gc24117fd552.el9.clyso ceph 1.1 M 2026-03-06T13:43:29.796 INFO:teuthology.orchestra.run.vm02.stdout: ceph-immutable-object-cache x86_64 2:19.2.3-47.gc24117fd552.el9.clyso ceph 145 k 2026-03-06T13:43:29.796 INFO:teuthology.orchestra.run.vm02.stdout: ceph-mgr x86_64 2:19.2.3-47.gc24117fd552.el9.clyso ceph 1.1 M 2026-03-06T13:43:29.796 INFO:teuthology.orchestra.run.vm02.stdout: ceph-mgr-cephadm noarch 2:19.2.3-47.gc24117fd552.el9.clyso ceph-noarch 150 k 2026-03-06T13:43:29.796 INFO:teuthology.orchestra.run.vm02.stdout: ceph-mgr-dashboard noarch 2:19.2.3-47.gc24117fd552.el9.clyso ceph-noarch 3.8 M 2026-03-06T13:43:29.796 INFO:teuthology.orchestra.run.vm02.stdout: ceph-mgr-diskprediction-local noarch 2:19.2.3-47.gc24117fd552.el9.clyso ceph-noarch 7.4 M 2026-03-06T13:43:29.796 INFO:teuthology.orchestra.run.vm02.stdout: ceph-mgr-rook noarch 2:19.2.3-47.gc24117fd552.el9.clyso ceph-noarch 49 k 2026-03-06T13:43:29.796 INFO:teuthology.orchestra.run.vm02.stdout: ceph-radosgw x86_64 2:19.2.3-47.gc24117fd552.el9.clyso ceph 11 M 2026-03-06T13:43:29.796 INFO:teuthology.orchestra.run.vm02.stdout: ceph-test x86_64 2:19.2.3-47.gc24117fd552.el9.clyso ceph 50 M 2026-03-06T13:43:29.796 INFO:teuthology.orchestra.run.vm02.stdout: ceph-volume noarch 2:19.2.3-47.gc24117fd552.el9.clyso ceph-noarch 299 k 2026-03-06T13:43:29.796 INFO:teuthology.orchestra.run.vm02.stdout: cephadm noarch 2:19.2.3-47.gc24117fd552.el9.clyso ceph-noarch 769 k 2026-03-06T13:43:29.796 INFO:teuthology.orchestra.run.vm02.stdout: libcephfs-devel x86_64 2:19.2.3-47.gc24117fd552.el9.clyso ceph 34 k 2026-03-06T13:43:29.796 INFO:teuthology.orchestra.run.vm02.stdout: libcephfs2 x86_64 2:19.2.3-47.gc24117fd552.el9.clyso ceph 998 k 2026-03-06T13:43:29.796 INFO:teuthology.orchestra.run.vm02.stdout: librados-devel x86_64 2:19.2.3-47.gc24117fd552.el9.clyso ceph 127 k 2026-03-06T13:43:29.796 INFO:teuthology.orchestra.run.vm02.stdout: perl-Test-Harness noarch 1:3.42-461.el9 appstream 295 k 2026-03-06T13:43:29.796 INFO:teuthology.orchestra.run.vm02.stdout: python3-cephfs x86_64 2:19.2.3-47.gc24117fd552.el9.clyso ceph 165 k 2026-03-06T13:43:29.796 INFO:teuthology.orchestra.run.vm02.stdout: python3-rados x86_64 2:19.2.3-47.gc24117fd552.el9.clyso ceph 322 k 2026-03-06T13:43:29.796 INFO:teuthology.orchestra.run.vm02.stdout: python3-rbd x86_64 2:19.2.3-47.gc24117fd552.el9.clyso ceph 303 k 2026-03-06T13:43:29.796 INFO:teuthology.orchestra.run.vm02.stdout: python3-rgw x86_64 2:19.2.3-47.gc24117fd552.el9.clyso ceph 100 k 2026-03-06T13:43:29.796 INFO:teuthology.orchestra.run.vm02.stdout: python3-xmltodict noarch 0.12.0-15.el9 epel 22 k 2026-03-06T13:43:29.796 INFO:teuthology.orchestra.run.vm02.stdout: rbd-fuse x86_64 2:19.2.3-47.gc24117fd552.el9.clyso ceph 85 k 2026-03-06T13:43:29.796 INFO:teuthology.orchestra.run.vm02.stdout: rbd-mirror x86_64 2:19.2.3-47.gc24117fd552.el9.clyso ceph 3.1 M 2026-03-06T13:43:29.796 INFO:teuthology.orchestra.run.vm02.stdout: rbd-nbd x86_64 2:19.2.3-47.gc24117fd552.el9.clyso ceph 171 k 2026-03-06T13:43:29.796 INFO:teuthology.orchestra.run.vm02.stdout: s3cmd noarch 2.4.0-1.el9 epel 206 k 2026-03-06T13:43:29.796 INFO:teuthology.orchestra.run.vm02.stdout:Upgrading: 2026-03-06T13:43:29.796 INFO:teuthology.orchestra.run.vm02.stdout: librados2 x86_64 2:19.2.3-47.gc24117fd552.el9.clyso ceph 3.4 M 2026-03-06T13:43:29.796 INFO:teuthology.orchestra.run.vm02.stdout: librbd1 x86_64 2:19.2.3-47.gc24117fd552.el9.clyso ceph 3.2 M 2026-03-06T13:43:29.796 INFO:teuthology.orchestra.run.vm02.stdout:Installing dependencies: 2026-03-06T13:43:29.796 INFO:teuthology.orchestra.run.vm02.stdout: abseil-cpp x86_64 20211102.0-4.el9 epel 551 k 2026-03-06T13:43:29.796 INFO:teuthology.orchestra.run.vm02.stdout: boost-program-options x86_64 1.75.0-13.el9 appstream 104 k 2026-03-06T13:43:29.796 INFO:teuthology.orchestra.run.vm02.stdout: ceph-common x86_64 2:19.2.3-47.gc24117fd552.el9.clyso ceph 22 M 2026-03-06T13:43:29.796 INFO:teuthology.orchestra.run.vm02.stdout: ceph-grafana-dashboards noarch 2:19.2.3-47.gc24117fd552.el9.clyso ceph-noarch 31 k 2026-03-06T13:43:29.796 INFO:teuthology.orchestra.run.vm02.stdout: ceph-mds x86_64 2:19.2.3-47.gc24117fd552.el9.clyso ceph 2.4 M 2026-03-06T13:43:29.796 INFO:teuthology.orchestra.run.vm02.stdout: ceph-mgr-modules-core noarch 2:19.2.3-47.gc24117fd552.el9.clyso ceph-noarch 252 k 2026-03-06T13:43:29.796 INFO:teuthology.orchestra.run.vm02.stdout: ceph-mon x86_64 2:19.2.3-47.gc24117fd552.el9.clyso ceph 4.7 M 2026-03-06T13:43:29.796 INFO:teuthology.orchestra.run.vm02.stdout: ceph-osd x86_64 2:19.2.3-47.gc24117fd552.el9.clyso ceph 17 M 2026-03-06T13:43:29.796 INFO:teuthology.orchestra.run.vm02.stdout: ceph-prometheus-alerts noarch 2:19.2.3-47.gc24117fd552.el9.clyso ceph-noarch 16 k 2026-03-06T13:43:29.796 INFO:teuthology.orchestra.run.vm02.stdout: ceph-selinux x86_64 2:19.2.3-47.gc24117fd552.el9.clyso ceph 25 k 2026-03-06T13:43:29.796 INFO:teuthology.orchestra.run.vm02.stdout: cryptsetup x86_64 2.8.1-3.el9 baseos 351 k 2026-03-06T13:43:29.796 INFO:teuthology.orchestra.run.vm02.stdout: flexiblas x86_64 3.0.4-9.el9 appstream 30 k 2026-03-06T13:43:29.796 INFO:teuthology.orchestra.run.vm02.stdout: flexiblas-netlib x86_64 3.0.4-9.el9 appstream 3.0 M 2026-03-06T13:43:29.796 INFO:teuthology.orchestra.run.vm02.stdout: flexiblas-openblas-openmp x86_64 3.0.4-9.el9 appstream 15 k 2026-03-06T13:43:29.796 INFO:teuthology.orchestra.run.vm02.stdout: fuse x86_64 2.9.9-17.el9 baseos 80 k 2026-03-06T13:43:29.796 INFO:teuthology.orchestra.run.vm02.stdout: gperftools-libs x86_64 2.9.1-3.el9 epel 308 k 2026-03-06T13:43:29.796 INFO:teuthology.orchestra.run.vm02.stdout: grpc-data noarch 1.46.7-10.el9 epel 19 k 2026-03-06T13:43:29.796 INFO:teuthology.orchestra.run.vm02.stdout: ledmon-libs x86_64 1.1.0-3.el9 baseos 40 k 2026-03-06T13:43:29.796 INFO:teuthology.orchestra.run.vm02.stdout: libarrow x86_64 9.0.0-15.el9 epel 4.4 M 2026-03-06T13:43:29.796 INFO:teuthology.orchestra.run.vm02.stdout: libarrow-doc noarch 9.0.0-15.el9 epel 25 k 2026-03-06T13:43:29.796 INFO:teuthology.orchestra.run.vm02.stdout: libcephsqlite x86_64 2:19.2.3-47.gc24117fd552.el9.clyso ceph 163 k 2026-03-06T13:43:29.796 INFO:teuthology.orchestra.run.vm02.stdout: libconfig x86_64 1.7.2-9.el9 baseos 72 k 2026-03-06T13:43:29.796 INFO:teuthology.orchestra.run.vm02.stdout: libgfortran x86_64 11.5.0-14.el9 baseos 794 k 2026-03-06T13:43:29.796 INFO:teuthology.orchestra.run.vm02.stdout: libnbd x86_64 1.20.3-4.el9 appstream 164 k 2026-03-06T13:43:29.796 INFO:teuthology.orchestra.run.vm02.stdout: liboath x86_64 2.6.12-1.el9 epel 49 k 2026-03-06T13:43:29.797 INFO:teuthology.orchestra.run.vm02.stdout: libpmemobj x86_64 1.12.1-1.el9 appstream 160 k 2026-03-06T13:43:29.797 INFO:teuthology.orchestra.run.vm02.stdout: libquadmath x86_64 11.5.0-14.el9 baseos 184 k 2026-03-06T13:43:29.797 INFO:teuthology.orchestra.run.vm02.stdout: librabbitmq x86_64 0.11.0-7.el9 appstream 45 k 2026-03-06T13:43:29.797 INFO:teuthology.orchestra.run.vm02.stdout: libradosstriper1 x86_64 2:19.2.3-47.gc24117fd552.el9.clyso ceph 503 k 2026-03-06T13:43:29.797 INFO:teuthology.orchestra.run.vm02.stdout: librdkafka x86_64 1.6.1-102.el9 appstream 662 k 2026-03-06T13:43:29.797 INFO:teuthology.orchestra.run.vm02.stdout: librgw2 x86_64 2:19.2.3-47.gc24117fd552.el9.clyso ceph 5.4 M 2026-03-06T13:43:29.797 INFO:teuthology.orchestra.run.vm02.stdout: libstoragemgmt x86_64 1.10.1-1.el9 appstream 246 k 2026-03-06T13:43:29.797 INFO:teuthology.orchestra.run.vm02.stdout: libunwind x86_64 1.6.2-1.el9 epel 67 k 2026-03-06T13:43:29.797 INFO:teuthology.orchestra.run.vm02.stdout: libxslt x86_64 1.1.34-12.el9 appstream 233 k 2026-03-06T13:43:29.797 INFO:teuthology.orchestra.run.vm02.stdout: lttng-ust x86_64 2.12.0-6.el9 appstream 292 k 2026-03-06T13:43:29.797 INFO:teuthology.orchestra.run.vm02.stdout: lua x86_64 5.4.4-4.el9 appstream 188 k 2026-03-06T13:43:29.797 INFO:teuthology.orchestra.run.vm02.stdout: lua-devel x86_64 5.4.4-4.el9 crb 22 k 2026-03-06T13:43:29.797 INFO:teuthology.orchestra.run.vm02.stdout: luarocks noarch 3.9.2-5.el9 epel 151 k 2026-03-06T13:43:29.797 INFO:teuthology.orchestra.run.vm02.stdout: mailcap noarch 2.1.49-5.el9 baseos 33 k 2026-03-06T13:43:29.797 INFO:teuthology.orchestra.run.vm02.stdout: openblas x86_64 0.3.29-1.el9 appstream 42 k 2026-03-06T13:43:29.797 INFO:teuthology.orchestra.run.vm02.stdout: openblas-openmp x86_64 0.3.29-1.el9 appstream 5.3 M 2026-03-06T13:43:29.797 INFO:teuthology.orchestra.run.vm02.stdout: parquet-libs x86_64 9.0.0-15.el9 epel 838 k 2026-03-06T13:43:29.797 INFO:teuthology.orchestra.run.vm02.stdout: pciutils x86_64 3.7.0-7.el9 baseos 93 k 2026-03-06T13:43:29.797 INFO:teuthology.orchestra.run.vm02.stdout: perl-Benchmark noarch 1.23-483.el9 appstream 26 k 2026-03-06T13:43:29.797 INFO:teuthology.orchestra.run.vm02.stdout: protobuf x86_64 3.14.0-17.el9 appstream 1.0 M 2026-03-06T13:43:29.797 INFO:teuthology.orchestra.run.vm02.stdout: protobuf-compiler x86_64 3.14.0-17.el9 crb 862 k 2026-03-06T13:43:29.797 INFO:teuthology.orchestra.run.vm02.stdout: python3-asyncssh noarch 2.13.2-5.el9 epel 548 k 2026-03-06T13:43:29.797 INFO:teuthology.orchestra.run.vm02.stdout: python3-autocommand noarch 2.2.2-8.el9 epel 29 k 2026-03-06T13:43:29.797 INFO:teuthology.orchestra.run.vm02.stdout: python3-babel noarch 2.9.1-2.el9 appstream 6.0 M 2026-03-06T13:43:29.797 INFO:teuthology.orchestra.run.vm02.stdout: python3-backports-tarfile noarch 1.2.0-1.el9 epel 60 k 2026-03-06T13:43:29.797 INFO:teuthology.orchestra.run.vm02.stdout: python3-bcrypt x86_64 3.2.2-1.el9 epel 43 k 2026-03-06T13:43:29.797 INFO:teuthology.orchestra.run.vm02.stdout: python3-cachetools noarch 4.2.4-1.el9 epel 32 k 2026-03-06T13:43:29.797 INFO:teuthology.orchestra.run.vm02.stdout: python3-ceph-argparse x86_64 2:19.2.3-47.gc24117fd552.el9.clyso ceph 45 k 2026-03-06T13:43:29.797 INFO:teuthology.orchestra.run.vm02.stdout: python3-ceph-common x86_64 2:19.2.3-47.gc24117fd552.el9.clyso ceph 142 k 2026-03-06T13:43:29.797 INFO:teuthology.orchestra.run.vm02.stdout: python3-certifi noarch 2023.05.07-4.el9 epel 14 k 2026-03-06T13:43:29.797 INFO:teuthology.orchestra.run.vm02.stdout: python3-cffi x86_64 1.14.5-5.el9 baseos 253 k 2026-03-06T13:43:29.797 INFO:teuthology.orchestra.run.vm02.stdout: python3-cheroot noarch 10.0.1-4.el9 epel 173 k 2026-03-06T13:43:29.797 INFO:teuthology.orchestra.run.vm02.stdout: python3-cherrypy noarch 18.6.1-2.el9 epel 358 k 2026-03-06T13:43:29.797 INFO:teuthology.orchestra.run.vm02.stdout: python3-cryptography x86_64 36.0.1-5.el9 baseos 1.2 M 2026-03-06T13:43:29.797 INFO:teuthology.orchestra.run.vm02.stdout: python3-devel x86_64 3.9.25-3.el9 appstream 244 k 2026-03-06T13:43:29.797 INFO:teuthology.orchestra.run.vm02.stdout: python3-google-auth noarch 1:2.45.0-1.el9 epel 254 k 2026-03-06T13:43:29.797 INFO:teuthology.orchestra.run.vm02.stdout: python3-grpcio x86_64 1.46.7-10.el9 epel 2.0 M 2026-03-06T13:43:29.797 INFO:teuthology.orchestra.run.vm02.stdout: python3-grpcio-tools x86_64 1.46.7-10.el9 epel 144 k 2026-03-06T13:43:29.797 INFO:teuthology.orchestra.run.vm02.stdout: python3-isodate noarch 0.6.1-3.el9 epel 56 k 2026-03-06T13:43:29.797 INFO:teuthology.orchestra.run.vm02.stdout: python3-jaraco noarch 8.2.1-3.el9 epel 11 k 2026-03-06T13:43:29.797 INFO:teuthology.orchestra.run.vm02.stdout: python3-jaraco-classes noarch 3.2.1-5.el9 epel 18 k 2026-03-06T13:43:29.797 INFO:teuthology.orchestra.run.vm02.stdout: python3-jaraco-collections noarch 3.0.0-8.el9 epel 23 k 2026-03-06T13:43:29.797 INFO:teuthology.orchestra.run.vm02.stdout: python3-jaraco-context noarch 6.0.1-3.el9 epel 20 k 2026-03-06T13:43:29.797 INFO:teuthology.orchestra.run.vm02.stdout: python3-jaraco-functools noarch 3.5.0-2.el9 epel 19 k 2026-03-06T13:43:29.797 INFO:teuthology.orchestra.run.vm02.stdout: python3-jaraco-text noarch 4.0.0-2.el9 epel 26 k 2026-03-06T13:43:29.797 INFO:teuthology.orchestra.run.vm02.stdout: python3-jinja2 noarch 2.11.3-8.el9 appstream 249 k 2026-03-06T13:43:29.797 INFO:teuthology.orchestra.run.vm02.stdout: python3-kubernetes noarch 1:26.1.0-3.el9 epel 1.0 M 2026-03-06T13:43:29.797 INFO:teuthology.orchestra.run.vm02.stdout: python3-libstoragemgmt x86_64 1.10.1-1.el9 appstream 177 k 2026-03-06T13:43:29.797 INFO:teuthology.orchestra.run.vm02.stdout: python3-logutils noarch 0.3.5-21.el9 epel 46 k 2026-03-06T13:43:29.797 INFO:teuthology.orchestra.run.vm02.stdout: python3-lxml x86_64 4.6.5-3.el9 appstream 1.2 M 2026-03-06T13:43:29.797 INFO:teuthology.orchestra.run.vm02.stdout: python3-mako noarch 1.1.4-6.el9 appstream 172 k 2026-03-06T13:43:29.797 INFO:teuthology.orchestra.run.vm02.stdout: python3-markupsafe x86_64 1.1.1-12.el9 appstream 35 k 2026-03-06T13:43:29.797 INFO:teuthology.orchestra.run.vm02.stdout: python3-more-itertools noarch 8.12.0-2.el9 epel 79 k 2026-03-06T13:43:29.797 INFO:teuthology.orchestra.run.vm02.stdout: python3-msgpack x86_64 1.0.3-2.el9 epel 86 k 2026-03-06T13:43:29.797 INFO:teuthology.orchestra.run.vm02.stdout: python3-natsort noarch 7.1.1-5.el9 epel 58 k 2026-03-06T13:43:29.797 INFO:teuthology.orchestra.run.vm02.stdout: python3-numpy x86_64 1:1.23.5-2.el9 appstream 6.1 M 2026-03-06T13:43:29.797 INFO:teuthology.orchestra.run.vm02.stdout: python3-numpy-f2py x86_64 1:1.23.5-2.el9 appstream 442 k 2026-03-06T13:43:29.797 INFO:teuthology.orchestra.run.vm02.stdout: python3-packaging noarch 20.9-5.el9 appstream 77 k 2026-03-06T13:43:29.798 INFO:teuthology.orchestra.run.vm02.stdout: python3-pecan noarch 1.4.2-3.el9 epel 272 k 2026-03-06T13:43:29.798 INFO:teuthology.orchestra.run.vm02.stdout: python3-ply noarch 3.11-14.el9 baseos 106 k 2026-03-06T13:43:29.798 INFO:teuthology.orchestra.run.vm02.stdout: python3-portend noarch 3.1.0-2.el9 epel 16 k 2026-03-06T13:43:29.798 INFO:teuthology.orchestra.run.vm02.stdout: python3-protobuf noarch 3.14.0-17.el9 appstream 267 k 2026-03-06T13:43:29.798 INFO:teuthology.orchestra.run.vm02.stdout: python3-pyOpenSSL noarch 21.0.0-1.el9 epel 90 k 2026-03-06T13:43:29.798 INFO:teuthology.orchestra.run.vm02.stdout: python3-pyasn1 noarch 0.4.8-7.el9 appstream 157 k 2026-03-06T13:43:29.798 INFO:teuthology.orchestra.run.vm02.stdout: python3-pyasn1-modules noarch 0.4.8-7.el9 appstream 277 k 2026-03-06T13:43:29.798 INFO:teuthology.orchestra.run.vm02.stdout: python3-pycparser noarch 2.20-6.el9 baseos 135 k 2026-03-06T13:43:29.798 INFO:teuthology.orchestra.run.vm02.stdout: python3-repoze-lru noarch 0.7-16.el9 epel 31 k 2026-03-06T13:43:29.798 INFO:teuthology.orchestra.run.vm02.stdout: python3-requests noarch 2.25.1-10.el9 baseos 126 k 2026-03-06T13:43:29.798 INFO:teuthology.orchestra.run.vm02.stdout: python3-requests-oauthlib noarch 1.3.0-12.el9 appstream 54 k 2026-03-06T13:43:29.798 INFO:teuthology.orchestra.run.vm02.stdout: python3-routes noarch 2.5.1-5.el9 epel 188 k 2026-03-06T13:43:29.798 INFO:teuthology.orchestra.run.vm02.stdout: python3-rsa noarch 4.9-2.el9 epel 59 k 2026-03-06T13:43:29.798 INFO:teuthology.orchestra.run.vm02.stdout: python3-scipy x86_64 1.9.3-2.el9 appstream 19 M 2026-03-06T13:43:29.798 INFO:teuthology.orchestra.run.vm02.stdout: python3-tempora noarch 5.0.0-2.el9 epel 36 k 2026-03-06T13:43:29.798 INFO:teuthology.orchestra.run.vm02.stdout: python3-toml noarch 0.10.2-6.el9 appstream 42 k 2026-03-06T13:43:29.798 INFO:teuthology.orchestra.run.vm02.stdout: python3-typing-extensions noarch 4.15.0-1.el9 epel 86 k 2026-03-06T13:43:29.798 INFO:teuthology.orchestra.run.vm02.stdout: python3-urllib3 noarch 1.26.5-7.el9 baseos 218 k 2026-03-06T13:43:29.798 INFO:teuthology.orchestra.run.vm02.stdout: python3-webob noarch 1.8.8-2.el9 epel 230 k 2026-03-06T13:43:29.798 INFO:teuthology.orchestra.run.vm02.stdout: python3-websocket-client noarch 1.2.3-2.el9 epel 90 k 2026-03-06T13:43:29.798 INFO:teuthology.orchestra.run.vm02.stdout: python3-werkzeug noarch 2.0.3-3.el9.1 epel 427 k 2026-03-06T13:43:29.798 INFO:teuthology.orchestra.run.vm02.stdout: python3-xmlsec x86_64 1.3.13-1.el9 epel 48 k 2026-03-06T13:43:29.798 INFO:teuthology.orchestra.run.vm02.stdout: python3-zc-lockfile noarch 2.0-10.el9 epel 20 k 2026-03-06T13:43:29.798 INFO:teuthology.orchestra.run.vm02.stdout: qatlib x86_64 25.08.0-2.el9 appstream 240 k 2026-03-06T13:43:29.798 INFO:teuthology.orchestra.run.vm02.stdout: qatzip-libs x86_64 1.3.1-1.el9 appstream 66 k 2026-03-06T13:43:29.798 INFO:teuthology.orchestra.run.vm02.stdout: re2 x86_64 1:20211101-20.el9 epel 191 k 2026-03-06T13:43:29.798 INFO:teuthology.orchestra.run.vm02.stdout: socat x86_64 1.7.4.1-8.el9 appstream 303 k 2026-03-06T13:43:29.798 INFO:teuthology.orchestra.run.vm02.stdout: thrift x86_64 0.15.0-4.el9 epel 1.6 M 2026-03-06T13:43:29.798 INFO:teuthology.orchestra.run.vm02.stdout: unzip x86_64 6.0-59.el9 baseos 182 k 2026-03-06T13:43:29.798 INFO:teuthology.orchestra.run.vm02.stdout: xmlsec1 x86_64 1.2.29-13.el9 appstream 189 k 2026-03-06T13:43:29.798 INFO:teuthology.orchestra.run.vm02.stdout: xmlsec1-openssl x86_64 1.2.29-13.el9 appstream 90 k 2026-03-06T13:43:29.798 INFO:teuthology.orchestra.run.vm02.stdout: xmlstarlet x86_64 1.6.1-20.el9 appstream 64 k 2026-03-06T13:43:29.798 INFO:teuthology.orchestra.run.vm02.stdout: zip x86_64 3.0-35.el9 baseos 266 k 2026-03-06T13:43:29.798 INFO:teuthology.orchestra.run.vm02.stdout:Installing weak dependencies: 2026-03-06T13:43:29.798 INFO:teuthology.orchestra.run.vm02.stdout: ceph-mgr-k8sevents noarch 2:19.2.3-47.gc24117fd552.el9.clyso ceph-noarch 22 k 2026-03-06T13:43:29.798 INFO:teuthology.orchestra.run.vm02.stdout: python3-influxdb noarch 5.3.1-1.el9 epel 139 k 2026-03-06T13:43:29.798 INFO:teuthology.orchestra.run.vm02.stdout: python3-saml noarch 1.16.0-1.el9 epel 125 k 2026-03-06T13:43:29.798 INFO:teuthology.orchestra.run.vm02.stdout: qatlib-service x86_64 25.08.0-2.el9 appstream 37 k 2026-03-06T13:43:29.798 INFO:teuthology.orchestra.run.vm02.stdout: smartmontools x86_64 1:7.2-10.el9 baseos 556 k 2026-03-06T13:43:29.798 INFO:teuthology.orchestra.run.vm02.stdout: 2026-03-06T13:43:29.798 INFO:teuthology.orchestra.run.vm02.stdout:Transaction Summary 2026-03-06T13:43:29.798 INFO:teuthology.orchestra.run.vm02.stdout:============================================================================================== 2026-03-06T13:43:29.798 INFO:teuthology.orchestra.run.vm02.stdout:Install 148 Packages 2026-03-06T13:43:29.798 INFO:teuthology.orchestra.run.vm02.stdout:Upgrade 2 Packages 2026-03-06T13:43:29.798 INFO:teuthology.orchestra.run.vm02.stdout: 2026-03-06T13:43:29.798 INFO:teuthology.orchestra.run.vm02.stdout:Total download size: 214 M 2026-03-06T13:43:29.798 INFO:teuthology.orchestra.run.vm02.stdout:Downloading Packages: 2026-03-06T13:43:30.684 INFO:teuthology.orchestra.run.vm04.stdout:(1/150): ceph-19.2.3-47.gc24117fd552.el9.clyso. 37 kB/s | 6.5 kB 00:00 2026-03-06T13:43:31.026 INFO:teuthology.orchestra.run.vm04.stdout:(2/150): ceph-base-19.2.3-47.gc24117fd552.el9.c 11 MB/s | 5.5 MB 00:00 2026-03-06T13:43:31.140 INFO:teuthology.orchestra.run.vm02.stdout:(1/150): ceph-19.2.3-47.gc24117fd552.el9.clyso. 80 kB/s | 6.5 kB 00:00 2026-03-06T13:43:31.190 INFO:teuthology.orchestra.run.vm04.stdout:(3/150): ceph-immutable-object-cache-19.2.3-47. 885 kB/s | 145 kB 00:00 2026-03-06T13:43:31.232 INFO:teuthology.orchestra.run.vm04.stdout:(4/150): ceph-fuse-19.2.3-47.gc24117fd552.el9.c 2.1 MB/s | 1.1 MB 00:00 2026-03-06T13:43:31.398 INFO:teuthology.orchestra.run.vm04.stdout:(5/150): ceph-mgr-19.2.3-47.gc24117fd552.el9.cl 6.5 MB/s | 1.1 MB 00:00 2026-03-06T13:43:31.427 INFO:teuthology.orchestra.run.vm02.stdout:(2/150): ceph-fuse-19.2.3-47.gc24117fd552.el9.c 4.0 MB/s | 1.1 MB 00:00 2026-03-06T13:43:31.575 INFO:teuthology.orchestra.run.vm06.stdout:(1/150): ceph-19.2.3-47.gc24117fd552.el9.clyso. 28 kB/s | 6.5 kB 00:00 2026-03-06T13:43:31.605 INFO:teuthology.orchestra.run.vm02.stdout:(3/150): ceph-immutable-object-cache-19.2.3-47. 819 kB/s | 145 kB 00:00 2026-03-06T13:43:31.687 INFO:teuthology.orchestra.run.vm04.stdout:(6/150): ceph-common-19.2.3-47.gc24117fd552.el9 19 MB/s | 22 MB 00:01 2026-03-06T13:43:31.707 INFO:teuthology.orchestra.run.vm02.stdout:(4/150): ceph-base-19.2.3-47.gc24117fd552.el9.c 8.5 MB/s | 5.5 MB 00:00 2026-03-06T13:43:31.797 INFO:teuthology.orchestra.run.vm04.stdout:(7/150): ceph-mds-19.2.3-47.gc24117fd552.el9.cl 4.0 MB/s | 2.4 MB 00:00 2026-03-06T13:43:31.903 INFO:teuthology.orchestra.run.vm06.stdout:(2/150): ceph-fuse-19.2.3-47.gc24117fd552.el9.c 3.5 MB/s | 1.1 MB 00:00 2026-03-06T13:43:31.913 INFO:teuthology.orchestra.run.vm02.stdout:(5/150): ceph-mgr-19.2.3-47.gc24117fd552.el9.cl 5.2 MB/s | 1.1 MB 00:00 2026-03-06T13:43:32.022 INFO:teuthology.orchestra.run.vm06.stdout:(3/150): ceph-immutable-object-cache-19.2.3-47. 1.2 MB/s | 145 kB 00:00 2026-03-06T13:43:32.023 INFO:teuthology.orchestra.run.vm02.stdout:(6/150): ceph-common-19.2.3-47.gc24117fd552.el9 23 MB/s | 22 MB 00:00 2026-03-06T13:43:32.030 INFO:teuthology.orchestra.run.vm04.stdout:(8/150): ceph-mon-19.2.3-47.gc24117fd552.el9.cl 7.5 MB/s | 4.7 MB 00:00 2026-03-06T13:43:32.079 INFO:teuthology.orchestra.run.vm04.stdout:(9/150): ceph-selinux-19.2.3-47.gc24117fd552.el 516 kB/s | 25 kB 00:00 2026-03-06T13:43:32.130 INFO:teuthology.orchestra.run.vm02.stdout:(7/150): ceph-mds-19.2.3-47.gc24117fd552.el9.cl 4.6 MB/s | 2.4 MB 00:00 2026-03-06T13:43:32.279 INFO:teuthology.orchestra.run.vm06.stdout:(4/150): ceph-base-19.2.3-47.gc24117fd552.el9.c 5.9 MB/s | 5.5 MB 00:00 2026-03-06T13:43:32.472 INFO:teuthology.orchestra.run.vm04.stdout:(10/150): ceph-radosgw-19.2.3-47.gc24117fd552.e 16 MB/s | 11 MB 00:00 2026-03-06T13:43:32.535 INFO:teuthology.orchestra.run.vm04.stdout:(11/150): libcephfs-devel-19.2.3-47.gc24117fd55 539 kB/s | 34 kB 00:00 2026-03-06T13:43:32.549 INFO:teuthology.orchestra.run.vm02.stdout:(8/150): ceph-mon-19.2.3-47.gc24117fd552.el9.cl 7.4 MB/s | 4.7 MB 00:00 2026-03-06T13:43:32.675 INFO:teuthology.orchestra.run.vm02.stdout:(9/150): ceph-selinux-19.2.3-47.gc24117fd552.el 199 kB/s | 25 kB 00:00 2026-03-06T13:43:32.696 INFO:teuthology.orchestra.run.vm04.stdout:(12/150): libcephfs2-19.2.3-47.gc24117fd552.el9 6.1 MB/s | 998 kB 00:00 2026-03-06T13:43:32.767 INFO:teuthology.orchestra.run.vm04.stdout:(13/150): libcephsqlite-19.2.3-47.gc24117fd552. 2.3 MB/s | 163 kB 00:00 2026-03-06T13:43:32.875 INFO:teuthology.orchestra.run.vm06.stdout:(5/150): ceph-common-19.2.3-47.gc24117fd552.el9 14 MB/s | 22 MB 00:01 2026-03-06T13:43:32.959 INFO:teuthology.orchestra.run.vm04.stdout:(14/150): librados-devel-19.2.3-47.gc24117fd552 660 kB/s | 127 kB 00:00 2026-03-06T13:43:32.973 INFO:teuthology.orchestra.run.vm06.stdout:(6/150): ceph-mgr-19.2.3-47.gc24117fd552.el9.cl 1.5 MB/s | 1.1 MB 00:00 2026-03-06T13:43:32.981 INFO:teuthology.orchestra.run.vm06.stdout:(7/150): ceph-mds-19.2.3-47.gc24117fd552.el9.cl 2.5 MB/s | 2.4 MB 00:00 2026-03-06T13:43:33.047 INFO:teuthology.orchestra.run.vm04.stdout:(15/150): ceph-osd-19.2.3-47.gc24117fd552.el9.c 13 MB/s | 17 MB 00:01 2026-03-06T13:43:33.154 INFO:teuthology.orchestra.run.vm04.stdout:(16/150): libradosstriper1-19.2.3-47.gc24117fd5 2.5 MB/s | 503 kB 00:00 2026-03-06T13:43:33.271 INFO:teuthology.orchestra.run.vm02.stdout:(10/150): ceph-radosgw-19.2.3-47.gc24117fd552.e 9.5 MB/s | 11 MB 00:01 2026-03-06T13:43:33.273 INFO:teuthology.orchestra.run.vm04.stdout:(17/150): python3-ceph-argparse-19.2.3-47.gc241 382 kB/s | 45 kB 00:00 2026-03-06T13:43:33.465 INFO:teuthology.orchestra.run.vm02.stdout:(11/150): libcephfs-devel-19.2.3-47.gc24117fd55 173 kB/s | 34 kB 00:00 2026-03-06T13:43:33.483 INFO:teuthology.orchestra.run.vm04.stdout:(18/150): python3-ceph-common-19.2.3-47.gc24117 676 kB/s | 142 kB 00:00 2026-03-06T13:43:33.551 INFO:teuthology.orchestra.run.vm04.stdout:(19/150): python3-cephfs-19.2.3-47.gc24117fd552 2.4 MB/s | 165 kB 00:00 2026-03-06T13:43:33.741 INFO:teuthology.orchestra.run.vm06.stdout:(8/150): ceph-mon-19.2.3-47.gc24117fd552.el9.cl 5.5 MB/s | 4.7 MB 00:00 2026-03-06T13:43:33.753 INFO:teuthology.orchestra.run.vm02.stdout:(12/150): libcephfs2-19.2.3-47.gc24117fd552.el9 3.4 MB/s | 998 kB 00:00 2026-03-06T13:43:33.794 INFO:teuthology.orchestra.run.vm06.stdout:(9/150): ceph-selinux-19.2.3-47.gc24117fd552.el 474 kB/s | 25 kB 00:00 2026-03-06T13:43:33.833 INFO:teuthology.orchestra.run.vm04.stdout:(20/150): python3-rados-19.2.3-47.gc24117fd552. 1.1 MB/s | 322 kB 00:00 2026-03-06T13:43:33.834 INFO:teuthology.orchestra.run.vm02.stdout:(13/150): ceph-osd-19.2.3-47.gc24117fd552.el9.c 9.4 MB/s | 17 MB 00:01 2026-03-06T13:43:33.835 INFO:teuthology.orchestra.run.vm02.stdout:(14/150): libcephsqlite-19.2.3-47.gc24117fd552. 1.9 MB/s | 163 kB 00:00 2026-03-06T13:43:34.005 INFO:teuthology.orchestra.run.vm02.stdout:(15/150): librados-devel-19.2.3-47.gc24117fd552 739 kB/s | 127 kB 00:00 2026-03-06T13:43:34.206 INFO:teuthology.orchestra.run.vm04.stdout:(21/150): librgw2-19.2.3-47.gc24117fd552.el9.cl 4.7 MB/s | 5.4 MB 00:01 2026-03-06T13:43:34.246 INFO:teuthology.orchestra.run.vm04.stdout:(22/150): python3-rgw-19.2.3-47.gc24117fd552.el 2.5 MB/s | 100 kB 00:00 2026-03-06T13:43:34.319 INFO:teuthology.orchestra.run.vm02.stdout:(16/150): libradosstriper1-19.2.3-47.gc24117fd5 1.0 MB/s | 503 kB 00:00 2026-03-06T13:43:34.435 INFO:teuthology.orchestra.run.vm06.stdout:(10/150): ceph-radosgw-19.2.3-47.gc24117fd552.e 7.4 MB/s | 11 MB 00:01 2026-03-06T13:43:34.436 INFO:teuthology.orchestra.run.vm04.stdout:(23/150): ceph-test-19.2.3-47.gc24117fd552.el9. 21 MB/s | 50 MB 00:02 2026-03-06T13:43:34.437 INFO:teuthology.orchestra.run.vm04.stdout:(24/150): rbd-fuse-19.2.3-47.gc24117fd552.el9.c 443 kB/s | 85 kB 00:00 2026-03-06T13:43:34.439 INFO:teuthology.orchestra.run.vm04.stdout:(25/150): python3-rbd-19.2.3-47.gc24117fd552.el 499 kB/s | 303 kB 00:00 2026-03-06T13:43:34.478 INFO:teuthology.orchestra.run.vm06.stdout:(11/150): libcephfs-devel-19.2.3-47.gc24117fd55 809 kB/s | 34 kB 00:00 2026-03-06T13:43:34.507 INFO:teuthology.orchestra.run.vm04.stdout:(26/150): ceph-grafana-dashboards-19.2.3-47.gc2 462 kB/s | 31 kB 00:00 2026-03-06T13:43:34.536 INFO:teuthology.orchestra.run.vm04.stdout:(27/150): rbd-nbd-19.2.3-47.gc24117fd552.el9.cl 1.7 MB/s | 171 kB 00:00 2026-03-06T13:43:34.571 INFO:teuthology.orchestra.run.vm04.stdout:(28/150): ceph-mgr-cephadm-19.2.3-47.gc24117fd5 2.3 MB/s | 150 kB 00:00 2026-03-06T13:43:34.589 INFO:teuthology.orchestra.run.vm02.stdout:(17/150): librgw2-19.2.3-47.gc24117fd552.el9.cl 9.3 MB/s | 5.4 MB 00:00 2026-03-06T13:43:34.638 INFO:teuthology.orchestra.run.vm06.stdout:(12/150): libcephfs2-19.2.3-47.gc24117fd552.el9 6.1 MB/s | 998 kB 00:00 2026-03-06T13:43:34.701 INFO:teuthology.orchestra.run.vm06.stdout:(13/150): libcephsqlite-19.2.3-47.gc24117fd552. 2.6 MB/s | 163 kB 00:00 2026-03-06T13:43:34.757 INFO:teuthology.orchestra.run.vm06.stdout:(14/150): librados-devel-19.2.3-47.gc24117fd552 2.2 MB/s | 127 kB 00:00 2026-03-06T13:43:34.762 INFO:teuthology.orchestra.run.vm02.stdout:(18/150): python3-ceph-argparse-19.2.3-47.gc241 102 kB/s | 45 kB 00:00 2026-03-06T13:43:34.829 INFO:teuthology.orchestra.run.vm06.stdout:(15/150): ceph-osd-19.2.3-47.gc24117fd552.el9.c 9.2 MB/s | 17 MB 00:01 2026-03-06T13:43:34.829 INFO:teuthology.orchestra.run.vm02.stdout:(19/150): python3-cephfs-19.2.3-47.gc24117fd552 2.4 MB/s | 165 kB 00:00 2026-03-06T13:43:34.836 INFO:teuthology.orchestra.run.vm06.stdout:(16/150): libradosstriper1-19.2.3-47.gc24117fd5 6.2 MB/s | 503 kB 00:00 2026-03-06T13:43:34.869 INFO:teuthology.orchestra.run.vm06.stdout:(17/150): python3-ceph-argparse-19.2.3-47.gc241 1.4 MB/s | 45 kB 00:00 2026-03-06T13:43:34.885 INFO:teuthology.orchestra.run.vm02.stdout:(20/150): python3-ceph-common-19.2.3-47.gc24117 482 kB/s | 142 kB 00:00 2026-03-06T13:43:35.186 INFO:teuthology.orchestra.run.vm06.stdout:(18/150): python3-ceph-common-19.2.3-47.gc24117 449 kB/s | 142 kB 00:00 2026-03-06T13:43:35.189 INFO:teuthology.orchestra.run.vm02.stdout:(21/150): python3-rbd-19.2.3-47.gc24117fd552.el 998 kB/s | 303 kB 00:00 2026-03-06T13:43:35.253 INFO:teuthology.orchestra.run.vm06.stdout:(19/150): python3-cephfs-19.2.3-47.gc24117fd552 2.4 MB/s | 165 kB 00:00 2026-03-06T13:43:35.311 INFO:teuthology.orchestra.run.vm02.stdout:(22/150): python3-rgw-19.2.3-47.gc24117fd552.el 826 kB/s | 100 kB 00:00 2026-03-06T13:43:35.378 INFO:teuthology.orchestra.run.vm02.stdout:(23/150): python3-rados-19.2.3-47.gc24117fd552. 587 kB/s | 322 kB 00:00 2026-03-06T13:43:35.391 INFO:teuthology.orchestra.run.vm02.stdout:(24/150): rbd-fuse-19.2.3-47.gc24117fd552.el9.c 1.0 MB/s | 85 kB 00:00 2026-03-06T13:43:35.493 INFO:teuthology.orchestra.run.vm02.stdout:(25/150): rbd-nbd-19.2.3-47.gc24117fd552.el9.cl 1.6 MB/s | 171 kB 00:00 2026-03-06T13:43:35.573 INFO:teuthology.orchestra.run.vm06.stdout:(20/150): librgw2-19.2.3-47.gc24117fd552.el9.cl 7.3 MB/s | 5.4 MB 00:00 2026-03-06T13:43:35.592 INFO:teuthology.orchestra.run.vm06.stdout:(21/150): python3-rados-19.2.3-47.gc24117fd552. 950 kB/s | 322 kB 00:00 2026-03-06T13:43:35.611 INFO:teuthology.orchestra.run.vm02.stdout:(26/150): ceph-grafana-dashboards-19.2.3-47.gc2 265 kB/s | 31 kB 00:00 2026-03-06T13:43:35.626 INFO:teuthology.orchestra.run.vm06.stdout:(22/150): python3-rgw-19.2.3-47.gc24117fd552.el 2.9 MB/s | 100 kB 00:00 2026-03-06T13:43:35.677 INFO:teuthology.orchestra.run.vm06.stdout:(23/150): rbd-fuse-19.2.3-47.gc24117fd552.el9.c 1.6 MB/s | 85 kB 00:00 2026-03-06T13:43:35.783 INFO:teuthology.orchestra.run.vm02.stdout:(27/150): ceph-test-19.2.3-47.gc24117fd552.el9. 16 MB/s | 50 MB 00:03 2026-03-06T13:43:35.784 INFO:teuthology.orchestra.run.vm04.stdout:(29/150): ceph-mgr-dashboard-19.2.3-47.gc24117f 3.0 MB/s | 3.8 MB 00:01 2026-03-06T13:43:35.798 INFO:teuthology.orchestra.run.vm02.stdout:(28/150): rbd-mirror-19.2.3-47.gc24117fd552.el9 7.4 MB/s | 3.1 MB 00:00 2026-03-06T13:43:35.798 INFO:teuthology.orchestra.run.vm06.stdout:(24/150): python3-rbd-19.2.3-47.gc24117fd552.el 1.3 MB/s | 303 kB 00:00 2026-03-06T13:43:35.896 INFO:teuthology.orchestra.run.vm02.stdout:(29/150): ceph-mgr-cephadm-19.2.3-47.gc24117fd5 527 kB/s | 150 kB 00:00 2026-03-06T13:43:35.921 INFO:teuthology.orchestra.run.vm06.stdout:(25/150): rbd-nbd-19.2.3-47.gc24117fd552.el9.cl 1.4 MB/s | 171 kB 00:00 2026-03-06T13:43:35.930 INFO:teuthology.orchestra.run.vm02.stdout:(30/150): ceph-mgr-k8sevents-19.2.3-47.gc24117f 668 kB/s | 22 kB 00:00 2026-03-06T13:43:35.935 INFO:teuthology.orchestra.run.vm04.stdout:(30/150): ceph-mgr-k8sevents-19.2.3-47.gc24117f 146 kB/s | 22 kB 00:00 2026-03-06T13:43:35.936 INFO:teuthology.orchestra.run.vm06.stdout:(26/150): rbd-mirror-19.2.3-47.gc24117fd552.el9 12 MB/s | 3.1 MB 00:00 2026-03-06T13:43:35.959 INFO:teuthology.orchestra.run.vm04.stdout:(31/150): ceph-mgr-diskprediction-local-19.2.3- 5.3 MB/s | 7.4 MB 00:01 2026-03-06T13:43:35.972 INFO:teuthology.orchestra.run.vm06.stdout:(27/150): ceph-mgr-cephadm-19.2.3-47.gc24117fd5 4.0 MB/s | 150 kB 00:00 2026-03-06T13:43:35.996 INFO:teuthology.orchestra.run.vm06.stdout:(28/150): ceph-grafana-dashboards-19.2.3-47.gc2 416 kB/s | 31 kB 00:00 2026-03-06T13:43:36.059 INFO:teuthology.orchestra.run.vm04.stdout:(32/150): ceph-mgr-modules-core-19.2.3-47.gc241 2.0 MB/s | 252 kB 00:00 2026-03-06T13:43:36.094 INFO:teuthology.orchestra.run.vm02.stdout:(31/150): ceph-mgr-dashboard-19.2.3-47.gc24117f 12 MB/s | 3.8 MB 00:00 2026-03-06T13:43:36.096 INFO:teuthology.orchestra.run.vm02.stdout:(32/150): ceph-mgr-modules-core-19.2.3-47.gc241 1.5 MB/s | 252 kB 00:00 2026-03-06T13:43:36.096 INFO:teuthology.orchestra.run.vm04.stdout:(33/150): ceph-mgr-rook-19.2.3-47.gc24117fd552. 360 kB/s | 49 kB 00:00 2026-03-06T13:43:36.126 INFO:teuthology.orchestra.run.vm02.stdout:(33/150): ceph-mgr-rook-19.2.3-47.gc24117fd552. 1.5 MB/s | 49 kB 00:00 2026-03-06T13:43:36.149 INFO:teuthology.orchestra.run.vm04.stdout:(34/150): ceph-volume-19.2.3-47.gc24117fd552.el 5.6 MB/s | 299 kB 00:00 2026-03-06T13:43:36.151 INFO:teuthology.orchestra.run.vm04.stdout:(35/150): ceph-prometheus-alerts-19.2.3-47.gc24 179 kB/s | 16 kB 00:00 2026-03-06T13:43:36.191 INFO:teuthology.orchestra.run.vm04.stdout:(36/150): rbd-mirror-19.2.3-47.gc24117fd552.el9 1.8 MB/s | 3.1 MB 00:01 2026-03-06T13:43:36.191 INFO:teuthology.orchestra.run.vm02.stdout:(34/150): ceph-mgr-diskprediction-local-19.2.3- 19 MB/s | 7.4 MB 00:00 2026-03-06T13:43:36.244 INFO:teuthology.orchestra.run.vm02.stdout:(35/150): ceph-prometheus-alerts-19.2.3-47.gc24 111 kB/s | 16 kB 00:00 2026-03-06T13:43:36.251 INFO:teuthology.orchestra.run.vm02.stdout:(36/150): cephadm-19.2.3-47.gc24117fd552.el9.cl 12 MB/s | 769 kB 00:00 2026-03-06T13:43:36.441 INFO:teuthology.orchestra.run.vm04.stdout:(37/150): cephadm-19.2.3-47.gc24117fd552.el9.cl 2.6 MB/s | 769 kB 00:00 2026-03-06T13:43:36.460 INFO:teuthology.orchestra.run.vm02.stdout:(37/150): ceph-volume-19.2.3-47.gc24117fd552.el 897 kB/s | 299 kB 00:00 2026-03-06T13:43:36.463 INFO:teuthology.orchestra.run.vm04.stdout:(38/150): bzip2-1.0.8-11.el9.x86_64.rpm 175 kB/s | 55 kB 00:00 2026-03-06T13:43:36.548 INFO:teuthology.orchestra.run.vm04.stdout:(39/150): cryptsetup-2.8.1-3.el9.x86_64.rpm 982 kB/s | 351 kB 00:00 2026-03-06T13:43:36.560 INFO:teuthology.orchestra.run.vm06.stdout:(29/150): ceph-mgr-dashboard-19.2.3-47.gc24117f 6.5 MB/s | 3.8 MB 00:00 2026-03-06T13:43:36.609 INFO:teuthology.orchestra.run.vm02.stdout:(38/150): bzip2-1.0.8-11.el9.x86_64.rpm 150 kB/s | 55 kB 00:00 2026-03-06T13:43:36.646 INFO:teuthology.orchestra.run.vm04.stdout:(40/150): ledmon-libs-1.1.0-3.el9.x86_64.rpm 222 kB/s | 40 kB 00:00 2026-03-06T13:43:36.715 INFO:teuthology.orchestra.run.vm02.stdout:(39/150): ledmon-libs-1.1.0-3.el9.x86_64.rpm 383 kB/s | 40 kB 00:00 2026-03-06T13:43:36.715 INFO:teuthology.orchestra.run.vm04.stdout:(41/150): libconfig-1.7.2-9.el9.x86_64.rpm 433 kB/s | 72 kB 00:00 2026-03-06T13:43:36.716 INFO:teuthology.orchestra.run.vm06.stdout:(30/150): ceph-test-19.2.3-47.gc24117fd552.el9. 17 MB/s | 50 MB 00:02 2026-03-06T13:43:36.728 INFO:teuthology.orchestra.run.vm06.stdout:(31/150): ceph-mgr-k8sevents-19.2.3-47.gc24117f 132 kB/s | 22 kB 00:00 2026-03-06T13:43:36.768 INFO:teuthology.orchestra.run.vm06.stdout:(32/150): ceph-mgr-diskprediction-local-19.2.3- 9.6 MB/s | 7.4 MB 00:00 2026-03-06T13:43:36.768 INFO:teuthology.orchestra.run.vm02.stdout:(40/150): cryptsetup-2.8.1-3.el9.x86_64.rpm 680 kB/s | 351 kB 00:00 2026-03-06T13:43:36.768 INFO:teuthology.orchestra.run.vm06.stdout:(33/150): ceph-mgr-rook-19.2.3-47.gc24117fd552. 1.2 MB/s | 49 kB 00:00 2026-03-06T13:43:36.770 INFO:teuthology.orchestra.run.vm06.stdout:(34/150): ceph-mgr-modules-core-19.2.3-47.gc241 4.6 MB/s | 252 kB 00:00 2026-03-06T13:43:36.783 INFO:teuthology.orchestra.run.vm04.stdout:(42/150): fuse-2.9.9-17.el9.x86_64.rpm 233 kB/s | 80 kB 00:00 2026-03-06T13:43:36.796 INFO:teuthology.orchestra.run.vm06.stdout:(35/150): ceph-prometheus-alerts-19.2.3-47.gc24 592 kB/s | 16 kB 00:00 2026-03-06T13:43:36.801 INFO:teuthology.orchestra.run.vm02.stdout:(41/150): libconfig-1.7.2-9.el9.x86_64.rpm 846 kB/s | 72 kB 00:00 2026-03-06T13:43:36.818 INFO:teuthology.orchestra.run.vm04.stdout:(43/150): mailcap-2.1.49-5.el9.noarch.rpm 961 kB/s | 33 kB 00:00 2026-03-06T13:43:36.832 INFO:teuthology.orchestra.run.vm04.stdout:(44/150): libquadmath-11.5.0-14.el9.x86_64.rpm 1.5 MB/s | 184 kB 00:00 2026-03-06T13:43:36.841 INFO:teuthology.orchestra.run.vm02.stdout:(42/150): fuse-2.9.9-17.el9.x86_64.rpm 209 kB/s | 80 kB 00:00 2026-03-06T13:43:36.842 INFO:teuthology.orchestra.run.vm04.stdout:(45/150): libgfortran-11.5.0-14.el9.x86_64.rpm 4.0 MB/s | 794 kB 00:00 2026-03-06T13:43:36.918 INFO:teuthology.orchestra.run.vm04.stdout:(46/150): python3-cffi-1.14.5-5.el9.x86_64.rpm 2.9 MB/s | 253 kB 00:00 2026-03-06T13:43:36.925 INFO:teuthology.orchestra.run.vm02.stdout:(43/150): libgfortran-11.5.0-14.el9.x86_64.rpm 5.0 MB/s | 794 kB 00:00 2026-03-06T13:43:36.925 INFO:teuthology.orchestra.run.vm02.stdout:(44/150): mailcap-2.1.49-5.el9.noarch.rpm 394 kB/s | 33 kB 00:00 2026-03-06T13:43:36.952 INFO:teuthology.orchestra.run.vm02.stdout:(45/150): libquadmath-11.5.0-14.el9.x86_64.rpm 1.2 MB/s | 184 kB 00:00 2026-03-06T13:43:36.956 INFO:teuthology.orchestra.run.vm06.stdout:(36/150): bzip2-1.0.8-11.el9.x86_64.rpm 341 kB/s | 55 kB 00:00 2026-03-06T13:43:36.967 INFO:teuthology.orchestra.run.vm04.stdout:(47/150): python3-ply-3.11-14.el9.noarch.rpm 2.2 MB/s | 106 kB 00:00 2026-03-06T13:43:37.009 INFO:teuthology.orchestra.run.vm06.stdout:(37/150): ceph-volume-19.2.3-47.gc24117fd552.el 1.2 MB/s | 299 kB 00:00 2026-03-06T13:43:37.010 INFO:teuthology.orchestra.run.vm02.stdout:(46/150): pciutils-3.7.0-7.el9.x86_64.rpm 1.1 MB/s | 93 kB 00:00 2026-03-06T13:43:37.027 INFO:teuthology.orchestra.run.vm04.stdout:(48/150): python3-cryptography-36.0.1-5.el9.x86 6.7 MB/s | 1.2 MB 00:00 2026-03-06T13:43:37.048 INFO:teuthology.orchestra.run.vm06.stdout:(38/150): cephadm-19.2.3-47.gc24117fd552.el9.cl 2.7 MB/s | 769 kB 00:00 2026-03-06T13:43:37.094 INFO:teuthology.orchestra.run.vm02.stdout:(47/150): python3-ply-3.11-14.el9.noarch.rpm 1.2 MB/s | 106 kB 00:00 2026-03-06T13:43:37.106 INFO:teuthology.orchestra.run.vm06.stdout:(39/150): cryptsetup-2.8.1-3.el9.x86_64.rpm 2.3 MB/s | 351 kB 00:00 2026-03-06T13:43:37.111 INFO:teuthology.orchestra.run.vm02.stdout:(48/150): python3-cffi-1.14.5-5.el9.x86_64.rpm 1.3 MB/s | 253 kB 00:00 2026-03-06T13:43:37.115 INFO:teuthology.orchestra.run.vm04.stdout:(49/150): python3-pycparser-2.20-6.el9.noarch.r 910 kB/s | 135 kB 00:00 2026-03-06T13:43:37.179 INFO:teuthology.orchestra.run.vm02.stdout:(49/150): python3-pycparser-2.20-6.el9.noarch.r 1.6 MB/s | 135 kB 00:00 2026-03-06T13:43:37.182 INFO:teuthology.orchestra.run.vm02.stdout:(50/150): python3-cryptography-36.0.1-5.el9.x86 5.4 MB/s | 1.2 MB 00:00 2026-03-06T13:43:37.186 INFO:teuthology.orchestra.run.vm06.stdout:(40/150): fuse-2.9.9-17.el9.x86_64.rpm 451 kB/s | 80 kB 00:00 2026-03-06T13:43:37.187 INFO:teuthology.orchestra.run.vm06.stdout:(41/150): libconfig-1.7.2-9.el9.x86_64.rpm 888 kB/s | 72 kB 00:00 2026-03-06T13:43:37.196 INFO:teuthology.orchestra.run.vm06.stdout:(42/150): ledmon-libs-1.1.0-3.el9.x86_64.rpm 274 kB/s | 40 kB 00:00 2026-03-06T13:43:37.233 INFO:teuthology.orchestra.run.vm02.stdout:(51/150): python3-requests-2.25.1-10.el9.noarch 1.0 MB/s | 126 kB 00:00 2026-03-06T13:43:37.233 INFO:teuthology.orchestra.run.vm06.stdout:(43/150): mailcap-2.1.49-5.el9.noarch.rpm 896 kB/s | 33 kB 00:00 2026-03-06T13:43:37.258 INFO:teuthology.orchestra.run.vm06.stdout:(44/150): libquadmath-11.5.0-14.el9.x86_64.rpm 2.6 MB/s | 184 kB 00:00 2026-03-06T13:43:37.263 INFO:teuthology.orchestra.run.vm02.stdout:(52/150): python3-urllib3-1.26.5-7.el9.noarch.r 2.5 MB/s | 218 kB 00:00 2026-03-06T13:43:37.284 INFO:teuthology.orchestra.run.vm02.stdout:(53/150): smartmontools-7.2-10.el9.x86_64.rpm 5.3 MB/s | 556 kB 00:00 2026-03-06T13:43:37.295 INFO:teuthology.orchestra.run.vm04.stdout:(50/150): pciutils-3.7.0-7.el9.x86_64.rpm 195 kB/s | 93 kB 00:00 2026-03-06T13:43:37.317 INFO:teuthology.orchestra.run.vm02.stdout:(54/150): unzip-6.0-59.el9.x86_64.rpm 2.1 MB/s | 182 kB 00:00 2026-03-06T13:43:37.336 INFO:teuthology.orchestra.run.vm06.stdout:(45/150): pciutils-3.7.0-7.el9.x86_64.rpm 908 kB/s | 93 kB 00:00 2026-03-06T13:43:37.344 INFO:teuthology.orchestra.run.vm04.stdout:(51/150): python3-requests-2.25.1-10.el9.noarch 399 kB/s | 126 kB 00:00 2026-03-06T13:43:37.349 INFO:teuthology.orchestra.run.vm06.stdout:(46/150): python3-cffi-1.14.5-5.el9.x86_64.rpm 2.7 MB/s | 253 kB 00:00 2026-03-06T13:43:37.351 INFO:teuthology.orchestra.run.vm02.stdout:(55/150): zip-3.0-35.el9.x86_64.rpm 3.0 MB/s | 266 kB 00:00 2026-03-06T13:43:37.366 INFO:teuthology.orchestra.run.vm04.stdout:(52/150): python3-urllib3-1.26.5-7.el9.noarch.r 868 kB/s | 218 kB 00:00 2026-03-06T13:43:37.383 INFO:teuthology.orchestra.run.vm02.stdout:(56/150): flexiblas-3.0.4-9.el9.x86_64.rpm 452 kB/s | 30 kB 00:00 2026-03-06T13:43:37.386 INFO:teuthology.orchestra.run.vm06.stdout:(47/150): libgfortran-11.5.0-14.el9.x86_64.rpm 3.9 MB/s | 794 kB 00:00 2026-03-06T13:43:37.388 INFO:teuthology.orchestra.run.vm06.stdout:(48/150): python3-ply-3.11-14.el9.noarch.rpm 2.7 MB/s | 106 kB 00:00 2026-03-06T13:43:37.391 INFO:teuthology.orchestra.run.vm02.stdout:(57/150): boost-program-options-1.75.0-13.el9.x 977 kB/s | 104 kB 00:00 2026-03-06T13:43:37.408 INFO:teuthology.orchestra.run.vm04.stdout:(53/150): unzip-6.0-59.el9.x86_64.rpm 2.8 MB/s | 182 kB 00:00 2026-03-06T13:43:37.426 INFO:teuthology.orchestra.run.vm02.stdout:(58/150): flexiblas-openblas-openmp-3.0.4-9.el9 347 kB/s | 15 kB 00:00 2026-03-06T13:43:37.430 INFO:teuthology.orchestra.run.vm06.stdout:(49/150): python3-pycparser-2.20-6.el9.noarch.r 3.1 MB/s | 135 kB 00:00 2026-03-06T13:43:37.438 INFO:teuthology.orchestra.run.vm04.stdout:(54/150): zip-3.0-35.el9.x86_64.rpm 3.6 MB/s | 266 kB 00:00 2026-03-06T13:43:37.452 INFO:teuthology.orchestra.run.vm02.stdout:(59/150): libnbd-1.20.3-4.el9.x86_64.rpm 2.7 MB/s | 164 kB 00:00 2026-03-06T13:43:37.463 INFO:teuthology.orchestra.run.vm06.stdout:(50/150): python3-requests-2.25.1-10.el9.noarch 1.7 MB/s | 126 kB 00:00 2026-03-06T13:43:37.484 INFO:teuthology.orchestra.run.vm06.stdout:(51/150): python3-cryptography-36.0.1-5.el9.x86 8.5 MB/s | 1.2 MB 00:00 2026-03-06T13:43:37.489 INFO:teuthology.orchestra.run.vm02.stdout:(60/150): librabbitmq-0.11.0-7.el9.x86_64.rpm 1.2 MB/s | 45 kB 00:00 2026-03-06T13:43:37.509 INFO:teuthology.orchestra.run.vm06.stdout:(52/150): python3-urllib3-1.26.5-7.el9.noarch.r 2.7 MB/s | 218 kB 00:00 2026-03-06T13:43:37.512 INFO:teuthology.orchestra.run.vm02.stdout:(61/150): libpmemobj-1.12.1-1.el9.x86_64.rpm 1.8 MB/s | 160 kB 00:00 2026-03-06T13:43:37.541 INFO:teuthology.orchestra.run.vm02.stdout:(62/150): flexiblas-netlib-3.0.4-9.el9.x86_64.r 16 MB/s | 3.0 MB 00:00 2026-03-06T13:43:37.546 INFO:teuthology.orchestra.run.vm06.stdout:(53/150): unzip-6.0-59.el9.x86_64.rpm 2.9 MB/s | 182 kB 00:00 2026-03-06T13:43:37.548 INFO:teuthology.orchestra.run.vm02.stdout:(63/150): librdkafka-1.6.1-102.el9.x86_64.rpm 11 MB/s | 662 kB 00:00 2026-03-06T13:43:37.561 INFO:teuthology.orchestra.run.vm02.stdout:(64/150): libxslt-1.1.34-12.el9.x86_64.rpm 12 MB/s | 233 kB 00:00 2026-03-06T13:43:37.564 INFO:teuthology.orchestra.run.vm02.stdout:(65/150): libstoragemgmt-1.10.1-1.el9.x86_64.rp 4.6 MB/s | 246 kB 00:00 2026-03-06T13:43:37.580 INFO:teuthology.orchestra.run.vm06.stdout:(54/150): smartmontools-7.2-10.el9.x86_64.rpm 4.6 MB/s | 556 kB 00:00 2026-03-06T13:43:37.581 INFO:teuthology.orchestra.run.vm04.stdout:(55/150): flexiblas-3.0.4-9.el9.x86_64.rpm 208 kB/s | 30 kB 00:00 2026-03-06T13:43:37.592 INFO:teuthology.orchestra.run.vm02.stdout:(66/150): lttng-ust-2.12.0-6.el9.x86_64.rpm 6.5 MB/s | 292 kB 00:00 2026-03-06T13:43:37.593 INFO:teuthology.orchestra.run.vm02.stdout:(67/150): lua-5.4.4-4.el9.x86_64.rpm 5.7 MB/s | 188 kB 00:00 2026-03-06T13:43:37.597 INFO:teuthology.orchestra.run.vm02.stdout:(68/150): openblas-0.3.29-1.el9.x86_64.rpm 1.3 MB/s | 42 kB 00:00 2026-03-06T13:43:37.623 INFO:teuthology.orchestra.run.vm04.stdout:(56/150): smartmontools-7.2-10.el9.x86_64.rpm 1.7 MB/s | 556 kB 00:00 2026-03-06T13:43:37.646 INFO:teuthology.orchestra.run.vm04.stdout:(57/150): boost-program-options-1.75.0-13.el9.x 438 kB/s | 104 kB 00:00 2026-03-06T13:43:37.655 INFO:teuthology.orchestra.run.vm06.stdout:(55/150): flexiblas-3.0.4-9.el9.x86_64.rpm 398 kB/s | 30 kB 00:00 2026-03-06T13:43:37.655 INFO:teuthology.orchestra.run.vm02.stdout:(69/150): perl-Test-Harness-3.42-461.el9.noarch 5.0 MB/s | 295 kB 00:00 2026-03-06T13:43:37.685 INFO:teuthology.orchestra.run.vm06.stdout:(56/150): zip-3.0-35.el9.x86_64.rpm 1.5 MB/s | 266 kB 00:00 2026-03-06T13:43:37.692 INFO:teuthology.orchestra.run.vm02.stdout:(70/150): perl-Benchmark-1.23-483.el9.noarch.rp 268 kB/s | 26 kB 00:00 2026-03-06T13:43:37.697 INFO:teuthology.orchestra.run.vm06.stdout:(57/150): boost-program-options-1.75.0-13.el9.x 691 kB/s | 104 kB 00:00 2026-03-06T13:43:37.723 INFO:teuthology.orchestra.run.vm02.stdout:(71/150): openblas-openmp-0.3.29-1.el9.x86_64.r 40 MB/s | 5.3 MB 00:00 2026-03-06T13:43:37.734 INFO:teuthology.orchestra.run.vm02.stdout:(72/150): protobuf-3.14.0-17.el9.x86_64.rpm 13 MB/s | 1.0 MB 00:00 2026-03-06T13:43:37.743 INFO:teuthology.orchestra.run.vm04.stdout:(58/150): libnbd-1.20.3-4.el9.x86_64.rpm 1.7 MB/s | 164 kB 00:00 2026-03-06T13:43:37.748 INFO:teuthology.orchestra.run.vm06.stdout:(58/150): libnbd-1.20.3-4.el9.x86_64.rpm 3.1 MB/s | 164 kB 00:00 2026-03-06T13:43:37.753 INFO:teuthology.orchestra.run.vm06.stdout:(59/150): flexiblas-openblas-openmp-3.0.4-9.el9 220 kB/s | 15 kB 00:00 2026-03-06T13:43:37.766 INFO:teuthology.orchestra.run.vm04.stdout:(59/150): flexiblas-openblas-openmp-3.0.4-9.el9 104 kB/s | 15 kB 00:00 2026-03-06T13:43:37.790 INFO:teuthology.orchestra.run.vm02.stdout:(73/150): python3-jinja2-2.11.3-8.el9.noarch.rp 4.4 MB/s | 249 kB 00:00 2026-03-06T13:43:37.791 INFO:teuthology.orchestra.run.vm06.stdout:(60/150): libpmemobj-1.12.1-1.el9.x86_64.rpm 3.7 MB/s | 160 kB 00:00 2026-03-06T13:43:37.803 INFO:teuthology.orchestra.run.vm04.stdout:(60/150): libpmemobj-1.12.1-1.el9.x86_64.rpm 2.6 MB/s | 160 kB 00:00 2026-03-06T13:43:37.810 INFO:teuthology.orchestra.run.vm06.stdout:(61/150): librabbitmq-0.11.0-7.el9.x86_64.rpm 795 kB/s | 45 kB 00:00 2026-03-06T13:43:37.857 INFO:teuthology.orchestra.run.vm02.stdout:(74/150): python3-devel-3.9.25-3.el9.x86_64.rpm 1.8 MB/s | 244 kB 00:00 2026-03-06T13:43:37.866 INFO:teuthology.orchestra.run.vm04.stdout:(61/150): librabbitmq-0.11.0-7.el9.x86_64.rpm 456 kB/s | 45 kB 00:00 2026-03-06T13:43:37.866 INFO:teuthology.orchestra.run.vm06.stdout:(62/150): librdkafka-1.6.1-102.el9.x86_64.rpm 8.7 MB/s | 662 kB 00:00 2026-03-06T13:43:37.878 INFO:teuthology.orchestra.run.vm06.stdout:(63/150): flexiblas-netlib-3.0.4-9.el9.x86_64.r 13 MB/s | 3.0 MB 00:00 2026-03-06T13:43:37.899 INFO:teuthology.orchestra.run.vm06.stdout:(64/150): libxslt-1.1.34-12.el9.x86_64.rpm 6.9 MB/s | 233 kB 00:00 2026-03-06T13:43:37.905 INFO:teuthology.orchestra.run.vm04.stdout:(62/150): librdkafka-1.6.1-102.el9.x86_64.rpm 6.4 MB/s | 662 kB 00:00 2026-03-06T13:43:37.922 INFO:teuthology.orchestra.run.vm02.stdout:(75/150): python3-babel-2.9.1-2.el9.noarch.rpm 26 MB/s | 6.0 MB 00:00 2026-03-06T13:43:37.949 INFO:teuthology.orchestra.run.vm06.stdout:(65/150): lttng-ust-2.12.0-6.el9.x86_64.rpm 4.1 MB/s | 292 kB 00:00 2026-03-06T13:43:37.957 INFO:teuthology.orchestra.run.vm04.stdout:(63/150): libxslt-1.1.34-12.el9.x86_64.rpm 4.4 MB/s | 233 kB 00:00 2026-03-06T13:43:37.960 INFO:teuthology.orchestra.run.vm02.stdout:(76/150): python3-libstoragemgmt-1.10.1-1.el9.x 1.0 MB/s | 177 kB 00:00 2026-03-06T13:43:37.976 INFO:teuthology.orchestra.run.vm06.stdout:(66/150): openblas-0.3.29-1.el9.x86_64.rpm 1.5 MB/s | 42 kB 00:00 2026-03-06T13:43:37.987 INFO:teuthology.orchestra.run.vm06.stdout:(67/150): libstoragemgmt-1.10.1-1.el9.x86_64.rp 1.4 MB/s | 246 kB 00:00 2026-03-06T13:43:37.997 INFO:teuthology.orchestra.run.vm02.stdout:(77/150): python3-markupsafe-1.1.1-12.el9.x86_6 974 kB/s | 35 kB 00:00 2026-03-06T13:43:38.013 INFO:teuthology.orchestra.run.vm04.stdout:(64/150): lttng-ust-2.12.0-6.el9.x86_64.rpm 5.2 MB/s | 292 kB 00:00 2026-03-06T13:43:38.014 INFO:teuthology.orchestra.run.vm04.stdout:(65/150): libstoragemgmt-1.10.1-1.el9.x86_64.rp 1.6 MB/s | 246 kB 00:00 2026-03-06T13:43:38.016 INFO:teuthology.orchestra.run.vm06.stdout:(68/150): lua-5.4.4-4.el9.x86_64.rpm 1.6 MB/s | 188 kB 00:00 2026-03-06T13:43:38.034 INFO:teuthology.orchestra.run.vm06.stdout:(69/150): perl-Benchmark-1.23-483.el9.noarch.rp 567 kB/s | 26 kB 00:00 2026-03-06T13:43:38.068 INFO:teuthology.orchestra.run.vm04.stdout:(66/150): lua-5.4.4-4.el9.x86_64.rpm 3.3 MB/s | 188 kB 00:00 2026-03-06T13:43:38.069 INFO:teuthology.orchestra.run.vm04.stdout:(67/150): openblas-0.3.29-1.el9.x86_64.rpm 773 kB/s | 42 kB 00:00 2026-03-06T13:43:38.072 INFO:teuthology.orchestra.run.vm06.stdout:(70/150): perl-Test-Harness-3.42-461.el9.noarch 5.2 MB/s | 295 kB 00:00 2026-03-06T13:43:38.081 INFO:teuthology.orchestra.run.vm02.stdout:(78/150): python3-lxml-4.6.5-3.el9.x86_64.rpm 5.5 MB/s | 1.2 MB 00:00 2026-03-06T13:43:38.096 INFO:teuthology.orchestra.run.vm06.stdout:(71/150): openblas-openmp-0.3.29-1.el9.x86_64.r 44 MB/s | 5.3 MB 00:00 2026-03-06T13:43:38.126 INFO:teuthology.orchestra.run.vm04.stdout:(68/150): perl-Benchmark-1.23-483.el9.noarch.rp 464 kB/s | 26 kB 00:00 2026-03-06T13:43:38.144 INFO:teuthology.orchestra.run.vm02.stdout:(79/150): python3-mako-1.1.4-6.el9.noarch.rpm 777 kB/s | 172 kB 00:00 2026-03-06T13:43:38.176 INFO:teuthology.orchestra.run.vm06.stdout:(72/150): protobuf-3.14.0-17.el9.x86_64.rpm 7.1 MB/s | 1.0 MB 00:00 2026-03-06T13:43:38.231 INFO:teuthology.orchestra.run.vm06.stdout:(73/150): python3-devel-3.9.25-3.el9.x86_64.rpm 1.8 MB/s | 244 kB 00:00 2026-03-06T13:43:38.233 INFO:teuthology.orchestra.run.vm06.stdout:(74/150): python3-jinja2-2.11.3-8.el9.noarch.rp 4.3 MB/s | 249 kB 00:00 2026-03-06T13:43:38.243 INFO:teuthology.orchestra.run.vm02.stdout:(80/150): python3-numpy-f2py-1.23.5-2.el9.x86_6 2.7 MB/s | 442 kB 00:00 2026-03-06T13:43:38.268 INFO:teuthology.orchestra.run.vm02.stdout:(81/150): python3-numpy-1.23.5-2.el9.x86_64.rpm 23 MB/s | 6.1 MB 00:00 2026-03-06T13:43:38.268 INFO:teuthology.orchestra.run.vm06.stdout:(75/150): python3-babel-2.9.1-2.el9.noarch.rpm 30 MB/s | 6.0 MB 00:00 2026-03-06T13:43:38.270 INFO:teuthology.orchestra.run.vm06.stdout:(76/150): python3-libstoragemgmt-1.10.1-1.el9.x 4.5 MB/s | 177 kB 00:00 2026-03-06T13:43:38.272 INFO:teuthology.orchestra.run.vm02.stdout:(82/150): python3-packaging-20.9-5.el9.noarch.r 605 kB/s | 77 kB 00:00 2026-03-06T13:43:38.286 INFO:teuthology.orchestra.run.vm06.stdout:(77/150): python3-lxml-4.6.5-3.el9.x86_64.rpm 23 MB/s | 1.2 MB 00:00 2026-03-06T13:43:38.298 INFO:teuthology.orchestra.run.vm06.stdout:(78/150): python3-markupsafe-1.1.1-12.el9.x86_6 1.2 MB/s | 35 kB 00:00 2026-03-06T13:43:38.306 INFO:teuthology.orchestra.run.vm04.stdout:(69/150): flexiblas-netlib-3.0.4-9.el9.x86_64.r 4.1 MB/s | 3.0 MB 00:00 2026-03-06T13:43:38.306 INFO:teuthology.orchestra.run.vm06.stdout:(79/150): python3-mako-1.1.4-6.el9.noarch.rpm 4.5 MB/s | 172 kB 00:00 2026-03-06T13:43:38.311 INFO:teuthology.orchestra.run.vm02.stdout:(83/150): python3-protobuf-3.14.0-17.el9.noarch 3.8 MB/s | 267 kB 00:00 2026-03-06T13:43:38.319 INFO:teuthology.orchestra.run.vm02.stdout:(84/150): python3-pyasn1-0.4.8-7.el9.noarch.rpm 3.0 MB/s | 157 kB 00:00 2026-03-06T13:43:38.334 INFO:teuthology.orchestra.run.vm06.stdout:(80/150): python3-packaging-20.9-5.el9.noarch.r 2.8 MB/s | 77 kB 00:00 2026-03-06T13:43:38.378 INFO:teuthology.orchestra.run.vm02.stdout:(85/150): python3-requests-oauthlib-1.3.0-12.el 806 kB/s | 54 kB 00:00 2026-03-06T13:43:38.388 INFO:teuthology.orchestra.run.vm02.stdout:(86/150): python3-pyasn1-modules-0.4.8-7.el9.no 2.3 MB/s | 277 kB 00:00 2026-03-06T13:43:38.398 INFO:teuthology.orchestra.run.vm06.stdout:(81/150): python3-protobuf-3.14.0-17.el9.noarch 4.1 MB/s | 267 kB 00:00 2026-03-06T13:43:38.417 INFO:teuthology.orchestra.run.vm06.stdout:(82/150): python3-numpy-1.23.5-2.el9.x86_64.rpm 47 MB/s | 6.1 MB 00:00 2026-03-06T13:43:38.418 INFO:teuthology.orchestra.run.vm02.stdout:(87/150): python3-toml-0.10.2-6.el9.noarch.rpm 1.0 MB/s | 42 kB 00:00 2026-03-06T13:43:38.419 INFO:teuthology.orchestra.run.vm06.stdout:(83/150): python3-numpy-f2py-1.23.5-2.el9.x86_6 3.6 MB/s | 442 kB 00:00 2026-03-06T13:43:38.424 INFO:teuthology.orchestra.run.vm06.stdout:(84/150): python3-pyasn1-0.4.8-7.el9.noarch.rpm 5.9 MB/s | 157 kB 00:00 2026-03-06T13:43:38.446 INFO:teuthology.orchestra.run.vm06.stdout:(85/150): python3-pyasn1-modules-0.4.8-7.el9.no 9.6 MB/s | 277 kB 00:00 2026-03-06T13:43:38.447 INFO:teuthology.orchestra.run.vm06.stdout:(86/150): python3-requests-oauthlib-1.3.0-12.el 1.9 MB/s | 54 kB 00:00 2026-03-06T13:43:38.460 INFO:teuthology.orchestra.run.vm02.stdout:(88/150): qatlib-25.08.0-2.el9.x86_64.rpm 3.2 MB/s | 240 kB 00:00 2026-03-06T13:43:38.489 INFO:teuthology.orchestra.run.vm02.stdout:(89/150): qatlib-service-25.08.0-2.el9.x86_64.r 519 kB/s | 37 kB 00:00 2026-03-06T13:43:38.502 INFO:teuthology.orchestra.run.vm02.stdout:(90/150): qatzip-libs-1.3.1-1.el9.x86_64.rpm 1.6 MB/s | 66 kB 00:00 2026-03-06T13:43:38.564 INFO:teuthology.orchestra.run.vm06.stdout:(87/150): qatlib-25.08.0-2.el9.x86_64.rpm 2.0 MB/s | 240 kB 00:00 2026-03-06T13:43:38.574 INFO:teuthology.orchestra.run.vm02.stdout:(91/150): socat-1.7.4.1-8.el9.x86_64.rpm 3.5 MB/s | 303 kB 00:00 2026-03-06T13:43:38.576 INFO:teuthology.orchestra.run.vm02.stdout:(92/150): xmlsec1-1.2.29-13.el9.x86_64.rpm 2.5 MB/s | 189 kB 00:00 2026-03-06T13:43:38.594 INFO:teuthology.orchestra.run.vm02.stdout:(93/150): xmlsec1-openssl-1.2.29-13.el9.x86_64. 4.5 MB/s | 90 kB 00:00 2026-03-06T13:43:38.606 INFO:teuthology.orchestra.run.vm04.stdout:(70/150): protobuf-3.14.0-17.el9.x86_64.rpm 3.4 MB/s | 1.0 MB 00:00 2026-03-06T13:43:38.663 INFO:teuthology.orchestra.run.vm06.stdout:(88/150): qatlib-service-25.08.0-2.el9.x86_64.r 374 kB/s | 37 kB 00:00 2026-03-06T13:43:38.739 INFO:teuthology.orchestra.run.vm02.stdout:(94/150): xmlstarlet-1.6.1-20.el9.x86_64.rpm 391 kB/s | 64 kB 00:00 2026-03-06T13:43:38.781 INFO:teuthology.orchestra.run.vm06.stdout:(89/150): python3-scipy-1.9.3-2.el9.x86_64.rpm 54 MB/s | 19 MB 00:00 2026-03-06T13:43:38.781 INFO:teuthology.orchestra.run.vm06.stdout:(90/150): qatzip-libs-1.3.1-1.el9.x86_64.rpm 563 kB/s | 66 kB 00:00 2026-03-06T13:43:38.792 INFO:teuthology.orchestra.run.vm02.stdout:(95/150): lua-devel-5.4.4-4.el9.x86_64.rpm 112 kB/s | 22 kB 00:00 2026-03-06T13:43:38.815 INFO:teuthology.orchestra.run.vm02.stdout:(96/150): abseil-cpp-20211102.0-4.el9.x86_64.rp 24 MB/s | 551 kB 00:00 2026-03-06T13:43:38.824 INFO:teuthology.orchestra.run.vm02.stdout:(97/150): gperftools-libs-2.9.1-3.el9.x86_64.rp 35 MB/s | 308 kB 00:00 2026-03-06T13:43:38.827 INFO:teuthology.orchestra.run.vm02.stdout:(98/150): grpc-data-1.46.7-10.el9.noarch.rpm 7.5 MB/s | 19 kB 00:00 2026-03-06T13:43:38.827 INFO:teuthology.orchestra.run.vm04.stdout:(71/150): perl-Test-Harness-3.42-461.el9.noarch 421 kB/s | 295 kB 00:00 2026-03-06T13:43:38.849 INFO:teuthology.orchestra.run.vm06.stdout:(91/150): python3-toml-0.10.2-6.el9.noarch.rpm 103 kB/s | 42 kB 00:00 2026-03-06T13:43:38.876 INFO:teuthology.orchestra.run.vm06.stdout:(92/150): socat-1.7.4.1-8.el9.x86_64.rpm 3.1 MB/s | 303 kB 00:00 2026-03-06T13:43:38.882 INFO:teuthology.orchestra.run.vm04.stdout:(72/150): python3-devel-3.9.25-3.el9.x86_64.rpm 4.4 MB/s | 244 kB 00:00 2026-03-06T13:43:38.884 INFO:teuthology.orchestra.run.vm06.stdout:(93/150): xmlsec1-openssl-1.2.29-13.el9.x86_64. 2.5 MB/s | 90 kB 00:00 2026-03-06T13:43:38.899 INFO:teuthology.orchestra.run.vm02.stdout:(99/150): libarrow-9.0.0-15.el9.x86_64.rpm 62 MB/s | 4.4 MB 00:00 2026-03-06T13:43:38.901 INFO:teuthology.orchestra.run.vm02.stdout:(100/150): libarrow-doc-9.0.0-15.el9.noarch.rpm 10 MB/s | 25 kB 00:00 2026-03-06T13:43:38.904 INFO:teuthology.orchestra.run.vm02.stdout:(101/150): liboath-2.6.12-1.el9.x86_64.rpm 17 MB/s | 49 kB 00:00 2026-03-06T13:43:38.907 INFO:teuthology.orchestra.run.vm02.stdout:(102/150): libunwind-1.6.2-1.el9.x86_64.rpm 24 MB/s | 67 kB 00:00 2026-03-06T13:43:38.912 INFO:teuthology.orchestra.run.vm02.stdout:(103/150): luarocks-3.9.2-5.el9.noarch.rpm 40 MB/s | 151 kB 00:00 2026-03-06T13:43:38.922 INFO:teuthology.orchestra.run.vm06.stdout:(94/150): xmlstarlet-1.6.1-20.el9.x86_64.rpm 1.3 MB/s | 64 kB 00:00 2026-03-06T13:43:38.925 INFO:teuthology.orchestra.run.vm02.stdout:(104/150): parquet-libs-9.0.0-15.el9.x86_64.rpm 62 MB/s | 838 kB 00:00 2026-03-06T13:43:38.935 INFO:teuthology.orchestra.run.vm02.stdout:(105/150): python3-asyncssh-2.13.2-5.el9.noarch 55 MB/s | 548 kB 00:00 2026-03-06T13:43:38.945 INFO:teuthology.orchestra.run.vm02.stdout:(106/150): python3-autocommand-2.2.2-8.el9.noar 3.1 MB/s | 29 kB 00:00 2026-03-06T13:43:38.948 INFO:teuthology.orchestra.run.vm02.stdout:(107/150): python3-backports-tarfile-1.2.0-1.el 18 MB/s | 60 kB 00:00 2026-03-06T13:43:38.951 INFO:teuthology.orchestra.run.vm02.stdout:(108/150): python3-bcrypt-3.2.2-1.el9.x86_64.rp 17 MB/s | 43 kB 00:00 2026-03-06T13:43:38.954 INFO:teuthology.orchestra.run.vm02.stdout:(109/150): python3-cachetools-4.2.4-1.el9.noarc 13 MB/s | 32 kB 00:00 2026-03-06T13:43:38.957 INFO:teuthology.orchestra.run.vm02.stdout:(110/150): python3-certifi-2023.05.07-4.el9.noa 5.7 MB/s | 14 kB 00:00 2026-03-06T13:43:38.966 INFO:teuthology.orchestra.run.vm04.stdout:(73/150): openblas-openmp-0.3.29-1.el9.x86_64.r 5.9 MB/s | 5.3 MB 00:00 2026-03-06T13:43:38.966 INFO:teuthology.orchestra.run.vm02.stdout:(111/150): python3-cheroot-10.0.1-4.el9.noarch. 19 MB/s | 173 kB 00:00 2026-03-06T13:43:38.973 INFO:teuthology.orchestra.run.vm02.stdout:(112/150): python3-cherrypy-18.6.1-2.el9.noarch 54 MB/s | 358 kB 00:00 2026-03-06T13:43:38.978 INFO:teuthology.orchestra.run.vm02.stdout:(113/150): python3-google-auth-2.45.0-1.el9.noa 51 MB/s | 254 kB 00:00 2026-03-06T13:43:38.979 INFO:teuthology.orchestra.run.vm06.stdout:(95/150): xmlsec1-1.2.29-13.el9.x86_64.rpm 958 kB/s | 189 kB 00:00 2026-03-06T13:43:38.979 INFO:teuthology.orchestra.run.vm04.stdout:(74/150): python3-jinja2-2.11.3-8.el9.noarch.rp 2.5 MB/s | 249 kB 00:00 2026-03-06T13:43:39.009 INFO:teuthology.orchestra.run.vm06.stdout:(96/150): abseil-cpp-20211102.0-4.el9.x86_64.rp 18 MB/s | 551 kB 00:00 2026-03-06T13:43:39.016 INFO:teuthology.orchestra.run.vm02.stdout:(114/150): python3-grpcio-1.46.7-10.el9.x86_64. 55 MB/s | 2.0 MB 00:00 2026-03-06T13:43:39.016 INFO:teuthology.orchestra.run.vm06.stdout:(97/150): gperftools-libs-2.9.1-3.el9.x86_64.rp 44 MB/s | 308 kB 00:00 2026-03-06T13:43:39.019 INFO:teuthology.orchestra.run.vm06.stdout:(98/150): grpc-data-1.46.7-10.el9.noarch.rpm 6.7 MB/s | 19 kB 00:00 2026-03-06T13:43:39.019 INFO:teuthology.orchestra.run.vm02.stdout:(115/150): python3-grpcio-tools-1.46.7-10.el9.x 39 MB/s | 144 kB 00:00 2026-03-06T13:43:39.027 INFO:teuthology.orchestra.run.vm02.stdout:(116/150): python3-influxdb-5.3.1-1.el9.noarch. 20 MB/s | 139 kB 00:00 2026-03-06T13:43:39.036 INFO:teuthology.orchestra.run.vm02.stdout:(117/150): python3-isodate-0.6.1-3.el9.noarch.r 6.0 MB/s | 56 kB 00:00 2026-03-06T13:43:39.044 INFO:teuthology.orchestra.run.vm02.stdout:(118/150): python3-jaraco-8.2.1-3.el9.noarch.rp 1.3 MB/s | 11 kB 00:00 2026-03-06T13:43:39.049 INFO:teuthology.orchestra.run.vm02.stdout:(119/150): python3-jaraco-classes-3.2.1-5.el9.n 3.6 MB/s | 18 kB 00:00 2026-03-06T13:43:39.054 INFO:teuthology.orchestra.run.vm02.stdout:(120/150): python3-jaraco-collections-3.0.0-8.e 5.0 MB/s | 23 kB 00:00 2026-03-06T13:43:39.059 INFO:teuthology.orchestra.run.vm02.stdout:(121/150): python3-jaraco-context-6.0.1-3.el9.n 4.2 MB/s | 20 kB 00:00 2026-03-06T13:43:39.064 INFO:teuthology.orchestra.run.vm02.stdout:(122/150): python3-jaraco-functools-3.5.0-2.el9 4.4 MB/s | 19 kB 00:00 2026-03-06T13:43:39.066 INFO:teuthology.orchestra.run.vm04.stdout:(75/150): python3-libstoragemgmt-1.10.1-1.el9.x 1.7 MB/s | 177 kB 00:00 2026-03-06T13:43:39.069 INFO:teuthology.orchestra.run.vm02.stdout:(123/150): python3-jaraco-text-4.0.0-2.el9.noar 4.8 MB/s | 26 kB 00:00 2026-03-06T13:43:39.086 INFO:teuthology.orchestra.run.vm06.stdout:(99/150): libarrow-9.0.0-15.el9.x86_64.rpm 66 MB/s | 4.4 MB 00:00 2026-03-06T13:43:39.088 INFO:teuthology.orchestra.run.vm02.stdout:(124/150): python3-kubernetes-26.1.0-3.el9.noar 56 MB/s | 1.0 MB 00:00 2026-03-06T13:43:39.089 INFO:teuthology.orchestra.run.vm06.stdout:(100/150): libarrow-doc-9.0.0-15.el9.noarch.rpm 11 MB/s | 25 kB 00:00 2026-03-06T13:43:39.091 INFO:teuthology.orchestra.run.vm02.stdout:(125/150): python3-logutils-0.3.5-21.el9.noarch 14 MB/s | 46 kB 00:00 2026-03-06T13:43:39.091 INFO:teuthology.orchestra.run.vm06.stdout:(101/150): liboath-2.6.12-1.el9.x86_64.rpm 18 MB/s | 49 kB 00:00 2026-03-06T13:43:39.094 INFO:teuthology.orchestra.run.vm06.stdout:(102/150): libunwind-1.6.2-1.el9.x86_64.rpm 25 MB/s | 67 kB 00:00 2026-03-06T13:43:39.095 INFO:teuthology.orchestra.run.vm02.stdout:(126/150): python3-more-itertools-8.12.0-2.el9. 22 MB/s | 79 kB 00:00 2026-03-06T13:43:39.098 INFO:teuthology.orchestra.run.vm06.stdout:(103/150): luarocks-3.9.2-5.el9.noarch.rpm 43 MB/s | 151 kB 00:00 2026-03-06T13:43:39.109 INFO:teuthology.orchestra.run.vm02.stdout:(127/150): python3-msgpack-1.0.3-2.el9.x86_64.r 6.5 MB/s | 86 kB 00:00 2026-03-06T13:43:39.111 INFO:teuthology.orchestra.run.vm06.stdout:(104/150): parquet-libs-9.0.0-15.el9.x86_64.rpm 64 MB/s | 838 kB 00:00 2026-03-06T13:43:39.112 INFO:teuthology.orchestra.run.vm02.stdout:(128/150): python3-natsort-7.1.1-5.el9.noarch.r 16 MB/s | 58 kB 00:00 2026-03-06T13:43:39.121 INFO:teuthology.orchestra.run.vm06.stdout:(105/150): python3-asyncssh-2.13.2-5.el9.noarch 57 MB/s | 548 kB 00:00 2026-03-06T13:43:39.123 INFO:teuthology.orchestra.run.vm06.stdout:(106/150): python3-autocommand-2.2.2-8.el9.noar 13 MB/s | 29 kB 00:00 2026-03-06T13:43:39.128 INFO:teuthology.orchestra.run.vm06.stdout:(107/150): python3-backports-tarfile-1.2.0-1.el 13 MB/s | 60 kB 00:00 2026-03-06T13:43:39.129 INFO:teuthology.orchestra.run.vm02.stdout:(129/150): python3-pecan-1.4.2-3.el9.noarch.rpm 16 MB/s | 272 kB 00:00 2026-03-06T13:43:39.131 INFO:teuthology.orchestra.run.vm06.stdout:(108/150): python3-bcrypt-3.2.2-1.el9.x86_64.rp 14 MB/s | 43 kB 00:00 2026-03-06T13:43:39.132 INFO:teuthology.orchestra.run.vm02.stdout:(130/150): python3-portend-3.1.0-2.el9.noarch.r 5.9 MB/s | 16 kB 00:00 2026-03-06T13:43:39.134 INFO:teuthology.orchestra.run.vm06.stdout:(109/150): python3-cachetools-4.2.4-1.el9.noarc 12 MB/s | 32 kB 00:00 2026-03-06T13:43:39.137 INFO:teuthology.orchestra.run.vm02.stdout:(131/150): python3-pyOpenSSL-21.0.0-1.el9.noarc 23 MB/s | 90 kB 00:00 2026-03-06T13:43:39.144 INFO:teuthology.orchestra.run.vm06.stdout:(110/150): python3-certifi-2023.05.07-4.el9.noa 1.4 MB/s | 14 kB 00:00 2026-03-06T13:43:39.144 INFO:teuthology.orchestra.run.vm02.stdout:(132/150): python3-repoze-lru-0.7-16.el9.noarch 3.9 MB/s | 31 kB 00:00 2026-03-06T13:43:39.149 INFO:teuthology.orchestra.run.vm06.stdout:(111/150): python3-cheroot-10.0.1-4.el9.noarch. 33 MB/s | 173 kB 00:00 2026-03-06T13:43:39.152 INFO:teuthology.orchestra.run.vm02.stdout:(133/150): python3-routes-2.5.1-5.el9.noarch.rp 28 MB/s | 188 kB 00:00 2026-03-06T13:43:39.155 INFO:teuthology.orchestra.run.vm06.stdout:(112/150): python3-cherrypy-18.6.1-2.el9.noarch 61 MB/s | 358 kB 00:00 2026-03-06T13:43:39.155 INFO:teuthology.orchestra.run.vm02.stdout:(134/150): python3-rsa-4.9-2.el9.noarch.rpm 18 MB/s | 59 kB 00:00 2026-03-06T13:43:39.160 INFO:teuthology.orchestra.run.vm06.stdout:(113/150): python3-google-auth-2.45.0-1.el9.noa 48 MB/s | 254 kB 00:00 2026-03-06T13:43:39.161 INFO:teuthology.orchestra.run.vm02.stdout:(135/150): python3-saml-1.16.0-1.el9.noarch.rpm 21 MB/s | 125 kB 00:00 2026-03-06T13:43:39.167 INFO:teuthology.orchestra.run.vm02.stdout:(136/150): python3-tempora-5.0.0-2.el9.noarch.r 7.1 MB/s | 36 kB 00:00 2026-03-06T13:43:39.167 INFO:teuthology.orchestra.run.vm04.stdout:(76/150): python3-mako-1.1.4-6.el9.noarch.rpm 1.7 MB/s | 172 kB 00:00 2026-03-06T13:43:39.174 INFO:teuthology.orchestra.run.vm02.stdout:(137/150): python3-typing-extensions-4.15.0-1.e 11 MB/s | 86 kB 00:00 2026-03-06T13:43:39.190 INFO:teuthology.orchestra.run.vm02.stdout:(138/150): python3-webob-1.8.8-2.el9.noarch.rpm 15 MB/s | 230 kB 00:00 2026-03-06T13:43:39.194 INFO:teuthology.orchestra.run.vm06.stdout:(114/150): python3-grpcio-1.46.7-10.el9.x86_64. 62 MB/s | 2.0 MB 00:00 2026-03-06T13:43:39.194 INFO:teuthology.orchestra.run.vm02.stdout:(139/150): python3-websocket-client-1.2.3-2.el9 19 MB/s | 90 kB 00:00 2026-03-06T13:43:39.205 INFO:teuthology.orchestra.run.vm06.stdout:(115/150): python3-grpcio-tools-1.46.7-10.el9.x 12 MB/s | 144 kB 00:00 2026-03-06T13:43:39.206 INFO:teuthology.orchestra.run.vm02.stdout:(140/150): protobuf-compiler-3.14.0-17.el9.x86_ 1.8 MB/s | 862 kB 00:00 2026-03-06T13:43:39.206 INFO:teuthology.orchestra.run.vm04.stdout:(77/150): python3-babel-2.9.1-2.el9.noarch.rpm 9.9 MB/s | 6.0 MB 00:00 2026-03-06T13:43:39.209 INFO:teuthology.orchestra.run.vm06.stdout:(116/150): python3-influxdb-5.3.1-1.el9.noarch. 34 MB/s | 139 kB 00:00 2026-03-06T13:43:39.214 INFO:teuthology.orchestra.run.vm06.stdout:(117/150): python3-isodate-0.6.1-3.el9.noarch.r 13 MB/s | 56 kB 00:00 2026-03-06T13:43:39.215 INFO:teuthology.orchestra.run.vm02.stdout:(141/150): python3-werkzeug-2.0.3-3.el9.1.noarc 21 MB/s | 427 kB 00:00 2026-03-06T13:43:39.215 INFO:teuthology.orchestra.run.vm02.stdout:(142/150): python3-xmlsec-1.3.13-1.el9.x86_64.r 4.9 MB/s | 48 kB 00:00 2026-03-06T13:43:39.216 INFO:teuthology.orchestra.run.vm04.stdout:(78/150): python3-markupsafe-1.1.1-12.el9.x86_6 707 kB/s | 35 kB 00:00 2026-03-06T13:43:39.216 INFO:teuthology.orchestra.run.vm06.stdout:(118/150): python3-jaraco-8.2.1-3.el9.noarch.rp 4.3 MB/s | 11 kB 00:00 2026-03-06T13:43:39.217 INFO:teuthology.orchestra.run.vm02.stdout:(143/150): python3-zc-lockfile-2.0-10.el9.noarc 9.4 MB/s | 20 kB 00:00 2026-03-06T13:43:39.218 INFO:teuthology.orchestra.run.vm02.stdout:(144/150): python3-xmltodict-0.12.0-15.el9.noar 8.0 MB/s | 22 kB 00:00 2026-03-06T13:43:39.219 INFO:teuthology.orchestra.run.vm06.stdout:(119/150): python3-jaraco-classes-3.2.1-5.el9.n 7.9 MB/s | 18 kB 00:00 2026-03-06T13:43:39.222 INFO:teuthology.orchestra.run.vm06.stdout:(120/150): python3-jaraco-collections-3.0.0-8.e 7.2 MB/s | 23 kB 00:00 2026-03-06T13:43:39.224 INFO:teuthology.orchestra.run.vm02.stdout:(145/150): s3cmd-2.4.0-1.el9.noarch.rpm 36 MB/s | 206 kB 00:00 2026-03-06T13:43:39.225 INFO:teuthology.orchestra.run.vm06.stdout:(121/150): python3-jaraco-context-6.0.1-3.el9.n 8.7 MB/s | 20 kB 00:00 2026-03-06T13:43:39.233 INFO:teuthology.orchestra.run.vm04.stdout:(79/150): python3-lxml-4.6.5-3.el9.x86_64.rpm 4.8 MB/s | 1.2 MB 00:00 2026-03-06T13:43:39.235 INFO:teuthology.orchestra.run.vm06.stdout:(122/150): python3-jaraco-functools-3.5.0-2.el9 2.0 MB/s | 19 kB 00:00 2026-03-06T13:43:39.235 INFO:teuthology.orchestra.run.vm02.stdout:(146/150): re2-20211101-20.el9.x86_64.rpm 11 MB/s | 191 kB 00:00 2026-03-06T13:43:39.238 INFO:teuthology.orchestra.run.vm06.stdout:(123/150): python3-jaraco-text-4.0.0-2.el9.noar 7.8 MB/s | 26 kB 00:00 2026-03-06T13:43:39.257 INFO:teuthology.orchestra.run.vm06.stdout:(124/150): python3-kubernetes-26.1.0-3.el9.noar 56 MB/s | 1.0 MB 00:00 2026-03-06T13:43:39.264 INFO:teuthology.orchestra.run.vm06.stdout:(125/150): python3-logutils-0.3.5-21.el9.noarch 7.8 MB/s | 46 kB 00:00 2026-03-06T13:43:39.265 INFO:teuthology.orchestra.run.vm02.stdout:(147/150): thrift-0.15.0-4.el9.x86_64.rpm 38 MB/s | 1.6 MB 00:00 2026-03-06T13:43:39.267 INFO:teuthology.orchestra.run.vm06.stdout:(126/150): python3-more-itertools-8.12.0-2.el9. 25 MB/s | 79 kB 00:00 2026-03-06T13:43:39.283 INFO:teuthology.orchestra.run.vm04.stdout:(80/150): python3-packaging-20.9-5.el9.noarch.r 1.5 MB/s | 77 kB 00:00 2026-03-06T13:43:39.288 INFO:teuthology.orchestra.run.vm06.stdout:(127/150): python3-msgpack-1.0.3-2.el9.x86_64.r 4.0 MB/s | 86 kB 00:00 2026-03-06T13:43:39.292 INFO:teuthology.orchestra.run.vm06.stdout:(128/150): python3-natsort-7.1.1-5.el9.noarch.r 17 MB/s | 58 kB 00:00 2026-03-06T13:43:39.300 INFO:teuthology.orchestra.run.vm06.stdout:(129/150): python3-pecan-1.4.2-3.el9.noarch.rpm 34 MB/s | 272 kB 00:00 2026-03-06T13:43:39.303 INFO:teuthology.orchestra.run.vm06.stdout:(130/150): python3-portend-3.1.0-2.el9.noarch.r 7.0 MB/s | 16 kB 00:00 2026-03-06T13:43:39.309 INFO:teuthology.orchestra.run.vm06.stdout:(131/150): python3-pyOpenSSL-21.0.0-1.el9.noarc 15 MB/s | 90 kB 00:00 2026-03-06T13:43:39.312 INFO:teuthology.orchestra.run.vm06.stdout:(132/150): python3-repoze-lru-0.7-16.el9.noarch 10 MB/s | 31 kB 00:00 2026-03-06T13:43:39.317 INFO:teuthology.orchestra.run.vm06.stdout:(133/150): python3-routes-2.5.1-5.el9.noarch.rp 36 MB/s | 188 kB 00:00 2026-03-06T13:43:39.320 INFO:teuthology.orchestra.run.vm06.stdout:(134/150): python3-rsa-4.9-2.el9.noarch.rpm 23 MB/s | 59 kB 00:00 2026-03-06T13:43:39.324 INFO:teuthology.orchestra.run.vm06.stdout:(135/150): python3-saml-1.16.0-1.el9.noarch.rpm 33 MB/s | 125 kB 00:00 2026-03-06T13:43:39.326 INFO:teuthology.orchestra.run.vm06.stdout:(136/150): python3-tempora-5.0.0-2.el9.noarch.r 15 MB/s | 36 kB 00:00 2026-03-06T13:43:39.329 INFO:teuthology.orchestra.run.vm06.stdout:(137/150): python3-typing-extensions-4.15.0-1.e 28 MB/s | 86 kB 00:00 2026-03-06T13:43:39.337 INFO:teuthology.orchestra.run.vm06.stdout:(138/150): python3-webob-1.8.8-2.el9.noarch.rpm 29 MB/s | 230 kB 00:00 2026-03-06T13:43:39.341 INFO:teuthology.orchestra.run.vm06.stdout:(139/150): python3-websocket-client-1.2.3-2.el9 25 MB/s | 90 kB 00:00 2026-03-06T13:43:39.349 INFO:teuthology.orchestra.run.vm06.stdout:(140/150): python3-werkzeug-2.0.3-3.el9.1.noarc 53 MB/s | 427 kB 00:00 2026-03-06T13:43:39.352 INFO:teuthology.orchestra.run.vm06.stdout:(141/150): python3-xmlsec-1.3.13-1.el9.x86_64.r 16 MB/s | 48 kB 00:00 2026-03-06T13:43:39.354 INFO:teuthology.orchestra.run.vm06.stdout:(142/150): python3-xmltodict-0.12.0-15.el9.noar 11 MB/s | 22 kB 00:00 2026-03-06T13:43:39.356 INFO:teuthology.orchestra.run.vm06.stdout:(143/150): python3-zc-lockfile-2.0-10.el9.noarc 10 MB/s | 20 kB 00:00 2026-03-06T13:43:39.361 INFO:teuthology.orchestra.run.vm06.stdout:(144/150): re2-20211101-20.el9.x86_64.rpm 41 MB/s | 191 kB 00:00 2026-03-06T13:43:39.365 INFO:teuthology.orchestra.run.vm06.stdout:(145/150): s3cmd-2.4.0-1.el9.noarch.rpm 47 MB/s | 206 kB 00:00 2026-03-06T13:43:39.381 INFO:teuthology.orchestra.run.vm04.stdout:(81/150): python3-protobuf-3.14.0-17.el9.noarch 2.7 MB/s | 267 kB 00:00 2026-03-06T13:43:39.388 INFO:teuthology.orchestra.run.vm06.stdout:(146/150): thrift-0.15.0-4.el9.x86_64.rpm 71 MB/s | 1.6 MB 00:00 2026-03-06T13:43:39.419 INFO:teuthology.orchestra.run.vm04.stdout:(82/150): python3-numpy-f2py-1.23.5-2.el9.x86_6 2.1 MB/s | 442 kB 00:00 2026-03-06T13:43:39.424 INFO:teuthology.orchestra.run.vm06.stdout:(147/150): lua-devel-5.4.4-4.el9.x86_64.rpm 41 kB/s | 22 kB 00:00 2026-03-06T13:43:39.432 INFO:teuthology.orchestra.run.vm04.stdout:(83/150): python3-pyasn1-0.4.8-7.el9.noarch.rpm 3.0 MB/s | 157 kB 00:00 2026-03-06T13:43:39.482 INFO:teuthology.orchestra.run.vm04.stdout:(84/150): python3-requests-oauthlib-1.3.0-12.el 1.1 MB/s | 54 kB 00:00 2026-03-06T13:43:39.518 INFO:teuthology.orchestra.run.vm04.stdout:(85/150): python3-pyasn1-modules-0.4.8-7.el9.no 2.7 MB/s | 277 kB 00:00 2026-03-06T13:43:39.568 INFO:teuthology.orchestra.run.vm04.stdout:(86/150): python3-toml-0.10.2-6.el9.noarch.rpm 839 kB/s | 42 kB 00:00 2026-03-06T13:43:39.621 INFO:teuthology.orchestra.run.vm02.stdout:(148/150): librados2-19.2.3-47.gc24117fd552.el9 8.9 MB/s | 3.4 MB 00:00 2026-03-06T13:43:39.631 INFO:teuthology.orchestra.run.vm02.stdout:(149/150): librbd1-19.2.3-47.gc24117fd552.el9.c 8.7 MB/s | 3.2 MB 00:00 2026-03-06T13:43:39.666 INFO:teuthology.orchestra.run.vm04.stdout:(87/150): qatlib-25.08.0-2.el9.x86_64.rpm 2.4 MB/s | 240 kB 00:00 2026-03-06T13:43:39.716 INFO:teuthology.orchestra.run.vm04.stdout:(88/150): qatlib-service-25.08.0-2.el9.x86_64.r 746 kB/s | 37 kB 00:00 2026-03-06T13:43:39.725 INFO:teuthology.orchestra.run.vm06.stdout:(148/150): librados2-19.2.3-47.gc24117fd552.el9 10 MB/s | 3.4 MB 00:00 2026-03-06T13:43:39.766 INFO:teuthology.orchestra.run.vm04.stdout:(89/150): qatzip-libs-1.3.1-1.el9.x86_64.rpm 1.3 MB/s | 66 kB 00:00 2026-03-06T13:43:39.909 INFO:teuthology.orchestra.run.vm04.stdout:(90/150): socat-1.7.4.1-8.el9.x86_64.rpm 2.1 MB/s | 303 kB 00:00 2026-03-06T13:43:39.929 INFO:teuthology.orchestra.run.vm04.stdout:(91/150): python3-numpy-1.23.5-2.el9.x86_64.rpm 8.5 MB/s | 6.1 MB 00:00 2026-03-06T13:43:39.979 INFO:teuthology.orchestra.run.vm04.stdout:(92/150): xmlsec1-openssl-1.2.29-13.el9.x86_64. 1.8 MB/s | 90 kB 00:00 2026-03-06T13:43:40.005 INFO:teuthology.orchestra.run.vm04.stdout:(93/150): xmlsec1-1.2.29-13.el9.x86_64.rpm 1.9 MB/s | 189 kB 00:00 2026-03-06T13:43:40.028 INFO:teuthology.orchestra.run.vm04.stdout:(94/150): xmlstarlet-1.6.1-20.el9.x86_64.rpm 1.3 MB/s | 64 kB 00:00 2026-03-06T13:43:40.227 INFO:teuthology.orchestra.run.vm06.stdout:(149/150): librbd1-19.2.3-47.gc24117fd552.el9.c 4.0 MB/s | 3.2 MB 00:00 2026-03-06T13:43:40.247 INFO:teuthology.orchestra.run.vm04.stdout:(95/150): lua-devel-5.4.4-4.el9.x86_64.rpm 92 kB/s | 22 kB 00:00 2026-03-06T13:43:40.261 INFO:teuthology.orchestra.run.vm04.stdout:(96/150): abseil-cpp-20211102.0-4.el9.x86_64.rp 39 MB/s | 551 kB 00:00 2026-03-06T13:43:40.267 INFO:teuthology.orchestra.run.vm04.stdout:(97/150): gperftools-libs-2.9.1-3.el9.x86_64.rp 49 MB/s | 308 kB 00:00 2026-03-06T13:43:40.269 INFO:teuthology.orchestra.run.vm04.stdout:(98/150): grpc-data-1.46.7-10.el9.noarch.rpm 9.1 MB/s | 19 kB 00:00 2026-03-06T13:43:40.329 INFO:teuthology.orchestra.run.vm04.stdout:(99/150): libarrow-9.0.0-15.el9.x86_64.rpm 74 MB/s | 4.4 MB 00:00 2026-03-06T13:43:40.332 INFO:teuthology.orchestra.run.vm04.stdout:(100/150): libarrow-doc-9.0.0-15.el9.noarch.rpm 9.3 MB/s | 25 kB 00:00 2026-03-06T13:43:40.335 INFO:teuthology.orchestra.run.vm04.stdout:(101/150): liboath-2.6.12-1.el9.x86_64.rpm 18 MB/s | 49 kB 00:00 2026-03-06T13:43:40.338 INFO:teuthology.orchestra.run.vm04.stdout:(102/150): libunwind-1.6.2-1.el9.x86_64.rpm 24 MB/s | 67 kB 00:00 2026-03-06T13:43:40.342 INFO:teuthology.orchestra.run.vm04.stdout:(103/150): luarocks-3.9.2-5.el9.noarch.rpm 42 MB/s | 151 kB 00:00 2026-03-06T13:43:40.354 INFO:teuthology.orchestra.run.vm04.stdout:(104/150): parquet-libs-9.0.0-15.el9.x86_64.rpm 67 MB/s | 838 kB 00:00 2026-03-06T13:43:40.363 INFO:teuthology.orchestra.run.vm04.stdout:(105/150): python3-asyncssh-2.13.2-5.el9.noarch 65 MB/s | 548 kB 00:00 2026-03-06T13:43:40.365 INFO:teuthology.orchestra.run.vm04.stdout:(106/150): python3-autocommand-2.2.2-8.el9.noar 13 MB/s | 29 kB 00:00 2026-03-06T13:43:40.368 INFO:teuthology.orchestra.run.vm04.stdout:(107/150): python3-backports-tarfile-1.2.0-1.el 19 MB/s | 60 kB 00:00 2026-03-06T13:43:40.370 INFO:teuthology.orchestra.run.vm04.stdout:(108/150): python3-bcrypt-3.2.2-1.el9.x86_64.rp 19 MB/s | 43 kB 00:00 2026-03-06T13:43:40.373 INFO:teuthology.orchestra.run.vm04.stdout:(109/150): python3-cachetools-4.2.4-1.el9.noarc 15 MB/s | 32 kB 00:00 2026-03-06T13:43:40.376 INFO:teuthology.orchestra.run.vm04.stdout:(110/150): python3-certifi-2023.05.07-4.el9.noa 5.3 MB/s | 14 kB 00:00 2026-03-06T13:43:40.380 INFO:teuthology.orchestra.run.vm04.stdout:(111/150): python3-cheroot-10.0.1-4.el9.noarch. 44 MB/s | 173 kB 00:00 2026-03-06T13:43:40.386 INFO:teuthology.orchestra.run.vm04.stdout:(112/150): python3-cherrypy-18.6.1-2.el9.noarch 60 MB/s | 358 kB 00:00 2026-03-06T13:43:40.391 INFO:teuthology.orchestra.run.vm04.stdout:(113/150): python3-google-auth-2.45.0-1.el9.noa 53 MB/s | 254 kB 00:00 2026-03-06T13:43:40.418 INFO:teuthology.orchestra.run.vm04.stdout:(114/150): python3-grpcio-1.46.7-10.el9.x86_64. 77 MB/s | 2.0 MB 00:00 2026-03-06T13:43:40.422 INFO:teuthology.orchestra.run.vm04.stdout:(115/150): python3-grpcio-tools-1.46.7-10.el9.x 36 MB/s | 144 kB 00:00 2026-03-06T13:43:40.426 INFO:teuthology.orchestra.run.vm04.stdout:(116/150): python3-influxdb-5.3.1-1.el9.noarch. 39 MB/s | 139 kB 00:00 2026-03-06T13:43:40.428 INFO:teuthology.orchestra.run.vm04.stdout:(117/150): python3-isodate-0.6.1-3.el9.noarch.r 23 MB/s | 56 kB 00:00 2026-03-06T13:43:40.430 INFO:teuthology.orchestra.run.vm04.stdout:(118/150): python3-jaraco-8.2.1-3.el9.noarch.rp 5.2 MB/s | 11 kB 00:00 2026-03-06T13:43:40.434 INFO:teuthology.orchestra.run.vm04.stdout:(119/150): python3-jaraco-classes-3.2.1-5.el9.n 5.6 MB/s | 18 kB 00:00 2026-03-06T13:43:40.436 INFO:teuthology.orchestra.run.vm04.stdout:(120/150): python3-jaraco-collections-3.0.0-8.e 9.5 MB/s | 23 kB 00:00 2026-03-06T13:43:40.439 INFO:teuthology.orchestra.run.vm04.stdout:(121/150): python3-jaraco-context-6.0.1-3.el9.n 8.2 MB/s | 20 kB 00:00 2026-03-06T13:43:40.441 INFO:teuthology.orchestra.run.vm04.stdout:(122/150): python3-jaraco-functools-3.5.0-2.el9 9.7 MB/s | 19 kB 00:00 2026-03-06T13:43:40.443 INFO:teuthology.orchestra.run.vm04.stdout:(123/150): python3-jaraco-text-4.0.0-2.el9.noar 12 MB/s | 26 kB 00:00 2026-03-06T13:43:40.458 INFO:teuthology.orchestra.run.vm04.stdout:(124/150): python3-kubernetes-26.1.0-3.el9.noar 73 MB/s | 1.0 MB 00:00 2026-03-06T13:43:40.460 INFO:teuthology.orchestra.run.vm04.stdout:(125/150): python3-logutils-0.3.5-21.el9.noarch 19 MB/s | 46 kB 00:00 2026-03-06T13:43:40.463 INFO:teuthology.orchestra.run.vm04.stdout:(126/150): python3-more-itertools-8.12.0-2.el9. 27 MB/s | 79 kB 00:00 2026-03-06T13:43:40.471 INFO:teuthology.orchestra.run.vm04.stdout:(127/150): python3-msgpack-1.0.3-2.el9.x86_64.r 12 MB/s | 86 kB 00:00 2026-03-06T13:43:40.475 INFO:teuthology.orchestra.run.vm04.stdout:(128/150): python3-natsort-7.1.1-5.el9.noarch.r 15 MB/s | 58 kB 00:00 2026-03-06T13:43:40.484 INFO:teuthology.orchestra.run.vm04.stdout:(129/150): python3-pecan-1.4.2-3.el9.noarch.rpm 30 MB/s | 272 kB 00:00 2026-03-06T13:43:40.486 INFO:teuthology.orchestra.run.vm04.stdout:(130/150): python3-portend-3.1.0-2.el9.noarch.r 7.1 MB/s | 16 kB 00:00 2026-03-06T13:43:40.490 INFO:teuthology.orchestra.run.vm04.stdout:(131/150): python3-pyOpenSSL-21.0.0-1.el9.noarc 25 MB/s | 90 kB 00:00 2026-03-06T13:43:40.492 INFO:teuthology.orchestra.run.vm04.stdout:(132/150): python3-repoze-lru-0.7-16.el9.noarch 14 MB/s | 31 kB 00:00 2026-03-06T13:43:40.498 INFO:teuthology.orchestra.run.vm04.stdout:(133/150): python3-routes-2.5.1-5.el9.noarch.rp 35 MB/s | 188 kB 00:00 2026-03-06T13:43:40.502 INFO:teuthology.orchestra.run.vm04.stdout:(134/150): python3-rsa-4.9-2.el9.noarch.rpm 17 MB/s | 59 kB 00:00 2026-03-06T13:43:40.506 INFO:teuthology.orchestra.run.vm04.stdout:(135/150): python3-saml-1.16.0-1.el9.noarch.rpm 31 MB/s | 125 kB 00:00 2026-03-06T13:43:40.508 INFO:teuthology.orchestra.run.vm04.stdout:(136/150): python3-tempora-5.0.0-2.el9.noarch.r 15 MB/s | 36 kB 00:00 2026-03-06T13:43:40.512 INFO:teuthology.orchestra.run.vm04.stdout:(137/150): python3-typing-extensions-4.15.0-1.e 26 MB/s | 86 kB 00:00 2026-03-06T13:43:40.519 INFO:teuthology.orchestra.run.vm04.stdout:(138/150): python3-webob-1.8.8-2.el9.noarch.rpm 31 MB/s | 230 kB 00:00 2026-03-06T13:43:40.522 INFO:teuthology.orchestra.run.vm04.stdout:(139/150): protobuf-compiler-3.14.0-17.el9.x86_ 1.7 MB/s | 862 kB 00:00 2026-03-06T13:43:40.524 INFO:teuthology.orchestra.run.vm04.stdout:(140/150): python3-websocket-client-1.2.3-2.el9 21 MB/s | 90 kB 00:00 2026-03-06T13:43:40.527 INFO:teuthology.orchestra.run.vm04.stdout:(141/150): python3-xmlsec-1.3.13-1.el9.x86_64.r 16 MB/s | 48 kB 00:00 2026-03-06T13:43:40.530 INFO:teuthology.orchestra.run.vm04.stdout:(142/150): python3-xmltodict-0.12.0-15.el9.noar 9.6 MB/s | 22 kB 00:00 2026-03-06T13:43:40.532 INFO:teuthology.orchestra.run.vm04.stdout:(143/150): python3-zc-lockfile-2.0-10.el9.noarc 9.4 MB/s | 20 kB 00:00 2026-03-06T13:43:40.536 INFO:teuthology.orchestra.run.vm04.stdout:(144/150): python3-werkzeug-2.0.3-3.el9.1.noarc 30 MB/s | 427 kB 00:00 2026-03-06T13:43:40.537 INFO:teuthology.orchestra.run.vm04.stdout:(145/150): re2-20211101-20.el9.x86_64.rpm 35 MB/s | 191 kB 00:00 2026-03-06T13:43:40.543 INFO:teuthology.orchestra.run.vm04.stdout:(146/150): s3cmd-2.4.0-1.el9.noarch.rpm 32 MB/s | 206 kB 00:00 2026-03-06T13:43:40.563 INFO:teuthology.orchestra.run.vm04.stdout:(147/150): thrift-0.15.0-4.el9.x86_64.rpm 64 MB/s | 1.6 MB 00:00 2026-03-06T13:43:40.765 INFO:teuthology.orchestra.run.vm02.stdout:(150/150): python3-scipy-1.9.3-2.el9.x86_64.rpm 7.9 MB/s | 19 MB 00:02 2026-03-06T13:43:40.768 INFO:teuthology.orchestra.run.vm02.stdout:-------------------------------------------------------------------------------- 2026-03-06T13:43:40.769 INFO:teuthology.orchestra.run.vm02.stdout:Total 19 MB/s | 214 MB 00:10 2026-03-06T13:43:40.946 INFO:teuthology.orchestra.run.vm06.stdout:(150/150): protobuf-compiler-3.14.0-17.el9.x86_ 426 kB/s | 862 kB 00:02 2026-03-06T13:43:40.948 INFO:teuthology.orchestra.run.vm06.stdout:-------------------------------------------------------------------------------- 2026-03-06T13:43:40.948 INFO:teuthology.orchestra.run.vm06.stdout:Total 19 MB/s | 214 MB 00:11 2026-03-06T13:43:41.409 INFO:teuthology.orchestra.run.vm04.stdout:(148/150): librbd1-19.2.3-47.gc24117fd552.el9.c 3.7 MB/s | 3.2 MB 00:00 2026-03-06T13:43:41.457 INFO:teuthology.orchestra.run.vm02.stdout:Running transaction check 2026-03-06T13:43:41.518 INFO:teuthology.orchestra.run.vm02.stdout:Transaction check succeeded. 2026-03-06T13:43:41.518 INFO:teuthology.orchestra.run.vm02.stdout:Running transaction test 2026-03-06T13:43:41.639 INFO:teuthology.orchestra.run.vm06.stdout:Running transaction check 2026-03-06T13:43:41.650 INFO:teuthology.orchestra.run.vm04.stdout:(149/150): librados2-19.2.3-47.gc24117fd552.el9 3.1 MB/s | 3.4 MB 00:01 2026-03-06T13:43:41.702 INFO:teuthology.orchestra.run.vm06.stdout:Transaction check succeeded. 2026-03-06T13:43:41.702 INFO:teuthology.orchestra.run.vm06.stdout:Running transaction test 2026-03-06T13:43:42.409 INFO:teuthology.orchestra.run.vm02.stdout:Transaction test succeeded. 2026-03-06T13:43:42.410 INFO:teuthology.orchestra.run.vm02.stdout:Running transaction 2026-03-06T13:43:42.467 INFO:teuthology.orchestra.run.vm04.stdout:(150/150): python3-scipy-1.9.3-2.el9.x86_64.rpm 6.5 MB/s | 19 MB 00:02 2026-03-06T13:43:42.471 INFO:teuthology.orchestra.run.vm04.stdout:-------------------------------------------------------------------------------- 2026-03-06T13:43:42.472 INFO:teuthology.orchestra.run.vm04.stdout:Total 17 MB/s | 214 MB 00:12 2026-03-06T13:43:42.635 INFO:teuthology.orchestra.run.vm06.stdout:Transaction test succeeded. 2026-03-06T13:43:42.635 INFO:teuthology.orchestra.run.vm06.stdout:Running transaction 2026-03-06T13:43:43.101 INFO:teuthology.orchestra.run.vm04.stdout:Running transaction check 2026-03-06T13:43:43.163 INFO:teuthology.orchestra.run.vm04.stdout:Transaction check succeeded. 2026-03-06T13:43:43.163 INFO:teuthology.orchestra.run.vm04.stdout:Running transaction test 2026-03-06T13:43:43.417 INFO:teuthology.orchestra.run.vm02.stdout: Preparing : 1/1 2026-03-06T13:43:43.431 INFO:teuthology.orchestra.run.vm02.stdout: Installing : python3-more-itertools-8.12.0-2.el9.noarch 1/152 2026-03-06T13:43:43.444 INFO:teuthology.orchestra.run.vm02.stdout: Installing : thrift-0.15.0-4.el9.x86_64 2/152 2026-03-06T13:43:43.627 INFO:teuthology.orchestra.run.vm02.stdout: Installing : lttng-ust-2.12.0-6.el9.x86_64 3/152 2026-03-06T13:43:43.630 INFO:teuthology.orchestra.run.vm02.stdout: Upgrading : librados2-2:19.2.3-47.gc24117fd552.el9.clyso.x86 4/152 2026-03-06T13:43:43.633 INFO:teuthology.orchestra.run.vm06.stdout: Preparing : 1/1 2026-03-06T13:43:43.647 INFO:teuthology.orchestra.run.vm06.stdout: Installing : python3-more-itertools-8.12.0-2.el9.noarch 1/152 2026-03-06T13:43:43.660 INFO:teuthology.orchestra.run.vm06.stdout: Installing : thrift-0.15.0-4.el9.x86_64 2/152 2026-03-06T13:43:43.699 INFO:teuthology.orchestra.run.vm02.stdout: Running scriptlet: librados2-2:19.2.3-47.gc24117fd552.el9.clyso.x86 4/152 2026-03-06T13:43:43.701 INFO:teuthology.orchestra.run.vm02.stdout: Installing : libcephfs2-2:19.2.3-47.gc24117fd552.el9.clyso.x8 5/152 2026-03-06T13:43:43.732 INFO:teuthology.orchestra.run.vm02.stdout: Running scriptlet: libcephfs2-2:19.2.3-47.gc24117fd552.el9.clyso.x8 5/152 2026-03-06T13:43:43.739 INFO:teuthology.orchestra.run.vm02.stdout: Installing : python3-rados-2:19.2.3-47.gc24117fd552.el9.clyso 6/152 2026-03-06T13:43:43.749 INFO:teuthology.orchestra.run.vm02.stdout: Installing : libxslt-1.1.34-12.el9.x86_64 7/152 2026-03-06T13:43:43.753 INFO:teuthology.orchestra.run.vm02.stdout: Installing : librdkafka-1.6.1-102.el9.x86_64 8/152 2026-03-06T13:43:43.756 INFO:teuthology.orchestra.run.vm02.stdout: Installing : librabbitmq-0.11.0-7.el9.x86_64 9/152 2026-03-06T13:43:43.761 INFO:teuthology.orchestra.run.vm02.stdout: Installing : python3-jaraco-8.2.1-3.el9.noarch 10/152 2026-03-06T13:43:43.803 INFO:teuthology.orchestra.run.vm02.stdout: Installing : libnbd-1.20.3-4.el9.x86_64 11/152 2026-03-06T13:43:43.812 INFO:teuthology.orchestra.run.vm02.stdout: Installing : python3-lxml-4.6.5-3.el9.x86_64 12/152 2026-03-06T13:43:43.822 INFO:teuthology.orchestra.run.vm02.stdout: Installing : xmlsec1-1.2.29-13.el9.x86_64 13/152 2026-03-06T13:43:43.824 INFO:teuthology.orchestra.run.vm02.stdout: Installing : libcephsqlite-2:19.2.3-47.gc24117fd552.el9.clyso 14/152 2026-03-06T13:43:43.843 INFO:teuthology.orchestra.run.vm06.stdout: Installing : lttng-ust-2.12.0-6.el9.x86_64 3/152 2026-03-06T13:43:43.845 INFO:teuthology.orchestra.run.vm06.stdout: Upgrading : librados2-2:19.2.3-47.gc24117fd552.el9.clyso.x86 4/152 2026-03-06T13:43:43.861 INFO:teuthology.orchestra.run.vm02.stdout: Running scriptlet: libcephsqlite-2:19.2.3-47.gc24117fd552.el9.clyso 14/152 2026-03-06T13:43:43.862 INFO:teuthology.orchestra.run.vm02.stdout: Installing : libradosstriper1-2:19.2.3-47.gc24117fd552.el9.cl 15/152 2026-03-06T13:43:43.876 INFO:teuthology.orchestra.run.vm02.stdout: Running scriptlet: libradosstriper1-2:19.2.3-47.gc24117fd552.el9.cl 15/152 2026-03-06T13:43:43.908 INFO:teuthology.orchestra.run.vm06.stdout: Running scriptlet: librados2-2:19.2.3-47.gc24117fd552.el9.clyso.x86 4/152 2026-03-06T13:43:43.910 INFO:teuthology.orchestra.run.vm06.stdout: Installing : libcephfs2-2:19.2.3-47.gc24117fd552.el9.clyso.x8 5/152 2026-03-06T13:43:43.911 INFO:teuthology.orchestra.run.vm02.stdout: Installing : re2-1:20211101-20.el9.x86_64 16/152 2026-03-06T13:43:43.941 INFO:teuthology.orchestra.run.vm06.stdout: Running scriptlet: libcephfs2-2:19.2.3-47.gc24117fd552.el9.clyso.x8 5/152 2026-03-06T13:43:43.948 INFO:teuthology.orchestra.run.vm06.stdout: Installing : python3-rados-2:19.2.3-47.gc24117fd552.el9.clyso 6/152 2026-03-06T13:43:43.950 INFO:teuthology.orchestra.run.vm02.stdout: Installing : libarrow-9.0.0-15.el9.x86_64 17/152 2026-03-06T13:43:43.956 INFO:teuthology.orchestra.run.vm02.stdout: Installing : python3-werkzeug-2.0.3-3.el9.1.noarch 18/152 2026-03-06T13:43:43.962 INFO:teuthology.orchestra.run.vm06.stdout: Installing : libxslt-1.1.34-12.el9.x86_64 7/152 2026-03-06T13:43:43.966 INFO:teuthology.orchestra.run.vm06.stdout: Installing : librdkafka-1.6.1-102.el9.x86_64 8/152 2026-03-06T13:43:43.969 INFO:teuthology.orchestra.run.vm06.stdout: Installing : librabbitmq-0.11.0-7.el9.x86_64 9/152 2026-03-06T13:43:43.975 INFO:teuthology.orchestra.run.vm06.stdout: Installing : python3-jaraco-8.2.1-3.el9.noarch 10/152 2026-03-06T13:43:43.994 INFO:teuthology.orchestra.run.vm02.stdout: Installing : liboath-2.6.12-1.el9.x86_64 19/152 2026-03-06T13:43:44.009 INFO:teuthology.orchestra.run.vm02.stdout: Installing : python3-pyasn1-0.4.8-7.el9.noarch 20/152 2026-03-06T13:43:44.018 INFO:teuthology.orchestra.run.vm02.stdout: Installing : python3-packaging-20.9-5.el9.noarch 21/152 2026-03-06T13:43:44.023 INFO:teuthology.orchestra.run.vm06.stdout: Installing : libnbd-1.20.3-4.el9.x86_64 11/152 2026-03-06T13:43:44.028 INFO:teuthology.orchestra.run.vm02.stdout: Installing : python3-markupsafe-1.1.1-12.el9.x86_64 22/152 2026-03-06T13:43:44.035 INFO:teuthology.orchestra.run.vm06.stdout: Installing : python3-lxml-4.6.5-3.el9.x86_64 12/152 2026-03-06T13:43:44.035 INFO:teuthology.orchestra.run.vm02.stdout: Installing : protobuf-3.14.0-17.el9.x86_64 23/152 2026-03-06T13:43:44.040 INFO:teuthology.orchestra.run.vm02.stdout: Installing : lua-5.4.4-4.el9.x86_64 24/152 2026-03-06T13:43:44.045 INFO:teuthology.orchestra.run.vm02.stdout: Installing : flexiblas-3.0.4-9.el9.x86_64 25/152 2026-03-06T13:43:44.047 INFO:teuthology.orchestra.run.vm06.stdout: Installing : xmlsec1-1.2.29-13.el9.x86_64 13/152 2026-03-06T13:43:44.047 INFO:teuthology.orchestra.run.vm04.stdout:Transaction test succeeded. 2026-03-06T13:43:44.048 INFO:teuthology.orchestra.run.vm04.stdout:Running transaction 2026-03-06T13:43:44.048 INFO:teuthology.orchestra.run.vm06.stdout: Installing : libcephsqlite-2:19.2.3-47.gc24117fd552.el9.clyso 14/152 2026-03-06T13:43:44.075 INFO:teuthology.orchestra.run.vm02.stdout: Installing : unzip-6.0-59.el9.x86_64 26/152 2026-03-06T13:43:44.092 INFO:teuthology.orchestra.run.vm02.stdout: Installing : python3-urllib3-1.26.5-7.el9.noarch 27/152 2026-03-06T13:43:44.095 INFO:teuthology.orchestra.run.vm06.stdout: Running scriptlet: libcephsqlite-2:19.2.3-47.gc24117fd552.el9.clyso 14/152 2026-03-06T13:43:44.096 INFO:teuthology.orchestra.run.vm02.stdout: Installing : python3-requests-2.25.1-10.el9.noarch 28/152 2026-03-06T13:43:44.099 INFO:teuthology.orchestra.run.vm06.stdout: Installing : libradosstriper1-2:19.2.3-47.gc24117fd552.el9.cl 15/152 2026-03-06T13:43:44.106 INFO:teuthology.orchestra.run.vm02.stdout: Installing : libquadmath-11.5.0-14.el9.x86_64 29/152 2026-03-06T13:43:44.108 INFO:teuthology.orchestra.run.vm02.stdout: Installing : libgfortran-11.5.0-14.el9.x86_64 30/152 2026-03-06T13:43:44.117 INFO:teuthology.orchestra.run.vm06.stdout: Running scriptlet: libradosstriper1-2:19.2.3-47.gc24117fd552.el9.cl 15/152 2026-03-06T13:43:44.141 INFO:teuthology.orchestra.run.vm02.stdout: Installing : ledmon-libs-1.1.0-3.el9.x86_64 31/152 2026-03-06T13:43:44.148 INFO:teuthology.orchestra.run.vm02.stdout: Installing : python3-ceph-common-2:19.2.3-47.gc24117fd552.el9 32/152 2026-03-06T13:43:44.152 INFO:teuthology.orchestra.run.vm06.stdout: Installing : re2-1:20211101-20.el9.x86_64 16/152 2026-03-06T13:43:44.160 INFO:teuthology.orchestra.run.vm02.stdout: Installing : python3-ceph-argparse-2:19.2.3-47.gc24117fd552.e 33/152 2026-03-06T13:43:44.175 INFO:teuthology.orchestra.run.vm02.stdout: Installing : python3-cephfs-2:19.2.3-47.gc24117fd552.el9.clys 34/152 2026-03-06T13:43:44.207 INFO:teuthology.orchestra.run.vm06.stdout: Installing : libarrow-9.0.0-15.el9.x86_64 17/152 2026-03-06T13:43:44.213 INFO:teuthology.orchestra.run.vm02.stdout: Installing : python3-requests-oauthlib-1.3.0-12.el9.noarch 35/152 2026-03-06T13:43:44.216 INFO:teuthology.orchestra.run.vm06.stdout: Installing : python3-werkzeug-2.0.3-3.el9.1.noarch 18/152 2026-03-06T13:43:44.245 INFO:teuthology.orchestra.run.vm06.stdout: Installing : liboath-2.6.12-1.el9.x86_64 19/152 2026-03-06T13:43:44.247 INFO:teuthology.orchestra.run.vm02.stdout: Installing : zip-3.0-35.el9.x86_64 36/152 2026-03-06T13:43:44.255 INFO:teuthology.orchestra.run.vm02.stdout: Installing : luarocks-3.9.2-5.el9.noarch 37/152 2026-03-06T13:43:44.263 INFO:teuthology.orchestra.run.vm06.stdout: Installing : python3-pyasn1-0.4.8-7.el9.noarch 20/152 2026-03-06T13:43:44.264 INFO:teuthology.orchestra.run.vm02.stdout: Installing : lua-devel-5.4.4-4.el9.x86_64 38/152 2026-03-06T13:43:44.272 INFO:teuthology.orchestra.run.vm06.stdout: Installing : python3-packaging-20.9-5.el9.noarch 21/152 2026-03-06T13:43:44.285 INFO:teuthology.orchestra.run.vm06.stdout: Installing : python3-markupsafe-1.1.1-12.el9.x86_64 22/152 2026-03-06T13:43:44.298 INFO:teuthology.orchestra.run.vm06.stdout: Installing : protobuf-3.14.0-17.el9.x86_64 23/152 2026-03-06T13:43:44.300 INFO:teuthology.orchestra.run.vm02.stdout: Installing : protobuf-compiler-3.14.0-17.el9.x86_64 39/152 2026-03-06T13:43:44.304 INFO:teuthology.orchestra.run.vm06.stdout: Installing : lua-5.4.4-4.el9.x86_64 24/152 2026-03-06T13:43:44.311 INFO:teuthology.orchestra.run.vm06.stdout: Installing : flexiblas-3.0.4-9.el9.x86_64 25/152 2026-03-06T13:43:44.345 INFO:teuthology.orchestra.run.vm06.stdout: Installing : unzip-6.0-59.el9.x86_64 26/152 2026-03-06T13:43:44.364 INFO:teuthology.orchestra.run.vm06.stdout: Installing : python3-urllib3-1.26.5-7.el9.noarch 27/152 2026-03-06T13:43:44.365 INFO:teuthology.orchestra.run.vm02.stdout: Installing : python3-mako-1.1.4-6.el9.noarch 40/152 2026-03-06T13:43:44.371 INFO:teuthology.orchestra.run.vm06.stdout: Installing : python3-requests-2.25.1-10.el9.noarch 28/152 2026-03-06T13:43:44.380 INFO:teuthology.orchestra.run.vm06.stdout: Installing : libquadmath-11.5.0-14.el9.x86_64 29/152 2026-03-06T13:43:44.383 INFO:teuthology.orchestra.run.vm06.stdout: Installing : libgfortran-11.5.0-14.el9.x86_64 30/152 2026-03-06T13:43:44.386 INFO:teuthology.orchestra.run.vm02.stdout: Installing : python3-pyasn1-modules-0.4.8-7.el9.noarch 41/152 2026-03-06T13:43:44.391 INFO:teuthology.orchestra.run.vm02.stdout: Installing : python3-rsa-4.9-2.el9.noarch 42/152 2026-03-06T13:43:44.397 INFO:teuthology.orchestra.run.vm02.stdout: Installing : xmlsec1-openssl-1.2.29-13.el9.x86_64 43/152 2026-03-06T13:43:44.420 INFO:teuthology.orchestra.run.vm06.stdout: Installing : ledmon-libs-1.1.0-3.el9.x86_64 31/152 2026-03-06T13:43:44.438 INFO:teuthology.orchestra.run.vm06.stdout: Installing : python3-ceph-common-2:19.2.3-47.gc24117fd552.el9 32/152 2026-03-06T13:43:44.441 INFO:teuthology.orchestra.run.vm02.stdout: Installing : python3-xmlsec-1.3.13-1.el9.x86_64 44/152 2026-03-06T13:43:44.447 INFO:teuthology.orchestra.run.vm02.stdout: Installing : python3-jaraco-classes-3.2.1-5.el9.noarch 45/152 2026-03-06T13:43:44.450 INFO:teuthology.orchestra.run.vm06.stdout: Installing : python3-ceph-argparse-2:19.2.3-47.gc24117fd552.e 33/152 2026-03-06T13:43:44.457 INFO:teuthology.orchestra.run.vm02.stdout: Installing : xmlstarlet-1.6.1-20.el9.x86_64 46/152 2026-03-06T13:43:44.463 INFO:teuthology.orchestra.run.vm02.stdout: Installing : librados-devel-2:19.2.3-47.gc24117fd552.el9.clys 47/152 2026-03-06T13:43:44.465 INFO:teuthology.orchestra.run.vm06.stdout: Installing : python3-cephfs-2:19.2.3-47.gc24117fd552.el9.clys 34/152 2026-03-06T13:43:44.468 INFO:teuthology.orchestra.run.vm02.stdout: Installing : python3-zc-lockfile-2.0-10.el9.noarch 48/152 2026-03-06T13:43:44.474 INFO:teuthology.orchestra.run.vm06.stdout: Installing : python3-requests-oauthlib-1.3.0-12.el9.noarch 35/152 2026-03-06T13:43:44.486 INFO:teuthology.orchestra.run.vm02.stdout: Installing : python3-xmltodict-0.12.0-15.el9.noarch 49/152 2026-03-06T13:43:44.508 INFO:teuthology.orchestra.run.vm06.stdout: Installing : zip-3.0-35.el9.x86_64 36/152 2026-03-06T13:43:44.512 INFO:teuthology.orchestra.run.vm02.stdout: Installing : python3-websocket-client-1.2.3-2.el9.noarch 50/152 2026-03-06T13:43:44.514 INFO:teuthology.orchestra.run.vm06.stdout: Installing : luarocks-3.9.2-5.el9.noarch 37/152 2026-03-06T13:43:44.520 INFO:teuthology.orchestra.run.vm02.stdout: Installing : python3-webob-1.8.8-2.el9.noarch 51/152 2026-03-06T13:43:44.523 INFO:teuthology.orchestra.run.vm06.stdout: Installing : lua-devel-5.4.4-4.el9.x86_64 38/152 2026-03-06T13:43:44.526 INFO:teuthology.orchestra.run.vm02.stdout: Installing : python3-typing-extensions-4.15.0-1.el9.noarch 52/152 2026-03-06T13:43:44.540 INFO:teuthology.orchestra.run.vm02.stdout: Installing : python3-repoze-lru-0.7-16.el9.noarch 53/152 2026-03-06T13:43:44.553 INFO:teuthology.orchestra.run.vm02.stdout: Installing : python3-routes-2.5.1-5.el9.noarch 54/152 2026-03-06T13:43:44.556 INFO:teuthology.orchestra.run.vm06.stdout: Installing : protobuf-compiler-3.14.0-17.el9.x86_64 39/152 2026-03-06T13:43:44.562 INFO:teuthology.orchestra.run.vm02.stdout: Installing : python3-natsort-7.1.1-5.el9.noarch 55/152 2026-03-06T13:43:44.586 INFO:teuthology.orchestra.run.vm02.stdout: Installing : python3-msgpack-1.0.3-2.el9.x86_64 56/152 2026-03-06T13:43:44.598 INFO:teuthology.orchestra.run.vm02.stdout: Installing : python3-influxdb-5.3.1-1.el9.noarch 57/152 2026-03-06T13:43:44.628 INFO:teuthology.orchestra.run.vm06.stdout: Installing : python3-mako-1.1.4-6.el9.noarch 40/152 2026-03-06T13:43:44.646 INFO:teuthology.orchestra.run.vm06.stdout: Installing : python3-pyasn1-modules-0.4.8-7.el9.noarch 41/152 2026-03-06T13:43:44.651 INFO:teuthology.orchestra.run.vm06.stdout: Installing : python3-rsa-4.9-2.el9.noarch 42/152 2026-03-06T13:43:44.658 INFO:teuthology.orchestra.run.vm06.stdout: Installing : xmlsec1-openssl-1.2.29-13.el9.x86_64 43/152 2026-03-06T13:43:44.665 INFO:teuthology.orchestra.run.vm06.stdout: Installing : python3-xmlsec-1.3.13-1.el9.x86_64 44/152 2026-03-06T13:43:44.669 INFO:teuthology.orchestra.run.vm02.stdout: Installing : python3-logutils-0.3.5-21.el9.noarch 58/152 2026-03-06T13:43:44.672 INFO:teuthology.orchestra.run.vm06.stdout: Installing : python3-jaraco-classes-3.2.1-5.el9.noarch 45/152 2026-03-06T13:43:44.684 INFO:teuthology.orchestra.run.vm06.stdout: Installing : xmlstarlet-1.6.1-20.el9.x86_64 46/152 2026-03-06T13:43:44.687 INFO:teuthology.orchestra.run.vm02.stdout: Installing : python3-pecan-1.4.2-3.el9.noarch 59/152 2026-03-06T13:43:44.692 INFO:teuthology.orchestra.run.vm06.stdout: Installing : librados-devel-2:19.2.3-47.gc24117fd552.el9.clys 47/152 2026-03-06T13:43:44.696 INFO:teuthology.orchestra.run.vm06.stdout: Installing : python3-zc-lockfile-2.0-10.el9.noarch 48/152 2026-03-06T13:43:44.709 INFO:teuthology.orchestra.run.vm02.stdout: Installing : python3-isodate-0.6.1-3.el9.noarch 60/152 2026-03-06T13:43:44.716 INFO:teuthology.orchestra.run.vm02.stdout: Installing : python3-saml-1.16.0-1.el9.noarch 61/152 2026-03-06T13:43:44.718 INFO:teuthology.orchestra.run.vm06.stdout: Installing : python3-xmltodict-0.12.0-15.el9.noarch 49/152 2026-03-06T13:43:44.727 INFO:teuthology.orchestra.run.vm02.stdout: Installing : python3-certifi-2023.05.07-4.el9.noarch 62/152 2026-03-06T13:43:44.749 INFO:teuthology.orchestra.run.vm06.stdout: Installing : python3-websocket-client-1.2.3-2.el9.noarch 50/152 2026-03-06T13:43:44.757 INFO:teuthology.orchestra.run.vm06.stdout: Installing : python3-webob-1.8.8-2.el9.noarch 51/152 2026-03-06T13:43:44.764 INFO:teuthology.orchestra.run.vm06.stdout: Installing : python3-typing-extensions-4.15.0-1.el9.noarch 52/152 2026-03-06T13:43:44.776 INFO:teuthology.orchestra.run.vm02.stdout: Installing : python3-cachetools-4.2.4-1.el9.noarch 63/152 2026-03-06T13:43:44.780 INFO:teuthology.orchestra.run.vm06.stdout: Installing : python3-repoze-lru-0.7-16.el9.noarch 53/152 2026-03-06T13:43:44.795 INFO:teuthology.orchestra.run.vm06.stdout: Installing : python3-routes-2.5.1-5.el9.noarch 54/152 2026-03-06T13:43:44.805 INFO:teuthology.orchestra.run.vm06.stdout: Installing : python3-natsort-7.1.1-5.el9.noarch 55/152 2026-03-06T13:43:44.833 INFO:teuthology.orchestra.run.vm06.stdout: Installing : python3-msgpack-1.0.3-2.el9.x86_64 56/152 2026-03-06T13:43:44.849 INFO:teuthology.orchestra.run.vm06.stdout: Installing : python3-influxdb-5.3.1-1.el9.noarch 57/152 2026-03-06T13:43:44.919 INFO:teuthology.orchestra.run.vm06.stdout: Installing : python3-logutils-0.3.5-21.el9.noarch 58/152 2026-03-06T13:43:44.937 INFO:teuthology.orchestra.run.vm06.stdout: Installing : python3-pecan-1.4.2-3.el9.noarch 59/152 2026-03-06T13:43:44.958 INFO:teuthology.orchestra.run.vm06.stdout: Installing : python3-isodate-0.6.1-3.el9.noarch 60/152 2026-03-06T13:43:44.967 INFO:teuthology.orchestra.run.vm06.stdout: Installing : python3-saml-1.16.0-1.el9.noarch 61/152 2026-03-06T13:43:44.980 INFO:teuthology.orchestra.run.vm06.stdout: Installing : python3-certifi-2023.05.07-4.el9.noarch 62/152 2026-03-06T13:43:45.024 INFO:teuthology.orchestra.run.vm04.stdout: Preparing : 1/1 2026-03-06T13:43:45.037 INFO:teuthology.orchestra.run.vm06.stdout: Installing : python3-cachetools-4.2.4-1.el9.noarch 63/152 2026-03-06T13:43:45.041 INFO:teuthology.orchestra.run.vm04.stdout: Installing : python3-more-itertools-8.12.0-2.el9.noarch 1/152 2026-03-06T13:43:45.065 INFO:teuthology.orchestra.run.vm04.stdout: Installing : thrift-0.15.0-4.el9.x86_64 2/152 2026-03-06T13:43:45.165 INFO:teuthology.orchestra.run.vm02.stdout: Installing : python3-google-auth-1:2.45.0-1.el9.noarch 64/152 2026-03-06T13:43:45.184 INFO:teuthology.orchestra.run.vm02.stdout: Installing : python3-kubernetes-1:26.1.0-3.el9.noarch 65/152 2026-03-06T13:43:45.191 INFO:teuthology.orchestra.run.vm02.stdout: Installing : python3-backports-tarfile-1.2.0-1.el9.noarch 66/152 2026-03-06T13:43:45.198 INFO:teuthology.orchestra.run.vm02.stdout: Installing : python3-jaraco-context-6.0.1-3.el9.noarch 67/152 2026-03-06T13:43:45.205 INFO:teuthology.orchestra.run.vm02.stdout: Installing : python3-autocommand-2.2.2-8.el9.noarch 68/152 2026-03-06T13:43:45.213 INFO:teuthology.orchestra.run.vm02.stdout: Installing : libunwind-1.6.2-1.el9.x86_64 69/152 2026-03-06T13:43:45.218 INFO:teuthology.orchestra.run.vm02.stdout: Installing : gperftools-libs-2.9.1-3.el9.x86_64 70/152 2026-03-06T13:43:45.221 INFO:teuthology.orchestra.run.vm02.stdout: Installing : libarrow-doc-9.0.0-15.el9.noarch 71/152 2026-03-06T13:43:45.245 INFO:teuthology.orchestra.run.vm04.stdout: Installing : lttng-ust-2.12.0-6.el9.x86_64 3/152 2026-03-06T13:43:45.247 INFO:teuthology.orchestra.run.vm04.stdout: Upgrading : librados2-2:19.2.3-47.gc24117fd552.el9.clyso.x86 4/152 2026-03-06T13:43:45.253 INFO:teuthology.orchestra.run.vm02.stdout: Installing : grpc-data-1.46.7-10.el9.noarch 72/152 2026-03-06T13:43:45.310 INFO:teuthology.orchestra.run.vm04.stdout: Running scriptlet: librados2-2:19.2.3-47.gc24117fd552.el9.clyso.x86 4/152 2026-03-06T13:43:45.311 INFO:teuthology.orchestra.run.vm04.stdout: Installing : libcephfs2-2:19.2.3-47.gc24117fd552.el9.clyso.x8 5/152 2026-03-06T13:43:45.312 INFO:teuthology.orchestra.run.vm02.stdout: Installing : abseil-cpp-20211102.0-4.el9.x86_64 73/152 2026-03-06T13:43:45.331 INFO:teuthology.orchestra.run.vm02.stdout: Installing : python3-grpcio-1.46.7-10.el9.x86_64 74/152 2026-03-06T13:43:45.341 INFO:teuthology.orchestra.run.vm04.stdout: Running scriptlet: libcephfs2-2:19.2.3-47.gc24117fd552.el9.clyso.x8 5/152 2026-03-06T13:43:45.343 INFO:teuthology.orchestra.run.vm02.stdout: Installing : socat-1.7.4.1-8.el9.x86_64 75/152 2026-03-06T13:43:45.347 INFO:teuthology.orchestra.run.vm04.stdout: Installing : python3-rados-2:19.2.3-47.gc24117fd552.el9.clyso 6/152 2026-03-06T13:43:45.351 INFO:teuthology.orchestra.run.vm02.stdout: Installing : python3-toml-0.10.2-6.el9.noarch 76/152 2026-03-06T13:43:45.358 INFO:teuthology.orchestra.run.vm04.stdout: Installing : libxslt-1.1.34-12.el9.x86_64 7/152 2026-03-06T13:43:45.362 INFO:teuthology.orchestra.run.vm04.stdout: Installing : librdkafka-1.6.1-102.el9.x86_64 8/152 2026-03-06T13:43:45.362 INFO:teuthology.orchestra.run.vm02.stdout: Installing : python3-jaraco-functools-3.5.0-2.el9.noarch 77/152 2026-03-06T13:43:45.364 INFO:teuthology.orchestra.run.vm04.stdout: Installing : librabbitmq-0.11.0-7.el9.x86_64 9/152 2026-03-06T13:43:45.368 INFO:teuthology.orchestra.run.vm02.stdout: Installing : python3-jaraco-text-4.0.0-2.el9.noarch 78/152 2026-03-06T13:43:45.369 INFO:teuthology.orchestra.run.vm04.stdout: Installing : python3-jaraco-8.2.1-3.el9.noarch 10/152 2026-03-06T13:43:45.378 INFO:teuthology.orchestra.run.vm02.stdout: Installing : python3-jaraco-collections-3.0.0-8.el9.noarch 79/152 2026-03-06T13:43:45.384 INFO:teuthology.orchestra.run.vm02.stdout: Installing : python3-tempora-5.0.0-2.el9.noarch 80/152 2026-03-06T13:43:45.410 INFO:teuthology.orchestra.run.vm04.stdout: Installing : libnbd-1.20.3-4.el9.x86_64 11/152 2026-03-06T13:43:45.419 INFO:teuthology.orchestra.run.vm04.stdout: Installing : python3-lxml-4.6.5-3.el9.x86_64 12/152 2026-03-06T13:43:45.419 INFO:teuthology.orchestra.run.vm02.stdout: Installing : python3-portend-3.1.0-2.el9.noarch 81/152 2026-03-06T13:43:45.429 INFO:teuthology.orchestra.run.vm04.stdout: Installing : xmlsec1-1.2.29-13.el9.x86_64 13/152 2026-03-06T13:43:45.430 INFO:teuthology.orchestra.run.vm04.stdout: Installing : libcephsqlite-2:19.2.3-47.gc24117fd552.el9.clyso 14/152 2026-03-06T13:43:45.432 INFO:teuthology.orchestra.run.vm02.stdout: Installing : python3-protobuf-3.14.0-17.el9.noarch 82/152 2026-03-06T13:43:45.468 INFO:teuthology.orchestra.run.vm04.stdout: Running scriptlet: libcephsqlite-2:19.2.3-47.gc24117fd552.el9.clyso 14/152 2026-03-06T13:43:45.470 INFO:teuthology.orchestra.run.vm04.stdout: Installing : libradosstriper1-2:19.2.3-47.gc24117fd552.el9.cl 15/152 2026-03-06T13:43:45.475 INFO:teuthology.orchestra.run.vm06.stdout: Installing : python3-google-auth-1:2.45.0-1.el9.noarch 64/152 2026-03-06T13:43:45.476 INFO:teuthology.orchestra.run.vm02.stdout: Installing : python3-grpcio-tools-1.46.7-10.el9.x86_64 83/152 2026-03-06T13:43:45.484 INFO:teuthology.orchestra.run.vm04.stdout: Running scriptlet: libradosstriper1-2:19.2.3-47.gc24117fd552.el9.cl 15/152 2026-03-06T13:43:45.492 INFO:teuthology.orchestra.run.vm06.stdout: Installing : python3-kubernetes-1:26.1.0-3.el9.noarch 65/152 2026-03-06T13:43:45.497 INFO:teuthology.orchestra.run.vm06.stdout: Installing : python3-backports-tarfile-1.2.0-1.el9.noarch 66/152 2026-03-06T13:43:45.506 INFO:teuthology.orchestra.run.vm06.stdout: Installing : python3-jaraco-context-6.0.1-3.el9.noarch 67/152 2026-03-06T13:43:45.510 INFO:teuthology.orchestra.run.vm06.stdout: Installing : python3-autocommand-2.2.2-8.el9.noarch 68/152 2026-03-06T13:43:45.519 INFO:teuthology.orchestra.run.vm04.stdout: Installing : re2-1:20211101-20.el9.x86_64 16/152 2026-03-06T13:43:45.519 INFO:teuthology.orchestra.run.vm06.stdout: Installing : libunwind-1.6.2-1.el9.x86_64 69/152 2026-03-06T13:43:45.523 INFO:teuthology.orchestra.run.vm06.stdout: Installing : gperftools-libs-2.9.1-3.el9.x86_64 70/152 2026-03-06T13:43:45.526 INFO:teuthology.orchestra.run.vm06.stdout: Installing : libarrow-doc-9.0.0-15.el9.noarch 71/152 2026-03-06T13:43:45.560 INFO:teuthology.orchestra.run.vm04.stdout: Installing : libarrow-9.0.0-15.el9.x86_64 17/152 2026-03-06T13:43:45.564 INFO:teuthology.orchestra.run.vm06.stdout: Installing : grpc-data-1.46.7-10.el9.noarch 72/152 2026-03-06T13:43:45.565 INFO:teuthology.orchestra.run.vm04.stdout: Installing : python3-werkzeug-2.0.3-3.el9.1.noarch 18/152 2026-03-06T13:43:45.593 INFO:teuthology.orchestra.run.vm04.stdout: Installing : liboath-2.6.12-1.el9.x86_64 19/152 2026-03-06T13:43:45.609 INFO:teuthology.orchestra.run.vm04.stdout: Installing : python3-pyasn1-0.4.8-7.el9.noarch 20/152 2026-03-06T13:43:45.618 INFO:teuthology.orchestra.run.vm04.stdout: Installing : python3-packaging-20.9-5.el9.noarch 21/152 2026-03-06T13:43:45.626 INFO:teuthology.orchestra.run.vm06.stdout: Installing : abseil-cpp-20211102.0-4.el9.x86_64 73/152 2026-03-06T13:43:45.629 INFO:teuthology.orchestra.run.vm04.stdout: Installing : python3-markupsafe-1.1.1-12.el9.x86_64 22/152 2026-03-06T13:43:45.637 INFO:teuthology.orchestra.run.vm04.stdout: Installing : protobuf-3.14.0-17.el9.x86_64 23/152 2026-03-06T13:43:45.640 INFO:teuthology.orchestra.run.vm04.stdout: Installing : lua-5.4.4-4.el9.x86_64 24/152 2026-03-06T13:43:45.641 INFO:teuthology.orchestra.run.vm06.stdout: Installing : python3-grpcio-1.46.7-10.el9.x86_64 74/152 2026-03-06T13:43:45.647 INFO:teuthology.orchestra.run.vm04.stdout: Installing : flexiblas-3.0.4-9.el9.x86_64 25/152 2026-03-06T13:43:45.650 INFO:teuthology.orchestra.run.vm06.stdout: Installing : socat-1.7.4.1-8.el9.x86_64 75/152 2026-03-06T13:43:45.656 INFO:teuthology.orchestra.run.vm06.stdout: Installing : python3-toml-0.10.2-6.el9.noarch 76/152 2026-03-06T13:43:45.668 INFO:teuthology.orchestra.run.vm06.stdout: Installing : python3-jaraco-functools-3.5.0-2.el9.noarch 77/152 2026-03-06T13:43:45.674 INFO:teuthology.orchestra.run.vm06.stdout: Installing : python3-jaraco-text-4.0.0-2.el9.noarch 78/152 2026-03-06T13:43:45.675 INFO:teuthology.orchestra.run.vm04.stdout: Installing : unzip-6.0-59.el9.x86_64 26/152 2026-03-06T13:43:45.685 INFO:teuthology.orchestra.run.vm06.stdout: Installing : python3-jaraco-collections-3.0.0-8.el9.noarch 79/152 2026-03-06T13:43:45.691 INFO:teuthology.orchestra.run.vm06.stdout: Installing : python3-tempora-5.0.0-2.el9.noarch 80/152 2026-03-06T13:43:45.692 INFO:teuthology.orchestra.run.vm04.stdout: Installing : python3-urllib3-1.26.5-7.el9.noarch 27/152 2026-03-06T13:43:45.697 INFO:teuthology.orchestra.run.vm04.stdout: Installing : python3-requests-2.25.1-10.el9.noarch 28/152 2026-03-06T13:43:45.706 INFO:teuthology.orchestra.run.vm04.stdout: Installing : libquadmath-11.5.0-14.el9.x86_64 29/152 2026-03-06T13:43:45.708 INFO:teuthology.orchestra.run.vm04.stdout: Installing : libgfortran-11.5.0-14.el9.x86_64 30/152 2026-03-06T13:43:45.732 INFO:teuthology.orchestra.run.vm06.stdout: Installing : python3-portend-3.1.0-2.el9.noarch 81/152 2026-03-06T13:43:45.740 INFO:teuthology.orchestra.run.vm04.stdout: Installing : ledmon-libs-1.1.0-3.el9.x86_64 31/152 2026-03-06T13:43:45.747 INFO:teuthology.orchestra.run.vm04.stdout: Installing : python3-ceph-common-2:19.2.3-47.gc24117fd552.el9 32/152 2026-03-06T13:43:45.747 INFO:teuthology.orchestra.run.vm06.stdout: Installing : python3-protobuf-3.14.0-17.el9.noarch 82/152 2026-03-06T13:43:45.755 INFO:teuthology.orchestra.run.vm02.stdout: Installing : python3-devel-3.9.25-3.el9.x86_64 84/152 2026-03-06T13:43:45.758 INFO:teuthology.orchestra.run.vm04.stdout: Installing : python3-ceph-argparse-2:19.2.3-47.gc24117fd552.e 33/152 2026-03-06T13:43:45.774 INFO:teuthology.orchestra.run.vm04.stdout: Installing : python3-cephfs-2:19.2.3-47.gc24117fd552.el9.clys 34/152 2026-03-06T13:43:45.784 INFO:teuthology.orchestra.run.vm04.stdout: Installing : python3-requests-oauthlib-1.3.0-12.el9.noarch 35/152 2026-03-06T13:43:45.787 INFO:teuthology.orchestra.run.vm02.stdout: Installing : python3-babel-2.9.1-2.el9.noarch 85/152 2026-03-06T13:43:45.790 INFO:teuthology.orchestra.run.vm02.stdout: Installing : python3-jinja2-2.11.3-8.el9.noarch 86/152 2026-03-06T13:43:45.794 INFO:teuthology.orchestra.run.vm02.stdout: Installing : perl-Benchmark-1.23-483.el9.noarch 87/152 2026-03-06T13:43:45.798 INFO:teuthology.orchestra.run.vm06.stdout: Installing : python3-grpcio-tools-1.46.7-10.el9.x86_64 83/152 2026-03-06T13:43:45.817 INFO:teuthology.orchestra.run.vm04.stdout: Installing : zip-3.0-35.el9.x86_64 36/152 2026-03-06T13:43:45.822 INFO:teuthology.orchestra.run.vm04.stdout: Installing : luarocks-3.9.2-5.el9.noarch 37/152 2026-03-06T13:43:45.832 INFO:teuthology.orchestra.run.vm04.stdout: Installing : lua-devel-5.4.4-4.el9.x86_64 38/152 2026-03-06T13:43:45.858 INFO:teuthology.orchestra.run.vm02.stdout: Installing : openblas-0.3.29-1.el9.x86_64 88/152 2026-03-06T13:43:45.861 INFO:teuthology.orchestra.run.vm02.stdout: Installing : openblas-openmp-0.3.29-1.el9.x86_64 89/152 2026-03-06T13:43:45.865 INFO:teuthology.orchestra.run.vm04.stdout: Installing : protobuf-compiler-3.14.0-17.el9.x86_64 39/152 2026-03-06T13:43:45.884 INFO:teuthology.orchestra.run.vm02.stdout: Installing : flexiblas-openblas-openmp-3.0.4-9.el9.x86_64 90/152 2026-03-06T13:43:45.931 INFO:teuthology.orchestra.run.vm04.stdout: Installing : python3-mako-1.1.4-6.el9.noarch 40/152 2026-03-06T13:43:45.953 INFO:teuthology.orchestra.run.vm04.stdout: Installing : python3-pyasn1-modules-0.4.8-7.el9.noarch 41/152 2026-03-06T13:43:45.957 INFO:teuthology.orchestra.run.vm04.stdout: Installing : python3-rsa-4.9-2.el9.noarch 42/152 2026-03-06T13:43:45.963 INFO:teuthology.orchestra.run.vm04.stdout: Installing : xmlsec1-openssl-1.2.29-13.el9.x86_64 43/152 2026-03-06T13:43:45.970 INFO:teuthology.orchestra.run.vm04.stdout: Installing : python3-xmlsec-1.3.13-1.el9.x86_64 44/152 2026-03-06T13:43:45.976 INFO:teuthology.orchestra.run.vm04.stdout: Installing : python3-jaraco-classes-3.2.1-5.el9.noarch 45/152 2026-03-06T13:43:45.986 INFO:teuthology.orchestra.run.vm04.stdout: Installing : xmlstarlet-1.6.1-20.el9.x86_64 46/152 2026-03-06T13:43:45.993 INFO:teuthology.orchestra.run.vm04.stdout: Installing : librados-devel-2:19.2.3-47.gc24117fd552.el9.clys 47/152 2026-03-06T13:43:45.997 INFO:teuthology.orchestra.run.vm04.stdout: Installing : python3-zc-lockfile-2.0-10.el9.noarch 48/152 2026-03-06T13:43:46.016 INFO:teuthology.orchestra.run.vm04.stdout: Installing : python3-xmltodict-0.12.0-15.el9.noarch 49/152 2026-03-06T13:43:46.045 INFO:teuthology.orchestra.run.vm04.stdout: Installing : python3-websocket-client-1.2.3-2.el9.noarch 50/152 2026-03-06T13:43:46.056 INFO:teuthology.orchestra.run.vm04.stdout: Installing : python3-webob-1.8.8-2.el9.noarch 51/152 2026-03-06T13:43:46.064 INFO:teuthology.orchestra.run.vm04.stdout: Installing : python3-typing-extensions-4.15.0-1.el9.noarch 52/152 2026-03-06T13:43:46.079 INFO:teuthology.orchestra.run.vm04.stdout: Installing : python3-repoze-lru-0.7-16.el9.noarch 53/152 2026-03-06T13:43:46.092 INFO:teuthology.orchestra.run.vm04.stdout: Installing : python3-routes-2.5.1-5.el9.noarch 54/152 2026-03-06T13:43:46.092 INFO:teuthology.orchestra.run.vm06.stdout: Installing : python3-devel-3.9.25-3.el9.x86_64 84/152 2026-03-06T13:43:46.101 INFO:teuthology.orchestra.run.vm04.stdout: Installing : python3-natsort-7.1.1-5.el9.noarch 55/152 2026-03-06T13:43:46.126 INFO:teuthology.orchestra.run.vm04.stdout: Installing : python3-msgpack-1.0.3-2.el9.x86_64 56/152 2026-03-06T13:43:46.130 INFO:teuthology.orchestra.run.vm06.stdout: Installing : python3-babel-2.9.1-2.el9.noarch 85/152 2026-03-06T13:43:46.135 INFO:teuthology.orchestra.run.vm06.stdout: Installing : python3-jinja2-2.11.3-8.el9.noarch 86/152 2026-03-06T13:43:46.138 INFO:teuthology.orchestra.run.vm04.stdout: Installing : python3-influxdb-5.3.1-1.el9.noarch 57/152 2026-03-06T13:43:46.139 INFO:teuthology.orchestra.run.vm06.stdout: Installing : perl-Benchmark-1.23-483.el9.noarch 87/152 2026-03-06T13:43:46.203 INFO:teuthology.orchestra.run.vm04.stdout: Installing : python3-logutils-0.3.5-21.el9.noarch 58/152 2026-03-06T13:43:46.208 INFO:teuthology.orchestra.run.vm06.stdout: Installing : openblas-0.3.29-1.el9.x86_64 88/152 2026-03-06T13:43:46.211 INFO:teuthology.orchestra.run.vm06.stdout: Installing : openblas-openmp-0.3.29-1.el9.x86_64 89/152 2026-03-06T13:43:46.220 INFO:teuthology.orchestra.run.vm04.stdout: Installing : python3-pecan-1.4.2-3.el9.noarch 59/152 2026-03-06T13:43:46.236 INFO:teuthology.orchestra.run.vm06.stdout: Installing : flexiblas-openblas-openmp-3.0.4-9.el9.x86_64 90/152 2026-03-06T13:43:46.239 INFO:teuthology.orchestra.run.vm04.stdout: Installing : python3-isodate-0.6.1-3.el9.noarch 60/152 2026-03-06T13:43:46.247 INFO:teuthology.orchestra.run.vm04.stdout: Installing : python3-saml-1.16.0-1.el9.noarch 61/152 2026-03-06T13:43:46.258 INFO:teuthology.orchestra.run.vm04.stdout: Installing : python3-certifi-2023.05.07-4.el9.noarch 62/152 2026-03-06T13:43:46.291 INFO:teuthology.orchestra.run.vm02.stdout: Installing : flexiblas-netlib-3.0.4-9.el9.x86_64 91/152 2026-03-06T13:43:46.306 INFO:teuthology.orchestra.run.vm04.stdout: Installing : python3-cachetools-4.2.4-1.el9.noarch 63/152 2026-03-06T13:43:46.401 INFO:teuthology.orchestra.run.vm02.stdout: Installing : python3-numpy-1:1.23.5-2.el9.x86_64 92/152 2026-03-06T13:43:46.692 INFO:teuthology.orchestra.run.vm06.stdout: Installing : flexiblas-netlib-3.0.4-9.el9.x86_64 91/152 2026-03-06T13:43:46.708 INFO:teuthology.orchestra.run.vm04.stdout: Installing : python3-google-auth-1:2.45.0-1.el9.noarch 64/152 2026-03-06T13:43:46.725 INFO:teuthology.orchestra.run.vm04.stdout: Installing : python3-kubernetes-1:26.1.0-3.el9.noarch 65/152 2026-03-06T13:43:46.731 INFO:teuthology.orchestra.run.vm04.stdout: Installing : python3-backports-tarfile-1.2.0-1.el9.noarch 66/152 2026-03-06T13:43:46.740 INFO:teuthology.orchestra.run.vm04.stdout: Installing : python3-jaraco-context-6.0.1-3.el9.noarch 67/152 2026-03-06T13:43:46.744 INFO:teuthology.orchestra.run.vm04.stdout: Installing : python3-autocommand-2.2.2-8.el9.noarch 68/152 2026-03-06T13:43:46.753 INFO:teuthology.orchestra.run.vm04.stdout: Installing : libunwind-1.6.2-1.el9.x86_64 69/152 2026-03-06T13:43:46.756 INFO:teuthology.orchestra.run.vm04.stdout: Installing : gperftools-libs-2.9.1-3.el9.x86_64 70/152 2026-03-06T13:43:46.758 INFO:teuthology.orchestra.run.vm04.stdout: Installing : libarrow-doc-9.0.0-15.el9.noarch 71/152 2026-03-06T13:43:46.789 INFO:teuthology.orchestra.run.vm04.stdout: Installing : grpc-data-1.46.7-10.el9.noarch 72/152 2026-03-06T13:43:46.800 INFO:teuthology.orchestra.run.vm06.stdout: Installing : python3-numpy-1:1.23.5-2.el9.x86_64 92/152 2026-03-06T13:43:46.843 INFO:teuthology.orchestra.run.vm04.stdout: Installing : abseil-cpp-20211102.0-4.el9.x86_64 73/152 2026-03-06T13:43:46.858 INFO:teuthology.orchestra.run.vm04.stdout: Installing : python3-grpcio-1.46.7-10.el9.x86_64 74/152 2026-03-06T13:43:46.867 INFO:teuthology.orchestra.run.vm04.stdout: Installing : socat-1.7.4.1-8.el9.x86_64 75/152 2026-03-06T13:43:46.876 INFO:teuthology.orchestra.run.vm04.stdout: Installing : python3-toml-0.10.2-6.el9.noarch 76/152 2026-03-06T13:43:46.888 INFO:teuthology.orchestra.run.vm04.stdout: Installing : python3-jaraco-functools-3.5.0-2.el9.noarch 77/152 2026-03-06T13:43:46.895 INFO:teuthology.orchestra.run.vm04.stdout: Installing : python3-jaraco-text-4.0.0-2.el9.noarch 78/152 2026-03-06T13:43:46.908 INFO:teuthology.orchestra.run.vm04.stdout: Installing : python3-jaraco-collections-3.0.0-8.el9.noarch 79/152 2026-03-06T13:43:46.914 INFO:teuthology.orchestra.run.vm04.stdout: Installing : python3-tempora-5.0.0-2.el9.noarch 80/152 2026-03-06T13:43:46.957 INFO:teuthology.orchestra.run.vm04.stdout: Installing : python3-portend-3.1.0-2.el9.noarch 81/152 2026-03-06T13:43:46.971 INFO:teuthology.orchestra.run.vm04.stdout: Installing : python3-protobuf-3.14.0-17.el9.noarch 82/152 2026-03-06T13:43:47.016 INFO:teuthology.orchestra.run.vm04.stdout: Installing : python3-grpcio-tools-1.46.7-10.el9.x86_64 83/152 2026-03-06T13:43:47.266 INFO:teuthology.orchestra.run.vm02.stdout: Installing : python3-numpy-f2py-1:1.23.5-2.el9.x86_64 93/152 2026-03-06T13:43:47.296 INFO:teuthology.orchestra.run.vm02.stdout: Installing : python3-scipy-1.9.3-2.el9.x86_64 94/152 2026-03-06T13:43:47.315 INFO:teuthology.orchestra.run.vm04.stdout: Installing : python3-devel-3.9.25-3.el9.x86_64 84/152 2026-03-06T13:43:47.351 INFO:teuthology.orchestra.run.vm04.stdout: Installing : python3-babel-2.9.1-2.el9.noarch 85/152 2026-03-06T13:43:47.355 INFO:teuthology.orchestra.run.vm04.stdout: Installing : python3-jinja2-2.11.3-8.el9.noarch 86/152 2026-03-06T13:43:47.360 INFO:teuthology.orchestra.run.vm04.stdout: Installing : perl-Benchmark-1.23-483.el9.noarch 87/152 2026-03-06T13:43:47.430 INFO:teuthology.orchestra.run.vm04.stdout: Installing : openblas-0.3.29-1.el9.x86_64 88/152 2026-03-06T13:43:47.433 INFO:teuthology.orchestra.run.vm04.stdout: Installing : openblas-openmp-0.3.29-1.el9.x86_64 89/152 2026-03-06T13:43:47.461 INFO:teuthology.orchestra.run.vm04.stdout: Installing : flexiblas-openblas-openmp-3.0.4-9.el9.x86_64 90/152 2026-03-06T13:43:47.472 INFO:teuthology.orchestra.run.vm02.stdout: Installing : libpmemobj-1.12.1-1.el9.x86_64 95/152 2026-03-06T13:43:47.477 INFO:teuthology.orchestra.run.vm02.stdout: Upgrading : librbd1-2:19.2.3-47.gc24117fd552.el9.clyso.x86_6 96/152 2026-03-06T13:43:47.512 INFO:teuthology.orchestra.run.vm02.stdout: Running scriptlet: librbd1-2:19.2.3-47.gc24117fd552.el9.clyso.x86_6 96/152 2026-03-06T13:43:47.517 INFO:teuthology.orchestra.run.vm02.stdout: Installing : python3-rbd-2:19.2.3-47.gc24117fd552.el9.clyso.x 97/152 2026-03-06T13:43:47.526 INFO:teuthology.orchestra.run.vm02.stdout: Installing : boost-program-options-1.75.0-13.el9.x86_64 98/152 2026-03-06T13:43:47.715 INFO:teuthology.orchestra.run.vm06.stdout: Installing : python3-numpy-f2py-1:1.23.5-2.el9.x86_64 93/152 2026-03-06T13:43:47.744 INFO:teuthology.orchestra.run.vm06.stdout: Installing : python3-scipy-1.9.3-2.el9.x86_64 94/152 2026-03-06T13:43:47.810 INFO:teuthology.orchestra.run.vm02.stdout: Installing : parquet-libs-9.0.0-15.el9.x86_64 99/152 2026-03-06T13:43:47.812 INFO:teuthology.orchestra.run.vm02.stdout: Installing : librgw2-2:19.2.3-47.gc24117fd552.el9.clyso.x86_6 100/152 2026-03-06T13:43:47.834 INFO:teuthology.orchestra.run.vm02.stdout: Running scriptlet: librgw2-2:19.2.3-47.gc24117fd552.el9.clyso.x86_6 100/152 2026-03-06T13:43:47.837 INFO:teuthology.orchestra.run.vm02.stdout: Installing : python3-rgw-2:19.2.3-47.gc24117fd552.el9.clyso.x 101/152 2026-03-06T13:43:47.878 INFO:teuthology.orchestra.run.vm04.stdout: Installing : flexiblas-netlib-3.0.4-9.el9.x86_64 91/152 2026-03-06T13:43:47.926 INFO:teuthology.orchestra.run.vm06.stdout: Installing : libpmemobj-1.12.1-1.el9.x86_64 95/152 2026-03-06T13:43:47.930 INFO:teuthology.orchestra.run.vm06.stdout: Upgrading : librbd1-2:19.2.3-47.gc24117fd552.el9.clyso.x86_6 96/152 2026-03-06T13:43:47.967 INFO:teuthology.orchestra.run.vm06.stdout: Running scriptlet: librbd1-2:19.2.3-47.gc24117fd552.el9.clyso.x86_6 96/152 2026-03-06T13:43:47.971 INFO:teuthology.orchestra.run.vm06.stdout: Installing : python3-rbd-2:19.2.3-47.gc24117fd552.el9.clyso.x 97/152 2026-03-06T13:43:47.974 INFO:teuthology.orchestra.run.vm04.stdout: Installing : python3-numpy-1:1.23.5-2.el9.x86_64 92/152 2026-03-06T13:43:47.980 INFO:teuthology.orchestra.run.vm06.stdout: Installing : boost-program-options-1.75.0-13.el9.x86_64 98/152 2026-03-06T13:43:48.284 INFO:teuthology.orchestra.run.vm06.stdout: Installing : parquet-libs-9.0.0-15.el9.x86_64 99/152 2026-03-06T13:43:48.287 INFO:teuthology.orchestra.run.vm06.stdout: Installing : librgw2-2:19.2.3-47.gc24117fd552.el9.clyso.x86_6 100/152 2026-03-06T13:43:48.310 INFO:teuthology.orchestra.run.vm06.stdout: Running scriptlet: librgw2-2:19.2.3-47.gc24117fd552.el9.clyso.x86_6 100/152 2026-03-06T13:43:48.313 INFO:teuthology.orchestra.run.vm06.stdout: Installing : python3-rgw-2:19.2.3-47.gc24117fd552.el9.clyso.x 101/152 2026-03-06T13:43:48.816 INFO:teuthology.orchestra.run.vm04.stdout: Installing : python3-numpy-f2py-1:1.23.5-2.el9.x86_64 93/152 2026-03-06T13:43:48.846 INFO:teuthology.orchestra.run.vm04.stdout: Installing : python3-scipy-1.9.3-2.el9.x86_64 94/152 2026-03-06T13:43:49.020 INFO:teuthology.orchestra.run.vm04.stdout: Installing : libpmemobj-1.12.1-1.el9.x86_64 95/152 2026-03-06T13:43:49.023 INFO:teuthology.orchestra.run.vm04.stdout: Upgrading : librbd1-2:19.2.3-47.gc24117fd552.el9.clyso.x86_6 96/152 2026-03-06T13:43:49.057 INFO:teuthology.orchestra.run.vm02.stdout: Running scriptlet: ceph-common-2:19.2.3-47.gc24117fd552.el9.clyso.x 102/152 2026-03-06T13:43:49.060 INFO:teuthology.orchestra.run.vm04.stdout: Running scriptlet: librbd1-2:19.2.3-47.gc24117fd552.el9.clyso.x86_6 96/152 2026-03-06T13:43:49.072 INFO:teuthology.orchestra.run.vm04.stdout: Installing : python3-rbd-2:19.2.3-47.gc24117fd552.el9.clyso.x 97/152 2026-03-06T13:43:49.080 INFO:teuthology.orchestra.run.vm02.stdout: Installing : ceph-common-2:19.2.3-47.gc24117fd552.el9.clyso.x 102/152 2026-03-06T13:43:49.084 INFO:teuthology.orchestra.run.vm04.stdout: Installing : boost-program-options-1.75.0-13.el9.x86_64 98/152 2026-03-06T13:43:49.115 INFO:teuthology.orchestra.run.vm02.stdout: Running scriptlet: ceph-common-2:19.2.3-47.gc24117fd552.el9.clyso.x 102/152 2026-03-06T13:43:49.119 INFO:teuthology.orchestra.run.vm02.stdout: Installing : smartmontools-1:7.2-10.el9.x86_64 103/152 2026-03-06T13:43:49.133 INFO:teuthology.orchestra.run.vm02.stdout: Running scriptlet: smartmontools-1:7.2-10.el9.x86_64 103/152 2026-03-06T13:43:49.133 INFO:teuthology.orchestra.run.vm02.stdout:Created symlink /etc/systemd/system/multi-user.target.wants/smartd.service → /usr/lib/systemd/system/smartd.service. 2026-03-06T13:43:49.133 INFO:teuthology.orchestra.run.vm02.stdout: 2026-03-06T13:43:49.162 INFO:teuthology.orchestra.run.vm02.stdout: Installing : python3-ply-3.11-14.el9.noarch 104/152 2026-03-06T13:43:49.184 INFO:teuthology.orchestra.run.vm02.stdout: Installing : python3-pycparser-2.20-6.el9.noarch 105/152 2026-03-06T13:43:49.279 INFO:teuthology.orchestra.run.vm02.stdout: Installing : python3-cffi-1.14.5-5.el9.x86_64 106/152 2026-03-06T13:43:49.295 INFO:teuthology.orchestra.run.vm02.stdout: Installing : python3-cryptography-36.0.1-5.el9.x86_64 107/152 2026-03-06T13:43:49.327 INFO:teuthology.orchestra.run.vm02.stdout: Installing : python3-pyOpenSSL-21.0.0-1.el9.noarch 108/152 2026-03-06T13:43:49.358 INFO:teuthology.orchestra.run.vm04.stdout: Installing : parquet-libs-9.0.0-15.el9.x86_64 99/152 2026-03-06T13:43:49.362 INFO:teuthology.orchestra.run.vm04.stdout: Installing : librgw2-2:19.2.3-47.gc24117fd552.el9.clyso.x86_6 100/152 2026-03-06T13:43:49.366 INFO:teuthology.orchestra.run.vm02.stdout: Installing : python3-cheroot-10.0.1-4.el9.noarch 109/152 2026-03-06T13:43:49.386 INFO:teuthology.orchestra.run.vm04.stdout: Running scriptlet: librgw2-2:19.2.3-47.gc24117fd552.el9.clyso.x86_6 100/152 2026-03-06T13:43:49.389 INFO:teuthology.orchestra.run.vm04.stdout: Installing : python3-rgw-2:19.2.3-47.gc24117fd552.el9.clyso.x 101/152 2026-03-06T13:43:49.430 INFO:teuthology.orchestra.run.vm02.stdout: Installing : python3-cherrypy-18.6.1-2.el9.noarch 110/152 2026-03-06T13:43:49.441 INFO:teuthology.orchestra.run.vm02.stdout: Installing : python3-asyncssh-2.13.2-5.el9.noarch 111/152 2026-03-06T13:43:49.447 INFO:teuthology.orchestra.run.vm02.stdout: Installing : python3-bcrypt-3.2.2-1.el9.x86_64 112/152 2026-03-06T13:43:49.454 INFO:teuthology.orchestra.run.vm02.stdout: Installing : pciutils-3.7.0-7.el9.x86_64 113/152 2026-03-06T13:43:49.459 INFO:teuthology.orchestra.run.vm02.stdout: Installing : qatlib-25.08.0-2.el9.x86_64 114/152 2026-03-06T13:43:49.462 INFO:teuthology.orchestra.run.vm02.stdout: Installing : qatlib-service-25.08.0-2.el9.x86_64 115/152 2026-03-06T13:43:49.482 INFO:teuthology.orchestra.run.vm02.stdout: Running scriptlet: qatlib-service-25.08.0-2.el9.x86_64 115/152 2026-03-06T13:43:49.590 INFO:teuthology.orchestra.run.vm06.stdout: Running scriptlet: ceph-common-2:19.2.3-47.gc24117fd552.el9.clyso.x 102/152 2026-03-06T13:43:49.612 INFO:teuthology.orchestra.run.vm06.stdout: Installing : ceph-common-2:19.2.3-47.gc24117fd552.el9.clyso.x 102/152 2026-03-06T13:43:49.640 INFO:teuthology.orchestra.run.vm06.stdout: Running scriptlet: ceph-common-2:19.2.3-47.gc24117fd552.el9.clyso.x 102/152 2026-03-06T13:43:49.644 INFO:teuthology.orchestra.run.vm06.stdout: Installing : smartmontools-1:7.2-10.el9.x86_64 103/152 2026-03-06T13:43:49.662 INFO:teuthology.orchestra.run.vm06.stdout: Running scriptlet: smartmontools-1:7.2-10.el9.x86_64 103/152 2026-03-06T13:43:49.662 INFO:teuthology.orchestra.run.vm06.stdout:Created symlink /etc/systemd/system/multi-user.target.wants/smartd.service → /usr/lib/systemd/system/smartd.service. 2026-03-06T13:43:49.662 INFO:teuthology.orchestra.run.vm06.stdout: 2026-03-06T13:43:49.691 INFO:teuthology.orchestra.run.vm06.stdout: Installing : python3-ply-3.11-14.el9.noarch 104/152 2026-03-06T13:43:49.713 INFO:teuthology.orchestra.run.vm06.stdout: Installing : python3-pycparser-2.20-6.el9.noarch 105/152 2026-03-06T13:43:49.813 INFO:teuthology.orchestra.run.vm06.stdout: Installing : python3-cffi-1.14.5-5.el9.x86_64 106/152 2026-03-06T13:43:49.820 INFO:teuthology.orchestra.run.vm02.stdout: Installing : qatzip-libs-1.3.1-1.el9.x86_64 116/152 2026-03-06T13:43:49.827 INFO:teuthology.orchestra.run.vm02.stdout: Installing : ceph-base-2:19.2.3-47.gc24117fd552.el9.clyso.x86 117/152 2026-03-06T13:43:49.829 INFO:teuthology.orchestra.run.vm06.stdout: Installing : python3-cryptography-36.0.1-5.el9.x86_64 107/152 2026-03-06T13:43:49.862 INFO:teuthology.orchestra.run.vm06.stdout: Installing : python3-pyOpenSSL-21.0.0-1.el9.noarch 108/152 2026-03-06T13:43:49.875 INFO:teuthology.orchestra.run.vm02.stdout: Running scriptlet: ceph-base-2:19.2.3-47.gc24117fd552.el9.clyso.x86 117/152 2026-03-06T13:43:49.875 INFO:teuthology.orchestra.run.vm02.stdout:Created symlink /etc/systemd/system/multi-user.target.wants/ceph.target → /usr/lib/systemd/system/ceph.target. 2026-03-06T13:43:49.875 INFO:teuthology.orchestra.run.vm02.stdout:Created symlink /etc/systemd/system/ceph.target.wants/ceph-crash.service → /usr/lib/systemd/system/ceph-crash.service. 2026-03-06T13:43:49.875 INFO:teuthology.orchestra.run.vm02.stdout: 2026-03-06T13:43:49.881 INFO:teuthology.orchestra.run.vm02.stdout: Installing : ceph-selinux-2:19.2.3-47.gc24117fd552.el9.clyso. 118/152 2026-03-06T13:43:49.905 INFO:teuthology.orchestra.run.vm06.stdout: Installing : python3-cheroot-10.0.1-4.el9.noarch 109/152 2026-03-06T13:43:49.979 INFO:teuthology.orchestra.run.vm06.stdout: Installing : python3-cherrypy-18.6.1-2.el9.noarch 110/152 2026-03-06T13:43:49.991 INFO:teuthology.orchestra.run.vm06.stdout: Installing : python3-asyncssh-2.13.2-5.el9.noarch 111/152 2026-03-06T13:43:49.998 INFO:teuthology.orchestra.run.vm06.stdout: Installing : python3-bcrypt-3.2.2-1.el9.x86_64 112/152 2026-03-06T13:43:50.006 INFO:teuthology.orchestra.run.vm06.stdout: Installing : pciutils-3.7.0-7.el9.x86_64 113/152 2026-03-06T13:43:50.012 INFO:teuthology.orchestra.run.vm06.stdout: Installing : qatlib-25.08.0-2.el9.x86_64 114/152 2026-03-06T13:43:50.043 INFO:teuthology.orchestra.run.vm06.stdout: Installing : qatlib-service-25.08.0-2.el9.x86_64 115/152 2026-03-06T13:43:50.074 INFO:teuthology.orchestra.run.vm06.stdout: Running scriptlet: qatlib-service-25.08.0-2.el9.x86_64 115/152 2026-03-06T13:43:50.418 INFO:teuthology.orchestra.run.vm06.stdout: Installing : qatzip-libs-1.3.1-1.el9.x86_64 116/152 2026-03-06T13:43:50.425 INFO:teuthology.orchestra.run.vm06.stdout: Installing : ceph-base-2:19.2.3-47.gc24117fd552.el9.clyso.x86 117/152 2026-03-06T13:43:50.477 INFO:teuthology.orchestra.run.vm06.stdout: Running scriptlet: ceph-base-2:19.2.3-47.gc24117fd552.el9.clyso.x86 117/152 2026-03-06T13:43:50.478 INFO:teuthology.orchestra.run.vm06.stdout:Created symlink /etc/systemd/system/multi-user.target.wants/ceph.target → /usr/lib/systemd/system/ceph.target. 2026-03-06T13:43:50.478 INFO:teuthology.orchestra.run.vm06.stdout:Created symlink /etc/systemd/system/ceph.target.wants/ceph-crash.service → /usr/lib/systemd/system/ceph-crash.service. 2026-03-06T13:43:50.478 INFO:teuthology.orchestra.run.vm06.stdout: 2026-03-06T13:43:50.486 INFO:teuthology.orchestra.run.vm06.stdout: Installing : ceph-selinux-2:19.2.3-47.gc24117fd552.el9.clyso. 118/152 2026-03-06T13:43:50.597 INFO:teuthology.orchestra.run.vm04.stdout: Running scriptlet: ceph-common-2:19.2.3-47.gc24117fd552.el9.clyso.x 102/152 2026-03-06T13:43:50.642 INFO:teuthology.orchestra.run.vm04.stdout: Installing : ceph-common-2:19.2.3-47.gc24117fd552.el9.clyso.x 102/152 2026-03-06T13:43:50.675 INFO:teuthology.orchestra.run.vm04.stdout: Running scriptlet: ceph-common-2:19.2.3-47.gc24117fd552.el9.clyso.x 102/152 2026-03-06T13:43:50.679 INFO:teuthology.orchestra.run.vm04.stdout: Installing : smartmontools-1:7.2-10.el9.x86_64 103/152 2026-03-06T13:43:50.695 INFO:teuthology.orchestra.run.vm04.stdout: Running scriptlet: smartmontools-1:7.2-10.el9.x86_64 103/152 2026-03-06T13:43:50.695 INFO:teuthology.orchestra.run.vm04.stdout:Created symlink /etc/systemd/system/multi-user.target.wants/smartd.service → /usr/lib/systemd/system/smartd.service. 2026-03-06T13:43:50.695 INFO:teuthology.orchestra.run.vm04.stdout: 2026-03-06T13:43:50.724 INFO:teuthology.orchestra.run.vm04.stdout: Installing : python3-ply-3.11-14.el9.noarch 104/152 2026-03-06T13:43:50.748 INFO:teuthology.orchestra.run.vm04.stdout: Installing : python3-pycparser-2.20-6.el9.noarch 105/152 2026-03-06T13:43:50.846 INFO:teuthology.orchestra.run.vm04.stdout: Installing : python3-cffi-1.14.5-5.el9.x86_64 106/152 2026-03-06T13:43:50.862 INFO:teuthology.orchestra.run.vm04.stdout: Installing : python3-cryptography-36.0.1-5.el9.x86_64 107/152 2026-03-06T13:43:50.895 INFO:teuthology.orchestra.run.vm04.stdout: Installing : python3-pyOpenSSL-21.0.0-1.el9.noarch 108/152 2026-03-06T13:43:50.934 INFO:teuthology.orchestra.run.vm04.stdout: Installing : python3-cheroot-10.0.1-4.el9.noarch 109/152 2026-03-06T13:43:50.998 INFO:teuthology.orchestra.run.vm04.stdout: Installing : python3-cherrypy-18.6.1-2.el9.noarch 110/152 2026-03-06T13:43:51.010 INFO:teuthology.orchestra.run.vm04.stdout: Installing : python3-asyncssh-2.13.2-5.el9.noarch 111/152 2026-03-06T13:43:51.017 INFO:teuthology.orchestra.run.vm04.stdout: Installing : python3-bcrypt-3.2.2-1.el9.x86_64 112/152 2026-03-06T13:43:51.023 INFO:teuthology.orchestra.run.vm04.stdout: Installing : pciutils-3.7.0-7.el9.x86_64 113/152 2026-03-06T13:43:51.027 INFO:teuthology.orchestra.run.vm04.stdout: Installing : qatlib-25.08.0-2.el9.x86_64 114/152 2026-03-06T13:43:51.030 INFO:teuthology.orchestra.run.vm04.stdout: Installing : qatlib-service-25.08.0-2.el9.x86_64 115/152 2026-03-06T13:43:51.046 INFO:teuthology.orchestra.run.vm04.stdout: Running scriptlet: qatlib-service-25.08.0-2.el9.x86_64 115/152 2026-03-06T13:43:51.362 INFO:teuthology.orchestra.run.vm04.stdout: Installing : qatzip-libs-1.3.1-1.el9.x86_64 116/152 2026-03-06T13:43:51.368 INFO:teuthology.orchestra.run.vm04.stdout: Installing : ceph-base-2:19.2.3-47.gc24117fd552.el9.clyso.x86 117/152 2026-03-06T13:43:51.420 INFO:teuthology.orchestra.run.vm04.stdout: Running scriptlet: ceph-base-2:19.2.3-47.gc24117fd552.el9.clyso.x86 117/152 2026-03-06T13:43:51.420 INFO:teuthology.orchestra.run.vm04.stdout:Created symlink /etc/systemd/system/multi-user.target.wants/ceph.target → /usr/lib/systemd/system/ceph.target. 2026-03-06T13:43:51.420 INFO:teuthology.orchestra.run.vm04.stdout:Created symlink /etc/systemd/system/ceph.target.wants/ceph-crash.service → /usr/lib/systemd/system/ceph-crash.service. 2026-03-06T13:43:51.420 INFO:teuthology.orchestra.run.vm04.stdout: 2026-03-06T13:43:51.425 INFO:teuthology.orchestra.run.vm04.stdout: Installing : ceph-selinux-2:19.2.3-47.gc24117fd552.el9.clyso. 118/152 2026-03-06T13:43:57.793 INFO:teuthology.orchestra.run.vm02.stdout: Running scriptlet: ceph-selinux-2:19.2.3-47.gc24117fd552.el9.clyso. 118/152 2026-03-06T13:43:57.793 INFO:teuthology.orchestra.run.vm02.stdout:skipping the directory /sys 2026-03-06T13:43:57.793 INFO:teuthology.orchestra.run.vm02.stdout:skipping the directory /proc 2026-03-06T13:43:57.793 INFO:teuthology.orchestra.run.vm02.stdout:skipping the directory /mnt 2026-03-06T13:43:57.793 INFO:teuthology.orchestra.run.vm02.stdout:skipping the directory /var/tmp 2026-03-06T13:43:57.793 INFO:teuthology.orchestra.run.vm02.stdout:skipping the directory /home 2026-03-06T13:43:57.793 INFO:teuthology.orchestra.run.vm02.stdout:skipping the directory /root 2026-03-06T13:43:57.793 INFO:teuthology.orchestra.run.vm02.stdout:skipping the directory /tmp 2026-03-06T13:43:57.793 INFO:teuthology.orchestra.run.vm02.stdout: 2026-03-06T13:43:57.927 INFO:teuthology.orchestra.run.vm02.stdout: Installing : ceph-mds-2:19.2.3-47.gc24117fd552.el9.clyso.x86_ 119/152 2026-03-06T13:43:57.955 INFO:teuthology.orchestra.run.vm02.stdout: Running scriptlet: ceph-mds-2:19.2.3-47.gc24117fd552.el9.clyso.x86_ 119/152 2026-03-06T13:43:57.955 INFO:teuthology.orchestra.run.vm02.stdout:Glob pattern passed to enable, but globs are not supported for this. 2026-03-06T13:43:57.955 INFO:teuthology.orchestra.run.vm02.stdout:Invalid unit name "ceph-mds@*.service" escaped as "ceph-mds@\x2a.service". 2026-03-06T13:43:57.955 INFO:teuthology.orchestra.run.vm02.stdout:Created symlink /etc/systemd/system/multi-user.target.wants/ceph-mds.target → /usr/lib/systemd/system/ceph-mds.target. 2026-03-06T13:43:57.955 INFO:teuthology.orchestra.run.vm02.stdout:Created symlink /etc/systemd/system/ceph.target.wants/ceph-mds.target → /usr/lib/systemd/system/ceph-mds.target. 2026-03-06T13:43:57.955 INFO:teuthology.orchestra.run.vm02.stdout: 2026-03-06T13:43:58.200 INFO:teuthology.orchestra.run.vm02.stdout: Installing : ceph-mon-2:19.2.3-47.gc24117fd552.el9.clyso.x86_ 120/152 2026-03-06T13:43:58.225 INFO:teuthology.orchestra.run.vm02.stdout: Running scriptlet: ceph-mon-2:19.2.3-47.gc24117fd552.el9.clyso.x86_ 120/152 2026-03-06T13:43:58.225 INFO:teuthology.orchestra.run.vm02.stdout:Glob pattern passed to enable, but globs are not supported for this. 2026-03-06T13:43:58.225 INFO:teuthology.orchestra.run.vm02.stdout:Invalid unit name "ceph-mon@*.service" escaped as "ceph-mon@\x2a.service". 2026-03-06T13:43:58.225 INFO:teuthology.orchestra.run.vm02.stdout:Created symlink /etc/systemd/system/multi-user.target.wants/ceph-mon.target → /usr/lib/systemd/system/ceph-mon.target. 2026-03-06T13:43:58.225 INFO:teuthology.orchestra.run.vm02.stdout:Created symlink /etc/systemd/system/ceph.target.wants/ceph-mon.target → /usr/lib/systemd/system/ceph-mon.target. 2026-03-06T13:43:58.225 INFO:teuthology.orchestra.run.vm02.stdout: 2026-03-06T13:43:58.235 INFO:teuthology.orchestra.run.vm02.stdout: Installing : mailcap-2.1.49-5.el9.noarch 121/152 2026-03-06T13:43:58.239 INFO:teuthology.orchestra.run.vm02.stdout: Installing : libconfig-1.7.2-9.el9.x86_64 122/152 2026-03-06T13:43:58.244 INFO:teuthology.orchestra.run.vm04.stdout: Running scriptlet: ceph-selinux-2:19.2.3-47.gc24117fd552.el9.clyso. 118/152 2026-03-06T13:43:58.244 INFO:teuthology.orchestra.run.vm04.stdout:skipping the directory /sys 2026-03-06T13:43:58.244 INFO:teuthology.orchestra.run.vm04.stdout:skipping the directory /proc 2026-03-06T13:43:58.244 INFO:teuthology.orchestra.run.vm04.stdout:skipping the directory /mnt 2026-03-06T13:43:58.244 INFO:teuthology.orchestra.run.vm04.stdout:skipping the directory /var/tmp 2026-03-06T13:43:58.244 INFO:teuthology.orchestra.run.vm04.stdout:skipping the directory /home 2026-03-06T13:43:58.244 INFO:teuthology.orchestra.run.vm04.stdout:skipping the directory /root 2026-03-06T13:43:58.244 INFO:teuthology.orchestra.run.vm04.stdout:skipping the directory /tmp 2026-03-06T13:43:58.244 INFO:teuthology.orchestra.run.vm04.stdout: 2026-03-06T13:43:58.258 INFO:teuthology.orchestra.run.vm02.stdout: Running scriptlet: libstoragemgmt-1.10.1-1.el9.x86_64 123/152 2026-03-06T13:43:58.258 INFO:teuthology.orchestra.run.vm02.stdout:Creating group 'qat' with GID 994. 2026-03-06T13:43:58.258 INFO:teuthology.orchestra.run.vm02.stdout:Creating group 'libstoragemgmt' with GID 993. 2026-03-06T13:43:58.258 INFO:teuthology.orchestra.run.vm02.stdout:Creating user 'libstoragemgmt' (daemon account for libstoragemgmt) with UID 993 and GID 993. 2026-03-06T13:43:58.258 INFO:teuthology.orchestra.run.vm02.stdout: 2026-03-06T13:43:58.269 INFO:teuthology.orchestra.run.vm02.stdout: Installing : libstoragemgmt-1.10.1-1.el9.x86_64 123/152 2026-03-06T13:43:58.294 INFO:teuthology.orchestra.run.vm06.stdout: Running scriptlet: ceph-selinux-2:19.2.3-47.gc24117fd552.el9.clyso. 118/152 2026-03-06T13:43:58.294 INFO:teuthology.orchestra.run.vm06.stdout:skipping the directory /sys 2026-03-06T13:43:58.294 INFO:teuthology.orchestra.run.vm06.stdout:skipping the directory /proc 2026-03-06T13:43:58.294 INFO:teuthology.orchestra.run.vm06.stdout:skipping the directory /mnt 2026-03-06T13:43:58.294 INFO:teuthology.orchestra.run.vm06.stdout:skipping the directory /var/tmp 2026-03-06T13:43:58.294 INFO:teuthology.orchestra.run.vm06.stdout:skipping the directory /home 2026-03-06T13:43:58.294 INFO:teuthology.orchestra.run.vm06.stdout:skipping the directory /root 2026-03-06T13:43:58.294 INFO:teuthology.orchestra.run.vm06.stdout:skipping the directory /tmp 2026-03-06T13:43:58.294 INFO:teuthology.orchestra.run.vm06.stdout: 2026-03-06T13:43:58.299 INFO:teuthology.orchestra.run.vm02.stdout: Running scriptlet: libstoragemgmt-1.10.1-1.el9.x86_64 123/152 2026-03-06T13:43:58.299 INFO:teuthology.orchestra.run.vm02.stdout:Created symlink /etc/systemd/system/multi-user.target.wants/libstoragemgmt.service → /usr/lib/systemd/system/libstoragemgmt.service. 2026-03-06T13:43:58.299 INFO:teuthology.orchestra.run.vm02.stdout: 2026-03-06T13:43:58.321 INFO:teuthology.orchestra.run.vm02.stdout: Installing : python3-libstoragemgmt-1.10.1-1.el9.x86_64 124/152 2026-03-06T13:43:58.348 INFO:teuthology.orchestra.run.vm02.stdout: Installing : fuse-2.9.9-17.el9.x86_64 125/152 2026-03-06T13:43:58.372 INFO:teuthology.orchestra.run.vm04.stdout: Installing : ceph-mds-2:19.2.3-47.gc24117fd552.el9.clyso.x86_ 119/152 2026-03-06T13:43:58.403 INFO:teuthology.orchestra.run.vm04.stdout: Running scriptlet: ceph-mds-2:19.2.3-47.gc24117fd552.el9.clyso.x86_ 119/152 2026-03-06T13:43:58.403 INFO:teuthology.orchestra.run.vm04.stdout:Glob pattern passed to enable, but globs are not supported for this. 2026-03-06T13:43:58.403 INFO:teuthology.orchestra.run.vm04.stdout:Invalid unit name "ceph-mds@*.service" escaped as "ceph-mds@\x2a.service". 2026-03-06T13:43:58.403 INFO:teuthology.orchestra.run.vm04.stdout:Created symlink /etc/systemd/system/multi-user.target.wants/ceph-mds.target → /usr/lib/systemd/system/ceph-mds.target. 2026-03-06T13:43:58.403 INFO:teuthology.orchestra.run.vm04.stdout:Created symlink /etc/systemd/system/ceph.target.wants/ceph-mds.target → /usr/lib/systemd/system/ceph-mds.target. 2026-03-06T13:43:58.403 INFO:teuthology.orchestra.run.vm04.stdout: 2026-03-06T13:43:58.424 INFO:teuthology.orchestra.run.vm02.stdout: Installing : cryptsetup-2.8.1-3.el9.x86_64 126/152 2026-03-06T13:43:58.429 INFO:teuthology.orchestra.run.vm02.stdout: Installing : ceph-volume-2:19.2.3-47.gc24117fd552.el9.clyso.n 127/152 2026-03-06T13:43:58.432 INFO:teuthology.orchestra.run.vm06.stdout: Installing : ceph-mds-2:19.2.3-47.gc24117fd552.el9.clyso.x86_ 119/152 2026-03-06T13:43:58.444 INFO:teuthology.orchestra.run.vm02.stdout: Running scriptlet: ceph-volume-2:19.2.3-47.gc24117fd552.el9.clyso.n 127/152 2026-03-06T13:43:58.444 INFO:teuthology.orchestra.run.vm02.stdout:Glob pattern passed to enable, but globs are not supported for this. 2026-03-06T13:43:58.444 INFO:teuthology.orchestra.run.vm02.stdout:Invalid unit name "ceph-volume@*.service" escaped as "ceph-volume@\x2a.service". 2026-03-06T13:43:58.444 INFO:teuthology.orchestra.run.vm02.stdout: 2026-03-06T13:43:58.460 INFO:teuthology.orchestra.run.vm06.stdout: Running scriptlet: ceph-mds-2:19.2.3-47.gc24117fd552.el9.clyso.x86_ 119/152 2026-03-06T13:43:58.460 INFO:teuthology.orchestra.run.vm06.stdout:Glob pattern passed to enable, but globs are not supported for this. 2026-03-06T13:43:58.460 INFO:teuthology.orchestra.run.vm06.stdout:Invalid unit name "ceph-mds@*.service" escaped as "ceph-mds@\x2a.service". 2026-03-06T13:43:58.460 INFO:teuthology.orchestra.run.vm06.stdout:Created symlink /etc/systemd/system/multi-user.target.wants/ceph-mds.target → /usr/lib/systemd/system/ceph-mds.target. 2026-03-06T13:43:58.460 INFO:teuthology.orchestra.run.vm06.stdout:Created symlink /etc/systemd/system/ceph.target.wants/ceph-mds.target → /usr/lib/systemd/system/ceph-mds.target. 2026-03-06T13:43:58.460 INFO:teuthology.orchestra.run.vm06.stdout: 2026-03-06T13:43:58.634 INFO:teuthology.orchestra.run.vm04.stdout: Installing : ceph-mon-2:19.2.3-47.gc24117fd552.el9.clyso.x86_ 120/152 2026-03-06T13:43:58.659 INFO:teuthology.orchestra.run.vm04.stdout: Running scriptlet: ceph-mon-2:19.2.3-47.gc24117fd552.el9.clyso.x86_ 120/152 2026-03-06T13:43:58.659 INFO:teuthology.orchestra.run.vm04.stdout:Glob pattern passed to enable, but globs are not supported for this. 2026-03-06T13:43:58.659 INFO:teuthology.orchestra.run.vm04.stdout:Invalid unit name "ceph-mon@*.service" escaped as "ceph-mon@\x2a.service". 2026-03-06T13:43:58.659 INFO:teuthology.orchestra.run.vm04.stdout:Created symlink /etc/systemd/system/multi-user.target.wants/ceph-mon.target → /usr/lib/systemd/system/ceph-mon.target. 2026-03-06T13:43:58.659 INFO:teuthology.orchestra.run.vm04.stdout:Created symlink /etc/systemd/system/ceph.target.wants/ceph-mon.target → /usr/lib/systemd/system/ceph-mon.target. 2026-03-06T13:43:58.659 INFO:teuthology.orchestra.run.vm04.stdout: 2026-03-06T13:43:58.668 INFO:teuthology.orchestra.run.vm04.stdout: Installing : mailcap-2.1.49-5.el9.noarch 121/152 2026-03-06T13:43:58.671 INFO:teuthology.orchestra.run.vm04.stdout: Installing : libconfig-1.7.2-9.el9.x86_64 122/152 2026-03-06T13:43:58.692 INFO:teuthology.orchestra.run.vm04.stdout: Running scriptlet: libstoragemgmt-1.10.1-1.el9.x86_64 123/152 2026-03-06T13:43:58.692 INFO:teuthology.orchestra.run.vm04.stdout:Creating group 'qat' with GID 994. 2026-03-06T13:43:58.692 INFO:teuthology.orchestra.run.vm04.stdout:Creating group 'libstoragemgmt' with GID 993. 2026-03-06T13:43:58.692 INFO:teuthology.orchestra.run.vm04.stdout:Creating user 'libstoragemgmt' (daemon account for libstoragemgmt) with UID 993 and GID 993. 2026-03-06T13:43:58.692 INFO:teuthology.orchestra.run.vm04.stdout: 2026-03-06T13:43:58.703 INFO:teuthology.orchestra.run.vm06.stdout: Installing : ceph-mon-2:19.2.3-47.gc24117fd552.el9.clyso.x86_ 120/152 2026-03-06T13:43:58.704 INFO:teuthology.orchestra.run.vm04.stdout: Installing : libstoragemgmt-1.10.1-1.el9.x86_64 123/152 2026-03-06T13:43:58.727 INFO:teuthology.orchestra.run.vm06.stdout: Running scriptlet: ceph-mon-2:19.2.3-47.gc24117fd552.el9.clyso.x86_ 120/152 2026-03-06T13:43:58.727 INFO:teuthology.orchestra.run.vm06.stdout:Glob pattern passed to enable, but globs are not supported for this. 2026-03-06T13:43:58.727 INFO:teuthology.orchestra.run.vm06.stdout:Invalid unit name "ceph-mon@*.service" escaped as "ceph-mon@\x2a.service". 2026-03-06T13:43:58.727 INFO:teuthology.orchestra.run.vm06.stdout:Created symlink /etc/systemd/system/multi-user.target.wants/ceph-mon.target → /usr/lib/systemd/system/ceph-mon.target. 2026-03-06T13:43:58.727 INFO:teuthology.orchestra.run.vm06.stdout:Created symlink /etc/systemd/system/ceph.target.wants/ceph-mon.target → /usr/lib/systemd/system/ceph-mon.target. 2026-03-06T13:43:58.727 INFO:teuthology.orchestra.run.vm06.stdout: 2026-03-06T13:43:58.731 INFO:teuthology.orchestra.run.vm04.stdout: Running scriptlet: libstoragemgmt-1.10.1-1.el9.x86_64 123/152 2026-03-06T13:43:58.731 INFO:teuthology.orchestra.run.vm04.stdout:Created symlink /etc/systemd/system/multi-user.target.wants/libstoragemgmt.service → /usr/lib/systemd/system/libstoragemgmt.service. 2026-03-06T13:43:58.731 INFO:teuthology.orchestra.run.vm04.stdout: 2026-03-06T13:43:58.736 INFO:teuthology.orchestra.run.vm06.stdout: Installing : mailcap-2.1.49-5.el9.noarch 121/152 2026-03-06T13:43:58.739 INFO:teuthology.orchestra.run.vm06.stdout: Installing : libconfig-1.7.2-9.el9.x86_64 122/152 2026-03-06T13:43:58.754 INFO:teuthology.orchestra.run.vm04.stdout: Installing : python3-libstoragemgmt-1.10.1-1.el9.x86_64 124/152 2026-03-06T13:43:58.759 INFO:teuthology.orchestra.run.vm06.stdout: Running scriptlet: libstoragemgmt-1.10.1-1.el9.x86_64 123/152 2026-03-06T13:43:58.759 INFO:teuthology.orchestra.run.vm06.stdout:Creating group 'qat' with GID 994. 2026-03-06T13:43:58.759 INFO:teuthology.orchestra.run.vm06.stdout:Creating group 'libstoragemgmt' with GID 993. 2026-03-06T13:43:58.759 INFO:teuthology.orchestra.run.vm06.stdout:Creating user 'libstoragemgmt' (daemon account for libstoragemgmt) with UID 993 and GID 993. 2026-03-06T13:43:58.759 INFO:teuthology.orchestra.run.vm06.stdout: 2026-03-06T13:43:58.770 INFO:teuthology.orchestra.run.vm06.stdout: Installing : libstoragemgmt-1.10.1-1.el9.x86_64 123/152 2026-03-06T13:43:58.783 INFO:teuthology.orchestra.run.vm04.stdout: Installing : fuse-2.9.9-17.el9.x86_64 125/152 2026-03-06T13:43:58.805 INFO:teuthology.orchestra.run.vm06.stdout: Running scriptlet: libstoragemgmt-1.10.1-1.el9.x86_64 123/152 2026-03-06T13:43:58.805 INFO:teuthology.orchestra.run.vm06.stdout:Created symlink /etc/systemd/system/multi-user.target.wants/libstoragemgmt.service → /usr/lib/systemd/system/libstoragemgmt.service. 2026-03-06T13:43:58.805 INFO:teuthology.orchestra.run.vm06.stdout: 2026-03-06T13:43:58.827 INFO:teuthology.orchestra.run.vm06.stdout: Installing : python3-libstoragemgmt-1.10.1-1.el9.x86_64 124/152 2026-03-06T13:43:58.857 INFO:teuthology.orchestra.run.vm06.stdout: Installing : fuse-2.9.9-17.el9.x86_64 125/152 2026-03-06T13:43:58.866 INFO:teuthology.orchestra.run.vm04.stdout: Installing : cryptsetup-2.8.1-3.el9.x86_64 126/152 2026-03-06T13:43:58.872 INFO:teuthology.orchestra.run.vm04.stdout: Installing : ceph-volume-2:19.2.3-47.gc24117fd552.el9.clyso.n 127/152 2026-03-06T13:43:58.884 INFO:teuthology.orchestra.run.vm04.stdout: Running scriptlet: ceph-volume-2:19.2.3-47.gc24117fd552.el9.clyso.n 127/152 2026-03-06T13:43:58.884 INFO:teuthology.orchestra.run.vm04.stdout:Glob pattern passed to enable, but globs are not supported for this. 2026-03-06T13:43:58.884 INFO:teuthology.orchestra.run.vm04.stdout:Invalid unit name "ceph-volume@*.service" escaped as "ceph-volume@\x2a.service". 2026-03-06T13:43:58.884 INFO:teuthology.orchestra.run.vm04.stdout: 2026-03-06T13:43:58.938 INFO:teuthology.orchestra.run.vm06.stdout: Installing : cryptsetup-2.8.1-3.el9.x86_64 126/152 2026-03-06T13:43:58.944 INFO:teuthology.orchestra.run.vm06.stdout: Installing : ceph-volume-2:19.2.3-47.gc24117fd552.el9.clyso.n 127/152 2026-03-06T13:43:58.959 INFO:teuthology.orchestra.run.vm06.stdout: Running scriptlet: ceph-volume-2:19.2.3-47.gc24117fd552.el9.clyso.n 127/152 2026-03-06T13:43:58.959 INFO:teuthology.orchestra.run.vm06.stdout:Glob pattern passed to enable, but globs are not supported for this. 2026-03-06T13:43:58.959 INFO:teuthology.orchestra.run.vm06.stdout:Invalid unit name "ceph-volume@*.service" escaped as "ceph-volume@\x2a.service". 2026-03-06T13:43:58.959 INFO:teuthology.orchestra.run.vm06.stdout: 2026-03-06T13:43:59.320 INFO:teuthology.orchestra.run.vm02.stdout: Installing : ceph-osd-2:19.2.3-47.gc24117fd552.el9.clyso.x86_ 128/152 2026-03-06T13:43:59.348 INFO:teuthology.orchestra.run.vm02.stdout: Running scriptlet: ceph-osd-2:19.2.3-47.gc24117fd552.el9.clyso.x86_ 128/152 2026-03-06T13:43:59.348 INFO:teuthology.orchestra.run.vm02.stdout:Glob pattern passed to enable, but globs are not supported for this. 2026-03-06T13:43:59.348 INFO:teuthology.orchestra.run.vm02.stdout:Invalid unit name "ceph-osd@*.service" escaped as "ceph-osd@\x2a.service". 2026-03-06T13:43:59.348 INFO:teuthology.orchestra.run.vm02.stdout:Created symlink /etc/systemd/system/multi-user.target.wants/ceph-osd.target → /usr/lib/systemd/system/ceph-osd.target. 2026-03-06T13:43:59.348 INFO:teuthology.orchestra.run.vm02.stdout:Created symlink /etc/systemd/system/ceph.target.wants/ceph-osd.target → /usr/lib/systemd/system/ceph-osd.target. 2026-03-06T13:43:59.348 INFO:teuthology.orchestra.run.vm02.stdout: 2026-03-06T13:43:59.416 INFO:teuthology.orchestra.run.vm02.stdout: Running scriptlet: cephadm-2:19.2.3-47.gc24117fd552.el9.clyso.noarc 129/152 2026-03-06T13:43:59.419 INFO:teuthology.orchestra.run.vm02.stdout: Installing : cephadm-2:19.2.3-47.gc24117fd552.el9.clyso.noarc 129/152 2026-03-06T13:43:59.426 INFO:teuthology.orchestra.run.vm02.stdout: Installing : ceph-prometheus-alerts-2:19.2.3-47.gc24117fd552. 130/152 2026-03-06T13:43:59.449 INFO:teuthology.orchestra.run.vm02.stdout: Installing : ceph-grafana-dashboards-2:19.2.3-47.gc24117fd552 131/152 2026-03-06T13:43:59.452 INFO:teuthology.orchestra.run.vm02.stdout: Installing : ceph-mgr-cephadm-2:19.2.3-47.gc24117fd552.el9.cl 132/152 2026-03-06T13:43:59.714 INFO:teuthology.orchestra.run.vm04.stdout: Installing : ceph-osd-2:19.2.3-47.gc24117fd552.el9.clyso.x86_ 128/152 2026-03-06T13:43:59.743 INFO:teuthology.orchestra.run.vm04.stdout: Running scriptlet: ceph-osd-2:19.2.3-47.gc24117fd552.el9.clyso.x86_ 128/152 2026-03-06T13:43:59.743 INFO:teuthology.orchestra.run.vm04.stdout:Glob pattern passed to enable, but globs are not supported for this. 2026-03-06T13:43:59.743 INFO:teuthology.orchestra.run.vm04.stdout:Invalid unit name "ceph-osd@*.service" escaped as "ceph-osd@\x2a.service". 2026-03-06T13:43:59.743 INFO:teuthology.orchestra.run.vm04.stdout:Created symlink /etc/systemd/system/multi-user.target.wants/ceph-osd.target → /usr/lib/systemd/system/ceph-osd.target. 2026-03-06T13:43:59.744 INFO:teuthology.orchestra.run.vm04.stdout:Created symlink /etc/systemd/system/ceph.target.wants/ceph-osd.target → /usr/lib/systemd/system/ceph-osd.target. 2026-03-06T13:43:59.744 INFO:teuthology.orchestra.run.vm04.stdout: 2026-03-06T13:43:59.817 INFO:teuthology.orchestra.run.vm04.stdout: Running scriptlet: cephadm-2:19.2.3-47.gc24117fd552.el9.clyso.noarc 129/152 2026-03-06T13:43:59.821 INFO:teuthology.orchestra.run.vm04.stdout: Installing : cephadm-2:19.2.3-47.gc24117fd552.el9.clyso.noarc 129/152 2026-03-06T13:43:59.827 INFO:teuthology.orchestra.run.vm04.stdout: Installing : ceph-prometheus-alerts-2:19.2.3-47.gc24117fd552. 130/152 2026-03-06T13:43:59.827 INFO:teuthology.orchestra.run.vm06.stdout: Installing : ceph-osd-2:19.2.3-47.gc24117fd552.el9.clyso.x86_ 128/152 2026-03-06T13:43:59.851 INFO:teuthology.orchestra.run.vm04.stdout: Installing : ceph-grafana-dashboards-2:19.2.3-47.gc24117fd552 131/152 2026-03-06T13:43:59.855 INFO:teuthology.orchestra.run.vm04.stdout: Installing : ceph-mgr-cephadm-2:19.2.3-47.gc24117fd552.el9.cl 132/152 2026-03-06T13:43:59.858 INFO:teuthology.orchestra.run.vm06.stdout: Running scriptlet: ceph-osd-2:19.2.3-47.gc24117fd552.el9.clyso.x86_ 128/152 2026-03-06T13:43:59.858 INFO:teuthology.orchestra.run.vm06.stdout:Glob pattern passed to enable, but globs are not supported for this. 2026-03-06T13:43:59.859 INFO:teuthology.orchestra.run.vm06.stdout:Invalid unit name "ceph-osd@*.service" escaped as "ceph-osd@\x2a.service". 2026-03-06T13:43:59.859 INFO:teuthology.orchestra.run.vm06.stdout:Created symlink /etc/systemd/system/multi-user.target.wants/ceph-osd.target → /usr/lib/systemd/system/ceph-osd.target. 2026-03-06T13:43:59.859 INFO:teuthology.orchestra.run.vm06.stdout:Created symlink /etc/systemd/system/ceph.target.wants/ceph-osd.target → /usr/lib/systemd/system/ceph-osd.target. 2026-03-06T13:43:59.859 INFO:teuthology.orchestra.run.vm06.stdout: 2026-03-06T13:43:59.932 INFO:teuthology.orchestra.run.vm06.stdout: Running scriptlet: cephadm-2:19.2.3-47.gc24117fd552.el9.clyso.noarc 129/152 2026-03-06T13:43:59.936 INFO:teuthology.orchestra.run.vm06.stdout: Installing : cephadm-2:19.2.3-47.gc24117fd552.el9.clyso.noarc 129/152 2026-03-06T13:43:59.943 INFO:teuthology.orchestra.run.vm06.stdout: Installing : ceph-prometheus-alerts-2:19.2.3-47.gc24117fd552. 130/152 2026-03-06T13:43:59.969 INFO:teuthology.orchestra.run.vm06.stdout: Installing : ceph-grafana-dashboards-2:19.2.3-47.gc24117fd552 131/152 2026-03-06T13:43:59.973 INFO:teuthology.orchestra.run.vm06.stdout: Installing : ceph-mgr-cephadm-2:19.2.3-47.gc24117fd552.el9.cl 132/152 2026-03-06T13:44:00.030 INFO:teuthology.orchestra.run.vm02.stdout: Running scriptlet: ceph-mgr-cephadm-2:19.2.3-47.gc24117fd552.el9.cl 132/152 2026-03-06T13:44:00.037 INFO:teuthology.orchestra.run.vm02.stdout: Installing : ceph-mgr-dashboard-2:19.2.3-47.gc24117fd552.el9. 133/152 2026-03-06T13:44:00.429 INFO:teuthology.orchestra.run.vm04.stdout: Running scriptlet: ceph-mgr-cephadm-2:19.2.3-47.gc24117fd552.el9.cl 132/152 2026-03-06T13:44:00.437 INFO:teuthology.orchestra.run.vm04.stdout: Installing : ceph-mgr-dashboard-2:19.2.3-47.gc24117fd552.el9. 133/152 2026-03-06T13:44:00.590 INFO:teuthology.orchestra.run.vm06.stdout: Running scriptlet: ceph-mgr-cephadm-2:19.2.3-47.gc24117fd552.el9.cl 132/152 2026-03-06T13:44:00.608 INFO:teuthology.orchestra.run.vm02.stdout: Running scriptlet: ceph-mgr-dashboard-2:19.2.3-47.gc24117fd552.el9. 133/152 2026-03-06T13:44:00.614 INFO:teuthology.orchestra.run.vm02.stdout: Installing : ceph-mgr-diskprediction-local-2:19.2.3-47.gc2411 134/152 2026-03-06T13:44:00.615 INFO:teuthology.orchestra.run.vm06.stdout: Installing : ceph-mgr-dashboard-2:19.2.3-47.gc24117fd552.el9. 133/152 2026-03-06T13:44:00.628 INFO:teuthology.orchestra.run.vm02.stdout: Running scriptlet: ceph-mgr-diskprediction-local-2:19.2.3-47.gc2411 134/152 2026-03-06T13:44:00.629 INFO:teuthology.orchestra.run.vm02.stdout: Installing : ceph-mgr-k8sevents-2:19.2.3-47.gc24117fd552.el9. 135/152 2026-03-06T13:44:00.691 INFO:teuthology.orchestra.run.vm02.stdout: Running scriptlet: ceph-mgr-k8sevents-2:19.2.3-47.gc24117fd552.el9. 135/152 2026-03-06T13:44:00.753 INFO:teuthology.orchestra.run.vm02.stdout: Installing : ceph-mgr-modules-core-2:19.2.3-47.gc24117fd552.e 136/152 2026-03-06T13:44:00.756 INFO:teuthology.orchestra.run.vm02.stdout: Installing : ceph-mgr-2:19.2.3-47.gc24117fd552.el9.clyso.x86_ 137/152 2026-03-06T13:44:00.778 INFO:teuthology.orchestra.run.vm02.stdout: Running scriptlet: ceph-mgr-2:19.2.3-47.gc24117fd552.el9.clyso.x86_ 137/152 2026-03-06T13:44:00.778 INFO:teuthology.orchestra.run.vm02.stdout:Glob pattern passed to enable, but globs are not supported for this. 2026-03-06T13:44:00.778 INFO:teuthology.orchestra.run.vm02.stdout:Invalid unit name "ceph-mgr@*.service" escaped as "ceph-mgr@\x2a.service". 2026-03-06T13:44:00.778 INFO:teuthology.orchestra.run.vm02.stdout:Created symlink /etc/systemd/system/multi-user.target.wants/ceph-mgr.target → /usr/lib/systemd/system/ceph-mgr.target. 2026-03-06T13:44:00.778 INFO:teuthology.orchestra.run.vm02.stdout:Created symlink /etc/systemd/system/ceph.target.wants/ceph-mgr.target → /usr/lib/systemd/system/ceph-mgr.target. 2026-03-06T13:44:00.778 INFO:teuthology.orchestra.run.vm02.stdout: 2026-03-06T13:44:00.792 INFO:teuthology.orchestra.run.vm02.stdout: Installing : ceph-mgr-rook-2:19.2.3-47.gc24117fd552.el9.clyso 138/152 2026-03-06T13:44:00.801 INFO:teuthology.orchestra.run.vm02.stdout: Running scriptlet: ceph-mgr-rook-2:19.2.3-47.gc24117fd552.el9.clyso 138/152 2026-03-06T13:44:00.862 INFO:teuthology.orchestra.run.vm02.stdout: Installing : ceph-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 139/152 2026-03-06T13:44:01.004 INFO:teuthology.orchestra.run.vm04.stdout: Running scriptlet: ceph-mgr-dashboard-2:19.2.3-47.gc24117fd552.el9. 133/152 2026-03-06T13:44:01.006 INFO:teuthology.orchestra.run.vm04.stdout: Installing : ceph-mgr-diskprediction-local-2:19.2.3-47.gc2411 134/152 2026-03-06T13:44:01.021 INFO:teuthology.orchestra.run.vm04.stdout: Running scriptlet: ceph-mgr-diskprediction-local-2:19.2.3-47.gc2411 134/152 2026-03-06T13:44:01.023 INFO:teuthology.orchestra.run.vm04.stdout: Installing : ceph-mgr-k8sevents-2:19.2.3-47.gc24117fd552.el9. 135/152 2026-03-06T13:44:01.087 INFO:teuthology.orchestra.run.vm04.stdout: Running scriptlet: ceph-mgr-k8sevents-2:19.2.3-47.gc24117fd552.el9. 135/152 2026-03-06T13:44:01.143 INFO:teuthology.orchestra.run.vm04.stdout: Installing : ceph-mgr-modules-core-2:19.2.3-47.gc24117fd552.e 136/152 2026-03-06T13:44:01.145 INFO:teuthology.orchestra.run.vm04.stdout: Installing : ceph-mgr-2:19.2.3-47.gc24117fd552.el9.clyso.x86_ 137/152 2026-03-06T13:44:01.170 INFO:teuthology.orchestra.run.vm04.stdout: Running scriptlet: ceph-mgr-2:19.2.3-47.gc24117fd552.el9.clyso.x86_ 137/152 2026-03-06T13:44:01.170 INFO:teuthology.orchestra.run.vm04.stdout:Glob pattern passed to enable, but globs are not supported for this. 2026-03-06T13:44:01.170 INFO:teuthology.orchestra.run.vm04.stdout:Invalid unit name "ceph-mgr@*.service" escaped as "ceph-mgr@\x2a.service". 2026-03-06T13:44:01.170 INFO:teuthology.orchestra.run.vm04.stdout:Created symlink /etc/systemd/system/multi-user.target.wants/ceph-mgr.target → /usr/lib/systemd/system/ceph-mgr.target. 2026-03-06T13:44:01.170 INFO:teuthology.orchestra.run.vm04.stdout:Created symlink /etc/systemd/system/ceph.target.wants/ceph-mgr.target → /usr/lib/systemd/system/ceph-mgr.target. 2026-03-06T13:44:01.170 INFO:teuthology.orchestra.run.vm04.stdout: 2026-03-06T13:44:01.184 INFO:teuthology.orchestra.run.vm04.stdout: Installing : ceph-mgr-rook-2:19.2.3-47.gc24117fd552.el9.clyso 138/152 2026-03-06T13:44:01.197 INFO:teuthology.orchestra.run.vm06.stdout: Running scriptlet: ceph-mgr-dashboard-2:19.2.3-47.gc24117fd552.el9. 133/152 2026-03-06T13:44:01.198 INFO:teuthology.orchestra.run.vm04.stdout: Running scriptlet: ceph-mgr-rook-2:19.2.3-47.gc24117fd552.el9.clyso 138/152 2026-03-06T13:44:01.200 INFO:teuthology.orchestra.run.vm06.stdout: Installing : ceph-mgr-diskprediction-local-2:19.2.3-47.gc2411 134/152 2026-03-06T13:44:01.214 INFO:teuthology.orchestra.run.vm06.stdout: Running scriptlet: ceph-mgr-diskprediction-local-2:19.2.3-47.gc2411 134/152 2026-03-06T13:44:01.216 INFO:teuthology.orchestra.run.vm06.stdout: Installing : ceph-mgr-k8sevents-2:19.2.3-47.gc24117fd552.el9. 135/152 2026-03-06T13:44:01.256 INFO:teuthology.orchestra.run.vm04.stdout: Installing : ceph-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 139/152 2026-03-06T13:44:01.284 INFO:teuthology.orchestra.run.vm06.stdout: Running scriptlet: ceph-mgr-k8sevents-2:19.2.3-47.gc24117fd552.el9. 135/152 2026-03-06T13:44:01.346 INFO:teuthology.orchestra.run.vm06.stdout: Installing : ceph-mgr-modules-core-2:19.2.3-47.gc24117fd552.e 136/152 2026-03-06T13:44:01.348 INFO:teuthology.orchestra.run.vm06.stdout: Installing : ceph-mgr-2:19.2.3-47.gc24117fd552.el9.clyso.x86_ 137/152 2026-03-06T13:44:01.373 INFO:teuthology.orchestra.run.vm06.stdout: Running scriptlet: ceph-mgr-2:19.2.3-47.gc24117fd552.el9.clyso.x86_ 137/152 2026-03-06T13:44:01.373 INFO:teuthology.orchestra.run.vm06.stdout:Glob pattern passed to enable, but globs are not supported for this. 2026-03-06T13:44:01.373 INFO:teuthology.orchestra.run.vm06.stdout:Invalid unit name "ceph-mgr@*.service" escaped as "ceph-mgr@\x2a.service". 2026-03-06T13:44:01.373 INFO:teuthology.orchestra.run.vm06.stdout:Created symlink /etc/systemd/system/multi-user.target.wants/ceph-mgr.target → /usr/lib/systemd/system/ceph-mgr.target. 2026-03-06T13:44:01.373 INFO:teuthology.orchestra.run.vm06.stdout:Created symlink /etc/systemd/system/ceph.target.wants/ceph-mgr.target → /usr/lib/systemd/system/ceph-mgr.target. 2026-03-06T13:44:01.373 INFO:teuthology.orchestra.run.vm06.stdout: 2026-03-06T13:44:01.388 INFO:teuthology.orchestra.run.vm06.stdout: Installing : ceph-mgr-rook-2:19.2.3-47.gc24117fd552.el9.clyso 138/152 2026-03-06T13:44:01.402 INFO:teuthology.orchestra.run.vm06.stdout: Running scriptlet: ceph-mgr-rook-2:19.2.3-47.gc24117fd552.el9.clyso 138/152 2026-03-06T13:44:01.421 INFO:teuthology.orchestra.run.vm02.stdout: Installing : ceph-fuse-2:19.2.3-47.gc24117fd552.el9.clyso.x86 140/152 2026-03-06T13:44:01.425 INFO:teuthology.orchestra.run.vm02.stdout: Installing : ceph-radosgw-2:19.2.3-47.gc24117fd552.el9.clyso. 141/152 2026-03-06T13:44:01.449 INFO:teuthology.orchestra.run.vm02.stdout: Running scriptlet: ceph-radosgw-2:19.2.3-47.gc24117fd552.el9.clyso. 141/152 2026-03-06T13:44:01.449 INFO:teuthology.orchestra.run.vm02.stdout:Glob pattern passed to enable, but globs are not supported for this. 2026-03-06T13:44:01.449 INFO:teuthology.orchestra.run.vm02.stdout:Invalid unit name "ceph-radosgw@*.service" escaped as "ceph-radosgw@\x2a.service". 2026-03-06T13:44:01.449 INFO:teuthology.orchestra.run.vm02.stdout:Created symlink /etc/systemd/system/multi-user.target.wants/ceph-radosgw.target → /usr/lib/systemd/system/ceph-radosgw.target. 2026-03-06T13:44:01.449 INFO:teuthology.orchestra.run.vm02.stdout:Created symlink /etc/systemd/system/ceph.target.wants/ceph-radosgw.target → /usr/lib/systemd/system/ceph-radosgw.target. 2026-03-06T13:44:01.449 INFO:teuthology.orchestra.run.vm02.stdout: 2026-03-06T13:44:01.462 INFO:teuthology.orchestra.run.vm02.stdout: Installing : ceph-immutable-object-cache-2:19.2.3-47.gc24117f 142/152 2026-03-06T13:44:01.463 INFO:teuthology.orchestra.run.vm06.stdout: Installing : ceph-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 139/152 2026-03-06T13:44:01.485 INFO:teuthology.orchestra.run.vm02.stdout: Running scriptlet: ceph-immutable-object-cache-2:19.2.3-47.gc24117f 142/152 2026-03-06T13:44:01.485 INFO:teuthology.orchestra.run.vm02.stdout:Glob pattern passed to enable, but globs are not supported for this. 2026-03-06T13:44:01.485 INFO:teuthology.orchestra.run.vm02.stdout:Invalid unit name "ceph-immutable-object-cache@*.service" escaped as "ceph-immutable-object-cache@\x2a.service". 2026-03-06T13:44:01.485 INFO:teuthology.orchestra.run.vm02.stdout: 2026-03-06T13:44:01.650 INFO:teuthology.orchestra.run.vm02.stdout: Installing : rbd-mirror-2:19.2.3-47.gc24117fd552.el9.clyso.x8 143/152 2026-03-06T13:44:01.677 INFO:teuthology.orchestra.run.vm02.stdout: Running scriptlet: rbd-mirror-2:19.2.3-47.gc24117fd552.el9.clyso.x8 143/152 2026-03-06T13:44:01.677 INFO:teuthology.orchestra.run.vm02.stdout:Glob pattern passed to enable, but globs are not supported for this. 2026-03-06T13:44:01.677 INFO:teuthology.orchestra.run.vm02.stdout:Invalid unit name "ceph-rbd-mirror@*.service" escaped as "ceph-rbd-mirror@\x2a.service". 2026-03-06T13:44:01.677 INFO:teuthology.orchestra.run.vm02.stdout:Created symlink /etc/systemd/system/multi-user.target.wants/ceph-rbd-mirror.target → /usr/lib/systemd/system/ceph-rbd-mirror.target. 2026-03-06T13:44:01.678 INFO:teuthology.orchestra.run.vm02.stdout:Created symlink /etc/systemd/system/ceph.target.wants/ceph-rbd-mirror.target → /usr/lib/systemd/system/ceph-rbd-mirror.target. 2026-03-06T13:44:01.678 INFO:teuthology.orchestra.run.vm02.stdout: 2026-03-06T13:44:01.787 INFO:teuthology.orchestra.run.vm04.stdout: Installing : ceph-fuse-2:19.2.3-47.gc24117fd552.el9.clyso.x86 140/152 2026-03-06T13:44:01.791 INFO:teuthology.orchestra.run.vm04.stdout: Installing : ceph-radosgw-2:19.2.3-47.gc24117fd552.el9.clyso. 141/152 2026-03-06T13:44:01.815 INFO:teuthology.orchestra.run.vm04.stdout: Running scriptlet: ceph-radosgw-2:19.2.3-47.gc24117fd552.el9.clyso. 141/152 2026-03-06T13:44:01.816 INFO:teuthology.orchestra.run.vm04.stdout:Glob pattern passed to enable, but globs are not supported for this. 2026-03-06T13:44:01.816 INFO:teuthology.orchestra.run.vm04.stdout:Invalid unit name "ceph-radosgw@*.service" escaped as "ceph-radosgw@\x2a.service". 2026-03-06T13:44:01.816 INFO:teuthology.orchestra.run.vm04.stdout:Created symlink /etc/systemd/system/multi-user.target.wants/ceph-radosgw.target → /usr/lib/systemd/system/ceph-radosgw.target. 2026-03-06T13:44:01.816 INFO:teuthology.orchestra.run.vm04.stdout:Created symlink /etc/systemd/system/ceph.target.wants/ceph-radosgw.target → /usr/lib/systemd/system/ceph-radosgw.target. 2026-03-06T13:44:01.816 INFO:teuthology.orchestra.run.vm04.stdout: 2026-03-06T13:44:01.827 INFO:teuthology.orchestra.run.vm04.stdout: Installing : ceph-immutable-object-cache-2:19.2.3-47.gc24117f 142/152 2026-03-06T13:44:01.851 INFO:teuthology.orchestra.run.vm04.stdout: Running scriptlet: ceph-immutable-object-cache-2:19.2.3-47.gc24117f 142/152 2026-03-06T13:44:01.851 INFO:teuthology.orchestra.run.vm04.stdout:Glob pattern passed to enable, but globs are not supported for this. 2026-03-06T13:44:01.851 INFO:teuthology.orchestra.run.vm04.stdout:Invalid unit name "ceph-immutable-object-cache@*.service" escaped as "ceph-immutable-object-cache@\x2a.service". 2026-03-06T13:44:01.851 INFO:teuthology.orchestra.run.vm04.stdout: 2026-03-06T13:44:02.011 INFO:teuthology.orchestra.run.vm04.stdout: Installing : rbd-mirror-2:19.2.3-47.gc24117fd552.el9.clyso.x8 143/152 2026-03-06T13:44:02.020 INFO:teuthology.orchestra.run.vm06.stdout: Installing : ceph-fuse-2:19.2.3-47.gc24117fd552.el9.clyso.x86 140/152 2026-03-06T13:44:02.023 INFO:teuthology.orchestra.run.vm06.stdout: Installing : ceph-radosgw-2:19.2.3-47.gc24117fd552.el9.clyso. 141/152 2026-03-06T13:44:02.038 INFO:teuthology.orchestra.run.vm04.stdout: Running scriptlet: rbd-mirror-2:19.2.3-47.gc24117fd552.el9.clyso.x8 143/152 2026-03-06T13:44:02.038 INFO:teuthology.orchestra.run.vm04.stdout:Glob pattern passed to enable, but globs are not supported for this. 2026-03-06T13:44:02.038 INFO:teuthology.orchestra.run.vm04.stdout:Invalid unit name "ceph-rbd-mirror@*.service" escaped as "ceph-rbd-mirror@\x2a.service". 2026-03-06T13:44:02.038 INFO:teuthology.orchestra.run.vm04.stdout:Created symlink /etc/systemd/system/multi-user.target.wants/ceph-rbd-mirror.target → /usr/lib/systemd/system/ceph-rbd-mirror.target. 2026-03-06T13:44:02.038 INFO:teuthology.orchestra.run.vm04.stdout:Created symlink /etc/systemd/system/ceph.target.wants/ceph-rbd-mirror.target → /usr/lib/systemd/system/ceph-rbd-mirror.target. 2026-03-06T13:44:02.038 INFO:teuthology.orchestra.run.vm04.stdout: 2026-03-06T13:44:02.047 INFO:teuthology.orchestra.run.vm06.stdout: Running scriptlet: ceph-radosgw-2:19.2.3-47.gc24117fd552.el9.clyso. 141/152 2026-03-06T13:44:02.047 INFO:teuthology.orchestra.run.vm06.stdout:Glob pattern passed to enable, but globs are not supported for this. 2026-03-06T13:44:02.047 INFO:teuthology.orchestra.run.vm06.stdout:Invalid unit name "ceph-radosgw@*.service" escaped as "ceph-radosgw@\x2a.service". 2026-03-06T13:44:02.047 INFO:teuthology.orchestra.run.vm06.stdout:Created symlink /etc/systemd/system/multi-user.target.wants/ceph-radosgw.target → /usr/lib/systemd/system/ceph-radosgw.target. 2026-03-06T13:44:02.047 INFO:teuthology.orchestra.run.vm06.stdout:Created symlink /etc/systemd/system/ceph.target.wants/ceph-radosgw.target → /usr/lib/systemd/system/ceph-radosgw.target. 2026-03-06T13:44:02.047 INFO:teuthology.orchestra.run.vm06.stdout: 2026-03-06T13:44:02.059 INFO:teuthology.orchestra.run.vm06.stdout: Installing : ceph-immutable-object-cache-2:19.2.3-47.gc24117f 142/152 2026-03-06T13:44:02.086 INFO:teuthology.orchestra.run.vm06.stdout: Running scriptlet: ceph-immutable-object-cache-2:19.2.3-47.gc24117f 142/152 2026-03-06T13:44:02.086 INFO:teuthology.orchestra.run.vm06.stdout:Glob pattern passed to enable, but globs are not supported for this. 2026-03-06T13:44:02.086 INFO:teuthology.orchestra.run.vm06.stdout:Invalid unit name "ceph-immutable-object-cache@*.service" escaped as "ceph-immutable-object-cache@\x2a.service". 2026-03-06T13:44:02.086 INFO:teuthology.orchestra.run.vm06.stdout: 2026-03-06T13:44:02.251 INFO:teuthology.orchestra.run.vm06.stdout: Installing : rbd-mirror-2:19.2.3-47.gc24117fd552.el9.clyso.x8 143/152 2026-03-06T13:44:02.275 INFO:teuthology.orchestra.run.vm06.stdout: Running scriptlet: rbd-mirror-2:19.2.3-47.gc24117fd552.el9.clyso.x8 143/152 2026-03-06T13:44:02.275 INFO:teuthology.orchestra.run.vm06.stdout:Glob pattern passed to enable, but globs are not supported for this. 2026-03-06T13:44:02.275 INFO:teuthology.orchestra.run.vm06.stdout:Invalid unit name "ceph-rbd-mirror@*.service" escaped as "ceph-rbd-mirror@\x2a.service". 2026-03-06T13:44:02.275 INFO:teuthology.orchestra.run.vm06.stdout:Created symlink /etc/systemd/system/multi-user.target.wants/ceph-rbd-mirror.target → /usr/lib/systemd/system/ceph-rbd-mirror.target. 2026-03-06T13:44:02.275 INFO:teuthology.orchestra.run.vm06.stdout:Created symlink /etc/systemd/system/ceph.target.wants/ceph-rbd-mirror.target → /usr/lib/systemd/system/ceph-rbd-mirror.target. 2026-03-06T13:44:02.275 INFO:teuthology.orchestra.run.vm06.stdout: 2026-03-06T13:44:04.394 INFO:teuthology.orchestra.run.vm02.stdout: Installing : ceph-test-2:19.2.3-47.gc24117fd552.el9.clyso.x86 144/152 2026-03-06T13:44:04.407 INFO:teuthology.orchestra.run.vm02.stdout: Installing : rbd-fuse-2:19.2.3-47.gc24117fd552.el9.clyso.x86_ 145/152 2026-03-06T13:44:04.442 INFO:teuthology.orchestra.run.vm02.stdout: Installing : rbd-nbd-2:19.2.3-47.gc24117fd552.el9.clyso.x86_6 146/152 2026-03-06T13:44:04.450 INFO:teuthology.orchestra.run.vm02.stdout: Installing : perl-Test-Harness-1:3.42-461.el9.noarch 147/152 2026-03-06T13:44:04.473 INFO:teuthology.orchestra.run.vm02.stdout: Installing : libcephfs-devel-2:19.2.3-47.gc24117fd552.el9.cly 148/152 2026-03-06T13:44:04.481 INFO:teuthology.orchestra.run.vm02.stdout: Installing : s3cmd-2.4.0-1.el9.noarch 149/152 2026-03-06T13:44:04.484 INFO:teuthology.orchestra.run.vm02.stdout: Installing : bzip2-1.0.8-11.el9.x86_64 150/152 2026-03-06T13:44:04.484 INFO:teuthology.orchestra.run.vm02.stdout: Cleanup : librbd1-2:16.2.4-5.el9.x86_64 151/152 2026-03-06T13:44:04.501 INFO:teuthology.orchestra.run.vm02.stdout: Running scriptlet: librbd1-2:16.2.4-5.el9.x86_64 151/152 2026-03-06T13:44:04.501 INFO:teuthology.orchestra.run.vm02.stdout: Cleanup : librados2-2:16.2.4-5.el9.x86_64 152/152 2026-03-06T13:44:04.674 INFO:teuthology.orchestra.run.vm04.stdout: Installing : ceph-test-2:19.2.3-47.gc24117fd552.el9.clyso.x86 144/152 2026-03-06T13:44:04.685 INFO:teuthology.orchestra.run.vm04.stdout: Installing : rbd-fuse-2:19.2.3-47.gc24117fd552.el9.clyso.x86_ 145/152 2026-03-06T13:44:04.721 INFO:teuthology.orchestra.run.vm04.stdout: Installing : rbd-nbd-2:19.2.3-47.gc24117fd552.el9.clyso.x86_6 146/152 2026-03-06T13:44:04.727 INFO:teuthology.orchestra.run.vm04.stdout: Installing : perl-Test-Harness-1:3.42-461.el9.noarch 147/152 2026-03-06T13:44:04.745 INFO:teuthology.orchestra.run.vm04.stdout: Installing : libcephfs-devel-2:19.2.3-47.gc24117fd552.el9.cly 148/152 2026-03-06T13:44:04.752 INFO:teuthology.orchestra.run.vm04.stdout: Installing : s3cmd-2.4.0-1.el9.noarch 149/152 2026-03-06T13:44:04.756 INFO:teuthology.orchestra.run.vm04.stdout: Installing : bzip2-1.0.8-11.el9.x86_64 150/152 2026-03-06T13:44:04.756 INFO:teuthology.orchestra.run.vm04.stdout: Cleanup : librbd1-2:16.2.4-5.el9.x86_64 151/152 2026-03-06T13:44:04.775 INFO:teuthology.orchestra.run.vm04.stdout: Running scriptlet: librbd1-2:16.2.4-5.el9.x86_64 151/152 2026-03-06T13:44:04.775 INFO:teuthology.orchestra.run.vm04.stdout: Cleanup : librados2-2:16.2.4-5.el9.x86_64 152/152 2026-03-06T13:44:05.017 INFO:teuthology.orchestra.run.vm06.stdout: Installing : ceph-test-2:19.2.3-47.gc24117fd552.el9.clyso.x86 144/152 2026-03-06T13:44:05.029 INFO:teuthology.orchestra.run.vm06.stdout: Installing : rbd-fuse-2:19.2.3-47.gc24117fd552.el9.clyso.x86_ 145/152 2026-03-06T13:44:05.070 INFO:teuthology.orchestra.run.vm06.stdout: Installing : rbd-nbd-2:19.2.3-47.gc24117fd552.el9.clyso.x86_6 146/152 2026-03-06T13:44:05.077 INFO:teuthology.orchestra.run.vm06.stdout: Installing : perl-Test-Harness-1:3.42-461.el9.noarch 147/152 2026-03-06T13:44:05.100 INFO:teuthology.orchestra.run.vm06.stdout: Installing : libcephfs-devel-2:19.2.3-47.gc24117fd552.el9.cly 148/152 2026-03-06T13:44:05.109 INFO:teuthology.orchestra.run.vm06.stdout: Installing : s3cmd-2.4.0-1.el9.noarch 149/152 2026-03-06T13:44:05.113 INFO:teuthology.orchestra.run.vm06.stdout: Installing : bzip2-1.0.8-11.el9.x86_64 150/152 2026-03-06T13:44:05.113 INFO:teuthology.orchestra.run.vm06.stdout: Cleanup : librbd1-2:16.2.4-5.el9.x86_64 151/152 2026-03-06T13:44:05.130 INFO:teuthology.orchestra.run.vm06.stdout: Running scriptlet: librbd1-2:16.2.4-5.el9.x86_64 151/152 2026-03-06T13:44:05.131 INFO:teuthology.orchestra.run.vm06.stdout: Cleanup : librados2-2:16.2.4-5.el9.x86_64 152/152 2026-03-06T13:44:06.288 INFO:teuthology.orchestra.run.vm02.stdout: Running scriptlet: librados2-2:16.2.4-5.el9.x86_64 152/152 2026-03-06T13:44:06.288 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : ceph-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 1/152 2026-03-06T13:44:06.288 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : ceph-base-2:19.2.3-47.gc24117fd552.el9.clyso.x86 2/152 2026-03-06T13:44:06.289 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : ceph-common-2:19.2.3-47.gc24117fd552.el9.clyso.x 3/152 2026-03-06T13:44:06.289 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : ceph-fuse-2:19.2.3-47.gc24117fd552.el9.clyso.x86 4/152 2026-03-06T13:44:06.289 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : ceph-immutable-object-cache-2:19.2.3-47.gc24117f 5/152 2026-03-06T13:44:06.289 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : ceph-mds-2:19.2.3-47.gc24117fd552.el9.clyso.x86_ 6/152 2026-03-06T13:44:06.289 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : ceph-mgr-2:19.2.3-47.gc24117fd552.el9.clyso.x86_ 7/152 2026-03-06T13:44:06.289 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : ceph-mon-2:19.2.3-47.gc24117fd552.el9.clyso.x86_ 8/152 2026-03-06T13:44:06.289 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : ceph-osd-2:19.2.3-47.gc24117fd552.el9.clyso.x86_ 9/152 2026-03-06T13:44:06.289 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : ceph-radosgw-2:19.2.3-47.gc24117fd552.el9.clyso. 10/152 2026-03-06T13:44:06.289 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : ceph-selinux-2:19.2.3-47.gc24117fd552.el9.clyso. 11/152 2026-03-06T13:44:06.289 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : ceph-test-2:19.2.3-47.gc24117fd552.el9.clyso.x86 12/152 2026-03-06T13:44:06.290 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : libcephfs-devel-2:19.2.3-47.gc24117fd552.el9.cly 13/152 2026-03-06T13:44:06.290 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : libcephfs2-2:19.2.3-47.gc24117fd552.el9.clyso.x8 14/152 2026-03-06T13:44:06.290 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : libcephsqlite-2:19.2.3-47.gc24117fd552.el9.clyso 15/152 2026-03-06T13:44:06.290 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : librados-devel-2:19.2.3-47.gc24117fd552.el9.clys 16/152 2026-03-06T13:44:06.290 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : libradosstriper1-2:19.2.3-47.gc24117fd552.el9.cl 17/152 2026-03-06T13:44:06.290 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : librgw2-2:19.2.3-47.gc24117fd552.el9.clyso.x86_6 18/152 2026-03-06T13:44:06.290 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : python3-ceph-argparse-2:19.2.3-47.gc24117fd552.e 19/152 2026-03-06T13:44:06.290 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : python3-ceph-common-2:19.2.3-47.gc24117fd552.el9 20/152 2026-03-06T13:44:06.290 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : python3-cephfs-2:19.2.3-47.gc24117fd552.el9.clys 21/152 2026-03-06T13:44:06.290 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : python3-rados-2:19.2.3-47.gc24117fd552.el9.clyso 22/152 2026-03-06T13:44:06.290 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : python3-rbd-2:19.2.3-47.gc24117fd552.el9.clyso.x 23/152 2026-03-06T13:44:06.290 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : python3-rgw-2:19.2.3-47.gc24117fd552.el9.clyso.x 24/152 2026-03-06T13:44:06.290 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : rbd-fuse-2:19.2.3-47.gc24117fd552.el9.clyso.x86_ 25/152 2026-03-06T13:44:06.290 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : rbd-mirror-2:19.2.3-47.gc24117fd552.el9.clyso.x8 26/152 2026-03-06T13:44:06.291 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : rbd-nbd-2:19.2.3-47.gc24117fd552.el9.clyso.x86_6 27/152 2026-03-06T13:44:06.291 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : ceph-grafana-dashboards-2:19.2.3-47.gc24117fd552 28/152 2026-03-06T13:44:06.291 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : ceph-mgr-cephadm-2:19.2.3-47.gc24117fd552.el9.cl 29/152 2026-03-06T13:44:06.291 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : ceph-mgr-dashboard-2:19.2.3-47.gc24117fd552.el9. 30/152 2026-03-06T13:44:06.291 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : ceph-mgr-diskprediction-local-2:19.2.3-47.gc2411 31/152 2026-03-06T13:44:06.291 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : ceph-mgr-k8sevents-2:19.2.3-47.gc24117fd552.el9. 32/152 2026-03-06T13:44:06.291 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : ceph-mgr-modules-core-2:19.2.3-47.gc24117fd552.e 33/152 2026-03-06T13:44:06.291 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : ceph-mgr-rook-2:19.2.3-47.gc24117fd552.el9.clyso 34/152 2026-03-06T13:44:06.291 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : ceph-prometheus-alerts-2:19.2.3-47.gc24117fd552. 35/152 2026-03-06T13:44:06.291 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : ceph-volume-2:19.2.3-47.gc24117fd552.el9.clyso.n 36/152 2026-03-06T13:44:06.291 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : cephadm-2:19.2.3-47.gc24117fd552.el9.clyso.noarc 37/152 2026-03-06T13:44:06.291 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : bzip2-1.0.8-11.el9.x86_64 38/152 2026-03-06T13:44:06.291 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : cryptsetup-2.8.1-3.el9.x86_64 39/152 2026-03-06T13:44:06.291 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : fuse-2.9.9-17.el9.x86_64 40/152 2026-03-06T13:44:06.291 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : ledmon-libs-1.1.0-3.el9.x86_64 41/152 2026-03-06T13:44:06.291 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : libconfig-1.7.2-9.el9.x86_64 42/152 2026-03-06T13:44:06.291 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : libgfortran-11.5.0-14.el9.x86_64 43/152 2026-03-06T13:44:06.291 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : libquadmath-11.5.0-14.el9.x86_64 44/152 2026-03-06T13:44:06.291 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : mailcap-2.1.49-5.el9.noarch 45/152 2026-03-06T13:44:06.291 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : pciutils-3.7.0-7.el9.x86_64 46/152 2026-03-06T13:44:06.291 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : python3-cffi-1.14.5-5.el9.x86_64 47/152 2026-03-06T13:44:06.291 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : python3-cryptography-36.0.1-5.el9.x86_64 48/152 2026-03-06T13:44:06.291 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : python3-ply-3.11-14.el9.noarch 49/152 2026-03-06T13:44:06.291 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : python3-pycparser-2.20-6.el9.noarch 50/152 2026-03-06T13:44:06.291 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : python3-requests-2.25.1-10.el9.noarch 51/152 2026-03-06T13:44:06.291 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : python3-urllib3-1.26.5-7.el9.noarch 52/152 2026-03-06T13:44:06.291 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : smartmontools-1:7.2-10.el9.x86_64 53/152 2026-03-06T13:44:06.291 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : unzip-6.0-59.el9.x86_64 54/152 2026-03-06T13:44:06.291 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : zip-3.0-35.el9.x86_64 55/152 2026-03-06T13:44:06.291 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : boost-program-options-1.75.0-13.el9.x86_64 56/152 2026-03-06T13:44:06.291 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : flexiblas-3.0.4-9.el9.x86_64 57/152 2026-03-06T13:44:06.291 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : flexiblas-netlib-3.0.4-9.el9.x86_64 58/152 2026-03-06T13:44:06.291 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : flexiblas-openblas-openmp-3.0.4-9.el9.x86_64 59/152 2026-03-06T13:44:06.292 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : libnbd-1.20.3-4.el9.x86_64 60/152 2026-03-06T13:44:06.292 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : libpmemobj-1.12.1-1.el9.x86_64 61/152 2026-03-06T13:44:06.292 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : librabbitmq-0.11.0-7.el9.x86_64 62/152 2026-03-06T13:44:06.292 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : librdkafka-1.6.1-102.el9.x86_64 63/152 2026-03-06T13:44:06.292 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : libstoragemgmt-1.10.1-1.el9.x86_64 64/152 2026-03-06T13:44:06.292 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : libxslt-1.1.34-12.el9.x86_64 65/152 2026-03-06T13:44:06.292 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : lttng-ust-2.12.0-6.el9.x86_64 66/152 2026-03-06T13:44:06.292 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : lua-5.4.4-4.el9.x86_64 67/152 2026-03-06T13:44:06.292 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : openblas-0.3.29-1.el9.x86_64 68/152 2026-03-06T13:44:06.292 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : openblas-openmp-0.3.29-1.el9.x86_64 69/152 2026-03-06T13:44:06.292 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : perl-Benchmark-1.23-483.el9.noarch 70/152 2026-03-06T13:44:06.292 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : perl-Test-Harness-1:3.42-461.el9.noarch 71/152 2026-03-06T13:44:06.293 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : protobuf-3.14.0-17.el9.x86_64 72/152 2026-03-06T13:44:06.293 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : python3-babel-2.9.1-2.el9.noarch 73/152 2026-03-06T13:44:06.293 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : python3-devel-3.9.25-3.el9.x86_64 74/152 2026-03-06T13:44:06.293 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : python3-jinja2-2.11.3-8.el9.noarch 75/152 2026-03-06T13:44:06.293 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : python3-libstoragemgmt-1.10.1-1.el9.x86_64 76/152 2026-03-06T13:44:06.293 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : python3-lxml-4.6.5-3.el9.x86_64 77/152 2026-03-06T13:44:06.293 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : python3-mako-1.1.4-6.el9.noarch 78/152 2026-03-06T13:44:06.293 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : python3-markupsafe-1.1.1-12.el9.x86_64 79/152 2026-03-06T13:44:06.293 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : python3-numpy-1:1.23.5-2.el9.x86_64 80/152 2026-03-06T13:44:06.293 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : python3-numpy-f2py-1:1.23.5-2.el9.x86_64 81/152 2026-03-06T13:44:06.293 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : python3-packaging-20.9-5.el9.noarch 82/152 2026-03-06T13:44:06.293 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : python3-protobuf-3.14.0-17.el9.noarch 83/152 2026-03-06T13:44:06.293 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : python3-pyasn1-0.4.8-7.el9.noarch 84/152 2026-03-06T13:44:06.293 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : python3-pyasn1-modules-0.4.8-7.el9.noarch 85/152 2026-03-06T13:44:06.293 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : python3-requests-oauthlib-1.3.0-12.el9.noarch 86/152 2026-03-06T13:44:06.293 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : python3-scipy-1.9.3-2.el9.x86_64 87/152 2026-03-06T13:44:06.293 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : python3-toml-0.10.2-6.el9.noarch 88/152 2026-03-06T13:44:06.293 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : qatlib-25.08.0-2.el9.x86_64 89/152 2026-03-06T13:44:06.293 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : qatlib-service-25.08.0-2.el9.x86_64 90/152 2026-03-06T13:44:06.293 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : qatzip-libs-1.3.1-1.el9.x86_64 91/152 2026-03-06T13:44:06.293 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : socat-1.7.4.1-8.el9.x86_64 92/152 2026-03-06T13:44:06.293 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : xmlsec1-1.2.29-13.el9.x86_64 93/152 2026-03-06T13:44:06.293 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : xmlsec1-openssl-1.2.29-13.el9.x86_64 94/152 2026-03-06T13:44:06.293 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : xmlstarlet-1.6.1-20.el9.x86_64 95/152 2026-03-06T13:44:06.293 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : lua-devel-5.4.4-4.el9.x86_64 96/152 2026-03-06T13:44:06.293 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : protobuf-compiler-3.14.0-17.el9.x86_64 97/152 2026-03-06T13:44:06.293 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : abseil-cpp-20211102.0-4.el9.x86_64 98/152 2026-03-06T13:44:06.293 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : gperftools-libs-2.9.1-3.el9.x86_64 99/152 2026-03-06T13:44:06.293 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : grpc-data-1.46.7-10.el9.noarch 100/152 2026-03-06T13:44:06.293 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : libarrow-9.0.0-15.el9.x86_64 101/152 2026-03-06T13:44:06.293 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : libarrow-doc-9.0.0-15.el9.noarch 102/152 2026-03-06T13:44:06.293 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : liboath-2.6.12-1.el9.x86_64 103/152 2026-03-06T13:44:06.293 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : libunwind-1.6.2-1.el9.x86_64 104/152 2026-03-06T13:44:06.293 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : luarocks-3.9.2-5.el9.noarch 105/152 2026-03-06T13:44:06.293 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : parquet-libs-9.0.0-15.el9.x86_64 106/152 2026-03-06T13:44:06.293 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : python3-asyncssh-2.13.2-5.el9.noarch 107/152 2026-03-06T13:44:06.293 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : python3-autocommand-2.2.2-8.el9.noarch 108/152 2026-03-06T13:44:06.294 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : python3-backports-tarfile-1.2.0-1.el9.noarch 109/152 2026-03-06T13:44:06.303 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : python3-bcrypt-3.2.2-1.el9.x86_64 110/152 2026-03-06T13:44:06.303 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : python3-cachetools-4.2.4-1.el9.noarch 111/152 2026-03-06T13:44:06.303 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : python3-certifi-2023.05.07-4.el9.noarch 112/152 2026-03-06T13:44:06.303 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : python3-cheroot-10.0.1-4.el9.noarch 113/152 2026-03-06T13:44:06.303 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : python3-cherrypy-18.6.1-2.el9.noarch 114/152 2026-03-06T13:44:06.303 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : python3-google-auth-1:2.45.0-1.el9.noarch 115/152 2026-03-06T13:44:06.303 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : python3-grpcio-1.46.7-10.el9.x86_64 116/152 2026-03-06T13:44:06.303 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : python3-grpcio-tools-1.46.7-10.el9.x86_64 117/152 2026-03-06T13:44:06.303 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : python3-influxdb-5.3.1-1.el9.noarch 118/152 2026-03-06T13:44:06.303 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : python3-isodate-0.6.1-3.el9.noarch 119/152 2026-03-06T13:44:06.303 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : python3-jaraco-8.2.1-3.el9.noarch 120/152 2026-03-06T13:44:06.303 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : python3-jaraco-classes-3.2.1-5.el9.noarch 121/152 2026-03-06T13:44:06.303 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : python3-jaraco-collections-3.0.0-8.el9.noarch 122/152 2026-03-06T13:44:06.303 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : python3-jaraco-context-6.0.1-3.el9.noarch 123/152 2026-03-06T13:44:06.303 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : python3-jaraco-functools-3.5.0-2.el9.noarch 124/152 2026-03-06T13:44:06.303 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : python3-jaraco-text-4.0.0-2.el9.noarch 125/152 2026-03-06T13:44:06.303 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : python3-kubernetes-1:26.1.0-3.el9.noarch 126/152 2026-03-06T13:44:06.303 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : python3-logutils-0.3.5-21.el9.noarch 127/152 2026-03-06T13:44:06.303 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : python3-more-itertools-8.12.0-2.el9.noarch 128/152 2026-03-06T13:44:06.303 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : python3-msgpack-1.0.3-2.el9.x86_64 129/152 2026-03-06T13:44:06.303 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : python3-natsort-7.1.1-5.el9.noarch 130/152 2026-03-06T13:44:06.303 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : python3-pecan-1.4.2-3.el9.noarch 131/152 2026-03-06T13:44:06.303 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : python3-portend-3.1.0-2.el9.noarch 132/152 2026-03-06T13:44:06.303 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : python3-pyOpenSSL-21.0.0-1.el9.noarch 133/152 2026-03-06T13:44:06.303 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : python3-repoze-lru-0.7-16.el9.noarch 134/152 2026-03-06T13:44:06.303 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : python3-routes-2.5.1-5.el9.noarch 135/152 2026-03-06T13:44:06.303 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : python3-rsa-4.9-2.el9.noarch 136/152 2026-03-06T13:44:06.303 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : python3-saml-1.16.0-1.el9.noarch 137/152 2026-03-06T13:44:06.303 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : python3-tempora-5.0.0-2.el9.noarch 138/152 2026-03-06T13:44:06.303 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : python3-typing-extensions-4.15.0-1.el9.noarch 139/152 2026-03-06T13:44:06.303 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : python3-webob-1.8.8-2.el9.noarch 140/152 2026-03-06T13:44:06.303 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : python3-websocket-client-1.2.3-2.el9.noarch 141/152 2026-03-06T13:44:06.303 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : python3-werkzeug-2.0.3-3.el9.1.noarch 142/152 2026-03-06T13:44:06.303 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : python3-xmlsec-1.3.13-1.el9.x86_64 143/152 2026-03-06T13:44:06.303 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : python3-xmltodict-0.12.0-15.el9.noarch 144/152 2026-03-06T13:44:06.304 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : python3-zc-lockfile-2.0-10.el9.noarch 145/152 2026-03-06T13:44:06.304 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : re2-1:20211101-20.el9.x86_64 146/152 2026-03-06T13:44:06.304 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : s3cmd-2.4.0-1.el9.noarch 147/152 2026-03-06T13:44:06.304 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : thrift-0.15.0-4.el9.x86_64 148/152 2026-03-06T13:44:06.304 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : librados2-2:19.2.3-47.gc24117fd552.el9.clyso.x86 149/152 2026-03-06T13:44:06.304 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : librados2-2:16.2.4-5.el9.x86_64 150/152 2026-03-06T13:44:06.304 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : librbd1-2:19.2.3-47.gc24117fd552.el9.clyso.x86_6 151/152 2026-03-06T13:44:07.174 INFO:teuthology.orchestra.run.vm04.stdout: Running scriptlet: librados2-2:16.2.4-5.el9.x86_64 152/152 2026-03-06T13:44:07.174 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : ceph-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 1/152 2026-03-06T13:44:07.174 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : ceph-base-2:19.2.3-47.gc24117fd552.el9.clyso.x86 2/152 2026-03-06T13:44:07.174 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : ceph-common-2:19.2.3-47.gc24117fd552.el9.clyso.x 3/152 2026-03-06T13:44:07.174 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : ceph-fuse-2:19.2.3-47.gc24117fd552.el9.clyso.x86 4/152 2026-03-06T13:44:07.174 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : ceph-immutable-object-cache-2:19.2.3-47.gc24117f 5/152 2026-03-06T13:44:07.174 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : ceph-mds-2:19.2.3-47.gc24117fd552.el9.clyso.x86_ 6/152 2026-03-06T13:44:07.174 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : ceph-mgr-2:19.2.3-47.gc24117fd552.el9.clyso.x86_ 7/152 2026-03-06T13:44:07.174 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : ceph-mon-2:19.2.3-47.gc24117fd552.el9.clyso.x86_ 8/152 2026-03-06T13:44:07.174 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : ceph-osd-2:19.2.3-47.gc24117fd552.el9.clyso.x86_ 9/152 2026-03-06T13:44:07.174 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : ceph-radosgw-2:19.2.3-47.gc24117fd552.el9.clyso. 10/152 2026-03-06T13:44:07.174 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : ceph-selinux-2:19.2.3-47.gc24117fd552.el9.clyso. 11/152 2026-03-06T13:44:07.174 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : ceph-test-2:19.2.3-47.gc24117fd552.el9.clyso.x86 12/152 2026-03-06T13:44:07.174 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : libcephfs-devel-2:19.2.3-47.gc24117fd552.el9.cly 13/152 2026-03-06T13:44:07.174 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : libcephfs2-2:19.2.3-47.gc24117fd552.el9.clyso.x8 14/152 2026-03-06T13:44:07.174 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : libcephsqlite-2:19.2.3-47.gc24117fd552.el9.clyso 15/152 2026-03-06T13:44:07.174 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : librados-devel-2:19.2.3-47.gc24117fd552.el9.clys 16/152 2026-03-06T13:44:07.174 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : libradosstriper1-2:19.2.3-47.gc24117fd552.el9.cl 17/152 2026-03-06T13:44:07.174 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : librgw2-2:19.2.3-47.gc24117fd552.el9.clyso.x86_6 18/152 2026-03-06T13:44:07.175 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-ceph-argparse-2:19.2.3-47.gc24117fd552.e 19/152 2026-03-06T13:44:07.175 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-ceph-common-2:19.2.3-47.gc24117fd552.el9 20/152 2026-03-06T13:44:07.175 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-cephfs-2:19.2.3-47.gc24117fd552.el9.clys 21/152 2026-03-06T13:44:07.175 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-rados-2:19.2.3-47.gc24117fd552.el9.clyso 22/152 2026-03-06T13:44:07.175 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-rbd-2:19.2.3-47.gc24117fd552.el9.clyso.x 23/152 2026-03-06T13:44:07.175 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-rgw-2:19.2.3-47.gc24117fd552.el9.clyso.x 24/152 2026-03-06T13:44:07.175 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : rbd-fuse-2:19.2.3-47.gc24117fd552.el9.clyso.x86_ 25/152 2026-03-06T13:44:07.175 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : rbd-mirror-2:19.2.3-47.gc24117fd552.el9.clyso.x8 26/152 2026-03-06T13:44:07.175 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : rbd-nbd-2:19.2.3-47.gc24117fd552.el9.clyso.x86_6 27/152 2026-03-06T13:44:07.175 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : ceph-grafana-dashboards-2:19.2.3-47.gc24117fd552 28/152 2026-03-06T13:44:07.175 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : ceph-mgr-cephadm-2:19.2.3-47.gc24117fd552.el9.cl 29/152 2026-03-06T13:44:07.177 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : ceph-mgr-dashboard-2:19.2.3-47.gc24117fd552.el9. 30/152 2026-03-06T13:44:07.177 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : ceph-mgr-diskprediction-local-2:19.2.3-47.gc2411 31/152 2026-03-06T13:44:07.177 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : ceph-mgr-k8sevents-2:19.2.3-47.gc24117fd552.el9. 32/152 2026-03-06T13:44:07.177 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : ceph-mgr-modules-core-2:19.2.3-47.gc24117fd552.e 33/152 2026-03-06T13:44:07.177 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : ceph-mgr-rook-2:19.2.3-47.gc24117fd552.el9.clyso 34/152 2026-03-06T13:44:07.177 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : ceph-prometheus-alerts-2:19.2.3-47.gc24117fd552. 35/152 2026-03-06T13:44:07.177 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : ceph-volume-2:19.2.3-47.gc24117fd552.el9.clyso.n 36/152 2026-03-06T13:44:07.177 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : cephadm-2:19.2.3-47.gc24117fd552.el9.clyso.noarc 37/152 2026-03-06T13:44:07.177 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : bzip2-1.0.8-11.el9.x86_64 38/152 2026-03-06T13:44:07.177 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : cryptsetup-2.8.1-3.el9.x86_64 39/152 2026-03-06T13:44:07.177 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : fuse-2.9.9-17.el9.x86_64 40/152 2026-03-06T13:44:07.177 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : ledmon-libs-1.1.0-3.el9.x86_64 41/152 2026-03-06T13:44:07.177 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : libconfig-1.7.2-9.el9.x86_64 42/152 2026-03-06T13:44:07.177 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : libgfortran-11.5.0-14.el9.x86_64 43/152 2026-03-06T13:44:07.177 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : libquadmath-11.5.0-14.el9.x86_64 44/152 2026-03-06T13:44:07.177 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : mailcap-2.1.49-5.el9.noarch 45/152 2026-03-06T13:44:07.177 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : pciutils-3.7.0-7.el9.x86_64 46/152 2026-03-06T13:44:07.177 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-cffi-1.14.5-5.el9.x86_64 47/152 2026-03-06T13:44:07.177 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-cryptography-36.0.1-5.el9.x86_64 48/152 2026-03-06T13:44:07.177 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-ply-3.11-14.el9.noarch 49/152 2026-03-06T13:44:07.177 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-pycparser-2.20-6.el9.noarch 50/152 2026-03-06T13:44:07.177 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-requests-2.25.1-10.el9.noarch 51/152 2026-03-06T13:44:07.177 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-urllib3-1.26.5-7.el9.noarch 52/152 2026-03-06T13:44:07.177 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : smartmontools-1:7.2-10.el9.x86_64 53/152 2026-03-06T13:44:07.177 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : unzip-6.0-59.el9.x86_64 54/152 2026-03-06T13:44:07.177 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : zip-3.0-35.el9.x86_64 55/152 2026-03-06T13:44:07.177 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : boost-program-options-1.75.0-13.el9.x86_64 56/152 2026-03-06T13:44:07.177 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : flexiblas-3.0.4-9.el9.x86_64 57/152 2026-03-06T13:44:07.177 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : flexiblas-netlib-3.0.4-9.el9.x86_64 58/152 2026-03-06T13:44:07.177 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : flexiblas-openblas-openmp-3.0.4-9.el9.x86_64 59/152 2026-03-06T13:44:07.177 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : libnbd-1.20.3-4.el9.x86_64 60/152 2026-03-06T13:44:07.177 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : libpmemobj-1.12.1-1.el9.x86_64 61/152 2026-03-06T13:44:07.178 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : librabbitmq-0.11.0-7.el9.x86_64 62/152 2026-03-06T13:44:07.178 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : librdkafka-1.6.1-102.el9.x86_64 63/152 2026-03-06T13:44:07.178 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : libstoragemgmt-1.10.1-1.el9.x86_64 64/152 2026-03-06T13:44:07.178 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : libxslt-1.1.34-12.el9.x86_64 65/152 2026-03-06T13:44:07.178 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : lttng-ust-2.12.0-6.el9.x86_64 66/152 2026-03-06T13:44:07.178 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : lua-5.4.4-4.el9.x86_64 67/152 2026-03-06T13:44:07.178 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : openblas-0.3.29-1.el9.x86_64 68/152 2026-03-06T13:44:07.178 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : openblas-openmp-0.3.29-1.el9.x86_64 69/152 2026-03-06T13:44:07.178 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : perl-Benchmark-1.23-483.el9.noarch 70/152 2026-03-06T13:44:07.178 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : perl-Test-Harness-1:3.42-461.el9.noarch 71/152 2026-03-06T13:44:07.178 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : protobuf-3.14.0-17.el9.x86_64 72/152 2026-03-06T13:44:07.178 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-babel-2.9.1-2.el9.noarch 73/152 2026-03-06T13:44:07.178 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-devel-3.9.25-3.el9.x86_64 74/152 2026-03-06T13:44:07.178 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-jinja2-2.11.3-8.el9.noarch 75/152 2026-03-06T13:44:07.178 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-libstoragemgmt-1.10.1-1.el9.x86_64 76/152 2026-03-06T13:44:07.178 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-lxml-4.6.5-3.el9.x86_64 77/152 2026-03-06T13:44:07.178 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-mako-1.1.4-6.el9.noarch 78/152 2026-03-06T13:44:07.178 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-markupsafe-1.1.1-12.el9.x86_64 79/152 2026-03-06T13:44:07.178 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-numpy-1:1.23.5-2.el9.x86_64 80/152 2026-03-06T13:44:07.178 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-numpy-f2py-1:1.23.5-2.el9.x86_64 81/152 2026-03-06T13:44:07.178 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-packaging-20.9-5.el9.noarch 82/152 2026-03-06T13:44:07.178 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-protobuf-3.14.0-17.el9.noarch 83/152 2026-03-06T13:44:07.178 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-pyasn1-0.4.8-7.el9.noarch 84/152 2026-03-06T13:44:07.178 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-pyasn1-modules-0.4.8-7.el9.noarch 85/152 2026-03-06T13:44:07.178 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-requests-oauthlib-1.3.0-12.el9.noarch 86/152 2026-03-06T13:44:07.178 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-scipy-1.9.3-2.el9.x86_64 87/152 2026-03-06T13:44:07.178 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-toml-0.10.2-6.el9.noarch 88/152 2026-03-06T13:44:07.178 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : qatlib-25.08.0-2.el9.x86_64 89/152 2026-03-06T13:44:07.178 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : qatlib-service-25.08.0-2.el9.x86_64 90/152 2026-03-06T13:44:07.178 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : qatzip-libs-1.3.1-1.el9.x86_64 91/152 2026-03-06T13:44:07.178 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : socat-1.7.4.1-8.el9.x86_64 92/152 2026-03-06T13:44:07.178 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : xmlsec1-1.2.29-13.el9.x86_64 93/152 2026-03-06T13:44:07.178 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : xmlsec1-openssl-1.2.29-13.el9.x86_64 94/152 2026-03-06T13:44:07.178 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : xmlstarlet-1.6.1-20.el9.x86_64 95/152 2026-03-06T13:44:07.178 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : lua-devel-5.4.4-4.el9.x86_64 96/152 2026-03-06T13:44:07.178 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : protobuf-compiler-3.14.0-17.el9.x86_64 97/152 2026-03-06T13:44:07.178 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : abseil-cpp-20211102.0-4.el9.x86_64 98/152 2026-03-06T13:44:07.178 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : gperftools-libs-2.9.1-3.el9.x86_64 99/152 2026-03-06T13:44:07.178 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : grpc-data-1.46.7-10.el9.noarch 100/152 2026-03-06T13:44:07.178 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : libarrow-9.0.0-15.el9.x86_64 101/152 2026-03-06T13:44:07.178 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : libarrow-doc-9.0.0-15.el9.noarch 102/152 2026-03-06T13:44:07.178 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : liboath-2.6.12-1.el9.x86_64 103/152 2026-03-06T13:44:07.178 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : libunwind-1.6.2-1.el9.x86_64 104/152 2026-03-06T13:44:07.178 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : luarocks-3.9.2-5.el9.noarch 105/152 2026-03-06T13:44:07.178 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : parquet-libs-9.0.0-15.el9.x86_64 106/152 2026-03-06T13:44:07.178 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-asyncssh-2.13.2-5.el9.noarch 107/152 2026-03-06T13:44:07.178 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-autocommand-2.2.2-8.el9.noarch 108/152 2026-03-06T13:44:07.178 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-backports-tarfile-1.2.0-1.el9.noarch 109/152 2026-03-06T13:44:07.178 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-bcrypt-3.2.2-1.el9.x86_64 110/152 2026-03-06T13:44:07.178 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-cachetools-4.2.4-1.el9.noarch 111/152 2026-03-06T13:44:07.178 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-certifi-2023.05.07-4.el9.noarch 112/152 2026-03-06T13:44:07.178 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-cheroot-10.0.1-4.el9.noarch 113/152 2026-03-06T13:44:07.178 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-cherrypy-18.6.1-2.el9.noarch 114/152 2026-03-06T13:44:07.178 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-google-auth-1:2.45.0-1.el9.noarch 115/152 2026-03-06T13:44:07.178 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-grpcio-1.46.7-10.el9.x86_64 116/152 2026-03-06T13:44:07.178 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-grpcio-tools-1.46.7-10.el9.x86_64 117/152 2026-03-06T13:44:07.178 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-influxdb-5.3.1-1.el9.noarch 118/152 2026-03-06T13:44:07.178 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-isodate-0.6.1-3.el9.noarch 119/152 2026-03-06T13:44:07.178 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-jaraco-8.2.1-3.el9.noarch 120/152 2026-03-06T13:44:07.178 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-jaraco-classes-3.2.1-5.el9.noarch 121/152 2026-03-06T13:44:07.178 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-jaraco-collections-3.0.0-8.el9.noarch 122/152 2026-03-06T13:44:07.178 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-jaraco-context-6.0.1-3.el9.noarch 123/152 2026-03-06T13:44:07.178 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-jaraco-functools-3.5.0-2.el9.noarch 124/152 2026-03-06T13:44:07.178 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-jaraco-text-4.0.0-2.el9.noarch 125/152 2026-03-06T13:44:07.178 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-kubernetes-1:26.1.0-3.el9.noarch 126/152 2026-03-06T13:44:07.179 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-logutils-0.3.5-21.el9.noarch 127/152 2026-03-06T13:44:07.179 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-more-itertools-8.12.0-2.el9.noarch 128/152 2026-03-06T13:44:07.179 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-msgpack-1.0.3-2.el9.x86_64 129/152 2026-03-06T13:44:07.179 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-natsort-7.1.1-5.el9.noarch 130/152 2026-03-06T13:44:07.179 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-pecan-1.4.2-3.el9.noarch 131/152 2026-03-06T13:44:07.179 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-portend-3.1.0-2.el9.noarch 132/152 2026-03-06T13:44:07.179 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-pyOpenSSL-21.0.0-1.el9.noarch 133/152 2026-03-06T13:44:07.179 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-repoze-lru-0.7-16.el9.noarch 134/152 2026-03-06T13:44:07.179 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-routes-2.5.1-5.el9.noarch 135/152 2026-03-06T13:44:07.179 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-rsa-4.9-2.el9.noarch 136/152 2026-03-06T13:44:07.179 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-saml-1.16.0-1.el9.noarch 137/152 2026-03-06T13:44:07.179 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-tempora-5.0.0-2.el9.noarch 138/152 2026-03-06T13:44:07.179 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-typing-extensions-4.15.0-1.el9.noarch 139/152 2026-03-06T13:44:07.179 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-webob-1.8.8-2.el9.noarch 140/152 2026-03-06T13:44:07.179 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-websocket-client-1.2.3-2.el9.noarch 141/152 2026-03-06T13:44:07.179 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-werkzeug-2.0.3-3.el9.1.noarch 142/152 2026-03-06T13:44:07.179 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-xmlsec-1.3.13-1.el9.x86_64 143/152 2026-03-06T13:44:07.179 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-xmltodict-0.12.0-15.el9.noarch 144/152 2026-03-06T13:44:07.179 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-zc-lockfile-2.0-10.el9.noarch 145/152 2026-03-06T13:44:07.179 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : re2-1:20211101-20.el9.x86_64 146/152 2026-03-06T13:44:07.179 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : s3cmd-2.4.0-1.el9.noarch 147/152 2026-03-06T13:44:07.179 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : thrift-0.15.0-4.el9.x86_64 148/152 2026-03-06T13:44:07.179 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : librados2-2:19.2.3-47.gc24117fd552.el9.clyso.x86 149/152 2026-03-06T13:44:07.179 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : librados2-2:16.2.4-5.el9.x86_64 150/152 2026-03-06T13:44:07.179 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : librbd1-2:19.2.3-47.gc24117fd552.el9.clyso.x86_6 151/152 2026-03-06T13:44:07.256 INFO:teuthology.orchestra.run.vm06.stdout: Running scriptlet: librados2-2:16.2.4-5.el9.x86_64 152/152 2026-03-06T13:44:07.256 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : ceph-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 1/152 2026-03-06T13:44:07.256 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : ceph-base-2:19.2.3-47.gc24117fd552.el9.clyso.x86 2/152 2026-03-06T13:44:07.256 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : ceph-common-2:19.2.3-47.gc24117fd552.el9.clyso.x 3/152 2026-03-06T13:44:07.256 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : ceph-fuse-2:19.2.3-47.gc24117fd552.el9.clyso.x86 4/152 2026-03-06T13:44:07.256 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : ceph-immutable-object-cache-2:19.2.3-47.gc24117f 5/152 2026-03-06T13:44:07.256 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : ceph-mds-2:19.2.3-47.gc24117fd552.el9.clyso.x86_ 6/152 2026-03-06T13:44:07.256 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : ceph-mgr-2:19.2.3-47.gc24117fd552.el9.clyso.x86_ 7/152 2026-03-06T13:44:07.256 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : ceph-mon-2:19.2.3-47.gc24117fd552.el9.clyso.x86_ 8/152 2026-03-06T13:44:07.256 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : ceph-osd-2:19.2.3-47.gc24117fd552.el9.clyso.x86_ 9/152 2026-03-06T13:44:07.256 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : ceph-radosgw-2:19.2.3-47.gc24117fd552.el9.clyso. 10/152 2026-03-06T13:44:07.256 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : ceph-selinux-2:19.2.3-47.gc24117fd552.el9.clyso. 11/152 2026-03-06T13:44:07.256 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : ceph-test-2:19.2.3-47.gc24117fd552.el9.clyso.x86 12/152 2026-03-06T13:44:07.256 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : libcephfs-devel-2:19.2.3-47.gc24117fd552.el9.cly 13/152 2026-03-06T13:44:07.256 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : libcephfs2-2:19.2.3-47.gc24117fd552.el9.clyso.x8 14/152 2026-03-06T13:44:07.256 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : libcephsqlite-2:19.2.3-47.gc24117fd552.el9.clyso 15/152 2026-03-06T13:44:07.256 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : librados-devel-2:19.2.3-47.gc24117fd552.el9.clys 16/152 2026-03-06T13:44:07.256 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : libradosstriper1-2:19.2.3-47.gc24117fd552.el9.cl 17/152 2026-03-06T13:44:07.256 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : librgw2-2:19.2.3-47.gc24117fd552.el9.clyso.x86_6 18/152 2026-03-06T13:44:07.257 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-ceph-argparse-2:19.2.3-47.gc24117fd552.e 19/152 2026-03-06T13:44:07.257 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-ceph-common-2:19.2.3-47.gc24117fd552.el9 20/152 2026-03-06T13:44:07.257 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-cephfs-2:19.2.3-47.gc24117fd552.el9.clys 21/152 2026-03-06T13:44:07.257 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-rados-2:19.2.3-47.gc24117fd552.el9.clyso 22/152 2026-03-06T13:44:07.257 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-rbd-2:19.2.3-47.gc24117fd552.el9.clyso.x 23/152 2026-03-06T13:44:07.257 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-rgw-2:19.2.3-47.gc24117fd552.el9.clyso.x 24/152 2026-03-06T13:44:07.257 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : rbd-fuse-2:19.2.3-47.gc24117fd552.el9.clyso.x86_ 25/152 2026-03-06T13:44:07.257 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : rbd-mirror-2:19.2.3-47.gc24117fd552.el9.clyso.x8 26/152 2026-03-06T13:44:07.257 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : rbd-nbd-2:19.2.3-47.gc24117fd552.el9.clyso.x86_6 27/152 2026-03-06T13:44:07.257 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : ceph-grafana-dashboards-2:19.2.3-47.gc24117fd552 28/152 2026-03-06T13:44:07.257 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : ceph-mgr-cephadm-2:19.2.3-47.gc24117fd552.el9.cl 29/152 2026-03-06T13:44:07.259 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : ceph-mgr-dashboard-2:19.2.3-47.gc24117fd552.el9. 30/152 2026-03-06T13:44:07.259 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : ceph-mgr-diskprediction-local-2:19.2.3-47.gc2411 31/152 2026-03-06T13:44:07.259 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : ceph-mgr-k8sevents-2:19.2.3-47.gc24117fd552.el9. 32/152 2026-03-06T13:44:07.259 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : ceph-mgr-modules-core-2:19.2.3-47.gc24117fd552.e 33/152 2026-03-06T13:44:07.259 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : ceph-mgr-rook-2:19.2.3-47.gc24117fd552.el9.clyso 34/152 2026-03-06T13:44:07.259 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : ceph-prometheus-alerts-2:19.2.3-47.gc24117fd552. 35/152 2026-03-06T13:44:07.259 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : ceph-volume-2:19.2.3-47.gc24117fd552.el9.clyso.n 36/152 2026-03-06T13:44:07.259 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : cephadm-2:19.2.3-47.gc24117fd552.el9.clyso.noarc 37/152 2026-03-06T13:44:07.259 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : bzip2-1.0.8-11.el9.x86_64 38/152 2026-03-06T13:44:07.259 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : cryptsetup-2.8.1-3.el9.x86_64 39/152 2026-03-06T13:44:07.260 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : fuse-2.9.9-17.el9.x86_64 40/152 2026-03-06T13:44:07.260 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : ledmon-libs-1.1.0-3.el9.x86_64 41/152 2026-03-06T13:44:07.260 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : libconfig-1.7.2-9.el9.x86_64 42/152 2026-03-06T13:44:07.260 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : libgfortran-11.5.0-14.el9.x86_64 43/152 2026-03-06T13:44:07.260 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : libquadmath-11.5.0-14.el9.x86_64 44/152 2026-03-06T13:44:07.260 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : mailcap-2.1.49-5.el9.noarch 45/152 2026-03-06T13:44:07.260 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : pciutils-3.7.0-7.el9.x86_64 46/152 2026-03-06T13:44:07.260 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-cffi-1.14.5-5.el9.x86_64 47/152 2026-03-06T13:44:07.260 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-cryptography-36.0.1-5.el9.x86_64 48/152 2026-03-06T13:44:07.260 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-ply-3.11-14.el9.noarch 49/152 2026-03-06T13:44:07.260 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-pycparser-2.20-6.el9.noarch 50/152 2026-03-06T13:44:07.260 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-requests-2.25.1-10.el9.noarch 51/152 2026-03-06T13:44:07.260 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-urllib3-1.26.5-7.el9.noarch 52/152 2026-03-06T13:44:07.260 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : smartmontools-1:7.2-10.el9.x86_64 53/152 2026-03-06T13:44:07.260 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : unzip-6.0-59.el9.x86_64 54/152 2026-03-06T13:44:07.260 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : zip-3.0-35.el9.x86_64 55/152 2026-03-06T13:44:07.260 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : boost-program-options-1.75.0-13.el9.x86_64 56/152 2026-03-06T13:44:07.260 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : flexiblas-3.0.4-9.el9.x86_64 57/152 2026-03-06T13:44:07.260 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : flexiblas-netlib-3.0.4-9.el9.x86_64 58/152 2026-03-06T13:44:07.261 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : flexiblas-openblas-openmp-3.0.4-9.el9.x86_64 59/152 2026-03-06T13:44:07.261 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : libnbd-1.20.3-4.el9.x86_64 60/152 2026-03-06T13:44:07.261 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : libpmemobj-1.12.1-1.el9.x86_64 61/152 2026-03-06T13:44:07.261 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : librabbitmq-0.11.0-7.el9.x86_64 62/152 2026-03-06T13:44:07.261 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : librdkafka-1.6.1-102.el9.x86_64 63/152 2026-03-06T13:44:07.261 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : libstoragemgmt-1.10.1-1.el9.x86_64 64/152 2026-03-06T13:44:07.261 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : libxslt-1.1.34-12.el9.x86_64 65/152 2026-03-06T13:44:07.261 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : lttng-ust-2.12.0-6.el9.x86_64 66/152 2026-03-06T13:44:07.261 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : lua-5.4.4-4.el9.x86_64 67/152 2026-03-06T13:44:07.261 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : openblas-0.3.29-1.el9.x86_64 68/152 2026-03-06T13:44:07.261 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : openblas-openmp-0.3.29-1.el9.x86_64 69/152 2026-03-06T13:44:07.262 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : perl-Benchmark-1.23-483.el9.noarch 70/152 2026-03-06T13:44:07.262 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : perl-Test-Harness-1:3.42-461.el9.noarch 71/152 2026-03-06T13:44:07.262 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : protobuf-3.14.0-17.el9.x86_64 72/152 2026-03-06T13:44:07.262 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-babel-2.9.1-2.el9.noarch 73/152 2026-03-06T13:44:07.262 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-devel-3.9.25-3.el9.x86_64 74/152 2026-03-06T13:44:07.262 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-jinja2-2.11.3-8.el9.noarch 75/152 2026-03-06T13:44:07.262 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-libstoragemgmt-1.10.1-1.el9.x86_64 76/152 2026-03-06T13:44:07.262 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-lxml-4.6.5-3.el9.x86_64 77/152 2026-03-06T13:44:07.262 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-mako-1.1.4-6.el9.noarch 78/152 2026-03-06T13:44:07.262 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-markupsafe-1.1.1-12.el9.x86_64 79/152 2026-03-06T13:44:07.262 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-numpy-1:1.23.5-2.el9.x86_64 80/152 2026-03-06T13:44:07.262 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-numpy-f2py-1:1.23.5-2.el9.x86_64 81/152 2026-03-06T13:44:07.262 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-packaging-20.9-5.el9.noarch 82/152 2026-03-06T13:44:07.262 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-protobuf-3.14.0-17.el9.noarch 83/152 2026-03-06T13:44:07.262 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-pyasn1-0.4.8-7.el9.noarch 84/152 2026-03-06T13:44:07.262 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-pyasn1-modules-0.4.8-7.el9.noarch 85/152 2026-03-06T13:44:07.262 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-requests-oauthlib-1.3.0-12.el9.noarch 86/152 2026-03-06T13:44:07.263 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-scipy-1.9.3-2.el9.x86_64 87/152 2026-03-06T13:44:07.263 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-toml-0.10.2-6.el9.noarch 88/152 2026-03-06T13:44:07.263 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : qatlib-25.08.0-2.el9.x86_64 89/152 2026-03-06T13:44:07.263 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : qatlib-service-25.08.0-2.el9.x86_64 90/152 2026-03-06T13:44:07.263 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : qatzip-libs-1.3.1-1.el9.x86_64 91/152 2026-03-06T13:44:07.263 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : socat-1.7.4.1-8.el9.x86_64 92/152 2026-03-06T13:44:07.263 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : xmlsec1-1.2.29-13.el9.x86_64 93/152 2026-03-06T13:44:07.263 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : xmlsec1-openssl-1.2.29-13.el9.x86_64 94/152 2026-03-06T13:44:07.263 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : xmlstarlet-1.6.1-20.el9.x86_64 95/152 2026-03-06T13:44:07.263 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : lua-devel-5.4.4-4.el9.x86_64 96/152 2026-03-06T13:44:07.263 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : protobuf-compiler-3.14.0-17.el9.x86_64 97/152 2026-03-06T13:44:07.263 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : abseil-cpp-20211102.0-4.el9.x86_64 98/152 2026-03-06T13:44:07.263 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : gperftools-libs-2.9.1-3.el9.x86_64 99/152 2026-03-06T13:44:07.263 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : grpc-data-1.46.7-10.el9.noarch 100/152 2026-03-06T13:44:07.263 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : libarrow-9.0.0-15.el9.x86_64 101/152 2026-03-06T13:44:07.263 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : libarrow-doc-9.0.0-15.el9.noarch 102/152 2026-03-06T13:44:07.263 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : liboath-2.6.12-1.el9.x86_64 103/152 2026-03-06T13:44:07.263 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : libunwind-1.6.2-1.el9.x86_64 104/152 2026-03-06T13:44:07.263 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : luarocks-3.9.2-5.el9.noarch 105/152 2026-03-06T13:44:07.263 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : parquet-libs-9.0.0-15.el9.x86_64 106/152 2026-03-06T13:44:07.263 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-asyncssh-2.13.2-5.el9.noarch 107/152 2026-03-06T13:44:07.263 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-autocommand-2.2.2-8.el9.noarch 108/152 2026-03-06T13:44:07.264 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-backports-tarfile-1.2.0-1.el9.noarch 109/152 2026-03-06T13:44:07.264 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-bcrypt-3.2.2-1.el9.x86_64 110/152 2026-03-06T13:44:07.264 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-cachetools-4.2.4-1.el9.noarch 111/152 2026-03-06T13:44:07.264 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-certifi-2023.05.07-4.el9.noarch 112/152 2026-03-06T13:44:07.264 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-cheroot-10.0.1-4.el9.noarch 113/152 2026-03-06T13:44:07.264 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-cherrypy-18.6.1-2.el9.noarch 114/152 2026-03-06T13:44:07.264 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-google-auth-1:2.45.0-1.el9.noarch 115/152 2026-03-06T13:44:07.264 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-grpcio-1.46.7-10.el9.x86_64 116/152 2026-03-06T13:44:07.264 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-grpcio-tools-1.46.7-10.el9.x86_64 117/152 2026-03-06T13:44:07.264 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-influxdb-5.3.1-1.el9.noarch 118/152 2026-03-06T13:44:07.264 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-isodate-0.6.1-3.el9.noarch 119/152 2026-03-06T13:44:07.264 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-jaraco-8.2.1-3.el9.noarch 120/152 2026-03-06T13:44:07.264 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-jaraco-classes-3.2.1-5.el9.noarch 121/152 2026-03-06T13:44:07.264 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-jaraco-collections-3.0.0-8.el9.noarch 122/152 2026-03-06T13:44:07.264 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-jaraco-context-6.0.1-3.el9.noarch 123/152 2026-03-06T13:44:07.264 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-jaraco-functools-3.5.0-2.el9.noarch 124/152 2026-03-06T13:44:07.264 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-jaraco-text-4.0.0-2.el9.noarch 125/152 2026-03-06T13:44:07.264 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-kubernetes-1:26.1.0-3.el9.noarch 126/152 2026-03-06T13:44:07.264 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-logutils-0.3.5-21.el9.noarch 127/152 2026-03-06T13:44:07.264 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-more-itertools-8.12.0-2.el9.noarch 128/152 2026-03-06T13:44:07.264 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-msgpack-1.0.3-2.el9.x86_64 129/152 2026-03-06T13:44:07.264 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-natsort-7.1.1-5.el9.noarch 130/152 2026-03-06T13:44:07.264 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-pecan-1.4.2-3.el9.noarch 131/152 2026-03-06T13:44:07.264 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-portend-3.1.0-2.el9.noarch 132/152 2026-03-06T13:44:07.264 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-pyOpenSSL-21.0.0-1.el9.noarch 133/152 2026-03-06T13:44:07.264 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-repoze-lru-0.7-16.el9.noarch 134/152 2026-03-06T13:44:07.264 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-routes-2.5.1-5.el9.noarch 135/152 2026-03-06T13:44:07.264 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-rsa-4.9-2.el9.noarch 136/152 2026-03-06T13:44:07.265 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-saml-1.16.0-1.el9.noarch 137/152 2026-03-06T13:44:07.265 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-tempora-5.0.0-2.el9.noarch 138/152 2026-03-06T13:44:07.265 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-typing-extensions-4.15.0-1.el9.noarch 139/152 2026-03-06T13:44:07.265 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-webob-1.8.8-2.el9.noarch 140/152 2026-03-06T13:44:07.265 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-websocket-client-1.2.3-2.el9.noarch 141/152 2026-03-06T13:44:07.265 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-werkzeug-2.0.3-3.el9.1.noarch 142/152 2026-03-06T13:44:07.265 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-xmlsec-1.3.13-1.el9.x86_64 143/152 2026-03-06T13:44:07.265 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-xmltodict-0.12.0-15.el9.noarch 144/152 2026-03-06T13:44:07.265 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-zc-lockfile-2.0-10.el9.noarch 145/152 2026-03-06T13:44:07.265 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : re2-1:20211101-20.el9.x86_64 146/152 2026-03-06T13:44:07.265 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : s3cmd-2.4.0-1.el9.noarch 147/152 2026-03-06T13:44:07.265 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : thrift-0.15.0-4.el9.x86_64 148/152 2026-03-06T13:44:07.265 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : librados2-2:19.2.3-47.gc24117fd552.el9.clyso.x86 149/152 2026-03-06T13:44:07.265 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : librados2-2:16.2.4-5.el9.x86_64 150/152 2026-03-06T13:44:07.265 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : librbd1-2:19.2.3-47.gc24117fd552.el9.clyso.x86_6 151/152 2026-03-06T13:44:07.274 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : librbd1-2:16.2.4-5.el9.x86_64 152/152 2026-03-06T13:44:07.274 INFO:teuthology.orchestra.run.vm02.stdout: 2026-03-06T13:44:07.274 INFO:teuthology.orchestra.run.vm02.stdout:Upgraded: 2026-03-06T13:44:07.274 INFO:teuthology.orchestra.run.vm02.stdout: librados2-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 2026-03-06T13:44:07.274 INFO:teuthology.orchestra.run.vm02.stdout: librbd1-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 2026-03-06T13:44:07.274 INFO:teuthology.orchestra.run.vm02.stdout:Installed: 2026-03-06T13:44:07.274 INFO:teuthology.orchestra.run.vm02.stdout: abseil-cpp-20211102.0-4.el9.x86_64 2026-03-06T13:44:07.274 INFO:teuthology.orchestra.run.vm02.stdout: boost-program-options-1.75.0-13.el9.x86_64 2026-03-06T13:44:07.275 INFO:teuthology.orchestra.run.vm02.stdout: bzip2-1.0.8-11.el9.x86_64 2026-03-06T13:44:07.275 INFO:teuthology.orchestra.run.vm02.stdout: ceph-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 2026-03-06T13:44:07.275 INFO:teuthology.orchestra.run.vm02.stdout: ceph-base-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 2026-03-06T13:44:07.275 INFO:teuthology.orchestra.run.vm02.stdout: ceph-common-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 2026-03-06T13:44:07.275 INFO:teuthology.orchestra.run.vm02.stdout: ceph-fuse-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 2026-03-06T13:44:07.275 INFO:teuthology.orchestra.run.vm02.stdout: ceph-grafana-dashboards-2:19.2.3-47.gc24117fd552.el9.clyso.noarch 2026-03-06T13:44:07.275 INFO:teuthology.orchestra.run.vm02.stdout: ceph-immutable-object-cache-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 2026-03-06T13:44:07.275 INFO:teuthology.orchestra.run.vm02.stdout: ceph-mds-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 2026-03-06T13:44:07.275 INFO:teuthology.orchestra.run.vm02.stdout: ceph-mgr-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 2026-03-06T13:44:07.275 INFO:teuthology.orchestra.run.vm02.stdout: ceph-mgr-cephadm-2:19.2.3-47.gc24117fd552.el9.clyso.noarch 2026-03-06T13:44:07.275 INFO:teuthology.orchestra.run.vm02.stdout: ceph-mgr-dashboard-2:19.2.3-47.gc24117fd552.el9.clyso.noarch 2026-03-06T13:44:07.275 INFO:teuthology.orchestra.run.vm02.stdout: ceph-mgr-diskprediction-local-2:19.2.3-47.gc24117fd552.el9.clyso.noarch 2026-03-06T13:44:07.275 INFO:teuthology.orchestra.run.vm02.stdout: ceph-mgr-k8sevents-2:19.2.3-47.gc24117fd552.el9.clyso.noarch 2026-03-06T13:44:07.275 INFO:teuthology.orchestra.run.vm02.stdout: ceph-mgr-modules-core-2:19.2.3-47.gc24117fd552.el9.clyso.noarch 2026-03-06T13:44:07.275 INFO:teuthology.orchestra.run.vm02.stdout: ceph-mgr-rook-2:19.2.3-47.gc24117fd552.el9.clyso.noarch 2026-03-06T13:44:07.275 INFO:teuthology.orchestra.run.vm02.stdout: ceph-mon-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 2026-03-06T13:44:07.275 INFO:teuthology.orchestra.run.vm02.stdout: ceph-osd-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 2026-03-06T13:44:07.275 INFO:teuthology.orchestra.run.vm02.stdout: ceph-prometheus-alerts-2:19.2.3-47.gc24117fd552.el9.clyso.noarch 2026-03-06T13:44:07.275 INFO:teuthology.orchestra.run.vm02.stdout: ceph-radosgw-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 2026-03-06T13:44:07.275 INFO:teuthology.orchestra.run.vm02.stdout: ceph-selinux-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 2026-03-06T13:44:07.275 INFO:teuthology.orchestra.run.vm02.stdout: ceph-test-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 2026-03-06T13:44:07.275 INFO:teuthology.orchestra.run.vm02.stdout: ceph-volume-2:19.2.3-47.gc24117fd552.el9.clyso.noarch 2026-03-06T13:44:07.275 INFO:teuthology.orchestra.run.vm02.stdout: cephadm-2:19.2.3-47.gc24117fd552.el9.clyso.noarch 2026-03-06T13:44:07.275 INFO:teuthology.orchestra.run.vm02.stdout: cryptsetup-2.8.1-3.el9.x86_64 2026-03-06T13:44:07.275 INFO:teuthology.orchestra.run.vm02.stdout: flexiblas-3.0.4-9.el9.x86_64 2026-03-06T13:44:07.275 INFO:teuthology.orchestra.run.vm02.stdout: flexiblas-netlib-3.0.4-9.el9.x86_64 2026-03-06T13:44:07.275 INFO:teuthology.orchestra.run.vm02.stdout: flexiblas-openblas-openmp-3.0.4-9.el9.x86_64 2026-03-06T13:44:07.275 INFO:teuthology.orchestra.run.vm02.stdout: fuse-2.9.9-17.el9.x86_64 2026-03-06T13:44:07.275 INFO:teuthology.orchestra.run.vm02.stdout: gperftools-libs-2.9.1-3.el9.x86_64 2026-03-06T13:44:07.275 INFO:teuthology.orchestra.run.vm02.stdout: grpc-data-1.46.7-10.el9.noarch 2026-03-06T13:44:07.275 INFO:teuthology.orchestra.run.vm02.stdout: ledmon-libs-1.1.0-3.el9.x86_64 2026-03-06T13:44:07.275 INFO:teuthology.orchestra.run.vm02.stdout: libarrow-9.0.0-15.el9.x86_64 2026-03-06T13:44:07.275 INFO:teuthology.orchestra.run.vm02.stdout: libarrow-doc-9.0.0-15.el9.noarch 2026-03-06T13:44:07.275 INFO:teuthology.orchestra.run.vm02.stdout: libcephfs-devel-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 2026-03-06T13:44:07.275 INFO:teuthology.orchestra.run.vm02.stdout: libcephfs2-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 2026-03-06T13:44:07.275 INFO:teuthology.orchestra.run.vm02.stdout: libcephsqlite-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 2026-03-06T13:44:07.275 INFO:teuthology.orchestra.run.vm02.stdout: libconfig-1.7.2-9.el9.x86_64 2026-03-06T13:44:07.275 INFO:teuthology.orchestra.run.vm02.stdout: libgfortran-11.5.0-14.el9.x86_64 2026-03-06T13:44:07.275 INFO:teuthology.orchestra.run.vm02.stdout: libnbd-1.20.3-4.el9.x86_64 2026-03-06T13:44:07.275 INFO:teuthology.orchestra.run.vm02.stdout: liboath-2.6.12-1.el9.x86_64 2026-03-06T13:44:07.275 INFO:teuthology.orchestra.run.vm02.stdout: libpmemobj-1.12.1-1.el9.x86_64 2026-03-06T13:44:07.275 INFO:teuthology.orchestra.run.vm02.stdout: libquadmath-11.5.0-14.el9.x86_64 2026-03-06T13:44:07.275 INFO:teuthology.orchestra.run.vm02.stdout: librabbitmq-0.11.0-7.el9.x86_64 2026-03-06T13:44:07.275 INFO:teuthology.orchestra.run.vm02.stdout: librados-devel-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 2026-03-06T13:44:07.275 INFO:teuthology.orchestra.run.vm02.stdout: libradosstriper1-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 2026-03-06T13:44:07.275 INFO:teuthology.orchestra.run.vm02.stdout: librdkafka-1.6.1-102.el9.x86_64 2026-03-06T13:44:07.275 INFO:teuthology.orchestra.run.vm02.stdout: librgw2-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 2026-03-06T13:44:07.275 INFO:teuthology.orchestra.run.vm02.stdout: libstoragemgmt-1.10.1-1.el9.x86_64 2026-03-06T13:44:07.275 INFO:teuthology.orchestra.run.vm02.stdout: libunwind-1.6.2-1.el9.x86_64 2026-03-06T13:44:07.275 INFO:teuthology.orchestra.run.vm02.stdout: libxslt-1.1.34-12.el9.x86_64 2026-03-06T13:44:07.275 INFO:teuthology.orchestra.run.vm02.stdout: lttng-ust-2.12.0-6.el9.x86_64 2026-03-06T13:44:07.275 INFO:teuthology.orchestra.run.vm02.stdout: lua-5.4.4-4.el9.x86_64 2026-03-06T13:44:07.275 INFO:teuthology.orchestra.run.vm02.stdout: lua-devel-5.4.4-4.el9.x86_64 2026-03-06T13:44:07.275 INFO:teuthology.orchestra.run.vm02.stdout: luarocks-3.9.2-5.el9.noarch 2026-03-06T13:44:07.275 INFO:teuthology.orchestra.run.vm02.stdout: mailcap-2.1.49-5.el9.noarch 2026-03-06T13:44:07.275 INFO:teuthology.orchestra.run.vm02.stdout: openblas-0.3.29-1.el9.x86_64 2026-03-06T13:44:07.275 INFO:teuthology.orchestra.run.vm02.stdout: openblas-openmp-0.3.29-1.el9.x86_64 2026-03-06T13:44:07.275 INFO:teuthology.orchestra.run.vm02.stdout: parquet-libs-9.0.0-15.el9.x86_64 2026-03-06T13:44:07.275 INFO:teuthology.orchestra.run.vm02.stdout: pciutils-3.7.0-7.el9.x86_64 2026-03-06T13:44:07.275 INFO:teuthology.orchestra.run.vm02.stdout: perl-Benchmark-1.23-483.el9.noarch 2026-03-06T13:44:07.275 INFO:teuthology.orchestra.run.vm02.stdout: perl-Test-Harness-1:3.42-461.el9.noarch 2026-03-06T13:44:07.276 INFO:teuthology.orchestra.run.vm02.stdout: protobuf-3.14.0-17.el9.x86_64 2026-03-06T13:44:07.276 INFO:teuthology.orchestra.run.vm02.stdout: protobuf-compiler-3.14.0-17.el9.x86_64 2026-03-06T13:44:07.276 INFO:teuthology.orchestra.run.vm02.stdout: python3-asyncssh-2.13.2-5.el9.noarch 2026-03-06T13:44:07.276 INFO:teuthology.orchestra.run.vm02.stdout: python3-autocommand-2.2.2-8.el9.noarch 2026-03-06T13:44:07.276 INFO:teuthology.orchestra.run.vm02.stdout: python3-babel-2.9.1-2.el9.noarch 2026-03-06T13:44:07.276 INFO:teuthology.orchestra.run.vm02.stdout: python3-backports-tarfile-1.2.0-1.el9.noarch 2026-03-06T13:44:07.276 INFO:teuthology.orchestra.run.vm02.stdout: python3-bcrypt-3.2.2-1.el9.x86_64 2026-03-06T13:44:07.276 INFO:teuthology.orchestra.run.vm02.stdout: python3-cachetools-4.2.4-1.el9.noarch 2026-03-06T13:44:07.276 INFO:teuthology.orchestra.run.vm02.stdout: python3-ceph-argparse-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 2026-03-06T13:44:07.276 INFO:teuthology.orchestra.run.vm02.stdout: python3-ceph-common-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 2026-03-06T13:44:07.276 INFO:teuthology.orchestra.run.vm02.stdout: python3-cephfs-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 2026-03-06T13:44:07.276 INFO:teuthology.orchestra.run.vm02.stdout: python3-certifi-2023.05.07-4.el9.noarch 2026-03-06T13:44:07.276 INFO:teuthology.orchestra.run.vm02.stdout: python3-cffi-1.14.5-5.el9.x86_64 2026-03-06T13:44:07.276 INFO:teuthology.orchestra.run.vm02.stdout: python3-cheroot-10.0.1-4.el9.noarch 2026-03-06T13:44:07.276 INFO:teuthology.orchestra.run.vm02.stdout: python3-cherrypy-18.6.1-2.el9.noarch 2026-03-06T13:44:07.276 INFO:teuthology.orchestra.run.vm02.stdout: python3-cryptography-36.0.1-5.el9.x86_64 2026-03-06T13:44:07.276 INFO:teuthology.orchestra.run.vm02.stdout: python3-devel-3.9.25-3.el9.x86_64 2026-03-06T13:44:07.276 INFO:teuthology.orchestra.run.vm02.stdout: python3-google-auth-1:2.45.0-1.el9.noarch 2026-03-06T13:44:07.276 INFO:teuthology.orchestra.run.vm02.stdout: python3-grpcio-1.46.7-10.el9.x86_64 2026-03-06T13:44:07.276 INFO:teuthology.orchestra.run.vm02.stdout: python3-grpcio-tools-1.46.7-10.el9.x86_64 2026-03-06T13:44:07.276 INFO:teuthology.orchestra.run.vm02.stdout: python3-influxdb-5.3.1-1.el9.noarch 2026-03-06T13:44:07.276 INFO:teuthology.orchestra.run.vm02.stdout: python3-isodate-0.6.1-3.el9.noarch 2026-03-06T13:44:07.276 INFO:teuthology.orchestra.run.vm02.stdout: python3-jaraco-8.2.1-3.el9.noarch 2026-03-06T13:44:07.276 INFO:teuthology.orchestra.run.vm02.stdout: python3-jaraco-classes-3.2.1-5.el9.noarch 2026-03-06T13:44:07.276 INFO:teuthology.orchestra.run.vm02.stdout: python3-jaraco-collections-3.0.0-8.el9.noarch 2026-03-06T13:44:07.276 INFO:teuthology.orchestra.run.vm02.stdout: python3-jaraco-context-6.0.1-3.el9.noarch 2026-03-06T13:44:07.276 INFO:teuthology.orchestra.run.vm02.stdout: python3-jaraco-functools-3.5.0-2.el9.noarch 2026-03-06T13:44:07.276 INFO:teuthology.orchestra.run.vm02.stdout: python3-jaraco-text-4.0.0-2.el9.noarch 2026-03-06T13:44:07.276 INFO:teuthology.orchestra.run.vm02.stdout: python3-jinja2-2.11.3-8.el9.noarch 2026-03-06T13:44:07.276 INFO:teuthology.orchestra.run.vm02.stdout: python3-kubernetes-1:26.1.0-3.el9.noarch 2026-03-06T13:44:07.276 INFO:teuthology.orchestra.run.vm02.stdout: python3-libstoragemgmt-1.10.1-1.el9.x86_64 2026-03-06T13:44:07.276 INFO:teuthology.orchestra.run.vm02.stdout: python3-logutils-0.3.5-21.el9.noarch 2026-03-06T13:44:07.276 INFO:teuthology.orchestra.run.vm02.stdout: python3-lxml-4.6.5-3.el9.x86_64 2026-03-06T13:44:07.276 INFO:teuthology.orchestra.run.vm02.stdout: python3-mako-1.1.4-6.el9.noarch 2026-03-06T13:44:07.276 INFO:teuthology.orchestra.run.vm02.stdout: python3-markupsafe-1.1.1-12.el9.x86_64 2026-03-06T13:44:07.276 INFO:teuthology.orchestra.run.vm02.stdout: python3-more-itertools-8.12.0-2.el9.noarch 2026-03-06T13:44:07.276 INFO:teuthology.orchestra.run.vm02.stdout: python3-msgpack-1.0.3-2.el9.x86_64 2026-03-06T13:44:07.276 INFO:teuthology.orchestra.run.vm02.stdout: python3-natsort-7.1.1-5.el9.noarch 2026-03-06T13:44:07.276 INFO:teuthology.orchestra.run.vm02.stdout: python3-numpy-1:1.23.5-2.el9.x86_64 2026-03-06T13:44:07.276 INFO:teuthology.orchestra.run.vm02.stdout: python3-numpy-f2py-1:1.23.5-2.el9.x86_64 2026-03-06T13:44:07.276 INFO:teuthology.orchestra.run.vm02.stdout: python3-packaging-20.9-5.el9.noarch 2026-03-06T13:44:07.276 INFO:teuthology.orchestra.run.vm02.stdout: python3-pecan-1.4.2-3.el9.noarch 2026-03-06T13:44:07.276 INFO:teuthology.orchestra.run.vm02.stdout: python3-ply-3.11-14.el9.noarch 2026-03-06T13:44:07.276 INFO:teuthology.orchestra.run.vm02.stdout: python3-portend-3.1.0-2.el9.noarch 2026-03-06T13:44:07.276 INFO:teuthology.orchestra.run.vm02.stdout: python3-protobuf-3.14.0-17.el9.noarch 2026-03-06T13:44:07.276 INFO:teuthology.orchestra.run.vm02.stdout: python3-pyOpenSSL-21.0.0-1.el9.noarch 2026-03-06T13:44:07.276 INFO:teuthology.orchestra.run.vm02.stdout: python3-pyasn1-0.4.8-7.el9.noarch 2026-03-06T13:44:07.276 INFO:teuthology.orchestra.run.vm02.stdout: python3-pyasn1-modules-0.4.8-7.el9.noarch 2026-03-06T13:44:07.276 INFO:teuthology.orchestra.run.vm02.stdout: python3-pycparser-2.20-6.el9.noarch 2026-03-06T13:44:07.276 INFO:teuthology.orchestra.run.vm02.stdout: python3-rados-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 2026-03-06T13:44:07.276 INFO:teuthology.orchestra.run.vm02.stdout: python3-rbd-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 2026-03-06T13:44:07.276 INFO:teuthology.orchestra.run.vm02.stdout: python3-repoze-lru-0.7-16.el9.noarch 2026-03-06T13:44:07.276 INFO:teuthology.orchestra.run.vm02.stdout: python3-requests-2.25.1-10.el9.noarch 2026-03-06T13:44:07.276 INFO:teuthology.orchestra.run.vm02.stdout: python3-requests-oauthlib-1.3.0-12.el9.noarch 2026-03-06T13:44:07.276 INFO:teuthology.orchestra.run.vm02.stdout: python3-rgw-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 2026-03-06T13:44:07.276 INFO:teuthology.orchestra.run.vm02.stdout: python3-routes-2.5.1-5.el9.noarch 2026-03-06T13:44:07.276 INFO:teuthology.orchestra.run.vm02.stdout: python3-rsa-4.9-2.el9.noarch 2026-03-06T13:44:07.276 INFO:teuthology.orchestra.run.vm02.stdout: python3-saml-1.16.0-1.el9.noarch 2026-03-06T13:44:07.276 INFO:teuthology.orchestra.run.vm02.stdout: python3-scipy-1.9.3-2.el9.x86_64 2026-03-06T13:44:07.277 INFO:teuthology.orchestra.run.vm02.stdout: python3-tempora-5.0.0-2.el9.noarch 2026-03-06T13:44:07.277 INFO:teuthology.orchestra.run.vm02.stdout: python3-toml-0.10.2-6.el9.noarch 2026-03-06T13:44:07.277 INFO:teuthology.orchestra.run.vm02.stdout: python3-typing-extensions-4.15.0-1.el9.noarch 2026-03-06T13:44:07.277 INFO:teuthology.orchestra.run.vm02.stdout: python3-urllib3-1.26.5-7.el9.noarch 2026-03-06T13:44:07.277 INFO:teuthology.orchestra.run.vm02.stdout: python3-webob-1.8.8-2.el9.noarch 2026-03-06T13:44:07.277 INFO:teuthology.orchestra.run.vm02.stdout: python3-websocket-client-1.2.3-2.el9.noarch 2026-03-06T13:44:07.277 INFO:teuthology.orchestra.run.vm02.stdout: python3-werkzeug-2.0.3-3.el9.1.noarch 2026-03-06T13:44:07.277 INFO:teuthology.orchestra.run.vm02.stdout: python3-xmlsec-1.3.13-1.el9.x86_64 2026-03-06T13:44:07.277 INFO:teuthology.orchestra.run.vm02.stdout: python3-xmltodict-0.12.0-15.el9.noarch 2026-03-06T13:44:07.277 INFO:teuthology.orchestra.run.vm02.stdout: python3-zc-lockfile-2.0-10.el9.noarch 2026-03-06T13:44:07.277 INFO:teuthology.orchestra.run.vm02.stdout: qatlib-25.08.0-2.el9.x86_64 2026-03-06T13:44:07.277 INFO:teuthology.orchestra.run.vm02.stdout: qatlib-service-25.08.0-2.el9.x86_64 2026-03-06T13:44:07.277 INFO:teuthology.orchestra.run.vm02.stdout: qatzip-libs-1.3.1-1.el9.x86_64 2026-03-06T13:44:07.277 INFO:teuthology.orchestra.run.vm02.stdout: rbd-fuse-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 2026-03-06T13:44:07.277 INFO:teuthology.orchestra.run.vm02.stdout: rbd-mirror-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 2026-03-06T13:44:07.277 INFO:teuthology.orchestra.run.vm02.stdout: rbd-nbd-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 2026-03-06T13:44:07.277 INFO:teuthology.orchestra.run.vm02.stdout: re2-1:20211101-20.el9.x86_64 2026-03-06T13:44:07.277 INFO:teuthology.orchestra.run.vm02.stdout: s3cmd-2.4.0-1.el9.noarch 2026-03-06T13:44:07.277 INFO:teuthology.orchestra.run.vm02.stdout: smartmontools-1:7.2-10.el9.x86_64 2026-03-06T13:44:07.277 INFO:teuthology.orchestra.run.vm02.stdout: socat-1.7.4.1-8.el9.x86_64 2026-03-06T13:44:07.277 INFO:teuthology.orchestra.run.vm02.stdout: thrift-0.15.0-4.el9.x86_64 2026-03-06T13:44:07.277 INFO:teuthology.orchestra.run.vm02.stdout: unzip-6.0-59.el9.x86_64 2026-03-06T13:44:07.277 INFO:teuthology.orchestra.run.vm02.stdout: xmlsec1-1.2.29-13.el9.x86_64 2026-03-06T13:44:07.277 INFO:teuthology.orchestra.run.vm02.stdout: xmlsec1-openssl-1.2.29-13.el9.x86_64 2026-03-06T13:44:07.277 INFO:teuthology.orchestra.run.vm02.stdout: xmlstarlet-1.6.1-20.el9.x86_64 2026-03-06T13:44:07.277 INFO:teuthology.orchestra.run.vm02.stdout: zip-3.0-35.el9.x86_64 2026-03-06T13:44:07.277 INFO:teuthology.orchestra.run.vm02.stdout: 2026-03-06T13:44:07.277 INFO:teuthology.orchestra.run.vm02.stdout:Complete! 2026-03-06T13:44:07.327 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : librbd1-2:16.2.4-5.el9.x86_64 152/152 2026-03-06T13:44:07.327 INFO:teuthology.orchestra.run.vm04.stdout: 2026-03-06T13:44:07.327 INFO:teuthology.orchestra.run.vm04.stdout:Upgraded: 2026-03-06T13:44:07.327 INFO:teuthology.orchestra.run.vm04.stdout: librados2-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 2026-03-06T13:44:07.327 INFO:teuthology.orchestra.run.vm04.stdout: librbd1-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 2026-03-06T13:44:07.327 INFO:teuthology.orchestra.run.vm04.stdout:Installed: 2026-03-06T13:44:07.327 INFO:teuthology.orchestra.run.vm04.stdout: abseil-cpp-20211102.0-4.el9.x86_64 2026-03-06T13:44:07.327 INFO:teuthology.orchestra.run.vm04.stdout: boost-program-options-1.75.0-13.el9.x86_64 2026-03-06T13:44:07.327 INFO:teuthology.orchestra.run.vm04.stdout: bzip2-1.0.8-11.el9.x86_64 2026-03-06T13:44:07.327 INFO:teuthology.orchestra.run.vm04.stdout: ceph-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 2026-03-06T13:44:07.327 INFO:teuthology.orchestra.run.vm04.stdout: ceph-base-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 2026-03-06T13:44:07.327 INFO:teuthology.orchestra.run.vm04.stdout: ceph-common-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 2026-03-06T13:44:07.327 INFO:teuthology.orchestra.run.vm04.stdout: ceph-fuse-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 2026-03-06T13:44:07.327 INFO:teuthology.orchestra.run.vm04.stdout: ceph-grafana-dashboards-2:19.2.3-47.gc24117fd552.el9.clyso.noarch 2026-03-06T13:44:07.327 INFO:teuthology.orchestra.run.vm04.stdout: ceph-immutable-object-cache-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 2026-03-06T13:44:07.327 INFO:teuthology.orchestra.run.vm04.stdout: ceph-mds-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 2026-03-06T13:44:07.327 INFO:teuthology.orchestra.run.vm04.stdout: ceph-mgr-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 2026-03-06T13:44:07.327 INFO:teuthology.orchestra.run.vm04.stdout: ceph-mgr-cephadm-2:19.2.3-47.gc24117fd552.el9.clyso.noarch 2026-03-06T13:44:07.327 INFO:teuthology.orchestra.run.vm04.stdout: ceph-mgr-dashboard-2:19.2.3-47.gc24117fd552.el9.clyso.noarch 2026-03-06T13:44:07.327 INFO:teuthology.orchestra.run.vm04.stdout: ceph-mgr-diskprediction-local-2:19.2.3-47.gc24117fd552.el9.clyso.noarch 2026-03-06T13:44:07.327 INFO:teuthology.orchestra.run.vm04.stdout: ceph-mgr-k8sevents-2:19.2.3-47.gc24117fd552.el9.clyso.noarch 2026-03-06T13:44:07.327 INFO:teuthology.orchestra.run.vm04.stdout: ceph-mgr-modules-core-2:19.2.3-47.gc24117fd552.el9.clyso.noarch 2026-03-06T13:44:07.327 INFO:teuthology.orchestra.run.vm04.stdout: ceph-mgr-rook-2:19.2.3-47.gc24117fd552.el9.clyso.noarch 2026-03-06T13:44:07.327 INFO:teuthology.orchestra.run.vm04.stdout: ceph-mon-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 2026-03-06T13:44:07.327 INFO:teuthology.orchestra.run.vm04.stdout: ceph-osd-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 2026-03-06T13:44:07.327 INFO:teuthology.orchestra.run.vm04.stdout: ceph-prometheus-alerts-2:19.2.3-47.gc24117fd552.el9.clyso.noarch 2026-03-06T13:44:07.327 INFO:teuthology.orchestra.run.vm04.stdout: ceph-radosgw-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 2026-03-06T13:44:07.328 INFO:teuthology.orchestra.run.vm04.stdout: ceph-selinux-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 2026-03-06T13:44:07.328 INFO:teuthology.orchestra.run.vm04.stdout: ceph-test-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 2026-03-06T13:44:07.328 INFO:teuthology.orchestra.run.vm04.stdout: ceph-volume-2:19.2.3-47.gc24117fd552.el9.clyso.noarch 2026-03-06T13:44:07.328 INFO:teuthology.orchestra.run.vm04.stdout: cephadm-2:19.2.3-47.gc24117fd552.el9.clyso.noarch 2026-03-06T13:44:07.328 INFO:teuthology.orchestra.run.vm04.stdout: cryptsetup-2.8.1-3.el9.x86_64 2026-03-06T13:44:07.328 INFO:teuthology.orchestra.run.vm04.stdout: flexiblas-3.0.4-9.el9.x86_64 2026-03-06T13:44:07.328 INFO:teuthology.orchestra.run.vm04.stdout: flexiblas-netlib-3.0.4-9.el9.x86_64 2026-03-06T13:44:07.328 INFO:teuthology.orchestra.run.vm04.stdout: flexiblas-openblas-openmp-3.0.4-9.el9.x86_64 2026-03-06T13:44:07.328 INFO:teuthology.orchestra.run.vm04.stdout: fuse-2.9.9-17.el9.x86_64 2026-03-06T13:44:07.328 INFO:teuthology.orchestra.run.vm04.stdout: gperftools-libs-2.9.1-3.el9.x86_64 2026-03-06T13:44:07.328 INFO:teuthology.orchestra.run.vm04.stdout: grpc-data-1.46.7-10.el9.noarch 2026-03-06T13:44:07.328 INFO:teuthology.orchestra.run.vm04.stdout: ledmon-libs-1.1.0-3.el9.x86_64 2026-03-06T13:44:07.328 INFO:teuthology.orchestra.run.vm04.stdout: libarrow-9.0.0-15.el9.x86_64 2026-03-06T13:44:07.328 INFO:teuthology.orchestra.run.vm04.stdout: libarrow-doc-9.0.0-15.el9.noarch 2026-03-06T13:44:07.328 INFO:teuthology.orchestra.run.vm04.stdout: libcephfs-devel-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 2026-03-06T13:44:07.328 INFO:teuthology.orchestra.run.vm04.stdout: libcephfs2-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 2026-03-06T13:44:07.328 INFO:teuthology.orchestra.run.vm04.stdout: libcephsqlite-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 2026-03-06T13:44:07.328 INFO:teuthology.orchestra.run.vm04.stdout: libconfig-1.7.2-9.el9.x86_64 2026-03-06T13:44:07.328 INFO:teuthology.orchestra.run.vm04.stdout: libgfortran-11.5.0-14.el9.x86_64 2026-03-06T13:44:07.328 INFO:teuthology.orchestra.run.vm04.stdout: libnbd-1.20.3-4.el9.x86_64 2026-03-06T13:44:07.328 INFO:teuthology.orchestra.run.vm04.stdout: liboath-2.6.12-1.el9.x86_64 2026-03-06T13:44:07.328 INFO:teuthology.orchestra.run.vm04.stdout: libpmemobj-1.12.1-1.el9.x86_64 2026-03-06T13:44:07.328 INFO:teuthology.orchestra.run.vm04.stdout: libquadmath-11.5.0-14.el9.x86_64 2026-03-06T13:44:07.328 INFO:teuthology.orchestra.run.vm04.stdout: librabbitmq-0.11.0-7.el9.x86_64 2026-03-06T13:44:07.328 INFO:teuthology.orchestra.run.vm04.stdout: librados-devel-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 2026-03-06T13:44:07.328 INFO:teuthology.orchestra.run.vm04.stdout: libradosstriper1-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 2026-03-06T13:44:07.328 INFO:teuthology.orchestra.run.vm04.stdout: librdkafka-1.6.1-102.el9.x86_64 2026-03-06T13:44:07.328 INFO:teuthology.orchestra.run.vm04.stdout: librgw2-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 2026-03-06T13:44:07.328 INFO:teuthology.orchestra.run.vm04.stdout: libstoragemgmt-1.10.1-1.el9.x86_64 2026-03-06T13:44:07.328 INFO:teuthology.orchestra.run.vm04.stdout: libunwind-1.6.2-1.el9.x86_64 2026-03-06T13:44:07.328 INFO:teuthology.orchestra.run.vm04.stdout: libxslt-1.1.34-12.el9.x86_64 2026-03-06T13:44:07.328 INFO:teuthology.orchestra.run.vm04.stdout: lttng-ust-2.12.0-6.el9.x86_64 2026-03-06T13:44:07.328 INFO:teuthology.orchestra.run.vm04.stdout: lua-5.4.4-4.el9.x86_64 2026-03-06T13:44:07.328 INFO:teuthology.orchestra.run.vm04.stdout: lua-devel-5.4.4-4.el9.x86_64 2026-03-06T13:44:07.328 INFO:teuthology.orchestra.run.vm04.stdout: luarocks-3.9.2-5.el9.noarch 2026-03-06T13:44:07.328 INFO:teuthology.orchestra.run.vm04.stdout: mailcap-2.1.49-5.el9.noarch 2026-03-06T13:44:07.328 INFO:teuthology.orchestra.run.vm04.stdout: openblas-0.3.29-1.el9.x86_64 2026-03-06T13:44:07.328 INFO:teuthology.orchestra.run.vm04.stdout: openblas-openmp-0.3.29-1.el9.x86_64 2026-03-06T13:44:07.328 INFO:teuthology.orchestra.run.vm04.stdout: parquet-libs-9.0.0-15.el9.x86_64 2026-03-06T13:44:07.328 INFO:teuthology.orchestra.run.vm04.stdout: pciutils-3.7.0-7.el9.x86_64 2026-03-06T13:44:07.328 INFO:teuthology.orchestra.run.vm04.stdout: perl-Benchmark-1.23-483.el9.noarch 2026-03-06T13:44:07.328 INFO:teuthology.orchestra.run.vm04.stdout: perl-Test-Harness-1:3.42-461.el9.noarch 2026-03-06T13:44:07.328 INFO:teuthology.orchestra.run.vm04.stdout: protobuf-3.14.0-17.el9.x86_64 2026-03-06T13:44:07.328 INFO:teuthology.orchestra.run.vm04.stdout: protobuf-compiler-3.14.0-17.el9.x86_64 2026-03-06T13:44:07.328 INFO:teuthology.orchestra.run.vm04.stdout: python3-asyncssh-2.13.2-5.el9.noarch 2026-03-06T13:44:07.328 INFO:teuthology.orchestra.run.vm04.stdout: python3-autocommand-2.2.2-8.el9.noarch 2026-03-06T13:44:07.328 INFO:teuthology.orchestra.run.vm04.stdout: python3-babel-2.9.1-2.el9.noarch 2026-03-06T13:44:07.328 INFO:teuthology.orchestra.run.vm04.stdout: python3-backports-tarfile-1.2.0-1.el9.noarch 2026-03-06T13:44:07.328 INFO:teuthology.orchestra.run.vm04.stdout: python3-bcrypt-3.2.2-1.el9.x86_64 2026-03-06T13:44:07.328 INFO:teuthology.orchestra.run.vm04.stdout: python3-cachetools-4.2.4-1.el9.noarch 2026-03-06T13:44:07.328 INFO:teuthology.orchestra.run.vm04.stdout: python3-ceph-argparse-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 2026-03-06T13:44:07.328 INFO:teuthology.orchestra.run.vm04.stdout: python3-ceph-common-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 2026-03-06T13:44:07.328 INFO:teuthology.orchestra.run.vm04.stdout: python3-cephfs-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 2026-03-06T13:44:07.328 INFO:teuthology.orchestra.run.vm04.stdout: python3-certifi-2023.05.07-4.el9.noarch 2026-03-06T13:44:07.329 INFO:teuthology.orchestra.run.vm04.stdout: python3-cffi-1.14.5-5.el9.x86_64 2026-03-06T13:44:07.329 INFO:teuthology.orchestra.run.vm04.stdout: python3-cheroot-10.0.1-4.el9.noarch 2026-03-06T13:44:07.329 INFO:teuthology.orchestra.run.vm04.stdout: python3-cherrypy-18.6.1-2.el9.noarch 2026-03-06T13:44:07.329 INFO:teuthology.orchestra.run.vm04.stdout: python3-cryptography-36.0.1-5.el9.x86_64 2026-03-06T13:44:07.329 INFO:teuthology.orchestra.run.vm04.stdout: python3-devel-3.9.25-3.el9.x86_64 2026-03-06T13:44:07.329 INFO:teuthology.orchestra.run.vm04.stdout: python3-google-auth-1:2.45.0-1.el9.noarch 2026-03-06T13:44:07.329 INFO:teuthology.orchestra.run.vm04.stdout: python3-grpcio-1.46.7-10.el9.x86_64 2026-03-06T13:44:07.329 INFO:teuthology.orchestra.run.vm04.stdout: python3-grpcio-tools-1.46.7-10.el9.x86_64 2026-03-06T13:44:07.329 INFO:teuthology.orchestra.run.vm04.stdout: python3-influxdb-5.3.1-1.el9.noarch 2026-03-06T13:44:07.329 INFO:teuthology.orchestra.run.vm04.stdout: python3-isodate-0.6.1-3.el9.noarch 2026-03-06T13:44:07.329 INFO:teuthology.orchestra.run.vm04.stdout: python3-jaraco-8.2.1-3.el9.noarch 2026-03-06T13:44:07.329 INFO:teuthology.orchestra.run.vm04.stdout: python3-jaraco-classes-3.2.1-5.el9.noarch 2026-03-06T13:44:07.329 INFO:teuthology.orchestra.run.vm04.stdout: python3-jaraco-collections-3.0.0-8.el9.noarch 2026-03-06T13:44:07.329 INFO:teuthology.orchestra.run.vm04.stdout: python3-jaraco-context-6.0.1-3.el9.noarch 2026-03-06T13:44:07.329 INFO:teuthology.orchestra.run.vm04.stdout: python3-jaraco-functools-3.5.0-2.el9.noarch 2026-03-06T13:44:07.329 INFO:teuthology.orchestra.run.vm04.stdout: python3-jaraco-text-4.0.0-2.el9.noarch 2026-03-06T13:44:07.329 INFO:teuthology.orchestra.run.vm04.stdout: python3-jinja2-2.11.3-8.el9.noarch 2026-03-06T13:44:07.329 INFO:teuthology.orchestra.run.vm04.stdout: python3-kubernetes-1:26.1.0-3.el9.noarch 2026-03-06T13:44:07.329 INFO:teuthology.orchestra.run.vm04.stdout: python3-libstoragemgmt-1.10.1-1.el9.x86_64 2026-03-06T13:44:07.329 INFO:teuthology.orchestra.run.vm04.stdout: python3-logutils-0.3.5-21.el9.noarch 2026-03-06T13:44:07.329 INFO:teuthology.orchestra.run.vm04.stdout: python3-lxml-4.6.5-3.el9.x86_64 2026-03-06T13:44:07.329 INFO:teuthology.orchestra.run.vm04.stdout: python3-mako-1.1.4-6.el9.noarch 2026-03-06T13:44:07.329 INFO:teuthology.orchestra.run.vm04.stdout: python3-markupsafe-1.1.1-12.el9.x86_64 2026-03-06T13:44:07.329 INFO:teuthology.orchestra.run.vm04.stdout: python3-more-itertools-8.12.0-2.el9.noarch 2026-03-06T13:44:07.329 INFO:teuthology.orchestra.run.vm04.stdout: python3-msgpack-1.0.3-2.el9.x86_64 2026-03-06T13:44:07.329 INFO:teuthology.orchestra.run.vm04.stdout: python3-natsort-7.1.1-5.el9.noarch 2026-03-06T13:44:07.329 INFO:teuthology.orchestra.run.vm04.stdout: python3-numpy-1:1.23.5-2.el9.x86_64 2026-03-06T13:44:07.329 INFO:teuthology.orchestra.run.vm04.stdout: python3-numpy-f2py-1:1.23.5-2.el9.x86_64 2026-03-06T13:44:07.329 INFO:teuthology.orchestra.run.vm04.stdout: python3-packaging-20.9-5.el9.noarch 2026-03-06T13:44:07.329 INFO:teuthology.orchestra.run.vm04.stdout: python3-pecan-1.4.2-3.el9.noarch 2026-03-06T13:44:07.329 INFO:teuthology.orchestra.run.vm04.stdout: python3-ply-3.11-14.el9.noarch 2026-03-06T13:44:07.329 INFO:teuthology.orchestra.run.vm04.stdout: python3-portend-3.1.0-2.el9.noarch 2026-03-06T13:44:07.329 INFO:teuthology.orchestra.run.vm04.stdout: python3-protobuf-3.14.0-17.el9.noarch 2026-03-06T13:44:07.329 INFO:teuthology.orchestra.run.vm04.stdout: python3-pyOpenSSL-21.0.0-1.el9.noarch 2026-03-06T13:44:07.329 INFO:teuthology.orchestra.run.vm04.stdout: python3-pyasn1-0.4.8-7.el9.noarch 2026-03-06T13:44:07.329 INFO:teuthology.orchestra.run.vm04.stdout: python3-pyasn1-modules-0.4.8-7.el9.noarch 2026-03-06T13:44:07.329 INFO:teuthology.orchestra.run.vm04.stdout: python3-pycparser-2.20-6.el9.noarch 2026-03-06T13:44:07.329 INFO:teuthology.orchestra.run.vm04.stdout: python3-rados-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 2026-03-06T13:44:07.329 INFO:teuthology.orchestra.run.vm04.stdout: python3-rbd-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 2026-03-06T13:44:07.329 INFO:teuthology.orchestra.run.vm04.stdout: python3-repoze-lru-0.7-16.el9.noarch 2026-03-06T13:44:07.329 INFO:teuthology.orchestra.run.vm04.stdout: python3-requests-2.25.1-10.el9.noarch 2026-03-06T13:44:07.329 INFO:teuthology.orchestra.run.vm04.stdout: python3-requests-oauthlib-1.3.0-12.el9.noarch 2026-03-06T13:44:07.329 INFO:teuthology.orchestra.run.vm04.stdout: python3-rgw-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 2026-03-06T13:44:07.329 INFO:teuthology.orchestra.run.vm04.stdout: python3-routes-2.5.1-5.el9.noarch 2026-03-06T13:44:07.329 INFO:teuthology.orchestra.run.vm04.stdout: python3-rsa-4.9-2.el9.noarch 2026-03-06T13:44:07.329 INFO:teuthology.orchestra.run.vm04.stdout: python3-saml-1.16.0-1.el9.noarch 2026-03-06T13:44:07.329 INFO:teuthology.orchestra.run.vm04.stdout: python3-scipy-1.9.3-2.el9.x86_64 2026-03-06T13:44:07.329 INFO:teuthology.orchestra.run.vm04.stdout: python3-tempora-5.0.0-2.el9.noarch 2026-03-06T13:44:07.329 INFO:teuthology.orchestra.run.vm04.stdout: python3-toml-0.10.2-6.el9.noarch 2026-03-06T13:44:07.329 INFO:teuthology.orchestra.run.vm04.stdout: python3-typing-extensions-4.15.0-1.el9.noarch 2026-03-06T13:44:07.329 INFO:teuthology.orchestra.run.vm04.stdout: python3-urllib3-1.26.5-7.el9.noarch 2026-03-06T13:44:07.329 INFO:teuthology.orchestra.run.vm04.stdout: python3-webob-1.8.8-2.el9.noarch 2026-03-06T13:44:07.329 INFO:teuthology.orchestra.run.vm04.stdout: python3-websocket-client-1.2.3-2.el9.noarch 2026-03-06T13:44:07.329 INFO:teuthology.orchestra.run.vm04.stdout: python3-werkzeug-2.0.3-3.el9.1.noarch 2026-03-06T13:44:07.329 INFO:teuthology.orchestra.run.vm04.stdout: python3-xmlsec-1.3.13-1.el9.x86_64 2026-03-06T13:44:07.329 INFO:teuthology.orchestra.run.vm04.stdout: python3-xmltodict-0.12.0-15.el9.noarch 2026-03-06T13:44:07.329 INFO:teuthology.orchestra.run.vm04.stdout: python3-zc-lockfile-2.0-10.el9.noarch 2026-03-06T13:44:07.329 INFO:teuthology.orchestra.run.vm04.stdout: qatlib-25.08.0-2.el9.x86_64 2026-03-06T13:44:07.329 INFO:teuthology.orchestra.run.vm04.stdout: qatlib-service-25.08.0-2.el9.x86_64 2026-03-06T13:44:07.329 INFO:teuthology.orchestra.run.vm04.stdout: qatzip-libs-1.3.1-1.el9.x86_64 2026-03-06T13:44:07.329 INFO:teuthology.orchestra.run.vm04.stdout: rbd-fuse-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 2026-03-06T13:44:07.330 INFO:teuthology.orchestra.run.vm04.stdout: rbd-mirror-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 2026-03-06T13:44:07.330 INFO:teuthology.orchestra.run.vm04.stdout: rbd-nbd-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 2026-03-06T13:44:07.330 INFO:teuthology.orchestra.run.vm04.stdout: re2-1:20211101-20.el9.x86_64 2026-03-06T13:44:07.330 INFO:teuthology.orchestra.run.vm04.stdout: s3cmd-2.4.0-1.el9.noarch 2026-03-06T13:44:07.330 INFO:teuthology.orchestra.run.vm04.stdout: smartmontools-1:7.2-10.el9.x86_64 2026-03-06T13:44:07.330 INFO:teuthology.orchestra.run.vm04.stdout: socat-1.7.4.1-8.el9.x86_64 2026-03-06T13:44:07.330 INFO:teuthology.orchestra.run.vm04.stdout: thrift-0.15.0-4.el9.x86_64 2026-03-06T13:44:07.330 INFO:teuthology.orchestra.run.vm04.stdout: unzip-6.0-59.el9.x86_64 2026-03-06T13:44:07.330 INFO:teuthology.orchestra.run.vm04.stdout: xmlsec1-1.2.29-13.el9.x86_64 2026-03-06T13:44:07.330 INFO:teuthology.orchestra.run.vm04.stdout: xmlsec1-openssl-1.2.29-13.el9.x86_64 2026-03-06T13:44:07.330 INFO:teuthology.orchestra.run.vm04.stdout: xmlstarlet-1.6.1-20.el9.x86_64 2026-03-06T13:44:07.330 INFO:teuthology.orchestra.run.vm04.stdout: zip-3.0-35.el9.x86_64 2026-03-06T13:44:07.330 INFO:teuthology.orchestra.run.vm04.stdout: 2026-03-06T13:44:07.330 INFO:teuthology.orchestra.run.vm04.stdout:Complete! 2026-03-06T13:44:07.379 DEBUG:teuthology.parallel:result is None 2026-03-06T13:44:07.423 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : librbd1-2:16.2.4-5.el9.x86_64 152/152 2026-03-06T13:44:07.423 INFO:teuthology.orchestra.run.vm06.stdout: 2026-03-06T13:44:07.423 INFO:teuthology.orchestra.run.vm06.stdout:Upgraded: 2026-03-06T13:44:07.424 INFO:teuthology.orchestra.run.vm06.stdout: librados2-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 2026-03-06T13:44:07.424 INFO:teuthology.orchestra.run.vm06.stdout: librbd1-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 2026-03-06T13:44:07.424 INFO:teuthology.orchestra.run.vm06.stdout:Installed: 2026-03-06T13:44:07.424 INFO:teuthology.orchestra.run.vm06.stdout: abseil-cpp-20211102.0-4.el9.x86_64 2026-03-06T13:44:07.424 INFO:teuthology.orchestra.run.vm06.stdout: boost-program-options-1.75.0-13.el9.x86_64 2026-03-06T13:44:07.424 INFO:teuthology.orchestra.run.vm06.stdout: bzip2-1.0.8-11.el9.x86_64 2026-03-06T13:44:07.424 INFO:teuthology.orchestra.run.vm06.stdout: ceph-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 2026-03-06T13:44:07.424 INFO:teuthology.orchestra.run.vm06.stdout: ceph-base-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 2026-03-06T13:44:07.424 INFO:teuthology.orchestra.run.vm06.stdout: ceph-common-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 2026-03-06T13:44:07.424 INFO:teuthology.orchestra.run.vm06.stdout: ceph-fuse-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 2026-03-06T13:44:07.424 INFO:teuthology.orchestra.run.vm06.stdout: ceph-grafana-dashboards-2:19.2.3-47.gc24117fd552.el9.clyso.noarch 2026-03-06T13:44:07.424 INFO:teuthology.orchestra.run.vm06.stdout: ceph-immutable-object-cache-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 2026-03-06T13:44:07.424 INFO:teuthology.orchestra.run.vm06.stdout: ceph-mds-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 2026-03-06T13:44:07.424 INFO:teuthology.orchestra.run.vm06.stdout: ceph-mgr-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 2026-03-06T13:44:07.424 INFO:teuthology.orchestra.run.vm06.stdout: ceph-mgr-cephadm-2:19.2.3-47.gc24117fd552.el9.clyso.noarch 2026-03-06T13:44:07.424 INFO:teuthology.orchestra.run.vm06.stdout: ceph-mgr-dashboard-2:19.2.3-47.gc24117fd552.el9.clyso.noarch 2026-03-06T13:44:07.424 INFO:teuthology.orchestra.run.vm06.stdout: ceph-mgr-diskprediction-local-2:19.2.3-47.gc24117fd552.el9.clyso.noarch 2026-03-06T13:44:07.424 INFO:teuthology.orchestra.run.vm06.stdout: ceph-mgr-k8sevents-2:19.2.3-47.gc24117fd552.el9.clyso.noarch 2026-03-06T13:44:07.424 INFO:teuthology.orchestra.run.vm06.stdout: ceph-mgr-modules-core-2:19.2.3-47.gc24117fd552.el9.clyso.noarch 2026-03-06T13:44:07.424 INFO:teuthology.orchestra.run.vm06.stdout: ceph-mgr-rook-2:19.2.3-47.gc24117fd552.el9.clyso.noarch 2026-03-06T13:44:07.424 INFO:teuthology.orchestra.run.vm06.stdout: ceph-mon-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 2026-03-06T13:44:07.424 INFO:teuthology.orchestra.run.vm06.stdout: ceph-osd-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 2026-03-06T13:44:07.424 INFO:teuthology.orchestra.run.vm06.stdout: ceph-prometheus-alerts-2:19.2.3-47.gc24117fd552.el9.clyso.noarch 2026-03-06T13:44:07.424 INFO:teuthology.orchestra.run.vm06.stdout: ceph-radosgw-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 2026-03-06T13:44:07.424 INFO:teuthology.orchestra.run.vm06.stdout: ceph-selinux-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 2026-03-06T13:44:07.424 INFO:teuthology.orchestra.run.vm06.stdout: ceph-test-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 2026-03-06T13:44:07.424 INFO:teuthology.orchestra.run.vm06.stdout: ceph-volume-2:19.2.3-47.gc24117fd552.el9.clyso.noarch 2026-03-06T13:44:07.424 INFO:teuthology.orchestra.run.vm06.stdout: cephadm-2:19.2.3-47.gc24117fd552.el9.clyso.noarch 2026-03-06T13:44:07.424 INFO:teuthology.orchestra.run.vm06.stdout: cryptsetup-2.8.1-3.el9.x86_64 2026-03-06T13:44:07.424 INFO:teuthology.orchestra.run.vm06.stdout: flexiblas-3.0.4-9.el9.x86_64 2026-03-06T13:44:07.424 INFO:teuthology.orchestra.run.vm06.stdout: flexiblas-netlib-3.0.4-9.el9.x86_64 2026-03-06T13:44:07.424 INFO:teuthology.orchestra.run.vm06.stdout: flexiblas-openblas-openmp-3.0.4-9.el9.x86_64 2026-03-06T13:44:07.424 INFO:teuthology.orchestra.run.vm06.stdout: fuse-2.9.9-17.el9.x86_64 2026-03-06T13:44:07.424 INFO:teuthology.orchestra.run.vm06.stdout: gperftools-libs-2.9.1-3.el9.x86_64 2026-03-06T13:44:07.424 INFO:teuthology.orchestra.run.vm06.stdout: grpc-data-1.46.7-10.el9.noarch 2026-03-06T13:44:07.424 INFO:teuthology.orchestra.run.vm06.stdout: ledmon-libs-1.1.0-3.el9.x86_64 2026-03-06T13:44:07.424 INFO:teuthology.orchestra.run.vm06.stdout: libarrow-9.0.0-15.el9.x86_64 2026-03-06T13:44:07.424 INFO:teuthology.orchestra.run.vm06.stdout: libarrow-doc-9.0.0-15.el9.noarch 2026-03-06T13:44:07.424 INFO:teuthology.orchestra.run.vm06.stdout: libcephfs-devel-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 2026-03-06T13:44:07.425 INFO:teuthology.orchestra.run.vm06.stdout: libcephfs2-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 2026-03-06T13:44:07.425 INFO:teuthology.orchestra.run.vm06.stdout: libcephsqlite-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 2026-03-06T13:44:07.425 INFO:teuthology.orchestra.run.vm06.stdout: libconfig-1.7.2-9.el9.x86_64 2026-03-06T13:44:07.425 INFO:teuthology.orchestra.run.vm06.stdout: libgfortran-11.5.0-14.el9.x86_64 2026-03-06T13:44:07.425 INFO:teuthology.orchestra.run.vm06.stdout: libnbd-1.20.3-4.el9.x86_64 2026-03-06T13:44:07.425 INFO:teuthology.orchestra.run.vm06.stdout: liboath-2.6.12-1.el9.x86_64 2026-03-06T13:44:07.425 INFO:teuthology.orchestra.run.vm06.stdout: libpmemobj-1.12.1-1.el9.x86_64 2026-03-06T13:44:07.425 INFO:teuthology.orchestra.run.vm06.stdout: libquadmath-11.5.0-14.el9.x86_64 2026-03-06T13:44:07.425 INFO:teuthology.orchestra.run.vm06.stdout: librabbitmq-0.11.0-7.el9.x86_64 2026-03-06T13:44:07.425 INFO:teuthology.orchestra.run.vm06.stdout: librados-devel-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 2026-03-06T13:44:07.425 INFO:teuthology.orchestra.run.vm06.stdout: libradosstriper1-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 2026-03-06T13:44:07.425 INFO:teuthology.orchestra.run.vm06.stdout: librdkafka-1.6.1-102.el9.x86_64 2026-03-06T13:44:07.425 INFO:teuthology.orchestra.run.vm06.stdout: librgw2-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 2026-03-06T13:44:07.425 INFO:teuthology.orchestra.run.vm06.stdout: libstoragemgmt-1.10.1-1.el9.x86_64 2026-03-06T13:44:07.425 INFO:teuthology.orchestra.run.vm06.stdout: libunwind-1.6.2-1.el9.x86_64 2026-03-06T13:44:07.425 INFO:teuthology.orchestra.run.vm06.stdout: libxslt-1.1.34-12.el9.x86_64 2026-03-06T13:44:07.425 INFO:teuthology.orchestra.run.vm06.stdout: lttng-ust-2.12.0-6.el9.x86_64 2026-03-06T13:44:07.425 INFO:teuthology.orchestra.run.vm06.stdout: lua-5.4.4-4.el9.x86_64 2026-03-06T13:44:07.425 INFO:teuthology.orchestra.run.vm06.stdout: lua-devel-5.4.4-4.el9.x86_64 2026-03-06T13:44:07.425 INFO:teuthology.orchestra.run.vm06.stdout: luarocks-3.9.2-5.el9.noarch 2026-03-06T13:44:07.425 INFO:teuthology.orchestra.run.vm06.stdout: mailcap-2.1.49-5.el9.noarch 2026-03-06T13:44:07.425 INFO:teuthology.orchestra.run.vm06.stdout: openblas-0.3.29-1.el9.x86_64 2026-03-06T13:44:07.425 INFO:teuthology.orchestra.run.vm06.stdout: openblas-openmp-0.3.29-1.el9.x86_64 2026-03-06T13:44:07.425 INFO:teuthology.orchestra.run.vm06.stdout: parquet-libs-9.0.0-15.el9.x86_64 2026-03-06T13:44:07.425 INFO:teuthology.orchestra.run.vm06.stdout: pciutils-3.7.0-7.el9.x86_64 2026-03-06T13:44:07.425 INFO:teuthology.orchestra.run.vm06.stdout: perl-Benchmark-1.23-483.el9.noarch 2026-03-06T13:44:07.425 INFO:teuthology.orchestra.run.vm06.stdout: perl-Test-Harness-1:3.42-461.el9.noarch 2026-03-06T13:44:07.425 INFO:teuthology.orchestra.run.vm06.stdout: protobuf-3.14.0-17.el9.x86_64 2026-03-06T13:44:07.425 INFO:teuthology.orchestra.run.vm06.stdout: protobuf-compiler-3.14.0-17.el9.x86_64 2026-03-06T13:44:07.425 INFO:teuthology.orchestra.run.vm06.stdout: python3-asyncssh-2.13.2-5.el9.noarch 2026-03-06T13:44:07.425 INFO:teuthology.orchestra.run.vm06.stdout: python3-autocommand-2.2.2-8.el9.noarch 2026-03-06T13:44:07.425 INFO:teuthology.orchestra.run.vm06.stdout: python3-babel-2.9.1-2.el9.noarch 2026-03-06T13:44:07.425 INFO:teuthology.orchestra.run.vm06.stdout: python3-backports-tarfile-1.2.0-1.el9.noarch 2026-03-06T13:44:07.425 INFO:teuthology.orchestra.run.vm06.stdout: python3-bcrypt-3.2.2-1.el9.x86_64 2026-03-06T13:44:07.425 INFO:teuthology.orchestra.run.vm06.stdout: python3-cachetools-4.2.4-1.el9.noarch 2026-03-06T13:44:07.425 INFO:teuthology.orchestra.run.vm06.stdout: python3-ceph-argparse-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 2026-03-06T13:44:07.425 INFO:teuthology.orchestra.run.vm06.stdout: python3-ceph-common-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 2026-03-06T13:44:07.425 INFO:teuthology.orchestra.run.vm06.stdout: python3-cephfs-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 2026-03-06T13:44:07.425 INFO:teuthology.orchestra.run.vm06.stdout: python3-certifi-2023.05.07-4.el9.noarch 2026-03-06T13:44:07.425 INFO:teuthology.orchestra.run.vm06.stdout: python3-cffi-1.14.5-5.el9.x86_64 2026-03-06T13:44:07.426 INFO:teuthology.orchestra.run.vm06.stdout: python3-cheroot-10.0.1-4.el9.noarch 2026-03-06T13:44:07.426 INFO:teuthology.orchestra.run.vm06.stdout: python3-cherrypy-18.6.1-2.el9.noarch 2026-03-06T13:44:07.426 INFO:teuthology.orchestra.run.vm06.stdout: python3-cryptography-36.0.1-5.el9.x86_64 2026-03-06T13:44:07.426 INFO:teuthology.orchestra.run.vm06.stdout: python3-devel-3.9.25-3.el9.x86_64 2026-03-06T13:44:07.426 INFO:teuthology.orchestra.run.vm06.stdout: python3-google-auth-1:2.45.0-1.el9.noarch 2026-03-06T13:44:07.426 INFO:teuthology.orchestra.run.vm06.stdout: python3-grpcio-1.46.7-10.el9.x86_64 2026-03-06T13:44:07.426 INFO:teuthology.orchestra.run.vm06.stdout: python3-grpcio-tools-1.46.7-10.el9.x86_64 2026-03-06T13:44:07.426 INFO:teuthology.orchestra.run.vm06.stdout: python3-influxdb-5.3.1-1.el9.noarch 2026-03-06T13:44:07.426 INFO:teuthology.orchestra.run.vm06.stdout: python3-isodate-0.6.1-3.el9.noarch 2026-03-06T13:44:07.426 INFO:teuthology.orchestra.run.vm06.stdout: python3-jaraco-8.2.1-3.el9.noarch 2026-03-06T13:44:07.426 INFO:teuthology.orchestra.run.vm06.stdout: python3-jaraco-classes-3.2.1-5.el9.noarch 2026-03-06T13:44:07.426 INFO:teuthology.orchestra.run.vm06.stdout: python3-jaraco-collections-3.0.0-8.el9.noarch 2026-03-06T13:44:07.426 INFO:teuthology.orchestra.run.vm06.stdout: python3-jaraco-context-6.0.1-3.el9.noarch 2026-03-06T13:44:07.426 INFO:teuthology.orchestra.run.vm06.stdout: python3-jaraco-functools-3.5.0-2.el9.noarch 2026-03-06T13:44:07.426 INFO:teuthology.orchestra.run.vm06.stdout: python3-jaraco-text-4.0.0-2.el9.noarch 2026-03-06T13:44:07.426 INFO:teuthology.orchestra.run.vm06.stdout: python3-jinja2-2.11.3-8.el9.noarch 2026-03-06T13:44:07.426 INFO:teuthology.orchestra.run.vm06.stdout: python3-kubernetes-1:26.1.0-3.el9.noarch 2026-03-06T13:44:07.426 INFO:teuthology.orchestra.run.vm06.stdout: python3-libstoragemgmt-1.10.1-1.el9.x86_64 2026-03-06T13:44:07.426 INFO:teuthology.orchestra.run.vm06.stdout: python3-logutils-0.3.5-21.el9.noarch 2026-03-06T13:44:07.426 INFO:teuthology.orchestra.run.vm06.stdout: python3-lxml-4.6.5-3.el9.x86_64 2026-03-06T13:44:07.426 INFO:teuthology.orchestra.run.vm06.stdout: python3-mako-1.1.4-6.el9.noarch 2026-03-06T13:44:07.426 INFO:teuthology.orchestra.run.vm06.stdout: python3-markupsafe-1.1.1-12.el9.x86_64 2026-03-06T13:44:07.426 INFO:teuthology.orchestra.run.vm06.stdout: python3-more-itertools-8.12.0-2.el9.noarch 2026-03-06T13:44:07.426 INFO:teuthology.orchestra.run.vm06.stdout: python3-msgpack-1.0.3-2.el9.x86_64 2026-03-06T13:44:07.426 INFO:teuthology.orchestra.run.vm06.stdout: python3-natsort-7.1.1-5.el9.noarch 2026-03-06T13:44:07.426 INFO:teuthology.orchestra.run.vm06.stdout: python3-numpy-1:1.23.5-2.el9.x86_64 2026-03-06T13:44:07.426 INFO:teuthology.orchestra.run.vm06.stdout: python3-numpy-f2py-1:1.23.5-2.el9.x86_64 2026-03-06T13:44:07.426 INFO:teuthology.orchestra.run.vm06.stdout: python3-packaging-20.9-5.el9.noarch 2026-03-06T13:44:07.426 INFO:teuthology.orchestra.run.vm06.stdout: python3-pecan-1.4.2-3.el9.noarch 2026-03-06T13:44:07.426 INFO:teuthology.orchestra.run.vm06.stdout: python3-ply-3.11-14.el9.noarch 2026-03-06T13:44:07.426 INFO:teuthology.orchestra.run.vm06.stdout: python3-portend-3.1.0-2.el9.noarch 2026-03-06T13:44:07.426 INFO:teuthology.orchestra.run.vm06.stdout: python3-protobuf-3.14.0-17.el9.noarch 2026-03-06T13:44:07.426 INFO:teuthology.orchestra.run.vm06.stdout: python3-pyOpenSSL-21.0.0-1.el9.noarch 2026-03-06T13:44:07.426 INFO:teuthology.orchestra.run.vm06.stdout: python3-pyasn1-0.4.8-7.el9.noarch 2026-03-06T13:44:07.426 INFO:teuthology.orchestra.run.vm06.stdout: python3-pyasn1-modules-0.4.8-7.el9.noarch 2026-03-06T13:44:07.426 INFO:teuthology.orchestra.run.vm06.stdout: python3-pycparser-2.20-6.el9.noarch 2026-03-06T13:44:07.426 INFO:teuthology.orchestra.run.vm06.stdout: python3-rados-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 2026-03-06T13:44:07.426 INFO:teuthology.orchestra.run.vm06.stdout: python3-rbd-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 2026-03-06T13:44:07.426 INFO:teuthology.orchestra.run.vm06.stdout: python3-repoze-lru-0.7-16.el9.noarch 2026-03-06T13:44:07.426 INFO:teuthology.orchestra.run.vm06.stdout: python3-requests-2.25.1-10.el9.noarch 2026-03-06T13:44:07.426 INFO:teuthology.orchestra.run.vm06.stdout: python3-requests-oauthlib-1.3.0-12.el9.noarch 2026-03-06T13:44:07.426 INFO:teuthology.orchestra.run.vm06.stdout: python3-rgw-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 2026-03-06T13:44:07.427 INFO:teuthology.orchestra.run.vm06.stdout: python3-routes-2.5.1-5.el9.noarch 2026-03-06T13:44:07.427 INFO:teuthology.orchestra.run.vm06.stdout: python3-rsa-4.9-2.el9.noarch 2026-03-06T13:44:07.427 INFO:teuthology.orchestra.run.vm06.stdout: python3-saml-1.16.0-1.el9.noarch 2026-03-06T13:44:07.427 INFO:teuthology.orchestra.run.vm06.stdout: python3-scipy-1.9.3-2.el9.x86_64 2026-03-06T13:44:07.427 INFO:teuthology.orchestra.run.vm06.stdout: python3-tempora-5.0.0-2.el9.noarch 2026-03-06T13:44:07.427 INFO:teuthology.orchestra.run.vm06.stdout: python3-toml-0.10.2-6.el9.noarch 2026-03-06T13:44:07.427 INFO:teuthology.orchestra.run.vm06.stdout: python3-typing-extensions-4.15.0-1.el9.noarch 2026-03-06T13:44:07.427 INFO:teuthology.orchestra.run.vm06.stdout: python3-urllib3-1.26.5-7.el9.noarch 2026-03-06T13:44:07.427 INFO:teuthology.orchestra.run.vm06.stdout: python3-webob-1.8.8-2.el9.noarch 2026-03-06T13:44:07.427 INFO:teuthology.orchestra.run.vm06.stdout: python3-websocket-client-1.2.3-2.el9.noarch 2026-03-06T13:44:07.427 INFO:teuthology.orchestra.run.vm06.stdout: python3-werkzeug-2.0.3-3.el9.1.noarch 2026-03-06T13:44:07.427 INFO:teuthology.orchestra.run.vm06.stdout: python3-xmlsec-1.3.13-1.el9.x86_64 2026-03-06T13:44:07.427 INFO:teuthology.orchestra.run.vm06.stdout: python3-xmltodict-0.12.0-15.el9.noarch 2026-03-06T13:44:07.427 INFO:teuthology.orchestra.run.vm06.stdout: python3-zc-lockfile-2.0-10.el9.noarch 2026-03-06T13:44:07.427 INFO:teuthology.orchestra.run.vm06.stdout: qatlib-25.08.0-2.el9.x86_64 2026-03-06T13:44:07.427 INFO:teuthology.orchestra.run.vm06.stdout: qatlib-service-25.08.0-2.el9.x86_64 2026-03-06T13:44:07.427 INFO:teuthology.orchestra.run.vm06.stdout: qatzip-libs-1.3.1-1.el9.x86_64 2026-03-06T13:44:07.427 INFO:teuthology.orchestra.run.vm06.stdout: rbd-fuse-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 2026-03-06T13:44:07.427 INFO:teuthology.orchestra.run.vm06.stdout: rbd-mirror-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 2026-03-06T13:44:07.427 INFO:teuthology.orchestra.run.vm06.stdout: rbd-nbd-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 2026-03-06T13:44:07.427 INFO:teuthology.orchestra.run.vm06.stdout: re2-1:20211101-20.el9.x86_64 2026-03-06T13:44:07.427 INFO:teuthology.orchestra.run.vm06.stdout: s3cmd-2.4.0-1.el9.noarch 2026-03-06T13:44:07.427 INFO:teuthology.orchestra.run.vm06.stdout: smartmontools-1:7.2-10.el9.x86_64 2026-03-06T13:44:07.427 INFO:teuthology.orchestra.run.vm06.stdout: socat-1.7.4.1-8.el9.x86_64 2026-03-06T13:44:07.427 INFO:teuthology.orchestra.run.vm06.stdout: thrift-0.15.0-4.el9.x86_64 2026-03-06T13:44:07.427 INFO:teuthology.orchestra.run.vm06.stdout: unzip-6.0-59.el9.x86_64 2026-03-06T13:44:07.427 INFO:teuthology.orchestra.run.vm06.stdout: xmlsec1-1.2.29-13.el9.x86_64 2026-03-06T13:44:07.427 INFO:teuthology.orchestra.run.vm06.stdout: xmlsec1-openssl-1.2.29-13.el9.x86_64 2026-03-06T13:44:07.427 INFO:teuthology.orchestra.run.vm06.stdout: xmlstarlet-1.6.1-20.el9.x86_64 2026-03-06T13:44:07.427 INFO:teuthology.orchestra.run.vm06.stdout: zip-3.0-35.el9.x86_64 2026-03-06T13:44:07.427 INFO:teuthology.orchestra.run.vm06.stdout: 2026-03-06T13:44:07.428 INFO:teuthology.orchestra.run.vm06.stdout:Complete! 2026-03-06T13:44:07.430 DEBUG:teuthology.parallel:result is None 2026-03-06T13:44:07.567 DEBUG:teuthology.parallel:result is None 2026-03-06T13:44:07.567 INFO:teuthology.task.install:Skipping version verification because we have custom repos... 2026-03-06T13:44:07.567 INFO:teuthology.task.install:Skipping version verification because we have custom repos... 2026-03-06T13:44:07.567 INFO:teuthology.task.install:Skipping version verification because we have custom repos... 2026-03-06T13:44:07.567 INFO:teuthology.task.install.util:Shipping valgrind.supp... 2026-03-06T13:44:07.567 DEBUG:teuthology.orchestra.run.vm02:> set -ex 2026-03-06T13:44:07.567 DEBUG:teuthology.orchestra.run.vm02:> sudo dd of=/home/ubuntu/cephtest/valgrind.supp 2026-03-06T13:44:07.602 DEBUG:teuthology.orchestra.run.vm04:> set -ex 2026-03-06T13:44:07.602 DEBUG:teuthology.orchestra.run.vm04:> sudo dd of=/home/ubuntu/cephtest/valgrind.supp 2026-03-06T13:44:07.631 DEBUG:teuthology.orchestra.run.vm06:> set -ex 2026-03-06T13:44:07.631 DEBUG:teuthology.orchestra.run.vm06:> sudo dd of=/home/ubuntu/cephtest/valgrind.supp 2026-03-06T13:44:07.699 INFO:teuthology.task.install.util:Shipping 'daemon-helper'... 2026-03-06T13:44:07.699 DEBUG:teuthology.orchestra.run.vm02:> set -ex 2026-03-06T13:44:07.699 DEBUG:teuthology.orchestra.run.vm02:> sudo dd of=/usr/bin/daemon-helper 2026-03-06T13:44:07.726 DEBUG:teuthology.orchestra.run.vm02:> sudo chmod a=rx -- /usr/bin/daemon-helper 2026-03-06T13:44:07.790 DEBUG:teuthology.orchestra.run.vm04:> set -ex 2026-03-06T13:44:07.791 DEBUG:teuthology.orchestra.run.vm04:> sudo dd of=/usr/bin/daemon-helper 2026-03-06T13:44:07.818 DEBUG:teuthology.orchestra.run.vm04:> sudo chmod a=rx -- /usr/bin/daemon-helper 2026-03-06T13:44:07.880 DEBUG:teuthology.orchestra.run.vm06:> set -ex 2026-03-06T13:44:07.880 DEBUG:teuthology.orchestra.run.vm06:> sudo dd of=/usr/bin/daemon-helper 2026-03-06T13:44:07.915 DEBUG:teuthology.orchestra.run.vm06:> sudo chmod a=rx -- /usr/bin/daemon-helper 2026-03-06T13:44:07.983 INFO:teuthology.task.install.util:Shipping 'adjust-ulimits'... 2026-03-06T13:44:07.983 DEBUG:teuthology.orchestra.run.vm02:> set -ex 2026-03-06T13:44:07.983 DEBUG:teuthology.orchestra.run.vm02:> sudo dd of=/usr/bin/adjust-ulimits 2026-03-06T13:44:08.011 DEBUG:teuthology.orchestra.run.vm02:> sudo chmod a=rx -- /usr/bin/adjust-ulimits 2026-03-06T13:44:08.074 DEBUG:teuthology.orchestra.run.vm04:> set -ex 2026-03-06T13:44:08.074 DEBUG:teuthology.orchestra.run.vm04:> sudo dd of=/usr/bin/adjust-ulimits 2026-03-06T13:44:08.100 DEBUG:teuthology.orchestra.run.vm04:> sudo chmod a=rx -- /usr/bin/adjust-ulimits 2026-03-06T13:44:08.169 DEBUG:teuthology.orchestra.run.vm06:> set -ex 2026-03-06T13:44:08.169 DEBUG:teuthology.orchestra.run.vm06:> sudo dd of=/usr/bin/adjust-ulimits 2026-03-06T13:44:08.202 DEBUG:teuthology.orchestra.run.vm06:> sudo chmod a=rx -- /usr/bin/adjust-ulimits 2026-03-06T13:44:08.277 INFO:teuthology.task.install.util:Shipping 'stdin-killer'... 2026-03-06T13:44:08.277 DEBUG:teuthology.orchestra.run.vm02:> set -ex 2026-03-06T13:44:08.277 DEBUG:teuthology.orchestra.run.vm02:> sudo dd of=/usr/bin/stdin-killer 2026-03-06T13:44:08.300 DEBUG:teuthology.orchestra.run.vm02:> sudo chmod a=rx -- /usr/bin/stdin-killer 2026-03-06T13:44:08.363 DEBUG:teuthology.orchestra.run.vm04:> set -ex 2026-03-06T13:44:08.363 DEBUG:teuthology.orchestra.run.vm04:> sudo dd of=/usr/bin/stdin-killer 2026-03-06T13:44:08.391 DEBUG:teuthology.orchestra.run.vm04:> sudo chmod a=rx -- /usr/bin/stdin-killer 2026-03-06T13:44:08.458 DEBUG:teuthology.orchestra.run.vm06:> set -ex 2026-03-06T13:44:08.460 DEBUG:teuthology.orchestra.run.vm06:> sudo dd of=/usr/bin/stdin-killer 2026-03-06T13:44:08.491 DEBUG:teuthology.orchestra.run.vm06:> sudo chmod a=rx -- /usr/bin/stdin-killer 2026-03-06T13:44:08.568 INFO:teuthology.run_tasks:Running task cephadm... 2026-03-06T13:44:08.615 INFO:tasks.cephadm:Config: {'conf': {'global': {'mon election default strategy': 3}, 'mgr': {'debug mgr': 20, 'debug ms': 1, 'mgr/cephadm/use_agent': True}, 'mon': {'debug mon': 20, 'debug ms': 1, 'debug paxos': 20}, 'osd': {'debug ms': 1, 'debug osd': 20, 'osd mclock iops capacity threshold hdd': 49000}}, 'flavor': 'default', 'log-ignorelist': ['\\(MDS_ALL_DOWN\\)', '\\(MDS_UP_LESS_THAN_MAX\\)', 'MON_DOWN', 'POOL_APP_NOT_ENABLED', 'mon down', 'mons down', 'out of quorum', 'CEPHADM_FAILED_DAEMON'], 'log-only-match': ['CEPHADM_'], 'sha1': 'c24117fd5525679b799527bc1bd1f1dd0a2db5e2', 'cephadm_binary_url': 'https://download.ceph.com/rpm-19.2.3/el9/noarch/cephadm', 'containers': {'image': 'harbor.clyso.com/custom-ceph/ceph/ceph:cobaltcore-storage-v19.2.3-fasttrack-3'}} 2026-03-06T13:44:08.615 INFO:tasks.cephadm:Provided image contains tag or digest, using it as is 2026-03-06T13:44:08.615 INFO:tasks.cephadm:Cluster image is harbor.clyso.com/custom-ceph/ceph/ceph:cobaltcore-storage-v19.2.3-fasttrack-3 2026-03-06T13:44:08.615 INFO:tasks.cephadm:Cluster fsid is 2b1b4a72-195a-11f1-b1df-bf1f800d676b 2026-03-06T13:44:08.615 INFO:tasks.cephadm:Choosing monitor IPs and ports... 2026-03-06T13:44:08.615 INFO:tasks.cephadm:Monitor IPs: {'mon.a': '192.168.123.102', 'mon.b': '192.168.123.104', 'mon.c': '192.168.123.106'} 2026-03-06T13:44:08.615 INFO:tasks.cephadm:First mon is mon.a on vm02 2026-03-06T13:44:08.615 INFO:tasks.cephadm:First mgr is a 2026-03-06T13:44:08.615 INFO:tasks.cephadm:Normalizing hostnames... 2026-03-06T13:44:08.615 DEBUG:teuthology.orchestra.run.vm02:> sudo hostname $(hostname -s) 2026-03-06T13:44:08.642 DEBUG:teuthology.orchestra.run.vm04:> sudo hostname $(hostname -s) 2026-03-06T13:44:08.672 DEBUG:teuthology.orchestra.run.vm06:> sudo hostname $(hostname -s) 2026-03-06T13:44:08.706 INFO:tasks.cephadm:Downloading cephadm from url: https://download.ceph.com/rpm-19.2.3/el9/noarch/cephadm 2026-03-06T13:44:08.706 DEBUG:teuthology.orchestra.run.vm02:> curl --silent -L https://download.ceph.com/rpm-19.2.3/el9/noarch/cephadm > /home/ubuntu/cephtest/cephadm && ls -l /home/ubuntu/cephtest/cephadm 2026-03-06T13:44:09.742 INFO:teuthology.orchestra.run.vm02.stdout:-rw-r--r--. 1 ubuntu ubuntu 787672 Mar 6 13:44 /home/ubuntu/cephtest/cephadm 2026-03-06T13:44:09.742 DEBUG:teuthology.orchestra.run.vm04:> curl --silent -L https://download.ceph.com/rpm-19.2.3/el9/noarch/cephadm > /home/ubuntu/cephtest/cephadm && ls -l /home/ubuntu/cephtest/cephadm 2026-03-06T13:44:10.934 INFO:teuthology.orchestra.run.vm04.stdout:-rw-r--r--. 1 ubuntu ubuntu 787672 Mar 6 13:44 /home/ubuntu/cephtest/cephadm 2026-03-06T13:44:10.934 DEBUG:teuthology.orchestra.run.vm06:> curl --silent -L https://download.ceph.com/rpm-19.2.3/el9/noarch/cephadm > /home/ubuntu/cephtest/cephadm && ls -l /home/ubuntu/cephtest/cephadm 2026-03-06T13:44:12.041 INFO:teuthology.orchestra.run.vm06.stdout:-rw-r--r--. 1 ubuntu ubuntu 787672 Mar 6 13:44 /home/ubuntu/cephtest/cephadm 2026-03-06T13:44:12.041 DEBUG:teuthology.orchestra.run.vm02:> test -s /home/ubuntu/cephtest/cephadm && test $(stat -c%s /home/ubuntu/cephtest/cephadm) -gt 1000 && chmod +x /home/ubuntu/cephtest/cephadm 2026-03-06T13:44:12.061 DEBUG:teuthology.orchestra.run.vm04:> test -s /home/ubuntu/cephtest/cephadm && test $(stat -c%s /home/ubuntu/cephtest/cephadm) -gt 1000 && chmod +x /home/ubuntu/cephtest/cephadm 2026-03-06T13:44:12.080 DEBUG:teuthology.orchestra.run.vm06:> test -s /home/ubuntu/cephtest/cephadm && test $(stat -c%s /home/ubuntu/cephtest/cephadm) -gt 1000 && chmod +x /home/ubuntu/cephtest/cephadm 2026-03-06T13:44:12.101 INFO:tasks.cephadm:Pulling image harbor.clyso.com/custom-ceph/ceph/ceph:cobaltcore-storage-v19.2.3-fasttrack-3 on all hosts... 2026-03-06T13:44:12.101 DEBUG:teuthology.orchestra.run.vm02:> sudo /home/ubuntu/cephtest/cephadm --image harbor.clyso.com/custom-ceph/ceph/ceph:cobaltcore-storage-v19.2.3-fasttrack-3 pull 2026-03-06T13:44:12.104 DEBUG:teuthology.orchestra.run.vm04:> sudo /home/ubuntu/cephtest/cephadm --image harbor.clyso.com/custom-ceph/ceph/ceph:cobaltcore-storage-v19.2.3-fasttrack-3 pull 2026-03-06T13:44:12.121 DEBUG:teuthology.orchestra.run.vm06:> sudo /home/ubuntu/cephtest/cephadm --image harbor.clyso.com/custom-ceph/ceph/ceph:cobaltcore-storage-v19.2.3-fasttrack-3 pull 2026-03-06T13:44:12.444 INFO:teuthology.orchestra.run.vm02.stderr:Pulling container image harbor.clyso.com/custom-ceph/ceph/ceph:cobaltcore-storage-v19.2.3-fasttrack-3... 2026-03-06T13:44:12.458 INFO:teuthology.orchestra.run.vm04.stderr:Pulling container image harbor.clyso.com/custom-ceph/ceph/ceph:cobaltcore-storage-v19.2.3-fasttrack-3... 2026-03-06T13:44:12.491 INFO:teuthology.orchestra.run.vm06.stderr:Pulling container image harbor.clyso.com/custom-ceph/ceph/ceph:cobaltcore-storage-v19.2.3-fasttrack-3... 2026-03-06T13:44:38.869 INFO:teuthology.orchestra.run.vm06.stdout:{ 2026-03-06T13:44:38.869 INFO:teuthology.orchestra.run.vm06.stdout: "ceph_version": "ceph version 19.2.3-47-gc24117fd552 (c24117fd5525679b799527bc1bd1f1dd0a2db5e2) squid (stable)", 2026-03-06T13:44:38.869 INFO:teuthology.orchestra.run.vm06.stdout: "image_id": "306e97de47e91c2b4b24d3dc09be3b3a12039b078f343d91220102acc6628a68", 2026-03-06T13:44:38.869 INFO:teuthology.orchestra.run.vm06.stdout: "repo_digests": [ 2026-03-06T13:44:38.869 INFO:teuthology.orchestra.run.vm06.stdout: "harbor.clyso.com/custom-ceph/ceph/ceph@sha256:26363c7a4eea9ef5a0148afc7b2a22b6f486596d87a30c2a9fdcda5db3eca62b" 2026-03-06T13:44:38.869 INFO:teuthology.orchestra.run.vm06.stdout: ] 2026-03-06T13:44:38.869 INFO:teuthology.orchestra.run.vm06.stdout:} 2026-03-06T13:44:39.008 INFO:teuthology.orchestra.run.vm04.stdout:{ 2026-03-06T13:44:39.008 INFO:teuthology.orchestra.run.vm04.stdout: "ceph_version": "ceph version 19.2.3-47-gc24117fd552 (c24117fd5525679b799527bc1bd1f1dd0a2db5e2) squid (stable)", 2026-03-06T13:44:39.008 INFO:teuthology.orchestra.run.vm04.stdout: "image_id": "306e97de47e91c2b4b24d3dc09be3b3a12039b078f343d91220102acc6628a68", 2026-03-06T13:44:39.008 INFO:teuthology.orchestra.run.vm04.stdout: "repo_digests": [ 2026-03-06T13:44:39.008 INFO:teuthology.orchestra.run.vm04.stdout: "harbor.clyso.com/custom-ceph/ceph/ceph@sha256:26363c7a4eea9ef5a0148afc7b2a22b6f486596d87a30c2a9fdcda5db3eca62b" 2026-03-06T13:44:39.008 INFO:teuthology.orchestra.run.vm04.stdout: ] 2026-03-06T13:44:39.008 INFO:teuthology.orchestra.run.vm04.stdout:} 2026-03-06T13:44:45.799 INFO:teuthology.orchestra.run.vm02.stdout:{ 2026-03-06T13:44:45.799 INFO:teuthology.orchestra.run.vm02.stdout: "ceph_version": "ceph version 19.2.3-47-gc24117fd552 (c24117fd5525679b799527bc1bd1f1dd0a2db5e2) squid (stable)", 2026-03-06T13:44:45.799 INFO:teuthology.orchestra.run.vm02.stdout: "image_id": "306e97de47e91c2b4b24d3dc09be3b3a12039b078f343d91220102acc6628a68", 2026-03-06T13:44:45.799 INFO:teuthology.orchestra.run.vm02.stdout: "repo_digests": [ 2026-03-06T13:44:45.799 INFO:teuthology.orchestra.run.vm02.stdout: "harbor.clyso.com/custom-ceph/ceph/ceph@sha256:26363c7a4eea9ef5a0148afc7b2a22b6f486596d87a30c2a9fdcda5db3eca62b" 2026-03-06T13:44:45.799 INFO:teuthology.orchestra.run.vm02.stdout: ] 2026-03-06T13:44:45.799 INFO:teuthology.orchestra.run.vm02.stdout:} 2026-03-06T13:44:45.818 DEBUG:teuthology.orchestra.run.vm02:> sudo mkdir -p /etc/ceph 2026-03-06T13:44:45.843 DEBUG:teuthology.orchestra.run.vm04:> sudo mkdir -p /etc/ceph 2026-03-06T13:44:45.870 DEBUG:teuthology.orchestra.run.vm06:> sudo mkdir -p /etc/ceph 2026-03-06T13:44:45.897 DEBUG:teuthology.orchestra.run.vm02:> sudo chmod 777 /etc/ceph 2026-03-06T13:44:45.924 DEBUG:teuthology.orchestra.run.vm04:> sudo chmod 777 /etc/ceph 2026-03-06T13:44:45.948 DEBUG:teuthology.orchestra.run.vm06:> sudo chmod 777 /etc/ceph 2026-03-06T13:44:45.973 INFO:tasks.cephadm:Writing seed config... 2026-03-06T13:44:45.974 INFO:tasks.cephadm: override: [global] mon election default strategy = 3 2026-03-06T13:44:45.974 INFO:tasks.cephadm: override: [mgr] debug mgr = 20 2026-03-06T13:44:45.974 INFO:tasks.cephadm: override: [mgr] debug ms = 1 2026-03-06T13:44:45.974 INFO:tasks.cephadm: override: [mgr] mgr/cephadm/use_agent = True 2026-03-06T13:44:45.974 INFO:tasks.cephadm: override: [mon] debug mon = 20 2026-03-06T13:44:45.974 INFO:tasks.cephadm: override: [mon] debug ms = 1 2026-03-06T13:44:45.974 INFO:tasks.cephadm: override: [mon] debug paxos = 20 2026-03-06T13:44:45.974 INFO:tasks.cephadm: override: [osd] debug ms = 1 2026-03-06T13:44:45.974 INFO:tasks.cephadm: override: [osd] debug osd = 20 2026-03-06T13:44:45.974 INFO:tasks.cephadm: override: [osd] osd mclock iops capacity threshold hdd = 49000 2026-03-06T13:44:45.974 DEBUG:teuthology.orchestra.run.vm02:> set -ex 2026-03-06T13:44:45.974 DEBUG:teuthology.orchestra.run.vm02:> dd of=/home/ubuntu/cephtest/seed.ceph.conf 2026-03-06T13:44:45.990 DEBUG:tasks.cephadm:Final config: [global] # make logging friendly to teuthology log_to_file = true log_to_stderr = false log to journald = false mon cluster log to file = true mon cluster log file level = debug mon clock drift allowed = 1.000 # replicate across OSDs, not hosts osd crush chooseleaf type = 0 #osd pool default size = 2 osd pool default erasure code profile = plugin=jerasure technique=reed_sol_van k=2 m=1 crush-failure-domain=osd # enable some debugging auth debug = true ms die on old message = true ms die on bug = true debug asserts on shutdown = true # adjust warnings mon max pg per osd = 10000# >= luminous mon pg warn max object skew = 0 mon osd allow primary affinity = true mon osd allow pg remap = true mon warn on legacy crush tunables = false mon warn on crush straw calc version zero = false mon warn on no sortbitwise = false mon warn on osd down out interval zero = false mon warn on too few osds = false mon_warn_on_pool_pg_num_not_power_of_two = false # disable pg_autoscaler by default for new pools osd_pool_default_pg_autoscale_mode = off # tests delete pools mon allow pool delete = true fsid = 2b1b4a72-195a-11f1-b1df-bf1f800d676b mon election default strategy = 3 [osd] osd scrub load threshold = 5.0 osd scrub max interval = 600 osd mclock profile = high_recovery_ops osd recover clone overlap = true osd recovery max chunk = 1048576 osd deep scrub update digest min age = 30 osd map max advance = 10 osd memory target autotune = true # debugging osd debug shutdown = true osd debug op order = true osd debug verify stray on activate = true osd debug pg log writeout = true osd debug verify cached snaps = true osd debug verify missing on start = true osd debug misdirected ops = true osd op queue = debug_random osd op queue cut off = debug_random osd shutdown pgref assert = true bdev debug aio = true osd sloppy crc = true debug ms = 1 debug osd = 20 osd mclock iops capacity threshold hdd = 49000 [mgr] mon reweight min pgs per osd = 4 mon reweight min bytes per osd = 10 mgr/telemetry/nag = false debug mgr = 20 debug ms = 1 mgr/cephadm/use_agent = True [mon] mon data avail warn = 5 mon mgr mkfs grace = 240 mon reweight min pgs per osd = 4 mon osd reporter subtree level = osd mon osd prime pg temp = true mon reweight min bytes per osd = 10 # rotate auth tickets quickly to exercise renewal paths auth mon ticket ttl = 660# 11m auth service ticket ttl = 240# 4m # don't complain about global id reclaim mon_warn_on_insecure_global_id_reclaim = false mon_warn_on_insecure_global_id_reclaim_allowed = false debug mon = 20 debug ms = 1 debug paxos = 20 [client.rgw] rgw cache enabled = true rgw enable ops log = true rgw enable usage log = true 2026-03-06T13:44:45.990 DEBUG:teuthology.orchestra.run.vm02:mon.a> sudo journalctl -f -n 0 -u ceph-2b1b4a72-195a-11f1-b1df-bf1f800d676b@mon.a.service 2026-03-06T13:44:46.032 DEBUG:teuthology.orchestra.run.vm02:mgr.a> sudo journalctl -f -n 0 -u ceph-2b1b4a72-195a-11f1-b1df-bf1f800d676b@mgr.a.service 2026-03-06T13:44:46.074 INFO:tasks.cephadm:Bootstrapping... 2026-03-06T13:44:46.074 DEBUG:teuthology.orchestra.run.vm02:> sudo /home/ubuntu/cephtest/cephadm --image harbor.clyso.com/custom-ceph/ceph/ceph:cobaltcore-storage-v19.2.3-fasttrack-3 -v bootstrap --fsid 2b1b4a72-195a-11f1-b1df-bf1f800d676b --config /home/ubuntu/cephtest/seed.ceph.conf --output-config /etc/ceph/ceph.conf --output-keyring /etc/ceph/ceph.client.admin.keyring --output-pub-ssh-key /home/ubuntu/cephtest/ceph.pub --mon-id a --mgr-id a --orphan-initial-daemons --skip-monitoring-stack --mon-ip 192.168.123.102 --skip-admin-label && sudo chmod +r /etc/ceph/ceph.client.admin.keyring 2026-03-06T13:44:46.358 INFO:teuthology.orchestra.run.vm02.stdout:-------------------------------------------------------------------------------- 2026-03-06T13:44:46.358 INFO:teuthology.orchestra.run.vm02.stdout:cephadm ['--image', 'harbor.clyso.com/custom-ceph/ceph/ceph:cobaltcore-storage-v19.2.3-fasttrack-3', '-v', 'bootstrap', '--fsid', '2b1b4a72-195a-11f1-b1df-bf1f800d676b', '--config', '/home/ubuntu/cephtest/seed.ceph.conf', '--output-config', '/etc/ceph/ceph.conf', '--output-keyring', '/etc/ceph/ceph.client.admin.keyring', '--output-pub-ssh-key', '/home/ubuntu/cephtest/ceph.pub', '--mon-id', 'a', '--mgr-id', 'a', '--orphan-initial-daemons', '--skip-monitoring-stack', '--mon-ip', '192.168.123.102', '--skip-admin-label'] 2026-03-06T13:44:46.358 INFO:teuthology.orchestra.run.vm02.stderr:Specifying an fsid for your cluster offers no advantages and may increase the likelihood of fsid conflicts. 2026-03-06T13:44:46.358 INFO:teuthology.orchestra.run.vm02.stdout:Verifying podman|docker is present... 2026-03-06T13:44:46.376 INFO:teuthology.orchestra.run.vm02.stdout:/bin/podman: stdout 5.8.0 2026-03-06T13:44:46.376 INFO:teuthology.orchestra.run.vm02.stdout:Verifying lvm2 is present... 2026-03-06T13:44:46.376 INFO:teuthology.orchestra.run.vm02.stdout:Verifying time synchronization is in place... 2026-03-06T13:44:46.382 INFO:teuthology.orchestra.run.vm02.stdout:Non-zero exit code 1 from systemctl is-enabled chrony.service 2026-03-06T13:44:46.382 INFO:teuthology.orchestra.run.vm02.stdout:systemctl: stderr Failed to get unit file state for chrony.service: No such file or directory 2026-03-06T13:44:46.387 INFO:teuthology.orchestra.run.vm02.stdout:Non-zero exit code 3 from systemctl is-active chrony.service 2026-03-06T13:44:46.387 INFO:teuthology.orchestra.run.vm02.stdout:systemctl: stdout inactive 2026-03-06T13:44:46.394 INFO:teuthology.orchestra.run.vm02.stdout:systemctl: stdout enabled 2026-03-06T13:44:46.398 INFO:teuthology.orchestra.run.vm02.stdout:systemctl: stdout active 2026-03-06T13:44:46.399 INFO:teuthology.orchestra.run.vm02.stdout:Unit chronyd.service is enabled and running 2026-03-06T13:44:46.399 INFO:teuthology.orchestra.run.vm02.stdout:Repeating the final host check... 2026-03-06T13:44:46.418 INFO:teuthology.orchestra.run.vm02.stdout:/bin/podman: stdout 5.8.0 2026-03-06T13:44:46.418 INFO:teuthology.orchestra.run.vm02.stdout:podman (/bin/podman) version 5.8.0 is present 2026-03-06T13:44:46.418 INFO:teuthology.orchestra.run.vm02.stdout:systemctl is present 2026-03-06T13:44:46.418 INFO:teuthology.orchestra.run.vm02.stdout:lvcreate is present 2026-03-06T13:44:46.424 INFO:teuthology.orchestra.run.vm02.stdout:Non-zero exit code 1 from systemctl is-enabled chrony.service 2026-03-06T13:44:46.424 INFO:teuthology.orchestra.run.vm02.stdout:systemctl: stderr Failed to get unit file state for chrony.service: No such file or directory 2026-03-06T13:44:46.429 INFO:teuthology.orchestra.run.vm02.stdout:Non-zero exit code 3 from systemctl is-active chrony.service 2026-03-06T13:44:46.429 INFO:teuthology.orchestra.run.vm02.stdout:systemctl: stdout inactive 2026-03-06T13:44:46.434 INFO:teuthology.orchestra.run.vm02.stdout:systemctl: stdout enabled 2026-03-06T13:44:46.439 INFO:teuthology.orchestra.run.vm02.stdout:systemctl: stdout active 2026-03-06T13:44:46.439 INFO:teuthology.orchestra.run.vm02.stdout:Unit chronyd.service is enabled and running 2026-03-06T13:44:46.439 INFO:teuthology.orchestra.run.vm02.stdout:Host looks OK 2026-03-06T13:44:46.439 INFO:teuthology.orchestra.run.vm02.stdout:Cluster fsid: 2b1b4a72-195a-11f1-b1df-bf1f800d676b 2026-03-06T13:44:46.439 INFO:teuthology.orchestra.run.vm02.stdout:Acquiring lock 139818858787600 on /run/cephadm/2b1b4a72-195a-11f1-b1df-bf1f800d676b.lock 2026-03-06T13:44:46.439 INFO:teuthology.orchestra.run.vm02.stdout:Lock 139818858787600 acquired on /run/cephadm/2b1b4a72-195a-11f1-b1df-bf1f800d676b.lock 2026-03-06T13:44:46.439 INFO:teuthology.orchestra.run.vm02.stdout:Verifying IP 192.168.123.102 port 3300 ... 2026-03-06T13:44:46.440 INFO:teuthology.orchestra.run.vm02.stdout:Verifying IP 192.168.123.102 port 6789 ... 2026-03-06T13:44:46.440 INFO:teuthology.orchestra.run.vm02.stdout:Base mon IP(s) is [192.168.123.102:3300, 192.168.123.102:6789], mon addrv is [v2:192.168.123.102:3300,v1:192.168.123.102:6789] 2026-03-06T13:44:46.443 INFO:teuthology.orchestra.run.vm02.stdout:/sbin/ip: stdout default via 192.168.123.1 dev eth0 proto dhcp src 192.168.123.102 metric 100 2026-03-06T13:44:46.443 INFO:teuthology.orchestra.run.vm02.stdout:/sbin/ip: stdout 192.168.123.0/24 dev eth0 proto kernel scope link src 192.168.123.102 metric 100 2026-03-06T13:44:46.445 INFO:teuthology.orchestra.run.vm02.stdout:/sbin/ip: stdout ::1 dev lo proto kernel metric 256 pref medium 2026-03-06T13:44:46.445 INFO:teuthology.orchestra.run.vm02.stdout:/sbin/ip: stdout fe80::/64 dev eth0 proto kernel metric 1024 pref medium 2026-03-06T13:44:46.447 INFO:teuthology.orchestra.run.vm02.stdout:/sbin/ip: stdout 1: lo: mtu 65536 state UNKNOWN qlen 1000 2026-03-06T13:44:46.447 INFO:teuthology.orchestra.run.vm02.stdout:/sbin/ip: stdout inet6 ::1/128 scope host 2026-03-06T13:44:46.447 INFO:teuthology.orchestra.run.vm02.stdout:/sbin/ip: stdout valid_lft forever preferred_lft forever 2026-03-06T13:44:46.447 INFO:teuthology.orchestra.run.vm02.stdout:/sbin/ip: stdout 2: eth0: mtu 1500 state UP qlen 1000 2026-03-06T13:44:46.447 INFO:teuthology.orchestra.run.vm02.stdout:/sbin/ip: stdout inet6 fe80::5055:ff:fe00:2/64 scope link noprefixroute 2026-03-06T13:44:46.447 INFO:teuthology.orchestra.run.vm02.stdout:/sbin/ip: stdout valid_lft forever preferred_lft forever 2026-03-06T13:44:46.448 INFO:teuthology.orchestra.run.vm02.stdout:Mon IP `192.168.123.102` is in CIDR network `192.168.123.0/24` 2026-03-06T13:44:46.448 INFO:teuthology.orchestra.run.vm02.stdout:Mon IP `192.168.123.102` is in CIDR network `192.168.123.0/24` 2026-03-06T13:44:46.448 INFO:teuthology.orchestra.run.vm02.stdout:Inferred mon public CIDR from local network configuration ['192.168.123.0/24', '192.168.123.0/24'] 2026-03-06T13:44:46.448 INFO:teuthology.orchestra.run.vm02.stdout:Internal network (--cluster-network) has not been provided, OSD replication will default to the public_network 2026-03-06T13:44:46.448 INFO:teuthology.orchestra.run.vm02.stdout:Pulling container image harbor.clyso.com/custom-ceph/ceph/ceph:cobaltcore-storage-v19.2.3-fasttrack-3... 2026-03-06T13:44:47.200 INFO:teuthology.orchestra.run.vm02.stdout:/bin/podman: stdout 306e97de47e91c2b4b24d3dc09be3b3a12039b078f343d91220102acc6628a68 2026-03-06T13:44:47.200 INFO:teuthology.orchestra.run.vm02.stdout:/bin/podman: stderr Trying to pull harbor.clyso.com/custom-ceph/ceph/ceph:cobaltcore-storage-v19.2.3-fasttrack-3... 2026-03-06T13:44:47.200 INFO:teuthology.orchestra.run.vm02.stdout:/bin/podman: stderr Getting image source signatures 2026-03-06T13:44:47.200 INFO:teuthology.orchestra.run.vm02.stdout:/bin/podman: stderr Copying blob sha256:d21d4233fd3d4dd2f376e5ef084c47891c860682c1de15a9c0357cea5defbc91 2026-03-06T13:44:47.200 INFO:teuthology.orchestra.run.vm02.stdout:/bin/podman: stderr Copying config sha256:306e97de47e91c2b4b24d3dc09be3b3a12039b078f343d91220102acc6628a68 2026-03-06T13:44:47.200 INFO:teuthology.orchestra.run.vm02.stdout:/bin/podman: stderr Writing manifest to image destination 2026-03-06T13:44:47.560 INFO:teuthology.orchestra.run.vm02.stdout:ceph: stdout ceph version 19.2.3-47-gc24117fd552 (c24117fd5525679b799527bc1bd1f1dd0a2db5e2) squid (stable) 2026-03-06T13:44:47.560 INFO:teuthology.orchestra.run.vm02.stdout:Ceph version: ceph version 19.2.3-47-gc24117fd552 (c24117fd5525679b799527bc1bd1f1dd0a2db5e2) squid (stable) 2026-03-06T13:44:47.560 INFO:teuthology.orchestra.run.vm02.stdout:Extracting ceph user uid/gid from container image... 2026-03-06T13:44:47.763 INFO:teuthology.orchestra.run.vm02.stdout:stat: stdout 167 167 2026-03-06T13:44:47.763 INFO:teuthology.orchestra.run.vm02.stdout:Creating initial keys... 2026-03-06T13:44:47.978 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-authtool: stdout AQA/zKppT5z7MhAAl+IsN4KYf+BosmNkDC3mTg== 2026-03-06T13:44:48.201 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-authtool: stdout AQBAzKppAR2cAxAAzxGMcjjeosOhSB+eE7M4uA== 2026-03-06T13:44:48.412 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-authtool: stdout AQBAzKppXbUzERAAbo6b49sEuenVEK1AIf3n4Q== 2026-03-06T13:44:48.412 INFO:teuthology.orchestra.run.vm02.stdout:Creating initial monmap... 2026-03-06T13:44:48.648 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/monmaptool: stdout /usr/bin/monmaptool: monmap file /tmp/monmap 2026-03-06T13:44:48.648 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/monmaptool: stdout setting min_mon_release = quincy 2026-03-06T13:44:48.648 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/monmaptool: stdout /usr/bin/monmaptool: set fsid to 2b1b4a72-195a-11f1-b1df-bf1f800d676b 2026-03-06T13:44:48.648 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/monmaptool: stdout /usr/bin/monmaptool: writing epoch 0 to /tmp/monmap (1 monitors) 2026-03-06T13:44:48.648 INFO:teuthology.orchestra.run.vm02.stdout:monmaptool for a [v2:192.168.123.102:3300,v1:192.168.123.102:6789] on /usr/bin/monmaptool: monmap file /tmp/monmap 2026-03-06T13:44:48.648 INFO:teuthology.orchestra.run.vm02.stdout:setting min_mon_release = quincy 2026-03-06T13:44:48.648 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/monmaptool: set fsid to 2b1b4a72-195a-11f1-b1df-bf1f800d676b 2026-03-06T13:44:48.648 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/monmaptool: writing epoch 0 to /tmp/monmap (1 monitors) 2026-03-06T13:44:48.648 INFO:teuthology.orchestra.run.vm02.stdout: 2026-03-06T13:44:48.648 INFO:teuthology.orchestra.run.vm02.stdout:Creating mon... 2026-03-06T13:44:48.893 INFO:teuthology.orchestra.run.vm02.stdout:create mon.a on 2026-03-06T13:44:49.037 INFO:teuthology.orchestra.run.vm02.stdout:systemctl: stderr Removed "/etc/systemd/system/multi-user.target.wants/ceph.target". 2026-03-06T13:44:49.152 INFO:teuthology.orchestra.run.vm02.stdout:systemctl: stderr Created symlink /etc/systemd/system/multi-user.target.wants/ceph.target → /etc/systemd/system/ceph.target. 2026-03-06T13:44:49.274 INFO:teuthology.orchestra.run.vm02.stdout:systemctl: stderr Created symlink /etc/systemd/system/multi-user.target.wants/ceph-2b1b4a72-195a-11f1-b1df-bf1f800d676b.target → /etc/systemd/system/ceph-2b1b4a72-195a-11f1-b1df-bf1f800d676b.target. 2026-03-06T13:44:49.274 INFO:teuthology.orchestra.run.vm02.stdout:systemctl: stderr Created symlink /etc/systemd/system/ceph.target.wants/ceph-2b1b4a72-195a-11f1-b1df-bf1f800d676b.target → /etc/systemd/system/ceph-2b1b4a72-195a-11f1-b1df-bf1f800d676b.target. 2026-03-06T13:44:49.416 INFO:teuthology.orchestra.run.vm02.stdout:Non-zero exit code 1 from systemctl reset-failed ceph-2b1b4a72-195a-11f1-b1df-bf1f800d676b@mon.a 2026-03-06T13:44:49.417 INFO:teuthology.orchestra.run.vm02.stdout:systemctl: stderr Failed to reset failed state of unit ceph-2b1b4a72-195a-11f1-b1df-bf1f800d676b@mon.a.service: Unit ceph-2b1b4a72-195a-11f1-b1df-bf1f800d676b@mon.a.service not loaded. 2026-03-06T13:44:49.542 INFO:teuthology.orchestra.run.vm02.stdout:systemctl: stderr Created symlink /etc/systemd/system/ceph-2b1b4a72-195a-11f1-b1df-bf1f800d676b.target.wants/ceph-2b1b4a72-195a-11f1-b1df-bf1f800d676b@mon.a.service → /etc/systemd/system/ceph-2b1b4a72-195a-11f1-b1df-bf1f800d676b@.service. 2026-03-06T13:44:49.683 INFO:teuthology.orchestra.run.vm02.stdout:firewalld does not appear to be present 2026-03-06T13:44:49.683 INFO:teuthology.orchestra.run.vm02.stdout:Not possible to enable service . firewalld.service is not available 2026-03-06T13:44:49.683 INFO:teuthology.orchestra.run.vm02.stdout:Waiting for mon to start... 2026-03-06T13:44:49.683 INFO:teuthology.orchestra.run.vm02.stdout:Waiting for mon... 2026-03-06T13:44:50.069 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout cluster: 2026-03-06T13:44:50.069 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout id: 2b1b4a72-195a-11f1-b1df-bf1f800d676b 2026-03-06T13:44:50.069 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout health: HEALTH_OK 2026-03-06T13:44:50.069 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout 2026-03-06T13:44:50.069 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout services: 2026-03-06T13:44:50.069 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout mon: 1 daemons, quorum a (age 0.225049s) 2026-03-06T13:44:50.069 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout mgr: no daemons active 2026-03-06T13:44:50.069 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout osd: 0 osds: 0 up, 0 in 2026-03-06T13:44:50.069 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout 2026-03-06T13:44:50.069 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout data: 2026-03-06T13:44:50.069 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout pools: 0 pools, 0 pgs 2026-03-06T13:44:50.069 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout objects: 0 objects, 0 B 2026-03-06T13:44:50.069 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout usage: 0 B used, 0 B / 0 B avail 2026-03-06T13:44:50.069 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout pgs: 2026-03-06T13:44:50.069 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout 2026-03-06T13:44:50.069 INFO:teuthology.orchestra.run.vm02.stdout:mon is available 2026-03-06T13:44:50.069 INFO:teuthology.orchestra.run.vm02.stdout:Assimilating anything we can from ceph.conf... 2026-03-06T13:44:50.463 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout 2026-03-06T13:44:50.463 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout [global] 2026-03-06T13:44:50.463 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout fsid = 2b1b4a72-195a-11f1-b1df-bf1f800d676b 2026-03-06T13:44:50.463 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout mon_cluster_log_file_level = debug 2026-03-06T13:44:50.463 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout mon_host = [v2:192.168.123.102:3300,v1:192.168.123.102:6789] 2026-03-06T13:44:50.463 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout mon_osd_allow_pg_remap = true 2026-03-06T13:44:50.463 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout mon_osd_allow_primary_affinity = true 2026-03-06T13:44:50.463 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout mon_warn_on_no_sortbitwise = false 2026-03-06T13:44:50.463 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout osd_crush_chooseleaf_type = 0 2026-03-06T13:44:50.463 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout 2026-03-06T13:44:50.463 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout [mgr] 2026-03-06T13:44:50.463 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout mgr/cephadm/use_agent = True 2026-03-06T13:44:50.463 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout mgr/telemetry/nag = false 2026-03-06T13:44:50.463 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout 2026-03-06T13:44:50.463 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout [osd] 2026-03-06T13:44:50.463 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout osd_map_max_advance = 10 2026-03-06T13:44:50.463 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout osd_sloppy_crc = true 2026-03-06T13:44:50.463 INFO:teuthology.orchestra.run.vm02.stdout:Generating new minimal ceph.conf... 2026-03-06T13:44:50.844 INFO:teuthology.orchestra.run.vm02.stdout:Restarting the monitor... 2026-03-06T13:44:51.102 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:44:50 vm02 ceph-2b1b4a72-195a-11f1-b1df-bf1f800d676b-mon-a[48796]: 2026-03-06T12:44:50.914+0000 7f680c813640 -1 mon.a@0(leader) e1 *** Got Signal Terminated *** 2026-03-06T13:44:51.362 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:44:51 vm02 podman[49093]: 2026-03-06 13:44:51.100487001 +0100 CET m=+0.198590545 container died 31819b49a49fab5d7dd7812dbb79d80536438f1c9d9e50a00f4026462a5ea970 (image=harbor.clyso.com/custom-ceph/ceph/ceph:cobaltcore-storage-v19.2.3-fasttrack-3, name=ceph-2b1b4a72-195a-11f1-b1df-bf1f800d676b-mon-a, CEPH_GIT_REPO=https://github.com/irq0/ceph.git, CEPH_REF=19.2.3-47-gc24117fd552, CEPH_SHA1=c24117fd5525679b799527bc1bd1f1dd0a2db5e2, FROM_IMAGE=rockylinux:9, GANESHA_REPO_BASEURL=https://buildlogs.centos.org/centos/$releasever-stream/storage/$basearch/nfsganesha-5/, OSD_FLAVOR=default, ceph=True, io.buildah.version=1.41.8) 2026-03-06T13:44:51.362 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:44:51 vm02 podman[49093]: 2026-03-06 13:44:51.217165691 +0100 CET m=+0.315269244 container remove 31819b49a49fab5d7dd7812dbb79d80536438f1c9d9e50a00f4026462a5ea970 (image=harbor.clyso.com/custom-ceph/ceph/ceph:cobaltcore-storage-v19.2.3-fasttrack-3, name=ceph-2b1b4a72-195a-11f1-b1df-bf1f800d676b-mon-a, ceph=True, io.buildah.version=1.41.8, CEPH_GIT_REPO=https://github.com/irq0/ceph.git, CEPH_REF=19.2.3-47-gc24117fd552, CEPH_SHA1=c24117fd5525679b799527bc1bd1f1dd0a2db5e2, FROM_IMAGE=rockylinux:9, GANESHA_REPO_BASEURL=https://buildlogs.centos.org/centos/$releasever-stream/storage/$basearch/nfsganesha-5/, OSD_FLAVOR=default) 2026-03-06T13:44:51.362 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:44:51 vm02 bash[49093]: ceph-2b1b4a72-195a-11f1-b1df-bf1f800d676b-mon-a 2026-03-06T13:44:51.362 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:44:51 vm02 systemd[1]: ceph-2b1b4a72-195a-11f1-b1df-bf1f800d676b@mon.a.service: Deactivated successfully. 2026-03-06T13:44:51.362 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:44:51 vm02 systemd[1]: Stopped Ceph mon.a for 2b1b4a72-195a-11f1-b1df-bf1f800d676b. 2026-03-06T13:44:51.362 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:44:51 vm02 systemd[1]: Starting Ceph mon.a for 2b1b4a72-195a-11f1-b1df-bf1f800d676b... 2026-03-06T13:44:51.362 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:44:51 vm02 podman[49174]: 2026-03-06 13:44:51.344599214 +0100 CET m=+0.013780574 container create 617d29b525ba3df76e8b6376fe6d32fa2c275cad6a2b8cec18b81a4f3993eee6 (image=harbor.clyso.com/custom-ceph/ceph/ceph:cobaltcore-storage-v19.2.3-fasttrack-3, name=ceph-2b1b4a72-195a-11f1-b1df-bf1f800d676b-mon-a, io.buildah.version=1.41.8, CEPH_GIT_REPO=https://github.com/irq0/ceph.git, CEPH_REF=19.2.3-47-gc24117fd552, CEPH_SHA1=c24117fd5525679b799527bc1bd1f1dd0a2db5e2, FROM_IMAGE=rockylinux:9, GANESHA_REPO_BASEURL=https://buildlogs.centos.org/centos/$releasever-stream/storage/$basearch/nfsganesha-5/, OSD_FLAVOR=default, ceph=True) 2026-03-06T13:44:51.388 INFO:teuthology.orchestra.run.vm02.stdout:Setting public_network to 192.168.123.0/24 in mon config section 2026-03-06T13:44:51.657 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:44:51 vm02 podman[49174]: 2026-03-06 13:44:51.376371139 +0100 CET m=+0.045552490 container init 617d29b525ba3df76e8b6376fe6d32fa2c275cad6a2b8cec18b81a4f3993eee6 (image=harbor.clyso.com/custom-ceph/ceph/ceph:cobaltcore-storage-v19.2.3-fasttrack-3, name=ceph-2b1b4a72-195a-11f1-b1df-bf1f800d676b-mon-a, OSD_FLAVOR=default, ceph=True, io.buildah.version=1.41.8, CEPH_GIT_REPO=https://github.com/irq0/ceph.git, CEPH_REF=19.2.3-47-gc24117fd552, CEPH_SHA1=c24117fd5525679b799527bc1bd1f1dd0a2db5e2, FROM_IMAGE=rockylinux:9, GANESHA_REPO_BASEURL=https://buildlogs.centos.org/centos/$releasever-stream/storage/$basearch/nfsganesha-5/) 2026-03-06T13:44:51.657 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:44:51 vm02 podman[49174]: 2026-03-06 13:44:51.380914663 +0100 CET m=+0.050096014 container start 617d29b525ba3df76e8b6376fe6d32fa2c275cad6a2b8cec18b81a4f3993eee6 (image=harbor.clyso.com/custom-ceph/ceph/ceph:cobaltcore-storage-v19.2.3-fasttrack-3, name=ceph-2b1b4a72-195a-11f1-b1df-bf1f800d676b-mon-a, io.buildah.version=1.41.8, CEPH_GIT_REPO=https://github.com/irq0/ceph.git, CEPH_REF=19.2.3-47-gc24117fd552, CEPH_SHA1=c24117fd5525679b799527bc1bd1f1dd0a2db5e2, FROM_IMAGE=rockylinux:9, GANESHA_REPO_BASEURL=https://buildlogs.centos.org/centos/$releasever-stream/storage/$basearch/nfsganesha-5/, OSD_FLAVOR=default, ceph=True) 2026-03-06T13:44:51.657 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:44:51 vm02 bash[49174]: 617d29b525ba3df76e8b6376fe6d32fa2c275cad6a2b8cec18b81a4f3993eee6 2026-03-06T13:44:51.657 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:44:51 vm02 podman[49174]: 2026-03-06 13:44:51.339072069 +0100 CET m=+0.008253429 image pull 306e97de47e91c2b4b24d3dc09be3b3a12039b078f343d91220102acc6628a68 harbor.clyso.com/custom-ceph/ceph/ceph:cobaltcore-storage-v19.2.3-fasttrack-3 2026-03-06T13:44:51.657 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:44:51 vm02 systemd[1]: Started Ceph mon.a for 2b1b4a72-195a-11f1-b1df-bf1f800d676b. 2026-03-06T13:44:51.657 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:44:51 vm02 ceph-mon[49207]: set uid:gid to 167:167 (ceph:ceph) 2026-03-06T13:44:51.658 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:44:51 vm02 ceph-mon[49207]: ceph version 19.2.3-47-gc24117fd552 (c24117fd5525679b799527bc1bd1f1dd0a2db5e2) squid (stable), process ceph-mon, pid 6 2026-03-06T13:44:51.658 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:44:51 vm02 ceph-mon[49207]: pidfile_write: ignore empty --pid-file 2026-03-06T13:44:51.658 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:44:51 vm02 ceph-mon[49207]: load: jerasure load: lrc 2026-03-06T13:44:51.658 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:44:51 vm02 ceph-mon[49207]: rocksdb: RocksDB version: 7.9.2 2026-03-06T13:44:51.658 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:44:51 vm02 ceph-mon[49207]: rocksdb: Git sha 0 2026-03-06T13:44:51.658 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:44:51 vm02 ceph-mon[49207]: rocksdb: Compile date 2026-03-03 21:08:28 2026-03-06T13:44:51.658 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:44:51 vm02 ceph-mon[49207]: rocksdb: DB SUMMARY 2026-03-06T13:44:51.658 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:44:51 vm02 ceph-mon[49207]: rocksdb: DB Session ID: WH4C4R8S6PTVEO2FL9OP 2026-03-06T13:44:51.658 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:44:51 vm02 ceph-mon[49207]: rocksdb: CURRENT file: CURRENT 2026-03-06T13:44:51.658 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:44:51 vm02 ceph-mon[49207]: rocksdb: IDENTITY file: IDENTITY 2026-03-06T13:44:51.658 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:44:51 vm02 ceph-mon[49207]: rocksdb: MANIFEST file: MANIFEST-000010 size: 179 Bytes 2026-03-06T13:44:51.658 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:44:51 vm02 ceph-mon[49207]: rocksdb: SST files in /var/lib/ceph/mon/ceph-a/store.db dir, Total Num: 1, files: 000008.sst 2026-03-06T13:44:51.658 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:44:51 vm02 ceph-mon[49207]: rocksdb: Write Ahead Log file in /var/lib/ceph/mon/ceph-a/store.db: 000009.log size: 88153 ; 2026-03-06T13:44:51.658 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:44:51 vm02 ceph-mon[49207]: rocksdb: Options.error_if_exists: 0 2026-03-06T13:44:51.658 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:44:51 vm02 ceph-mon[49207]: rocksdb: Options.create_if_missing: 0 2026-03-06T13:44:51.658 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:44:51 vm02 ceph-mon[49207]: rocksdb: Options.paranoid_checks: 1 2026-03-06T13:44:51.658 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:44:51 vm02 ceph-mon[49207]: rocksdb: Options.flush_verify_memtable_count: 1 2026-03-06T13:44:51.658 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:44:51 vm02 ceph-mon[49207]: rocksdb: Options.track_and_verify_wals_in_manifest: 0 2026-03-06T13:44:51.658 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:44:51 vm02 ceph-mon[49207]: rocksdb: Options.verify_sst_unique_id_in_manifest: 1 2026-03-06T13:44:51.658 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:44:51 vm02 ceph-mon[49207]: rocksdb: Options.env: 0x55d543bbcca0 2026-03-06T13:44:51.658 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:44:51 vm02 ceph-mon[49207]: rocksdb: Options.fs: PosixFileSystem 2026-03-06T13:44:51.658 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:44:51 vm02 ceph-mon[49207]: rocksdb: Options.info_log: 0x55d545ba3700 2026-03-06T13:44:51.658 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:44:51 vm02 ceph-mon[49207]: rocksdb: Options.max_file_opening_threads: 16 2026-03-06T13:44:51.658 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:44:51 vm02 ceph-mon[49207]: rocksdb: Options.statistics: (nil) 2026-03-06T13:44:51.658 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:44:51 vm02 ceph-mon[49207]: rocksdb: Options.use_fsync: 0 2026-03-06T13:44:51.658 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:44:51 vm02 ceph-mon[49207]: rocksdb: Options.max_log_file_size: 0 2026-03-06T13:44:51.658 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:44:51 vm02 ceph-mon[49207]: rocksdb: Options.max_manifest_file_size: 1073741824 2026-03-06T13:44:51.658 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:44:51 vm02 ceph-mon[49207]: rocksdb: Options.log_file_time_to_roll: 0 2026-03-06T13:44:51.658 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:44:51 vm02 ceph-mon[49207]: rocksdb: Options.keep_log_file_num: 1000 2026-03-06T13:44:51.658 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:44:51 vm02 ceph-mon[49207]: rocksdb: Options.recycle_log_file_num: 0 2026-03-06T13:44:51.658 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:44:51 vm02 ceph-mon[49207]: rocksdb: Options.allow_fallocate: 1 2026-03-06T13:44:51.658 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:44:51 vm02 ceph-mon[49207]: rocksdb: Options.allow_mmap_reads: 0 2026-03-06T13:44:51.658 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:44:51 vm02 ceph-mon[49207]: rocksdb: Options.allow_mmap_writes: 0 2026-03-06T13:44:51.658 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:44:51 vm02 ceph-mon[49207]: rocksdb: Options.use_direct_reads: 0 2026-03-06T13:44:51.658 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:44:51 vm02 ceph-mon[49207]: rocksdb: Options.use_direct_io_for_flush_and_compaction: 0 2026-03-06T13:44:51.658 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:44:51 vm02 ceph-mon[49207]: rocksdb: Options.create_missing_column_families: 0 2026-03-06T13:44:51.658 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:44:51 vm02 ceph-mon[49207]: rocksdb: Options.db_log_dir: 2026-03-06T13:44:51.658 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:44:51 vm02 ceph-mon[49207]: rocksdb: Options.wal_dir: 2026-03-06T13:44:51.658 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:44:51 vm02 ceph-mon[49207]: rocksdb: Options.table_cache_numshardbits: 6 2026-03-06T13:44:51.658 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:44:51 vm02 ceph-mon[49207]: rocksdb: Options.WAL_ttl_seconds: 0 2026-03-06T13:44:51.658 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:44:51 vm02 ceph-mon[49207]: rocksdb: Options.WAL_size_limit_MB: 0 2026-03-06T13:44:51.658 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:44:51 vm02 ceph-mon[49207]: rocksdb: Options.max_write_batch_group_size_bytes: 1048576 2026-03-06T13:44:51.658 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:44:51 vm02 ceph-mon[49207]: rocksdb: Options.manifest_preallocation_size: 4194304 2026-03-06T13:44:51.658 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:44:51 vm02 ceph-mon[49207]: rocksdb: Options.is_fd_close_on_exec: 1 2026-03-06T13:44:51.658 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:44:51 vm02 ceph-mon[49207]: rocksdb: Options.advise_random_on_open: 1 2026-03-06T13:44:51.658 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:44:51 vm02 ceph-mon[49207]: rocksdb: Options.db_write_buffer_size: 0 2026-03-06T13:44:51.658 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:44:51 vm02 ceph-mon[49207]: rocksdb: Options.write_buffer_manager: 0x55d545ba7900 2026-03-06T13:44:51.658 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:44:51 vm02 ceph-mon[49207]: rocksdb: Options.access_hint_on_compaction_start: 1 2026-03-06T13:44:51.658 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:44:51 vm02 ceph-mon[49207]: rocksdb: Options.random_access_max_buffer_size: 1048576 2026-03-06T13:44:51.658 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:44:51 vm02 ceph-mon[49207]: rocksdb: Options.use_adaptive_mutex: 0 2026-03-06T13:44:51.658 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:44:51 vm02 ceph-mon[49207]: rocksdb: Options.rate_limiter: (nil) 2026-03-06T13:44:51.658 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:44:51 vm02 ceph-mon[49207]: rocksdb: Options.sst_file_manager.rate_bytes_per_sec: 0 2026-03-06T13:44:51.658 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:44:51 vm02 ceph-mon[49207]: rocksdb: Options.wal_recovery_mode: 2 2026-03-06T13:44:51.658 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:44:51 vm02 ceph-mon[49207]: rocksdb: Options.enable_thread_tracking: 0 2026-03-06T13:44:51.658 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:44:51 vm02 ceph-mon[49207]: rocksdb: Options.enable_pipelined_write: 0 2026-03-06T13:44:51.658 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:44:51 vm02 ceph-mon[49207]: rocksdb: Options.unordered_write: 0 2026-03-06T13:44:51.658 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:44:51 vm02 ceph-mon[49207]: rocksdb: Options.allow_concurrent_memtable_write: 1 2026-03-06T13:44:51.658 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:44:51 vm02 ceph-mon[49207]: rocksdb: Options.enable_write_thread_adaptive_yield: 1 2026-03-06T13:44:51.658 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:44:51 vm02 ceph-mon[49207]: rocksdb: Options.write_thread_max_yield_usec: 100 2026-03-06T13:44:51.658 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:44:51 vm02 ceph-mon[49207]: rocksdb: Options.write_thread_slow_yield_usec: 3 2026-03-06T13:44:51.658 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:44:51 vm02 ceph-mon[49207]: rocksdb: Options.row_cache: None 2026-03-06T13:44:51.658 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:44:51 vm02 ceph-mon[49207]: rocksdb: Options.wal_filter: None 2026-03-06T13:44:51.658 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:44:51 vm02 ceph-mon[49207]: rocksdb: Options.avoid_flush_during_recovery: 0 2026-03-06T13:44:51.658 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:44:51 vm02 ceph-mon[49207]: rocksdb: Options.allow_ingest_behind: 0 2026-03-06T13:44:51.659 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:44:51 vm02 ceph-mon[49207]: rocksdb: Options.two_write_queues: 0 2026-03-06T13:44:51.659 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:44:51 vm02 ceph-mon[49207]: rocksdb: Options.manual_wal_flush: 0 2026-03-06T13:44:51.659 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:44:51 vm02 ceph-mon[49207]: rocksdb: Options.wal_compression: 0 2026-03-06T13:44:51.659 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:44:51 vm02 ceph-mon[49207]: rocksdb: Options.atomic_flush: 0 2026-03-06T13:44:51.659 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:44:51 vm02 ceph-mon[49207]: rocksdb: Options.avoid_unnecessary_blocking_io: 0 2026-03-06T13:44:51.659 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:44:51 vm02 ceph-mon[49207]: rocksdb: Options.persist_stats_to_disk: 0 2026-03-06T13:44:51.659 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:44:51 vm02 ceph-mon[49207]: rocksdb: Options.write_dbid_to_manifest: 0 2026-03-06T13:44:51.659 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:44:51 vm02 ceph-mon[49207]: rocksdb: Options.log_readahead_size: 0 2026-03-06T13:44:51.659 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:44:51 vm02 ceph-mon[49207]: rocksdb: Options.file_checksum_gen_factory: Unknown 2026-03-06T13:44:51.659 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:44:51 vm02 ceph-mon[49207]: rocksdb: Options.best_efforts_recovery: 0 2026-03-06T13:44:51.659 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:44:51 vm02 ceph-mon[49207]: rocksdb: Options.max_bgerror_resume_count: 2147483647 2026-03-06T13:44:51.659 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:44:51 vm02 ceph-mon[49207]: rocksdb: Options.bgerror_resume_retry_interval: 1000000 2026-03-06T13:44:51.659 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:44:51 vm02 ceph-mon[49207]: rocksdb: Options.allow_data_in_errors: 0 2026-03-06T13:44:51.659 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:44:51 vm02 ceph-mon[49207]: rocksdb: Options.db_host_id: __hostname__ 2026-03-06T13:44:51.659 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:44:51 vm02 ceph-mon[49207]: rocksdb: Options.enforce_single_del_contracts: true 2026-03-06T13:44:51.659 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:44:51 vm02 ceph-mon[49207]: rocksdb: Options.max_background_jobs: 2 2026-03-06T13:44:51.659 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:44:51 vm02 ceph-mon[49207]: rocksdb: Options.max_background_compactions: -1 2026-03-06T13:44:51.659 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:44:51 vm02 ceph-mon[49207]: rocksdb: Options.max_subcompactions: 1 2026-03-06T13:44:51.659 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:44:51 vm02 ceph-mon[49207]: rocksdb: Options.avoid_flush_during_shutdown: 0 2026-03-06T13:44:51.659 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:44:51 vm02 ceph-mon[49207]: rocksdb: Options.writable_file_max_buffer_size: 1048576 2026-03-06T13:44:51.659 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:44:51 vm02 ceph-mon[49207]: rocksdb: Options.delayed_write_rate : 16777216 2026-03-06T13:44:51.659 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:44:51 vm02 ceph-mon[49207]: rocksdb: Options.max_total_wal_size: 0 2026-03-06T13:44:51.659 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:44:51 vm02 ceph-mon[49207]: rocksdb: Options.delete_obsolete_files_period_micros: 21600000000 2026-03-06T13:44:51.659 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:44:51 vm02 ceph-mon[49207]: rocksdb: Options.stats_dump_period_sec: 600 2026-03-06T13:44:51.659 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:44:51 vm02 ceph-mon[49207]: rocksdb: Options.stats_persist_period_sec: 600 2026-03-06T13:44:51.659 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:44:51 vm02 ceph-mon[49207]: rocksdb: Options.stats_history_buffer_size: 1048576 2026-03-06T13:44:51.659 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:44:51 vm02 ceph-mon[49207]: rocksdb: Options.max_open_files: -1 2026-03-06T13:44:51.659 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:44:51 vm02 ceph-mon[49207]: rocksdb: Options.bytes_per_sync: 0 2026-03-06T13:44:51.659 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:44:51 vm02 ceph-mon[49207]: rocksdb: Options.wal_bytes_per_sync: 0 2026-03-06T13:44:51.659 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:44:51 vm02 ceph-mon[49207]: rocksdb: Options.strict_bytes_per_sync: 0 2026-03-06T13:44:51.659 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:44:51 vm02 ceph-mon[49207]: rocksdb: Options.compaction_readahead_size: 0 2026-03-06T13:44:51.659 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:44:51 vm02 ceph-mon[49207]: rocksdb: Options.max_background_flushes: -1 2026-03-06T13:44:51.659 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:44:51 vm02 ceph-mon[49207]: rocksdb: Compression algorithms supported: 2026-03-06T13:44:51.659 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:44:51 vm02 ceph-mon[49207]: rocksdb: kZSTD supported: 0 2026-03-06T13:44:51.659 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:44:51 vm02 ceph-mon[49207]: rocksdb: kXpressCompression supported: 0 2026-03-06T13:44:51.659 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:44:51 vm02 ceph-mon[49207]: rocksdb: kBZip2Compression supported: 0 2026-03-06T13:44:51.659 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:44:51 vm02 ceph-mon[49207]: rocksdb: kZSTDNotFinalCompression supported: 0 2026-03-06T13:44:51.659 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:44:51 vm02 ceph-mon[49207]: rocksdb: kLZ4Compression supported: 1 2026-03-06T13:44:51.659 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:44:51 vm02 ceph-mon[49207]: rocksdb: kZlibCompression supported: 1 2026-03-06T13:44:51.659 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:44:51 vm02 ceph-mon[49207]: rocksdb: kLZ4HCCompression supported: 1 2026-03-06T13:44:51.659 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:44:51 vm02 ceph-mon[49207]: rocksdb: kSnappyCompression supported: 1 2026-03-06T13:44:51.659 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:44:51 vm02 ceph-mon[49207]: rocksdb: Fast CRC32 supported: Supported on x86 2026-03-06T13:44:51.659 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:44:51 vm02 ceph-mon[49207]: rocksdb: DMutex implementation: pthread_mutex_t 2026-03-06T13:44:51.659 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:44:51 vm02 ceph-mon[49207]: rocksdb: [db/version_set.cc:5527] Recovering from manifest file: /var/lib/ceph/mon/ceph-a/store.db/MANIFEST-000010 2026-03-06T13:44:51.659 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:44:51 vm02 ceph-mon[49207]: rocksdb: [db/column_family.cc:630] --------------- Options for column family [default]: 2026-03-06T13:44:51.659 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:44:51 vm02 ceph-mon[49207]: rocksdb: Options.comparator: leveldb.BytewiseComparator 2026-03-06T13:44:51.659 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:44:51 vm02 ceph-mon[49207]: rocksdb: Options.merge_operator: 2026-03-06T13:44:51.659 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:44:51 vm02 ceph-mon[49207]: rocksdb: Options.compaction_filter: None 2026-03-06T13:44:51.659 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:44:51 vm02 ceph-mon[49207]: rocksdb: Options.compaction_filter_factory: None 2026-03-06T13:44:51.659 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:44:51 vm02 ceph-mon[49207]: rocksdb: Options.sst_partitioner_factory: None 2026-03-06T13:44:51.659 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:44:51 vm02 ceph-mon[49207]: rocksdb: Options.memtable_factory: SkipListFactory 2026-03-06T13:44:51.659 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:44:51 vm02 ceph-mon[49207]: rocksdb: Options.table_factory: BlockBasedTable 2026-03-06T13:44:51.659 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:44:51 vm02 ceph-mon[49207]: rocksdb: table_factory options: flush_block_policy_factory: FlushBlockBySizePolicyFactory (0x55d545b5df40) 2026-03-06T13:44:51.659 INFO:journalctl@ceph.mon.a.vm02.stdout: cache_index_and_filter_blocks: 1 2026-03-06T13:44:51.659 INFO:journalctl@ceph.mon.a.vm02.stdout: cache_index_and_filter_blocks_with_high_priority: 0 2026-03-06T13:44:51.659 INFO:journalctl@ceph.mon.a.vm02.stdout: pin_l0_filter_and_index_blocks_in_cache: 0 2026-03-06T13:44:51.659 INFO:journalctl@ceph.mon.a.vm02.stdout: pin_top_level_index_and_filter: 1 2026-03-06T13:44:51.659 INFO:journalctl@ceph.mon.a.vm02.stdout: index_type: 0 2026-03-06T13:44:51.659 INFO:journalctl@ceph.mon.a.vm02.stdout: data_block_index_type: 0 2026-03-06T13:44:51.659 INFO:journalctl@ceph.mon.a.vm02.stdout: index_shortening: 1 2026-03-06T13:44:51.659 INFO:journalctl@ceph.mon.a.vm02.stdout: data_block_hash_table_util_ratio: 0.750000 2026-03-06T13:44:51.659 INFO:journalctl@ceph.mon.a.vm02.stdout: checksum: 4 2026-03-06T13:44:51.659 INFO:journalctl@ceph.mon.a.vm02.stdout: no_block_cache: 0 2026-03-06T13:44:51.659 INFO:journalctl@ceph.mon.a.vm02.stdout: block_cache: 0x55d545bc71f0 2026-03-06T13:44:51.659 INFO:journalctl@ceph.mon.a.vm02.stdout: block_cache_name: BinnedLRUCache 2026-03-06T13:44:51.659 INFO:journalctl@ceph.mon.a.vm02.stdout: block_cache_options: 2026-03-06T13:44:51.659 INFO:journalctl@ceph.mon.a.vm02.stdout: capacity : 536870912 2026-03-06T13:44:51.660 INFO:journalctl@ceph.mon.a.vm02.stdout: num_shard_bits : 4 2026-03-06T13:44:51.660 INFO:journalctl@ceph.mon.a.vm02.stdout: strict_capacity_limit : 0 2026-03-06T13:44:51.660 INFO:journalctl@ceph.mon.a.vm02.stdout: high_pri_pool_ratio: 0.000 2026-03-06T13:44:51.660 INFO:journalctl@ceph.mon.a.vm02.stdout: block_cache_compressed: (nil) 2026-03-06T13:44:51.660 INFO:journalctl@ceph.mon.a.vm02.stdout: persistent_cache: (nil) 2026-03-06T13:44:51.660 INFO:journalctl@ceph.mon.a.vm02.stdout: block_size: 4096 2026-03-06T13:44:51.660 INFO:journalctl@ceph.mon.a.vm02.stdout: block_size_deviation: 10 2026-03-06T13:44:51.660 INFO:journalctl@ceph.mon.a.vm02.stdout: block_restart_interval: 16 2026-03-06T13:44:51.660 INFO:journalctl@ceph.mon.a.vm02.stdout: index_block_restart_interval: 1 2026-03-06T13:44:51.660 INFO:journalctl@ceph.mon.a.vm02.stdout: metadata_block_size: 4096 2026-03-06T13:44:51.660 INFO:journalctl@ceph.mon.a.vm02.stdout: partition_filters: 0 2026-03-06T13:44:51.660 INFO:journalctl@ceph.mon.a.vm02.stdout: use_delta_encoding: 1 2026-03-06T13:44:51.660 INFO:journalctl@ceph.mon.a.vm02.stdout: filter_policy: bloomfilter 2026-03-06T13:44:51.660 INFO:journalctl@ceph.mon.a.vm02.stdout: whole_key_filtering: 1 2026-03-06T13:44:51.660 INFO:journalctl@ceph.mon.a.vm02.stdout: verify_compression: 0 2026-03-06T13:44:51.660 INFO:journalctl@ceph.mon.a.vm02.stdout: read_amp_bytes_per_bit: 0 2026-03-06T13:44:51.660 INFO:journalctl@ceph.mon.a.vm02.stdout: format_version: 5 2026-03-06T13:44:51.660 INFO:journalctl@ceph.mon.a.vm02.stdout: enable_index_compression: 1 2026-03-06T13:44:51.660 INFO:journalctl@ceph.mon.a.vm02.stdout: block_align: 0 2026-03-06T13:44:51.660 INFO:journalctl@ceph.mon.a.vm02.stdout: max_auto_readahead_size: 262144 2026-03-06T13:44:51.660 INFO:journalctl@ceph.mon.a.vm02.stdout: prepopulate_block_cache: 0 2026-03-06T13:44:51.660 INFO:journalctl@ceph.mon.a.vm02.stdout: initial_auto_readahead_size: 8192 2026-03-06T13:44:51.660 INFO:journalctl@ceph.mon.a.vm02.stdout: num_file_reads_for_auto_readahead: 2 2026-03-06T13:44:51.660 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:44:51 vm02 ceph-mon[49207]: rocksdb: Options.write_buffer_size: 33554432 2026-03-06T13:44:51.660 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:44:51 vm02 ceph-mon[49207]: rocksdb: Options.max_write_buffer_number: 2 2026-03-06T13:44:51.660 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:44:51 vm02 ceph-mon[49207]: rocksdb: Options.compression: NoCompression 2026-03-06T13:44:51.660 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:44:51 vm02 ceph-mon[49207]: rocksdb: Options.bottommost_compression: Disabled 2026-03-06T13:44:51.660 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:44:51 vm02 ceph-mon[49207]: rocksdb: Options.prefix_extractor: nullptr 2026-03-06T13:44:51.660 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:44:51 vm02 ceph-mon[49207]: rocksdb: Options.memtable_insert_with_hint_prefix_extractor: nullptr 2026-03-06T13:44:51.660 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:44:51 vm02 ceph-mon[49207]: rocksdb: Options.num_levels: 7 2026-03-06T13:44:51.660 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:44:51 vm02 ceph-mon[49207]: rocksdb: Options.min_write_buffer_number_to_merge: 1 2026-03-06T13:44:51.660 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:44:51 vm02 ceph-mon[49207]: rocksdb: Options.max_write_buffer_number_to_maintain: 0 2026-03-06T13:44:51.660 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:44:51 vm02 ceph-mon[49207]: rocksdb: Options.max_write_buffer_size_to_maintain: 0 2026-03-06T13:44:51.660 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:44:51 vm02 ceph-mon[49207]: rocksdb: Options.bottommost_compression_opts.window_bits: -14 2026-03-06T13:44:51.660 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:44:51 vm02 ceph-mon[49207]: rocksdb: Options.bottommost_compression_opts.level: 32767 2026-03-06T13:44:51.660 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:44:51 vm02 ceph-mon[49207]: rocksdb: Options.bottommost_compression_opts.strategy: 0 2026-03-06T13:44:51.660 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:44:51 vm02 ceph-mon[49207]: rocksdb: Options.bottommost_compression_opts.max_dict_bytes: 0 2026-03-06T13:44:51.660 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:44:51 vm02 ceph-mon[49207]: rocksdb: Options.bottommost_compression_opts.zstd_max_train_bytes: 0 2026-03-06T13:44:51.660 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:44:51 vm02 ceph-mon[49207]: rocksdb: Options.bottommost_compression_opts.parallel_threads: 1 2026-03-06T13:44:51.660 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:44:51 vm02 ceph-mon[49207]: rocksdb: Options.bottommost_compression_opts.enabled: false 2026-03-06T13:44:51.660 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:44:51 vm02 ceph-mon[49207]: rocksdb: Options.bottommost_compression_opts.max_dict_buffer_bytes: 0 2026-03-06T13:44:51.660 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:44:51 vm02 ceph-mon[49207]: rocksdb: Options.bottommost_compression_opts.use_zstd_dict_trainer: true 2026-03-06T13:44:51.660 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:44:51 vm02 ceph-mon[49207]: rocksdb: Options.compression_opts.window_bits: -14 2026-03-06T13:44:51.660 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:44:51 vm02 ceph-mon[49207]: rocksdb: Options.compression_opts.level: 32767 2026-03-06T13:44:51.660 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:44:51 vm02 ceph-mon[49207]: rocksdb: Options.compression_opts.strategy: 0 2026-03-06T13:44:51.660 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:44:51 vm02 ceph-mon[49207]: rocksdb: Options.compression_opts.max_dict_bytes: 0 2026-03-06T13:44:51.660 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:44:51 vm02 ceph-mon[49207]: rocksdb: Options.compression_opts.zstd_max_train_bytes: 0 2026-03-06T13:44:51.660 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:44:51 vm02 ceph-mon[49207]: rocksdb: Options.compression_opts.use_zstd_dict_trainer: true 2026-03-06T13:44:51.660 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:44:51 vm02 ceph-mon[49207]: rocksdb: Options.compression_opts.parallel_threads: 1 2026-03-06T13:44:51.660 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:44:51 vm02 ceph-mon[49207]: rocksdb: Options.compression_opts.enabled: false 2026-03-06T13:44:51.660 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:44:51 vm02 ceph-mon[49207]: rocksdb: Options.compression_opts.max_dict_buffer_bytes: 0 2026-03-06T13:44:51.660 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:44:51 vm02 ceph-mon[49207]: rocksdb: Options.level0_file_num_compaction_trigger: 4 2026-03-06T13:44:51.660 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:44:51 vm02 ceph-mon[49207]: rocksdb: Options.level0_slowdown_writes_trigger: 20 2026-03-06T13:44:51.660 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:44:51 vm02 ceph-mon[49207]: rocksdb: Options.level0_stop_writes_trigger: 36 2026-03-06T13:44:51.660 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:44:51 vm02 ceph-mon[49207]: rocksdb: Options.target_file_size_base: 67108864 2026-03-06T13:44:51.660 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:44:51 vm02 ceph-mon[49207]: rocksdb: Options.target_file_size_multiplier: 1 2026-03-06T13:44:51.660 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:44:51 vm02 ceph-mon[49207]: rocksdb: Options.max_bytes_for_level_base: 268435456 2026-03-06T13:44:51.660 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:44:51 vm02 ceph-mon[49207]: rocksdb: Options.level_compaction_dynamic_level_bytes: 1 2026-03-06T13:44:51.660 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:44:51 vm02 ceph-mon[49207]: rocksdb: Options.max_bytes_for_level_multiplier: 10.000000 2026-03-06T13:44:51.660 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:44:51 vm02 ceph-mon[49207]: rocksdb: Options.max_bytes_for_level_multiplier_addtl[0]: 1 2026-03-06T13:44:51.660 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:44:51 vm02 ceph-mon[49207]: rocksdb: Options.max_bytes_for_level_multiplier_addtl[1]: 1 2026-03-06T13:44:51.660 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:44:51 vm02 ceph-mon[49207]: rocksdb: Options.max_bytes_for_level_multiplier_addtl[2]: 1 2026-03-06T13:44:51.660 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:44:51 vm02 ceph-mon[49207]: rocksdb: Options.max_bytes_for_level_multiplier_addtl[3]: 1 2026-03-06T13:44:51.660 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:44:51 vm02 ceph-mon[49207]: rocksdb: Options.max_bytes_for_level_multiplier_addtl[4]: 1 2026-03-06T13:44:51.660 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:44:51 vm02 ceph-mon[49207]: rocksdb: Options.max_bytes_for_level_multiplier_addtl[5]: 1 2026-03-06T13:44:51.660 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:44:51 vm02 ceph-mon[49207]: rocksdb: Options.max_bytes_for_level_multiplier_addtl[6]: 1 2026-03-06T13:44:51.660 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:44:51 vm02 ceph-mon[49207]: rocksdb: Options.max_sequential_skip_in_iterations: 8 2026-03-06T13:44:51.661 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:44:51 vm02 ceph-mon[49207]: rocksdb: Options.max_compaction_bytes: 1677721600 2026-03-06T13:44:51.661 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:44:51 vm02 ceph-mon[49207]: rocksdb: Options.ignore_max_compaction_bytes_for_input: true 2026-03-06T13:44:51.661 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:44:51 vm02 ceph-mon[49207]: rocksdb: Options.arena_block_size: 1048576 2026-03-06T13:44:51.661 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:44:51 vm02 ceph-mon[49207]: rocksdb: Options.soft_pending_compaction_bytes_limit: 68719476736 2026-03-06T13:44:51.661 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:44:51 vm02 ceph-mon[49207]: rocksdb: Options.hard_pending_compaction_bytes_limit: 274877906944 2026-03-06T13:44:51.661 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:44:51 vm02 ceph-mon[49207]: rocksdb: Options.disable_auto_compactions: 0 2026-03-06T13:44:51.661 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:44:51 vm02 ceph-mon[49207]: rocksdb: Options.compaction_style: kCompactionStyleLevel 2026-03-06T13:44:51.661 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:44:51 vm02 ceph-mon[49207]: rocksdb: Options.compaction_pri: kMinOverlappingRatio 2026-03-06T13:44:51.661 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:44:51 vm02 ceph-mon[49207]: rocksdb: Options.compaction_options_universal.size_ratio: 1 2026-03-06T13:44:51.661 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:44:51 vm02 ceph-mon[49207]: rocksdb: Options.compaction_options_universal.min_merge_width: 2 2026-03-06T13:44:51.661 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:44:51 vm02 ceph-mon[49207]: rocksdb: Options.compaction_options_universal.max_merge_width: 4294967295 2026-03-06T13:44:51.661 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:44:51 vm02 ceph-mon[49207]: rocksdb: Options.compaction_options_universal.max_size_amplification_percent: 200 2026-03-06T13:44:51.661 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:44:51 vm02 ceph-mon[49207]: rocksdb: Options.compaction_options_universal.compression_size_percent: -1 2026-03-06T13:44:51.661 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:44:51 vm02 ceph-mon[49207]: rocksdb: Options.compaction_options_universal.stop_style: kCompactionStopStyleTotalSize 2026-03-06T13:44:51.661 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:44:51 vm02 ceph-mon[49207]: rocksdb: Options.compaction_options_fifo.max_table_files_size: 1073741824 2026-03-06T13:44:51.661 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:44:51 vm02 ceph-mon[49207]: rocksdb: Options.compaction_options_fifo.allow_compaction: 0 2026-03-06T13:44:51.661 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:44:51 vm02 ceph-mon[49207]: rocksdb: Options.table_properties_collectors: CompactOnDeletionCollector (Sliding window size = 32768 Deletion trigger = 16384 Deletion ratio = 0); 2026-03-06T13:44:51.661 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:44:51 vm02 ceph-mon[49207]: rocksdb: Options.inplace_update_support: 0 2026-03-06T13:44:51.661 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:44:51 vm02 ceph-mon[49207]: rocksdb: Options.inplace_update_num_locks: 10000 2026-03-06T13:44:51.661 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:44:51 vm02 ceph-mon[49207]: rocksdb: Options.memtable_prefix_bloom_size_ratio: 0.000000 2026-03-06T13:44:51.661 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:44:51 vm02 ceph-mon[49207]: rocksdb: Options.memtable_whole_key_filtering: 0 2026-03-06T13:44:51.661 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:44:51 vm02 ceph-mon[49207]: rocksdb: Options.memtable_huge_page_size: 0 2026-03-06T13:44:51.661 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:44:51 vm02 ceph-mon[49207]: rocksdb: Options.bloom_locality: 0 2026-03-06T13:44:51.661 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:44:51 vm02 ceph-mon[49207]: rocksdb: Options.max_successive_merges: 0 2026-03-06T13:44:51.661 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:44:51 vm02 ceph-mon[49207]: rocksdb: Options.optimize_filters_for_hits: 0 2026-03-06T13:44:51.661 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:44:51 vm02 ceph-mon[49207]: rocksdb: Options.paranoid_file_checks: 0 2026-03-06T13:44:51.661 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:44:51 vm02 ceph-mon[49207]: rocksdb: Options.force_consistency_checks: 1 2026-03-06T13:44:51.661 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:44:51 vm02 ceph-mon[49207]: rocksdb: Options.report_bg_io_stats: 0 2026-03-06T13:44:51.661 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:44:51 vm02 ceph-mon[49207]: rocksdb: Options.ttl: 2592000 2026-03-06T13:44:51.661 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:44:51 vm02 ceph-mon[49207]: rocksdb: Options.periodic_compaction_seconds: 0 2026-03-06T13:44:51.661 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:44:51 vm02 ceph-mon[49207]: rocksdb: Options.preclude_last_level_data_seconds: 0 2026-03-06T13:44:51.661 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:44:51 vm02 ceph-mon[49207]: rocksdb: Options.preserve_internal_time_seconds: 0 2026-03-06T13:44:51.661 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:44:51 vm02 ceph-mon[49207]: rocksdb: Options.enable_blob_files: false 2026-03-06T13:44:51.661 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:44:51 vm02 ceph-mon[49207]: rocksdb: Options.min_blob_size: 0 2026-03-06T13:44:51.661 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:44:51 vm02 ceph-mon[49207]: rocksdb: Options.blob_file_size: 268435456 2026-03-06T13:44:51.661 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:44:51 vm02 ceph-mon[49207]: rocksdb: Options.blob_compression_type: NoCompression 2026-03-06T13:44:51.661 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:44:51 vm02 ceph-mon[49207]: rocksdb: Options.enable_blob_garbage_collection: false 2026-03-06T13:44:51.661 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:44:51 vm02 ceph-mon[49207]: rocksdb: Options.blob_garbage_collection_age_cutoff: 0.250000 2026-03-06T13:44:51.661 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:44:51 vm02 ceph-mon[49207]: rocksdb: Options.blob_garbage_collection_force_threshold: 1.000000 2026-03-06T13:44:51.661 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:44:51 vm02 ceph-mon[49207]: rocksdb: Options.blob_compaction_readahead_size: 0 2026-03-06T13:44:51.661 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:44:51 vm02 ceph-mon[49207]: rocksdb: Options.blob_file_starting_level: 0 2026-03-06T13:44:51.661 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:44:51 vm02 ceph-mon[49207]: rocksdb: Options.experimental_mempurge_threshold: 0.000000 2026-03-06T13:44:51.661 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:44:51 vm02 ceph-mon[49207]: rocksdb: [db/version_set.cc:5566] Recovered from manifest file:/var/lib/ceph/mon/ceph-a/store.db/MANIFEST-000010 succeeded,manifest_file_number is 10, next_file_number is 12, last_sequence is 5, log_number is 5,prev_log_number is 0,max_column_family is 0,min_log_number_to_keep is 5 2026-03-06T13:44:51.661 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:44:51 vm02 ceph-mon[49207]: rocksdb: [db/version_set.cc:5581] Column family [default] (ID 0), log number is 5 2026-03-06T13:44:51.661 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:44:51 vm02 ceph-mon[49207]: rocksdb: [db/db_impl/db_impl_open.cc:539] DB ID: 50d0409a-a440-4dc9-92a7-26d9b406b438 2026-03-06T13:44:51.661 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:44:51 vm02 ceph-mon[49207]: rocksdb: EVENT_LOG_v1 {"time_micros": 1772801091407625, "job": 1, "event": "recovery_started", "wal_files": [9]} 2026-03-06T13:44:51.661 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:44:51 vm02 ceph-mon[49207]: rocksdb: [db/db_impl/db_impl_open.cc:1043] Recovering log #9 mode 2 2026-03-06T13:44:51.661 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:44:51 vm02 ceph-mon[49207]: rocksdb: EVENT_LOG_v1 {"time_micros": 1772801091409451, "cf_name": "default", "job": 1, "event": "table_file_creation", "file_number": 13, "file_size": 85131, "file_checksum": "", "file_checksum_func_name": "Unknown", "smallest_seqno": 8, "largest_seqno": 246, "table_properties": {"data_size": 83283, "index_size": 237, "index_partitions": 0, "top_level_index_size": 0, "index_key_is_user_key": 1, "index_value_is_delta_encoded": 1, "filter_size": 581, "raw_key_size": 10161, "raw_average_key_size": 47, "raw_value_size": 77452, "raw_average_value_size": 363, "num_data_blocks": 11, "num_entries": 213, "num_filter_entries": 213, "num_deletions": 3, "num_merge_operands": 0, "num_range_deletions": 0, "format_version": 0, "fixed_key_len": 0, "filter_policy": "bloomfilter", "column_family_name": "default", "column_family_id": 0, "comparator": "leveldb.BytewiseComparator", "merge_operator": "", "prefix_extractor_name": "nullptr", "property_collectors": "[CompactOnDeletionCollector]", "compression": "NoCompression", "compression_options": "window_bits=-14; level=32767; strategy=0; max_dict_bytes=0; zstd_max_train_bytes=0; enabled=0; max_dict_buffer_bytes=0; use_zstd_dict_trainer=1; ", "creation_time": 1772801091, "oldest_key_time": 0, "file_creation_time": 0, "slow_compression_estimated_data_size": 0, "fast_compression_estimated_data_size": 0, "db_id": "50d0409a-a440-4dc9-92a7-26d9b406b438", "db_session_id": "WH4C4R8S6PTVEO2FL9OP", "orig_file_number": 13, "seqno_to_time_mapping": "N/A"}} 2026-03-06T13:44:51.661 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:44:51 vm02 ceph-mon[49207]: rocksdb: EVENT_LOG_v1 {"time_micros": 1772801091409501, "job": 1, "event": "recovery_finished"} 2026-03-06T13:44:51.661 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:44:51 vm02 ceph-mon[49207]: rocksdb: [db/version_set.cc:5047] Creating manifest 15 2026-03-06T13:44:51.661 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:44:51 vm02 ceph-mon[49207]: rocksdb: [file/delete_scheduler.cc:74] Deleted file /var/lib/ceph/mon/ceph-a/store.db/000009.log immediately, rate_bytes_per_sec 0, total_trash_size 0 max_trash_db_ratio 0.250000 2026-03-06T13:44:51.661 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:44:51 vm02 ceph-mon[49207]: rocksdb: [db/db_impl/db_impl_open.cc:1987] SstFileManager instance 0x55d545bc8e00 2026-03-06T13:44:51.661 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:44:51 vm02 ceph-mon[49207]: rocksdb: DB pointer 0x55d545cd4000 2026-03-06T13:44:51.661 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:44:51 vm02 ceph-mon[49207]: starting mon.a rank 0 at public addrs [v2:192.168.123.102:3300/0,v1:192.168.123.102:6789/0] at bind addrs [v2:192.168.123.102:3300/0,v1:192.168.123.102:6789/0] mon_data /var/lib/ceph/mon/ceph-a fsid 2b1b4a72-195a-11f1-b1df-bf1f800d676b 2026-03-06T13:44:51.661 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:44:51 vm02 ceph-mon[49207]: mon.a@-1(???) e1 preinit fsid 2b1b4a72-195a-11f1-b1df-bf1f800d676b 2026-03-06T13:44:51.661 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:44:51 vm02 ceph-mon[49207]: mon.a@-1(???).mds e1 new map 2026-03-06T13:44:51.661 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:44:51 vm02 ceph-mon[49207]: mon.a@-1(???).mds e1 print_map 2026-03-06T13:44:51.661 INFO:journalctl@ceph.mon.a.vm02.stdout: e1 2026-03-06T13:44:51.661 INFO:journalctl@ceph.mon.a.vm02.stdout: btime 2026-03-06T12:44:49:710370+0000 2026-03-06T13:44:51.661 INFO:journalctl@ceph.mon.a.vm02.stdout: enable_multiple, ever_enabled_multiple: 1,1 2026-03-06T13:44:51.661 INFO:journalctl@ceph.mon.a.vm02.stdout: default compat: compat={},rocompat={},incompat={1=base v0.20,2=client writeable ranges,3=default file layouts on dirs,4=dir inode in separate object,5=mds uses versioned encoding,6=dirfrag is stored in omap,8=no anchor table,9=file layout v2,10=snaprealm v2,11=minor log segments,12=quiesce subvolumes} 2026-03-06T13:44:51.661 INFO:journalctl@ceph.mon.a.vm02.stdout: legacy client fscid: -1 2026-03-06T13:44:51.661 INFO:journalctl@ceph.mon.a.vm02.stdout: 2026-03-06T13:44:51.661 INFO:journalctl@ceph.mon.a.vm02.stdout: No filesystems configured 2026-03-06T13:44:51.661 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:44:51 vm02 ceph-mon[49207]: mon.a@-1(???).osd e1 crush map has features 3314932999778484224, adjusting msgr requires 2026-03-06T13:44:51.661 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:44:51 vm02 ceph-mon[49207]: mon.a@-1(???).osd e1 crush map has features 288514050185494528, adjusting msgr requires 2026-03-06T13:44:51.662 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:44:51 vm02 ceph-mon[49207]: mon.a@-1(???).osd e1 crush map has features 288514050185494528, adjusting msgr requires 2026-03-06T13:44:51.662 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:44:51 vm02 ceph-mon[49207]: mon.a@-1(???).osd e1 crush map has features 288514050185494528, adjusting msgr requires 2026-03-06T13:44:51.662 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:44:51 vm02 ceph-mon[49207]: mon.a@-1(???).paxosservice(auth 1..2) refresh upgraded, format 0 -> 3 2026-03-06T13:44:51.662 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:44:51 vm02 ceph-mon[49207]: rocksdb: [db/db_impl/db_impl.cc:1109] ------- DUMPING STATS ------- 2026-03-06T13:44:51.662 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:44:51 vm02 ceph-mon[49207]: rocksdb: [db/db_impl/db_impl.cc:1111] 2026-03-06T13:44:51.662 INFO:journalctl@ceph.mon.a.vm02.stdout: ** DB Stats ** 2026-03-06T13:44:51.662 INFO:journalctl@ceph.mon.a.vm02.stdout: Uptime(secs): 0.0 total, 0.0 interval 2026-03-06T13:44:51.662 INFO:journalctl@ceph.mon.a.vm02.stdout: Cumulative writes: 0 writes, 0 keys, 0 commit groups, 0.0 writes per commit group, ingest: 0.00 GB, 0.00 MB/s 2026-03-06T13:44:51.662 INFO:journalctl@ceph.mon.a.vm02.stdout: Cumulative WAL: 0 writes, 0 syncs, 0.00 writes per sync, written: 0.00 GB, 0.00 MB/s 2026-03-06T13:44:51.662 INFO:journalctl@ceph.mon.a.vm02.stdout: Cumulative stall: 00:00:0.000 H:M:S, 0.0 percent 2026-03-06T13:44:51.662 INFO:journalctl@ceph.mon.a.vm02.stdout: Interval writes: 0 writes, 0 keys, 0 commit groups, 0.0 writes per commit group, ingest: 0.00 MB, 0.00 MB/s 2026-03-06T13:44:51.662 INFO:journalctl@ceph.mon.a.vm02.stdout: Interval WAL: 0 writes, 0 syncs, 0.00 writes per sync, written: 0.00 GB, 0.00 MB/s 2026-03-06T13:44:51.662 INFO:journalctl@ceph.mon.a.vm02.stdout: Interval stall: 00:00:0.000 H:M:S, 0.0 percent 2026-03-06T13:44:51.662 INFO:journalctl@ceph.mon.a.vm02.stdout: 2026-03-06T13:44:51.662 INFO:journalctl@ceph.mon.a.vm02.stdout: ** Compaction Stats [default] ** 2026-03-06T13:44:51.662 INFO:journalctl@ceph.mon.a.vm02.stdout: Level Files Size Score Read(GB) Rn(GB) Rnp1(GB) Write(GB) Wnew(GB) Moved(GB) W-Amp Rd(MB/s) Wr(MB/s) Comp(sec) CompMergeCPU(sec) Comp(cnt) Avg(sec) KeyIn KeyDrop Rblob(GB) Wblob(GB) 2026-03-06T13:44:51.662 INFO:journalctl@ceph.mon.a.vm02.stdout: ------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------ 2026-03-06T13:44:51.662 INFO:journalctl@ceph.mon.a.vm02.stdout: L0 2/0 84.99 KB 0.5 0.0 0.0 0.0 0.0 0.0 0.0 1.0 0.0 53.2 0.00 0.00 1 0.002 0 0 0.0 0.0 2026-03-06T13:44:51.662 INFO:journalctl@ceph.mon.a.vm02.stdout: Sum 2/0 84.99 KB 0.0 0.0 0.0 0.0 0.0 0.0 0.0 1.0 0.0 53.2 0.00 0.00 1 0.002 0 0 0.0 0.0 2026-03-06T13:44:51.662 INFO:journalctl@ceph.mon.a.vm02.stdout: Int 0/0 0.00 KB 0.0 0.0 0.0 0.0 0.0 0.0 0.0 1.0 0.0 53.2 0.00 0.00 1 0.002 0 0 0.0 0.0 2026-03-06T13:44:51.662 INFO:journalctl@ceph.mon.a.vm02.stdout: 2026-03-06T13:44:51.662 INFO:journalctl@ceph.mon.a.vm02.stdout: ** Compaction Stats [default] ** 2026-03-06T13:44:51.662 INFO:journalctl@ceph.mon.a.vm02.stdout: Priority Files Size Score Read(GB) Rn(GB) Rnp1(GB) Write(GB) Wnew(GB) Moved(GB) W-Amp Rd(MB/s) Wr(MB/s) Comp(sec) CompMergeCPU(sec) Comp(cnt) Avg(sec) KeyIn KeyDrop Rblob(GB) Wblob(GB) 2026-03-06T13:44:51.662 INFO:journalctl@ceph.mon.a.vm02.stdout: --------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------- 2026-03-06T13:44:51.662 INFO:journalctl@ceph.mon.a.vm02.stdout: User 0/0 0.00 KB 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0.0 53.2 0.00 0.00 1 0.002 0 0 0.0 0.0 2026-03-06T13:44:51.662 INFO:journalctl@ceph.mon.a.vm02.stdout: 2026-03-06T13:44:51.662 INFO:journalctl@ceph.mon.a.vm02.stdout: Blob file count: 0, total size: 0.0 GB, garbage size: 0.0 GB, space amp: 0.0 2026-03-06T13:44:51.662 INFO:journalctl@ceph.mon.a.vm02.stdout: 2026-03-06T13:44:51.662 INFO:journalctl@ceph.mon.a.vm02.stdout: Uptime(secs): 0.0 total, 0.0 interval 2026-03-06T13:44:51.662 INFO:journalctl@ceph.mon.a.vm02.stdout: Flush(GB): cumulative 0.000, interval 0.000 2026-03-06T13:44:51.662 INFO:journalctl@ceph.mon.a.vm02.stdout: AddFile(GB): cumulative 0.000, interval 0.000 2026-03-06T13:44:51.662 INFO:journalctl@ceph.mon.a.vm02.stdout: AddFile(Total Files): cumulative 0, interval 0 2026-03-06T13:44:51.662 INFO:journalctl@ceph.mon.a.vm02.stdout: AddFile(L0 Files): cumulative 0, interval 0 2026-03-06T13:44:51.662 INFO:journalctl@ceph.mon.a.vm02.stdout: AddFile(Keys): cumulative 0, interval 0 2026-03-06T13:44:51.662 INFO:journalctl@ceph.mon.a.vm02.stdout: Cumulative compaction: 0.00 GB write, 11.18 MB/s write, 0.00 GB read, 0.00 MB/s read, 0.0 seconds 2026-03-06T13:44:51.662 INFO:journalctl@ceph.mon.a.vm02.stdout: Interval compaction: 0.00 GB write, 11.18 MB/s write, 0.00 GB read, 0.00 MB/s read, 0.0 seconds 2026-03-06T13:44:51.662 INFO:journalctl@ceph.mon.a.vm02.stdout: Stalls(count): 0 level0_slowdown, 0 level0_slowdown_with_compaction, 0 level0_numfiles, 0 level0_numfiles_with_compaction, 0 stop for pending_compaction_bytes, 0 slowdown for pending_compaction_bytes, 0 memtable_compaction, 0 memtable_slowdown, interval 0 total count 2026-03-06T13:44:51.662 INFO:journalctl@ceph.mon.a.vm02.stdout: Block cache BinnedLRUCache@0x55d545bc71f0#6 capacity: 512.00 MB usage: 47.83 KB table_size: 0 occupancy: 18446744073709551615 collections: 1 last_copies: 0 last_secs: 0.001127 secs_since: 0 2026-03-06T13:44:51.662 INFO:journalctl@ceph.mon.a.vm02.stdout: Block cache entry stats(count,size,portion): DataBlock(8,46.64 KB,0.00889599%) FilterBlock(2,0.77 KB,0.000146031%) IndexBlock(2,0.42 KB,8.04663e-05%) Misc(1,0.00 KB,0%) 2026-03-06T13:44:51.662 INFO:journalctl@ceph.mon.a.vm02.stdout: 2026-03-06T13:44:51.662 INFO:journalctl@ceph.mon.a.vm02.stdout: ** File Read Latency Histogram By Level [default] ** 2026-03-06T13:44:51.662 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:44:51 vm02 ceph-mon[49207]: mon.a is new leader, mons a in quorum (ranks 0) 2026-03-06T13:44:51.662 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:44:51 vm02 ceph-mon[49207]: monmap epoch 1 2026-03-06T13:44:51.662 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:44:51 vm02 ceph-mon[49207]: fsid 2b1b4a72-195a-11f1-b1df-bf1f800d676b 2026-03-06T13:44:51.662 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:44:51 vm02 ceph-mon[49207]: last_changed 2026-03-06T12:44:48.508617+0000 2026-03-06T13:44:51.662 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:44:51 vm02 ceph-mon[49207]: created 2026-03-06T12:44:48.508617+0000 2026-03-06T13:44:51.662 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:44:51 vm02 ceph-mon[49207]: min_mon_release 19 (squid) 2026-03-06T13:44:51.662 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:44:51 vm02 ceph-mon[49207]: election_strategy: 1 2026-03-06T13:44:51.662 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:44:51 vm02 ceph-mon[49207]: 0: [v2:192.168.123.102:3300/0,v1:192.168.123.102:6789/0] mon.a 2026-03-06T13:44:51.662 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:44:51 vm02 ceph-mon[49207]: fsmap 2026-03-06T13:44:51.662 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:44:51 vm02 ceph-mon[49207]: osdmap e1: 0 total, 0 up, 0 in 2026-03-06T13:44:51.662 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:44:51 vm02 ceph-mon[49207]: mgrmap e1: no daemons active 2026-03-06T13:44:51.782 INFO:teuthology.orchestra.run.vm02.stdout:Wrote config to /etc/ceph/ceph.conf 2026-03-06T13:44:51.784 INFO:teuthology.orchestra.run.vm02.stdout:Wrote keyring to /etc/ceph/ceph.client.admin.keyring 2026-03-06T13:44:51.784 INFO:teuthology.orchestra.run.vm02.stdout:Creating mgr... 2026-03-06T13:44:51.784 INFO:teuthology.orchestra.run.vm02.stdout:Verifying port 0.0.0.0:9283 ... 2026-03-06T13:44:51.784 INFO:teuthology.orchestra.run.vm02.stdout:Verifying port 0.0.0.0:8765 ... 2026-03-06T13:44:51.923 INFO:teuthology.orchestra.run.vm02.stdout:Non-zero exit code 1 from systemctl reset-failed ceph-2b1b4a72-195a-11f1-b1df-bf1f800d676b@mgr.a 2026-03-06T13:44:51.923 INFO:teuthology.orchestra.run.vm02.stdout:systemctl: stderr Failed to reset failed state of unit ceph-2b1b4a72-195a-11f1-b1df-bf1f800d676b@mgr.a.service: Unit ceph-2b1b4a72-195a-11f1-b1df-bf1f800d676b@mgr.a.service not loaded. 2026-03-06T13:44:52.043 INFO:teuthology.orchestra.run.vm02.stdout:systemctl: stderr Created symlink /etc/systemd/system/ceph-2b1b4a72-195a-11f1-b1df-bf1f800d676b.target.wants/ceph-2b1b4a72-195a-11f1-b1df-bf1f800d676b@mgr.a.service → /etc/systemd/system/ceph-2b1b4a72-195a-11f1-b1df-bf1f800d676b@.service. 2026-03-06T13:44:52.189 INFO:journalctl@ceph.mgr.a.vm02.stdout:Mar 06 13:44:52 vm02 podman[49429]: 2026-03-06 13:44:52.155848381 +0100 CET m=+0.031035379 container create 6b7b3227c2471ec9cc79a4394e408b9ae7046c8aa6026fcc1dcbf32f8c16a854 (image=harbor.clyso.com/custom-ceph/ceph/ceph:cobaltcore-storage-v19.2.3-fasttrack-3, name=ceph-2b1b4a72-195a-11f1-b1df-bf1f800d676b-mgr-a, ceph=True, io.buildah.version=1.41.8, CEPH_GIT_REPO=https://github.com/irq0/ceph.git, CEPH_REF=19.2.3-47-gc24117fd552, CEPH_SHA1=c24117fd5525679b799527bc1bd1f1dd0a2db5e2, FROM_IMAGE=rockylinux:9, GANESHA_REPO_BASEURL=https://buildlogs.centos.org/centos/$releasever-stream/storage/$basearch/nfsganesha-5/, OSD_FLAVOR=default) 2026-03-06T13:44:52.215 INFO:teuthology.orchestra.run.vm02.stdout:firewalld does not appear to be present 2026-03-06T13:44:52.215 INFO:teuthology.orchestra.run.vm02.stdout:Not possible to enable service . firewalld.service is not available 2026-03-06T13:44:52.215 INFO:teuthology.orchestra.run.vm02.stdout:firewalld does not appear to be present 2026-03-06T13:44:52.215 INFO:teuthology.orchestra.run.vm02.stdout:Not possible to open ports <[9283, 8765]>. firewalld.service is not available 2026-03-06T13:44:52.215 INFO:teuthology.orchestra.run.vm02.stdout:Waiting for mgr to start... 2026-03-06T13:44:52.215 INFO:teuthology.orchestra.run.vm02.stdout:Waiting for mgr... 2026-03-06T13:44:52.485 INFO:journalctl@ceph.mgr.a.vm02.stdout:Mar 06 13:44:52 vm02 podman[49429]: 2026-03-06 13:44:52.203537127 +0100 CET m=+0.078724125 container init 6b7b3227c2471ec9cc79a4394e408b9ae7046c8aa6026fcc1dcbf32f8c16a854 (image=harbor.clyso.com/custom-ceph/ceph/ceph:cobaltcore-storage-v19.2.3-fasttrack-3, name=ceph-2b1b4a72-195a-11f1-b1df-bf1f800d676b-mgr-a, CEPH_GIT_REPO=https://github.com/irq0/ceph.git, CEPH_REF=19.2.3-47-gc24117fd552, CEPH_SHA1=c24117fd5525679b799527bc1bd1f1dd0a2db5e2, FROM_IMAGE=rockylinux:9, GANESHA_REPO_BASEURL=https://buildlogs.centos.org/centos/$releasever-stream/storage/$basearch/nfsganesha-5/, OSD_FLAVOR=default, ceph=True, io.buildah.version=1.41.8) 2026-03-06T13:44:52.486 INFO:journalctl@ceph.mgr.a.vm02.stdout:Mar 06 13:44:52 vm02 podman[49429]: 2026-03-06 13:44:52.207548826 +0100 CET m=+0.082735824 container start 6b7b3227c2471ec9cc79a4394e408b9ae7046c8aa6026fcc1dcbf32f8c16a854 (image=harbor.clyso.com/custom-ceph/ceph/ceph:cobaltcore-storage-v19.2.3-fasttrack-3, name=ceph-2b1b4a72-195a-11f1-b1df-bf1f800d676b-mgr-a, FROM_IMAGE=rockylinux:9, GANESHA_REPO_BASEURL=https://buildlogs.centos.org/centos/$releasever-stream/storage/$basearch/nfsganesha-5/, OSD_FLAVOR=default, ceph=True, io.buildah.version=1.41.8, CEPH_GIT_REPO=https://github.com/irq0/ceph.git, CEPH_REF=19.2.3-47-gc24117fd552, CEPH_SHA1=c24117fd5525679b799527bc1bd1f1dd0a2db5e2) 2026-03-06T13:44:52.486 INFO:journalctl@ceph.mgr.a.vm02.stdout:Mar 06 13:44:52 vm02 bash[49429]: 6b7b3227c2471ec9cc79a4394e408b9ae7046c8aa6026fcc1dcbf32f8c16a854 2026-03-06T13:44:52.486 INFO:journalctl@ceph.mgr.a.vm02.stdout:Mar 06 13:44:52 vm02 podman[49429]: 2026-03-06 13:44:52.132873229 +0100 CET m=+0.008060227 image pull 306e97de47e91c2b4b24d3dc09be3b3a12039b078f343d91220102acc6628a68 harbor.clyso.com/custom-ceph/ceph/ceph:cobaltcore-storage-v19.2.3-fasttrack-3 2026-03-06T13:44:52.486 INFO:journalctl@ceph.mgr.a.vm02.stdout:Mar 06 13:44:52 vm02 systemd[1]: Started Ceph mgr.a for 2b1b4a72-195a-11f1-b1df-bf1f800d676b. 2026-03-06T13:44:52.486 INFO:journalctl@ceph.mgr.a.vm02.stdout:Mar 06 13:44:52 vm02 ceph-2b1b4a72-195a-11f1-b1df-bf1f800d676b-mgr-a[49440]: 2026-03-06T12:44:52.439+0000 7f0c83333100 -1 mgr[py] Module alerts has missing NOTIFY_TYPES member 2026-03-06T13:44:52.651 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout 2026-03-06T13:44:52.651 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout { 2026-03-06T13:44:52.651 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "fsid": "2b1b4a72-195a-11f1-b1df-bf1f800d676b", 2026-03-06T13:44:52.651 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "health": { 2026-03-06T13:44:52.651 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "status": "HEALTH_OK", 2026-03-06T13:44:52.651 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "checks": {}, 2026-03-06T13:44:52.651 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "mutes": [] 2026-03-06T13:44:52.651 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout }, 2026-03-06T13:44:52.651 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "election_epoch": 5, 2026-03-06T13:44:52.651 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "quorum": [ 2026-03-06T13:44:52.651 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout 0 2026-03-06T13:44:52.651 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout ], 2026-03-06T13:44:52.651 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "quorum_names": [ 2026-03-06T13:44:52.651 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "a" 2026-03-06T13:44:52.651 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout ], 2026-03-06T13:44:52.651 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "quorum_age": 1, 2026-03-06T13:44:52.651 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "monmap": { 2026-03-06T13:44:52.651 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "epoch": 1, 2026-03-06T13:44:52.651 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "min_mon_release_name": "squid", 2026-03-06T13:44:52.651 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "num_mons": 1 2026-03-06T13:44:52.651 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout }, 2026-03-06T13:44:52.651 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "osdmap": { 2026-03-06T13:44:52.651 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "epoch": 1, 2026-03-06T13:44:52.652 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "num_osds": 0, 2026-03-06T13:44:52.652 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "num_up_osds": 0, 2026-03-06T13:44:52.652 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "osd_up_since": 0, 2026-03-06T13:44:52.652 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "num_in_osds": 0, 2026-03-06T13:44:52.652 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "osd_in_since": 0, 2026-03-06T13:44:52.652 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "num_remapped_pgs": 0 2026-03-06T13:44:52.652 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout }, 2026-03-06T13:44:52.652 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "pgmap": { 2026-03-06T13:44:52.652 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "pgs_by_state": [], 2026-03-06T13:44:52.652 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "num_pgs": 0, 2026-03-06T13:44:52.652 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "num_pools": 0, 2026-03-06T13:44:52.652 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "num_objects": 0, 2026-03-06T13:44:52.652 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "data_bytes": 0, 2026-03-06T13:44:52.652 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "bytes_used": 0, 2026-03-06T13:44:52.652 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "bytes_avail": 0, 2026-03-06T13:44:52.652 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "bytes_total": 0 2026-03-06T13:44:52.652 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout }, 2026-03-06T13:44:52.652 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "fsmap": { 2026-03-06T13:44:52.652 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "epoch": 1, 2026-03-06T13:44:52.652 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "btime": "2026-03-06T12:44:49:710370+0000", 2026-03-06T13:44:52.652 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "by_rank": [], 2026-03-06T13:44:52.652 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "up:standby": 0 2026-03-06T13:44:52.652 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout }, 2026-03-06T13:44:52.652 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "mgrmap": { 2026-03-06T13:44:52.652 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "available": false, 2026-03-06T13:44:52.652 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "num_standbys": 0, 2026-03-06T13:44:52.652 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "modules": [ 2026-03-06T13:44:52.652 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "iostat", 2026-03-06T13:44:52.652 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "nfs", 2026-03-06T13:44:52.652 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "restful" 2026-03-06T13:44:52.652 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout ], 2026-03-06T13:44:52.652 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "services": {} 2026-03-06T13:44:52.652 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout }, 2026-03-06T13:44:52.652 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "servicemap": { 2026-03-06T13:44:52.652 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "epoch": 1, 2026-03-06T13:44:52.652 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "modified": "2026-03-06T12:44:49.710875+0000", 2026-03-06T13:44:52.652 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "services": {} 2026-03-06T13:44:52.652 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout }, 2026-03-06T13:44:52.652 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "progress_events": {} 2026-03-06T13:44:52.652 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout } 2026-03-06T13:44:52.652 INFO:teuthology.orchestra.run.vm02.stdout:mgr not available, waiting (1/15)... 2026-03-06T13:44:52.862 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:44:52 vm02 ceph-mon[49207]: from='client.? 192.168.123.102:0/1479433379' entity='client.admin' 2026-03-06T13:44:52.862 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:44:52 vm02 ceph-mon[49207]: from='client.? 192.168.123.102:0/2396099408' entity='client.admin' cmd=[{"prefix": "status", "format": "json-pretty"}]: dispatch 2026-03-06T13:44:52.862 INFO:journalctl@ceph.mgr.a.vm02.stdout:Mar 06 13:44:52 vm02 ceph-2b1b4a72-195a-11f1-b1df-bf1f800d676b-mgr-a[49440]: 2026-03-06T12:44:52.566+0000 7f0c83333100 -1 mgr[py] Module balancer has missing NOTIFY_TYPES member 2026-03-06T13:44:54.112 INFO:journalctl@ceph.mgr.a.vm02.stdout:Mar 06 13:44:53 vm02 ceph-2b1b4a72-195a-11f1-b1df-bf1f800d676b-mgr-a[49440]: 2026-03-06T12:44:53.620+0000 7f0c83333100 -1 mgr[py] Module crash has missing NOTIFY_TYPES member 2026-03-06T13:44:54.718 INFO:journalctl@ceph.mgr.a.vm02.stdout:Mar 06 13:44:54 vm02 ceph-2b1b4a72-195a-11f1-b1df-bf1f800d676b-mgr-a[49440]: 2026-03-06T12:44:54.455+0000 7f0c83333100 -1 mgr[py] Module devicehealth has missing NOTIFY_TYPES member 2026-03-06T13:44:54.718 INFO:journalctl@ceph.mgr.a.vm02.stdout:Mar 06 13:44:54 vm02 ceph-2b1b4a72-195a-11f1-b1df-bf1f800d676b-mgr-a[49440]: 2026-03-06T12:44:54.576+0000 7f0c83333100 -1 mgr[py] Module influx has missing NOTIFY_TYPES member 2026-03-06T13:44:54.972 INFO:journalctl@ceph.mgr.a.vm02.stdout:Mar 06 13:44:54 vm02 ceph-2b1b4a72-195a-11f1-b1df-bf1f800d676b-mgr-a[49440]: 2026-03-06T12:44:54.835+0000 7f0c83333100 -1 mgr[py] Module iostat has missing NOTIFY_TYPES member 2026-03-06T13:44:55.098 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout 2026-03-06T13:44:55.098 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout { 2026-03-06T13:44:55.099 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "fsid": "2b1b4a72-195a-11f1-b1df-bf1f800d676b", 2026-03-06T13:44:55.099 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "health": { 2026-03-06T13:44:55.099 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "status": "HEALTH_OK", 2026-03-06T13:44:55.099 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "checks": {}, 2026-03-06T13:44:55.099 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "mutes": [] 2026-03-06T13:44:55.099 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout }, 2026-03-06T13:44:55.099 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "election_epoch": 5, 2026-03-06T13:44:55.099 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "quorum": [ 2026-03-06T13:44:55.099 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout 0 2026-03-06T13:44:55.099 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout ], 2026-03-06T13:44:55.099 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "quorum_names": [ 2026-03-06T13:44:55.099 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "a" 2026-03-06T13:44:55.099 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout ], 2026-03-06T13:44:55.099 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "quorum_age": 3, 2026-03-06T13:44:55.099 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "monmap": { 2026-03-06T13:44:55.099 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "epoch": 1, 2026-03-06T13:44:55.099 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "min_mon_release_name": "squid", 2026-03-06T13:44:55.099 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "num_mons": 1 2026-03-06T13:44:55.099 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout }, 2026-03-06T13:44:55.099 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "osdmap": { 2026-03-06T13:44:55.099 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "epoch": 1, 2026-03-06T13:44:55.099 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "num_osds": 0, 2026-03-06T13:44:55.099 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "num_up_osds": 0, 2026-03-06T13:44:55.099 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "osd_up_since": 0, 2026-03-06T13:44:55.099 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "num_in_osds": 0, 2026-03-06T13:44:55.099 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "osd_in_since": 0, 2026-03-06T13:44:55.099 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "num_remapped_pgs": 0 2026-03-06T13:44:55.100 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout }, 2026-03-06T13:44:55.100 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "pgmap": { 2026-03-06T13:44:55.100 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "pgs_by_state": [], 2026-03-06T13:44:55.100 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "num_pgs": 0, 2026-03-06T13:44:55.100 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "num_pools": 0, 2026-03-06T13:44:55.100 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "num_objects": 0, 2026-03-06T13:44:55.100 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "data_bytes": 0, 2026-03-06T13:44:55.100 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "bytes_used": 0, 2026-03-06T13:44:55.100 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "bytes_avail": 0, 2026-03-06T13:44:55.100 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "bytes_total": 0 2026-03-06T13:44:55.100 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout }, 2026-03-06T13:44:55.100 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "fsmap": { 2026-03-06T13:44:55.100 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "epoch": 1, 2026-03-06T13:44:55.100 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "btime": "2026-03-06T12:44:49:710370+0000", 2026-03-06T13:44:55.100 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "by_rank": [], 2026-03-06T13:44:55.100 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "up:standby": 0 2026-03-06T13:44:55.100 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout }, 2026-03-06T13:44:55.100 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "mgrmap": { 2026-03-06T13:44:55.100 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "available": false, 2026-03-06T13:44:55.100 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "num_standbys": 0, 2026-03-06T13:44:55.100 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "modules": [ 2026-03-06T13:44:55.100 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "iostat", 2026-03-06T13:44:55.100 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "nfs", 2026-03-06T13:44:55.100 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "restful" 2026-03-06T13:44:55.100 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout ], 2026-03-06T13:44:55.100 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "services": {} 2026-03-06T13:44:55.100 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout }, 2026-03-06T13:44:55.100 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "servicemap": { 2026-03-06T13:44:55.100 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "epoch": 1, 2026-03-06T13:44:55.100 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "modified": "2026-03-06T12:44:49.710875+0000", 2026-03-06T13:44:55.100 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "services": {} 2026-03-06T13:44:55.100 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout }, 2026-03-06T13:44:55.100 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "progress_events": {} 2026-03-06T13:44:55.100 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout } 2026-03-06T13:44:55.100 INFO:teuthology.orchestra.run.vm02.stdout:mgr not available, waiting (2/15)... 2026-03-06T13:44:55.362 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:44:54 vm02 ceph-mon[49207]: from='client.? 192.168.123.102:0/2515410826' entity='client.admin' cmd=[{"prefix": "status", "format": "json-pretty"}]: dispatch 2026-03-06T13:44:56.862 INFO:journalctl@ceph.mgr.a.vm02.stdout:Mar 06 13:44:56 vm02 ceph-2b1b4a72-195a-11f1-b1df-bf1f800d676b-mgr-a[49440]: 2026-03-06T12:44:56.588+0000 7f0c83333100 -1 mgr[py] Module nfs has missing NOTIFY_TYPES member 2026-03-06T13:44:57.173 INFO:journalctl@ceph.mgr.a.vm02.stdout:Mar 06 13:44:56 vm02 ceph-2b1b4a72-195a-11f1-b1df-bf1f800d676b-mgr-a[49440]: 2026-03-06T12:44:56.910+0000 7f0c83333100 -1 mgr[py] Module orchestrator has missing NOTIFY_TYPES member 2026-03-06T13:44:57.173 INFO:journalctl@ceph.mgr.a.vm02.stdout:Mar 06 13:44:57 vm02 ceph-2b1b4a72-195a-11f1-b1df-bf1f800d676b-mgr-a[49440]: 2026-03-06T12:44:57.043+0000 7f0c83333100 -1 mgr[py] Module osd_perf_query has missing NOTIFY_TYPES member 2026-03-06T13:44:57.431 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:44:57 vm02 ceph-mon[49207]: from='client.? 192.168.123.102:0/3612432979' entity='client.admin' cmd=[{"prefix": "status", "format": "json-pretty"}]: dispatch 2026-03-06T13:44:57.431 INFO:journalctl@ceph.mgr.a.vm02.stdout:Mar 06 13:44:57 vm02 ceph-2b1b4a72-195a-11f1-b1df-bf1f800d676b-mgr-a[49440]: 2026-03-06T12:44:57.172+0000 7f0c83333100 -1 mgr[py] Module osd_support has missing NOTIFY_TYPES member 2026-03-06T13:44:57.431 INFO:journalctl@ceph.mgr.a.vm02.stdout:Mar 06 13:44:57 vm02 ceph-2b1b4a72-195a-11f1-b1df-bf1f800d676b-mgr-a[49440]: 2026-03-06T12:44:57.311+0000 7f0c83333100 -1 mgr[py] Module pg_autoscaler has missing NOTIFY_TYPES member 2026-03-06T13:44:57.522 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout 2026-03-06T13:44:57.522 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout { 2026-03-06T13:44:57.522 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "fsid": "2b1b4a72-195a-11f1-b1df-bf1f800d676b", 2026-03-06T13:44:57.522 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "health": { 2026-03-06T13:44:57.522 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "status": "HEALTH_OK", 2026-03-06T13:44:57.522 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "checks": {}, 2026-03-06T13:44:57.522 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "mutes": [] 2026-03-06T13:44:57.522 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout }, 2026-03-06T13:44:57.522 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "election_epoch": 5, 2026-03-06T13:44:57.522 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "quorum": [ 2026-03-06T13:44:57.522 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout 0 2026-03-06T13:44:57.522 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout ], 2026-03-06T13:44:57.522 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "quorum_names": [ 2026-03-06T13:44:57.522 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "a" 2026-03-06T13:44:57.523 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout ], 2026-03-06T13:44:57.523 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "quorum_age": 5, 2026-03-06T13:44:57.523 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "monmap": { 2026-03-06T13:44:57.523 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "epoch": 1, 2026-03-06T13:44:57.523 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "min_mon_release_name": "squid", 2026-03-06T13:44:57.523 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "num_mons": 1 2026-03-06T13:44:57.523 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout }, 2026-03-06T13:44:57.523 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "osdmap": { 2026-03-06T13:44:57.523 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "epoch": 1, 2026-03-06T13:44:57.523 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "num_osds": 0, 2026-03-06T13:44:57.523 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "num_up_osds": 0, 2026-03-06T13:44:57.523 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "osd_up_since": 0, 2026-03-06T13:44:57.523 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "num_in_osds": 0, 2026-03-06T13:44:57.523 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "osd_in_since": 0, 2026-03-06T13:44:57.523 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "num_remapped_pgs": 0 2026-03-06T13:44:57.524 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout }, 2026-03-06T13:44:57.524 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "pgmap": { 2026-03-06T13:44:57.524 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "pgs_by_state": [], 2026-03-06T13:44:57.524 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "num_pgs": 0, 2026-03-06T13:44:57.524 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "num_pools": 0, 2026-03-06T13:44:57.524 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "num_objects": 0, 2026-03-06T13:44:57.524 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "data_bytes": 0, 2026-03-06T13:44:57.524 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "bytes_used": 0, 2026-03-06T13:44:57.524 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "bytes_avail": 0, 2026-03-06T13:44:57.524 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "bytes_total": 0 2026-03-06T13:44:57.524 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout }, 2026-03-06T13:44:57.524 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "fsmap": { 2026-03-06T13:44:57.524 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "epoch": 1, 2026-03-06T13:44:57.524 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "btime": "2026-03-06T12:44:49:710370+0000", 2026-03-06T13:44:57.524 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "by_rank": [], 2026-03-06T13:44:57.524 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "up:standby": 0 2026-03-06T13:44:57.524 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout }, 2026-03-06T13:44:57.524 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "mgrmap": { 2026-03-06T13:44:57.524 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "available": false, 2026-03-06T13:44:57.524 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "num_standbys": 0, 2026-03-06T13:44:57.524 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "modules": [ 2026-03-06T13:44:57.524 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "iostat", 2026-03-06T13:44:57.524 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "nfs", 2026-03-06T13:44:57.524 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "restful" 2026-03-06T13:44:57.524 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout ], 2026-03-06T13:44:57.524 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "services": {} 2026-03-06T13:44:57.524 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout }, 2026-03-06T13:44:57.524 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "servicemap": { 2026-03-06T13:44:57.524 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "epoch": 1, 2026-03-06T13:44:57.524 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "modified": "2026-03-06T12:44:49.710875+0000", 2026-03-06T13:44:57.524 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "services": {} 2026-03-06T13:44:57.524 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout }, 2026-03-06T13:44:57.525 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "progress_events": {} 2026-03-06T13:44:57.525 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout } 2026-03-06T13:44:57.525 INFO:teuthology.orchestra.run.vm02.stdout:mgr not available, waiting (3/15)... 2026-03-06T13:44:57.862 INFO:journalctl@ceph.mgr.a.vm02.stdout:Mar 06 13:44:57 vm02 ceph-2b1b4a72-195a-11f1-b1df-bf1f800d676b-mgr-a[49440]: 2026-03-06T12:44:57.439+0000 7f0c83333100 -1 mgr[py] Module progress has missing NOTIFY_TYPES member 2026-03-06T13:44:58.362 INFO:journalctl@ceph.mgr.a.vm02.stdout:Mar 06 13:44:57 vm02 ceph-2b1b4a72-195a-11f1-b1df-bf1f800d676b-mgr-a[49440]: 2026-03-06T12:44:57.977+0000 7f0c83333100 -1 mgr[py] Module prometheus has missing NOTIFY_TYPES member 2026-03-06T13:44:58.362 INFO:journalctl@ceph.mgr.a.vm02.stdout:Mar 06 13:44:58 vm02 ceph-2b1b4a72-195a-11f1-b1df-bf1f800d676b-mgr-a[49440]: 2026-03-06T12:44:58.131+0000 7f0c83333100 -1 mgr[py] Module rbd_support has missing NOTIFY_TYPES member 2026-03-06T13:44:59.362 INFO:journalctl@ceph.mgr.a.vm02.stdout:Mar 06 13:44:58 vm02 ceph-2b1b4a72-195a-11f1-b1df-bf1f800d676b-mgr-a[49440]: 2026-03-06T12:44:58.860+0000 7f0c83333100 -1 mgr[py] Module rgw has missing NOTIFY_TYPES member 2026-03-06T13:45:00.065 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout 2026-03-06T13:45:00.065 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout { 2026-03-06T13:45:00.065 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "fsid": "2b1b4a72-195a-11f1-b1df-bf1f800d676b", 2026-03-06T13:45:00.065 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "health": { 2026-03-06T13:45:00.065 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "status": "HEALTH_OK", 2026-03-06T13:45:00.065 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "checks": {}, 2026-03-06T13:45:00.065 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "mutes": [] 2026-03-06T13:45:00.065 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout }, 2026-03-06T13:45:00.065 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "election_epoch": 5, 2026-03-06T13:45:00.065 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "quorum": [ 2026-03-06T13:45:00.065 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout 0 2026-03-06T13:45:00.065 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout ], 2026-03-06T13:45:00.065 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "quorum_names": [ 2026-03-06T13:45:00.065 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "a" 2026-03-06T13:45:00.065 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout ], 2026-03-06T13:45:00.065 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "quorum_age": 8, 2026-03-06T13:45:00.065 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "monmap": { 2026-03-06T13:45:00.065 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "epoch": 1, 2026-03-06T13:45:00.065 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "min_mon_release_name": "squid", 2026-03-06T13:45:00.065 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "num_mons": 1 2026-03-06T13:45:00.065 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout }, 2026-03-06T13:45:00.065 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "osdmap": { 2026-03-06T13:45:00.065 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "epoch": 1, 2026-03-06T13:45:00.065 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "num_osds": 0, 2026-03-06T13:45:00.065 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "num_up_osds": 0, 2026-03-06T13:45:00.065 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "osd_up_since": 0, 2026-03-06T13:45:00.065 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "num_in_osds": 0, 2026-03-06T13:45:00.066 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "osd_in_since": 0, 2026-03-06T13:45:00.066 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "num_remapped_pgs": 0 2026-03-06T13:45:00.067 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout }, 2026-03-06T13:45:00.067 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "pgmap": { 2026-03-06T13:45:00.067 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "pgs_by_state": [], 2026-03-06T13:45:00.067 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "num_pgs": 0, 2026-03-06T13:45:00.067 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "num_pools": 0, 2026-03-06T13:45:00.067 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "num_objects": 0, 2026-03-06T13:45:00.067 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "data_bytes": 0, 2026-03-06T13:45:00.067 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "bytes_used": 0, 2026-03-06T13:45:00.067 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "bytes_avail": 0, 2026-03-06T13:45:00.067 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "bytes_total": 0 2026-03-06T13:45:00.067 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout }, 2026-03-06T13:45:00.067 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "fsmap": { 2026-03-06T13:45:00.067 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "epoch": 1, 2026-03-06T13:45:00.067 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "btime": "2026-03-06T12:44:49:710370+0000", 2026-03-06T13:45:00.067 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "by_rank": [], 2026-03-06T13:45:00.067 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "up:standby": 0 2026-03-06T13:45:00.067 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout }, 2026-03-06T13:45:00.067 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "mgrmap": { 2026-03-06T13:45:00.067 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "available": false, 2026-03-06T13:45:00.067 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "num_standbys": 0, 2026-03-06T13:45:00.067 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "modules": [ 2026-03-06T13:45:00.067 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "iostat", 2026-03-06T13:45:00.067 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "nfs", 2026-03-06T13:45:00.067 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "restful" 2026-03-06T13:45:00.067 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout ], 2026-03-06T13:45:00.067 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "services": {} 2026-03-06T13:45:00.067 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout }, 2026-03-06T13:45:00.067 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "servicemap": { 2026-03-06T13:45:00.067 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "epoch": 1, 2026-03-06T13:45:00.067 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "modified": "2026-03-06T12:44:49.710875+0000", 2026-03-06T13:45:00.067 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "services": {} 2026-03-06T13:45:00.067 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout }, 2026-03-06T13:45:00.067 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "progress_events": {} 2026-03-06T13:45:00.067 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout } 2026-03-06T13:45:00.067 INFO:teuthology.orchestra.run.vm02.stdout:mgr not available, waiting (4/15)... 2026-03-06T13:45:00.111 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:44:59 vm02 ceph-mon[49207]: from='client.? 192.168.123.102:0/59197504' entity='client.admin' cmd=[{"prefix": "status", "format": "json-pretty"}]: dispatch 2026-03-06T13:45:00.111 INFO:journalctl@ceph.mgr.a.vm02.stdout:Mar 06 13:44:59 vm02 ceph-2b1b4a72-195a-11f1-b1df-bf1f800d676b-mgr-a[49440]: 2026-03-06T12:44:59.966+0000 7f0c83333100 -1 mgr[py] Module rook has missing NOTIFY_TYPES member 2026-03-06T13:45:00.111 INFO:journalctl@ceph.mgr.a.vm02.stdout:Mar 06 13:45:00 vm02 ceph-2b1b4a72-195a-11f1-b1df-bf1f800d676b-mgr-a[49440]: 2026-03-06T12:45:00.108+0000 7f0c83333100 -1 mgr[py] Module selftest has missing NOTIFY_TYPES member 2026-03-06T13:45:00.362 INFO:journalctl@ceph.mgr.a.vm02.stdout:Mar 06 13:45:00 vm02 ceph-2b1b4a72-195a-11f1-b1df-bf1f800d676b-mgr-a[49440]: 2026-03-06T12:45:00.239+0000 7f0c83333100 -1 mgr[py] Module snap_schedule has missing NOTIFY_TYPES member 2026-03-06T13:45:00.862 INFO:journalctl@ceph.mgr.a.vm02.stdout:Mar 06 13:45:00 vm02 ceph-2b1b4a72-195a-11f1-b1df-bf1f800d676b-mgr-a[49440]: 2026-03-06T12:45:00.503+0000 7f0c83333100 -1 mgr[py] Module status has missing NOTIFY_TYPES member 2026-03-06T13:45:00.862 INFO:journalctl@ceph.mgr.a.vm02.stdout:Mar 06 13:45:00 vm02 ceph-2b1b4a72-195a-11f1-b1df-bf1f800d676b-mgr-a[49440]: 2026-03-06T12:45:00.625+0000 7f0c83333100 -1 mgr[py] Module telegraf has missing NOTIFY_TYPES member 2026-03-06T13:45:01.266 INFO:journalctl@ceph.mgr.a.vm02.stdout:Mar 06 13:45:00 vm02 ceph-2b1b4a72-195a-11f1-b1df-bf1f800d676b-mgr-a[49440]: 2026-03-06T12:45:00.928+0000 7f0c83333100 -1 mgr[py] Module telemetry has missing NOTIFY_TYPES member 2026-03-06T13:45:01.266 INFO:journalctl@ceph.mgr.a.vm02.stdout:Mar 06 13:45:01 vm02 ceph-2b1b4a72-195a-11f1-b1df-bf1f800d676b-mgr-a[49440]: 2026-03-06T12:45:01.263+0000 7f0c83333100 -1 mgr[py] Module test_orchestrator has missing NOTIFY_TYPES member 2026-03-06T13:45:02.093 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:45:01 vm02 ceph-mon[49207]: Activating manager daemon a 2026-03-06T13:45:02.094 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:45:01 vm02 ceph-mon[49207]: mgrmap e2: a(active, starting, since 0.00553259s) 2026-03-06T13:45:02.094 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:45:01 vm02 ceph-mon[49207]: from='mgr.14100 192.168.123.102:0/345396643' entity='mgr.a' cmd=[{"prefix": "mds metadata"}]: dispatch 2026-03-06T13:45:02.094 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:45:01 vm02 ceph-mon[49207]: from='mgr.14100 192.168.123.102:0/345396643' entity='mgr.a' cmd=[{"prefix": "osd metadata"}]: dispatch 2026-03-06T13:45:02.094 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:45:01 vm02 ceph-mon[49207]: from='mgr.14100 192.168.123.102:0/345396643' entity='mgr.a' cmd=[{"prefix": "mon metadata"}]: dispatch 2026-03-06T13:45:02.094 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:45:01 vm02 ceph-mon[49207]: from='mgr.14100 192.168.123.102:0/345396643' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "a"}]: dispatch 2026-03-06T13:45:02.094 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:45:01 vm02 ceph-mon[49207]: from='mgr.14100 192.168.123.102:0/345396643' entity='mgr.a' cmd=[{"prefix": "mgr metadata", "who": "a", "id": "a"}]: dispatch 2026-03-06T13:45:02.094 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:45:01 vm02 ceph-mon[49207]: Manager daemon a is now available 2026-03-06T13:45:02.094 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:45:01 vm02 ceph-mon[49207]: from='mgr.14100 192.168.123.102:0/345396643' entity='mgr.a' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/a/mirror_snapshot_schedule"}]: dispatch 2026-03-06T13:45:02.094 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:45:01 vm02 ceph-mon[49207]: from='mgr.14100 192.168.123.102:0/345396643' entity='mgr.a' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/a/trash_purge_schedule"}]: dispatch 2026-03-06T13:45:02.094 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:45:01 vm02 ceph-mon[49207]: from='mgr.14100 192.168.123.102:0/345396643' entity='mgr.a' 2026-03-06T13:45:02.094 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:45:01 vm02 ceph-mon[49207]: from='mgr.14100 192.168.123.102:0/345396643' entity='mgr.a' 2026-03-06T13:45:02.094 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:45:01 vm02 ceph-mon[49207]: from='mgr.14100 192.168.123.102:0/345396643' entity='mgr.a' 2026-03-06T13:45:02.094 INFO:journalctl@ceph.mgr.a.vm02.stdout:Mar 06 13:45:01 vm02 ceph-2b1b4a72-195a-11f1-b1df-bf1f800d676b-mgr-a[49440]: 2026-03-06T12:45:01.632+0000 7f0c83333100 -1 mgr[py] Module volumes has missing NOTIFY_TYPES member 2026-03-06T13:45:02.094 INFO:journalctl@ceph.mgr.a.vm02.stdout:Mar 06 13:45:01 vm02 ceph-2b1b4a72-195a-11f1-b1df-bf1f800d676b-mgr-a[49440]: 2026-03-06T12:45:01.754+0000 7f0c83333100 -1 mgr[py] Module zabbix has missing NOTIFY_TYPES member 2026-03-06T13:45:02.498 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout 2026-03-06T13:45:02.498 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout { 2026-03-06T13:45:02.498 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "fsid": "2b1b4a72-195a-11f1-b1df-bf1f800d676b", 2026-03-06T13:45:02.498 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "health": { 2026-03-06T13:45:02.498 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "status": "HEALTH_OK", 2026-03-06T13:45:02.498 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "checks": {}, 2026-03-06T13:45:02.498 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "mutes": [] 2026-03-06T13:45:02.498 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout }, 2026-03-06T13:45:02.498 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "election_epoch": 5, 2026-03-06T13:45:02.498 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "quorum": [ 2026-03-06T13:45:02.498 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout 0 2026-03-06T13:45:02.499 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout ], 2026-03-06T13:45:02.499 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "quorum_names": [ 2026-03-06T13:45:02.499 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "a" 2026-03-06T13:45:02.499 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout ], 2026-03-06T13:45:02.499 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "quorum_age": 10, 2026-03-06T13:45:02.499 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "monmap": { 2026-03-06T13:45:02.499 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "epoch": 1, 2026-03-06T13:45:02.499 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "min_mon_release_name": "squid", 2026-03-06T13:45:02.499 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "num_mons": 1 2026-03-06T13:45:02.499 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout }, 2026-03-06T13:45:02.499 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "osdmap": { 2026-03-06T13:45:02.499 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "epoch": 1, 2026-03-06T13:45:02.499 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "num_osds": 0, 2026-03-06T13:45:02.499 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "num_up_osds": 0, 2026-03-06T13:45:02.499 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "osd_up_since": 0, 2026-03-06T13:45:02.499 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "num_in_osds": 0, 2026-03-06T13:45:02.499 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "osd_in_since": 0, 2026-03-06T13:45:02.499 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "num_remapped_pgs": 0 2026-03-06T13:45:02.500 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout }, 2026-03-06T13:45:02.500 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "pgmap": { 2026-03-06T13:45:02.500 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "pgs_by_state": [], 2026-03-06T13:45:02.500 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "num_pgs": 0, 2026-03-06T13:45:02.500 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "num_pools": 0, 2026-03-06T13:45:02.500 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "num_objects": 0, 2026-03-06T13:45:02.500 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "data_bytes": 0, 2026-03-06T13:45:02.500 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "bytes_used": 0, 2026-03-06T13:45:02.500 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "bytes_avail": 0, 2026-03-06T13:45:02.500 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "bytes_total": 0 2026-03-06T13:45:02.500 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout }, 2026-03-06T13:45:02.500 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "fsmap": { 2026-03-06T13:45:02.500 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "epoch": 1, 2026-03-06T13:45:02.501 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "btime": "2026-03-06T12:44:49:710370+0000", 2026-03-06T13:45:02.501 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "by_rank": [], 2026-03-06T13:45:02.501 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "up:standby": 0 2026-03-06T13:45:02.501 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout }, 2026-03-06T13:45:02.501 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "mgrmap": { 2026-03-06T13:45:02.501 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "available": false, 2026-03-06T13:45:02.501 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "num_standbys": 0, 2026-03-06T13:45:02.501 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "modules": [ 2026-03-06T13:45:02.501 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "iostat", 2026-03-06T13:45:02.501 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "nfs", 2026-03-06T13:45:02.501 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "restful" 2026-03-06T13:45:02.501 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout ], 2026-03-06T13:45:02.501 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "services": {} 2026-03-06T13:45:02.501 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout }, 2026-03-06T13:45:02.501 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "servicemap": { 2026-03-06T13:45:02.501 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "epoch": 1, 2026-03-06T13:45:02.501 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "modified": "2026-03-06T12:44:49.710875+0000", 2026-03-06T13:45:02.501 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "services": {} 2026-03-06T13:45:02.501 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout }, 2026-03-06T13:45:02.501 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "progress_events": {} 2026-03-06T13:45:02.501 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout } 2026-03-06T13:45:02.501 INFO:teuthology.orchestra.run.vm02.stdout:mgr not available, waiting (5/15)... 2026-03-06T13:45:03.112 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:45:02 vm02 ceph-mon[49207]: from='client.? 192.168.123.102:0/338634930' entity='client.admin' cmd=[{"prefix": "status", "format": "json-pretty"}]: dispatch 2026-03-06T13:45:03.112 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:45:02 vm02 ceph-mon[49207]: mgrmap e3: a(active, since 1.01187s) 2026-03-06T13:45:05.039 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout 2026-03-06T13:45:05.039 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout { 2026-03-06T13:45:05.039 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "fsid": "2b1b4a72-195a-11f1-b1df-bf1f800d676b", 2026-03-06T13:45:05.039 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "health": { 2026-03-06T13:45:05.039 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "status": "HEALTH_OK", 2026-03-06T13:45:05.039 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "checks": {}, 2026-03-06T13:45:05.040 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "mutes": [] 2026-03-06T13:45:05.040 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout }, 2026-03-06T13:45:05.040 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "election_epoch": 5, 2026-03-06T13:45:05.040 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "quorum": [ 2026-03-06T13:45:05.040 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout 0 2026-03-06T13:45:05.040 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout ], 2026-03-06T13:45:05.040 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "quorum_names": [ 2026-03-06T13:45:05.040 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "a" 2026-03-06T13:45:05.040 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout ], 2026-03-06T13:45:05.040 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "quorum_age": 13, 2026-03-06T13:45:05.040 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "monmap": { 2026-03-06T13:45:05.040 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "epoch": 1, 2026-03-06T13:45:05.040 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "min_mon_release_name": "squid", 2026-03-06T13:45:05.040 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "num_mons": 1 2026-03-06T13:45:05.040 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout }, 2026-03-06T13:45:05.040 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "osdmap": { 2026-03-06T13:45:05.041 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "epoch": 1, 2026-03-06T13:45:05.041 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "num_osds": 0, 2026-03-06T13:45:05.041 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "num_up_osds": 0, 2026-03-06T13:45:05.041 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "osd_up_since": 0, 2026-03-06T13:45:05.041 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "num_in_osds": 0, 2026-03-06T13:45:05.041 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "osd_in_since": 0, 2026-03-06T13:45:05.041 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "num_remapped_pgs": 0 2026-03-06T13:45:05.041 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout }, 2026-03-06T13:45:05.041 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "pgmap": { 2026-03-06T13:45:05.041 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "pgs_by_state": [], 2026-03-06T13:45:05.041 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "num_pgs": 0, 2026-03-06T13:45:05.041 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "num_pools": 0, 2026-03-06T13:45:05.041 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "num_objects": 0, 2026-03-06T13:45:05.041 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "data_bytes": 0, 2026-03-06T13:45:05.041 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "bytes_used": 0, 2026-03-06T13:45:05.041 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "bytes_avail": 0, 2026-03-06T13:45:05.041 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "bytes_total": 0 2026-03-06T13:45:05.041 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout }, 2026-03-06T13:45:05.041 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "fsmap": { 2026-03-06T13:45:05.041 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "epoch": 1, 2026-03-06T13:45:05.041 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "btime": "2026-03-06T12:44:49:710370+0000", 2026-03-06T13:45:05.041 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "by_rank": [], 2026-03-06T13:45:05.041 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "up:standby": 0 2026-03-06T13:45:05.041 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout }, 2026-03-06T13:45:05.041 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "mgrmap": { 2026-03-06T13:45:05.041 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "available": true, 2026-03-06T13:45:05.041 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "num_standbys": 0, 2026-03-06T13:45:05.041 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "modules": [ 2026-03-06T13:45:05.041 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "iostat", 2026-03-06T13:45:05.041 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "nfs", 2026-03-06T13:45:05.042 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "restful" 2026-03-06T13:45:05.042 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout ], 2026-03-06T13:45:05.042 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "services": {} 2026-03-06T13:45:05.042 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout }, 2026-03-06T13:45:05.042 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "servicemap": { 2026-03-06T13:45:05.042 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "epoch": 1, 2026-03-06T13:45:05.042 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "modified": "2026-03-06T12:44:49.710875+0000", 2026-03-06T13:45:05.042 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "services": {} 2026-03-06T13:45:05.042 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout }, 2026-03-06T13:45:05.042 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "progress_events": {} 2026-03-06T13:45:05.042 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout } 2026-03-06T13:45:05.042 INFO:teuthology.orchestra.run.vm02.stdout:mgr is available 2026-03-06T13:45:05.079 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:45:04 vm02 ceph-mon[49207]: mgrmap e4: a(active, since 2s) 2026-03-06T13:45:05.522 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout 2026-03-06T13:45:05.522 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout [global] 2026-03-06T13:45:05.522 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout fsid = 2b1b4a72-195a-11f1-b1df-bf1f800d676b 2026-03-06T13:45:05.522 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout mon_cluster_log_file_level = debug 2026-03-06T13:45:05.522 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout mon_host = [v2:192.168.123.102:3300,v1:192.168.123.102:6789] 2026-03-06T13:45:05.523 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout mon_osd_allow_pg_remap = true 2026-03-06T13:45:05.523 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout mon_osd_allow_primary_affinity = true 2026-03-06T13:45:05.523 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout mon_warn_on_no_sortbitwise = false 2026-03-06T13:45:05.523 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout osd_crush_chooseleaf_type = 0 2026-03-06T13:45:05.523 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout 2026-03-06T13:45:05.523 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout [mgr] 2026-03-06T13:45:05.523 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout mgr/telemetry/nag = false 2026-03-06T13:45:05.523 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout 2026-03-06T13:45:05.523 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout [osd] 2026-03-06T13:45:05.523 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout osd_map_max_advance = 10 2026-03-06T13:45:05.523 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout osd_sloppy_crc = true 2026-03-06T13:45:05.523 INFO:teuthology.orchestra.run.vm02.stdout:Enabling cephadm module... 2026-03-06T13:45:06.112 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:45:05 vm02 ceph-mon[49207]: from='client.? 192.168.123.102:0/633537950' entity='client.admin' cmd=[{"prefix": "status", "format": "json-pretty"}]: dispatch 2026-03-06T13:45:06.112 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:45:05 vm02 ceph-mon[49207]: from='client.? 192.168.123.102:0/2654559664' entity='client.admin' cmd=[{"prefix": "config assimilate-conf"}]: dispatch 2026-03-06T13:45:06.112 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:45:05 vm02 ceph-mon[49207]: from='client.? 192.168.123.102:0/2654559664' entity='client.admin' cmd='[{"prefix": "config assimilate-conf"}]': finished 2026-03-06T13:45:07.084 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:45:06 vm02 ceph-mon[49207]: from='client.? 192.168.123.102:0/4248162015' entity='client.admin' cmd=[{"prefix": "mgr module enable", "module": "cephadm"}]: dispatch 2026-03-06T13:45:07.085 INFO:journalctl@ceph.mgr.a.vm02.stdout:Mar 06 13:45:06 vm02 ceph-2b1b4a72-195a-11f1-b1df-bf1f800d676b-mgr-a[49440]: ignoring --setuser ceph since I am not root 2026-03-06T13:45:07.085 INFO:journalctl@ceph.mgr.a.vm02.stdout:Mar 06 13:45:06 vm02 ceph-2b1b4a72-195a-11f1-b1df-bf1f800d676b-mgr-a[49440]: ignoring --setgroup ceph since I am not root 2026-03-06T13:45:07.362 INFO:journalctl@ceph.mgr.a.vm02.stdout:Mar 06 13:45:07 vm02 ceph-2b1b4a72-195a-11f1-b1df-bf1f800d676b-mgr-a[49440]: 2026-03-06T12:45:07.143+0000 7fb523f38100 -1 mgr[py] Module alerts has missing NOTIFY_TYPES member 2026-03-06T13:45:07.362 INFO:journalctl@ceph.mgr.a.vm02.stdout:Mar 06 13:45:07 vm02 ceph-2b1b4a72-195a-11f1-b1df-bf1f800d676b-mgr-a[49440]: 2026-03-06T12:45:07.304+0000 7fb523f38100 -1 mgr[py] Module balancer has missing NOTIFY_TYPES member 2026-03-06T13:45:07.683 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout { 2026-03-06T13:45:07.683 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "epoch": 5, 2026-03-06T13:45:07.683 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "available": true, 2026-03-06T13:45:07.683 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "active_name": "a", 2026-03-06T13:45:07.683 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "num_standby": 0 2026-03-06T13:45:07.683 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout } 2026-03-06T13:45:07.683 INFO:teuthology.orchestra.run.vm02.stdout:Waiting for the mgr to restart... 2026-03-06T13:45:07.683 INFO:teuthology.orchestra.run.vm02.stdout:Waiting for mgr epoch 5... 2026-03-06T13:45:08.112 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:45:07 vm02 ceph-mon[49207]: from='client.? 192.168.123.102:0/4248162015' entity='client.admin' cmd='[{"prefix": "mgr module enable", "module": "cephadm"}]': finished 2026-03-06T13:45:08.112 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:45:07 vm02 ceph-mon[49207]: mgrmap e5: a(active, since 5s) 2026-03-06T13:45:08.112 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:45:07 vm02 ceph-mon[49207]: from='client.? 192.168.123.102:0/4217248049' entity='client.admin' cmd=[{"prefix": "mgr stat"}]: dispatch 2026-03-06T13:45:09.112 INFO:journalctl@ceph.mgr.a.vm02.stdout:Mar 06 13:45:08 vm02 ceph-2b1b4a72-195a-11f1-b1df-bf1f800d676b-mgr-a[49440]: 2026-03-06T12:45:08.611+0000 7fb523f38100 -1 mgr[py] Module crash has missing NOTIFY_TYPES member 2026-03-06T13:45:09.862 INFO:journalctl@ceph.mgr.a.vm02.stdout:Mar 06 13:45:09 vm02 ceph-2b1b4a72-195a-11f1-b1df-bf1f800d676b-mgr-a[49440]: 2026-03-06T12:45:09.498+0000 7fb523f38100 -1 mgr[py] Module devicehealth has missing NOTIFY_TYPES member 2026-03-06T13:45:09.862 INFO:journalctl@ceph.mgr.a.vm02.stdout:Mar 06 13:45:09 vm02 ceph-2b1b4a72-195a-11f1-b1df-bf1f800d676b-mgr-a[49440]: 2026-03-06T12:45:09.625+0000 7fb523f38100 -1 mgr[py] Module influx has missing NOTIFY_TYPES member 2026-03-06T13:45:10.362 INFO:journalctl@ceph.mgr.a.vm02.stdout:Mar 06 13:45:09 vm02 ceph-2b1b4a72-195a-11f1-b1df-bf1f800d676b-mgr-a[49440]: 2026-03-06T12:45:09.880+0000 7fb523f38100 -1 mgr[py] Module iostat has missing NOTIFY_TYPES member 2026-03-06T13:45:12.112 INFO:journalctl@ceph.mgr.a.vm02.stdout:Mar 06 13:45:11 vm02 ceph-2b1b4a72-195a-11f1-b1df-bf1f800d676b-mgr-a[49440]: 2026-03-06T12:45:11.777+0000 7fb523f38100 -1 mgr[py] Module nfs has missing NOTIFY_TYPES member 2026-03-06T13:45:12.402 INFO:journalctl@ceph.mgr.a.vm02.stdout:Mar 06 13:45:12 vm02 ceph-2b1b4a72-195a-11f1-b1df-bf1f800d676b-mgr-a[49440]: 2026-03-06T12:45:12.137+0000 7fb523f38100 -1 mgr[py] Module orchestrator has missing NOTIFY_TYPES member 2026-03-06T13:45:12.402 INFO:journalctl@ceph.mgr.a.vm02.stdout:Mar 06 13:45:12 vm02 ceph-2b1b4a72-195a-11f1-b1df-bf1f800d676b-mgr-a[49440]: 2026-03-06T12:45:12.277+0000 7fb523f38100 -1 mgr[py] Module osd_perf_query has missing NOTIFY_TYPES member 2026-03-06T13:45:12.670 INFO:journalctl@ceph.mgr.a.vm02.stdout:Mar 06 13:45:12 vm02 ceph-2b1b4a72-195a-11f1-b1df-bf1f800d676b-mgr-a[49440]: 2026-03-06T12:45:12.400+0000 7fb523f38100 -1 mgr[py] Module osd_support has missing NOTIFY_TYPES member 2026-03-06T13:45:12.670 INFO:journalctl@ceph.mgr.a.vm02.stdout:Mar 06 13:45:12 vm02 ceph-2b1b4a72-195a-11f1-b1df-bf1f800d676b-mgr-a[49440]: 2026-03-06T12:45:12.542+0000 7fb523f38100 -1 mgr[py] Module pg_autoscaler has missing NOTIFY_TYPES member 2026-03-06T13:45:13.112 INFO:journalctl@ceph.mgr.a.vm02.stdout:Mar 06 13:45:12 vm02 ceph-2b1b4a72-195a-11f1-b1df-bf1f800d676b-mgr-a[49440]: 2026-03-06T12:45:12.667+0000 7fb523f38100 -1 mgr[py] Module progress has missing NOTIFY_TYPES member 2026-03-06T13:45:13.612 INFO:journalctl@ceph.mgr.a.vm02.stdout:Mar 06 13:45:13 vm02 ceph-2b1b4a72-195a-11f1-b1df-bf1f800d676b-mgr-a[49440]: 2026-03-06T12:45:13.216+0000 7fb523f38100 -1 mgr[py] Module prometheus has missing NOTIFY_TYPES member 2026-03-06T13:45:13.612 INFO:journalctl@ceph.mgr.a.vm02.stdout:Mar 06 13:45:13 vm02 ceph-2b1b4a72-195a-11f1-b1df-bf1f800d676b-mgr-a[49440]: 2026-03-06T12:45:13.372+0000 7fb523f38100 -1 mgr[py] Module rbd_support has missing NOTIFY_TYPES member 2026-03-06T13:45:14.362 INFO:journalctl@ceph.mgr.a.vm02.stdout:Mar 06 13:45:14 vm02 ceph-2b1b4a72-195a-11f1-b1df-bf1f800d676b-mgr-a[49440]: 2026-03-06T12:45:14.103+0000 7fb523f38100 -1 mgr[py] Module rgw has missing NOTIFY_TYPES member 2026-03-06T13:45:15.426 INFO:journalctl@ceph.mgr.a.vm02.stdout:Mar 06 13:45:15 vm02 ceph-2b1b4a72-195a-11f1-b1df-bf1f800d676b-mgr-a[49440]: 2026-03-06T12:45:15.172+0000 7fb523f38100 -1 mgr[py] Module rook has missing NOTIFY_TYPES member 2026-03-06T13:45:15.426 INFO:journalctl@ceph.mgr.a.vm02.stdout:Mar 06 13:45:15 vm02 ceph-2b1b4a72-195a-11f1-b1df-bf1f800d676b-mgr-a[49440]: 2026-03-06T12:45:15.296+0000 7fb523f38100 -1 mgr[py] Module selftest has missing NOTIFY_TYPES member 2026-03-06T13:45:15.683 INFO:journalctl@ceph.mgr.a.vm02.stdout:Mar 06 13:45:15 vm02 ceph-2b1b4a72-195a-11f1-b1df-bf1f800d676b-mgr-a[49440]: 2026-03-06T12:45:15.424+0000 7fb523f38100 -1 mgr[py] Module snap_schedule has missing NOTIFY_TYPES member 2026-03-06T13:45:16.092 INFO:journalctl@ceph.mgr.a.vm02.stdout:Mar 06 13:45:15 vm02 ceph-2b1b4a72-195a-11f1-b1df-bf1f800d676b-mgr-a[49440]: 2026-03-06T12:45:15.681+0000 7fb523f38100 -1 mgr[py] Module status has missing NOTIFY_TYPES member 2026-03-06T13:45:16.092 INFO:journalctl@ceph.mgr.a.vm02.stdout:Mar 06 13:45:15 vm02 ceph-2b1b4a72-195a-11f1-b1df-bf1f800d676b-mgr-a[49440]: 2026-03-06T12:45:15.800+0000 7fb523f38100 -1 mgr[py] Module telegraf has missing NOTIFY_TYPES member 2026-03-06T13:45:16.362 INFO:journalctl@ceph.mgr.a.vm02.stdout:Mar 06 13:45:16 vm02 ceph-2b1b4a72-195a-11f1-b1df-bf1f800d676b-mgr-a[49440]: 2026-03-06T12:45:16.089+0000 7fb523f38100 -1 mgr[py] Module telemetry has missing NOTIFY_TYPES member 2026-03-06T13:45:16.769 INFO:journalctl@ceph.mgr.a.vm02.stdout:Mar 06 13:45:16 vm02 ceph-2b1b4a72-195a-11f1-b1df-bf1f800d676b-mgr-a[49440]: 2026-03-06T12:45:16.421+0000 7fb523f38100 -1 mgr[py] Module test_orchestrator has missing NOTIFY_TYPES member 2026-03-06T13:45:17.112 INFO:journalctl@ceph.mgr.a.vm02.stdout:Mar 06 13:45:16 vm02 ceph-2b1b4a72-195a-11f1-b1df-bf1f800d676b-mgr-a[49440]: 2026-03-06T12:45:16.766+0000 7fb523f38100 -1 mgr[py] Module volumes has missing NOTIFY_TYPES member 2026-03-06T13:45:17.112 INFO:journalctl@ceph.mgr.a.vm02.stdout:Mar 06 13:45:16 vm02 ceph-2b1b4a72-195a-11f1-b1df-bf1f800d676b-mgr-a[49440]: 2026-03-06T12:45:16.882+0000 7fb523f38100 -1 mgr[py] Module zabbix has missing NOTIFY_TYPES member 2026-03-06T13:45:17.862 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:45:17 vm02 ceph-mon[49207]: Active manager daemon a restarted 2026-03-06T13:45:17.862 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:45:17 vm02 ceph-mon[49207]: Activating manager daemon a 2026-03-06T13:45:17.862 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:45:17 vm02 ceph-mon[49207]: osdmap e2: 0 total, 0 up, 0 in 2026-03-06T13:45:17.862 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:45:17 vm02 ceph-mon[49207]: mgrmap e6: a(active, starting, since 0.653286s) 2026-03-06T13:45:17.862 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:45:17 vm02 ceph-mon[49207]: from='mgr.14124 192.168.123.102:0/3391264294' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "a"}]: dispatch 2026-03-06T13:45:17.862 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:45:17 vm02 ceph-mon[49207]: from='mgr.14124 192.168.123.102:0/3391264294' entity='mgr.a' cmd=[{"prefix": "mgr metadata", "who": "a", "id": "a"}]: dispatch 2026-03-06T13:45:17.862 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:45:17 vm02 ceph-mon[49207]: from='mgr.14124 192.168.123.102:0/3391264294' entity='mgr.a' cmd=[{"prefix": "mds metadata"}]: dispatch 2026-03-06T13:45:17.862 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:45:17 vm02 ceph-mon[49207]: from='mgr.14124 192.168.123.102:0/3391264294' entity='mgr.a' cmd=[{"prefix": "osd metadata"}]: dispatch 2026-03-06T13:45:17.862 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:45:17 vm02 ceph-mon[49207]: from='mgr.14124 192.168.123.102:0/3391264294' entity='mgr.a' cmd=[{"prefix": "mon metadata"}]: dispatch 2026-03-06T13:45:17.862 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:45:17 vm02 ceph-mon[49207]: Manager daemon a is now available 2026-03-06T13:45:17.862 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:45:17 vm02 ceph-mon[49207]: from='mgr.14124 192.168.123.102:0/3391264294' entity='mgr.a' 2026-03-06T13:45:17.862 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:45:17 vm02 ceph-mon[49207]: from='mgr.14124 192.168.123.102:0/3391264294' entity='mgr.a' 2026-03-06T13:45:17.862 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:45:17 vm02 ceph-mon[49207]: from='mgr.14124 192.168.123.102:0/3391264294' entity='mgr.a' 2026-03-06T13:45:17.862 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:45:17 vm02 ceph-mon[49207]: from='mgr.14124 192.168.123.102:0/3391264294' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-06T13:45:17.862 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:45:17 vm02 ceph-mon[49207]: from='mgr.14124 192.168.123.102:0/3391264294' entity='mgr.a' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/a/mirror_snapshot_schedule"}]: dispatch 2026-03-06T13:45:17.862 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:45:17 vm02 ceph-mon[49207]: from='mgr.14124 192.168.123.102:0/3391264294' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-06T13:45:18.692 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout { 2026-03-06T13:45:18.692 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "mgrmap_epoch": 7, 2026-03-06T13:45:18.692 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "initialized": true 2026-03-06T13:45:18.692 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout } 2026-03-06T13:45:18.692 INFO:teuthology.orchestra.run.vm02.stdout:mgr epoch 5 is available 2026-03-06T13:45:18.692 INFO:teuthology.orchestra.run.vm02.stdout:Setting orchestrator backend to cephadm... 2026-03-06T13:45:18.799 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:45:18 vm02 ceph-mon[49207]: Found migration_current of "None". Setting to last migration. 2026-03-06T13:45:18.799 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:45:18 vm02 ceph-mon[49207]: from='mgr.14124 192.168.123.102:0/3391264294' entity='mgr.a' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/a/trash_purge_schedule"}]: dispatch 2026-03-06T13:45:18.799 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:45:18 vm02 ceph-mon[49207]: from='mgr.14124 192.168.123.102:0/3391264294' entity='mgr.a' 2026-03-06T13:45:18.799 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:45:18 vm02 ceph-mon[49207]: from='mgr.14124 192.168.123.102:0/3391264294' entity='mgr.a' 2026-03-06T13:45:18.799 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:45:18 vm02 ceph-mon[49207]: mgrmap e7: a(active, since 1.65628s) 2026-03-06T13:45:19.694 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout value unchanged 2026-03-06T13:45:19.694 INFO:teuthology.orchestra.run.vm02.stdout:Generating ssh key... 2026-03-06T13:45:19.808 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:45:19 vm02 ceph-mon[49207]: from='client.14128 -' entity='client.admin' cmd=[{"prefix": "get_command_descriptions"}]: dispatch 2026-03-06T13:45:19.808 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:45:19 vm02 ceph-mon[49207]: from='client.14128 -' entity='client.admin' cmd=[{"prefix": "mgr_status"}]: dispatch 2026-03-06T13:45:19.808 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:45:19 vm02 ceph-mon[49207]: [06/Mar/2026:12:45:18] ENGINE Bus STARTING 2026-03-06T13:45:19.808 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:45:19 vm02 ceph-mon[49207]: from='mgr.14124 192.168.123.102:0/3391264294' entity='mgr.a' 2026-03-06T13:45:19.808 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:45:19 vm02 ceph-mon[49207]: from='mgr.14124 192.168.123.102:0/3391264294' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-06T13:45:19.808 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:45:19 vm02 ceph-mon[49207]: from='mgr.14124 192.168.123.102:0/3391264294' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-06T13:45:20.362 INFO:journalctl@ceph.mgr.a.vm02.stdout:Mar 06 13:45:20 vm02 ceph-2b1b4a72-195a-11f1-b1df-bf1f800d676b-mgr-a[49440]: Generating public/private rsa key pair. 2026-03-06T13:45:20.362 INFO:journalctl@ceph.mgr.a.vm02.stdout:Mar 06 13:45:20 vm02 ceph-2b1b4a72-195a-11f1-b1df-bf1f800d676b-mgr-a[49440]: Your identification has been saved in /tmp/tmp6zl0u4p6/key 2026-03-06T13:45:20.362 INFO:journalctl@ceph.mgr.a.vm02.stdout:Mar 06 13:45:20 vm02 ceph-2b1b4a72-195a-11f1-b1df-bf1f800d676b-mgr-a[49440]: Your public key has been saved in /tmp/tmp6zl0u4p6/key.pub 2026-03-06T13:45:20.362 INFO:journalctl@ceph.mgr.a.vm02.stdout:Mar 06 13:45:20 vm02 ceph-2b1b4a72-195a-11f1-b1df-bf1f800d676b-mgr-a[49440]: The key fingerprint is: 2026-03-06T13:45:20.362 INFO:journalctl@ceph.mgr.a.vm02.stdout:Mar 06 13:45:20 vm02 ceph-2b1b4a72-195a-11f1-b1df-bf1f800d676b-mgr-a[49440]: SHA256:9ox3bTVql3wvyEw5dmTH1yoGo3IS+621Jlg+RknlZ8Q ceph-2b1b4a72-195a-11f1-b1df-bf1f800d676b 2026-03-06T13:45:20.362 INFO:journalctl@ceph.mgr.a.vm02.stdout:Mar 06 13:45:20 vm02 ceph-2b1b4a72-195a-11f1-b1df-bf1f800d676b-mgr-a[49440]: The key's randomart image is: 2026-03-06T13:45:20.362 INFO:journalctl@ceph.mgr.a.vm02.stdout:Mar 06 13:45:20 vm02 ceph-2b1b4a72-195a-11f1-b1df-bf1f800d676b-mgr-a[49440]: +---[RSA 3072]----+ 2026-03-06T13:45:20.362 INFO:journalctl@ceph.mgr.a.vm02.stdout:Mar 06 13:45:20 vm02 ceph-2b1b4a72-195a-11f1-b1df-bf1f800d676b-mgr-a[49440]: | . | 2026-03-06T13:45:20.362 INFO:journalctl@ceph.mgr.a.vm02.stdout:Mar 06 13:45:20 vm02 ceph-2b1b4a72-195a-11f1-b1df-bf1f800d676b-mgr-a[49440]: | . E | 2026-03-06T13:45:20.362 INFO:journalctl@ceph.mgr.a.vm02.stdout:Mar 06 13:45:20 vm02 ceph-2b1b4a72-195a-11f1-b1df-bf1f800d676b-mgr-a[49440]: | o . ..| 2026-03-06T13:45:20.362 INFO:journalctl@ceph.mgr.a.vm02.stdout:Mar 06 13:45:20 vm02 ceph-2b1b4a72-195a-11f1-b1df-bf1f800d676b-mgr-a[49440]: | . . + o o =| 2026-03-06T13:45:20.362 INFO:journalctl@ceph.mgr.a.vm02.stdout:Mar 06 13:45:20 vm02 ceph-2b1b4a72-195a-11f1-b1df-bf1f800d676b-mgr-a[49440]: | +So = + +o| 2026-03-06T13:45:20.362 INFO:journalctl@ceph.mgr.a.vm02.stdout:Mar 06 13:45:20 vm02 ceph-2b1b4a72-195a-11f1-b1df-bf1f800d676b-mgr-a[49440]: | +.*+ B.=.o| 2026-03-06T13:45:20.362 INFO:journalctl@ceph.mgr.a.vm02.stdout:Mar 06 13:45:20 vm02 ceph-2b1b4a72-195a-11f1-b1df-bf1f800d676b-mgr-a[49440]: | X..=*.Bo+.| 2026-03-06T13:45:20.362 INFO:journalctl@ceph.mgr.a.vm02.stdout:Mar 06 13:45:20 vm02 ceph-2b1b4a72-195a-11f1-b1df-bf1f800d676b-mgr-a[49440]: | . *ooo=.o o| 2026-03-06T13:45:20.362 INFO:journalctl@ceph.mgr.a.vm02.stdout:Mar 06 13:45:20 vm02 ceph-2b1b4a72-195a-11f1-b1df-bf1f800d676b-mgr-a[49440]: | ..=. ..| 2026-03-06T13:45:20.362 INFO:journalctl@ceph.mgr.a.vm02.stdout:Mar 06 13:45:20 vm02 ceph-2b1b4a72-195a-11f1-b1df-bf1f800d676b-mgr-a[49440]: +----[SHA256]-----+ 2026-03-06T13:45:20.650 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:45:20 vm02 ceph-mon[49207]: [06/Mar/2026:12:45:18] ENGINE Serving on http://192.168.123.102:8765 2026-03-06T13:45:20.650 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:45:20 vm02 ceph-mon[49207]: from='client.14136 -' entity='client.admin' cmd=[{"prefix": "orch set backend", "module_name": "cephadm", "target": ["mon-mgr", ""]}]: dispatch 2026-03-06T13:45:20.651 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:45:20 vm02 ceph-mon[49207]: [06/Mar/2026:12:45:19] ENGINE Serving on https://192.168.123.102:7150 2026-03-06T13:45:20.651 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:45:20 vm02 ceph-mon[49207]: [06/Mar/2026:12:45:19] ENGINE Bus STARTED 2026-03-06T13:45:20.651 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:45:20 vm02 ceph-mon[49207]: [06/Mar/2026:12:45:19] ENGINE Client ('192.168.123.102', 36630) lost — peer dropped the TLS connection suddenly, during handshake: (6, 'TLS/SSL connection has been closed (EOF) (_ssl.c:1147)') 2026-03-06T13:45:20.651 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:45:20 vm02 ceph-mon[49207]: from='client.14138 -' entity='client.admin' cmd=[{"prefix": "cephadm set-user", "user": "root", "target": ["mon-mgr", ""]}]: dispatch 2026-03-06T13:45:20.651 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:45:20 vm02 ceph-mon[49207]: mgrmap e8: a(active, since 2s) 2026-03-06T13:45:20.651 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:45:20 vm02 ceph-mon[49207]: from='mgr.14124 192.168.123.102:0/3391264294' entity='mgr.a' 2026-03-06T13:45:20.651 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:45:20 vm02 ceph-mon[49207]: from='mgr.14124 192.168.123.102:0/3391264294' entity='mgr.a' 2026-03-06T13:45:20.805 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout ssh-rsa AAAAB3NzaC1yc2EAAAADAQABAAABgQClATzkQHLVv+lvBgYX2LelwgSx3fQOZQpm8sL5mO3SJ3d03vYuUX2QSOxnja2HZhWInnOqLeQzl2fsxAYrGdjZvuP0hm+poU5JRPtQuADk4IOQTCosP3e12eApZcoLI+Nq/PKOoKwaKOQILzt3FqC0l8CmfNDOgK5it9XriUWkzAXDAcYys4boaaPfvhZIK1Ul4dLDJtleZ3uMK+gLegnUc2fywR/iYn7/ZuGWkb35m2JKwobJqYv9Da0tJ480cnvExUeEsk8+lqvpOhyEmIjm+6rRjiMgb6iajjzN0P32AKkoBEC7YdsNwXL4Ur8eZKNf5GZu6NoVXOnQoXAhFkEb035NJ1NZiuKVPs2ksGmtVvAPWjvM4Y5msq+JQdoc5yUk3mCYl30bP79NfMgqPXOyyJwl/T2zUdkrGsGoQE5CMsjyTgpAZr7ehVsEaEvFk67dQf8rha4S8SDcS2GxzLeiV2nue3vLtH6JlD+pAMaz4BSOToxPwteuZJ/uP5yuIvM= ceph-2b1b4a72-195a-11f1-b1df-bf1f800d676b 2026-03-06T13:45:20.805 INFO:teuthology.orchestra.run.vm02.stdout:Wrote public SSH key to /home/ubuntu/cephtest/ceph.pub 2026-03-06T13:45:20.805 INFO:teuthology.orchestra.run.vm02.stdout:Adding key to root@localhost authorized_keys... 2026-03-06T13:45:20.805 INFO:teuthology.orchestra.run.vm02.stdout:Adding host vm02... 2026-03-06T13:45:21.862 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:45:21 vm02 ceph-mon[49207]: from='client.14140 -' entity='client.admin' cmd=[{"prefix": "cephadm generate-key", "target": ["mon-mgr", ""]}]: dispatch 2026-03-06T13:45:21.862 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:45:21 vm02 ceph-mon[49207]: Generating ssh key... 2026-03-06T13:45:21.862 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:45:21 vm02 ceph-mon[49207]: from='client.14142 -' entity='client.admin' cmd=[{"prefix": "cephadm get-pub-key", "target": ["mon-mgr", ""]}]: dispatch 2026-03-06T13:45:22.862 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:45:22 vm02 ceph-mon[49207]: from='client.14144 -' entity='client.admin' cmd=[{"prefix": "orch host add", "hostname": "vm02", "addr": "192.168.123.102", "target": ["mon-mgr", ""]}]: dispatch 2026-03-06T13:45:22.862 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:45:22 vm02 ceph-mon[49207]: Deploying cephadm binary to vm02 2026-03-06T13:45:23.715 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout Added host 'vm02' with addr '192.168.123.102' 2026-03-06T13:45:23.715 INFO:teuthology.orchestra.run.vm02.stdout:Deploying unmanaged mon service... 2026-03-06T13:45:24.252 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout Scheduled mon update... 2026-03-06T13:45:24.253 INFO:teuthology.orchestra.run.vm02.stdout:Deploying unmanaged mgr service... 2026-03-06T13:45:24.626 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:45:24 vm02 ceph-mon[49207]: from='mgr.14124 192.168.123.102:0/3391264294' entity='mgr.a' 2026-03-06T13:45:24.627 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:45:24 vm02 ceph-mon[49207]: Added host vm02 2026-03-06T13:45:24.627 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:45:24 vm02 ceph-mon[49207]: from='mgr.14124 192.168.123.102:0/3391264294' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-06T13:45:24.627 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:45:24 vm02 ceph-mon[49207]: from='mgr.14124 192.168.123.102:0/3391264294' entity='mgr.a' 2026-03-06T13:45:24.738 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout Scheduled mgr update... 2026-03-06T13:45:25.595 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:45:25 vm02 ceph-mon[49207]: from='client.14146 -' entity='client.admin' cmd=[{"prefix": "orch apply", "service_type": "mon", "unmanaged": true, "target": ["mon-mgr", ""]}]: dispatch 2026-03-06T13:45:25.595 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:45:25 vm02 ceph-mon[49207]: Saving service mon spec with placement count:5 2026-03-06T13:45:25.595 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:45:25 vm02 ceph-mon[49207]: from='client.14148 -' entity='client.admin' cmd=[{"prefix": "orch apply", "service_type": "mgr", "unmanaged": true, "target": ["mon-mgr", ""]}]: dispatch 2026-03-06T13:45:25.595 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:45:25 vm02 ceph-mon[49207]: Saving service mgr spec with placement count:2 2026-03-06T13:45:25.595 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:45:25 vm02 ceph-mon[49207]: from='mgr.14124 192.168.123.102:0/3391264294' entity='mgr.a' 2026-03-06T13:45:25.595 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:45:25 vm02 ceph-mon[49207]: from='mgr.14124 192.168.123.102:0/3391264294' entity='mgr.a' 2026-03-06T13:45:25.595 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:45:25 vm02 ceph-mon[49207]: from='client.? 192.168.123.102:0/436004875' entity='client.admin' 2026-03-06T13:45:25.595 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:45:25 vm02 ceph-mon[49207]: from='mgr.14124 192.168.123.102:0/3391264294' entity='mgr.a' 2026-03-06T13:45:25.777 INFO:teuthology.orchestra.run.vm02.stdout:Enabling the dashboard module... 2026-03-06T13:45:26.916 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:45:26 vm02 ceph-mon[49207]: from='client.? 192.168.123.102:0/1080732264' entity='client.admin' 2026-03-06T13:45:26.916 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:45:26 vm02 ceph-mon[49207]: from='client.? 192.168.123.102:0/137948838' entity='client.admin' cmd=[{"prefix": "mgr module enable", "module": "dashboard"}]: dispatch 2026-03-06T13:45:26.916 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:45:26 vm02 ceph-mon[49207]: from='mgr.14124 192.168.123.102:0/3391264294' entity='mgr.a' 2026-03-06T13:45:26.916 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:45:26 vm02 ceph-mon[49207]: from='mgr.14124 192.168.123.102:0/3391264294' entity='mgr.a' 2026-03-06T13:45:26.916 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:45:26 vm02 ceph-mon[49207]: from='mgr.14124 192.168.123.102:0/3391264294' entity='mgr.a' cmd=[{"prefix": "config rm", "who": "osd/host:vm02", "name": "osd_memory_target"}]: dispatch 2026-03-06T13:45:26.916 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:45:26 vm02 ceph-mon[49207]: from='mgr.14124 192.168.123.102:0/3391264294' entity='mgr.a' 2026-03-06T13:45:26.916 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:45:26 vm02 ceph-mon[49207]: from='mgr.14124 192.168.123.102:0/3391264294' entity='mgr.a' cmd=[{"prefix": "auth get-or-create", "entity": "client.agent.vm02", "caps": []}]: dispatch 2026-03-06T13:45:26.916 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:45:26 vm02 ceph-mon[49207]: from='mgr.14124 192.168.123.102:0/3391264294' entity='mgr.a' cmd='[{"prefix": "auth get-or-create", "entity": "client.agent.vm02", "caps": []}]': finished 2026-03-06T13:45:26.917 INFO:journalctl@ceph.mgr.a.vm02.stdout:Mar 06 13:45:26 vm02 ceph-2b1b4a72-195a-11f1-b1df-bf1f800d676b-mgr-a[49440]: ignoring --setuser ceph since I am not root 2026-03-06T13:45:26.917 INFO:journalctl@ceph.mgr.a.vm02.stdout:Mar 06 13:45:26 vm02 ceph-2b1b4a72-195a-11f1-b1df-bf1f800d676b-mgr-a[49440]: ignoring --setgroup ceph since I am not root 2026-03-06T13:45:27.218 INFO:journalctl@ceph.mgr.a.vm02.stdout:Mar 06 13:45:26 vm02 ceph-2b1b4a72-195a-11f1-b1df-bf1f800d676b-mgr-a[49440]: 2026-03-06T12:45:26.973+0000 7f3e8b27f100 -1 mgr[py] Module alerts has missing NOTIFY_TYPES member 2026-03-06T13:45:27.218 INFO:journalctl@ceph.mgr.a.vm02.stdout:Mar 06 13:45:27 vm02 ceph-2b1b4a72-195a-11f1-b1df-bf1f800d676b-mgr-a[49440]: 2026-03-06T12:45:27.114+0000 7f3e8b27f100 -1 mgr[py] Module balancer has missing NOTIFY_TYPES member 2026-03-06T13:45:27.361 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout { 2026-03-06T13:45:27.361 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "epoch": 9, 2026-03-06T13:45:27.361 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "available": true, 2026-03-06T13:45:27.361 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "active_name": "a", 2026-03-06T13:45:27.361 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "num_standby": 0 2026-03-06T13:45:27.361 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout } 2026-03-06T13:45:27.361 INFO:teuthology.orchestra.run.vm02.stdout:Waiting for the mgr to restart... 2026-03-06T13:45:27.361 INFO:teuthology.orchestra.run.vm02.stdout:Waiting for mgr epoch 9... 2026-03-06T13:45:27.862 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:45:27 vm02 ceph-mon[49207]: from='client.? 192.168.123.102:0/137948838' entity='client.admin' cmd='[{"prefix": "mgr module enable", "module": "dashboard"}]': finished 2026-03-06T13:45:27.862 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:45:27 vm02 ceph-mon[49207]: mgrmap e9: a(active, since 9s) 2026-03-06T13:45:27.862 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:45:27 vm02 ceph-mon[49207]: from='client.? 192.168.123.102:0/844318749' entity='client.admin' cmd=[{"prefix": "mgr stat"}]: dispatch 2026-03-06T13:45:28.612 INFO:journalctl@ceph.mgr.a.vm02.stdout:Mar 06 13:45:28 vm02 ceph-2b1b4a72-195a-11f1-b1df-bf1f800d676b-mgr-a[49440]: 2026-03-06T12:45:28.274+0000 7f3e8b27f100 -1 mgr[py] Module crash has missing NOTIFY_TYPES member 2026-03-06T13:45:29.524 INFO:journalctl@ceph.mgr.a.vm02.stdout:Mar 06 13:45:29 vm02 ceph-2b1b4a72-195a-11f1-b1df-bf1f800d676b-mgr-a[49440]: 2026-03-06T12:45:29.156+0000 7f3e8b27f100 -1 mgr[py] Module devicehealth has missing NOTIFY_TYPES member 2026-03-06T13:45:29.524 INFO:journalctl@ceph.mgr.a.vm02.stdout:Mar 06 13:45:29 vm02 ceph-2b1b4a72-195a-11f1-b1df-bf1f800d676b-mgr-a[49440]: 2026-03-06T12:45:29.278+0000 7f3e8b27f100 -1 mgr[py] Module influx has missing NOTIFY_TYPES member 2026-03-06T13:45:29.862 INFO:journalctl@ceph.mgr.a.vm02.stdout:Mar 06 13:45:29 vm02 ceph-2b1b4a72-195a-11f1-b1df-bf1f800d676b-mgr-a[49440]: 2026-03-06T12:45:29.522+0000 7f3e8b27f100 -1 mgr[py] Module iostat has missing NOTIFY_TYPES member 2026-03-06T13:45:31.739 INFO:journalctl@ceph.mgr.a.vm02.stdout:Mar 06 13:45:31 vm02 ceph-2b1b4a72-195a-11f1-b1df-bf1f800d676b-mgr-a[49440]: 2026-03-06T12:45:31.373+0000 7f3e8b27f100 -1 mgr[py] Module nfs has missing NOTIFY_TYPES member 2026-03-06T13:45:31.739 INFO:journalctl@ceph.mgr.a.vm02.stdout:Mar 06 13:45:31 vm02 ceph-2b1b4a72-195a-11f1-b1df-bf1f800d676b-mgr-a[49440]: 2026-03-06T12:45:31.736+0000 7f3e8b27f100 -1 mgr[py] Module orchestrator has missing NOTIFY_TYPES member 2026-03-06T13:45:32.001 INFO:journalctl@ceph.mgr.a.vm02.stdout:Mar 06 13:45:31 vm02 ceph-2b1b4a72-195a-11f1-b1df-bf1f800d676b-mgr-a[49440]: 2026-03-06T12:45:31.877+0000 7f3e8b27f100 -1 mgr[py] Module osd_perf_query has missing NOTIFY_TYPES member 2026-03-06T13:45:32.001 INFO:journalctl@ceph.mgr.a.vm02.stdout:Mar 06 13:45:31 vm02 ceph-2b1b4a72-195a-11f1-b1df-bf1f800d676b-mgr-a[49440]: 2026-03-06T12:45:31.998+0000 7f3e8b27f100 -1 mgr[py] Module osd_support has missing NOTIFY_TYPES member 2026-03-06T13:45:32.268 INFO:journalctl@ceph.mgr.a.vm02.stdout:Mar 06 13:45:32 vm02 ceph-2b1b4a72-195a-11f1-b1df-bf1f800d676b-mgr-a[49440]: 2026-03-06T12:45:32.140+0000 7f3e8b27f100 -1 mgr[py] Module pg_autoscaler has missing NOTIFY_TYPES member 2026-03-06T13:45:32.268 INFO:journalctl@ceph.mgr.a.vm02.stdout:Mar 06 13:45:32 vm02 ceph-2b1b4a72-195a-11f1-b1df-bf1f800d676b-mgr-a[49440]: 2026-03-06T12:45:32.265+0000 7f3e8b27f100 -1 mgr[py] Module progress has missing NOTIFY_TYPES member 2026-03-06T13:45:33.112 INFO:journalctl@ceph.mgr.a.vm02.stdout:Mar 06 13:45:32 vm02 ceph-2b1b4a72-195a-11f1-b1df-bf1f800d676b-mgr-a[49440]: 2026-03-06T12:45:32.778+0000 7f3e8b27f100 -1 mgr[py] Module prometheus has missing NOTIFY_TYPES member 2026-03-06T13:45:33.112 INFO:journalctl@ceph.mgr.a.vm02.stdout:Mar 06 13:45:32 vm02 ceph-2b1b4a72-195a-11f1-b1df-bf1f800d676b-mgr-a[49440]: 2026-03-06T12:45:32.943+0000 7f3e8b27f100 -1 mgr[py] Module rbd_support has missing NOTIFY_TYPES member 2026-03-06T13:45:34.112 INFO:journalctl@ceph.mgr.a.vm02.stdout:Mar 06 13:45:33 vm02 ceph-2b1b4a72-195a-11f1-b1df-bf1f800d676b-mgr-a[49440]: 2026-03-06T12:45:33.695+0000 7f3e8b27f100 -1 mgr[py] Module rgw has missing NOTIFY_TYPES member 2026-03-06T13:45:35.019 INFO:journalctl@ceph.mgr.a.vm02.stdout:Mar 06 13:45:34 vm02 ceph-2b1b4a72-195a-11f1-b1df-bf1f800d676b-mgr-a[49440]: 2026-03-06T12:45:34.759+0000 7f3e8b27f100 -1 mgr[py] Module rook has missing NOTIFY_TYPES member 2026-03-06T13:45:35.019 INFO:journalctl@ceph.mgr.a.vm02.stdout:Mar 06 13:45:34 vm02 ceph-2b1b4a72-195a-11f1-b1df-bf1f800d676b-mgr-a[49440]: 2026-03-06T12:45:34.888+0000 7f3e8b27f100 -1 mgr[py] Module selftest has missing NOTIFY_TYPES member 2026-03-06T13:45:35.019 INFO:journalctl@ceph.mgr.a.vm02.stdout:Mar 06 13:45:35 vm02 ceph-2b1b4a72-195a-11f1-b1df-bf1f800d676b-mgr-a[49440]: 2026-03-06T12:45:35.016+0000 7f3e8b27f100 -1 mgr[py] Module snap_schedule has missing NOTIFY_TYPES member 2026-03-06T13:45:35.288 INFO:journalctl@ceph.mgr.a.vm02.stdout:Mar 06 13:45:35 vm02 ceph-2b1b4a72-195a-11f1-b1df-bf1f800d676b-mgr-a[49440]: 2026-03-06T12:45:35.286+0000 7f3e8b27f100 -1 mgr[py] Module status has missing NOTIFY_TYPES member 2026-03-06T13:45:35.612 INFO:journalctl@ceph.mgr.a.vm02.stdout:Mar 06 13:45:35 vm02 ceph-2b1b4a72-195a-11f1-b1df-bf1f800d676b-mgr-a[49440]: 2026-03-06T12:45:35.414+0000 7f3e8b27f100 -1 mgr[py] Module telegraf has missing NOTIFY_TYPES member 2026-03-06T13:45:36.066 INFO:journalctl@ceph.mgr.a.vm02.stdout:Mar 06 13:45:35 vm02 ceph-2b1b4a72-195a-11f1-b1df-bf1f800d676b-mgr-a[49440]: 2026-03-06T12:45:35.732+0000 7f3e8b27f100 -1 mgr[py] Module telemetry has missing NOTIFY_TYPES member 2026-03-06T13:45:36.066 INFO:journalctl@ceph.mgr.a.vm02.stdout:Mar 06 13:45:36 vm02 ceph-2b1b4a72-195a-11f1-b1df-bf1f800d676b-mgr-a[49440]: 2026-03-06T12:45:36.063+0000 7f3e8b27f100 -1 mgr[py] Module test_orchestrator has missing NOTIFY_TYPES member 2026-03-06T13:45:36.724 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:45:36 vm02 ceph-mon[49207]: Active manager daemon a restarted 2026-03-06T13:45:36.724 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:45:36 vm02 ceph-mon[49207]: Activating manager daemon a 2026-03-06T13:45:36.724 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:45:36 vm02 ceph-mon[49207]: osdmap e3: 0 total, 0 up, 0 in 2026-03-06T13:45:36.724 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:45:36 vm02 ceph-mon[49207]: mgrmap e10: a(active, starting, since 0.017421s) 2026-03-06T13:45:36.724 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:45:36 vm02 ceph-mon[49207]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "a"}]: dispatch 2026-03-06T13:45:36.724 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:45:36 vm02 ceph-mon[49207]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' cmd=[{"prefix": "mgr metadata", "who": "a", "id": "a"}]: dispatch 2026-03-06T13:45:36.724 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:45:36 vm02 ceph-mon[49207]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' cmd=[{"prefix": "mds metadata"}]: dispatch 2026-03-06T13:45:36.725 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:45:36 vm02 ceph-mon[49207]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' cmd=[{"prefix": "osd metadata"}]: dispatch 2026-03-06T13:45:36.725 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:45:36 vm02 ceph-mon[49207]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' cmd=[{"prefix": "mon metadata"}]: dispatch 2026-03-06T13:45:36.725 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:45:36 vm02 ceph-mon[49207]: Manager daemon a is now available 2026-03-06T13:45:36.725 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:45:36 vm02 ceph-mon[49207]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:45:36.725 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:45:36 vm02 ceph-mon[49207]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-06T13:45:36.725 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:45:36 vm02 ceph-mon[49207]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/a/mirror_snapshot_schedule"}]: dispatch 2026-03-06T13:45:36.725 INFO:journalctl@ceph.mgr.a.vm02.stdout:Mar 06 13:45:36 vm02 ceph-2b1b4a72-195a-11f1-b1df-bf1f800d676b-mgr-a[49440]: 2026-03-06T12:45:36.441+0000 7f3e8b27f100 -1 mgr[py] Module volumes has missing NOTIFY_TYPES member 2026-03-06T13:45:36.725 INFO:journalctl@ceph.mgr.a.vm02.stdout:Mar 06 13:45:36 vm02 ceph-2b1b4a72-195a-11f1-b1df-bf1f800d676b-mgr-a[49440]: 2026-03-06T12:45:36.561+0000 7f3e8b27f100 -1 mgr[py] Module zabbix has missing NOTIFY_TYPES member 2026-03-06T13:45:37.731 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout { 2026-03-06T13:45:37.731 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "mgrmap_epoch": 10, 2026-03-06T13:45:37.731 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "initialized": true 2026-03-06T13:45:37.731 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout } 2026-03-06T13:45:37.731 INFO:teuthology.orchestra.run.vm02.stdout:mgr epoch 9 is available 2026-03-06T13:45:37.731 INFO:teuthology.orchestra.run.vm02.stdout:Generating a dashboard self-signed certificate... 2026-03-06T13:45:37.850 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:45:37 vm02 ceph-mon[49207]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/a/trash_purge_schedule"}]: dispatch 2026-03-06T13:45:37.850 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:45:37 vm02 ceph-mon[49207]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:45:37.850 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:45:37 vm02 ceph-mon[49207]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' cmd=[{"prefix": "config rm", "who": "osd/host:vm02", "name": "osd_memory_target"}]: dispatch 2026-03-06T13:45:37.850 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:45:37 vm02 ceph-mon[49207]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:45:37.851 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:45:37 vm02 ceph-mon[49207]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' cmd=[{"prefix": "auth get-or-create", "entity": "client.agent.vm02", "caps": []}]: dispatch 2026-03-06T13:45:37.851 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:45:37 vm02 ceph-mon[49207]: mgrmap e11: a(active, since 1.03014s) 2026-03-06T13:45:38.321 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout Self-signed certificate created 2026-03-06T13:45:38.321 INFO:teuthology.orchestra.run.vm02.stdout:Creating initial admin user... 2026-03-06T13:45:38.818 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:45:38 vm02 ceph-mon[49207]: [06/Mar/2026:12:45:37] ENGINE Bus STARTING 2026-03-06T13:45:38.818 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:45:38 vm02 ceph-mon[49207]: [06/Mar/2026:12:45:37] ENGINE Serving on https://192.168.123.102:7150 2026-03-06T13:45:38.818 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:45:38 vm02 ceph-mon[49207]: [06/Mar/2026:12:45:37] ENGINE Client ('192.168.123.102', 46858) lost — peer dropped the TLS connection suddenly, during handshake: (6, 'TLS/SSL connection has been closed (EOF) (_ssl.c:1147)') 2026-03-06T13:45:38.818 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:45:38 vm02 ceph-mon[49207]: [06/Mar/2026:12:45:37] ENGINE Serving on http://192.168.123.102:8765 2026-03-06T13:45:38.818 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:45:38 vm02 ceph-mon[49207]: [06/Mar/2026:12:45:37] ENGINE Bus STARTED 2026-03-06T13:45:38.818 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:45:38 vm02 ceph-mon[49207]: from='client.14160 -' entity='client.admin' cmd=[{"prefix": "get_command_descriptions"}]: dispatch 2026-03-06T13:45:38.818 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:45:38 vm02 ceph-mon[49207]: from='client.14160 -' entity='client.admin' cmd=[{"prefix": "mgr_status"}]: dispatch 2026-03-06T13:45:38.818 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:45:38 vm02 ceph-mon[49207]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:45:38.818 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:45:38 vm02 ceph-mon[49207]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:45:38.956 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout {"username": "admin", "password": "$2b$12$tcNmr1iAzEt.OGIVpH8HGePUzXCjvChYULU77/qwFs/g3qvU6ijd.", "roles": ["administrator"], "name": null, "email": null, "lastUpdate": 1772801138, "enabled": true, "pwdExpirationDate": null, "pwdUpdateRequired": true} 2026-03-06T13:45:38.956 INFO:teuthology.orchestra.run.vm02.stdout:Fetching dashboard port number... 2026-03-06T13:45:39.450 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout 8443 2026-03-06T13:45:39.450 INFO:teuthology.orchestra.run.vm02.stdout:firewalld does not appear to be present 2026-03-06T13:45:39.450 INFO:teuthology.orchestra.run.vm02.stdout:Not possible to open ports <[8443]>. firewalld.service is not available 2026-03-06T13:45:39.452 INFO:teuthology.orchestra.run.vm02.stdout:Ceph Dashboard is now available at: 2026-03-06T13:45:39.452 INFO:teuthology.orchestra.run.vm02.stdout: 2026-03-06T13:45:39.452 INFO:teuthology.orchestra.run.vm02.stdout: URL: https://vm02.local:8443/ 2026-03-06T13:45:39.452 INFO:teuthology.orchestra.run.vm02.stdout: User: admin 2026-03-06T13:45:39.452 INFO:teuthology.orchestra.run.vm02.stdout: Password: ztfzl032zz 2026-03-06T13:45:39.452 INFO:teuthology.orchestra.run.vm02.stdout: 2026-03-06T13:45:39.452 INFO:teuthology.orchestra.run.vm02.stdout:Saving cluster configuration to /var/lib/ceph/2b1b4a72-195a-11f1-b1df-bf1f800d676b/config directory 2026-03-06T13:45:40.036 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:45:39 vm02 ceph-mon[49207]: Deploying daemon agent.vm02 on vm02 2026-03-06T13:45:40.036 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:45:39 vm02 ceph-mon[49207]: from='client.14168 -' entity='client.admin' cmd=[{"prefix": "dashboard create-self-signed-cert", "target": ["mon-mgr", ""]}]: dispatch 2026-03-06T13:45:40.036 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:45:39 vm02 ceph-mon[49207]: from='client.14170 -' entity='client.admin' cmd=[{"prefix": "dashboard ac-user-create", "username": "admin", "rolename": "administrator", "force_password": true, "pwd_update_required": true, "target": ["mon-mgr", ""]}]: dispatch 2026-03-06T13:45:40.036 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:45:39 vm02 ceph-mon[49207]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:45:40.036 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:45:39 vm02 ceph-mon[49207]: mgrmap e12: a(active, since 2s) 2026-03-06T13:45:40.036 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:45:39 vm02 ceph-mon[49207]: from='client.? 192.168.123.102:0/2801111873' entity='client.admin' cmd=[{"prefix": "config get", "who": "mgr", "key": "mgr/dashboard/ssl_server_port"}]: dispatch 2026-03-06T13:45:40.164 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stderr set mgr/dashboard/cluster/status 2026-03-06T13:45:40.164 INFO:teuthology.orchestra.run.vm02.stdout:You can access the Ceph CLI as following in case of multi-cluster or non-default config: 2026-03-06T13:45:40.164 INFO:teuthology.orchestra.run.vm02.stdout: 2026-03-06T13:45:40.164 INFO:teuthology.orchestra.run.vm02.stdout: sudo /home/ubuntu/cephtest/cephadm shell --fsid 2b1b4a72-195a-11f1-b1df-bf1f800d676b -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring 2026-03-06T13:45:40.164 INFO:teuthology.orchestra.run.vm02.stdout: 2026-03-06T13:45:40.164 INFO:teuthology.orchestra.run.vm02.stdout:Or, if you are only running a single cluster on this host: 2026-03-06T13:45:40.164 INFO:teuthology.orchestra.run.vm02.stdout: 2026-03-06T13:45:40.164 INFO:teuthology.orchestra.run.vm02.stdout: sudo /home/ubuntu/cephtest/cephadm shell 2026-03-06T13:45:40.164 INFO:teuthology.orchestra.run.vm02.stdout: 2026-03-06T13:45:40.164 INFO:teuthology.orchestra.run.vm02.stdout:Please consider enabling telemetry to help improve Ceph: 2026-03-06T13:45:40.164 INFO:teuthology.orchestra.run.vm02.stdout: 2026-03-06T13:45:40.164 INFO:teuthology.orchestra.run.vm02.stdout: ceph telemetry on 2026-03-06T13:45:40.164 INFO:teuthology.orchestra.run.vm02.stdout: 2026-03-06T13:45:40.164 INFO:teuthology.orchestra.run.vm02.stdout:For more information see: 2026-03-06T13:45:40.164 INFO:teuthology.orchestra.run.vm02.stdout: 2026-03-06T13:45:40.164 INFO:teuthology.orchestra.run.vm02.stdout: https://docs.ceph.com/en/latest/mgr/telemetry/ 2026-03-06T13:45:40.164 INFO:teuthology.orchestra.run.vm02.stdout: 2026-03-06T13:45:40.164 INFO:teuthology.orchestra.run.vm02.stdout:Bootstrap complete. 2026-03-06T13:45:40.200 INFO:tasks.cephadm:Fetching config... 2026-03-06T13:45:40.200 DEBUG:teuthology.orchestra.run.vm02:> set -ex 2026-03-06T13:45:40.200 DEBUG:teuthology.orchestra.run.vm02:> dd if=/etc/ceph/ceph.conf of=/dev/stdout 2026-03-06T13:45:40.215 INFO:tasks.cephadm:Fetching client.admin keyring... 2026-03-06T13:45:40.215 DEBUG:teuthology.orchestra.run.vm02:> set -ex 2026-03-06T13:45:40.215 DEBUG:teuthology.orchestra.run.vm02:> dd if=/etc/ceph/ceph.client.admin.keyring of=/dev/stdout 2026-03-06T13:45:40.289 INFO:tasks.cephadm:Fetching mon keyring... 2026-03-06T13:45:40.289 DEBUG:teuthology.orchestra.run.vm02:> set -ex 2026-03-06T13:45:40.289 DEBUG:teuthology.orchestra.run.vm02:> sudo dd if=/var/lib/ceph/2b1b4a72-195a-11f1-b1df-bf1f800d676b/mon.a/keyring of=/dev/stdout 2026-03-06T13:45:40.382 INFO:tasks.cephadm:Fetching pub ssh key... 2026-03-06T13:45:40.382 DEBUG:teuthology.orchestra.run.vm02:> set -ex 2026-03-06T13:45:40.382 DEBUG:teuthology.orchestra.run.vm02:> dd if=/home/ubuntu/cephtest/ceph.pub of=/dev/stdout 2026-03-06T13:45:40.452 INFO:tasks.cephadm:Installing pub ssh key for root users... 2026-03-06T13:45:40.452 DEBUG:teuthology.orchestra.run.vm02:> sudo install -d -m 0700 /root/.ssh && echo 'ssh-rsa AAAAB3NzaC1yc2EAAAADAQABAAABgQClATzkQHLVv+lvBgYX2LelwgSx3fQOZQpm8sL5mO3SJ3d03vYuUX2QSOxnja2HZhWInnOqLeQzl2fsxAYrGdjZvuP0hm+poU5JRPtQuADk4IOQTCosP3e12eApZcoLI+Nq/PKOoKwaKOQILzt3FqC0l8CmfNDOgK5it9XriUWkzAXDAcYys4boaaPfvhZIK1Ul4dLDJtleZ3uMK+gLegnUc2fywR/iYn7/ZuGWkb35m2JKwobJqYv9Da0tJ480cnvExUeEsk8+lqvpOhyEmIjm+6rRjiMgb6iajjzN0P32AKkoBEC7YdsNwXL4Ur8eZKNf5GZu6NoVXOnQoXAhFkEb035NJ1NZiuKVPs2ksGmtVvAPWjvM4Y5msq+JQdoc5yUk3mCYl30bP79NfMgqPXOyyJwl/T2zUdkrGsGoQE5CMsjyTgpAZr7ehVsEaEvFk67dQf8rha4S8SDcS2GxzLeiV2nue3vLtH6JlD+pAMaz4BSOToxPwteuZJ/uP5yuIvM= ceph-2b1b4a72-195a-11f1-b1df-bf1f800d676b' | sudo tee -a /root/.ssh/authorized_keys && sudo chmod 0600 /root/.ssh/authorized_keys 2026-03-06T13:45:40.580 INFO:teuthology.orchestra.run.vm02.stdout:ssh-rsa AAAAB3NzaC1yc2EAAAADAQABAAABgQClATzkQHLVv+lvBgYX2LelwgSx3fQOZQpm8sL5mO3SJ3d03vYuUX2QSOxnja2HZhWInnOqLeQzl2fsxAYrGdjZvuP0hm+poU5JRPtQuADk4IOQTCosP3e12eApZcoLI+Nq/PKOoKwaKOQILzt3FqC0l8CmfNDOgK5it9XriUWkzAXDAcYys4boaaPfvhZIK1Ul4dLDJtleZ3uMK+gLegnUc2fywR/iYn7/ZuGWkb35m2JKwobJqYv9Da0tJ480cnvExUeEsk8+lqvpOhyEmIjm+6rRjiMgb6iajjzN0P32AKkoBEC7YdsNwXL4Ur8eZKNf5GZu6NoVXOnQoXAhFkEb035NJ1NZiuKVPs2ksGmtVvAPWjvM4Y5msq+JQdoc5yUk3mCYl30bP79NfMgqPXOyyJwl/T2zUdkrGsGoQE5CMsjyTgpAZr7ehVsEaEvFk67dQf8rha4S8SDcS2GxzLeiV2nue3vLtH6JlD+pAMaz4BSOToxPwteuZJ/uP5yuIvM= ceph-2b1b4a72-195a-11f1-b1df-bf1f800d676b 2026-03-06T13:45:40.597 DEBUG:teuthology.orchestra.run.vm04:> sudo install -d -m 0700 /root/.ssh && echo 'ssh-rsa AAAAB3NzaC1yc2EAAAADAQABAAABgQClATzkQHLVv+lvBgYX2LelwgSx3fQOZQpm8sL5mO3SJ3d03vYuUX2QSOxnja2HZhWInnOqLeQzl2fsxAYrGdjZvuP0hm+poU5JRPtQuADk4IOQTCosP3e12eApZcoLI+Nq/PKOoKwaKOQILzt3FqC0l8CmfNDOgK5it9XriUWkzAXDAcYys4boaaPfvhZIK1Ul4dLDJtleZ3uMK+gLegnUc2fywR/iYn7/ZuGWkb35m2JKwobJqYv9Da0tJ480cnvExUeEsk8+lqvpOhyEmIjm+6rRjiMgb6iajjzN0P32AKkoBEC7YdsNwXL4Ur8eZKNf5GZu6NoVXOnQoXAhFkEb035NJ1NZiuKVPs2ksGmtVvAPWjvM4Y5msq+JQdoc5yUk3mCYl30bP79NfMgqPXOyyJwl/T2zUdkrGsGoQE5CMsjyTgpAZr7ehVsEaEvFk67dQf8rha4S8SDcS2GxzLeiV2nue3vLtH6JlD+pAMaz4BSOToxPwteuZJ/uP5yuIvM= ceph-2b1b4a72-195a-11f1-b1df-bf1f800d676b' | sudo tee -a /root/.ssh/authorized_keys && sudo chmod 0600 /root/.ssh/authorized_keys 2026-03-06T13:45:40.629 INFO:teuthology.orchestra.run.vm04.stdout:ssh-rsa AAAAB3NzaC1yc2EAAAADAQABAAABgQClATzkQHLVv+lvBgYX2LelwgSx3fQOZQpm8sL5mO3SJ3d03vYuUX2QSOxnja2HZhWInnOqLeQzl2fsxAYrGdjZvuP0hm+poU5JRPtQuADk4IOQTCosP3e12eApZcoLI+Nq/PKOoKwaKOQILzt3FqC0l8CmfNDOgK5it9XriUWkzAXDAcYys4boaaPfvhZIK1Ul4dLDJtleZ3uMK+gLegnUc2fywR/iYn7/ZuGWkb35m2JKwobJqYv9Da0tJ480cnvExUeEsk8+lqvpOhyEmIjm+6rRjiMgb6iajjzN0P32AKkoBEC7YdsNwXL4Ur8eZKNf5GZu6NoVXOnQoXAhFkEb035NJ1NZiuKVPs2ksGmtVvAPWjvM4Y5msq+JQdoc5yUk3mCYl30bP79NfMgqPXOyyJwl/T2zUdkrGsGoQE5CMsjyTgpAZr7ehVsEaEvFk67dQf8rha4S8SDcS2GxzLeiV2nue3vLtH6JlD+pAMaz4BSOToxPwteuZJ/uP5yuIvM= ceph-2b1b4a72-195a-11f1-b1df-bf1f800d676b 2026-03-06T13:45:40.641 DEBUG:teuthology.orchestra.run.vm06:> sudo install -d -m 0700 /root/.ssh && echo 'ssh-rsa AAAAB3NzaC1yc2EAAAADAQABAAABgQClATzkQHLVv+lvBgYX2LelwgSx3fQOZQpm8sL5mO3SJ3d03vYuUX2QSOxnja2HZhWInnOqLeQzl2fsxAYrGdjZvuP0hm+poU5JRPtQuADk4IOQTCosP3e12eApZcoLI+Nq/PKOoKwaKOQILzt3FqC0l8CmfNDOgK5it9XriUWkzAXDAcYys4boaaPfvhZIK1Ul4dLDJtleZ3uMK+gLegnUc2fywR/iYn7/ZuGWkb35m2JKwobJqYv9Da0tJ480cnvExUeEsk8+lqvpOhyEmIjm+6rRjiMgb6iajjzN0P32AKkoBEC7YdsNwXL4Ur8eZKNf5GZu6NoVXOnQoXAhFkEb035NJ1NZiuKVPs2ksGmtVvAPWjvM4Y5msq+JQdoc5yUk3mCYl30bP79NfMgqPXOyyJwl/T2zUdkrGsGoQE5CMsjyTgpAZr7ehVsEaEvFk67dQf8rha4S8SDcS2GxzLeiV2nue3vLtH6JlD+pAMaz4BSOToxPwteuZJ/uP5yuIvM= ceph-2b1b4a72-195a-11f1-b1df-bf1f800d676b' | sudo tee -a /root/.ssh/authorized_keys && sudo chmod 0600 /root/.ssh/authorized_keys 2026-03-06T13:45:40.678 INFO:teuthology.orchestra.run.vm06.stdout:ssh-rsa AAAAB3NzaC1yc2EAAAADAQABAAABgQClATzkQHLVv+lvBgYX2LelwgSx3fQOZQpm8sL5mO3SJ3d03vYuUX2QSOxnja2HZhWInnOqLeQzl2fsxAYrGdjZvuP0hm+poU5JRPtQuADk4IOQTCosP3e12eApZcoLI+Nq/PKOoKwaKOQILzt3FqC0l8CmfNDOgK5it9XriUWkzAXDAcYys4boaaPfvhZIK1Ul4dLDJtleZ3uMK+gLegnUc2fywR/iYn7/ZuGWkb35m2JKwobJqYv9Da0tJ480cnvExUeEsk8+lqvpOhyEmIjm+6rRjiMgb6iajjzN0P32AKkoBEC7YdsNwXL4Ur8eZKNf5GZu6NoVXOnQoXAhFkEb035NJ1NZiuKVPs2ksGmtVvAPWjvM4Y5msq+JQdoc5yUk3mCYl30bP79NfMgqPXOyyJwl/T2zUdkrGsGoQE5CMsjyTgpAZr7ehVsEaEvFk67dQf8rha4S8SDcS2GxzLeiV2nue3vLtH6JlD+pAMaz4BSOToxPwteuZJ/uP5yuIvM= ceph-2b1b4a72-195a-11f1-b1df-bf1f800d676b 2026-03-06T13:45:40.690 DEBUG:teuthology.orchestra.run.vm02:> sudo /home/ubuntu/cephtest/cephadm --image harbor.clyso.com/custom-ceph/ceph/ceph:cobaltcore-storage-v19.2.3-fasttrack-3 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 2b1b4a72-195a-11f1-b1df-bf1f800d676b -- ceph config set mgr mgr/cephadm/allow_ptrace true 2026-03-06T13:45:40.934 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:45:40 vm02 ceph-mon[49207]: from='client.? 192.168.123.102:0/521031786' entity='client.admin' 2026-03-06T13:45:41.240 INFO:teuthology.orchestra.run.vm02.stderr:Inferring config /var/lib/ceph/2b1b4a72-195a-11f1-b1df-bf1f800d676b/mon.a/config 2026-03-06T13:45:41.879 INFO:tasks.cephadm:Distributing conf and client.admin keyring to all hosts + 0755 2026-03-06T13:45:41.879 DEBUG:teuthology.orchestra.run.vm02:> sudo /home/ubuntu/cephtest/cephadm --image harbor.clyso.com/custom-ceph/ceph/ceph:cobaltcore-storage-v19.2.3-fasttrack-3 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 2b1b4a72-195a-11f1-b1df-bf1f800d676b -- ceph orch client-keyring set client.admin '*' --mode 0755 2026-03-06T13:45:42.300 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:45:42 vm02 ceph-mon[49207]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:45:42.300 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:45:42 vm02 ceph-mon[49207]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:45:42.300 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:45:42 vm02 ceph-mon[49207]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:45:42.300 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:45:42 vm02 ceph-mon[49207]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-06T13:45:42.300 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:45:42 vm02 ceph-mon[49207]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:45:42.300 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:45:42 vm02 ceph-mon[49207]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:45:42.300 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:45:42 vm02 ceph-mon[49207]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:45:42.300 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:45:42 vm02 ceph-mon[49207]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:45:42.300 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:45:42 vm02 ceph-mon[49207]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:45:42.300 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:45:42 vm02 ceph-mon[49207]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-06T13:45:42.300 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:45:42 vm02 ceph-mon[49207]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:45:42.300 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:45:42 vm02 ceph-mon[49207]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:45:42.300 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:45:42 vm02 ceph-mon[49207]: from='client.? 192.168.123.102:0/928896413' entity='client.admin' 2026-03-06T13:45:42.300 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:45:42 vm02 ceph-mon[49207]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-06T13:45:42.300 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:45:42 vm02 ceph-mon[49207]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:45:42.300 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:45:42 vm02 ceph-mon[49207]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:45:42.334 INFO:teuthology.orchestra.run.vm02.stderr:Inferring config /var/lib/ceph/2b1b4a72-195a-11f1-b1df-bf1f800d676b/mon.a/config 2026-03-06T13:45:42.962 INFO:tasks.cephadm:Writing (initial) conf and keyring to vm04 2026-03-06T13:45:42.963 DEBUG:teuthology.orchestra.run.vm04:> set -ex 2026-03-06T13:45:42.963 DEBUG:teuthology.orchestra.run.vm04:> dd of=/etc/ceph/ceph.conf 2026-03-06T13:45:42.983 DEBUG:teuthology.orchestra.run.vm04:> set -ex 2026-03-06T13:45:42.983 DEBUG:teuthology.orchestra.run.vm04:> dd of=/etc/ceph/ceph.client.admin.keyring 2026-03-06T13:45:43.040 INFO:tasks.cephadm:Adding host vm04 to orchestrator... 2026-03-06T13:45:43.040 DEBUG:teuthology.orchestra.run.vm02:> sudo /home/ubuntu/cephtest/cephadm --image harbor.clyso.com/custom-ceph/ceph/ceph:cobaltcore-storage-v19.2.3-fasttrack-3 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 2b1b4a72-195a-11f1-b1df-bf1f800d676b -- ceph orch host add vm04 2026-03-06T13:45:43.653 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:45:43 vm02 ceph-mon[49207]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:45:43.653 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:45:43 vm02 ceph-mon[49207]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:45:43.653 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:45:43 vm02 ceph-mon[49207]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:45:43.653 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:45:43 vm02 ceph-mon[49207]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-06T13:45:43.653 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:45:43 vm02 ceph-mon[49207]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-06T13:45:43.653 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:45:43 vm02 ceph-mon[49207]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-06T13:45:43.699 INFO:teuthology.orchestra.run.vm02.stderr:Inferring config /var/lib/ceph/2b1b4a72-195a-11f1-b1df-bf1f800d676b/mon.a/config 2026-03-06T13:45:44.862 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:45:44 vm02 ceph-mon[49207]: from='client.14178 -' entity='client.admin' cmd=[{"prefix": "orch client-keyring set", "entity": "client.admin", "placement": "*", "mode": "0755", "target": ["mon-mgr", ""]}]: dispatch 2026-03-06T13:45:44.862 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:45:44 vm02 ceph-mon[49207]: Updating vm02:/etc/ceph/ceph.conf 2026-03-06T13:45:44.862 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:45:44 vm02 ceph-mon[49207]: Updating vm02:/var/lib/ceph/2b1b4a72-195a-11f1-b1df-bf1f800d676b/config/ceph.conf 2026-03-06T13:45:44.862 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:45:44 vm02 ceph-mon[49207]: Updating vm02:/etc/ceph/ceph.client.admin.keyring 2026-03-06T13:45:44.862 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:45:44 vm02 ceph-mon[49207]: Updating vm02:/var/lib/ceph/2b1b4a72-195a-11f1-b1df-bf1f800d676b/config/ceph.client.admin.keyring 2026-03-06T13:45:44.862 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:45:44 vm02 ceph-mon[49207]: mgrmap e13: a(active, since 6s) 2026-03-06T13:45:44.862 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:45:44 vm02 ceph-mon[49207]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:45:44.862 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:45:44 vm02 ceph-mon[49207]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:45:44.862 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:45:44 vm02 ceph-mon[49207]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:45:44.862 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:45:44 vm02 ceph-mon[49207]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:45:45.862 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:45:45 vm02 ceph-mon[49207]: from='client.14180 -' entity='client.admin' cmd=[{"prefix": "orch host add", "hostname": "vm04", "target": ["mon-mgr", ""]}]: dispatch 2026-03-06T13:45:46.463 INFO:teuthology.orchestra.run.vm02.stdout:Added host 'vm04' with addr '192.168.123.104' 2026-03-06T13:45:46.621 DEBUG:teuthology.orchestra.run.vm02:> sudo /home/ubuntu/cephtest/cephadm --image harbor.clyso.com/custom-ceph/ceph/ceph:cobaltcore-storage-v19.2.3-fasttrack-3 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 2b1b4a72-195a-11f1-b1df-bf1f800d676b -- ceph orch host ls --format=json 2026-03-06T13:45:46.862 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:45:46 vm02 ceph-mon[49207]: Deploying cephadm binary to vm04 2026-03-06T13:45:46.862 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:45:46 vm02 ceph-mon[49207]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:45:46.862 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:45:46 vm02 ceph-mon[49207]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-06T13:45:46.957 INFO:teuthology.orchestra.run.vm02.stderr:Inferring config /var/lib/ceph/2b1b4a72-195a-11f1-b1df-bf1f800d676b/mon.a/config 2026-03-06T13:45:47.315 INFO:teuthology.orchestra.run.vm02.stdout: 2026-03-06T13:45:47.315 INFO:teuthology.orchestra.run.vm02.stdout:[{"addr": "192.168.123.102", "hostname": "vm02", "labels": [], "status": ""}, {"addr": "192.168.123.104", "hostname": "vm04", "labels": [], "status": ""}] 2026-03-06T13:45:47.478 INFO:tasks.cephadm:Writing (initial) conf and keyring to vm06 2026-03-06T13:45:47.478 DEBUG:teuthology.orchestra.run.vm06:> set -ex 2026-03-06T13:45:47.478 DEBUG:teuthology.orchestra.run.vm06:> dd of=/etc/ceph/ceph.conf 2026-03-06T13:45:47.494 DEBUG:teuthology.orchestra.run.vm06:> set -ex 2026-03-06T13:45:47.494 DEBUG:teuthology.orchestra.run.vm06:> dd of=/etc/ceph/ceph.client.admin.keyring 2026-03-06T13:45:47.552 INFO:tasks.cephadm:Adding host vm06 to orchestrator... 2026-03-06T13:45:47.552 DEBUG:teuthology.orchestra.run.vm02:> sudo /home/ubuntu/cephtest/cephadm --image harbor.clyso.com/custom-ceph/ceph/ceph:cobaltcore-storage-v19.2.3-fasttrack-3 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 2b1b4a72-195a-11f1-b1df-bf1f800d676b -- ceph orch host add vm06 2026-03-06T13:45:47.579 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:45:47 vm02 ceph-mon[49207]: Added host vm04 2026-03-06T13:45:47.579 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:45:47 vm02 ceph-mon[49207]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:45:47.579 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:45:47 vm02 ceph-mon[49207]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:45:47.579 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:45:47 vm02 ceph-mon[49207]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:45:47.579 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:45:47 vm02 ceph-mon[49207]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' cmd=[{"prefix": "config rm", "who": "osd/host:vm04", "name": "osd_memory_target"}]: dispatch 2026-03-06T13:45:47.579 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:45:47 vm02 ceph-mon[49207]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-06T13:45:47.579 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:45:47 vm02 ceph-mon[49207]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-06T13:45:47.892 INFO:teuthology.orchestra.run.vm02.stderr:Inferring config /var/lib/ceph/2b1b4a72-195a-11f1-b1df-bf1f800d676b/mon.a/config 2026-03-06T13:45:49.111 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:45:48 vm02 ceph-mon[49207]: Updating vm04:/etc/ceph/ceph.conf 2026-03-06T13:45:49.112 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:45:48 vm02 ceph-mon[49207]: Updating vm04:/var/lib/ceph/2b1b4a72-195a-11f1-b1df-bf1f800d676b/config/ceph.conf 2026-03-06T13:45:49.112 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:45:48 vm02 ceph-mon[49207]: from='client.14182 -' entity='client.admin' cmd=[{"prefix": "orch host ls", "target": ["mon-mgr", ""], "format": "json"}]: dispatch 2026-03-06T13:45:49.112 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:45:48 vm02 ceph-mon[49207]: Updating vm04:/etc/ceph/ceph.client.admin.keyring 2026-03-06T13:45:49.112 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:45:48 vm02 ceph-mon[49207]: Updating vm04:/var/lib/ceph/2b1b4a72-195a-11f1-b1df-bf1f800d676b/config/ceph.client.admin.keyring 2026-03-06T13:45:49.112 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:45:48 vm02 ceph-mon[49207]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:45:49.112 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:45:48 vm02 ceph-mon[49207]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:45:49.112 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:45:48 vm02 ceph-mon[49207]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' cmd=[{"prefix": "auth get-or-create", "entity": "client.agent.vm04", "caps": []}]: dispatch 2026-03-06T13:45:49.112 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:45:48 vm02 ceph-mon[49207]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' cmd='[{"prefix": "auth get-or-create", "entity": "client.agent.vm04", "caps": []}]': finished 2026-03-06T13:45:49.112 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:45:48 vm02 ceph-mon[49207]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:45:49.112 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:45:48 vm02 ceph-mon[49207]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:45:49.112 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:45:48 vm02 ceph-mon[49207]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:45:50.112 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:45:49 vm02 ceph-mon[49207]: Deploying daemon agent.vm04 on vm04 2026-03-06T13:45:50.112 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:45:49 vm02 ceph-mon[49207]: from='client.14184 -' entity='client.admin' cmd=[{"prefix": "orch host add", "hostname": "vm06", "target": ["mon-mgr", ""]}]: dispatch 2026-03-06T13:45:50.112 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:45:49 vm02 ceph-mon[49207]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-06T13:45:50.112 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:45:49 vm02 ceph-mon[49207]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-06T13:45:50.112 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:45:49 vm02 ceph-mon[49207]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-06T13:45:50.112 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:45:49 vm02 ceph-mon[49207]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:45:50.112 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:45:49 vm02 ceph-mon[49207]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-06T13:45:50.112 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:45:49 vm02 ceph-mon[49207]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-06T13:45:50.112 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:45:49 vm02 ceph-mon[49207]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-06T13:45:50.112 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:45:49 vm02 ceph-mon[49207]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:45:50.112 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:45:49 vm02 ceph-mon[49207]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:45:50.112 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:45:49 vm02 ceph-mon[49207]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:45:50.112 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:45:49 vm02 ceph-mon[49207]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:45:50.640 INFO:teuthology.orchestra.run.vm02.stdout:Added host 'vm06' with addr '192.168.123.106' 2026-03-06T13:45:50.793 DEBUG:teuthology.orchestra.run.vm02:> sudo /home/ubuntu/cephtest/cephadm --image harbor.clyso.com/custom-ceph/ceph/ceph:cobaltcore-storage-v19.2.3-fasttrack-3 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 2b1b4a72-195a-11f1-b1df-bf1f800d676b -- ceph orch host ls --format=json 2026-03-06T13:45:51.112 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:45:50 vm02 ceph-mon[49207]: Deploying cephadm binary to vm06 2026-03-06T13:45:51.172 INFO:teuthology.orchestra.run.vm02.stderr:Inferring config /var/lib/ceph/2b1b4a72-195a-11f1-b1df-bf1f800d676b/mon.a/config 2026-03-06T13:45:51.519 INFO:teuthology.orchestra.run.vm02.stdout: 2026-03-06T13:45:51.519 INFO:teuthology.orchestra.run.vm02.stdout:[{"addr": "192.168.123.102", "hostname": "vm02", "labels": [], "status": ""}, {"addr": "192.168.123.104", "hostname": "vm04", "labels": [], "status": ""}, {"addr": "192.168.123.106", "hostname": "vm06", "labels": [], "status": ""}] 2026-03-06T13:45:51.687 INFO:tasks.cephadm:Setting crush tunables to default 2026-03-06T13:45:51.687 DEBUG:teuthology.orchestra.run.vm02:> sudo /home/ubuntu/cephtest/cephadm --image harbor.clyso.com/custom-ceph/ceph/ceph:cobaltcore-storage-v19.2.3-fasttrack-3 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 2b1b4a72-195a-11f1-b1df-bf1f800d676b -- ceph osd crush tunables default 2026-03-06T13:45:51.862 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:45:51 vm02 ceph-mon[49207]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:45:51.862 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:45:51 vm02 ceph-mon[49207]: Added host vm06 2026-03-06T13:45:51.862 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:45:51 vm02 ceph-mon[49207]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-06T13:45:51.862 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:45:51 vm02 ceph-mon[49207]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:45:51.862 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:45:51 vm02 ceph-mon[49207]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:45:51.862 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:45:51 vm02 ceph-mon[49207]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:45:51.862 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:45:51 vm02 ceph-mon[49207]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:45:51.862 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:45:51 vm02 ceph-mon[49207]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:45:51.862 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:45:51 vm02 ceph-mon[49207]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' cmd=[{"prefix": "config rm", "who": "osd/host:vm06", "name": "osd_memory_target"}]: dispatch 2026-03-06T13:45:51.862 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:45:51 vm02 ceph-mon[49207]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-06T13:45:51.862 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:45:51 vm02 ceph-mon[49207]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-06T13:45:51.862 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:45:51 vm02 ceph-mon[49207]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:45:52.037 INFO:teuthology.orchestra.run.vm02.stderr:Inferring config /var/lib/ceph/2b1b4a72-195a-11f1-b1df-bf1f800d676b/mon.a/config 2026-03-06T13:45:52.840 INFO:teuthology.orchestra.run.vm02.stderr:adjusted tunables profile to default 2026-03-06T13:45:52.965 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:45:52 vm02 ceph-mon[49207]: Updating vm06:/etc/ceph/ceph.conf 2026-03-06T13:45:52.965 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:45:52 vm02 ceph-mon[49207]: Updating vm06:/var/lib/ceph/2b1b4a72-195a-11f1-b1df-bf1f800d676b/config/ceph.conf 2026-03-06T13:45:52.965 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:45:52 vm02 ceph-mon[49207]: from='client.14186 -' entity='client.admin' cmd=[{"prefix": "orch host ls", "target": ["mon-mgr", ""], "format": "json"}]: dispatch 2026-03-06T13:45:52.965 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:45:52 vm02 ceph-mon[49207]: Updating vm06:/etc/ceph/ceph.client.admin.keyring 2026-03-06T13:45:52.965 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:45:52 vm02 ceph-mon[49207]: Updating vm06:/var/lib/ceph/2b1b4a72-195a-11f1-b1df-bf1f800d676b/config/ceph.client.admin.keyring 2026-03-06T13:45:52.965 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:45:52 vm02 ceph-mon[49207]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:45:52.965 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:45:52 vm02 ceph-mon[49207]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:45:52.965 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:45:52 vm02 ceph-mon[49207]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' cmd=[{"prefix": "auth get-or-create", "entity": "client.agent.vm06", "caps": []}]: dispatch 2026-03-06T13:45:52.965 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:45:52 vm02 ceph-mon[49207]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' cmd='[{"prefix": "auth get-or-create", "entity": "client.agent.vm06", "caps": []}]': finished 2026-03-06T13:45:52.965 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:45:52 vm02 ceph-mon[49207]: from='client.? 192.168.123.102:0/2305221987' entity='client.admin' cmd=[{"prefix": "osd crush tunables", "profile": "default"}]: dispatch 2026-03-06T13:45:52.998 INFO:tasks.cephadm:Adding mon.a on vm02 2026-03-06T13:45:52.998 INFO:tasks.cephadm:Adding mon.b on vm04 2026-03-06T13:45:52.998 INFO:tasks.cephadm:Adding mon.c on vm06 2026-03-06T13:45:52.998 DEBUG:teuthology.orchestra.run.vm06:> sudo /home/ubuntu/cephtest/cephadm --image harbor.clyso.com/custom-ceph/ceph/ceph:cobaltcore-storage-v19.2.3-fasttrack-3 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 2b1b4a72-195a-11f1-b1df-bf1f800d676b -- ceph orch apply mon '3;vm02:192.168.123.102=a;vm04:192.168.123.104=b;vm06:192.168.123.106=c' 2026-03-06T13:45:53.371 INFO:teuthology.orchestra.run.vm06.stderr:Inferring config /var/lib/ceph/2b1b4a72-195a-11f1-b1df-bf1f800d676b/config/ceph.conf 2026-03-06T13:45:53.902 INFO:teuthology.orchestra.run.vm06.stdout:Scheduled mon update... 2026-03-06T13:45:54.098 DEBUG:teuthology.orchestra.run.vm04:mon.b> sudo journalctl -f -n 0 -u ceph-2b1b4a72-195a-11f1-b1df-bf1f800d676b@mon.b.service 2026-03-06T13:45:54.100 DEBUG:teuthology.orchestra.run.vm06:mon.c> sudo journalctl -f -n 0 -u ceph-2b1b4a72-195a-11f1-b1df-bf1f800d676b@mon.c.service 2026-03-06T13:45:54.101 INFO:tasks.cephadm:Waiting for 3 mons in monmap... 2026-03-06T13:45:54.101 DEBUG:teuthology.orchestra.run.vm06:> sudo /home/ubuntu/cephtest/cephadm --image harbor.clyso.com/custom-ceph/ceph/ceph:cobaltcore-storage-v19.2.3-fasttrack-3 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 2b1b4a72-195a-11f1-b1df-bf1f800d676b -- ceph mon dump -f json 2026-03-06T13:45:54.112 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:45:53 vm02 ceph-mon[49207]: Deploying daemon agent.vm06 on vm06 2026-03-06T13:45:54.112 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:45:53 vm02 ceph-mon[49207]: from='client.? 192.168.123.102:0/2305221987' entity='client.admin' cmd='[{"prefix": "osd crush tunables", "profile": "default"}]': finished 2026-03-06T13:45:54.112 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:45:53 vm02 ceph-mon[49207]: osdmap e4: 0 total, 0 up, 0 in 2026-03-06T13:45:54.112 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:45:53 vm02 ceph-mon[49207]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:45:54.112 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:45:53 vm02 ceph-mon[49207]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:45:54.112 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:45:53 vm02 ceph-mon[49207]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:45:54.112 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:45:53 vm02 ceph-mon[49207]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-06T13:45:54.112 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:45:53 vm02 ceph-mon[49207]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-06T13:45:54.112 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:45:53 vm02 ceph-mon[49207]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-06T13:45:54.112 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:45:53 vm02 ceph-mon[49207]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:45:54.112 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:45:53 vm02 ceph-mon[49207]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-06T13:45:54.112 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:45:53 vm02 ceph-mon[49207]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-06T13:45:54.112 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:45:53 vm02 ceph-mon[49207]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-06T13:45:54.112 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:45:53 vm02 ceph-mon[49207]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:45:54.711 INFO:teuthology.orchestra.run.vm06.stderr:Inferring config /var/lib/ceph/2b1b4a72-195a-11f1-b1df-bf1f800d676b/config/ceph.conf 2026-03-06T13:45:55.128 INFO:teuthology.orchestra.run.vm06.stdout: 2026-03-06T13:45:55.128 INFO:teuthology.orchestra.run.vm06.stdout:{"epoch":1,"fsid":"2b1b4a72-195a-11f1-b1df-bf1f800d676b","modified":"2026-03-06T12:44:48.508617Z","created":"2026-03-06T12:44:48.508617Z","min_mon_release":19,"min_mon_release_name":"squid","election_strategy":1,"disallowed_leaders":"","stretch_mode":false,"tiebreaker_mon":"","removed_ranks":"","features":{"persistent":["kraken","luminous","mimic","osdmap-prune","nautilus","octopus","pacific","elector-pinging","quincy","reef","squid"],"optional":[]},"mons":[{"rank":0,"name":"a","public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.102:3300","nonce":0},{"type":"v1","addr":"192.168.123.102:6789","nonce":0}]},"addr":"192.168.123.102:6789/0","public_addr":"192.168.123.102:6789/0","priority":0,"weight":0,"crush_location":"{}"}],"quorum":[0]} 2026-03-06T13:45:55.128 INFO:teuthology.orchestra.run.vm06.stderr:dumped monmap epoch 1 2026-03-06T13:45:55.362 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:45:54 vm02 ceph-mon[49207]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:45:55.362 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:45:54 vm02 ceph-mon[49207]: from='client.14190 -' entity='client.admin' cmd=[{"prefix": "orch apply", "service_type": "mon", "placement": "3;vm02:192.168.123.102=a;vm04:192.168.123.104=b;vm06:192.168.123.106=c", "target": ["mon-mgr", ""]}]: dispatch 2026-03-06T13:45:55.362 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:45:54 vm02 ceph-mon[49207]: Saving service mon spec with placement vm02:192.168.123.102=a;vm04:192.168.123.104=b;vm06:192.168.123.106=c;count:3 2026-03-06T13:45:55.362 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:45:54 vm02 ceph-mon[49207]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:45:55.362 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:45:54 vm02 ceph-mon[49207]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-06T13:45:55.362 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:45:54 vm02 ceph-mon[49207]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-06T13:45:55.362 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:45:54 vm02 ceph-mon[49207]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-06T13:45:55.362 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:45:54 vm02 ceph-mon[49207]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:45:55.362 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:45:54 vm02 ceph-mon[49207]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "mon."}]: dispatch 2026-03-06T13:45:55.362 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:45:54 vm02 ceph-mon[49207]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-06T13:45:55.362 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:45:54 vm02 ceph-mon[49207]: Deploying daemon mon.c on vm06 2026-03-06T13:45:55.362 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:45:54 vm02 ceph-mon[49207]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:45:55.362 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:45:54 vm02 ceph-mon[49207]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:45:56.316 INFO:tasks.cephadm:Waiting for 3 mons in monmap... 2026-03-06T13:45:56.316 DEBUG:teuthology.orchestra.run.vm06:> sudo /home/ubuntu/cephtest/cephadm --image harbor.clyso.com/custom-ceph/ceph/ceph:cobaltcore-storage-v19.2.3-fasttrack-3 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 2b1b4a72-195a-11f1-b1df-bf1f800d676b -- ceph mon dump -f json 2026-03-06T13:45:56.362 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:45:55 vm02 ceph-mon[49207]: from='client.? 192.168.123.106:0/1418622913' entity='client.admin' cmd=[{"prefix": "mon dump", "format": "json"}]: dispatch 2026-03-06T13:45:56.658 INFO:teuthology.orchestra.run.vm06.stderr:Inferring config /var/lib/ceph/2b1b4a72-195a-11f1-b1df-bf1f800d676b/mon.c/config 2026-03-06T13:45:58.049 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:45:57 vm04 ceph-mon[53170]: mon.b@-1(synchronizing).mgr e13 mkfs or daemon transitioned to available, loading commands 2026-03-06T13:46:01.362 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:46:01 vm02 ceph-mon[49207]: Deploying daemon mon.b on vm04 2026-03-06T13:46:01.362 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:46:01 vm02 ceph-mon[49207]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "a"}]: dispatch 2026-03-06T13:46:01.363 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:46:01 vm02 ceph-mon[49207]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "c"}]: dispatch 2026-03-06T13:46:01.363 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:46:01 vm02 ceph-mon[49207]: mon.a calling monitor election 2026-03-06T13:46:01.363 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:46:01 vm02 ceph-mon[49207]: pgmap v4: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-06T13:46:01.363 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:46:01 vm02 ceph-mon[49207]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "c"}]: dispatch 2026-03-06T13:46:01.363 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:46:01 vm02 ceph-mon[49207]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-06T13:46:01.363 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:46:01 vm02 ceph-mon[49207]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "c"}]: dispatch 2026-03-06T13:46:01.363 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:46:01 vm02 ceph-mon[49207]: mon.c calling monitor election 2026-03-06T13:46:01.363 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:46:01 vm02 ceph-mon[49207]: pgmap v5: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-06T13:46:01.363 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:46:01 vm02 ceph-mon[49207]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-06T13:46:01.363 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:46:01 vm02 ceph-mon[49207]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "c"}]: dispatch 2026-03-06T13:46:01.363 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:46:01 vm02 ceph-mon[49207]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-06T13:46:01.363 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:46:01 vm02 ceph-mon[49207]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "c"}]: dispatch 2026-03-06T13:46:01.363 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:46:01 vm02 ceph-mon[49207]: pgmap v6: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-06T13:46:01.363 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:46:01 vm02 ceph-mon[49207]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-06T13:46:01.363 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:46:01 vm02 ceph-mon[49207]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "c"}]: dispatch 2026-03-06T13:46:01.363 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:46:01 vm02 ceph-mon[49207]: mon.a is new leader, mons a,c in quorum (ranks 0,1) 2026-03-06T13:46:01.363 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:46:01 vm02 ceph-mon[49207]: monmap epoch 2 2026-03-06T13:46:01.363 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:46:01 vm02 ceph-mon[49207]: fsid 2b1b4a72-195a-11f1-b1df-bf1f800d676b 2026-03-06T13:46:01.363 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:46:01 vm02 ceph-mon[49207]: last_changed 2026-03-06T12:45:56.006464+0000 2026-03-06T13:46:01.363 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:46:01 vm02 ceph-mon[49207]: created 2026-03-06T12:44:48.508617+0000 2026-03-06T13:46:01.363 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:46:01 vm02 ceph-mon[49207]: min_mon_release 19 (squid) 2026-03-06T13:46:01.363 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:46:01 vm02 ceph-mon[49207]: election_strategy: 1 2026-03-06T13:46:01.363 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:46:01 vm02 ceph-mon[49207]: 0: [v2:192.168.123.102:3300/0,v1:192.168.123.102:6789/0] mon.a 2026-03-06T13:46:01.363 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:46:01 vm02 ceph-mon[49207]: 1: [v2:192.168.123.106:3300/0,v1:192.168.123.106:6789/0] mon.c 2026-03-06T13:46:01.363 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:46:01 vm02 ceph-mon[49207]: fsmap 2026-03-06T13:46:01.363 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:46:01 vm02 ceph-mon[49207]: osdmap e4: 0 total, 0 up, 0 in 2026-03-06T13:46:01.363 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:46:01 vm02 ceph-mon[49207]: mgrmap e13: a(active, since 24s) 2026-03-06T13:46:01.363 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:46:01 vm02 ceph-mon[49207]: overall HEALTH_OK 2026-03-06T13:46:01.363 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:46:01 vm02 ceph-mon[49207]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:46:01.363 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:46:01 vm02 ceph-mon[49207]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:46:01.363 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:46:01 vm02 ceph-mon[49207]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:46:01.363 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:46:01 vm02 ceph-mon[49207]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:46:01.363 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:46:01 vm02 ceph-mon[49207]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:46:01.363 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:46:01 vm02 ceph-mon[49207]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:46:01.363 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:46:01 vm02 ceph-mon[49207]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:46:01.491 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:46:01 vm06 ceph-mon[56528]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:46:01.493 INFO:teuthology.orchestra.run.vm06.stdout: 2026-03-06T13:46:01.493 INFO:teuthology.orchestra.run.vm06.stdout:{"epoch":2,"fsid":"2b1b4a72-195a-11f1-b1df-bf1f800d676b","modified":"2026-03-06T12:45:56.006464Z","created":"2026-03-06T12:44:48.508617Z","min_mon_release":19,"min_mon_release_name":"squid","election_strategy":1,"disallowed_leaders":"","stretch_mode":false,"tiebreaker_mon":"","removed_ranks":"","features":{"persistent":["kraken","luminous","mimic","osdmap-prune","nautilus","octopus","pacific","elector-pinging","quincy","reef","squid"],"optional":[]},"mons":[{"rank":0,"name":"a","public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.102:3300","nonce":0},{"type":"v1","addr":"192.168.123.102:6789","nonce":0}]},"addr":"192.168.123.102:6789/0","public_addr":"192.168.123.102:6789/0","priority":0,"weight":0,"crush_location":"{}"},{"rank":1,"name":"c","public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.106:3300","nonce":0},{"type":"v1","addr":"192.168.123.106:6789","nonce":0}]},"addr":"192.168.123.106:6789/0","public_addr":"192.168.123.106:6789/0","priority":0,"weight":0,"crush_location":"{}"}],"quorum":[0,1]} 2026-03-06T13:46:01.493 INFO:teuthology.orchestra.run.vm06.stderr:dumped monmap epoch 2 2026-03-06T13:46:02.050 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:46:01 vm04 ceph-mon[53170]: Deploying daemon mon.b on vm04 2026-03-06T13:46:02.050 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:46:01 vm04 ceph-mon[53170]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "a"}]: dispatch 2026-03-06T13:46:02.050 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:46:01 vm04 ceph-mon[53170]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "c"}]: dispatch 2026-03-06T13:46:02.050 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:46:01 vm04 ceph-mon[53170]: mon.a calling monitor election 2026-03-06T13:46:02.050 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:46:01 vm04 ceph-mon[53170]: pgmap v4: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-06T13:46:02.050 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:46:01 vm04 ceph-mon[53170]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "c"}]: dispatch 2026-03-06T13:46:02.050 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:46:01 vm04 ceph-mon[53170]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-06T13:46:02.050 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:46:01 vm04 ceph-mon[53170]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "c"}]: dispatch 2026-03-06T13:46:02.050 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:46:01 vm04 ceph-mon[53170]: mon.c calling monitor election 2026-03-06T13:46:02.050 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:46:01 vm04 ceph-mon[53170]: pgmap v5: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-06T13:46:02.050 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:46:01 vm04 ceph-mon[53170]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-06T13:46:02.050 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:46:01 vm04 ceph-mon[53170]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "c"}]: dispatch 2026-03-06T13:46:02.050 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:46:01 vm04 ceph-mon[53170]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-06T13:46:02.050 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:46:01 vm04 ceph-mon[53170]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "c"}]: dispatch 2026-03-06T13:46:02.050 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:46:01 vm04 ceph-mon[53170]: pgmap v6: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-06T13:46:02.050 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:46:01 vm04 ceph-mon[53170]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-06T13:46:02.050 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:46:01 vm04 ceph-mon[53170]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "c"}]: dispatch 2026-03-06T13:46:02.050 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:46:01 vm04 ceph-mon[53170]: mon.a is new leader, mons a,c in quorum (ranks 0,1) 2026-03-06T13:46:02.050 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:46:01 vm04 ceph-mon[53170]: monmap epoch 2 2026-03-06T13:46:02.050 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:46:01 vm04 ceph-mon[53170]: fsid 2b1b4a72-195a-11f1-b1df-bf1f800d676b 2026-03-06T13:46:02.050 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:46:01 vm04 ceph-mon[53170]: last_changed 2026-03-06T12:45:56.006464+0000 2026-03-06T13:46:02.050 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:46:01 vm04 ceph-mon[53170]: created 2026-03-06T12:44:48.508617+0000 2026-03-06T13:46:02.050 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:46:01 vm04 ceph-mon[53170]: min_mon_release 19 (squid) 2026-03-06T13:46:02.050 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:46:01 vm04 ceph-mon[53170]: election_strategy: 1 2026-03-06T13:46:02.050 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:46:01 vm04 ceph-mon[53170]: 0: [v2:192.168.123.102:3300/0,v1:192.168.123.102:6789/0] mon.a 2026-03-06T13:46:02.050 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:46:01 vm04 ceph-mon[53170]: 1: [v2:192.168.123.106:3300/0,v1:192.168.123.106:6789/0] mon.c 2026-03-06T13:46:02.050 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:46:01 vm04 ceph-mon[53170]: fsmap 2026-03-06T13:46:02.050 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:46:01 vm04 ceph-mon[53170]: osdmap e4: 0 total, 0 up, 0 in 2026-03-06T13:46:02.050 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:46:01 vm04 ceph-mon[53170]: mgrmap e13: a(active, since 24s) 2026-03-06T13:46:02.050 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:46:01 vm04 ceph-mon[53170]: overall HEALTH_OK 2026-03-06T13:46:02.050 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:46:01 vm04 ceph-mon[53170]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:46:02.050 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:46:01 vm04 ceph-mon[53170]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:46:02.050 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:46:01 vm04 ceph-mon[53170]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:46:02.050 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:46:01 vm04 ceph-mon[53170]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:46:02.050 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:46:01 vm04 ceph-mon[53170]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:46:02.050 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:46:01 vm04 ceph-mon[53170]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:46:02.050 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:46:01 vm04 ceph-mon[53170]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:46:02.069 INFO:journalctl@ceph.mgr.a.vm02.stdout:Mar 06 13:46:02 vm02 ceph-2b1b4a72-195a-11f1-b1df-bf1f800d676b-mgr-a[49440]: 2026-03-06T12:46:02.005+0000 7f3e6c292640 -1 mgr.server handle_report got status from non-daemon mon.c 2026-03-06T13:46:02.670 INFO:tasks.cephadm:Waiting for 3 mons in monmap... 2026-03-06T13:46:02.670 DEBUG:teuthology.orchestra.run.vm06:> sudo /home/ubuntu/cephtest/cephadm --image harbor.clyso.com/custom-ceph/ceph/ceph:cobaltcore-storage-v19.2.3-fasttrack-3 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 2b1b4a72-195a-11f1-b1df-bf1f800d676b -- ceph mon dump -f json 2026-03-06T13:46:03.023 INFO:teuthology.orchestra.run.vm06.stderr:Inferring config /var/lib/ceph/2b1b4a72-195a-11f1-b1df-bf1f800d676b/mon.c/config 2026-03-06T13:46:07.077 INFO:teuthology.orchestra.run.vm06.stdout: 2026-03-06T13:46:07.078 INFO:teuthology.orchestra.run.vm06.stdout:{"epoch":3,"fsid":"2b1b4a72-195a-11f1-b1df-bf1f800d676b","modified":"2026-03-06T12:46:01.790449Z","created":"2026-03-06T12:44:48.508617Z","min_mon_release":19,"min_mon_release_name":"squid","election_strategy":1,"disallowed_leaders":"","stretch_mode":false,"tiebreaker_mon":"","removed_ranks":"","features":{"persistent":["kraken","luminous","mimic","osdmap-prune","nautilus","octopus","pacific","elector-pinging","quincy","reef","squid"],"optional":[]},"mons":[{"rank":0,"name":"a","public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.102:3300","nonce":0},{"type":"v1","addr":"192.168.123.102:6789","nonce":0}]},"addr":"192.168.123.102:6789/0","public_addr":"192.168.123.102:6789/0","priority":0,"weight":0,"crush_location":"{}"},{"rank":1,"name":"c","public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.106:3300","nonce":0},{"type":"v1","addr":"192.168.123.106:6789","nonce":0}]},"addr":"192.168.123.106:6789/0","public_addr":"192.168.123.106:6789/0","priority":0,"weight":0,"crush_location":"{}"},{"rank":2,"name":"b","public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.104:3300","nonce":0},{"type":"v1","addr":"192.168.123.104:6789","nonce":0}]},"addr":"192.168.123.104:6789/0","public_addr":"192.168.123.104:6789/0","priority":0,"weight":0,"crush_location":"{}"}],"quorum":[0,1,2]} 2026-03-06T13:46:07.078 INFO:teuthology.orchestra.run.vm06.stderr:dumped monmap epoch 3 2026-03-06T13:46:07.112 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:46:06 vm02 ceph-mon[49207]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "a"}]: dispatch 2026-03-06T13:46:07.112 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:46:06 vm02 ceph-mon[49207]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-06T13:46:07.112 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:46:06 vm02 ceph-mon[49207]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "c"}]: dispatch 2026-03-06T13:46:07.112 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:46:06 vm02 ceph-mon[49207]: mon.a calling monitor election 2026-03-06T13:46:07.112 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:46:06 vm02 ceph-mon[49207]: mon.c calling monitor election 2026-03-06T13:46:07.112 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:46:06 vm02 ceph-mon[49207]: pgmap v7: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-06T13:46:07.112 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:46:06 vm02 ceph-mon[49207]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-06T13:46:07.112 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:46:06 vm02 ceph-mon[49207]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-06T13:46:07.112 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:46:06 vm02 ceph-mon[49207]: mon.b calling monitor election 2026-03-06T13:46:07.112 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:46:06 vm02 ceph-mon[49207]: pgmap v8: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-06T13:46:07.112 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:46:06 vm02 ceph-mon[49207]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-06T13:46:07.112 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:46:06 vm02 ceph-mon[49207]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-06T13:46:07.112 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:46:06 vm02 ceph-mon[49207]: pgmap v9: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-06T13:46:07.112 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:46:06 vm02 ceph-mon[49207]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-06T13:46:07.112 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:46:06 vm02 ceph-mon[49207]: mon.a is new leader, mons a,c,b in quorum (ranks 0,1,2) 2026-03-06T13:46:07.112 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:46:06 vm02 ceph-mon[49207]: monmap epoch 3 2026-03-06T13:46:07.112 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:46:06 vm02 ceph-mon[49207]: fsid 2b1b4a72-195a-11f1-b1df-bf1f800d676b 2026-03-06T13:46:07.112 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:46:06 vm02 ceph-mon[49207]: last_changed 2026-03-06T12:46:01.790449+0000 2026-03-06T13:46:07.112 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:46:06 vm02 ceph-mon[49207]: created 2026-03-06T12:44:48.508617+0000 2026-03-06T13:46:07.112 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:46:06 vm02 ceph-mon[49207]: min_mon_release 19 (squid) 2026-03-06T13:46:07.112 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:46:06 vm02 ceph-mon[49207]: election_strategy: 1 2026-03-06T13:46:07.112 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:46:06 vm02 ceph-mon[49207]: 0: [v2:192.168.123.102:3300/0,v1:192.168.123.102:6789/0] mon.a 2026-03-06T13:46:07.112 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:46:06 vm02 ceph-mon[49207]: 1: [v2:192.168.123.106:3300/0,v1:192.168.123.106:6789/0] mon.c 2026-03-06T13:46:07.112 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:46:06 vm02 ceph-mon[49207]: 2: [v2:192.168.123.104:3300/0,v1:192.168.123.104:6789/0] mon.b 2026-03-06T13:46:07.112 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:46:06 vm02 ceph-mon[49207]: fsmap 2026-03-06T13:46:07.112 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:46:06 vm02 ceph-mon[49207]: osdmap e4: 0 total, 0 up, 0 in 2026-03-06T13:46:07.112 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:46:06 vm02 ceph-mon[49207]: mgrmap e13: a(active, since 30s) 2026-03-06T13:46:07.112 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:46:06 vm02 ceph-mon[49207]: overall HEALTH_OK 2026-03-06T13:46:07.112 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:46:06 vm02 ceph-mon[49207]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:46:07.112 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:46:06 vm02 ceph-mon[49207]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:46:07.112 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:46:06 vm02 ceph-mon[49207]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:46:07.112 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:46:06 vm02 ceph-mon[49207]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:46:07.112 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:46:06 vm02 ceph-mon[49207]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:46:07.112 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:46:06 vm02 ceph-mon[49207]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:46:07.112 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:46:06 vm02 ceph-mon[49207]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:46:07.112 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:46:06 vm02 ceph-mon[49207]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "mon."}]: dispatch 2026-03-06T13:46:07.112 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:46:06 vm02 ceph-mon[49207]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' cmd=[{"prefix": "config get", "who": "mon", "key": "public_network"}]: dispatch 2026-03-06T13:46:07.112 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:46:06 vm02 ceph-mon[49207]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-06T13:46:07.112 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:46:06 vm02 ceph-mon[49207]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:46:07.113 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:46:06 vm02 ceph-mon[49207]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:46:07.113 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:46:06 vm02 ceph-mon[49207]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:46:07.113 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:46:06 vm02 ceph-mon[49207]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:46:07.113 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:46:06 vm02 ceph-mon[49207]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:46:07.113 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:46:06 vm06 ceph-mon[56528]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "a"}]: dispatch 2026-03-06T13:46:07.113 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:46:06 vm06 ceph-mon[56528]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-06T13:46:07.113 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:46:06 vm06 ceph-mon[56528]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "c"}]: dispatch 2026-03-06T13:46:07.113 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:46:06 vm06 ceph-mon[56528]: mon.a calling monitor election 2026-03-06T13:46:07.113 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:46:06 vm06 ceph-mon[56528]: mon.c calling monitor election 2026-03-06T13:46:07.113 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:46:06 vm06 ceph-mon[56528]: pgmap v7: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-06T13:46:07.113 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:46:06 vm06 ceph-mon[56528]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-06T13:46:07.113 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:46:06 vm06 ceph-mon[56528]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-06T13:46:07.113 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:46:06 vm06 ceph-mon[56528]: mon.b calling monitor election 2026-03-06T13:46:07.113 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:46:06 vm06 ceph-mon[56528]: pgmap v8: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-06T13:46:07.113 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:46:06 vm06 ceph-mon[56528]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-06T13:46:07.113 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:46:06 vm06 ceph-mon[56528]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-06T13:46:07.113 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:46:06 vm06 ceph-mon[56528]: pgmap v9: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-06T13:46:07.113 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:46:06 vm06 ceph-mon[56528]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-06T13:46:07.113 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:46:06 vm06 ceph-mon[56528]: mon.a is new leader, mons a,c,b in quorum (ranks 0,1,2) 2026-03-06T13:46:07.113 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:46:06 vm06 ceph-mon[56528]: monmap epoch 3 2026-03-06T13:46:07.113 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:46:06 vm06 ceph-mon[56528]: fsid 2b1b4a72-195a-11f1-b1df-bf1f800d676b 2026-03-06T13:46:07.113 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:46:06 vm06 ceph-mon[56528]: last_changed 2026-03-06T12:46:01.790449+0000 2026-03-06T13:46:07.113 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:46:06 vm06 ceph-mon[56528]: created 2026-03-06T12:44:48.508617+0000 2026-03-06T13:46:07.113 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:46:06 vm06 ceph-mon[56528]: min_mon_release 19 (squid) 2026-03-06T13:46:07.113 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:46:06 vm06 ceph-mon[56528]: election_strategy: 1 2026-03-06T13:46:07.113 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:46:06 vm06 ceph-mon[56528]: 0: [v2:192.168.123.102:3300/0,v1:192.168.123.102:6789/0] mon.a 2026-03-06T13:46:07.113 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:46:06 vm06 ceph-mon[56528]: 1: [v2:192.168.123.106:3300/0,v1:192.168.123.106:6789/0] mon.c 2026-03-06T13:46:07.113 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:46:06 vm06 ceph-mon[56528]: 2: [v2:192.168.123.104:3300/0,v1:192.168.123.104:6789/0] mon.b 2026-03-06T13:46:07.113 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:46:06 vm06 ceph-mon[56528]: fsmap 2026-03-06T13:46:07.113 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:46:06 vm06 ceph-mon[56528]: osdmap e4: 0 total, 0 up, 0 in 2026-03-06T13:46:07.113 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:46:06 vm06 ceph-mon[56528]: mgrmap e13: a(active, since 30s) 2026-03-06T13:46:07.113 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:46:06 vm06 ceph-mon[56528]: overall HEALTH_OK 2026-03-06T13:46:07.113 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:46:06 vm06 ceph-mon[56528]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:46:07.113 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:46:06 vm06 ceph-mon[56528]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:46:07.113 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:46:06 vm06 ceph-mon[56528]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:46:07.113 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:46:06 vm06 ceph-mon[56528]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:46:07.113 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:46:06 vm06 ceph-mon[56528]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:46:07.113 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:46:06 vm06 ceph-mon[56528]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:46:07.113 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:46:06 vm06 ceph-mon[56528]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:46:07.113 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:46:06 vm06 ceph-mon[56528]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "mon."}]: dispatch 2026-03-06T13:46:07.113 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:46:06 vm06 ceph-mon[56528]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' cmd=[{"prefix": "config get", "who": "mon", "key": "public_network"}]: dispatch 2026-03-06T13:46:07.113 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:46:06 vm06 ceph-mon[56528]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-06T13:46:07.113 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:46:06 vm06 ceph-mon[56528]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:46:07.113 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:46:06 vm06 ceph-mon[56528]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:46:07.113 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:46:06 vm06 ceph-mon[56528]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:46:07.113 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:46:06 vm06 ceph-mon[56528]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:46:07.113 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:46:06 vm06 ceph-mon[56528]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:46:07.171 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:46:06 vm04 ceph-mon[53170]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "a"}]: dispatch 2026-03-06T13:46:07.171 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:46:06 vm04 ceph-mon[53170]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-06T13:46:07.171 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:46:06 vm04 ceph-mon[53170]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "c"}]: dispatch 2026-03-06T13:46:07.171 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:46:06 vm04 ceph-mon[53170]: mon.a calling monitor election 2026-03-06T13:46:07.171 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:46:06 vm04 ceph-mon[53170]: mon.c calling monitor election 2026-03-06T13:46:07.171 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:46:06 vm04 ceph-mon[53170]: pgmap v7: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-06T13:46:07.171 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:46:06 vm04 ceph-mon[53170]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-06T13:46:07.171 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:46:06 vm04 ceph-mon[53170]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-06T13:46:07.171 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:46:06 vm04 ceph-mon[53170]: mon.b calling monitor election 2026-03-06T13:46:07.171 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:46:06 vm04 ceph-mon[53170]: pgmap v8: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-06T13:46:07.171 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:46:06 vm04 ceph-mon[53170]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-06T13:46:07.171 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:46:06 vm04 ceph-mon[53170]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-06T13:46:07.171 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:46:06 vm04 ceph-mon[53170]: pgmap v9: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-06T13:46:07.171 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:46:06 vm04 ceph-mon[53170]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-06T13:46:07.171 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:46:06 vm04 ceph-mon[53170]: mon.a is new leader, mons a,c,b in quorum (ranks 0,1,2) 2026-03-06T13:46:07.171 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:46:06 vm04 ceph-mon[53170]: monmap epoch 3 2026-03-06T13:46:07.171 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:46:06 vm04 ceph-mon[53170]: fsid 2b1b4a72-195a-11f1-b1df-bf1f800d676b 2026-03-06T13:46:07.171 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:46:06 vm04 ceph-mon[53170]: last_changed 2026-03-06T12:46:01.790449+0000 2026-03-06T13:46:07.171 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:46:06 vm04 ceph-mon[53170]: created 2026-03-06T12:44:48.508617+0000 2026-03-06T13:46:07.171 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:46:06 vm04 ceph-mon[53170]: min_mon_release 19 (squid) 2026-03-06T13:46:07.171 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:46:06 vm04 ceph-mon[53170]: election_strategy: 1 2026-03-06T13:46:07.172 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:46:06 vm04 ceph-mon[53170]: 0: [v2:192.168.123.102:3300/0,v1:192.168.123.102:6789/0] mon.a 2026-03-06T13:46:07.172 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:46:06 vm04 ceph-mon[53170]: 1: [v2:192.168.123.106:3300/0,v1:192.168.123.106:6789/0] mon.c 2026-03-06T13:46:07.172 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:46:06 vm04 ceph-mon[53170]: 2: [v2:192.168.123.104:3300/0,v1:192.168.123.104:6789/0] mon.b 2026-03-06T13:46:07.172 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:46:06 vm04 ceph-mon[53170]: fsmap 2026-03-06T13:46:07.172 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:46:06 vm04 ceph-mon[53170]: osdmap e4: 0 total, 0 up, 0 in 2026-03-06T13:46:07.172 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:46:06 vm04 ceph-mon[53170]: mgrmap e13: a(active, since 30s) 2026-03-06T13:46:07.172 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:46:06 vm04 ceph-mon[53170]: overall HEALTH_OK 2026-03-06T13:46:07.172 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:46:06 vm04 ceph-mon[53170]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:46:07.172 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:46:06 vm04 ceph-mon[53170]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:46:07.172 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:46:06 vm04 ceph-mon[53170]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:46:07.172 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:46:06 vm04 ceph-mon[53170]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:46:07.172 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:46:06 vm04 ceph-mon[53170]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:46:07.172 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:46:06 vm04 ceph-mon[53170]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:46:07.172 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:46:06 vm04 ceph-mon[53170]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:46:07.172 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:46:06 vm04 ceph-mon[53170]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "mon."}]: dispatch 2026-03-06T13:46:07.172 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:46:06 vm04 ceph-mon[53170]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' cmd=[{"prefix": "config get", "who": "mon", "key": "public_network"}]: dispatch 2026-03-06T13:46:07.172 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:46:06 vm04 ceph-mon[53170]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-06T13:46:07.172 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:46:06 vm04 ceph-mon[53170]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:46:07.172 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:46:06 vm04 ceph-mon[53170]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:46:07.172 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:46:06 vm04 ceph-mon[53170]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:46:07.172 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:46:06 vm04 ceph-mon[53170]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:46:07.172 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:46:06 vm04 ceph-mon[53170]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:46:07.259 INFO:tasks.cephadm:Generating final ceph.conf file... 2026-03-06T13:46:07.259 DEBUG:teuthology.orchestra.run.vm02:> sudo /home/ubuntu/cephtest/cephadm --image harbor.clyso.com/custom-ceph/ceph/ceph:cobaltcore-storage-v19.2.3-fasttrack-3 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 2b1b4a72-195a-11f1-b1df-bf1f800d676b -- ceph config generate-minimal-conf 2026-03-06T13:46:07.614 INFO:teuthology.orchestra.run.vm02.stderr:Inferring config /var/lib/ceph/2b1b4a72-195a-11f1-b1df-bf1f800d676b/mon.a/config 2026-03-06T13:46:07.947 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:46:07 vm02 ceph-mon[49207]: Reconfiguring mon.b (monmap changed)... 2026-03-06T13:46:07.947 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:46:07 vm02 ceph-mon[49207]: Reconfiguring daemon mon.b on vm04 2026-03-06T13:46:07.947 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:46:07 vm02 ceph-mon[49207]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:46:07.947 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:46:07 vm02 ceph-mon[49207]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:46:07.947 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:46:07 vm02 ceph-mon[49207]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:46:07.947 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:46:07 vm02 ceph-mon[49207]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:46:07.947 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:46:07 vm02 ceph-mon[49207]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:46:07.947 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:46:07 vm02 ceph-mon[49207]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:46:07.947 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:46:07 vm02 ceph-mon[49207]: from='client.? 192.168.123.106:0/436511624' entity='client.admin' cmd=[{"prefix": "mon dump", "format": "json"}]: dispatch 2026-03-06T13:46:07.947 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:46:07 vm02 ceph-mon[49207]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:46:07.947 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:46:07 vm02 ceph-mon[49207]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:46:07.947 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:46:07 vm02 ceph-mon[49207]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-06T13:46:07.947 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:46:07 vm02 ceph-mon[49207]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-06T13:46:07.947 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:46:07 vm02 ceph-mon[49207]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-06T13:46:07.947 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:46:07 vm02 ceph-mon[49207]: Updating vm02:/etc/ceph/ceph.conf 2026-03-06T13:46:07.947 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:46:07 vm02 ceph-mon[49207]: Updating vm04:/etc/ceph/ceph.conf 2026-03-06T13:46:07.947 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:46:07 vm02 ceph-mon[49207]: Updating vm06:/etc/ceph/ceph.conf 2026-03-06T13:46:07.947 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:46:07 vm02 ceph-mon[49207]: Updating vm04:/var/lib/ceph/2b1b4a72-195a-11f1-b1df-bf1f800d676b/config/ceph.conf 2026-03-06T13:46:07.947 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:46:07 vm02 ceph-mon[49207]: Updating vm06:/var/lib/ceph/2b1b4a72-195a-11f1-b1df-bf1f800d676b/config/ceph.conf 2026-03-06T13:46:07.947 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:46:07 vm02 ceph-mon[49207]: Updating vm02:/var/lib/ceph/2b1b4a72-195a-11f1-b1df-bf1f800d676b/config/ceph.conf 2026-03-06T13:46:07.947 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:46:07 vm02 ceph-mon[49207]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:46:07.947 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:46:07 vm02 ceph-mon[49207]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:46:07.947 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:46:07 vm02 ceph-mon[49207]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:46:07.947 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:46:07 vm02 ceph-mon[49207]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:46:07.947 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:46:07 vm02 ceph-mon[49207]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-06T13:46:07.947 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:46:07 vm02 ceph-mon[49207]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:46:07.947 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:46:07 vm02 ceph-mon[49207]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:46:07.947 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:46:07 vm02 ceph-mon[49207]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:46:07.947 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:46:07 vm02 ceph-mon[49207]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "mon."}]: dispatch 2026-03-06T13:46:07.947 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:46:07 vm02 ceph-mon[49207]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' cmd=[{"prefix": "config get", "who": "mon", "key": "public_network"}]: dispatch 2026-03-06T13:46:07.947 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:46:07 vm02 ceph-mon[49207]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-06T13:46:08.000 INFO:teuthology.orchestra.run.vm02.stdout:# minimal ceph.conf for 2b1b4a72-195a-11f1-b1df-bf1f800d676b 2026-03-06T13:46:08.000 INFO:teuthology.orchestra.run.vm02.stdout:[global] 2026-03-06T13:46:08.000 INFO:teuthology.orchestra.run.vm02.stdout: fsid = 2b1b4a72-195a-11f1-b1df-bf1f800d676b 2026-03-06T13:46:08.000 INFO:teuthology.orchestra.run.vm02.stdout: mon_host = [v2:192.168.123.102:3300/0,v1:192.168.123.102:6789/0] [v2:192.168.123.104:3300/0,v1:192.168.123.104:6789/0] [v2:192.168.123.106:3300/0,v1:192.168.123.106:6789/0] 2026-03-06T13:46:08.183 INFO:tasks.cephadm:Distributing (final) config and client.admin keyring... 2026-03-06T13:46:08.183 DEBUG:teuthology.orchestra.run.vm02:> set -ex 2026-03-06T13:46:08.183 DEBUG:teuthology.orchestra.run.vm02:> sudo dd of=/etc/ceph/ceph.conf 2026-03-06T13:46:08.212 DEBUG:teuthology.orchestra.run.vm02:> set -ex 2026-03-06T13:46:08.213 DEBUG:teuthology.orchestra.run.vm02:> sudo dd of=/etc/ceph/ceph.client.admin.keyring 2026-03-06T13:46:08.260 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:46:07 vm06 ceph-mon[56528]: Reconfiguring mon.b (monmap changed)... 2026-03-06T13:46:08.318 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:46:07 vm06 ceph-mon[56528]: Reconfiguring daemon mon.b on vm04 2026-03-06T13:46:08.318 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:46:07 vm06 ceph-mon[56528]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:46:08.318 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:46:07 vm06 ceph-mon[56528]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:46:08.318 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:46:07 vm06 ceph-mon[56528]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:46:08.318 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:46:07 vm06 ceph-mon[56528]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:46:08.318 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:46:07 vm06 ceph-mon[56528]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:46:08.318 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:46:07 vm06 ceph-mon[56528]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:46:08.318 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:46:07 vm06 ceph-mon[56528]: from='client.? 192.168.123.106:0/436511624' entity='client.admin' cmd=[{"prefix": "mon dump", "format": "json"}]: dispatch 2026-03-06T13:46:08.318 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:46:07 vm06 ceph-mon[56528]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:46:08.318 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:46:07 vm06 ceph-mon[56528]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:46:08.318 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:46:07 vm06 ceph-mon[56528]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-06T13:46:08.318 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:46:07 vm06 ceph-mon[56528]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-06T13:46:08.318 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:46:07 vm06 ceph-mon[56528]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-06T13:46:08.318 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:46:07 vm06 ceph-mon[56528]: Updating vm02:/etc/ceph/ceph.conf 2026-03-06T13:46:08.318 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:46:07 vm06 ceph-mon[56528]: Updating vm04:/etc/ceph/ceph.conf 2026-03-06T13:46:08.318 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:46:07 vm06 ceph-mon[56528]: Updating vm06:/etc/ceph/ceph.conf 2026-03-06T13:46:08.318 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:46:07 vm06 ceph-mon[56528]: Updating vm04:/var/lib/ceph/2b1b4a72-195a-11f1-b1df-bf1f800d676b/config/ceph.conf 2026-03-06T13:46:08.318 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:46:07 vm06 ceph-mon[56528]: Updating vm06:/var/lib/ceph/2b1b4a72-195a-11f1-b1df-bf1f800d676b/config/ceph.conf 2026-03-06T13:46:08.318 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:46:07 vm06 ceph-mon[56528]: Updating vm02:/var/lib/ceph/2b1b4a72-195a-11f1-b1df-bf1f800d676b/config/ceph.conf 2026-03-06T13:46:08.318 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:46:07 vm06 ceph-mon[56528]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:46:08.318 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:46:07 vm06 ceph-mon[56528]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:46:08.318 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:46:07 vm06 ceph-mon[56528]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:46:08.318 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:46:07 vm06 ceph-mon[56528]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:46:08.318 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:46:07 vm06 ceph-mon[56528]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-06T13:46:08.318 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:46:07 vm06 ceph-mon[56528]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:46:08.318 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:46:07 vm06 ceph-mon[56528]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:46:08.318 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:46:07 vm06 ceph-mon[56528]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:46:08.318 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:46:07 vm06 ceph-mon[56528]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "mon."}]: dispatch 2026-03-06T13:46:08.318 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:46:07 vm06 ceph-mon[56528]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' cmd=[{"prefix": "config get", "who": "mon", "key": "public_network"}]: dispatch 2026-03-06T13:46:08.318 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:46:07 vm06 ceph-mon[56528]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-06T13:46:08.319 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:46:07 vm04 ceph-mon[53170]: Reconfiguring mon.b (monmap changed)... 2026-03-06T13:46:08.319 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:46:07 vm04 ceph-mon[53170]: Reconfiguring daemon mon.b on vm04 2026-03-06T13:46:08.319 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:46:07 vm04 ceph-mon[53170]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:46:08.319 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:46:07 vm04 ceph-mon[53170]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:46:08.319 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:46:07 vm04 ceph-mon[53170]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:46:08.319 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:46:07 vm04 ceph-mon[53170]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:46:08.319 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:46:07 vm04 ceph-mon[53170]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:46:08.319 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:46:07 vm04 ceph-mon[53170]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:46:08.319 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:46:07 vm04 ceph-mon[53170]: from='client.? 192.168.123.106:0/436511624' entity='client.admin' cmd=[{"prefix": "mon dump", "format": "json"}]: dispatch 2026-03-06T13:46:08.319 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:46:07 vm04 ceph-mon[53170]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:46:08.319 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:46:07 vm04 ceph-mon[53170]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:46:08.319 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:46:07 vm04 ceph-mon[53170]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-06T13:46:08.319 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:46:07 vm04 ceph-mon[53170]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-06T13:46:08.319 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:46:07 vm04 ceph-mon[53170]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-06T13:46:08.319 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:46:07 vm04 ceph-mon[53170]: Updating vm02:/etc/ceph/ceph.conf 2026-03-06T13:46:08.319 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:46:07 vm04 ceph-mon[53170]: Updating vm04:/etc/ceph/ceph.conf 2026-03-06T13:46:08.319 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:46:07 vm04 ceph-mon[53170]: Updating vm06:/etc/ceph/ceph.conf 2026-03-06T13:46:08.319 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:46:07 vm04 ceph-mon[53170]: Updating vm04:/var/lib/ceph/2b1b4a72-195a-11f1-b1df-bf1f800d676b/config/ceph.conf 2026-03-06T13:46:08.319 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:46:07 vm04 ceph-mon[53170]: Updating vm06:/var/lib/ceph/2b1b4a72-195a-11f1-b1df-bf1f800d676b/config/ceph.conf 2026-03-06T13:46:08.319 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:46:07 vm04 ceph-mon[53170]: Updating vm02:/var/lib/ceph/2b1b4a72-195a-11f1-b1df-bf1f800d676b/config/ceph.conf 2026-03-06T13:46:08.319 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:46:07 vm04 ceph-mon[53170]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:46:08.319 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:46:07 vm04 ceph-mon[53170]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:46:08.319 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:46:07 vm04 ceph-mon[53170]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:46:08.319 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:46:07 vm04 ceph-mon[53170]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:46:08.319 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:46:07 vm04 ceph-mon[53170]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-06T13:46:08.319 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:46:07 vm04 ceph-mon[53170]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:46:08.319 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:46:07 vm04 ceph-mon[53170]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:46:08.319 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:46:07 vm04 ceph-mon[53170]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:46:08.319 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:46:07 vm04 ceph-mon[53170]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "mon."}]: dispatch 2026-03-06T13:46:08.319 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:46:07 vm04 ceph-mon[53170]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' cmd=[{"prefix": "config get", "who": "mon", "key": "public_network"}]: dispatch 2026-03-06T13:46:08.319 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:46:07 vm04 ceph-mon[53170]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-06T13:46:08.319 DEBUG:teuthology.orchestra.run.vm04:> set -ex 2026-03-06T13:46:08.319 DEBUG:teuthology.orchestra.run.vm04:> sudo dd of=/etc/ceph/ceph.conf 2026-03-06T13:46:08.345 DEBUG:teuthology.orchestra.run.vm04:> set -ex 2026-03-06T13:46:08.345 DEBUG:teuthology.orchestra.run.vm04:> sudo dd of=/etc/ceph/ceph.client.admin.keyring 2026-03-06T13:46:08.408 DEBUG:teuthology.orchestra.run.vm06:> set -ex 2026-03-06T13:46:08.408 DEBUG:teuthology.orchestra.run.vm06:> sudo dd of=/etc/ceph/ceph.conf 2026-03-06T13:46:08.433 DEBUG:teuthology.orchestra.run.vm06:> set -ex 2026-03-06T13:46:08.433 DEBUG:teuthology.orchestra.run.vm06:> sudo dd of=/etc/ceph/ceph.client.admin.keyring 2026-03-06T13:46:08.499 INFO:tasks.cephadm:Adding mgr.a on vm02 2026-03-06T13:46:08.499 INFO:tasks.cephadm:Adding mgr.b on vm04 2026-03-06T13:46:08.499 DEBUG:teuthology.orchestra.run.vm06:> sudo /home/ubuntu/cephtest/cephadm --image harbor.clyso.com/custom-ceph/ceph/ceph:cobaltcore-storage-v19.2.3-fasttrack-3 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 2b1b4a72-195a-11f1-b1df-bf1f800d676b -- ceph orch apply mgr '2;vm02=a;vm04=b' 2026-03-06T13:46:08.886 INFO:teuthology.orchestra.run.vm06.stderr:Inferring config /var/lib/ceph/2b1b4a72-195a-11f1-b1df-bf1f800d676b/mon.c/config 2026-03-06T13:46:09.112 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:46:08 vm02 ceph-mon[49207]: Reconfiguring mon.a (monmap changed)... 2026-03-06T13:46:09.112 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:46:08 vm02 ceph-mon[49207]: Reconfiguring daemon mon.a on vm02 2026-03-06T13:46:09.112 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:46:08 vm02 ceph-mon[49207]: from='client.? 192.168.123.102:0/1161045360' entity='client.admin' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-06T13:46:09.112 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:46:08 vm02 ceph-mon[49207]: pgmap v10: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-06T13:46:09.112 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:46:08 vm02 ceph-mon[49207]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:46:09.112 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:46:08 vm02 ceph-mon[49207]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:46:09.112 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:46:08 vm02 ceph-mon[49207]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "mon."}]: dispatch 2026-03-06T13:46:09.112 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:46:08 vm02 ceph-mon[49207]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' cmd=[{"prefix": "config get", "who": "mon", "key": "public_network"}]: dispatch 2026-03-06T13:46:09.112 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:46:08 vm02 ceph-mon[49207]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-06T13:46:09.112 INFO:journalctl@ceph.mgr.a.vm02.stdout:Mar 06 13:46:08 vm02 ceph-2b1b4a72-195a-11f1-b1df-bf1f800d676b-mgr-a[49440]: 2026-03-06T12:46:08.779+0000 7f3e6c292640 -1 mgr.server handle_report got status from non-daemon mon.b 2026-03-06T13:46:09.171 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:46:08 vm06 ceph-mon[56528]: Reconfiguring mon.a (monmap changed)... 2026-03-06T13:46:09.171 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:46:08 vm06 ceph-mon[56528]: Reconfiguring daemon mon.a on vm02 2026-03-06T13:46:09.172 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:46:08 vm06 ceph-mon[56528]: from='client.? 192.168.123.102:0/1161045360' entity='client.admin' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-06T13:46:09.172 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:46:08 vm06 ceph-mon[56528]: pgmap v10: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-06T13:46:09.172 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:46:08 vm06 ceph-mon[56528]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:46:09.172 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:46:08 vm06 ceph-mon[56528]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:46:09.172 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:46:08 vm06 ceph-mon[56528]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "mon."}]: dispatch 2026-03-06T13:46:09.172 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:46:08 vm06 ceph-mon[56528]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' cmd=[{"prefix": "config get", "who": "mon", "key": "public_network"}]: dispatch 2026-03-06T13:46:09.172 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:46:08 vm06 ceph-mon[56528]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-06T13:46:09.288 INFO:teuthology.orchestra.run.vm06.stdout:Scheduled mgr update... 2026-03-06T13:46:09.299 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:46:08 vm04 ceph-mon[53170]: Reconfiguring mon.a (monmap changed)... 2026-03-06T13:46:09.300 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:46:08 vm04 ceph-mon[53170]: Reconfiguring daemon mon.a on vm02 2026-03-06T13:46:09.300 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:46:08 vm04 ceph-mon[53170]: from='client.? 192.168.123.102:0/1161045360' entity='client.admin' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-06T13:46:09.300 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:46:08 vm04 ceph-mon[53170]: pgmap v10: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-06T13:46:09.300 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:46:08 vm04 ceph-mon[53170]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:46:09.300 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:46:08 vm04 ceph-mon[53170]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:46:09.300 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:46:08 vm04 ceph-mon[53170]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "mon."}]: dispatch 2026-03-06T13:46:09.300 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:46:08 vm04 ceph-mon[53170]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' cmd=[{"prefix": "config get", "who": "mon", "key": "public_network"}]: dispatch 2026-03-06T13:46:09.300 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:46:08 vm04 ceph-mon[53170]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-06T13:46:09.448 DEBUG:teuthology.orchestra.run.vm04:mgr.b> sudo journalctl -f -n 0 -u ceph-2b1b4a72-195a-11f1-b1df-bf1f800d676b@mgr.b.service 2026-03-06T13:46:09.450 INFO:tasks.cephadm:Deploying OSDs... 2026-03-06T13:46:09.450 DEBUG:teuthology.orchestra.run.vm02:> set -ex 2026-03-06T13:46:09.450 DEBUG:teuthology.orchestra.run.vm02:> dd if=/scratch_devs of=/dev/stdout 2026-03-06T13:46:09.465 DEBUG:teuthology.orchestra.run:got remote process result: 1 2026-03-06T13:46:09.466 DEBUG:teuthology.orchestra.run.vm02:> ls /dev/[sv]d? 2026-03-06T13:46:09.523 INFO:teuthology.orchestra.run.vm02.stdout:/dev/vda 2026-03-06T13:46:09.523 INFO:teuthology.orchestra.run.vm02.stdout:/dev/vdb 2026-03-06T13:46:09.523 INFO:teuthology.orchestra.run.vm02.stdout:/dev/vdc 2026-03-06T13:46:09.523 INFO:teuthology.orchestra.run.vm02.stdout:/dev/vdd 2026-03-06T13:46:09.523 INFO:teuthology.orchestra.run.vm02.stdout:/dev/vde 2026-03-06T13:46:09.523 WARNING:teuthology.misc:Removing root device: /dev/vda from device list 2026-03-06T13:46:09.523 DEBUG:teuthology.misc:devs=['/dev/vdb', '/dev/vdc', '/dev/vdd', '/dev/vde'] 2026-03-06T13:46:09.523 DEBUG:teuthology.orchestra.run.vm02:> stat /dev/vdb 2026-03-06T13:46:09.582 INFO:teuthology.orchestra.run.vm02.stdout: File: /dev/vdb 2026-03-06T13:46:09.582 INFO:teuthology.orchestra.run.vm02.stdout: Size: 0 Blocks: 0 IO Block: 512 block special file 2026-03-06T13:46:09.582 INFO:teuthology.orchestra.run.vm02.stdout:Device: 6h/6d Inode: 254 Links: 1 Device type: fc,10 2026-03-06T13:46:09.582 INFO:teuthology.orchestra.run.vm02.stdout:Access: (0660/brw-rw----) Uid: ( 0/ root) Gid: ( 6/ disk) 2026-03-06T13:46:09.582 INFO:teuthology.orchestra.run.vm02.stdout:Context: system_u:object_r:fixed_disk_device_t:s0 2026-03-06T13:46:09.582 INFO:teuthology.orchestra.run.vm02.stdout:Access: 2026-03-06 13:46:03.303947386 +0100 2026-03-06T13:46:09.583 INFO:teuthology.orchestra.run.vm02.stdout:Modify: 2026-03-06 13:43:10.683000635 +0100 2026-03-06T13:46:09.583 INFO:teuthology.orchestra.run.vm02.stdout:Change: 2026-03-06 13:43:10.683000635 +0100 2026-03-06T13:46:09.583 INFO:teuthology.orchestra.run.vm02.stdout: Birth: 2026-03-06 13:38:25.248000000 +0100 2026-03-06T13:46:09.583 DEBUG:teuthology.orchestra.run.vm02:> sudo dd if=/dev/vdb of=/dev/null count=1 2026-03-06T13:46:09.649 INFO:teuthology.orchestra.run.vm02.stderr:1+0 records in 2026-03-06T13:46:09.649 INFO:teuthology.orchestra.run.vm02.stderr:1+0 records out 2026-03-06T13:46:09.649 INFO:teuthology.orchestra.run.vm02.stderr:512 bytes copied, 0.000160019 s, 3.2 MB/s 2026-03-06T13:46:09.650 DEBUG:teuthology.orchestra.run.vm02:> ! mount | grep -v devtmpfs | grep -q /dev/vdb 2026-03-06T13:46:09.707 DEBUG:teuthology.orchestra.run.vm02:> stat /dev/vdc 2026-03-06T13:46:09.763 INFO:teuthology.orchestra.run.vm02.stdout: File: /dev/vdc 2026-03-06T13:46:09.763 INFO:teuthology.orchestra.run.vm02.stdout: Size: 0 Blocks: 0 IO Block: 512 block special file 2026-03-06T13:46:09.763 INFO:teuthology.orchestra.run.vm02.stdout:Device: 6h/6d Inode: 255 Links: 1 Device type: fc,20 2026-03-06T13:46:09.763 INFO:teuthology.orchestra.run.vm02.stdout:Access: (0660/brw-rw----) Uid: ( 0/ root) Gid: ( 6/ disk) 2026-03-06T13:46:09.763 INFO:teuthology.orchestra.run.vm02.stdout:Context: system_u:object_r:fixed_disk_device_t:s0 2026-03-06T13:46:09.763 INFO:teuthology.orchestra.run.vm02.stdout:Access: 2026-03-06 13:46:03.308947388 +0100 2026-03-06T13:46:09.763 INFO:teuthology.orchestra.run.vm02.stdout:Modify: 2026-03-06 13:43:10.700000644 +0100 2026-03-06T13:46:09.763 INFO:teuthology.orchestra.run.vm02.stdout:Change: 2026-03-06 13:43:10.700000644 +0100 2026-03-06T13:46:09.763 INFO:teuthology.orchestra.run.vm02.stdout: Birth: 2026-03-06 13:38:25.255000000 +0100 2026-03-06T13:46:09.763 DEBUG:teuthology.orchestra.run.vm02:> sudo dd if=/dev/vdc of=/dev/null count=1 2026-03-06T13:46:09.826 INFO:teuthology.orchestra.run.vm02.stderr:1+0 records in 2026-03-06T13:46:09.826 INFO:teuthology.orchestra.run.vm02.stderr:1+0 records out 2026-03-06T13:46:09.826 INFO:teuthology.orchestra.run.vm02.stderr:512 bytes copied, 0.000130224 s, 3.9 MB/s 2026-03-06T13:46:09.827 DEBUG:teuthology.orchestra.run.vm02:> ! mount | grep -v devtmpfs | grep -q /dev/vdc 2026-03-06T13:46:09.884 DEBUG:teuthology.orchestra.run.vm02:> stat /dev/vdd 2026-03-06T13:46:09.942 INFO:teuthology.orchestra.run.vm02.stdout: File: /dev/vdd 2026-03-06T13:46:09.942 INFO:teuthology.orchestra.run.vm02.stdout: Size: 0 Blocks: 0 IO Block: 512 block special file 2026-03-06T13:46:09.942 INFO:teuthology.orchestra.run.vm02.stdout:Device: 6h/6d Inode: 256 Links: 1 Device type: fc,30 2026-03-06T13:46:09.942 INFO:teuthology.orchestra.run.vm02.stdout:Access: (0660/brw-rw----) Uid: ( 0/ root) Gid: ( 6/ disk) 2026-03-06T13:46:09.942 INFO:teuthology.orchestra.run.vm02.stdout:Context: system_u:object_r:fixed_disk_device_t:s0 2026-03-06T13:46:09.942 INFO:teuthology.orchestra.run.vm02.stdout:Access: 2026-03-06 13:46:03.312947389 +0100 2026-03-06T13:46:09.942 INFO:teuthology.orchestra.run.vm02.stdout:Modify: 2026-03-06 13:43:10.692000640 +0100 2026-03-06T13:46:09.942 INFO:teuthology.orchestra.run.vm02.stdout:Change: 2026-03-06 13:43:10.692000640 +0100 2026-03-06T13:46:09.942 INFO:teuthology.orchestra.run.vm02.stdout: Birth: 2026-03-06 13:38:25.262000000 +0100 2026-03-06T13:46:09.942 DEBUG:teuthology.orchestra.run.vm02:> sudo dd if=/dev/vdd of=/dev/null count=1 2026-03-06T13:46:10.006 INFO:teuthology.orchestra.run.vm02.stderr:1+0 records in 2026-03-06T13:46:10.006 INFO:teuthology.orchestra.run.vm02.stderr:1+0 records out 2026-03-06T13:46:10.006 INFO:teuthology.orchestra.run.vm02.stderr:512 bytes copied, 0.000189715 s, 2.7 MB/s 2026-03-06T13:46:10.007 DEBUG:teuthology.orchestra.run.vm02:> ! mount | grep -v devtmpfs | grep -q /dev/vdd 2026-03-06T13:46:10.065 DEBUG:teuthology.orchestra.run.vm02:> stat /dev/vde 2026-03-06T13:46:10.125 INFO:teuthology.orchestra.run.vm02.stdout: File: /dev/vde 2026-03-06T13:46:10.125 INFO:teuthology.orchestra.run.vm02.stdout: Size: 0 Blocks: 0 IO Block: 512 block special file 2026-03-06T13:46:10.126 INFO:teuthology.orchestra.run.vm02.stdout:Device: 6h/6d Inode: 257 Links: 1 Device type: fc,40 2026-03-06T13:46:10.126 INFO:teuthology.orchestra.run.vm02.stdout:Access: (0660/brw-rw----) Uid: ( 0/ root) Gid: ( 6/ disk) 2026-03-06T13:46:10.126 INFO:teuthology.orchestra.run.vm02.stdout:Context: system_u:object_r:fixed_disk_device_t:s0 2026-03-06T13:46:10.126 INFO:teuthology.orchestra.run.vm02.stdout:Access: 2026-03-06 13:46:03.316947391 +0100 2026-03-06T13:46:10.126 INFO:teuthology.orchestra.run.vm02.stdout:Modify: 2026-03-06 13:43:10.691000639 +0100 2026-03-06T13:46:10.126 INFO:teuthology.orchestra.run.vm02.stdout:Change: 2026-03-06 13:43:10.691000639 +0100 2026-03-06T13:46:10.126 INFO:teuthology.orchestra.run.vm02.stdout: Birth: 2026-03-06 13:38:25.327000000 +0100 2026-03-06T13:46:10.126 DEBUG:teuthology.orchestra.run.vm02:> sudo dd if=/dev/vde of=/dev/null count=1 2026-03-06T13:46:10.192 INFO:teuthology.orchestra.run.vm02.stderr:1+0 records in 2026-03-06T13:46:10.192 INFO:teuthology.orchestra.run.vm02.stderr:1+0 records out 2026-03-06T13:46:10.193 INFO:teuthology.orchestra.run.vm02.stderr:512 bytes copied, 0.00020778 s, 2.5 MB/s 2026-03-06T13:46:10.194 DEBUG:teuthology.orchestra.run.vm02:> ! mount | grep -v devtmpfs | grep -q /dev/vde 2026-03-06T13:46:10.254 DEBUG:teuthology.orchestra.run.vm04:> set -ex 2026-03-06T13:46:10.254 DEBUG:teuthology.orchestra.run.vm04:> dd if=/scratch_devs of=/dev/stdout 2026-03-06T13:46:10.273 DEBUG:teuthology.orchestra.run:got remote process result: 1 2026-03-06T13:46:10.273 DEBUG:teuthology.orchestra.run.vm04:> ls /dev/[sv]d? 2026-03-06T13:46:10.289 INFO:journalctl@ceph.mgr.b.vm04.stdout:Mar 06 13:46:10 vm04 systemd[1]: Starting Ceph mgr.b for 2b1b4a72-195a-11f1-b1df-bf1f800d676b... 2026-03-06T13:46:10.310 INFO:teuthology.orchestra.run.vm04.stdout:/dev/vda 2026-03-06T13:46:10.310 INFO:teuthology.orchestra.run.vm04.stdout:/dev/vdb 2026-03-06T13:46:10.310 INFO:teuthology.orchestra.run.vm04.stdout:/dev/vdc 2026-03-06T13:46:10.310 INFO:teuthology.orchestra.run.vm04.stdout:/dev/vdd 2026-03-06T13:46:10.310 INFO:teuthology.orchestra.run.vm04.stdout:/dev/vde 2026-03-06T13:46:10.311 WARNING:teuthology.misc:Removing root device: /dev/vda from device list 2026-03-06T13:46:10.311 DEBUG:teuthology.misc:devs=['/dev/vdb', '/dev/vdc', '/dev/vdd', '/dev/vde'] 2026-03-06T13:46:10.311 DEBUG:teuthology.orchestra.run.vm04:> stat /dev/vdb 2026-03-06T13:46:10.374 INFO:teuthology.orchestra.run.vm04.stdout: File: /dev/vdb 2026-03-06T13:46:10.374 INFO:teuthology.orchestra.run.vm04.stdout: Size: 0 Blocks: 0 IO Block: 512 block special file 2026-03-06T13:46:10.374 INFO:teuthology.orchestra.run.vm04.stdout:Device: 6h/6d Inode: 221 Links: 1 Device type: fc,10 2026-03-06T13:46:10.374 INFO:teuthology.orchestra.run.vm04.stdout:Access: (0660/brw-rw----) Uid: ( 0/ root) Gid: ( 6/ disk) 2026-03-06T13:46:10.374 INFO:teuthology.orchestra.run.vm04.stdout:Context: system_u:object_r:fixed_disk_device_t:s0 2026-03-06T13:46:10.374 INFO:teuthology.orchestra.run.vm04.stdout:Access: 2026-03-06 13:46:03.524741548 +0100 2026-03-06T13:46:10.374 INFO:teuthology.orchestra.run.vm04.stdout:Modify: 2026-03-06 13:43:10.535776134 +0100 2026-03-06T13:46:10.374 INFO:teuthology.orchestra.run.vm04.stdout:Change: 2026-03-06 13:43:10.535776134 +0100 2026-03-06T13:46:10.374 INFO:teuthology.orchestra.run.vm04.stdout: Birth: 2026-03-06 13:38:56.218000000 +0100 2026-03-06T13:46:10.374 DEBUG:teuthology.orchestra.run.vm04:> sudo dd if=/dev/vdb of=/dev/null count=1 2026-03-06T13:46:10.467 INFO:teuthology.orchestra.run.vm04.stderr:1+0 records in 2026-03-06T13:46:10.467 INFO:teuthology.orchestra.run.vm04.stderr:1+0 records out 2026-03-06T13:46:10.467 INFO:teuthology.orchestra.run.vm04.stderr:512 bytes copied, 0.000104765 s, 4.9 MB/s 2026-03-06T13:46:10.467 DEBUG:teuthology.orchestra.run.vm04:> ! mount | grep -v devtmpfs | grep -q /dev/vdb 2026-03-06T13:46:10.495 DEBUG:teuthology.orchestra.run.vm04:> stat /dev/vdc 2026-03-06T13:46:10.552 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:46:10 vm04 ceph-mon[53170]: Reconfiguring mon.c (monmap changed)... 2026-03-06T13:46:10.552 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:46:10 vm04 ceph-mon[53170]: Reconfiguring daemon mon.c on vm06 2026-03-06T13:46:10.552 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:46:10 vm04 ceph-mon[53170]: from='client.14214 -' entity='client.admin' cmd=[{"prefix": "orch apply", "service_type": "mgr", "placement": "2;vm02=a;vm04=b", "target": ["mon-mgr", ""]}]: dispatch 2026-03-06T13:46:10.552 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:46:10 vm04 ceph-mon[53170]: Saving service mgr spec with placement vm02=a;vm04=b;count:2 2026-03-06T13:46:10.552 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:46:10 vm04 ceph-mon[53170]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:46:10.552 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:46:10 vm04 ceph-mon[53170]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:46:10.552 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:46:10 vm04 ceph-mon[53170]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:46:10.552 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:46:10 vm04 ceph-mon[53170]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-06T13:46:10.552 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:46:10 vm04 ceph-mon[53170]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-06T13:46:10.553 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:46:10 vm04 ceph-mon[53170]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-06T13:46:10.553 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:46:10 vm04 ceph-mon[53170]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:46:10.553 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:46:10 vm04 ceph-mon[53170]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' cmd=[{"prefix": "auth get-or-create", "entity": "mgr.b", "caps": ["mon", "profile mgr", "osd", "allow *", "mds", "allow *"]}]: dispatch 2026-03-06T13:46:10.553 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:46:10 vm04 ceph-mon[53170]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' cmd='[{"prefix": "auth get-or-create", "entity": "mgr.b", "caps": ["mon", "profile mgr", "osd", "allow *", "mds", "allow *"]}]': finished 2026-03-06T13:46:10.553 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:46:10 vm04 ceph-mon[53170]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' cmd=[{"prefix": "mgr services"}]: dispatch 2026-03-06T13:46:10.553 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:46:10 vm04 ceph-mon[53170]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-06T13:46:10.553 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:46:10 vm04 ceph-mon[53170]: Deploying daemon mgr.b on vm04 2026-03-06T13:46:10.553 INFO:journalctl@ceph.mgr.b.vm04.stdout:Mar 06 13:46:10 vm04 podman[54758]: 2026-03-06 13:46:10.349986249 +0100 CET m=+0.020513368 container create dfe14d6f23c7f94e074b71a0ce7b950025bfb78dd4fc4ac1f471d375e36511f5 (image=harbor.clyso.com/custom-ceph/ceph/ceph@sha256:26363c7a4eea9ef5a0148afc7b2a22b6f486596d87a30c2a9fdcda5db3eca62b, name=ceph-2b1b4a72-195a-11f1-b1df-bf1f800d676b-mgr-b, CEPH_GIT_REPO=https://github.com/irq0/ceph.git, CEPH_REF=19.2.3-47-gc24117fd552, CEPH_SHA1=c24117fd5525679b799527bc1bd1f1dd0a2db5e2, FROM_IMAGE=rockylinux:9, GANESHA_REPO_BASEURL=https://buildlogs.centos.org/centos/$releasever-stream/storage/$basearch/nfsganesha-5/, OSD_FLAVOR=default, ceph=True, io.buildah.version=1.41.8) 2026-03-06T13:46:10.553 INFO:journalctl@ceph.mgr.b.vm04.stdout:Mar 06 13:46:10 vm04 podman[54758]: 2026-03-06 13:46:10.410811927 +0100 CET m=+0.081339046 container init dfe14d6f23c7f94e074b71a0ce7b950025bfb78dd4fc4ac1f471d375e36511f5 (image=harbor.clyso.com/custom-ceph/ceph/ceph@sha256:26363c7a4eea9ef5a0148afc7b2a22b6f486596d87a30c2a9fdcda5db3eca62b, name=ceph-2b1b4a72-195a-11f1-b1df-bf1f800d676b-mgr-b, CEPH_REF=19.2.3-47-gc24117fd552, CEPH_SHA1=c24117fd5525679b799527bc1bd1f1dd0a2db5e2, FROM_IMAGE=rockylinux:9, GANESHA_REPO_BASEURL=https://buildlogs.centos.org/centos/$releasever-stream/storage/$basearch/nfsganesha-5/, OSD_FLAVOR=default, ceph=True, io.buildah.version=1.41.8, CEPH_GIT_REPO=https://github.com/irq0/ceph.git) 2026-03-06T13:46:10.553 INFO:journalctl@ceph.mgr.b.vm04.stdout:Mar 06 13:46:10 vm04 podman[54758]: 2026-03-06 13:46:10.417337409 +0100 CET m=+0.087864528 container start dfe14d6f23c7f94e074b71a0ce7b950025bfb78dd4fc4ac1f471d375e36511f5 (image=harbor.clyso.com/custom-ceph/ceph/ceph@sha256:26363c7a4eea9ef5a0148afc7b2a22b6f486596d87a30c2a9fdcda5db3eca62b, name=ceph-2b1b4a72-195a-11f1-b1df-bf1f800d676b-mgr-b, CEPH_GIT_REPO=https://github.com/irq0/ceph.git, CEPH_REF=19.2.3-47-gc24117fd552, CEPH_SHA1=c24117fd5525679b799527bc1bd1f1dd0a2db5e2, FROM_IMAGE=rockylinux:9, GANESHA_REPO_BASEURL=https://buildlogs.centos.org/centos/$releasever-stream/storage/$basearch/nfsganesha-5/, OSD_FLAVOR=default, ceph=True, io.buildah.version=1.41.8) 2026-03-06T13:46:10.553 INFO:journalctl@ceph.mgr.b.vm04.stdout:Mar 06 13:46:10 vm04 bash[54758]: dfe14d6f23c7f94e074b71a0ce7b950025bfb78dd4fc4ac1f471d375e36511f5 2026-03-06T13:46:10.553 INFO:journalctl@ceph.mgr.b.vm04.stdout:Mar 06 13:46:10 vm04 podman[54758]: 2026-03-06 13:46:10.339499305 +0100 CET m=+0.010026424 image pull 306e97de47e91c2b4b24d3dc09be3b3a12039b078f343d91220102acc6628a68 harbor.clyso.com/custom-ceph/ceph/ceph@sha256:26363c7a4eea9ef5a0148afc7b2a22b6f486596d87a30c2a9fdcda5db3eca62b 2026-03-06T13:46:10.553 INFO:journalctl@ceph.mgr.b.vm04.stdout:Mar 06 13:46:10 vm04 systemd[1]: Started Ceph mgr.b for 2b1b4a72-195a-11f1-b1df-bf1f800d676b. 2026-03-06T13:46:10.558 INFO:teuthology.orchestra.run.vm04.stdout: File: /dev/vdc 2026-03-06T13:46:10.558 INFO:teuthology.orchestra.run.vm04.stdout: Size: 0 Blocks: 0 IO Block: 512 block special file 2026-03-06T13:46:10.558 INFO:teuthology.orchestra.run.vm04.stdout:Device: 6h/6d Inode: 222 Links: 1 Device type: fc,20 2026-03-06T13:46:10.558 INFO:teuthology.orchestra.run.vm04.stdout:Access: (0660/brw-rw----) Uid: ( 0/ root) Gid: ( 6/ disk) 2026-03-06T13:46:10.559 INFO:teuthology.orchestra.run.vm04.stdout:Context: system_u:object_r:fixed_disk_device_t:s0 2026-03-06T13:46:10.559 INFO:teuthology.orchestra.run.vm04.stdout:Access: 2026-03-06 13:46:03.528741548 +0100 2026-03-06T13:46:10.559 INFO:teuthology.orchestra.run.vm04.stdout:Modify: 2026-03-06 13:43:10.499776116 +0100 2026-03-06T13:46:10.559 INFO:teuthology.orchestra.run.vm04.stdout:Change: 2026-03-06 13:43:10.499776116 +0100 2026-03-06T13:46:10.559 INFO:teuthology.orchestra.run.vm04.stdout: Birth: 2026-03-06 13:38:56.221000000 +0100 2026-03-06T13:46:10.559 DEBUG:teuthology.orchestra.run.vm04:> sudo dd if=/dev/vdc of=/dev/null count=1 2026-03-06T13:46:10.612 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:46:10 vm02 ceph-mon[49207]: Reconfiguring mon.c (monmap changed)... 2026-03-06T13:46:10.612 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:46:10 vm02 ceph-mon[49207]: Reconfiguring daemon mon.c on vm06 2026-03-06T13:46:10.612 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:46:10 vm02 ceph-mon[49207]: from='client.14214 -' entity='client.admin' cmd=[{"prefix": "orch apply", "service_type": "mgr", "placement": "2;vm02=a;vm04=b", "target": ["mon-mgr", ""]}]: dispatch 2026-03-06T13:46:10.612 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:46:10 vm02 ceph-mon[49207]: Saving service mgr spec with placement vm02=a;vm04=b;count:2 2026-03-06T13:46:10.612 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:46:10 vm02 ceph-mon[49207]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:46:10.612 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:46:10 vm02 ceph-mon[49207]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:46:10.612 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:46:10 vm02 ceph-mon[49207]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:46:10.612 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:46:10 vm02 ceph-mon[49207]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-06T13:46:10.612 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:46:10 vm02 ceph-mon[49207]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-06T13:46:10.612 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:46:10 vm02 ceph-mon[49207]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-06T13:46:10.612 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:46:10 vm02 ceph-mon[49207]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:46:10.612 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:46:10 vm02 ceph-mon[49207]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' cmd=[{"prefix": "auth get-or-create", "entity": "mgr.b", "caps": ["mon", "profile mgr", "osd", "allow *", "mds", "allow *"]}]: dispatch 2026-03-06T13:46:10.612 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:46:10 vm02 ceph-mon[49207]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' cmd='[{"prefix": "auth get-or-create", "entity": "mgr.b", "caps": ["mon", "profile mgr", "osd", "allow *", "mds", "allow *"]}]': finished 2026-03-06T13:46:10.612 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:46:10 vm02 ceph-mon[49207]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' cmd=[{"prefix": "mgr services"}]: dispatch 2026-03-06T13:46:10.612 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:46:10 vm02 ceph-mon[49207]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-06T13:46:10.612 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:46:10 vm02 ceph-mon[49207]: Deploying daemon mgr.b on vm04 2026-03-06T13:46:10.625 INFO:teuthology.orchestra.run.vm04.stderr:1+0 records in 2026-03-06T13:46:10.625 INFO:teuthology.orchestra.run.vm04.stderr:1+0 records out 2026-03-06T13:46:10.625 INFO:teuthology.orchestra.run.vm04.stderr:512 bytes copied, 0.000155772 s, 3.3 MB/s 2026-03-06T13:46:10.626 DEBUG:teuthology.orchestra.run.vm04:> ! mount | grep -v devtmpfs | grep -q /dev/vdc 2026-03-06T13:46:10.690 DEBUG:teuthology.orchestra.run.vm04:> stat /dev/vdd 2026-03-06T13:46:10.749 INFO:teuthology.orchestra.run.vm04.stdout: File: /dev/vdd 2026-03-06T13:46:10.749 INFO:teuthology.orchestra.run.vm04.stdout: Size: 0 Blocks: 0 IO Block: 512 block special file 2026-03-06T13:46:10.749 INFO:teuthology.orchestra.run.vm04.stdout:Device: 6h/6d Inode: 223 Links: 1 Device type: fc,30 2026-03-06T13:46:10.749 INFO:teuthology.orchestra.run.vm04.stdout:Access: (0660/brw-rw----) Uid: ( 0/ root) Gid: ( 6/ disk) 2026-03-06T13:46:10.749 INFO:teuthology.orchestra.run.vm04.stdout:Context: system_u:object_r:fixed_disk_device_t:s0 2026-03-06T13:46:10.749 INFO:teuthology.orchestra.run.vm04.stdout:Access: 2026-03-06 13:46:03.532741548 +0100 2026-03-06T13:46:10.749 INFO:teuthology.orchestra.run.vm04.stdout:Modify: 2026-03-06 13:43:10.496776114 +0100 2026-03-06T13:46:10.749 INFO:teuthology.orchestra.run.vm04.stdout:Change: 2026-03-06 13:43:10.496776114 +0100 2026-03-06T13:46:10.749 INFO:teuthology.orchestra.run.vm04.stdout: Birth: 2026-03-06 13:38:56.222000000 +0100 2026-03-06T13:46:10.749 DEBUG:teuthology.orchestra.run.vm04:> sudo dd if=/dev/vdd of=/dev/null count=1 2026-03-06T13:46:10.760 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:46:10 vm06 ceph-mon[56528]: Reconfiguring mon.c (monmap changed)... 2026-03-06T13:46:10.760 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:46:10 vm06 ceph-mon[56528]: Reconfiguring daemon mon.c on vm06 2026-03-06T13:46:10.760 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:46:10 vm06 ceph-mon[56528]: from='client.14214 -' entity='client.admin' cmd=[{"prefix": "orch apply", "service_type": "mgr", "placement": "2;vm02=a;vm04=b", "target": ["mon-mgr", ""]}]: dispatch 2026-03-06T13:46:10.760 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:46:10 vm06 ceph-mon[56528]: Saving service mgr spec with placement vm02=a;vm04=b;count:2 2026-03-06T13:46:10.760 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:46:10 vm06 ceph-mon[56528]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:46:10.760 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:46:10 vm06 ceph-mon[56528]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:46:10.760 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:46:10 vm06 ceph-mon[56528]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:46:10.761 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:46:10 vm06 ceph-mon[56528]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-06T13:46:10.761 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:46:10 vm06 ceph-mon[56528]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-06T13:46:10.761 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:46:10 vm06 ceph-mon[56528]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-06T13:46:10.761 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:46:10 vm06 ceph-mon[56528]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:46:10.761 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:46:10 vm06 ceph-mon[56528]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' cmd=[{"prefix": "auth get-or-create", "entity": "mgr.b", "caps": ["mon", "profile mgr", "osd", "allow *", "mds", "allow *"]}]: dispatch 2026-03-06T13:46:10.761 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:46:10 vm06 ceph-mon[56528]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' cmd='[{"prefix": "auth get-or-create", "entity": "mgr.b", "caps": ["mon", "profile mgr", "osd", "allow *", "mds", "allow *"]}]': finished 2026-03-06T13:46:10.761 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:46:10 vm06 ceph-mon[56528]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' cmd=[{"prefix": "mgr services"}]: dispatch 2026-03-06T13:46:10.761 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:46:10 vm06 ceph-mon[56528]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-06T13:46:10.761 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:46:10 vm06 ceph-mon[56528]: Deploying daemon mgr.b on vm04 2026-03-06T13:46:10.818 INFO:teuthology.orchestra.run.vm04.stderr:1+0 records in 2026-03-06T13:46:10.818 INFO:teuthology.orchestra.run.vm04.stderr:1+0 records out 2026-03-06T13:46:10.818 INFO:teuthology.orchestra.run.vm04.stderr:512 bytes copied, 0.000132809 s, 3.9 MB/s 2026-03-06T13:46:10.819 DEBUG:teuthology.orchestra.run.vm04:> ! mount | grep -v devtmpfs | grep -q /dev/vdd 2026-03-06T13:46:10.881 DEBUG:teuthology.orchestra.run.vm04:> stat /dev/vde 2026-03-06T13:46:10.945 INFO:teuthology.orchestra.run.vm04.stdout: File: /dev/vde 2026-03-06T13:46:10.945 INFO:teuthology.orchestra.run.vm04.stdout: Size: 0 Blocks: 0 IO Block: 512 block special file 2026-03-06T13:46:10.945 INFO:teuthology.orchestra.run.vm04.stdout:Device: 6h/6d Inode: 224 Links: 1 Device type: fc,40 2026-03-06T13:46:10.945 INFO:teuthology.orchestra.run.vm04.stdout:Access: (0660/brw-rw----) Uid: ( 0/ root) Gid: ( 6/ disk) 2026-03-06T13:46:10.945 INFO:teuthology.orchestra.run.vm04.stdout:Context: system_u:object_r:fixed_disk_device_t:s0 2026-03-06T13:46:10.945 INFO:teuthology.orchestra.run.vm04.stdout:Access: 2026-03-06 13:46:03.536741547 +0100 2026-03-06T13:46:10.945 INFO:teuthology.orchestra.run.vm04.stdout:Modify: 2026-03-06 13:43:10.509776121 +0100 2026-03-06T13:46:10.945 INFO:teuthology.orchestra.run.vm04.stdout:Change: 2026-03-06 13:43:10.509776121 +0100 2026-03-06T13:46:10.945 INFO:teuthology.orchestra.run.vm04.stdout: Birth: 2026-03-06 13:38:56.225000000 +0100 2026-03-06T13:46:10.945 DEBUG:teuthology.orchestra.run.vm04:> sudo dd if=/dev/vde of=/dev/null count=1 2026-03-06T13:46:11.009 INFO:journalctl@ceph.mgr.b.vm04.stdout:Mar 06 13:46:10 vm04 ceph-2b1b4a72-195a-11f1-b1df-bf1f800d676b-mgr-b[54791]: 2026-03-06T12:46:10.696+0000 7ff9bb54a100 -1 mgr[py] Module alerts has missing NOTIFY_TYPES member 2026-03-06T13:46:11.009 INFO:journalctl@ceph.mgr.b.vm04.stdout:Mar 06 13:46:10 vm04 ceph-2b1b4a72-195a-11f1-b1df-bf1f800d676b-mgr-b[54791]: 2026-03-06T12:46:10.835+0000 7ff9bb54a100 -1 mgr[py] Module balancer has missing NOTIFY_TYPES member 2026-03-06T13:46:11.011 INFO:teuthology.orchestra.run.vm04.stderr:1+0 records in 2026-03-06T13:46:11.011 INFO:teuthology.orchestra.run.vm04.stderr:1+0 records out 2026-03-06T13:46:11.011 INFO:teuthology.orchestra.run.vm04.stderr:512 bytes copied, 0.000134612 s, 3.8 MB/s 2026-03-06T13:46:11.012 DEBUG:teuthology.orchestra.run.vm04:> ! mount | grep -v devtmpfs | grep -q /dev/vde 2026-03-06T13:46:11.075 DEBUG:teuthology.orchestra.run.vm06:> set -ex 2026-03-06T13:46:11.075 DEBUG:teuthology.orchestra.run.vm06:> dd if=/scratch_devs of=/dev/stdout 2026-03-06T13:46:11.092 DEBUG:teuthology.orchestra.run:got remote process result: 1 2026-03-06T13:46:11.093 DEBUG:teuthology.orchestra.run.vm06:> ls /dev/[sv]d? 2026-03-06T13:46:11.151 INFO:teuthology.orchestra.run.vm06.stdout:/dev/vda 2026-03-06T13:46:11.151 INFO:teuthology.orchestra.run.vm06.stdout:/dev/vdb 2026-03-06T13:46:11.151 INFO:teuthology.orchestra.run.vm06.stdout:/dev/vdc 2026-03-06T13:46:11.151 INFO:teuthology.orchestra.run.vm06.stdout:/dev/vdd 2026-03-06T13:46:11.151 INFO:teuthology.orchestra.run.vm06.stdout:/dev/vde 2026-03-06T13:46:11.151 WARNING:teuthology.misc:Removing root device: /dev/vda from device list 2026-03-06T13:46:11.151 DEBUG:teuthology.misc:devs=['/dev/vdb', '/dev/vdc', '/dev/vdd', '/dev/vde'] 2026-03-06T13:46:11.151 DEBUG:teuthology.orchestra.run.vm06:> stat /dev/vdb 2026-03-06T13:46:11.214 INFO:teuthology.orchestra.run.vm06.stdout: File: /dev/vdb 2026-03-06T13:46:11.214 INFO:teuthology.orchestra.run.vm06.stdout: Size: 0 Blocks: 0 IO Block: 512 block special file 2026-03-06T13:46:11.214 INFO:teuthology.orchestra.run.vm06.stdout:Device: 6h/6d Inode: 233 Links: 1 Device type: fc,10 2026-03-06T13:46:11.214 INFO:teuthology.orchestra.run.vm06.stdout:Access: (0660/brw-rw----) Uid: ( 0/ root) Gid: ( 6/ disk) 2026-03-06T13:46:11.214 INFO:teuthology.orchestra.run.vm06.stdout:Context: system_u:object_r:fixed_disk_device_t:s0 2026-03-06T13:46:11.214 INFO:teuthology.orchestra.run.vm06.stdout:Access: 2026-03-06 13:46:04.362685584 +0100 2026-03-06T13:46:11.214 INFO:teuthology.orchestra.run.vm06.stdout:Modify: 2026-03-06 13:43:10.850782891 +0100 2026-03-06T13:46:11.214 INFO:teuthology.orchestra.run.vm06.stdout:Change: 2026-03-06 13:43:10.850782891 +0100 2026-03-06T13:46:11.214 INFO:teuthology.orchestra.run.vm06.stdout: Birth: 2026-03-06 13:39:27.237000000 +0100 2026-03-06T13:46:11.214 DEBUG:teuthology.orchestra.run.vm06:> sudo dd if=/dev/vdb of=/dev/null count=1 2026-03-06T13:46:11.280 INFO:teuthology.orchestra.run.vm06.stderr:1+0 records in 2026-03-06T13:46:11.280 INFO:teuthology.orchestra.run.vm06.stderr:1+0 records out 2026-03-06T13:46:11.280 INFO:teuthology.orchestra.run.vm06.stderr:512 bytes copied, 0.000148026 s, 3.5 MB/s 2026-03-06T13:46:11.281 DEBUG:teuthology.orchestra.run.vm06:> ! mount | grep -v devtmpfs | grep -q /dev/vdb 2026-03-06T13:46:11.337 DEBUG:teuthology.orchestra.run.vm06:> stat /dev/vdc 2026-03-06T13:46:11.394 INFO:teuthology.orchestra.run.vm06.stdout: File: /dev/vdc 2026-03-06T13:46:11.394 INFO:teuthology.orchestra.run.vm06.stdout: Size: 0 Blocks: 0 IO Block: 512 block special file 2026-03-06T13:46:11.394 INFO:teuthology.orchestra.run.vm06.stdout:Device: 6h/6d Inode: 255 Links: 1 Device type: fc,20 2026-03-06T13:46:11.394 INFO:teuthology.orchestra.run.vm06.stdout:Access: (0660/brw-rw----) Uid: ( 0/ root) Gid: ( 6/ disk) 2026-03-06T13:46:11.394 INFO:teuthology.orchestra.run.vm06.stdout:Context: system_u:object_r:fixed_disk_device_t:s0 2026-03-06T13:46:11.394 INFO:teuthology.orchestra.run.vm06.stdout:Access: 2026-03-06 13:46:04.366685585 +0100 2026-03-06T13:46:11.394 INFO:teuthology.orchestra.run.vm06.stdout:Modify: 2026-03-06 13:43:10.816782877 +0100 2026-03-06T13:46:11.394 INFO:teuthology.orchestra.run.vm06.stdout:Change: 2026-03-06 13:43:10.816782877 +0100 2026-03-06T13:46:11.394 INFO:teuthology.orchestra.run.vm06.stdout: Birth: 2026-03-06 13:39:27.245000000 +0100 2026-03-06T13:46:11.395 DEBUG:teuthology.orchestra.run.vm06:> sudo dd if=/dev/vdc of=/dev/null count=1 2026-03-06T13:46:11.457 INFO:teuthology.orchestra.run.vm06.stderr:1+0 records in 2026-03-06T13:46:11.457 INFO:teuthology.orchestra.run.vm06.stderr:1+0 records out 2026-03-06T13:46:11.457 INFO:teuthology.orchestra.run.vm06.stderr:512 bytes copied, 0.00014388 s, 3.6 MB/s 2026-03-06T13:46:11.458 DEBUG:teuthology.orchestra.run.vm06:> ! mount | grep -v devtmpfs | grep -q /dev/vdc 2026-03-06T13:46:11.515 DEBUG:teuthology.orchestra.run.vm06:> stat /dev/vdd 2026-03-06T13:46:11.572 INFO:teuthology.orchestra.run.vm06.stdout: File: /dev/vdd 2026-03-06T13:46:11.572 INFO:teuthology.orchestra.run.vm06.stdout: Size: 0 Blocks: 0 IO Block: 512 block special file 2026-03-06T13:46:11.572 INFO:teuthology.orchestra.run.vm06.stdout:Device: 6h/6d Inode: 256 Links: 1 Device type: fc,30 2026-03-06T13:46:11.572 INFO:teuthology.orchestra.run.vm06.stdout:Access: (0660/brw-rw----) Uid: ( 0/ root) Gid: ( 6/ disk) 2026-03-06T13:46:11.572 INFO:teuthology.orchestra.run.vm06.stdout:Context: system_u:object_r:fixed_disk_device_t:s0 2026-03-06T13:46:11.572 INFO:teuthology.orchestra.run.vm06.stdout:Access: 2026-03-06 13:46:04.370685586 +0100 2026-03-06T13:46:11.572 INFO:teuthology.orchestra.run.vm06.stdout:Modify: 2026-03-06 13:43:10.870782899 +0100 2026-03-06T13:46:11.572 INFO:teuthology.orchestra.run.vm06.stdout:Change: 2026-03-06 13:43:10.870782899 +0100 2026-03-06T13:46:11.572 INFO:teuthology.orchestra.run.vm06.stdout: Birth: 2026-03-06 13:39:27.249000000 +0100 2026-03-06T13:46:11.573 DEBUG:teuthology.orchestra.run.vm06:> sudo dd if=/dev/vdd of=/dev/null count=1 2026-03-06T13:46:11.637 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:46:11 vm06 ceph-mon[56528]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:46:11.637 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:46:11 vm06 ceph-mon[56528]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:46:11.637 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:46:11 vm06 ceph-mon[56528]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:46:11.637 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:46:11 vm06 ceph-mon[56528]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:46:11.637 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:46:11 vm06 ceph-mon[56528]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-06T13:46:11.637 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:46:11 vm06 ceph-mon[56528]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-06T13:46:11.637 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:46:11 vm06 ceph-mon[56528]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-06T13:46:11.637 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:46:11 vm06 ceph-mon[56528]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:46:11.637 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:46:11 vm06 ceph-mon[56528]: Metadata not up to date on all hosts. Skipping non agent specs 2026-03-06T13:46:11.637 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:46:11 vm06 ceph-mon[56528]: pgmap v11: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-06T13:46:11.637 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:46:11 vm06 ceph-mon[56528]: Reconfiguring mgr.a (unknown last config time)... 2026-03-06T13:46:11.637 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:46:11 vm06 ceph-mon[56528]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' cmd=[{"prefix": "auth get-or-create", "entity": "mgr.a", "caps": ["mon", "profile mgr", "osd", "allow *", "mds", "allow *"]}]: dispatch 2026-03-06T13:46:11.637 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:46:11 vm06 ceph-mon[56528]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' cmd=[{"prefix": "mgr services"}]: dispatch 2026-03-06T13:46:11.637 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:46:11 vm06 ceph-mon[56528]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-06T13:46:11.637 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:46:11 vm06 ceph-mon[56528]: Reconfiguring daemon mgr.a on vm02 2026-03-06T13:46:11.637 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:46:11 vm06 ceph-mon[56528]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:46:11.637 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:46:11 vm06 ceph-mon[56528]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:46:11.637 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:46:11 vm06 ceph-mon[56528]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-06T13:46:11.637 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:46:11 vm06 ceph-mon[56528]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-06T13:46:11.637 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:46:11 vm06 ceph-mon[56528]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-06T13:46:11.637 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:46:11 vm06 ceph-mon[56528]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:46:11.640 INFO:teuthology.orchestra.run.vm06.stderr:1+0 records in 2026-03-06T13:46:11.640 INFO:teuthology.orchestra.run.vm06.stderr:1+0 records out 2026-03-06T13:46:11.640 INFO:teuthology.orchestra.run.vm06.stderr:512 bytes copied, 0.00016011 s, 3.2 MB/s 2026-03-06T13:46:11.641 DEBUG:teuthology.orchestra.run.vm06:> ! mount | grep -v devtmpfs | grep -q /dev/vdd 2026-03-06T13:46:11.680 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:46:11 vm04 ceph-mon[53170]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:46:11.680 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:46:11 vm04 ceph-mon[53170]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:46:11.680 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:46:11 vm04 ceph-mon[53170]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:46:11.680 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:46:11 vm04 ceph-mon[53170]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:46:11.680 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:46:11 vm04 ceph-mon[53170]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-06T13:46:11.680 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:46:11 vm04 ceph-mon[53170]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-06T13:46:11.680 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:46:11 vm04 ceph-mon[53170]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-06T13:46:11.680 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:46:11 vm04 ceph-mon[53170]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:46:11.680 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:46:11 vm04 ceph-mon[53170]: Metadata not up to date on all hosts. Skipping non agent specs 2026-03-06T13:46:11.680 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:46:11 vm04 ceph-mon[53170]: pgmap v11: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-06T13:46:11.680 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:46:11 vm04 ceph-mon[53170]: Reconfiguring mgr.a (unknown last config time)... 2026-03-06T13:46:11.680 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:46:11 vm04 ceph-mon[53170]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' cmd=[{"prefix": "auth get-or-create", "entity": "mgr.a", "caps": ["mon", "profile mgr", "osd", "allow *", "mds", "allow *"]}]: dispatch 2026-03-06T13:46:11.680 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:46:11 vm04 ceph-mon[53170]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' cmd=[{"prefix": "mgr services"}]: dispatch 2026-03-06T13:46:11.680 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:46:11 vm04 ceph-mon[53170]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-06T13:46:11.680 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:46:11 vm04 ceph-mon[53170]: Reconfiguring daemon mgr.a on vm02 2026-03-06T13:46:11.680 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:46:11 vm04 ceph-mon[53170]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:46:11.680 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:46:11 vm04 ceph-mon[53170]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:46:11.680 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:46:11 vm04 ceph-mon[53170]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-06T13:46:11.680 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:46:11 vm04 ceph-mon[53170]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-06T13:46:11.680 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:46:11 vm04 ceph-mon[53170]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-06T13:46:11.680 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:46:11 vm04 ceph-mon[53170]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:46:11.702 DEBUG:teuthology.orchestra.run.vm06:> stat /dev/vde 2026-03-06T13:46:11.759 INFO:teuthology.orchestra.run.vm06.stdout: File: /dev/vde 2026-03-06T13:46:11.759 INFO:teuthology.orchestra.run.vm06.stdout: Size: 0 Blocks: 0 IO Block: 512 block special file 2026-03-06T13:46:11.759 INFO:teuthology.orchestra.run.vm06.stdout:Device: 6h/6d Inode: 257 Links: 1 Device type: fc,40 2026-03-06T13:46:11.759 INFO:teuthology.orchestra.run.vm06.stdout:Access: (0660/brw-rw----) Uid: ( 0/ root) Gid: ( 6/ disk) 2026-03-06T13:46:11.759 INFO:teuthology.orchestra.run.vm06.stdout:Context: system_u:object_r:fixed_disk_device_t:s0 2026-03-06T13:46:11.759 INFO:teuthology.orchestra.run.vm06.stdout:Access: 2026-03-06 13:46:04.374685587 +0100 2026-03-06T13:46:11.759 INFO:teuthology.orchestra.run.vm06.stdout:Modify: 2026-03-06 13:43:10.835782885 +0100 2026-03-06T13:46:11.759 INFO:teuthology.orchestra.run.vm06.stdout:Change: 2026-03-06 13:43:10.835782885 +0100 2026-03-06T13:46:11.759 INFO:teuthology.orchestra.run.vm06.stdout: Birth: 2026-03-06 13:39:27.253000000 +0100 2026-03-06T13:46:11.759 DEBUG:teuthology.orchestra.run.vm06:> sudo dd if=/dev/vde of=/dev/null count=1 2026-03-06T13:46:11.822 INFO:teuthology.orchestra.run.vm06.stderr:1+0 records in 2026-03-06T13:46:11.822 INFO:teuthology.orchestra.run.vm06.stderr:1+0 records out 2026-03-06T13:46:11.822 INFO:teuthology.orchestra.run.vm06.stderr:512 bytes copied, 0.000140543 s, 3.6 MB/s 2026-03-06T13:46:11.823 DEBUG:teuthology.orchestra.run.vm06:> ! mount | grep -v devtmpfs | grep -q /dev/vde 2026-03-06T13:46:11.862 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:46:11 vm02 ceph-mon[49207]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:46:11.862 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:46:11 vm02 ceph-mon[49207]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:46:11.862 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:46:11 vm02 ceph-mon[49207]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:46:11.862 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:46:11 vm02 ceph-mon[49207]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:46:11.862 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:46:11 vm02 ceph-mon[49207]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-06T13:46:11.862 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:46:11 vm02 ceph-mon[49207]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-06T13:46:11.862 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:46:11 vm02 ceph-mon[49207]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-06T13:46:11.862 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:46:11 vm02 ceph-mon[49207]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:46:11.862 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:46:11 vm02 ceph-mon[49207]: Metadata not up to date on all hosts. Skipping non agent specs 2026-03-06T13:46:11.862 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:46:11 vm02 ceph-mon[49207]: pgmap v11: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-06T13:46:11.862 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:46:11 vm02 ceph-mon[49207]: Reconfiguring mgr.a (unknown last config time)... 2026-03-06T13:46:11.862 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:46:11 vm02 ceph-mon[49207]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' cmd=[{"prefix": "auth get-or-create", "entity": "mgr.a", "caps": ["mon", "profile mgr", "osd", "allow *", "mds", "allow *"]}]: dispatch 2026-03-06T13:46:11.862 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:46:11 vm02 ceph-mon[49207]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' cmd=[{"prefix": "mgr services"}]: dispatch 2026-03-06T13:46:11.862 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:46:11 vm02 ceph-mon[49207]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-06T13:46:11.862 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:46:11 vm02 ceph-mon[49207]: Reconfiguring daemon mgr.a on vm02 2026-03-06T13:46:11.862 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:46:11 vm02 ceph-mon[49207]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:46:11.862 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:46:11 vm02 ceph-mon[49207]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:46:11.862 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:46:11 vm02 ceph-mon[49207]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-06T13:46:11.862 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:46:11 vm02 ceph-mon[49207]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-06T13:46:11.862 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:46:11 vm02 ceph-mon[49207]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-06T13:46:11.862 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:46:11 vm02 ceph-mon[49207]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:46:11.881 INFO:tasks.cephadm:Deploying osd.0 on vm02 with /dev/vde... 2026-03-06T13:46:11.881 DEBUG:teuthology.orchestra.run.vm02:> sudo /home/ubuntu/cephtest/cephadm --image harbor.clyso.com/custom-ceph/ceph/ceph:cobaltcore-storage-v19.2.3-fasttrack-3 ceph-volume -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 2b1b4a72-195a-11f1-b1df-bf1f800d676b -- lvm zap /dev/vde 2026-03-06T13:46:12.205 INFO:teuthology.orchestra.run.vm02.stderr:Inferring config /var/lib/ceph/2b1b4a72-195a-11f1-b1df-bf1f800d676b/mon.a/config 2026-03-06T13:46:12.550 INFO:journalctl@ceph.mgr.b.vm04.stdout:Mar 06 13:46:12 vm04 ceph-2b1b4a72-195a-11f1-b1df-bf1f800d676b-mgr-b[54791]: 2026-03-06T12:46:12.260+0000 7ff9bb54a100 -1 mgr[py] Module crash has missing NOTIFY_TYPES member 2026-03-06T13:46:12.849 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:46:12 vm04 ceph-mon[53170]: Metadata not up to date on all hosts. Skipping non agent specs 2026-03-06T13:46:12.849 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:46:12 vm04 ceph-mon[53170]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:46:12.862 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:46:12 vm02 ceph-mon[49207]: Metadata not up to date on all hosts. Skipping non agent specs 2026-03-06T13:46:12.862 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:46:12 vm02 ceph-mon[49207]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:46:13.010 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:46:12 vm06 ceph-mon[56528]: Metadata not up to date on all hosts. Skipping non agent specs 2026-03-06T13:46:13.011 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:46:12 vm06 ceph-mon[56528]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:46:13.494 INFO:journalctl@ceph.mgr.b.vm04.stdout:Mar 06 13:46:13 vm04 ceph-2b1b4a72-195a-11f1-b1df-bf1f800d676b-mgr-b[54791]: 2026-03-06T12:46:13.358+0000 7ff9bb54a100 -1 mgr[py] Module devicehealth has missing NOTIFY_TYPES member 2026-03-06T13:46:13.563 INFO:teuthology.orchestra.run.vm02.stdout: 2026-03-06T13:46:13.583 DEBUG:teuthology.orchestra.run.vm02:> sudo /home/ubuntu/cephtest/cephadm --image harbor.clyso.com/custom-ceph/ceph/ceph:cobaltcore-storage-v19.2.3-fasttrack-3 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 2b1b4a72-195a-11f1-b1df-bf1f800d676b -- ceph orch daemon add osd vm02:/dev/vde 2026-03-06T13:46:13.759 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:46:13 vm04 ceph-mon[53170]: pgmap v12: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-06T13:46:13.759 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:46:13 vm04 ceph-mon[53170]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:46:13.759 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:46:13 vm04 ceph-mon[53170]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:46:13.759 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:46:13 vm04 ceph-mon[53170]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-06T13:46:13.760 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:46:13 vm04 ceph-mon[53170]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:46:13.760 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:46:13 vm04 ceph-mon[53170]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-06T13:46:13.760 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:46:13 vm04 ceph-mon[53170]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-06T13:46:13.760 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:46:13 vm04 ceph-mon[53170]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:46:13.760 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:46:13 vm04 ceph-mon[53170]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:46:13.760 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:46:13 vm04 ceph-mon[53170]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:46:13.760 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:46:13 vm04 ceph-mon[53170]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:46:13.760 INFO:journalctl@ceph.mgr.b.vm04.stdout:Mar 06 13:46:13 vm04 ceph-2b1b4a72-195a-11f1-b1df-bf1f800d676b-mgr-b[54791]: 2026-03-06T12:46:13.491+0000 7ff9bb54a100 -1 mgr[py] Module influx has missing NOTIFY_TYPES member 2026-03-06T13:46:13.862 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:46:13 vm02 ceph-mon[49207]: pgmap v12: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-06T13:46:13.862 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:46:13 vm02 ceph-mon[49207]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:46:13.862 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:46:13 vm02 ceph-mon[49207]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:46:13.862 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:46:13 vm02 ceph-mon[49207]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-06T13:46:13.862 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:46:13 vm02 ceph-mon[49207]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:46:13.862 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:46:13 vm02 ceph-mon[49207]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-06T13:46:13.862 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:46:13 vm02 ceph-mon[49207]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-06T13:46:13.862 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:46:13 vm02 ceph-mon[49207]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:46:13.862 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:46:13 vm02 ceph-mon[49207]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:46:13.862 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:46:13 vm02 ceph-mon[49207]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:46:13.862 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:46:13 vm02 ceph-mon[49207]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:46:13.928 INFO:teuthology.orchestra.run.vm02.stderr:Inferring config /var/lib/ceph/2b1b4a72-195a-11f1-b1df-bf1f800d676b/mon.a/config 2026-03-06T13:46:14.010 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:46:13 vm06 ceph-mon[56528]: pgmap v12: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-06T13:46:14.010 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:46:13 vm06 ceph-mon[56528]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:46:14.010 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:46:13 vm06 ceph-mon[56528]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:46:14.010 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:46:13 vm06 ceph-mon[56528]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-06T13:46:14.011 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:46:13 vm06 ceph-mon[56528]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:46:14.011 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:46:13 vm06 ceph-mon[56528]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-06T13:46:14.011 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:46:13 vm06 ceph-mon[56528]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-06T13:46:14.011 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:46:13 vm06 ceph-mon[56528]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:46:14.011 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:46:13 vm06 ceph-mon[56528]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:46:14.011 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:46:13 vm06 ceph-mon[56528]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:46:14.011 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:46:13 vm06 ceph-mon[56528]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:46:14.049 INFO:journalctl@ceph.mgr.b.vm04.stdout:Mar 06 13:46:13 vm04 ceph-2b1b4a72-195a-11f1-b1df-bf1f800d676b-mgr-b[54791]: 2026-03-06T12:46:13.756+0000 7ff9bb54a100 -1 mgr[py] Module iostat has missing NOTIFY_TYPES member 2026-03-06T13:46:14.921 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:46:14 vm02 ceph-mon[49207]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' cmd=[{"prefix": "osd tree", "states": ["destroyed"], "format": "json"}]: dispatch 2026-03-06T13:46:14.921 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:46:14 vm02 ceph-mon[49207]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.bootstrap-osd"}]: dispatch 2026-03-06T13:46:14.921 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:46:14 vm02 ceph-mon[49207]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-06T13:46:15.010 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:46:14 vm06 ceph-mon[56528]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' cmd=[{"prefix": "osd tree", "states": ["destroyed"], "format": "json"}]: dispatch 2026-03-06T13:46:15.010 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:46:14 vm06 ceph-mon[56528]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.bootstrap-osd"}]: dispatch 2026-03-06T13:46:15.010 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:46:14 vm06 ceph-mon[56528]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-06T13:46:15.049 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:46:14 vm04 ceph-mon[53170]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' cmd=[{"prefix": "osd tree", "states": ["destroyed"], "format": "json"}]: dispatch 2026-03-06T13:46:15.049 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:46:14 vm04 ceph-mon[53170]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.bootstrap-osd"}]: dispatch 2026-03-06T13:46:15.049 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:46:14 vm04 ceph-mon[53170]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-06T13:46:15.737 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:46:15 vm02 ceph-mon[49207]: from='client.14223 -' entity='client.admin' cmd=[{"prefix": "orch daemon add osd", "svc_arg": "vm02:/dev/vde", "target": ["mon-mgr", ""]}]: dispatch 2026-03-06T13:46:15.738 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:46:15 vm02 ceph-mon[49207]: pgmap v13: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-06T13:46:15.738 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:46:15 vm02 ceph-mon[49207]: from='client.? 192.168.123.102:0/1383120545' entity='client.bootstrap-osd' cmd=[{"prefix": "osd new", "uuid": "9b067403-3d7b-4d89-a7e1-1cd83ba7845f"}]: dispatch 2026-03-06T13:46:15.738 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:46:15 vm02 ceph-mon[49207]: from='client.? 192.168.123.102:0/1383120545' entity='client.bootstrap-osd' cmd='[{"prefix": "osd new", "uuid": "9b067403-3d7b-4d89-a7e1-1cd83ba7845f"}]': finished 2026-03-06T13:46:15.738 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:46:15 vm02 ceph-mon[49207]: osdmap e5: 1 total, 0 up, 1 in 2026-03-06T13:46:15.738 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:46:15 vm02 ceph-mon[49207]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 0}]: dispatch 2026-03-06T13:46:16.010 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:46:15 vm06 ceph-mon[56528]: from='client.14223 -' entity='client.admin' cmd=[{"prefix": "orch daemon add osd", "svc_arg": "vm02:/dev/vde", "target": ["mon-mgr", ""]}]: dispatch 2026-03-06T13:46:16.010 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:46:15 vm06 ceph-mon[56528]: pgmap v13: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-06T13:46:16.010 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:46:15 vm06 ceph-mon[56528]: from='client.? 192.168.123.102:0/1383120545' entity='client.bootstrap-osd' cmd=[{"prefix": "osd new", "uuid": "9b067403-3d7b-4d89-a7e1-1cd83ba7845f"}]: dispatch 2026-03-06T13:46:16.010 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:46:15 vm06 ceph-mon[56528]: from='client.? 192.168.123.102:0/1383120545' entity='client.bootstrap-osd' cmd='[{"prefix": "osd new", "uuid": "9b067403-3d7b-4d89-a7e1-1cd83ba7845f"}]': finished 2026-03-06T13:46:16.010 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:46:15 vm06 ceph-mon[56528]: osdmap e5: 1 total, 0 up, 1 in 2026-03-06T13:46:16.010 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:46:15 vm06 ceph-mon[56528]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 0}]: dispatch 2026-03-06T13:46:16.049 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:46:15 vm04 ceph-mon[53170]: from='client.14223 -' entity='client.admin' cmd=[{"prefix": "orch daemon add osd", "svc_arg": "vm02:/dev/vde", "target": ["mon-mgr", ""]}]: dispatch 2026-03-06T13:46:16.050 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:46:15 vm04 ceph-mon[53170]: pgmap v13: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-06T13:46:16.050 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:46:15 vm04 ceph-mon[53170]: from='client.? 192.168.123.102:0/1383120545' entity='client.bootstrap-osd' cmd=[{"prefix": "osd new", "uuid": "9b067403-3d7b-4d89-a7e1-1cd83ba7845f"}]: dispatch 2026-03-06T13:46:16.050 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:46:15 vm04 ceph-mon[53170]: from='client.? 192.168.123.102:0/1383120545' entity='client.bootstrap-osd' cmd='[{"prefix": "osd new", "uuid": "9b067403-3d7b-4d89-a7e1-1cd83ba7845f"}]': finished 2026-03-06T13:46:16.050 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:46:15 vm04 ceph-mon[53170]: osdmap e5: 1 total, 0 up, 1 in 2026-03-06T13:46:16.050 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:46:15 vm04 ceph-mon[53170]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 0}]: dispatch 2026-03-06T13:46:16.050 INFO:journalctl@ceph.mgr.b.vm04.stdout:Mar 06 13:46:15 vm04 ceph-2b1b4a72-195a-11f1-b1df-bf1f800d676b-mgr-b[54791]: 2026-03-06T12:46:15.707+0000 7ff9bb54a100 -1 mgr[py] Module nfs has missing NOTIFY_TYPES member 2026-03-06T13:46:16.363 INFO:journalctl@ceph.mgr.b.vm04.stdout:Mar 06 13:46:16 vm04 ceph-2b1b4a72-195a-11f1-b1df-bf1f800d676b-mgr-b[54791]: 2026-03-06T12:46:16.083+0000 7ff9bb54a100 -1 mgr[py] Module orchestrator has missing NOTIFY_TYPES member 2026-03-06T13:46:16.363 INFO:journalctl@ceph.mgr.b.vm04.stdout:Mar 06 13:46:16 vm04 ceph-2b1b4a72-195a-11f1-b1df-bf1f800d676b-mgr-b[54791]: 2026-03-06T12:46:16.229+0000 7ff9bb54a100 -1 mgr[py] Module osd_perf_query has missing NOTIFY_TYPES member 2026-03-06T13:46:16.363 INFO:journalctl@ceph.mgr.b.vm04.stdout:Mar 06 13:46:16 vm04 ceph-2b1b4a72-195a-11f1-b1df-bf1f800d676b-mgr-b[54791]: 2026-03-06T12:46:16.360+0000 7ff9bb54a100 -1 mgr[py] Module osd_support has missing NOTIFY_TYPES member 2026-03-06T13:46:16.642 INFO:journalctl@ceph.mgr.b.vm04.stdout:Mar 06 13:46:16 vm04 ceph-2b1b4a72-195a-11f1-b1df-bf1f800d676b-mgr-b[54791]: 2026-03-06T12:46:16.513+0000 7ff9bb54a100 -1 mgr[py] Module pg_autoscaler has missing NOTIFY_TYPES member 2026-03-06T13:46:17.010 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:46:16 vm06 ceph-mon[56528]: from='client.? 192.168.123.102:0/3306834046' entity='client.bootstrap-osd' cmd=[{"prefix": "mon getmap"}]: dispatch 2026-03-06T13:46:17.049 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:46:16 vm04 ceph-mon[53170]: from='client.? 192.168.123.102:0/3306834046' entity='client.bootstrap-osd' cmd=[{"prefix": "mon getmap"}]: dispatch 2026-03-06T13:46:17.050 INFO:journalctl@ceph.mgr.b.vm04.stdout:Mar 06 13:46:16 vm04 ceph-2b1b4a72-195a-11f1-b1df-bf1f800d676b-mgr-b[54791]: 2026-03-06T12:46:16.639+0000 7ff9bb54a100 -1 mgr[py] Module progress has missing NOTIFY_TYPES member 2026-03-06T13:46:17.112 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:46:16 vm02 ceph-mon[49207]: from='client.? 192.168.123.102:0/3306834046' entity='client.bootstrap-osd' cmd=[{"prefix": "mon getmap"}]: dispatch 2026-03-06T13:46:17.550 INFO:journalctl@ceph.mgr.b.vm04.stdout:Mar 06 13:46:17 vm04 ceph-2b1b4a72-195a-11f1-b1df-bf1f800d676b-mgr-b[54791]: 2026-03-06T12:46:17.175+0000 7ff9bb54a100 -1 mgr[py] Module prometheus has missing NOTIFY_TYPES member 2026-03-06T13:46:17.550 INFO:journalctl@ceph.mgr.b.vm04.stdout:Mar 06 13:46:17 vm04 ceph-2b1b4a72-195a-11f1-b1df-bf1f800d676b-mgr-b[54791]: 2026-03-06T12:46:17.353+0000 7ff9bb54a100 -1 mgr[py] Module rbd_support has missing NOTIFY_TYPES member 2026-03-06T13:46:18.010 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:46:17 vm06 ceph-mon[56528]: pgmap v15: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-06T13:46:18.049 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:46:17 vm04 ceph-mon[53170]: pgmap v15: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-06T13:46:18.112 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:46:17 vm02 ceph-mon[49207]: pgmap v15: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-06T13:46:18.550 INFO:journalctl@ceph.mgr.b.vm04.stdout:Mar 06 13:46:18 vm04 ceph-2b1b4a72-195a-11f1-b1df-bf1f800d676b-mgr-b[54791]: 2026-03-06T12:46:18.146+0000 7ff9bb54a100 -1 mgr[py] Module rgw has missing NOTIFY_TYPES member 2026-03-06T13:46:19.550 INFO:journalctl@ceph.mgr.b.vm04.stdout:Mar 06 13:46:19 vm04 ceph-2b1b4a72-195a-11f1-b1df-bf1f800d676b-mgr-b[54791]: 2026-03-06T12:46:19.276+0000 7ff9bb54a100 -1 mgr[py] Module rook has missing NOTIFY_TYPES member 2026-03-06T13:46:19.550 INFO:journalctl@ceph.mgr.b.vm04.stdout:Mar 06 13:46:19 vm04 ceph-2b1b4a72-195a-11f1-b1df-bf1f800d676b-mgr-b[54791]: 2026-03-06T12:46:19.414+0000 7ff9bb54a100 -1 mgr[py] Module selftest has missing NOTIFY_TYPES member 2026-03-06T13:46:19.841 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:46:19 vm04 ceph-mon[53170]: pgmap v16: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-06T13:46:19.841 INFO:journalctl@ceph.mgr.b.vm04.stdout:Mar 06 13:46:19 vm04 ceph-2b1b4a72-195a-11f1-b1df-bf1f800d676b-mgr-b[54791]: 2026-03-06T12:46:19.556+0000 7ff9bb54a100 -1 mgr[py] Module snap_schedule has missing NOTIFY_TYPES member 2026-03-06T13:46:20.010 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:46:19 vm06 ceph-mon[56528]: pgmap v16: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-06T13:46:20.112 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:46:19 vm02 ceph-mon[49207]: pgmap v16: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-06T13:46:20.300 INFO:journalctl@ceph.mgr.b.vm04.stdout:Mar 06 13:46:19 vm04 ceph-2b1b4a72-195a-11f1-b1df-bf1f800d676b-mgr-b[54791]: 2026-03-06T12:46:19.838+0000 7ff9bb54a100 -1 mgr[py] Module status has missing NOTIFY_TYPES member 2026-03-06T13:46:20.300 INFO:journalctl@ceph.mgr.b.vm04.stdout:Mar 06 13:46:19 vm04 ceph-2b1b4a72-195a-11f1-b1df-bf1f800d676b-mgr-b[54791]: 2026-03-06T12:46:19.973+0000 7ff9bb54a100 -1 mgr[py] Module telegraf has missing NOTIFY_TYPES member 2026-03-06T13:46:20.670 INFO:journalctl@ceph.mgr.b.vm04.stdout:Mar 06 13:46:20 vm04 ceph-2b1b4a72-195a-11f1-b1df-bf1f800d676b-mgr-b[54791]: 2026-03-06T12:46:20.302+0000 7ff9bb54a100 -1 mgr[py] Module telemetry has missing NOTIFY_TYPES member 2026-03-06T13:46:21.010 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:46:20 vm06 ceph-mon[56528]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "osd.0"}]: dispatch 2026-03-06T13:46:21.010 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:46:20 vm06 ceph-mon[56528]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-06T13:46:21.021 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:46:20 vm02 ceph-mon[49207]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "osd.0"}]: dispatch 2026-03-06T13:46:21.021 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:46:20 vm02 ceph-mon[49207]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-06T13:46:21.050 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:46:20 vm04 ceph-mon[53170]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "osd.0"}]: dispatch 2026-03-06T13:46:21.050 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:46:20 vm04 ceph-mon[53170]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-06T13:46:21.050 INFO:journalctl@ceph.mgr.b.vm04.stdout:Mar 06 13:46:20 vm04 ceph-2b1b4a72-195a-11f1-b1df-bf1f800d676b-mgr-b[54791]: 2026-03-06T12:46:20.667+0000 7ff9bb54a100 -1 mgr[py] Module test_orchestrator has missing NOTIFY_TYPES member 2026-03-06T13:46:21.550 INFO:journalctl@ceph.mgr.b.vm04.stdout:Mar 06 13:46:21 vm04 ceph-2b1b4a72-195a-11f1-b1df-bf1f800d676b-mgr-b[54791]: 2026-03-06T12:46:21.056+0000 7ff9bb54a100 -1 mgr[py] Module volumes has missing NOTIFY_TYPES member 2026-03-06T13:46:21.550 INFO:journalctl@ceph.mgr.b.vm04.stdout:Mar 06 13:46:21 vm04 ceph-2b1b4a72-195a-11f1-b1df-bf1f800d676b-mgr-b[54791]: 2026-03-06T12:46:21.185+0000 7ff9bb54a100 -1 mgr[py] Module zabbix has missing NOTIFY_TYPES member 2026-03-06T13:46:21.715 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:46:21 vm02 ceph-mon[49207]: Deploying daemon osd.0 on vm02 2026-03-06T13:46:21.715 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:46:21 vm02 ceph-mon[49207]: pgmap v17: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-06T13:46:21.991 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:46:21 vm02 ceph-mon[49207]: Standby manager daemon b started 2026-03-06T13:46:21.991 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:46:21 vm02 ceph-mon[49207]: from='mgr.? 192.168.123.104:0/1088279745' entity='mgr.b' cmd=[{"prefix": "config-key get", "key": "mgr/dashboard/b/crt"}]: dispatch 2026-03-06T13:46:21.991 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:46:21 vm02 ceph-mon[49207]: from='mgr.? 192.168.123.104:0/1088279745' entity='mgr.b' cmd=[{"prefix": "config-key get", "key": "mgr/dashboard/crt"}]: dispatch 2026-03-06T13:46:21.991 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:46:21 vm02 ceph-mon[49207]: from='mgr.? 192.168.123.104:0/1088279745' entity='mgr.b' cmd=[{"prefix": "config-key get", "key": "mgr/dashboard/b/key"}]: dispatch 2026-03-06T13:46:21.991 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:46:21 vm02 ceph-mon[49207]: from='mgr.? 192.168.123.104:0/1088279745' entity='mgr.b' cmd=[{"prefix": "config-key get", "key": "mgr/dashboard/key"}]: dispatch 2026-03-06T13:46:22.010 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:46:21 vm06 ceph-mon[56528]: Deploying daemon osd.0 on vm02 2026-03-06T13:46:22.010 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:46:21 vm06 ceph-mon[56528]: pgmap v17: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-06T13:46:22.010 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:46:21 vm06 ceph-mon[56528]: Standby manager daemon b started 2026-03-06T13:46:22.010 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:46:21 vm06 ceph-mon[56528]: from='mgr.? 192.168.123.104:0/1088279745' entity='mgr.b' cmd=[{"prefix": "config-key get", "key": "mgr/dashboard/b/crt"}]: dispatch 2026-03-06T13:46:22.010 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:46:21 vm06 ceph-mon[56528]: from='mgr.? 192.168.123.104:0/1088279745' entity='mgr.b' cmd=[{"prefix": "config-key get", "key": "mgr/dashboard/crt"}]: dispatch 2026-03-06T13:46:22.010 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:46:21 vm06 ceph-mon[56528]: from='mgr.? 192.168.123.104:0/1088279745' entity='mgr.b' cmd=[{"prefix": "config-key get", "key": "mgr/dashboard/b/key"}]: dispatch 2026-03-06T13:46:22.010 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:46:21 vm06 ceph-mon[56528]: from='mgr.? 192.168.123.104:0/1088279745' entity='mgr.b' cmd=[{"prefix": "config-key get", "key": "mgr/dashboard/key"}]: dispatch 2026-03-06T13:46:22.050 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:46:21 vm04 ceph-mon[53170]: Deploying daemon osd.0 on vm02 2026-03-06T13:46:22.050 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:46:21 vm04 ceph-mon[53170]: pgmap v17: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-06T13:46:22.050 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:46:21 vm04 ceph-mon[53170]: Standby manager daemon b started 2026-03-06T13:46:22.050 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:46:21 vm04 ceph-mon[53170]: from='mgr.? 192.168.123.104:0/1088279745' entity='mgr.b' cmd=[{"prefix": "config-key get", "key": "mgr/dashboard/b/crt"}]: dispatch 2026-03-06T13:46:22.050 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:46:21 vm04 ceph-mon[53170]: from='mgr.? 192.168.123.104:0/1088279745' entity='mgr.b' cmd=[{"prefix": "config-key get", "key": "mgr/dashboard/crt"}]: dispatch 2026-03-06T13:46:22.050 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:46:21 vm04 ceph-mon[53170]: from='mgr.? 192.168.123.104:0/1088279745' entity='mgr.b' cmd=[{"prefix": "config-key get", "key": "mgr/dashboard/b/key"}]: dispatch 2026-03-06T13:46:22.050 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:46:21 vm04 ceph-mon[53170]: from='mgr.? 192.168.123.104:0/1088279745' entity='mgr.b' cmd=[{"prefix": "config-key get", "key": "mgr/dashboard/key"}]: dispatch 2026-03-06T13:46:22.837 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:46:22 vm06 ceph-mon[56528]: mgrmap e14: a(active, since 45s), standbys: b 2026-03-06T13:46:22.837 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:46:22 vm06 ceph-mon[56528]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' cmd=[{"prefix": "mgr metadata", "who": "b", "id": "b"}]: dispatch 2026-03-06T13:46:22.837 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:46:22 vm06 ceph-mon[56528]: pgmap v18: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-06T13:46:23.007 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:46:22 vm02 ceph-mon[49207]: mgrmap e14: a(active, since 45s), standbys: b 2026-03-06T13:46:23.007 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:46:22 vm02 ceph-mon[49207]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' cmd=[{"prefix": "mgr metadata", "who": "b", "id": "b"}]: dispatch 2026-03-06T13:46:23.007 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:46:22 vm02 ceph-mon[49207]: pgmap v18: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-06T13:46:23.050 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:46:22 vm04 ceph-mon[53170]: mgrmap e14: a(active, since 45s), standbys: b 2026-03-06T13:46:23.050 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:46:22 vm04 ceph-mon[53170]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' cmd=[{"prefix": "mgr metadata", "who": "b", "id": "b"}]: dispatch 2026-03-06T13:46:23.050 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:46:22 vm04 ceph-mon[53170]: pgmap v18: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-06T13:46:24.561 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:46:24 vm02 ceph-mon[49207]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:46:24.562 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:46:24 vm02 ceph-mon[49207]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:46:24.562 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:46:24 vm02 ceph-mon[49207]: pgmap v19: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-06T13:46:24.562 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:46:24 vm02 ceph-mon[49207]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-06T13:46:24.562 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:46:24 vm02 ceph-mon[49207]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-06T13:46:24.562 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:46:24 vm02 ceph-mon[49207]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:46:24.562 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:46:24 vm02 ceph-mon[49207]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-06T13:46:24.562 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:46:24 vm02 ceph-mon[49207]: pgmap v20: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-06T13:46:24.562 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:46:24 vm02 ceph-mon[49207]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:46:24.562 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:46:24 vm02 ceph-mon[49207]: Reconfiguring osd.0 (unknown last config time)... 2026-03-06T13:46:24.562 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:46:24 vm02 ceph-mon[49207]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "osd.0"}]: dispatch 2026-03-06T13:46:24.562 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:46:24 vm02 ceph-mon[49207]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-06T13:46:24.562 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:46:24 vm02 ceph-mon[49207]: Reconfiguring daemon osd.0 on vm02 2026-03-06T13:46:24.562 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:46:24 vm02 ceph-mon[49207]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:46:24.562 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:46:24 vm02 ceph-mon[49207]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:46:24.562 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:46:24 vm02 ceph-mon[49207]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:46:24.562 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:46:24 vm02 ceph-mon[49207]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:46:24.562 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:46:24 vm02 ceph-mon[49207]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:46:24.760 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:46:24 vm06 ceph-mon[56528]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:46:24.761 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:46:24 vm06 ceph-mon[56528]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:46:24.761 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:46:24 vm06 ceph-mon[56528]: pgmap v19: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-06T13:46:24.761 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:46:24 vm06 ceph-mon[56528]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-06T13:46:24.761 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:46:24 vm06 ceph-mon[56528]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-06T13:46:24.761 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:46:24 vm06 ceph-mon[56528]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:46:24.761 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:46:24 vm06 ceph-mon[56528]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-06T13:46:24.761 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:46:24 vm06 ceph-mon[56528]: pgmap v20: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-06T13:46:24.761 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:46:24 vm06 ceph-mon[56528]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:46:24.761 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:46:24 vm06 ceph-mon[56528]: Reconfiguring osd.0 (unknown last config time)... 2026-03-06T13:46:24.761 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:46:24 vm06 ceph-mon[56528]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "osd.0"}]: dispatch 2026-03-06T13:46:24.761 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:46:24 vm06 ceph-mon[56528]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-06T13:46:24.761 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:46:24 vm06 ceph-mon[56528]: Reconfiguring daemon osd.0 on vm02 2026-03-06T13:46:24.761 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:46:24 vm06 ceph-mon[56528]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:46:24.761 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:46:24 vm06 ceph-mon[56528]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:46:24.761 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:46:24 vm06 ceph-mon[56528]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:46:24.761 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:46:24 vm06 ceph-mon[56528]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:46:24.761 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:46:24 vm06 ceph-mon[56528]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:46:24.800 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:46:24 vm04 ceph-mon[53170]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:46:24.800 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:46:24 vm04 ceph-mon[53170]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:46:24.800 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:46:24 vm04 ceph-mon[53170]: pgmap v19: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-06T13:46:24.800 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:46:24 vm04 ceph-mon[53170]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-06T13:46:24.800 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:46:24 vm04 ceph-mon[53170]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-06T13:46:24.800 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:46:24 vm04 ceph-mon[53170]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:46:24.800 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:46:24 vm04 ceph-mon[53170]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-06T13:46:24.800 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:46:24 vm04 ceph-mon[53170]: pgmap v20: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-06T13:46:24.800 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:46:24 vm04 ceph-mon[53170]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:46:24.800 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:46:24 vm04 ceph-mon[53170]: Reconfiguring osd.0 (unknown last config time)... 2026-03-06T13:46:24.800 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:46:24 vm04 ceph-mon[53170]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "osd.0"}]: dispatch 2026-03-06T13:46:24.800 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:46:24 vm04 ceph-mon[53170]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-06T13:46:24.800 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:46:24 vm04 ceph-mon[53170]: Reconfiguring daemon osd.0 on vm02 2026-03-06T13:46:24.800 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:46:24 vm04 ceph-mon[53170]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:46:24.800 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:46:24 vm04 ceph-mon[53170]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:46:24.800 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:46:24 vm04 ceph-mon[53170]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:46:24.800 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:46:24 vm04 ceph-mon[53170]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:46:24.800 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:46:24 vm04 ceph-mon[53170]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:46:25.612 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:46:25 vm02 ceph-mon[49207]: pgmap v21: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-06T13:46:25.612 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:46:25 vm02 ceph-mon[49207]: Health check failed: 1 failed cephadm daemon(s) (CEPHADM_FAILED_DAEMON) 2026-03-06T13:46:25.612 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:46:25 vm02 ceph-mon[49207]: from='osd.0 [v2:192.168.123.102:6802/3358520121,v1:192.168.123.102:6803/3358520121]' entity='osd.0' cmd=[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["0"]}]: dispatch 2026-03-06T13:46:25.612 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:46:25 vm02 ceph-mon[49207]: from='osd.0 ' entity='osd.0' cmd=[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["0"]}]: dispatch 2026-03-06T13:46:25.761 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:46:25 vm06 ceph-mon[56528]: pgmap v21: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-06T13:46:25.761 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:46:25 vm06 ceph-mon[56528]: Health check failed: 1 failed cephadm daemon(s) (CEPHADM_FAILED_DAEMON) 2026-03-06T13:46:25.761 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:46:25 vm06 ceph-mon[56528]: from='osd.0 [v2:192.168.123.102:6802/3358520121,v1:192.168.123.102:6803/3358520121]' entity='osd.0' cmd=[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["0"]}]: dispatch 2026-03-06T13:46:25.761 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:46:25 vm06 ceph-mon[56528]: from='osd.0 ' entity='osd.0' cmd=[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["0"]}]: dispatch 2026-03-06T13:46:25.799 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:46:25 vm04 ceph-mon[53170]: pgmap v21: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-06T13:46:25.800 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:46:25 vm04 ceph-mon[53170]: Health check failed: 1 failed cephadm daemon(s) (CEPHADM_FAILED_DAEMON) 2026-03-06T13:46:25.800 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:46:25 vm04 ceph-mon[53170]: from='osd.0 [v2:192.168.123.102:6802/3358520121,v1:192.168.123.102:6803/3358520121]' entity='osd.0' cmd=[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["0"]}]: dispatch 2026-03-06T13:46:25.800 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:46:25 vm04 ceph-mon[53170]: from='osd.0 ' entity='osd.0' cmd=[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["0"]}]: dispatch 2026-03-06T13:46:26.368 INFO:teuthology.orchestra.run.vm02.stdout:Created osd(s) 0 on host 'vm02' 2026-03-06T13:46:26.546 DEBUG:teuthology.orchestra.run.vm02:osd.0> sudo journalctl -f -n 0 -u ceph-2b1b4a72-195a-11f1-b1df-bf1f800d676b@osd.0.service 2026-03-06T13:46:26.548 INFO:tasks.cephadm:Deploying osd.1 on vm04 with /dev/vde... 2026-03-06T13:46:26.548 DEBUG:teuthology.orchestra.run.vm04:> sudo /home/ubuntu/cephtest/cephadm --image harbor.clyso.com/custom-ceph/ceph/ceph:cobaltcore-storage-v19.2.3-fasttrack-3 ceph-volume -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 2b1b4a72-195a-11f1-b1df-bf1f800d676b -- lvm zap /dev/vde 2026-03-06T13:46:26.760 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:46:26 vm06 ceph-mon[56528]: from='osd.0 ' entity='osd.0' cmd='[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["0"]}]': finished 2026-03-06T13:46:26.760 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:46:26 vm06 ceph-mon[56528]: osdmap e6: 1 total, 0 up, 1 in 2026-03-06T13:46:26.760 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:46:26 vm06 ceph-mon[56528]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 0}]: dispatch 2026-03-06T13:46:26.760 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:46:26 vm06 ceph-mon[56528]: from='osd.0 [v2:192.168.123.102:6802/3358520121,v1:192.168.123.102:6803/3358520121]' entity='osd.0' cmd=[{"prefix": "osd crush create-or-move", "id": 0, "weight":0.0195, "args": ["host=vm02", "root=default"]}]: dispatch 2026-03-06T13:46:26.760 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:46:26 vm06 ceph-mon[56528]: from='osd.0 ' entity='osd.0' cmd=[{"prefix": "osd crush create-or-move", "id": 0, "weight":0.0195, "args": ["host=vm02", "root=default"]}]: dispatch 2026-03-06T13:46:26.760 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:46:26 vm06 ceph-mon[56528]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:46:26.760 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:46:26 vm06 ceph-mon[56528]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:46:26.800 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:46:26 vm04 ceph-mon[53170]: from='osd.0 ' entity='osd.0' cmd='[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["0"]}]': finished 2026-03-06T13:46:26.800 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:46:26 vm04 ceph-mon[53170]: osdmap e6: 1 total, 0 up, 1 in 2026-03-06T13:46:26.800 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:46:26 vm04 ceph-mon[53170]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 0}]: dispatch 2026-03-06T13:46:26.800 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:46:26 vm04 ceph-mon[53170]: from='osd.0 [v2:192.168.123.102:6802/3358520121,v1:192.168.123.102:6803/3358520121]' entity='osd.0' cmd=[{"prefix": "osd crush create-or-move", "id": 0, "weight":0.0195, "args": ["host=vm02", "root=default"]}]: dispatch 2026-03-06T13:46:26.800 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:46:26 vm04 ceph-mon[53170]: from='osd.0 ' entity='osd.0' cmd=[{"prefix": "osd crush create-or-move", "id": 0, "weight":0.0195, "args": ["host=vm02", "root=default"]}]: dispatch 2026-03-06T13:46:26.800 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:46:26 vm04 ceph-mon[53170]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:46:26.800 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:46:26 vm04 ceph-mon[53170]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:46:26.862 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:46:26 vm02 ceph-mon[49207]: from='osd.0 ' entity='osd.0' cmd='[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["0"]}]': finished 2026-03-06T13:46:26.862 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:46:26 vm02 ceph-mon[49207]: osdmap e6: 1 total, 0 up, 1 in 2026-03-06T13:46:26.862 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:46:26 vm02 ceph-mon[49207]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 0}]: dispatch 2026-03-06T13:46:26.862 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:46:26 vm02 ceph-mon[49207]: from='osd.0 [v2:192.168.123.102:6802/3358520121,v1:192.168.123.102:6803/3358520121]' entity='osd.0' cmd=[{"prefix": "osd crush create-or-move", "id": 0, "weight":0.0195, "args": ["host=vm02", "root=default"]}]: dispatch 2026-03-06T13:46:26.862 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:46:26 vm02 ceph-mon[49207]: from='osd.0 ' entity='osd.0' cmd=[{"prefix": "osd crush create-or-move", "id": 0, "weight":0.0195, "args": ["host=vm02", "root=default"]}]: dispatch 2026-03-06T13:46:26.862 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:46:26 vm02 ceph-mon[49207]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:46:26.862 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:46:26 vm02 ceph-mon[49207]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:46:26.896 INFO:teuthology.orchestra.run.vm04.stderr:Inferring config /var/lib/ceph/2b1b4a72-195a-11f1-b1df-bf1f800d676b/mon.b/config 2026-03-06T13:46:27.512 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:46:27 vm04 ceph-mon[53170]: pgmap v23: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-06T13:46:27.512 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:46:27 vm04 ceph-mon[53170]: from='osd.0 ' entity='osd.0' cmd='[{"prefix": "osd crush create-or-move", "id": 0, "weight":0.0195, "args": ["host=vm02", "root=default"]}]': finished 2026-03-06T13:46:27.512 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:46:27 vm04 ceph-mon[53170]: osdmap e7: 1 total, 0 up, 1 in 2026-03-06T13:46:27.512 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:46:27 vm04 ceph-mon[53170]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 0}]: dispatch 2026-03-06T13:46:27.512 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:46:27 vm04 ceph-mon[53170]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 0}]: dispatch 2026-03-06T13:46:27.512 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:46:27 vm04 ceph-mon[53170]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:46:27.512 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:46:27 vm04 ceph-mon[53170]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:46:27.512 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:46:27 vm04 ceph-mon[53170]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-06T13:46:27.513 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:46:27 vm04 ceph-mon[53170]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' cmd=[{"prefix": "config rm", "who": "osd/host:vm02", "name": "osd_memory_target"}]: dispatch 2026-03-06T13:46:27.513 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:46:27 vm04 ceph-mon[53170]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-06T13:46:27.513 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:46:27 vm04 ceph-mon[53170]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-06T13:46:27.513 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:46:27 vm04 ceph-mon[53170]: pgmap v25: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-06T13:46:27.513 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:46:27 vm04 ceph-mon[53170]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:46:27.513 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:46:27 vm04 ceph-mon[53170]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 0}]: dispatch 2026-03-06T13:46:27.760 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:46:27 vm06 ceph-mon[56528]: pgmap v23: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-06T13:46:27.761 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:46:27 vm06 ceph-mon[56528]: from='osd.0 ' entity='osd.0' cmd='[{"prefix": "osd crush create-or-move", "id": 0, "weight":0.0195, "args": ["host=vm02", "root=default"]}]': finished 2026-03-06T13:46:27.761 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:46:27 vm06 ceph-mon[56528]: osdmap e7: 1 total, 0 up, 1 in 2026-03-06T13:46:27.761 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:46:27 vm06 ceph-mon[56528]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 0}]: dispatch 2026-03-06T13:46:27.761 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:46:27 vm06 ceph-mon[56528]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 0}]: dispatch 2026-03-06T13:46:27.761 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:46:27 vm06 ceph-mon[56528]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:46:27.761 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:46:27 vm06 ceph-mon[56528]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:46:27.761 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:46:27 vm06 ceph-mon[56528]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-06T13:46:27.761 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:46:27 vm06 ceph-mon[56528]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' cmd=[{"prefix": "config rm", "who": "osd/host:vm02", "name": "osd_memory_target"}]: dispatch 2026-03-06T13:46:27.761 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:46:27 vm06 ceph-mon[56528]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-06T13:46:27.761 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:46:27 vm06 ceph-mon[56528]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-06T13:46:27.761 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:46:27 vm06 ceph-mon[56528]: pgmap v25: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-06T13:46:27.761 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:46:27 vm06 ceph-mon[56528]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:46:27.761 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:46:27 vm06 ceph-mon[56528]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 0}]: dispatch 2026-03-06T13:46:27.862 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:46:27 vm02 ceph-mon[49207]: pgmap v23: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-06T13:46:27.862 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:46:27 vm02 ceph-mon[49207]: from='osd.0 ' entity='osd.0' cmd='[{"prefix": "osd crush create-or-move", "id": 0, "weight":0.0195, "args": ["host=vm02", "root=default"]}]': finished 2026-03-06T13:46:27.862 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:46:27 vm02 ceph-mon[49207]: osdmap e7: 1 total, 0 up, 1 in 2026-03-06T13:46:27.862 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:46:27 vm02 ceph-mon[49207]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 0}]: dispatch 2026-03-06T13:46:27.862 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:46:27 vm02 ceph-mon[49207]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 0}]: dispatch 2026-03-06T13:46:27.862 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:46:27 vm02 ceph-mon[49207]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:46:27.862 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:46:27 vm02 ceph-mon[49207]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:46:27.862 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:46:27 vm02 ceph-mon[49207]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-06T13:46:27.862 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:46:27 vm02 ceph-mon[49207]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' cmd=[{"prefix": "config rm", "who": "osd/host:vm02", "name": "osd_memory_target"}]: dispatch 2026-03-06T13:46:27.862 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:46:27 vm02 ceph-mon[49207]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-06T13:46:27.862 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:46:27 vm02 ceph-mon[49207]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-06T13:46:27.862 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:46:27 vm02 ceph-mon[49207]: pgmap v25: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-06T13:46:27.862 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:46:27 vm02 ceph-mon[49207]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:46:27.862 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:46:27 vm02 ceph-mon[49207]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 0}]: dispatch 2026-03-06T13:46:28.214 INFO:teuthology.orchestra.run.vm04.stdout: 2026-03-06T13:46:28.234 DEBUG:teuthology.orchestra.run.vm04:> sudo /home/ubuntu/cephtest/cephadm --image harbor.clyso.com/custom-ceph/ceph/ceph:cobaltcore-storage-v19.2.3-fasttrack-3 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 2b1b4a72-195a-11f1-b1df-bf1f800d676b -- ceph orch daemon add osd vm04:/dev/vde 2026-03-06T13:46:28.563 INFO:teuthology.orchestra.run.vm04.stderr:Inferring config /var/lib/ceph/2b1b4a72-195a-11f1-b1df-bf1f800d676b/mon.b/config 2026-03-06T13:46:28.800 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:46:28 vm04 ceph-mon[53170]: purged_snaps scrub starts 2026-03-06T13:46:28.800 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:46:28 vm04 ceph-mon[53170]: purged_snaps scrub ok 2026-03-06T13:46:28.800 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:46:28 vm04 ceph-mon[53170]: osd.0 [v2:192.168.123.102:6802/3358520121,v1:192.168.123.102:6803/3358520121] boot 2026-03-06T13:46:28.800 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:46:28 vm04 ceph-mon[53170]: osdmap e8: 1 total, 1 up, 1 in 2026-03-06T13:46:28.800 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:46:28 vm04 ceph-mon[53170]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 0}]: dispatch 2026-03-06T13:46:28.800 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:46:28 vm04 ceph-mon[53170]: Health check cleared: CEPHADM_FAILED_DAEMON (was: 1 failed cephadm daemon(s)) 2026-03-06T13:46:28.800 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:46:28 vm04 ceph-mon[53170]: Cluster is now healthy 2026-03-06T13:46:28.862 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:46:28 vm02 ceph-mon[49207]: purged_snaps scrub starts 2026-03-06T13:46:28.862 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:46:28 vm02 ceph-mon[49207]: purged_snaps scrub ok 2026-03-06T13:46:28.862 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:46:28 vm02 ceph-mon[49207]: osd.0 [v2:192.168.123.102:6802/3358520121,v1:192.168.123.102:6803/3358520121] boot 2026-03-06T13:46:28.862 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:46:28 vm02 ceph-mon[49207]: osdmap e8: 1 total, 1 up, 1 in 2026-03-06T13:46:28.862 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:46:28 vm02 ceph-mon[49207]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 0}]: dispatch 2026-03-06T13:46:28.862 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:46:28 vm02 ceph-mon[49207]: Health check cleared: CEPHADM_FAILED_DAEMON (was: 1 failed cephadm daemon(s)) 2026-03-06T13:46:28.862 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:46:28 vm02 ceph-mon[49207]: Cluster is now healthy 2026-03-06T13:46:29.010 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:46:28 vm06 ceph-mon[56528]: purged_snaps scrub starts 2026-03-06T13:46:29.010 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:46:28 vm06 ceph-mon[56528]: purged_snaps scrub ok 2026-03-06T13:46:29.010 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:46:28 vm06 ceph-mon[56528]: osd.0 [v2:192.168.123.102:6802/3358520121,v1:192.168.123.102:6803/3358520121] boot 2026-03-06T13:46:29.010 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:46:28 vm06 ceph-mon[56528]: osdmap e8: 1 total, 1 up, 1 in 2026-03-06T13:46:29.010 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:46:28 vm06 ceph-mon[56528]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 0}]: dispatch 2026-03-06T13:46:29.010 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:46:28 vm06 ceph-mon[56528]: Health check cleared: CEPHADM_FAILED_DAEMON (was: 1 failed cephadm daemon(s)) 2026-03-06T13:46:29.011 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:46:28 vm06 ceph-mon[56528]: Cluster is now healthy 2026-03-06T13:46:30.010 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:46:29 vm06 ceph-mon[56528]: pgmap v27: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-06T13:46:30.010 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:46:29 vm06 ceph-mon[56528]: osdmap e9: 1 total, 1 up, 1 in 2026-03-06T13:46:30.010 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:46:29 vm06 ceph-mon[56528]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' cmd=[{"prefix": "osd tree", "states": ["destroyed"], "format": "json"}]: dispatch 2026-03-06T13:46:30.010 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:46:29 vm06 ceph-mon[56528]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.bootstrap-osd"}]: dispatch 2026-03-06T13:46:30.010 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:46:29 vm06 ceph-mon[56528]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-06T13:46:30.050 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:46:29 vm04 ceph-mon[53170]: pgmap v27: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-06T13:46:30.050 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:46:29 vm04 ceph-mon[53170]: osdmap e9: 1 total, 1 up, 1 in 2026-03-06T13:46:30.050 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:46:29 vm04 ceph-mon[53170]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' cmd=[{"prefix": "osd tree", "states": ["destroyed"], "format": "json"}]: dispatch 2026-03-06T13:46:30.050 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:46:29 vm04 ceph-mon[53170]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.bootstrap-osd"}]: dispatch 2026-03-06T13:46:30.050 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:46:29 vm04 ceph-mon[53170]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-06T13:46:30.111 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:46:29 vm02 ceph-mon[49207]: pgmap v27: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-06T13:46:30.112 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:46:29 vm02 ceph-mon[49207]: osdmap e9: 1 total, 1 up, 1 in 2026-03-06T13:46:30.112 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:46:29 vm02 ceph-mon[49207]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' cmd=[{"prefix": "osd tree", "states": ["destroyed"], "format": "json"}]: dispatch 2026-03-06T13:46:30.112 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:46:29 vm02 ceph-mon[49207]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.bootstrap-osd"}]: dispatch 2026-03-06T13:46:30.112 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:46:29 vm02 ceph-mon[49207]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-06T13:46:31.010 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:46:30 vm06 ceph-mon[56528]: from='client.24131 -' entity='client.admin' cmd=[{"prefix": "orch daemon add osd", "svc_arg": "vm04:/dev/vde", "target": ["mon-mgr", ""]}]: dispatch 2026-03-06T13:46:31.010 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:46:30 vm06 ceph-mon[56528]: from='client.? 192.168.123.104:0/421070526' entity='client.bootstrap-osd' cmd=[{"prefix": "osd new", "uuid": "1b4f83a6-1a7a-4b65-95a3-d2d4a61115b6"}]: dispatch 2026-03-06T13:46:31.010 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:46:30 vm06 ceph-mon[56528]: from='client.? ' entity='client.bootstrap-osd' cmd=[{"prefix": "osd new", "uuid": "1b4f83a6-1a7a-4b65-95a3-d2d4a61115b6"}]: dispatch 2026-03-06T13:46:31.010 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:46:30 vm06 ceph-mon[56528]: from='client.? ' entity='client.bootstrap-osd' cmd='[{"prefix": "osd new", "uuid": "1b4f83a6-1a7a-4b65-95a3-d2d4a61115b6"}]': finished 2026-03-06T13:46:31.010 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:46:30 vm06 ceph-mon[56528]: osdmap e10: 2 total, 1 up, 2 in 2026-03-06T13:46:31.010 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:46:30 vm06 ceph-mon[56528]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 1}]: dispatch 2026-03-06T13:46:31.050 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:46:30 vm04 ceph-mon[53170]: from='client.24131 -' entity='client.admin' cmd=[{"prefix": "orch daemon add osd", "svc_arg": "vm04:/dev/vde", "target": ["mon-mgr", ""]}]: dispatch 2026-03-06T13:46:31.050 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:46:30 vm04 ceph-mon[53170]: from='client.? 192.168.123.104:0/421070526' entity='client.bootstrap-osd' cmd=[{"prefix": "osd new", "uuid": "1b4f83a6-1a7a-4b65-95a3-d2d4a61115b6"}]: dispatch 2026-03-06T13:46:31.050 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:46:30 vm04 ceph-mon[53170]: from='client.? ' entity='client.bootstrap-osd' cmd=[{"prefix": "osd new", "uuid": "1b4f83a6-1a7a-4b65-95a3-d2d4a61115b6"}]: dispatch 2026-03-06T13:46:31.050 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:46:30 vm04 ceph-mon[53170]: from='client.? ' entity='client.bootstrap-osd' cmd='[{"prefix": "osd new", "uuid": "1b4f83a6-1a7a-4b65-95a3-d2d4a61115b6"}]': finished 2026-03-06T13:46:31.050 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:46:30 vm04 ceph-mon[53170]: osdmap e10: 2 total, 1 up, 2 in 2026-03-06T13:46:31.050 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:46:30 vm04 ceph-mon[53170]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 1}]: dispatch 2026-03-06T13:46:31.112 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:46:30 vm02 ceph-mon[49207]: from='client.24131 -' entity='client.admin' cmd=[{"prefix": "orch daemon add osd", "svc_arg": "vm04:/dev/vde", "target": ["mon-mgr", ""]}]: dispatch 2026-03-06T13:46:31.112 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:46:30 vm02 ceph-mon[49207]: from='client.? 192.168.123.104:0/421070526' entity='client.bootstrap-osd' cmd=[{"prefix": "osd new", "uuid": "1b4f83a6-1a7a-4b65-95a3-d2d4a61115b6"}]: dispatch 2026-03-06T13:46:31.112 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:46:30 vm02 ceph-mon[49207]: from='client.? ' entity='client.bootstrap-osd' cmd=[{"prefix": "osd new", "uuid": "1b4f83a6-1a7a-4b65-95a3-d2d4a61115b6"}]: dispatch 2026-03-06T13:46:31.112 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:46:30 vm02 ceph-mon[49207]: from='client.? ' entity='client.bootstrap-osd' cmd='[{"prefix": "osd new", "uuid": "1b4f83a6-1a7a-4b65-95a3-d2d4a61115b6"}]': finished 2026-03-06T13:46:31.112 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:46:30 vm02 ceph-mon[49207]: osdmap e10: 2 total, 1 up, 2 in 2026-03-06T13:46:31.112 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:46:30 vm02 ceph-mon[49207]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 1}]: dispatch 2026-03-06T13:46:32.010 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:46:31 vm06 ceph-mon[56528]: pgmap v30: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-06T13:46:32.010 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:46:31 vm06 ceph-mon[56528]: from='client.? 192.168.123.104:0/2086023674' entity='client.bootstrap-osd' cmd=[{"prefix": "mon getmap"}]: dispatch 2026-03-06T13:46:32.047 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:46:31 vm04 ceph-mon[53170]: pgmap v30: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-06T13:46:32.047 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:46:31 vm04 ceph-mon[53170]: from='client.? 192.168.123.104:0/2086023674' entity='client.bootstrap-osd' cmd=[{"prefix": "mon getmap"}]: dispatch 2026-03-06T13:46:32.111 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:46:31 vm02 ceph-mon[49207]: pgmap v30: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-06T13:46:32.111 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:46:31 vm02 ceph-mon[49207]: from='client.? 192.168.123.104:0/2086023674' entity='client.bootstrap-osd' cmd=[{"prefix": "mon getmap"}]: dispatch 2026-03-06T13:46:34.010 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:46:33 vm06 ceph-mon[56528]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:46:34.010 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:46:33 vm06 ceph-mon[56528]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:46:34.010 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:46:33 vm06 ceph-mon[56528]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:46:34.010 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:46:33 vm06 ceph-mon[56528]: pgmap v31: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-06T13:46:34.049 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:46:33 vm04 ceph-mon[53170]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:46:34.049 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:46:33 vm04 ceph-mon[53170]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:46:34.049 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:46:33 vm04 ceph-mon[53170]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:46:34.050 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:46:33 vm04 ceph-mon[53170]: pgmap v31: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-06T13:46:34.112 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:46:33 vm02 ceph-mon[49207]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:46:34.112 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:46:33 vm02 ceph-mon[49207]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:46:34.112 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:46:33 vm02 ceph-mon[49207]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:46:34.112 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:46:33 vm02 ceph-mon[49207]: pgmap v31: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-06T13:46:35.050 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:46:34 vm04 ceph-mon[53170]: pgmap v32: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-06T13:46:35.112 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:46:34 vm02 ceph-mon[49207]: pgmap v32: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-06T13:46:35.260 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:46:34 vm06 ceph-mon[56528]: pgmap v32: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-06T13:46:36.173 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:46:36 vm04 ceph-mon[53170]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "osd.1"}]: dispatch 2026-03-06T13:46:36.173 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:46:36 vm04 ceph-mon[53170]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-06T13:46:36.510 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:46:36 vm06 ceph-mon[56528]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "osd.1"}]: dispatch 2026-03-06T13:46:36.510 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:46:36 vm06 ceph-mon[56528]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-06T13:46:36.611 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:46:36 vm02 ceph-mon[49207]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "osd.1"}]: dispatch 2026-03-06T13:46:36.612 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:46:36 vm02 ceph-mon[49207]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-06T13:46:37.181 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:46:37 vm04 ceph-mon[53170]: Deploying daemon osd.1 on vm04 2026-03-06T13:46:37.181 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:46:37 vm04 ceph-mon[53170]: pgmap v33: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-06T13:46:37.510 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:46:37 vm06 ceph-mon[56528]: Deploying daemon osd.1 on vm04 2026-03-06T13:46:37.510 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:46:37 vm06 ceph-mon[56528]: pgmap v33: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-06T13:46:37.611 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:46:37 vm02 ceph-mon[49207]: Deploying daemon osd.1 on vm04 2026-03-06T13:46:37.612 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:46:37 vm02 ceph-mon[49207]: pgmap v33: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-06T13:46:38.778 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:46:38 vm04 ceph-mon[53170]: pgmap v34: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-06T13:46:39.112 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:46:38 vm02 ceph-mon[49207]: pgmap v34: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-06T13:46:39.260 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:46:38 vm06 ceph-mon[56528]: pgmap v34: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-06T13:46:40.050 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:46:39 vm04 ceph-mon[53170]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-06T13:46:40.050 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:46:39 vm04 ceph-mon[53170]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-06T13:46:40.050 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:46:39 vm04 ceph-mon[53170]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-06T13:46:40.050 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:46:39 vm04 ceph-mon[53170]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:46:40.050 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:46:39 vm04 ceph-mon[53170]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:46:40.050 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:46:39 vm04 ceph-mon[53170]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:46:40.111 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:46:39 vm02 ceph-mon[49207]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-06T13:46:40.112 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:46:39 vm02 ceph-mon[49207]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-06T13:46:40.112 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:46:39 vm02 ceph-mon[49207]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-06T13:46:40.112 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:46:39 vm02 ceph-mon[49207]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:46:40.112 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:46:39 vm02 ceph-mon[49207]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:46:40.112 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:46:39 vm02 ceph-mon[49207]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:46:40.260 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:46:39 vm06 ceph-mon[56528]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-06T13:46:40.260 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:46:39 vm06 ceph-mon[56528]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-06T13:46:40.260 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:46:39 vm06 ceph-mon[56528]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-06T13:46:40.260 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:46:39 vm06 ceph-mon[56528]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:46:40.260 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:46:39 vm06 ceph-mon[56528]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:46:40.260 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:46:39 vm06 ceph-mon[56528]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:46:41.050 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:46:40 vm04 ceph-mon[53170]: pgmap v35: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-06T13:46:41.111 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:46:40 vm02 ceph-mon[49207]: pgmap v35: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-06T13:46:41.260 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:46:40 vm06 ceph-mon[56528]: pgmap v35: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-06T13:46:41.567 INFO:teuthology.orchestra.run.vm04.stdout:Created osd(s) 1 on host 'vm04' 2026-03-06T13:46:41.736 DEBUG:teuthology.orchestra.run.vm04:osd.1> sudo journalctl -f -n 0 -u ceph-2b1b4a72-195a-11f1-b1df-bf1f800d676b@osd.1.service 2026-03-06T13:46:41.737 INFO:tasks.cephadm:Deploying osd.2 on vm06 with /dev/vde... 2026-03-06T13:46:41.737 DEBUG:teuthology.orchestra.run.vm06:> sudo /home/ubuntu/cephtest/cephadm --image harbor.clyso.com/custom-ceph/ceph/ceph:cobaltcore-storage-v19.2.3-fasttrack-3 ceph-volume -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 2b1b4a72-195a-11f1-b1df-bf1f800d676b -- lvm zap /dev/vde 2026-03-06T13:46:42.050 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:46:41 vm04 ceph-mon[53170]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-06T13:46:42.050 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:46:41 vm04 ceph-mon[53170]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' cmd=[{"prefix": "config rm", "who": "osd/host:vm04", "name": "osd_memory_target"}]: dispatch 2026-03-06T13:46:42.050 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:46:41 vm04 ceph-mon[53170]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:46:42.050 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:46:41 vm04 ceph-mon[53170]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-06T13:46:42.050 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:46:41 vm04 ceph-mon[53170]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-06T13:46:42.050 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:46:41 vm04 ceph-mon[53170]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:46:42.050 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:46:41 vm04 ceph-mon[53170]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:46:42.050 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:46:41 vm04 ceph-mon[53170]: from='osd.1 [v2:192.168.123.104:6800/1947797507,v1:192.168.123.104:6801/1947797507]' entity='osd.1' cmd=[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["1"]}]: dispatch 2026-03-06T13:46:42.050 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:46:41 vm04 ceph-mon[53170]: from='osd.1 ' entity='osd.1' cmd=[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["1"]}]: dispatch 2026-03-06T13:46:42.058 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:46:41 vm06 ceph-mon[56528]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-06T13:46:42.058 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:46:41 vm06 ceph-mon[56528]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' cmd=[{"prefix": "config rm", "who": "osd/host:vm04", "name": "osd_memory_target"}]: dispatch 2026-03-06T13:46:42.058 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:46:41 vm06 ceph-mon[56528]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:46:42.058 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:46:41 vm06 ceph-mon[56528]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-06T13:46:42.058 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:46:41 vm06 ceph-mon[56528]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-06T13:46:42.058 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:46:41 vm06 ceph-mon[56528]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:46:42.058 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:46:41 vm06 ceph-mon[56528]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:46:42.058 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:46:41 vm06 ceph-mon[56528]: from='osd.1 [v2:192.168.123.104:6800/1947797507,v1:192.168.123.104:6801/1947797507]' entity='osd.1' cmd=[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["1"]}]: dispatch 2026-03-06T13:46:42.058 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:46:41 vm06 ceph-mon[56528]: from='osd.1 ' entity='osd.1' cmd=[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["1"]}]: dispatch 2026-03-06T13:46:42.112 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:46:41 vm02 ceph-mon[49207]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-06T13:46:42.112 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:46:41 vm02 ceph-mon[49207]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' cmd=[{"prefix": "config rm", "who": "osd/host:vm04", "name": "osd_memory_target"}]: dispatch 2026-03-06T13:46:42.112 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:46:41 vm02 ceph-mon[49207]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:46:42.112 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:46:41 vm02 ceph-mon[49207]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-06T13:46:42.112 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:46:41 vm02 ceph-mon[49207]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-06T13:46:42.112 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:46:41 vm02 ceph-mon[49207]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:46:42.112 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:46:41 vm02 ceph-mon[49207]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:46:42.112 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:46:41 vm02 ceph-mon[49207]: from='osd.1 [v2:192.168.123.104:6800/1947797507,v1:192.168.123.104:6801/1947797507]' entity='osd.1' cmd=[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["1"]}]: dispatch 2026-03-06T13:46:42.112 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:46:41 vm02 ceph-mon[49207]: from='osd.1 ' entity='osd.1' cmd=[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["1"]}]: dispatch 2026-03-06T13:46:42.282 INFO:teuthology.orchestra.run.vm06.stderr:Inferring config /var/lib/ceph/2b1b4a72-195a-11f1-b1df-bf1f800d676b/mon.c/config 2026-03-06T13:46:43.776 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:46:43 vm02 ceph-mon[49207]: from='osd.1 ' entity='osd.1' cmd='[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["1"]}]': finished 2026-03-06T13:46:43.776 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:46:43 vm02 ceph-mon[49207]: osdmap e11: 2 total, 1 up, 2 in 2026-03-06T13:46:43.776 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:46:43 vm02 ceph-mon[49207]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 1}]: dispatch 2026-03-06T13:46:43.776 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:46:43 vm02 ceph-mon[49207]: from='osd.1 [v2:192.168.123.104:6800/1947797507,v1:192.168.123.104:6801/1947797507]' entity='osd.1' cmd=[{"prefix": "osd crush create-or-move", "id": 1, "weight":0.0195, "args": ["host=vm04", "root=default"]}]: dispatch 2026-03-06T13:46:43.776 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:46:43 vm02 ceph-mon[49207]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:46:43.776 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:46:43 vm02 ceph-mon[49207]: from='osd.1 ' entity='osd.1' cmd=[{"prefix": "osd crush create-or-move", "id": 1, "weight":0.0195, "args": ["host=vm04", "root=default"]}]: dispatch 2026-03-06T13:46:43.776 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:46:43 vm02 ceph-mon[49207]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:46:43.776 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:46:43 vm02 ceph-mon[49207]: pgmap v37: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-06T13:46:43.776 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:46:43 vm02 ceph-mon[49207]: Detected new or changed devices on vm02 2026-03-06T13:46:43.776 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:46:43 vm02 ceph-mon[49207]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-06T13:46:43.776 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:46:43 vm02 ceph-mon[49207]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:46:43.776 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:46:43 vm02 ceph-mon[49207]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-06T13:46:43.776 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:46:43 vm02 ceph-mon[49207]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-06T13:46:43.776 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:46:43 vm02 ceph-mon[49207]: pgmap v38: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-06T13:46:43.776 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:46:43 vm02 ceph-mon[49207]: pgmap v39: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-06T13:46:43.776 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:46:43 vm02 ceph-mon[49207]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:46:43.776 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:46:43 vm02 ceph-mon[49207]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:46:43.776 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:46:43 vm02 ceph-mon[49207]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:46:43.776 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:46:43 vm02 ceph-mon[49207]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:46:43.800 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:46:43 vm06 ceph-mon[56528]: from='osd.1 ' entity='osd.1' cmd='[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["1"]}]': finished 2026-03-06T13:46:43.800 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:46:43 vm06 ceph-mon[56528]: osdmap e11: 2 total, 1 up, 2 in 2026-03-06T13:46:43.800 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:46:43 vm06 ceph-mon[56528]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 1}]: dispatch 2026-03-06T13:46:43.800 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:46:43 vm06 ceph-mon[56528]: from='osd.1 [v2:192.168.123.104:6800/1947797507,v1:192.168.123.104:6801/1947797507]' entity='osd.1' cmd=[{"prefix": "osd crush create-or-move", "id": 1, "weight":0.0195, "args": ["host=vm04", "root=default"]}]: dispatch 2026-03-06T13:46:43.800 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:46:43 vm06 ceph-mon[56528]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:46:43.800 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:46:43 vm06 ceph-mon[56528]: from='osd.1 ' entity='osd.1' cmd=[{"prefix": "osd crush create-or-move", "id": 1, "weight":0.0195, "args": ["host=vm04", "root=default"]}]: dispatch 2026-03-06T13:46:43.800 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:46:43 vm06 ceph-mon[56528]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:46:43.800 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:46:43 vm06 ceph-mon[56528]: pgmap v37: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-06T13:46:43.800 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:46:43 vm06 ceph-mon[56528]: Detected new or changed devices on vm02 2026-03-06T13:46:43.800 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:46:43 vm06 ceph-mon[56528]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-06T13:46:43.800 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:46:43 vm06 ceph-mon[56528]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:46:43.800 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:46:43 vm06 ceph-mon[56528]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-06T13:46:43.800 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:46:43 vm06 ceph-mon[56528]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-06T13:46:43.800 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:46:43 vm06 ceph-mon[56528]: pgmap v38: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-06T13:46:43.800 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:46:43 vm06 ceph-mon[56528]: pgmap v39: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-06T13:46:43.800 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:46:43 vm06 ceph-mon[56528]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:46:43.800 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:46:43 vm06 ceph-mon[56528]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:46:43.800 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:46:43 vm06 ceph-mon[56528]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:46:43.800 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:46:43 vm06 ceph-mon[56528]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:46:44.050 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:46:43 vm04 ceph-mon[53170]: from='osd.1 ' entity='osd.1' cmd='[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["1"]}]': finished 2026-03-06T13:46:44.050 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:46:43 vm04 ceph-mon[53170]: osdmap e11: 2 total, 1 up, 2 in 2026-03-06T13:46:44.050 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:46:43 vm04 ceph-mon[53170]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 1}]: dispatch 2026-03-06T13:46:44.050 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:46:43 vm04 ceph-mon[53170]: from='osd.1 [v2:192.168.123.104:6800/1947797507,v1:192.168.123.104:6801/1947797507]' entity='osd.1' cmd=[{"prefix": "osd crush create-or-move", "id": 1, "weight":0.0195, "args": ["host=vm04", "root=default"]}]: dispatch 2026-03-06T13:46:44.050 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:46:43 vm04 ceph-mon[53170]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:46:44.050 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:46:43 vm04 ceph-mon[53170]: from='osd.1 ' entity='osd.1' cmd=[{"prefix": "osd crush create-or-move", "id": 1, "weight":0.0195, "args": ["host=vm04", "root=default"]}]: dispatch 2026-03-06T13:46:44.050 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:46:43 vm04 ceph-mon[53170]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:46:44.050 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:46:43 vm04 ceph-mon[53170]: pgmap v37: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-06T13:46:44.050 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:46:43 vm04 ceph-mon[53170]: Detected new or changed devices on vm02 2026-03-06T13:46:44.050 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:46:43 vm04 ceph-mon[53170]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-06T13:46:44.050 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:46:43 vm04 ceph-mon[53170]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:46:44.050 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:46:43 vm04 ceph-mon[53170]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-06T13:46:44.050 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:46:43 vm04 ceph-mon[53170]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-06T13:46:44.050 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:46:43 vm04 ceph-mon[53170]: pgmap v38: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-06T13:46:44.050 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:46:43 vm04 ceph-mon[53170]: pgmap v39: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-06T13:46:44.050 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:46:43 vm04 ceph-mon[53170]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:46:44.051 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:46:43 vm04 ceph-mon[53170]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:46:44.051 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:46:43 vm04 ceph-mon[53170]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:46:44.051 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:46:43 vm04 ceph-mon[53170]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:46:44.051 INFO:journalctl@ceph.osd.1.vm04.stdout:Mar 06 13:46:43 vm04 ceph-2b1b4a72-195a-11f1-b1df-bf1f800d676b-osd-1[58203]: 2026-03-06T12:46:43.604+0000 7f78f1299640 -1 osd.1 0 waiting for initial osdmap 2026-03-06T13:46:44.051 INFO:journalctl@ceph.osd.1.vm04.stdout:Mar 06 13:46:43 vm04 ceph-2b1b4a72-195a-11f1-b1df-bf1f800d676b-osd-1[58203]: 2026-03-06T12:46:43.608+0000 7f78ec8b0640 -1 osd.1 12 set_numa_affinity unable to identify public interface '' numa node: (2) No such file or directory 2026-03-06T13:46:44.682 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:46:44 vm06 ceph-mon[56528]: pgmap v40: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-06T13:46:44.682 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:46:44 vm06 ceph-mon[56528]: Health check failed: 1 failed cephadm daemon(s) (CEPHADM_FAILED_DAEMON) 2026-03-06T13:46:44.682 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:46:44 vm06 ceph-mon[56528]: from='osd.1 ' entity='osd.1' cmd='[{"prefix": "osd crush create-or-move", "id": 1, "weight":0.0195, "args": ["host=vm04", "root=default"]}]': finished 2026-03-06T13:46:44.682 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:46:44 vm06 ceph-mon[56528]: osdmap e12: 2 total, 1 up, 2 in 2026-03-06T13:46:44.682 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:46:44 vm06 ceph-mon[56528]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 1}]: dispatch 2026-03-06T13:46:44.682 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:46:44 vm06 ceph-mon[56528]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 1}]: dispatch 2026-03-06T13:46:44.682 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:46:44 vm06 ceph-mon[56528]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:46:44.862 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:46:44 vm02 ceph-mon[49207]: pgmap v40: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-06T13:46:44.862 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:46:44 vm02 ceph-mon[49207]: Health check failed: 1 failed cephadm daemon(s) (CEPHADM_FAILED_DAEMON) 2026-03-06T13:46:44.863 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:46:44 vm02 ceph-mon[49207]: from='osd.1 ' entity='osd.1' cmd='[{"prefix": "osd crush create-or-move", "id": 1, "weight":0.0195, "args": ["host=vm04", "root=default"]}]': finished 2026-03-06T13:46:44.863 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:46:44 vm02 ceph-mon[49207]: osdmap e12: 2 total, 1 up, 2 in 2026-03-06T13:46:44.863 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:46:44 vm02 ceph-mon[49207]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 1}]: dispatch 2026-03-06T13:46:44.863 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:46:44 vm02 ceph-mon[49207]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 1}]: dispatch 2026-03-06T13:46:44.863 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:46:44 vm02 ceph-mon[49207]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:46:44.863 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:46:44 vm02 ceph-mon[49207]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:46:44.863 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:46:44 vm02 ceph-mon[49207]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-06T13:46:44.863 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:46:44 vm02 ceph-mon[49207]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-06T13:46:44.863 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:46:44 vm02 ceph-mon[49207]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-06T13:46:44.863 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:46:44 vm02 ceph-mon[49207]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:46:44.863 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:46:44 vm02 ceph-mon[49207]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:46:45.011 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:46:44 vm06 ceph-mon[56528]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:46:45.011 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:46:44 vm06 ceph-mon[56528]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-06T13:46:45.011 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:46:44 vm06 ceph-mon[56528]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-06T13:46:45.011 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:46:44 vm06 ceph-mon[56528]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-06T13:46:45.011 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:46:44 vm06 ceph-mon[56528]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:46:45.011 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:46:44 vm06 ceph-mon[56528]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:46:45.045 INFO:teuthology.orchestra.run.vm06.stdout: 2026-03-06T13:46:45.049 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:46:44 vm04 ceph-mon[53170]: pgmap v40: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-06T13:46:45.050 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:46:44 vm04 ceph-mon[53170]: Health check failed: 1 failed cephadm daemon(s) (CEPHADM_FAILED_DAEMON) 2026-03-06T13:46:45.050 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:46:44 vm04 ceph-mon[53170]: from='osd.1 ' entity='osd.1' cmd='[{"prefix": "osd crush create-or-move", "id": 1, "weight":0.0195, "args": ["host=vm04", "root=default"]}]': finished 2026-03-06T13:46:45.050 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:46:44 vm04 ceph-mon[53170]: osdmap e12: 2 total, 1 up, 2 in 2026-03-06T13:46:45.050 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:46:44 vm04 ceph-mon[53170]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 1}]: dispatch 2026-03-06T13:46:45.050 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:46:44 vm04 ceph-mon[53170]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 1}]: dispatch 2026-03-06T13:46:45.050 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:46:44 vm04 ceph-mon[53170]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:46:45.050 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:46:44 vm04 ceph-mon[53170]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:46:45.050 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:46:44 vm04 ceph-mon[53170]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-06T13:46:45.050 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:46:44 vm04 ceph-mon[53170]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-06T13:46:45.050 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:46:44 vm04 ceph-mon[53170]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-06T13:46:45.050 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:46:44 vm04 ceph-mon[53170]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:46:45.050 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:46:44 vm04 ceph-mon[53170]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:46:45.065 DEBUG:teuthology.orchestra.run.vm06:> sudo /home/ubuntu/cephtest/cephadm --image harbor.clyso.com/custom-ceph/ceph/ceph:cobaltcore-storage-v19.2.3-fasttrack-3 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 2b1b4a72-195a-11f1-b1df-bf1f800d676b -- ceph orch daemon add osd vm06:/dev/vde 2026-03-06T13:46:45.418 INFO:teuthology.orchestra.run.vm06.stderr:Inferring config /var/lib/ceph/2b1b4a72-195a-11f1-b1df-bf1f800d676b/mon.c/config 2026-03-06T13:46:45.760 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:46:45 vm06 ceph-mon[56528]: purged_snaps scrub starts 2026-03-06T13:46:45.761 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:46:45 vm06 ceph-mon[56528]: purged_snaps scrub ok 2026-03-06T13:46:45.761 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:46:45 vm06 ceph-mon[56528]: pgmap v42: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-06T13:46:45.761 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:46:45 vm06 ceph-mon[56528]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 1}]: dispatch 2026-03-06T13:46:45.761 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:46:45 vm06 ceph-mon[56528]: osd.1 [v2:192.168.123.104:6800/1947797507,v1:192.168.123.104:6801/1947797507] boot 2026-03-06T13:46:45.761 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:46:45 vm06 ceph-mon[56528]: osdmap e13: 2 total, 2 up, 2 in 2026-03-06T13:46:45.761 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:46:45 vm06 ceph-mon[56528]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 1}]: dispatch 2026-03-06T13:46:45.761 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:46:45 vm06 ceph-mon[56528]: Health check cleared: CEPHADM_FAILED_DAEMON (was: 1 failed cephadm daemon(s)) 2026-03-06T13:46:45.761 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:46:45 vm06 ceph-mon[56528]: Cluster is now healthy 2026-03-06T13:46:46.050 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:46:45 vm04 ceph-mon[53170]: purged_snaps scrub starts 2026-03-06T13:46:46.050 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:46:45 vm04 ceph-mon[53170]: purged_snaps scrub ok 2026-03-06T13:46:46.050 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:46:45 vm04 ceph-mon[53170]: pgmap v42: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-06T13:46:46.050 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:46:45 vm04 ceph-mon[53170]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 1}]: dispatch 2026-03-06T13:46:46.050 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:46:45 vm04 ceph-mon[53170]: osd.1 [v2:192.168.123.104:6800/1947797507,v1:192.168.123.104:6801/1947797507] boot 2026-03-06T13:46:46.050 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:46:45 vm04 ceph-mon[53170]: osdmap e13: 2 total, 2 up, 2 in 2026-03-06T13:46:46.050 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:46:45 vm04 ceph-mon[53170]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 1}]: dispatch 2026-03-06T13:46:46.050 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:46:45 vm04 ceph-mon[53170]: Health check cleared: CEPHADM_FAILED_DAEMON (was: 1 failed cephadm daemon(s)) 2026-03-06T13:46:46.050 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:46:45 vm04 ceph-mon[53170]: Cluster is now healthy 2026-03-06T13:46:46.112 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:46:45 vm02 ceph-mon[49207]: purged_snaps scrub starts 2026-03-06T13:46:46.112 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:46:45 vm02 ceph-mon[49207]: purged_snaps scrub ok 2026-03-06T13:46:46.112 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:46:45 vm02 ceph-mon[49207]: pgmap v42: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-06T13:46:46.112 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:46:45 vm02 ceph-mon[49207]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 1}]: dispatch 2026-03-06T13:46:46.112 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:46:45 vm02 ceph-mon[49207]: osd.1 [v2:192.168.123.104:6800/1947797507,v1:192.168.123.104:6801/1947797507] boot 2026-03-06T13:46:46.112 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:46:45 vm02 ceph-mon[49207]: osdmap e13: 2 total, 2 up, 2 in 2026-03-06T13:46:46.112 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:46:45 vm02 ceph-mon[49207]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 1}]: dispatch 2026-03-06T13:46:46.112 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:46:45 vm02 ceph-mon[49207]: Health check cleared: CEPHADM_FAILED_DAEMON (was: 1 failed cephadm daemon(s)) 2026-03-06T13:46:46.112 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:46:45 vm02 ceph-mon[49207]: Cluster is now healthy 2026-03-06T13:46:46.694 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:46:46 vm06 ceph-mon[56528]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' cmd=[{"prefix": "osd tree", "states": ["destroyed"], "format": "json"}]: dispatch 2026-03-06T13:46:46.694 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:46:46 vm06 ceph-mon[56528]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.bootstrap-osd"}]: dispatch 2026-03-06T13:46:46.694 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:46:46 vm06 ceph-mon[56528]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-06T13:46:47.049 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:46:46 vm04 ceph-mon[53170]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' cmd=[{"prefix": "osd tree", "states": ["destroyed"], "format": "json"}]: dispatch 2026-03-06T13:46:47.049 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:46:46 vm04 ceph-mon[53170]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.bootstrap-osd"}]: dispatch 2026-03-06T13:46:47.050 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:46:46 vm04 ceph-mon[53170]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-06T13:46:47.112 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:46:46 vm02 ceph-mon[49207]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' cmd=[{"prefix": "osd tree", "states": ["destroyed"], "format": "json"}]: dispatch 2026-03-06T13:46:47.112 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:46:46 vm02 ceph-mon[49207]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.bootstrap-osd"}]: dispatch 2026-03-06T13:46:47.112 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:46:46 vm02 ceph-mon[49207]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-06T13:46:47.697 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:46:47 vm06 ceph-mon[56528]: from='client.24145 -' entity='client.admin' cmd=[{"prefix": "orch daemon add osd", "svc_arg": "vm06:/dev/vde", "target": ["mon-mgr", ""]}]: dispatch 2026-03-06T13:46:47.697 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:46:47 vm06 ceph-mon[56528]: pgmap v44: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-06T13:46:47.697 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:46:47 vm06 ceph-mon[56528]: osdmap e14: 2 total, 2 up, 2 in 2026-03-06T13:46:47.697 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:46:47 vm06 ceph-mon[56528]: from='client.? 192.168.123.106:0/3493237259' entity='client.bootstrap-osd' cmd=[{"prefix": "osd new", "uuid": "6a09eea0-8ddc-4889-9cc9-9f59732bfa8c"}]: dispatch 2026-03-06T13:46:47.697 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:46:47 vm06 ceph-mon[56528]: from='client.? ' entity='client.bootstrap-osd' cmd=[{"prefix": "osd new", "uuid": "6a09eea0-8ddc-4889-9cc9-9f59732bfa8c"}]: dispatch 2026-03-06T13:46:47.697 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:46:47 vm06 ceph-mon[56528]: from='client.? ' entity='client.bootstrap-osd' cmd='[{"prefix": "osd new", "uuid": "6a09eea0-8ddc-4889-9cc9-9f59732bfa8c"}]': finished 2026-03-06T13:46:47.697 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:46:47 vm06 ceph-mon[56528]: osdmap e15: 3 total, 2 up, 3 in 2026-03-06T13:46:47.697 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:46:47 vm06 ceph-mon[56528]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 2}]: dispatch 2026-03-06T13:46:48.050 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:46:47 vm04 ceph-mon[53170]: from='client.24145 -' entity='client.admin' cmd=[{"prefix": "orch daemon add osd", "svc_arg": "vm06:/dev/vde", "target": ["mon-mgr", ""]}]: dispatch 2026-03-06T13:46:48.050 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:46:47 vm04 ceph-mon[53170]: pgmap v44: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-06T13:46:48.050 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:46:47 vm04 ceph-mon[53170]: osdmap e14: 2 total, 2 up, 2 in 2026-03-06T13:46:48.050 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:46:47 vm04 ceph-mon[53170]: from='client.? 192.168.123.106:0/3493237259' entity='client.bootstrap-osd' cmd=[{"prefix": "osd new", "uuid": "6a09eea0-8ddc-4889-9cc9-9f59732bfa8c"}]: dispatch 2026-03-06T13:46:48.050 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:46:47 vm04 ceph-mon[53170]: from='client.? ' entity='client.bootstrap-osd' cmd=[{"prefix": "osd new", "uuid": "6a09eea0-8ddc-4889-9cc9-9f59732bfa8c"}]: dispatch 2026-03-06T13:46:48.050 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:46:47 vm04 ceph-mon[53170]: from='client.? ' entity='client.bootstrap-osd' cmd='[{"prefix": "osd new", "uuid": "6a09eea0-8ddc-4889-9cc9-9f59732bfa8c"}]': finished 2026-03-06T13:46:48.050 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:46:47 vm04 ceph-mon[53170]: osdmap e15: 3 total, 2 up, 3 in 2026-03-06T13:46:48.050 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:46:47 vm04 ceph-mon[53170]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 2}]: dispatch 2026-03-06T13:46:48.112 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:46:47 vm02 ceph-mon[49207]: from='client.24145 -' entity='client.admin' cmd=[{"prefix": "orch daemon add osd", "svc_arg": "vm06:/dev/vde", "target": ["mon-mgr", ""]}]: dispatch 2026-03-06T13:46:48.112 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:46:47 vm02 ceph-mon[49207]: pgmap v44: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-06T13:46:48.112 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:46:47 vm02 ceph-mon[49207]: osdmap e14: 2 total, 2 up, 2 in 2026-03-06T13:46:48.112 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:46:47 vm02 ceph-mon[49207]: from='client.? 192.168.123.106:0/3493237259' entity='client.bootstrap-osd' cmd=[{"prefix": "osd new", "uuid": "6a09eea0-8ddc-4889-9cc9-9f59732bfa8c"}]: dispatch 2026-03-06T13:46:48.112 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:46:47 vm02 ceph-mon[49207]: from='client.? ' entity='client.bootstrap-osd' cmd=[{"prefix": "osd new", "uuid": "6a09eea0-8ddc-4889-9cc9-9f59732bfa8c"}]: dispatch 2026-03-06T13:46:48.112 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:46:47 vm02 ceph-mon[49207]: from='client.? ' entity='client.bootstrap-osd' cmd='[{"prefix": "osd new", "uuid": "6a09eea0-8ddc-4889-9cc9-9f59732bfa8c"}]': finished 2026-03-06T13:46:48.112 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:46:47 vm02 ceph-mon[49207]: osdmap e15: 3 total, 2 up, 3 in 2026-03-06T13:46:48.112 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:46:47 vm02 ceph-mon[49207]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 2}]: dispatch 2026-03-06T13:46:49.010 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:46:48 vm06 ceph-mon[56528]: from='client.? 192.168.123.106:0/1261412365' entity='client.bootstrap-osd' cmd=[{"prefix": "mon getmap"}]: dispatch 2026-03-06T13:46:49.049 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:46:48 vm04 ceph-mon[53170]: from='client.? 192.168.123.106:0/1261412365' entity='client.bootstrap-osd' cmd=[{"prefix": "mon getmap"}]: dispatch 2026-03-06T13:46:49.112 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:46:48 vm02 ceph-mon[49207]: from='client.? 192.168.123.106:0/1261412365' entity='client.bootstrap-osd' cmd=[{"prefix": "mon getmap"}]: dispatch 2026-03-06T13:46:50.010 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:46:49 vm06 ceph-mon[56528]: pgmap v47: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-06T13:46:50.049 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:46:49 vm04 ceph-mon[53170]: pgmap v47: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-06T13:46:50.111 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:46:49 vm02 ceph-mon[49207]: pgmap v47: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-06T13:46:51.976 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:46:51 vm06 ceph-mon[56528]: pgmap v48: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-06T13:46:52.050 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:46:51 vm04 ceph-mon[53170]: pgmap v48: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-06T13:46:52.112 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:46:51 vm02 ceph-mon[49207]: pgmap v48: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-06T13:46:52.983 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:46:52 vm04 ceph-mon[53170]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "osd.2"}]: dispatch 2026-03-06T13:46:52.983 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:46:52 vm04 ceph-mon[53170]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-06T13:46:53.011 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:46:52 vm06 ceph-mon[56528]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "osd.2"}]: dispatch 2026-03-06T13:46:53.011 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:46:52 vm06 ceph-mon[56528]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-06T13:46:53.112 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:46:52 vm02 ceph-mon[49207]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "osd.2"}]: dispatch 2026-03-06T13:46:53.112 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:46:52 vm02 ceph-mon[49207]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-06T13:46:54.011 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:46:53 vm06 ceph-mon[56528]: Deploying daemon osd.2 on vm06 2026-03-06T13:46:54.011 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:46:53 vm06 ceph-mon[56528]: pgmap v49: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-06T13:46:54.011 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:46:53 vm06 ceph-mon[56528]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:46:54.011 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:46:53 vm06 ceph-mon[56528]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:46:54.011 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:46:53 vm06 ceph-mon[56528]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-06T13:46:54.011 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:46:53 vm06 ceph-mon[56528]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-06T13:46:54.011 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:46:53 vm06 ceph-mon[56528]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:46:54.011 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:46:53 vm06 ceph-mon[56528]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-06T13:46:54.011 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:46:53 vm06 ceph-mon[56528]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:46:54.011 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:46:53 vm06 ceph-mon[56528]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:46:54.011 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:46:53 vm06 ceph-mon[56528]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:46:54.011 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:46:53 vm06 ceph-mon[56528]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-06T13:46:54.011 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:46:53 vm06 ceph-mon[56528]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-06T13:46:54.011 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:46:53 vm06 ceph-mon[56528]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:46:54.011 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:46:53 vm06 ceph-mon[56528]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-06T13:46:54.011 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:46:53 vm06 ceph-mon[56528]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:46:54.050 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:46:53 vm04 ceph-mon[53170]: Deploying daemon osd.2 on vm06 2026-03-06T13:46:54.050 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:46:53 vm04 ceph-mon[53170]: pgmap v49: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-06T13:46:54.050 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:46:53 vm04 ceph-mon[53170]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:46:54.050 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:46:53 vm04 ceph-mon[53170]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:46:54.050 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:46:53 vm04 ceph-mon[53170]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-06T13:46:54.050 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:46:53 vm04 ceph-mon[53170]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-06T13:46:54.050 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:46:53 vm04 ceph-mon[53170]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:46:54.050 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:46:53 vm04 ceph-mon[53170]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-06T13:46:54.050 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:46:53 vm04 ceph-mon[53170]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:46:54.050 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:46:53 vm04 ceph-mon[53170]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:46:54.050 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:46:53 vm04 ceph-mon[53170]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:46:54.050 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:46:53 vm04 ceph-mon[53170]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-06T13:46:54.050 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:46:53 vm04 ceph-mon[53170]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-06T13:46:54.050 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:46:53 vm04 ceph-mon[53170]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:46:54.050 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:46:53 vm04 ceph-mon[53170]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-06T13:46:54.050 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:46:53 vm04 ceph-mon[53170]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:46:54.112 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:46:53 vm02 ceph-mon[49207]: Deploying daemon osd.2 on vm06 2026-03-06T13:46:54.112 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:46:53 vm02 ceph-mon[49207]: pgmap v49: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-06T13:46:54.112 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:46:53 vm02 ceph-mon[49207]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:46:54.112 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:46:53 vm02 ceph-mon[49207]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:46:54.112 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:46:53 vm02 ceph-mon[49207]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-06T13:46:54.112 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:46:53 vm02 ceph-mon[49207]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-06T13:46:54.112 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:46:53 vm02 ceph-mon[49207]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:46:54.112 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:46:53 vm02 ceph-mon[49207]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-06T13:46:54.112 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:46:53 vm02 ceph-mon[49207]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:46:54.112 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:46:53 vm02 ceph-mon[49207]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:46:54.112 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:46:53 vm02 ceph-mon[49207]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:46:54.112 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:46:53 vm02 ceph-mon[49207]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-06T13:46:54.112 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:46:53 vm02 ceph-mon[49207]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-06T13:46:54.112 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:46:53 vm02 ceph-mon[49207]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:46:54.112 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:46:53 vm02 ceph-mon[49207]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-06T13:46:54.112 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:46:53 vm02 ceph-mon[49207]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:46:55.011 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:46:54 vm06 ceph-mon[56528]: Detected new or changed devices on vm04 2026-03-06T13:46:55.012 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:46:54 vm06 ceph-mon[56528]: pgmap v50: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-06T13:46:55.050 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:46:54 vm04 ceph-mon[53170]: Detected new or changed devices on vm04 2026-03-06T13:46:55.050 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:46:54 vm04 ceph-mon[53170]: pgmap v50: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-06T13:46:55.112 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:46:54 vm02 ceph-mon[49207]: Detected new or changed devices on vm04 2026-03-06T13:46:55.112 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:46:54 vm02 ceph-mon[49207]: pgmap v50: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-06T13:46:55.761 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:46:55 vm06 ceph-mon[56528]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-06T13:46:55.761 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:46:55 vm06 ceph-mon[56528]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-06T13:46:55.761 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:46:55 vm06 ceph-mon[56528]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-06T13:46:55.761 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:46:55 vm06 ceph-mon[56528]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:46:55.761 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:46:55 vm06 ceph-mon[56528]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:46:55.761 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:46:55 vm06 ceph-mon[56528]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:46:56.050 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:46:55 vm04 ceph-mon[53170]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-06T13:46:56.050 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:46:55 vm04 ceph-mon[53170]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-06T13:46:56.050 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:46:55 vm04 ceph-mon[53170]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-06T13:46:56.050 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:46:55 vm04 ceph-mon[53170]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:46:56.050 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:46:55 vm04 ceph-mon[53170]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:46:56.050 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:46:55 vm04 ceph-mon[53170]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:46:56.112 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:46:55 vm02 ceph-mon[49207]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-06T13:46:56.112 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:46:55 vm02 ceph-mon[49207]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-06T13:46:56.112 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:46:55 vm02 ceph-mon[49207]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-06T13:46:56.112 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:46:55 vm02 ceph-mon[49207]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:46:56.112 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:46:55 vm02 ceph-mon[49207]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:46:56.112 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:46:55 vm02 ceph-mon[49207]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:46:57.011 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:46:56 vm06 ceph-mon[56528]: pgmap v51: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-06T13:46:57.050 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:46:56 vm04 ceph-mon[53170]: pgmap v51: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-06T13:46:57.112 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:46:56 vm02 ceph-mon[49207]: pgmap v51: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-06T13:46:57.579 INFO:teuthology.orchestra.run.vm06.stdout:Created osd(s) 2 on host 'vm06' 2026-03-06T13:46:57.763 DEBUG:teuthology.orchestra.run.vm06:osd.2> sudo journalctl -f -n 0 -u ceph-2b1b4a72-195a-11f1-b1df-bf1f800d676b@osd.2.service 2026-03-06T13:46:57.764 INFO:tasks.cephadm:Waiting for 3 OSDs to come up... 2026-03-06T13:46:57.764 DEBUG:teuthology.orchestra.run.vm02:> sudo /home/ubuntu/cephtest/cephadm --image harbor.clyso.com/custom-ceph/ceph/ceph:cobaltcore-storage-v19.2.3-fasttrack-3 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 2b1b4a72-195a-11f1-b1df-bf1f800d676b -- ceph osd stat -f json 2026-03-06T13:46:58.011 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:46:57 vm06 ceph-mon[56528]: from='osd.2 [v2:192.168.123.106:6800/613497499,v1:192.168.123.106:6801/613497499]' entity='osd.2' cmd=[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["2"]}]: dispatch 2026-03-06T13:46:58.011 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:46:57 vm06 ceph-mon[56528]: from='osd.2 ' entity='osd.2' cmd=[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["2"]}]: dispatch 2026-03-06T13:46:58.011 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:46:57 vm06 ceph-mon[56528]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-06T13:46:58.011 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:46:57 vm06 ceph-mon[56528]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' cmd=[{"prefix": "config rm", "who": "osd.2", "name": "osd_memory_target"}]: dispatch 2026-03-06T13:46:58.011 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:46:57 vm06 ceph-mon[56528]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:46:58.011 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:46:57 vm06 ceph-mon[56528]: Adjusting osd_memory_target on vm06 to 3329M 2026-03-06T13:46:58.011 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:46:57 vm06 ceph-mon[56528]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:46:58.011 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:46:57 vm06 ceph-mon[56528]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:46:58.011 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:46:57 vm06 ceph-mon[56528]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-06T13:46:58.011 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:46:57 vm06 ceph-mon[56528]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-06T13:46:58.011 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:46:57 vm06 ceph-mon[56528]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:46:58.050 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:46:57 vm04 ceph-mon[53170]: from='osd.2 [v2:192.168.123.106:6800/613497499,v1:192.168.123.106:6801/613497499]' entity='osd.2' cmd=[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["2"]}]: dispatch 2026-03-06T13:46:58.050 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:46:57 vm04 ceph-mon[53170]: from='osd.2 ' entity='osd.2' cmd=[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["2"]}]: dispatch 2026-03-06T13:46:58.050 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:46:57 vm04 ceph-mon[53170]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-06T13:46:58.050 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:46:57 vm04 ceph-mon[53170]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' cmd=[{"prefix": "config rm", "who": "osd.2", "name": "osd_memory_target"}]: dispatch 2026-03-06T13:46:58.050 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:46:57 vm04 ceph-mon[53170]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:46:58.050 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:46:57 vm04 ceph-mon[53170]: Adjusting osd_memory_target on vm06 to 3329M 2026-03-06T13:46:58.050 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:46:57 vm04 ceph-mon[53170]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:46:58.050 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:46:57 vm04 ceph-mon[53170]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:46:58.050 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:46:57 vm04 ceph-mon[53170]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-06T13:46:58.050 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:46:57 vm04 ceph-mon[53170]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-06T13:46:58.050 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:46:57 vm04 ceph-mon[53170]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:46:58.100 INFO:teuthology.orchestra.run.vm02.stderr:Inferring config /var/lib/ceph/2b1b4a72-195a-11f1-b1df-bf1f800d676b/mon.a/config 2026-03-06T13:46:58.112 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:46:57 vm02 ceph-mon[49207]: from='osd.2 [v2:192.168.123.106:6800/613497499,v1:192.168.123.106:6801/613497499]' entity='osd.2' cmd=[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["2"]}]: dispatch 2026-03-06T13:46:58.112 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:46:57 vm02 ceph-mon[49207]: from='osd.2 ' entity='osd.2' cmd=[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["2"]}]: dispatch 2026-03-06T13:46:58.112 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:46:57 vm02 ceph-mon[49207]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-06T13:46:58.112 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:46:57 vm02 ceph-mon[49207]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' cmd=[{"prefix": "config rm", "who": "osd.2", "name": "osd_memory_target"}]: dispatch 2026-03-06T13:46:58.112 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:46:57 vm02 ceph-mon[49207]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:46:58.112 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:46:57 vm02 ceph-mon[49207]: Adjusting osd_memory_target on vm06 to 3329M 2026-03-06T13:46:58.112 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:46:57 vm02 ceph-mon[49207]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:46:58.112 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:46:57 vm02 ceph-mon[49207]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:46:58.112 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:46:57 vm02 ceph-mon[49207]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-06T13:46:58.112 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:46:57 vm02 ceph-mon[49207]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-06T13:46:58.112 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:46:57 vm02 ceph-mon[49207]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:46:58.474 INFO:teuthology.orchestra.run.vm02.stdout: 2026-03-06T13:46:58.651 INFO:teuthology.orchestra.run.vm02.stdout:{"epoch":16,"num_osds":3,"num_up_osds":2,"osd_up_since":1772801204,"num_in_osds":3,"osd_in_since":1772801207,"num_remapped_pgs":0} 2026-03-06T13:46:58.762 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:46:58 vm02 ceph-mon[49207]: from='osd.2 ' entity='osd.2' cmd='[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["2"]}]': finished 2026-03-06T13:46:59.050 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:46:58 vm04 ceph-mon[53170]: from='osd.2 ' entity='osd.2' cmd='[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["2"]}]': finished 2026-03-06T13:46:59.050 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:46:58 vm04 ceph-mon[53170]: osdmap e16: 3 total, 2 up, 3 in 2026-03-06T13:46:59.050 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:46:58 vm04 ceph-mon[53170]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 2}]: dispatch 2026-03-06T13:46:59.050 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:46:58 vm04 ceph-mon[53170]: from='osd.2 [v2:192.168.123.106:6800/613497499,v1:192.168.123.106:6801/613497499]' entity='osd.2' cmd=[{"prefix": "osd crush create-or-move", "id": 2, "weight":0.0195, "args": ["host=vm06", "root=default"]}]: dispatch 2026-03-06T13:46:59.050 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:46:58 vm04 ceph-mon[53170]: from='osd.2 ' entity='osd.2' cmd=[{"prefix": "osd crush create-or-move", "id": 2, "weight":0.0195, "args": ["host=vm06", "root=default"]}]: dispatch 2026-03-06T13:46:59.050 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:46:58 vm04 ceph-mon[53170]: pgmap v53: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-06T13:46:59.050 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:46:58 vm04 ceph-mon[53170]: from='client.? 192.168.123.102:0/2963035490' entity='client.admin' cmd=[{"prefix": "osd stat", "format": "json"}]: dispatch 2026-03-06T13:46:59.112 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:46:58 vm02 ceph-mon[49207]: osdmap e16: 3 total, 2 up, 3 in 2026-03-06T13:46:59.112 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:46:58 vm02 ceph-mon[49207]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 2}]: dispatch 2026-03-06T13:46:59.112 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:46:58 vm02 ceph-mon[49207]: from='osd.2 [v2:192.168.123.106:6800/613497499,v1:192.168.123.106:6801/613497499]' entity='osd.2' cmd=[{"prefix": "osd crush create-or-move", "id": 2, "weight":0.0195, "args": ["host=vm06", "root=default"]}]: dispatch 2026-03-06T13:46:59.112 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:46:58 vm02 ceph-mon[49207]: from='osd.2 ' entity='osd.2' cmd=[{"prefix": "osd crush create-or-move", "id": 2, "weight":0.0195, "args": ["host=vm06", "root=default"]}]: dispatch 2026-03-06T13:46:59.112 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:46:58 vm02 ceph-mon[49207]: pgmap v53: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-06T13:46:59.112 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:46:58 vm02 ceph-mon[49207]: from='client.? 192.168.123.102:0/2963035490' entity='client.admin' cmd=[{"prefix": "osd stat", "format": "json"}]: dispatch 2026-03-06T13:46:59.260 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:46:58 vm06 ceph-mon[56528]: from='osd.2 ' entity='osd.2' cmd='[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["2"]}]': finished 2026-03-06T13:46:59.260 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:46:58 vm06 ceph-mon[56528]: osdmap e16: 3 total, 2 up, 3 in 2026-03-06T13:46:59.260 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:46:58 vm06 ceph-mon[56528]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 2}]: dispatch 2026-03-06T13:46:59.260 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:46:58 vm06 ceph-mon[56528]: from='osd.2 [v2:192.168.123.106:6800/613497499,v1:192.168.123.106:6801/613497499]' entity='osd.2' cmd=[{"prefix": "osd crush create-or-move", "id": 2, "weight":0.0195, "args": ["host=vm06", "root=default"]}]: dispatch 2026-03-06T13:46:59.260 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:46:58 vm06 ceph-mon[56528]: from='osd.2 ' entity='osd.2' cmd=[{"prefix": "osd crush create-or-move", "id": 2, "weight":0.0195, "args": ["host=vm06", "root=default"]}]: dispatch 2026-03-06T13:46:59.260 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:46:58 vm06 ceph-mon[56528]: pgmap v53: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-06T13:46:59.261 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:46:58 vm06 ceph-mon[56528]: from='client.? 192.168.123.102:0/2963035490' entity='client.admin' cmd=[{"prefix": "osd stat", "format": "json"}]: dispatch 2026-03-06T13:46:59.652 DEBUG:teuthology.orchestra.run.vm02:> sudo /home/ubuntu/cephtest/cephadm --image harbor.clyso.com/custom-ceph/ceph/ceph:cobaltcore-storage-v19.2.3-fasttrack-3 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 2b1b4a72-195a-11f1-b1df-bf1f800d676b -- ceph osd stat -f json 2026-03-06T13:46:59.998 INFO:teuthology.orchestra.run.vm02.stderr:Inferring config /var/lib/ceph/2b1b4a72-195a-11f1-b1df-bf1f800d676b/mon.a/config 2026-03-06T13:47:00.011 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:46:59 vm06 ceph-mon[56528]: from='osd.2 ' entity='osd.2' cmd='[{"prefix": "osd crush create-or-move", "id": 2, "weight":0.0195, "args": ["host=vm06", "root=default"]}]': finished 2026-03-06T13:47:00.011 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:46:59 vm06 ceph-mon[56528]: osdmap e17: 3 total, 2 up, 3 in 2026-03-06T13:47:00.011 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:46:59 vm06 ceph-mon[56528]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 2}]: dispatch 2026-03-06T13:47:00.011 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:46:59 vm06 ceph-mon[56528]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 2}]: dispatch 2026-03-06T13:47:00.011 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:46:59 vm06 ceph-mon[56528]: from='osd.2 ' entity='osd.2' 2026-03-06T13:47:00.011 INFO:journalctl@ceph.osd.2.vm06.stdout:Mar 06 13:46:59 vm06 ceph-2b1b4a72-195a-11f1-b1df-bf1f800d676b-osd-2[61819]: 2026-03-06T12:46:59.611+0000 7fa11a968640 -1 osd.2 0 waiting for initial osdmap 2026-03-06T13:47:00.011 INFO:journalctl@ceph.osd.2.vm06.stdout:Mar 06 13:46:59 vm06 ceph-2b1b4a72-195a-11f1-b1df-bf1f800d676b-osd-2[61819]: 2026-03-06T12:46:59.618+0000 7fa11577e640 -1 osd.2 17 set_numa_affinity unable to identify public interface '' numa node: (2) No such file or directory 2026-03-06T13:47:00.025 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:46:59 vm02 ceph-mon[49207]: from='osd.2 ' entity='osd.2' cmd='[{"prefix": "osd crush create-or-move", "id": 2, "weight":0.0195, "args": ["host=vm06", "root=default"]}]': finished 2026-03-06T13:47:00.025 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:46:59 vm02 ceph-mon[49207]: osdmap e17: 3 total, 2 up, 3 in 2026-03-06T13:47:00.025 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:46:59 vm02 ceph-mon[49207]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 2}]: dispatch 2026-03-06T13:47:00.025 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:46:59 vm02 ceph-mon[49207]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 2}]: dispatch 2026-03-06T13:47:00.025 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:46:59 vm02 ceph-mon[49207]: from='osd.2 ' entity='osd.2' 2026-03-06T13:47:00.050 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:46:59 vm04 ceph-mon[53170]: from='osd.2 ' entity='osd.2' cmd='[{"prefix": "osd crush create-or-move", "id": 2, "weight":0.0195, "args": ["host=vm06", "root=default"]}]': finished 2026-03-06T13:47:00.050 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:46:59 vm04 ceph-mon[53170]: osdmap e17: 3 total, 2 up, 3 in 2026-03-06T13:47:00.050 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:46:59 vm04 ceph-mon[53170]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 2}]: dispatch 2026-03-06T13:47:00.050 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:46:59 vm04 ceph-mon[53170]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 2}]: dispatch 2026-03-06T13:47:00.050 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:46:59 vm04 ceph-mon[53170]: from='osd.2 ' entity='osd.2' 2026-03-06T13:47:00.367 INFO:teuthology.orchestra.run.vm02.stdout: 2026-03-06T13:47:00.535 INFO:teuthology.orchestra.run.vm02.stdout:{"epoch":17,"num_osds":3,"num_up_osds":2,"osd_up_since":1772801204,"num_in_osds":3,"osd_in_since":1772801207,"num_remapped_pgs":0} 2026-03-06T13:47:01.050 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:47:00 vm04 ceph-mon[53170]: purged_snaps scrub starts 2026-03-06T13:47:01.050 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:47:00 vm04 ceph-mon[53170]: purged_snaps scrub ok 2026-03-06T13:47:01.050 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:47:00 vm04 ceph-mon[53170]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 2}]: dispatch 2026-03-06T13:47:01.050 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:47:00 vm04 ceph-mon[53170]: pgmap v55: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-06T13:47:01.050 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:47:00 vm04 ceph-mon[53170]: from='client.? 192.168.123.102:0/2168909318' entity='client.admin' cmd=[{"prefix": "osd stat", "format": "json"}]: dispatch 2026-03-06T13:47:01.050 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:47:00 vm04 ceph-mon[53170]: osd.2 [v2:192.168.123.106:6800/613497499,v1:192.168.123.106:6801/613497499] boot 2026-03-06T13:47:01.050 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:47:00 vm04 ceph-mon[53170]: osdmap e18: 3 total, 3 up, 3 in 2026-03-06T13:47:01.050 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:47:00 vm04 ceph-mon[53170]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 2}]: dispatch 2026-03-06T13:47:01.050 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:47:00 vm04 ceph-mon[53170]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' cmd=[{"prefix": "osd pool create", "format": "json", "pool": ".mgr", "pg_num": 1, "pg_num_min": 1, "pg_num_max": 32, "yes_i_really_mean_it": true}]: dispatch 2026-03-06T13:47:01.112 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:47:00 vm02 ceph-mon[49207]: purged_snaps scrub starts 2026-03-06T13:47:01.112 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:47:00 vm02 ceph-mon[49207]: purged_snaps scrub ok 2026-03-06T13:47:01.112 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:47:00 vm02 ceph-mon[49207]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 2}]: dispatch 2026-03-06T13:47:01.112 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:47:00 vm02 ceph-mon[49207]: pgmap v55: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-06T13:47:01.112 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:47:00 vm02 ceph-mon[49207]: from='client.? 192.168.123.102:0/2168909318' entity='client.admin' cmd=[{"prefix": "osd stat", "format": "json"}]: dispatch 2026-03-06T13:47:01.112 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:47:00 vm02 ceph-mon[49207]: osd.2 [v2:192.168.123.106:6800/613497499,v1:192.168.123.106:6801/613497499] boot 2026-03-06T13:47:01.112 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:47:00 vm02 ceph-mon[49207]: osdmap e18: 3 total, 3 up, 3 in 2026-03-06T13:47:01.112 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:47:00 vm02 ceph-mon[49207]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 2}]: dispatch 2026-03-06T13:47:01.112 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:47:00 vm02 ceph-mon[49207]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' cmd=[{"prefix": "osd pool create", "format": "json", "pool": ".mgr", "pg_num": 1, "pg_num_min": 1, "pg_num_max": 32, "yes_i_really_mean_it": true}]: dispatch 2026-03-06T13:47:01.260 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:47:00 vm06 ceph-mon[56528]: purged_snaps scrub starts 2026-03-06T13:47:01.260 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:47:00 vm06 ceph-mon[56528]: purged_snaps scrub ok 2026-03-06T13:47:01.260 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:47:00 vm06 ceph-mon[56528]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 2}]: dispatch 2026-03-06T13:47:01.260 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:47:00 vm06 ceph-mon[56528]: pgmap v55: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-06T13:47:01.260 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:47:00 vm06 ceph-mon[56528]: from='client.? 192.168.123.102:0/2168909318' entity='client.admin' cmd=[{"prefix": "osd stat", "format": "json"}]: dispatch 2026-03-06T13:47:01.260 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:47:00 vm06 ceph-mon[56528]: osd.2 [v2:192.168.123.106:6800/613497499,v1:192.168.123.106:6801/613497499] boot 2026-03-06T13:47:01.260 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:47:00 vm06 ceph-mon[56528]: osdmap e18: 3 total, 3 up, 3 in 2026-03-06T13:47:01.260 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:47:00 vm06 ceph-mon[56528]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 2}]: dispatch 2026-03-06T13:47:01.260 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:47:00 vm06 ceph-mon[56528]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' cmd=[{"prefix": "osd pool create", "format": "json", "pool": ".mgr", "pg_num": 1, "pg_num_min": 1, "pg_num_max": 32, "yes_i_really_mean_it": true}]: dispatch 2026-03-06T13:47:01.535 DEBUG:teuthology.orchestra.run.vm02:> sudo /home/ubuntu/cephtest/cephadm --image harbor.clyso.com/custom-ceph/ceph/ceph:cobaltcore-storage-v19.2.3-fasttrack-3 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 2b1b4a72-195a-11f1-b1df-bf1f800d676b -- ceph osd stat -f json 2026-03-06T13:47:01.871 INFO:teuthology.orchestra.run.vm02.stderr:Inferring config /var/lib/ceph/2b1b4a72-195a-11f1-b1df-bf1f800d676b/mon.a/config 2026-03-06T13:47:02.234 INFO:teuthology.orchestra.run.vm02.stdout: 2026-03-06T13:47:02.425 INFO:teuthology.orchestra.run.vm02.stdout:{"epoch":19,"num_osds":3,"num_up_osds":3,"osd_up_since":1772801220,"num_in_osds":3,"osd_in_since":1772801207,"num_remapped_pgs":0} 2026-03-06T13:47:02.425 DEBUG:teuthology.orchestra.run.vm02:> sudo /home/ubuntu/cephtest/cephadm --image harbor.clyso.com/custom-ceph/ceph/ceph:cobaltcore-storage-v19.2.3-fasttrack-3 shell --fsid 2b1b4a72-195a-11f1-b1df-bf1f800d676b -- ceph osd dump --format=json 2026-03-06T13:47:02.836 INFO:teuthology.orchestra.run.vm02.stderr:Inferring config /var/lib/ceph/2b1b4a72-195a-11f1-b1df-bf1f800d676b/mon.a/config 2026-03-06T13:47:02.844 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:47:02 vm02 ceph-mon[49207]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' cmd='[{"prefix": "osd pool create", "format": "json", "pool": ".mgr", "pg_num": 1, "pg_num_min": 1, "pg_num_max": 32, "yes_i_really_mean_it": true}]': finished 2026-03-06T13:47:02.844 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:47:02 vm02 ceph-mon[49207]: osdmap e19: 3 total, 3 up, 3 in 2026-03-06T13:47:02.844 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:47:02 vm02 ceph-mon[49207]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' cmd=[{"prefix": "osd pool application enable", "format": "json", "pool": ".mgr", "app": "mgr", "yes_i_really_mean_it": true}]: dispatch 2026-03-06T13:47:02.844 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:47:02 vm02 ceph-mon[49207]: from='client.? 192.168.123.102:0/3363935478' entity='client.admin' cmd=[{"prefix": "osd stat", "format": "json"}]: dispatch 2026-03-06T13:47:02.844 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:47:02 vm02 sudo[61839]: ceph : PWD=/ ; USER=root ; COMMAND=/usr/sbin/smartctl -x --json=o /dev/vda 2026-03-06T13:47:02.845 INFO:journalctl@ceph.osd.0.vm02.stdout:Mar 06 13:47:02 vm02 sudo[61816]: ceph : PWD=/ ; USER=root ; COMMAND=/usr/sbin/smartctl -x --json=o /dev/vde 2026-03-06T13:47:02.845 INFO:journalctl@ceph.osd.0.vm02.stdout:Mar 06 13:47:02 vm02 sudo[61816]: pam_systemd(sudo:session): Failed to connect to system bus: No such file or directory 2026-03-06T13:47:02.845 INFO:journalctl@ceph.osd.0.vm02.stdout:Mar 06 13:47:02 vm02 sudo[61816]: pam_unix(sudo:session): session opened for user root(uid=0) by (uid=167) 2026-03-06T13:47:02.845 INFO:journalctl@ceph.osd.0.vm02.stdout:Mar 06 13:47:02 vm02 sudo[61816]: pam_unix(sudo:session): session closed for user root 2026-03-06T13:47:02.896 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:47:02 vm06 ceph-mon[56528]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' cmd='[{"prefix": "osd pool create", "format": "json", "pool": ".mgr", "pg_num": 1, "pg_num_min": 1, "pg_num_max": 32, "yes_i_really_mean_it": true}]': finished 2026-03-06T13:47:02.896 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:47:02 vm06 ceph-mon[56528]: osdmap e19: 3 total, 3 up, 3 in 2026-03-06T13:47:02.896 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:47:02 vm06 ceph-mon[56528]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' cmd=[{"prefix": "osd pool application enable", "format": "json", "pool": ".mgr", "app": "mgr", "yes_i_really_mean_it": true}]: dispatch 2026-03-06T13:47:02.896 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:47:02 vm06 ceph-mon[56528]: from='client.? 192.168.123.102:0/3363935478' entity='client.admin' cmd=[{"prefix": "osd stat", "format": "json"}]: dispatch 2026-03-06T13:47:02.896 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:47:02 vm06 sudo[62777]: ceph : PWD=/ ; USER=root ; COMMAND=/usr/sbin/smartctl -x --json=o /dev/vda 2026-03-06T13:47:02.896 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:47:02 vm06 sudo[62777]: pam_systemd(sudo:session): Failed to connect to system bus: No such file or directory 2026-03-06T13:47:02.896 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:47:02 vm06 sudo[62777]: pam_unix(sudo:session): session opened for user root(uid=0) by (uid=167) 2026-03-06T13:47:02.896 INFO:journalctl@ceph.osd.2.vm06.stdout:Mar 06 13:47:02 vm06 sudo[62760]: ceph : PWD=/ ; USER=root ; COMMAND=/usr/sbin/smartctl -x --json=o /dev/vde 2026-03-06T13:47:02.896 INFO:journalctl@ceph.osd.2.vm06.stdout:Mar 06 13:47:02 vm06 sudo[62760]: pam_systemd(sudo:session): Failed to connect to system bus: No such file or directory 2026-03-06T13:47:02.896 INFO:journalctl@ceph.osd.2.vm06.stdout:Mar 06 13:47:02 vm06 sudo[62760]: pam_unix(sudo:session): session opened for user root(uid=0) by (uid=167) 2026-03-06T13:47:02.896 INFO:journalctl@ceph.osd.2.vm06.stdout:Mar 06 13:47:02 vm06 sudo[62760]: pam_unix(sudo:session): session closed for user root 2026-03-06T13:47:02.909 INFO:journalctl@ceph.osd.1.vm04.stdout:Mar 06 13:47:02 vm04 sudo[59609]: ceph : PWD=/ ; USER=root ; COMMAND=/usr/sbin/smartctl -x --json=o /dev/vde 2026-03-06T13:47:02.909 INFO:journalctl@ceph.osd.1.vm04.stdout:Mar 06 13:47:02 vm04 sudo[59609]: pam_systemd(sudo:session): Failed to connect to system bus: No such file or directory 2026-03-06T13:47:02.909 INFO:journalctl@ceph.osd.1.vm04.stdout:Mar 06 13:47:02 vm04 sudo[59609]: pam_unix(sudo:session): session opened for user root(uid=0) by (uid=167) 2026-03-06T13:47:02.909 INFO:journalctl@ceph.osd.1.vm04.stdout:Mar 06 13:47:02 vm04 sudo[59609]: pam_unix(sudo:session): session closed for user root 2026-03-06T13:47:02.909 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:47:02 vm04 ceph-mon[53170]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' cmd='[{"prefix": "osd pool create", "format": "json", "pool": ".mgr", "pg_num": 1, "pg_num_min": 1, "pg_num_max": 32, "yes_i_really_mean_it": true}]': finished 2026-03-06T13:47:02.909 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:47:02 vm04 ceph-mon[53170]: osdmap e19: 3 total, 3 up, 3 in 2026-03-06T13:47:02.909 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:47:02 vm04 ceph-mon[53170]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' cmd=[{"prefix": "osd pool application enable", "format": "json", "pool": ".mgr", "app": "mgr", "yes_i_really_mean_it": true}]: dispatch 2026-03-06T13:47:02.909 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:47:02 vm04 ceph-mon[53170]: from='client.? 192.168.123.102:0/3363935478' entity='client.admin' cmd=[{"prefix": "osd stat", "format": "json"}]: dispatch 2026-03-06T13:47:03.112 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:47:02 vm02 sudo[61839]: pam_systemd(sudo:session): Failed to connect to system bus: No such file or directory 2026-03-06T13:47:03.112 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:47:02 vm02 sudo[61839]: pam_unix(sudo:session): session opened for user root(uid=0) by (uid=167) 2026-03-06T13:47:03.112 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:47:02 vm02 sudo[61839]: pam_unix(sudo:session): session closed for user root 2026-03-06T13:47:03.179 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:47:02 vm06 sudo[62777]: pam_unix(sudo:session): session closed for user root 2026-03-06T13:47:03.243 INFO:teuthology.orchestra.run.vm02.stdout: 2026-03-06T13:47:03.243 INFO:teuthology.orchestra.run.vm02.stdout:{"epoch":20,"fsid":"2b1b4a72-195a-11f1-b1df-bf1f800d676b","created":"2026-03-06T12:44:49.710645+0000","modified":"2026-03-06T12:47:02.629332+0000","last_up_change":"2026-03-06T12:47:00.617832+0000","last_in_change":"2026-03-06T12:46:47.212348+0000","flags":"sortbitwise,recovery_deletes,purged_snapdirs,pglog_hardlimit","flags_num":5799936,"flags_set":["pglog_hardlimit","purged_snapdirs","recovery_deletes","sortbitwise"],"crush_version":8,"full_ratio":0.94999998807907104,"backfillfull_ratio":0.89999997615814209,"nearfull_ratio":0.85000002384185791,"cluster_snapshot":"","pool_max":1,"max_osd":3,"require_min_compat_client":"luminous","min_compat_client":"jewel","require_osd_release":"squid","allow_crimson":false,"pools":[{"pool":1,"pool_name":".mgr","create_time":"2026-03-06T12:47:00.638787+0000","flags":32769,"flags_names":"hashpspool,creating","type":1,"size":3,"min_size":2,"crush_rule":0,"peering_crush_bucket_count":0,"peering_crush_bucket_target":0,"peering_crush_bucket_barrier":0,"peering_crush_bucket_mandatory_member":2147483647,"is_stretch_pool":false,"object_hash":2,"pg_autoscale_mode":"off","pg_num":1,"pg_placement_num":1,"pg_placement_num_target":1,"pg_num_target":1,"pg_num_pending":1,"last_pg_merge_meta":{"source_pgid":"0.0","ready_epoch":0,"last_epoch_started":0,"last_epoch_clean":0,"source_version":"0'0","target_version":"0'0"},"last_change":"20","last_force_op_resend":"0","last_force_op_resend_prenautilus":"0","last_force_op_resend_preluminous":"0","auid":0,"snap_mode":"selfmanaged","snap_seq":0,"snap_epoch":0,"pool_snaps":[],"removed_snaps":"[]","quota_max_bytes":0,"quota_max_objects":0,"tiers":[],"tier_of":-1,"read_tier":-1,"write_tier":-1,"cache_mode":"none","target_max_bytes":0,"target_max_objects":0,"cache_target_dirty_ratio_micro":400000,"cache_target_dirty_high_ratio_micro":600000,"cache_target_full_ratio_micro":800000,"cache_min_flush_age":0,"cache_min_evict_age":0,"erasure_code_profile":"","hit_set_params":{"type":"none"},"hit_set_period":0,"hit_set_count":0,"use_gmt_hitset":true,"min_read_recency_for_promote":0,"min_write_recency_for_promote":0,"hit_set_grade_decay_rate":0,"hit_set_search_last_n":0,"grade_table":[],"stripe_width":0,"expected_num_objects":0,"fast_read":false,"options":{"pg_num_max":32,"pg_num_min":1},"application_metadata":{"mgr":{}},"read_balance":{"score_type":"Fair distribution","score_acting":3,"score_stable":3,"optimal_score":1,"raw_score_acting":3,"raw_score_stable":3,"primary_affinity_weighted":1,"average_primary_affinity":1,"average_primary_affinity_weighted":1}}],"osds":[{"osd":0,"uuid":"9b067403-3d7b-4d89-a7e1-1cd83ba7845f","up":1,"in":1,"weight":1,"primary_affinity":1,"last_clean_begin":0,"last_clean_end":0,"up_from":8,"up_thru":0,"down_at":0,"lost_at":0,"public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.102:6802","nonce":3358520121},{"type":"v1","addr":"192.168.123.102:6803","nonce":3358520121}]},"cluster_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.102:6804","nonce":3358520121},{"type":"v1","addr":"192.168.123.102:6805","nonce":3358520121}]},"heartbeat_back_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.102:6808","nonce":3358520121},{"type":"v1","addr":"192.168.123.102:6809","nonce":3358520121}]},"heartbeat_front_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.102:6806","nonce":3358520121},{"type":"v1","addr":"192.168.123.102:6807","nonce":3358520121}]},"public_addr":"192.168.123.102:6803/3358520121","cluster_addr":"192.168.123.102:6805/3358520121","heartbeat_back_addr":"192.168.123.102:6809/3358520121","heartbeat_front_addr":"192.168.123.102:6807/3358520121","state":["exists","up"]},{"osd":1,"uuid":"1b4f83a6-1a7a-4b65-95a3-d2d4a61115b6","up":1,"in":1,"weight":1,"primary_affinity":1,"last_clean_begin":0,"last_clean_end":0,"up_from":13,"up_thru":19,"down_at":0,"lost_at":0,"public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.104:6800","nonce":1947797507},{"type":"v1","addr":"192.168.123.104:6801","nonce":1947797507}]},"cluster_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.104:6802","nonce":1947797507},{"type":"v1","addr":"192.168.123.104:6803","nonce":1947797507}]},"heartbeat_back_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.104:6806","nonce":1947797507},{"type":"v1","addr":"192.168.123.104:6807","nonce":1947797507}]},"heartbeat_front_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.104:6804","nonce":1947797507},{"type":"v1","addr":"192.168.123.104:6805","nonce":1947797507}]},"public_addr":"192.168.123.104:6801/1947797507","cluster_addr":"192.168.123.104:6803/1947797507","heartbeat_back_addr":"192.168.123.104:6807/1947797507","heartbeat_front_addr":"192.168.123.104:6805/1947797507","state":["exists","up"]},{"osd":2,"uuid":"6a09eea0-8ddc-4889-9cc9-9f59732bfa8c","up":1,"in":1,"weight":1,"primary_affinity":1,"last_clean_begin":0,"last_clean_end":0,"up_from":18,"up_thru":0,"down_at":0,"lost_at":0,"public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.106:6800","nonce":613497499},{"type":"v1","addr":"192.168.123.106:6801","nonce":613497499}]},"cluster_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.106:6802","nonce":613497499},{"type":"v1","addr":"192.168.123.106:6803","nonce":613497499}]},"heartbeat_back_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.106:6806","nonce":613497499},{"type":"v1","addr":"192.168.123.106:6807","nonce":613497499}]},"heartbeat_front_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.106:6804","nonce":613497499},{"type":"v1","addr":"192.168.123.106:6805","nonce":613497499}]},"public_addr":"192.168.123.106:6801/613497499","cluster_addr":"192.168.123.106:6803/613497499","heartbeat_back_addr":"192.168.123.106:6807/613497499","heartbeat_front_addr":"192.168.123.106:6805/613497499","state":["exists","up"]}],"osd_xinfo":[{"osd":0,"down_stamp":"0.000000","laggy_probability":0,"laggy_interval":0,"features":4540701547738038271,"old_weight":0,"last_purged_snaps_scrub":"2026-03-06T12:46:26.304156+0000","dead_epoch":0},{"osd":1,"down_stamp":"0.000000","laggy_probability":0,"laggy_interval":0,"features":4540701547738038271,"old_weight":0,"last_purged_snaps_scrub":"2026-03-06T12:46:42.644905+0000","dead_epoch":0},{"osd":2,"down_stamp":"0.000000","laggy_probability":0,"laggy_interval":0,"features":4540701547738038271,"old_weight":0,"last_purged_snaps_scrub":"2026-03-06T12:46:57.951366+0000","dead_epoch":0}],"pg_upmap":[],"pg_upmap_items":[],"pg_upmap_primaries":[],"pg_temp":[],"primary_temp":[],"blocklist":{"192.168.123.102:0/1357185787":"2026-03-07T12:45:36.564123+0000","192.168.123.102:6801/3504819941":"2026-03-07T12:45:36.564123+0000","192.168.123.102:6800/3504819941":"2026-03-07T12:45:36.564123+0000","192.168.123.102:0/517145126":"2026-03-07T12:45:36.564123+0000","192.168.123.102:0/1479063576":"2026-03-07T12:45:36.564123+0000","192.168.123.102:0/2955285738":"2026-03-07T12:45:16.884814+0000","192.168.123.102:0/775143890":"2026-03-07T12:45:16.884814+0000","192.168.123.102:0/2841567175":"2026-03-07T12:45:16.884814+0000","192.168.123.102:6801/3168015872":"2026-03-07T12:45:16.884814+0000","192.168.123.102:6800/3168015872":"2026-03-07T12:45:16.884814+0000"},"range_blocklist":{},"erasure_code_profiles":{"default":{"crush-failure-domain":"osd","k":"2","m":"1","plugin":"jerasure","technique":"reed_sol_van"}},"removed_snaps_queue":[],"new_removed_snaps":[],"new_purged_snaps":[],"crush_node_flags":{},"device_class_flags":{},"stretch_mode":{"stretch_mode_enabled":false,"stretch_bucket_count":0,"degraded_stretch_mode":0,"recovering_stretch_mode":0,"stretch_mode_bucket":0}} 2026-03-06T13:47:03.300 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:47:02 vm04 sudo[59612]: ceph : PWD=/ ; USER=root ; COMMAND=/usr/sbin/smartctl -x --json=o /dev/vda 2026-03-06T13:47:03.300 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:47:02 vm04 sudo[59612]: pam_systemd(sudo:session): Failed to connect to system bus: No such file or directory 2026-03-06T13:47:03.300 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:47:02 vm04 sudo[59612]: pam_unix(sudo:session): session opened for user root(uid=0) by (uid=167) 2026-03-06T13:47:03.300 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:47:02 vm04 sudo[59612]: pam_unix(sudo:session): session closed for user root 2026-03-06T13:47:03.415 INFO:tasks.cephadm.ceph_manager.ceph:[{'pool': 1, 'pool_name': '.mgr', 'create_time': '2026-03-06T12:47:00.638787+0000', 'flags': 32769, 'flags_names': 'hashpspool,creating', 'type': 1, 'size': 3, 'min_size': 2, 'crush_rule': 0, 'peering_crush_bucket_count': 0, 'peering_crush_bucket_target': 0, 'peering_crush_bucket_barrier': 0, 'peering_crush_bucket_mandatory_member': 2147483647, 'is_stretch_pool': False, 'object_hash': 2, 'pg_autoscale_mode': 'off', 'pg_num': 1, 'pg_placement_num': 1, 'pg_placement_num_target': 1, 'pg_num_target': 1, 'pg_num_pending': 1, 'last_pg_merge_meta': {'source_pgid': '0.0', 'ready_epoch': 0, 'last_epoch_started': 0, 'last_epoch_clean': 0, 'source_version': "0'0", 'target_version': "0'0"}, 'last_change': '20', 'last_force_op_resend': '0', 'last_force_op_resend_prenautilus': '0', 'last_force_op_resend_preluminous': '0', 'auid': 0, 'snap_mode': 'selfmanaged', 'snap_seq': 0, 'snap_epoch': 0, 'pool_snaps': [], 'removed_snaps': '[]', 'quota_max_bytes': 0, 'quota_max_objects': 0, 'tiers': [], 'tier_of': -1, 'read_tier': -1, 'write_tier': -1, 'cache_mode': 'none', 'target_max_bytes': 0, 'target_max_objects': 0, 'cache_target_dirty_ratio_micro': 400000, 'cache_target_dirty_high_ratio_micro': 600000, 'cache_target_full_ratio_micro': 800000, 'cache_min_flush_age': 0, 'cache_min_evict_age': 0, 'erasure_code_profile': '', 'hit_set_params': {'type': 'none'}, 'hit_set_period': 0, 'hit_set_count': 0, 'use_gmt_hitset': True, 'min_read_recency_for_promote': 0, 'min_write_recency_for_promote': 0, 'hit_set_grade_decay_rate': 0, 'hit_set_search_last_n': 0, 'grade_table': [], 'stripe_width': 0, 'expected_num_objects': 0, 'fast_read': False, 'options': {'pg_num_max': 32, 'pg_num_min': 1}, 'application_metadata': {'mgr': {}}, 'read_balance': {'score_type': 'Fair distribution', 'score_acting': 3, 'score_stable': 3, 'optimal_score': 1, 'raw_score_acting': 3, 'raw_score_stable': 3, 'primary_affinity_weighted': 1, 'average_primary_affinity': 1, 'average_primary_affinity_weighted': 1}}] 2026-03-06T13:47:03.416 DEBUG:teuthology.orchestra.run.vm02:> sudo /home/ubuntu/cephtest/cephadm --image harbor.clyso.com/custom-ceph/ceph/ceph:cobaltcore-storage-v19.2.3-fasttrack-3 shell --fsid 2b1b4a72-195a-11f1-b1df-bf1f800d676b -- ceph osd pool get .mgr pg_num 2026-03-06T13:47:03.901 INFO:teuthology.orchestra.run.vm02.stderr:Inferring config /var/lib/ceph/2b1b4a72-195a-11f1-b1df-bf1f800d676b/mon.a/config 2026-03-06T13:47:03.926 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:47:03 vm02 ceph-mon[49207]: pgmap v58: 1 pgs: 1 unknown; 0 B data, 79 MiB used, 60 GiB / 60 GiB avail 2026-03-06T13:47:03.926 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:47:03 vm02 ceph-mon[49207]: Health check failed: 1 pool(s) do not have an application enabled (POOL_APP_NOT_ENABLED) 2026-03-06T13:47:03.926 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:47:03 vm02 ceph-mon[49207]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' cmd='[{"prefix": "osd pool application enable", "format": "json", "pool": ".mgr", "app": "mgr", "yes_i_really_mean_it": true}]': finished 2026-03-06T13:47:03.926 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:47:03 vm02 ceph-mon[49207]: osdmap e20: 3 total, 3 up, 3 in 2026-03-06T13:47:03.926 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:47:03 vm02 ceph-mon[49207]: from='admin socket' entity='admin socket' cmd='smart' args=[json]: dispatch 2026-03-06T13:47:03.926 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:47:03 vm02 ceph-mon[49207]: from='admin socket' entity='admin socket' cmd=smart args=[json]: finished 2026-03-06T13:47:03.926 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:47:03 vm02 ceph-mon[49207]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "a"}]: dispatch 2026-03-06T13:47:03.926 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:47:03 vm02 ceph-mon[49207]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-06T13:47:03.926 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:47:03 vm02 ceph-mon[49207]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "c"}]: dispatch 2026-03-06T13:47:03.926 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:47:03 vm02 ceph-mon[49207]: from='admin socket' entity='admin socket' cmd='smart' args=[json]: dispatch 2026-03-06T13:47:03.926 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:47:03 vm02 ceph-mon[49207]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "a"}]: dispatch 2026-03-06T13:47:03.926 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:47:03 vm02 ceph-mon[49207]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-06T13:47:03.926 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:47:03 vm02 ceph-mon[49207]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "c"}]: dispatch 2026-03-06T13:47:03.926 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:47:03 vm02 ceph-mon[49207]: from='admin socket' entity='admin socket' cmd=smart args=[json]: finished 2026-03-06T13:47:03.926 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:47:03 vm02 ceph-mon[49207]: from='admin socket' entity='admin socket' cmd='smart' args=[json]: dispatch 2026-03-06T13:47:03.926 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:47:03 vm02 ceph-mon[49207]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "a"}]: dispatch 2026-03-06T13:47:03.926 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:47:03 vm02 ceph-mon[49207]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-06T13:47:03.926 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:47:03 vm02 ceph-mon[49207]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "c"}]: dispatch 2026-03-06T13:47:03.926 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:47:03 vm02 ceph-mon[49207]: from='admin socket' entity='admin socket' cmd=smart args=[json]: finished 2026-03-06T13:47:03.926 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:47:03 vm02 ceph-mon[49207]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:47:03.926 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:47:03 vm02 ceph-mon[49207]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:47:03.926 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:47:03 vm02 ceph-mon[49207]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-06T13:47:03.926 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:47:03 vm02 ceph-mon[49207]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:47:03.926 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:47:03 vm02 ceph-mon[49207]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-06T13:47:03.926 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:47:03 vm02 ceph-mon[49207]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-06T13:47:03.926 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:47:03 vm02 ceph-mon[49207]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:47:03.926 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:47:03 vm02 ceph-mon[49207]: from='client.? 192.168.123.102:0/1663317696' entity='client.admin' cmd=[{"prefix": "osd dump", "format": "json"}]: dispatch 2026-03-06T13:47:03.986 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:47:03 vm06 ceph-mon[56528]: pgmap v58: 1 pgs: 1 unknown; 0 B data, 79 MiB used, 60 GiB / 60 GiB avail 2026-03-06T13:47:03.986 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:47:03 vm06 ceph-mon[56528]: Health check failed: 1 pool(s) do not have an application enabled (POOL_APP_NOT_ENABLED) 2026-03-06T13:47:03.986 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:47:03 vm06 ceph-mon[56528]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' cmd='[{"prefix": "osd pool application enable", "format": "json", "pool": ".mgr", "app": "mgr", "yes_i_really_mean_it": true}]': finished 2026-03-06T13:47:03.986 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:47:03 vm06 ceph-mon[56528]: osdmap e20: 3 total, 3 up, 3 in 2026-03-06T13:47:03.986 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:47:03 vm06 ceph-mon[56528]: from='admin socket' entity='admin socket' cmd='smart' args=[json]: dispatch 2026-03-06T13:47:03.986 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:47:03 vm06 ceph-mon[56528]: from='admin socket' entity='admin socket' cmd=smart args=[json]: finished 2026-03-06T13:47:03.986 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:47:03 vm06 ceph-mon[56528]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "a"}]: dispatch 2026-03-06T13:47:03.986 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:47:03 vm06 ceph-mon[56528]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-06T13:47:03.986 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:47:03 vm06 ceph-mon[56528]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "c"}]: dispatch 2026-03-06T13:47:03.986 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:47:03 vm06 ceph-mon[56528]: from='admin socket' entity='admin socket' cmd='smart' args=[json]: dispatch 2026-03-06T13:47:03.986 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:47:03 vm06 ceph-mon[56528]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "a"}]: dispatch 2026-03-06T13:47:03.986 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:47:03 vm06 ceph-mon[56528]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-06T13:47:03.986 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:47:03 vm06 ceph-mon[56528]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "c"}]: dispatch 2026-03-06T13:47:03.986 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:47:03 vm06 ceph-mon[56528]: from='admin socket' entity='admin socket' cmd=smart args=[json]: finished 2026-03-06T13:47:03.986 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:47:03 vm06 ceph-mon[56528]: from='admin socket' entity='admin socket' cmd='smart' args=[json]: dispatch 2026-03-06T13:47:03.986 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:47:03 vm06 ceph-mon[56528]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "a"}]: dispatch 2026-03-06T13:47:03.986 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:47:03 vm06 ceph-mon[56528]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-06T13:47:03.986 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:47:03 vm06 ceph-mon[56528]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "c"}]: dispatch 2026-03-06T13:47:03.986 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:47:03 vm06 ceph-mon[56528]: from='admin socket' entity='admin socket' cmd=smart args=[json]: finished 2026-03-06T13:47:03.986 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:47:03 vm06 ceph-mon[56528]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:47:03.986 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:47:03 vm06 ceph-mon[56528]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:47:03.986 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:47:03 vm06 ceph-mon[56528]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-06T13:47:03.986 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:47:03 vm06 ceph-mon[56528]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:47:03.986 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:47:03 vm06 ceph-mon[56528]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-06T13:47:03.986 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:47:03 vm06 ceph-mon[56528]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-06T13:47:03.986 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:47:03 vm06 ceph-mon[56528]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:47:03.986 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:47:03 vm06 ceph-mon[56528]: from='client.? 192.168.123.102:0/1663317696' entity='client.admin' cmd=[{"prefix": "osd dump", "format": "json"}]: dispatch 2026-03-06T13:47:04.050 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:47:03 vm04 ceph-mon[53170]: pgmap v58: 1 pgs: 1 unknown; 0 B data, 79 MiB used, 60 GiB / 60 GiB avail 2026-03-06T13:47:04.050 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:47:03 vm04 ceph-mon[53170]: Health check failed: 1 pool(s) do not have an application enabled (POOL_APP_NOT_ENABLED) 2026-03-06T13:47:04.050 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:47:03 vm04 ceph-mon[53170]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' cmd='[{"prefix": "osd pool application enable", "format": "json", "pool": ".mgr", "app": "mgr", "yes_i_really_mean_it": true}]': finished 2026-03-06T13:47:04.050 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:47:03 vm04 ceph-mon[53170]: osdmap e20: 3 total, 3 up, 3 in 2026-03-06T13:47:04.050 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:47:03 vm04 ceph-mon[53170]: from='admin socket' entity='admin socket' cmd='smart' args=[json]: dispatch 2026-03-06T13:47:04.050 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:47:03 vm04 ceph-mon[53170]: from='admin socket' entity='admin socket' cmd=smart args=[json]: finished 2026-03-06T13:47:04.050 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:47:03 vm04 ceph-mon[53170]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "a"}]: dispatch 2026-03-06T13:47:04.050 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:47:03 vm04 ceph-mon[53170]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-06T13:47:04.050 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:47:03 vm04 ceph-mon[53170]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "c"}]: dispatch 2026-03-06T13:47:04.050 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:47:03 vm04 ceph-mon[53170]: from='admin socket' entity='admin socket' cmd='smart' args=[json]: dispatch 2026-03-06T13:47:04.050 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:47:03 vm04 ceph-mon[53170]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "a"}]: dispatch 2026-03-06T13:47:04.050 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:47:03 vm04 ceph-mon[53170]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-06T13:47:04.050 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:47:03 vm04 ceph-mon[53170]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "c"}]: dispatch 2026-03-06T13:47:04.050 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:47:03 vm04 ceph-mon[53170]: from='admin socket' entity='admin socket' cmd=smart args=[json]: finished 2026-03-06T13:47:04.050 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:47:03 vm04 ceph-mon[53170]: from='admin socket' entity='admin socket' cmd='smart' args=[json]: dispatch 2026-03-06T13:47:04.050 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:47:03 vm04 ceph-mon[53170]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "a"}]: dispatch 2026-03-06T13:47:04.050 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:47:03 vm04 ceph-mon[53170]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-06T13:47:04.050 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:47:03 vm04 ceph-mon[53170]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "c"}]: dispatch 2026-03-06T13:47:04.050 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:47:03 vm04 ceph-mon[53170]: from='admin socket' entity='admin socket' cmd=smart args=[json]: finished 2026-03-06T13:47:04.050 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:47:03 vm04 ceph-mon[53170]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:47:04.050 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:47:03 vm04 ceph-mon[53170]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:47:04.050 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:47:03 vm04 ceph-mon[53170]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-06T13:47:04.050 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:47:03 vm04 ceph-mon[53170]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:47:04.050 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:47:03 vm04 ceph-mon[53170]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-06T13:47:04.050 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:47:03 vm04 ceph-mon[53170]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-06T13:47:04.050 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:47:03 vm04 ceph-mon[53170]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:47:04.050 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:47:03 vm04 ceph-mon[53170]: from='client.? 192.168.123.102:0/1663317696' entity='client.admin' cmd=[{"prefix": "osd dump", "format": "json"}]: dispatch 2026-03-06T13:47:04.300 INFO:teuthology.orchestra.run.vm02.stdout:pg_num: 1 2026-03-06T13:47:04.469 INFO:tasks.cephadm:Setting up client nodes... 2026-03-06T13:47:04.469 INFO:tasks.ceph:Waiting until ceph daemons up and pgs clean... 2026-03-06T13:47:04.469 INFO:tasks.cephadm.ceph_manager.ceph:waiting for mgr available 2026-03-06T13:47:04.469 DEBUG:teuthology.orchestra.run.vm02:> sudo /home/ubuntu/cephtest/cephadm --image harbor.clyso.com/custom-ceph/ceph/ceph:cobaltcore-storage-v19.2.3-fasttrack-3 shell --fsid 2b1b4a72-195a-11f1-b1df-bf1f800d676b -- ceph mgr dump --format=json 2026-03-06T13:47:04.653 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:47:04 vm06 ceph-mon[56528]: osdmap e21: 3 total, 3 up, 3 in 2026-03-06T13:47:04.654 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:47:04 vm06 ceph-mon[56528]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:47:04.654 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:47:04 vm06 ceph-mon[56528]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:47:04.654 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:47:04 vm06 ceph-mon[56528]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:47:04.654 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:47:04 vm06 ceph-mon[56528]: from='client.? 192.168.123.102:0/1556804825' entity='client.admin' cmd=[{"prefix": "osd pool get", "pool": ".mgr", "var": "pg_num"}]: dispatch 2026-03-06T13:47:04.654 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:47:04 vm06 ceph-mon[56528]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:47:04.654 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:47:04 vm06 ceph-mon[56528]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:47:04.654 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:47:04 vm06 ceph-mon[56528]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-06T13:47:04.654 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:47:04 vm06 ceph-mon[56528]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-06T13:47:04.654 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:47:04 vm06 ceph-mon[56528]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:47:04.654 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:47:04 vm06 ceph-mon[56528]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-06T13:47:04.654 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:47:04 vm06 ceph-mon[56528]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:47:04.828 INFO:teuthology.orchestra.run.vm02.stderr:Inferring config /var/lib/ceph/2b1b4a72-195a-11f1-b1df-bf1f800d676b/mon.a/config 2026-03-06T13:47:04.921 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:47:04 vm02 ceph-mon[49207]: osdmap e21: 3 total, 3 up, 3 in 2026-03-06T13:47:04.921 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:47:04 vm02 ceph-mon[49207]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:47:04.921 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:47:04 vm02 ceph-mon[49207]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:47:04.921 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:47:04 vm02 ceph-mon[49207]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:47:04.921 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:47:04 vm02 ceph-mon[49207]: from='client.? 192.168.123.102:0/1556804825' entity='client.admin' cmd=[{"prefix": "osd pool get", "pool": ".mgr", "var": "pg_num"}]: dispatch 2026-03-06T13:47:04.921 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:47:04 vm02 ceph-mon[49207]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:47:04.922 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:47:04 vm02 ceph-mon[49207]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:47:04.922 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:47:04 vm02 ceph-mon[49207]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-06T13:47:04.922 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:47:04 vm02 ceph-mon[49207]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-06T13:47:04.922 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:47:04 vm02 ceph-mon[49207]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:47:04.922 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:47:04 vm02 ceph-mon[49207]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-06T13:47:04.922 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:47:04 vm02 ceph-mon[49207]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:47:05.050 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:47:04 vm04 ceph-mon[53170]: osdmap e21: 3 total, 3 up, 3 in 2026-03-06T13:47:05.050 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:47:04 vm04 ceph-mon[53170]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:47:05.050 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:47:04 vm04 ceph-mon[53170]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:47:05.050 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:47:04 vm04 ceph-mon[53170]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:47:05.050 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:47:04 vm04 ceph-mon[53170]: from='client.? 192.168.123.102:0/1556804825' entity='client.admin' cmd=[{"prefix": "osd pool get", "pool": ".mgr", "var": "pg_num"}]: dispatch 2026-03-06T13:47:05.050 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:47:04 vm04 ceph-mon[53170]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:47:05.050 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:47:04 vm04 ceph-mon[53170]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:47:05.050 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:47:04 vm04 ceph-mon[53170]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-06T13:47:05.050 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:47:04 vm04 ceph-mon[53170]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-06T13:47:05.050 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:47:04 vm04 ceph-mon[53170]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:47:05.050 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:47:04 vm04 ceph-mon[53170]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-06T13:47:05.050 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:47:04 vm04 ceph-mon[53170]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:47:05.251 INFO:teuthology.orchestra.run.vm02.stdout: 2026-03-06T13:47:05.422 INFO:teuthology.orchestra.run.vm02.stdout:{"epoch":14,"flags":0,"active_gid":14156,"active_name":"a","active_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.102:6800","nonce":3376296353},{"type":"v1","addr":"192.168.123.102:6801","nonce":3376296353}]},"active_addr":"192.168.123.102:6801/3376296353","active_change":"2026-03-06T12:45:36.564214+0000","active_mgr_features":4540701547738038271,"available":true,"standbys":[{"gid":24107,"name":"b","mgr_features":4540701547738038271,"available_modules":[{"name":"alerts","can_run":true,"error_string":"","module_options":{"interval":{"name":"interval","type":"secs","level":"advanced","flags":1,"default_value":"60","min":"","max":"","enum_allowed":[],"desc":"How frequently to reexamine health status","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"smtp_destination":{"name":"smtp_destination","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"Email address to send alerts to","long_desc":"","tags":[],"see_also":[]},"smtp_from_name":{"name":"smtp_from_name","type":"str","level":"advanced","flags":1,"default_value":"Ceph","min":"","max":"","enum_allowed":[],"desc":"Email From: name","long_desc":"","tags":[],"see_also":[]},"smtp_host":{"name":"smtp_host","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"SMTP server","long_desc":"","tags":[],"see_also":[]},"smtp_password":{"name":"smtp_password","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"Password to authenticate with","long_desc":"","tags":[],"see_also":[]},"smtp_port":{"name":"smtp_port","type":"int","level":"advanced","flags":1,"default_value":"465","min":"","max":"","enum_allowed":[],"desc":"SMTP port","long_desc":"","tags":[],"see_also":[]},"smtp_sender":{"name":"smtp_sender","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"SMTP envelope sender","long_desc":"","tags":[],"see_also":[]},"smtp_ssl":{"name":"smtp_ssl","type":"bool","level":"advanced","flags":1,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"Use SSL to connect to SMTP server","long_desc":"","tags":[],"see_also":[]},"smtp_user":{"name":"smtp_user","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"User to authenticate as","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"balancer","can_run":true,"error_string":"","module_options":{"active":{"name":"active","type":"bool","level":"advanced","flags":1,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"automatically balance PGs across cluster","long_desc":"","tags":[],"see_also":[]},"begin_time":{"name":"begin_time","type":"str","level":"advanced","flags":1,"default_value":"0000","min":"","max":"","enum_allowed":[],"desc":"beginning time of day to automatically balance","long_desc":"This is a time of day in the format HHMM.","tags":[],"see_also":[]},"begin_weekday":{"name":"begin_weekday","type":"uint","level":"advanced","flags":1,"default_value":"0","min":"0","max":"6","enum_allowed":[],"desc":"Restrict automatic balancing to this day of the week or later","long_desc":"0 = Sunday, 1 = Monday, etc.","tags":[],"see_also":[]},"crush_compat_max_iterations":{"name":"crush_compat_max_iterations","type":"uint","level":"advanced","flags":1,"default_value":"25","min":"1","max":"250","enum_allowed":[],"desc":"maximum number of iterations to attempt optimization","long_desc":"","tags":[],"see_also":[]},"crush_compat_metrics":{"name":"crush_compat_metrics","type":"str","level":"advanced","flags":1,"default_value":"pgs,objects,bytes","min":"","max":"","enum_allowed":[],"desc":"metrics with which to calculate OSD utilization","long_desc":"Value is a list of one or more of \"pgs\", \"objects\", or \"bytes\", and indicates which metrics to use to balance utilization.","tags":[],"see_also":[]},"crush_compat_step":{"name":"crush_compat_step","type":"float","level":"advanced","flags":1,"default_value":"0.5","min":"0.001","max":"0.999","enum_allowed":[],"desc":"aggressiveness of optimization","long_desc":".99 is very aggressive, .01 is less aggressive","tags":[],"see_also":[]},"end_time":{"name":"end_time","type":"str","level":"advanced","flags":1,"default_value":"2359","min":"","max":"","enum_allowed":[],"desc":"ending time of day to automatically balance","long_desc":"This is a time of day in the format HHMM.","tags":[],"see_also":[]},"end_weekday":{"name":"end_weekday","type":"uint","level":"advanced","flags":1,"default_value":"0","min":"0","max":"6","enum_allowed":[],"desc":"Restrict automatic balancing to days of the week earlier than this","long_desc":"0 = Sunday, 1 = Monday, etc.","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"min_score":{"name":"min_score","type":"float","level":"advanced","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"minimum score, below which no optimization is attempted","long_desc":"","tags":[],"see_also":[]},"mode":{"name":"mode","type":"str","level":"advanced","flags":1,"default_value":"upmap","min":"","max":"","enum_allowed":["crush-compat","none","read","upmap","upmap-read"],"desc":"Balancer mode","long_desc":"","tags":[],"see_also":[]},"pool_ids":{"name":"pool_ids","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"pools which the automatic balancing will be limited to","long_desc":"","tags":[],"see_also":[]},"sleep_interval":{"name":"sleep_interval","type":"secs","level":"advanced","flags":1,"default_value":"60","min":"","max":"","enum_allowed":[],"desc":"how frequently to wake up and attempt optimization","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"update_pg_upmap_activity":{"name":"update_pg_upmap_activity","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Updates pg_upmap activity stats to be used in `balancer status detail`","long_desc":"","tags":[],"see_also":[]},"upmap_max_deviation":{"name":"upmap_max_deviation","type":"int","level":"advanced","flags":1,"default_value":"5","min":"1","max":"","enum_allowed":[],"desc":"deviation below which no optimization is attempted","long_desc":"If the number of PGs are within this count then no optimization is attempted","tags":[],"see_also":[]},"upmap_max_optimizations":{"name":"upmap_max_optimizations","type":"uint","level":"advanced","flags":1,"default_value":"10","min":"","max":"","enum_allowed":[],"desc":"maximum upmap optimizations to make per attempt","long_desc":"","tags":[],"see_also":[]}}},{"name":"cephadm","can_run":true,"error_string":"","module_options":{"agent_down_multiplier":{"name":"agent_down_multiplier","type":"float","level":"advanced","flags":0,"default_value":"3.0","min":"","max":"","enum_allowed":[],"desc":"Multiplied by agent refresh rate to calculate how long agent must not report before being marked down","long_desc":"","tags":[],"see_also":[]},"agent_refresh_rate":{"name":"agent_refresh_rate","type":"secs","level":"advanced","flags":0,"default_value":"20","min":"","max":"","enum_allowed":[],"desc":"How often agent on each host will try to gather and send metadata","long_desc":"","tags":[],"see_also":[]},"agent_starting_port":{"name":"agent_starting_port","type":"int","level":"advanced","flags":0,"default_value":"4721","min":"","max":"","enum_allowed":[],"desc":"First port agent will try to bind to (will also try up to next 1000 subsequent ports if blocked)","long_desc":"","tags":[],"see_also":[]},"allow_ptrace":{"name":"allow_ptrace","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"allow SYS_PTRACE capability on ceph containers","long_desc":"The SYS_PTRACE capability is needed to attach to a process with gdb or strace. Enabling this options can allow debugging daemons that encounter problems at runtime.","tags":[],"see_also":[]},"autotune_interval":{"name":"autotune_interval","type":"secs","level":"advanced","flags":0,"default_value":"600","min":"","max":"","enum_allowed":[],"desc":"how frequently to autotune daemon memory","long_desc":"","tags":[],"see_also":[]},"autotune_memory_target_ratio":{"name":"autotune_memory_target_ratio","type":"float","level":"advanced","flags":0,"default_value":"0.7","min":"","max":"","enum_allowed":[],"desc":"ratio of total system memory to divide amongst autotuned daemons","long_desc":"","tags":[],"see_also":[]},"cephadm_log_destination":{"name":"cephadm_log_destination","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":["file","file,syslog","syslog"],"desc":"Destination for cephadm command's persistent logging","long_desc":"","tags":[],"see_also":[]},"cgroups_split":{"name":"cgroups_split","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"Pass --cgroups=split when cephadm creates containers (currently podman only)","long_desc":"","tags":[],"see_also":[]},"config_checks_enabled":{"name":"config_checks_enabled","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Enable or disable the cephadm configuration analysis","long_desc":"","tags":[],"see_also":[]},"config_dashboard":{"name":"config_dashboard","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"manage configs like API endpoints in Dashboard.","long_desc":"","tags":[],"see_also":[]},"container_image_alertmanager":{"name":"container_image_alertmanager","type":"str","level":"advanced","flags":0,"default_value":"quay.io/prometheus/alertmanager:v0.25.0","min":"","max":"","enum_allowed":[],"desc":"Prometheus container image","long_desc":"","tags":[],"see_also":[]},"container_image_base":{"name":"container_image_base","type":"str","level":"advanced","flags":1,"default_value":"quay.io/ceph/ceph","min":"","max":"","enum_allowed":[],"desc":"Container image name, without the tag","long_desc":"","tags":[],"see_also":[]},"container_image_elasticsearch":{"name":"container_image_elasticsearch","type":"str","level":"advanced","flags":0,"default_value":"quay.io/omrizeneva/elasticsearch:6.8.23","min":"","max":"","enum_allowed":[],"desc":"elasticsearch container image","long_desc":"","tags":[],"see_also":[]},"container_image_grafana":{"name":"container_image_grafana","type":"str","level":"advanced","flags":0,"default_value":"quay.io/ceph/grafana:10.4.0","min":"","max":"","enum_allowed":[],"desc":"Prometheus container image","long_desc":"","tags":[],"see_also":[]},"container_image_haproxy":{"name":"container_image_haproxy","type":"str","level":"advanced","flags":0,"default_value":"quay.io/ceph/haproxy:2.3","min":"","max":"","enum_allowed":[],"desc":"HAproxy container image","long_desc":"","tags":[],"see_also":[]},"container_image_jaeger_agent":{"name":"container_image_jaeger_agent","type":"str","level":"advanced","flags":0,"default_value":"quay.io/jaegertracing/jaeger-agent:1.29","min":"","max":"","enum_allowed":[],"desc":"Jaeger agent container image","long_desc":"","tags":[],"see_also":[]},"container_image_jaeger_collector":{"name":"container_image_jaeger_collector","type":"str","level":"advanced","flags":0,"default_value":"quay.io/jaegertracing/jaeger-collector:1.29","min":"","max":"","enum_allowed":[],"desc":"Jaeger collector container image","long_desc":"","tags":[],"see_also":[]},"container_image_jaeger_query":{"name":"container_image_jaeger_query","type":"str","level":"advanced","flags":0,"default_value":"quay.io/jaegertracing/jaeger-query:1.29","min":"","max":"","enum_allowed":[],"desc":"Jaeger query container image","long_desc":"","tags":[],"see_also":[]},"container_image_keepalived":{"name":"container_image_keepalived","type":"str","level":"advanced","flags":0,"default_value":"quay.io/ceph/keepalived:2.2.4","min":"","max":"","enum_allowed":[],"desc":"Keepalived container image","long_desc":"","tags":[],"see_also":[]},"container_image_loki":{"name":"container_image_loki","type":"str","level":"advanced","flags":0,"default_value":"quay.io/ceph/loki:3.0.0","min":"","max":"","enum_allowed":[],"desc":"Loki container image","long_desc":"","tags":[],"see_also":[]},"container_image_node_exporter":{"name":"container_image_node_exporter","type":"str","level":"advanced","flags":0,"default_value":"quay.io/prometheus/node-exporter:v1.7.0","min":"","max":"","enum_allowed":[],"desc":"Prometheus container image","long_desc":"","tags":[],"see_also":[]},"container_image_nvmeof":{"name":"container_image_nvmeof","type":"str","level":"advanced","flags":0,"default_value":"quay.io/ceph/nvmeof:1.2.5","min":"","max":"","enum_allowed":[],"desc":"Nvme-of container image","long_desc":"","tags":[],"see_also":[]},"container_image_prometheus":{"name":"container_image_prometheus","type":"str","level":"advanced","flags":0,"default_value":"quay.io/prometheus/prometheus:v2.51.0","min":"","max":"","enum_allowed":[],"desc":"Prometheus container image","long_desc":"","tags":[],"see_also":[]},"container_image_promtail":{"name":"container_image_promtail","type":"str","level":"advanced","flags":0,"default_value":"quay.io/ceph/promtail:3.0.0","min":"","max":"","enum_allowed":[],"desc":"Promtail container image","long_desc":"","tags":[],"see_also":[]},"container_image_samba":{"name":"container_image_samba","type":"str","level":"advanced","flags":0,"default_value":"quay.io/samba.org/samba-server:devbuilds-centos-amd64","min":"","max":"","enum_allowed":[],"desc":"Samba/SMB container image","long_desc":"","tags":[],"see_also":[]},"container_image_snmp_gateway":{"name":"container_image_snmp_gateway","type":"str","level":"advanced","flags":0,"default_value":"quay.io/ceph/snmp-notifier:v1.2.1","min":"","max":"","enum_allowed":[],"desc":"SNMP Gateway container image","long_desc":"","tags":[],"see_also":[]},"container_init":{"name":"container_init","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"Run podman/docker with `--init`","long_desc":"","tags":[],"see_also":[]},"daemon_cache_timeout":{"name":"daemon_cache_timeout","type":"secs","level":"advanced","flags":0,"default_value":"600","min":"","max":"","enum_allowed":[],"desc":"seconds to cache service (daemon) inventory","long_desc":"","tags":[],"see_also":[]},"default_cephadm_command_timeout":{"name":"default_cephadm_command_timeout","type":"int","level":"advanced","flags":0,"default_value":"900","min":"","max":"","enum_allowed":[],"desc":"Default timeout applied to cephadm commands run directly on the host (in seconds)","long_desc":"","tags":[],"see_also":[]},"default_registry":{"name":"default_registry","type":"str","level":"advanced","flags":0,"default_value":"quay.io","min":"","max":"","enum_allowed":[],"desc":"Search-registry to which we should normalize unqualified image names. This is not the default registry","long_desc":"","tags":[],"see_also":[]},"device_cache_timeout":{"name":"device_cache_timeout","type":"secs","level":"advanced","flags":0,"default_value":"1800","min":"","max":"","enum_allowed":[],"desc":"seconds to cache device inventory","long_desc":"","tags":[],"see_also":[]},"device_enhanced_scan":{"name":"device_enhanced_scan","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Use libstoragemgmt during device scans","long_desc":"","tags":[],"see_also":[]},"facts_cache_timeout":{"name":"facts_cache_timeout","type":"secs","level":"advanced","flags":0,"default_value":"60","min":"","max":"","enum_allowed":[],"desc":"seconds to cache host facts data","long_desc":"","tags":[],"see_also":[]},"grafana_dashboards_path":{"name":"grafana_dashboards_path","type":"str","level":"advanced","flags":0,"default_value":"/etc/grafana/dashboards/ceph-dashboard/","min":"","max":"","enum_allowed":[],"desc":"location of dashboards to include in grafana deployments","long_desc":"","tags":[],"see_also":[]},"host_check_interval":{"name":"host_check_interval","type":"secs","level":"advanced","flags":0,"default_value":"600","min":"","max":"","enum_allowed":[],"desc":"how frequently to perform a host check","long_desc":"","tags":[],"see_also":[]},"hw_monitoring":{"name":"hw_monitoring","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Deploy hw monitoring daemon on every host.","long_desc":"","tags":[],"see_also":[]},"inventory_list_all":{"name":"inventory_list_all","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Whether ceph-volume inventory should report more devices (mostly mappers (LVs / mpaths), partitions...)","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_refresh_metadata":{"name":"log_refresh_metadata","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Log all refresh metadata. Includes daemon, device, and host info collected regularly. Only has effect if logging at debug level","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"log to the \"cephadm\" cluster log channel\"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"manage_etc_ceph_ceph_conf":{"name":"manage_etc_ceph_ceph_conf","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Manage and own /etc/ceph/ceph.conf on the hosts.","long_desc":"","tags":[],"see_also":[]},"manage_etc_ceph_ceph_conf_hosts":{"name":"manage_etc_ceph_ceph_conf_hosts","type":"str","level":"advanced","flags":0,"default_value":"*","min":"","max":"","enum_allowed":[],"desc":"PlacementSpec describing on which hosts to manage /etc/ceph/ceph.conf","long_desc":"","tags":[],"see_also":[]},"max_count_per_host":{"name":"max_count_per_host","type":"int","level":"advanced","flags":0,"default_value":"10","min":"","max":"","enum_allowed":[],"desc":"max number of daemons per service per host","long_desc":"","tags":[],"see_also":[]},"max_osd_draining_count":{"name":"max_osd_draining_count","type":"int","level":"advanced","flags":0,"default_value":"10","min":"","max":"","enum_allowed":[],"desc":"max number of osds that will be drained simultaneously when osds are removed","long_desc":"","tags":[],"see_also":[]},"migration_current":{"name":"migration_current","type":"int","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"internal - do not modify","long_desc":"","tags":[],"see_also":[]},"mode":{"name":"mode","type":"str","level":"advanced","flags":0,"default_value":"root","min":"","max":"","enum_allowed":["cephadm-package","root"],"desc":"mode for remote execution of cephadm","long_desc":"","tags":[],"see_also":[]},"oob_default_addr":{"name":"oob_default_addr","type":"str","level":"advanced","flags":0,"default_value":"169.254.1.1","min":"","max":"","enum_allowed":[],"desc":"Default address for RedFish API (oob management).","long_desc":"","tags":[],"see_also":[]},"prometheus_alerts_path":{"name":"prometheus_alerts_path","type":"str","level":"advanced","flags":0,"default_value":"/etc/prometheus/ceph/ceph_default_alerts.yml","min":"","max":"","enum_allowed":[],"desc":"location of alerts to include in prometheus deployments","long_desc":"","tags":[],"see_also":[]},"registry_insecure":{"name":"registry_insecure","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Registry is to be considered insecure (no TLS available). Only for development purposes.","long_desc":"","tags":[],"see_also":[]},"registry_password":{"name":"registry_password","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"Custom repository password. Only used for logging into a registry.","long_desc":"","tags":[],"see_also":[]},"registry_url":{"name":"registry_url","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"Registry url for login purposes. This is not the default registry","long_desc":"","tags":[],"see_also":[]},"registry_username":{"name":"registry_username","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"Custom repository username. Only used for logging into a registry.","long_desc":"","tags":[],"see_also":[]},"secure_monitoring_stack":{"name":"secure_monitoring_stack","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Enable TLS security for all the monitoring stack daemons","long_desc":"","tags":[],"see_also":[]},"service_discovery_port":{"name":"service_discovery_port","type":"int","level":"advanced","flags":0,"default_value":"8765","min":"","max":"","enum_allowed":[],"desc":"cephadm service discovery port","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"ssh_config_file":{"name":"ssh_config_file","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"customized SSH config file to connect to managed hosts","long_desc":"","tags":[],"see_also":[]},"ssh_keepalive_count_max":{"name":"ssh_keepalive_count_max","type":"int","level":"advanced","flags":0,"default_value":"3","min":"","max":"","enum_allowed":[],"desc":"How many times ssh connections can fail liveness checks before the host is marked offline","long_desc":"","tags":[],"see_also":[]},"ssh_keepalive_interval":{"name":"ssh_keepalive_interval","type":"int","level":"advanced","flags":0,"default_value":"7","min":"","max":"","enum_allowed":[],"desc":"How often ssh connections are checked for liveness","long_desc":"","tags":[],"see_also":[]},"use_agent":{"name":"use_agent","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Use cephadm agent on each host to gather and send metadata","long_desc":"","tags":[],"see_also":[]},"use_repo_digest":{"name":"use_repo_digest","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"Automatically convert image tags to image digest. Make sure all daemons use the same image","long_desc":"","tags":[],"see_also":[]},"warn_on_failed_host_check":{"name":"warn_on_failed_host_check","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"raise a health warning if the host check fails","long_desc":"","tags":[],"see_also":[]},"warn_on_stray_daemons":{"name":"warn_on_stray_daemons","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"raise a health warning if daemons are detected that are not managed by cephadm","long_desc":"","tags":[],"see_also":[]},"warn_on_stray_hosts":{"name":"warn_on_stray_hosts","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"raise a health warning if daemons are detected on a host that is not managed by cephadm","long_desc":"","tags":[],"see_also":[]}}},{"name":"crash","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"retain_interval":{"name":"retain_interval","type":"secs","level":"advanced","flags":1,"default_value":"31536000","min":"","max":"","enum_allowed":[],"desc":"how long to retain crashes before pruning them","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"warn_recent_interval":{"name":"warn_recent_interval","type":"secs","level":"advanced","flags":1,"default_value":"1209600","min":"","max":"","enum_allowed":[],"desc":"time interval in which to warn about recent crashes","long_desc":"","tags":[],"see_also":[]}}},{"name":"dashboard","can_run":true,"error_string":"","module_options":{"ACCOUNT_LOCKOUT_ATTEMPTS":{"name":"ACCOUNT_LOCKOUT_ATTEMPTS","type":"int","level":"advanced","flags":0,"default_value":"10","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"ALERTMANAGER_API_HOST":{"name":"ALERTMANAGER_API_HOST","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"ALERTMANAGER_API_SSL_VERIFY":{"name":"ALERTMANAGER_API_SSL_VERIFY","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"AUDIT_API_ENABLED":{"name":"AUDIT_API_ENABLED","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"AUDIT_API_LOG_PAYLOAD":{"name":"AUDIT_API_LOG_PAYLOAD","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"ENABLE_BROWSABLE_API":{"name":"ENABLE_BROWSABLE_API","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"FEATURE_TOGGLE_CEPHFS":{"name":"FEATURE_TOGGLE_CEPHFS","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"FEATURE_TOGGLE_DASHBOARD":{"name":"FEATURE_TOGGLE_DASHBOARD","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"FEATURE_TOGGLE_ISCSI":{"name":"FEATURE_TOGGLE_ISCSI","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"FEATURE_TOGGLE_MIRRORING":{"name":"FEATURE_TOGGLE_MIRRORING","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"FEATURE_TOGGLE_NFS":{"name":"FEATURE_TOGGLE_NFS","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"FEATURE_TOGGLE_RBD":{"name":"FEATURE_TOGGLE_RBD","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"FEATURE_TOGGLE_RGW":{"name":"FEATURE_TOGGLE_RGW","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"GANESHA_CLUSTERS_RADOS_POOL_NAMESPACE":{"name":"GANESHA_CLUSTERS_RADOS_POOL_NAMESPACE","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"GRAFANA_API_PASSWORD":{"name":"GRAFANA_API_PASSWORD","type":"str","level":"advanced","flags":0,"default_value":"admin","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"GRAFANA_API_SSL_VERIFY":{"name":"GRAFANA_API_SSL_VERIFY","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"GRAFANA_API_URL":{"name":"GRAFANA_API_URL","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"GRAFANA_API_USERNAME":{"name":"GRAFANA_API_USERNAME","type":"str","level":"advanced","flags":0,"default_value":"admin","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"GRAFANA_FRONTEND_API_URL":{"name":"GRAFANA_FRONTEND_API_URL","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"GRAFANA_UPDATE_DASHBOARDS":{"name":"GRAFANA_UPDATE_DASHBOARDS","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"ISCSI_API_SSL_VERIFICATION":{"name":"ISCSI_API_SSL_VERIFICATION","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"ISSUE_TRACKER_API_KEY":{"name":"ISSUE_TRACKER_API_KEY","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PROMETHEUS_API_HOST":{"name":"PROMETHEUS_API_HOST","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PROMETHEUS_API_SSL_VERIFY":{"name":"PROMETHEUS_API_SSL_VERIFY","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PWD_POLICY_CHECK_COMPLEXITY_ENABLED":{"name":"PWD_POLICY_CHECK_COMPLEXITY_ENABLED","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PWD_POLICY_CHECK_EXCLUSION_LIST_ENABLED":{"name":"PWD_POLICY_CHECK_EXCLUSION_LIST_ENABLED","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PWD_POLICY_CHECK_LENGTH_ENABLED":{"name":"PWD_POLICY_CHECK_LENGTH_ENABLED","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PWD_POLICY_CHECK_OLDPWD_ENABLED":{"name":"PWD_POLICY_CHECK_OLDPWD_ENABLED","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PWD_POLICY_CHECK_REPETITIVE_CHARS_ENABLED":{"name":"PWD_POLICY_CHECK_REPETITIVE_CHARS_ENABLED","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PWD_POLICY_CHECK_SEQUENTIAL_CHARS_ENABLED":{"name":"PWD_POLICY_CHECK_SEQUENTIAL_CHARS_ENABLED","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PWD_POLICY_CHECK_USERNAME_ENABLED":{"name":"PWD_POLICY_CHECK_USERNAME_ENABLED","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PWD_POLICY_ENABLED":{"name":"PWD_POLICY_ENABLED","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PWD_POLICY_EXCLUSION_LIST":{"name":"PWD_POLICY_EXCLUSION_LIST","type":"str","level":"advanced","flags":0,"default_value":"osd,host,dashboard,pool,block,nfs,ceph,monitors,gateway,logs,crush,maps","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PWD_POLICY_MIN_COMPLEXITY":{"name":"PWD_POLICY_MIN_COMPLEXITY","type":"int","level":"advanced","flags":0,"default_value":"10","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PWD_POLICY_MIN_LENGTH":{"name":"PWD_POLICY_MIN_LENGTH","type":"int","level":"advanced","flags":0,"default_value":"8","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"REST_REQUESTS_TIMEOUT":{"name":"REST_REQUESTS_TIMEOUT","type":"int","level":"advanced","flags":0,"default_value":"45","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"RGW_API_ACCESS_KEY":{"name":"RGW_API_ACCESS_KEY","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"RGW_API_ADMIN_RESOURCE":{"name":"RGW_API_ADMIN_RESOURCE","type":"str","level":"advanced","flags":0,"default_value":"admin","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"RGW_API_SECRET_KEY":{"name":"RGW_API_SECRET_KEY","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"RGW_API_SSL_VERIFY":{"name":"RGW_API_SSL_VERIFY","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"USER_PWD_EXPIRATION_SPAN":{"name":"USER_PWD_EXPIRATION_SPAN","type":"int","level":"advanced","flags":0,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"USER_PWD_EXPIRATION_WARNING_1":{"name":"USER_PWD_EXPIRATION_WARNING_1","type":"int","level":"advanced","flags":0,"default_value":"10","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"USER_PWD_EXPIRATION_WARNING_2":{"name":"USER_PWD_EXPIRATION_WARNING_2","type":"int","level":"advanced","flags":0,"default_value":"5","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"cross_origin_url":{"name":"cross_origin_url","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"crt_file":{"name":"crt_file","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"debug":{"name":"debug","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Enable/disable debug options","long_desc":"","tags":[],"see_also":[]},"jwt_token_ttl":{"name":"jwt_token_ttl","type":"int","level":"advanced","flags":0,"default_value":"28800","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"key_file":{"name":"key_file","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"motd":{"name":"motd","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"The message of the day","long_desc":"","tags":[],"see_also":[]},"redirect_resolve_ip_addr":{"name":"redirect_resolve_ip_addr","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"server_addr":{"name":"server_addr","type":"str","level":"advanced","flags":0,"default_value":"::","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"server_port":{"name":"server_port","type":"int","level":"advanced","flags":0,"default_value":"8080","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"ssl":{"name":"ssl","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"ssl_server_port":{"name":"ssl_server_port","type":"int","level":"advanced","flags":0,"default_value":"8443","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"standby_behaviour":{"name":"standby_behaviour","type":"str","level":"advanced","flags":0,"default_value":"redirect","min":"","max":"","enum_allowed":["error","redirect"],"desc":"","long_desc":"","tags":[],"see_also":[]},"standby_error_status_code":{"name":"standby_error_status_code","type":"int","level":"advanced","flags":0,"default_value":"500","min":"400","max":"599","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"url_prefix":{"name":"url_prefix","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"devicehealth","can_run":true,"error_string":"","module_options":{"enable_monitoring":{"name":"enable_monitoring","type":"bool","level":"advanced","flags":1,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"monitor device health metrics","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"mark_out_threshold":{"name":"mark_out_threshold","type":"secs","level":"advanced","flags":1,"default_value":"2419200","min":"","max":"","enum_allowed":[],"desc":"automatically mark OSD if it may fail before this long","long_desc":"","tags":[],"see_also":[]},"pool_name":{"name":"pool_name","type":"str","level":"advanced","flags":1,"default_value":"device_health_metrics","min":"","max":"","enum_allowed":[],"desc":"name of pool in which to store device health metrics","long_desc":"","tags":[],"see_also":[]},"retention_period":{"name":"retention_period","type":"secs","level":"advanced","flags":1,"default_value":"15552000","min":"","max":"","enum_allowed":[],"desc":"how long to retain device health metrics","long_desc":"","tags":[],"see_also":[]},"scrape_frequency":{"name":"scrape_frequency","type":"secs","level":"advanced","flags":1,"default_value":"86400","min":"","max":"","enum_allowed":[],"desc":"how frequently to scrape device health metrics","long_desc":"","tags":[],"see_also":[]},"self_heal":{"name":"self_heal","type":"bool","level":"advanced","flags":1,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"preemptively heal cluster around devices that may fail","long_desc":"","tags":[],"see_also":[]},"sleep_interval":{"name":"sleep_interval","type":"secs","level":"advanced","flags":1,"default_value":"600","min":"","max":"","enum_allowed":[],"desc":"how frequently to wake up and check device health","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"warn_threshold":{"name":"warn_threshold","type":"secs","level":"advanced","flags":1,"default_value":"7257600","min":"","max":"","enum_allowed":[],"desc":"raise health warning if OSD may fail before this long","long_desc":"","tags":[],"see_also":[]}}},{"name":"influx","can_run":false,"error_string":"influxdb python module not found","module_options":{"batch_size":{"name":"batch_size","type":"int","level":"advanced","flags":0,"default_value":"5000","min":"","max":"","enum_allowed":[],"desc":"How big batches of data points should be when sending to InfluxDB.","long_desc":"","tags":[],"see_also":[]},"database":{"name":"database","type":"str","level":"advanced","flags":0,"default_value":"ceph","min":"","max":"","enum_allowed":[],"desc":"InfluxDB database name. You will need to create this database and grant write privileges to the configured username or the username must have admin privileges to create it.","long_desc":"","tags":[],"see_also":[]},"hostname":{"name":"hostname","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"InfluxDB server hostname","long_desc":"","tags":[],"see_also":[]},"interval":{"name":"interval","type":"secs","level":"advanced","flags":0,"default_value":"30","min":"5","max":"","enum_allowed":[],"desc":"Time between reports to InfluxDB. Default 30 seconds.","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"password":{"name":"password","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"password of InfluxDB server user","long_desc":"","tags":[],"see_also":[]},"port":{"name":"port","type":"int","level":"advanced","flags":0,"default_value":"8086","min":"","max":"","enum_allowed":[],"desc":"InfluxDB server port","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"ssl":{"name":"ssl","type":"str","level":"advanced","flags":0,"default_value":"false","min":"","max":"","enum_allowed":[],"desc":"Use https connection for InfluxDB server. Use \"true\" or \"false\".","long_desc":"","tags":[],"see_also":[]},"threads":{"name":"threads","type":"int","level":"advanced","flags":0,"default_value":"5","min":"1","max":"32","enum_allowed":[],"desc":"How many worker threads should be spawned for sending data to InfluxDB.","long_desc":"","tags":[],"see_also":[]},"username":{"name":"username","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"username of InfluxDB server user","long_desc":"","tags":[],"see_also":[]},"verify_ssl":{"name":"verify_ssl","type":"str","level":"advanced","flags":0,"default_value":"true","min":"","max":"","enum_allowed":[],"desc":"Verify https cert for InfluxDB server. Use \"true\" or \"false\".","long_desc":"","tags":[],"see_also":[]}}},{"name":"insights","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"iostat","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"k8sevents","can_run":true,"error_string":"","module_options":{"ceph_event_retention_days":{"name":"ceph_event_retention_days","type":"int","level":"advanced","flags":0,"default_value":"7","min":"","max":"","enum_allowed":[],"desc":"Days to hold ceph event information within local cache","long_desc":"","tags":[],"see_also":[]},"config_check_secs":{"name":"config_check_secs","type":"int","level":"advanced","flags":0,"default_value":"10","min":"10","max":"","enum_allowed":[],"desc":"interval (secs) to check for cluster configuration changes","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"localpool","can_run":true,"error_string":"","module_options":{"failure_domain":{"name":"failure_domain","type":"str","level":"advanced","flags":1,"default_value":"host","min":"","max":"","enum_allowed":[],"desc":"failure domain for any created local pool","long_desc":"what failure domain we should separate data replicas across.","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"min_size":{"name":"min_size","type":"int","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"default min_size for any created local pool","long_desc":"value to set min_size to (unchanged from Ceph's default if this option is not set)","tags":[],"see_also":[]},"num_rep":{"name":"num_rep","type":"int","level":"advanced","flags":1,"default_value":"3","min":"","max":"","enum_allowed":[],"desc":"default replica count for any created local pool","long_desc":"","tags":[],"see_also":[]},"pg_num":{"name":"pg_num","type":"int","level":"advanced","flags":1,"default_value":"128","min":"","max":"","enum_allowed":[],"desc":"default pg_num for any created local pool","long_desc":"","tags":[],"see_also":[]},"prefix":{"name":"prefix","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"name prefix for any created local pool","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"subtree":{"name":"subtree","type":"str","level":"advanced","flags":1,"default_value":"rack","min":"","max":"","enum_allowed":[],"desc":"CRUSH level for which to create a local pool","long_desc":"which CRUSH subtree type the module should create a pool for.","tags":[],"see_also":[]}}},{"name":"mds_autoscaler","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"mirroring","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"nfs","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"orchestrator","can_run":true,"error_string":"","module_options":{"fail_fs":{"name":"fail_fs","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Fail filesystem for rapid multi-rank mds upgrade","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"orchestrator":{"name":"orchestrator","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["cephadm","rook","test_orchestrator"],"desc":"Orchestrator backend","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"osd_perf_query","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"osd_support","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"pg_autoscaler","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sleep_interval":{"name":"sleep_interval","type":"secs","level":"advanced","flags":0,"default_value":"60","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"threshold":{"name":"threshold","type":"float","level":"advanced","flags":0,"default_value":"3.0","min":"1.0","max":"","enum_allowed":[],"desc":"scaling threshold","long_desc":"The factor by which the `NEW PG_NUM` must vary from the current`PG_NUM` before being accepted. Cannot be less than 1.0","tags":[],"see_also":[]}}},{"name":"progress","can_run":true,"error_string":"","module_options":{"allow_pg_recovery_event":{"name":"allow_pg_recovery_event","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"allow the module to show pg recovery progress","long_desc":"","tags":[],"see_also":[]},"enabled":{"name":"enabled","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"max_completed_events":{"name":"max_completed_events","type":"int","level":"advanced","flags":1,"default_value":"50","min":"","max":"","enum_allowed":[],"desc":"number of past completed events to remember","long_desc":"","tags":[],"see_also":[]},"sleep_interval":{"name":"sleep_interval","type":"secs","level":"advanced","flags":1,"default_value":"5","min":"","max":"","enum_allowed":[],"desc":"how long the module is going to sleep","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"prometheus","can_run":true,"error_string":"","module_options":{"cache":{"name":"cache","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"exclude_perf_counters":{"name":"exclude_perf_counters","type":"bool","level":"advanced","flags":1,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"Do not include perf-counters in the metrics output","long_desc":"Gathering perf-counters from a single Prometheus exporter can degrade ceph-mgr performance, especially in large clusters. Instead, Ceph-exporter daemons are now used by default for perf-counter gathering. This should only be disabled when no ceph-exporters are deployed.","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"rbd_stats_pools":{"name":"rbd_stats_pools","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"rbd_stats_pools_refresh_interval":{"name":"rbd_stats_pools_refresh_interval","type":"int","level":"advanced","flags":0,"default_value":"300","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"scrape_interval":{"name":"scrape_interval","type":"float","level":"advanced","flags":0,"default_value":"15.0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"server_addr":{"name":"server_addr","type":"str","level":"advanced","flags":0,"default_value":"::","min":"","max":"","enum_allowed":[],"desc":"the IPv4 or IPv6 address on which the module listens for HTTP requests","long_desc":"","tags":[],"see_also":[]},"server_port":{"name":"server_port","type":"int","level":"advanced","flags":1,"default_value":"9283","min":"","max":"","enum_allowed":[],"desc":"the port on which the module listens for HTTP requests","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"stale_cache_strategy":{"name":"stale_cache_strategy","type":"str","level":"advanced","flags":0,"default_value":"log","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"standby_behaviour":{"name":"standby_behaviour","type":"str","level":"advanced","flags":1,"default_value":"default","min":"","max":"","enum_allowed":["default","error"],"desc":"","long_desc":"","tags":[],"see_also":[]},"standby_error_status_code":{"name":"standby_error_status_code","type":"int","level":"advanced","flags":1,"default_value":"500","min":"400","max":"599","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"rbd_support","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"max_concurrent_snap_create":{"name":"max_concurrent_snap_create","type":"int","level":"advanced","flags":0,"default_value":"10","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"mirror_snapshot_schedule":{"name":"mirror_snapshot_schedule","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"trash_purge_schedule":{"name":"trash_purge_schedule","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"restful","can_run":true,"error_string":"","module_options":{"enable_auth":{"name":"enable_auth","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"key_file":{"name":"key_file","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"max_requests":{"name":"max_requests","type":"int","level":"advanced","flags":0,"default_value":"500","min":"","max":"","enum_allowed":[],"desc":"Maximum number of requests to keep in memory. When new request comes in, the oldest request will be removed if the number of requests exceeds the max request number. if un-finished request is removed, error message will be logged in the ceph-mgr log.","long_desc":"","tags":[],"see_also":[]},"server_addr":{"name":"server_addr","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"server_port":{"name":"server_port","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"rgw","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"secondary_zone_period_retry_limit":{"name":"secondary_zone_period_retry_limit","type":"int","level":"advanced","flags":0,"default_value":"5","min":"","max":"","enum_allowed":[],"desc":"RGW module period update retry limit for secondary site","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"rook","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"storage_class":{"name":"storage_class","type":"str","level":"advanced","flags":0,"default_value":"local","min":"","max":"","enum_allowed":[],"desc":"storage class name for LSO-discovered PVs","long_desc":"","tags":[],"see_also":[]}}},{"name":"selftest","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"roption1":{"name":"roption1","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"roption2":{"name":"roption2","type":"str","level":"advanced","flags":0,"default_value":"xyz","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"rwoption1":{"name":"rwoption1","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"rwoption2":{"name":"rwoption2","type":"int","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"rwoption3":{"name":"rwoption3","type":"float","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"rwoption4":{"name":"rwoption4","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"rwoption5":{"name":"rwoption5","type":"bool","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"rwoption6":{"name":"rwoption6","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"rwoption7":{"name":"rwoption7","type":"int","level":"advanced","flags":0,"default_value":"","min":"1","max":"42","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"testkey":{"name":"testkey","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"testlkey":{"name":"testlkey","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"testnewline":{"name":"testnewline","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"snap_schedule","can_run":true,"error_string":"","module_options":{"allow_m_granularity":{"name":"allow_m_granularity","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"allow minute scheduled snapshots","long_desc":"","tags":[],"see_also":[]},"dump_on_update":{"name":"dump_on_update","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"dump database to debug log on update","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"stats","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"status","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"telegraf","can_run":true,"error_string":"","module_options":{"address":{"name":"address","type":"str","level":"advanced","flags":0,"default_value":"unixgram:///tmp/telegraf.sock","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"interval":{"name":"interval","type":"secs","level":"advanced","flags":0,"default_value":"15","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"telemetry","can_run":true,"error_string":"","module_options":{"channel_basic":{"name":"channel_basic","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"Share basic cluster information (size, version)","long_desc":"","tags":[],"see_also":[]},"channel_crash":{"name":"channel_crash","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"Share metadata about Ceph daemon crashes (version, stack straces, etc)","long_desc":"","tags":[],"see_also":[]},"channel_device":{"name":"channel_device","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"Share device health metrics (e.g., SMART data, minus potentially identifying info like serial numbers)","long_desc":"","tags":[],"see_also":[]},"channel_ident":{"name":"channel_ident","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Share a user-provided description and/or contact email for the cluster","long_desc":"","tags":[],"see_also":[]},"channel_perf":{"name":"channel_perf","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Share various performance metrics of a cluster","long_desc":"","tags":[],"see_also":[]},"contact":{"name":"contact","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"description":{"name":"description","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"device_url":{"name":"device_url","type":"str","level":"advanced","flags":0,"default_value":"https://telemetry.ceph.com/device","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"enabled":{"name":"enabled","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"interval":{"name":"interval","type":"int","level":"advanced","flags":0,"default_value":"24","min":"8","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"last_opt_revision":{"name":"last_opt_revision","type":"int","level":"advanced","flags":0,"default_value":"1","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"leaderboard":{"name":"leaderboard","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"leaderboard_description":{"name":"leaderboard_description","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"organization":{"name":"organization","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"proxy":{"name":"proxy","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"url":{"name":"url","type":"str","level":"advanced","flags":0,"default_value":"https://telemetry.ceph.com/report","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"test_orchestrator","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"volumes","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"max_concurrent_clones":{"name":"max_concurrent_clones","type":"int","level":"advanced","flags":0,"default_value":"4","min":"","max":"","enum_allowed":[],"desc":"Number of asynchronous cloner threads","long_desc":"","tags":[],"see_also":[]},"periodic_async_work":{"name":"periodic_async_work","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Periodically check for async work","long_desc":"","tags":[],"see_also":[]},"snapshot_clone_delay":{"name":"snapshot_clone_delay","type":"int","level":"advanced","flags":0,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"Delay clone begin operation by snapshot_clone_delay seconds","long_desc":"","tags":[],"see_also":[]},"snapshot_clone_no_wait":{"name":"snapshot_clone_no_wait","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"Reject subvolume clone request when cloner threads are busy","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"zabbix","can_run":true,"error_string":"","module_options":{"discovery_interval":{"name":"discovery_interval","type":"uint","level":"advanced","flags":0,"default_value":"100","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"identifier":{"name":"identifier","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"interval":{"name":"interval","type":"secs","level":"advanced","flags":0,"default_value":"60","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"zabbix_host":{"name":"zabbix_host","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"zabbix_port":{"name":"zabbix_port","type":"int","level":"advanced","flags":0,"default_value":"10051","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"zabbix_sender":{"name":"zabbix_sender","type":"str","level":"advanced","flags":0,"default_value":"/usr/bin/zabbix_sender","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}}]}],"modules":["cephadm","dashboard","iostat","nfs","restful"],"available_modules":[{"name":"alerts","can_run":true,"error_string":"","module_options":{"interval":{"name":"interval","type":"secs","level":"advanced","flags":1,"default_value":"60","min":"","max":"","enum_allowed":[],"desc":"How frequently to reexamine health status","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"smtp_destination":{"name":"smtp_destination","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"Email address to send alerts to","long_desc":"","tags":[],"see_also":[]},"smtp_from_name":{"name":"smtp_from_name","type":"str","level":"advanced","flags":1,"default_value":"Ceph","min":"","max":"","enum_allowed":[],"desc":"Email From: name","long_desc":"","tags":[],"see_also":[]},"smtp_host":{"name":"smtp_host","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"SMTP server","long_desc":"","tags":[],"see_also":[]},"smtp_password":{"name":"smtp_password","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"Password to authenticate with","long_desc":"","tags":[],"see_also":[]},"smtp_port":{"name":"smtp_port","type":"int","level":"advanced","flags":1,"default_value":"465","min":"","max":"","enum_allowed":[],"desc":"SMTP port","long_desc":"","tags":[],"see_also":[]},"smtp_sender":{"name":"smtp_sender","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"SMTP envelope sender","long_desc":"","tags":[],"see_also":[]},"smtp_ssl":{"name":"smtp_ssl","type":"bool","level":"advanced","flags":1,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"Use SSL to connect to SMTP server","long_desc":"","tags":[],"see_also":[]},"smtp_user":{"name":"smtp_user","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"User to authenticate as","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"balancer","can_run":true,"error_string":"","module_options":{"active":{"name":"active","type":"bool","level":"advanced","flags":1,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"automatically balance PGs across cluster","long_desc":"","tags":[],"see_also":[]},"begin_time":{"name":"begin_time","type":"str","level":"advanced","flags":1,"default_value":"0000","min":"","max":"","enum_allowed":[],"desc":"beginning time of day to automatically balance","long_desc":"This is a time of day in the format HHMM.","tags":[],"see_also":[]},"begin_weekday":{"name":"begin_weekday","type":"uint","level":"advanced","flags":1,"default_value":"0","min":"0","max":"6","enum_allowed":[],"desc":"Restrict automatic balancing to this day of the week or later","long_desc":"0 = Sunday, 1 = Monday, etc.","tags":[],"see_also":[]},"crush_compat_max_iterations":{"name":"crush_compat_max_iterations","type":"uint","level":"advanced","flags":1,"default_value":"25","min":"1","max":"250","enum_allowed":[],"desc":"maximum number of iterations to attempt optimization","long_desc":"","tags":[],"see_also":[]},"crush_compat_metrics":{"name":"crush_compat_metrics","type":"str","level":"advanced","flags":1,"default_value":"pgs,objects,bytes","min":"","max":"","enum_allowed":[],"desc":"metrics with which to calculate OSD utilization","long_desc":"Value is a list of one or more of \"pgs\", \"objects\", or \"bytes\", and indicates which metrics to use to balance utilization.","tags":[],"see_also":[]},"crush_compat_step":{"name":"crush_compat_step","type":"float","level":"advanced","flags":1,"default_value":"0.5","min":"0.001","max":"0.999","enum_allowed":[],"desc":"aggressiveness of optimization","long_desc":".99 is very aggressive, .01 is less aggressive","tags":[],"see_also":[]},"end_time":{"name":"end_time","type":"str","level":"advanced","flags":1,"default_value":"2359","min":"","max":"","enum_allowed":[],"desc":"ending time of day to automatically balance","long_desc":"This is a time of day in the format HHMM.","tags":[],"see_also":[]},"end_weekday":{"name":"end_weekday","type":"uint","level":"advanced","flags":1,"default_value":"0","min":"0","max":"6","enum_allowed":[],"desc":"Restrict automatic balancing to days of the week earlier than this","long_desc":"0 = Sunday, 1 = Monday, etc.","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"min_score":{"name":"min_score","type":"float","level":"advanced","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"minimum score, below which no optimization is attempted","long_desc":"","tags":[],"see_also":[]},"mode":{"name":"mode","type":"str","level":"advanced","flags":1,"default_value":"upmap","min":"","max":"","enum_allowed":["crush-compat","none","read","upmap","upmap-read"],"desc":"Balancer mode","long_desc":"","tags":[],"see_also":[]},"pool_ids":{"name":"pool_ids","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"pools which the automatic balancing will be limited to","long_desc":"","tags":[],"see_also":[]},"sleep_interval":{"name":"sleep_interval","type":"secs","level":"advanced","flags":1,"default_value":"60","min":"","max":"","enum_allowed":[],"desc":"how frequently to wake up and attempt optimization","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"update_pg_upmap_activity":{"name":"update_pg_upmap_activity","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Updates pg_upmap activity stats to be used in `balancer status detail`","long_desc":"","tags":[],"see_also":[]},"upmap_max_deviation":{"name":"upmap_max_deviation","type":"int","level":"advanced","flags":1,"default_value":"5","min":"1","max":"","enum_allowed":[],"desc":"deviation below which no optimization is attempted","long_desc":"If the number of PGs are within this count then no optimization is attempted","tags":[],"see_also":[]},"upmap_max_optimizations":{"name":"upmap_max_optimizations","type":"uint","level":"advanced","flags":1,"default_value":"10","min":"","max":"","enum_allowed":[],"desc":"maximum upmap optimizations to make per attempt","long_desc":"","tags":[],"see_also":[]}}},{"name":"cephadm","can_run":true,"error_string":"","module_options":{"agent_down_multiplier":{"name":"agent_down_multiplier","type":"float","level":"advanced","flags":0,"default_value":"3.0","min":"","max":"","enum_allowed":[],"desc":"Multiplied by agent refresh rate to calculate how long agent must not report before being marked down","long_desc":"","tags":[],"see_also":[]},"agent_refresh_rate":{"name":"agent_refresh_rate","type":"secs","level":"advanced","flags":0,"default_value":"20","min":"","max":"","enum_allowed":[],"desc":"How often agent on each host will try to gather and send metadata","long_desc":"","tags":[],"see_also":[]},"agent_starting_port":{"name":"agent_starting_port","type":"int","level":"advanced","flags":0,"default_value":"4721","min":"","max":"","enum_allowed":[],"desc":"First port agent will try to bind to (will also try up to next 1000 subsequent ports if blocked)","long_desc":"","tags":[],"see_also":[]},"allow_ptrace":{"name":"allow_ptrace","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"allow SYS_PTRACE capability on ceph containers","long_desc":"The SYS_PTRACE capability is needed to attach to a process with gdb or strace. Enabling this options can allow debugging daemons that encounter problems at runtime.","tags":[],"see_also":[]},"autotune_interval":{"name":"autotune_interval","type":"secs","level":"advanced","flags":0,"default_value":"600","min":"","max":"","enum_allowed":[],"desc":"how frequently to autotune daemon memory","long_desc":"","tags":[],"see_also":[]},"autotune_memory_target_ratio":{"name":"autotune_memory_target_ratio","type":"float","level":"advanced","flags":0,"default_value":"0.7","min":"","max":"","enum_allowed":[],"desc":"ratio of total system memory to divide amongst autotuned daemons","long_desc":"","tags":[],"see_also":[]},"cephadm_log_destination":{"name":"cephadm_log_destination","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":["file","file,syslog","syslog"],"desc":"Destination for cephadm command's persistent logging","long_desc":"","tags":[],"see_also":[]},"cgroups_split":{"name":"cgroups_split","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"Pass --cgroups=split when cephadm creates containers (currently podman only)","long_desc":"","tags":[],"see_also":[]},"config_checks_enabled":{"name":"config_checks_enabled","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Enable or disable the cephadm configuration analysis","long_desc":"","tags":[],"see_also":[]},"config_dashboard":{"name":"config_dashboard","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"manage configs like API endpoints in Dashboard.","long_desc":"","tags":[],"see_also":[]},"container_image_alertmanager":{"name":"container_image_alertmanager","type":"str","level":"advanced","flags":0,"default_value":"quay.io/prometheus/alertmanager:v0.25.0","min":"","max":"","enum_allowed":[],"desc":"Prometheus container image","long_desc":"","tags":[],"see_also":[]},"container_image_base":{"name":"container_image_base","type":"str","level":"advanced","flags":1,"default_value":"quay.io/ceph/ceph","min":"","max":"","enum_allowed":[],"desc":"Container image name, without the tag","long_desc":"","tags":[],"see_also":[]},"container_image_elasticsearch":{"name":"container_image_elasticsearch","type":"str","level":"advanced","flags":0,"default_value":"quay.io/omrizeneva/elasticsearch:6.8.23","min":"","max":"","enum_allowed":[],"desc":"elasticsearch container image","long_desc":"","tags":[],"see_also":[]},"container_image_grafana":{"name":"container_image_grafana","type":"str","level":"advanced","flags":0,"default_value":"quay.io/ceph/grafana:10.4.0","min":"","max":"","enum_allowed":[],"desc":"Prometheus container image","long_desc":"","tags":[],"see_also":[]},"container_image_haproxy":{"name":"container_image_haproxy","type":"str","level":"advanced","flags":0,"default_value":"quay.io/ceph/haproxy:2.3","min":"","max":"","enum_allowed":[],"desc":"HAproxy container image","long_desc":"","tags":[],"see_also":[]},"container_image_jaeger_agent":{"name":"container_image_jaeger_agent","type":"str","level":"advanced","flags":0,"default_value":"quay.io/jaegertracing/jaeger-agent:1.29","min":"","max":"","enum_allowed":[],"desc":"Jaeger agent container image","long_desc":"","tags":[],"see_also":[]},"container_image_jaeger_collector":{"name":"container_image_jaeger_collector","type":"str","level":"advanced","flags":0,"default_value":"quay.io/jaegertracing/jaeger-collector:1.29","min":"","max":"","enum_allowed":[],"desc":"Jaeger collector container image","long_desc":"","tags":[],"see_also":[]},"container_image_jaeger_query":{"name":"container_image_jaeger_query","type":"str","level":"advanced","flags":0,"default_value":"quay.io/jaegertracing/jaeger-query:1.29","min":"","max":"","enum_allowed":[],"desc":"Jaeger query container image","long_desc":"","tags":[],"see_also":[]},"container_image_keepalived":{"name":"container_image_keepalived","type":"str","level":"advanced","flags":0,"default_value":"quay.io/ceph/keepalived:2.2.4","min":"","max":"","enum_allowed":[],"desc":"Keepalived container image","long_desc":"","tags":[],"see_also":[]},"container_image_loki":{"name":"container_image_loki","type":"str","level":"advanced","flags":0,"default_value":"quay.io/ceph/loki:3.0.0","min":"","max":"","enum_allowed":[],"desc":"Loki container image","long_desc":"","tags":[],"see_also":[]},"container_image_node_exporter":{"name":"container_image_node_exporter","type":"str","level":"advanced","flags":0,"default_value":"quay.io/prometheus/node-exporter:v1.7.0","min":"","max":"","enum_allowed":[],"desc":"Prometheus container image","long_desc":"","tags":[],"see_also":[]},"container_image_nvmeof":{"name":"container_image_nvmeof","type":"str","level":"advanced","flags":0,"default_value":"quay.io/ceph/nvmeof:1.2.5","min":"","max":"","enum_allowed":[],"desc":"Nvme-of container image","long_desc":"","tags":[],"see_also":[]},"container_image_prometheus":{"name":"container_image_prometheus","type":"str","level":"advanced","flags":0,"default_value":"quay.io/prometheus/prometheus:v2.51.0","min":"","max":"","enum_allowed":[],"desc":"Prometheus container image","long_desc":"","tags":[],"see_also":[]},"container_image_promtail":{"name":"container_image_promtail","type":"str","level":"advanced","flags":0,"default_value":"quay.io/ceph/promtail:3.0.0","min":"","max":"","enum_allowed":[],"desc":"Promtail container image","long_desc":"","tags":[],"see_also":[]},"container_image_samba":{"name":"container_image_samba","type":"str","level":"advanced","flags":0,"default_value":"quay.io/samba.org/samba-server:devbuilds-centos-amd64","min":"","max":"","enum_allowed":[],"desc":"Samba/SMB container image","long_desc":"","tags":[],"see_also":[]},"container_image_snmp_gateway":{"name":"container_image_snmp_gateway","type":"str","level":"advanced","flags":0,"default_value":"quay.io/ceph/snmp-notifier:v1.2.1","min":"","max":"","enum_allowed":[],"desc":"SNMP Gateway container image","long_desc":"","tags":[],"see_also":[]},"container_init":{"name":"container_init","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"Run podman/docker with `--init`","long_desc":"","tags":[],"see_also":[]},"daemon_cache_timeout":{"name":"daemon_cache_timeout","type":"secs","level":"advanced","flags":0,"default_value":"600","min":"","max":"","enum_allowed":[],"desc":"seconds to cache service (daemon) inventory","long_desc":"","tags":[],"see_also":[]},"default_cephadm_command_timeout":{"name":"default_cephadm_command_timeout","type":"int","level":"advanced","flags":0,"default_value":"900","min":"","max":"","enum_allowed":[],"desc":"Default timeout applied to cephadm commands run directly on the host (in seconds)","long_desc":"","tags":[],"see_also":[]},"default_registry":{"name":"default_registry","type":"str","level":"advanced","flags":0,"default_value":"quay.io","min":"","max":"","enum_allowed":[],"desc":"Search-registry to which we should normalize unqualified image names. This is not the default registry","long_desc":"","tags":[],"see_also":[]},"device_cache_timeout":{"name":"device_cache_timeout","type":"secs","level":"advanced","flags":0,"default_value":"1800","min":"","max":"","enum_allowed":[],"desc":"seconds to cache device inventory","long_desc":"","tags":[],"see_also":[]},"device_enhanced_scan":{"name":"device_enhanced_scan","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Use libstoragemgmt during device scans","long_desc":"","tags":[],"see_also":[]},"facts_cache_timeout":{"name":"facts_cache_timeout","type":"secs","level":"advanced","flags":0,"default_value":"60","min":"","max":"","enum_allowed":[],"desc":"seconds to cache host facts data","long_desc":"","tags":[],"see_also":[]},"grafana_dashboards_path":{"name":"grafana_dashboards_path","type":"str","level":"advanced","flags":0,"default_value":"/etc/grafana/dashboards/ceph-dashboard/","min":"","max":"","enum_allowed":[],"desc":"location of dashboards to include in grafana deployments","long_desc":"","tags":[],"see_also":[]},"host_check_interval":{"name":"host_check_interval","type":"secs","level":"advanced","flags":0,"default_value":"600","min":"","max":"","enum_allowed":[],"desc":"how frequently to perform a host check","long_desc":"","tags":[],"see_also":[]},"hw_monitoring":{"name":"hw_monitoring","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Deploy hw monitoring daemon on every host.","long_desc":"","tags":[],"see_also":[]},"inventory_list_all":{"name":"inventory_list_all","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Whether ceph-volume inventory should report more devices (mostly mappers (LVs / mpaths), partitions...)","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_refresh_metadata":{"name":"log_refresh_metadata","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Log all refresh metadata. Includes daemon, device, and host info collected regularly. Only has effect if logging at debug level","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"log to the \"cephadm\" cluster log channel\"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"manage_etc_ceph_ceph_conf":{"name":"manage_etc_ceph_ceph_conf","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Manage and own /etc/ceph/ceph.conf on the hosts.","long_desc":"","tags":[],"see_also":[]},"manage_etc_ceph_ceph_conf_hosts":{"name":"manage_etc_ceph_ceph_conf_hosts","type":"str","level":"advanced","flags":0,"default_value":"*","min":"","max":"","enum_allowed":[],"desc":"PlacementSpec describing on which hosts to manage /etc/ceph/ceph.conf","long_desc":"","tags":[],"see_also":[]},"max_count_per_host":{"name":"max_count_per_host","type":"int","level":"advanced","flags":0,"default_value":"10","min":"","max":"","enum_allowed":[],"desc":"max number of daemons per service per host","long_desc":"","tags":[],"see_also":[]},"max_osd_draining_count":{"name":"max_osd_draining_count","type":"int","level":"advanced","flags":0,"default_value":"10","min":"","max":"","enum_allowed":[],"desc":"max number of osds that will be drained simultaneously when osds are removed","long_desc":"","tags":[],"see_also":[]},"migration_current":{"name":"migration_current","type":"int","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"internal - do not modify","long_desc":"","tags":[],"see_also":[]},"mode":{"name":"mode","type":"str","level":"advanced","flags":0,"default_value":"root","min":"","max":"","enum_allowed":["cephadm-package","root"],"desc":"mode for remote execution of cephadm","long_desc":"","tags":[],"see_also":[]},"oob_default_addr":{"name":"oob_default_addr","type":"str","level":"advanced","flags":0,"default_value":"169.254.1.1","min":"","max":"","enum_allowed":[],"desc":"Default address for RedFish API (oob management).","long_desc":"","tags":[],"see_also":[]},"prometheus_alerts_path":{"name":"prometheus_alerts_path","type":"str","level":"advanced","flags":0,"default_value":"/etc/prometheus/ceph/ceph_default_alerts.yml","min":"","max":"","enum_allowed":[],"desc":"location of alerts to include in prometheus deployments","long_desc":"","tags":[],"see_also":[]},"registry_insecure":{"name":"registry_insecure","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Registry is to be considered insecure (no TLS available). Only for development purposes.","long_desc":"","tags":[],"see_also":[]},"registry_password":{"name":"registry_password","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"Custom repository password. Only used for logging into a registry.","long_desc":"","tags":[],"see_also":[]},"registry_url":{"name":"registry_url","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"Registry url for login purposes. This is not the default registry","long_desc":"","tags":[],"see_also":[]},"registry_username":{"name":"registry_username","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"Custom repository username. Only used for logging into a registry.","long_desc":"","tags":[],"see_also":[]},"secure_monitoring_stack":{"name":"secure_monitoring_stack","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Enable TLS security for all the monitoring stack daemons","long_desc":"","tags":[],"see_also":[]},"service_discovery_port":{"name":"service_discovery_port","type":"int","level":"advanced","flags":0,"default_value":"8765","min":"","max":"","enum_allowed":[],"desc":"cephadm service discovery port","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"ssh_config_file":{"name":"ssh_config_file","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"customized SSH config file to connect to managed hosts","long_desc":"","tags":[],"see_also":[]},"ssh_keepalive_count_max":{"name":"ssh_keepalive_count_max","type":"int","level":"advanced","flags":0,"default_value":"3","min":"","max":"","enum_allowed":[],"desc":"How many times ssh connections can fail liveness checks before the host is marked offline","long_desc":"","tags":[],"see_also":[]},"ssh_keepalive_interval":{"name":"ssh_keepalive_interval","type":"int","level":"advanced","flags":0,"default_value":"7","min":"","max":"","enum_allowed":[],"desc":"How often ssh connections are checked for liveness","long_desc":"","tags":[],"see_also":[]},"use_agent":{"name":"use_agent","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Use cephadm agent on each host to gather and send metadata","long_desc":"","tags":[],"see_also":[]},"use_repo_digest":{"name":"use_repo_digest","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"Automatically convert image tags to image digest. Make sure all daemons use the same image","long_desc":"","tags":[],"see_also":[]},"warn_on_failed_host_check":{"name":"warn_on_failed_host_check","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"raise a health warning if the host check fails","long_desc":"","tags":[],"see_also":[]},"warn_on_stray_daemons":{"name":"warn_on_stray_daemons","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"raise a health warning if daemons are detected that are not managed by cephadm","long_desc":"","tags":[],"see_also":[]},"warn_on_stray_hosts":{"name":"warn_on_stray_hosts","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"raise a health warning if daemons are detected on a host that is not managed by cephadm","long_desc":"","tags":[],"see_also":[]}}},{"name":"crash","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"retain_interval":{"name":"retain_interval","type":"secs","level":"advanced","flags":1,"default_value":"31536000","min":"","max":"","enum_allowed":[],"desc":"how long to retain crashes before pruning them","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"warn_recent_interval":{"name":"warn_recent_interval","type":"secs","level":"advanced","flags":1,"default_value":"1209600","min":"","max":"","enum_allowed":[],"desc":"time interval in which to warn about recent crashes","long_desc":"","tags":[],"see_also":[]}}},{"name":"dashboard","can_run":true,"error_string":"","module_options":{"ACCOUNT_LOCKOUT_ATTEMPTS":{"name":"ACCOUNT_LOCKOUT_ATTEMPTS","type":"int","level":"advanced","flags":0,"default_value":"10","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"ALERTMANAGER_API_HOST":{"name":"ALERTMANAGER_API_HOST","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"ALERTMANAGER_API_SSL_VERIFY":{"name":"ALERTMANAGER_API_SSL_VERIFY","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"AUDIT_API_ENABLED":{"name":"AUDIT_API_ENABLED","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"AUDIT_API_LOG_PAYLOAD":{"name":"AUDIT_API_LOG_PAYLOAD","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"ENABLE_BROWSABLE_API":{"name":"ENABLE_BROWSABLE_API","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"FEATURE_TOGGLE_CEPHFS":{"name":"FEATURE_TOGGLE_CEPHFS","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"FEATURE_TOGGLE_DASHBOARD":{"name":"FEATURE_TOGGLE_DASHBOARD","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"FEATURE_TOGGLE_ISCSI":{"name":"FEATURE_TOGGLE_ISCSI","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"FEATURE_TOGGLE_MIRRORING":{"name":"FEATURE_TOGGLE_MIRRORING","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"FEATURE_TOGGLE_NFS":{"name":"FEATURE_TOGGLE_NFS","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"FEATURE_TOGGLE_RBD":{"name":"FEATURE_TOGGLE_RBD","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"FEATURE_TOGGLE_RGW":{"name":"FEATURE_TOGGLE_RGW","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"GANESHA_CLUSTERS_RADOS_POOL_NAMESPACE":{"name":"GANESHA_CLUSTERS_RADOS_POOL_NAMESPACE","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"GRAFANA_API_PASSWORD":{"name":"GRAFANA_API_PASSWORD","type":"str","level":"advanced","flags":0,"default_value":"admin","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"GRAFANA_API_SSL_VERIFY":{"name":"GRAFANA_API_SSL_VERIFY","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"GRAFANA_API_URL":{"name":"GRAFANA_API_URL","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"GRAFANA_API_USERNAME":{"name":"GRAFANA_API_USERNAME","type":"str","level":"advanced","flags":0,"default_value":"admin","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"GRAFANA_FRONTEND_API_URL":{"name":"GRAFANA_FRONTEND_API_URL","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"GRAFANA_UPDATE_DASHBOARDS":{"name":"GRAFANA_UPDATE_DASHBOARDS","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"ISCSI_API_SSL_VERIFICATION":{"name":"ISCSI_API_SSL_VERIFICATION","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"ISSUE_TRACKER_API_KEY":{"name":"ISSUE_TRACKER_API_KEY","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PROMETHEUS_API_HOST":{"name":"PROMETHEUS_API_HOST","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PROMETHEUS_API_SSL_VERIFY":{"name":"PROMETHEUS_API_SSL_VERIFY","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PWD_POLICY_CHECK_COMPLEXITY_ENABLED":{"name":"PWD_POLICY_CHECK_COMPLEXITY_ENABLED","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PWD_POLICY_CHECK_EXCLUSION_LIST_ENABLED":{"name":"PWD_POLICY_CHECK_EXCLUSION_LIST_ENABLED","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PWD_POLICY_CHECK_LENGTH_ENABLED":{"name":"PWD_POLICY_CHECK_LENGTH_ENABLED","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PWD_POLICY_CHECK_OLDPWD_ENABLED":{"name":"PWD_POLICY_CHECK_OLDPWD_ENABLED","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PWD_POLICY_CHECK_REPETITIVE_CHARS_ENABLED":{"name":"PWD_POLICY_CHECK_REPETITIVE_CHARS_ENABLED","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PWD_POLICY_CHECK_SEQUENTIAL_CHARS_ENABLED":{"name":"PWD_POLICY_CHECK_SEQUENTIAL_CHARS_ENABLED","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PWD_POLICY_CHECK_USERNAME_ENABLED":{"name":"PWD_POLICY_CHECK_USERNAME_ENABLED","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PWD_POLICY_ENABLED":{"name":"PWD_POLICY_ENABLED","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PWD_POLICY_EXCLUSION_LIST":{"name":"PWD_POLICY_EXCLUSION_LIST","type":"str","level":"advanced","flags":0,"default_value":"osd,host,dashboard,pool,block,nfs,ceph,monitors,gateway,logs,crush,maps","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PWD_POLICY_MIN_COMPLEXITY":{"name":"PWD_POLICY_MIN_COMPLEXITY","type":"int","level":"advanced","flags":0,"default_value":"10","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PWD_POLICY_MIN_LENGTH":{"name":"PWD_POLICY_MIN_LENGTH","type":"int","level":"advanced","flags":0,"default_value":"8","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"REST_REQUESTS_TIMEOUT":{"name":"REST_REQUESTS_TIMEOUT","type":"int","level":"advanced","flags":0,"default_value":"45","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"RGW_API_ACCESS_KEY":{"name":"RGW_API_ACCESS_KEY","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"RGW_API_ADMIN_RESOURCE":{"name":"RGW_API_ADMIN_RESOURCE","type":"str","level":"advanced","flags":0,"default_value":"admin","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"RGW_API_SECRET_KEY":{"name":"RGW_API_SECRET_KEY","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"RGW_API_SSL_VERIFY":{"name":"RGW_API_SSL_VERIFY","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"USER_PWD_EXPIRATION_SPAN":{"name":"USER_PWD_EXPIRATION_SPAN","type":"int","level":"advanced","flags":0,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"USER_PWD_EXPIRATION_WARNING_1":{"name":"USER_PWD_EXPIRATION_WARNING_1","type":"int","level":"advanced","flags":0,"default_value":"10","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"USER_PWD_EXPIRATION_WARNING_2":{"name":"USER_PWD_EXPIRATION_WARNING_2","type":"int","level":"advanced","flags":0,"default_value":"5","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"cross_origin_url":{"name":"cross_origin_url","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"crt_file":{"name":"crt_file","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"debug":{"name":"debug","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Enable/disable debug options","long_desc":"","tags":[],"see_also":[]},"jwt_token_ttl":{"name":"jwt_token_ttl","type":"int","level":"advanced","flags":0,"default_value":"28800","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"key_file":{"name":"key_file","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"motd":{"name":"motd","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"The message of the day","long_desc":"","tags":[],"see_also":[]},"redirect_resolve_ip_addr":{"name":"redirect_resolve_ip_addr","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"server_addr":{"name":"server_addr","type":"str","level":"advanced","flags":0,"default_value":"::","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"server_port":{"name":"server_port","type":"int","level":"advanced","flags":0,"default_value":"8080","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"ssl":{"name":"ssl","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"ssl_server_port":{"name":"ssl_server_port","type":"int","level":"advanced","flags":0,"default_value":"8443","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"standby_behaviour":{"name":"standby_behaviour","type":"str","level":"advanced","flags":0,"default_value":"redirect","min":"","max":"","enum_allowed":["error","redirect"],"desc":"","long_desc":"","tags":[],"see_also":[]},"standby_error_status_code":{"name":"standby_error_status_code","type":"int","level":"advanced","flags":0,"default_value":"500","min":"400","max":"599","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"url_prefix":{"name":"url_prefix","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"devicehealth","can_run":true,"error_string":"","module_options":{"enable_monitoring":{"name":"enable_monitoring","type":"bool","level":"advanced","flags":1,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"monitor device health metrics","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"mark_out_threshold":{"name":"mark_out_threshold","type":"secs","level":"advanced","flags":1,"default_value":"2419200","min":"","max":"","enum_allowed":[],"desc":"automatically mark OSD if it may fail before this long","long_desc":"","tags":[],"see_also":[]},"pool_name":{"name":"pool_name","type":"str","level":"advanced","flags":1,"default_value":"device_health_metrics","min":"","max":"","enum_allowed":[],"desc":"name of pool in which to store device health metrics","long_desc":"","tags":[],"see_also":[]},"retention_period":{"name":"retention_period","type":"secs","level":"advanced","flags":1,"default_value":"15552000","min":"","max":"","enum_allowed":[],"desc":"how long to retain device health metrics","long_desc":"","tags":[],"see_also":[]},"scrape_frequency":{"name":"scrape_frequency","type":"secs","level":"advanced","flags":1,"default_value":"86400","min":"","max":"","enum_allowed":[],"desc":"how frequently to scrape device health metrics","long_desc":"","tags":[],"see_also":[]},"self_heal":{"name":"self_heal","type":"bool","level":"advanced","flags":1,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"preemptively heal cluster around devices that may fail","long_desc":"","tags":[],"see_also":[]},"sleep_interval":{"name":"sleep_interval","type":"secs","level":"advanced","flags":1,"default_value":"600","min":"","max":"","enum_allowed":[],"desc":"how frequently to wake up and check device health","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"warn_threshold":{"name":"warn_threshold","type":"secs","level":"advanced","flags":1,"default_value":"7257600","min":"","max":"","enum_allowed":[],"desc":"raise health warning if OSD may fail before this long","long_desc":"","tags":[],"see_also":[]}}},{"name":"influx","can_run":false,"error_string":"influxdb python module not found","module_options":{"batch_size":{"name":"batch_size","type":"int","level":"advanced","flags":0,"default_value":"5000","min":"","max":"","enum_allowed":[],"desc":"How big batches of data points should be when sending to InfluxDB.","long_desc":"","tags":[],"see_also":[]},"database":{"name":"database","type":"str","level":"advanced","flags":0,"default_value":"ceph","min":"","max":"","enum_allowed":[],"desc":"InfluxDB database name. You will need to create this database and grant write privileges to the configured username or the username must have admin privileges to create it.","long_desc":"","tags":[],"see_also":[]},"hostname":{"name":"hostname","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"InfluxDB server hostname","long_desc":"","tags":[],"see_also":[]},"interval":{"name":"interval","type":"secs","level":"advanced","flags":0,"default_value":"30","min":"5","max":"","enum_allowed":[],"desc":"Time between reports to InfluxDB. Default 30 seconds.","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"password":{"name":"password","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"password of InfluxDB server user","long_desc":"","tags":[],"see_also":[]},"port":{"name":"port","type":"int","level":"advanced","flags":0,"default_value":"8086","min":"","max":"","enum_allowed":[],"desc":"InfluxDB server port","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"ssl":{"name":"ssl","type":"str","level":"advanced","flags":0,"default_value":"false","min":"","max":"","enum_allowed":[],"desc":"Use https connection for InfluxDB server. Use \"true\" or \"false\".","long_desc":"","tags":[],"see_also":[]},"threads":{"name":"threads","type":"int","level":"advanced","flags":0,"default_value":"5","min":"1","max":"32","enum_allowed":[],"desc":"How many worker threads should be spawned for sending data to InfluxDB.","long_desc":"","tags":[],"see_also":[]},"username":{"name":"username","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"username of InfluxDB server user","long_desc":"","tags":[],"see_also":[]},"verify_ssl":{"name":"verify_ssl","type":"str","level":"advanced","flags":0,"default_value":"true","min":"","max":"","enum_allowed":[],"desc":"Verify https cert for InfluxDB server. Use \"true\" or \"false\".","long_desc":"","tags":[],"see_also":[]}}},{"name":"insights","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"iostat","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"k8sevents","can_run":true,"error_string":"","module_options":{"ceph_event_retention_days":{"name":"ceph_event_retention_days","type":"int","level":"advanced","flags":0,"default_value":"7","min":"","max":"","enum_allowed":[],"desc":"Days to hold ceph event information within local cache","long_desc":"","tags":[],"see_also":[]},"config_check_secs":{"name":"config_check_secs","type":"int","level":"advanced","flags":0,"default_value":"10","min":"10","max":"","enum_allowed":[],"desc":"interval (secs) to check for cluster configuration changes","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"localpool","can_run":true,"error_string":"","module_options":{"failure_domain":{"name":"failure_domain","type":"str","level":"advanced","flags":1,"default_value":"host","min":"","max":"","enum_allowed":[],"desc":"failure domain for any created local pool","long_desc":"what failure domain we should separate data replicas across.","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"min_size":{"name":"min_size","type":"int","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"default min_size for any created local pool","long_desc":"value to set min_size to (unchanged from Ceph's default if this option is not set)","tags":[],"see_also":[]},"num_rep":{"name":"num_rep","type":"int","level":"advanced","flags":1,"default_value":"3","min":"","max":"","enum_allowed":[],"desc":"default replica count for any created local pool","long_desc":"","tags":[],"see_also":[]},"pg_num":{"name":"pg_num","type":"int","level":"advanced","flags":1,"default_value":"128","min":"","max":"","enum_allowed":[],"desc":"default pg_num for any created local pool","long_desc":"","tags":[],"see_also":[]},"prefix":{"name":"prefix","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"name prefix for any created local pool","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"subtree":{"name":"subtree","type":"str","level":"advanced","flags":1,"default_value":"rack","min":"","max":"","enum_allowed":[],"desc":"CRUSH level for which to create a local pool","long_desc":"which CRUSH subtree type the module should create a pool for.","tags":[],"see_also":[]}}},{"name":"mds_autoscaler","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"mirroring","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"nfs","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"orchestrator","can_run":true,"error_string":"","module_options":{"fail_fs":{"name":"fail_fs","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Fail filesystem for rapid multi-rank mds upgrade","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"orchestrator":{"name":"orchestrator","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["cephadm","rook","test_orchestrator"],"desc":"Orchestrator backend","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"osd_perf_query","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"osd_support","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"pg_autoscaler","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sleep_interval":{"name":"sleep_interval","type":"secs","level":"advanced","flags":0,"default_value":"60","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"threshold":{"name":"threshold","type":"float","level":"advanced","flags":0,"default_value":"3.0","min":"1.0","max":"","enum_allowed":[],"desc":"scaling threshold","long_desc":"The factor by which the `NEW PG_NUM` must vary from the current`PG_NUM` before being accepted. Cannot be less than 1.0","tags":[],"see_also":[]}}},{"name":"progress","can_run":true,"error_string":"","module_options":{"allow_pg_recovery_event":{"name":"allow_pg_recovery_event","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"allow the module to show pg recovery progress","long_desc":"","tags":[],"see_also":[]},"enabled":{"name":"enabled","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"max_completed_events":{"name":"max_completed_events","type":"int","level":"advanced","flags":1,"default_value":"50","min":"","max":"","enum_allowed":[],"desc":"number of past completed events to remember","long_desc":"","tags":[],"see_also":[]},"sleep_interval":{"name":"sleep_interval","type":"secs","level":"advanced","flags":1,"default_value":"5","min":"","max":"","enum_allowed":[],"desc":"how long the module is going to sleep","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"prometheus","can_run":true,"error_string":"","module_options":{"cache":{"name":"cache","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"exclude_perf_counters":{"name":"exclude_perf_counters","type":"bool","level":"advanced","flags":1,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"Do not include perf-counters in the metrics output","long_desc":"Gathering perf-counters from a single Prometheus exporter can degrade ceph-mgr performance, especially in large clusters. Instead, Ceph-exporter daemons are now used by default for perf-counter gathering. This should only be disabled when no ceph-exporters are deployed.","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"rbd_stats_pools":{"name":"rbd_stats_pools","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"rbd_stats_pools_refresh_interval":{"name":"rbd_stats_pools_refresh_interval","type":"int","level":"advanced","flags":0,"default_value":"300","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"scrape_interval":{"name":"scrape_interval","type":"float","level":"advanced","flags":0,"default_value":"15.0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"server_addr":{"name":"server_addr","type":"str","level":"advanced","flags":0,"default_value":"::","min":"","max":"","enum_allowed":[],"desc":"the IPv4 or IPv6 address on which the module listens for HTTP requests","long_desc":"","tags":[],"see_also":[]},"server_port":{"name":"server_port","type":"int","level":"advanced","flags":1,"default_value":"9283","min":"","max":"","enum_allowed":[],"desc":"the port on which the module listens for HTTP requests","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"stale_cache_strategy":{"name":"stale_cache_strategy","type":"str","level":"advanced","flags":0,"default_value":"log","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"standby_behaviour":{"name":"standby_behaviour","type":"str","level":"advanced","flags":1,"default_value":"default","min":"","max":"","enum_allowed":["default","error"],"desc":"","long_desc":"","tags":[],"see_also":[]},"standby_error_status_code":{"name":"standby_error_status_code","type":"int","level":"advanced","flags":1,"default_value":"500","min":"400","max":"599","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"rbd_support","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"max_concurrent_snap_create":{"name":"max_concurrent_snap_create","type":"int","level":"advanced","flags":0,"default_value":"10","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"mirror_snapshot_schedule":{"name":"mirror_snapshot_schedule","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"trash_purge_schedule":{"name":"trash_purge_schedule","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"restful","can_run":true,"error_string":"","module_options":{"enable_auth":{"name":"enable_auth","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"key_file":{"name":"key_file","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"max_requests":{"name":"max_requests","type":"int","level":"advanced","flags":0,"default_value":"500","min":"","max":"","enum_allowed":[],"desc":"Maximum number of requests to keep in memory. When new request comes in, the oldest request will be removed if the number of requests exceeds the max request number. if un-finished request is removed, error message will be logged in the ceph-mgr log.","long_desc":"","tags":[],"see_also":[]},"server_addr":{"name":"server_addr","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"server_port":{"name":"server_port","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"rgw","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"secondary_zone_period_retry_limit":{"name":"secondary_zone_period_retry_limit","type":"int","level":"advanced","flags":0,"default_value":"5","min":"","max":"","enum_allowed":[],"desc":"RGW module period update retry limit for secondary site","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"rook","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"storage_class":{"name":"storage_class","type":"str","level":"advanced","flags":0,"default_value":"local","min":"","max":"","enum_allowed":[],"desc":"storage class name for LSO-discovered PVs","long_desc":"","tags":[],"see_also":[]}}},{"name":"selftest","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"roption1":{"name":"roption1","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"roption2":{"name":"roption2","type":"str","level":"advanced","flags":0,"default_value":"xyz","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"rwoption1":{"name":"rwoption1","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"rwoption2":{"name":"rwoption2","type":"int","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"rwoption3":{"name":"rwoption3","type":"float","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"rwoption4":{"name":"rwoption4","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"rwoption5":{"name":"rwoption5","type":"bool","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"rwoption6":{"name":"rwoption6","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"rwoption7":{"name":"rwoption7","type":"int","level":"advanced","flags":0,"default_value":"","min":"1","max":"42","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"testkey":{"name":"testkey","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"testlkey":{"name":"testlkey","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"testnewline":{"name":"testnewline","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"snap_schedule","can_run":true,"error_string":"","module_options":{"allow_m_granularity":{"name":"allow_m_granularity","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"allow minute scheduled snapshots","long_desc":"","tags":[],"see_also":[]},"dump_on_update":{"name":"dump_on_update","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"dump database to debug log on update","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"stats","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"status","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"telegraf","can_run":true,"error_string":"","module_options":{"address":{"name":"address","type":"str","level":"advanced","flags":0,"default_value":"unixgram:///tmp/telegraf.sock","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"interval":{"name":"interval","type":"secs","level":"advanced","flags":0,"default_value":"15","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"telemetry","can_run":true,"error_string":"","module_options":{"channel_basic":{"name":"channel_basic","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"Share basic cluster information (size, version)","long_desc":"","tags":[],"see_also":[]},"channel_crash":{"name":"channel_crash","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"Share metadata about Ceph daemon crashes (version, stack straces, etc)","long_desc":"","tags":[],"see_also":[]},"channel_device":{"name":"channel_device","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"Share device health metrics (e.g., SMART data, minus potentially identifying info like serial numbers)","long_desc":"","tags":[],"see_also":[]},"channel_ident":{"name":"channel_ident","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Share a user-provided description and/or contact email for the cluster","long_desc":"","tags":[],"see_also":[]},"channel_perf":{"name":"channel_perf","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Share various performance metrics of a cluster","long_desc":"","tags":[],"see_also":[]},"contact":{"name":"contact","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"description":{"name":"description","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"device_url":{"name":"device_url","type":"str","level":"advanced","flags":0,"default_value":"https://telemetry.ceph.com/device","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"enabled":{"name":"enabled","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"interval":{"name":"interval","type":"int","level":"advanced","flags":0,"default_value":"24","min":"8","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"last_opt_revision":{"name":"last_opt_revision","type":"int","level":"advanced","flags":0,"default_value":"1","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"leaderboard":{"name":"leaderboard","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"leaderboard_description":{"name":"leaderboard_description","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"organization":{"name":"organization","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"proxy":{"name":"proxy","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"url":{"name":"url","type":"str","level":"advanced","flags":0,"default_value":"https://telemetry.ceph.com/report","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"test_orchestrator","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"volumes","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"max_concurrent_clones":{"name":"max_concurrent_clones","type":"int","level":"advanced","flags":0,"default_value":"4","min":"","max":"","enum_allowed":[],"desc":"Number of asynchronous cloner threads","long_desc":"","tags":[],"see_also":[]},"periodic_async_work":{"name":"periodic_async_work","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Periodically check for async work","long_desc":"","tags":[],"see_also":[]},"snapshot_clone_delay":{"name":"snapshot_clone_delay","type":"int","level":"advanced","flags":0,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"Delay clone begin operation by snapshot_clone_delay seconds","long_desc":"","tags":[],"see_also":[]},"snapshot_clone_no_wait":{"name":"snapshot_clone_no_wait","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"Reject subvolume clone request when cloner threads are busy","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"zabbix","can_run":true,"error_string":"","module_options":{"discovery_interval":{"name":"discovery_interval","type":"uint","level":"advanced","flags":0,"default_value":"100","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"identifier":{"name":"identifier","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"interval":{"name":"interval","type":"secs","level":"advanced","flags":0,"default_value":"60","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"zabbix_host":{"name":"zabbix_host","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"zabbix_port":{"name":"zabbix_port","type":"int","level":"advanced","flags":0,"default_value":"10051","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"zabbix_sender":{"name":"zabbix_sender","type":"str","level":"advanced","flags":0,"default_value":"/usr/bin/zabbix_sender","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}}],"services":{"dashboard":"https://192.168.123.102:8443/"},"always_on_modules":{"octopus":["balancer","crash","devicehealth","orchestrator","pg_autoscaler","progress","rbd_support","status","telemetry","volumes"],"pacific":["balancer","crash","devicehealth","orchestrator","pg_autoscaler","progress","rbd_support","status","telemetry","volumes"],"quincy":["balancer","crash","devicehealth","orchestrator","pg_autoscaler","progress","rbd_support","status","telemetry","volumes"],"reef":["balancer","crash","devicehealth","orchestrator","pg_autoscaler","progress","rbd_support","status","telemetry","volumes"],"squid":["balancer","crash","devicehealth","orchestrator","pg_autoscaler","progress","rbd_support","status","telemetry","volumes"]},"force_disabled_modules":{},"last_failure_osd_epoch":3,"active_clients":[{"name":"libcephsqlite","addrvec":[{"type":"v2","addr":"192.168.123.102:0","nonce":1081699145}]},{"name":"rbd_support","addrvec":[{"type":"v2","addr":"192.168.123.102:0","nonce":3033973711}]},{"name":"volumes","addrvec":[{"type":"v2","addr":"192.168.123.102:0","nonce":2631827095}]}]} 2026-03-06T13:47:05.423 INFO:tasks.cephadm.ceph_manager.ceph:mgr available! 2026-03-06T13:47:05.423 INFO:tasks.cephadm.ceph_manager.ceph:waiting for all up 2026-03-06T13:47:05.424 DEBUG:teuthology.orchestra.run.vm02:> sudo /home/ubuntu/cephtest/cephadm --image harbor.clyso.com/custom-ceph/ceph/ceph:cobaltcore-storage-v19.2.3-fasttrack-3 shell --fsid 2b1b4a72-195a-11f1-b1df-bf1f800d676b -- ceph osd dump --format=json 2026-03-06T13:47:05.774 INFO:teuthology.orchestra.run.vm02.stderr:Inferring config /var/lib/ceph/2b1b4a72-195a-11f1-b1df-bf1f800d676b/mon.a/config 2026-03-06T13:47:06.010 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:47:05 vm06 ceph-mon[56528]: pgmap v61: 1 pgs: 1 unknown; 0 B data, 80 MiB used, 60 GiB / 60 GiB avail 2026-03-06T13:47:06.010 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:47:05 vm06 ceph-mon[56528]: Detected new or changed devices on vm06 2026-03-06T13:47:06.010 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:47:05 vm06 ceph-mon[56528]: Health check cleared: POOL_APP_NOT_ENABLED (was: 1 pool(s) do not have an application enabled) 2026-03-06T13:47:06.010 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:47:05 vm06 ceph-mon[56528]: Cluster is now healthy 2026-03-06T13:47:06.010 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:47:05 vm06 ceph-mon[56528]: from='client.? 192.168.123.102:0/2333212563' entity='client.admin' cmd=[{"prefix": "mgr dump", "format": "json"}]: dispatch 2026-03-06T13:47:06.010 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:47:05 vm06 ceph-mon[56528]: mgrmap e15: a(active, since 88s), standbys: b 2026-03-06T13:47:06.049 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:47:05 vm04 ceph-mon[53170]: pgmap v61: 1 pgs: 1 unknown; 0 B data, 80 MiB used, 60 GiB / 60 GiB avail 2026-03-06T13:47:06.049 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:47:05 vm04 ceph-mon[53170]: Detected new or changed devices on vm06 2026-03-06T13:47:06.049 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:47:05 vm04 ceph-mon[53170]: Health check cleared: POOL_APP_NOT_ENABLED (was: 1 pool(s) do not have an application enabled) 2026-03-06T13:47:06.049 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:47:05 vm04 ceph-mon[53170]: Cluster is now healthy 2026-03-06T13:47:06.050 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:47:05 vm04 ceph-mon[53170]: from='client.? 192.168.123.102:0/2333212563' entity='client.admin' cmd=[{"prefix": "mgr dump", "format": "json"}]: dispatch 2026-03-06T13:47:06.050 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:47:05 vm04 ceph-mon[53170]: mgrmap e15: a(active, since 88s), standbys: b 2026-03-06T13:47:06.110 INFO:teuthology.orchestra.run.vm02.stdout: 2026-03-06T13:47:06.111 INFO:teuthology.orchestra.run.vm02.stdout:{"epoch":21,"fsid":"2b1b4a72-195a-11f1-b1df-bf1f800d676b","created":"2026-03-06T12:44:49.710645+0000","modified":"2026-03-06T12:47:03.639935+0000","last_up_change":"2026-03-06T12:47:00.617832+0000","last_in_change":"2026-03-06T12:46:47.212348+0000","flags":"sortbitwise,recovery_deletes,purged_snapdirs,pglog_hardlimit","flags_num":5799936,"flags_set":["pglog_hardlimit","purged_snapdirs","recovery_deletes","sortbitwise"],"crush_version":8,"full_ratio":0.94999998807907104,"backfillfull_ratio":0.89999997615814209,"nearfull_ratio":0.85000002384185791,"cluster_snapshot":"","pool_max":1,"max_osd":3,"require_min_compat_client":"luminous","min_compat_client":"jewel","require_osd_release":"squid","allow_crimson":false,"pools":[{"pool":1,"pool_name":".mgr","create_time":"2026-03-06T12:47:00.638787+0000","flags":1,"flags_names":"hashpspool","type":1,"size":3,"min_size":2,"crush_rule":0,"peering_crush_bucket_count":0,"peering_crush_bucket_target":0,"peering_crush_bucket_barrier":0,"peering_crush_bucket_mandatory_member":2147483647,"is_stretch_pool":false,"object_hash":2,"pg_autoscale_mode":"off","pg_num":1,"pg_placement_num":1,"pg_placement_num_target":1,"pg_num_target":1,"pg_num_pending":1,"last_pg_merge_meta":{"source_pgid":"0.0","ready_epoch":0,"last_epoch_started":0,"last_epoch_clean":0,"source_version":"0'0","target_version":"0'0"},"last_change":"21","last_force_op_resend":"0","last_force_op_resend_prenautilus":"0","last_force_op_resend_preluminous":"0","auid":0,"snap_mode":"selfmanaged","snap_seq":0,"snap_epoch":0,"pool_snaps":[],"removed_snaps":"[]","quota_max_bytes":0,"quota_max_objects":0,"tiers":[],"tier_of":-1,"read_tier":-1,"write_tier":-1,"cache_mode":"none","target_max_bytes":0,"target_max_objects":0,"cache_target_dirty_ratio_micro":400000,"cache_target_dirty_high_ratio_micro":600000,"cache_target_full_ratio_micro":800000,"cache_min_flush_age":0,"cache_min_evict_age":0,"erasure_code_profile":"","hit_set_params":{"type":"none"},"hit_set_period":0,"hit_set_count":0,"use_gmt_hitset":true,"min_read_recency_for_promote":0,"min_write_recency_for_promote":0,"hit_set_grade_decay_rate":0,"hit_set_search_last_n":0,"grade_table":[],"stripe_width":0,"expected_num_objects":0,"fast_read":false,"options":{"pg_num_max":32,"pg_num_min":1},"application_metadata":{"mgr":{}},"read_balance":{"score_type":"Fair distribution","score_acting":3,"score_stable":3,"optimal_score":1,"raw_score_acting":3,"raw_score_stable":3,"primary_affinity_weighted":1,"average_primary_affinity":1,"average_primary_affinity_weighted":1}}],"osds":[{"osd":0,"uuid":"9b067403-3d7b-4d89-a7e1-1cd83ba7845f","up":1,"in":1,"weight":1,"primary_affinity":1,"last_clean_begin":0,"last_clean_end":0,"up_from":8,"up_thru":0,"down_at":0,"lost_at":0,"public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.102:6802","nonce":3358520121},{"type":"v1","addr":"192.168.123.102:6803","nonce":3358520121}]},"cluster_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.102:6804","nonce":3358520121},{"type":"v1","addr":"192.168.123.102:6805","nonce":3358520121}]},"heartbeat_back_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.102:6808","nonce":3358520121},{"type":"v1","addr":"192.168.123.102:6809","nonce":3358520121}]},"heartbeat_front_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.102:6806","nonce":3358520121},{"type":"v1","addr":"192.168.123.102:6807","nonce":3358520121}]},"public_addr":"192.168.123.102:6803/3358520121","cluster_addr":"192.168.123.102:6805/3358520121","heartbeat_back_addr":"192.168.123.102:6809/3358520121","heartbeat_front_addr":"192.168.123.102:6807/3358520121","state":["exists","up"]},{"osd":1,"uuid":"1b4f83a6-1a7a-4b65-95a3-d2d4a61115b6","up":1,"in":1,"weight":1,"primary_affinity":1,"last_clean_begin":0,"last_clean_end":0,"up_from":13,"up_thru":19,"down_at":0,"lost_at":0,"public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.104:6800","nonce":1947797507},{"type":"v1","addr":"192.168.123.104:6801","nonce":1947797507}]},"cluster_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.104:6802","nonce":1947797507},{"type":"v1","addr":"192.168.123.104:6803","nonce":1947797507}]},"heartbeat_back_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.104:6806","nonce":1947797507},{"type":"v1","addr":"192.168.123.104:6807","nonce":1947797507}]},"heartbeat_front_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.104:6804","nonce":1947797507},{"type":"v1","addr":"192.168.123.104:6805","nonce":1947797507}]},"public_addr":"192.168.123.104:6801/1947797507","cluster_addr":"192.168.123.104:6803/1947797507","heartbeat_back_addr":"192.168.123.104:6807/1947797507","heartbeat_front_addr":"192.168.123.104:6805/1947797507","state":["exists","up"]},{"osd":2,"uuid":"6a09eea0-8ddc-4889-9cc9-9f59732bfa8c","up":1,"in":1,"weight":1,"primary_affinity":1,"last_clean_begin":0,"last_clean_end":0,"up_from":18,"up_thru":0,"down_at":0,"lost_at":0,"public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.106:6800","nonce":613497499},{"type":"v1","addr":"192.168.123.106:6801","nonce":613497499}]},"cluster_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.106:6802","nonce":613497499},{"type":"v1","addr":"192.168.123.106:6803","nonce":613497499}]},"heartbeat_back_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.106:6806","nonce":613497499},{"type":"v1","addr":"192.168.123.106:6807","nonce":613497499}]},"heartbeat_front_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.106:6804","nonce":613497499},{"type":"v1","addr":"192.168.123.106:6805","nonce":613497499}]},"public_addr":"192.168.123.106:6801/613497499","cluster_addr":"192.168.123.106:6803/613497499","heartbeat_back_addr":"192.168.123.106:6807/613497499","heartbeat_front_addr":"192.168.123.106:6805/613497499","state":["exists","up"]}],"osd_xinfo":[{"osd":0,"down_stamp":"0.000000","laggy_probability":0,"laggy_interval":0,"features":4540701547738038271,"old_weight":0,"last_purged_snaps_scrub":"2026-03-06T12:46:26.304156+0000","dead_epoch":0},{"osd":1,"down_stamp":"0.000000","laggy_probability":0,"laggy_interval":0,"features":4540701547738038271,"old_weight":0,"last_purged_snaps_scrub":"2026-03-06T12:46:42.644905+0000","dead_epoch":0},{"osd":2,"down_stamp":"0.000000","laggy_probability":0,"laggy_interval":0,"features":4540701547738038271,"old_weight":0,"last_purged_snaps_scrub":"2026-03-06T12:46:57.951366+0000","dead_epoch":0}],"pg_upmap":[],"pg_upmap_items":[],"pg_upmap_primaries":[],"pg_temp":[],"primary_temp":[],"blocklist":{"192.168.123.102:0/1357185787":"2026-03-07T12:45:36.564123+0000","192.168.123.102:6801/3504819941":"2026-03-07T12:45:36.564123+0000","192.168.123.102:6800/3504819941":"2026-03-07T12:45:36.564123+0000","192.168.123.102:0/517145126":"2026-03-07T12:45:36.564123+0000","192.168.123.102:0/1479063576":"2026-03-07T12:45:36.564123+0000","192.168.123.102:0/2955285738":"2026-03-07T12:45:16.884814+0000","192.168.123.102:0/775143890":"2026-03-07T12:45:16.884814+0000","192.168.123.102:0/2841567175":"2026-03-07T12:45:16.884814+0000","192.168.123.102:6801/3168015872":"2026-03-07T12:45:16.884814+0000","192.168.123.102:6800/3168015872":"2026-03-07T12:45:16.884814+0000"},"range_blocklist":{},"erasure_code_profiles":{"default":{"crush-failure-domain":"osd","k":"2","m":"1","plugin":"jerasure","technique":"reed_sol_van"}},"removed_snaps_queue":[],"new_removed_snaps":[],"new_purged_snaps":[],"crush_node_flags":{},"device_class_flags":{},"stretch_mode":{"stretch_mode_enabled":false,"stretch_bucket_count":0,"degraded_stretch_mode":0,"recovering_stretch_mode":0,"stretch_mode_bucket":0}} 2026-03-06T13:47:06.111 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:47:05 vm02 ceph-mon[49207]: pgmap v61: 1 pgs: 1 unknown; 0 B data, 80 MiB used, 60 GiB / 60 GiB avail 2026-03-06T13:47:06.111 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:47:05 vm02 ceph-mon[49207]: Detected new or changed devices on vm06 2026-03-06T13:47:06.111 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:47:05 vm02 ceph-mon[49207]: Health check cleared: POOL_APP_NOT_ENABLED (was: 1 pool(s) do not have an application enabled) 2026-03-06T13:47:06.111 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:47:05 vm02 ceph-mon[49207]: Cluster is now healthy 2026-03-06T13:47:06.111 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:47:05 vm02 ceph-mon[49207]: from='client.? 192.168.123.102:0/2333212563' entity='client.admin' cmd=[{"prefix": "mgr dump", "format": "json"}]: dispatch 2026-03-06T13:47:06.111 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:47:05 vm02 ceph-mon[49207]: mgrmap e15: a(active, since 88s), standbys: b 2026-03-06T13:47:06.291 INFO:tasks.cephadm.ceph_manager.ceph:all up! 2026-03-06T13:47:06.291 DEBUG:teuthology.orchestra.run.vm02:> sudo /home/ubuntu/cephtest/cephadm --image harbor.clyso.com/custom-ceph/ceph/ceph:cobaltcore-storage-v19.2.3-fasttrack-3 shell --fsid 2b1b4a72-195a-11f1-b1df-bf1f800d676b -- ceph osd dump --format=json 2026-03-06T13:47:06.616 INFO:teuthology.orchestra.run.vm02.stderr:Inferring config /var/lib/ceph/2b1b4a72-195a-11f1-b1df-bf1f800d676b/mon.a/config 2026-03-06T13:47:06.957 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:47:06 vm02 ceph-mon[49207]: from='client.? 192.168.123.102:0/4111544637' entity='client.admin' cmd=[{"prefix": "osd dump", "format": "json"}]: dispatch 2026-03-06T13:47:06.957 INFO:teuthology.orchestra.run.vm02.stdout: 2026-03-06T13:47:06.957 INFO:teuthology.orchestra.run.vm02.stdout:{"epoch":21,"fsid":"2b1b4a72-195a-11f1-b1df-bf1f800d676b","created":"2026-03-06T12:44:49.710645+0000","modified":"2026-03-06T12:47:03.639935+0000","last_up_change":"2026-03-06T12:47:00.617832+0000","last_in_change":"2026-03-06T12:46:47.212348+0000","flags":"sortbitwise,recovery_deletes,purged_snapdirs,pglog_hardlimit","flags_num":5799936,"flags_set":["pglog_hardlimit","purged_snapdirs","recovery_deletes","sortbitwise"],"crush_version":8,"full_ratio":0.94999998807907104,"backfillfull_ratio":0.89999997615814209,"nearfull_ratio":0.85000002384185791,"cluster_snapshot":"","pool_max":1,"max_osd":3,"require_min_compat_client":"luminous","min_compat_client":"jewel","require_osd_release":"squid","allow_crimson":false,"pools":[{"pool":1,"pool_name":".mgr","create_time":"2026-03-06T12:47:00.638787+0000","flags":1,"flags_names":"hashpspool","type":1,"size":3,"min_size":2,"crush_rule":0,"peering_crush_bucket_count":0,"peering_crush_bucket_target":0,"peering_crush_bucket_barrier":0,"peering_crush_bucket_mandatory_member":2147483647,"is_stretch_pool":false,"object_hash":2,"pg_autoscale_mode":"off","pg_num":1,"pg_placement_num":1,"pg_placement_num_target":1,"pg_num_target":1,"pg_num_pending":1,"last_pg_merge_meta":{"source_pgid":"0.0","ready_epoch":0,"last_epoch_started":0,"last_epoch_clean":0,"source_version":"0'0","target_version":"0'0"},"last_change":"21","last_force_op_resend":"0","last_force_op_resend_prenautilus":"0","last_force_op_resend_preluminous":"0","auid":0,"snap_mode":"selfmanaged","snap_seq":0,"snap_epoch":0,"pool_snaps":[],"removed_snaps":"[]","quota_max_bytes":0,"quota_max_objects":0,"tiers":[],"tier_of":-1,"read_tier":-1,"write_tier":-1,"cache_mode":"none","target_max_bytes":0,"target_max_objects":0,"cache_target_dirty_ratio_micro":400000,"cache_target_dirty_high_ratio_micro":600000,"cache_target_full_ratio_micro":800000,"cache_min_flush_age":0,"cache_min_evict_age":0,"erasure_code_profile":"","hit_set_params":{"type":"none"},"hit_set_period":0,"hit_set_count":0,"use_gmt_hitset":true,"min_read_recency_for_promote":0,"min_write_recency_for_promote":0,"hit_set_grade_decay_rate":0,"hit_set_search_last_n":0,"grade_table":[],"stripe_width":0,"expected_num_objects":0,"fast_read":false,"options":{"pg_num_max":32,"pg_num_min":1},"application_metadata":{"mgr":{}},"read_balance":{"score_type":"Fair distribution","score_acting":3,"score_stable":3,"optimal_score":1,"raw_score_acting":3,"raw_score_stable":3,"primary_affinity_weighted":1,"average_primary_affinity":1,"average_primary_affinity_weighted":1}}],"osds":[{"osd":0,"uuid":"9b067403-3d7b-4d89-a7e1-1cd83ba7845f","up":1,"in":1,"weight":1,"primary_affinity":1,"last_clean_begin":0,"last_clean_end":0,"up_from":8,"up_thru":0,"down_at":0,"lost_at":0,"public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.102:6802","nonce":3358520121},{"type":"v1","addr":"192.168.123.102:6803","nonce":3358520121}]},"cluster_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.102:6804","nonce":3358520121},{"type":"v1","addr":"192.168.123.102:6805","nonce":3358520121}]},"heartbeat_back_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.102:6808","nonce":3358520121},{"type":"v1","addr":"192.168.123.102:6809","nonce":3358520121}]},"heartbeat_front_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.102:6806","nonce":3358520121},{"type":"v1","addr":"192.168.123.102:6807","nonce":3358520121}]},"public_addr":"192.168.123.102:6803/3358520121","cluster_addr":"192.168.123.102:6805/3358520121","heartbeat_back_addr":"192.168.123.102:6809/3358520121","heartbeat_front_addr":"192.168.123.102:6807/3358520121","state":["exists","up"]},{"osd":1,"uuid":"1b4f83a6-1a7a-4b65-95a3-d2d4a61115b6","up":1,"in":1,"weight":1,"primary_affinity":1,"last_clean_begin":0,"last_clean_end":0,"up_from":13,"up_thru":19,"down_at":0,"lost_at":0,"public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.104:6800","nonce":1947797507},{"type":"v1","addr":"192.168.123.104:6801","nonce":1947797507}]},"cluster_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.104:6802","nonce":1947797507},{"type":"v1","addr":"192.168.123.104:6803","nonce":1947797507}]},"heartbeat_back_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.104:6806","nonce":1947797507},{"type":"v1","addr":"192.168.123.104:6807","nonce":1947797507}]},"heartbeat_front_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.104:6804","nonce":1947797507},{"type":"v1","addr":"192.168.123.104:6805","nonce":1947797507}]},"public_addr":"192.168.123.104:6801/1947797507","cluster_addr":"192.168.123.104:6803/1947797507","heartbeat_back_addr":"192.168.123.104:6807/1947797507","heartbeat_front_addr":"192.168.123.104:6805/1947797507","state":["exists","up"]},{"osd":2,"uuid":"6a09eea0-8ddc-4889-9cc9-9f59732bfa8c","up":1,"in":1,"weight":1,"primary_affinity":1,"last_clean_begin":0,"last_clean_end":0,"up_from":18,"up_thru":0,"down_at":0,"lost_at":0,"public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.106:6800","nonce":613497499},{"type":"v1","addr":"192.168.123.106:6801","nonce":613497499}]},"cluster_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.106:6802","nonce":613497499},{"type":"v1","addr":"192.168.123.106:6803","nonce":613497499}]},"heartbeat_back_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.106:6806","nonce":613497499},{"type":"v1","addr":"192.168.123.106:6807","nonce":613497499}]},"heartbeat_front_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.106:6804","nonce":613497499},{"type":"v1","addr":"192.168.123.106:6805","nonce":613497499}]},"public_addr":"192.168.123.106:6801/613497499","cluster_addr":"192.168.123.106:6803/613497499","heartbeat_back_addr":"192.168.123.106:6807/613497499","heartbeat_front_addr":"192.168.123.106:6805/613497499","state":["exists","up"]}],"osd_xinfo":[{"osd":0,"down_stamp":"0.000000","laggy_probability":0,"laggy_interval":0,"features":4540701547738038271,"old_weight":0,"last_purged_snaps_scrub":"2026-03-06T12:46:26.304156+0000","dead_epoch":0},{"osd":1,"down_stamp":"0.000000","laggy_probability":0,"laggy_interval":0,"features":4540701547738038271,"old_weight":0,"last_purged_snaps_scrub":"2026-03-06T12:46:42.644905+0000","dead_epoch":0},{"osd":2,"down_stamp":"0.000000","laggy_probability":0,"laggy_interval":0,"features":4540701547738038271,"old_weight":0,"last_purged_snaps_scrub":"2026-03-06T12:46:57.951366+0000","dead_epoch":0}],"pg_upmap":[],"pg_upmap_items":[],"pg_upmap_primaries":[],"pg_temp":[],"primary_temp":[],"blocklist":{"192.168.123.102:0/1357185787":"2026-03-07T12:45:36.564123+0000","192.168.123.102:6801/3504819941":"2026-03-07T12:45:36.564123+0000","192.168.123.102:6800/3504819941":"2026-03-07T12:45:36.564123+0000","192.168.123.102:0/517145126":"2026-03-07T12:45:36.564123+0000","192.168.123.102:0/1479063576":"2026-03-07T12:45:36.564123+0000","192.168.123.102:0/2955285738":"2026-03-07T12:45:16.884814+0000","192.168.123.102:0/775143890":"2026-03-07T12:45:16.884814+0000","192.168.123.102:0/2841567175":"2026-03-07T12:45:16.884814+0000","192.168.123.102:6801/3168015872":"2026-03-07T12:45:16.884814+0000","192.168.123.102:6800/3168015872":"2026-03-07T12:45:16.884814+0000"},"range_blocklist":{},"erasure_code_profiles":{"default":{"crush-failure-domain":"osd","k":"2","m":"1","plugin":"jerasure","technique":"reed_sol_van"}},"removed_snaps_queue":[],"new_removed_snaps":[],"new_purged_snaps":[],"crush_node_flags":{},"device_class_flags":{},"stretch_mode":{"stretch_mode_enabled":false,"stretch_bucket_count":0,"degraded_stretch_mode":0,"recovering_stretch_mode":0,"stretch_mode_bucket":0}} 2026-03-06T13:47:07.010 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:47:06 vm06 ceph-mon[56528]: from='client.? 192.168.123.102:0/4111544637' entity='client.admin' cmd=[{"prefix": "osd dump", "format": "json"}]: dispatch 2026-03-06T13:47:07.049 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:47:06 vm04 ceph-mon[53170]: from='client.? 192.168.123.102:0/4111544637' entity='client.admin' cmd=[{"prefix": "osd dump", "format": "json"}]: dispatch 2026-03-06T13:47:07.133 DEBUG:teuthology.orchestra.run.vm02:> sudo /home/ubuntu/cephtest/cephadm --image harbor.clyso.com/custom-ceph/ceph/ceph:cobaltcore-storage-v19.2.3-fasttrack-3 shell --fsid 2b1b4a72-195a-11f1-b1df-bf1f800d676b -- ceph tell osd.0 flush_pg_stats 2026-03-06T13:47:07.133 DEBUG:teuthology.orchestra.run.vm02:> sudo /home/ubuntu/cephtest/cephadm --image harbor.clyso.com/custom-ceph/ceph/ceph:cobaltcore-storage-v19.2.3-fasttrack-3 shell --fsid 2b1b4a72-195a-11f1-b1df-bf1f800d676b -- ceph tell osd.1 flush_pg_stats 2026-03-06T13:47:07.133 DEBUG:teuthology.orchestra.run.vm02:> sudo /home/ubuntu/cephtest/cephadm --image harbor.clyso.com/custom-ceph/ceph/ceph:cobaltcore-storage-v19.2.3-fasttrack-3 shell --fsid 2b1b4a72-195a-11f1-b1df-bf1f800d676b -- ceph tell osd.2 flush_pg_stats 2026-03-06T13:47:07.630 INFO:teuthology.orchestra.run.vm02.stderr:Inferring config /var/lib/ceph/2b1b4a72-195a-11f1-b1df-bf1f800d676b/mon.a/config 2026-03-06T13:47:07.657 INFO:teuthology.orchestra.run.vm02.stderr:Inferring config /var/lib/ceph/2b1b4a72-195a-11f1-b1df-bf1f800d676b/mon.a/config 2026-03-06T13:47:07.783 INFO:teuthology.orchestra.run.vm02.stderr:Inferring config /var/lib/ceph/2b1b4a72-195a-11f1-b1df-bf1f800d676b/mon.a/config 2026-03-06T13:47:08.010 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:47:07 vm06 ceph-mon[56528]: pgmap v62: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-06T13:47:08.010 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:47:07 vm06 ceph-mon[56528]: from='client.? 192.168.123.102:0/2305346915' entity='client.admin' cmd=[{"prefix": "osd dump", "format": "json"}]: dispatch 2026-03-06T13:47:08.049 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:47:07 vm04 ceph-mon[53170]: pgmap v62: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-06T13:47:08.049 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:47:07 vm04 ceph-mon[53170]: from='client.? 192.168.123.102:0/2305346915' entity='client.admin' cmd=[{"prefix": "osd dump", "format": "json"}]: dispatch 2026-03-06T13:47:08.059 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:47:07 vm02 ceph-mon[49207]: pgmap v62: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-06T13:47:08.059 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:47:07 vm02 ceph-mon[49207]: from='client.? 192.168.123.102:0/2305346915' entity='client.admin' cmd=[{"prefix": "osd dump", "format": "json"}]: dispatch 2026-03-06T13:47:08.225 INFO:teuthology.orchestra.run.vm02.stdout:55834574854 2026-03-06T13:47:08.225 DEBUG:teuthology.orchestra.run.vm02:> sudo /home/ubuntu/cephtest/cephadm --image harbor.clyso.com/custom-ceph/ceph/ceph:cobaltcore-storage-v19.2.3-fasttrack-3 shell --fsid 2b1b4a72-195a-11f1-b1df-bf1f800d676b -- ceph osd last-stat-seq osd.1 2026-03-06T13:47:08.248 INFO:teuthology.orchestra.run.vm02.stdout:34359738377 2026-03-06T13:47:08.248 DEBUG:teuthology.orchestra.run.vm02:> sudo /home/ubuntu/cephtest/cephadm --image harbor.clyso.com/custom-ceph/ceph/ceph:cobaltcore-storage-v19.2.3-fasttrack-3 shell --fsid 2b1b4a72-195a-11f1-b1df-bf1f800d676b -- ceph osd last-stat-seq osd.0 2026-03-06T13:47:08.404 INFO:teuthology.orchestra.run.vm02.stdout:77309411331 2026-03-06T13:47:08.405 DEBUG:teuthology.orchestra.run.vm02:> sudo /home/ubuntu/cephtest/cephadm --image harbor.clyso.com/custom-ceph/ceph/ceph:cobaltcore-storage-v19.2.3-fasttrack-3 shell --fsid 2b1b4a72-195a-11f1-b1df-bf1f800d676b -- ceph osd last-stat-seq osd.2 2026-03-06T13:47:08.699 INFO:teuthology.orchestra.run.vm02.stderr:Inferring config /var/lib/ceph/2b1b4a72-195a-11f1-b1df-bf1f800d676b/mon.a/config 2026-03-06T13:47:08.734 INFO:teuthology.orchestra.run.vm02.stderr:Inferring config /var/lib/ceph/2b1b4a72-195a-11f1-b1df-bf1f800d676b/mon.a/config 2026-03-06T13:47:09.006 INFO:teuthology.orchestra.run.vm02.stderr:Inferring config /var/lib/ceph/2b1b4a72-195a-11f1-b1df-bf1f800d676b/mon.a/config 2026-03-06T13:47:09.264 INFO:teuthology.orchestra.run.vm02.stdout:34359738377 2026-03-06T13:47:09.331 INFO:teuthology.orchestra.run.vm02.stdout:55834574854 2026-03-06T13:47:09.437 INFO:tasks.cephadm.ceph_manager.ceph:need seq 34359738377 got 34359738377 for osd.0 2026-03-06T13:47:09.437 DEBUG:teuthology.parallel:result is None 2026-03-06T13:47:09.487 INFO:teuthology.orchestra.run.vm02.stdout:77309411331 2026-03-06T13:47:09.528 INFO:tasks.cephadm.ceph_manager.ceph:need seq 55834574854 got 55834574854 for osd.1 2026-03-06T13:47:09.528 DEBUG:teuthology.parallel:result is None 2026-03-06T13:47:09.660 INFO:tasks.cephadm.ceph_manager.ceph:need seq 77309411331 got 77309411331 for osd.2 2026-03-06T13:47:09.660 DEBUG:teuthology.parallel:result is None 2026-03-06T13:47:09.660 INFO:tasks.cephadm.ceph_manager.ceph:waiting for clean 2026-03-06T13:47:09.660 DEBUG:teuthology.orchestra.run.vm02:> sudo /home/ubuntu/cephtest/cephadm --image harbor.clyso.com/custom-ceph/ceph/ceph:cobaltcore-storage-v19.2.3-fasttrack-3 shell --fsid 2b1b4a72-195a-11f1-b1df-bf1f800d676b -- ceph pg dump --format=json 2026-03-06T13:47:09.999 INFO:teuthology.orchestra.run.vm02.stderr:Inferring config /var/lib/ceph/2b1b4a72-195a-11f1-b1df-bf1f800d676b/mon.a/config 2026-03-06T13:47:10.010 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:47:09 vm06 ceph-mon[56528]: pgmap v63: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-06T13:47:10.010 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:47:09 vm06 ceph-mon[56528]: from='client.? 192.168.123.102:0/44913122' entity='client.admin' cmd=[{"prefix": "osd last-stat-seq", "id": 0}]: dispatch 2026-03-06T13:47:10.010 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:47:09 vm06 ceph-mon[56528]: from='client.? 192.168.123.102:0/2665180684' entity='client.admin' cmd=[{"prefix": "osd last-stat-seq", "id": 1}]: dispatch 2026-03-06T13:47:10.010 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:47:09 vm06 ceph-mon[56528]: from='client.? 192.168.123.102:0/2176623382' entity='client.admin' cmd=[{"prefix": "osd last-stat-seq", "id": 2}]: dispatch 2026-03-06T13:47:10.022 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:47:09 vm02 ceph-mon[49207]: pgmap v63: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-06T13:47:10.022 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:47:09 vm02 ceph-mon[49207]: from='client.? 192.168.123.102:0/44913122' entity='client.admin' cmd=[{"prefix": "osd last-stat-seq", "id": 0}]: dispatch 2026-03-06T13:47:10.022 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:47:09 vm02 ceph-mon[49207]: from='client.? 192.168.123.102:0/2665180684' entity='client.admin' cmd=[{"prefix": "osd last-stat-seq", "id": 1}]: dispatch 2026-03-06T13:47:10.022 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:47:09 vm02 ceph-mon[49207]: from='client.? 192.168.123.102:0/2176623382' entity='client.admin' cmd=[{"prefix": "osd last-stat-seq", "id": 2}]: dispatch 2026-03-06T13:47:10.050 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:47:09 vm04 ceph-mon[53170]: pgmap v63: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-06T13:47:10.050 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:47:09 vm04 ceph-mon[53170]: from='client.? 192.168.123.102:0/44913122' entity='client.admin' cmd=[{"prefix": "osd last-stat-seq", "id": 0}]: dispatch 2026-03-06T13:47:10.050 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:47:09 vm04 ceph-mon[53170]: from='client.? 192.168.123.102:0/2665180684' entity='client.admin' cmd=[{"prefix": "osd last-stat-seq", "id": 1}]: dispatch 2026-03-06T13:47:10.050 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:47:09 vm04 ceph-mon[53170]: from='client.? 192.168.123.102:0/2176623382' entity='client.admin' cmd=[{"prefix": "osd last-stat-seq", "id": 2}]: dispatch 2026-03-06T13:47:10.330 INFO:teuthology.orchestra.run.vm02.stdout: 2026-03-06T13:47:10.330 INFO:teuthology.orchestra.run.vm02.stderr:dumped all 2026-03-06T13:47:10.486 INFO:teuthology.orchestra.run.vm02.stdout:{"pg_ready":true,"pg_map":{"version":64,"stamp":"2026-03-06T12:47:10.188135+0000","last_osdmap_epoch":0,"last_pg_scan":0,"pg_stats_sum":{"stat_sum":{"num_bytes":459280,"num_objects":2,"num_object_clones":0,"num_object_copies":6,"num_objects_missing_on_primary":0,"num_objects_missing":0,"num_objects_degraded":0,"num_objects_misplaced":0,"num_objects_unfound":0,"num_objects_dirty":2,"num_whiteouts":0,"num_read":46,"num_read_kb":37,"num_write":57,"num_write_kb":584,"num_scrub_errors":0,"num_shallow_scrub_errors":0,"num_deep_scrub_errors":0,"num_objects_recovered":0,"num_bytes_recovered":0,"num_keys_recovered":0,"num_objects_omap":0,"num_objects_hit_set_archive":0,"num_bytes_hit_set_archive":0,"num_flush":0,"num_flush_kb":0,"num_evict":0,"num_evict_kb":0,"num_promote":0,"num_flush_mode_high":0,"num_flush_mode_low":0,"num_evict_mode_some":0,"num_evict_mode_full":0,"num_objects_pinned":0,"num_legacy_snapsets":0,"num_large_omap_objects":0,"num_objects_manifest":0,"num_omap_bytes":0,"num_omap_keys":0,"num_objects_repaired":0},"store_stats":{"total":0,"available":0,"internally_reserved":0,"allocated":0,"data_stored":0,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":0},"log_size":32,"ondisk_log_size":32,"up":3,"acting":3,"num_store_stats":0},"osd_stats_sum":{"up_from":0,"seq":0,"num_pgs":3,"num_osds":3,"num_per_pool_osds":3,"num_per_pool_omap_osds":3,"kb":62902272,"kb_used":82796,"kb_used_data":1884,"kb_used_omap":4,"kb_used_meta":80443,"kb_avail":62819476,"statfs":{"total":64411926528,"available":64327143424,"internally_reserved":0,"allocated":1929216,"data_stored":1541172,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":4770,"internal_metadata":82373982},"hb_peers":[],"snap_trim_queue_len":0,"num_snap_trimming":0,"num_shards_repaired":0,"op_queue_age_hist":{"histogram":[],"upper_bound":1},"perf_stat":{"commit_latency_ms":0,"apply_latency_ms":0,"commit_latency_ns":0,"apply_latency_ns":0},"alerts":[],"network_ping_times":[]},"pg_stats_delta":{"stat_sum":{"num_bytes":0,"num_objects":0,"num_object_clones":0,"num_object_copies":0,"num_objects_missing_on_primary":0,"num_objects_missing":0,"num_objects_degraded":0,"num_objects_misplaced":0,"num_objects_unfound":0,"num_objects_dirty":0,"num_whiteouts":0,"num_read":0,"num_read_kb":0,"num_write":0,"num_write_kb":0,"num_scrub_errors":0,"num_shallow_scrub_errors":0,"num_deep_scrub_errors":0,"num_objects_recovered":0,"num_bytes_recovered":0,"num_keys_recovered":0,"num_objects_omap":0,"num_objects_hit_set_archive":0,"num_bytes_hit_set_archive":0,"num_flush":0,"num_flush_kb":0,"num_evict":0,"num_evict_kb":0,"num_promote":0,"num_flush_mode_high":0,"num_flush_mode_low":0,"num_evict_mode_some":0,"num_evict_mode_full":0,"num_objects_pinned":0,"num_legacy_snapsets":0,"num_large_omap_objects":0,"num_objects_manifest":0,"num_omap_bytes":0,"num_omap_keys":0,"num_objects_repaired":0},"store_stats":{"total":0,"available":0,"internally_reserved":0,"allocated":0,"data_stored":0,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":0},"log_size":0,"ondisk_log_size":0,"up":0,"acting":0,"num_store_stats":0,"stamp_delta":"4.000541"},"pg_stats":[{"pgid":"1.0","version":"20'32","reported_seq":57,"reported_epoch":21,"state":"active+clean","last_fresh":"2026-03-06T12:47:03.646123+0000","last_change":"2026-03-06T12:47:02.650769+0000","last_active":"2026-03-06T12:47:03.646123+0000","last_peered":"2026-03-06T12:47:03.646123+0000","last_clean":"2026-03-06T12:47:03.646123+0000","last_became_active":"2026-03-06T12:47:02.650643+0000","last_became_peered":"2026-03-06T12:47:02.650643+0000","last_unstale":"2026-03-06T12:47:03.646123+0000","last_undegraded":"2026-03-06T12:47:03.646123+0000","last_fullsized":"2026-03-06T12:47:03.646123+0000","mapping_epoch":19,"log_start":"0'0","ondisk_log_start":"0'0","created":19,"last_epoch_clean":20,"parent":"0.0","parent_split_bits":0,"last_scrub":"0'0","last_scrub_stamp":"2026-03-06T12:47:01.620483+0000","last_deep_scrub":"0'0","last_deep_scrub_stamp":"2026-03-06T12:47:01.620483+0000","last_clean_scrub_stamp":"2026-03-06T12:47:01.620483+0000","objects_scrubbed":0,"log_size":32,"log_dups_size":0,"ondisk_log_size":32,"stats_invalid":false,"dirty_stats_invalid":false,"omap_stats_invalid":false,"hitset_stats_invalid":false,"hitset_bytes_stats_invalid":false,"pin_stats_invalid":false,"manifest_stats_invalid":false,"snaptrimq_len":0,"last_scrub_duration":0,"scrub_schedule":"periodic scrub scheduled @ 2026-03-07T13:00:47.538414+0000","scrub_duration":0,"objects_trimmed":0,"snaptrim_duration":0,"stat_sum":{"num_bytes":459280,"num_objects":2,"num_object_clones":0,"num_object_copies":6,"num_objects_missing_on_primary":0,"num_objects_missing":0,"num_objects_degraded":0,"num_objects_misplaced":0,"num_objects_unfound":0,"num_objects_dirty":2,"num_whiteouts":0,"num_read":46,"num_read_kb":37,"num_write":57,"num_write_kb":584,"num_scrub_errors":0,"num_shallow_scrub_errors":0,"num_deep_scrub_errors":0,"num_objects_recovered":0,"num_bytes_recovered":0,"num_keys_recovered":0,"num_objects_omap":0,"num_objects_hit_set_archive":0,"num_bytes_hit_set_archive":0,"num_flush":0,"num_flush_kb":0,"num_evict":0,"num_evict_kb":0,"num_promote":0,"num_flush_mode_high":0,"num_flush_mode_low":0,"num_evict_mode_some":0,"num_evict_mode_full":0,"num_objects_pinned":0,"num_legacy_snapsets":0,"num_large_omap_objects":0,"num_objects_manifest":0,"num_omap_bytes":0,"num_omap_keys":0,"num_objects_repaired":0},"up":[1,2,0],"acting":[1,2,0],"avail_no_missing":[],"object_location_counts":[],"blocked_by":[],"up_primary":1,"acting_primary":1,"purged_snaps":[]}],"pool_stats":[{"poolid":1,"num_pg":1,"stat_sum":{"num_bytes":459280,"num_objects":2,"num_object_clones":0,"num_object_copies":6,"num_objects_missing_on_primary":0,"num_objects_missing":0,"num_objects_degraded":0,"num_objects_misplaced":0,"num_objects_unfound":0,"num_objects_dirty":2,"num_whiteouts":0,"num_read":46,"num_read_kb":37,"num_write":57,"num_write_kb":584,"num_scrub_errors":0,"num_shallow_scrub_errors":0,"num_deep_scrub_errors":0,"num_objects_recovered":0,"num_bytes_recovered":0,"num_keys_recovered":0,"num_objects_omap":0,"num_objects_hit_set_archive":0,"num_bytes_hit_set_archive":0,"num_flush":0,"num_flush_kb":0,"num_evict":0,"num_evict_kb":0,"num_promote":0,"num_flush_mode_high":0,"num_flush_mode_low":0,"num_evict_mode_some":0,"num_evict_mode_full":0,"num_objects_pinned":0,"num_legacy_snapsets":0,"num_large_omap_objects":0,"num_objects_manifest":0,"num_omap_bytes":0,"num_omap_keys":0,"num_objects_repaired":0},"store_stats":{"total":0,"available":0,"internally_reserved":0,"allocated":1388544,"data_stored":1377840,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":0},"log_size":32,"ondisk_log_size":32,"up":3,"acting":3,"num_store_stats":3}],"osd_stats":[{"osd":2,"up_from":18,"seq":77309411331,"num_pgs":1,"num_osds":1,"num_per_pool_osds":1,"num_per_pool_omap_osds":1,"kb":20967424,"kb_used":27604,"kb_used_data":628,"kb_used_omap":1,"kb_used_meta":26814,"kb_avail":20939820,"statfs":{"total":21470642176,"available":21442375680,"internally_reserved":0,"allocated":643072,"data_stored":513724,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":1590,"internal_metadata":27457994},"hb_peers":[0,1],"snap_trim_queue_len":0,"num_snap_trimming":0,"num_shards_repaired":0,"op_queue_age_hist":{"histogram":[],"upper_bound":1},"perf_stat":{"commit_latency_ms":0,"apply_latency_ms":0,"commit_latency_ns":0,"apply_latency_ns":0},"alerts":[]},{"osd":1,"up_from":13,"seq":55834574854,"num_pgs":1,"num_osds":1,"num_per_pool_osds":1,"num_per_pool_omap_osds":1,"kb":20967424,"kb_used":27592,"kb_used_data":628,"kb_used_omap":1,"kb_used_meta":26814,"kb_avail":20939832,"statfs":{"total":21470642176,"available":21442387968,"internally_reserved":0,"allocated":643072,"data_stored":513724,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":1590,"internal_metadata":27457994},"hb_peers":[0,2],"snap_trim_queue_len":0,"num_snap_trimming":0,"num_shards_repaired":0,"op_queue_age_hist":{"histogram":[],"upper_bound":1},"perf_stat":{"commit_latency_ms":0,"apply_latency_ms":0,"commit_latency_ns":0,"apply_latency_ns":0},"alerts":[]},{"osd":0,"up_from":8,"seq":34359738378,"num_pgs":1,"num_osds":1,"num_per_pool_osds":1,"num_per_pool_omap_osds":1,"kb":20967424,"kb_used":27600,"kb_used_data":628,"kb_used_omap":1,"kb_used_meta":26814,"kb_avail":20939824,"statfs":{"total":21470642176,"available":21442379776,"internally_reserved":0,"allocated":643072,"data_stored":513724,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":1590,"internal_metadata":27457994},"hb_peers":[1,2],"snap_trim_queue_len":0,"num_snap_trimming":0,"num_shards_repaired":0,"op_queue_age_hist":{"histogram":[],"upper_bound":1},"perf_stat":{"commit_latency_ms":0,"apply_latency_ms":0,"commit_latency_ns":0,"apply_latency_ns":0},"alerts":[]}],"pool_statfs":[{"poolid":1,"osd":0,"total":0,"available":0,"internally_reserved":0,"allocated":462848,"data_stored":459280,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":0},{"poolid":1,"osd":1,"total":0,"available":0,"internally_reserved":0,"allocated":462848,"data_stored":459280,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":0},{"poolid":1,"osd":2,"total":0,"available":0,"internally_reserved":0,"allocated":462848,"data_stored":459280,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":0}]}} 2026-03-06T13:47:10.486 DEBUG:teuthology.orchestra.run.vm02:> sudo /home/ubuntu/cephtest/cephadm --image harbor.clyso.com/custom-ceph/ceph/ceph:cobaltcore-storage-v19.2.3-fasttrack-3 shell --fsid 2b1b4a72-195a-11f1-b1df-bf1f800d676b -- ceph pg dump --format=json 2026-03-06T13:47:10.827 INFO:teuthology.orchestra.run.vm02.stderr:Inferring config /var/lib/ceph/2b1b4a72-195a-11f1-b1df-bf1f800d676b/mon.a/config 2026-03-06T13:47:11.161 INFO:teuthology.orchestra.run.vm02.stdout: 2026-03-06T13:47:11.161 INFO:teuthology.orchestra.run.vm02.stderr:dumped all 2026-03-06T13:47:11.350 INFO:teuthology.orchestra.run.vm02.stdout:{"pg_ready":true,"pg_map":{"version":64,"stamp":"2026-03-06T12:47:10.188135+0000","last_osdmap_epoch":0,"last_pg_scan":0,"pg_stats_sum":{"stat_sum":{"num_bytes":459280,"num_objects":2,"num_object_clones":0,"num_object_copies":6,"num_objects_missing_on_primary":0,"num_objects_missing":0,"num_objects_degraded":0,"num_objects_misplaced":0,"num_objects_unfound":0,"num_objects_dirty":2,"num_whiteouts":0,"num_read":46,"num_read_kb":37,"num_write":57,"num_write_kb":584,"num_scrub_errors":0,"num_shallow_scrub_errors":0,"num_deep_scrub_errors":0,"num_objects_recovered":0,"num_bytes_recovered":0,"num_keys_recovered":0,"num_objects_omap":0,"num_objects_hit_set_archive":0,"num_bytes_hit_set_archive":0,"num_flush":0,"num_flush_kb":0,"num_evict":0,"num_evict_kb":0,"num_promote":0,"num_flush_mode_high":0,"num_flush_mode_low":0,"num_evict_mode_some":0,"num_evict_mode_full":0,"num_objects_pinned":0,"num_legacy_snapsets":0,"num_large_omap_objects":0,"num_objects_manifest":0,"num_omap_bytes":0,"num_omap_keys":0,"num_objects_repaired":0},"store_stats":{"total":0,"available":0,"internally_reserved":0,"allocated":0,"data_stored":0,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":0},"log_size":32,"ondisk_log_size":32,"up":3,"acting":3,"num_store_stats":0},"osd_stats_sum":{"up_from":0,"seq":0,"num_pgs":3,"num_osds":3,"num_per_pool_osds":3,"num_per_pool_omap_osds":3,"kb":62902272,"kb_used":82796,"kb_used_data":1884,"kb_used_omap":4,"kb_used_meta":80443,"kb_avail":62819476,"statfs":{"total":64411926528,"available":64327143424,"internally_reserved":0,"allocated":1929216,"data_stored":1541172,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":4770,"internal_metadata":82373982},"hb_peers":[],"snap_trim_queue_len":0,"num_snap_trimming":0,"num_shards_repaired":0,"op_queue_age_hist":{"histogram":[],"upper_bound":1},"perf_stat":{"commit_latency_ms":0,"apply_latency_ms":0,"commit_latency_ns":0,"apply_latency_ns":0},"alerts":[],"network_ping_times":[]},"pg_stats_delta":{"stat_sum":{"num_bytes":0,"num_objects":0,"num_object_clones":0,"num_object_copies":0,"num_objects_missing_on_primary":0,"num_objects_missing":0,"num_objects_degraded":0,"num_objects_misplaced":0,"num_objects_unfound":0,"num_objects_dirty":0,"num_whiteouts":0,"num_read":0,"num_read_kb":0,"num_write":0,"num_write_kb":0,"num_scrub_errors":0,"num_shallow_scrub_errors":0,"num_deep_scrub_errors":0,"num_objects_recovered":0,"num_bytes_recovered":0,"num_keys_recovered":0,"num_objects_omap":0,"num_objects_hit_set_archive":0,"num_bytes_hit_set_archive":0,"num_flush":0,"num_flush_kb":0,"num_evict":0,"num_evict_kb":0,"num_promote":0,"num_flush_mode_high":0,"num_flush_mode_low":0,"num_evict_mode_some":0,"num_evict_mode_full":0,"num_objects_pinned":0,"num_legacy_snapsets":0,"num_large_omap_objects":0,"num_objects_manifest":0,"num_omap_bytes":0,"num_omap_keys":0,"num_objects_repaired":0},"store_stats":{"total":0,"available":0,"internally_reserved":0,"allocated":0,"data_stored":0,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":0},"log_size":0,"ondisk_log_size":0,"up":0,"acting":0,"num_store_stats":0,"stamp_delta":"4.000541"},"pg_stats":[{"pgid":"1.0","version":"20'32","reported_seq":57,"reported_epoch":21,"state":"active+clean","last_fresh":"2026-03-06T12:47:03.646123+0000","last_change":"2026-03-06T12:47:02.650769+0000","last_active":"2026-03-06T12:47:03.646123+0000","last_peered":"2026-03-06T12:47:03.646123+0000","last_clean":"2026-03-06T12:47:03.646123+0000","last_became_active":"2026-03-06T12:47:02.650643+0000","last_became_peered":"2026-03-06T12:47:02.650643+0000","last_unstale":"2026-03-06T12:47:03.646123+0000","last_undegraded":"2026-03-06T12:47:03.646123+0000","last_fullsized":"2026-03-06T12:47:03.646123+0000","mapping_epoch":19,"log_start":"0'0","ondisk_log_start":"0'0","created":19,"last_epoch_clean":20,"parent":"0.0","parent_split_bits":0,"last_scrub":"0'0","last_scrub_stamp":"2026-03-06T12:47:01.620483+0000","last_deep_scrub":"0'0","last_deep_scrub_stamp":"2026-03-06T12:47:01.620483+0000","last_clean_scrub_stamp":"2026-03-06T12:47:01.620483+0000","objects_scrubbed":0,"log_size":32,"log_dups_size":0,"ondisk_log_size":32,"stats_invalid":false,"dirty_stats_invalid":false,"omap_stats_invalid":false,"hitset_stats_invalid":false,"hitset_bytes_stats_invalid":false,"pin_stats_invalid":false,"manifest_stats_invalid":false,"snaptrimq_len":0,"last_scrub_duration":0,"scrub_schedule":"periodic scrub scheduled @ 2026-03-07T13:00:47.538414+0000","scrub_duration":0,"objects_trimmed":0,"snaptrim_duration":0,"stat_sum":{"num_bytes":459280,"num_objects":2,"num_object_clones":0,"num_object_copies":6,"num_objects_missing_on_primary":0,"num_objects_missing":0,"num_objects_degraded":0,"num_objects_misplaced":0,"num_objects_unfound":0,"num_objects_dirty":2,"num_whiteouts":0,"num_read":46,"num_read_kb":37,"num_write":57,"num_write_kb":584,"num_scrub_errors":0,"num_shallow_scrub_errors":0,"num_deep_scrub_errors":0,"num_objects_recovered":0,"num_bytes_recovered":0,"num_keys_recovered":0,"num_objects_omap":0,"num_objects_hit_set_archive":0,"num_bytes_hit_set_archive":0,"num_flush":0,"num_flush_kb":0,"num_evict":0,"num_evict_kb":0,"num_promote":0,"num_flush_mode_high":0,"num_flush_mode_low":0,"num_evict_mode_some":0,"num_evict_mode_full":0,"num_objects_pinned":0,"num_legacy_snapsets":0,"num_large_omap_objects":0,"num_objects_manifest":0,"num_omap_bytes":0,"num_omap_keys":0,"num_objects_repaired":0},"up":[1,2,0],"acting":[1,2,0],"avail_no_missing":[],"object_location_counts":[],"blocked_by":[],"up_primary":1,"acting_primary":1,"purged_snaps":[]}],"pool_stats":[{"poolid":1,"num_pg":1,"stat_sum":{"num_bytes":459280,"num_objects":2,"num_object_clones":0,"num_object_copies":6,"num_objects_missing_on_primary":0,"num_objects_missing":0,"num_objects_degraded":0,"num_objects_misplaced":0,"num_objects_unfound":0,"num_objects_dirty":2,"num_whiteouts":0,"num_read":46,"num_read_kb":37,"num_write":57,"num_write_kb":584,"num_scrub_errors":0,"num_shallow_scrub_errors":0,"num_deep_scrub_errors":0,"num_objects_recovered":0,"num_bytes_recovered":0,"num_keys_recovered":0,"num_objects_omap":0,"num_objects_hit_set_archive":0,"num_bytes_hit_set_archive":0,"num_flush":0,"num_flush_kb":0,"num_evict":0,"num_evict_kb":0,"num_promote":0,"num_flush_mode_high":0,"num_flush_mode_low":0,"num_evict_mode_some":0,"num_evict_mode_full":0,"num_objects_pinned":0,"num_legacy_snapsets":0,"num_large_omap_objects":0,"num_objects_manifest":0,"num_omap_bytes":0,"num_omap_keys":0,"num_objects_repaired":0},"store_stats":{"total":0,"available":0,"internally_reserved":0,"allocated":1388544,"data_stored":1377840,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":0},"log_size":32,"ondisk_log_size":32,"up":3,"acting":3,"num_store_stats":3}],"osd_stats":[{"osd":2,"up_from":18,"seq":77309411331,"num_pgs":1,"num_osds":1,"num_per_pool_osds":1,"num_per_pool_omap_osds":1,"kb":20967424,"kb_used":27604,"kb_used_data":628,"kb_used_omap":1,"kb_used_meta":26814,"kb_avail":20939820,"statfs":{"total":21470642176,"available":21442375680,"internally_reserved":0,"allocated":643072,"data_stored":513724,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":1590,"internal_metadata":27457994},"hb_peers":[0,1],"snap_trim_queue_len":0,"num_snap_trimming":0,"num_shards_repaired":0,"op_queue_age_hist":{"histogram":[],"upper_bound":1},"perf_stat":{"commit_latency_ms":0,"apply_latency_ms":0,"commit_latency_ns":0,"apply_latency_ns":0},"alerts":[]},{"osd":1,"up_from":13,"seq":55834574854,"num_pgs":1,"num_osds":1,"num_per_pool_osds":1,"num_per_pool_omap_osds":1,"kb":20967424,"kb_used":27592,"kb_used_data":628,"kb_used_omap":1,"kb_used_meta":26814,"kb_avail":20939832,"statfs":{"total":21470642176,"available":21442387968,"internally_reserved":0,"allocated":643072,"data_stored":513724,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":1590,"internal_metadata":27457994},"hb_peers":[0,2],"snap_trim_queue_len":0,"num_snap_trimming":0,"num_shards_repaired":0,"op_queue_age_hist":{"histogram":[],"upper_bound":1},"perf_stat":{"commit_latency_ms":0,"apply_latency_ms":0,"commit_latency_ns":0,"apply_latency_ns":0},"alerts":[]},{"osd":0,"up_from":8,"seq":34359738378,"num_pgs":1,"num_osds":1,"num_per_pool_osds":1,"num_per_pool_omap_osds":1,"kb":20967424,"kb_used":27600,"kb_used_data":628,"kb_used_omap":1,"kb_used_meta":26814,"kb_avail":20939824,"statfs":{"total":21470642176,"available":21442379776,"internally_reserved":0,"allocated":643072,"data_stored":513724,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":1590,"internal_metadata":27457994},"hb_peers":[1,2],"snap_trim_queue_len":0,"num_snap_trimming":0,"num_shards_repaired":0,"op_queue_age_hist":{"histogram":[],"upper_bound":1},"perf_stat":{"commit_latency_ms":0,"apply_latency_ms":0,"commit_latency_ns":0,"apply_latency_ns":0},"alerts":[]}],"pool_statfs":[{"poolid":1,"osd":0,"total":0,"available":0,"internally_reserved":0,"allocated":462848,"data_stored":459280,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":0},{"poolid":1,"osd":1,"total":0,"available":0,"internally_reserved":0,"allocated":462848,"data_stored":459280,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":0},{"poolid":1,"osd":2,"total":0,"available":0,"internally_reserved":0,"allocated":462848,"data_stored":459280,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":0}]}} 2026-03-06T13:47:11.351 INFO:tasks.cephadm.ceph_manager.ceph:clean! 2026-03-06T13:47:11.351 INFO:tasks.ceph:Waiting until ceph cluster ceph is healthy... 2026-03-06T13:47:11.351 INFO:tasks.cephadm.ceph_manager.ceph:wait_until_healthy 2026-03-06T13:47:11.351 DEBUG:teuthology.orchestra.run.vm02:> sudo /home/ubuntu/cephtest/cephadm --image harbor.clyso.com/custom-ceph/ceph/ceph:cobaltcore-storage-v19.2.3-fasttrack-3 shell --fsid 2b1b4a72-195a-11f1-b1df-bf1f800d676b -- ceph health --format=json 2026-03-06T13:47:11.700 INFO:teuthology.orchestra.run.vm02.stderr:Inferring config /var/lib/ceph/2b1b4a72-195a-11f1-b1df-bf1f800d676b/mon.a/config 2026-03-06T13:47:12.010 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:47:11 vm06 ceph-mon[56528]: pgmap v64: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-06T13:47:12.010 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:47:11 vm06 ceph-mon[56528]: from='client.24248 -' entity='client.admin' cmd=[{"prefix": "pg dump", "target": ["mon-mgr", ""], "format": "json"}]: dispatch 2026-03-06T13:47:12.050 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:47:11 vm04 ceph-mon[53170]: pgmap v64: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-06T13:47:12.050 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:47:11 vm04 ceph-mon[53170]: from='client.24248 -' entity='client.admin' cmd=[{"prefix": "pg dump", "target": ["mon-mgr", ""], "format": "json"}]: dispatch 2026-03-06T13:47:12.056 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:47:11 vm02 ceph-mon[49207]: pgmap v64: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-06T13:47:12.056 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:47:11 vm02 ceph-mon[49207]: from='client.24248 -' entity='client.admin' cmd=[{"prefix": "pg dump", "target": ["mon-mgr", ""], "format": "json"}]: dispatch 2026-03-06T13:47:12.056 INFO:teuthology.orchestra.run.vm02.stdout: 2026-03-06T13:47:12.056 INFO:teuthology.orchestra.run.vm02.stdout:{"status":"HEALTH_OK","checks":{},"mutes":[]} 2026-03-06T13:47:12.237 INFO:tasks.cephadm.ceph_manager.ceph:wait_until_healthy done 2026-03-06T13:47:12.237 INFO:tasks.cephadm:Setup complete, yielding 2026-03-06T13:47:12.237 INFO:teuthology.run_tasks:Running task cephadm.apply... 2026-03-06T13:47:12.240 INFO:tasks.cephadm:Applying spec(s): placement: count: 3 service_id: foo service_type: mon spec: crush_locations: host.a: - datacenter=a host.b: - datacenter=b - rack=2 host.c: - datacenter=a - rack=3 2026-03-06T13:47:12.240 DEBUG:teuthology.orchestra.run.vm02:> sudo /home/ubuntu/cephtest/cephadm --image harbor.clyso.com/custom-ceph/ceph/ceph:cobaltcore-storage-v19.2.3-fasttrack-3 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 2b1b4a72-195a-11f1-b1df-bf1f800d676b -- ceph orch apply -i - 2026-03-06T13:47:12.577 INFO:teuthology.orchestra.run.vm02.stderr:Inferring config /var/lib/ceph/2b1b4a72-195a-11f1-b1df-bf1f800d676b/mon.a/config 2026-03-06T13:47:12.820 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:47:12 vm04 ceph-mon[53170]: from='client.24241 -' entity='client.admin' cmd=[{"prefix": "pg dump", "target": ["mon-mgr", ""], "format": "json"}]: dispatch 2026-03-06T13:47:12.820 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:47:12 vm04 ceph-mon[53170]: from='client.? 192.168.123.102:0/3838694518' entity='client.admin' cmd=[{"prefix": "health", "format": "json"}]: dispatch 2026-03-06T13:47:12.862 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:47:12 vm02 ceph-mon[49207]: from='client.24241 -' entity='client.admin' cmd=[{"prefix": "pg dump", "target": ["mon-mgr", ""], "format": "json"}]: dispatch 2026-03-06T13:47:12.862 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:47:12 vm02 ceph-mon[49207]: from='client.? 192.168.123.102:0/3838694518' entity='client.admin' cmd=[{"prefix": "health", "format": "json"}]: dispatch 2026-03-06T13:47:12.968 INFO:teuthology.orchestra.run.vm02.stdout:Scheduled mon update... 2026-03-06T13:47:13.010 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:47:12 vm06 ceph-mon[56528]: from='client.24241 -' entity='client.admin' cmd=[{"prefix": "pg dump", "target": ["mon-mgr", ""], "format": "json"}]: dispatch 2026-03-06T13:47:13.010 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:47:12 vm06 ceph-mon[56528]: from='client.? 192.168.123.102:0/3838694518' entity='client.admin' cmd=[{"prefix": "health", "format": "json"}]: dispatch 2026-03-06T13:47:13.136 INFO:teuthology.run_tasks:Running task cephadm.shell... 2026-03-06T13:47:13.138 INFO:tasks.cephadm:Running commands on role host.a host ubuntu@vm02.local 2026-03-06T13:47:13.138 DEBUG:teuthology.orchestra.run.vm02:> sudo /home/ubuntu/cephtest/cephadm --image harbor.clyso.com/custom-ceph/ceph/ceph:cobaltcore-storage-v19.2.3-fasttrack-3 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 2b1b4a72-195a-11f1-b1df-bf1f800d676b -- bash -c 'set -ex 2026-03-06T13:47:13.139 DEBUG:teuthology.orchestra.run.vm02:> # since we don'"'"'t know the real hostnames before the test, the next 2026-03-06T13:47:13.139 DEBUG:teuthology.orchestra.run.vm02:> # bit is in order to replace the fake hostnames "host.a/b/c" with 2026-03-06T13:47:13.139 DEBUG:teuthology.orchestra.run.vm02:> # the actual names cephadm knows the host by within the mon spec 2026-03-06T13:47:13.139 DEBUG:teuthology.orchestra.run.vm02:> ceph orch host ls --format json | jq -r '"'"'.[] | .hostname'"'"' > realnames 2026-03-06T13:47:13.139 DEBUG:teuthology.orchestra.run.vm02:> echo $'"'"'host.a\nhost.b\nhost.c'"'"' > fakenames 2026-03-06T13:47:13.139 DEBUG:teuthology.orchestra.run.vm02:> echo $'"'"'a\nb\nc'"'"' > mon_ids 2026-03-06T13:47:13.139 DEBUG:teuthology.orchestra.run.vm02:> echo $'"'"'{datacenter=a}\n{datacenter=b,rack=2}\n{datacenter=a,rack=3}'"'"' > crush_locs 2026-03-06T13:47:13.139 DEBUG:teuthology.orchestra.run.vm02:> ceph orch ls --service-name mon --export > mon.yaml 2026-03-06T13:47:13.139 DEBUG:teuthology.orchestra.run.vm02:> MONSPEC=`cat mon.yaml` 2026-03-06T13:47:13.139 DEBUG:teuthology.orchestra.run.vm02:> echo "$MONSPEC" 2026-03-06T13:47:13.139 DEBUG:teuthology.orchestra.run.vm02:> while read realname <&3 && read fakename <&4; do 2026-03-06T13:47:13.139 DEBUG:teuthology.orchestra.run.vm02:> MONSPEC="${MONSPEC//$fakename/$realname}" 2026-03-06T13:47:13.139 DEBUG:teuthology.orchestra.run.vm02:> done 3 echo "$MONSPEC" > mon.yaml 2026-03-06T13:47:13.139 DEBUG:teuthology.orchestra.run.vm02:> cat mon.yaml 2026-03-06T13:47:13.139 DEBUG:teuthology.orchestra.run.vm02:> # now the spec should have the real hostnames, so let'"'"'s re-apply 2026-03-06T13:47:13.139 DEBUG:teuthology.orchestra.run.vm02:> ceph orch apply -i mon.yaml 2026-03-06T13:47:13.139 DEBUG:teuthology.orchestra.run.vm02:> sleep 90 2026-03-06T13:47:13.139 DEBUG:teuthology.orchestra.run.vm02:> ceph orch ps --refresh 2026-03-06T13:47:13.139 DEBUG:teuthology.orchestra.run.vm02:> ceph orch ls --service-name mon --export > mon.yaml; ceph orch apply -i mon.yaml 2026-03-06T13:47:13.139 DEBUG:teuthology.orchestra.run.vm02:> sleep 90 2026-03-06T13:47:13.139 DEBUG:teuthology.orchestra.run.vm02:> ceph mon dump 2026-03-06T13:47:13.139 DEBUG:teuthology.orchestra.run.vm02:> ceph mon dump --format json 2026-03-06T13:47:13.139 DEBUG:teuthology.orchestra.run.vm02:> # verify all the crush locations got set from "ceph mon dump" output 2026-03-06T13:47:13.139 DEBUG:teuthology.orchestra.run.vm02:> while read monid <&3 && read crushloc <&4; do 2026-03-06T13:47:13.139 DEBUG:teuthology.orchestra.run.vm02:> ceph mon dump --format json | jq --arg monid "$monid" --arg crushloc "$crushloc" -e '"'"'.mons | .[] | select(.name == $monid) | .crush_location == $crushloc'"'"' 2026-03-06T13:47:13.139 DEBUG:teuthology.orchestra.run.vm02:> done 3 ' 2026-03-06T13:47:13.477 INFO:teuthology.orchestra.run.vm02.stderr:Inferring config /var/lib/ceph/2b1b4a72-195a-11f1-b1df-bf1f800d676b/mon.a/config 2026-03-06T13:47:13.577 INFO:teuthology.orchestra.run.vm02.stderr:+ ceph orch host ls --format json 2026-03-06T13:47:13.577 INFO:teuthology.orchestra.run.vm02.stderr:+ jq -r '.[] | .hostname' 2026-03-06T13:47:13.800 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:47:13 vm04 ceph-mon[53170]: pgmap v65: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-06T13:47:13.800 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:47:13 vm04 ceph-mon[53170]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:47:13.800 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:47:13 vm04 ceph-mon[53170]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-06T13:47:13.800 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:47:13 vm04 ceph-mon[53170]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-06T13:47:13.800 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:47:13 vm04 ceph-mon[53170]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-06T13:47:13.800 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:47:13 vm04 ceph-mon[53170]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:47:13.800 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:47:13 vm04 ceph-mon[53170]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' cmd=[{"prefix": "quorum_status"}]: dispatch 2026-03-06T13:47:13.800 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:47:13 vm04 ceph-mon[53170]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:47:13.800 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:47:13 vm04 ceph-mon[53170]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:47:13.800 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:47:13 vm04 ceph-mon[53170]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:47:13.800 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:47:13 vm04 ceph-mon[53170]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:47:13.800 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:47:13 vm04 ceph-mon[53170]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:47:13.800 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:47:13 vm04 ceph-mon[53170]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:47:13.801 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:47:13 vm04 ceph-mon[53170]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:47:13.857 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:47:13 vm02 ceph-mon[49207]: pgmap v65: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-06T13:47:13.857 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:47:13 vm02 ceph-mon[49207]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:47:13.857 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:47:13 vm02 ceph-mon[49207]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-06T13:47:13.857 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:47:13 vm02 ceph-mon[49207]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-06T13:47:13.857 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:47:13 vm02 ceph-mon[49207]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-06T13:47:13.857 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:47:13 vm02 ceph-mon[49207]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:47:13.857 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:47:13 vm02 ceph-mon[49207]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' cmd=[{"prefix": "quorum_status"}]: dispatch 2026-03-06T13:47:13.857 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:47:13 vm02 ceph-mon[49207]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:47:13.857 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:47:13 vm02 ceph-mon[49207]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:47:13.857 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:47:13 vm02 ceph-mon[49207]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:47:13.857 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:47:13 vm02 ceph-mon[49207]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:47:13.857 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:47:13 vm02 ceph-mon[49207]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:47:13.858 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:47:13 vm02 ceph-mon[49207]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:47:13.858 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:47:13 vm02 ceph-mon[49207]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:47:13.858 INFO:teuthology.orchestra.run.vm02.stderr:+ echo 'host.a 2026-03-06T13:47:13.858 INFO:teuthology.orchestra.run.vm02.stderr:host.b 2026-03-06T13:47:13.858 INFO:teuthology.orchestra.run.vm02.stderr:host.c' 2026-03-06T13:47:13.858 INFO:teuthology.orchestra.run.vm02.stderr:+ echo 'a 2026-03-06T13:47:13.858 INFO:teuthology.orchestra.run.vm02.stderr:b 2026-03-06T13:47:13.858 INFO:teuthology.orchestra.run.vm02.stderr:c' 2026-03-06T13:47:13.858 INFO:teuthology.orchestra.run.vm02.stderr:+ echo '{datacenter=a} 2026-03-06T13:47:13.858 INFO:teuthology.orchestra.run.vm02.stderr:{datacenter=b,rack=2} 2026-03-06T13:47:13.858 INFO:teuthology.orchestra.run.vm02.stderr:{datacenter=a,rack=3}' 2026-03-06T13:47:13.858 INFO:teuthology.orchestra.run.vm02.stderr:+ ceph orch ls --service-name mon --export 2026-03-06T13:47:14.010 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:47:13 vm06 ceph-mon[56528]: pgmap v65: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-06T13:47:14.010 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:47:13 vm06 ceph-mon[56528]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:47:14.010 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:47:13 vm06 ceph-mon[56528]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-06T13:47:14.010 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:47:13 vm06 ceph-mon[56528]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-06T13:47:14.010 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:47:13 vm06 ceph-mon[56528]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-06T13:47:14.010 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:47:13 vm06 ceph-mon[56528]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:47:14.010 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:47:13 vm06 ceph-mon[56528]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' cmd=[{"prefix": "quorum_status"}]: dispatch 2026-03-06T13:47:14.010 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:47:13 vm06 ceph-mon[56528]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:47:14.010 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:47:13 vm06 ceph-mon[56528]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:47:14.010 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:47:13 vm06 ceph-mon[56528]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:47:14.010 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:47:13 vm06 ceph-mon[56528]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:47:14.010 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:47:13 vm06 ceph-mon[56528]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:47:14.010 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:47:13 vm06 ceph-mon[56528]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:47:14.010 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:47:13 vm06 ceph-mon[56528]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:47:14.029 INFO:teuthology.orchestra.run.vm02.stderr:++ cat mon.yaml 2026-03-06T13:47:14.031 INFO:teuthology.orchestra.run.vm02.stderr:+ MONSPEC='service_type: mon 2026-03-06T13:47:14.031 INFO:teuthology.orchestra.run.vm02.stderr:service_name: mon 2026-03-06T13:47:14.031 INFO:teuthology.orchestra.run.vm02.stderr:placement: 2026-03-06T13:47:14.031 INFO:teuthology.orchestra.run.vm02.stderr: count: 3 2026-03-06T13:47:14.031 INFO:teuthology.orchestra.run.vm02.stderr:spec: 2026-03-06T13:47:14.031 INFO:teuthology.orchestra.run.vm02.stderr: crush_locations: 2026-03-06T13:47:14.031 INFO:teuthology.orchestra.run.vm02.stderr: host.a: 2026-03-06T13:47:14.031 INFO:teuthology.orchestra.run.vm02.stderr: - datacenter=a 2026-03-06T13:47:14.031 INFO:teuthology.orchestra.run.vm02.stderr: host.b: 2026-03-06T13:47:14.031 INFO:teuthology.orchestra.run.vm02.stderr: - datacenter=b 2026-03-06T13:47:14.031 INFO:teuthology.orchestra.run.vm02.stderr: - rack=2 2026-03-06T13:47:14.031 INFO:teuthology.orchestra.run.vm02.stderr: host.c: 2026-03-06T13:47:14.031 INFO:teuthology.orchestra.run.vm02.stderr: - datacenter=a 2026-03-06T13:47:14.031 INFO:teuthology.orchestra.run.vm02.stderr: - rack=3' 2026-03-06T13:47:14.031 INFO:teuthology.orchestra.run.vm02.stderr:+ echo 'service_type: mon 2026-03-06T13:47:14.031 INFO:teuthology.orchestra.run.vm02.stderr:service_name: mon 2026-03-06T13:47:14.032 INFO:teuthology.orchestra.run.vm02.stderr:placement: 2026-03-06T13:47:14.032 INFO:teuthology.orchestra.run.vm02.stderr: count: 3 2026-03-06T13:47:14.032 INFO:teuthology.orchestra.run.vm02.stderr:spec: 2026-03-06T13:47:14.032 INFO:teuthology.orchestra.run.vm02.stderr: crush_locations: 2026-03-06T13:47:14.032 INFO:teuthology.orchestra.run.vm02.stderr: host.a: 2026-03-06T13:47:14.032 INFO:teuthology.orchestra.run.vm02.stderr: - datacenter=a 2026-03-06T13:47:14.032 INFO:teuthology.orchestra.run.vm02.stderr: host.b: 2026-03-06T13:47:14.032 INFO:teuthology.orchestra.run.vm02.stderr: - datacenter=b 2026-03-06T13:47:14.032 INFO:teuthology.orchestra.run.vm02.stderr: - rack=2 2026-03-06T13:47:14.032 INFO:teuthology.orchestra.run.vm02.stderr: host.c: 2026-03-06T13:47:14.032 INFO:teuthology.orchestra.run.vm02.stderr: - datacenter=a 2026-03-06T13:47:14.032 INFO:teuthology.orchestra.run.vm02.stderr: - rack=3' 2026-03-06T13:47:14.032 INFO:teuthology.orchestra.run.vm02.stdout:service_type: mon 2026-03-06T13:47:14.032 INFO:teuthology.orchestra.run.vm02.stdout:service_name: mon 2026-03-06T13:47:14.032 INFO:teuthology.orchestra.run.vm02.stdout:placement: 2026-03-06T13:47:14.032 INFO:teuthology.orchestra.run.vm02.stdout: count: 3 2026-03-06T13:47:14.032 INFO:teuthology.orchestra.run.vm02.stdout:spec: 2026-03-06T13:47:14.032 INFO:teuthology.orchestra.run.vm02.stdout: crush_locations: 2026-03-06T13:47:14.032 INFO:teuthology.orchestra.run.vm02.stdout: host.a: 2026-03-06T13:47:14.032 INFO:teuthology.orchestra.run.vm02.stdout: - datacenter=a 2026-03-06T13:47:14.032 INFO:teuthology.orchestra.run.vm02.stdout: host.b: 2026-03-06T13:47:14.032 INFO:teuthology.orchestra.run.vm02.stdout: - datacenter=b 2026-03-06T13:47:14.032 INFO:teuthology.orchestra.run.vm02.stdout: - rack=2 2026-03-06T13:47:14.032 INFO:teuthology.orchestra.run.vm02.stdout: host.c: 2026-03-06T13:47:14.032 INFO:teuthology.orchestra.run.vm02.stdout: - datacenter=a 2026-03-06T13:47:14.032 INFO:teuthology.orchestra.run.vm02.stdout: - rack=3 2026-03-06T13:47:14.033 INFO:teuthology.orchestra.run.vm02.stderr:+ read realname 2026-03-06T13:47:14.033 INFO:teuthology.orchestra.run.vm02.stderr:+ read fakename 2026-03-06T13:47:14.033 INFO:teuthology.orchestra.run.vm02.stderr:+ MONSPEC='service_type: mon 2026-03-06T13:47:14.033 INFO:teuthology.orchestra.run.vm02.stderr:service_name: mon 2026-03-06T13:47:14.033 INFO:teuthology.orchestra.run.vm02.stderr:placement: 2026-03-06T13:47:14.033 INFO:teuthology.orchestra.run.vm02.stderr: count: 3 2026-03-06T13:47:14.033 INFO:teuthology.orchestra.run.vm02.stderr:spec: 2026-03-06T13:47:14.033 INFO:teuthology.orchestra.run.vm02.stderr: crush_locations: 2026-03-06T13:47:14.033 INFO:teuthology.orchestra.run.vm02.stderr: vm02: 2026-03-06T13:47:14.033 INFO:teuthology.orchestra.run.vm02.stderr: - datacenter=a 2026-03-06T13:47:14.033 INFO:teuthology.orchestra.run.vm02.stderr: host.b: 2026-03-06T13:47:14.033 INFO:teuthology.orchestra.run.vm02.stderr: - datacenter=b 2026-03-06T13:47:14.033 INFO:teuthology.orchestra.run.vm02.stderr: - rack=2 2026-03-06T13:47:14.033 INFO:teuthology.orchestra.run.vm02.stderr: host.c: 2026-03-06T13:47:14.033 INFO:teuthology.orchestra.run.vm02.stderr: - datacenter=a 2026-03-06T13:47:14.034 INFO:teuthology.orchestra.run.vm02.stderr: - rack=3' 2026-03-06T13:47:14.034 INFO:teuthology.orchestra.run.vm02.stderr:+ read realname 2026-03-06T13:47:14.034 INFO:teuthology.orchestra.run.vm02.stderr:+ read fakename 2026-03-06T13:47:14.034 INFO:teuthology.orchestra.run.vm02.stderr:+ MONSPEC='service_type: mon 2026-03-06T13:47:14.034 INFO:teuthology.orchestra.run.vm02.stderr:service_name: mon 2026-03-06T13:47:14.034 INFO:teuthology.orchestra.run.vm02.stderr:placement: 2026-03-06T13:47:14.034 INFO:teuthology.orchestra.run.vm02.stderr: count: 3 2026-03-06T13:47:14.034 INFO:teuthology.orchestra.run.vm02.stderr:spec: 2026-03-06T13:47:14.034 INFO:teuthology.orchestra.run.vm02.stderr: crush_locations: 2026-03-06T13:47:14.034 INFO:teuthology.orchestra.run.vm02.stderr: vm02: 2026-03-06T13:47:14.034 INFO:teuthology.orchestra.run.vm02.stderr: - datacenter=a 2026-03-06T13:47:14.034 INFO:teuthology.orchestra.run.vm02.stderr: vm04: 2026-03-06T13:47:14.034 INFO:teuthology.orchestra.run.vm02.stderr: - datacenter=b 2026-03-06T13:47:14.034 INFO:teuthology.orchestra.run.vm02.stderr: - rack=2 2026-03-06T13:47:14.034 INFO:teuthology.orchestra.run.vm02.stderr: host.c: 2026-03-06T13:47:14.034 INFO:teuthology.orchestra.run.vm02.stderr: - datacenter=a 2026-03-06T13:47:14.034 INFO:teuthology.orchestra.run.vm02.stderr: - rack=3' 2026-03-06T13:47:14.034 INFO:teuthology.orchestra.run.vm02.stderr:+ read realname 2026-03-06T13:47:14.034 INFO:teuthology.orchestra.run.vm02.stderr:+ read fakename 2026-03-06T13:47:14.034 INFO:teuthology.orchestra.run.vm02.stderr:+ MONSPEC='service_type: mon 2026-03-06T13:47:14.034 INFO:teuthology.orchestra.run.vm02.stderr:service_name: mon 2026-03-06T13:47:14.034 INFO:teuthology.orchestra.run.vm02.stderr:placement: 2026-03-06T13:47:14.034 INFO:teuthology.orchestra.run.vm02.stderr: count: 3 2026-03-06T13:47:14.034 INFO:teuthology.orchestra.run.vm02.stderr:spec: 2026-03-06T13:47:14.034 INFO:teuthology.orchestra.run.vm02.stderr: crush_locations: 2026-03-06T13:47:14.034 INFO:teuthology.orchestra.run.vm02.stderr: vm02: 2026-03-06T13:47:14.034 INFO:teuthology.orchestra.run.vm02.stderr: - datacenter=a 2026-03-06T13:47:14.034 INFO:teuthology.orchestra.run.vm02.stderr: vm04: 2026-03-06T13:47:14.034 INFO:teuthology.orchestra.run.vm02.stderr: - datacenter=b 2026-03-06T13:47:14.034 INFO:teuthology.orchestra.run.vm02.stderr: - rack=2 2026-03-06T13:47:14.034 INFO:teuthology.orchestra.run.vm02.stderr: vm06: 2026-03-06T13:47:14.034 INFO:teuthology.orchestra.run.vm02.stderr: - datacenter=a 2026-03-06T13:47:14.034 INFO:teuthology.orchestra.run.vm02.stderr: - rack=3' 2026-03-06T13:47:14.034 INFO:teuthology.orchestra.run.vm02.stderr:+ read realname 2026-03-06T13:47:14.034 INFO:teuthology.orchestra.run.vm02.stderr:+ echo 'service_type: mon 2026-03-06T13:47:14.034 INFO:teuthology.orchestra.run.vm02.stderr:service_name: mon 2026-03-06T13:47:14.034 INFO:teuthology.orchestra.run.vm02.stderr:placement: 2026-03-06T13:47:14.034 INFO:teuthology.orchestra.run.vm02.stderr: count: 3 2026-03-06T13:47:14.034 INFO:teuthology.orchestra.run.vm02.stderr:spec: 2026-03-06T13:47:14.034 INFO:teuthology.orchestra.run.vm02.stderr: crush_locations: 2026-03-06T13:47:14.034 INFO:teuthology.orchestra.run.vm02.stderr: vm02: 2026-03-06T13:47:14.034 INFO:teuthology.orchestra.run.vm02.stderr: - datacenter=a 2026-03-06T13:47:14.034 INFO:teuthology.orchestra.run.vm02.stderr: vm04: 2026-03-06T13:47:14.034 INFO:teuthology.orchestra.run.vm02.stderr: - datacenter=b 2026-03-06T13:47:14.034 INFO:teuthology.orchestra.run.vm02.stderr: - rack=2 2026-03-06T13:47:14.034 INFO:teuthology.orchestra.run.vm02.stderr: vm06: 2026-03-06T13:47:14.034 INFO:teuthology.orchestra.run.vm02.stderr: - datacenter=a 2026-03-06T13:47:14.034 INFO:teuthology.orchestra.run.vm02.stderr: - rack=3' 2026-03-06T13:47:14.034 INFO:teuthology.orchestra.run.vm02.stderr:+ cat mon.yaml 2026-03-06T13:47:14.034 INFO:teuthology.orchestra.run.vm02.stderr:+ ceph orch apply -i mon.yaml 2026-03-06T13:47:14.034 INFO:teuthology.orchestra.run.vm02.stdout:service_type: mon 2026-03-06T13:47:14.034 INFO:teuthology.orchestra.run.vm02.stdout:service_name: mon 2026-03-06T13:47:14.034 INFO:teuthology.orchestra.run.vm02.stdout:placement: 2026-03-06T13:47:14.034 INFO:teuthology.orchestra.run.vm02.stdout: count: 3 2026-03-06T13:47:14.035 INFO:teuthology.orchestra.run.vm02.stdout:spec: 2026-03-06T13:47:14.035 INFO:teuthology.orchestra.run.vm02.stdout: crush_locations: 2026-03-06T13:47:14.035 INFO:teuthology.orchestra.run.vm02.stdout: vm02: 2026-03-06T13:47:14.035 INFO:teuthology.orchestra.run.vm02.stdout: - datacenter=a 2026-03-06T13:47:14.035 INFO:teuthology.orchestra.run.vm02.stdout: vm04: 2026-03-06T13:47:14.035 INFO:teuthology.orchestra.run.vm02.stdout: - datacenter=b 2026-03-06T13:47:14.035 INFO:teuthology.orchestra.run.vm02.stdout: - rack=2 2026-03-06T13:47:14.035 INFO:teuthology.orchestra.run.vm02.stdout: vm06: 2026-03-06T13:47:14.035 INFO:teuthology.orchestra.run.vm02.stdout: - datacenter=a 2026-03-06T13:47:14.035 INFO:teuthology.orchestra.run.vm02.stdout: - rack=3 2026-03-06T13:47:14.201 INFO:teuthology.orchestra.run.vm02.stdout:Scheduled mon update... 2026-03-06T13:47:14.212 INFO:teuthology.orchestra.run.vm02.stderr:+ sleep 90 2026-03-06T13:47:15.550 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:47:15 vm04 ceph-mon[53170]: from='client.24253 -' entity='client.admin' cmd=[{"prefix": "orch host ls", "target": ["mon-mgr", ""], "format": "json"}]: dispatch 2026-03-06T13:47:15.550 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:47:15 vm04 ceph-mon[53170]: from='client.14403 -' entity='client.admin' cmd=[{"prefix": "orch ls", "service_name": "mon", "export": true, "target": ["mon-mgr", ""]}]: dispatch 2026-03-06T13:47:15.550 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:47:15 vm04 ceph-mon[53170]: pgmap v66: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-06T13:47:15.550 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:47:15 vm04 ceph-mon[53170]: from='client.14409 -' entity='client.admin' cmd=[{"prefix": "orch apply", "target": ["mon-mgr", ""]}]: dispatch 2026-03-06T13:47:15.550 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:47:15 vm04 ceph-mon[53170]: Saving service mon spec with placement count:3 2026-03-06T13:47:15.550 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:47:15 vm04 ceph-mon[53170]: Setting crush location for mon a to {datacenter=a} 2026-03-06T13:47:15.550 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:47:15 vm04 ceph-mon[53170]: Setting crush location for mon b to {datacenter=b,rack=2} 2026-03-06T13:47:15.550 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:47:15 vm04 ceph-mon[53170]: Setting crush location for mon c to {datacenter=a,rack=3} 2026-03-06T13:47:15.550 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:47:15 vm04 ceph-mon[53170]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' cmd='[{"prefix": "mon set_location", "name": "c", "args": ["datacenter=a", "rack=3"]}]': finished 2026-03-06T13:47:15.550 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:47:15 vm04 ceph-mon[53170]: mon.c calling monitor election 2026-03-06T13:47:15.550 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:47:15 vm04 ceph-mon[53170]: mon.b calling monitor election 2026-03-06T13:47:15.550 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:47:15 vm04 ceph-mon[53170]: mon.a calling monitor election 2026-03-06T13:47:15.550 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:47:15 vm04 ceph-mon[53170]: mon.a is new leader, mons a,c,b in quorum (ranks 0,1,2) 2026-03-06T13:47:15.550 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:47:15 vm04 ceph-mon[53170]: monmap epoch 6 2026-03-06T13:47:15.550 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:47:15 vm04 ceph-mon[53170]: fsid 2b1b4a72-195a-11f1-b1df-bf1f800d676b 2026-03-06T13:47:15.550 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:47:15 vm04 ceph-mon[53170]: last_changed 2026-03-06T12:47:14.260805+0000 2026-03-06T13:47:15.550 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:47:15 vm04 ceph-mon[53170]: created 2026-03-06T12:44:48.508617+0000 2026-03-06T13:47:15.550 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:47:15 vm04 ceph-mon[53170]: min_mon_release 19 (squid) 2026-03-06T13:47:15.550 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:47:15 vm04 ceph-mon[53170]: election_strategy: 1 2026-03-06T13:47:15.551 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:47:15 vm04 ceph-mon[53170]: 0: [v2:192.168.123.102:3300/0,v1:192.168.123.102:6789/0] mon.a; crush_location {datacenter=a} 2026-03-06T13:47:15.551 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:47:15 vm04 ceph-mon[53170]: 1: [v2:192.168.123.106:3300/0,v1:192.168.123.106:6789/0] mon.c; crush_location {datacenter=a,rack=3} 2026-03-06T13:47:15.551 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:47:15 vm04 ceph-mon[53170]: 2: [v2:192.168.123.104:3300/0,v1:192.168.123.104:6789/0] mon.b; crush_location {datacenter=b,rack=2} 2026-03-06T13:47:15.551 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:47:15 vm04 ceph-mon[53170]: fsmap 2026-03-06T13:47:15.551 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:47:15 vm04 ceph-mon[53170]: osdmap e21: 3 total, 3 up, 3 in 2026-03-06T13:47:15.551 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:47:15 vm04 ceph-mon[53170]: mgrmap e15: a(active, since 97s), standbys: b 2026-03-06T13:47:15.551 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:47:15 vm04 ceph-mon[53170]: overall HEALTH_OK 2026-03-06T13:47:15.551 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:47:15 vm04 ceph-mon[53170]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "a"}]: dispatch 2026-03-06T13:47:15.551 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:47:15 vm04 ceph-mon[53170]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-06T13:47:15.551 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:47:15 vm04 ceph-mon[53170]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "c"}]: dispatch 2026-03-06T13:47:15.551 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:47:15 vm04 ceph-mon[53170]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:47:15.551 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:47:15 vm04 ceph-mon[53170]: Reconfiguring mon.a (monmap changed)... 2026-03-06T13:47:15.551 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:47:15 vm04 ceph-mon[53170]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "mon."}]: dispatch 2026-03-06T13:47:15.551 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:47:15 vm04 ceph-mon[53170]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' cmd=[{"prefix": "config get", "who": "mon", "key": "public_network"}]: dispatch 2026-03-06T13:47:15.551 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:47:15 vm04 ceph-mon[53170]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-06T13:47:15.551 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:47:15 vm04 ceph-mon[53170]: Reconfiguring daemon mon.a on vm02 2026-03-06T13:47:15.551 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:47:15 vm04 ceph-mon[53170]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:47:15.551 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:47:15 vm04 ceph-mon[53170]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:47:15.551 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:47:15 vm04 ceph-mon[53170]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' cmd=[{"prefix": "auth get-or-create", "entity": "mgr.a", "caps": ["mon", "profile mgr", "osd", "allow *", "mds", "allow *"]}]: dispatch 2026-03-06T13:47:15.551 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:47:15 vm04 ceph-mon[53170]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' cmd=[{"prefix": "mgr services"}]: dispatch 2026-03-06T13:47:15.551 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:47:15 vm04 ceph-mon[53170]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-06T13:47:15.613 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:47:15 vm02 ceph-mon[49207]: from='client.24253 -' entity='client.admin' cmd=[{"prefix": "orch host ls", "target": ["mon-mgr", ""], "format": "json"}]: dispatch 2026-03-06T13:47:15.613 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:47:15 vm02 ceph-mon[49207]: from='client.14403 -' entity='client.admin' cmd=[{"prefix": "orch ls", "service_name": "mon", "export": true, "target": ["mon-mgr", ""]}]: dispatch 2026-03-06T13:47:15.613 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:47:15 vm02 ceph-mon[49207]: pgmap v66: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-06T13:47:15.613 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:47:15 vm02 ceph-mon[49207]: from='client.14409 -' entity='client.admin' cmd=[{"prefix": "orch apply", "target": ["mon-mgr", ""]}]: dispatch 2026-03-06T13:47:15.613 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:47:15 vm02 ceph-mon[49207]: Saving service mon spec with placement count:3 2026-03-06T13:47:15.613 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:47:15 vm02 ceph-mon[49207]: Setting crush location for mon a to {datacenter=a} 2026-03-06T13:47:15.613 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:47:15 vm02 ceph-mon[49207]: Setting crush location for mon b to {datacenter=b,rack=2} 2026-03-06T13:47:15.613 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:47:15 vm02 ceph-mon[49207]: Setting crush location for mon c to {datacenter=a,rack=3} 2026-03-06T13:47:15.613 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:47:15 vm02 ceph-mon[49207]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' cmd='[{"prefix": "mon set_location", "name": "c", "args": ["datacenter=a", "rack=3"]}]': finished 2026-03-06T13:47:15.613 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:47:15 vm02 ceph-mon[49207]: mon.c calling monitor election 2026-03-06T13:47:15.613 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:47:15 vm02 ceph-mon[49207]: mon.b calling monitor election 2026-03-06T13:47:15.613 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:47:15 vm02 ceph-mon[49207]: mon.a calling monitor election 2026-03-06T13:47:15.613 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:47:15 vm02 ceph-mon[49207]: mon.a is new leader, mons a,c,b in quorum (ranks 0,1,2) 2026-03-06T13:47:15.613 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:47:15 vm02 ceph-mon[49207]: monmap epoch 6 2026-03-06T13:47:15.613 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:47:15 vm02 ceph-mon[49207]: fsid 2b1b4a72-195a-11f1-b1df-bf1f800d676b 2026-03-06T13:47:15.613 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:47:15 vm02 ceph-mon[49207]: last_changed 2026-03-06T12:47:14.260805+0000 2026-03-06T13:47:15.613 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:47:15 vm02 ceph-mon[49207]: created 2026-03-06T12:44:48.508617+0000 2026-03-06T13:47:15.613 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:47:15 vm02 ceph-mon[49207]: min_mon_release 19 (squid) 2026-03-06T13:47:15.613 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:47:15 vm02 ceph-mon[49207]: election_strategy: 1 2026-03-06T13:47:15.613 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:47:15 vm02 ceph-mon[49207]: 0: [v2:192.168.123.102:3300/0,v1:192.168.123.102:6789/0] mon.a; crush_location {datacenter=a} 2026-03-06T13:47:15.614 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:47:15 vm02 ceph-mon[49207]: 1: [v2:192.168.123.106:3300/0,v1:192.168.123.106:6789/0] mon.c; crush_location {datacenter=a,rack=3} 2026-03-06T13:47:15.614 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:47:15 vm02 ceph-mon[49207]: 2: [v2:192.168.123.104:3300/0,v1:192.168.123.104:6789/0] mon.b; crush_location {datacenter=b,rack=2} 2026-03-06T13:47:15.614 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:47:15 vm02 ceph-mon[49207]: fsmap 2026-03-06T13:47:15.614 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:47:15 vm02 ceph-mon[49207]: osdmap e21: 3 total, 3 up, 3 in 2026-03-06T13:47:15.614 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:47:15 vm02 ceph-mon[49207]: mgrmap e15: a(active, since 97s), standbys: b 2026-03-06T13:47:15.614 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:47:15 vm02 ceph-mon[49207]: overall HEALTH_OK 2026-03-06T13:47:15.614 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:47:15 vm02 ceph-mon[49207]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "a"}]: dispatch 2026-03-06T13:47:15.614 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:47:15 vm02 ceph-mon[49207]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-06T13:47:15.614 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:47:15 vm02 ceph-mon[49207]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "c"}]: dispatch 2026-03-06T13:47:15.614 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:47:15 vm02 ceph-mon[49207]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:47:15.614 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:47:15 vm02 ceph-mon[49207]: Reconfiguring mon.a (monmap changed)... 2026-03-06T13:47:15.614 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:47:15 vm02 ceph-mon[49207]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "mon."}]: dispatch 2026-03-06T13:47:15.614 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:47:15 vm02 ceph-mon[49207]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' cmd=[{"prefix": "config get", "who": "mon", "key": "public_network"}]: dispatch 2026-03-06T13:47:15.614 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:47:15 vm02 ceph-mon[49207]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-06T13:47:15.614 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:47:15 vm02 ceph-mon[49207]: Reconfiguring daemon mon.a on vm02 2026-03-06T13:47:15.614 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:47:15 vm02 ceph-mon[49207]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:47:15.614 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:47:15 vm02 ceph-mon[49207]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:47:15.614 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:47:15 vm02 ceph-mon[49207]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' cmd=[{"prefix": "auth get-or-create", "entity": "mgr.a", "caps": ["mon", "profile mgr", "osd", "allow *", "mds", "allow *"]}]: dispatch 2026-03-06T13:47:15.614 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:47:15 vm02 ceph-mon[49207]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' cmd=[{"prefix": "mgr services"}]: dispatch 2026-03-06T13:47:15.614 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:47:15 vm02 ceph-mon[49207]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-06T13:47:15.760 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:47:15 vm06 ceph-mon[56528]: from='client.24253 -' entity='client.admin' cmd=[{"prefix": "orch host ls", "target": ["mon-mgr", ""], "format": "json"}]: dispatch 2026-03-06T13:47:15.761 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:47:15 vm06 ceph-mon[56528]: from='client.14403 -' entity='client.admin' cmd=[{"prefix": "orch ls", "service_name": "mon", "export": true, "target": ["mon-mgr", ""]}]: dispatch 2026-03-06T13:47:15.761 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:47:15 vm06 ceph-mon[56528]: pgmap v66: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-06T13:47:15.761 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:47:15 vm06 ceph-mon[56528]: from='client.14409 -' entity='client.admin' cmd=[{"prefix": "orch apply", "target": ["mon-mgr", ""]}]: dispatch 2026-03-06T13:47:15.761 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:47:15 vm06 ceph-mon[56528]: Saving service mon spec with placement count:3 2026-03-06T13:47:15.761 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:47:15 vm06 ceph-mon[56528]: Setting crush location for mon a to {datacenter=a} 2026-03-06T13:47:15.761 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:47:15 vm06 ceph-mon[56528]: Setting crush location for mon b to {datacenter=b,rack=2} 2026-03-06T13:47:15.761 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:47:15 vm06 ceph-mon[56528]: Setting crush location for mon c to {datacenter=a,rack=3} 2026-03-06T13:47:15.761 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:47:15 vm06 ceph-mon[56528]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' cmd='[{"prefix": "mon set_location", "name": "c", "args": ["datacenter=a", "rack=3"]}]': finished 2026-03-06T13:47:15.761 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:47:15 vm06 ceph-mon[56528]: mon.c calling monitor election 2026-03-06T13:47:15.761 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:47:15 vm06 ceph-mon[56528]: mon.b calling monitor election 2026-03-06T13:47:15.761 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:47:15 vm06 ceph-mon[56528]: mon.a calling monitor election 2026-03-06T13:47:15.761 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:47:15 vm06 ceph-mon[56528]: mon.a is new leader, mons a,c,b in quorum (ranks 0,1,2) 2026-03-06T13:47:15.761 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:47:15 vm06 ceph-mon[56528]: monmap epoch 6 2026-03-06T13:47:15.761 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:47:15 vm06 ceph-mon[56528]: fsid 2b1b4a72-195a-11f1-b1df-bf1f800d676b 2026-03-06T13:47:15.761 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:47:15 vm06 ceph-mon[56528]: last_changed 2026-03-06T12:47:14.260805+0000 2026-03-06T13:47:15.761 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:47:15 vm06 ceph-mon[56528]: created 2026-03-06T12:44:48.508617+0000 2026-03-06T13:47:15.761 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:47:15 vm06 ceph-mon[56528]: min_mon_release 19 (squid) 2026-03-06T13:47:15.761 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:47:15 vm06 ceph-mon[56528]: election_strategy: 1 2026-03-06T13:47:15.761 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:47:15 vm06 ceph-mon[56528]: 0: [v2:192.168.123.102:3300/0,v1:192.168.123.102:6789/0] mon.a; crush_location {datacenter=a} 2026-03-06T13:47:15.761 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:47:15 vm06 ceph-mon[56528]: 1: [v2:192.168.123.106:3300/0,v1:192.168.123.106:6789/0] mon.c; crush_location {datacenter=a,rack=3} 2026-03-06T13:47:15.761 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:47:15 vm06 ceph-mon[56528]: 2: [v2:192.168.123.104:3300/0,v1:192.168.123.104:6789/0] mon.b; crush_location {datacenter=b,rack=2} 2026-03-06T13:47:15.761 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:47:15 vm06 ceph-mon[56528]: fsmap 2026-03-06T13:47:15.761 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:47:15 vm06 ceph-mon[56528]: osdmap e21: 3 total, 3 up, 3 in 2026-03-06T13:47:15.761 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:47:15 vm06 ceph-mon[56528]: mgrmap e15: a(active, since 97s), standbys: b 2026-03-06T13:47:15.761 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:47:15 vm06 ceph-mon[56528]: overall HEALTH_OK 2026-03-06T13:47:15.761 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:47:15 vm06 ceph-mon[56528]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "a"}]: dispatch 2026-03-06T13:47:15.761 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:47:15 vm06 ceph-mon[56528]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-06T13:47:15.761 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:47:15 vm06 ceph-mon[56528]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "c"}]: dispatch 2026-03-06T13:47:15.761 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:47:15 vm06 ceph-mon[56528]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:47:15.761 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:47:15 vm06 ceph-mon[56528]: Reconfiguring mon.a (monmap changed)... 2026-03-06T13:47:15.761 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:47:15 vm06 ceph-mon[56528]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "mon."}]: dispatch 2026-03-06T13:47:15.761 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:47:15 vm06 ceph-mon[56528]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' cmd=[{"prefix": "config get", "who": "mon", "key": "public_network"}]: dispatch 2026-03-06T13:47:15.761 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:47:15 vm06 ceph-mon[56528]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-06T13:47:15.761 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:47:15 vm06 ceph-mon[56528]: Reconfiguring daemon mon.a on vm02 2026-03-06T13:47:15.761 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:47:15 vm06 ceph-mon[56528]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:47:15.761 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:47:15 vm06 ceph-mon[56528]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:47:15.761 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:47:15 vm06 ceph-mon[56528]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' cmd=[{"prefix": "auth get-or-create", "entity": "mgr.a", "caps": ["mon", "profile mgr", "osd", "allow *", "mds", "allow *"]}]: dispatch 2026-03-06T13:47:15.761 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:47:15 vm06 ceph-mon[56528]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' cmd=[{"prefix": "mgr services"}]: dispatch 2026-03-06T13:47:15.761 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:47:15 vm06 ceph-mon[56528]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-06T13:47:16.688 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:47:16 vm04 ceph-mon[53170]: Reconfiguring mgr.a (monmap changed)... 2026-03-06T13:47:16.705 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:47:16 vm04 ceph-mon[53170]: Reconfiguring daemon mgr.a on vm02 2026-03-06T13:47:16.705 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:47:16 vm04 ceph-mon[53170]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:47:16.705 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:47:16 vm04 ceph-mon[53170]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:47:16.705 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:47:16 vm04 ceph-mon[53170]: Reconfiguring osd.0 (monmap changed)... 2026-03-06T13:47:16.705 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:47:16 vm04 ceph-mon[53170]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "osd.0"}]: dispatch 2026-03-06T13:47:16.705 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:47:16 vm04 ceph-mon[53170]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-06T13:47:16.705 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:47:16 vm04 ceph-mon[53170]: Reconfiguring daemon osd.0 on vm02 2026-03-06T13:47:16.705 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:47:16 vm04 ceph-mon[53170]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:47:16.705 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:47:16 vm04 ceph-mon[53170]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:47:16.706 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:47:16 vm04 ceph-mon[53170]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "mon."}]: dispatch 2026-03-06T13:47:16.706 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:47:16 vm04 ceph-mon[53170]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' cmd=[{"prefix": "config get", "who": "mon", "key": "public_network"}]: dispatch 2026-03-06T13:47:16.706 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:47:16 vm04 ceph-mon[53170]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-06T13:47:16.760 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:47:16 vm06 ceph-mon[56528]: Reconfiguring mgr.a (monmap changed)... 2026-03-06T13:47:16.760 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:47:16 vm06 ceph-mon[56528]: Reconfiguring daemon mgr.a on vm02 2026-03-06T13:47:16.760 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:47:16 vm06 ceph-mon[56528]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:47:16.760 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:47:16 vm06 ceph-mon[56528]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:47:16.760 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:47:16 vm06 ceph-mon[56528]: Reconfiguring osd.0 (monmap changed)... 2026-03-06T13:47:16.760 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:47:16 vm06 ceph-mon[56528]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "osd.0"}]: dispatch 2026-03-06T13:47:16.760 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:47:16 vm06 ceph-mon[56528]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-06T13:47:16.760 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:47:16 vm06 ceph-mon[56528]: Reconfiguring daemon osd.0 on vm02 2026-03-06T13:47:16.760 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:47:16 vm06 ceph-mon[56528]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:47:16.760 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:47:16 vm06 ceph-mon[56528]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:47:16.760 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:47:16 vm06 ceph-mon[56528]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "mon."}]: dispatch 2026-03-06T13:47:16.760 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:47:16 vm06 ceph-mon[56528]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' cmd=[{"prefix": "config get", "who": "mon", "key": "public_network"}]: dispatch 2026-03-06T13:47:16.760 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:47:16 vm06 ceph-mon[56528]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-06T13:47:16.862 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:47:16 vm02 ceph-mon[49207]: Reconfiguring mgr.a (monmap changed)... 2026-03-06T13:47:16.862 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:47:16 vm02 ceph-mon[49207]: Reconfiguring daemon mgr.a on vm02 2026-03-06T13:47:16.862 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:47:16 vm02 ceph-mon[49207]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:47:16.862 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:47:16 vm02 ceph-mon[49207]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:47:16.862 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:47:16 vm02 ceph-mon[49207]: Reconfiguring osd.0 (monmap changed)... 2026-03-06T13:47:16.862 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:47:16 vm02 ceph-mon[49207]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "osd.0"}]: dispatch 2026-03-06T13:47:16.862 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:47:16 vm02 ceph-mon[49207]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-06T13:47:16.862 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:47:16 vm02 ceph-mon[49207]: Reconfiguring daemon osd.0 on vm02 2026-03-06T13:47:16.862 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:47:16 vm02 ceph-mon[49207]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:47:16.862 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:47:16 vm02 ceph-mon[49207]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:47:16.862 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:47:16 vm02 ceph-mon[49207]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "mon."}]: dispatch 2026-03-06T13:47:16.862 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:47:16 vm02 ceph-mon[49207]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' cmd=[{"prefix": "config get", "who": "mon", "key": "public_network"}]: dispatch 2026-03-06T13:47:16.862 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:47:16 vm02 ceph-mon[49207]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-06T13:47:17.473 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:47:17 vm04 ceph-mon[53170]: Reconfiguring mon.b (monmap changed)... 2026-03-06T13:47:17.473 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:47:17 vm04 ceph-mon[53170]: Reconfiguring daemon mon.b on vm04 2026-03-06T13:47:17.473 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:47:17 vm04 ceph-mon[53170]: pgmap v67: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-06T13:47:17.473 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:47:17 vm04 ceph-mon[53170]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:47:17.473 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:47:17 vm04 ceph-mon[53170]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:47:17.473 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:47:17 vm04 ceph-mon[53170]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' cmd=[{"prefix": "auth get-or-create", "entity": "mgr.b", "caps": ["mon", "profile mgr", "osd", "allow *", "mds", "allow *"]}]: dispatch 2026-03-06T13:47:17.473 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:47:17 vm04 ceph-mon[53170]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' cmd=[{"prefix": "mgr services"}]: dispatch 2026-03-06T13:47:17.473 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:47:17 vm04 ceph-mon[53170]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-06T13:47:17.473 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:47:17 vm04 ceph-mon[53170]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:47:17.760 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:47:17 vm06 ceph-mon[56528]: Reconfiguring mon.b (monmap changed)... 2026-03-06T13:47:17.760 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:47:17 vm06 ceph-mon[56528]: Reconfiguring daemon mon.b on vm04 2026-03-06T13:47:17.760 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:47:17 vm06 ceph-mon[56528]: pgmap v67: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-06T13:47:17.760 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:47:17 vm06 ceph-mon[56528]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:47:17.760 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:47:17 vm06 ceph-mon[56528]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:47:17.760 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:47:17 vm06 ceph-mon[56528]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' cmd=[{"prefix": "auth get-or-create", "entity": "mgr.b", "caps": ["mon", "profile mgr", "osd", "allow *", "mds", "allow *"]}]: dispatch 2026-03-06T13:47:17.760 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:47:17 vm06 ceph-mon[56528]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' cmd=[{"prefix": "mgr services"}]: dispatch 2026-03-06T13:47:17.760 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:47:17 vm06 ceph-mon[56528]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-06T13:47:17.760 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:47:17 vm06 ceph-mon[56528]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:47:17.761 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:47:17 vm06 ceph-mon[56528]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:47:17.761 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:47:17 vm06 ceph-mon[56528]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "osd.1"}]: dispatch 2026-03-06T13:47:17.761 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:47:17 vm06 ceph-mon[56528]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-06T13:47:17.799 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:47:17 vm04 ceph-mon[53170]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:47:17.799 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:47:17 vm04 ceph-mon[53170]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "osd.1"}]: dispatch 2026-03-06T13:47:17.799 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:47:17 vm04 ceph-mon[53170]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-06T13:47:17.862 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:47:17 vm02 ceph-mon[49207]: Reconfiguring mon.b (monmap changed)... 2026-03-06T13:47:17.862 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:47:17 vm02 ceph-mon[49207]: Reconfiguring daemon mon.b on vm04 2026-03-06T13:47:17.862 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:47:17 vm02 ceph-mon[49207]: pgmap v67: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-06T13:47:17.862 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:47:17 vm02 ceph-mon[49207]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:47:17.862 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:47:17 vm02 ceph-mon[49207]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:47:17.862 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:47:17 vm02 ceph-mon[49207]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' cmd=[{"prefix": "auth get-or-create", "entity": "mgr.b", "caps": ["mon", "profile mgr", "osd", "allow *", "mds", "allow *"]}]: dispatch 2026-03-06T13:47:17.862 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:47:17 vm02 ceph-mon[49207]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' cmd=[{"prefix": "mgr services"}]: dispatch 2026-03-06T13:47:17.862 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:47:17 vm02 ceph-mon[49207]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-06T13:47:17.862 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:47:17 vm02 ceph-mon[49207]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:47:17.862 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:47:17 vm02 ceph-mon[49207]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:47:17.862 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:47:17 vm02 ceph-mon[49207]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "osd.1"}]: dispatch 2026-03-06T13:47:17.862 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:47:17 vm02 ceph-mon[49207]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-06T13:47:18.760 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:47:18 vm06 ceph-mon[56528]: Reconfiguring mgr.b (monmap changed)... 2026-03-06T13:47:18.760 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:47:18 vm06 ceph-mon[56528]: Reconfiguring daemon mgr.b on vm04 2026-03-06T13:47:18.761 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:47:18 vm06 ceph-mon[56528]: Reconfiguring osd.1 (monmap changed)... 2026-03-06T13:47:18.761 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:47:18 vm06 ceph-mon[56528]: Reconfiguring daemon osd.1 on vm04 2026-03-06T13:47:18.761 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:47:18 vm06 ceph-mon[56528]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:47:18.761 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:47:18 vm06 ceph-mon[56528]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:47:18.761 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:47:18 vm06 ceph-mon[56528]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "mon."}]: dispatch 2026-03-06T13:47:18.761 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:47:18 vm06 ceph-mon[56528]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' cmd=[{"prefix": "config get", "who": "mon", "key": "public_network"}]: dispatch 2026-03-06T13:47:18.761 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:47:18 vm06 ceph-mon[56528]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-06T13:47:18.799 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:47:18 vm04 ceph-mon[53170]: Reconfiguring mgr.b (monmap changed)... 2026-03-06T13:47:18.800 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:47:18 vm04 ceph-mon[53170]: Reconfiguring daemon mgr.b on vm04 2026-03-06T13:47:18.800 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:47:18 vm04 ceph-mon[53170]: Reconfiguring osd.1 (monmap changed)... 2026-03-06T13:47:18.800 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:47:18 vm04 ceph-mon[53170]: Reconfiguring daemon osd.1 on vm04 2026-03-06T13:47:18.800 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:47:18 vm04 ceph-mon[53170]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:47:18.800 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:47:18 vm04 ceph-mon[53170]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:47:18.800 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:47:18 vm04 ceph-mon[53170]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "mon."}]: dispatch 2026-03-06T13:47:18.800 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:47:18 vm04 ceph-mon[53170]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' cmd=[{"prefix": "config get", "who": "mon", "key": "public_network"}]: dispatch 2026-03-06T13:47:18.800 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:47:18 vm04 ceph-mon[53170]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-06T13:47:18.862 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:47:18 vm02 ceph-mon[49207]: Reconfiguring mgr.b (monmap changed)... 2026-03-06T13:47:18.862 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:47:18 vm02 ceph-mon[49207]: Reconfiguring daemon mgr.b on vm04 2026-03-06T13:47:18.862 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:47:18 vm02 ceph-mon[49207]: Reconfiguring osd.1 (monmap changed)... 2026-03-06T13:47:18.862 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:47:18 vm02 ceph-mon[49207]: Reconfiguring daemon osd.1 on vm04 2026-03-06T13:47:18.862 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:47:18 vm02 ceph-mon[49207]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:47:18.862 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:47:18 vm02 ceph-mon[49207]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:47:18.862 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:47:18 vm02 ceph-mon[49207]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "mon."}]: dispatch 2026-03-06T13:47:18.862 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:47:18 vm02 ceph-mon[49207]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' cmd=[{"prefix": "config get", "who": "mon", "key": "public_network"}]: dispatch 2026-03-06T13:47:18.862 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:47:18 vm02 ceph-mon[49207]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-06T13:47:20.010 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:47:19 vm06 ceph-mon[56528]: Reconfiguring mon.c (monmap changed)... 2026-03-06T13:47:20.010 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:47:19 vm06 ceph-mon[56528]: Reconfiguring daemon mon.c on vm06 2026-03-06T13:47:20.010 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:47:19 vm06 ceph-mon[56528]: pgmap v68: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-06T13:47:20.010 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:47:19 vm06 ceph-mon[56528]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:47:20.010 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:47:19 vm06 ceph-mon[56528]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:47:20.010 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:47:19 vm06 ceph-mon[56528]: Reconfiguring osd.2 (monmap changed)... 2026-03-06T13:47:20.010 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:47:19 vm06 ceph-mon[56528]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "osd.2"}]: dispatch 2026-03-06T13:47:20.010 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:47:19 vm06 ceph-mon[56528]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-06T13:47:20.010 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:47:19 vm06 ceph-mon[56528]: Reconfiguring daemon osd.2 on vm06 2026-03-06T13:47:20.010 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:47:19 vm06 ceph-mon[56528]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:47:20.011 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:47:19 vm06 ceph-mon[56528]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:47:20.049 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:47:19 vm04 ceph-mon[53170]: Reconfiguring mon.c (monmap changed)... 2026-03-06T13:47:20.050 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:47:19 vm04 ceph-mon[53170]: Reconfiguring daemon mon.c on vm06 2026-03-06T13:47:20.050 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:47:19 vm04 ceph-mon[53170]: pgmap v68: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-06T13:47:20.050 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:47:19 vm04 ceph-mon[53170]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:47:20.050 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:47:19 vm04 ceph-mon[53170]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:47:20.050 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:47:19 vm04 ceph-mon[53170]: Reconfiguring osd.2 (monmap changed)... 2026-03-06T13:47:20.050 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:47:19 vm04 ceph-mon[53170]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "osd.2"}]: dispatch 2026-03-06T13:47:20.050 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:47:19 vm04 ceph-mon[53170]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-06T13:47:20.050 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:47:19 vm04 ceph-mon[53170]: Reconfiguring daemon osd.2 on vm06 2026-03-06T13:47:20.050 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:47:19 vm04 ceph-mon[53170]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:47:20.050 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:47:19 vm04 ceph-mon[53170]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:47:20.112 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:47:19 vm02 ceph-mon[49207]: Reconfiguring mon.c (monmap changed)... 2026-03-06T13:47:20.112 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:47:19 vm02 ceph-mon[49207]: Reconfiguring daemon mon.c on vm06 2026-03-06T13:47:20.112 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:47:19 vm02 ceph-mon[49207]: pgmap v68: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-06T13:47:20.112 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:47:19 vm02 ceph-mon[49207]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:47:20.112 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:47:19 vm02 ceph-mon[49207]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:47:20.112 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:47:19 vm02 ceph-mon[49207]: Reconfiguring osd.2 (monmap changed)... 2026-03-06T13:47:20.112 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:47:19 vm02 ceph-mon[49207]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "osd.2"}]: dispatch 2026-03-06T13:47:20.112 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:47:19 vm02 ceph-mon[49207]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-06T13:47:20.112 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:47:19 vm02 ceph-mon[49207]: Reconfiguring daemon osd.2 on vm06 2026-03-06T13:47:20.112 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:47:19 vm02 ceph-mon[49207]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:47:20.112 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:47:19 vm02 ceph-mon[49207]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:47:21.901 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:47:21 vm02 ceph-mon[49207]: pgmap v69: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-06T13:47:22.010 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:47:21 vm06 ceph-mon[56528]: pgmap v69: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-06T13:47:22.049 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:47:21 vm04 ceph-mon[53170]: pgmap v69: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-06T13:47:23.885 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:47:23 vm06 ceph-mon[56528]: pgmap v70: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-06T13:47:24.049 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:47:23 vm04 ceph-mon[53170]: pgmap v70: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-06T13:47:24.112 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:47:23 vm02 ceph-mon[49207]: pgmap v70: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-06T13:47:25.760 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:47:25 vm06 ceph-mon[56528]: pgmap v71: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-06T13:47:25.760 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:47:25 vm06 ceph-mon[56528]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:47:25.760 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:47:25 vm06 ceph-mon[56528]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:47:25.760 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:47:25 vm06 ceph-mon[56528]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:47:25.760 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:47:25 vm06 ceph-mon[56528]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:47:25.760 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:47:25 vm06 ceph-mon[56528]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:47:25.760 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:47:25 vm06 ceph-mon[56528]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:47:25.800 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:47:25 vm04 ceph-mon[53170]: pgmap v71: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-06T13:47:25.800 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:47:25 vm04 ceph-mon[53170]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:47:25.800 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:47:25 vm04 ceph-mon[53170]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:47:25.800 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:47:25 vm04 ceph-mon[53170]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:47:25.800 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:47:25 vm04 ceph-mon[53170]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:47:25.800 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:47:25 vm04 ceph-mon[53170]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:47:25.800 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:47:25 vm04 ceph-mon[53170]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:47:25.861 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:47:25 vm02 ceph-mon[49207]: pgmap v71: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-06T13:47:25.862 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:47:25 vm02 ceph-mon[49207]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:47:25.862 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:47:25 vm02 ceph-mon[49207]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:47:25.862 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:47:25 vm02 ceph-mon[49207]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:47:25.862 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:47:25 vm02 ceph-mon[49207]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:47:25.862 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:47:25 vm02 ceph-mon[49207]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:47:25.862 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:47:25 vm02 ceph-mon[49207]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:47:27.760 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:47:27 vm06 ceph-mon[56528]: pgmap v72: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-06T13:47:27.799 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:47:27 vm04 ceph-mon[53170]: pgmap v72: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-06T13:47:27.861 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:47:27 vm02 ceph-mon[49207]: pgmap v72: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-06T13:47:29.760 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:47:29 vm06 ceph-mon[56528]: pgmap v73: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-06T13:47:29.800 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:47:29 vm04 ceph-mon[53170]: pgmap v73: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-06T13:47:29.861 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:47:29 vm02 ceph-mon[49207]: pgmap v73: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-06T13:47:31.760 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:47:31 vm06 ceph-mon[56528]: pgmap v74: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-06T13:47:31.799 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:47:31 vm04 ceph-mon[53170]: pgmap v74: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-06T13:47:31.861 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:47:31 vm02 ceph-mon[49207]: pgmap v74: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-06T13:47:33.631 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:47:33 vm04 ceph-mon[53170]: pgmap v75: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-06T13:47:33.760 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:47:33 vm06 ceph-mon[56528]: pgmap v75: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-06T13:47:33.862 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:47:33 vm02 ceph-mon[49207]: pgmap v75: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-06T13:47:35.010 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:47:34 vm06 ceph-mon[56528]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:47:35.010 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:47:34 vm06 ceph-mon[56528]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:47:35.010 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:47:34 vm06 ceph-mon[56528]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:47:35.050 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:47:34 vm04 ceph-mon[53170]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:47:35.050 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:47:34 vm04 ceph-mon[53170]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:47:35.050 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:47:34 vm04 ceph-mon[53170]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:47:35.111 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:47:34 vm02 ceph-mon[49207]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:47:35.111 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:47:34 vm02 ceph-mon[49207]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:47:35.111 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:47:34 vm02 ceph-mon[49207]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:47:36.010 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:47:35 vm06 ceph-mon[56528]: pgmap v76: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-06T13:47:36.049 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:47:35 vm04 ceph-mon[53170]: pgmap v76: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-06T13:47:36.111 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:47:35 vm02 ceph-mon[49207]: pgmap v76: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-06T13:47:38.010 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:47:37 vm06 ceph-mon[56528]: pgmap v77: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-06T13:47:38.049 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:47:37 vm04 ceph-mon[53170]: pgmap v77: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-06T13:47:38.111 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:47:37 vm02 ceph-mon[49207]: pgmap v77: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-06T13:47:40.010 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:47:39 vm06 ceph-mon[56528]: pgmap v78: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-06T13:47:40.049 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:47:39 vm04 ceph-mon[53170]: pgmap v78: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-06T13:47:40.111 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:47:39 vm02 ceph-mon[49207]: pgmap v78: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-06T13:47:42.010 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:47:41 vm06 ceph-mon[56528]: pgmap v79: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-06T13:47:42.034 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:47:41 vm02 ceph-mon[49207]: pgmap v79: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-06T13:47:42.049 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:47:41 vm04 ceph-mon[53170]: pgmap v79: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-06T13:47:43.655 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:47:43 vm06 ceph-mon[56528]: pgmap v80: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-06T13:47:44.050 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:47:43 vm04 ceph-mon[53170]: pgmap v80: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-06T13:47:44.112 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:47:43 vm02 ceph-mon[49207]: pgmap v80: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-06T13:47:45.761 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:47:45 vm06 ceph-mon[56528]: pgmap v81: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-06T13:47:45.761 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:47:45 vm06 ceph-mon[56528]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:47:45.761 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:47:45 vm06 ceph-mon[56528]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:47:45.761 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:47:45 vm06 ceph-mon[56528]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:47:45.761 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:47:45 vm06 ceph-mon[56528]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:47:45.761 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:47:45 vm06 ceph-mon[56528]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:47:45.761 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:47:45 vm06 ceph-mon[56528]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:47:45.799 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:47:45 vm04 ceph-mon[53170]: pgmap v81: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-06T13:47:45.800 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:47:45 vm04 ceph-mon[53170]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:47:45.800 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:47:45 vm04 ceph-mon[53170]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:47:45.800 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:47:45 vm04 ceph-mon[53170]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:47:45.800 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:47:45 vm04 ceph-mon[53170]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:47:45.800 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:47:45 vm04 ceph-mon[53170]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:47:45.800 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:47:45 vm04 ceph-mon[53170]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:47:45.862 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:47:45 vm02 ceph-mon[49207]: pgmap v81: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-06T13:47:45.862 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:47:45 vm02 ceph-mon[49207]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:47:45.862 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:47:45 vm02 ceph-mon[49207]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:47:45.862 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:47:45 vm02 ceph-mon[49207]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:47:45.862 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:47:45 vm02 ceph-mon[49207]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:47:45.862 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:47:45 vm02 ceph-mon[49207]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:47:45.862 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:47:45 vm02 ceph-mon[49207]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:47:47.760 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:47:47 vm06 ceph-mon[56528]: pgmap v82: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-06T13:47:47.800 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:47:47 vm04 ceph-mon[53170]: pgmap v82: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-06T13:47:47.862 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:47:47 vm02 ceph-mon[49207]: pgmap v82: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-06T13:47:49.760 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:47:49 vm06 ceph-mon[56528]: pgmap v83: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-06T13:47:49.799 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:47:49 vm04 ceph-mon[53170]: pgmap v83: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-06T13:47:49.862 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:47:49 vm02 ceph-mon[49207]: pgmap v83: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-06T13:47:51.760 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:47:51 vm06 ceph-mon[56528]: pgmap v84: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-06T13:47:51.799 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:47:51 vm04 ceph-mon[53170]: pgmap v84: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-06T13:47:51.862 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:47:51 vm02 ceph-mon[49207]: pgmap v84: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-06T13:47:53.728 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:47:53 vm04 ceph-mon[53170]: pgmap v85: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-06T13:47:53.760 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:47:53 vm06 ceph-mon[56528]: pgmap v85: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-06T13:47:53.862 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:47:53 vm02 ceph-mon[49207]: pgmap v85: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-06T13:47:55.010 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:47:54 vm06 ceph-mon[56528]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:47:55.010 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:47:54 vm06 ceph-mon[56528]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:47:55.010 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:47:54 vm06 ceph-mon[56528]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:47:55.049 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:47:54 vm04 ceph-mon[53170]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:47:55.049 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:47:54 vm04 ceph-mon[53170]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:47:55.049 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:47:54 vm04 ceph-mon[53170]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:47:55.112 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:47:54 vm02 ceph-mon[49207]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:47:55.112 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:47:54 vm02 ceph-mon[49207]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:47:55.112 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:47:54 vm02 ceph-mon[49207]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:47:56.010 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:47:55 vm06 ceph-mon[56528]: pgmap v86: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-06T13:47:56.049 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:47:55 vm04 ceph-mon[53170]: pgmap v86: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-06T13:47:56.112 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:47:55 vm02 ceph-mon[49207]: pgmap v86: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-06T13:47:58.010 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:47:57 vm06 ceph-mon[56528]: pgmap v87: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-06T13:47:58.050 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:47:57 vm04 ceph-mon[53170]: pgmap v87: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-06T13:47:58.111 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:47:57 vm02 ceph-mon[49207]: pgmap v87: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-06T13:47:59.010 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:47:58 vm06 ceph-mon[56528]: pgmap v88: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-06T13:47:59.049 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:47:58 vm04 ceph-mon[53170]: pgmap v88: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-06T13:47:59.111 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:47:58 vm02 ceph-mon[49207]: pgmap v88: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-06T13:48:01.510 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:48:01 vm06 ceph-mon[56528]: pgmap v89: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-06T13:48:01.549 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:48:01 vm04 ceph-mon[53170]: pgmap v89: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-06T13:48:01.611 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:48:01 vm02 ceph-mon[49207]: pgmap v89: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-06T13:48:03.455 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:48:03 vm06 ceph-mon[56528]: pgmap v90: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-06T13:48:03.549 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:48:03 vm04 ceph-mon[53170]: pgmap v90: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-06T13:48:03.612 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:48:03 vm02 ceph-mon[49207]: pgmap v90: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-06T13:48:05.862 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:48:05 vm02 ceph-mon[49207]: pgmap v91: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-06T13:48:05.862 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:48:05 vm02 ceph-mon[49207]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:48:05.862 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:48:05 vm02 ceph-mon[49207]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:48:05.862 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:48:05 vm02 ceph-mon[49207]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:48:05.862 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:48:05 vm02 ceph-mon[49207]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:48:05.862 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:48:05 vm02 ceph-mon[49207]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:48:05.862 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:48:05 vm02 ceph-mon[49207]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:48:06.010 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:48:05 vm06 ceph-mon[56528]: pgmap v91: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-06T13:48:06.010 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:48:05 vm06 ceph-mon[56528]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:48:06.010 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:48:05 vm06 ceph-mon[56528]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:48:06.010 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:48:05 vm06 ceph-mon[56528]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:48:06.010 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:48:05 vm06 ceph-mon[56528]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:48:06.010 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:48:05 vm06 ceph-mon[56528]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:48:06.011 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:48:05 vm06 ceph-mon[56528]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:48:06.049 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:48:05 vm04 ceph-mon[53170]: pgmap v91: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-06T13:48:06.049 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:48:05 vm04 ceph-mon[53170]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:48:06.049 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:48:05 vm04 ceph-mon[53170]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:48:06.050 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:48:05 vm04 ceph-mon[53170]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:48:06.050 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:48:05 vm04 ceph-mon[53170]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:48:06.050 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:48:05 vm04 ceph-mon[53170]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:48:06.050 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:48:05 vm04 ceph-mon[53170]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:48:07.862 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:48:07 vm02 ceph-mon[49207]: pgmap v92: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-06T13:48:08.010 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:48:07 vm06 ceph-mon[56528]: pgmap v92: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-06T13:48:08.049 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:48:07 vm04 ceph-mon[53170]: pgmap v92: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-06T13:48:09.862 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:48:09 vm02 ceph-mon[49207]: pgmap v93: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-06T13:48:10.010 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:48:09 vm06 ceph-mon[56528]: pgmap v93: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-06T13:48:10.049 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:48:09 vm04 ceph-mon[53170]: pgmap v93: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-06T13:48:11.862 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:48:11 vm02 ceph-mon[49207]: pgmap v94: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-06T13:48:12.010 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:48:11 vm06 ceph-mon[56528]: pgmap v94: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-06T13:48:12.049 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:48:11 vm04 ceph-mon[53170]: pgmap v94: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-06T13:48:13.960 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:48:13 vm04 ceph-mon[53170]: pgmap v95: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-06T13:48:14.010 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:48:13 vm06 ceph-mon[56528]: pgmap v95: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-06T13:48:14.111 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:48:13 vm02 ceph-mon[49207]: pgmap v95: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-06T13:48:15.112 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:48:14 vm02 ceph-mon[49207]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:48:15.112 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:48:14 vm02 ceph-mon[49207]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:48:15.112 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:48:14 vm02 ceph-mon[49207]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:48:15.112 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:48:14 vm02 ceph-mon[49207]: pgmap v96: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-06T13:48:15.260 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:48:14 vm06 ceph-mon[56528]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:48:15.260 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:48:14 vm06 ceph-mon[56528]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:48:15.260 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:48:14 vm06 ceph-mon[56528]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:48:15.260 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:48:14 vm06 ceph-mon[56528]: pgmap v96: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-06T13:48:15.300 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:48:14 vm04 ceph-mon[53170]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:48:15.300 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:48:14 vm04 ceph-mon[53170]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:48:15.300 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:48:14 vm04 ceph-mon[53170]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:48:15.300 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:48:14 vm04 ceph-mon[53170]: pgmap v96: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-06T13:48:17.510 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:48:17 vm06 ceph-mon[56528]: pgmap v97: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-06T13:48:17.549 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:48:17 vm04 ceph-mon[53170]: pgmap v97: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-06T13:48:17.611 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:48:17 vm02 ceph-mon[49207]: pgmap v97: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-06T13:48:19.510 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:48:19 vm06 ceph-mon[56528]: pgmap v98: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-06T13:48:19.510 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:48:19 vm06 ceph-mon[56528]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-06T13:48:19.510 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:48:19 vm06 ceph-mon[56528]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-06T13:48:19.510 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:48:19 vm06 ceph-mon[56528]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-06T13:48:19.510 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:48:19 vm06 ceph-mon[56528]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:48:19.550 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:48:19 vm04 ceph-mon[53170]: pgmap v98: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-06T13:48:19.550 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:48:19 vm04 ceph-mon[53170]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-06T13:48:19.550 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:48:19 vm04 ceph-mon[53170]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-06T13:48:19.550 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:48:19 vm04 ceph-mon[53170]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-06T13:48:19.550 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:48:19 vm04 ceph-mon[53170]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:48:19.612 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:48:19 vm02 ceph-mon[49207]: pgmap v98: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-06T13:48:19.612 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:48:19 vm02 ceph-mon[49207]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-06T13:48:19.612 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:48:19 vm02 ceph-mon[49207]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-06T13:48:19.612 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:48:19 vm02 ceph-mon[49207]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-06T13:48:19.612 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:48:19 vm02 ceph-mon[49207]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:48:21.549 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:48:21 vm04 ceph-mon[53170]: pgmap v99: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-06T13:48:21.611 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:48:21 vm02 ceph-mon[49207]: pgmap v99: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-06T13:48:21.760 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:48:21 vm06 ceph-mon[56528]: pgmap v99: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-06T13:48:23.549 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:48:23 vm04 ceph-mon[53170]: pgmap v100: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-06T13:48:23.578 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:48:23 vm02 ceph-mon[49207]: pgmap v100: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-06T13:48:23.580 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:48:23 vm06 ceph-mon[56528]: pgmap v100: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-06T13:48:26.010 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:48:25 vm06 ceph-mon[56528]: pgmap v101: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-06T13:48:26.010 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:48:25 vm06 ceph-mon[56528]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:48:26.010 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:48:25 vm06 ceph-mon[56528]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:48:26.011 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:48:25 vm06 ceph-mon[56528]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:48:26.011 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:48:25 vm06 ceph-mon[56528]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:48:26.011 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:48:25 vm06 ceph-mon[56528]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:48:26.011 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:48:25 vm06 ceph-mon[56528]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:48:26.049 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:48:25 vm04 ceph-mon[53170]: pgmap v101: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-06T13:48:26.049 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:48:25 vm04 ceph-mon[53170]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:48:26.049 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:48:25 vm04 ceph-mon[53170]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:48:26.050 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:48:25 vm04 ceph-mon[53170]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:48:26.050 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:48:25 vm04 ceph-mon[53170]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:48:26.050 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:48:25 vm04 ceph-mon[53170]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:48:26.050 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:48:25 vm04 ceph-mon[53170]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:48:26.112 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:48:25 vm02 ceph-mon[49207]: pgmap v101: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-06T13:48:26.112 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:48:25 vm02 ceph-mon[49207]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:48:26.112 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:48:25 vm02 ceph-mon[49207]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:48:26.112 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:48:25 vm02 ceph-mon[49207]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:48:26.112 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:48:25 vm02 ceph-mon[49207]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:48:26.112 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:48:25 vm02 ceph-mon[49207]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:48:26.112 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:48:25 vm02 ceph-mon[49207]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:48:28.010 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:48:27 vm06 ceph-mon[56528]: pgmap v102: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-06T13:48:28.049 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:48:27 vm04 ceph-mon[53170]: pgmap v102: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-06T13:48:28.111 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:48:27 vm02 ceph-mon[49207]: pgmap v102: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-06T13:48:30.010 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:48:29 vm06 ceph-mon[56528]: pgmap v103: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-06T13:48:30.049 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:48:29 vm04 ceph-mon[53170]: pgmap v103: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-06T13:48:30.111 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:48:29 vm02 ceph-mon[49207]: pgmap v103: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-06T13:48:32.010 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:48:31 vm06 ceph-mon[56528]: pgmap v104: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-06T13:48:32.049 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:48:31 vm04 ceph-mon[53170]: pgmap v104: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-06T13:48:32.111 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:48:31 vm02 ceph-mon[49207]: pgmap v104: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-06T13:48:34.010 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:48:33 vm06 ceph-mon[56528]: pgmap v105: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-06T13:48:34.048 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:48:33 vm04 ceph-mon[53170]: pgmap v105: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-06T13:48:34.111 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:48:33 vm02 ceph-mon[49207]: pgmap v105: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-06T13:48:35.260 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:48:34 vm06 ceph-mon[56528]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:48:35.260 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:48:34 vm06 ceph-mon[56528]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:48:35.260 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:48:34 vm06 ceph-mon[56528]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:48:35.260 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:48:34 vm06 ceph-mon[56528]: pgmap v106: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-06T13:48:35.300 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:48:34 vm04 ceph-mon[53170]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:48:35.300 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:48:34 vm04 ceph-mon[53170]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:48:35.300 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:48:34 vm04 ceph-mon[53170]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:48:35.300 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:48:34 vm04 ceph-mon[53170]: pgmap v106: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-06T13:48:35.361 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:48:34 vm02 ceph-mon[49207]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:48:35.361 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:48:34 vm02 ceph-mon[49207]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:48:35.362 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:48:34 vm02 ceph-mon[49207]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:48:35.362 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:48:34 vm02 ceph-mon[49207]: pgmap v106: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-06T13:48:37.510 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:48:37 vm06 ceph-mon[56528]: pgmap v107: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-06T13:48:37.550 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:48:37 vm04 ceph-mon[53170]: pgmap v107: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-06T13:48:37.611 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:48:37 vm02 ceph-mon[49207]: pgmap v107: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-06T13:48:39.510 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:48:39 vm06 ceph-mon[56528]: pgmap v108: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-06T13:48:39.550 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:48:39 vm04 ceph-mon[53170]: pgmap v108: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-06T13:48:39.611 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:48:39 vm02 ceph-mon[49207]: pgmap v108: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-06T13:48:41.550 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:48:41 vm04 ceph-mon[53170]: pgmap v109: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-06T13:48:41.611 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:48:41 vm02 ceph-mon[49207]: pgmap v109: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-06T13:48:41.760 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:48:41 vm06 ceph-mon[56528]: pgmap v109: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-06T13:48:43.518 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:48:43 vm02 ceph-mon[49207]: pgmap v110: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-06T13:48:43.525 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:48:43 vm06 ceph-mon[56528]: pgmap v110: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-06T13:48:43.550 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:48:43 vm04 ceph-mon[53170]: pgmap v110: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-06T13:48:44.215 INFO:teuthology.orchestra.run.vm02.stderr:+ ceph orch ps --refresh 2026-03-06T13:48:44.379 INFO:teuthology.orchestra.run.vm02.stdout:NAME HOST PORTS STATUS REFRESHED AGE MEM USE MEM LIM VERSION IMAGE ID CONTAINER ID 2026-03-06T13:48:44.379 INFO:teuthology.orchestra.run.vm02.stdout:agent.vm02 vm02 running 19s ago 3m - - 2026-03-06T13:48:44.379 INFO:teuthology.orchestra.run.vm02.stdout:agent.vm04 vm04 running 10s ago 2m - - 2026-03-06T13:48:44.379 INFO:teuthology.orchestra.run.vm02.stdout:agent.vm06 vm06 running 19s ago 2m - - 2026-03-06T13:48:44.379 INFO:teuthology.orchestra.run.vm02.stdout:mgr.a vm02 *:9283,8765 running (3m) 19s ago 3m 548M - 19.2.3-47-gc24117fd552 306e97de47e9 6b7b3227c247 2026-03-06T13:48:44.379 INFO:teuthology.orchestra.run.vm02.stdout:mgr.b vm04 *:8443,8765 running (2m) 10s ago 2m 475M - 19.2.3-47-gc24117fd552 306e97de47e9 dfe14d6f23c7 2026-03-06T13:48:44.379 INFO:teuthology.orchestra.run.vm02.stdout:mon.a vm02 running (3m) 19s ago 3m 46.9M 2048M 19.2.3-47-gc24117fd552 306e97de47e9 617d29b525ba 2026-03-06T13:48:44.379 INFO:teuthology.orchestra.run.vm02.stdout:mon.b vm04 running (2m) 10s ago 2m 43.2M 2048M 19.2.3-47-gc24117fd552 306e97de47e9 3b09de05b312 2026-03-06T13:48:44.379 INFO:teuthology.orchestra.run.vm02.stdout:mon.c vm06 running (2m) 19s ago 2m 46.7M 2048M 19.2.3-47-gc24117fd552 306e97de47e9 8defa8e110bb 2026-03-06T13:48:44.379 INFO:teuthology.orchestra.run.vm02.stdout:osd.0 vm02 running (2m) 19s ago 2m 37.9M 4096M 19.2.3-47-gc24117fd552 306e97de47e9 23e922edba97 2026-03-06T13:48:44.379 INFO:teuthology.orchestra.run.vm02.stdout:osd.1 vm04 running (2m) 10s ago 2m 39.8M 4096M 19.2.3-47-gc24117fd552 306e97de47e9 7adc2cdd55ae 2026-03-06T13:48:44.379 INFO:teuthology.orchestra.run.vm02.stdout:osd.2 vm06 running (108s) 19s ago 108s 59.1M 3329M 19.2.3-47-gc24117fd552 306e97de47e9 954e3d67a9dc 2026-03-06T13:48:44.402 INFO:teuthology.orchestra.run.vm02.stderr:+ ceph orch ls --service-name mon --export 2026-03-06T13:48:44.568 INFO:teuthology.orchestra.run.vm02.stderr:+ ceph orch apply -i mon.yaml 2026-03-06T13:48:44.746 INFO:teuthology.orchestra.run.vm02.stdout:Scheduled mon update... 2026-03-06T13:48:44.756 INFO:teuthology.orchestra.run.vm02.stderr:+ sleep 90 2026-03-06T13:48:45.550 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:48:45 vm04 ceph-mon[53170]: pgmap v111: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-06T13:48:45.550 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:48:45 vm04 ceph-mon[53170]: from='client.14415 -' entity='client.admin' cmd=[{"prefix": "orch ps", "refresh": true, "target": ["mon-mgr", ""]}]: dispatch 2026-03-06T13:48:45.550 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:48:45 vm04 ceph-mon[53170]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-06T13:48:45.550 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:48:45 vm04 ceph-mon[53170]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-06T13:48:45.550 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:48:45 vm04 ceph-mon[53170]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-06T13:48:45.550 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:48:45 vm04 ceph-mon[53170]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:48:45.550 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:48:45 vm04 ceph-mon[53170]: from='client.14421 -' entity='client.admin' cmd=[{"prefix": "orch ls", "service_name": "mon", "export": true, "target": ["mon-mgr", ""]}]: dispatch 2026-03-06T13:48:45.550 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:48:45 vm04 ceph-mon[53170]: from='client.14427 -' entity='client.admin' cmd=[{"prefix": "orch apply", "target": ["mon-mgr", ""]}]: dispatch 2026-03-06T13:48:45.550 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:48:45 vm04 ceph-mon[53170]: Saving service mon spec with placement count:3 2026-03-06T13:48:45.550 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:48:45 vm04 ceph-mon[53170]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:48:45.550 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:48:45 vm04 ceph-mon[53170]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-06T13:48:45.550 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:48:45 vm04 ceph-mon[53170]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-06T13:48:45.550 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:48:45 vm04 ceph-mon[53170]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-06T13:48:45.550 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:48:45 vm04 ceph-mon[53170]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:48:45.550 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:48:45 vm04 ceph-mon[53170]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' cmd=[{"prefix": "quorum_status"}]: dispatch 2026-03-06T13:48:45.550 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:48:45 vm04 ceph-mon[53170]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:48:45.550 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:48:45 vm04 ceph-mon[53170]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:48:45.550 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:48:45 vm04 ceph-mon[53170]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:48:45.550 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:48:45 vm04 ceph-mon[53170]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:48:45.550 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:48:45 vm04 ceph-mon[53170]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:48:45.550 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:48:45 vm04 ceph-mon[53170]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:48:45.550 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:48:45 vm04 ceph-mon[53170]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:48:45.612 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:48:45 vm02 ceph-mon[49207]: pgmap v111: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-06T13:48:45.612 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:48:45 vm02 ceph-mon[49207]: from='client.14415 -' entity='client.admin' cmd=[{"prefix": "orch ps", "refresh": true, "target": ["mon-mgr", ""]}]: dispatch 2026-03-06T13:48:45.612 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:48:45 vm02 ceph-mon[49207]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-06T13:48:45.612 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:48:45 vm02 ceph-mon[49207]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-06T13:48:45.612 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:48:45 vm02 ceph-mon[49207]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-06T13:48:45.612 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:48:45 vm02 ceph-mon[49207]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:48:45.612 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:48:45 vm02 ceph-mon[49207]: from='client.14421 -' entity='client.admin' cmd=[{"prefix": "orch ls", "service_name": "mon", "export": true, "target": ["mon-mgr", ""]}]: dispatch 2026-03-06T13:48:45.612 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:48:45 vm02 ceph-mon[49207]: from='client.14427 -' entity='client.admin' cmd=[{"prefix": "orch apply", "target": ["mon-mgr", ""]}]: dispatch 2026-03-06T13:48:45.612 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:48:45 vm02 ceph-mon[49207]: Saving service mon spec with placement count:3 2026-03-06T13:48:45.612 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:48:45 vm02 ceph-mon[49207]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:48:45.612 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:48:45 vm02 ceph-mon[49207]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-06T13:48:45.612 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:48:45 vm02 ceph-mon[49207]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-06T13:48:45.612 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:48:45 vm02 ceph-mon[49207]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-06T13:48:45.612 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:48:45 vm02 ceph-mon[49207]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:48:45.612 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:48:45 vm02 ceph-mon[49207]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' cmd=[{"prefix": "quorum_status"}]: dispatch 2026-03-06T13:48:45.612 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:48:45 vm02 ceph-mon[49207]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:48:45.612 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:48:45 vm02 ceph-mon[49207]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:48:45.612 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:48:45 vm02 ceph-mon[49207]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:48:45.612 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:48:45 vm02 ceph-mon[49207]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:48:45.612 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:48:45 vm02 ceph-mon[49207]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:48:45.612 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:48:45 vm02 ceph-mon[49207]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:48:45.612 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:48:45 vm02 ceph-mon[49207]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:48:45.760 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:48:45 vm06 ceph-mon[56528]: pgmap v111: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-06T13:48:45.760 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:48:45 vm06 ceph-mon[56528]: from='client.14415 -' entity='client.admin' cmd=[{"prefix": "orch ps", "refresh": true, "target": ["mon-mgr", ""]}]: dispatch 2026-03-06T13:48:45.760 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:48:45 vm06 ceph-mon[56528]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-06T13:48:45.760 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:48:45 vm06 ceph-mon[56528]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-06T13:48:45.760 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:48:45 vm06 ceph-mon[56528]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-06T13:48:45.760 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:48:45 vm06 ceph-mon[56528]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:48:45.760 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:48:45 vm06 ceph-mon[56528]: from='client.14421 -' entity='client.admin' cmd=[{"prefix": "orch ls", "service_name": "mon", "export": true, "target": ["mon-mgr", ""]}]: dispatch 2026-03-06T13:48:45.760 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:48:45 vm06 ceph-mon[56528]: from='client.14427 -' entity='client.admin' cmd=[{"prefix": "orch apply", "target": ["mon-mgr", ""]}]: dispatch 2026-03-06T13:48:45.760 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:48:45 vm06 ceph-mon[56528]: Saving service mon spec with placement count:3 2026-03-06T13:48:45.761 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:48:45 vm06 ceph-mon[56528]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:48:45.761 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:48:45 vm06 ceph-mon[56528]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-06T13:48:45.761 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:48:45 vm06 ceph-mon[56528]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-06T13:48:45.761 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:48:45 vm06 ceph-mon[56528]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-06T13:48:45.761 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:48:45 vm06 ceph-mon[56528]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:48:45.761 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:48:45 vm06 ceph-mon[56528]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' cmd=[{"prefix": "quorum_status"}]: dispatch 2026-03-06T13:48:45.761 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:48:45 vm06 ceph-mon[56528]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:48:45.761 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:48:45 vm06 ceph-mon[56528]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:48:45.761 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:48:45 vm06 ceph-mon[56528]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:48:45.761 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:48:45 vm06 ceph-mon[56528]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:48:45.761 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:48:45 vm06 ceph-mon[56528]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:48:45.761 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:48:45 vm06 ceph-mon[56528]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:48:45.761 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:48:45 vm06 ceph-mon[56528]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:48:47.549 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:48:47 vm04 ceph-mon[53170]: pgmap v112: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-06T13:48:47.611 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:48:47 vm02 ceph-mon[49207]: pgmap v112: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-06T13:48:47.760 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:48:47 vm06 ceph-mon[56528]: pgmap v112: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-06T13:48:49.550 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:48:49 vm04 ceph-mon[53170]: pgmap v113: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-06T13:48:49.611 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:48:49 vm02 ceph-mon[49207]: pgmap v113: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-06T13:48:49.760 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:48:49 vm06 ceph-mon[56528]: pgmap v113: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-06T13:48:51.550 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:48:51 vm04 ceph-mon[53170]: pgmap v114: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-06T13:48:51.611 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:48:51 vm02 ceph-mon[49207]: pgmap v114: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-06T13:48:51.760 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:48:51 vm06 ceph-mon[56528]: pgmap v114: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-06T13:48:53.475 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:48:53 vm04 ceph-mon[53170]: pgmap v115: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-06T13:48:53.611 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:48:53 vm02 ceph-mon[49207]: pgmap v115: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-06T13:48:53.760 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:48:53 vm06 ceph-mon[56528]: pgmap v115: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-06T13:48:55.361 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:48:55 vm02 ceph-mon[49207]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:48:55.362 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:48:55 vm02 ceph-mon[49207]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:48:55.362 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:48:55 vm02 ceph-mon[49207]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:48:55.362 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:48:55 vm02 ceph-mon[49207]: pgmap v116: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-06T13:48:55.510 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:48:55 vm06 ceph-mon[56528]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:48:55.510 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:48:55 vm06 ceph-mon[56528]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:48:55.511 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:48:55 vm06 ceph-mon[56528]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:48:55.511 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:48:55 vm06 ceph-mon[56528]: pgmap v116: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-06T13:48:55.550 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:48:55 vm04 ceph-mon[53170]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:48:55.550 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:48:55 vm04 ceph-mon[53170]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:48:55.550 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:48:55 vm04 ceph-mon[53170]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:48:55.550 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:48:55 vm04 ceph-mon[53170]: pgmap v116: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-06T13:48:57.510 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:48:57 vm06 ceph-mon[56528]: pgmap v117: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-06T13:48:57.550 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:48:57 vm04 ceph-mon[53170]: pgmap v117: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-06T13:48:57.612 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:48:57 vm02 ceph-mon[49207]: pgmap v117: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-06T13:48:59.550 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:48:59 vm04 ceph-mon[53170]: pgmap v118: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-06T13:48:59.612 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:48:59 vm02 ceph-mon[49207]: pgmap v118: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-06T13:48:59.760 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:48:59 vm06 ceph-mon[56528]: pgmap v118: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-06T13:49:01.550 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:49:01 vm04 ceph-mon[53170]: pgmap v119: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-06T13:49:01.611 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:49:01 vm02 ceph-mon[49207]: pgmap v119: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-06T13:49:01.760 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:49:01 vm06 ceph-mon[56528]: pgmap v119: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-06T13:49:03.276 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:49:03 vm02 ceph-mon[49207]: pgmap v120: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-06T13:49:03.550 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:49:03 vm04 ceph-mon[53170]: pgmap v120: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-06T13:49:03.556 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:49:03 vm06 ceph-mon[56528]: pgmap v120: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-06T13:49:05.550 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:49:05 vm04 ceph-mon[53170]: pgmap v121: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-06T13:49:05.550 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:49:05 vm04 ceph-mon[53170]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:49:05.550 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:49:05 vm04 ceph-mon[53170]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:49:05.550 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:49:05 vm04 ceph-mon[53170]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:49:05.550 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:49:05 vm04 ceph-mon[53170]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:49:05.550 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:49:05 vm04 ceph-mon[53170]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:49:05.550 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:49:05 vm04 ceph-mon[53170]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:49:05.612 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:49:05 vm02 ceph-mon[49207]: pgmap v121: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-06T13:49:05.612 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:49:05 vm02 ceph-mon[49207]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:49:05.612 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:49:05 vm02 ceph-mon[49207]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:49:05.612 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:49:05 vm02 ceph-mon[49207]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:49:05.612 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:49:05 vm02 ceph-mon[49207]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:49:05.612 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:49:05 vm02 ceph-mon[49207]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:49:05.612 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:49:05 vm02 ceph-mon[49207]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:49:05.760 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:49:05 vm06 ceph-mon[56528]: pgmap v121: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-06T13:49:05.760 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:49:05 vm06 ceph-mon[56528]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:49:05.760 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:49:05 vm06 ceph-mon[56528]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:49:05.760 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:49:05 vm06 ceph-mon[56528]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:49:05.760 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:49:05 vm06 ceph-mon[56528]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:49:05.760 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:49:05 vm06 ceph-mon[56528]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:49:05.760 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:49:05 vm06 ceph-mon[56528]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:49:07.612 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:49:07 vm02 ceph-mon[49207]: pgmap v122: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-06T13:49:07.760 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:49:07 vm06 ceph-mon[56528]: pgmap v122: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-06T13:49:07.799 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:49:07 vm04 ceph-mon[53170]: pgmap v122: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-06T13:49:09.612 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:49:09 vm02 ceph-mon[49207]: pgmap v123: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-06T13:49:09.760 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:49:09 vm06 ceph-mon[56528]: pgmap v123: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-06T13:49:09.799 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:49:09 vm04 ceph-mon[53170]: pgmap v123: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-06T13:49:11.612 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:49:11 vm02 ceph-mon[49207]: pgmap v124: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-06T13:49:11.760 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:49:11 vm06 ceph-mon[56528]: pgmap v124: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-06T13:49:11.799 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:49:11 vm04 ceph-mon[53170]: pgmap v124: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-06T13:49:13.549 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:49:13 vm04 ceph-mon[53170]: pgmap v125: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-06T13:49:13.611 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:49:13 vm02 ceph-mon[49207]: pgmap v125: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-06T13:49:13.760 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:49:13 vm06 ceph-mon[56528]: pgmap v125: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-06T13:49:15.612 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:49:15 vm02 ceph-mon[49207]: pgmap v126: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-06T13:49:15.612 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:49:15 vm02 ceph-mon[49207]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:49:15.612 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:49:15 vm02 ceph-mon[49207]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:49:15.612 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:49:15 vm02 ceph-mon[49207]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:49:15.760 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:49:15 vm06 ceph-mon[56528]: pgmap v126: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-06T13:49:15.760 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:49:15 vm06 ceph-mon[56528]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:49:15.760 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:49:15 vm06 ceph-mon[56528]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:49:15.760 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:49:15 vm06 ceph-mon[56528]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:49:15.800 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:49:15 vm04 ceph-mon[53170]: pgmap v126: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-06T13:49:15.800 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:49:15 vm04 ceph-mon[53170]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:49:15.800 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:49:15 vm04 ceph-mon[53170]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:49:15.800 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:49:15 vm04 ceph-mon[53170]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:49:17.611 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:49:17 vm02 ceph-mon[49207]: pgmap v127: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-06T13:49:17.760 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:49:17 vm06 ceph-mon[56528]: pgmap v127: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-06T13:49:17.799 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:49:17 vm04 ceph-mon[53170]: pgmap v127: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-06T13:49:19.760 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:49:19 vm06 ceph-mon[56528]: pgmap v128: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-06T13:49:19.799 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:49:19 vm04 ceph-mon[53170]: pgmap v128: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-06T13:49:19.862 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:49:19 vm02 ceph-mon[49207]: pgmap v128: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-06T13:49:21.760 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:49:21 vm06 ceph-mon[56528]: pgmap v129: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-06T13:49:21.799 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:49:21 vm04 ceph-mon[53170]: pgmap v129: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-06T13:49:21.862 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:49:21 vm02 ceph-mon[49207]: pgmap v129: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-06T13:49:23.669 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:49:23 vm02 ceph-mon[49207]: pgmap v130: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-06T13:49:23.760 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:49:23 vm06 ceph-mon[56528]: pgmap v130: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-06T13:49:23.799 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:49:23 vm04 ceph-mon[53170]: pgmap v130: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-06T13:49:25.510 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:49:25 vm06 ceph-mon[56528]: pgmap v131: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-06T13:49:25.510 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:49:25 vm06 ceph-mon[56528]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:49:25.510 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:49:25 vm06 ceph-mon[56528]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:49:25.510 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:49:25 vm06 ceph-mon[56528]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:49:25.510 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:49:25 vm06 ceph-mon[56528]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:49:25.510 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:49:25 vm06 ceph-mon[56528]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:49:25.510 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:49:25 vm06 ceph-mon[56528]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:49:25.799 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:49:25 vm04 ceph-mon[53170]: pgmap v131: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-06T13:49:25.799 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:49:25 vm04 ceph-mon[53170]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:49:25.800 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:49:25 vm04 ceph-mon[53170]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:49:25.800 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:49:25 vm04 ceph-mon[53170]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:49:25.800 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:49:25 vm04 ceph-mon[53170]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:49:25.800 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:49:25 vm04 ceph-mon[53170]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:49:25.800 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:49:25 vm04 ceph-mon[53170]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:49:25.862 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:49:25 vm02 ceph-mon[49207]: pgmap v131: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-06T13:49:25.862 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:49:25 vm02 ceph-mon[49207]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:49:25.862 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:49:25 vm02 ceph-mon[49207]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:49:25.862 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:49:25 vm02 ceph-mon[49207]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:49:25.862 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:49:25 vm02 ceph-mon[49207]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:49:25.862 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:49:25 vm02 ceph-mon[49207]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:49:25.862 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:49:25 vm02 ceph-mon[49207]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:49:27.760 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:49:27 vm06 ceph-mon[56528]: pgmap v132: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-06T13:49:27.799 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:49:27 vm04 ceph-mon[53170]: pgmap v132: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-06T13:49:27.861 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:49:27 vm02 ceph-mon[49207]: pgmap v132: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-06T13:49:29.760 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:49:29 vm06 ceph-mon[56528]: pgmap v133: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-06T13:49:29.799 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:49:29 vm04 ceph-mon[53170]: pgmap v133: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-06T13:49:29.861 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:49:29 vm02 ceph-mon[49207]: pgmap v133: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-06T13:49:31.760 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:49:31 vm06 ceph-mon[56528]: pgmap v134: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-06T13:49:31.799 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:49:31 vm04 ceph-mon[53170]: pgmap v134: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-06T13:49:31.861 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:49:31 vm02 ceph-mon[49207]: pgmap v134: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-06T13:49:33.764 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:49:33 vm04 ceph-mon[53170]: pgmap v135: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-06T13:49:33.862 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:49:33 vm02 ceph-mon[49207]: pgmap v135: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-06T13:49:34.010 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:49:33 vm06 ceph-mon[56528]: pgmap v135: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-06T13:49:36.010 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:49:35 vm06 ceph-mon[56528]: pgmap v136: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-06T13:49:36.010 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:49:35 vm06 ceph-mon[56528]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:49:36.010 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:49:35 vm06 ceph-mon[56528]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:49:36.010 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:49:35 vm06 ceph-mon[56528]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:49:36.050 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:49:35 vm04 ceph-mon[53170]: pgmap v136: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-06T13:49:36.050 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:49:35 vm04 ceph-mon[53170]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:49:36.050 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:49:35 vm04 ceph-mon[53170]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:49:36.050 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:49:35 vm04 ceph-mon[53170]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:49:36.112 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:49:35 vm02 ceph-mon[49207]: pgmap v136: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-06T13:49:36.112 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:49:35 vm02 ceph-mon[49207]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:49:36.112 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:49:35 vm02 ceph-mon[49207]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:49:36.112 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:49:35 vm02 ceph-mon[49207]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:49:38.010 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:49:37 vm06 ceph-mon[56528]: pgmap v137: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-06T13:49:38.049 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:49:37 vm04 ceph-mon[53170]: pgmap v137: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-06T13:49:38.111 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:49:37 vm02 ceph-mon[49207]: pgmap v137: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-06T13:49:40.010 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:49:39 vm06 ceph-mon[56528]: pgmap v138: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-06T13:49:40.049 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:49:39 vm04 ceph-mon[53170]: pgmap v138: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-06T13:49:40.112 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:49:39 vm02 ceph-mon[49207]: pgmap v138: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-06T13:49:42.010 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:49:41 vm06 ceph-mon[56528]: pgmap v139: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-06T13:49:42.049 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:49:41 vm04 ceph-mon[53170]: pgmap v139: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-06T13:49:42.112 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:49:41 vm02 ceph-mon[49207]: pgmap v139: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-06T13:49:43.862 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:49:43 vm02 ceph-mon[49207]: pgmap v140: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-06T13:49:43.959 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:49:43 vm06 ceph-mon[56528]: pgmap v140: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-06T13:49:44.049 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:49:43 vm04 ceph-mon[53170]: pgmap v140: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-06T13:49:46.010 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:49:45 vm06 ceph-mon[56528]: pgmap v141: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-06T13:49:46.010 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:49:45 vm06 ceph-mon[56528]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-06T13:49:46.010 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:49:45 vm06 ceph-mon[56528]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-06T13:49:46.010 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:49:45 vm06 ceph-mon[56528]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-06T13:49:46.010 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:49:45 vm06 ceph-mon[56528]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:49:46.010 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:49:45 vm06 ceph-mon[56528]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:49:46.010 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:49:45 vm06 ceph-mon[56528]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:49:46.010 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:49:45 vm06 ceph-mon[56528]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:49:46.010 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:49:45 vm06 ceph-mon[56528]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:49:46.010 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:49:45 vm06 ceph-mon[56528]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:49:46.010 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:49:45 vm06 ceph-mon[56528]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:49:46.049 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:49:45 vm04 ceph-mon[53170]: pgmap v141: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-06T13:49:46.049 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:49:45 vm04 ceph-mon[53170]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-06T13:49:46.049 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:49:45 vm04 ceph-mon[53170]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-06T13:49:46.049 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:49:45 vm04 ceph-mon[53170]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-06T13:49:46.049 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:49:45 vm04 ceph-mon[53170]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:49:46.049 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:49:45 vm04 ceph-mon[53170]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:49:46.049 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:49:45 vm04 ceph-mon[53170]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:49:46.049 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:49:45 vm04 ceph-mon[53170]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:49:46.049 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:49:45 vm04 ceph-mon[53170]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:49:46.049 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:49:45 vm04 ceph-mon[53170]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:49:46.050 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:49:45 vm04 ceph-mon[53170]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:49:46.112 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:49:45 vm02 ceph-mon[49207]: pgmap v141: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-06T13:49:46.112 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:49:45 vm02 ceph-mon[49207]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-06T13:49:46.112 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:49:45 vm02 ceph-mon[49207]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-06T13:49:46.112 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:49:45 vm02 ceph-mon[49207]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-06T13:49:46.112 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:49:45 vm02 ceph-mon[49207]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:49:46.112 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:49:45 vm02 ceph-mon[49207]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:49:46.112 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:49:45 vm02 ceph-mon[49207]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:49:46.112 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:49:45 vm02 ceph-mon[49207]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:49:46.112 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:49:45 vm02 ceph-mon[49207]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:49:46.112 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:49:45 vm02 ceph-mon[49207]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:49:46.112 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:49:45 vm02 ceph-mon[49207]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:49:48.010 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:49:47 vm06 ceph-mon[56528]: pgmap v142: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-06T13:49:48.049 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:49:47 vm04 ceph-mon[53170]: pgmap v142: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-06T13:49:48.111 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:49:47 vm02 ceph-mon[49207]: pgmap v142: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-06T13:49:50.010 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:49:49 vm06 ceph-mon[56528]: pgmap v143: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-06T13:49:50.049 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:49:49 vm04 ceph-mon[53170]: pgmap v143: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-06T13:49:50.111 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:49:49 vm02 ceph-mon[49207]: pgmap v143: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-06T13:49:52.010 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:49:51 vm06 ceph-mon[56528]: pgmap v144: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-06T13:49:52.049 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:49:51 vm04 ceph-mon[53170]: pgmap v144: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-06T13:49:52.111 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:49:51 vm02 ceph-mon[49207]: pgmap v144: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-06T13:49:53.915 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:49:53 vm04 ceph-mon[53170]: pgmap v145: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-06T13:49:54.010 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:49:53 vm06 ceph-mon[56528]: pgmap v145: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-06T13:49:54.111 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:49:53 vm02 ceph-mon[49207]: pgmap v145: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-06T13:49:56.010 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:49:55 vm06 ceph-mon[56528]: pgmap v146: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-06T13:49:56.010 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:49:55 vm06 ceph-mon[56528]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:49:56.010 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:49:55 vm06 ceph-mon[56528]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:49:56.010 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:49:55 vm06 ceph-mon[56528]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:49:56.049 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:49:55 vm04 ceph-mon[53170]: pgmap v146: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-06T13:49:56.050 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:49:55 vm04 ceph-mon[53170]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:49:56.050 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:49:55 vm04 ceph-mon[53170]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:49:56.050 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:49:55 vm04 ceph-mon[53170]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:49:56.112 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:49:55 vm02 ceph-mon[49207]: pgmap v146: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-06T13:49:56.112 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:49:55 vm02 ceph-mon[49207]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:49:56.112 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:49:55 vm02 ceph-mon[49207]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:49:56.112 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:49:55 vm02 ceph-mon[49207]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:49:58.010 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:49:57 vm06 ceph-mon[56528]: pgmap v147: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-06T13:49:58.049 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:49:57 vm04 ceph-mon[53170]: pgmap v147: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-06T13:49:58.111 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:49:57 vm02 ceph-mon[49207]: pgmap v147: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-06T13:50:00.010 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:49:59 vm06 ceph-mon[56528]: pgmap v148: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-06T13:50:00.049 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:49:59 vm04 ceph-mon[53170]: pgmap v148: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-06T13:50:00.111 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:49:59 vm02 ceph-mon[49207]: pgmap v148: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-06T13:50:01.010 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:50:00 vm06 ceph-mon[56528]: overall HEALTH_OK 2026-03-06T13:50:01.049 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:50:00 vm04 ceph-mon[53170]: overall HEALTH_OK 2026-03-06T13:50:01.111 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:50:00 vm02 ceph-mon[49207]: overall HEALTH_OK 2026-03-06T13:50:02.010 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:50:01 vm06 ceph-mon[56528]: pgmap v149: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-06T13:50:02.049 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:50:01 vm04 ceph-mon[53170]: pgmap v149: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-06T13:50:02.111 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:50:01 vm02 ceph-mon[49207]: pgmap v149: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-06T13:50:03.862 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:50:03 vm02 ceph-mon[49207]: pgmap v150: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-06T13:50:04.010 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:50:03 vm06 ceph-mon[56528]: pgmap v150: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-06T13:50:04.049 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:50:03 vm04 ceph-mon[53170]: pgmap v150: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-06T13:50:06.010 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:50:05 vm06 ceph-mon[56528]: pgmap v151: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-06T13:50:06.010 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:50:05 vm06 ceph-mon[56528]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:50:06.010 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:50:05 vm06 ceph-mon[56528]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:50:06.010 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:50:05 vm06 ceph-mon[56528]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:50:06.010 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:50:05 vm06 ceph-mon[56528]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:50:06.010 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:50:05 vm06 ceph-mon[56528]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:50:06.010 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:50:05 vm06 ceph-mon[56528]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:50:06.050 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:50:05 vm04 ceph-mon[53170]: pgmap v151: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-06T13:50:06.050 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:50:05 vm04 ceph-mon[53170]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:50:06.050 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:50:05 vm04 ceph-mon[53170]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:50:06.050 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:50:05 vm04 ceph-mon[53170]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:50:06.050 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:50:05 vm04 ceph-mon[53170]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:50:06.050 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:50:05 vm04 ceph-mon[53170]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:50:06.050 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:50:05 vm04 ceph-mon[53170]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:50:06.112 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:50:05 vm02 ceph-mon[49207]: pgmap v151: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-06T13:50:06.112 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:50:05 vm02 ceph-mon[49207]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:50:06.112 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:50:05 vm02 ceph-mon[49207]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:50:06.112 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:50:05 vm02 ceph-mon[49207]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:50:06.112 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:50:05 vm02 ceph-mon[49207]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:50:06.112 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:50:05 vm02 ceph-mon[49207]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:50:06.112 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:50:05 vm02 ceph-mon[49207]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:50:07.112 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:50:06 vm02 ceph-mon[49207]: pgmap v152: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-06T13:50:07.260 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:50:06 vm06 ceph-mon[56528]: pgmap v152: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-06T13:50:07.300 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:50:06 vm04 ceph-mon[53170]: pgmap v152: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-06T13:50:09.550 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:50:09 vm04 ceph-mon[53170]: pgmap v153: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-06T13:50:09.612 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:50:09 vm02 ceph-mon[49207]: pgmap v153: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-06T13:50:09.760 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:50:09 vm06 ceph-mon[56528]: pgmap v153: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-06T13:50:11.549 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:50:11 vm04 ceph-mon[53170]: pgmap v154: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-06T13:50:11.612 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:50:11 vm02 ceph-mon[49207]: pgmap v154: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-06T13:50:11.760 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:50:11 vm06 ceph-mon[56528]: pgmap v154: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-06T13:50:13.549 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:50:13 vm04 ceph-mon[53170]: pgmap v155: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-06T13:50:13.612 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:50:13 vm02 ceph-mon[49207]: pgmap v155: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-06T13:50:13.760 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:50:13 vm06 ceph-mon[56528]: pgmap v155: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-06T13:50:14.758 INFO:teuthology.orchestra.run.vm02.stderr:+ ceph mon dump 2026-03-06T13:50:14.937 INFO:teuthology.orchestra.run.vm02.stdout:epoch 6 2026-03-06T13:50:14.937 INFO:teuthology.orchestra.run.vm02.stdout:fsid 2b1b4a72-195a-11f1-b1df-bf1f800d676b 2026-03-06T13:50:14.937 INFO:teuthology.orchestra.run.vm02.stdout:last_changed 2026-03-06T12:47:14.260805+0000 2026-03-06T13:50:14.937 INFO:teuthology.orchestra.run.vm02.stdout:created 2026-03-06T12:44:48.508617+0000 2026-03-06T13:50:14.937 INFO:teuthology.orchestra.run.vm02.stdout:min_mon_release 19 (squid) 2026-03-06T13:50:14.937 INFO:teuthology.orchestra.run.vm02.stdout:election_strategy: 1 2026-03-06T13:50:14.937 INFO:teuthology.orchestra.run.vm02.stdout:0: [v2:192.168.123.102:3300/0,v1:192.168.123.102:6789/0] mon.a; crush_location {datacenter=a} 2026-03-06T13:50:14.937 INFO:teuthology.orchestra.run.vm02.stdout:1: [v2:192.168.123.106:3300/0,v1:192.168.123.106:6789/0] mon.c; crush_location {datacenter=a,rack=3} 2026-03-06T13:50:14.937 INFO:teuthology.orchestra.run.vm02.stdout:2: [v2:192.168.123.104:3300/0,v1:192.168.123.104:6789/0] mon.b; crush_location {datacenter=b,rack=2} 2026-03-06T13:50:14.937 INFO:teuthology.orchestra.run.vm02.stderr:dumped monmap epoch 6 2026-03-06T13:50:14.946 INFO:teuthology.orchestra.run.vm02.stderr:+ ceph mon dump --format json 2026-03-06T13:50:15.114 INFO:teuthology.orchestra.run.vm02.stdout: 2026-03-06T13:50:15.114 INFO:teuthology.orchestra.run.vm02.stdout:{"epoch":6,"fsid":"2b1b4a72-195a-11f1-b1df-bf1f800d676b","modified":"2026-03-06T12:47:14.260805Z","created":"2026-03-06T12:44:48.508617Z","min_mon_release":19,"min_mon_release_name":"squid","election_strategy":1,"disallowed_leaders":"","stretch_mode":false,"tiebreaker_mon":"","removed_ranks":"","features":{"persistent":["kraken","luminous","mimic","osdmap-prune","nautilus","octopus","pacific","elector-pinging","quincy","reef","squid"],"optional":[]},"mons":[{"rank":0,"name":"a","public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.102:3300","nonce":0},{"type":"v1","addr":"192.168.123.102:6789","nonce":0}]},"addr":"192.168.123.102:6789/0","public_addr":"192.168.123.102:6789/0","priority":0,"weight":0,"crush_location":"{datacenter=a}"},{"rank":1,"name":"c","public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.106:3300","nonce":0},{"type":"v1","addr":"192.168.123.106:6789","nonce":0}]},"addr":"192.168.123.106:6789/0","public_addr":"192.168.123.106:6789/0","priority":0,"weight":0,"crush_location":"{datacenter=a,rack=3}"},{"rank":2,"name":"b","public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.104:3300","nonce":0},{"type":"v1","addr":"192.168.123.104:6789","nonce":0}]},"addr":"192.168.123.104:6789/0","public_addr":"192.168.123.104:6789/0","priority":0,"weight":0,"crush_location":"{datacenter=b,rack=2}"}],"quorum":[0,1,2]} 2026-03-06T13:50:15.114 INFO:teuthology.orchestra.run.vm02.stderr:dumped monmap epoch 6 2026-03-06T13:50:15.122 INFO:teuthology.orchestra.run.vm02.stderr:+ read monid 2026-03-06T13:50:15.123 INFO:teuthology.orchestra.run.vm02.stderr:+ read crushloc 2026-03-06T13:50:15.123 INFO:teuthology.orchestra.run.vm02.stderr:+ jq --arg monid a --arg crushloc '{datacenter=a}' -e '.mons | .[] | select(.name == $monid) | .crush_location == $crushloc' 2026-03-06T13:50:15.123 INFO:teuthology.orchestra.run.vm02.stderr:+ ceph mon dump --format json 2026-03-06T13:50:15.293 INFO:teuthology.orchestra.run.vm02.stderr:dumped monmap epoch 6 2026-03-06T13:50:15.301 INFO:teuthology.orchestra.run.vm02.stdout:true 2026-03-06T13:50:15.302 INFO:teuthology.orchestra.run.vm02.stderr:+ read monid 2026-03-06T13:50:15.302 INFO:teuthology.orchestra.run.vm02.stderr:+ read crushloc 2026-03-06T13:50:15.302 INFO:teuthology.orchestra.run.vm02.stderr:+ ceph mon dump --format json 2026-03-06T13:50:15.302 INFO:teuthology.orchestra.run.vm02.stderr:+ jq --arg monid b --arg crushloc '{datacenter=b,rack=2}' -e '.mons | .[] | select(.name == $monid) | .crush_location == $crushloc' 2026-03-06T13:50:15.469 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:50:15 vm02 ceph-mon[49207]: pgmap v156: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-06T13:50:15.469 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:50:15 vm02 ceph-mon[49207]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:50:15.469 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:50:15 vm02 ceph-mon[49207]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:50:15.469 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:50:15 vm02 ceph-mon[49207]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:50:15.469 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:50:15 vm02 ceph-mon[49207]: from='client.? 192.168.123.102:0/3796242468' entity='client.admin' cmd=[{"prefix": "mon dump"}]: dispatch 2026-03-06T13:50:15.469 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:50:15 vm02 ceph-mon[49207]: from='client.? 192.168.123.102:0/3507920006' entity='client.admin' cmd=[{"prefix": "mon dump", "format": "json"}]: dispatch 2026-03-06T13:50:15.469 INFO:teuthology.orchestra.run.vm02.stderr:dumped monmap epoch 6 2026-03-06T13:50:15.476 INFO:teuthology.orchestra.run.vm02.stdout:true 2026-03-06T13:50:15.476 INFO:teuthology.orchestra.run.vm02.stderr:+ read monid 2026-03-06T13:50:15.477 INFO:teuthology.orchestra.run.vm02.stderr:+ read crushloc 2026-03-06T13:50:15.477 INFO:teuthology.orchestra.run.vm02.stderr:+ ceph mon dump --format json 2026-03-06T13:50:15.477 INFO:teuthology.orchestra.run.vm02.stderr:+ jq --arg monid c --arg crushloc '{datacenter=a,rack=3}' -e '.mons | .[] | select(.name == $monid) | .crush_location == $crushloc' 2026-03-06T13:50:15.550 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:50:15 vm04 ceph-mon[53170]: pgmap v156: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-06T13:50:15.550 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:50:15 vm04 ceph-mon[53170]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:50:15.550 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:50:15 vm04 ceph-mon[53170]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:50:15.550 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:50:15 vm04 ceph-mon[53170]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:50:15.550 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:50:15 vm04 ceph-mon[53170]: from='client.? 192.168.123.102:0/3796242468' entity='client.admin' cmd=[{"prefix": "mon dump"}]: dispatch 2026-03-06T13:50:15.550 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:50:15 vm04 ceph-mon[53170]: from='client.? 192.168.123.102:0/3507920006' entity='client.admin' cmd=[{"prefix": "mon dump", "format": "json"}]: dispatch 2026-03-06T13:50:15.644 INFO:teuthology.orchestra.run.vm02.stderr:dumped monmap epoch 6 2026-03-06T13:50:15.651 INFO:teuthology.orchestra.run.vm02.stdout:true 2026-03-06T13:50:15.651 INFO:teuthology.orchestra.run.vm02.stderr:+ read monid 2026-03-06T13:50:15.760 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:50:15 vm06 ceph-mon[56528]: pgmap v156: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-06T13:50:15.760 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:50:15 vm06 ceph-mon[56528]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:50:15.760 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:50:15 vm06 ceph-mon[56528]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:50:15.760 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:50:15 vm06 ceph-mon[56528]: from='mgr.14156 192.168.123.102:0/3001667563' entity='mgr.a' 2026-03-06T13:50:15.760 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:50:15 vm06 ceph-mon[56528]: from='client.? 192.168.123.102:0/3796242468' entity='client.admin' cmd=[{"prefix": "mon dump"}]: dispatch 2026-03-06T13:50:15.760 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:50:15 vm06 ceph-mon[56528]: from='client.? 192.168.123.102:0/3507920006' entity='client.admin' cmd=[{"prefix": "mon dump", "format": "json"}]: dispatch 2026-03-06T13:50:15.812 DEBUG:teuthology.run_tasks:Unwinding manager cephadm 2026-03-06T13:50:15.814 INFO:tasks.cephadm:Teardown begin 2026-03-06T13:50:15.814 DEBUG:teuthology.orchestra.run.vm02:> sudo rm -f /etc/ceph/ceph.conf /etc/ceph/ceph.client.admin.keyring 2026-03-06T13:50:15.840 DEBUG:teuthology.orchestra.run.vm04:> sudo rm -f /etc/ceph/ceph.conf /etc/ceph/ceph.client.admin.keyring 2026-03-06T13:50:15.870 DEBUG:teuthology.orchestra.run.vm06:> sudo rm -f /etc/ceph/ceph.conf /etc/ceph/ceph.client.admin.keyring 2026-03-06T13:50:15.896 INFO:tasks.cephadm:Cleaning up testdir ceph.* files... 2026-03-06T13:50:15.896 DEBUG:teuthology.orchestra.run.vm02:> rm -f /home/ubuntu/cephtest/seed.ceph.conf /home/ubuntu/cephtest/ceph.pub 2026-03-06T13:50:15.913 DEBUG:teuthology.orchestra.run.vm04:> rm -f /home/ubuntu/cephtest/seed.ceph.conf /home/ubuntu/cephtest/ceph.pub 2026-03-06T13:50:15.928 DEBUG:teuthology.orchestra.run.vm06:> rm -f /home/ubuntu/cephtest/seed.ceph.conf /home/ubuntu/cephtest/ceph.pub 2026-03-06T13:50:15.950 INFO:tasks.cephadm:Stopping all daemons... 2026-03-06T13:50:15.950 INFO:tasks.cephadm.mon.a:Stopping mon.a... 2026-03-06T13:50:15.950 DEBUG:teuthology.orchestra.run.vm02:> sudo systemctl stop ceph-2b1b4a72-195a-11f1-b1df-bf1f800d676b@mon.a 2026-03-06T13:50:16.054 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 06 13:50:15 vm02 systemd[1]: Stopping Ceph mon.a for 2b1b4a72-195a-11f1-b1df-bf1f800d676b... 2026-03-06T13:50:16.274 DEBUG:teuthology.orchestra.run.vm02:> sudo pkill -f 'journalctl -f -n 0 -u ceph-2b1b4a72-195a-11f1-b1df-bf1f800d676b@mon.a.service' 2026-03-06T13:50:16.304 DEBUG:teuthology.orchestra.run:got remote process result: None 2026-03-06T13:50:16.304 INFO:tasks.cephadm.mon.a:Stopped mon.a 2026-03-06T13:50:16.304 INFO:tasks.cephadm.mon.c:Stopping mon.b... 2026-03-06T13:50:16.304 DEBUG:teuthology.orchestra.run.vm04:> sudo systemctl stop ceph-2b1b4a72-195a-11f1-b1df-bf1f800d676b@mon.b 2026-03-06T13:50:16.601 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:50:16 vm04 systemd[1]: Stopping Ceph mon.b for 2b1b4a72-195a-11f1-b1df-bf1f800d676b... 2026-03-06T13:50:16.601 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:50:16 vm04 ceph-2b1b4a72-195a-11f1-b1df-bf1f800d676b-mon-b[53147]: 2026-03-06T12:50:16.394+0000 7fd276f1b640 -1 received signal: Terminated from /run/podman-init -- /usr/bin/ceph-mon -n mon.b -f --setuser ceph --setgroup ceph --default-log-to-file=false --default-log-to-journald=true --default-log-to-stderr=false --default-mon-cluster-log-to-file=false --default-mon-cluster-log-to-journald=true --default-mon-cluster-log-to-stderr=false (PID: 1) UID: 0 2026-03-06T13:50:16.601 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:50:16 vm04 ceph-2b1b4a72-195a-11f1-b1df-bf1f800d676b-mon-b[53147]: 2026-03-06T12:50:16.394+0000 7fd276f1b640 -1 mon.b@2(peon) e6 *** Got Signal Terminated *** 2026-03-06T13:50:16.601 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:50:16 vm04 podman[62500]: 2026-03-06 13:50:16.428166444 +0100 CET m=+0.046333255 container died 3b09de05b312406a780dcfccf7052b987f4ef80124786542910405225017a5e5 (image=harbor.clyso.com/custom-ceph/ceph/ceph@sha256:26363c7a4eea9ef5a0148afc7b2a22b6f486596d87a30c2a9fdcda5db3eca62b, name=ceph-2b1b4a72-195a-11f1-b1df-bf1f800d676b-mon-b, CEPH_SHA1=c24117fd5525679b799527bc1bd1f1dd0a2db5e2, FROM_IMAGE=rockylinux:9, GANESHA_REPO_BASEURL=https://buildlogs.centos.org/centos/$releasever-stream/storage/$basearch/nfsganesha-5/, OSD_FLAVOR=default, ceph=True, io.buildah.version=1.41.8, CEPH_GIT_REPO=https://github.com/irq0/ceph.git, CEPH_REF=19.2.3-47-gc24117fd552) 2026-03-06T13:50:16.601 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:50:16 vm04 podman[62500]: 2026-03-06 13:50:16.554171135 +0100 CET m=+0.172337946 container remove 3b09de05b312406a780dcfccf7052b987f4ef80124786542910405225017a5e5 (image=harbor.clyso.com/custom-ceph/ceph/ceph@sha256:26363c7a4eea9ef5a0148afc7b2a22b6f486596d87a30c2a9fdcda5db3eca62b, name=ceph-2b1b4a72-195a-11f1-b1df-bf1f800d676b-mon-b, FROM_IMAGE=rockylinux:9, GANESHA_REPO_BASEURL=https://buildlogs.centos.org/centos/$releasever-stream/storage/$basearch/nfsganesha-5/, OSD_FLAVOR=default, ceph=True, io.buildah.version=1.41.8, CEPH_GIT_REPO=https://github.com/irq0/ceph.git, CEPH_REF=19.2.3-47-gc24117fd552, CEPH_SHA1=c24117fd5525679b799527bc1bd1f1dd0a2db5e2) 2026-03-06T13:50:16.601 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 06 13:50:16 vm04 bash[62500]: ceph-2b1b4a72-195a-11f1-b1df-bf1f800d676b-mon-b 2026-03-06T13:50:16.615 DEBUG:teuthology.orchestra.run.vm04:> sudo pkill -f 'journalctl -f -n 0 -u ceph-2b1b4a72-195a-11f1-b1df-bf1f800d676b@mon.b.service' 2026-03-06T13:50:16.645 DEBUG:teuthology.orchestra.run:got remote process result: None 2026-03-06T13:50:16.645 INFO:tasks.cephadm.mon.c:Stopped mon.b 2026-03-06T13:50:16.645 INFO:tasks.cephadm.mon.c:Stopping mon.c... 2026-03-06T13:50:16.645 DEBUG:teuthology.orchestra.run.vm06:> sudo systemctl stop ceph-2b1b4a72-195a-11f1-b1df-bf1f800d676b@mon.c 2026-03-06T13:50:16.920 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:50:16 vm06 systemd[1]: Stopping Ceph mon.c for 2b1b4a72-195a-11f1-b1df-bf1f800d676b... 2026-03-06T13:50:16.920 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:50:16 vm06 ceph-2b1b4a72-195a-11f1-b1df-bf1f800d676b-mon-c[56504]: 2026-03-06T12:50:16.744+0000 7fed4105d640 -1 received signal: Terminated from /run/podman-init -- /usr/bin/ceph-mon -n mon.c -f --setuser ceph --setgroup ceph --default-log-to-file=false --default-log-to-journald=true --default-log-to-stderr=false --default-mon-cluster-log-to-file=false --default-mon-cluster-log-to-journald=true --default-mon-cluster-log-to-stderr=false (PID: 1) UID: 0 2026-03-06T13:50:16.920 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 06 13:50:16 vm06 ceph-2b1b4a72-195a-11f1-b1df-bf1f800d676b-mon-c[56504]: 2026-03-06T12:50:16.744+0000 7fed4105d640 -1 mon.c@1(peon) e6 *** Got Signal Terminated *** 2026-03-06T13:50:17.096 DEBUG:teuthology.orchestra.run.vm06:> sudo pkill -f 'journalctl -f -n 0 -u ceph-2b1b4a72-195a-11f1-b1df-bf1f800d676b@mon.c.service' 2026-03-06T13:50:17.126 DEBUG:teuthology.orchestra.run:got remote process result: None 2026-03-06T13:50:17.126 INFO:tasks.cephadm.mon.c:Stopped mon.c 2026-03-06T13:50:17.126 INFO:tasks.cephadm.mgr.a:Stopping mgr.a... 2026-03-06T13:50:17.127 DEBUG:teuthology.orchestra.run.vm02:> sudo systemctl stop ceph-2b1b4a72-195a-11f1-b1df-bf1f800d676b@mgr.a 2026-03-06T13:50:17.405 INFO:journalctl@ceph.mgr.a.vm02.stdout:Mar 06 13:50:17 vm02 systemd[1]: Stopping Ceph mgr.a for 2b1b4a72-195a-11f1-b1df-bf1f800d676b... 2026-03-06T13:50:17.405 INFO:journalctl@ceph.mgr.a.vm02.stdout:Mar 06 13:50:17 vm02 podman[66893]: 2026-03-06 13:50:17.268026206 +0100 CET m=+0.058029860 container died 6b7b3227c2471ec9cc79a4394e408b9ae7046c8aa6026fcc1dcbf32f8c16a854 (image=harbor.clyso.com/custom-ceph/ceph/ceph:cobaltcore-storage-v19.2.3-fasttrack-3, name=ceph-2b1b4a72-195a-11f1-b1df-bf1f800d676b-mgr-a, CEPH_SHA1=c24117fd5525679b799527bc1bd1f1dd0a2db5e2, FROM_IMAGE=rockylinux:9, GANESHA_REPO_BASEURL=https://buildlogs.centos.org/centos/$releasever-stream/storage/$basearch/nfsganesha-5/, OSD_FLAVOR=default, ceph=True, io.buildah.version=1.41.8, CEPH_GIT_REPO=https://github.com/irq0/ceph.git, CEPH_REF=19.2.3-47-gc24117fd552) 2026-03-06T13:50:17.465 DEBUG:teuthology.orchestra.run.vm02:> sudo pkill -f 'journalctl -f -n 0 -u ceph-2b1b4a72-195a-11f1-b1df-bf1f800d676b@mgr.a.service' 2026-03-06T13:50:17.492 DEBUG:teuthology.orchestra.run:got remote process result: None 2026-03-06T13:50:17.492 INFO:tasks.cephadm.mgr.a:Stopped mgr.a 2026-03-06T13:50:17.492 INFO:tasks.cephadm.mgr.b:Stopping mgr.b... 2026-03-06T13:50:17.492 DEBUG:teuthology.orchestra.run.vm04:> sudo systemctl stop ceph-2b1b4a72-195a-11f1-b1df-bf1f800d676b@mgr.b 2026-03-06T13:50:17.794 INFO:journalctl@ceph.mgr.b.vm04.stdout:Mar 06 13:50:17 vm04 systemd[1]: Stopping Ceph mgr.b for 2b1b4a72-195a-11f1-b1df-bf1f800d676b... 2026-03-06T13:50:17.794 INFO:journalctl@ceph.mgr.b.vm04.stdout:Mar 06 13:50:17 vm04 podman[62613]: 2026-03-06 13:50:17.619653877 +0100 CET m=+0.047488368 container died dfe14d6f23c7f94e074b71a0ce7b950025bfb78dd4fc4ac1f471d375e36511f5 (image=harbor.clyso.com/custom-ceph/ceph/ceph@sha256:26363c7a4eea9ef5a0148afc7b2a22b6f486596d87a30c2a9fdcda5db3eca62b, name=ceph-2b1b4a72-195a-11f1-b1df-bf1f800d676b-mgr-b, io.buildah.version=1.41.8, CEPH_GIT_REPO=https://github.com/irq0/ceph.git, CEPH_REF=19.2.3-47-gc24117fd552, CEPH_SHA1=c24117fd5525679b799527bc1bd1f1dd0a2db5e2, FROM_IMAGE=rockylinux:9, GANESHA_REPO_BASEURL=https://buildlogs.centos.org/centos/$releasever-stream/storage/$basearch/nfsganesha-5/, OSD_FLAVOR=default, ceph=True) 2026-03-06T13:50:17.794 INFO:journalctl@ceph.mgr.b.vm04.stdout:Mar 06 13:50:17 vm04 podman[62613]: 2026-03-06 13:50:17.747571859 +0100 CET m=+0.175406350 container remove dfe14d6f23c7f94e074b71a0ce7b950025bfb78dd4fc4ac1f471d375e36511f5 (image=harbor.clyso.com/custom-ceph/ceph/ceph@sha256:26363c7a4eea9ef5a0148afc7b2a22b6f486596d87a30c2a9fdcda5db3eca62b, name=ceph-2b1b4a72-195a-11f1-b1df-bf1f800d676b-mgr-b, OSD_FLAVOR=default, ceph=True, io.buildah.version=1.41.8, CEPH_GIT_REPO=https://github.com/irq0/ceph.git, CEPH_REF=19.2.3-47-gc24117fd552, CEPH_SHA1=c24117fd5525679b799527bc1bd1f1dd0a2db5e2, FROM_IMAGE=rockylinux:9, GANESHA_REPO_BASEURL=https://buildlogs.centos.org/centos/$releasever-stream/storage/$basearch/nfsganesha-5/) 2026-03-06T13:50:17.794 INFO:journalctl@ceph.mgr.b.vm04.stdout:Mar 06 13:50:17 vm04 bash[62613]: ceph-2b1b4a72-195a-11f1-b1df-bf1f800d676b-mgr-b 2026-03-06T13:50:17.794 INFO:journalctl@ceph.mgr.b.vm04.stdout:Mar 06 13:50:17 vm04 systemd[1]: ceph-2b1b4a72-195a-11f1-b1df-bf1f800d676b@mgr.b.service: Main process exited, code=exited, status=143/n/a 2026-03-06T13:50:17.801 DEBUG:teuthology.orchestra.run.vm04:> sudo pkill -f 'journalctl -f -n 0 -u ceph-2b1b4a72-195a-11f1-b1df-bf1f800d676b@mgr.b.service' 2026-03-06T13:50:17.828 DEBUG:teuthology.orchestra.run:got remote process result: None 2026-03-06T13:50:17.828 INFO:tasks.cephadm.mgr.b:Stopped mgr.b 2026-03-06T13:50:17.828 INFO:tasks.cephadm.osd.0:Stopping osd.0... 2026-03-06T13:50:17.828 DEBUG:teuthology.orchestra.run.vm02:> sudo systemctl stop ceph-2b1b4a72-195a-11f1-b1df-bf1f800d676b@osd.0 2026-03-06T13:50:18.112 INFO:journalctl@ceph.osd.0.vm02.stdout:Mar 06 13:50:17 vm02 systemd[1]: Stopping Ceph osd.0 for 2b1b4a72-195a-11f1-b1df-bf1f800d676b... 2026-03-06T13:50:18.112 INFO:journalctl@ceph.osd.0.vm02.stdout:Mar 06 13:50:17 vm02 ceph-2b1b4a72-195a-11f1-b1df-bf1f800d676b-osd-0[59732]: 2026-03-06T12:50:17.918+0000 7f08a65b3640 -1 received signal: Terminated from /run/podman-init -- /usr/bin/ceph-osd -n osd.0 -f --setuser ceph --setgroup ceph --default-log-to-file=false --default-log-to-journald=true --default-log-to-stderr=false (PID: 1) UID: 0 2026-03-06T13:50:18.112 INFO:journalctl@ceph.osd.0.vm02.stdout:Mar 06 13:50:17 vm02 ceph-2b1b4a72-195a-11f1-b1df-bf1f800d676b-osd-0[59732]: 2026-03-06T12:50:17.918+0000 7f08a65b3640 -1 osd.0 21 *** Got signal Terminated *** 2026-03-06T13:50:18.112 INFO:journalctl@ceph.osd.0.vm02.stdout:Mar 06 13:50:17 vm02 ceph-2b1b4a72-195a-11f1-b1df-bf1f800d676b-osd-0[59732]: 2026-03-06T12:50:17.918+0000 7f08a65b3640 -1 osd.0 21 *** Immediate shutdown (osd_fast_shutdown=true) *** 2026-03-06T13:50:22.978 INFO:journalctl@ceph.osd.0.vm02.stdout:Mar 06 13:50:22 vm02 podman[67009]: 2026-03-06 13:50:22.937658228 +0100 CET m=+5.033175357 container died 23e922edba9778f82a5e6d31c45db69de388469b9c89ea14f9b177f0554211b0 (image=harbor.clyso.com/custom-ceph/ceph/ceph@sha256:26363c7a4eea9ef5a0148afc7b2a22b6f486596d87a30c2a9fdcda5db3eca62b, name=ceph-2b1b4a72-195a-11f1-b1df-bf1f800d676b-osd-0, CEPH_GIT_REPO=https://github.com/irq0/ceph.git, CEPH_REF=19.2.3-47-gc24117fd552, CEPH_SHA1=c24117fd5525679b799527bc1bd1f1dd0a2db5e2, FROM_IMAGE=rockylinux:9, GANESHA_REPO_BASEURL=https://buildlogs.centos.org/centos/$releasever-stream/storage/$basearch/nfsganesha-5/, OSD_FLAVOR=default, ceph=True, io.buildah.version=1.41.8) 2026-03-06T13:50:23.233 INFO:journalctl@ceph.osd.0.vm02.stdout:Mar 06 13:50:23 vm02 podman[67009]: 2026-03-06 13:50:23.0611502 +0100 CET m=+5.156667329 container remove 23e922edba9778f82a5e6d31c45db69de388469b9c89ea14f9b177f0554211b0 (image=harbor.clyso.com/custom-ceph/ceph/ceph@sha256:26363c7a4eea9ef5a0148afc7b2a22b6f486596d87a30c2a9fdcda5db3eca62b, name=ceph-2b1b4a72-195a-11f1-b1df-bf1f800d676b-osd-0, CEPH_REF=19.2.3-47-gc24117fd552, CEPH_SHA1=c24117fd5525679b799527bc1bd1f1dd0a2db5e2, FROM_IMAGE=rockylinux:9, GANESHA_REPO_BASEURL=https://buildlogs.centos.org/centos/$releasever-stream/storage/$basearch/nfsganesha-5/, OSD_FLAVOR=default, ceph=True, io.buildah.version=1.41.8, CEPH_GIT_REPO=https://github.com/irq0/ceph.git) 2026-03-06T13:50:23.233 INFO:journalctl@ceph.osd.0.vm02.stdout:Mar 06 13:50:23 vm02 bash[67009]: ceph-2b1b4a72-195a-11f1-b1df-bf1f800d676b-osd-0 2026-03-06T13:50:23.612 INFO:journalctl@ceph.osd.0.vm02.stdout:Mar 06 13:50:23 vm02 podman[67179]: 2026-03-06 13:50:23.232330925 +0100 CET m=+0.021601475 container create c96542f9cfbd36966ffe8eb0519040610e5fea65f196089ddaa24efadf68c660 (image=harbor.clyso.com/custom-ceph/ceph/ceph@sha256:26363c7a4eea9ef5a0148afc7b2a22b6f486596d87a30c2a9fdcda5db3eca62b, name=ceph-2b1b4a72-195a-11f1-b1df-bf1f800d676b-osd-0-deactivate, CEPH_REF=19.2.3-47-gc24117fd552, CEPH_SHA1=c24117fd5525679b799527bc1bd1f1dd0a2db5e2, FROM_IMAGE=rockylinux:9, GANESHA_REPO_BASEURL=https://buildlogs.centos.org/centos/$releasever-stream/storage/$basearch/nfsganesha-5/, OSD_FLAVOR=default, ceph=True, io.buildah.version=1.41.8, CEPH_GIT_REPO=https://github.com/irq0/ceph.git) 2026-03-06T13:50:23.612 INFO:journalctl@ceph.osd.0.vm02.stdout:Mar 06 13:50:23 vm02 podman[67179]: 2026-03-06 13:50:23.279749864 +0100 CET m=+0.069020403 container init c96542f9cfbd36966ffe8eb0519040610e5fea65f196089ddaa24efadf68c660 (image=harbor.clyso.com/custom-ceph/ceph/ceph@sha256:26363c7a4eea9ef5a0148afc7b2a22b6f486596d87a30c2a9fdcda5db3eca62b, name=ceph-2b1b4a72-195a-11f1-b1df-bf1f800d676b-osd-0-deactivate, CEPH_GIT_REPO=https://github.com/irq0/ceph.git, CEPH_REF=19.2.3-47-gc24117fd552, CEPH_SHA1=c24117fd5525679b799527bc1bd1f1dd0a2db5e2, FROM_IMAGE=rockylinux:9, GANESHA_REPO_BASEURL=https://buildlogs.centos.org/centos/$releasever-stream/storage/$basearch/nfsganesha-5/, OSD_FLAVOR=default, ceph=True, io.buildah.version=1.41.8) 2026-03-06T13:50:23.612 INFO:journalctl@ceph.osd.0.vm02.stdout:Mar 06 13:50:23 vm02 podman[67179]: 2026-03-06 13:50:23.296703767 +0100 CET m=+0.085974317 container start c96542f9cfbd36966ffe8eb0519040610e5fea65f196089ddaa24efadf68c660 (image=harbor.clyso.com/custom-ceph/ceph/ceph@sha256:26363c7a4eea9ef5a0148afc7b2a22b6f486596d87a30c2a9fdcda5db3eca62b, name=ceph-2b1b4a72-195a-11f1-b1df-bf1f800d676b-osd-0-deactivate, OSD_FLAVOR=default, ceph=True, io.buildah.version=1.41.8, CEPH_GIT_REPO=https://github.com/irq0/ceph.git, CEPH_REF=19.2.3-47-gc24117fd552, CEPH_SHA1=c24117fd5525679b799527bc1bd1f1dd0a2db5e2, FROM_IMAGE=rockylinux:9, GANESHA_REPO_BASEURL=https://buildlogs.centos.org/centos/$releasever-stream/storage/$basearch/nfsganesha-5/) 2026-03-06T13:50:23.612 INFO:journalctl@ceph.osd.0.vm02.stdout:Mar 06 13:50:23 vm02 podman[67179]: 2026-03-06 13:50:23.297903563 +0100 CET m=+0.087174113 container attach c96542f9cfbd36966ffe8eb0519040610e5fea65f196089ddaa24efadf68c660 (image=harbor.clyso.com/custom-ceph/ceph/ceph@sha256:26363c7a4eea9ef5a0148afc7b2a22b6f486596d87a30c2a9fdcda5db3eca62b, name=ceph-2b1b4a72-195a-11f1-b1df-bf1f800d676b-osd-0-deactivate, CEPH_GIT_REPO=https://github.com/irq0/ceph.git, CEPH_REF=19.2.3-47-gc24117fd552, CEPH_SHA1=c24117fd5525679b799527bc1bd1f1dd0a2db5e2, FROM_IMAGE=rockylinux:9, GANESHA_REPO_BASEURL=https://buildlogs.centos.org/centos/$releasever-stream/storage/$basearch/nfsganesha-5/, OSD_FLAVOR=default, ceph=True, io.buildah.version=1.41.8) 2026-03-06T13:50:23.612 INFO:journalctl@ceph.osd.0.vm02.stdout:Mar 06 13:50:23 vm02 podman[67179]: 2026-03-06 13:50:23.220498998 +0100 CET m=+0.009769538 image pull 306e97de47e91c2b4b24d3dc09be3b3a12039b078f343d91220102acc6628a68 harbor.clyso.com/custom-ceph/ceph/ceph@sha256:26363c7a4eea9ef5a0148afc7b2a22b6f486596d87a30c2a9fdcda5db3eca62b 2026-03-06T13:50:23.799 DEBUG:teuthology.orchestra.run.vm02:> sudo pkill -f 'journalctl -f -n 0 -u ceph-2b1b4a72-195a-11f1-b1df-bf1f800d676b@osd.0.service' 2026-03-06T13:50:23.829 DEBUG:teuthology.orchestra.run:got remote process result: None 2026-03-06T13:50:23.829 INFO:tasks.cephadm.osd.0:Stopped osd.0 2026-03-06T13:50:23.829 INFO:tasks.cephadm.osd.1:Stopping osd.1... 2026-03-06T13:50:23.829 DEBUG:teuthology.orchestra.run.vm04:> sudo systemctl stop ceph-2b1b4a72-195a-11f1-b1df-bf1f800d676b@osd.1 2026-03-06T13:50:24.300 INFO:journalctl@ceph.osd.1.vm04.stdout:Mar 06 13:50:23 vm04 systemd[1]: Stopping Ceph osd.1 for 2b1b4a72-195a-11f1-b1df-bf1f800d676b... 2026-03-06T13:50:24.300 INFO:journalctl@ceph.osd.1.vm04.stdout:Mar 06 13:50:23 vm04 ceph-2b1b4a72-195a-11f1-b1df-bf1f800d676b-osd-1[58203]: 2026-03-06T12:50:23.916+0000 7f78f1a9a640 -1 received signal: Terminated from /run/podman-init -- /usr/bin/ceph-osd -n osd.1 -f --setuser ceph --setgroup ceph --default-log-to-file=false --default-log-to-journald=true --default-log-to-stderr=false (PID: 1) UID: 0 2026-03-06T13:50:24.300 INFO:journalctl@ceph.osd.1.vm04.stdout:Mar 06 13:50:23 vm04 ceph-2b1b4a72-195a-11f1-b1df-bf1f800d676b-osd-1[58203]: 2026-03-06T12:50:23.916+0000 7f78f1a9a640 -1 osd.1 21 *** Got signal Terminated *** 2026-03-06T13:50:24.300 INFO:journalctl@ceph.osd.1.vm04.stdout:Mar 06 13:50:23 vm04 ceph-2b1b4a72-195a-11f1-b1df-bf1f800d676b-osd-1[58203]: 2026-03-06T12:50:23.916+0000 7f78f1a9a640 -1 osd.1 21 *** Immediate shutdown (osd_fast_shutdown=true) *** 2026-03-06T13:50:29.201 INFO:journalctl@ceph.osd.1.vm04.stdout:Mar 06 13:50:28 vm04 podman[62727]: 2026-03-06 13:50:28.944870512 +0100 CET m=+5.040130293 container died 7adc2cdd55ae42ef10a5a33d28535016cc0531c2da268bde8b603bdc7d4bdb6c (image=harbor.clyso.com/custom-ceph/ceph/ceph@sha256:26363c7a4eea9ef5a0148afc7b2a22b6f486596d87a30c2a9fdcda5db3eca62b, name=ceph-2b1b4a72-195a-11f1-b1df-bf1f800d676b-osd-1, OSD_FLAVOR=default, ceph=True, io.buildah.version=1.41.8, CEPH_GIT_REPO=https://github.com/irq0/ceph.git, CEPH_REF=19.2.3-47-gc24117fd552, CEPH_SHA1=c24117fd5525679b799527bc1bd1f1dd0a2db5e2, FROM_IMAGE=rockylinux:9, GANESHA_REPO_BASEURL=https://buildlogs.centos.org/centos/$releasever-stream/storage/$basearch/nfsganesha-5/) 2026-03-06T13:50:29.201 INFO:journalctl@ceph.osd.1.vm04.stdout:Mar 06 13:50:29 vm04 podman[62727]: 2026-03-06 13:50:29.069351128 +0100 CET m=+5.164610900 container remove 7adc2cdd55ae42ef10a5a33d28535016cc0531c2da268bde8b603bdc7d4bdb6c (image=harbor.clyso.com/custom-ceph/ceph/ceph@sha256:26363c7a4eea9ef5a0148afc7b2a22b6f486596d87a30c2a9fdcda5db3eca62b, name=ceph-2b1b4a72-195a-11f1-b1df-bf1f800d676b-osd-1, io.buildah.version=1.41.8, CEPH_GIT_REPO=https://github.com/irq0/ceph.git, CEPH_REF=19.2.3-47-gc24117fd552, CEPH_SHA1=c24117fd5525679b799527bc1bd1f1dd0a2db5e2, FROM_IMAGE=rockylinux:9, GANESHA_REPO_BASEURL=https://buildlogs.centos.org/centos/$releasever-stream/storage/$basearch/nfsganesha-5/, OSD_FLAVOR=default, ceph=True) 2026-03-06T13:50:29.201 INFO:journalctl@ceph.osd.1.vm04.stdout:Mar 06 13:50:29 vm04 bash[62727]: ceph-2b1b4a72-195a-11f1-b1df-bf1f800d676b-osd-1 2026-03-06T13:50:29.550 INFO:journalctl@ceph.osd.1.vm04.stdout:Mar 06 13:50:29 vm04 podman[62805]: 2026-03-06 13:50:29.199528869 +0100 CET m=+0.015210190 container create ce8ba485bf757cb38b8d6789925c1b1abc45a54b192a1d7e7c9143f79920c908 (image=harbor.clyso.com/custom-ceph/ceph/ceph@sha256:26363c7a4eea9ef5a0148afc7b2a22b6f486596d87a30c2a9fdcda5db3eca62b, name=ceph-2b1b4a72-195a-11f1-b1df-bf1f800d676b-osd-1-deactivate, CEPH_SHA1=c24117fd5525679b799527bc1bd1f1dd0a2db5e2, FROM_IMAGE=rockylinux:9, GANESHA_REPO_BASEURL=https://buildlogs.centos.org/centos/$releasever-stream/storage/$basearch/nfsganesha-5/, OSD_FLAVOR=default, ceph=True, io.buildah.version=1.41.8, CEPH_GIT_REPO=https://github.com/irq0/ceph.git, CEPH_REF=19.2.3-47-gc24117fd552) 2026-03-06T13:50:29.550 INFO:journalctl@ceph.osd.1.vm04.stdout:Mar 06 13:50:29 vm04 podman[62805]: 2026-03-06 13:50:29.235103457 +0100 CET m=+0.050784788 container init ce8ba485bf757cb38b8d6789925c1b1abc45a54b192a1d7e7c9143f79920c908 (image=harbor.clyso.com/custom-ceph/ceph/ceph@sha256:26363c7a4eea9ef5a0148afc7b2a22b6f486596d87a30c2a9fdcda5db3eca62b, name=ceph-2b1b4a72-195a-11f1-b1df-bf1f800d676b-osd-1-deactivate, CEPH_REF=19.2.3-47-gc24117fd552, CEPH_SHA1=c24117fd5525679b799527bc1bd1f1dd0a2db5e2, FROM_IMAGE=rockylinux:9, GANESHA_REPO_BASEURL=https://buildlogs.centos.org/centos/$releasever-stream/storage/$basearch/nfsganesha-5/, OSD_FLAVOR=default, ceph=True, io.buildah.version=1.41.8, CEPH_GIT_REPO=https://github.com/irq0/ceph.git) 2026-03-06T13:50:29.550 INFO:journalctl@ceph.osd.1.vm04.stdout:Mar 06 13:50:29 vm04 podman[62805]: 2026-03-06 13:50:29.240134443 +0100 CET m=+0.055815754 container start ce8ba485bf757cb38b8d6789925c1b1abc45a54b192a1d7e7c9143f79920c908 (image=harbor.clyso.com/custom-ceph/ceph/ceph@sha256:26363c7a4eea9ef5a0148afc7b2a22b6f486596d87a30c2a9fdcda5db3eca62b, name=ceph-2b1b4a72-195a-11f1-b1df-bf1f800d676b-osd-1-deactivate, CEPH_REF=19.2.3-47-gc24117fd552, CEPH_SHA1=c24117fd5525679b799527bc1bd1f1dd0a2db5e2, FROM_IMAGE=rockylinux:9, GANESHA_REPO_BASEURL=https://buildlogs.centos.org/centos/$releasever-stream/storage/$basearch/nfsganesha-5/, OSD_FLAVOR=default, ceph=True, io.buildah.version=1.41.8, CEPH_GIT_REPO=https://github.com/irq0/ceph.git) 2026-03-06T13:50:29.550 INFO:journalctl@ceph.osd.1.vm04.stdout:Mar 06 13:50:29 vm04 podman[62805]: 2026-03-06 13:50:29.242767181 +0100 CET m=+0.058448502 container attach ce8ba485bf757cb38b8d6789925c1b1abc45a54b192a1d7e7c9143f79920c908 (image=harbor.clyso.com/custom-ceph/ceph/ceph@sha256:26363c7a4eea9ef5a0148afc7b2a22b6f486596d87a30c2a9fdcda5db3eca62b, name=ceph-2b1b4a72-195a-11f1-b1df-bf1f800d676b-osd-1-deactivate, GANESHA_REPO_BASEURL=https://buildlogs.centos.org/centos/$releasever-stream/storage/$basearch/nfsganesha-5/, OSD_FLAVOR=default, ceph=True, io.buildah.version=1.41.8, CEPH_GIT_REPO=https://github.com/irq0/ceph.git, CEPH_REF=19.2.3-47-gc24117fd552, CEPH_SHA1=c24117fd5525679b799527bc1bd1f1dd0a2db5e2, FROM_IMAGE=rockylinux:9) 2026-03-06T13:50:29.550 INFO:journalctl@ceph.osd.1.vm04.stdout:Mar 06 13:50:29 vm04 podman[62805]: 2026-03-06 13:50:29.193713976 +0100 CET m=+0.009395297 image pull 306e97de47e91c2b4b24d3dc09be3b3a12039b078f343d91220102acc6628a68 harbor.clyso.com/custom-ceph/ceph/ceph@sha256:26363c7a4eea9ef5a0148afc7b2a22b6f486596d87a30c2a9fdcda5db3eca62b 2026-03-06T13:50:29.742 DEBUG:teuthology.orchestra.run.vm04:> sudo pkill -f 'journalctl -f -n 0 -u ceph-2b1b4a72-195a-11f1-b1df-bf1f800d676b@osd.1.service' 2026-03-06T13:50:29.773 DEBUG:teuthology.orchestra.run:got remote process result: None 2026-03-06T13:50:29.773 INFO:tasks.cephadm.osd.1:Stopped osd.1 2026-03-06T13:50:29.773 INFO:tasks.cephadm.osd.2:Stopping osd.2... 2026-03-06T13:50:29.773 DEBUG:teuthology.orchestra.run.vm06:> sudo systemctl stop ceph-2b1b4a72-195a-11f1-b1df-bf1f800d676b@osd.2 2026-03-06T13:50:30.260 INFO:journalctl@ceph.osd.2.vm06.stdout:Mar 06 13:50:29 vm06 systemd[1]: Stopping Ceph osd.2 for 2b1b4a72-195a-11f1-b1df-bf1f800d676b... 2026-03-06T13:50:30.260 INFO:journalctl@ceph.osd.2.vm06.stdout:Mar 06 13:50:29 vm06 ceph-2b1b4a72-195a-11f1-b1df-bf1f800d676b-osd-2[61819]: 2026-03-06T12:50:29.874+0000 7fa11b169640 -1 received signal: Terminated from /run/podman-init -- /usr/bin/ceph-osd -n osd.2 -f --setuser ceph --setgroup ceph --default-log-to-file=false --default-log-to-journald=true --default-log-to-stderr=false (PID: 1) UID: 0 2026-03-06T13:50:30.260 INFO:journalctl@ceph.osd.2.vm06.stdout:Mar 06 13:50:29 vm06 ceph-2b1b4a72-195a-11f1-b1df-bf1f800d676b-osd-2[61819]: 2026-03-06T12:50:29.874+0000 7fa11b169640 -1 osd.2 21 *** Got signal Terminated *** 2026-03-06T13:50:30.260 INFO:journalctl@ceph.osd.2.vm06.stdout:Mar 06 13:50:29 vm06 ceph-2b1b4a72-195a-11f1-b1df-bf1f800d676b-osd-2[61819]: 2026-03-06T12:50:29.874+0000 7fa11b169640 -1 osd.2 21 *** Immediate shutdown (osd_fast_shutdown=true) *** 2026-03-06T13:50:35.153 INFO:journalctl@ceph.osd.2.vm06.stdout:Mar 06 13:50:34 vm06 podman[66045]: 2026-03-06 13:50:34.899943216 +0100 CET m=+5.043050218 container died 954e3d67a9dc330495831b63063f687a1b727542faa84ec2ff6b3e4b89e43304 (image=harbor.clyso.com/custom-ceph/ceph/ceph@sha256:26363c7a4eea9ef5a0148afc7b2a22b6f486596d87a30c2a9fdcda5db3eca62b, name=ceph-2b1b4a72-195a-11f1-b1df-bf1f800d676b-osd-2, CEPH_REF=19.2.3-47-gc24117fd552, CEPH_SHA1=c24117fd5525679b799527bc1bd1f1dd0a2db5e2, FROM_IMAGE=rockylinux:9, GANESHA_REPO_BASEURL=https://buildlogs.centos.org/centos/$releasever-stream/storage/$basearch/nfsganesha-5/, OSD_FLAVOR=default, ceph=True, io.buildah.version=1.41.8, CEPH_GIT_REPO=https://github.com/irq0/ceph.git) 2026-03-06T13:50:35.153 INFO:journalctl@ceph.osd.2.vm06.stdout:Mar 06 13:50:35 vm06 podman[66045]: 2026-03-06 13:50:35.027340364 +0100 CET m=+5.170447366 container remove 954e3d67a9dc330495831b63063f687a1b727542faa84ec2ff6b3e4b89e43304 (image=harbor.clyso.com/custom-ceph/ceph/ceph@sha256:26363c7a4eea9ef5a0148afc7b2a22b6f486596d87a30c2a9fdcda5db3eca62b, name=ceph-2b1b4a72-195a-11f1-b1df-bf1f800d676b-osd-2, ceph=True, io.buildah.version=1.41.8, CEPH_GIT_REPO=https://github.com/irq0/ceph.git, CEPH_REF=19.2.3-47-gc24117fd552, CEPH_SHA1=c24117fd5525679b799527bc1bd1f1dd0a2db5e2, FROM_IMAGE=rockylinux:9, GANESHA_REPO_BASEURL=https://buildlogs.centos.org/centos/$releasever-stream/storage/$basearch/nfsganesha-5/, OSD_FLAVOR=default) 2026-03-06T13:50:35.154 INFO:journalctl@ceph.osd.2.vm06.stdout:Mar 06 13:50:35 vm06 bash[66045]: ceph-2b1b4a72-195a-11f1-b1df-bf1f800d676b-osd-2 2026-03-06T13:50:35.510 INFO:journalctl@ceph.osd.2.vm06.stdout:Mar 06 13:50:35 vm06 podman[66135]: 2026-03-06 13:50:35.152755431 +0100 CET m=+0.014496935 container create d2c7af1e9dd518a66a5f197e60b5c09d10b064c60b5c5d7990c4cdef53a39e92 (image=harbor.clyso.com/custom-ceph/ceph/ceph@sha256:26363c7a4eea9ef5a0148afc7b2a22b6f486596d87a30c2a9fdcda5db3eca62b, name=ceph-2b1b4a72-195a-11f1-b1df-bf1f800d676b-osd-2-deactivate, CEPH_GIT_REPO=https://github.com/irq0/ceph.git, CEPH_REF=19.2.3-47-gc24117fd552, CEPH_SHA1=c24117fd5525679b799527bc1bd1f1dd0a2db5e2, FROM_IMAGE=rockylinux:9, GANESHA_REPO_BASEURL=https://buildlogs.centos.org/centos/$releasever-stream/storage/$basearch/nfsganesha-5/, OSD_FLAVOR=default, ceph=True, io.buildah.version=1.41.8) 2026-03-06T13:50:35.510 INFO:journalctl@ceph.osd.2.vm06.stdout:Mar 06 13:50:35 vm06 podman[66135]: 2026-03-06 13:50:35.181710489 +0100 CET m=+0.043451993 container init d2c7af1e9dd518a66a5f197e60b5c09d10b064c60b5c5d7990c4cdef53a39e92 (image=harbor.clyso.com/custom-ceph/ceph/ceph@sha256:26363c7a4eea9ef5a0148afc7b2a22b6f486596d87a30c2a9fdcda5db3eca62b, name=ceph-2b1b4a72-195a-11f1-b1df-bf1f800d676b-osd-2-deactivate, GANESHA_REPO_BASEURL=https://buildlogs.centos.org/centos/$releasever-stream/storage/$basearch/nfsganesha-5/, OSD_FLAVOR=default, ceph=True, io.buildah.version=1.41.8, CEPH_GIT_REPO=https://github.com/irq0/ceph.git, CEPH_REF=19.2.3-47-gc24117fd552, CEPH_SHA1=c24117fd5525679b799527bc1bd1f1dd0a2db5e2, FROM_IMAGE=rockylinux:9) 2026-03-06T13:50:35.510 INFO:journalctl@ceph.osd.2.vm06.stdout:Mar 06 13:50:35 vm06 podman[66135]: 2026-03-06 13:50:35.186042798 +0100 CET m=+0.047784302 container start d2c7af1e9dd518a66a5f197e60b5c09d10b064c60b5c5d7990c4cdef53a39e92 (image=harbor.clyso.com/custom-ceph/ceph/ceph@sha256:26363c7a4eea9ef5a0148afc7b2a22b6f486596d87a30c2a9fdcda5db3eca62b, name=ceph-2b1b4a72-195a-11f1-b1df-bf1f800d676b-osd-2-deactivate, CEPH_SHA1=c24117fd5525679b799527bc1bd1f1dd0a2db5e2, FROM_IMAGE=rockylinux:9, GANESHA_REPO_BASEURL=https://buildlogs.centos.org/centos/$releasever-stream/storage/$basearch/nfsganesha-5/, OSD_FLAVOR=default, ceph=True, io.buildah.version=1.41.8, CEPH_GIT_REPO=https://github.com/irq0/ceph.git, CEPH_REF=19.2.3-47-gc24117fd552) 2026-03-06T13:50:35.511 INFO:journalctl@ceph.osd.2.vm06.stdout:Mar 06 13:50:35 vm06 podman[66135]: 2026-03-06 13:50:35.187002524 +0100 CET m=+0.048744018 container attach d2c7af1e9dd518a66a5f197e60b5c09d10b064c60b5c5d7990c4cdef53a39e92 (image=harbor.clyso.com/custom-ceph/ceph/ceph@sha256:26363c7a4eea9ef5a0148afc7b2a22b6f486596d87a30c2a9fdcda5db3eca62b, name=ceph-2b1b4a72-195a-11f1-b1df-bf1f800d676b-osd-2-deactivate, io.buildah.version=1.41.8, CEPH_GIT_REPO=https://github.com/irq0/ceph.git, CEPH_REF=19.2.3-47-gc24117fd552, CEPH_SHA1=c24117fd5525679b799527bc1bd1f1dd0a2db5e2, FROM_IMAGE=rockylinux:9, GANESHA_REPO_BASEURL=https://buildlogs.centos.org/centos/$releasever-stream/storage/$basearch/nfsganesha-5/, OSD_FLAVOR=default, ceph=True) 2026-03-06T13:50:35.511 INFO:journalctl@ceph.osd.2.vm06.stdout:Mar 06 13:50:35 vm06 podman[66135]: 2026-03-06 13:50:35.146899099 +0100 CET m=+0.008640603 image pull 306e97de47e91c2b4b24d3dc09be3b3a12039b078f343d91220102acc6628a68 harbor.clyso.com/custom-ceph/ceph/ceph@sha256:26363c7a4eea9ef5a0148afc7b2a22b6f486596d87a30c2a9fdcda5db3eca62b 2026-03-06T13:50:35.682 DEBUG:teuthology.orchestra.run.vm06:> sudo pkill -f 'journalctl -f -n 0 -u ceph-2b1b4a72-195a-11f1-b1df-bf1f800d676b@osd.2.service' 2026-03-06T13:50:35.711 DEBUG:teuthology.orchestra.run:got remote process result: None 2026-03-06T13:50:35.711 INFO:tasks.cephadm.osd.2:Stopped osd.2 2026-03-06T13:50:35.711 DEBUG:teuthology.orchestra.run.vm02:> sudo /home/ubuntu/cephtest/cephadm rm-cluster --fsid 2b1b4a72-195a-11f1-b1df-bf1f800d676b --force --keep-logs 2026-03-06T13:50:35.988 INFO:teuthology.orchestra.run.vm02.stdout:Deleting cluster with fsid: 2b1b4a72-195a-11f1-b1df-bf1f800d676b 2026-03-06T13:50:37.119 DEBUG:teuthology.orchestra.run.vm04:> sudo /home/ubuntu/cephtest/cephadm rm-cluster --fsid 2b1b4a72-195a-11f1-b1df-bf1f800d676b --force --keep-logs 2026-03-06T13:50:37.390 INFO:teuthology.orchestra.run.vm04.stdout:Deleting cluster with fsid: 2b1b4a72-195a-11f1-b1df-bf1f800d676b 2026-03-06T13:50:38.421 DEBUG:teuthology.orchestra.run.vm06:> sudo /home/ubuntu/cephtest/cephadm rm-cluster --fsid 2b1b4a72-195a-11f1-b1df-bf1f800d676b --force --keep-logs 2026-03-06T13:50:38.689 INFO:teuthology.orchestra.run.vm06.stdout:Deleting cluster with fsid: 2b1b4a72-195a-11f1-b1df-bf1f800d676b 2026-03-06T13:50:39.563 DEBUG:teuthology.orchestra.run.vm02:> sudo rm -f /etc/ceph/ceph.conf /etc/ceph/ceph.client.admin.keyring 2026-03-06T13:50:39.587 DEBUG:teuthology.orchestra.run.vm04:> sudo rm -f /etc/ceph/ceph.conf /etc/ceph/ceph.client.admin.keyring 2026-03-06T13:50:39.614 DEBUG:teuthology.orchestra.run.vm06:> sudo rm -f /etc/ceph/ceph.conf /etc/ceph/ceph.client.admin.keyring 2026-03-06T13:50:39.638 INFO:tasks.cephadm:Archiving crash dumps... 2026-03-06T13:50:39.638 DEBUG:teuthology.misc:Transferring archived files from vm02:/var/lib/ceph/2b1b4a72-195a-11f1-b1df-bf1f800d676b/crash to /archive/irq0-2026-03-06_13:20:18-orch:cephadm:workunits-cobaltcore-storage-v19.2.3-fasttrack-3-none-default-vps/274/remote/vm02/crash 2026-03-06T13:50:39.638 DEBUG:teuthology.orchestra.run.vm02:> sudo tar c -f - -C /var/lib/ceph/2b1b4a72-195a-11f1-b1df-bf1f800d676b/crash -- . 2026-03-06T13:50:39.662 INFO:teuthology.orchestra.run.vm02.stderr:tar: /var/lib/ceph/2b1b4a72-195a-11f1-b1df-bf1f800d676b/crash: Cannot open: No such file or directory 2026-03-06T13:50:39.662 INFO:teuthology.orchestra.run.vm02.stderr:tar: Error is not recoverable: exiting now 2026-03-06T13:50:39.663 DEBUG:teuthology.misc:Transferring archived files from vm04:/var/lib/ceph/2b1b4a72-195a-11f1-b1df-bf1f800d676b/crash to /archive/irq0-2026-03-06_13:20:18-orch:cephadm:workunits-cobaltcore-storage-v19.2.3-fasttrack-3-none-default-vps/274/remote/vm04/crash 2026-03-06T13:50:39.663 DEBUG:teuthology.orchestra.run.vm04:> sudo tar c -f - -C /var/lib/ceph/2b1b4a72-195a-11f1-b1df-bf1f800d676b/crash -- . 2026-03-06T13:50:39.687 INFO:teuthology.orchestra.run.vm04.stderr:tar: /var/lib/ceph/2b1b4a72-195a-11f1-b1df-bf1f800d676b/crash: Cannot open: No such file or directory 2026-03-06T13:50:39.687 INFO:teuthology.orchestra.run.vm04.stderr:tar: Error is not recoverable: exiting now 2026-03-06T13:50:39.688 DEBUG:teuthology.misc:Transferring archived files from vm06:/var/lib/ceph/2b1b4a72-195a-11f1-b1df-bf1f800d676b/crash to /archive/irq0-2026-03-06_13:20:18-orch:cephadm:workunits-cobaltcore-storage-v19.2.3-fasttrack-3-none-default-vps/274/remote/vm06/crash 2026-03-06T13:50:39.688 DEBUG:teuthology.orchestra.run.vm06:> sudo tar c -f - -C /var/lib/ceph/2b1b4a72-195a-11f1-b1df-bf1f800d676b/crash -- . 2026-03-06T13:50:39.712 INFO:teuthology.orchestra.run.vm06.stderr:tar: /var/lib/ceph/2b1b4a72-195a-11f1-b1df-bf1f800d676b/crash: Cannot open: No such file or directory 2026-03-06T13:50:39.712 INFO:teuthology.orchestra.run.vm06.stderr:tar: Error is not recoverable: exiting now 2026-03-06T13:50:39.713 INFO:tasks.cephadm:Checking cluster log for badness... 2026-03-06T13:50:39.713 DEBUG:teuthology.orchestra.run.vm02:> sudo egrep '\[ERR\]|\[WRN\]|\[SEC\]' /var/log/ceph/2b1b4a72-195a-11f1-b1df-bf1f800d676b/ceph.log | egrep CEPHADM_ | egrep -v '\(MDS_ALL_DOWN\)' | egrep -v '\(MDS_UP_LESS_THAN_MAX\)' | egrep -v MON_DOWN | egrep -v POOL_APP_NOT_ENABLED | egrep -v 'mon down' | egrep -v 'mons down' | egrep -v 'out of quorum' | egrep -v CEPHADM_FAILED_DAEMON | head -n 1 2026-03-06T13:50:39.741 INFO:tasks.cephadm:Compressing logs... 2026-03-06T13:50:39.741 DEBUG:teuthology.orchestra.run.vm02:> time sudo find /var/log/ceph /var/log/rbd-target-api -name '*.log' -print0 | sudo xargs --max-args=1 --max-procs=0 --verbose -0 --no-run-if-empty -- gzip -5 --verbose -- 2026-03-06T13:50:39.784 DEBUG:teuthology.orchestra.run.vm04:> time sudo find /var/log/ceph /var/log/rbd-target-api -name '*.log' -print0 | sudo xargs --max-args=1 --max-procs=0 --verbose -0 --no-run-if-empty -- gzip -5 --verbose -- 2026-03-06T13:50:39.786 DEBUG:teuthology.orchestra.run.vm06:> time sudo find /var/log/ceph /var/log/rbd-target-api -name '*.log' -print0 | sudo xargs --max-args=1 --max-procs=0 --verbose -0 --no-run-if-empty -- gzip -5 --verbose -- 2026-03-06T13:50:39.805 INFO:teuthology.orchestra.run.vm02.stderr:find: ‘/var/log/rbd-target-api’: No such file or directory 2026-03-06T13:50:39.806 INFO:teuthology.orchestra.run.vm02.stderr:gzip -5 --verbose -- /var/log/ceph/cephadm.log 2026-03-06T13:50:39.807 INFO:teuthology.orchestra.run.vm02.stderr:gzip -5 --verbose -- /var/log/ceph/2b1b4a72-195a-11f1-b1df-bf1f800d676b/ceph-mon.a.log 2026-03-06T13:50:39.808 INFO:teuthology.orchestra.run.vm02.stderr:/var/log/ceph/cephadm.log: 85.7% -- replaced with /var/log/ceph/cephadm.log.gz 2026-03-06T13:50:39.808 INFO:teuthology.orchestra.run.vm02.stderr:gzip -5 --verbose -- /var/log/ceph/2b1b4a72-195a-11f1-b1df-bf1f800d676b/ceph.log 2026-03-06T13:50:39.809 INFO:teuthology.orchestra.run.vm06.stderr:find: ‘/var/log/rbd-target-api’: No such file or directory 2026-03-06T13:50:39.810 INFO:teuthology.orchestra.run.vm06.stderr:gzip -5 --verbose -- /var/log/ceph/cephadm.log 2026-03-06T13:50:39.811 INFO:teuthology.orchestra.run.vm06.stderr:gzip -5 --verbose -- /var/log/ceph/2b1b4a72-195a-11f1-b1df-bf1f800d676b/ceph-mon.c.log 2026-03-06T13:50:39.811 INFO:teuthology.orchestra.run.vm04.stderr:find: gzip -5 --verbose -- /var/log/ceph/cephadm.log 2026-03-06T13:50:39.811 INFO:teuthology.orchestra.run.vm04.stderr:‘/var/log/rbd-target-api’: No such file or directory 2026-03-06T13:50:39.811 INFO:teuthology.orchestra.run.vm06.stderr:/var/log/ceph/cephadm.log: gzip -5 --verbose -- /var/log/ceph/2b1b4a72-195a-11f1-b1df-bf1f800d676b/ceph-volume.log 2026-03-06T13:50:39.812 INFO:teuthology.orchestra.run.vm06.stderr: 81.9% -- replaced with /var/log/ceph/cephadm.log.gz 2026-03-06T13:50:39.812 INFO:teuthology.orchestra.run.vm02.stderr:/var/log/ceph/2b1b4a72-195a-11f1-b1df-bf1f800d676b/ceph-mon.a.log: gzip -5 --verbose -- /var/log/ceph/2b1b4a72-195a-11f1-b1df-bf1f800d676b/ceph.audit.log 2026-03-06T13:50:39.812 INFO:teuthology.orchestra.run.vm06.stderr:/var/log/ceph/2b1b4a72-195a-11f1-b1df-bf1f800d676b/ceph-mon.c.log: gzip -5 --verbose -- /var/log/ceph/2b1b4a72-195a-11f1-b1df-bf1f800d676b/ceph.cephadm.log 2026-03-06T13:50:39.812 INFO:teuthology.orchestra.run.vm02.stderr:/var/log/ceph/2b1b4a72-195a-11f1-b1df-bf1f800d676b/ceph.log: 86.6% -- replaced with /var/log/ceph/2b1b4a72-195a-11f1-b1df-bf1f800d676b/ceph.log.gz 2026-03-06T13:50:39.812 INFO:teuthology.orchestra.run.vm04.stderr:gzip -5 --verbose -- /var/log/ceph/2b1b4a72-195a-11f1-b1df-bf1f800d676b/ceph-volume.log 2026-03-06T13:50:39.812 INFO:teuthology.orchestra.run.vm02.stderr:gzip -5 --verbose -- /var/log/ceph/2b1b4a72-195a-11f1-b1df-bf1f800d676b/ceph-mgr.a.log 2026-03-06T13:50:39.813 INFO:teuthology.orchestra.run.vm04.stderr:gzip -5 --verbose -- /var/log/ceph/2b1b4a72-195a-11f1-b1df-bf1f800d676b/ceph-mon.b.log 2026-03-06T13:50:39.813 INFO:teuthology.orchestra.run.vm04.stderr:/var/log/ceph/cephadm.log: /var/log/ceph/2b1b4a72-195a-11f1-b1df-bf1f800d676b/ceph-volume.log: 81.9% -- replaced with /var/log/ceph/cephadm.log.gz 2026-03-06T13:50:39.814 INFO:teuthology.orchestra.run.vm02.stderr:/var/log/ceph/2b1b4a72-195a-11f1-b1df-bf1f800d676b/ceph.audit.log: 90.5% -- replaced with /var/log/ceph/2b1b4a72-195a-11f1-b1df-bf1f800d676b/ceph.audit.log.gz 2026-03-06T13:50:39.814 INFO:teuthology.orchestra.run.vm04.stderr:gzip -5 --verbose -- /var/log/ceph/2b1b4a72-195a-11f1-b1df-bf1f800d676b/ceph.cephadm.log 2026-03-06T13:50:39.814 INFO:teuthology.orchestra.run.vm02.stderr:gzip -5 --verbose -- /var/log/ceph/2b1b4a72-195a-11f1-b1df-bf1f800d676b/ceph.cephadm.log 2026-03-06T13:50:39.818 INFO:teuthology.orchestra.run.vm06.stderr:/var/log/ceph/2b1b4a72-195a-11f1-b1df-bf1f800d676b/ceph-volume.log: gzip -5 --verbose -- /var/log/ceph/2b1b4a72-195a-11f1-b1df-bf1f800d676b/ceph.audit.log 2026-03-06T13:50:39.818 INFO:teuthology.orchestra.run.vm06.stderr:/var/log/ceph/2b1b4a72-195a-11f1-b1df-bf1f800d676b/ceph.cephadm.log: 81.2% -- replaced with /var/log/ceph/2b1b4a72-195a-11f1-b1df-bf1f800d676b/ceph.cephadm.log.gz 2026-03-06T13:50:39.818 INFO:teuthology.orchestra.run.vm04.stderr:/var/log/ceph/2b1b4a72-195a-11f1-b1df-bf1f800d676b/ceph-mon.b.log: gzip -5 --verbose -- /var/log/ceph/2b1b4a72-195a-11f1-b1df-bf1f800d676b/ceph.audit.log 2026-03-06T13:50:39.819 INFO:teuthology.orchestra.run.vm04.stderr:/var/log/ceph/2b1b4a72-195a-11f1-b1df-bf1f800d676b/ceph.cephadm.log: 81.2% -- replaced with /var/log/ceph/2b1b4a72-195a-11f1-b1df-bf1f800d676b/ceph.cephadm.log.gz 2026-03-06T13:50:39.821 INFO:teuthology.orchestra.run.vm06.stderr: 92.7% -- replaced with /var/log/ceph/2b1b4a72-195a-11f1-b1df-bf1f800d676b/ceph-volume.log.gz 2026-03-06T13:50:39.821 INFO:teuthology.orchestra.run.vm02.stderr:/var/log/ceph/2b1b4a72-195a-11f1-b1df-bf1f800d676b/ceph-mgr.a.log: gzip -5 --verbose -- /var/log/ceph/2b1b4a72-195a-11f1-b1df-bf1f800d676b/ceph-volume.log 2026-03-06T13:50:39.821 INFO:teuthology.orchestra.run.vm04.stderr: 92.6% -- replaced with /var/log/ceph/2b1b4a72-195a-11f1-b1df-bf1f800d676b/ceph-volume.log.gz 2026-03-06T13:50:39.821 INFO:teuthology.orchestra.run.vm02.stderr:/var/log/ceph/2b1b4a72-195a-11f1-b1df-bf1f800d676b/ceph.cephadm.log: 81.7% -- replaced with /var/log/ceph/2b1b4a72-195a-11f1-b1df-bf1f800d676b/ceph.cephadm.log.gz 2026-03-06T13:50:39.822 INFO:teuthology.orchestra.run.vm06.stderr:gzip -5 --verbose -- /var/log/ceph/2b1b4a72-195a-11f1-b1df-bf1f800d676b/ceph.log 2026-03-06T13:50:39.822 INFO:teuthology.orchestra.run.vm06.stderr:/var/log/ceph/2b1b4a72-195a-11f1-b1df-bf1f800d676b/ceph.audit.log: 90.7% -- replaced with /var/log/ceph/2b1b4a72-195a-11f1-b1df-bf1f800d676b/ceph.audit.log.gz 2026-03-06T13:50:39.822 INFO:teuthology.orchestra.run.vm04.stderr:gzip -5 --verbose -- /var/log/ceph/2b1b4a72-195a-11f1-b1df-bf1f800d676b/ceph.log 2026-03-06T13:50:39.822 INFO:teuthology.orchestra.run.vm06.stderr:gzip -5 --verbose -- /var/log/ceph/2b1b4a72-195a-11f1-b1df-bf1f800d676b/ceph-osd.2.log 2026-03-06T13:50:39.823 INFO:teuthology.orchestra.run.vm06.stderr:/var/log/ceph/2b1b4a72-195a-11f1-b1df-bf1f800d676b/ceph.log: 86.7% -- replaced with /var/log/ceph/2b1b4a72-195a-11f1-b1df-bf1f800d676b/ceph.log.gz 2026-03-06T13:50:39.823 INFO:teuthology.orchestra.run.vm04.stderr:/var/log/ceph/2b1b4a72-195a-11f1-b1df-bf1f800d676b/ceph.audit.log: 90.7% -- replaced with /var/log/ceph/2b1b4a72-195a-11f1-b1df-bf1f800d676b/ceph.audit.log.gz 2026-03-06T13:50:39.824 INFO:teuthology.orchestra.run.vm02.stderr:gzip -5 --verbose -- /var/log/ceph/2b1b4a72-195a-11f1-b1df-bf1f800d676b/ceph-osd.0.log 2026-03-06T13:50:39.824 INFO:teuthology.orchestra.run.vm04.stderr:gzip -5 --verbose -- /var/log/ceph/2b1b4a72-195a-11f1-b1df-bf1f800d676b/ceph-mgr.b.log 2026-03-06T13:50:39.825 INFO:teuthology.orchestra.run.vm04.stderr:/var/log/ceph/2b1b4a72-195a-11f1-b1df-bf1f800d676b/ceph.log: 86.7% -- replaced with /var/log/ceph/2b1b4a72-195a-11f1-b1df-bf1f800d676b/ceph.log.gz 2026-03-06T13:50:39.827 INFO:teuthology.orchestra.run.vm04.stderr:gzip -5 --verbose -- /var/log/ceph/2b1b4a72-195a-11f1-b1df-bf1f800d676b/ceph-osd.1.log 2026-03-06T13:50:39.829 INFO:teuthology.orchestra.run.vm04.stderr:/var/log/ceph/2b1b4a72-195a-11f1-b1df-bf1f800d676b/ceph-mgr.b.log: 91.0% -- replaced with /var/log/ceph/2b1b4a72-195a-11f1-b1df-bf1f800d676b/ceph-mgr.b.log.gz 2026-03-06T13:50:39.839 INFO:teuthology.orchestra.run.vm06.stderr:/var/log/ceph/2b1b4a72-195a-11f1-b1df-bf1f800d676b/ceph-osd.2.log: 93.3% -- replaced with /var/log/ceph/2b1b4a72-195a-11f1-b1df-bf1f800d676b/ceph-osd.2.log.gz 2026-03-06T13:50:39.847 INFO:teuthology.orchestra.run.vm02.stderr:/var/log/ceph/2b1b4a72-195a-11f1-b1df-bf1f800d676b/ceph-volume.log: /var/log/ceph/2b1b4a72-195a-11f1-b1df-bf1f800d676b/ceph-osd.0.log: 92.6% -- replaced with /var/log/ceph/2b1b4a72-195a-11f1-b1df-bf1f800d676b/ceph-volume.log.gz 2026-03-06T13:50:39.852 INFO:teuthology.orchestra.run.vm04.stderr:/var/log/ceph/2b1b4a72-195a-11f1-b1df-bf1f800d676b/ceph-osd.1.log: 93.4% -- replaced with /var/log/ceph/2b1b4a72-195a-11f1-b1df-bf1f800d676b/ceph-osd.1.log.gz 2026-03-06T13:50:39.854 INFO:teuthology.orchestra.run.vm06.stderr: 92.8% -- replaced with /var/log/ceph/2b1b4a72-195a-11f1-b1df-bf1f800d676b/ceph-mon.c.log.gz 2026-03-06T13:50:39.856 INFO:teuthology.orchestra.run.vm06.stderr: 2026-03-06T13:50:39.856 INFO:teuthology.orchestra.run.vm06.stderr:real 0m0.056s 2026-03-06T13:50:39.856 INFO:teuthology.orchestra.run.vm06.stderr:user 0m0.069s 2026-03-06T13:50:39.856 INFO:teuthology.orchestra.run.vm06.stderr:sys 0m0.020s 2026-03-06T13:50:39.857 INFO:teuthology.orchestra.run.vm04.stderr: 92.5% -- replaced with /var/log/ceph/2b1b4a72-195a-11f1-b1df-bf1f800d676b/ceph-mon.b.log.gz 2026-03-06T13:50:39.858 INFO:teuthology.orchestra.run.vm02.stderr: 88.8% -- replaced with /var/log/ceph/2b1b4a72-195a-11f1-b1df-bf1f800d676b/ceph-mgr.a.log.gz 2026-03-06T13:50:39.859 INFO:teuthology.orchestra.run.vm04.stderr: 2026-03-06T13:50:39.859 INFO:teuthology.orchestra.run.vm04.stderr:real 0m0.058s 2026-03-06T13:50:39.859 INFO:teuthology.orchestra.run.vm04.stderr:user 0m0.078s 2026-03-06T13:50:39.859 INFO:teuthology.orchestra.run.vm04.stderr:sys 0m0.022s 2026-03-06T13:50:39.864 INFO:teuthology.orchestra.run.vm02.stderr: 93.2% -- replaced with /var/log/ceph/2b1b4a72-195a-11f1-b1df-bf1f800d676b/ceph-osd.0.log.gz 2026-03-06T13:50:39.957 INFO:teuthology.orchestra.run.vm02.stderr: 91.5% -- replaced with /var/log/ceph/2b1b4a72-195a-11f1-b1df-bf1f800d676b/ceph-mon.a.log.gz 2026-03-06T13:50:39.959 INFO:teuthology.orchestra.run.vm02.stderr: 2026-03-06T13:50:39.959 INFO:teuthology.orchestra.run.vm02.stderr:real 0m0.163s 2026-03-06T13:50:39.959 INFO:teuthology.orchestra.run.vm02.stderr:user 0m0.193s 2026-03-06T13:50:39.959 INFO:teuthology.orchestra.run.vm02.stderr:sys 0m0.025s 2026-03-06T13:50:39.959 INFO:tasks.cephadm:Archiving logs... 2026-03-06T13:50:39.959 DEBUG:teuthology.misc:Transferring archived files from vm02:/var/log/ceph to /archive/irq0-2026-03-06_13:20:18-orch:cephadm:workunits-cobaltcore-storage-v19.2.3-fasttrack-3-none-default-vps/274/remote/vm02/log 2026-03-06T13:50:39.959 DEBUG:teuthology.orchestra.run.vm02:> sudo tar c -f - -C /var/log/ceph -- . 2026-03-06T13:50:40.042 DEBUG:teuthology.misc:Transferring archived files from vm04:/var/log/ceph to /archive/irq0-2026-03-06_13:20:18-orch:cephadm:workunits-cobaltcore-storage-v19.2.3-fasttrack-3-none-default-vps/274/remote/vm04/log 2026-03-06T13:50:40.042 DEBUG:teuthology.orchestra.run.vm04:> sudo tar c -f - -C /var/log/ceph -- . 2026-03-06T13:50:40.071 DEBUG:teuthology.misc:Transferring archived files from vm06:/var/log/ceph to /archive/irq0-2026-03-06_13:20:18-orch:cephadm:workunits-cobaltcore-storage-v19.2.3-fasttrack-3-none-default-vps/274/remote/vm06/log 2026-03-06T13:50:40.071 DEBUG:teuthology.orchestra.run.vm06:> sudo tar c -f - -C /var/log/ceph -- . 2026-03-06T13:50:40.098 INFO:tasks.cephadm:Removing cluster... 2026-03-06T13:50:40.098 DEBUG:teuthology.orchestra.run.vm02:> sudo /home/ubuntu/cephtest/cephadm rm-cluster --fsid 2b1b4a72-195a-11f1-b1df-bf1f800d676b --force 2026-03-06T13:50:40.359 INFO:teuthology.orchestra.run.vm02.stdout:Deleting cluster with fsid: 2b1b4a72-195a-11f1-b1df-bf1f800d676b 2026-03-06T13:50:40.577 DEBUG:teuthology.orchestra.run.vm04:> sudo /home/ubuntu/cephtest/cephadm rm-cluster --fsid 2b1b4a72-195a-11f1-b1df-bf1f800d676b --force 2026-03-06T13:50:40.843 INFO:teuthology.orchestra.run.vm04.stdout:Deleting cluster with fsid: 2b1b4a72-195a-11f1-b1df-bf1f800d676b 2026-03-06T13:50:41.045 DEBUG:teuthology.orchestra.run.vm06:> sudo /home/ubuntu/cephtest/cephadm rm-cluster --fsid 2b1b4a72-195a-11f1-b1df-bf1f800d676b --force 2026-03-06T13:50:41.308 INFO:teuthology.orchestra.run.vm06.stdout:Deleting cluster with fsid: 2b1b4a72-195a-11f1-b1df-bf1f800d676b 2026-03-06T13:50:41.501 INFO:tasks.cephadm:Removing cephadm ... 2026-03-06T13:50:41.501 DEBUG:teuthology.orchestra.run.vm02:> rm -rf /home/ubuntu/cephtest/cephadm 2026-03-06T13:50:41.516 DEBUG:teuthology.orchestra.run.vm04:> rm -rf /home/ubuntu/cephtest/cephadm 2026-03-06T13:50:41.531 DEBUG:teuthology.orchestra.run.vm06:> rm -rf /home/ubuntu/cephtest/cephadm 2026-03-06T13:50:41.544 INFO:tasks.cephadm:Teardown complete 2026-03-06T13:50:41.544 DEBUG:teuthology.run_tasks:Unwinding manager install 2026-03-06T13:50:41.547 INFO:teuthology.task.install.util:Removing shipped files: /home/ubuntu/cephtest/valgrind.supp /usr/bin/daemon-helper /usr/bin/adjust-ulimits /usr/bin/stdin-killer... 2026-03-06T13:50:41.547 DEBUG:teuthology.orchestra.run.vm02:> sudo rm -f -- /home/ubuntu/cephtest/valgrind.supp /usr/bin/daemon-helper /usr/bin/adjust-ulimits /usr/bin/stdin-killer 2026-03-06T13:50:41.558 DEBUG:teuthology.orchestra.run.vm04:> sudo rm -f -- /home/ubuntu/cephtest/valgrind.supp /usr/bin/daemon-helper /usr/bin/adjust-ulimits /usr/bin/stdin-killer 2026-03-06T13:50:41.573 DEBUG:teuthology.orchestra.run.vm06:> sudo rm -f -- /home/ubuntu/cephtest/valgrind.supp /usr/bin/daemon-helper /usr/bin/adjust-ulimits /usr/bin/stdin-killer 2026-03-06T13:50:41.616 INFO:teuthology.task.install.rpm:Removing packages: ceph-radosgw, ceph-test, ceph, ceph-base, cephadm, ceph-immutable-object-cache, ceph-mgr, ceph-mgr-dashboard, ceph-mgr-diskprediction-local, ceph-mgr-rook, ceph-mgr-cephadm, ceph-fuse, ceph-volume, librados-devel, libcephfs2, libcephfs-devel, librados2, librbd1, python3-rados, python3-rgw, python3-cephfs, python3-rbd, rbd-fuse, rbd-mirror, rbd-nbd on rpm system. 2026-03-06T13:50:41.616 DEBUG:teuthology.orchestra.run.vm02:> 2026-03-06T13:50:41.616 DEBUG:teuthology.orchestra.run.vm02:> for d in ceph-radosgw ceph-test ceph ceph-base cephadm ceph-immutable-object-cache ceph-mgr ceph-mgr-dashboard ceph-mgr-diskprediction-local ceph-mgr-rook ceph-mgr-cephadm ceph-fuse ceph-volume librados-devel libcephfs2 libcephfs-devel librados2 librbd1 python3-rados python3-rgw python3-cephfs python3-rbd rbd-fuse rbd-mirror rbd-nbd ; do 2026-03-06T13:50:41.616 DEBUG:teuthology.orchestra.run.vm02:> sudo yum -y remove $d || true 2026-03-06T13:50:41.616 DEBUG:teuthology.orchestra.run.vm02:> done 2026-03-06T13:50:41.622 INFO:teuthology.task.install.rpm:Removing packages: ceph-radosgw, ceph-test, ceph, ceph-base, cephadm, ceph-immutable-object-cache, ceph-mgr, ceph-mgr-dashboard, ceph-mgr-diskprediction-local, ceph-mgr-rook, ceph-mgr-cephadm, ceph-fuse, ceph-volume, librados-devel, libcephfs2, libcephfs-devel, librados2, librbd1, python3-rados, python3-rgw, python3-cephfs, python3-rbd, rbd-fuse, rbd-mirror, rbd-nbd on rpm system. 2026-03-06T13:50:41.622 DEBUG:teuthology.orchestra.run.vm04:> 2026-03-06T13:50:41.622 DEBUG:teuthology.orchestra.run.vm04:> for d in ceph-radosgw ceph-test ceph ceph-base cephadm ceph-immutable-object-cache ceph-mgr ceph-mgr-dashboard ceph-mgr-diskprediction-local ceph-mgr-rook ceph-mgr-cephadm ceph-fuse ceph-volume librados-devel libcephfs2 libcephfs-devel librados2 librbd1 python3-rados python3-rgw python3-cephfs python3-rbd rbd-fuse rbd-mirror rbd-nbd ; do 2026-03-06T13:50:41.622 DEBUG:teuthology.orchestra.run.vm04:> sudo yum -y remove $d || true 2026-03-06T13:50:41.622 DEBUG:teuthology.orchestra.run.vm04:> done 2026-03-06T13:50:41.627 INFO:teuthology.task.install.rpm:Removing packages: ceph-radosgw, ceph-test, ceph, ceph-base, cephadm, ceph-immutable-object-cache, ceph-mgr, ceph-mgr-dashboard, ceph-mgr-diskprediction-local, ceph-mgr-rook, ceph-mgr-cephadm, ceph-fuse, ceph-volume, librados-devel, libcephfs2, libcephfs-devel, librados2, librbd1, python3-rados, python3-rgw, python3-cephfs, python3-rbd, rbd-fuse, rbd-mirror, rbd-nbd on rpm system. 2026-03-06T13:50:41.627 DEBUG:teuthology.orchestra.run.vm06:> 2026-03-06T13:50:41.627 DEBUG:teuthology.orchestra.run.vm06:> for d in ceph-radosgw ceph-test ceph ceph-base cephadm ceph-immutable-object-cache ceph-mgr ceph-mgr-dashboard ceph-mgr-diskprediction-local ceph-mgr-rook ceph-mgr-cephadm ceph-fuse ceph-volume librados-devel libcephfs2 libcephfs-devel librados2 librbd1 python3-rados python3-rgw python3-cephfs python3-rbd rbd-fuse rbd-mirror rbd-nbd ; do 2026-03-06T13:50:41.627 DEBUG:teuthology.orchestra.run.vm06:> sudo yum -y remove $d || true 2026-03-06T13:50:41.627 DEBUG:teuthology.orchestra.run.vm06:> done 2026-03-06T13:50:41.796 INFO:teuthology.orchestra.run.vm02.stdout:Dependencies resolved. 2026-03-06T13:50:41.796 INFO:teuthology.orchestra.run.vm02.stdout:================================================================================ 2026-03-06T13:50:41.796 INFO:teuthology.orchestra.run.vm02.stdout: Package Arch Version Repo Size 2026-03-06T13:50:41.796 INFO:teuthology.orchestra.run.vm02.stdout:================================================================================ 2026-03-06T13:50:41.796 INFO:teuthology.orchestra.run.vm02.stdout:Removing: 2026-03-06T13:50:41.797 INFO:teuthology.orchestra.run.vm02.stdout: ceph-radosgw x86_64 2:19.2.3-47.gc24117fd552.el9.clyso @ceph 39 M 2026-03-06T13:50:41.797 INFO:teuthology.orchestra.run.vm02.stdout:Removing unused dependencies: 2026-03-06T13:50:41.797 INFO:teuthology.orchestra.run.vm02.stdout: mailcap noarch 2.1.49-5.el9 @baseos 78 k 2026-03-06T13:50:41.797 INFO:teuthology.orchestra.run.vm02.stdout: 2026-03-06T13:50:41.797 INFO:teuthology.orchestra.run.vm02.stdout:Transaction Summary 2026-03-06T13:50:41.797 INFO:teuthology.orchestra.run.vm02.stdout:================================================================================ 2026-03-06T13:50:41.797 INFO:teuthology.orchestra.run.vm02.stdout:Remove 2 Packages 2026-03-06T13:50:41.797 INFO:teuthology.orchestra.run.vm02.stdout: 2026-03-06T13:50:41.797 INFO:teuthology.orchestra.run.vm02.stdout:Freed space: 39 M 2026-03-06T13:50:41.797 INFO:teuthology.orchestra.run.vm02.stdout:Running transaction check 2026-03-06T13:50:41.799 INFO:teuthology.orchestra.run.vm02.stdout:Transaction check succeeded. 2026-03-06T13:50:41.799 INFO:teuthology.orchestra.run.vm02.stdout:Running transaction test 2026-03-06T13:50:41.811 INFO:teuthology.orchestra.run.vm02.stdout:Transaction test succeeded. 2026-03-06T13:50:41.812 INFO:teuthology.orchestra.run.vm02.stdout:Running transaction 2026-03-06T13:50:41.814 INFO:teuthology.orchestra.run.vm04.stdout:Dependencies resolved. 2026-03-06T13:50:41.814 INFO:teuthology.orchestra.run.vm04.stdout:================================================================================ 2026-03-06T13:50:41.814 INFO:teuthology.orchestra.run.vm04.stdout: Package Arch Version Repo Size 2026-03-06T13:50:41.814 INFO:teuthology.orchestra.run.vm04.stdout:================================================================================ 2026-03-06T13:50:41.814 INFO:teuthology.orchestra.run.vm04.stdout:Removing: 2026-03-06T13:50:41.814 INFO:teuthology.orchestra.run.vm04.stdout: ceph-radosgw x86_64 2:19.2.3-47.gc24117fd552.el9.clyso @ceph 39 M 2026-03-06T13:50:41.814 INFO:teuthology.orchestra.run.vm04.stdout:Removing unused dependencies: 2026-03-06T13:50:41.814 INFO:teuthology.orchestra.run.vm04.stdout: mailcap noarch 2.1.49-5.el9 @baseos 78 k 2026-03-06T13:50:41.814 INFO:teuthology.orchestra.run.vm04.stdout: 2026-03-06T13:50:41.814 INFO:teuthology.orchestra.run.vm04.stdout:Transaction Summary 2026-03-06T13:50:41.814 INFO:teuthology.orchestra.run.vm04.stdout:================================================================================ 2026-03-06T13:50:41.814 INFO:teuthology.orchestra.run.vm04.stdout:Remove 2 Packages 2026-03-06T13:50:41.814 INFO:teuthology.orchestra.run.vm04.stdout: 2026-03-06T13:50:41.814 INFO:teuthology.orchestra.run.vm04.stdout:Freed space: 39 M 2026-03-06T13:50:41.814 INFO:teuthology.orchestra.run.vm04.stdout:Running transaction check 2026-03-06T13:50:41.817 INFO:teuthology.orchestra.run.vm04.stdout:Transaction check succeeded. 2026-03-06T13:50:41.817 INFO:teuthology.orchestra.run.vm04.stdout:Running transaction test 2026-03-06T13:50:41.828 INFO:teuthology.orchestra.run.vm06.stdout:Dependencies resolved. 2026-03-06T13:50:41.829 INFO:teuthology.orchestra.run.vm06.stdout:================================================================================ 2026-03-06T13:50:41.829 INFO:teuthology.orchestra.run.vm06.stdout: Package Arch Version Repo Size 2026-03-06T13:50:41.829 INFO:teuthology.orchestra.run.vm06.stdout:================================================================================ 2026-03-06T13:50:41.829 INFO:teuthology.orchestra.run.vm06.stdout:Removing: 2026-03-06T13:50:41.829 INFO:teuthology.orchestra.run.vm06.stdout: ceph-radosgw x86_64 2:19.2.3-47.gc24117fd552.el9.clyso @ceph 39 M 2026-03-06T13:50:41.829 INFO:teuthology.orchestra.run.vm06.stdout:Removing unused dependencies: 2026-03-06T13:50:41.829 INFO:teuthology.orchestra.run.vm06.stdout: mailcap noarch 2.1.49-5.el9 @baseos 78 k 2026-03-06T13:50:41.829 INFO:teuthology.orchestra.run.vm06.stdout: 2026-03-06T13:50:41.829 INFO:teuthology.orchestra.run.vm06.stdout:Transaction Summary 2026-03-06T13:50:41.829 INFO:teuthology.orchestra.run.vm06.stdout:================================================================================ 2026-03-06T13:50:41.829 INFO:teuthology.orchestra.run.vm06.stdout:Remove 2 Packages 2026-03-06T13:50:41.829 INFO:teuthology.orchestra.run.vm06.stdout: 2026-03-06T13:50:41.829 INFO:teuthology.orchestra.run.vm06.stdout:Freed space: 39 M 2026-03-06T13:50:41.829 INFO:teuthology.orchestra.run.vm06.stdout:Running transaction check 2026-03-06T13:50:41.830 INFO:teuthology.orchestra.run.vm04.stdout:Transaction test succeeded. 2026-03-06T13:50:41.830 INFO:teuthology.orchestra.run.vm04.stdout:Running transaction 2026-03-06T13:50:41.831 INFO:teuthology.orchestra.run.vm06.stdout:Transaction check succeeded. 2026-03-06T13:50:41.831 INFO:teuthology.orchestra.run.vm06.stdout:Running transaction test 2026-03-06T13:50:41.841 INFO:teuthology.orchestra.run.vm02.stdout: Preparing : 1/1 2026-03-06T13:50:41.844 INFO:teuthology.orchestra.run.vm06.stdout:Transaction test succeeded. 2026-03-06T13:50:41.845 INFO:teuthology.orchestra.run.vm06.stdout:Running transaction 2026-03-06T13:50:41.865 INFO:teuthology.orchestra.run.vm04.stdout: Preparing : 1/1 2026-03-06T13:50:41.866 INFO:teuthology.orchestra.run.vm02.stdout: Running scriptlet: ceph-radosgw-2:19.2.3-47.gc24117fd552.el9.clyso.x86_ 1/2 2026-03-06T13:50:41.866 INFO:teuthology.orchestra.run.vm02.stdout:Glob pattern passed to enable, but globs are not supported for this. 2026-03-06T13:50:41.866 INFO:teuthology.orchestra.run.vm02.stdout:Invalid unit name "ceph-radosgw@*.service" escaped as "ceph-radosgw@\x2a.service". 2026-03-06T13:50:41.866 INFO:teuthology.orchestra.run.vm02.stdout:Removed "/etc/systemd/system/multi-user.target.wants/ceph-radosgw.target". 2026-03-06T13:50:41.866 INFO:teuthology.orchestra.run.vm02.stdout:Removed "/etc/systemd/system/ceph.target.wants/ceph-radosgw.target". 2026-03-06T13:50:41.866 INFO:teuthology.orchestra.run.vm02.stdout: 2026-03-06T13:50:41.869 INFO:teuthology.orchestra.run.vm02.stdout: Erasing : ceph-radosgw-2:19.2.3-47.gc24117fd552.el9.clyso.x86_ 1/2 2026-03-06T13:50:41.874 INFO:teuthology.orchestra.run.vm06.stdout: Preparing : 1/1 2026-03-06T13:50:41.878 INFO:teuthology.orchestra.run.vm02.stdout: Running scriptlet: ceph-radosgw-2:19.2.3-47.gc24117fd552.el9.clyso.x86_ 1/2 2026-03-06T13:50:41.888 INFO:teuthology.orchestra.run.vm04.stdout: Running scriptlet: ceph-radosgw-2:19.2.3-47.gc24117fd552.el9.clyso.x86_ 1/2 2026-03-06T13:50:41.888 INFO:teuthology.orchestra.run.vm04.stdout:Glob pattern passed to enable, but globs are not supported for this. 2026-03-06T13:50:41.888 INFO:teuthology.orchestra.run.vm04.stdout:Invalid unit name "ceph-radosgw@*.service" escaped as "ceph-radosgw@\x2a.service". 2026-03-06T13:50:41.888 INFO:teuthology.orchestra.run.vm04.stdout:Removed "/etc/systemd/system/multi-user.target.wants/ceph-radosgw.target". 2026-03-06T13:50:41.888 INFO:teuthology.orchestra.run.vm04.stdout:Removed "/etc/systemd/system/ceph.target.wants/ceph-radosgw.target". 2026-03-06T13:50:41.888 INFO:teuthology.orchestra.run.vm04.stdout: 2026-03-06T13:50:41.891 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : ceph-radosgw-2:19.2.3-47.gc24117fd552.el9.clyso.x86_ 1/2 2026-03-06T13:50:41.893 INFO:teuthology.orchestra.run.vm02.stdout: Erasing : mailcap-2.1.49-5.el9.noarch 2/2 2026-03-06T13:50:41.893 INFO:teuthology.orchestra.run.vm06.stdout: Running scriptlet: ceph-radosgw-2:19.2.3-47.gc24117fd552.el9.clyso.x86_ 1/2 2026-03-06T13:50:41.893 INFO:teuthology.orchestra.run.vm06.stdout:Glob pattern passed to enable, but globs are not supported for this. 2026-03-06T13:50:41.893 INFO:teuthology.orchestra.run.vm06.stdout:Invalid unit name "ceph-radosgw@*.service" escaped as "ceph-radosgw@\x2a.service". 2026-03-06T13:50:41.893 INFO:teuthology.orchestra.run.vm06.stdout:Removed "/etc/systemd/system/multi-user.target.wants/ceph-radosgw.target". 2026-03-06T13:50:41.893 INFO:teuthology.orchestra.run.vm06.stdout:Removed "/etc/systemd/system/ceph.target.wants/ceph-radosgw.target". 2026-03-06T13:50:41.893 INFO:teuthology.orchestra.run.vm06.stdout: 2026-03-06T13:50:41.896 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : ceph-radosgw-2:19.2.3-47.gc24117fd552.el9.clyso.x86_ 1/2 2026-03-06T13:50:41.900 INFO:teuthology.orchestra.run.vm04.stdout: Running scriptlet: ceph-radosgw-2:19.2.3-47.gc24117fd552.el9.clyso.x86_ 1/2 2026-03-06T13:50:41.905 INFO:teuthology.orchestra.run.vm06.stdout: Running scriptlet: ceph-radosgw-2:19.2.3-47.gc24117fd552.el9.clyso.x86_ 1/2 2026-03-06T13:50:41.914 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : mailcap-2.1.49-5.el9.noarch 2/2 2026-03-06T13:50:41.919 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : mailcap-2.1.49-5.el9.noarch 2/2 2026-03-06T13:50:41.962 INFO:teuthology.orchestra.run.vm02.stdout: Running scriptlet: mailcap-2.1.49-5.el9.noarch 2/2 2026-03-06T13:50:41.963 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : ceph-radosgw-2:19.2.3-47.gc24117fd552.el9.clyso.x86_ 1/2 2026-03-06T13:50:41.978 INFO:teuthology.orchestra.run.vm04.stdout: Running scriptlet: mailcap-2.1.49-5.el9.noarch 2/2 2026-03-06T13:50:41.978 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : ceph-radosgw-2:19.2.3-47.gc24117fd552.el9.clyso.x86_ 1/2 2026-03-06T13:50:41.984 INFO:teuthology.orchestra.run.vm06.stdout: Running scriptlet: mailcap-2.1.49-5.el9.noarch 2/2 2026-03-06T13:50:41.985 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : ceph-radosgw-2:19.2.3-47.gc24117fd552.el9.clyso.x86_ 1/2 2026-03-06T13:50:42.018 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : mailcap-2.1.49-5.el9.noarch 2/2 2026-03-06T13:50:42.018 INFO:teuthology.orchestra.run.vm02.stdout: 2026-03-06T13:50:42.018 INFO:teuthology.orchestra.run.vm02.stdout:Removed: 2026-03-06T13:50:42.018 INFO:teuthology.orchestra.run.vm02.stdout: ceph-radosgw-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 2026-03-06T13:50:42.018 INFO:teuthology.orchestra.run.vm02.stdout: mailcap-2.1.49-5.el9.noarch 2026-03-06T13:50:42.018 INFO:teuthology.orchestra.run.vm02.stdout: 2026-03-06T13:50:42.018 INFO:teuthology.orchestra.run.vm02.stdout:Complete! 2026-03-06T13:50:42.025 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : mailcap-2.1.49-5.el9.noarch 2/2 2026-03-06T13:50:42.025 INFO:teuthology.orchestra.run.vm04.stdout: 2026-03-06T13:50:42.026 INFO:teuthology.orchestra.run.vm04.stdout:Removed: 2026-03-06T13:50:42.026 INFO:teuthology.orchestra.run.vm04.stdout: ceph-radosgw-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 2026-03-06T13:50:42.026 INFO:teuthology.orchestra.run.vm04.stdout: mailcap-2.1.49-5.el9.noarch 2026-03-06T13:50:42.026 INFO:teuthology.orchestra.run.vm04.stdout: 2026-03-06T13:50:42.026 INFO:teuthology.orchestra.run.vm04.stdout:Complete! 2026-03-06T13:50:42.034 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : mailcap-2.1.49-5.el9.noarch 2/2 2026-03-06T13:50:42.034 INFO:teuthology.orchestra.run.vm06.stdout: 2026-03-06T13:50:42.034 INFO:teuthology.orchestra.run.vm06.stdout:Removed: 2026-03-06T13:50:42.034 INFO:teuthology.orchestra.run.vm06.stdout: ceph-radosgw-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 2026-03-06T13:50:42.034 INFO:teuthology.orchestra.run.vm06.stdout: mailcap-2.1.49-5.el9.noarch 2026-03-06T13:50:42.034 INFO:teuthology.orchestra.run.vm06.stdout: 2026-03-06T13:50:42.034 INFO:teuthology.orchestra.run.vm06.stdout:Complete! 2026-03-06T13:50:42.213 INFO:teuthology.orchestra.run.vm02.stdout:Dependencies resolved. 2026-03-06T13:50:42.213 INFO:teuthology.orchestra.run.vm02.stdout:================================================================================ 2026-03-06T13:50:42.213 INFO:teuthology.orchestra.run.vm02.stdout: Package Arch Version Repository Size 2026-03-06T13:50:42.213 INFO:teuthology.orchestra.run.vm02.stdout:================================================================================ 2026-03-06T13:50:42.213 INFO:teuthology.orchestra.run.vm02.stdout:Removing: 2026-03-06T13:50:42.213 INFO:teuthology.orchestra.run.vm02.stdout: ceph-test x86_64 2:19.2.3-47.gc24117fd552.el9.clyso @ceph 210 M 2026-03-06T13:50:42.213 INFO:teuthology.orchestra.run.vm02.stdout:Removing unused dependencies: 2026-03-06T13:50:42.213 INFO:teuthology.orchestra.run.vm02.stdout: socat x86_64 1.7.4.1-8.el9 @appstream 1.1 M 2026-03-06T13:50:42.213 INFO:teuthology.orchestra.run.vm02.stdout: xmlstarlet x86_64 1.6.1-20.el9 @appstream 195 k 2026-03-06T13:50:42.213 INFO:teuthology.orchestra.run.vm02.stdout: 2026-03-06T13:50:42.213 INFO:teuthology.orchestra.run.vm02.stdout:Transaction Summary 2026-03-06T13:50:42.214 INFO:teuthology.orchestra.run.vm02.stdout:================================================================================ 2026-03-06T13:50:42.214 INFO:teuthology.orchestra.run.vm02.stdout:Remove 3 Packages 2026-03-06T13:50:42.214 INFO:teuthology.orchestra.run.vm02.stdout: 2026-03-06T13:50:42.214 INFO:teuthology.orchestra.run.vm02.stdout:Freed space: 211 M 2026-03-06T13:50:42.214 INFO:teuthology.orchestra.run.vm02.stdout:Running transaction check 2026-03-06T13:50:42.216 INFO:teuthology.orchestra.run.vm02.stdout:Transaction check succeeded. 2026-03-06T13:50:42.216 INFO:teuthology.orchestra.run.vm02.stdout:Running transaction test 2026-03-06T13:50:42.222 INFO:teuthology.orchestra.run.vm04.stdout:Dependencies resolved. 2026-03-06T13:50:42.223 INFO:teuthology.orchestra.run.vm04.stdout:================================================================================ 2026-03-06T13:50:42.223 INFO:teuthology.orchestra.run.vm04.stdout: Package Arch Version Repository Size 2026-03-06T13:50:42.223 INFO:teuthology.orchestra.run.vm04.stdout:================================================================================ 2026-03-06T13:50:42.223 INFO:teuthology.orchestra.run.vm04.stdout:Removing: 2026-03-06T13:50:42.223 INFO:teuthology.orchestra.run.vm04.stdout: ceph-test x86_64 2:19.2.3-47.gc24117fd552.el9.clyso @ceph 210 M 2026-03-06T13:50:42.223 INFO:teuthology.orchestra.run.vm04.stdout:Removing unused dependencies: 2026-03-06T13:50:42.223 INFO:teuthology.orchestra.run.vm04.stdout: socat x86_64 1.7.4.1-8.el9 @appstream 1.1 M 2026-03-06T13:50:42.223 INFO:teuthology.orchestra.run.vm04.stdout: xmlstarlet x86_64 1.6.1-20.el9 @appstream 195 k 2026-03-06T13:50:42.223 INFO:teuthology.orchestra.run.vm04.stdout: 2026-03-06T13:50:42.223 INFO:teuthology.orchestra.run.vm04.stdout:Transaction Summary 2026-03-06T13:50:42.223 INFO:teuthology.orchestra.run.vm04.stdout:================================================================================ 2026-03-06T13:50:42.223 INFO:teuthology.orchestra.run.vm04.stdout:Remove 3 Packages 2026-03-06T13:50:42.223 INFO:teuthology.orchestra.run.vm04.stdout: 2026-03-06T13:50:42.223 INFO:teuthology.orchestra.run.vm04.stdout:Freed space: 211 M 2026-03-06T13:50:42.223 INFO:teuthology.orchestra.run.vm04.stdout:Running transaction check 2026-03-06T13:50:42.225 INFO:teuthology.orchestra.run.vm04.stdout:Transaction check succeeded. 2026-03-06T13:50:42.226 INFO:teuthology.orchestra.run.vm04.stdout:Running transaction test 2026-03-06T13:50:42.229 INFO:teuthology.orchestra.run.vm06.stdout:Dependencies resolved. 2026-03-06T13:50:42.230 INFO:teuthology.orchestra.run.vm06.stdout:================================================================================ 2026-03-06T13:50:42.230 INFO:teuthology.orchestra.run.vm06.stdout: Package Arch Version Repository Size 2026-03-06T13:50:42.230 INFO:teuthology.orchestra.run.vm06.stdout:================================================================================ 2026-03-06T13:50:42.230 INFO:teuthology.orchestra.run.vm06.stdout:Removing: 2026-03-06T13:50:42.230 INFO:teuthology.orchestra.run.vm06.stdout: ceph-test x86_64 2:19.2.3-47.gc24117fd552.el9.clyso @ceph 210 M 2026-03-06T13:50:42.230 INFO:teuthology.orchestra.run.vm06.stdout:Removing unused dependencies: 2026-03-06T13:50:42.230 INFO:teuthology.orchestra.run.vm06.stdout: socat x86_64 1.7.4.1-8.el9 @appstream 1.1 M 2026-03-06T13:50:42.230 INFO:teuthology.orchestra.run.vm06.stdout: xmlstarlet x86_64 1.6.1-20.el9 @appstream 195 k 2026-03-06T13:50:42.230 INFO:teuthology.orchestra.run.vm06.stdout: 2026-03-06T13:50:42.230 INFO:teuthology.orchestra.run.vm06.stdout:Transaction Summary 2026-03-06T13:50:42.230 INFO:teuthology.orchestra.run.vm06.stdout:================================================================================ 2026-03-06T13:50:42.230 INFO:teuthology.orchestra.run.vm06.stdout:Remove 3 Packages 2026-03-06T13:50:42.230 INFO:teuthology.orchestra.run.vm06.stdout: 2026-03-06T13:50:42.230 INFO:teuthology.orchestra.run.vm06.stdout:Freed space: 211 M 2026-03-06T13:50:42.230 INFO:teuthology.orchestra.run.vm06.stdout:Running transaction check 2026-03-06T13:50:42.233 INFO:teuthology.orchestra.run.vm06.stdout:Transaction check succeeded. 2026-03-06T13:50:42.233 INFO:teuthology.orchestra.run.vm06.stdout:Running transaction test 2026-03-06T13:50:42.237 INFO:teuthology.orchestra.run.vm02.stdout:Transaction test succeeded. 2026-03-06T13:50:42.237 INFO:teuthology.orchestra.run.vm02.stdout:Running transaction 2026-03-06T13:50:42.248 INFO:teuthology.orchestra.run.vm04.stdout:Transaction test succeeded. 2026-03-06T13:50:42.248 INFO:teuthology.orchestra.run.vm04.stdout:Running transaction 2026-03-06T13:50:42.255 INFO:teuthology.orchestra.run.vm06.stdout:Transaction test succeeded. 2026-03-06T13:50:42.255 INFO:teuthology.orchestra.run.vm06.stdout:Running transaction 2026-03-06T13:50:42.296 INFO:teuthology.orchestra.run.vm02.stdout: Preparing : 1/1 2026-03-06T13:50:42.301 INFO:teuthology.orchestra.run.vm02.stdout: Erasing : ceph-test-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 1/3 2026-03-06T13:50:42.304 INFO:teuthology.orchestra.run.vm02.stdout: Erasing : socat-1.7.4.1-8.el9.x86_64 2/3 2026-03-06T13:50:42.307 INFO:teuthology.orchestra.run.vm04.stdout: Preparing : 1/1 2026-03-06T13:50:42.313 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : ceph-test-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 1/3 2026-03-06T13:50:42.314 INFO:teuthology.orchestra.run.vm06.stdout: Preparing : 1/1 2026-03-06T13:50:42.315 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : socat-1.7.4.1-8.el9.x86_64 2/3 2026-03-06T13:50:42.318 INFO:teuthology.orchestra.run.vm02.stdout: Erasing : xmlstarlet-1.6.1-20.el9.x86_64 3/3 2026-03-06T13:50:42.320 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : ceph-test-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 1/3 2026-03-06T13:50:42.322 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : socat-1.7.4.1-8.el9.x86_64 2/3 2026-03-06T13:50:42.329 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : xmlstarlet-1.6.1-20.el9.x86_64 3/3 2026-03-06T13:50:42.336 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : xmlstarlet-1.6.1-20.el9.x86_64 3/3 2026-03-06T13:50:42.379 INFO:teuthology.orchestra.run.vm02.stdout: Running scriptlet: xmlstarlet-1.6.1-20.el9.x86_64 3/3 2026-03-06T13:50:42.380 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : ceph-test-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 1/3 2026-03-06T13:50:42.380 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : socat-1.7.4.1-8.el9.x86_64 2/3 2026-03-06T13:50:42.390 INFO:teuthology.orchestra.run.vm04.stdout: Running scriptlet: xmlstarlet-1.6.1-20.el9.x86_64 3/3 2026-03-06T13:50:42.390 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : ceph-test-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 1/3 2026-03-06T13:50:42.390 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : socat-1.7.4.1-8.el9.x86_64 2/3 2026-03-06T13:50:42.404 INFO:teuthology.orchestra.run.vm06.stdout: Running scriptlet: xmlstarlet-1.6.1-20.el9.x86_64 3/3 2026-03-06T13:50:42.404 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : ceph-test-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 1/3 2026-03-06T13:50:42.404 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : socat-1.7.4.1-8.el9.x86_64 2/3 2026-03-06T13:50:42.431 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : xmlstarlet-1.6.1-20.el9.x86_64 3/3 2026-03-06T13:50:42.431 INFO:teuthology.orchestra.run.vm02.stdout: 2026-03-06T13:50:42.431 INFO:teuthology.orchestra.run.vm02.stdout:Removed: 2026-03-06T13:50:42.431 INFO:teuthology.orchestra.run.vm02.stdout: ceph-test-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 2026-03-06T13:50:42.431 INFO:teuthology.orchestra.run.vm02.stdout: socat-1.7.4.1-8.el9.x86_64 2026-03-06T13:50:42.431 INFO:teuthology.orchestra.run.vm02.stdout: xmlstarlet-1.6.1-20.el9.x86_64 2026-03-06T13:50:42.431 INFO:teuthology.orchestra.run.vm02.stdout: 2026-03-06T13:50:42.431 INFO:teuthology.orchestra.run.vm02.stdout:Complete! 2026-03-06T13:50:42.439 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : xmlstarlet-1.6.1-20.el9.x86_64 3/3 2026-03-06T13:50:42.440 INFO:teuthology.orchestra.run.vm04.stdout: 2026-03-06T13:50:42.440 INFO:teuthology.orchestra.run.vm04.stdout:Removed: 2026-03-06T13:50:42.440 INFO:teuthology.orchestra.run.vm04.stdout: ceph-test-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 2026-03-06T13:50:42.440 INFO:teuthology.orchestra.run.vm04.stdout: socat-1.7.4.1-8.el9.x86_64 2026-03-06T13:50:42.440 INFO:teuthology.orchestra.run.vm04.stdout: xmlstarlet-1.6.1-20.el9.x86_64 2026-03-06T13:50:42.440 INFO:teuthology.orchestra.run.vm04.stdout: 2026-03-06T13:50:42.440 INFO:teuthology.orchestra.run.vm04.stdout:Complete! 2026-03-06T13:50:42.452 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : xmlstarlet-1.6.1-20.el9.x86_64 3/3 2026-03-06T13:50:42.452 INFO:teuthology.orchestra.run.vm06.stdout: 2026-03-06T13:50:42.452 INFO:teuthology.orchestra.run.vm06.stdout:Removed: 2026-03-06T13:50:42.452 INFO:teuthology.orchestra.run.vm06.stdout: ceph-test-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 2026-03-06T13:50:42.452 INFO:teuthology.orchestra.run.vm06.stdout: socat-1.7.4.1-8.el9.x86_64 2026-03-06T13:50:42.452 INFO:teuthology.orchestra.run.vm06.stdout: xmlstarlet-1.6.1-20.el9.x86_64 2026-03-06T13:50:42.452 INFO:teuthology.orchestra.run.vm06.stdout: 2026-03-06T13:50:42.452 INFO:teuthology.orchestra.run.vm06.stdout:Complete! 2026-03-06T13:50:42.635 INFO:teuthology.orchestra.run.vm02.stdout:Dependencies resolved. 2026-03-06T13:50:42.636 INFO:teuthology.orchestra.run.vm02.stdout:================================================================================ 2026-03-06T13:50:42.636 INFO:teuthology.orchestra.run.vm02.stdout: Package Arch Version Repository Size 2026-03-06T13:50:42.636 INFO:teuthology.orchestra.run.vm02.stdout:================================================================================ 2026-03-06T13:50:42.636 INFO:teuthology.orchestra.run.vm02.stdout:Removing: 2026-03-06T13:50:42.636 INFO:teuthology.orchestra.run.vm02.stdout: ceph x86_64 2:19.2.3-47.gc24117fd552.el9.clyso @ceph 0 2026-03-06T13:50:42.636 INFO:teuthology.orchestra.run.vm02.stdout:Removing unused dependencies: 2026-03-06T13:50:42.636 INFO:teuthology.orchestra.run.vm02.stdout: ceph-mds x86_64 2:19.2.3-47.gc24117fd552.el9.clyso @ceph 7.4 M 2026-03-06T13:50:42.636 INFO:teuthology.orchestra.run.vm02.stdout: ceph-mon x86_64 2:19.2.3-47.gc24117fd552.el9.clyso @ceph 18 M 2026-03-06T13:50:42.636 INFO:teuthology.orchestra.run.vm02.stdout: lua x86_64 5.4.4-4.el9 @appstream 593 k 2026-03-06T13:50:42.636 INFO:teuthology.orchestra.run.vm02.stdout: lua-devel x86_64 5.4.4-4.el9 @crb 49 k 2026-03-06T13:50:42.636 INFO:teuthology.orchestra.run.vm02.stdout: luarocks noarch 3.9.2-5.el9 @epel 692 k 2026-03-06T13:50:42.636 INFO:teuthology.orchestra.run.vm02.stdout: unzip x86_64 6.0-59.el9 @baseos 389 k 2026-03-06T13:50:42.636 INFO:teuthology.orchestra.run.vm02.stdout: zip x86_64 3.0-35.el9 @baseos 724 k 2026-03-06T13:50:42.636 INFO:teuthology.orchestra.run.vm02.stdout: 2026-03-06T13:50:42.636 INFO:teuthology.orchestra.run.vm02.stdout:Transaction Summary 2026-03-06T13:50:42.636 INFO:teuthology.orchestra.run.vm02.stdout:================================================================================ 2026-03-06T13:50:42.636 INFO:teuthology.orchestra.run.vm02.stdout:Remove 8 Packages 2026-03-06T13:50:42.637 INFO:teuthology.orchestra.run.vm02.stdout: 2026-03-06T13:50:42.637 INFO:teuthology.orchestra.run.vm02.stdout:Freed space: 28 M 2026-03-06T13:50:42.637 INFO:teuthology.orchestra.run.vm02.stdout:Running transaction check 2026-03-06T13:50:42.639 INFO:teuthology.orchestra.run.vm02.stdout:Transaction check succeeded. 2026-03-06T13:50:42.639 INFO:teuthology.orchestra.run.vm02.stdout:Running transaction test 2026-03-06T13:50:42.646 INFO:teuthology.orchestra.run.vm04.stdout:Dependencies resolved. 2026-03-06T13:50:42.646 INFO:teuthology.orchestra.run.vm04.stdout:================================================================================ 2026-03-06T13:50:42.646 INFO:teuthology.orchestra.run.vm04.stdout: Package Arch Version Repository Size 2026-03-06T13:50:42.646 INFO:teuthology.orchestra.run.vm04.stdout:================================================================================ 2026-03-06T13:50:42.646 INFO:teuthology.orchestra.run.vm04.stdout:Removing: 2026-03-06T13:50:42.646 INFO:teuthology.orchestra.run.vm04.stdout: ceph x86_64 2:19.2.3-47.gc24117fd552.el9.clyso @ceph 0 2026-03-06T13:50:42.646 INFO:teuthology.orchestra.run.vm04.stdout:Removing unused dependencies: 2026-03-06T13:50:42.646 INFO:teuthology.orchestra.run.vm04.stdout: ceph-mds x86_64 2:19.2.3-47.gc24117fd552.el9.clyso @ceph 7.4 M 2026-03-06T13:50:42.646 INFO:teuthology.orchestra.run.vm04.stdout: ceph-mon x86_64 2:19.2.3-47.gc24117fd552.el9.clyso @ceph 18 M 2026-03-06T13:50:42.646 INFO:teuthology.orchestra.run.vm04.stdout: lua x86_64 5.4.4-4.el9 @appstream 593 k 2026-03-06T13:50:42.646 INFO:teuthology.orchestra.run.vm04.stdout: lua-devel x86_64 5.4.4-4.el9 @crb 49 k 2026-03-06T13:50:42.646 INFO:teuthology.orchestra.run.vm04.stdout: luarocks noarch 3.9.2-5.el9 @epel 692 k 2026-03-06T13:50:42.646 INFO:teuthology.orchestra.run.vm04.stdout: unzip x86_64 6.0-59.el9 @baseos 389 k 2026-03-06T13:50:42.646 INFO:teuthology.orchestra.run.vm04.stdout: zip x86_64 3.0-35.el9 @baseos 724 k 2026-03-06T13:50:42.646 INFO:teuthology.orchestra.run.vm04.stdout: 2026-03-06T13:50:42.646 INFO:teuthology.orchestra.run.vm04.stdout:Transaction Summary 2026-03-06T13:50:42.646 INFO:teuthology.orchestra.run.vm04.stdout:================================================================================ 2026-03-06T13:50:42.646 INFO:teuthology.orchestra.run.vm04.stdout:Remove 8 Packages 2026-03-06T13:50:42.646 INFO:teuthology.orchestra.run.vm04.stdout: 2026-03-06T13:50:42.647 INFO:teuthology.orchestra.run.vm04.stdout:Freed space: 28 M 2026-03-06T13:50:42.647 INFO:teuthology.orchestra.run.vm04.stdout:Running transaction check 2026-03-06T13:50:42.649 INFO:teuthology.orchestra.run.vm04.stdout:Transaction check succeeded. 2026-03-06T13:50:42.649 INFO:teuthology.orchestra.run.vm04.stdout:Running transaction test 2026-03-06T13:50:42.652 INFO:teuthology.orchestra.run.vm06.stdout:Dependencies resolved. 2026-03-06T13:50:42.653 INFO:teuthology.orchestra.run.vm06.stdout:================================================================================ 2026-03-06T13:50:42.653 INFO:teuthology.orchestra.run.vm06.stdout: Package Arch Version Repository Size 2026-03-06T13:50:42.653 INFO:teuthology.orchestra.run.vm06.stdout:================================================================================ 2026-03-06T13:50:42.653 INFO:teuthology.orchestra.run.vm06.stdout:Removing: 2026-03-06T13:50:42.653 INFO:teuthology.orchestra.run.vm06.stdout: ceph x86_64 2:19.2.3-47.gc24117fd552.el9.clyso @ceph 0 2026-03-06T13:50:42.653 INFO:teuthology.orchestra.run.vm06.stdout:Removing unused dependencies: 2026-03-06T13:50:42.653 INFO:teuthology.orchestra.run.vm06.stdout: ceph-mds x86_64 2:19.2.3-47.gc24117fd552.el9.clyso @ceph 7.4 M 2026-03-06T13:50:42.653 INFO:teuthology.orchestra.run.vm06.stdout: ceph-mon x86_64 2:19.2.3-47.gc24117fd552.el9.clyso @ceph 18 M 2026-03-06T13:50:42.653 INFO:teuthology.orchestra.run.vm06.stdout: lua x86_64 5.4.4-4.el9 @appstream 593 k 2026-03-06T13:50:42.653 INFO:teuthology.orchestra.run.vm06.stdout: lua-devel x86_64 5.4.4-4.el9 @crb 49 k 2026-03-06T13:50:42.653 INFO:teuthology.orchestra.run.vm06.stdout: luarocks noarch 3.9.2-5.el9 @epel 692 k 2026-03-06T13:50:42.653 INFO:teuthology.orchestra.run.vm06.stdout: unzip x86_64 6.0-59.el9 @baseos 389 k 2026-03-06T13:50:42.653 INFO:teuthology.orchestra.run.vm06.stdout: zip x86_64 3.0-35.el9 @baseos 724 k 2026-03-06T13:50:42.653 INFO:teuthology.orchestra.run.vm06.stdout: 2026-03-06T13:50:42.653 INFO:teuthology.orchestra.run.vm06.stdout:Transaction Summary 2026-03-06T13:50:42.653 INFO:teuthology.orchestra.run.vm06.stdout:================================================================================ 2026-03-06T13:50:42.653 INFO:teuthology.orchestra.run.vm06.stdout:Remove 8 Packages 2026-03-06T13:50:42.653 INFO:teuthology.orchestra.run.vm06.stdout: 2026-03-06T13:50:42.653 INFO:teuthology.orchestra.run.vm06.stdout:Freed space: 28 M 2026-03-06T13:50:42.653 INFO:teuthology.orchestra.run.vm06.stdout:Running transaction check 2026-03-06T13:50:42.656 INFO:teuthology.orchestra.run.vm06.stdout:Transaction check succeeded. 2026-03-06T13:50:42.656 INFO:teuthology.orchestra.run.vm06.stdout:Running transaction test 2026-03-06T13:50:42.660 INFO:teuthology.orchestra.run.vm02.stdout:Transaction test succeeded. 2026-03-06T13:50:42.660 INFO:teuthology.orchestra.run.vm02.stdout:Running transaction 2026-03-06T13:50:42.671 INFO:teuthology.orchestra.run.vm04.stdout:Transaction test succeeded. 2026-03-06T13:50:42.672 INFO:teuthology.orchestra.run.vm04.stdout:Running transaction 2026-03-06T13:50:42.678 INFO:teuthology.orchestra.run.vm06.stdout:Transaction test succeeded. 2026-03-06T13:50:42.678 INFO:teuthology.orchestra.run.vm06.stdout:Running transaction 2026-03-06T13:50:42.699 INFO:teuthology.orchestra.run.vm02.stdout: Preparing : 1/1 2026-03-06T13:50:42.704 INFO:teuthology.orchestra.run.vm02.stdout: Erasing : ceph-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 1/8 2026-03-06T13:50:42.706 INFO:teuthology.orchestra.run.vm02.stdout: Erasing : luarocks-3.9.2-5.el9.noarch 2/8 2026-03-06T13:50:42.708 INFO:teuthology.orchestra.run.vm02.stdout: Erasing : lua-devel-5.4.4-4.el9.x86_64 3/8 2026-03-06T13:50:42.711 INFO:teuthology.orchestra.run.vm02.stdout: Erasing : zip-3.0-35.el9.x86_64 4/8 2026-03-06T13:50:42.712 INFO:teuthology.orchestra.run.vm04.stdout: Preparing : 1/1 2026-03-06T13:50:42.713 INFO:teuthology.orchestra.run.vm02.stdout: Erasing : unzip-6.0-59.el9.x86_64 5/8 2026-03-06T13:50:42.715 INFO:teuthology.orchestra.run.vm02.stdout: Erasing : lua-5.4.4-4.el9.x86_64 6/8 2026-03-06T13:50:42.717 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : ceph-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 1/8 2026-03-06T13:50:42.717 INFO:teuthology.orchestra.run.vm06.stdout: Preparing : 1/1 2026-03-06T13:50:42.720 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : luarocks-3.9.2-5.el9.noarch 2/8 2026-03-06T13:50:42.721 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : lua-devel-5.4.4-4.el9.x86_64 3/8 2026-03-06T13:50:42.722 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : ceph-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 1/8 2026-03-06T13:50:42.724 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : zip-3.0-35.el9.x86_64 4/8 2026-03-06T13:50:42.725 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : luarocks-3.9.2-5.el9.noarch 2/8 2026-03-06T13:50:42.726 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : unzip-6.0-59.el9.x86_64 5/8 2026-03-06T13:50:42.727 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : lua-devel-5.4.4-4.el9.x86_64 3/8 2026-03-06T13:50:42.728 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : lua-5.4.4-4.el9.x86_64 6/8 2026-03-06T13:50:42.729 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : zip-3.0-35.el9.x86_64 4/8 2026-03-06T13:50:42.731 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : unzip-6.0-59.el9.x86_64 5/8 2026-03-06T13:50:42.733 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : lua-5.4.4-4.el9.x86_64 6/8 2026-03-06T13:50:42.734 INFO:teuthology.orchestra.run.vm02.stdout: Running scriptlet: ceph-mds-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 7/8 2026-03-06T13:50:42.734 INFO:teuthology.orchestra.run.vm02.stdout:Glob pattern passed to enable, but globs are not supported for this. 2026-03-06T13:50:42.734 INFO:teuthology.orchestra.run.vm02.stdout:Invalid unit name "ceph-mds@*.service" escaped as "ceph-mds@\x2a.service". 2026-03-06T13:50:42.734 INFO:teuthology.orchestra.run.vm02.stdout:Removed "/etc/systemd/system/multi-user.target.wants/ceph-mds.target". 2026-03-06T13:50:42.734 INFO:teuthology.orchestra.run.vm02.stdout:Removed "/etc/systemd/system/ceph.target.wants/ceph-mds.target". 2026-03-06T13:50:42.734 INFO:teuthology.orchestra.run.vm02.stdout: 2026-03-06T13:50:42.735 INFO:teuthology.orchestra.run.vm02.stdout: Erasing : ceph-mds-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 7/8 2026-03-06T13:50:42.741 INFO:teuthology.orchestra.run.vm02.stdout: Running scriptlet: ceph-mds-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 7/8 2026-03-06T13:50:42.748 INFO:teuthology.orchestra.run.vm04.stdout: Running scriptlet: ceph-mds-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 7/8 2026-03-06T13:50:42.748 INFO:teuthology.orchestra.run.vm04.stdout:Glob pattern passed to enable, but globs are not supported for this. 2026-03-06T13:50:42.748 INFO:teuthology.orchestra.run.vm04.stdout:Invalid unit name "ceph-mds@*.service" escaped as "ceph-mds@\x2a.service". 2026-03-06T13:50:42.748 INFO:teuthology.orchestra.run.vm04.stdout:Removed "/etc/systemd/system/multi-user.target.wants/ceph-mds.target". 2026-03-06T13:50:42.748 INFO:teuthology.orchestra.run.vm04.stdout:Removed "/etc/systemd/system/ceph.target.wants/ceph-mds.target". 2026-03-06T13:50:42.748 INFO:teuthology.orchestra.run.vm04.stdout: 2026-03-06T13:50:42.749 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : ceph-mds-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 7/8 2026-03-06T13:50:42.752 INFO:teuthology.orchestra.run.vm06.stdout: Running scriptlet: ceph-mds-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 7/8 2026-03-06T13:50:42.752 INFO:teuthology.orchestra.run.vm06.stdout:Glob pattern passed to enable, but globs are not supported for this. 2026-03-06T13:50:42.752 INFO:teuthology.orchestra.run.vm06.stdout:Invalid unit name "ceph-mds@*.service" escaped as "ceph-mds@\x2a.service". 2026-03-06T13:50:42.752 INFO:teuthology.orchestra.run.vm06.stdout:Removed "/etc/systemd/system/multi-user.target.wants/ceph-mds.target". 2026-03-06T13:50:42.752 INFO:teuthology.orchestra.run.vm06.stdout:Removed "/etc/systemd/system/ceph.target.wants/ceph-mds.target". 2026-03-06T13:50:42.752 INFO:teuthology.orchestra.run.vm06.stdout: 2026-03-06T13:50:42.753 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : ceph-mds-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 7/8 2026-03-06T13:50:42.755 INFO:teuthology.orchestra.run.vm04.stdout: Running scriptlet: ceph-mds-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 7/8 2026-03-06T13:50:42.760 INFO:teuthology.orchestra.run.vm06.stdout: Running scriptlet: ceph-mds-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 7/8 2026-03-06T13:50:42.761 INFO:teuthology.orchestra.run.vm02.stdout: Running scriptlet: ceph-mon-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 8/8 2026-03-06T13:50:42.761 INFO:teuthology.orchestra.run.vm02.stdout:Glob pattern passed to enable, but globs are not supported for this. 2026-03-06T13:50:42.761 INFO:teuthology.orchestra.run.vm02.stdout:Invalid unit name "ceph-mon@*.service" escaped as "ceph-mon@\x2a.service". 2026-03-06T13:50:42.761 INFO:teuthology.orchestra.run.vm02.stdout:Removed "/etc/systemd/system/multi-user.target.wants/ceph-mon.target". 2026-03-06T13:50:42.761 INFO:teuthology.orchestra.run.vm02.stdout:Removed "/etc/systemd/system/ceph.target.wants/ceph-mon.target". 2026-03-06T13:50:42.761 INFO:teuthology.orchestra.run.vm02.stdout: 2026-03-06T13:50:42.762 INFO:teuthology.orchestra.run.vm02.stdout: Erasing : ceph-mon-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 8/8 2026-03-06T13:50:42.774 INFO:teuthology.orchestra.run.vm04.stdout: Running scriptlet: ceph-mon-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 8/8 2026-03-06T13:50:42.774 INFO:teuthology.orchestra.run.vm04.stdout:Glob pattern passed to enable, but globs are not supported for this. 2026-03-06T13:50:42.775 INFO:teuthology.orchestra.run.vm04.stdout:Invalid unit name "ceph-mon@*.service" escaped as "ceph-mon@\x2a.service". 2026-03-06T13:50:42.775 INFO:teuthology.orchestra.run.vm04.stdout:Removed "/etc/systemd/system/multi-user.target.wants/ceph-mon.target". 2026-03-06T13:50:42.775 INFO:teuthology.orchestra.run.vm04.stdout:Removed "/etc/systemd/system/ceph.target.wants/ceph-mon.target". 2026-03-06T13:50:42.775 INFO:teuthology.orchestra.run.vm04.stdout: 2026-03-06T13:50:42.776 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : ceph-mon-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 8/8 2026-03-06T13:50:42.779 INFO:teuthology.orchestra.run.vm06.stdout: Running scriptlet: ceph-mon-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 8/8 2026-03-06T13:50:42.779 INFO:teuthology.orchestra.run.vm06.stdout:Glob pattern passed to enable, but globs are not supported for this. 2026-03-06T13:50:42.779 INFO:teuthology.orchestra.run.vm06.stdout:Invalid unit name "ceph-mon@*.service" escaped as "ceph-mon@\x2a.service". 2026-03-06T13:50:42.779 INFO:teuthology.orchestra.run.vm06.stdout:Removed "/etc/systemd/system/multi-user.target.wants/ceph-mon.target". 2026-03-06T13:50:42.779 INFO:teuthology.orchestra.run.vm06.stdout:Removed "/etc/systemd/system/ceph.target.wants/ceph-mon.target". 2026-03-06T13:50:42.779 INFO:teuthology.orchestra.run.vm06.stdout: 2026-03-06T13:50:42.781 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : ceph-mon-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 8/8 2026-03-06T13:50:42.853 INFO:teuthology.orchestra.run.vm02.stdout: Running scriptlet: ceph-mon-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 8/8 2026-03-06T13:50:42.853 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : ceph-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 1/8 2026-03-06T13:50:42.853 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : ceph-mds-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 2/8 2026-03-06T13:50:42.853 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : ceph-mon-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 3/8 2026-03-06T13:50:42.853 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : lua-5.4.4-4.el9.x86_64 4/8 2026-03-06T13:50:42.853 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : lua-devel-5.4.4-4.el9.x86_64 5/8 2026-03-06T13:50:42.853 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : luarocks-3.9.2-5.el9.noarch 6/8 2026-03-06T13:50:42.853 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : unzip-6.0-59.el9.x86_64 7/8 2026-03-06T13:50:42.862 INFO:teuthology.orchestra.run.vm04.stdout: Running scriptlet: ceph-mon-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 8/8 2026-03-06T13:50:42.862 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : ceph-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 1/8 2026-03-06T13:50:42.863 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : ceph-mds-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 2/8 2026-03-06T13:50:42.863 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : ceph-mon-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 3/8 2026-03-06T13:50:42.863 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : lua-5.4.4-4.el9.x86_64 4/8 2026-03-06T13:50:42.863 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : lua-devel-5.4.4-4.el9.x86_64 5/8 2026-03-06T13:50:42.863 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : luarocks-3.9.2-5.el9.noarch 6/8 2026-03-06T13:50:42.863 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : unzip-6.0-59.el9.x86_64 7/8 2026-03-06T13:50:42.868 INFO:teuthology.orchestra.run.vm06.stdout: Running scriptlet: ceph-mon-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 8/8 2026-03-06T13:50:42.868 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : ceph-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 1/8 2026-03-06T13:50:42.868 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : ceph-mds-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 2/8 2026-03-06T13:50:42.868 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : ceph-mon-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 3/8 2026-03-06T13:50:42.868 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : lua-5.4.4-4.el9.x86_64 4/8 2026-03-06T13:50:42.868 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : lua-devel-5.4.4-4.el9.x86_64 5/8 2026-03-06T13:50:42.868 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : luarocks-3.9.2-5.el9.noarch 6/8 2026-03-06T13:50:42.868 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : unzip-6.0-59.el9.x86_64 7/8 2026-03-06T13:50:42.906 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : zip-3.0-35.el9.x86_64 8/8 2026-03-06T13:50:42.906 INFO:teuthology.orchestra.run.vm02.stdout: 2026-03-06T13:50:42.906 INFO:teuthology.orchestra.run.vm02.stdout:Removed: 2026-03-06T13:50:42.906 INFO:teuthology.orchestra.run.vm02.stdout: ceph-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 2026-03-06T13:50:42.906 INFO:teuthology.orchestra.run.vm02.stdout: ceph-mds-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 2026-03-06T13:50:42.907 INFO:teuthology.orchestra.run.vm02.stdout: ceph-mon-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 2026-03-06T13:50:42.907 INFO:teuthology.orchestra.run.vm02.stdout: lua-5.4.4-4.el9.x86_64 2026-03-06T13:50:42.907 INFO:teuthology.orchestra.run.vm02.stdout: lua-devel-5.4.4-4.el9.x86_64 2026-03-06T13:50:42.907 INFO:teuthology.orchestra.run.vm02.stdout: luarocks-3.9.2-5.el9.noarch 2026-03-06T13:50:42.907 INFO:teuthology.orchestra.run.vm02.stdout: unzip-6.0-59.el9.x86_64 2026-03-06T13:50:42.907 INFO:teuthology.orchestra.run.vm02.stdout: zip-3.0-35.el9.x86_64 2026-03-06T13:50:42.907 INFO:teuthology.orchestra.run.vm02.stdout: 2026-03-06T13:50:42.907 INFO:teuthology.orchestra.run.vm02.stdout:Complete! 2026-03-06T13:50:42.919 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : zip-3.0-35.el9.x86_64 8/8 2026-03-06T13:50:42.919 INFO:teuthology.orchestra.run.vm06.stdout: 2026-03-06T13:50:42.919 INFO:teuthology.orchestra.run.vm06.stdout:Removed: 2026-03-06T13:50:42.920 INFO:teuthology.orchestra.run.vm06.stdout: ceph-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 2026-03-06T13:50:42.920 INFO:teuthology.orchestra.run.vm06.stdout: ceph-mds-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 2026-03-06T13:50:42.920 INFO:teuthology.orchestra.run.vm06.stdout: ceph-mon-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 2026-03-06T13:50:42.920 INFO:teuthology.orchestra.run.vm06.stdout: lua-5.4.4-4.el9.x86_64 2026-03-06T13:50:42.920 INFO:teuthology.orchestra.run.vm06.stdout: lua-devel-5.4.4-4.el9.x86_64 2026-03-06T13:50:42.920 INFO:teuthology.orchestra.run.vm06.stdout: luarocks-3.9.2-5.el9.noarch 2026-03-06T13:50:42.920 INFO:teuthology.orchestra.run.vm06.stdout: unzip-6.0-59.el9.x86_64 2026-03-06T13:50:42.920 INFO:teuthology.orchestra.run.vm06.stdout: zip-3.0-35.el9.x86_64 2026-03-06T13:50:42.920 INFO:teuthology.orchestra.run.vm06.stdout: 2026-03-06T13:50:42.920 INFO:teuthology.orchestra.run.vm06.stdout:Complete! 2026-03-06T13:50:42.920 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : zip-3.0-35.el9.x86_64 8/8 2026-03-06T13:50:42.920 INFO:teuthology.orchestra.run.vm04.stdout: 2026-03-06T13:50:42.920 INFO:teuthology.orchestra.run.vm04.stdout:Removed: 2026-03-06T13:50:42.920 INFO:teuthology.orchestra.run.vm04.stdout: ceph-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 2026-03-06T13:50:42.920 INFO:teuthology.orchestra.run.vm04.stdout: ceph-mds-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 2026-03-06T13:50:42.920 INFO:teuthology.orchestra.run.vm04.stdout: ceph-mon-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 2026-03-06T13:50:42.920 INFO:teuthology.orchestra.run.vm04.stdout: lua-5.4.4-4.el9.x86_64 2026-03-06T13:50:42.920 INFO:teuthology.orchestra.run.vm04.stdout: lua-devel-5.4.4-4.el9.x86_64 2026-03-06T13:50:42.920 INFO:teuthology.orchestra.run.vm04.stdout: luarocks-3.9.2-5.el9.noarch 2026-03-06T13:50:42.920 INFO:teuthology.orchestra.run.vm04.stdout: unzip-6.0-59.el9.x86_64 2026-03-06T13:50:42.920 INFO:teuthology.orchestra.run.vm04.stdout: zip-3.0-35.el9.x86_64 2026-03-06T13:50:42.920 INFO:teuthology.orchestra.run.vm04.stdout: 2026-03-06T13:50:42.920 INFO:teuthology.orchestra.run.vm04.stdout:Complete! 2026-03-06T13:50:43.115 INFO:teuthology.orchestra.run.vm02.stdout:Dependencies resolved. 2026-03-06T13:50:43.121 INFO:teuthology.orchestra.run.vm02.stdout:=================================================================================================== 2026-03-06T13:50:43.121 INFO:teuthology.orchestra.run.vm02.stdout: Package Arch Version Repository Size 2026-03-06T13:50:43.121 INFO:teuthology.orchestra.run.vm02.stdout:=================================================================================================== 2026-03-06T13:50:43.121 INFO:teuthology.orchestra.run.vm02.stdout:Removing: 2026-03-06T13:50:43.121 INFO:teuthology.orchestra.run.vm02.stdout: ceph-base x86_64 2:19.2.3-47.gc24117fd552.el9.clyso @ceph 23 M 2026-03-06T13:50:43.121 INFO:teuthology.orchestra.run.vm02.stdout:Removing dependent packages: 2026-03-06T13:50:43.121 INFO:teuthology.orchestra.run.vm02.stdout: ceph-immutable-object-cache x86_64 2:19.2.3-47.gc24117fd552.el9.clyso @ceph 431 k 2026-03-06T13:50:43.121 INFO:teuthology.orchestra.run.vm02.stdout: ceph-mgr x86_64 2:19.2.3-47.gc24117fd552.el9.clyso @ceph 3.4 M 2026-03-06T13:50:43.121 INFO:teuthology.orchestra.run.vm02.stdout: ceph-mgr-cephadm noarch 2:19.2.3-47.gc24117fd552.el9.clyso @ceph-noarch 803 k 2026-03-06T13:50:43.121 INFO:teuthology.orchestra.run.vm02.stdout: ceph-mgr-dashboard noarch 2:19.2.3-47.gc24117fd552.el9.clyso @ceph-noarch 88 M 2026-03-06T13:50:43.121 INFO:teuthology.orchestra.run.vm02.stdout: ceph-mgr-diskprediction-local noarch 2:19.2.3-47.gc24117fd552.el9.clyso @ceph-noarch 66 M 2026-03-06T13:50:43.121 INFO:teuthology.orchestra.run.vm02.stdout: ceph-mgr-rook noarch 2:19.2.3-47.gc24117fd552.el9.clyso @ceph-noarch 563 k 2026-03-06T13:50:43.121 INFO:teuthology.orchestra.run.vm02.stdout: ceph-osd x86_64 2:19.2.3-47.gc24117fd552.el9.clyso @ceph 59 M 2026-03-06T13:50:43.121 INFO:teuthology.orchestra.run.vm02.stdout: ceph-volume noarch 2:19.2.3-47.gc24117fd552.el9.clyso @ceph-noarch 1.4 M 2026-03-06T13:50:43.122 INFO:teuthology.orchestra.run.vm02.stdout: rbd-mirror x86_64 2:19.2.3-47.gc24117fd552.el9.clyso @ceph 13 M 2026-03-06T13:50:43.122 INFO:teuthology.orchestra.run.vm02.stdout:Removing unused dependencies: 2026-03-06T13:50:43.122 INFO:teuthology.orchestra.run.vm02.stdout: abseil-cpp x86_64 20211102.0-4.el9 @epel 1.9 M 2026-03-06T13:50:43.122 INFO:teuthology.orchestra.run.vm02.stdout: ceph-common x86_64 2:19.2.3-47.gc24117fd552.el9.clyso @ceph 85 M 2026-03-06T13:50:43.122 INFO:teuthology.orchestra.run.vm02.stdout: ceph-grafana-dashboards noarch 2:19.2.3-47.gc24117fd552.el9.clyso @ceph-noarch 626 k 2026-03-06T13:50:43.122 INFO:teuthology.orchestra.run.vm02.stdout: ceph-mgr-k8sevents noarch 2:19.2.3-47.gc24117fd552.el9.clyso @ceph-noarch 60 k 2026-03-06T13:50:43.122 INFO:teuthology.orchestra.run.vm02.stdout: ceph-mgr-modules-core noarch 2:19.2.3-47.gc24117fd552.el9.clyso @ceph-noarch 1.5 M 2026-03-06T13:50:43.122 INFO:teuthology.orchestra.run.vm02.stdout: ceph-prometheus-alerts noarch 2:19.2.3-47.gc24117fd552.el9.clyso @ceph-noarch 51 k 2026-03-06T13:50:43.122 INFO:teuthology.orchestra.run.vm02.stdout: ceph-selinux x86_64 2:19.2.3-47.gc24117fd552.el9.clyso @ceph 138 k 2026-03-06T13:50:43.122 INFO:teuthology.orchestra.run.vm02.stdout: cryptsetup x86_64 2.8.1-3.el9 @baseos 770 k 2026-03-06T13:50:43.122 INFO:teuthology.orchestra.run.vm02.stdout: flexiblas x86_64 3.0.4-9.el9 @appstream 68 k 2026-03-06T13:50:43.122 INFO:teuthology.orchestra.run.vm02.stdout: flexiblas-netlib x86_64 3.0.4-9.el9 @appstream 11 M 2026-03-06T13:50:43.122 INFO:teuthology.orchestra.run.vm02.stdout: flexiblas-openblas-openmp x86_64 3.0.4-9.el9 @appstream 39 k 2026-03-06T13:50:43.122 INFO:teuthology.orchestra.run.vm02.stdout: gperftools-libs x86_64 2.9.1-3.el9 @epel 1.4 M 2026-03-06T13:50:43.122 INFO:teuthology.orchestra.run.vm02.stdout: grpc-data noarch 1.46.7-10.el9 @epel 13 k 2026-03-06T13:50:43.122 INFO:teuthology.orchestra.run.vm02.stdout: ledmon-libs x86_64 1.1.0-3.el9 @baseos 80 k 2026-03-06T13:50:43.122 INFO:teuthology.orchestra.run.vm02.stdout: libcephsqlite x86_64 2:19.2.3-47.gc24117fd552.el9.clyso @ceph 425 k 2026-03-06T13:50:43.122 INFO:teuthology.orchestra.run.vm02.stdout: libconfig x86_64 1.7.2-9.el9 @baseos 220 k 2026-03-06T13:50:43.122 INFO:teuthology.orchestra.run.vm02.stdout: libgfortran x86_64 11.5.0-14.el9 @baseos 2.8 M 2026-03-06T13:50:43.122 INFO:teuthology.orchestra.run.vm02.stdout: liboath x86_64 2.6.12-1.el9 @epel 94 k 2026-03-06T13:50:43.122 INFO:teuthology.orchestra.run.vm02.stdout: libquadmath x86_64 11.5.0-14.el9 @baseos 330 k 2026-03-06T13:50:43.122 INFO:teuthology.orchestra.run.vm02.stdout: libradosstriper1 x86_64 2:19.2.3-47.gc24117fd552.el9.clyso @ceph 1.6 M 2026-03-06T13:50:43.122 INFO:teuthology.orchestra.run.vm02.stdout: libstoragemgmt x86_64 1.10.1-1.el9 @appstream 685 k 2026-03-06T13:50:43.122 INFO:teuthology.orchestra.run.vm02.stdout: libunwind x86_64 1.6.2-1.el9 @epel 170 k 2026-03-06T13:50:43.122 INFO:teuthology.orchestra.run.vm02.stdout: libxslt x86_64 1.1.34-12.el9 @appstream 743 k 2026-03-06T13:50:43.122 INFO:teuthology.orchestra.run.vm02.stdout: openblas x86_64 0.3.29-1.el9 @appstream 112 k 2026-03-06T13:50:43.122 INFO:teuthology.orchestra.run.vm02.stdout: openblas-openmp x86_64 0.3.29-1.el9 @appstream 46 M 2026-03-06T13:50:43.122 INFO:teuthology.orchestra.run.vm02.stdout: pciutils x86_64 3.7.0-7.el9 @baseos 216 k 2026-03-06T13:50:43.122 INFO:teuthology.orchestra.run.vm02.stdout: protobuf x86_64 3.14.0-17.el9 @appstream 3.5 M 2026-03-06T13:50:43.122 INFO:teuthology.orchestra.run.vm02.stdout: protobuf-compiler x86_64 3.14.0-17.el9 @crb 2.9 M 2026-03-06T13:50:43.122 INFO:teuthology.orchestra.run.vm02.stdout: python3-asyncssh noarch 2.13.2-5.el9 @epel 3.9 M 2026-03-06T13:50:43.122 INFO:teuthology.orchestra.run.vm02.stdout: python3-autocommand noarch 2.2.2-8.el9 @epel 82 k 2026-03-06T13:50:43.122 INFO:teuthology.orchestra.run.vm02.stdout: python3-babel noarch 2.9.1-2.el9 @appstream 27 M 2026-03-06T13:50:43.122 INFO:teuthology.orchestra.run.vm02.stdout: python3-backports-tarfile noarch 1.2.0-1.el9 @epel 254 k 2026-03-06T13:50:43.122 INFO:teuthology.orchestra.run.vm02.stdout: python3-bcrypt x86_64 3.2.2-1.el9 @epel 87 k 2026-03-06T13:50:43.122 INFO:teuthology.orchestra.run.vm02.stdout: python3-cachetools noarch 4.2.4-1.el9 @epel 93 k 2026-03-06T13:50:43.122 INFO:teuthology.orchestra.run.vm02.stdout: python3-ceph-common x86_64 2:19.2.3-47.gc24117fd552.el9.clyso @ceph 702 k 2026-03-06T13:50:43.122 INFO:teuthology.orchestra.run.vm02.stdout: python3-certifi noarch 2023.05.07-4.el9 @epel 6.3 k 2026-03-06T13:50:43.122 INFO:teuthology.orchestra.run.vm02.stdout: python3-cffi x86_64 1.14.5-5.el9 @baseos 1.0 M 2026-03-06T13:50:43.122 INFO:teuthology.orchestra.run.vm02.stdout: python3-chardet noarch 4.0.0-5.el9 @anaconda 1.4 M 2026-03-06T13:50:43.122 INFO:teuthology.orchestra.run.vm02.stdout: python3-cheroot noarch 10.0.1-4.el9 @epel 682 k 2026-03-06T13:50:43.122 INFO:teuthology.orchestra.run.vm02.stdout: python3-cherrypy noarch 18.6.1-2.el9 @epel 1.1 M 2026-03-06T13:50:43.122 INFO:teuthology.orchestra.run.vm02.stdout: python3-cryptography x86_64 36.0.1-5.el9 @baseos 4.5 M 2026-03-06T13:50:43.122 INFO:teuthology.orchestra.run.vm02.stdout: python3-devel x86_64 3.9.25-3.el9 @appstream 765 k 2026-03-06T13:50:43.122 INFO:teuthology.orchestra.run.vm02.stdout: python3-google-auth noarch 1:2.45.0-1.el9 @epel 1.4 M 2026-03-06T13:50:43.122 INFO:teuthology.orchestra.run.vm02.stdout: python3-grpcio x86_64 1.46.7-10.el9 @epel 6.7 M 2026-03-06T13:50:43.122 INFO:teuthology.orchestra.run.vm02.stdout: python3-grpcio-tools x86_64 1.46.7-10.el9 @epel 418 k 2026-03-06T13:50:43.122 INFO:teuthology.orchestra.run.vm02.stdout: python3-idna noarch 2.10-7.el9.1 @anaconda 513 k 2026-03-06T13:50:43.122 INFO:teuthology.orchestra.run.vm02.stdout: python3-influxdb noarch 5.3.1-1.el9 @epel 747 k 2026-03-06T13:50:43.122 INFO:teuthology.orchestra.run.vm02.stdout: python3-isodate noarch 0.6.1-3.el9 @epel 203 k 2026-03-06T13:50:43.122 INFO:teuthology.orchestra.run.vm02.stdout: python3-jaraco noarch 8.2.1-3.el9 @epel 3.7 k 2026-03-06T13:50:43.122 INFO:teuthology.orchestra.run.vm02.stdout: python3-jaraco-classes noarch 3.2.1-5.el9 @epel 24 k 2026-03-06T13:50:43.122 INFO:teuthology.orchestra.run.vm02.stdout: python3-jaraco-collections noarch 3.0.0-8.el9 @epel 55 k 2026-03-06T13:50:43.122 INFO:teuthology.orchestra.run.vm02.stdout: python3-jaraco-context noarch 6.0.1-3.el9 @epel 31 k 2026-03-06T13:50:43.122 INFO:teuthology.orchestra.run.vm02.stdout: python3-jaraco-functools noarch 3.5.0-2.el9 @epel 33 k 2026-03-06T13:50:43.122 INFO:teuthology.orchestra.run.vm02.stdout: python3-jaraco-text noarch 4.0.0-2.el9 @epel 51 k 2026-03-06T13:50:43.122 INFO:teuthology.orchestra.run.vm02.stdout: python3-jinja2 noarch 2.11.3-8.el9 @appstream 1.1 M 2026-03-06T13:50:43.122 INFO:teuthology.orchestra.run.vm02.stdout: python3-jsonpatch noarch 1.21-16.el9 @koji-override-0 55 k 2026-03-06T13:50:43.122 INFO:teuthology.orchestra.run.vm02.stdout: python3-jsonpointer noarch 2.0-4.el9 @koji-override-0 34 k 2026-03-06T13:50:43.123 INFO:teuthology.orchestra.run.vm02.stdout: python3-kubernetes noarch 1:26.1.0-3.el9 @epel 21 M 2026-03-06T13:50:43.123 INFO:teuthology.orchestra.run.vm02.stdout: python3-libstoragemgmt x86_64 1.10.1-1.el9 @appstream 832 k 2026-03-06T13:50:43.123 INFO:teuthology.orchestra.run.vm02.stdout: python3-logutils noarch 0.3.5-21.el9 @epel 126 k 2026-03-06T13:50:43.123 INFO:teuthology.orchestra.run.vm02.stdout: python3-lxml x86_64 4.6.5-3.el9 @appstream 4.2 M 2026-03-06T13:50:43.123 INFO:teuthology.orchestra.run.vm02.stdout: python3-mako noarch 1.1.4-6.el9 @appstream 534 k 2026-03-06T13:50:43.123 INFO:teuthology.orchestra.run.vm02.stdout: python3-markupsafe x86_64 1.1.1-12.el9 @appstream 60 k 2026-03-06T13:50:43.123 INFO:teuthology.orchestra.run.vm02.stdout: python3-more-itertools noarch 8.12.0-2.el9 @epel 378 k 2026-03-06T13:50:43.123 INFO:teuthology.orchestra.run.vm02.stdout: python3-msgpack x86_64 1.0.3-2.el9 @epel 264 k 2026-03-06T13:50:43.123 INFO:teuthology.orchestra.run.vm02.stdout: python3-natsort noarch 7.1.1-5.el9 @epel 215 k 2026-03-06T13:50:43.123 INFO:teuthology.orchestra.run.vm02.stdout: python3-numpy x86_64 1:1.23.5-2.el9 @appstream 30 M 2026-03-06T13:50:43.123 INFO:teuthology.orchestra.run.vm02.stdout: python3-numpy-f2py x86_64 1:1.23.5-2.el9 @appstream 1.7 M 2026-03-06T13:50:43.123 INFO:teuthology.orchestra.run.vm02.stdout: python3-oauthlib noarch 3.1.1-5.el9 @koji-override-0 888 k 2026-03-06T13:50:43.123 INFO:teuthology.orchestra.run.vm02.stdout: python3-packaging noarch 20.9-5.el9 @appstream 248 k 2026-03-06T13:50:43.123 INFO:teuthology.orchestra.run.vm02.stdout: python3-pecan noarch 1.4.2-3.el9 @epel 1.3 M 2026-03-06T13:50:43.123 INFO:teuthology.orchestra.run.vm02.stdout: python3-ply noarch 3.11-14.el9 @baseos 430 k 2026-03-06T13:50:43.123 INFO:teuthology.orchestra.run.vm02.stdout: python3-portend noarch 3.1.0-2.el9 @epel 20 k 2026-03-06T13:50:43.123 INFO:teuthology.orchestra.run.vm02.stdout: python3-prettytable noarch 0.7.2-27.el9 @koji-override-0 166 k 2026-03-06T13:50:43.123 INFO:teuthology.orchestra.run.vm02.stdout: python3-protobuf noarch 3.14.0-17.el9 @appstream 1.4 M 2026-03-06T13:50:43.123 INFO:teuthology.orchestra.run.vm02.stdout: python3-pyOpenSSL noarch 21.0.0-1.el9 @epel 389 k 2026-03-06T13:50:43.123 INFO:teuthology.orchestra.run.vm02.stdout: python3-pyasn1 noarch 0.4.8-7.el9 @appstream 622 k 2026-03-06T13:50:43.123 INFO:teuthology.orchestra.run.vm02.stdout: python3-pyasn1-modules noarch 0.4.8-7.el9 @appstream 1.0 M 2026-03-06T13:50:43.123 INFO:teuthology.orchestra.run.vm02.stdout: python3-pycparser noarch 2.20-6.el9 @baseos 745 k 2026-03-06T13:50:43.123 INFO:teuthology.orchestra.run.vm02.stdout: python3-pysocks noarch 1.7.1-12.el9 @anaconda 88 k 2026-03-06T13:50:43.123 INFO:teuthology.orchestra.run.vm02.stdout: python3-pytz noarch 2021.1-5.el9 @koji-override-0 176 k 2026-03-06T13:50:43.123 INFO:teuthology.orchestra.run.vm02.stdout: python3-repoze-lru noarch 0.7-16.el9 @epel 83 k 2026-03-06T13:50:43.123 INFO:teuthology.orchestra.run.vm02.stdout: python3-requests noarch 2.25.1-10.el9 @baseos 405 k 2026-03-06T13:50:43.123 INFO:teuthology.orchestra.run.vm02.stdout: python3-requests-oauthlib noarch 1.3.0-12.el9 @appstream 119 k 2026-03-06T13:50:43.123 INFO:teuthology.orchestra.run.vm02.stdout: python3-routes noarch 2.5.1-5.el9 @epel 459 k 2026-03-06T13:50:43.123 INFO:teuthology.orchestra.run.vm02.stdout: python3-rsa noarch 4.9-2.el9 @epel 202 k 2026-03-06T13:50:43.123 INFO:teuthology.orchestra.run.vm02.stdout: python3-saml noarch 1.16.0-1.el9 @epel 730 k 2026-03-06T13:50:43.123 INFO:teuthology.orchestra.run.vm02.stdout: python3-scipy x86_64 1.9.3-2.el9 @appstream 76 M 2026-03-06T13:50:43.123 INFO:teuthology.orchestra.run.vm02.stdout: python3-tempora noarch 5.0.0-2.el9 @epel 96 k 2026-03-06T13:50:43.123 INFO:teuthology.orchestra.run.vm02.stdout: python3-toml noarch 0.10.2-6.el9 @appstream 99 k 2026-03-06T13:50:43.123 INFO:teuthology.orchestra.run.vm02.stdout: python3-typing-extensions noarch 4.15.0-1.el9 @epel 447 k 2026-03-06T13:50:43.123 INFO:teuthology.orchestra.run.vm02.stdout: python3-urllib3 noarch 1.26.5-7.el9 @baseos 746 k 2026-03-06T13:50:43.123 INFO:teuthology.orchestra.run.vm02.stdout: python3-webob noarch 1.8.8-2.el9 @epel 1.2 M 2026-03-06T13:50:43.123 INFO:teuthology.orchestra.run.vm02.stdout: python3-websocket-client noarch 1.2.3-2.el9 @epel 319 k 2026-03-06T13:50:43.123 INFO:teuthology.orchestra.run.vm02.stdout: python3-werkzeug noarch 2.0.3-3.el9.1 @epel 1.9 M 2026-03-06T13:50:43.123 INFO:teuthology.orchestra.run.vm02.stdout: python3-xmlsec x86_64 1.3.13-1.el9 @epel 158 k 2026-03-06T13:50:43.123 INFO:teuthology.orchestra.run.vm02.stdout: python3-zc-lockfile noarch 2.0-10.el9 @epel 35 k 2026-03-06T13:50:43.123 INFO:teuthology.orchestra.run.vm02.stdout: qatlib x86_64 25.08.0-2.el9 @appstream 639 k 2026-03-06T13:50:43.123 INFO:teuthology.orchestra.run.vm02.stdout: qatlib-service x86_64 25.08.0-2.el9 @appstream 69 k 2026-03-06T13:50:43.123 INFO:teuthology.orchestra.run.vm02.stdout: qatzip-libs x86_64 1.3.1-1.el9 @appstream 148 k 2026-03-06T13:50:43.123 INFO:teuthology.orchestra.run.vm02.stdout: smartmontools x86_64 1:7.2-10.el9 @baseos 1.9 M 2026-03-06T13:50:43.123 INFO:teuthology.orchestra.run.vm02.stdout: xmlsec1 x86_64 1.2.29-13.el9 @appstream 596 k 2026-03-06T13:50:43.123 INFO:teuthology.orchestra.run.vm02.stdout: xmlsec1-openssl x86_64 1.2.29-13.el9 @appstream 281 k 2026-03-06T13:50:43.123 INFO:teuthology.orchestra.run.vm02.stdout: 2026-03-06T13:50:43.123 INFO:teuthology.orchestra.run.vm02.stdout:Transaction Summary 2026-03-06T13:50:43.123 INFO:teuthology.orchestra.run.vm02.stdout:=================================================================================================== 2026-03-06T13:50:43.123 INFO:teuthology.orchestra.run.vm02.stdout:Remove 113 Packages 2026-03-06T13:50:43.123 INFO:teuthology.orchestra.run.vm02.stdout: 2026-03-06T13:50:43.123 INFO:teuthology.orchestra.run.vm02.stdout:Freed space: 623 M 2026-03-06T13:50:43.123 INFO:teuthology.orchestra.run.vm02.stdout:Running transaction check 2026-03-06T13:50:43.125 INFO:teuthology.orchestra.run.vm04.stdout:Dependencies resolved. 2026-03-06T13:50:43.128 INFO:teuthology.orchestra.run.vm06.stdout:Dependencies resolved. 2026-03-06T13:50:43.131 INFO:teuthology.orchestra.run.vm04.stdout:=================================================================================================== 2026-03-06T13:50:43.131 INFO:teuthology.orchestra.run.vm04.stdout: Package Arch Version Repository Size 2026-03-06T13:50:43.131 INFO:teuthology.orchestra.run.vm04.stdout:=================================================================================================== 2026-03-06T13:50:43.131 INFO:teuthology.orchestra.run.vm04.stdout:Removing: 2026-03-06T13:50:43.131 INFO:teuthology.orchestra.run.vm04.stdout: ceph-base x86_64 2:19.2.3-47.gc24117fd552.el9.clyso @ceph 23 M 2026-03-06T13:50:43.131 INFO:teuthology.orchestra.run.vm04.stdout:Removing dependent packages: 2026-03-06T13:50:43.131 INFO:teuthology.orchestra.run.vm04.stdout: ceph-immutable-object-cache x86_64 2:19.2.3-47.gc24117fd552.el9.clyso @ceph 431 k 2026-03-06T13:50:43.131 INFO:teuthology.orchestra.run.vm04.stdout: ceph-mgr x86_64 2:19.2.3-47.gc24117fd552.el9.clyso @ceph 3.4 M 2026-03-06T13:50:43.132 INFO:teuthology.orchestra.run.vm04.stdout: ceph-mgr-cephadm noarch 2:19.2.3-47.gc24117fd552.el9.clyso @ceph-noarch 803 k 2026-03-06T13:50:43.132 INFO:teuthology.orchestra.run.vm04.stdout: ceph-mgr-dashboard noarch 2:19.2.3-47.gc24117fd552.el9.clyso @ceph-noarch 88 M 2026-03-06T13:50:43.132 INFO:teuthology.orchestra.run.vm04.stdout: ceph-mgr-diskprediction-local noarch 2:19.2.3-47.gc24117fd552.el9.clyso @ceph-noarch 66 M 2026-03-06T13:50:43.132 INFO:teuthology.orchestra.run.vm04.stdout: ceph-mgr-rook noarch 2:19.2.3-47.gc24117fd552.el9.clyso @ceph-noarch 563 k 2026-03-06T13:50:43.132 INFO:teuthology.orchestra.run.vm04.stdout: ceph-osd x86_64 2:19.2.3-47.gc24117fd552.el9.clyso @ceph 59 M 2026-03-06T13:50:43.132 INFO:teuthology.orchestra.run.vm04.stdout: ceph-volume noarch 2:19.2.3-47.gc24117fd552.el9.clyso @ceph-noarch 1.4 M 2026-03-06T13:50:43.132 INFO:teuthology.orchestra.run.vm04.stdout: rbd-mirror x86_64 2:19.2.3-47.gc24117fd552.el9.clyso @ceph 13 M 2026-03-06T13:50:43.132 INFO:teuthology.orchestra.run.vm04.stdout:Removing unused dependencies: 2026-03-06T13:50:43.132 INFO:teuthology.orchestra.run.vm04.stdout: abseil-cpp x86_64 20211102.0-4.el9 @epel 1.9 M 2026-03-06T13:50:43.132 INFO:teuthology.orchestra.run.vm04.stdout: ceph-common x86_64 2:19.2.3-47.gc24117fd552.el9.clyso @ceph 85 M 2026-03-06T13:50:43.132 INFO:teuthology.orchestra.run.vm04.stdout: ceph-grafana-dashboards noarch 2:19.2.3-47.gc24117fd552.el9.clyso @ceph-noarch 626 k 2026-03-06T13:50:43.132 INFO:teuthology.orchestra.run.vm04.stdout: ceph-mgr-k8sevents noarch 2:19.2.3-47.gc24117fd552.el9.clyso @ceph-noarch 60 k 2026-03-06T13:50:43.132 INFO:teuthology.orchestra.run.vm04.stdout: ceph-mgr-modules-core noarch 2:19.2.3-47.gc24117fd552.el9.clyso @ceph-noarch 1.5 M 2026-03-06T13:50:43.132 INFO:teuthology.orchestra.run.vm04.stdout: ceph-prometheus-alerts noarch 2:19.2.3-47.gc24117fd552.el9.clyso @ceph-noarch 51 k 2026-03-06T13:50:43.132 INFO:teuthology.orchestra.run.vm04.stdout: ceph-selinux x86_64 2:19.2.3-47.gc24117fd552.el9.clyso @ceph 138 k 2026-03-06T13:50:43.132 INFO:teuthology.orchestra.run.vm04.stdout: cryptsetup x86_64 2.8.1-3.el9 @baseos 770 k 2026-03-06T13:50:43.132 INFO:teuthology.orchestra.run.vm04.stdout: flexiblas x86_64 3.0.4-9.el9 @appstream 68 k 2026-03-06T13:50:43.132 INFO:teuthology.orchestra.run.vm04.stdout: flexiblas-netlib x86_64 3.0.4-9.el9 @appstream 11 M 2026-03-06T13:50:43.132 INFO:teuthology.orchestra.run.vm04.stdout: flexiblas-openblas-openmp x86_64 3.0.4-9.el9 @appstream 39 k 2026-03-06T13:50:43.132 INFO:teuthology.orchestra.run.vm04.stdout: gperftools-libs x86_64 2.9.1-3.el9 @epel 1.4 M 2026-03-06T13:50:43.132 INFO:teuthology.orchestra.run.vm04.stdout: grpc-data noarch 1.46.7-10.el9 @epel 13 k 2026-03-06T13:50:43.132 INFO:teuthology.orchestra.run.vm04.stdout: ledmon-libs x86_64 1.1.0-3.el9 @baseos 80 k 2026-03-06T13:50:43.132 INFO:teuthology.orchestra.run.vm04.stdout: libcephsqlite x86_64 2:19.2.3-47.gc24117fd552.el9.clyso @ceph 425 k 2026-03-06T13:50:43.132 INFO:teuthology.orchestra.run.vm04.stdout: libconfig x86_64 1.7.2-9.el9 @baseos 220 k 2026-03-06T13:50:43.132 INFO:teuthology.orchestra.run.vm04.stdout: libgfortran x86_64 11.5.0-14.el9 @baseos 2.8 M 2026-03-06T13:50:43.132 INFO:teuthology.orchestra.run.vm04.stdout: liboath x86_64 2.6.12-1.el9 @epel 94 k 2026-03-06T13:50:43.132 INFO:teuthology.orchestra.run.vm04.stdout: libquadmath x86_64 11.5.0-14.el9 @baseos 330 k 2026-03-06T13:50:43.132 INFO:teuthology.orchestra.run.vm04.stdout: libradosstriper1 x86_64 2:19.2.3-47.gc24117fd552.el9.clyso @ceph 1.6 M 2026-03-06T13:50:43.132 INFO:teuthology.orchestra.run.vm04.stdout: libstoragemgmt x86_64 1.10.1-1.el9 @appstream 685 k 2026-03-06T13:50:43.132 INFO:teuthology.orchestra.run.vm04.stdout: libunwind x86_64 1.6.2-1.el9 @epel 170 k 2026-03-06T13:50:43.132 INFO:teuthology.orchestra.run.vm04.stdout: libxslt x86_64 1.1.34-12.el9 @appstream 743 k 2026-03-06T13:50:43.132 INFO:teuthology.orchestra.run.vm04.stdout: openblas x86_64 0.3.29-1.el9 @appstream 112 k 2026-03-06T13:50:43.132 INFO:teuthology.orchestra.run.vm04.stdout: openblas-openmp x86_64 0.3.29-1.el9 @appstream 46 M 2026-03-06T13:50:43.132 INFO:teuthology.orchestra.run.vm04.stdout: pciutils x86_64 3.7.0-7.el9 @baseos 216 k 2026-03-06T13:50:43.132 INFO:teuthology.orchestra.run.vm04.stdout: protobuf x86_64 3.14.0-17.el9 @appstream 3.5 M 2026-03-06T13:50:43.132 INFO:teuthology.orchestra.run.vm04.stdout: protobuf-compiler x86_64 3.14.0-17.el9 @crb 2.9 M 2026-03-06T13:50:43.132 INFO:teuthology.orchestra.run.vm04.stdout: python3-asyncssh noarch 2.13.2-5.el9 @epel 3.9 M 2026-03-06T13:50:43.132 INFO:teuthology.orchestra.run.vm04.stdout: python3-autocommand noarch 2.2.2-8.el9 @epel 82 k 2026-03-06T13:50:43.133 INFO:teuthology.orchestra.run.vm04.stdout: python3-babel noarch 2.9.1-2.el9 @appstream 27 M 2026-03-06T13:50:43.133 INFO:teuthology.orchestra.run.vm04.stdout: python3-backports-tarfile noarch 1.2.0-1.el9 @epel 254 k 2026-03-06T13:50:43.133 INFO:teuthology.orchestra.run.vm04.stdout: python3-bcrypt x86_64 3.2.2-1.el9 @epel 87 k 2026-03-06T13:50:43.133 INFO:teuthology.orchestra.run.vm04.stdout: python3-cachetools noarch 4.2.4-1.el9 @epel 93 k 2026-03-06T13:50:43.133 INFO:teuthology.orchestra.run.vm04.stdout: python3-ceph-common x86_64 2:19.2.3-47.gc24117fd552.el9.clyso @ceph 702 k 2026-03-06T13:50:43.133 INFO:teuthology.orchestra.run.vm04.stdout: python3-certifi noarch 2023.05.07-4.el9 @epel 6.3 k 2026-03-06T13:50:43.133 INFO:teuthology.orchestra.run.vm04.stdout: python3-cffi x86_64 1.14.5-5.el9 @baseos 1.0 M 2026-03-06T13:50:43.133 INFO:teuthology.orchestra.run.vm04.stdout: python3-chardet noarch 4.0.0-5.el9 @anaconda 1.4 M 2026-03-06T13:50:43.133 INFO:teuthology.orchestra.run.vm04.stdout: python3-cheroot noarch 10.0.1-4.el9 @epel 682 k 2026-03-06T13:50:43.133 INFO:teuthology.orchestra.run.vm04.stdout: python3-cherrypy noarch 18.6.1-2.el9 @epel 1.1 M 2026-03-06T13:50:43.133 INFO:teuthology.orchestra.run.vm04.stdout: python3-cryptography x86_64 36.0.1-5.el9 @baseos 4.5 M 2026-03-06T13:50:43.133 INFO:teuthology.orchestra.run.vm04.stdout: python3-devel x86_64 3.9.25-3.el9 @appstream 765 k 2026-03-06T13:50:43.133 INFO:teuthology.orchestra.run.vm04.stdout: python3-google-auth noarch 1:2.45.0-1.el9 @epel 1.4 M 2026-03-06T13:50:43.133 INFO:teuthology.orchestra.run.vm04.stdout: python3-grpcio x86_64 1.46.7-10.el9 @epel 6.7 M 2026-03-06T13:50:43.133 INFO:teuthology.orchestra.run.vm04.stdout: python3-grpcio-tools x86_64 1.46.7-10.el9 @epel 418 k 2026-03-06T13:50:43.133 INFO:teuthology.orchestra.run.vm04.stdout: python3-idna noarch 2.10-7.el9.1 @anaconda 513 k 2026-03-06T13:50:43.133 INFO:teuthology.orchestra.run.vm04.stdout: python3-influxdb noarch 5.3.1-1.el9 @epel 747 k 2026-03-06T13:50:43.133 INFO:teuthology.orchestra.run.vm04.stdout: python3-isodate noarch 0.6.1-3.el9 @epel 203 k 2026-03-06T13:50:43.133 INFO:teuthology.orchestra.run.vm04.stdout: python3-jaraco noarch 8.2.1-3.el9 @epel 3.7 k 2026-03-06T13:50:43.133 INFO:teuthology.orchestra.run.vm04.stdout: python3-jaraco-classes noarch 3.2.1-5.el9 @epel 24 k 2026-03-06T13:50:43.133 INFO:teuthology.orchestra.run.vm04.stdout: python3-jaraco-collections noarch 3.0.0-8.el9 @epel 55 k 2026-03-06T13:50:43.133 INFO:teuthology.orchestra.run.vm04.stdout: python3-jaraco-context noarch 6.0.1-3.el9 @epel 31 k 2026-03-06T13:50:43.133 INFO:teuthology.orchestra.run.vm04.stdout: python3-jaraco-functools noarch 3.5.0-2.el9 @epel 33 k 2026-03-06T13:50:43.133 INFO:teuthology.orchestra.run.vm04.stdout: python3-jaraco-text noarch 4.0.0-2.el9 @epel 51 k 2026-03-06T13:50:43.133 INFO:teuthology.orchestra.run.vm04.stdout: python3-jinja2 noarch 2.11.3-8.el9 @appstream 1.1 M 2026-03-06T13:50:43.133 INFO:teuthology.orchestra.run.vm04.stdout: python3-jsonpatch noarch 1.21-16.el9 @koji-override-0 55 k 2026-03-06T13:50:43.133 INFO:teuthology.orchestra.run.vm04.stdout: python3-jsonpointer noarch 2.0-4.el9 @koji-override-0 34 k 2026-03-06T13:50:43.133 INFO:teuthology.orchestra.run.vm04.stdout: python3-kubernetes noarch 1:26.1.0-3.el9 @epel 21 M 2026-03-06T13:50:43.133 INFO:teuthology.orchestra.run.vm04.stdout: python3-libstoragemgmt x86_64 1.10.1-1.el9 @appstream 832 k 2026-03-06T13:50:43.133 INFO:teuthology.orchestra.run.vm04.stdout: python3-logutils noarch 0.3.5-21.el9 @epel 126 k 2026-03-06T13:50:43.133 INFO:teuthology.orchestra.run.vm04.stdout: python3-lxml x86_64 4.6.5-3.el9 @appstream 4.2 M 2026-03-06T13:50:43.133 INFO:teuthology.orchestra.run.vm04.stdout: python3-mako noarch 1.1.4-6.el9 @appstream 534 k 2026-03-06T13:50:43.133 INFO:teuthology.orchestra.run.vm04.stdout: python3-markupsafe x86_64 1.1.1-12.el9 @appstream 60 k 2026-03-06T13:50:43.133 INFO:teuthology.orchestra.run.vm04.stdout: python3-more-itertools noarch 8.12.0-2.el9 @epel 378 k 2026-03-06T13:50:43.133 INFO:teuthology.orchestra.run.vm04.stdout: python3-msgpack x86_64 1.0.3-2.el9 @epel 264 k 2026-03-06T13:50:43.133 INFO:teuthology.orchestra.run.vm04.stdout: python3-natsort noarch 7.1.1-5.el9 @epel 215 k 2026-03-06T13:50:43.133 INFO:teuthology.orchestra.run.vm04.stdout: python3-numpy x86_64 1:1.23.5-2.el9 @appstream 30 M 2026-03-06T13:50:43.134 INFO:teuthology.orchestra.run.vm04.stdout: python3-numpy-f2py x86_64 1:1.23.5-2.el9 @appstream 1.7 M 2026-03-06T13:50:43.134 INFO:teuthology.orchestra.run.vm04.stdout: python3-oauthlib noarch 3.1.1-5.el9 @koji-override-0 888 k 2026-03-06T13:50:43.134 INFO:teuthology.orchestra.run.vm04.stdout: python3-packaging noarch 20.9-5.el9 @appstream 248 k 2026-03-06T13:50:43.134 INFO:teuthology.orchestra.run.vm04.stdout: python3-pecan noarch 1.4.2-3.el9 @epel 1.3 M 2026-03-06T13:50:43.134 INFO:teuthology.orchestra.run.vm04.stdout: python3-ply noarch 3.11-14.el9 @baseos 430 k 2026-03-06T13:50:43.134 INFO:teuthology.orchestra.run.vm04.stdout: python3-portend noarch 3.1.0-2.el9 @epel 20 k 2026-03-06T13:50:43.134 INFO:teuthology.orchestra.run.vm04.stdout: python3-prettytable noarch 0.7.2-27.el9 @koji-override-0 166 k 2026-03-06T13:50:43.134 INFO:teuthology.orchestra.run.vm04.stdout: python3-protobuf noarch 3.14.0-17.el9 @appstream 1.4 M 2026-03-06T13:50:43.134 INFO:teuthology.orchestra.run.vm04.stdout: python3-pyOpenSSL noarch 21.0.0-1.el9 @epel 389 k 2026-03-06T13:50:43.134 INFO:teuthology.orchestra.run.vm04.stdout: python3-pyasn1 noarch 0.4.8-7.el9 @appstream 622 k 2026-03-06T13:50:43.134 INFO:teuthology.orchestra.run.vm04.stdout: python3-pyasn1-modules noarch 0.4.8-7.el9 @appstream 1.0 M 2026-03-06T13:50:43.134 INFO:teuthology.orchestra.run.vm04.stdout: python3-pycparser noarch 2.20-6.el9 @baseos 745 k 2026-03-06T13:50:43.134 INFO:teuthology.orchestra.run.vm04.stdout: python3-pysocks noarch 1.7.1-12.el9 @anaconda 88 k 2026-03-06T13:50:43.134 INFO:teuthology.orchestra.run.vm04.stdout: python3-pytz noarch 2021.1-5.el9 @koji-override-0 176 k 2026-03-06T13:50:43.134 INFO:teuthology.orchestra.run.vm04.stdout: python3-repoze-lru noarch 0.7-16.el9 @epel 83 k 2026-03-06T13:50:43.134 INFO:teuthology.orchestra.run.vm04.stdout: python3-requests noarch 2.25.1-10.el9 @baseos 405 k 2026-03-06T13:50:43.134 INFO:teuthology.orchestra.run.vm04.stdout: python3-requests-oauthlib noarch 1.3.0-12.el9 @appstream 119 k 2026-03-06T13:50:43.134 INFO:teuthology.orchestra.run.vm04.stdout: python3-routes noarch 2.5.1-5.el9 @epel 459 k 2026-03-06T13:50:43.134 INFO:teuthology.orchestra.run.vm04.stdout: python3-rsa noarch 4.9-2.el9 @epel 202 k 2026-03-06T13:50:43.134 INFO:teuthology.orchestra.run.vm04.stdout: python3-saml noarch 1.16.0-1.el9 @epel 730 k 2026-03-06T13:50:43.134 INFO:teuthology.orchestra.run.vm04.stdout: python3-scipy x86_64 1.9.3-2.el9 @appstream 76 M 2026-03-06T13:50:43.134 INFO:teuthology.orchestra.run.vm04.stdout: python3-tempora noarch 5.0.0-2.el9 @epel 96 k 2026-03-06T13:50:43.134 INFO:teuthology.orchestra.run.vm04.stdout: python3-toml noarch 0.10.2-6.el9 @appstream 99 k 2026-03-06T13:50:43.134 INFO:teuthology.orchestra.run.vm04.stdout: python3-typing-extensions noarch 4.15.0-1.el9 @epel 447 k 2026-03-06T13:50:43.134 INFO:teuthology.orchestra.run.vm04.stdout: python3-urllib3 noarch 1.26.5-7.el9 @baseos 746 k 2026-03-06T13:50:43.134 INFO:teuthology.orchestra.run.vm04.stdout: python3-webob noarch 1.8.8-2.el9 @epel 1.2 M 2026-03-06T13:50:43.134 INFO:teuthology.orchestra.run.vm04.stdout: python3-websocket-client noarch 1.2.3-2.el9 @epel 319 k 2026-03-06T13:50:43.134 INFO:teuthology.orchestra.run.vm04.stdout: python3-werkzeug noarch 2.0.3-3.el9.1 @epel 1.9 M 2026-03-06T13:50:43.134 INFO:teuthology.orchestra.run.vm04.stdout: python3-xmlsec x86_64 1.3.13-1.el9 @epel 158 k 2026-03-06T13:50:43.134 INFO:teuthology.orchestra.run.vm04.stdout: python3-zc-lockfile noarch 2.0-10.el9 @epel 35 k 2026-03-06T13:50:43.134 INFO:teuthology.orchestra.run.vm04.stdout: qatlib x86_64 25.08.0-2.el9 @appstream 639 k 2026-03-06T13:50:43.134 INFO:teuthology.orchestra.run.vm04.stdout: qatlib-service x86_64 25.08.0-2.el9 @appstream 69 k 2026-03-06T13:50:43.134 INFO:teuthology.orchestra.run.vm04.stdout: qatzip-libs x86_64 1.3.1-1.el9 @appstream 148 k 2026-03-06T13:50:43.134 INFO:teuthology.orchestra.run.vm04.stdout: smartmontools x86_64 1:7.2-10.el9 @baseos 1.9 M 2026-03-06T13:50:43.134 INFO:teuthology.orchestra.run.vm04.stdout: xmlsec1 x86_64 1.2.29-13.el9 @appstream 596 k 2026-03-06T13:50:43.134 INFO:teuthology.orchestra.run.vm04.stdout: xmlsec1-openssl x86_64 1.2.29-13.el9 @appstream 281 k 2026-03-06T13:50:43.134 INFO:teuthology.orchestra.run.vm04.stdout: 2026-03-06T13:50:43.134 INFO:teuthology.orchestra.run.vm04.stdout:Transaction Summary 2026-03-06T13:50:43.134 INFO:teuthology.orchestra.run.vm04.stdout:=================================================================================================== 2026-03-06T13:50:43.134 INFO:teuthology.orchestra.run.vm04.stdout:Remove 113 Packages 2026-03-06T13:50:43.135 INFO:teuthology.orchestra.run.vm04.stdout: 2026-03-06T13:50:43.135 INFO:teuthology.orchestra.run.vm04.stdout:Freed space: 623 M 2026-03-06T13:50:43.135 INFO:teuthology.orchestra.run.vm04.stdout:Running transaction check 2026-03-06T13:50:43.136 INFO:teuthology.orchestra.run.vm06.stdout:=================================================================================================== 2026-03-06T13:50:43.136 INFO:teuthology.orchestra.run.vm06.stdout: Package Arch Version Repository Size 2026-03-06T13:50:43.136 INFO:teuthology.orchestra.run.vm06.stdout:=================================================================================================== 2026-03-06T13:50:43.136 INFO:teuthology.orchestra.run.vm06.stdout:Removing: 2026-03-06T13:50:43.136 INFO:teuthology.orchestra.run.vm06.stdout: ceph-base x86_64 2:19.2.3-47.gc24117fd552.el9.clyso @ceph 23 M 2026-03-06T13:50:43.136 INFO:teuthology.orchestra.run.vm06.stdout:Removing dependent packages: 2026-03-06T13:50:43.136 INFO:teuthology.orchestra.run.vm06.stdout: ceph-immutable-object-cache x86_64 2:19.2.3-47.gc24117fd552.el9.clyso @ceph 431 k 2026-03-06T13:50:43.136 INFO:teuthology.orchestra.run.vm06.stdout: ceph-mgr x86_64 2:19.2.3-47.gc24117fd552.el9.clyso @ceph 3.4 M 2026-03-06T13:50:43.136 INFO:teuthology.orchestra.run.vm06.stdout: ceph-mgr-cephadm noarch 2:19.2.3-47.gc24117fd552.el9.clyso @ceph-noarch 803 k 2026-03-06T13:50:43.136 INFO:teuthology.orchestra.run.vm06.stdout: ceph-mgr-dashboard noarch 2:19.2.3-47.gc24117fd552.el9.clyso @ceph-noarch 88 M 2026-03-06T13:50:43.136 INFO:teuthology.orchestra.run.vm06.stdout: ceph-mgr-diskprediction-local noarch 2:19.2.3-47.gc24117fd552.el9.clyso @ceph-noarch 66 M 2026-03-06T13:50:43.136 INFO:teuthology.orchestra.run.vm06.stdout: ceph-mgr-rook noarch 2:19.2.3-47.gc24117fd552.el9.clyso @ceph-noarch 563 k 2026-03-06T13:50:43.136 INFO:teuthology.orchestra.run.vm06.stdout: ceph-osd x86_64 2:19.2.3-47.gc24117fd552.el9.clyso @ceph 59 M 2026-03-06T13:50:43.136 INFO:teuthology.orchestra.run.vm06.stdout: ceph-volume noarch 2:19.2.3-47.gc24117fd552.el9.clyso @ceph-noarch 1.4 M 2026-03-06T13:50:43.136 INFO:teuthology.orchestra.run.vm06.stdout: rbd-mirror x86_64 2:19.2.3-47.gc24117fd552.el9.clyso @ceph 13 M 2026-03-06T13:50:43.136 INFO:teuthology.orchestra.run.vm06.stdout:Removing unused dependencies: 2026-03-06T13:50:43.136 INFO:teuthology.orchestra.run.vm06.stdout: abseil-cpp x86_64 20211102.0-4.el9 @epel 1.9 M 2026-03-06T13:50:43.136 INFO:teuthology.orchestra.run.vm06.stdout: ceph-common x86_64 2:19.2.3-47.gc24117fd552.el9.clyso @ceph 85 M 2026-03-06T13:50:43.137 INFO:teuthology.orchestra.run.vm06.stdout: ceph-grafana-dashboards noarch 2:19.2.3-47.gc24117fd552.el9.clyso @ceph-noarch 626 k 2026-03-06T13:50:43.137 INFO:teuthology.orchestra.run.vm06.stdout: ceph-mgr-k8sevents noarch 2:19.2.3-47.gc24117fd552.el9.clyso @ceph-noarch 60 k 2026-03-06T13:50:43.137 INFO:teuthology.orchestra.run.vm06.stdout: ceph-mgr-modules-core noarch 2:19.2.3-47.gc24117fd552.el9.clyso @ceph-noarch 1.5 M 2026-03-06T13:50:43.137 INFO:teuthology.orchestra.run.vm06.stdout: ceph-prometheus-alerts noarch 2:19.2.3-47.gc24117fd552.el9.clyso @ceph-noarch 51 k 2026-03-06T13:50:43.137 INFO:teuthology.orchestra.run.vm06.stdout: ceph-selinux x86_64 2:19.2.3-47.gc24117fd552.el9.clyso @ceph 138 k 2026-03-06T13:50:43.137 INFO:teuthology.orchestra.run.vm06.stdout: cryptsetup x86_64 2.8.1-3.el9 @baseos 770 k 2026-03-06T13:50:43.137 INFO:teuthology.orchestra.run.vm06.stdout: flexiblas x86_64 3.0.4-9.el9 @appstream 68 k 2026-03-06T13:50:43.137 INFO:teuthology.orchestra.run.vm06.stdout: flexiblas-netlib x86_64 3.0.4-9.el9 @appstream 11 M 2026-03-06T13:50:43.137 INFO:teuthology.orchestra.run.vm06.stdout: flexiblas-openblas-openmp x86_64 3.0.4-9.el9 @appstream 39 k 2026-03-06T13:50:43.137 INFO:teuthology.orchestra.run.vm06.stdout: gperftools-libs x86_64 2.9.1-3.el9 @epel 1.4 M 2026-03-06T13:50:43.137 INFO:teuthology.orchestra.run.vm06.stdout: grpc-data noarch 1.46.7-10.el9 @epel 13 k 2026-03-06T13:50:43.137 INFO:teuthology.orchestra.run.vm06.stdout: ledmon-libs x86_64 1.1.0-3.el9 @baseos 80 k 2026-03-06T13:50:43.137 INFO:teuthology.orchestra.run.vm06.stdout: libcephsqlite x86_64 2:19.2.3-47.gc24117fd552.el9.clyso @ceph 425 k 2026-03-06T13:50:43.137 INFO:teuthology.orchestra.run.vm06.stdout: libconfig x86_64 1.7.2-9.el9 @baseos 220 k 2026-03-06T13:50:43.137 INFO:teuthology.orchestra.run.vm06.stdout: libgfortran x86_64 11.5.0-14.el9 @baseos 2.8 M 2026-03-06T13:50:43.137 INFO:teuthology.orchestra.run.vm06.stdout: liboath x86_64 2.6.12-1.el9 @epel 94 k 2026-03-06T13:50:43.137 INFO:teuthology.orchestra.run.vm06.stdout: libquadmath x86_64 11.5.0-14.el9 @baseos 330 k 2026-03-06T13:50:43.137 INFO:teuthology.orchestra.run.vm06.stdout: libradosstriper1 x86_64 2:19.2.3-47.gc24117fd552.el9.clyso @ceph 1.6 M 2026-03-06T13:50:43.137 INFO:teuthology.orchestra.run.vm06.stdout: libstoragemgmt x86_64 1.10.1-1.el9 @appstream 685 k 2026-03-06T13:50:43.137 INFO:teuthology.orchestra.run.vm06.stdout: libunwind x86_64 1.6.2-1.el9 @epel 170 k 2026-03-06T13:50:43.137 INFO:teuthology.orchestra.run.vm06.stdout: libxslt x86_64 1.1.34-12.el9 @appstream 743 k 2026-03-06T13:50:43.137 INFO:teuthology.orchestra.run.vm06.stdout: openblas x86_64 0.3.29-1.el9 @appstream 112 k 2026-03-06T13:50:43.137 INFO:teuthology.orchestra.run.vm06.stdout: openblas-openmp x86_64 0.3.29-1.el9 @appstream 46 M 2026-03-06T13:50:43.137 INFO:teuthology.orchestra.run.vm06.stdout: pciutils x86_64 3.7.0-7.el9 @baseos 216 k 2026-03-06T13:50:43.137 INFO:teuthology.orchestra.run.vm06.stdout: protobuf x86_64 3.14.0-17.el9 @appstream 3.5 M 2026-03-06T13:50:43.137 INFO:teuthology.orchestra.run.vm06.stdout: protobuf-compiler x86_64 3.14.0-17.el9 @crb 2.9 M 2026-03-06T13:50:43.137 INFO:teuthology.orchestra.run.vm06.stdout: python3-asyncssh noarch 2.13.2-5.el9 @epel 3.9 M 2026-03-06T13:50:43.137 INFO:teuthology.orchestra.run.vm06.stdout: python3-autocommand noarch 2.2.2-8.el9 @epel 82 k 2026-03-06T13:50:43.137 INFO:teuthology.orchestra.run.vm06.stdout: python3-babel noarch 2.9.1-2.el9 @appstream 27 M 2026-03-06T13:50:43.137 INFO:teuthology.orchestra.run.vm06.stdout: python3-backports-tarfile noarch 1.2.0-1.el9 @epel 254 k 2026-03-06T13:50:43.137 INFO:teuthology.orchestra.run.vm06.stdout: python3-bcrypt x86_64 3.2.2-1.el9 @epel 87 k 2026-03-06T13:50:43.137 INFO:teuthology.orchestra.run.vm06.stdout: python3-cachetools noarch 4.2.4-1.el9 @epel 93 k 2026-03-06T13:50:43.137 INFO:teuthology.orchestra.run.vm06.stdout: python3-ceph-common x86_64 2:19.2.3-47.gc24117fd552.el9.clyso @ceph 702 k 2026-03-06T13:50:43.137 INFO:teuthology.orchestra.run.vm06.stdout: python3-certifi noarch 2023.05.07-4.el9 @epel 6.3 k 2026-03-06T13:50:43.137 INFO:teuthology.orchestra.run.vm06.stdout: python3-cffi x86_64 1.14.5-5.el9 @baseos 1.0 M 2026-03-06T13:50:43.137 INFO:teuthology.orchestra.run.vm06.stdout: python3-chardet noarch 4.0.0-5.el9 @anaconda 1.4 M 2026-03-06T13:50:43.137 INFO:teuthology.orchestra.run.vm06.stdout: python3-cheroot noarch 10.0.1-4.el9 @epel 682 k 2026-03-06T13:50:43.137 INFO:teuthology.orchestra.run.vm06.stdout: python3-cherrypy noarch 18.6.1-2.el9 @epel 1.1 M 2026-03-06T13:50:43.137 INFO:teuthology.orchestra.run.vm06.stdout: python3-cryptography x86_64 36.0.1-5.el9 @baseos 4.5 M 2026-03-06T13:50:43.138 INFO:teuthology.orchestra.run.vm06.stdout: python3-devel x86_64 3.9.25-3.el9 @appstream 765 k 2026-03-06T13:50:43.138 INFO:teuthology.orchestra.run.vm06.stdout: python3-google-auth noarch 1:2.45.0-1.el9 @epel 1.4 M 2026-03-06T13:50:43.138 INFO:teuthology.orchestra.run.vm06.stdout: python3-grpcio x86_64 1.46.7-10.el9 @epel 6.7 M 2026-03-06T13:50:43.138 INFO:teuthology.orchestra.run.vm06.stdout: python3-grpcio-tools x86_64 1.46.7-10.el9 @epel 418 k 2026-03-06T13:50:43.138 INFO:teuthology.orchestra.run.vm06.stdout: python3-idna noarch 2.10-7.el9.1 @anaconda 513 k 2026-03-06T13:50:43.138 INFO:teuthology.orchestra.run.vm06.stdout: python3-influxdb noarch 5.3.1-1.el9 @epel 747 k 2026-03-06T13:50:43.138 INFO:teuthology.orchestra.run.vm06.stdout: python3-isodate noarch 0.6.1-3.el9 @epel 203 k 2026-03-06T13:50:43.138 INFO:teuthology.orchestra.run.vm06.stdout: python3-jaraco noarch 8.2.1-3.el9 @epel 3.7 k 2026-03-06T13:50:43.138 INFO:teuthology.orchestra.run.vm06.stdout: python3-jaraco-classes noarch 3.2.1-5.el9 @epel 24 k 2026-03-06T13:50:43.138 INFO:teuthology.orchestra.run.vm06.stdout: python3-jaraco-collections noarch 3.0.0-8.el9 @epel 55 k 2026-03-06T13:50:43.138 INFO:teuthology.orchestra.run.vm06.stdout: python3-jaraco-context noarch 6.0.1-3.el9 @epel 31 k 2026-03-06T13:50:43.138 INFO:teuthology.orchestra.run.vm06.stdout: python3-jaraco-functools noarch 3.5.0-2.el9 @epel 33 k 2026-03-06T13:50:43.138 INFO:teuthology.orchestra.run.vm06.stdout: python3-jaraco-text noarch 4.0.0-2.el9 @epel 51 k 2026-03-06T13:50:43.138 INFO:teuthology.orchestra.run.vm06.stdout: python3-jinja2 noarch 2.11.3-8.el9 @appstream 1.1 M 2026-03-06T13:50:43.138 INFO:teuthology.orchestra.run.vm06.stdout: python3-jsonpatch noarch 1.21-16.el9 @koji-override-0 55 k 2026-03-06T13:50:43.138 INFO:teuthology.orchestra.run.vm06.stdout: python3-jsonpointer noarch 2.0-4.el9 @koji-override-0 34 k 2026-03-06T13:50:43.138 INFO:teuthology.orchestra.run.vm06.stdout: python3-kubernetes noarch 1:26.1.0-3.el9 @epel 21 M 2026-03-06T13:50:43.138 INFO:teuthology.orchestra.run.vm06.stdout: python3-libstoragemgmt x86_64 1.10.1-1.el9 @appstream 832 k 2026-03-06T13:50:43.138 INFO:teuthology.orchestra.run.vm06.stdout: python3-logutils noarch 0.3.5-21.el9 @epel 126 k 2026-03-06T13:50:43.138 INFO:teuthology.orchestra.run.vm06.stdout: python3-lxml x86_64 4.6.5-3.el9 @appstream 4.2 M 2026-03-06T13:50:43.138 INFO:teuthology.orchestra.run.vm06.stdout: python3-mako noarch 1.1.4-6.el9 @appstream 534 k 2026-03-06T13:50:43.138 INFO:teuthology.orchestra.run.vm06.stdout: python3-markupsafe x86_64 1.1.1-12.el9 @appstream 60 k 2026-03-06T13:50:43.138 INFO:teuthology.orchestra.run.vm06.stdout: python3-more-itertools noarch 8.12.0-2.el9 @epel 378 k 2026-03-06T13:50:43.138 INFO:teuthology.orchestra.run.vm06.stdout: python3-msgpack x86_64 1.0.3-2.el9 @epel 264 k 2026-03-06T13:50:43.138 INFO:teuthology.orchestra.run.vm06.stdout: python3-natsort noarch 7.1.1-5.el9 @epel 215 k 2026-03-06T13:50:43.138 INFO:teuthology.orchestra.run.vm06.stdout: python3-numpy x86_64 1:1.23.5-2.el9 @appstream 30 M 2026-03-06T13:50:43.138 INFO:teuthology.orchestra.run.vm06.stdout: python3-numpy-f2py x86_64 1:1.23.5-2.el9 @appstream 1.7 M 2026-03-06T13:50:43.138 INFO:teuthology.orchestra.run.vm06.stdout: python3-oauthlib noarch 3.1.1-5.el9 @koji-override-0 888 k 2026-03-06T13:50:43.138 INFO:teuthology.orchestra.run.vm06.stdout: python3-packaging noarch 20.9-5.el9 @appstream 248 k 2026-03-06T13:50:43.138 INFO:teuthology.orchestra.run.vm06.stdout: python3-pecan noarch 1.4.2-3.el9 @epel 1.3 M 2026-03-06T13:50:43.138 INFO:teuthology.orchestra.run.vm06.stdout: python3-ply noarch 3.11-14.el9 @baseos 430 k 2026-03-06T13:50:43.138 INFO:teuthology.orchestra.run.vm06.stdout: python3-portend noarch 3.1.0-2.el9 @epel 20 k 2026-03-06T13:50:43.138 INFO:teuthology.orchestra.run.vm06.stdout: python3-prettytable noarch 0.7.2-27.el9 @koji-override-0 166 k 2026-03-06T13:50:43.138 INFO:teuthology.orchestra.run.vm06.stdout: python3-protobuf noarch 3.14.0-17.el9 @appstream 1.4 M 2026-03-06T13:50:43.138 INFO:teuthology.orchestra.run.vm06.stdout: python3-pyOpenSSL noarch 21.0.0-1.el9 @epel 389 k 2026-03-06T13:50:43.138 INFO:teuthology.orchestra.run.vm06.stdout: python3-pyasn1 noarch 0.4.8-7.el9 @appstream 622 k 2026-03-06T13:50:43.138 INFO:teuthology.orchestra.run.vm06.stdout: python3-pyasn1-modules noarch 0.4.8-7.el9 @appstream 1.0 M 2026-03-06T13:50:43.138 INFO:teuthology.orchestra.run.vm06.stdout: python3-pycparser noarch 2.20-6.el9 @baseos 745 k 2026-03-06T13:50:43.138 INFO:teuthology.orchestra.run.vm06.stdout: python3-pysocks noarch 1.7.1-12.el9 @anaconda 88 k 2026-03-06T13:50:43.138 INFO:teuthology.orchestra.run.vm06.stdout: python3-pytz noarch 2021.1-5.el9 @koji-override-0 176 k 2026-03-06T13:50:43.138 INFO:teuthology.orchestra.run.vm06.stdout: python3-repoze-lru noarch 0.7-16.el9 @epel 83 k 2026-03-06T13:50:43.138 INFO:teuthology.orchestra.run.vm06.stdout: python3-requests noarch 2.25.1-10.el9 @baseos 405 k 2026-03-06T13:50:43.138 INFO:teuthology.orchestra.run.vm06.stdout: python3-requests-oauthlib noarch 1.3.0-12.el9 @appstream 119 k 2026-03-06T13:50:43.138 INFO:teuthology.orchestra.run.vm06.stdout: python3-routes noarch 2.5.1-5.el9 @epel 459 k 2026-03-06T13:50:43.139 INFO:teuthology.orchestra.run.vm06.stdout: python3-rsa noarch 4.9-2.el9 @epel 202 k 2026-03-06T13:50:43.139 INFO:teuthology.orchestra.run.vm06.stdout: python3-saml noarch 1.16.0-1.el9 @epel 730 k 2026-03-06T13:50:43.139 INFO:teuthology.orchestra.run.vm06.stdout: python3-scipy x86_64 1.9.3-2.el9 @appstream 76 M 2026-03-06T13:50:43.139 INFO:teuthology.orchestra.run.vm06.stdout: python3-tempora noarch 5.0.0-2.el9 @epel 96 k 2026-03-06T13:50:43.139 INFO:teuthology.orchestra.run.vm06.stdout: python3-toml noarch 0.10.2-6.el9 @appstream 99 k 2026-03-06T13:50:43.139 INFO:teuthology.orchestra.run.vm06.stdout: python3-typing-extensions noarch 4.15.0-1.el9 @epel 447 k 2026-03-06T13:50:43.139 INFO:teuthology.orchestra.run.vm06.stdout: python3-urllib3 noarch 1.26.5-7.el9 @baseos 746 k 2026-03-06T13:50:43.139 INFO:teuthology.orchestra.run.vm06.stdout: python3-webob noarch 1.8.8-2.el9 @epel 1.2 M 2026-03-06T13:50:43.139 INFO:teuthology.orchestra.run.vm06.stdout: python3-websocket-client noarch 1.2.3-2.el9 @epel 319 k 2026-03-06T13:50:43.139 INFO:teuthology.orchestra.run.vm06.stdout: python3-werkzeug noarch 2.0.3-3.el9.1 @epel 1.9 M 2026-03-06T13:50:43.139 INFO:teuthology.orchestra.run.vm06.stdout: python3-xmlsec x86_64 1.3.13-1.el9 @epel 158 k 2026-03-06T13:50:43.139 INFO:teuthology.orchestra.run.vm06.stdout: python3-zc-lockfile noarch 2.0-10.el9 @epel 35 k 2026-03-06T13:50:43.139 INFO:teuthology.orchestra.run.vm06.stdout: qatlib x86_64 25.08.0-2.el9 @appstream 639 k 2026-03-06T13:50:43.139 INFO:teuthology.orchestra.run.vm06.stdout: qatlib-service x86_64 25.08.0-2.el9 @appstream 69 k 2026-03-06T13:50:43.139 INFO:teuthology.orchestra.run.vm06.stdout: qatzip-libs x86_64 1.3.1-1.el9 @appstream 148 k 2026-03-06T13:50:43.139 INFO:teuthology.orchestra.run.vm06.stdout: smartmontools x86_64 1:7.2-10.el9 @baseos 1.9 M 2026-03-06T13:50:43.139 INFO:teuthology.orchestra.run.vm06.stdout: xmlsec1 x86_64 1.2.29-13.el9 @appstream 596 k 2026-03-06T13:50:43.139 INFO:teuthology.orchestra.run.vm06.stdout: xmlsec1-openssl x86_64 1.2.29-13.el9 @appstream 281 k 2026-03-06T13:50:43.139 INFO:teuthology.orchestra.run.vm06.stdout: 2026-03-06T13:50:43.139 INFO:teuthology.orchestra.run.vm06.stdout:Transaction Summary 2026-03-06T13:50:43.139 INFO:teuthology.orchestra.run.vm06.stdout:=================================================================================================== 2026-03-06T13:50:43.139 INFO:teuthology.orchestra.run.vm06.stdout:Remove 113 Packages 2026-03-06T13:50:43.139 INFO:teuthology.orchestra.run.vm06.stdout: 2026-03-06T13:50:43.139 INFO:teuthology.orchestra.run.vm06.stdout:Freed space: 623 M 2026-03-06T13:50:43.139 INFO:teuthology.orchestra.run.vm06.stdout:Running transaction check 2026-03-06T13:50:43.150 INFO:teuthology.orchestra.run.vm02.stdout:Transaction check succeeded. 2026-03-06T13:50:43.150 INFO:teuthology.orchestra.run.vm02.stdout:Running transaction test 2026-03-06T13:50:43.160 INFO:teuthology.orchestra.run.vm04.stdout:Transaction check succeeded. 2026-03-06T13:50:43.160 INFO:teuthology.orchestra.run.vm04.stdout:Running transaction test 2026-03-06T13:50:43.166 INFO:teuthology.orchestra.run.vm06.stdout:Transaction check succeeded. 2026-03-06T13:50:43.167 INFO:teuthology.orchestra.run.vm06.stdout:Running transaction test 2026-03-06T13:50:43.259 INFO:teuthology.orchestra.run.vm02.stdout:Transaction test succeeded. 2026-03-06T13:50:43.259 INFO:teuthology.orchestra.run.vm02.stdout:Running transaction 2026-03-06T13:50:43.264 INFO:teuthology.orchestra.run.vm04.stdout:Transaction test succeeded. 2026-03-06T13:50:43.265 INFO:teuthology.orchestra.run.vm04.stdout:Running transaction 2026-03-06T13:50:43.270 INFO:teuthology.orchestra.run.vm06.stdout:Transaction test succeeded. 2026-03-06T13:50:43.271 INFO:teuthology.orchestra.run.vm06.stdout:Running transaction 2026-03-06T13:50:43.405 INFO:teuthology.orchestra.run.vm02.stdout: Preparing : 1/1 2026-03-06T13:50:43.406 INFO:teuthology.orchestra.run.vm02.stdout: Erasing : ceph-mgr-rook-2:19.2.3-47.gc24117fd552.el9.clyso 1/113 2026-03-06T13:50:43.408 INFO:teuthology.orchestra.run.vm04.stdout: Preparing : 1/1 2026-03-06T13:50:43.408 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : ceph-mgr-rook-2:19.2.3-47.gc24117fd552.el9.clyso 1/113 2026-03-06T13:50:43.413 INFO:teuthology.orchestra.run.vm02.stdout: Running scriptlet: ceph-mgr-rook-2:19.2.3-47.gc24117fd552.el9.clyso 1/113 2026-03-06T13:50:43.416 INFO:teuthology.orchestra.run.vm04.stdout: Running scriptlet: ceph-mgr-rook-2:19.2.3-47.gc24117fd552.el9.clyso 1/113 2026-03-06T13:50:43.417 INFO:teuthology.orchestra.run.vm06.stdout: Preparing : 1/1 2026-03-06T13:50:43.417 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : ceph-mgr-rook-2:19.2.3-47.gc24117fd552.el9.clyso 1/113 2026-03-06T13:50:43.424 INFO:teuthology.orchestra.run.vm06.stdout: Running scriptlet: ceph-mgr-rook-2:19.2.3-47.gc24117fd552.el9.clyso 1/113 2026-03-06T13:50:43.428 INFO:teuthology.orchestra.run.vm02.stdout: Running scriptlet: ceph-mgr-2:19.2.3-47.gc24117fd552.el9.clyso.x86_ 2/113 2026-03-06T13:50:43.428 INFO:teuthology.orchestra.run.vm02.stdout:Glob pattern passed to enable, but globs are not supported for this. 2026-03-06T13:50:43.428 INFO:teuthology.orchestra.run.vm02.stdout:Invalid unit name "ceph-mgr@*.service" escaped as "ceph-mgr@\x2a.service". 2026-03-06T13:50:43.428 INFO:teuthology.orchestra.run.vm02.stdout:Removed "/etc/systemd/system/multi-user.target.wants/ceph-mgr.target". 2026-03-06T13:50:43.428 INFO:teuthology.orchestra.run.vm02.stdout:Removed "/etc/systemd/system/ceph.target.wants/ceph-mgr.target". 2026-03-06T13:50:43.429 INFO:teuthology.orchestra.run.vm02.stdout: 2026-03-06T13:50:43.429 INFO:teuthology.orchestra.run.vm02.stdout: Erasing : ceph-mgr-2:19.2.3-47.gc24117fd552.el9.clyso.x86_ 2/113 2026-03-06T13:50:43.433 INFO:teuthology.orchestra.run.vm04.stdout: Running scriptlet: ceph-mgr-2:19.2.3-47.gc24117fd552.el9.clyso.x86_ 2/113 2026-03-06T13:50:43.433 INFO:teuthology.orchestra.run.vm04.stdout:Glob pattern passed to enable, but globs are not supported for this. 2026-03-06T13:50:43.433 INFO:teuthology.orchestra.run.vm04.stdout:Invalid unit name "ceph-mgr@*.service" escaped as "ceph-mgr@\x2a.service". 2026-03-06T13:50:43.433 INFO:teuthology.orchestra.run.vm04.stdout:Removed "/etc/systemd/system/multi-user.target.wants/ceph-mgr.target". 2026-03-06T13:50:43.433 INFO:teuthology.orchestra.run.vm04.stdout:Removed "/etc/systemd/system/ceph.target.wants/ceph-mgr.target". 2026-03-06T13:50:43.433 INFO:teuthology.orchestra.run.vm04.stdout: 2026-03-06T13:50:43.434 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : ceph-mgr-2:19.2.3-47.gc24117fd552.el9.clyso.x86_ 2/113 2026-03-06T13:50:43.442 INFO:teuthology.orchestra.run.vm06.stdout: Running scriptlet: ceph-mgr-2:19.2.3-47.gc24117fd552.el9.clyso.x86_ 2/113 2026-03-06T13:50:43.442 INFO:teuthology.orchestra.run.vm06.stdout:Glob pattern passed to enable, but globs are not supported for this. 2026-03-06T13:50:43.442 INFO:teuthology.orchestra.run.vm06.stdout:Invalid unit name "ceph-mgr@*.service" escaped as "ceph-mgr@\x2a.service". 2026-03-06T13:50:43.442 INFO:teuthology.orchestra.run.vm06.stdout:Removed "/etc/systemd/system/multi-user.target.wants/ceph-mgr.target". 2026-03-06T13:50:43.442 INFO:teuthology.orchestra.run.vm06.stdout:Removed "/etc/systemd/system/ceph.target.wants/ceph-mgr.target". 2026-03-06T13:50:43.442 INFO:teuthology.orchestra.run.vm06.stdout: 2026-03-06T13:50:43.443 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : ceph-mgr-2:19.2.3-47.gc24117fd552.el9.clyso.x86_ 2/113 2026-03-06T13:50:43.444 INFO:teuthology.orchestra.run.vm02.stdout: Running scriptlet: ceph-mgr-2:19.2.3-47.gc24117fd552.el9.clyso.x86_ 2/113 2026-03-06T13:50:43.448 INFO:teuthology.orchestra.run.vm04.stdout: Running scriptlet: ceph-mgr-2:19.2.3-47.gc24117fd552.el9.clyso.x86_ 2/113 2026-03-06T13:50:43.455 INFO:teuthology.orchestra.run.vm06.stdout: Running scriptlet: ceph-mgr-2:19.2.3-47.gc24117fd552.el9.clyso.x86_ 2/113 2026-03-06T13:50:43.466 INFO:teuthology.orchestra.run.vm02.stdout: Erasing : ceph-mgr-modules-core-2:19.2.3-47.gc24117fd552.e 3/113 2026-03-06T13:50:43.467 INFO:teuthology.orchestra.run.vm02.stdout: Erasing : ceph-mgr-dashboard-2:19.2.3-47.gc24117fd552.el9. 4/113 2026-03-06T13:50:43.471 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : ceph-mgr-modules-core-2:19.2.3-47.gc24117fd552.e 3/113 2026-03-06T13:50:43.472 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : ceph-mgr-dashboard-2:19.2.3-47.gc24117fd552.el9. 4/113 2026-03-06T13:50:43.478 INFO:teuthology.orchestra.run.vm02.stdout: Running scriptlet: ceph-mgr-dashboard-2:19.2.3-47.gc24117fd552.el9. 4/113 2026-03-06T13:50:43.478 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : ceph-mgr-modules-core-2:19.2.3-47.gc24117fd552.e 3/113 2026-03-06T13:50:43.478 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : ceph-mgr-dashboard-2:19.2.3-47.gc24117fd552.el9. 4/113 2026-03-06T13:50:43.482 INFO:teuthology.orchestra.run.vm02.stdout: Erasing : python3-influxdb-5.3.1-1.el9.noarch 5/113 2026-03-06T13:50:43.483 INFO:teuthology.orchestra.run.vm02.stdout: Erasing : ceph-mgr-cephadm-2:19.2.3-47.gc24117fd552.el9.cl 6/113 2026-03-06T13:50:43.483 INFO:teuthology.orchestra.run.vm04.stdout: Running scriptlet: ceph-mgr-dashboard-2:19.2.3-47.gc24117fd552.el9. 4/113 2026-03-06T13:50:43.488 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : python3-influxdb-5.3.1-1.el9.noarch 5/113 2026-03-06T13:50:43.488 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : ceph-mgr-cephadm-2:19.2.3-47.gc24117fd552.el9.cl 6/113 2026-03-06T13:50:43.490 INFO:teuthology.orchestra.run.vm06.stdout: Running scriptlet: ceph-mgr-dashboard-2:19.2.3-47.gc24117fd552.el9. 4/113 2026-03-06T13:50:43.494 INFO:teuthology.orchestra.run.vm02.stdout: Running scriptlet: ceph-mgr-cephadm-2:19.2.3-47.gc24117fd552.el9.cl 6/113 2026-03-06T13:50:43.494 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : python3-influxdb-5.3.1-1.el9.noarch 5/113 2026-03-06T13:50:43.494 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : ceph-mgr-cephadm-2:19.2.3-47.gc24117fd552.el9.cl 6/113 2026-03-06T13:50:43.499 INFO:teuthology.orchestra.run.vm04.stdout: Running scriptlet: ceph-mgr-cephadm-2:19.2.3-47.gc24117fd552.el9.cl 6/113 2026-03-06T13:50:43.500 INFO:teuthology.orchestra.run.vm02.stdout: Erasing : python3-cherrypy-18.6.1-2.el9.noarch 7/113 2026-03-06T13:50:43.504 INFO:teuthology.orchestra.run.vm02.stdout: Erasing : python3-cheroot-10.0.1-4.el9.noarch 8/113 2026-03-06T13:50:43.505 INFO:teuthology.orchestra.run.vm06.stdout: Running scriptlet: ceph-mgr-cephadm-2:19.2.3-47.gc24117fd552.el9.cl 6/113 2026-03-06T13:50:43.506 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : python3-cherrypy-18.6.1-2.el9.noarch 7/113 2026-03-06T13:50:43.510 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : python3-cheroot-10.0.1-4.el9.noarch 8/113 2026-03-06T13:50:43.512 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : python3-cherrypy-18.6.1-2.el9.noarch 7/113 2026-03-06T13:50:43.512 INFO:teuthology.orchestra.run.vm02.stdout: Erasing : python3-grpcio-tools-1.46.7-10.el9.x86_64 9/113 2026-03-06T13:50:43.515 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : python3-cheroot-10.0.1-4.el9.noarch 8/113 2026-03-06T13:50:43.516 INFO:teuthology.orchestra.run.vm02.stdout: Erasing : python3-grpcio-1.46.7-10.el9.x86_64 10/113 2026-03-06T13:50:43.518 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : python3-grpcio-tools-1.46.7-10.el9.x86_64 9/113 2026-03-06T13:50:43.522 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : python3-grpcio-1.46.7-10.el9.x86_64 10/113 2026-03-06T13:50:43.523 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : python3-grpcio-tools-1.46.7-10.el9.x86_64 9/113 2026-03-06T13:50:43.527 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : python3-grpcio-1.46.7-10.el9.x86_64 10/113 2026-03-06T13:50:43.534 INFO:teuthology.orchestra.run.vm02.stdout: Running scriptlet: ceph-osd-2:19.2.3-47.gc24117fd552.el9.clyso.x86_ 11/113 2026-03-06T13:50:43.534 INFO:teuthology.orchestra.run.vm02.stdout:Glob pattern passed to enable, but globs are not supported for this. 2026-03-06T13:50:43.534 INFO:teuthology.orchestra.run.vm02.stdout:Invalid unit name "ceph-osd@*.service" escaped as "ceph-osd@\x2a.service". 2026-03-06T13:50:43.534 INFO:teuthology.orchestra.run.vm02.stdout:Removed "/etc/systemd/system/multi-user.target.wants/ceph-osd.target". 2026-03-06T13:50:43.534 INFO:teuthology.orchestra.run.vm02.stdout:Removed "/etc/systemd/system/ceph.target.wants/ceph-osd.target". 2026-03-06T13:50:43.534 INFO:teuthology.orchestra.run.vm02.stdout: 2026-03-06T13:50:43.539 INFO:teuthology.orchestra.run.vm02.stdout: Erasing : ceph-osd-2:19.2.3-47.gc24117fd552.el9.clyso.x86_ 11/113 2026-03-06T13:50:43.541 INFO:teuthology.orchestra.run.vm04.stdout: Running scriptlet: ceph-osd-2:19.2.3-47.gc24117fd552.el9.clyso.x86_ 11/113 2026-03-06T13:50:43.541 INFO:teuthology.orchestra.run.vm04.stdout:Glob pattern passed to enable, but globs are not supported for this. 2026-03-06T13:50:43.541 INFO:teuthology.orchestra.run.vm04.stdout:Invalid unit name "ceph-osd@*.service" escaped as "ceph-osd@\x2a.service". 2026-03-06T13:50:43.541 INFO:teuthology.orchestra.run.vm04.stdout:Removed "/etc/systemd/system/multi-user.target.wants/ceph-osd.target". 2026-03-06T13:50:43.541 INFO:teuthology.orchestra.run.vm04.stdout:Removed "/etc/systemd/system/ceph.target.wants/ceph-osd.target". 2026-03-06T13:50:43.541 INFO:teuthology.orchestra.run.vm04.stdout: 2026-03-06T13:50:43.546 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : ceph-osd-2:19.2.3-47.gc24117fd552.el9.clyso.x86_ 11/113 2026-03-06T13:50:43.547 INFO:teuthology.orchestra.run.vm02.stdout: Running scriptlet: ceph-osd-2:19.2.3-47.gc24117fd552.el9.clyso.x86_ 11/113 2026-03-06T13:50:43.548 INFO:teuthology.orchestra.run.vm06.stdout: Running scriptlet: ceph-osd-2:19.2.3-47.gc24117fd552.el9.clyso.x86_ 11/113 2026-03-06T13:50:43.548 INFO:teuthology.orchestra.run.vm06.stdout:Glob pattern passed to enable, but globs are not supported for this. 2026-03-06T13:50:43.548 INFO:teuthology.orchestra.run.vm06.stdout:Invalid unit name "ceph-osd@*.service" escaped as "ceph-osd@\x2a.service". 2026-03-06T13:50:43.548 INFO:teuthology.orchestra.run.vm06.stdout:Removed "/etc/systemd/system/multi-user.target.wants/ceph-osd.target". 2026-03-06T13:50:43.548 INFO:teuthology.orchestra.run.vm06.stdout:Removed "/etc/systemd/system/ceph.target.wants/ceph-osd.target". 2026-03-06T13:50:43.548 INFO:teuthology.orchestra.run.vm06.stdout: 2026-03-06T13:50:43.553 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : ceph-osd-2:19.2.3-47.gc24117fd552.el9.clyso.x86_ 11/113 2026-03-06T13:50:43.555 INFO:teuthology.orchestra.run.vm04.stdout: Running scriptlet: ceph-osd-2:19.2.3-47.gc24117fd552.el9.clyso.x86_ 11/113 2026-03-06T13:50:43.561 INFO:teuthology.orchestra.run.vm02.stdout: Running scriptlet: ceph-volume-2:19.2.3-47.gc24117fd552.el9.clyso.n 12/113 2026-03-06T13:50:43.561 INFO:teuthology.orchestra.run.vm02.stdout:Glob pattern passed to enable, but globs are not supported for this. 2026-03-06T13:50:43.561 INFO:teuthology.orchestra.run.vm02.stdout:Invalid unit name "ceph-volume@*.service" escaped as "ceph-volume@\x2a.service". 2026-03-06T13:50:43.561 INFO:teuthology.orchestra.run.vm02.stdout: 2026-03-06T13:50:43.562 INFO:teuthology.orchestra.run.vm06.stdout: Running scriptlet: ceph-osd-2:19.2.3-47.gc24117fd552.el9.clyso.x86_ 11/113 2026-03-06T13:50:43.570 INFO:teuthology.orchestra.run.vm02.stdout: Erasing : ceph-volume-2:19.2.3-47.gc24117fd552.el9.clyso.n 12/113 2026-03-06T13:50:43.571 INFO:teuthology.orchestra.run.vm04.stdout: Running scriptlet: ceph-volume-2:19.2.3-47.gc24117fd552.el9.clyso.n 12/113 2026-03-06T13:50:43.571 INFO:teuthology.orchestra.run.vm04.stdout:Glob pattern passed to enable, but globs are not supported for this. 2026-03-06T13:50:43.571 INFO:teuthology.orchestra.run.vm04.stdout:Invalid unit name "ceph-volume@*.service" escaped as "ceph-volume@\x2a.service". 2026-03-06T13:50:43.571 INFO:teuthology.orchestra.run.vm04.stdout: 2026-03-06T13:50:43.577 INFO:teuthology.orchestra.run.vm06.stdout: Running scriptlet: ceph-volume-2:19.2.3-47.gc24117fd552.el9.clyso.n 12/113 2026-03-06T13:50:43.577 INFO:teuthology.orchestra.run.vm06.stdout:Glob pattern passed to enable, but globs are not supported for this. 2026-03-06T13:50:43.577 INFO:teuthology.orchestra.run.vm06.stdout:Invalid unit name "ceph-volume@*.service" escaped as "ceph-volume@\x2a.service". 2026-03-06T13:50:43.577 INFO:teuthology.orchestra.run.vm06.stdout: 2026-03-06T13:50:43.579 INFO:teuthology.orchestra.run.vm02.stdout: Running scriptlet: ceph-volume-2:19.2.3-47.gc24117fd552.el9.clyso.n 12/113 2026-03-06T13:50:43.579 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : ceph-volume-2:19.2.3-47.gc24117fd552.el9.clyso.n 12/113 2026-03-06T13:50:43.581 INFO:teuthology.orchestra.run.vm02.stdout: Erasing : python3-jaraco-collections-3.0.0-8.el9.noarch 13/113 2026-03-06T13:50:43.585 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : ceph-volume-2:19.2.3-47.gc24117fd552.el9.clyso.n 12/113 2026-03-06T13:50:43.586 INFO:teuthology.orchestra.run.vm02.stdout: Erasing : python3-jaraco-text-4.0.0-2.el9.noarch 14/113 2026-03-06T13:50:43.588 INFO:teuthology.orchestra.run.vm04.stdout: Running scriptlet: ceph-volume-2:19.2.3-47.gc24117fd552.el9.clyso.n 12/113 2026-03-06T13:50:43.590 INFO:teuthology.orchestra.run.vm02.stdout: Erasing : python3-jinja2-2.11.3-8.el9.noarch 15/113 2026-03-06T13:50:43.591 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : python3-jaraco-collections-3.0.0-8.el9.noarch 13/113 2026-03-06T13:50:43.594 INFO:teuthology.orchestra.run.vm06.stdout: Running scriptlet: ceph-volume-2:19.2.3-47.gc24117fd552.el9.clyso.n 12/113 2026-03-06T13:50:43.596 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : python3-jaraco-text-4.0.0-2.el9.noarch 14/113 2026-03-06T13:50:43.596 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : python3-jaraco-collections-3.0.0-8.el9.noarch 13/113 2026-03-06T13:50:43.600 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : python3-jinja2-2.11.3-8.el9.noarch 15/113 2026-03-06T13:50:43.601 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : python3-jaraco-text-4.0.0-2.el9.noarch 14/113 2026-03-06T13:50:43.602 INFO:teuthology.orchestra.run.vm02.stdout: Erasing : python3-saml-1.16.0-1.el9.noarch 16/113 2026-03-06T13:50:43.605 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : python3-jinja2-2.11.3-8.el9.noarch 15/113 2026-03-06T13:50:43.611 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : python3-saml-1.16.0-1.el9.noarch 16/113 2026-03-06T13:50:43.616 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : python3-saml-1.16.0-1.el9.noarch 16/113 2026-03-06T13:50:43.632 INFO:teuthology.orchestra.run.vm02.stdout: Erasing : python3-pecan-1.4.2-3.el9.noarch 17/113 2026-03-06T13:50:43.638 INFO:teuthology.orchestra.run.vm02.stdout: Erasing : python3-babel-2.9.1-2.el9.noarch 18/113 2026-03-06T13:50:43.640 INFO:teuthology.orchestra.run.vm02.stdout: Erasing : python3-jaraco-classes-3.2.1-5.el9.noarch 19/113 2026-03-06T13:50:43.641 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : python3-pecan-1.4.2-3.el9.noarch 17/113 2026-03-06T13:50:43.645 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : python3-pecan-1.4.2-3.el9.noarch 17/113 2026-03-06T13:50:43.647 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : python3-babel-2.9.1-2.el9.noarch 18/113 2026-03-06T13:50:43.649 INFO:teuthology.orchestra.run.vm02.stdout: Erasing : python3-pyOpenSSL-21.0.0-1.el9.noarch 20/113 2026-03-06T13:50:43.651 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : python3-jaraco-classes-3.2.1-5.el9.noarch 19/113 2026-03-06T13:50:43.651 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : python3-babel-2.9.1-2.el9.noarch 18/113 2026-03-06T13:50:43.654 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : python3-jaraco-classes-3.2.1-5.el9.noarch 19/113 2026-03-06T13:50:43.659 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : python3-pyOpenSSL-21.0.0-1.el9.noarch 20/113 2026-03-06T13:50:43.659 INFO:teuthology.orchestra.run.vm02.stdout: Erasing : python3-asyncssh-2.13.2-5.el9.noarch 21/113 2026-03-06T13:50:43.660 INFO:teuthology.orchestra.run.vm02.stdout: Erasing : ceph-mgr-diskprediction-local-2:19.2.3-47.gc2411 22/113 2026-03-06T13:50:43.662 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : python3-pyOpenSSL-21.0.0-1.el9.noarch 20/113 2026-03-06T13:50:43.667 INFO:teuthology.orchestra.run.vm02.stdout: Running scriptlet: ceph-mgr-diskprediction-local-2:19.2.3-47.gc2411 22/113 2026-03-06T13:50:43.670 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : python3-asyncssh-2.13.2-5.el9.noarch 21/113 2026-03-06T13:50:43.670 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : ceph-mgr-diskprediction-local-2:19.2.3-47.gc2411 22/113 2026-03-06T13:50:43.673 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : python3-asyncssh-2.13.2-5.el9.noarch 21/113 2026-03-06T13:50:43.673 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : ceph-mgr-diskprediction-local-2:19.2.3-47.gc2411 22/113 2026-03-06T13:50:43.677 INFO:teuthology.orchestra.run.vm04.stdout: Running scriptlet: ceph-mgr-diskprediction-local-2:19.2.3-47.gc2411 22/113 2026-03-06T13:50:43.679 INFO:teuthology.orchestra.run.vm06.stdout: Running scriptlet: ceph-mgr-diskprediction-local-2:19.2.3-47.gc2411 22/113 2026-03-06T13:50:43.757 INFO:teuthology.orchestra.run.vm02.stdout: Erasing : python3-jsonpatch-1.21-16.el9.noarch 23/113 2026-03-06T13:50:43.767 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : python3-jsonpatch-1.21-16.el9.noarch 23/113 2026-03-06T13:50:43.767 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : python3-jsonpatch-1.21-16.el9.noarch 23/113 2026-03-06T13:50:43.772 INFO:teuthology.orchestra.run.vm02.stdout: Erasing : python3-scipy-1.9.3-2.el9.x86_64 24/113 2026-03-06T13:50:43.779 INFO:teuthology.orchestra.run.vm02.stdout: Erasing : python3-xmlsec-1.3.13-1.el9.x86_64 25/113 2026-03-06T13:50:43.783 INFO:teuthology.orchestra.run.vm02.stdout: Erasing : python3-lxml-4.6.5-3.el9.x86_64 26/113 2026-03-06T13:50:43.783 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : python3-scipy-1.9.3-2.el9.x86_64 24/113 2026-03-06T13:50:43.784 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : python3-scipy-1.9.3-2.el9.x86_64 24/113 2026-03-06T13:50:43.790 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : python3-xmlsec-1.3.13-1.el9.x86_64 25/113 2026-03-06T13:50:43.790 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : python3-xmlsec-1.3.13-1.el9.x86_64 25/113 2026-03-06T13:50:43.794 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : python3-lxml-4.6.5-3.el9.x86_64 26/113 2026-03-06T13:50:43.794 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : python3-lxml-4.6.5-3.el9.x86_64 26/113 2026-03-06T13:50:43.795 INFO:teuthology.orchestra.run.vm02.stdout: Running scriptlet: libstoragemgmt-1.10.1-1.el9.x86_64 27/113 2026-03-06T13:50:43.795 INFO:teuthology.orchestra.run.vm02.stdout:Removed "/etc/systemd/system/multi-user.target.wants/libstoragemgmt.service". 2026-03-06T13:50:43.795 INFO:teuthology.orchestra.run.vm02.stdout: 2026-03-06T13:50:43.796 INFO:teuthology.orchestra.run.vm02.stdout: Erasing : libstoragemgmt-1.10.1-1.el9.x86_64 27/113 2026-03-06T13:50:43.807 INFO:teuthology.orchestra.run.vm04.stdout: Running scriptlet: libstoragemgmt-1.10.1-1.el9.x86_64 27/113 2026-03-06T13:50:43.807 INFO:teuthology.orchestra.run.vm04.stdout:Removed "/etc/systemd/system/multi-user.target.wants/libstoragemgmt.service". 2026-03-06T13:50:43.807 INFO:teuthology.orchestra.run.vm04.stdout: 2026-03-06T13:50:43.807 INFO:teuthology.orchestra.run.vm06.stdout: Running scriptlet: libstoragemgmt-1.10.1-1.el9.x86_64 27/113 2026-03-06T13:50:43.807 INFO:teuthology.orchestra.run.vm06.stdout:Removed "/etc/systemd/system/multi-user.target.wants/libstoragemgmt.service". 2026-03-06T13:50:43.807 INFO:teuthology.orchestra.run.vm06.stdout: 2026-03-06T13:50:43.808 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : libstoragemgmt-1.10.1-1.el9.x86_64 27/113 2026-03-06T13:50:43.808 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : libstoragemgmt-1.10.1-1.el9.x86_64 27/113 2026-03-06T13:50:43.822 INFO:teuthology.orchestra.run.vm02.stdout: Running scriptlet: libstoragemgmt-1.10.1-1.el9.x86_64 27/113 2026-03-06T13:50:43.826 INFO:teuthology.orchestra.run.vm02.stdout: Erasing : python3-numpy-f2py-1:1.23.5-2.el9.x86_64 28/113 2026-03-06T13:50:43.828 INFO:teuthology.orchestra.run.vm02.stdout: Erasing : xmlsec1-openssl-1.2.29-13.el9.x86_64 29/113 2026-03-06T13:50:43.832 INFO:teuthology.orchestra.run.vm04.stdout: Running scriptlet: libstoragemgmt-1.10.1-1.el9.x86_64 27/113 2026-03-06T13:50:43.833 INFO:teuthology.orchestra.run.vm06.stdout: Running scriptlet: libstoragemgmt-1.10.1-1.el9.x86_64 27/113 2026-03-06T13:50:43.838 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : python3-numpy-f2py-1:1.23.5-2.el9.x86_64 28/113 2026-03-06T13:50:43.838 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : python3-numpy-f2py-1:1.23.5-2.el9.x86_64 28/113 2026-03-06T13:50:43.840 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : xmlsec1-openssl-1.2.29-13.el9.x86_64 29/113 2026-03-06T13:50:43.840 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : xmlsec1-openssl-1.2.29-13.el9.x86_64 29/113 2026-03-06T13:50:43.841 INFO:teuthology.orchestra.run.vm02.stdout: Erasing : xmlsec1-1.2.29-13.el9.x86_64 30/113 2026-03-06T13:50:43.847 INFO:teuthology.orchestra.run.vm02.stdout: Erasing : python3-cryptography-36.0.1-5.el9.x86_64 31/113 2026-03-06T13:50:43.849 INFO:teuthology.orchestra.run.vm02.stdout: Erasing : protobuf-compiler-3.14.0-17.el9.x86_64 32/113 2026-03-06T13:50:43.851 INFO:teuthology.orchestra.run.vm02.stdout: Erasing : python3-bcrypt-3.2.2-1.el9.x86_64 33/113 2026-03-06T13:50:43.852 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : xmlsec1-1.2.29-13.el9.x86_64 30/113 2026-03-06T13:50:43.853 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : xmlsec1-1.2.29-13.el9.x86_64 30/113 2026-03-06T13:50:43.857 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : python3-cryptography-36.0.1-5.el9.x86_64 31/113 2026-03-06T13:50:43.858 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : python3-cryptography-36.0.1-5.el9.x86_64 31/113 2026-03-06T13:50:43.860 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : protobuf-compiler-3.14.0-17.el9.x86_64 32/113 2026-03-06T13:50:43.861 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : protobuf-compiler-3.14.0-17.el9.x86_64 32/113 2026-03-06T13:50:43.862 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : python3-bcrypt-3.2.2-1.el9.x86_64 33/113 2026-03-06T13:50:43.863 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : python3-bcrypt-3.2.2-1.el9.x86_64 33/113 2026-03-06T13:50:43.871 INFO:teuthology.orchestra.run.vm02.stdout: Running scriptlet: rbd-mirror-2:19.2.3-47.gc24117fd552.el9.clyso.x8 34/113 2026-03-06T13:50:43.872 INFO:teuthology.orchestra.run.vm02.stdout:Glob pattern passed to enable, but globs are not supported for this. 2026-03-06T13:50:43.872 INFO:teuthology.orchestra.run.vm02.stdout:Invalid unit name "ceph-rbd-mirror@*.service" escaped as "ceph-rbd-mirror@\x2a.service". 2026-03-06T13:50:43.872 INFO:teuthology.orchestra.run.vm02.stdout:Removed "/etc/systemd/system/multi-user.target.wants/ceph-rbd-mirror.target". 2026-03-06T13:50:43.872 INFO:teuthology.orchestra.run.vm02.stdout:Removed "/etc/systemd/system/ceph.target.wants/ceph-rbd-mirror.target". 2026-03-06T13:50:43.872 INFO:teuthology.orchestra.run.vm02.stdout: 2026-03-06T13:50:43.873 INFO:teuthology.orchestra.run.vm02.stdout: Erasing : rbd-mirror-2:19.2.3-47.gc24117fd552.el9.clyso.x8 34/113 2026-03-06T13:50:43.883 INFO:teuthology.orchestra.run.vm04.stdout: Running scriptlet: rbd-mirror-2:19.2.3-47.gc24117fd552.el9.clyso.x8 34/113 2026-03-06T13:50:43.883 INFO:teuthology.orchestra.run.vm04.stdout:Glob pattern passed to enable, but globs are not supported for this. 2026-03-06T13:50:43.883 INFO:teuthology.orchestra.run.vm04.stdout:Invalid unit name "ceph-rbd-mirror@*.service" escaped as "ceph-rbd-mirror@\x2a.service". 2026-03-06T13:50:43.883 INFO:teuthology.orchestra.run.vm04.stdout:Removed "/etc/systemd/system/multi-user.target.wants/ceph-rbd-mirror.target". 2026-03-06T13:50:43.883 INFO:teuthology.orchestra.run.vm04.stdout:Removed "/etc/systemd/system/ceph.target.wants/ceph-rbd-mirror.target". 2026-03-06T13:50:43.883 INFO:teuthology.orchestra.run.vm04.stdout: 2026-03-06T13:50:43.883 INFO:teuthology.orchestra.run.vm06.stdout: Running scriptlet: rbd-mirror-2:19.2.3-47.gc24117fd552.el9.clyso.x8 34/113 2026-03-06T13:50:43.883 INFO:teuthology.orchestra.run.vm06.stdout:Glob pattern passed to enable, but globs are not supported for this. 2026-03-06T13:50:43.883 INFO:teuthology.orchestra.run.vm06.stdout:Invalid unit name "ceph-rbd-mirror@*.service" escaped as "ceph-rbd-mirror@\x2a.service". 2026-03-06T13:50:43.883 INFO:teuthology.orchestra.run.vm06.stdout:Removed "/etc/systemd/system/multi-user.target.wants/ceph-rbd-mirror.target". 2026-03-06T13:50:43.883 INFO:teuthology.orchestra.run.vm06.stdout:Removed "/etc/systemd/system/ceph.target.wants/ceph-rbd-mirror.target". 2026-03-06T13:50:43.883 INFO:teuthology.orchestra.run.vm06.stdout: 2026-03-06T13:50:43.884 INFO:teuthology.orchestra.run.vm02.stdout: Running scriptlet: rbd-mirror-2:19.2.3-47.gc24117fd552.el9.clyso.x8 34/113 2026-03-06T13:50:43.884 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : rbd-mirror-2:19.2.3-47.gc24117fd552.el9.clyso.x8 34/113 2026-03-06T13:50:43.885 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : rbd-mirror-2:19.2.3-47.gc24117fd552.el9.clyso.x8 34/113 2026-03-06T13:50:43.888 INFO:teuthology.orchestra.run.vm02.stdout: Erasing : python3-mako-1.1.4-6.el9.noarch 35/113 2026-03-06T13:50:43.890 INFO:teuthology.orchestra.run.vm02.stdout: Erasing : python3-jaraco-context-6.0.1-3.el9.noarch 36/113 2026-03-06T13:50:43.893 INFO:teuthology.orchestra.run.vm02.stdout: Erasing : python3-portend-3.1.0-2.el9.noarch 37/113 2026-03-06T13:50:43.895 INFO:teuthology.orchestra.run.vm02.stdout: Erasing : python3-tempora-5.0.0-2.el9.noarch 38/113 2026-03-06T13:50:43.896 INFO:teuthology.orchestra.run.vm04.stdout: Running scriptlet: rbd-mirror-2:19.2.3-47.gc24117fd552.el9.clyso.x8 34/113 2026-03-06T13:50:43.896 INFO:teuthology.orchestra.run.vm06.stdout: Running scriptlet: rbd-mirror-2:19.2.3-47.gc24117fd552.el9.clyso.x8 34/113 2026-03-06T13:50:43.899 INFO:teuthology.orchestra.run.vm02.stdout: Erasing : python3-jaraco-functools-3.5.0-2.el9.noarch 39/113 2026-03-06T13:50:43.900 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : python3-mako-1.1.4-6.el9.noarch 35/113 2026-03-06T13:50:43.900 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : python3-mako-1.1.4-6.el9.noarch 35/113 2026-03-06T13:50:43.901 INFO:teuthology.orchestra.run.vm02.stdout: Erasing : python3-routes-2.5.1-5.el9.noarch 40/113 2026-03-06T13:50:43.901 INFO:teuthology.orchestra.run.vm02.stdout: Erasing : ceph-mgr-k8sevents-2:19.2.3-47.gc24117fd552.el9. 41/113 2026-03-06T13:50:43.902 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : python3-jaraco-context-6.0.1-3.el9.noarch 36/113 2026-03-06T13:50:43.902 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : python3-jaraco-context-6.0.1-3.el9.noarch 36/113 2026-03-06T13:50:43.905 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : python3-portend-3.1.0-2.el9.noarch 37/113 2026-03-06T13:50:43.905 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : python3-portend-3.1.0-2.el9.noarch 37/113 2026-03-06T13:50:43.908 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : python3-tempora-5.0.0-2.el9.noarch 38/113 2026-03-06T13:50:43.908 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : python3-tempora-5.0.0-2.el9.noarch 38/113 2026-03-06T13:50:43.911 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : python3-jaraco-functools-3.5.0-2.el9.noarch 39/113 2026-03-06T13:50:43.911 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : python3-jaraco-functools-3.5.0-2.el9.noarch 39/113 2026-03-06T13:50:43.913 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : python3-routes-2.5.1-5.el9.noarch 40/113 2026-03-06T13:50:43.914 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : ceph-mgr-k8sevents-2:19.2.3-47.gc24117fd552.el9. 41/113 2026-03-06T13:50:43.914 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : python3-routes-2.5.1-5.el9.noarch 40/113 2026-03-06T13:50:43.914 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : ceph-mgr-k8sevents-2:19.2.3-47.gc24117fd552.el9. 41/113 2026-03-06T13:50:43.953 INFO:teuthology.orchestra.run.vm02.stdout: Running scriptlet: ceph-mgr-k8sevents-2:19.2.3-47.gc24117fd552.el9. 41/113 2026-03-06T13:50:43.961 INFO:teuthology.orchestra.run.vm02.stdout: Erasing : python3-kubernetes-1:26.1.0-3.el9.noarch 42/113 2026-03-06T13:50:43.963 INFO:teuthology.orchestra.run.vm06.stdout: Running scriptlet: ceph-mgr-k8sevents-2:19.2.3-47.gc24117fd552.el9. 41/113 2026-03-06T13:50:43.964 INFO:teuthology.orchestra.run.vm02.stdout: Erasing : python3-requests-oauthlib-1.3.0-12.el9.noarch 43/113 2026-03-06T13:50:43.965 INFO:teuthology.orchestra.run.vm04.stdout: Running scriptlet: ceph-mgr-k8sevents-2:19.2.3-47.gc24117fd552.el9. 41/113 2026-03-06T13:50:43.972 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : python3-kubernetes-1:26.1.0-3.el9.noarch 42/113 2026-03-06T13:50:43.973 INFO:teuthology.orchestra.run.vm02.stdout: Erasing : python3-requests-2.25.1-10.el9.noarch 44/113 2026-03-06T13:50:43.974 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : python3-kubernetes-1:26.1.0-3.el9.noarch 42/113 2026-03-06T13:50:43.977 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : python3-requests-oauthlib-1.3.0-12.el9.noarch 43/113 2026-03-06T13:50:43.978 INFO:teuthology.orchestra.run.vm02.stdout: Erasing : python3-google-auth-1:2.45.0-1.el9.noarch 45/113 2026-03-06T13:50:43.979 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : python3-requests-oauthlib-1.3.0-12.el9.noarch 43/113 2026-03-06T13:50:43.987 INFO:teuthology.orchestra.run.vm02.stdout: Erasing : python3-rsa-4.9-2.el9.noarch 46/113 2026-03-06T13:50:43.989 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : python3-requests-2.25.1-10.el9.noarch 44/113 2026-03-06T13:50:43.990 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : python3-requests-2.25.1-10.el9.noarch 44/113 2026-03-06T13:50:43.994 INFO:teuthology.orchestra.run.vm02.stdout: Erasing : python3-pyasn1-modules-0.4.8-7.el9.noarch 47/113 2026-03-06T13:50:43.994 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : python3-google-auth-1:2.45.0-1.el9.noarch 45/113 2026-03-06T13:50:43.995 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : python3-google-auth-1:2.45.0-1.el9.noarch 45/113 2026-03-06T13:50:43.998 INFO:teuthology.orchestra.run.vm02.stdout: Erasing : python3-urllib3-1.26.5-7.el9.noarch 48/113 2026-03-06T13:50:44.002 INFO:teuthology.orchestra.run.vm02.stdout: Erasing : python3-cffi-1.14.5-5.el9.x86_64 49/113 2026-03-06T13:50:44.004 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : python3-rsa-4.9-2.el9.noarch 46/113 2026-03-06T13:50:44.005 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : python3-rsa-4.9-2.el9.noarch 46/113 2026-03-06T13:50:44.010 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : python3-pyasn1-modules-0.4.8-7.el9.noarch 47/113 2026-03-06T13:50:44.011 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : python3-pyasn1-modules-0.4.8-7.el9.noarch 47/113 2026-03-06T13:50:44.014 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : python3-urllib3-1.26.5-7.el9.noarch 48/113 2026-03-06T13:50:44.015 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : python3-urllib3-1.26.5-7.el9.noarch 48/113 2026-03-06T13:50:44.018 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : python3-cffi-1.14.5-5.el9.x86_64 49/113 2026-03-06T13:50:44.020 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : python3-cffi-1.14.5-5.el9.x86_64 49/113 2026-03-06T13:50:44.047 INFO:teuthology.orchestra.run.vm02.stdout: Erasing : python3-pycparser-2.20-6.el9.noarch 50/113 2026-03-06T13:50:44.057 INFO:teuthology.orchestra.run.vm02.stdout: Erasing : python3-numpy-1:1.23.5-2.el9.x86_64 51/113 2026-03-06T13:50:44.060 INFO:teuthology.orchestra.run.vm02.stdout: Erasing : flexiblas-netlib-3.0.4-9.el9.x86_64 52/113 2026-03-06T13:50:44.062 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : python3-pycparser-2.20-6.el9.noarch 50/113 2026-03-06T13:50:44.064 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : python3-pycparser-2.20-6.el9.noarch 50/113 2026-03-06T13:50:44.065 INFO:teuthology.orchestra.run.vm02.stdout: Erasing : flexiblas-openblas-openmp-3.0.4-9.el9.x86_64 53/113 2026-03-06T13:50:44.067 INFO:teuthology.orchestra.run.vm02.stdout: Erasing : openblas-openmp-0.3.29-1.el9.x86_64 54/113 2026-03-06T13:50:44.071 INFO:teuthology.orchestra.run.vm02.stdout: Erasing : libgfortran-11.5.0-14.el9.x86_64 55/113 2026-03-06T13:50:44.073 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : python3-numpy-1:1.23.5-2.el9.x86_64 51/113 2026-03-06T13:50:44.073 INFO:teuthology.orchestra.run.vm02.stdout: Erasing : python3-libstoragemgmt-1.10.1-1.el9.x86_64 56/113 2026-03-06T13:50:44.075 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : flexiblas-netlib-3.0.4-9.el9.x86_64 52/113 2026-03-06T13:50:44.076 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : python3-numpy-1:1.23.5-2.el9.x86_64 51/113 2026-03-06T13:50:44.079 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : flexiblas-netlib-3.0.4-9.el9.x86_64 52/113 2026-03-06T13:50:44.081 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : flexiblas-openblas-openmp-3.0.4-9.el9.x86_64 53/113 2026-03-06T13:50:44.083 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : openblas-openmp-0.3.29-1.el9.x86_64 54/113 2026-03-06T13:50:44.084 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : flexiblas-openblas-openmp-3.0.4-9.el9.x86_64 53/113 2026-03-06T13:50:44.086 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : openblas-openmp-0.3.29-1.el9.x86_64 54/113 2026-03-06T13:50:44.086 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : libgfortran-11.5.0-14.el9.x86_64 55/113 2026-03-06T13:50:44.089 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : python3-libstoragemgmt-1.10.1-1.el9.x86_64 56/113 2026-03-06T13:50:44.090 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : libgfortran-11.5.0-14.el9.x86_64 55/113 2026-03-06T13:50:44.092 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : python3-libstoragemgmt-1.10.1-1.el9.x86_64 56/113 2026-03-06T13:50:44.095 INFO:teuthology.orchestra.run.vm02.stdout: Running scriptlet: ceph-immutable-object-cache-2:19.2.3-47.gc24117f 57/113 2026-03-06T13:50:44.095 INFO:teuthology.orchestra.run.vm02.stdout:Glob pattern passed to enable, but globs are not supported for this. 2026-03-06T13:50:44.095 INFO:teuthology.orchestra.run.vm02.stdout:Invalid unit name "ceph-immutable-object-cache@*.service" escaped as "ceph-immutable-object-cache@\x2a.service". 2026-03-06T13:50:44.095 INFO:teuthology.orchestra.run.vm02.stdout: 2026-03-06T13:50:44.095 INFO:teuthology.orchestra.run.vm02.stdout: Erasing : ceph-immutable-object-cache-2:19.2.3-47.gc24117f 57/113 2026-03-06T13:50:44.103 INFO:teuthology.orchestra.run.vm02.stdout: Running scriptlet: ceph-immutable-object-cache-2:19.2.3-47.gc24117f 57/113 2026-03-06T13:50:44.104 INFO:teuthology.orchestra.run.vm02.stdout: Erasing : openblas-0.3.29-1.el9.x86_64 58/113 2026-03-06T13:50:44.106 INFO:teuthology.orchestra.run.vm02.stdout: Erasing : flexiblas-3.0.4-9.el9.x86_64 59/113 2026-03-06T13:50:44.109 INFO:teuthology.orchestra.run.vm02.stdout: Erasing : python3-ply-3.11-14.el9.noarch 60/113 2026-03-06T13:50:44.109 INFO:teuthology.orchestra.run.vm06.stdout: Running scriptlet: ceph-immutable-object-cache-2:19.2.3-47.gc24117f 57/113 2026-03-06T13:50:44.109 INFO:teuthology.orchestra.run.vm06.stdout:Glob pattern passed to enable, but globs are not supported for this. 2026-03-06T13:50:44.109 INFO:teuthology.orchestra.run.vm06.stdout:Invalid unit name "ceph-immutable-object-cache@*.service" escaped as "ceph-immutable-object-cache@\x2a.service". 2026-03-06T13:50:44.109 INFO:teuthology.orchestra.run.vm06.stdout: 2026-03-06T13:50:44.110 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : ceph-immutable-object-cache-2:19.2.3-47.gc24117f 57/113 2026-03-06T13:50:44.111 INFO:teuthology.orchestra.run.vm04.stdout: Running scriptlet: ceph-immutable-object-cache-2:19.2.3-47.gc24117f 57/113 2026-03-06T13:50:44.111 INFO:teuthology.orchestra.run.vm04.stdout:Glob pattern passed to enable, but globs are not supported for this. 2026-03-06T13:50:44.111 INFO:teuthology.orchestra.run.vm04.stdout:Invalid unit name "ceph-immutable-object-cache@*.service" escaped as "ceph-immutable-object-cache@\x2a.service". 2026-03-06T13:50:44.111 INFO:teuthology.orchestra.run.vm04.stdout: 2026-03-06T13:50:44.111 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : ceph-immutable-object-cache-2:19.2.3-47.gc24117f 57/113 2026-03-06T13:50:44.112 INFO:teuthology.orchestra.run.vm02.stdout: Erasing : python3-idna-2.10-7.el9.1.noarch 61/113 2026-03-06T13:50:44.117 INFO:teuthology.orchestra.run.vm02.stdout: Erasing : python3-pysocks-1.7.1-12.el9.noarch 62/113 2026-03-06T13:50:44.117 INFO:teuthology.orchestra.run.vm06.stdout: Running scriptlet: ceph-immutable-object-cache-2:19.2.3-47.gc24117f 57/113 2026-03-06T13:50:44.119 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : openblas-0.3.29-1.el9.x86_64 58/113 2026-03-06T13:50:44.119 INFO:teuthology.orchestra.run.vm04.stdout: Running scriptlet: ceph-immutable-object-cache-2:19.2.3-47.gc24117f 57/113 2026-03-06T13:50:44.121 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : openblas-0.3.29-1.el9.x86_64 58/113 2026-03-06T13:50:44.121 INFO:teuthology.orchestra.run.vm02.stdout: Erasing : python3-pyasn1-0.4.8-7.el9.noarch 63/113 2026-03-06T13:50:44.121 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : flexiblas-3.0.4-9.el9.x86_64 59/113 2026-03-06T13:50:44.123 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : flexiblas-3.0.4-9.el9.x86_64 59/113 2026-03-06T13:50:44.124 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : python3-ply-3.11-14.el9.noarch 60/113 2026-03-06T13:50:44.125 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : python3-ply-3.11-14.el9.noarch 60/113 2026-03-06T13:50:44.126 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : python3-idna-2.10-7.el9.1.noarch 61/113 2026-03-06T13:50:44.126 INFO:teuthology.orchestra.run.vm02.stdout: Erasing : python3-cachetools-4.2.4-1.el9.noarch 64/113 2026-03-06T13:50:44.128 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : python3-idna-2.10-7.el9.1.noarch 61/113 2026-03-06T13:50:44.131 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : python3-pysocks-1.7.1-12.el9.noarch 62/113 2026-03-06T13:50:44.133 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : python3-pysocks-1.7.1-12.el9.noarch 62/113 2026-03-06T13:50:44.134 INFO:teuthology.orchestra.run.vm02.stdout: Erasing : python3-chardet-4.0.0-5.el9.noarch 65/113 2026-03-06T13:50:44.134 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : python3-pyasn1-0.4.8-7.el9.noarch 63/113 2026-03-06T13:50:44.137 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : python3-pyasn1-0.4.8-7.el9.noarch 63/113 2026-03-06T13:50:44.139 INFO:teuthology.orchestra.run.vm02.stdout: Erasing : python3-oauthlib-3.1.1-5.el9.noarch 66/113 2026-03-06T13:50:44.140 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : python3-cachetools-4.2.4-1.el9.noarch 64/113 2026-03-06T13:50:44.142 INFO:teuthology.orchestra.run.vm02.stdout: Erasing : python3-websocket-client-1.2.3-2.el9.noarch 67/113 2026-03-06T13:50:44.143 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : python3-cachetools-4.2.4-1.el9.noarch 64/113 2026-03-06T13:50:44.145 INFO:teuthology.orchestra.run.vm02.stdout: Erasing : python3-certifi-2023.05.07-4.el9.noarch 68/113 2026-03-06T13:50:44.147 INFO:teuthology.orchestra.run.vm02.stdout: Erasing : python3-repoze-lru-0.7-16.el9.noarch 69/113 2026-03-06T13:50:44.147 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : python3-chardet-4.0.0-5.el9.noarch 65/113 2026-03-06T13:50:44.149 INFO:teuthology.orchestra.run.vm02.stdout: Erasing : python3-jaraco-8.2.1-3.el9.noarch 70/113 2026-03-06T13:50:44.151 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : python3-chardet-4.0.0-5.el9.noarch 65/113 2026-03-06T13:50:44.151 INFO:teuthology.orchestra.run.vm02.stdout: Erasing : python3-more-itertools-8.12.0-2.el9.noarch 71/113 2026-03-06T13:50:44.152 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : python3-oauthlib-3.1.1-5.el9.noarch 66/113 2026-03-06T13:50:44.154 INFO:teuthology.orchestra.run.vm02.stdout: Erasing : python3-toml-0.10.2-6.el9.noarch 72/113 2026-03-06T13:50:44.155 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : python3-websocket-client-1.2.3-2.el9.noarch 67/113 2026-03-06T13:50:44.156 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : python3-oauthlib-3.1.1-5.el9.noarch 66/113 2026-03-06T13:50:44.157 INFO:teuthology.orchestra.run.vm02.stdout: Erasing : python3-pytz-2021.1-5.el9.noarch 73/113 2026-03-06T13:50:44.157 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : python3-certifi-2023.05.07-4.el9.noarch 68/113 2026-03-06T13:50:44.159 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : python3-websocket-client-1.2.3-2.el9.noarch 67/113 2026-03-06T13:50:44.159 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : python3-repoze-lru-0.7-16.el9.noarch 69/113 2026-03-06T13:50:44.162 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : python3-jaraco-8.2.1-3.el9.noarch 70/113 2026-03-06T13:50:44.162 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : python3-certifi-2023.05.07-4.el9.noarch 68/113 2026-03-06T13:50:44.164 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : python3-repoze-lru-0.7-16.el9.noarch 69/113 2026-03-06T13:50:44.164 INFO:teuthology.orchestra.run.vm02.stdout: Erasing : python3-backports-tarfile-1.2.0-1.el9.noarch 74/113 2026-03-06T13:50:44.164 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : python3-more-itertools-8.12.0-2.el9.noarch 71/113 2026-03-06T13:50:44.166 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : python3-jaraco-8.2.1-3.el9.noarch 70/113 2026-03-06T13:50:44.167 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : python3-toml-0.10.2-6.el9.noarch 72/113 2026-03-06T13:50:44.168 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : python3-more-itertools-8.12.0-2.el9.noarch 71/113 2026-03-06T13:50:44.168 INFO:teuthology.orchestra.run.vm02.stdout: Erasing : python3-devel-3.9.25-3.el9.x86_64 75/113 2026-03-06T13:50:44.169 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : python3-pytz-2021.1-5.el9.noarch 73/113 2026-03-06T13:50:44.170 INFO:teuthology.orchestra.run.vm02.stdout: Erasing : python3-jsonpointer-2.0-4.el9.noarch 76/113 2026-03-06T13:50:44.171 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : python3-toml-0.10.2-6.el9.noarch 72/113 2026-03-06T13:50:44.173 INFO:teuthology.orchestra.run.vm02.stdout: Erasing : python3-typing-extensions-4.15.0-1.el9.noarch 77/113 2026-03-06T13:50:44.175 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : python3-pytz-2021.1-5.el9.noarch 73/113 2026-03-06T13:50:44.176 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : python3-backports-tarfile-1.2.0-1.el9.noarch 74/113 2026-03-06T13:50:44.178 INFO:teuthology.orchestra.run.vm02.stdout: Erasing : python3-logutils-0.3.5-21.el9.noarch 78/113 2026-03-06T13:50:44.181 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : python3-devel-3.9.25-3.el9.x86_64 75/113 2026-03-06T13:50:44.182 INFO:teuthology.orchestra.run.vm02.stdout: Erasing : python3-webob-1.8.8-2.el9.noarch 79/113 2026-03-06T13:50:44.182 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : python3-jsonpointer-2.0-4.el9.noarch 76/113 2026-03-06T13:50:44.183 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : python3-backports-tarfile-1.2.0-1.el9.noarch 74/113 2026-03-06T13:50:44.185 INFO:teuthology.orchestra.run.vm02.stdout: Erasing : python3-isodate-0.6.1-3.el9.noarch 80/113 2026-03-06T13:50:44.185 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : python3-typing-extensions-4.15.0-1.el9.noarch 77/113 2026-03-06T13:50:44.187 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : python3-devel-3.9.25-3.el9.x86_64 75/113 2026-03-06T13:50:44.188 INFO:teuthology.orchestra.run.vm02.stdout: Erasing : python3-autocommand-2.2.2-8.el9.noarch 81/113 2026-03-06T13:50:44.190 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : python3-jsonpointer-2.0-4.el9.noarch 76/113 2026-03-06T13:50:44.190 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : python3-logutils-0.3.5-21.el9.noarch 78/113 2026-03-06T13:50:44.191 INFO:teuthology.orchestra.run.vm02.stdout: Erasing : python3-packaging-20.9-5.el9.noarch 82/113 2026-03-06T13:50:44.193 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : python3-typing-extensions-4.15.0-1.el9.noarch 77/113 2026-03-06T13:50:44.194 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : python3-webob-1.8.8-2.el9.noarch 79/113 2026-03-06T13:50:44.196 INFO:teuthology.orchestra.run.vm02.stdout: Erasing : grpc-data-1.46.7-10.el9.noarch 83/113 2026-03-06T13:50:44.197 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : python3-isodate-0.6.1-3.el9.noarch 80/113 2026-03-06T13:50:44.198 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : python3-logutils-0.3.5-21.el9.noarch 78/113 2026-03-06T13:50:44.199 INFO:teuthology.orchestra.run.vm02.stdout: Erasing : python3-protobuf-3.14.0-17.el9.noarch 84/113 2026-03-06T13:50:44.200 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : python3-autocommand-2.2.2-8.el9.noarch 81/113 2026-03-06T13:50:44.202 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : python3-webob-1.8.8-2.el9.noarch 79/113 2026-03-06T13:50:44.202 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : python3-packaging-20.9-5.el9.noarch 82/113 2026-03-06T13:50:44.203 INFO:teuthology.orchestra.run.vm02.stdout: Erasing : python3-zc-lockfile-2.0-10.el9.noarch 85/113 2026-03-06T13:50:44.204 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : python3-isodate-0.6.1-3.el9.noarch 80/113 2026-03-06T13:50:44.205 INFO:teuthology.orchestra.run.vm02.stdout: Erasing : python3-natsort-7.1.1-5.el9.noarch 86/113 2026-03-06T13:50:44.206 INFO:teuthology.orchestra.run.vm02.stdout: Erasing : ceph-grafana-dashboards-2:19.2.3-47.gc24117fd552 87/113 2026-03-06T13:50:44.208 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : grpc-data-1.46.7-10.el9.noarch 83/113 2026-03-06T13:50:44.208 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : python3-autocommand-2.2.2-8.el9.noarch 81/113 2026-03-06T13:50:44.210 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : python3-packaging-20.9-5.el9.noarch 82/113 2026-03-06T13:50:44.211 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : python3-protobuf-3.14.0-17.el9.noarch 84/113 2026-03-06T13:50:44.212 INFO:teuthology.orchestra.run.vm02.stdout: Erasing : ceph-prometheus-alerts-2:19.2.3-47.gc24117fd552. 88/113 2026-03-06T13:50:44.214 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : python3-zc-lockfile-2.0-10.el9.noarch 85/113 2026-03-06T13:50:44.215 INFO:teuthology.orchestra.run.vm02.stdout: Erasing : python3-werkzeug-2.0.3-3.el9.1.noarch 89/113 2026-03-06T13:50:44.216 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : grpc-data-1.46.7-10.el9.noarch 83/113 2026-03-06T13:50:44.217 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : python3-natsort-7.1.1-5.el9.noarch 86/113 2026-03-06T13:50:44.218 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : ceph-grafana-dashboards-2:19.2.3-47.gc24117fd552 87/113 2026-03-06T13:50:44.219 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : python3-protobuf-3.14.0-17.el9.noarch 84/113 2026-03-06T13:50:44.222 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : python3-zc-lockfile-2.0-10.el9.noarch 85/113 2026-03-06T13:50:44.223 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : ceph-prometheus-alerts-2:19.2.3-47.gc24117fd552. 88/113 2026-03-06T13:50:44.225 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : python3-natsort-7.1.1-5.el9.noarch 86/113 2026-03-06T13:50:44.226 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : ceph-grafana-dashboards-2:19.2.3-47.gc24117fd552 87/113 2026-03-06T13:50:44.227 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : python3-werkzeug-2.0.3-3.el9.1.noarch 89/113 2026-03-06T13:50:44.232 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : ceph-prometheus-alerts-2:19.2.3-47.gc24117fd552. 88/113 2026-03-06T13:50:44.234 INFO:teuthology.orchestra.run.vm02.stdout: Running scriptlet: ceph-base-2:19.2.3-47.gc24117fd552.el9.clyso.x86 90/113 2026-03-06T13:50:44.234 INFO:teuthology.orchestra.run.vm02.stdout:Removed "/etc/systemd/system/ceph.target.wants/ceph-crash.service". 2026-03-06T13:50:44.234 INFO:teuthology.orchestra.run.vm02.stdout: 2026-03-06T13:50:44.236 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : python3-werkzeug-2.0.3-3.el9.1.noarch 89/113 2026-03-06T13:50:44.242 INFO:teuthology.orchestra.run.vm02.stdout: Erasing : ceph-base-2:19.2.3-47.gc24117fd552.el9.clyso.x86 90/113 2026-03-06T13:50:44.245 INFO:teuthology.orchestra.run.vm06.stdout: Running scriptlet: ceph-base-2:19.2.3-47.gc24117fd552.el9.clyso.x86 90/113 2026-03-06T13:50:44.245 INFO:teuthology.orchestra.run.vm06.stdout:Removed "/etc/systemd/system/ceph.target.wants/ceph-crash.service". 2026-03-06T13:50:44.245 INFO:teuthology.orchestra.run.vm06.stdout: 2026-03-06T13:50:44.252 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : ceph-base-2:19.2.3-47.gc24117fd552.el9.clyso.x86 90/113 2026-03-06T13:50:44.254 INFO:teuthology.orchestra.run.vm04.stdout: Running scriptlet: ceph-base-2:19.2.3-47.gc24117fd552.el9.clyso.x86 90/113 2026-03-06T13:50:44.254 INFO:teuthology.orchestra.run.vm04.stdout:Removed "/etc/systemd/system/ceph.target.wants/ceph-crash.service". 2026-03-06T13:50:44.254 INFO:teuthology.orchestra.run.vm04.stdout: 2026-03-06T13:50:44.262 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : ceph-base-2:19.2.3-47.gc24117fd552.el9.clyso.x86 90/113 2026-03-06T13:50:44.270 INFO:teuthology.orchestra.run.vm02.stdout: Running scriptlet: ceph-base-2:19.2.3-47.gc24117fd552.el9.clyso.x86 90/113 2026-03-06T13:50:44.270 INFO:teuthology.orchestra.run.vm02.stdout: Erasing : ceph-common-2:19.2.3-47.gc24117fd552.el9.clyso.x 91/113 2026-03-06T13:50:44.278 INFO:teuthology.orchestra.run.vm06.stdout: Running scriptlet: ceph-base-2:19.2.3-47.gc24117fd552.el9.clyso.x86 90/113 2026-03-06T13:50:44.278 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : ceph-common-2:19.2.3-47.gc24117fd552.el9.clyso.x 91/113 2026-03-06T13:50:44.281 INFO:teuthology.orchestra.run.vm02.stdout: Running scriptlet: ceph-common-2:19.2.3-47.gc24117fd552.el9.clyso.x 91/113 2026-03-06T13:50:44.286 INFO:teuthology.orchestra.run.vm02.stdout: Erasing : qatzip-libs-1.3.1-1.el9.x86_64 92/113 2026-03-06T13:50:44.288 INFO:teuthology.orchestra.run.vm06.stdout: Running scriptlet: ceph-common-2:19.2.3-47.gc24117fd552.el9.clyso.x 91/113 2026-03-06T13:50:44.288 INFO:teuthology.orchestra.run.vm02.stdout: Erasing : python3-ceph-common-2:19.2.3-47.gc24117fd552.el9 93/113 2026-03-06T13:50:44.289 INFO:teuthology.orchestra.run.vm04.stdout: Running scriptlet: ceph-base-2:19.2.3-47.gc24117fd552.el9.clyso.x86 90/113 2026-03-06T13:50:44.289 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : ceph-common-2:19.2.3-47.gc24117fd552.el9.clyso.x 91/113 2026-03-06T13:50:44.290 INFO:teuthology.orchestra.run.vm02.stdout: Erasing : python3-prettytable-0.7.2-27.el9.noarch 94/113 2026-03-06T13:50:44.290 INFO:teuthology.orchestra.run.vm02.stdout: Erasing : ceph-selinux-2:19.2.3-47.gc24117fd552.el9.clyso. 95/113 2026-03-06T13:50:44.293 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : qatzip-libs-1.3.1-1.el9.x86_64 92/113 2026-03-06T13:50:44.296 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : python3-ceph-common-2:19.2.3-47.gc24117fd552.el9 93/113 2026-03-06T13:50:44.298 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : python3-prettytable-0.7.2-27.el9.noarch 94/113 2026-03-06T13:50:44.298 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : ceph-selinux-2:19.2.3-47.gc24117fd552.el9.clyso. 95/113 2026-03-06T13:50:44.300 INFO:teuthology.orchestra.run.vm04.stdout: Running scriptlet: ceph-common-2:19.2.3-47.gc24117fd552.el9.clyso.x 91/113 2026-03-06T13:50:44.305 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : qatzip-libs-1.3.1-1.el9.x86_64 92/113 2026-03-06T13:50:44.308 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : python3-ceph-common-2:19.2.3-47.gc24117fd552.el9 93/113 2026-03-06T13:50:44.310 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : python3-prettytable-0.7.2-27.el9.noarch 94/113 2026-03-06T13:50:44.310 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : ceph-selinux-2:19.2.3-47.gc24117fd552.el9.clyso. 95/113 2026-03-06T13:50:49.766 INFO:teuthology.orchestra.run.vm06.stdout: Running scriptlet: ceph-selinux-2:19.2.3-47.gc24117fd552.el9.clyso. 95/113 2026-03-06T13:50:49.766 INFO:teuthology.orchestra.run.vm06.stdout:skipping the directory /sys 2026-03-06T13:50:49.766 INFO:teuthology.orchestra.run.vm06.stdout:skipping the directory /proc 2026-03-06T13:50:49.766 INFO:teuthology.orchestra.run.vm06.stdout:skipping the directory /mnt 2026-03-06T13:50:49.766 INFO:teuthology.orchestra.run.vm06.stdout:skipping the directory /var/tmp 2026-03-06T13:50:49.766 INFO:teuthology.orchestra.run.vm06.stdout:skipping the directory /home 2026-03-06T13:50:49.766 INFO:teuthology.orchestra.run.vm06.stdout:skipping the directory /root 2026-03-06T13:50:49.766 INFO:teuthology.orchestra.run.vm06.stdout:skipping the directory /tmp 2026-03-06T13:50:49.766 INFO:teuthology.orchestra.run.vm06.stdout: 2026-03-06T13:50:49.775 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : qatlib-25.08.0-2.el9.x86_64 96/113 2026-03-06T13:50:49.795 INFO:teuthology.orchestra.run.vm06.stdout: Running scriptlet: qatlib-service-25.08.0-2.el9.x86_64 97/113 2026-03-06T13:50:49.795 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : qatlib-service-25.08.0-2.el9.x86_64 97/113 2026-03-06T13:50:49.803 INFO:teuthology.orchestra.run.vm06.stdout: Running scriptlet: qatlib-service-25.08.0-2.el9.x86_64 97/113 2026-03-06T13:50:49.806 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : gperftools-libs-2.9.1-3.el9.x86_64 98/113 2026-03-06T13:50:49.808 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : libunwind-1.6.2-1.el9.x86_64 99/113 2026-03-06T13:50:49.811 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : pciutils-3.7.0-7.el9.x86_64 100/113 2026-03-06T13:50:49.813 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : liboath-2.6.12-1.el9.x86_64 101/113 2026-03-06T13:50:49.813 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : libradosstriper1-2:19.2.3-47.gc24117fd552.el9.cl 102/113 2026-03-06T13:50:49.829 INFO:teuthology.orchestra.run.vm06.stdout: Running scriptlet: libradosstriper1-2:19.2.3-47.gc24117fd552.el9.cl 102/113 2026-03-06T13:50:49.830 INFO:teuthology.orchestra.run.vm02.stdout: Running scriptlet: ceph-selinux-2:19.2.3-47.gc24117fd552.el9.clyso. 95/113 2026-03-06T13:50:49.830 INFO:teuthology.orchestra.run.vm02.stdout:skipping the directory /sys 2026-03-06T13:50:49.830 INFO:teuthology.orchestra.run.vm02.stdout:skipping the directory /proc 2026-03-06T13:50:49.830 INFO:teuthology.orchestra.run.vm02.stdout:skipping the directory /mnt 2026-03-06T13:50:49.830 INFO:teuthology.orchestra.run.vm02.stdout:skipping the directory /var/tmp 2026-03-06T13:50:49.830 INFO:teuthology.orchestra.run.vm02.stdout:skipping the directory /home 2026-03-06T13:50:49.830 INFO:teuthology.orchestra.run.vm02.stdout:skipping the directory /root 2026-03-06T13:50:49.830 INFO:teuthology.orchestra.run.vm02.stdout:skipping the directory /tmp 2026-03-06T13:50:49.830 INFO:teuthology.orchestra.run.vm02.stdout: 2026-03-06T13:50:49.839 INFO:teuthology.orchestra.run.vm02.stdout: Erasing : qatlib-25.08.0-2.el9.x86_64 96/113 2026-03-06T13:50:49.844 INFO:teuthology.orchestra.run.vm06.stdout: Running scriptlet: smartmontools-1:7.2-10.el9.x86_64 103/113 2026-03-06T13:50:49.844 INFO:teuthology.orchestra.run.vm06.stdout:Removed "/etc/systemd/system/multi-user.target.wants/smartd.service". 2026-03-06T13:50:49.844 INFO:teuthology.orchestra.run.vm06.stdout: 2026-03-06T13:50:49.846 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : smartmontools-1:7.2-10.el9.x86_64 103/113 2026-03-06T13:50:49.855 INFO:teuthology.orchestra.run.vm02.stdout: Running scriptlet: qatlib-service-25.08.0-2.el9.x86_64 97/113 2026-03-06T13:50:49.856 INFO:teuthology.orchestra.run.vm02.stdout: Erasing : qatlib-service-25.08.0-2.el9.x86_64 97/113 2026-03-06T13:50:49.857 INFO:teuthology.orchestra.run.vm06.stdout: Running scriptlet: smartmontools-1:7.2-10.el9.x86_64 103/113 2026-03-06T13:50:49.859 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : ledmon-libs-1.1.0-3.el9.x86_64 104/113 2026-03-06T13:50:49.861 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : libquadmath-11.5.0-14.el9.x86_64 105/113 2026-03-06T13:50:49.863 INFO:teuthology.orchestra.run.vm02.stdout: Running scriptlet: qatlib-service-25.08.0-2.el9.x86_64 97/113 2026-03-06T13:50:49.864 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : python3-markupsafe-1.1.1-12.el9.x86_64 106/113 2026-03-06T13:50:49.866 INFO:teuthology.orchestra.run.vm02.stdout: Erasing : gperftools-libs-2.9.1-3.el9.x86_64 98/113 2026-03-06T13:50:49.866 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : protobuf-3.14.0-17.el9.x86_64 107/113 2026-03-06T13:50:49.869 INFO:teuthology.orchestra.run.vm02.stdout: Erasing : libunwind-1.6.2-1.el9.x86_64 99/113 2026-03-06T13:50:49.869 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : libxslt-1.1.34-12.el9.x86_64 108/113 2026-03-06T13:50:49.871 INFO:teuthology.orchestra.run.vm02.stdout: Erasing : pciutils-3.7.0-7.el9.x86_64 100/113 2026-03-06T13:50:49.873 INFO:teuthology.orchestra.run.vm02.stdout: Erasing : liboath-2.6.12-1.el9.x86_64 101/113 2026-03-06T13:50:49.873 INFO:teuthology.orchestra.run.vm02.stdout: Erasing : libradosstriper1-2:19.2.3-47.gc24117fd552.el9.cl 102/113 2026-03-06T13:50:49.875 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : libconfig-1.7.2-9.el9.x86_64 109/113 2026-03-06T13:50:49.884 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : cryptsetup-2.8.1-3.el9.x86_64 110/113 2026-03-06T13:50:49.888 INFO:teuthology.orchestra.run.vm02.stdout: Running scriptlet: libradosstriper1-2:19.2.3-47.gc24117fd552.el9.cl 102/113 2026-03-06T13:50:49.889 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : abseil-cpp-20211102.0-4.el9.x86_64 111/113 2026-03-06T13:50:49.891 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : python3-msgpack-1.0.3-2.el9.x86_64 112/113 2026-03-06T13:50:49.891 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : libcephsqlite-2:19.2.3-47.gc24117fd552.el9.clyso 113/113 2026-03-06T13:50:49.901 INFO:teuthology.orchestra.run.vm02.stdout: Running scriptlet: smartmontools-1:7.2-10.el9.x86_64 103/113 2026-03-06T13:50:49.901 INFO:teuthology.orchestra.run.vm02.stdout:Removed "/etc/systemd/system/multi-user.target.wants/smartd.service". 2026-03-06T13:50:49.901 INFO:teuthology.orchestra.run.vm02.stdout: 2026-03-06T13:50:49.903 INFO:teuthology.orchestra.run.vm02.stdout: Erasing : smartmontools-1:7.2-10.el9.x86_64 103/113 2026-03-06T13:50:49.904 INFO:teuthology.orchestra.run.vm04.stdout: Running scriptlet: ceph-selinux-2:19.2.3-47.gc24117fd552.el9.clyso. 95/113 2026-03-06T13:50:49.905 INFO:teuthology.orchestra.run.vm04.stdout:skipping the directory /sys 2026-03-06T13:50:49.905 INFO:teuthology.orchestra.run.vm04.stdout:skipping the directory /proc 2026-03-06T13:50:49.905 INFO:teuthology.orchestra.run.vm04.stdout:skipping the directory /mnt 2026-03-06T13:50:49.905 INFO:teuthology.orchestra.run.vm04.stdout:skipping the directory /var/tmp 2026-03-06T13:50:49.905 INFO:teuthology.orchestra.run.vm04.stdout:skipping the directory /home 2026-03-06T13:50:49.905 INFO:teuthology.orchestra.run.vm04.stdout:skipping the directory /root 2026-03-06T13:50:49.905 INFO:teuthology.orchestra.run.vm04.stdout:skipping the directory /tmp 2026-03-06T13:50:49.905 INFO:teuthology.orchestra.run.vm04.stdout: 2026-03-06T13:50:49.913 INFO:teuthology.orchestra.run.vm02.stdout: Running scriptlet: smartmontools-1:7.2-10.el9.x86_64 103/113 2026-03-06T13:50:49.913 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : qatlib-25.08.0-2.el9.x86_64 96/113 2026-03-06T13:50:49.915 INFO:teuthology.orchestra.run.vm02.stdout: Erasing : ledmon-libs-1.1.0-3.el9.x86_64 104/113 2026-03-06T13:50:49.917 INFO:teuthology.orchestra.run.vm02.stdout: Erasing : libquadmath-11.5.0-14.el9.x86_64 105/113 2026-03-06T13:50:49.920 INFO:teuthology.orchestra.run.vm02.stdout: Erasing : python3-markupsafe-1.1.1-12.el9.x86_64 106/113 2026-03-06T13:50:49.923 INFO:teuthology.orchestra.run.vm02.stdout: Erasing : protobuf-3.14.0-17.el9.x86_64 107/113 2026-03-06T13:50:49.925 INFO:teuthology.orchestra.run.vm02.stdout: Erasing : libxslt-1.1.34-12.el9.x86_64 108/113 2026-03-06T13:50:49.931 INFO:teuthology.orchestra.run.vm02.stdout: Erasing : libconfig-1.7.2-9.el9.x86_64 109/113 2026-03-06T13:50:49.932 INFO:teuthology.orchestra.run.vm04.stdout: Running scriptlet: qatlib-service-25.08.0-2.el9.x86_64 97/113 2026-03-06T13:50:49.932 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : qatlib-service-25.08.0-2.el9.x86_64 97/113 2026-03-06T13:50:49.938 INFO:teuthology.orchestra.run.vm02.stdout: Erasing : cryptsetup-2.8.1-3.el9.x86_64 110/113 2026-03-06T13:50:49.941 INFO:teuthology.orchestra.run.vm04.stdout: Running scriptlet: qatlib-service-25.08.0-2.el9.x86_64 97/113 2026-03-06T13:50:49.944 INFO:teuthology.orchestra.run.vm02.stdout: Erasing : abseil-cpp-20211102.0-4.el9.x86_64 111/113 2026-03-06T13:50:49.944 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : gperftools-libs-2.9.1-3.el9.x86_64 98/113 2026-03-06T13:50:49.947 INFO:teuthology.orchestra.run.vm02.stdout: Erasing : python3-msgpack-1.0.3-2.el9.x86_64 112/113 2026-03-06T13:50:49.947 INFO:teuthology.orchestra.run.vm02.stdout: Erasing : libcephsqlite-2:19.2.3-47.gc24117fd552.el9.clyso 113/113 2026-03-06T13:50:49.947 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : libunwind-1.6.2-1.el9.x86_64 99/113 2026-03-06T13:50:49.950 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : pciutils-3.7.0-7.el9.x86_64 100/113 2026-03-06T13:50:49.952 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : liboath-2.6.12-1.el9.x86_64 101/113 2026-03-06T13:50:49.952 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : libradosstriper1-2:19.2.3-47.gc24117fd552.el9.cl 102/113 2026-03-06T13:50:49.966 INFO:teuthology.orchestra.run.vm04.stdout: Running scriptlet: libradosstriper1-2:19.2.3-47.gc24117fd552.el9.cl 102/113 2026-03-06T13:50:49.981 INFO:teuthology.orchestra.run.vm04.stdout: Running scriptlet: smartmontools-1:7.2-10.el9.x86_64 103/113 2026-03-06T13:50:49.981 INFO:teuthology.orchestra.run.vm04.stdout:Removed "/etc/systemd/system/multi-user.target.wants/smartd.service". 2026-03-06T13:50:49.981 INFO:teuthology.orchestra.run.vm04.stdout: 2026-03-06T13:50:49.983 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : smartmontools-1:7.2-10.el9.x86_64 103/113 2026-03-06T13:50:49.988 INFO:teuthology.orchestra.run.vm06.stdout: Running scriptlet: libcephsqlite-2:19.2.3-47.gc24117fd552.el9.clyso 113/113 2026-03-06T13:50:49.988 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : abseil-cpp-20211102.0-4.el9.x86_64 1/113 2026-03-06T13:50:49.988 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : ceph-base-2:19.2.3-47.gc24117fd552.el9.clyso.x86 2/113 2026-03-06T13:50:49.988 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : ceph-common-2:19.2.3-47.gc24117fd552.el9.clyso.x 3/113 2026-03-06T13:50:49.988 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : ceph-grafana-dashboards-2:19.2.3-47.gc24117fd552 4/113 2026-03-06T13:50:49.988 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : ceph-immutable-object-cache-2:19.2.3-47.gc24117f 5/113 2026-03-06T13:50:49.988 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : ceph-mgr-2:19.2.3-47.gc24117fd552.el9.clyso.x86_ 6/113 2026-03-06T13:50:49.988 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : ceph-mgr-cephadm-2:19.2.3-47.gc24117fd552.el9.cl 7/113 2026-03-06T13:50:49.988 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : ceph-mgr-dashboard-2:19.2.3-47.gc24117fd552.el9. 8/113 2026-03-06T13:50:49.988 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : ceph-mgr-diskprediction-local-2:19.2.3-47.gc2411 9/113 2026-03-06T13:50:49.989 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : ceph-mgr-k8sevents-2:19.2.3-47.gc24117fd552.el9. 10/113 2026-03-06T13:50:49.989 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : ceph-mgr-modules-core-2:19.2.3-47.gc24117fd552.e 11/113 2026-03-06T13:50:49.989 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : ceph-mgr-rook-2:19.2.3-47.gc24117fd552.el9.clyso 12/113 2026-03-06T13:50:49.989 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : ceph-osd-2:19.2.3-47.gc24117fd552.el9.clyso.x86_ 13/113 2026-03-06T13:50:49.989 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : ceph-prometheus-alerts-2:19.2.3-47.gc24117fd552. 14/113 2026-03-06T13:50:49.989 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : ceph-selinux-2:19.2.3-47.gc24117fd552.el9.clyso. 15/113 2026-03-06T13:50:49.989 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : ceph-volume-2:19.2.3-47.gc24117fd552.el9.clyso.n 16/113 2026-03-06T13:50:49.989 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : cryptsetup-2.8.1-3.el9.x86_64 17/113 2026-03-06T13:50:49.989 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : flexiblas-3.0.4-9.el9.x86_64 18/113 2026-03-06T13:50:49.989 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : flexiblas-netlib-3.0.4-9.el9.x86_64 19/113 2026-03-06T13:50:49.989 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : flexiblas-openblas-openmp-3.0.4-9.el9.x86_64 20/113 2026-03-06T13:50:49.989 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : gperftools-libs-2.9.1-3.el9.x86_64 21/113 2026-03-06T13:50:49.989 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : grpc-data-1.46.7-10.el9.noarch 22/113 2026-03-06T13:50:49.989 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : ledmon-libs-1.1.0-3.el9.x86_64 23/113 2026-03-06T13:50:49.989 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : libcephsqlite-2:19.2.3-47.gc24117fd552.el9.clyso 24/113 2026-03-06T13:50:49.989 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : libconfig-1.7.2-9.el9.x86_64 25/113 2026-03-06T13:50:49.989 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : libgfortran-11.5.0-14.el9.x86_64 26/113 2026-03-06T13:50:49.989 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : liboath-2.6.12-1.el9.x86_64 27/113 2026-03-06T13:50:49.989 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : libquadmath-11.5.0-14.el9.x86_64 28/113 2026-03-06T13:50:49.989 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : libradosstriper1-2:19.2.3-47.gc24117fd552.el9.cl 29/113 2026-03-06T13:50:49.989 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : libstoragemgmt-1.10.1-1.el9.x86_64 30/113 2026-03-06T13:50:49.989 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : libunwind-1.6.2-1.el9.x86_64 31/113 2026-03-06T13:50:49.989 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : libxslt-1.1.34-12.el9.x86_64 32/113 2026-03-06T13:50:49.989 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : openblas-0.3.29-1.el9.x86_64 33/113 2026-03-06T13:50:49.989 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : openblas-openmp-0.3.29-1.el9.x86_64 34/113 2026-03-06T13:50:49.989 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : pciutils-3.7.0-7.el9.x86_64 35/113 2026-03-06T13:50:49.989 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : protobuf-3.14.0-17.el9.x86_64 36/113 2026-03-06T13:50:49.989 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : protobuf-compiler-3.14.0-17.el9.x86_64 37/113 2026-03-06T13:50:49.990 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-asyncssh-2.13.2-5.el9.noarch 38/113 2026-03-06T13:50:49.990 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-autocommand-2.2.2-8.el9.noarch 39/113 2026-03-06T13:50:49.990 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-babel-2.9.1-2.el9.noarch 40/113 2026-03-06T13:50:49.990 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-backports-tarfile-1.2.0-1.el9.noarch 41/113 2026-03-06T13:50:49.990 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-bcrypt-3.2.2-1.el9.x86_64 42/113 2026-03-06T13:50:49.990 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-cachetools-4.2.4-1.el9.noarch 43/113 2026-03-06T13:50:49.991 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-ceph-common-2:19.2.3-47.gc24117fd552.el9 44/113 2026-03-06T13:50:49.991 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-certifi-2023.05.07-4.el9.noarch 45/113 2026-03-06T13:50:49.991 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-cffi-1.14.5-5.el9.x86_64 46/113 2026-03-06T13:50:49.991 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-chardet-4.0.0-5.el9.noarch 47/113 2026-03-06T13:50:49.991 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-cheroot-10.0.1-4.el9.noarch 48/113 2026-03-06T13:50:49.991 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-cherrypy-18.6.1-2.el9.noarch 49/113 2026-03-06T13:50:49.991 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-cryptography-36.0.1-5.el9.x86_64 50/113 2026-03-06T13:50:49.991 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-devel-3.9.25-3.el9.x86_64 51/113 2026-03-06T13:50:49.991 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-google-auth-1:2.45.0-1.el9.noarch 52/113 2026-03-06T13:50:49.991 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-grpcio-1.46.7-10.el9.x86_64 53/113 2026-03-06T13:50:49.991 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-grpcio-tools-1.46.7-10.el9.x86_64 54/113 2026-03-06T13:50:49.991 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-idna-2.10-7.el9.1.noarch 55/113 2026-03-06T13:50:49.991 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-influxdb-5.3.1-1.el9.noarch 56/113 2026-03-06T13:50:49.991 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-isodate-0.6.1-3.el9.noarch 57/113 2026-03-06T13:50:49.991 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-jaraco-8.2.1-3.el9.noarch 58/113 2026-03-06T13:50:49.991 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-jaraco-classes-3.2.1-5.el9.noarch 59/113 2026-03-06T13:50:49.991 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-jaraco-collections-3.0.0-8.el9.noarch 60/113 2026-03-06T13:50:49.991 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-jaraco-context-6.0.1-3.el9.noarch 61/113 2026-03-06T13:50:49.991 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-jaraco-functools-3.5.0-2.el9.noarch 62/113 2026-03-06T13:50:49.991 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-jaraco-text-4.0.0-2.el9.noarch 63/113 2026-03-06T13:50:49.991 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-jinja2-2.11.3-8.el9.noarch 64/113 2026-03-06T13:50:49.991 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-jsonpatch-1.21-16.el9.noarch 65/113 2026-03-06T13:50:49.991 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-jsonpointer-2.0-4.el9.noarch 66/113 2026-03-06T13:50:49.991 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-kubernetes-1:26.1.0-3.el9.noarch 67/113 2026-03-06T13:50:49.991 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-libstoragemgmt-1.10.1-1.el9.x86_64 68/113 2026-03-06T13:50:49.991 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-logutils-0.3.5-21.el9.noarch 69/113 2026-03-06T13:50:49.991 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-lxml-4.6.5-3.el9.x86_64 70/113 2026-03-06T13:50:49.991 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-mako-1.1.4-6.el9.noarch 71/113 2026-03-06T13:50:49.991 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-markupsafe-1.1.1-12.el9.x86_64 72/113 2026-03-06T13:50:49.991 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-more-itertools-8.12.0-2.el9.noarch 73/113 2026-03-06T13:50:49.991 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-msgpack-1.0.3-2.el9.x86_64 74/113 2026-03-06T13:50:49.991 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-natsort-7.1.1-5.el9.noarch 75/113 2026-03-06T13:50:49.991 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-numpy-1:1.23.5-2.el9.x86_64 76/113 2026-03-06T13:50:49.991 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-numpy-f2py-1:1.23.5-2.el9.x86_64 77/113 2026-03-06T13:50:49.991 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-oauthlib-3.1.1-5.el9.noarch 78/113 2026-03-06T13:50:49.991 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-packaging-20.9-5.el9.noarch 79/113 2026-03-06T13:50:49.991 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-pecan-1.4.2-3.el9.noarch 80/113 2026-03-06T13:50:49.991 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-ply-3.11-14.el9.noarch 81/113 2026-03-06T13:50:49.991 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-portend-3.1.0-2.el9.noarch 82/113 2026-03-06T13:50:49.991 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-prettytable-0.7.2-27.el9.noarch 83/113 2026-03-06T13:50:49.991 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-protobuf-3.14.0-17.el9.noarch 84/113 2026-03-06T13:50:49.991 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-pyOpenSSL-21.0.0-1.el9.noarch 85/113 2026-03-06T13:50:49.991 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-pyasn1-0.4.8-7.el9.noarch 86/113 2026-03-06T13:50:49.991 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-pyasn1-modules-0.4.8-7.el9.noarch 87/113 2026-03-06T13:50:49.991 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-pycparser-2.20-6.el9.noarch 88/113 2026-03-06T13:50:49.991 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-pysocks-1.7.1-12.el9.noarch 89/113 2026-03-06T13:50:49.992 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-pytz-2021.1-5.el9.noarch 90/113 2026-03-06T13:50:49.992 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-repoze-lru-0.7-16.el9.noarch 91/113 2026-03-06T13:50:49.992 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-requests-2.25.1-10.el9.noarch 92/113 2026-03-06T13:50:49.992 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-requests-oauthlib-1.3.0-12.el9.noarch 93/113 2026-03-06T13:50:49.992 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-routes-2.5.1-5.el9.noarch 94/113 2026-03-06T13:50:49.992 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-rsa-4.9-2.el9.noarch 95/113 2026-03-06T13:50:49.992 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-saml-1.16.0-1.el9.noarch 96/113 2026-03-06T13:50:49.992 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-scipy-1.9.3-2.el9.x86_64 97/113 2026-03-06T13:50:49.992 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-tempora-5.0.0-2.el9.noarch 98/113 2026-03-06T13:50:49.992 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-toml-0.10.2-6.el9.noarch 99/113 2026-03-06T13:50:49.992 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-typing-extensions-4.15.0-1.el9.noarch 100/113 2026-03-06T13:50:49.992 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-urllib3-1.26.5-7.el9.noarch 101/113 2026-03-06T13:50:49.992 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-webob-1.8.8-2.el9.noarch 102/113 2026-03-06T13:50:49.992 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-websocket-client-1.2.3-2.el9.noarch 103/113 2026-03-06T13:50:49.992 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-werkzeug-2.0.3-3.el9.1.noarch 104/113 2026-03-06T13:50:49.992 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-xmlsec-1.3.13-1.el9.x86_64 105/113 2026-03-06T13:50:49.992 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-zc-lockfile-2.0-10.el9.noarch 106/113 2026-03-06T13:50:49.992 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : qatlib-25.08.0-2.el9.x86_64 107/113 2026-03-06T13:50:49.992 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : qatlib-service-25.08.0-2.el9.x86_64 108/113 2026-03-06T13:50:49.992 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : qatzip-libs-1.3.1-1.el9.x86_64 109/113 2026-03-06T13:50:49.992 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : rbd-mirror-2:19.2.3-47.gc24117fd552.el9.clyso.x8 110/113 2026-03-06T13:50:49.992 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : smartmontools-1:7.2-10.el9.x86_64 111/113 2026-03-06T13:50:49.992 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : xmlsec1-1.2.29-13.el9.x86_64 112/113 2026-03-06T13:50:49.992 INFO:teuthology.orchestra.run.vm04.stdout: Running scriptlet: smartmontools-1:7.2-10.el9.x86_64 103/113 2026-03-06T13:50:49.995 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : ledmon-libs-1.1.0-3.el9.x86_64 104/113 2026-03-06T13:50:49.997 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : libquadmath-11.5.0-14.el9.x86_64 105/113 2026-03-06T13:50:50.000 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : python3-markupsafe-1.1.1-12.el9.x86_64 106/113 2026-03-06T13:50:50.002 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : protobuf-3.14.0-17.el9.x86_64 107/113 2026-03-06T13:50:50.005 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : libxslt-1.1.34-12.el9.x86_64 108/113 2026-03-06T13:50:50.010 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : libconfig-1.7.2-9.el9.x86_64 109/113 2026-03-06T13:50:50.018 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : cryptsetup-2.8.1-3.el9.x86_64 110/113 2026-03-06T13:50:50.023 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : abseil-cpp-20211102.0-4.el9.x86_64 111/113 2026-03-06T13:50:50.025 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : python3-msgpack-1.0.3-2.el9.x86_64 112/113 2026-03-06T13:50:50.025 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : libcephsqlite-2:19.2.3-47.gc24117fd552.el9.clyso 113/113 2026-03-06T13:50:50.065 INFO:teuthology.orchestra.run.vm02.stdout: Running scriptlet: libcephsqlite-2:19.2.3-47.gc24117fd552.el9.clyso 113/113 2026-03-06T13:50:50.065 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : abseil-cpp-20211102.0-4.el9.x86_64 1/113 2026-03-06T13:50:50.065 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : ceph-base-2:19.2.3-47.gc24117fd552.el9.clyso.x86 2/113 2026-03-06T13:50:50.065 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : ceph-common-2:19.2.3-47.gc24117fd552.el9.clyso.x 3/113 2026-03-06T13:50:50.065 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : ceph-grafana-dashboards-2:19.2.3-47.gc24117fd552 4/113 2026-03-06T13:50:50.065 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : ceph-immutable-object-cache-2:19.2.3-47.gc24117f 5/113 2026-03-06T13:50:50.065 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : ceph-mgr-2:19.2.3-47.gc24117fd552.el9.clyso.x86_ 6/113 2026-03-06T13:50:50.065 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : ceph-mgr-cephadm-2:19.2.3-47.gc24117fd552.el9.cl 7/113 2026-03-06T13:50:50.065 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : ceph-mgr-dashboard-2:19.2.3-47.gc24117fd552.el9. 8/113 2026-03-06T13:50:50.065 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : ceph-mgr-diskprediction-local-2:19.2.3-47.gc2411 9/113 2026-03-06T13:50:50.065 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : ceph-mgr-k8sevents-2:19.2.3-47.gc24117fd552.el9. 10/113 2026-03-06T13:50:50.065 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : ceph-mgr-modules-core-2:19.2.3-47.gc24117fd552.e 11/113 2026-03-06T13:50:50.065 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : ceph-mgr-rook-2:19.2.3-47.gc24117fd552.el9.clyso 12/113 2026-03-06T13:50:50.065 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : ceph-osd-2:19.2.3-47.gc24117fd552.el9.clyso.x86_ 13/113 2026-03-06T13:50:50.065 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : ceph-prometheus-alerts-2:19.2.3-47.gc24117fd552. 14/113 2026-03-06T13:50:50.066 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : ceph-selinux-2:19.2.3-47.gc24117fd552.el9.clyso. 15/113 2026-03-06T13:50:50.066 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : ceph-volume-2:19.2.3-47.gc24117fd552.el9.clyso.n 16/113 2026-03-06T13:50:50.066 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : cryptsetup-2.8.1-3.el9.x86_64 17/113 2026-03-06T13:50:50.066 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : flexiblas-3.0.4-9.el9.x86_64 18/113 2026-03-06T13:50:50.066 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : flexiblas-netlib-3.0.4-9.el9.x86_64 19/113 2026-03-06T13:50:50.066 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : flexiblas-openblas-openmp-3.0.4-9.el9.x86_64 20/113 2026-03-06T13:50:50.066 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : gperftools-libs-2.9.1-3.el9.x86_64 21/113 2026-03-06T13:50:50.066 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : grpc-data-1.46.7-10.el9.noarch 22/113 2026-03-06T13:50:50.066 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : ledmon-libs-1.1.0-3.el9.x86_64 23/113 2026-03-06T13:50:50.066 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : libcephsqlite-2:19.2.3-47.gc24117fd552.el9.clyso 24/113 2026-03-06T13:50:50.066 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : libconfig-1.7.2-9.el9.x86_64 25/113 2026-03-06T13:50:50.066 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : libgfortran-11.5.0-14.el9.x86_64 26/113 2026-03-06T13:50:50.066 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : liboath-2.6.12-1.el9.x86_64 27/113 2026-03-06T13:50:50.066 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : libquadmath-11.5.0-14.el9.x86_64 28/113 2026-03-06T13:50:50.066 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : libradosstriper1-2:19.2.3-47.gc24117fd552.el9.cl 29/113 2026-03-06T13:50:50.066 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : libstoragemgmt-1.10.1-1.el9.x86_64 30/113 2026-03-06T13:50:50.066 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : libunwind-1.6.2-1.el9.x86_64 31/113 2026-03-06T13:50:50.066 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : libxslt-1.1.34-12.el9.x86_64 32/113 2026-03-06T13:50:50.066 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : openblas-0.3.29-1.el9.x86_64 33/113 2026-03-06T13:50:50.066 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : openblas-openmp-0.3.29-1.el9.x86_64 34/113 2026-03-06T13:50:50.066 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : pciutils-3.7.0-7.el9.x86_64 35/113 2026-03-06T13:50:50.066 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : protobuf-3.14.0-17.el9.x86_64 36/113 2026-03-06T13:50:50.066 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : protobuf-compiler-3.14.0-17.el9.x86_64 37/113 2026-03-06T13:50:50.066 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : python3-asyncssh-2.13.2-5.el9.noarch 38/113 2026-03-06T13:50:50.066 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : python3-autocommand-2.2.2-8.el9.noarch 39/113 2026-03-06T13:50:50.066 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : python3-babel-2.9.1-2.el9.noarch 40/113 2026-03-06T13:50:50.066 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : python3-backports-tarfile-1.2.0-1.el9.noarch 41/113 2026-03-06T13:50:50.066 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : python3-bcrypt-3.2.2-1.el9.x86_64 42/113 2026-03-06T13:50:50.066 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : python3-cachetools-4.2.4-1.el9.noarch 43/113 2026-03-06T13:50:50.066 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : python3-ceph-common-2:19.2.3-47.gc24117fd552.el9 44/113 2026-03-06T13:50:50.066 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : python3-certifi-2023.05.07-4.el9.noarch 45/113 2026-03-06T13:50:50.066 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : python3-cffi-1.14.5-5.el9.x86_64 46/113 2026-03-06T13:50:50.066 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : python3-chardet-4.0.0-5.el9.noarch 47/113 2026-03-06T13:50:50.066 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : python3-cheroot-10.0.1-4.el9.noarch 48/113 2026-03-06T13:50:50.066 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : python3-cherrypy-18.6.1-2.el9.noarch 49/113 2026-03-06T13:50:50.066 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : python3-cryptography-36.0.1-5.el9.x86_64 50/113 2026-03-06T13:50:50.066 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : python3-devel-3.9.25-3.el9.x86_64 51/113 2026-03-06T13:50:50.066 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : python3-google-auth-1:2.45.0-1.el9.noarch 52/113 2026-03-06T13:50:50.066 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : python3-grpcio-1.46.7-10.el9.x86_64 53/113 2026-03-06T13:50:50.066 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : python3-grpcio-tools-1.46.7-10.el9.x86_64 54/113 2026-03-06T13:50:50.066 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : python3-idna-2.10-7.el9.1.noarch 55/113 2026-03-06T13:50:50.066 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : python3-influxdb-5.3.1-1.el9.noarch 56/113 2026-03-06T13:50:50.066 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : python3-isodate-0.6.1-3.el9.noarch 57/113 2026-03-06T13:50:50.067 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : python3-jaraco-8.2.1-3.el9.noarch 58/113 2026-03-06T13:50:50.067 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : python3-jaraco-classes-3.2.1-5.el9.noarch 59/113 2026-03-06T13:50:50.067 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : python3-jaraco-collections-3.0.0-8.el9.noarch 60/113 2026-03-06T13:50:50.067 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : python3-jaraco-context-6.0.1-3.el9.noarch 61/113 2026-03-06T13:50:50.067 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : python3-jaraco-functools-3.5.0-2.el9.noarch 62/113 2026-03-06T13:50:50.067 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : python3-jaraco-text-4.0.0-2.el9.noarch 63/113 2026-03-06T13:50:50.067 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : python3-jinja2-2.11.3-8.el9.noarch 64/113 2026-03-06T13:50:50.067 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : python3-jsonpatch-1.21-16.el9.noarch 65/113 2026-03-06T13:50:50.067 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : python3-jsonpointer-2.0-4.el9.noarch 66/113 2026-03-06T13:50:50.067 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : python3-kubernetes-1:26.1.0-3.el9.noarch 67/113 2026-03-06T13:50:50.067 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : python3-libstoragemgmt-1.10.1-1.el9.x86_64 68/113 2026-03-06T13:50:50.067 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : python3-logutils-0.3.5-21.el9.noarch 69/113 2026-03-06T13:50:50.067 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : python3-lxml-4.6.5-3.el9.x86_64 70/113 2026-03-06T13:50:50.067 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : python3-mako-1.1.4-6.el9.noarch 71/113 2026-03-06T13:50:50.067 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : python3-markupsafe-1.1.1-12.el9.x86_64 72/113 2026-03-06T13:50:50.067 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : python3-more-itertools-8.12.0-2.el9.noarch 73/113 2026-03-06T13:50:50.067 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : python3-msgpack-1.0.3-2.el9.x86_64 74/113 2026-03-06T13:50:50.067 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : python3-natsort-7.1.1-5.el9.noarch 75/113 2026-03-06T13:50:50.067 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : python3-numpy-1:1.23.5-2.el9.x86_64 76/113 2026-03-06T13:50:50.067 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : python3-numpy-f2py-1:1.23.5-2.el9.x86_64 77/113 2026-03-06T13:50:50.067 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : python3-oauthlib-3.1.1-5.el9.noarch 78/113 2026-03-06T13:50:50.067 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : python3-packaging-20.9-5.el9.noarch 79/113 2026-03-06T13:50:50.067 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : python3-pecan-1.4.2-3.el9.noarch 80/113 2026-03-06T13:50:50.067 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : python3-ply-3.11-14.el9.noarch 81/113 2026-03-06T13:50:50.067 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : python3-portend-3.1.0-2.el9.noarch 82/113 2026-03-06T13:50:50.067 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : python3-prettytable-0.7.2-27.el9.noarch 83/113 2026-03-06T13:50:50.067 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : python3-protobuf-3.14.0-17.el9.noarch 84/113 2026-03-06T13:50:50.067 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : python3-pyOpenSSL-21.0.0-1.el9.noarch 85/113 2026-03-06T13:50:50.067 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : python3-pyasn1-0.4.8-7.el9.noarch 86/113 2026-03-06T13:50:50.067 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : python3-pyasn1-modules-0.4.8-7.el9.noarch 87/113 2026-03-06T13:50:50.067 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : python3-pycparser-2.20-6.el9.noarch 88/113 2026-03-06T13:50:50.067 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : python3-pysocks-1.7.1-12.el9.noarch 89/113 2026-03-06T13:50:50.067 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : python3-pytz-2021.1-5.el9.noarch 90/113 2026-03-06T13:50:50.067 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : python3-repoze-lru-0.7-16.el9.noarch 91/113 2026-03-06T13:50:50.067 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : python3-requests-2.25.1-10.el9.noarch 92/113 2026-03-06T13:50:50.067 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : python3-requests-oauthlib-1.3.0-12.el9.noarch 93/113 2026-03-06T13:50:50.067 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : python3-routes-2.5.1-5.el9.noarch 94/113 2026-03-06T13:50:50.067 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : python3-rsa-4.9-2.el9.noarch 95/113 2026-03-06T13:50:50.067 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : python3-saml-1.16.0-1.el9.noarch 96/113 2026-03-06T13:50:50.067 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : python3-scipy-1.9.3-2.el9.x86_64 97/113 2026-03-06T13:50:50.067 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : python3-tempora-5.0.0-2.el9.noarch 98/113 2026-03-06T13:50:50.067 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : python3-toml-0.10.2-6.el9.noarch 99/113 2026-03-06T13:50:50.067 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : python3-typing-extensions-4.15.0-1.el9.noarch 100/113 2026-03-06T13:50:50.067 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : python3-urllib3-1.26.5-7.el9.noarch 101/113 2026-03-06T13:50:50.068 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : python3-webob-1.8.8-2.el9.noarch 102/113 2026-03-06T13:50:50.068 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : python3-websocket-client-1.2.3-2.el9.noarch 103/113 2026-03-06T13:50:50.068 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : python3-werkzeug-2.0.3-3.el9.1.noarch 104/113 2026-03-06T13:50:50.068 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : python3-xmlsec-1.3.13-1.el9.x86_64 105/113 2026-03-06T13:50:50.068 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : python3-zc-lockfile-2.0-10.el9.noarch 106/113 2026-03-06T13:50:50.068 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : qatlib-25.08.0-2.el9.x86_64 107/113 2026-03-06T13:50:50.068 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : qatlib-service-25.08.0-2.el9.x86_64 108/113 2026-03-06T13:50:50.068 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : qatzip-libs-1.3.1-1.el9.x86_64 109/113 2026-03-06T13:50:50.068 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : rbd-mirror-2:19.2.3-47.gc24117fd552.el9.clyso.x8 110/113 2026-03-06T13:50:50.068 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : smartmontools-1:7.2-10.el9.x86_64 111/113 2026-03-06T13:50:50.068 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : xmlsec1-1.2.29-13.el9.x86_64 112/113 2026-03-06T13:50:50.082 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : xmlsec1-openssl-1.2.29-13.el9.x86_64 113/113 2026-03-06T13:50:50.083 INFO:teuthology.orchestra.run.vm06.stdout: 2026-03-06T13:50:50.083 INFO:teuthology.orchestra.run.vm06.stdout:Removed: 2026-03-06T13:50:50.083 INFO:teuthology.orchestra.run.vm06.stdout: abseil-cpp-20211102.0-4.el9.x86_64 2026-03-06T13:50:50.083 INFO:teuthology.orchestra.run.vm06.stdout: ceph-base-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 2026-03-06T13:50:50.083 INFO:teuthology.orchestra.run.vm06.stdout: ceph-common-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 2026-03-06T13:50:50.083 INFO:teuthology.orchestra.run.vm06.stdout: ceph-grafana-dashboards-2:19.2.3-47.gc24117fd552.el9.clyso.noarch 2026-03-06T13:50:50.083 INFO:teuthology.orchestra.run.vm06.stdout: ceph-immutable-object-cache-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 2026-03-06T13:50:50.083 INFO:teuthology.orchestra.run.vm06.stdout: ceph-mgr-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 2026-03-06T13:50:50.083 INFO:teuthology.orchestra.run.vm06.stdout: ceph-mgr-cephadm-2:19.2.3-47.gc24117fd552.el9.clyso.noarch 2026-03-06T13:50:50.083 INFO:teuthology.orchestra.run.vm06.stdout: ceph-mgr-dashboard-2:19.2.3-47.gc24117fd552.el9.clyso.noarch 2026-03-06T13:50:50.083 INFO:teuthology.orchestra.run.vm06.stdout: ceph-mgr-diskprediction-local-2:19.2.3-47.gc24117fd552.el9.clyso.noarch 2026-03-06T13:50:50.083 INFO:teuthology.orchestra.run.vm06.stdout: ceph-mgr-k8sevents-2:19.2.3-47.gc24117fd552.el9.clyso.noarch 2026-03-06T13:50:50.083 INFO:teuthology.orchestra.run.vm06.stdout: ceph-mgr-modules-core-2:19.2.3-47.gc24117fd552.el9.clyso.noarch 2026-03-06T13:50:50.083 INFO:teuthology.orchestra.run.vm06.stdout: ceph-mgr-rook-2:19.2.3-47.gc24117fd552.el9.clyso.noarch 2026-03-06T13:50:50.083 INFO:teuthology.orchestra.run.vm06.stdout: ceph-osd-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 2026-03-06T13:50:50.083 INFO:teuthology.orchestra.run.vm06.stdout: ceph-prometheus-alerts-2:19.2.3-47.gc24117fd552.el9.clyso.noarch 2026-03-06T13:50:50.083 INFO:teuthology.orchestra.run.vm06.stdout: ceph-selinux-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 2026-03-06T13:50:50.083 INFO:teuthology.orchestra.run.vm06.stdout: ceph-volume-2:19.2.3-47.gc24117fd552.el9.clyso.noarch 2026-03-06T13:50:50.083 INFO:teuthology.orchestra.run.vm06.stdout: cryptsetup-2.8.1-3.el9.x86_64 2026-03-06T13:50:50.083 INFO:teuthology.orchestra.run.vm06.stdout: flexiblas-3.0.4-9.el9.x86_64 2026-03-06T13:50:50.083 INFO:teuthology.orchestra.run.vm06.stdout: flexiblas-netlib-3.0.4-9.el9.x86_64 2026-03-06T13:50:50.083 INFO:teuthology.orchestra.run.vm06.stdout: flexiblas-openblas-openmp-3.0.4-9.el9.x86_64 2026-03-06T13:50:50.083 INFO:teuthology.orchestra.run.vm06.stdout: gperftools-libs-2.9.1-3.el9.x86_64 2026-03-06T13:50:50.083 INFO:teuthology.orchestra.run.vm06.stdout: grpc-data-1.46.7-10.el9.noarch 2026-03-06T13:50:50.083 INFO:teuthology.orchestra.run.vm06.stdout: ledmon-libs-1.1.0-3.el9.x86_64 2026-03-06T13:50:50.083 INFO:teuthology.orchestra.run.vm06.stdout: libcephsqlite-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 2026-03-06T13:50:50.083 INFO:teuthology.orchestra.run.vm06.stdout: libconfig-1.7.2-9.el9.x86_64 2026-03-06T13:50:50.083 INFO:teuthology.orchestra.run.vm06.stdout: libgfortran-11.5.0-14.el9.x86_64 2026-03-06T13:50:50.083 INFO:teuthology.orchestra.run.vm06.stdout: liboath-2.6.12-1.el9.x86_64 2026-03-06T13:50:50.083 INFO:teuthology.orchestra.run.vm06.stdout: libquadmath-11.5.0-14.el9.x86_64 2026-03-06T13:50:50.083 INFO:teuthology.orchestra.run.vm06.stdout: libradosstriper1-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 2026-03-06T13:50:50.083 INFO:teuthology.orchestra.run.vm06.stdout: libstoragemgmt-1.10.1-1.el9.x86_64 2026-03-06T13:50:50.083 INFO:teuthology.orchestra.run.vm06.stdout: libunwind-1.6.2-1.el9.x86_64 2026-03-06T13:50:50.083 INFO:teuthology.orchestra.run.vm06.stdout: libxslt-1.1.34-12.el9.x86_64 2026-03-06T13:50:50.083 INFO:teuthology.orchestra.run.vm06.stdout: openblas-0.3.29-1.el9.x86_64 2026-03-06T13:50:50.083 INFO:teuthology.orchestra.run.vm06.stdout: openblas-openmp-0.3.29-1.el9.x86_64 2026-03-06T13:50:50.083 INFO:teuthology.orchestra.run.vm06.stdout: pciutils-3.7.0-7.el9.x86_64 2026-03-06T13:50:50.083 INFO:teuthology.orchestra.run.vm06.stdout: protobuf-3.14.0-17.el9.x86_64 2026-03-06T13:50:50.083 INFO:teuthology.orchestra.run.vm06.stdout: protobuf-compiler-3.14.0-17.el9.x86_64 2026-03-06T13:50:50.083 INFO:teuthology.orchestra.run.vm06.stdout: python3-asyncssh-2.13.2-5.el9.noarch 2026-03-06T13:50:50.083 INFO:teuthology.orchestra.run.vm06.stdout: python3-autocommand-2.2.2-8.el9.noarch 2026-03-06T13:50:50.083 INFO:teuthology.orchestra.run.vm06.stdout: python3-babel-2.9.1-2.el9.noarch 2026-03-06T13:50:50.083 INFO:teuthology.orchestra.run.vm06.stdout: python3-backports-tarfile-1.2.0-1.el9.noarch 2026-03-06T13:50:50.083 INFO:teuthology.orchestra.run.vm06.stdout: python3-bcrypt-3.2.2-1.el9.x86_64 2026-03-06T13:50:50.083 INFO:teuthology.orchestra.run.vm06.stdout: python3-cachetools-4.2.4-1.el9.noarch 2026-03-06T13:50:50.083 INFO:teuthology.orchestra.run.vm06.stdout: python3-ceph-common-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 2026-03-06T13:50:50.083 INFO:teuthology.orchestra.run.vm06.stdout: python3-certifi-2023.05.07-4.el9.noarch 2026-03-06T13:50:50.083 INFO:teuthology.orchestra.run.vm06.stdout: python3-cffi-1.14.5-5.el9.x86_64 2026-03-06T13:50:50.083 INFO:teuthology.orchestra.run.vm06.stdout: python3-chardet-4.0.0-5.el9.noarch 2026-03-06T13:50:50.083 INFO:teuthology.orchestra.run.vm06.stdout: python3-cheroot-10.0.1-4.el9.noarch 2026-03-06T13:50:50.083 INFO:teuthology.orchestra.run.vm06.stdout: python3-cherrypy-18.6.1-2.el9.noarch 2026-03-06T13:50:50.083 INFO:teuthology.orchestra.run.vm06.stdout: python3-cryptography-36.0.1-5.el9.x86_64 2026-03-06T13:50:50.083 INFO:teuthology.orchestra.run.vm06.stdout: python3-devel-3.9.25-3.el9.x86_64 2026-03-06T13:50:50.083 INFO:teuthology.orchestra.run.vm06.stdout: python3-google-auth-1:2.45.0-1.el9.noarch 2026-03-06T13:50:50.083 INFO:teuthology.orchestra.run.vm06.stdout: python3-grpcio-1.46.7-10.el9.x86_64 2026-03-06T13:50:50.083 INFO:teuthology.orchestra.run.vm06.stdout: python3-grpcio-tools-1.46.7-10.el9.x86_64 2026-03-06T13:50:50.083 INFO:teuthology.orchestra.run.vm06.stdout: python3-idna-2.10-7.el9.1.noarch 2026-03-06T13:50:50.083 INFO:teuthology.orchestra.run.vm06.stdout: python3-influxdb-5.3.1-1.el9.noarch 2026-03-06T13:50:50.084 INFO:teuthology.orchestra.run.vm06.stdout: python3-isodate-0.6.1-3.el9.noarch 2026-03-06T13:50:50.084 INFO:teuthology.orchestra.run.vm06.stdout: python3-jaraco-8.2.1-3.el9.noarch 2026-03-06T13:50:50.084 INFO:teuthology.orchestra.run.vm06.stdout: python3-jaraco-classes-3.2.1-5.el9.noarch 2026-03-06T13:50:50.084 INFO:teuthology.orchestra.run.vm06.stdout: python3-jaraco-collections-3.0.0-8.el9.noarch 2026-03-06T13:50:50.084 INFO:teuthology.orchestra.run.vm06.stdout: python3-jaraco-context-6.0.1-3.el9.noarch 2026-03-06T13:50:50.084 INFO:teuthology.orchestra.run.vm06.stdout: python3-jaraco-functools-3.5.0-2.el9.noarch 2026-03-06T13:50:50.084 INFO:teuthology.orchestra.run.vm06.stdout: python3-jaraco-text-4.0.0-2.el9.noarch 2026-03-06T13:50:50.084 INFO:teuthology.orchestra.run.vm06.stdout: python3-jinja2-2.11.3-8.el9.noarch 2026-03-06T13:50:50.084 INFO:teuthology.orchestra.run.vm06.stdout: python3-jsonpatch-1.21-16.el9.noarch 2026-03-06T13:50:50.084 INFO:teuthology.orchestra.run.vm06.stdout: python3-jsonpointer-2.0-4.el9.noarch 2026-03-06T13:50:50.084 INFO:teuthology.orchestra.run.vm06.stdout: python3-kubernetes-1:26.1.0-3.el9.noarch 2026-03-06T13:50:50.084 INFO:teuthology.orchestra.run.vm06.stdout: python3-libstoragemgmt-1.10.1-1.el9.x86_64 2026-03-06T13:50:50.084 INFO:teuthology.orchestra.run.vm06.stdout: python3-logutils-0.3.5-21.el9.noarch 2026-03-06T13:50:50.084 INFO:teuthology.orchestra.run.vm06.stdout: python3-lxml-4.6.5-3.el9.x86_64 2026-03-06T13:50:50.084 INFO:teuthology.orchestra.run.vm06.stdout: python3-mako-1.1.4-6.el9.noarch 2026-03-06T13:50:50.084 INFO:teuthology.orchestra.run.vm06.stdout: python3-markupsafe-1.1.1-12.el9.x86_64 2026-03-06T13:50:50.084 INFO:teuthology.orchestra.run.vm06.stdout: python3-more-itertools-8.12.0-2.el9.noarch 2026-03-06T13:50:50.084 INFO:teuthology.orchestra.run.vm06.stdout: python3-msgpack-1.0.3-2.el9.x86_64 2026-03-06T13:50:50.084 INFO:teuthology.orchestra.run.vm06.stdout: python3-natsort-7.1.1-5.el9.noarch 2026-03-06T13:50:50.084 INFO:teuthology.orchestra.run.vm06.stdout: python3-numpy-1:1.23.5-2.el9.x86_64 2026-03-06T13:50:50.084 INFO:teuthology.orchestra.run.vm06.stdout: python3-numpy-f2py-1:1.23.5-2.el9.x86_64 2026-03-06T13:50:50.084 INFO:teuthology.orchestra.run.vm06.stdout: python3-oauthlib-3.1.1-5.el9.noarch 2026-03-06T13:50:50.084 INFO:teuthology.orchestra.run.vm06.stdout: python3-packaging-20.9-5.el9.noarch 2026-03-06T13:50:50.084 INFO:teuthology.orchestra.run.vm06.stdout: python3-pecan-1.4.2-3.el9.noarch 2026-03-06T13:50:50.084 INFO:teuthology.orchestra.run.vm06.stdout: python3-ply-3.11-14.el9.noarch 2026-03-06T13:50:50.084 INFO:teuthology.orchestra.run.vm06.stdout: python3-portend-3.1.0-2.el9.noarch 2026-03-06T13:50:50.084 INFO:teuthology.orchestra.run.vm06.stdout: python3-prettytable-0.7.2-27.el9.noarch 2026-03-06T13:50:50.084 INFO:teuthology.orchestra.run.vm06.stdout: python3-protobuf-3.14.0-17.el9.noarch 2026-03-06T13:50:50.084 INFO:teuthology.orchestra.run.vm06.stdout: python3-pyOpenSSL-21.0.0-1.el9.noarch 2026-03-06T13:50:50.084 INFO:teuthology.orchestra.run.vm06.stdout: python3-pyasn1-0.4.8-7.el9.noarch 2026-03-06T13:50:50.084 INFO:teuthology.orchestra.run.vm06.stdout: python3-pyasn1-modules-0.4.8-7.el9.noarch 2026-03-06T13:50:50.084 INFO:teuthology.orchestra.run.vm06.stdout: python3-pycparser-2.20-6.el9.noarch 2026-03-06T13:50:50.084 INFO:teuthology.orchestra.run.vm06.stdout: python3-pysocks-1.7.1-12.el9.noarch 2026-03-06T13:50:50.084 INFO:teuthology.orchestra.run.vm06.stdout: python3-pytz-2021.1-5.el9.noarch 2026-03-06T13:50:50.084 INFO:teuthology.orchestra.run.vm06.stdout: python3-repoze-lru-0.7-16.el9.noarch 2026-03-06T13:50:50.084 INFO:teuthology.orchestra.run.vm06.stdout: python3-requests-2.25.1-10.el9.noarch 2026-03-06T13:50:50.084 INFO:teuthology.orchestra.run.vm06.stdout: python3-requests-oauthlib-1.3.0-12.el9.noarch 2026-03-06T13:50:50.084 INFO:teuthology.orchestra.run.vm06.stdout: python3-routes-2.5.1-5.el9.noarch 2026-03-06T13:50:50.084 INFO:teuthology.orchestra.run.vm06.stdout: python3-rsa-4.9-2.el9.noarch 2026-03-06T13:50:50.084 INFO:teuthology.orchestra.run.vm06.stdout: python3-saml-1.16.0-1.el9.noarch 2026-03-06T13:50:50.084 INFO:teuthology.orchestra.run.vm06.stdout: python3-scipy-1.9.3-2.el9.x86_64 2026-03-06T13:50:50.084 INFO:teuthology.orchestra.run.vm06.stdout: python3-tempora-5.0.0-2.el9.noarch 2026-03-06T13:50:50.084 INFO:teuthology.orchestra.run.vm06.stdout: python3-toml-0.10.2-6.el9.noarch 2026-03-06T13:50:50.084 INFO:teuthology.orchestra.run.vm06.stdout: python3-typing-extensions-4.15.0-1.el9.noarch 2026-03-06T13:50:50.084 INFO:teuthology.orchestra.run.vm06.stdout: python3-urllib3-1.26.5-7.el9.noarch 2026-03-06T13:50:50.084 INFO:teuthology.orchestra.run.vm06.stdout: python3-webob-1.8.8-2.el9.noarch 2026-03-06T13:50:50.084 INFO:teuthology.orchestra.run.vm06.stdout: python3-websocket-client-1.2.3-2.el9.noarch 2026-03-06T13:50:50.084 INFO:teuthology.orchestra.run.vm06.stdout: python3-werkzeug-2.0.3-3.el9.1.noarch 2026-03-06T13:50:50.084 INFO:teuthology.orchestra.run.vm06.stdout: python3-xmlsec-1.3.13-1.el9.x86_64 2026-03-06T13:50:50.084 INFO:teuthology.orchestra.run.vm06.stdout: python3-zc-lockfile-2.0-10.el9.noarch 2026-03-06T13:50:50.084 INFO:teuthology.orchestra.run.vm06.stdout: qatlib-25.08.0-2.el9.x86_64 2026-03-06T13:50:50.084 INFO:teuthology.orchestra.run.vm06.stdout: qatlib-service-25.08.0-2.el9.x86_64 2026-03-06T13:50:50.084 INFO:teuthology.orchestra.run.vm06.stdout: qatzip-libs-1.3.1-1.el9.x86_64 2026-03-06T13:50:50.084 INFO:teuthology.orchestra.run.vm06.stdout: rbd-mirror-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 2026-03-06T13:50:50.084 INFO:teuthology.orchestra.run.vm06.stdout: smartmontools-1:7.2-10.el9.x86_64 2026-03-06T13:50:50.084 INFO:teuthology.orchestra.run.vm06.stdout: xmlsec1-1.2.29-13.el9.x86_64 2026-03-06T13:50:50.085 INFO:teuthology.orchestra.run.vm06.stdout: xmlsec1-openssl-1.2.29-13.el9.x86_64 2026-03-06T13:50:50.085 INFO:teuthology.orchestra.run.vm06.stdout: 2026-03-06T13:50:50.085 INFO:teuthology.orchestra.run.vm06.stdout:Complete! 2026-03-06T13:50:50.149 INFO:teuthology.orchestra.run.vm04.stdout: Running scriptlet: libcephsqlite-2:19.2.3-47.gc24117fd552.el9.clyso 113/113 2026-03-06T13:50:50.149 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : abseil-cpp-20211102.0-4.el9.x86_64 1/113 2026-03-06T13:50:50.149 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : ceph-base-2:19.2.3-47.gc24117fd552.el9.clyso.x86 2/113 2026-03-06T13:50:50.149 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : ceph-common-2:19.2.3-47.gc24117fd552.el9.clyso.x 3/113 2026-03-06T13:50:50.149 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : ceph-grafana-dashboards-2:19.2.3-47.gc24117fd552 4/113 2026-03-06T13:50:50.149 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : ceph-immutable-object-cache-2:19.2.3-47.gc24117f 5/113 2026-03-06T13:50:50.149 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : ceph-mgr-2:19.2.3-47.gc24117fd552.el9.clyso.x86_ 6/113 2026-03-06T13:50:50.149 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : ceph-mgr-cephadm-2:19.2.3-47.gc24117fd552.el9.cl 7/113 2026-03-06T13:50:50.150 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : ceph-mgr-dashboard-2:19.2.3-47.gc24117fd552.el9. 8/113 2026-03-06T13:50:50.150 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : ceph-mgr-diskprediction-local-2:19.2.3-47.gc2411 9/113 2026-03-06T13:50:50.151 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : ceph-mgr-k8sevents-2:19.2.3-47.gc24117fd552.el9. 10/113 2026-03-06T13:50:50.151 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : ceph-mgr-modules-core-2:19.2.3-47.gc24117fd552.e 11/113 2026-03-06T13:50:50.151 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : ceph-mgr-rook-2:19.2.3-47.gc24117fd552.el9.clyso 12/113 2026-03-06T13:50:50.151 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : ceph-osd-2:19.2.3-47.gc24117fd552.el9.clyso.x86_ 13/113 2026-03-06T13:50:50.151 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : ceph-prometheus-alerts-2:19.2.3-47.gc24117fd552. 14/113 2026-03-06T13:50:50.151 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : ceph-selinux-2:19.2.3-47.gc24117fd552.el9.clyso. 15/113 2026-03-06T13:50:50.151 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : ceph-volume-2:19.2.3-47.gc24117fd552.el9.clyso.n 16/113 2026-03-06T13:50:50.151 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : cryptsetup-2.8.1-3.el9.x86_64 17/113 2026-03-06T13:50:50.151 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : flexiblas-3.0.4-9.el9.x86_64 18/113 2026-03-06T13:50:50.151 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : flexiblas-netlib-3.0.4-9.el9.x86_64 19/113 2026-03-06T13:50:50.151 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : flexiblas-openblas-openmp-3.0.4-9.el9.x86_64 20/113 2026-03-06T13:50:50.151 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : gperftools-libs-2.9.1-3.el9.x86_64 21/113 2026-03-06T13:50:50.151 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : grpc-data-1.46.7-10.el9.noarch 22/113 2026-03-06T13:50:50.151 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : ledmon-libs-1.1.0-3.el9.x86_64 23/113 2026-03-06T13:50:50.151 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : libcephsqlite-2:19.2.3-47.gc24117fd552.el9.clyso 24/113 2026-03-06T13:50:50.151 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : libconfig-1.7.2-9.el9.x86_64 25/113 2026-03-06T13:50:50.151 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : libgfortran-11.5.0-14.el9.x86_64 26/113 2026-03-06T13:50:50.151 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : liboath-2.6.12-1.el9.x86_64 27/113 2026-03-06T13:50:50.151 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : libquadmath-11.5.0-14.el9.x86_64 28/113 2026-03-06T13:50:50.151 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : libradosstriper1-2:19.2.3-47.gc24117fd552.el9.cl 29/113 2026-03-06T13:50:50.151 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : libstoragemgmt-1.10.1-1.el9.x86_64 30/113 2026-03-06T13:50:50.151 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : libunwind-1.6.2-1.el9.x86_64 31/113 2026-03-06T13:50:50.151 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : libxslt-1.1.34-12.el9.x86_64 32/113 2026-03-06T13:50:50.151 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : openblas-0.3.29-1.el9.x86_64 33/113 2026-03-06T13:50:50.151 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : openblas-openmp-0.3.29-1.el9.x86_64 34/113 2026-03-06T13:50:50.151 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : pciutils-3.7.0-7.el9.x86_64 35/113 2026-03-06T13:50:50.152 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : protobuf-3.14.0-17.el9.x86_64 36/113 2026-03-06T13:50:50.152 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : protobuf-compiler-3.14.0-17.el9.x86_64 37/113 2026-03-06T13:50:50.152 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-asyncssh-2.13.2-5.el9.noarch 38/113 2026-03-06T13:50:50.152 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-autocommand-2.2.2-8.el9.noarch 39/113 2026-03-06T13:50:50.152 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-babel-2.9.1-2.el9.noarch 40/113 2026-03-06T13:50:50.152 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-backports-tarfile-1.2.0-1.el9.noarch 41/113 2026-03-06T13:50:50.152 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-bcrypt-3.2.2-1.el9.x86_64 42/113 2026-03-06T13:50:50.152 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-cachetools-4.2.4-1.el9.noarch 43/113 2026-03-06T13:50:50.152 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-ceph-common-2:19.2.3-47.gc24117fd552.el9 44/113 2026-03-06T13:50:50.152 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-certifi-2023.05.07-4.el9.noarch 45/113 2026-03-06T13:50:50.152 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-cffi-1.14.5-5.el9.x86_64 46/113 2026-03-06T13:50:50.154 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-chardet-4.0.0-5.el9.noarch 47/113 2026-03-06T13:50:50.154 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-cheroot-10.0.1-4.el9.noarch 48/113 2026-03-06T13:50:50.154 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-cherrypy-18.6.1-2.el9.noarch 49/113 2026-03-06T13:50:50.154 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-cryptography-36.0.1-5.el9.x86_64 50/113 2026-03-06T13:50:50.154 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-devel-3.9.25-3.el9.x86_64 51/113 2026-03-06T13:50:50.154 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-google-auth-1:2.45.0-1.el9.noarch 52/113 2026-03-06T13:50:50.154 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-grpcio-1.46.7-10.el9.x86_64 53/113 2026-03-06T13:50:50.154 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-grpcio-tools-1.46.7-10.el9.x86_64 54/113 2026-03-06T13:50:50.154 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-idna-2.10-7.el9.1.noarch 55/113 2026-03-06T13:50:50.154 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-influxdb-5.3.1-1.el9.noarch 56/113 2026-03-06T13:50:50.154 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-isodate-0.6.1-3.el9.noarch 57/113 2026-03-06T13:50:50.154 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-jaraco-8.2.1-3.el9.noarch 58/113 2026-03-06T13:50:50.154 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-jaraco-classes-3.2.1-5.el9.noarch 59/113 2026-03-06T13:50:50.154 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-jaraco-collections-3.0.0-8.el9.noarch 60/113 2026-03-06T13:50:50.154 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-jaraco-context-6.0.1-3.el9.noarch 61/113 2026-03-06T13:50:50.154 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-jaraco-functools-3.5.0-2.el9.noarch 62/113 2026-03-06T13:50:50.154 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-jaraco-text-4.0.0-2.el9.noarch 63/113 2026-03-06T13:50:50.154 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-jinja2-2.11.3-8.el9.noarch 64/113 2026-03-06T13:50:50.154 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-jsonpatch-1.21-16.el9.noarch 65/113 2026-03-06T13:50:50.154 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-jsonpointer-2.0-4.el9.noarch 66/113 2026-03-06T13:50:50.154 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-kubernetes-1:26.1.0-3.el9.noarch 67/113 2026-03-06T13:50:50.154 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-libstoragemgmt-1.10.1-1.el9.x86_64 68/113 2026-03-06T13:50:50.154 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-logutils-0.3.5-21.el9.noarch 69/113 2026-03-06T13:50:50.154 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-lxml-4.6.5-3.el9.x86_64 70/113 2026-03-06T13:50:50.154 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-mako-1.1.4-6.el9.noarch 71/113 2026-03-06T13:50:50.154 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-markupsafe-1.1.1-12.el9.x86_64 72/113 2026-03-06T13:50:50.154 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-more-itertools-8.12.0-2.el9.noarch 73/113 2026-03-06T13:50:50.154 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-msgpack-1.0.3-2.el9.x86_64 74/113 2026-03-06T13:50:50.154 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-natsort-7.1.1-5.el9.noarch 75/113 2026-03-06T13:50:50.154 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-numpy-1:1.23.5-2.el9.x86_64 76/113 2026-03-06T13:50:50.155 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-numpy-f2py-1:1.23.5-2.el9.x86_64 77/113 2026-03-06T13:50:50.155 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-oauthlib-3.1.1-5.el9.noarch 78/113 2026-03-06T13:50:50.155 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-packaging-20.9-5.el9.noarch 79/113 2026-03-06T13:50:50.155 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-pecan-1.4.2-3.el9.noarch 80/113 2026-03-06T13:50:50.155 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-ply-3.11-14.el9.noarch 81/113 2026-03-06T13:50:50.155 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-portend-3.1.0-2.el9.noarch 82/113 2026-03-06T13:50:50.155 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-prettytable-0.7.2-27.el9.noarch 83/113 2026-03-06T13:50:50.155 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-protobuf-3.14.0-17.el9.noarch 84/113 2026-03-06T13:50:50.155 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-pyOpenSSL-21.0.0-1.el9.noarch 85/113 2026-03-06T13:50:50.155 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-pyasn1-0.4.8-7.el9.noarch 86/113 2026-03-06T13:50:50.155 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-pyasn1-modules-0.4.8-7.el9.noarch 87/113 2026-03-06T13:50:50.155 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-pycparser-2.20-6.el9.noarch 88/113 2026-03-06T13:50:50.155 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-pysocks-1.7.1-12.el9.noarch 89/113 2026-03-06T13:50:50.155 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-pytz-2021.1-5.el9.noarch 90/113 2026-03-06T13:50:50.155 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-repoze-lru-0.7-16.el9.noarch 91/113 2026-03-06T13:50:50.155 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-requests-2.25.1-10.el9.noarch 92/113 2026-03-06T13:50:50.155 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-requests-oauthlib-1.3.0-12.el9.noarch 93/113 2026-03-06T13:50:50.155 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-routes-2.5.1-5.el9.noarch 94/113 2026-03-06T13:50:50.155 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-rsa-4.9-2.el9.noarch 95/113 2026-03-06T13:50:50.155 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-saml-1.16.0-1.el9.noarch 96/113 2026-03-06T13:50:50.155 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-scipy-1.9.3-2.el9.x86_64 97/113 2026-03-06T13:50:50.155 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-tempora-5.0.0-2.el9.noarch 98/113 2026-03-06T13:50:50.155 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-toml-0.10.2-6.el9.noarch 99/113 2026-03-06T13:50:50.155 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-typing-extensions-4.15.0-1.el9.noarch 100/113 2026-03-06T13:50:50.155 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-urllib3-1.26.5-7.el9.noarch 101/113 2026-03-06T13:50:50.155 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-webob-1.8.8-2.el9.noarch 102/113 2026-03-06T13:50:50.155 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-websocket-client-1.2.3-2.el9.noarch 103/113 2026-03-06T13:50:50.155 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-werkzeug-2.0.3-3.el9.1.noarch 104/113 2026-03-06T13:50:50.155 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-xmlsec-1.3.13-1.el9.x86_64 105/113 2026-03-06T13:50:50.155 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-zc-lockfile-2.0-10.el9.noarch 106/113 2026-03-06T13:50:50.155 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : xmlsec1-openssl-1.2.29-13.el9.x86_64 113/113 2026-03-06T13:50:50.156 INFO:teuthology.orchestra.run.vm02.stdout: 2026-03-06T13:50:50.156 INFO:teuthology.orchestra.run.vm02.stdout:Removed: 2026-03-06T13:50:50.156 INFO:teuthology.orchestra.run.vm02.stdout: abseil-cpp-20211102.0-4.el9.x86_64 2026-03-06T13:50:50.156 INFO:teuthology.orchestra.run.vm02.stdout: ceph-base-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 2026-03-06T13:50:50.156 INFO:teuthology.orchestra.run.vm02.stdout: ceph-common-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 2026-03-06T13:50:50.156 INFO:teuthology.orchestra.run.vm02.stdout: ceph-grafana-dashboards-2:19.2.3-47.gc24117fd552.el9.clyso.noarch 2026-03-06T13:50:50.156 INFO:teuthology.orchestra.run.vm02.stdout: ceph-immutable-object-cache-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 2026-03-06T13:50:50.156 INFO:teuthology.orchestra.run.vm02.stdout: ceph-mgr-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 2026-03-06T13:50:50.156 INFO:teuthology.orchestra.run.vm02.stdout: ceph-mgr-cephadm-2:19.2.3-47.gc24117fd552.el9.clyso.noarch 2026-03-06T13:50:50.156 INFO:teuthology.orchestra.run.vm02.stdout: ceph-mgr-dashboard-2:19.2.3-47.gc24117fd552.el9.clyso.noarch 2026-03-06T13:50:50.156 INFO:teuthology.orchestra.run.vm02.stdout: ceph-mgr-diskprediction-local-2:19.2.3-47.gc24117fd552.el9.clyso.noarch 2026-03-06T13:50:50.156 INFO:teuthology.orchestra.run.vm02.stdout: ceph-mgr-k8sevents-2:19.2.3-47.gc24117fd552.el9.clyso.noarch 2026-03-06T13:50:50.156 INFO:teuthology.orchestra.run.vm02.stdout: ceph-mgr-modules-core-2:19.2.3-47.gc24117fd552.el9.clyso.noarch 2026-03-06T13:50:50.156 INFO:teuthology.orchestra.run.vm02.stdout: ceph-mgr-rook-2:19.2.3-47.gc24117fd552.el9.clyso.noarch 2026-03-06T13:50:50.156 INFO:teuthology.orchestra.run.vm02.stdout: ceph-osd-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 2026-03-06T13:50:50.156 INFO:teuthology.orchestra.run.vm02.stdout: ceph-prometheus-alerts-2:19.2.3-47.gc24117fd552.el9.clyso.noarch 2026-03-06T13:50:50.156 INFO:teuthology.orchestra.run.vm02.stdout: ceph-selinux-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 2026-03-06T13:50:50.156 INFO:teuthology.orchestra.run.vm02.stdout: ceph-volume-2:19.2.3-47.gc24117fd552.el9.clyso.noarch 2026-03-06T13:50:50.156 INFO:teuthology.orchestra.run.vm02.stdout: cryptsetup-2.8.1-3.el9.x86_64 2026-03-06T13:50:50.156 INFO:teuthology.orchestra.run.vm02.stdout: flexiblas-3.0.4-9.el9.x86_64 2026-03-06T13:50:50.156 INFO:teuthology.orchestra.run.vm02.stdout: flexiblas-netlib-3.0.4-9.el9.x86_64 2026-03-06T13:50:50.156 INFO:teuthology.orchestra.run.vm02.stdout: flexiblas-openblas-openmp-3.0.4-9.el9.x86_64 2026-03-06T13:50:50.156 INFO:teuthology.orchestra.run.vm02.stdout: gperftools-libs-2.9.1-3.el9.x86_64 2026-03-06T13:50:50.156 INFO:teuthology.orchestra.run.vm02.stdout: grpc-data-1.46.7-10.el9.noarch 2026-03-06T13:50:50.156 INFO:teuthology.orchestra.run.vm02.stdout: ledmon-libs-1.1.0-3.el9.x86_64 2026-03-06T13:50:50.156 INFO:teuthology.orchestra.run.vm02.stdout: libcephsqlite-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 2026-03-06T13:50:50.156 INFO:teuthology.orchestra.run.vm02.stdout: libconfig-1.7.2-9.el9.x86_64 2026-03-06T13:50:50.156 INFO:teuthology.orchestra.run.vm02.stdout: libgfortran-11.5.0-14.el9.x86_64 2026-03-06T13:50:50.156 INFO:teuthology.orchestra.run.vm02.stdout: liboath-2.6.12-1.el9.x86_64 2026-03-06T13:50:50.156 INFO:teuthology.orchestra.run.vm02.stdout: libquadmath-11.5.0-14.el9.x86_64 2026-03-06T13:50:50.156 INFO:teuthology.orchestra.run.vm02.stdout: libradosstriper1-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 2026-03-06T13:50:50.156 INFO:teuthology.orchestra.run.vm02.stdout: libstoragemgmt-1.10.1-1.el9.x86_64 2026-03-06T13:50:50.156 INFO:teuthology.orchestra.run.vm02.stdout: libunwind-1.6.2-1.el9.x86_64 2026-03-06T13:50:50.156 INFO:teuthology.orchestra.run.vm02.stdout: libxslt-1.1.34-12.el9.x86_64 2026-03-06T13:50:50.156 INFO:teuthology.orchestra.run.vm02.stdout: openblas-0.3.29-1.el9.x86_64 2026-03-06T13:50:50.156 INFO:teuthology.orchestra.run.vm02.stdout: openblas-openmp-0.3.29-1.el9.x86_64 2026-03-06T13:50:50.156 INFO:teuthology.orchestra.run.vm02.stdout: pciutils-3.7.0-7.el9.x86_64 2026-03-06T13:50:50.156 INFO:teuthology.orchestra.run.vm02.stdout: protobuf-3.14.0-17.el9.x86_64 2026-03-06T13:50:50.156 INFO:teuthology.orchestra.run.vm02.stdout: protobuf-compiler-3.14.0-17.el9.x86_64 2026-03-06T13:50:50.156 INFO:teuthology.orchestra.run.vm02.stdout: python3-asyncssh-2.13.2-5.el9.noarch 2026-03-06T13:50:50.156 INFO:teuthology.orchestra.run.vm02.stdout: python3-autocommand-2.2.2-8.el9.noarch 2026-03-06T13:50:50.156 INFO:teuthology.orchestra.run.vm02.stdout: python3-babel-2.9.1-2.el9.noarch 2026-03-06T13:50:50.157 INFO:teuthology.orchestra.run.vm02.stdout: python3-backports-tarfile-1.2.0-1.el9.noarch 2026-03-06T13:50:50.157 INFO:teuthology.orchestra.run.vm02.stdout: python3-bcrypt-3.2.2-1.el9.x86_64 2026-03-06T13:50:50.157 INFO:teuthology.orchestra.run.vm02.stdout: python3-cachetools-4.2.4-1.el9.noarch 2026-03-06T13:50:50.157 INFO:teuthology.orchestra.run.vm02.stdout: python3-ceph-common-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 2026-03-06T13:50:50.157 INFO:teuthology.orchestra.run.vm02.stdout: python3-certifi-2023.05.07-4.el9.noarch 2026-03-06T13:50:50.157 INFO:teuthology.orchestra.run.vm02.stdout: python3-cffi-1.14.5-5.el9.x86_64 2026-03-06T13:50:50.157 INFO:teuthology.orchestra.run.vm02.stdout: python3-chardet-4.0.0-5.el9.noarch 2026-03-06T13:50:50.157 INFO:teuthology.orchestra.run.vm02.stdout: python3-cheroot-10.0.1-4.el9.noarch 2026-03-06T13:50:50.157 INFO:teuthology.orchestra.run.vm02.stdout: python3-cherrypy-18.6.1-2.el9.noarch 2026-03-06T13:50:50.157 INFO:teuthology.orchestra.run.vm02.stdout: python3-cryptography-36.0.1-5.el9.x86_64 2026-03-06T13:50:50.157 INFO:teuthology.orchestra.run.vm02.stdout: python3-devel-3.9.25-3.el9.x86_64 2026-03-06T13:50:50.157 INFO:teuthology.orchestra.run.vm02.stdout: python3-google-auth-1:2.45.0-1.el9.noarch 2026-03-06T13:50:50.157 INFO:teuthology.orchestra.run.vm02.stdout: python3-grpcio-1.46.7-10.el9.x86_64 2026-03-06T13:50:50.157 INFO:teuthology.orchestra.run.vm02.stdout: python3-grpcio-tools-1.46.7-10.el9.x86_64 2026-03-06T13:50:50.157 INFO:teuthology.orchestra.run.vm02.stdout: python3-idna-2.10-7.el9.1.noarch 2026-03-06T13:50:50.157 INFO:teuthology.orchestra.run.vm02.stdout: python3-influxdb-5.3.1-1.el9.noarch 2026-03-06T13:50:50.157 INFO:teuthology.orchestra.run.vm02.stdout: python3-isodate-0.6.1-3.el9.noarch 2026-03-06T13:50:50.157 INFO:teuthology.orchestra.run.vm02.stdout: python3-jaraco-8.2.1-3.el9.noarch 2026-03-06T13:50:50.157 INFO:teuthology.orchestra.run.vm02.stdout: python3-jaraco-classes-3.2.1-5.el9.noarch 2026-03-06T13:50:50.157 INFO:teuthology.orchestra.run.vm02.stdout: python3-jaraco-collections-3.0.0-8.el9.noarch 2026-03-06T13:50:50.157 INFO:teuthology.orchestra.run.vm02.stdout: python3-jaraco-context-6.0.1-3.el9.noarch 2026-03-06T13:50:50.157 INFO:teuthology.orchestra.run.vm02.stdout: python3-jaraco-functools-3.5.0-2.el9.noarch 2026-03-06T13:50:50.157 INFO:teuthology.orchestra.run.vm02.stdout: python3-jaraco-text-4.0.0-2.el9.noarch 2026-03-06T13:50:50.157 INFO:teuthology.orchestra.run.vm02.stdout: python3-jinja2-2.11.3-8.el9.noarch 2026-03-06T13:50:50.157 INFO:teuthology.orchestra.run.vm02.stdout: python3-jsonpatch-1.21-16.el9.noarch 2026-03-06T13:50:50.157 INFO:teuthology.orchestra.run.vm02.stdout: python3-jsonpointer-2.0-4.el9.noarch 2026-03-06T13:50:50.157 INFO:teuthology.orchestra.run.vm02.stdout: python3-kubernetes-1:26.1.0-3.el9.noarch 2026-03-06T13:50:50.157 INFO:teuthology.orchestra.run.vm02.stdout: python3-libstoragemgmt-1.10.1-1.el9.x86_64 2026-03-06T13:50:50.157 INFO:teuthology.orchestra.run.vm02.stdout: python3-logutils-0.3.5-21.el9.noarch 2026-03-06T13:50:50.157 INFO:teuthology.orchestra.run.vm02.stdout: python3-lxml-4.6.5-3.el9.x86_64 2026-03-06T13:50:50.157 INFO:teuthology.orchestra.run.vm02.stdout: python3-mako-1.1.4-6.el9.noarch 2026-03-06T13:50:50.157 INFO:teuthology.orchestra.run.vm02.stdout: python3-markupsafe-1.1.1-12.el9.x86_64 2026-03-06T13:50:50.157 INFO:teuthology.orchestra.run.vm02.stdout: python3-more-itertools-8.12.0-2.el9.noarch 2026-03-06T13:50:50.157 INFO:teuthology.orchestra.run.vm02.stdout: python3-msgpack-1.0.3-2.el9.x86_64 2026-03-06T13:50:50.157 INFO:teuthology.orchestra.run.vm02.stdout: python3-natsort-7.1.1-5.el9.noarch 2026-03-06T13:50:50.157 INFO:teuthology.orchestra.run.vm02.stdout: python3-numpy-1:1.23.5-2.el9.x86_64 2026-03-06T13:50:50.157 INFO:teuthology.orchestra.run.vm02.stdout: python3-numpy-f2py-1:1.23.5-2.el9.x86_64 2026-03-06T13:50:50.157 INFO:teuthology.orchestra.run.vm02.stdout: python3-oauthlib-3.1.1-5.el9.noarch 2026-03-06T13:50:50.157 INFO:teuthology.orchestra.run.vm02.stdout: python3-packaging-20.9-5.el9.noarch 2026-03-06T13:50:50.157 INFO:teuthology.orchestra.run.vm02.stdout: python3-pecan-1.4.2-3.el9.noarch 2026-03-06T13:50:50.157 INFO:teuthology.orchestra.run.vm02.stdout: python3-ply-3.11-14.el9.noarch 2026-03-06T13:50:50.157 INFO:teuthology.orchestra.run.vm02.stdout: python3-portend-3.1.0-2.el9.noarch 2026-03-06T13:50:50.158 INFO:teuthology.orchestra.run.vm02.stdout: python3-prettytable-0.7.2-27.el9.noarch 2026-03-06T13:50:50.158 INFO:teuthology.orchestra.run.vm02.stdout: python3-protobuf-3.14.0-17.el9.noarch 2026-03-06T13:50:50.158 INFO:teuthology.orchestra.run.vm02.stdout: python3-pyOpenSSL-21.0.0-1.el9.noarch 2026-03-06T13:50:50.158 INFO:teuthology.orchestra.run.vm02.stdout: python3-pyasn1-0.4.8-7.el9.noarch 2026-03-06T13:50:50.158 INFO:teuthology.orchestra.run.vm02.stdout: python3-pyasn1-modules-0.4.8-7.el9.noarch 2026-03-06T13:50:50.158 INFO:teuthology.orchestra.run.vm02.stdout: python3-pycparser-2.20-6.el9.noarch 2026-03-06T13:50:50.158 INFO:teuthology.orchestra.run.vm02.stdout: python3-pysocks-1.7.1-12.el9.noarch 2026-03-06T13:50:50.158 INFO:teuthology.orchestra.run.vm02.stdout: python3-pytz-2021.1-5.el9.noarch 2026-03-06T13:50:50.158 INFO:teuthology.orchestra.run.vm02.stdout: python3-repoze-lru-0.7-16.el9.noarch 2026-03-06T13:50:50.158 INFO:teuthology.orchestra.run.vm02.stdout: python3-requests-2.25.1-10.el9.noarch 2026-03-06T13:50:50.158 INFO:teuthology.orchestra.run.vm02.stdout: python3-requests-oauthlib-1.3.0-12.el9.noarch 2026-03-06T13:50:50.158 INFO:teuthology.orchestra.run.vm02.stdout: python3-routes-2.5.1-5.el9.noarch 2026-03-06T13:50:50.158 INFO:teuthology.orchestra.run.vm02.stdout: python3-rsa-4.9-2.el9.noarch 2026-03-06T13:50:50.158 INFO:teuthology.orchestra.run.vm02.stdout: python3-saml-1.16.0-1.el9.noarch 2026-03-06T13:50:50.158 INFO:teuthology.orchestra.run.vm02.stdout: python3-scipy-1.9.3-2.el9.x86_64 2026-03-06T13:50:50.158 INFO:teuthology.orchestra.run.vm02.stdout: python3-tempora-5.0.0-2.el9.noarch 2026-03-06T13:50:50.158 INFO:teuthology.orchestra.run.vm02.stdout: python3-toml-0.10.2-6.el9.noarch 2026-03-06T13:50:50.158 INFO:teuthology.orchestra.run.vm02.stdout: python3-typing-extensions-4.15.0-1.el9.noarch 2026-03-06T13:50:50.158 INFO:teuthology.orchestra.run.vm02.stdout: python3-urllib3-1.26.5-7.el9.noarch 2026-03-06T13:50:50.158 INFO:teuthology.orchestra.run.vm02.stdout: python3-webob-1.8.8-2.el9.noarch 2026-03-06T13:50:50.158 INFO:teuthology.orchestra.run.vm02.stdout: python3-websocket-client-1.2.3-2.el9.noarch 2026-03-06T13:50:50.158 INFO:teuthology.orchestra.run.vm02.stdout: python3-werkzeug-2.0.3-3.el9.1.noarch 2026-03-06T13:50:50.158 INFO:teuthology.orchestra.run.vm02.stdout: python3-xmlsec-1.3.13-1.el9.x86_64 2026-03-06T13:50:50.158 INFO:teuthology.orchestra.run.vm02.stdout: python3-zc-lockfile-2.0-10.el9.noarch 2026-03-06T13:50:50.158 INFO:teuthology.orchestra.run.vm02.stdout: qatlib-25.08.0-2.el9.x86_64 2026-03-06T13:50:50.158 INFO:teuthology.orchestra.run.vm02.stdout: qatlib-service-25.08.0-2.el9.x86_64 2026-03-06T13:50:50.158 INFO:teuthology.orchestra.run.vm02.stdout: qatzip-libs-1.3.1-1.el9.x86_64 2026-03-06T13:50:50.158 INFO:teuthology.orchestra.run.vm02.stdout: rbd-mirror-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 2026-03-06T13:50:50.158 INFO:teuthology.orchestra.run.vm02.stdout: smartmontools-1:7.2-10.el9.x86_64 2026-03-06T13:50:50.158 INFO:teuthology.orchestra.run.vm02.stdout: xmlsec1-1.2.29-13.el9.x86_64 2026-03-06T13:50:50.158 INFO:teuthology.orchestra.run.vm02.stdout: xmlsec1-openssl-1.2.29-13.el9.x86_64 2026-03-06T13:50:50.158 INFO:teuthology.orchestra.run.vm02.stdout: 2026-03-06T13:50:50.158 INFO:teuthology.orchestra.run.vm02.stdout:Complete! 2026-03-06T13:50:50.158 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : qatlib-25.08.0-2.el9.x86_64 107/113 2026-03-06T13:50:50.159 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : qatlib-service-25.08.0-2.el9.x86_64 108/113 2026-03-06T13:50:50.159 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : qatzip-libs-1.3.1-1.el9.x86_64 109/113 2026-03-06T13:50:50.159 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : rbd-mirror-2:19.2.3-47.gc24117fd552.el9.clyso.x8 110/113 2026-03-06T13:50:50.159 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : smartmontools-1:7.2-10.el9.x86_64 111/113 2026-03-06T13:50:50.159 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : xmlsec1-1.2.29-13.el9.x86_64 112/113 2026-03-06T13:50:50.244 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : xmlsec1-openssl-1.2.29-13.el9.x86_64 113/113 2026-03-06T13:50:50.244 INFO:teuthology.orchestra.run.vm04.stdout: 2026-03-06T13:50:50.244 INFO:teuthology.orchestra.run.vm04.stdout:Removed: 2026-03-06T13:50:50.244 INFO:teuthology.orchestra.run.vm04.stdout: abseil-cpp-20211102.0-4.el9.x86_64 2026-03-06T13:50:50.244 INFO:teuthology.orchestra.run.vm04.stdout: ceph-base-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 2026-03-06T13:50:50.244 INFO:teuthology.orchestra.run.vm04.stdout: ceph-common-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 2026-03-06T13:50:50.244 INFO:teuthology.orchestra.run.vm04.stdout: ceph-grafana-dashboards-2:19.2.3-47.gc24117fd552.el9.clyso.noarch 2026-03-06T13:50:50.244 INFO:teuthology.orchestra.run.vm04.stdout: ceph-immutable-object-cache-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 2026-03-06T13:50:50.244 INFO:teuthology.orchestra.run.vm04.stdout: ceph-mgr-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 2026-03-06T13:50:50.244 INFO:teuthology.orchestra.run.vm04.stdout: ceph-mgr-cephadm-2:19.2.3-47.gc24117fd552.el9.clyso.noarch 2026-03-06T13:50:50.244 INFO:teuthology.orchestra.run.vm04.stdout: ceph-mgr-dashboard-2:19.2.3-47.gc24117fd552.el9.clyso.noarch 2026-03-06T13:50:50.244 INFO:teuthology.orchestra.run.vm04.stdout: ceph-mgr-diskprediction-local-2:19.2.3-47.gc24117fd552.el9.clyso.noarch 2026-03-06T13:50:50.244 INFO:teuthology.orchestra.run.vm04.stdout: ceph-mgr-k8sevents-2:19.2.3-47.gc24117fd552.el9.clyso.noarch 2026-03-06T13:50:50.244 INFO:teuthology.orchestra.run.vm04.stdout: ceph-mgr-modules-core-2:19.2.3-47.gc24117fd552.el9.clyso.noarch 2026-03-06T13:50:50.244 INFO:teuthology.orchestra.run.vm04.stdout: ceph-mgr-rook-2:19.2.3-47.gc24117fd552.el9.clyso.noarch 2026-03-06T13:50:50.244 INFO:teuthology.orchestra.run.vm04.stdout: ceph-osd-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 2026-03-06T13:50:50.244 INFO:teuthology.orchestra.run.vm04.stdout: ceph-prometheus-alerts-2:19.2.3-47.gc24117fd552.el9.clyso.noarch 2026-03-06T13:50:50.244 INFO:teuthology.orchestra.run.vm04.stdout: ceph-selinux-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 2026-03-06T13:50:50.244 INFO:teuthology.orchestra.run.vm04.stdout: ceph-volume-2:19.2.3-47.gc24117fd552.el9.clyso.noarch 2026-03-06T13:50:50.244 INFO:teuthology.orchestra.run.vm04.stdout: cryptsetup-2.8.1-3.el9.x86_64 2026-03-06T13:50:50.244 INFO:teuthology.orchestra.run.vm04.stdout: flexiblas-3.0.4-9.el9.x86_64 2026-03-06T13:50:50.244 INFO:teuthology.orchestra.run.vm04.stdout: flexiblas-netlib-3.0.4-9.el9.x86_64 2026-03-06T13:50:50.244 INFO:teuthology.orchestra.run.vm04.stdout: flexiblas-openblas-openmp-3.0.4-9.el9.x86_64 2026-03-06T13:50:50.244 INFO:teuthology.orchestra.run.vm04.stdout: gperftools-libs-2.9.1-3.el9.x86_64 2026-03-06T13:50:50.244 INFO:teuthology.orchestra.run.vm04.stdout: grpc-data-1.46.7-10.el9.noarch 2026-03-06T13:50:50.244 INFO:teuthology.orchestra.run.vm04.stdout: ledmon-libs-1.1.0-3.el9.x86_64 2026-03-06T13:50:50.244 INFO:teuthology.orchestra.run.vm04.stdout: libcephsqlite-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 2026-03-06T13:50:50.244 INFO:teuthology.orchestra.run.vm04.stdout: libconfig-1.7.2-9.el9.x86_64 2026-03-06T13:50:50.244 INFO:teuthology.orchestra.run.vm04.stdout: libgfortran-11.5.0-14.el9.x86_64 2026-03-06T13:50:50.244 INFO:teuthology.orchestra.run.vm04.stdout: liboath-2.6.12-1.el9.x86_64 2026-03-06T13:50:50.244 INFO:teuthology.orchestra.run.vm04.stdout: libquadmath-11.5.0-14.el9.x86_64 2026-03-06T13:50:50.244 INFO:teuthology.orchestra.run.vm04.stdout: libradosstriper1-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 2026-03-06T13:50:50.245 INFO:teuthology.orchestra.run.vm04.stdout: libstoragemgmt-1.10.1-1.el9.x86_64 2026-03-06T13:50:50.245 INFO:teuthology.orchestra.run.vm04.stdout: libunwind-1.6.2-1.el9.x86_64 2026-03-06T13:50:50.245 INFO:teuthology.orchestra.run.vm04.stdout: libxslt-1.1.34-12.el9.x86_64 2026-03-06T13:50:50.245 INFO:teuthology.orchestra.run.vm04.stdout: openblas-0.3.29-1.el9.x86_64 2026-03-06T13:50:50.245 INFO:teuthology.orchestra.run.vm04.stdout: openblas-openmp-0.3.29-1.el9.x86_64 2026-03-06T13:50:50.245 INFO:teuthology.orchestra.run.vm04.stdout: pciutils-3.7.0-7.el9.x86_64 2026-03-06T13:50:50.245 INFO:teuthology.orchestra.run.vm04.stdout: protobuf-3.14.0-17.el9.x86_64 2026-03-06T13:50:50.245 INFO:teuthology.orchestra.run.vm04.stdout: protobuf-compiler-3.14.0-17.el9.x86_64 2026-03-06T13:50:50.245 INFO:teuthology.orchestra.run.vm04.stdout: python3-asyncssh-2.13.2-5.el9.noarch 2026-03-06T13:50:50.245 INFO:teuthology.orchestra.run.vm04.stdout: python3-autocommand-2.2.2-8.el9.noarch 2026-03-06T13:50:50.245 INFO:teuthology.orchestra.run.vm04.stdout: python3-babel-2.9.1-2.el9.noarch 2026-03-06T13:50:50.245 INFO:teuthology.orchestra.run.vm04.stdout: python3-backports-tarfile-1.2.0-1.el9.noarch 2026-03-06T13:50:50.245 INFO:teuthology.orchestra.run.vm04.stdout: python3-bcrypt-3.2.2-1.el9.x86_64 2026-03-06T13:50:50.245 INFO:teuthology.orchestra.run.vm04.stdout: python3-cachetools-4.2.4-1.el9.noarch 2026-03-06T13:50:50.245 INFO:teuthology.orchestra.run.vm04.stdout: python3-ceph-common-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 2026-03-06T13:50:50.245 INFO:teuthology.orchestra.run.vm04.stdout: python3-certifi-2023.05.07-4.el9.noarch 2026-03-06T13:50:50.245 INFO:teuthology.orchestra.run.vm04.stdout: python3-cffi-1.14.5-5.el9.x86_64 2026-03-06T13:50:50.245 INFO:teuthology.orchestra.run.vm04.stdout: python3-chardet-4.0.0-5.el9.noarch 2026-03-06T13:50:50.245 INFO:teuthology.orchestra.run.vm04.stdout: python3-cheroot-10.0.1-4.el9.noarch 2026-03-06T13:50:50.245 INFO:teuthology.orchestra.run.vm04.stdout: python3-cherrypy-18.6.1-2.el9.noarch 2026-03-06T13:50:50.245 INFO:teuthology.orchestra.run.vm04.stdout: python3-cryptography-36.0.1-5.el9.x86_64 2026-03-06T13:50:50.245 INFO:teuthology.orchestra.run.vm04.stdout: python3-devel-3.9.25-3.el9.x86_64 2026-03-06T13:50:50.245 INFO:teuthology.orchestra.run.vm04.stdout: python3-google-auth-1:2.45.0-1.el9.noarch 2026-03-06T13:50:50.245 INFO:teuthology.orchestra.run.vm04.stdout: python3-grpcio-1.46.7-10.el9.x86_64 2026-03-06T13:50:50.245 INFO:teuthology.orchestra.run.vm04.stdout: python3-grpcio-tools-1.46.7-10.el9.x86_64 2026-03-06T13:50:50.245 INFO:teuthology.orchestra.run.vm04.stdout: python3-idna-2.10-7.el9.1.noarch 2026-03-06T13:50:50.245 INFO:teuthology.orchestra.run.vm04.stdout: python3-influxdb-5.3.1-1.el9.noarch 2026-03-06T13:50:50.245 INFO:teuthology.orchestra.run.vm04.stdout: python3-isodate-0.6.1-3.el9.noarch 2026-03-06T13:50:50.245 INFO:teuthology.orchestra.run.vm04.stdout: python3-jaraco-8.2.1-3.el9.noarch 2026-03-06T13:50:50.245 INFO:teuthology.orchestra.run.vm04.stdout: python3-jaraco-classes-3.2.1-5.el9.noarch 2026-03-06T13:50:50.245 INFO:teuthology.orchestra.run.vm04.stdout: python3-jaraco-collections-3.0.0-8.el9.noarch 2026-03-06T13:50:50.245 INFO:teuthology.orchestra.run.vm04.stdout: python3-jaraco-context-6.0.1-3.el9.noarch 2026-03-06T13:50:50.245 INFO:teuthology.orchestra.run.vm04.stdout: python3-jaraco-functools-3.5.0-2.el9.noarch 2026-03-06T13:50:50.245 INFO:teuthology.orchestra.run.vm04.stdout: python3-jaraco-text-4.0.0-2.el9.noarch 2026-03-06T13:50:50.245 INFO:teuthology.orchestra.run.vm04.stdout: python3-jinja2-2.11.3-8.el9.noarch 2026-03-06T13:50:50.245 INFO:teuthology.orchestra.run.vm04.stdout: python3-jsonpatch-1.21-16.el9.noarch 2026-03-06T13:50:50.245 INFO:teuthology.orchestra.run.vm04.stdout: python3-jsonpointer-2.0-4.el9.noarch 2026-03-06T13:50:50.245 INFO:teuthology.orchestra.run.vm04.stdout: python3-kubernetes-1:26.1.0-3.el9.noarch 2026-03-06T13:50:50.245 INFO:teuthology.orchestra.run.vm04.stdout: python3-libstoragemgmt-1.10.1-1.el9.x86_64 2026-03-06T13:50:50.245 INFO:teuthology.orchestra.run.vm04.stdout: python3-logutils-0.3.5-21.el9.noarch 2026-03-06T13:50:50.245 INFO:teuthology.orchestra.run.vm04.stdout: python3-lxml-4.6.5-3.el9.x86_64 2026-03-06T13:50:50.245 INFO:teuthology.orchestra.run.vm04.stdout: python3-mako-1.1.4-6.el9.noarch 2026-03-06T13:50:50.246 INFO:teuthology.orchestra.run.vm04.stdout: python3-markupsafe-1.1.1-12.el9.x86_64 2026-03-06T13:50:50.246 INFO:teuthology.orchestra.run.vm04.stdout: python3-more-itertools-8.12.0-2.el9.noarch 2026-03-06T13:50:50.246 INFO:teuthology.orchestra.run.vm04.stdout: python3-msgpack-1.0.3-2.el9.x86_64 2026-03-06T13:50:50.246 INFO:teuthology.orchestra.run.vm04.stdout: python3-natsort-7.1.1-5.el9.noarch 2026-03-06T13:50:50.246 INFO:teuthology.orchestra.run.vm04.stdout: python3-numpy-1:1.23.5-2.el9.x86_64 2026-03-06T13:50:50.246 INFO:teuthology.orchestra.run.vm04.stdout: python3-numpy-f2py-1:1.23.5-2.el9.x86_64 2026-03-06T13:50:50.246 INFO:teuthology.orchestra.run.vm04.stdout: python3-oauthlib-3.1.1-5.el9.noarch 2026-03-06T13:50:50.246 INFO:teuthology.orchestra.run.vm04.stdout: python3-packaging-20.9-5.el9.noarch 2026-03-06T13:50:50.246 INFO:teuthology.orchestra.run.vm04.stdout: python3-pecan-1.4.2-3.el9.noarch 2026-03-06T13:50:50.246 INFO:teuthology.orchestra.run.vm04.stdout: python3-ply-3.11-14.el9.noarch 2026-03-06T13:50:50.246 INFO:teuthology.orchestra.run.vm04.stdout: python3-portend-3.1.0-2.el9.noarch 2026-03-06T13:50:50.246 INFO:teuthology.orchestra.run.vm04.stdout: python3-prettytable-0.7.2-27.el9.noarch 2026-03-06T13:50:50.246 INFO:teuthology.orchestra.run.vm04.stdout: python3-protobuf-3.14.0-17.el9.noarch 2026-03-06T13:50:50.246 INFO:teuthology.orchestra.run.vm04.stdout: python3-pyOpenSSL-21.0.0-1.el9.noarch 2026-03-06T13:50:50.246 INFO:teuthology.orchestra.run.vm04.stdout: python3-pyasn1-0.4.8-7.el9.noarch 2026-03-06T13:50:50.246 INFO:teuthology.orchestra.run.vm04.stdout: python3-pyasn1-modules-0.4.8-7.el9.noarch 2026-03-06T13:50:50.246 INFO:teuthology.orchestra.run.vm04.stdout: python3-pycparser-2.20-6.el9.noarch 2026-03-06T13:50:50.246 INFO:teuthology.orchestra.run.vm04.stdout: python3-pysocks-1.7.1-12.el9.noarch 2026-03-06T13:50:50.246 INFO:teuthology.orchestra.run.vm04.stdout: python3-pytz-2021.1-5.el9.noarch 2026-03-06T13:50:50.246 INFO:teuthology.orchestra.run.vm04.stdout: python3-repoze-lru-0.7-16.el9.noarch 2026-03-06T13:50:50.246 INFO:teuthology.orchestra.run.vm04.stdout: python3-requests-2.25.1-10.el9.noarch 2026-03-06T13:50:50.246 INFO:teuthology.orchestra.run.vm04.stdout: python3-requests-oauthlib-1.3.0-12.el9.noarch 2026-03-06T13:50:50.246 INFO:teuthology.orchestra.run.vm04.stdout: python3-routes-2.5.1-5.el9.noarch 2026-03-06T13:50:50.246 INFO:teuthology.orchestra.run.vm04.stdout: python3-rsa-4.9-2.el9.noarch 2026-03-06T13:50:50.246 INFO:teuthology.orchestra.run.vm04.stdout: python3-saml-1.16.0-1.el9.noarch 2026-03-06T13:50:50.246 INFO:teuthology.orchestra.run.vm04.stdout: python3-scipy-1.9.3-2.el9.x86_64 2026-03-06T13:50:50.246 INFO:teuthology.orchestra.run.vm04.stdout: python3-tempora-5.0.0-2.el9.noarch 2026-03-06T13:50:50.246 INFO:teuthology.orchestra.run.vm04.stdout: python3-toml-0.10.2-6.el9.noarch 2026-03-06T13:50:50.246 INFO:teuthology.orchestra.run.vm04.stdout: python3-typing-extensions-4.15.0-1.el9.noarch 2026-03-06T13:50:50.246 INFO:teuthology.orchestra.run.vm04.stdout: python3-urllib3-1.26.5-7.el9.noarch 2026-03-06T13:50:50.246 INFO:teuthology.orchestra.run.vm04.stdout: python3-webob-1.8.8-2.el9.noarch 2026-03-06T13:50:50.246 INFO:teuthology.orchestra.run.vm04.stdout: python3-websocket-client-1.2.3-2.el9.noarch 2026-03-06T13:50:50.246 INFO:teuthology.orchestra.run.vm04.stdout: python3-werkzeug-2.0.3-3.el9.1.noarch 2026-03-06T13:50:50.246 INFO:teuthology.orchestra.run.vm04.stdout: python3-xmlsec-1.3.13-1.el9.x86_64 2026-03-06T13:50:50.246 INFO:teuthology.orchestra.run.vm04.stdout: python3-zc-lockfile-2.0-10.el9.noarch 2026-03-06T13:50:50.246 INFO:teuthology.orchestra.run.vm04.stdout: qatlib-25.08.0-2.el9.x86_64 2026-03-06T13:50:50.246 INFO:teuthology.orchestra.run.vm04.stdout: qatlib-service-25.08.0-2.el9.x86_64 2026-03-06T13:50:50.246 INFO:teuthology.orchestra.run.vm04.stdout: qatzip-libs-1.3.1-1.el9.x86_64 2026-03-06T13:50:50.246 INFO:teuthology.orchestra.run.vm04.stdout: rbd-mirror-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 2026-03-06T13:50:50.246 INFO:teuthology.orchestra.run.vm04.stdout: smartmontools-1:7.2-10.el9.x86_64 2026-03-06T13:50:50.246 INFO:teuthology.orchestra.run.vm04.stdout: xmlsec1-1.2.29-13.el9.x86_64 2026-03-06T13:50:50.247 INFO:teuthology.orchestra.run.vm04.stdout: xmlsec1-openssl-1.2.29-13.el9.x86_64 2026-03-06T13:50:50.247 INFO:teuthology.orchestra.run.vm04.stdout: 2026-03-06T13:50:50.247 INFO:teuthology.orchestra.run.vm04.stdout:Complete! 2026-03-06T13:50:50.306 INFO:teuthology.orchestra.run.vm06.stdout:Dependencies resolved. 2026-03-06T13:50:50.307 INFO:teuthology.orchestra.run.vm06.stdout:================================================================================ 2026-03-06T13:50:50.307 INFO:teuthology.orchestra.run.vm06.stdout: Package Arch Version Repository Size 2026-03-06T13:50:50.307 INFO:teuthology.orchestra.run.vm06.stdout:================================================================================ 2026-03-06T13:50:50.307 INFO:teuthology.orchestra.run.vm06.stdout:Removing: 2026-03-06T13:50:50.307 INFO:teuthology.orchestra.run.vm06.stdout: cephadm noarch 2:19.2.3-47.gc24117fd552.el9.clyso @ceph-noarch 775 k 2026-03-06T13:50:50.307 INFO:teuthology.orchestra.run.vm06.stdout: 2026-03-06T13:50:50.307 INFO:teuthology.orchestra.run.vm06.stdout:Transaction Summary 2026-03-06T13:50:50.307 INFO:teuthology.orchestra.run.vm06.stdout:================================================================================ 2026-03-06T13:50:50.307 INFO:teuthology.orchestra.run.vm06.stdout:Remove 1 Package 2026-03-06T13:50:50.307 INFO:teuthology.orchestra.run.vm06.stdout: 2026-03-06T13:50:50.307 INFO:teuthology.orchestra.run.vm06.stdout:Freed space: 775 k 2026-03-06T13:50:50.307 INFO:teuthology.orchestra.run.vm06.stdout:Running transaction check 2026-03-06T13:50:50.309 INFO:teuthology.orchestra.run.vm06.stdout:Transaction check succeeded. 2026-03-06T13:50:50.309 INFO:teuthology.orchestra.run.vm06.stdout:Running transaction test 2026-03-06T13:50:50.310 INFO:teuthology.orchestra.run.vm06.stdout:Transaction test succeeded. 2026-03-06T13:50:50.310 INFO:teuthology.orchestra.run.vm06.stdout:Running transaction 2026-03-06T13:50:50.327 INFO:teuthology.orchestra.run.vm06.stdout: Preparing : 1/1 2026-03-06T13:50:50.328 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : cephadm-2:19.2.3-47.gc24117fd552.el9.clyso.noarch 1/1 2026-03-06T13:50:50.380 INFO:teuthology.orchestra.run.vm02.stdout:Dependencies resolved. 2026-03-06T13:50:50.381 INFO:teuthology.orchestra.run.vm02.stdout:================================================================================ 2026-03-06T13:50:50.381 INFO:teuthology.orchestra.run.vm02.stdout: Package Arch Version Repository Size 2026-03-06T13:50:50.381 INFO:teuthology.orchestra.run.vm02.stdout:================================================================================ 2026-03-06T13:50:50.381 INFO:teuthology.orchestra.run.vm02.stdout:Removing: 2026-03-06T13:50:50.381 INFO:teuthology.orchestra.run.vm02.stdout: cephadm noarch 2:19.2.3-47.gc24117fd552.el9.clyso @ceph-noarch 775 k 2026-03-06T13:50:50.381 INFO:teuthology.orchestra.run.vm02.stdout: 2026-03-06T13:50:50.381 INFO:teuthology.orchestra.run.vm02.stdout:Transaction Summary 2026-03-06T13:50:50.381 INFO:teuthology.orchestra.run.vm02.stdout:================================================================================ 2026-03-06T13:50:50.381 INFO:teuthology.orchestra.run.vm02.stdout:Remove 1 Package 2026-03-06T13:50:50.381 INFO:teuthology.orchestra.run.vm02.stdout: 2026-03-06T13:50:50.381 INFO:teuthology.orchestra.run.vm02.stdout:Freed space: 775 k 2026-03-06T13:50:50.381 INFO:teuthology.orchestra.run.vm02.stdout:Running transaction check 2026-03-06T13:50:50.383 INFO:teuthology.orchestra.run.vm02.stdout:Transaction check succeeded. 2026-03-06T13:50:50.383 INFO:teuthology.orchestra.run.vm02.stdout:Running transaction test 2026-03-06T13:50:50.386 INFO:teuthology.orchestra.run.vm02.stdout:Transaction test succeeded. 2026-03-06T13:50:50.386 INFO:teuthology.orchestra.run.vm02.stdout:Running transaction 2026-03-06T13:50:50.404 INFO:teuthology.orchestra.run.vm02.stdout: Preparing : 1/1 2026-03-06T13:50:50.404 INFO:teuthology.orchestra.run.vm02.stdout: Erasing : cephadm-2:19.2.3-47.gc24117fd552.el9.clyso.noarch 1/1 2026-03-06T13:50:50.438 INFO:teuthology.orchestra.run.vm06.stdout: Running scriptlet: cephadm-2:19.2.3-47.gc24117fd552.el9.clyso.noarch 1/1 2026-03-06T13:50:50.481 INFO:teuthology.orchestra.run.vm04.stdout:Dependencies resolved. 2026-03-06T13:50:50.481 INFO:teuthology.orchestra.run.vm04.stdout:================================================================================ 2026-03-06T13:50:50.481 INFO:teuthology.orchestra.run.vm04.stdout: Package Arch Version Repository Size 2026-03-06T13:50:50.481 INFO:teuthology.orchestra.run.vm04.stdout:================================================================================ 2026-03-06T13:50:50.481 INFO:teuthology.orchestra.run.vm04.stdout:Removing: 2026-03-06T13:50:50.481 INFO:teuthology.orchestra.run.vm04.stdout: cephadm noarch 2:19.2.3-47.gc24117fd552.el9.clyso @ceph-noarch 775 k 2026-03-06T13:50:50.481 INFO:teuthology.orchestra.run.vm04.stdout: 2026-03-06T13:50:50.481 INFO:teuthology.orchestra.run.vm04.stdout:Transaction Summary 2026-03-06T13:50:50.481 INFO:teuthology.orchestra.run.vm04.stdout:================================================================================ 2026-03-06T13:50:50.481 INFO:teuthology.orchestra.run.vm04.stdout:Remove 1 Package 2026-03-06T13:50:50.481 INFO:teuthology.orchestra.run.vm04.stdout: 2026-03-06T13:50:50.481 INFO:teuthology.orchestra.run.vm04.stdout:Freed space: 775 k 2026-03-06T13:50:50.481 INFO:teuthology.orchestra.run.vm04.stdout:Running transaction check 2026-03-06T13:50:50.482 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : cephadm-2:19.2.3-47.gc24117fd552.el9.clyso.noarch 1/1 2026-03-06T13:50:50.482 INFO:teuthology.orchestra.run.vm06.stdout: 2026-03-06T13:50:50.482 INFO:teuthology.orchestra.run.vm06.stdout:Removed: 2026-03-06T13:50:50.482 INFO:teuthology.orchestra.run.vm06.stdout: cephadm-2:19.2.3-47.gc24117fd552.el9.clyso.noarch 2026-03-06T13:50:50.482 INFO:teuthology.orchestra.run.vm06.stdout: 2026-03-06T13:50:50.482 INFO:teuthology.orchestra.run.vm06.stdout:Complete! 2026-03-06T13:50:50.483 INFO:teuthology.orchestra.run.vm04.stdout:Transaction check succeeded. 2026-03-06T13:50:50.483 INFO:teuthology.orchestra.run.vm04.stdout:Running transaction test 2026-03-06T13:50:50.484 INFO:teuthology.orchestra.run.vm04.stdout:Transaction test succeeded. 2026-03-06T13:50:50.484 INFO:teuthology.orchestra.run.vm04.stdout:Running transaction 2026-03-06T13:50:50.501 INFO:teuthology.orchestra.run.vm04.stdout: Preparing : 1/1 2026-03-06T13:50:50.501 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : cephadm-2:19.2.3-47.gc24117fd552.el9.clyso.noarch 1/1 2026-03-06T13:50:50.509 INFO:teuthology.orchestra.run.vm02.stdout: Running scriptlet: cephadm-2:19.2.3-47.gc24117fd552.el9.clyso.noarch 1/1 2026-03-06T13:50:50.561 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : cephadm-2:19.2.3-47.gc24117fd552.el9.clyso.noarch 1/1 2026-03-06T13:50:50.562 INFO:teuthology.orchestra.run.vm02.stdout: 2026-03-06T13:50:50.562 INFO:teuthology.orchestra.run.vm02.stdout:Removed: 2026-03-06T13:50:50.562 INFO:teuthology.orchestra.run.vm02.stdout: cephadm-2:19.2.3-47.gc24117fd552.el9.clyso.noarch 2026-03-06T13:50:50.562 INFO:teuthology.orchestra.run.vm02.stdout: 2026-03-06T13:50:50.562 INFO:teuthology.orchestra.run.vm02.stdout:Complete! 2026-03-06T13:50:50.608 INFO:teuthology.orchestra.run.vm04.stdout: Running scriptlet: cephadm-2:19.2.3-47.gc24117fd552.el9.clyso.noarch 1/1 2026-03-06T13:50:50.645 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : cephadm-2:19.2.3-47.gc24117fd552.el9.clyso.noarch 1/1 2026-03-06T13:50:50.645 INFO:teuthology.orchestra.run.vm04.stdout: 2026-03-06T13:50:50.645 INFO:teuthology.orchestra.run.vm04.stdout:Removed: 2026-03-06T13:50:50.645 INFO:teuthology.orchestra.run.vm04.stdout: cephadm-2:19.2.3-47.gc24117fd552.el9.clyso.noarch 2026-03-06T13:50:50.645 INFO:teuthology.orchestra.run.vm04.stdout: 2026-03-06T13:50:50.645 INFO:teuthology.orchestra.run.vm04.stdout:Complete! 2026-03-06T13:50:50.678 INFO:teuthology.orchestra.run.vm06.stdout:No match for argument: ceph-immutable-object-cache 2026-03-06T13:50:50.678 INFO:teuthology.orchestra.run.vm06.stderr:No packages marked for removal. 2026-03-06T13:50:50.680 INFO:teuthology.orchestra.run.vm06.stdout:Dependencies resolved. 2026-03-06T13:50:50.681 INFO:teuthology.orchestra.run.vm06.stdout:Nothing to do. 2026-03-06T13:50:50.681 INFO:teuthology.orchestra.run.vm06.stdout:Complete! 2026-03-06T13:50:50.749 INFO:teuthology.orchestra.run.vm02.stdout:No match for argument: ceph-immutable-object-cache 2026-03-06T13:50:50.749 INFO:teuthology.orchestra.run.vm02.stderr:No packages marked for removal. 2026-03-06T13:50:50.751 INFO:teuthology.orchestra.run.vm02.stdout:Dependencies resolved. 2026-03-06T13:50:50.752 INFO:teuthology.orchestra.run.vm02.stdout:Nothing to do. 2026-03-06T13:50:50.752 INFO:teuthology.orchestra.run.vm02.stdout:Complete! 2026-03-06T13:50:50.839 INFO:teuthology.orchestra.run.vm04.stdout:No match for argument: ceph-immutable-object-cache 2026-03-06T13:50:50.839 INFO:teuthology.orchestra.run.vm04.stderr:No packages marked for removal. 2026-03-06T13:50:50.841 INFO:teuthology.orchestra.run.vm04.stdout:Dependencies resolved. 2026-03-06T13:50:50.842 INFO:teuthology.orchestra.run.vm04.stdout:Nothing to do. 2026-03-06T13:50:50.842 INFO:teuthology.orchestra.run.vm04.stdout:Complete! 2026-03-06T13:50:50.860 INFO:teuthology.orchestra.run.vm06.stdout:No match for argument: ceph-mgr 2026-03-06T13:50:50.860 INFO:teuthology.orchestra.run.vm06.stderr:No packages marked for removal. 2026-03-06T13:50:50.862 INFO:teuthology.orchestra.run.vm06.stdout:Dependencies resolved. 2026-03-06T13:50:50.863 INFO:teuthology.orchestra.run.vm06.stdout:Nothing to do. 2026-03-06T13:50:50.863 INFO:teuthology.orchestra.run.vm06.stdout:Complete! 2026-03-06T13:50:50.930 INFO:teuthology.orchestra.run.vm02.stdout:No match for argument: ceph-mgr 2026-03-06T13:50:50.930 INFO:teuthology.orchestra.run.vm02.stderr:No packages marked for removal. 2026-03-06T13:50:50.932 INFO:teuthology.orchestra.run.vm02.stdout:Dependencies resolved. 2026-03-06T13:50:50.932 INFO:teuthology.orchestra.run.vm02.stdout:Nothing to do. 2026-03-06T13:50:50.932 INFO:teuthology.orchestra.run.vm02.stdout:Complete! 2026-03-06T13:50:51.014 INFO:teuthology.orchestra.run.vm04.stdout:No match for argument: ceph-mgr 2026-03-06T13:50:51.014 INFO:teuthology.orchestra.run.vm04.stderr:No packages marked for removal. 2026-03-06T13:50:51.016 INFO:teuthology.orchestra.run.vm04.stdout:Dependencies resolved. 2026-03-06T13:50:51.017 INFO:teuthology.orchestra.run.vm04.stdout:Nothing to do. 2026-03-06T13:50:51.017 INFO:teuthology.orchestra.run.vm04.stdout:Complete! 2026-03-06T13:50:51.036 INFO:teuthology.orchestra.run.vm06.stdout:No match for argument: ceph-mgr-dashboard 2026-03-06T13:50:51.036 INFO:teuthology.orchestra.run.vm06.stderr:No packages marked for removal. 2026-03-06T13:50:51.039 INFO:teuthology.orchestra.run.vm06.stdout:Dependencies resolved. 2026-03-06T13:50:51.039 INFO:teuthology.orchestra.run.vm06.stdout:Nothing to do. 2026-03-06T13:50:51.039 INFO:teuthology.orchestra.run.vm06.stdout:Complete! 2026-03-06T13:50:51.105 INFO:teuthology.orchestra.run.vm02.stdout:No match for argument: ceph-mgr-dashboard 2026-03-06T13:50:51.105 INFO:teuthology.orchestra.run.vm02.stderr:No packages marked for removal. 2026-03-06T13:50:51.107 INFO:teuthology.orchestra.run.vm02.stdout:Dependencies resolved. 2026-03-06T13:50:51.108 INFO:teuthology.orchestra.run.vm02.stdout:Nothing to do. 2026-03-06T13:50:51.108 INFO:teuthology.orchestra.run.vm02.stdout:Complete! 2026-03-06T13:50:51.190 INFO:teuthology.orchestra.run.vm04.stdout:No match for argument: ceph-mgr-dashboard 2026-03-06T13:50:51.190 INFO:teuthology.orchestra.run.vm04.stderr:No packages marked for removal. 2026-03-06T13:50:51.192 INFO:teuthology.orchestra.run.vm04.stdout:Dependencies resolved. 2026-03-06T13:50:51.193 INFO:teuthology.orchestra.run.vm04.stdout:Nothing to do. 2026-03-06T13:50:51.193 INFO:teuthology.orchestra.run.vm04.stdout:Complete! 2026-03-06T13:50:51.214 INFO:teuthology.orchestra.run.vm06.stdout:No match for argument: ceph-mgr-diskprediction-local 2026-03-06T13:50:51.214 INFO:teuthology.orchestra.run.vm06.stderr:No packages marked for removal. 2026-03-06T13:50:51.216 INFO:teuthology.orchestra.run.vm06.stdout:Dependencies resolved. 2026-03-06T13:50:51.216 INFO:teuthology.orchestra.run.vm06.stdout:Nothing to do. 2026-03-06T13:50:51.217 INFO:teuthology.orchestra.run.vm06.stdout:Complete! 2026-03-06T13:50:51.279 INFO:teuthology.orchestra.run.vm02.stdout:No match for argument: ceph-mgr-diskprediction-local 2026-03-06T13:50:51.279 INFO:teuthology.orchestra.run.vm02.stderr:No packages marked for removal. 2026-03-06T13:50:51.281 INFO:teuthology.orchestra.run.vm02.stdout:Dependencies resolved. 2026-03-06T13:50:51.282 INFO:teuthology.orchestra.run.vm02.stdout:Nothing to do. 2026-03-06T13:50:51.282 INFO:teuthology.orchestra.run.vm02.stdout:Complete! 2026-03-06T13:50:51.364 INFO:teuthology.orchestra.run.vm04.stdout:No match for argument: ceph-mgr-diskprediction-local 2026-03-06T13:50:51.364 INFO:teuthology.orchestra.run.vm04.stderr:No packages marked for removal. 2026-03-06T13:50:51.367 INFO:teuthology.orchestra.run.vm04.stdout:Dependencies resolved. 2026-03-06T13:50:51.367 INFO:teuthology.orchestra.run.vm04.stdout:Nothing to do. 2026-03-06T13:50:51.367 INFO:teuthology.orchestra.run.vm04.stdout:Complete! 2026-03-06T13:50:51.387 INFO:teuthology.orchestra.run.vm06.stdout:No match for argument: ceph-mgr-rook 2026-03-06T13:50:51.387 INFO:teuthology.orchestra.run.vm06.stderr:No packages marked for removal. 2026-03-06T13:50:51.390 INFO:teuthology.orchestra.run.vm06.stdout:Dependencies resolved. 2026-03-06T13:50:51.390 INFO:teuthology.orchestra.run.vm06.stdout:Nothing to do. 2026-03-06T13:50:51.390 INFO:teuthology.orchestra.run.vm06.stdout:Complete! 2026-03-06T13:50:51.451 INFO:teuthology.orchestra.run.vm02.stdout:No match for argument: ceph-mgr-rook 2026-03-06T13:50:51.451 INFO:teuthology.orchestra.run.vm02.stderr:No packages marked for removal. 2026-03-06T13:50:51.453 INFO:teuthology.orchestra.run.vm02.stdout:Dependencies resolved. 2026-03-06T13:50:51.454 INFO:teuthology.orchestra.run.vm02.stdout:Nothing to do. 2026-03-06T13:50:51.454 INFO:teuthology.orchestra.run.vm02.stdout:Complete! 2026-03-06T13:50:51.539 INFO:teuthology.orchestra.run.vm04.stdout:No match for argument: ceph-mgr-rook 2026-03-06T13:50:51.539 INFO:teuthology.orchestra.run.vm04.stderr:No packages marked for removal. 2026-03-06T13:50:51.541 INFO:teuthology.orchestra.run.vm04.stdout:Dependencies resolved. 2026-03-06T13:50:51.542 INFO:teuthology.orchestra.run.vm04.stdout:Nothing to do. 2026-03-06T13:50:51.542 INFO:teuthology.orchestra.run.vm04.stdout:Complete! 2026-03-06T13:50:51.564 INFO:teuthology.orchestra.run.vm06.stdout:No match for argument: ceph-mgr-cephadm 2026-03-06T13:50:51.564 INFO:teuthology.orchestra.run.vm06.stderr:No packages marked for removal. 2026-03-06T13:50:51.566 INFO:teuthology.orchestra.run.vm06.stdout:Dependencies resolved. 2026-03-06T13:50:51.567 INFO:teuthology.orchestra.run.vm06.stdout:Nothing to do. 2026-03-06T13:50:51.567 INFO:teuthology.orchestra.run.vm06.stdout:Complete! 2026-03-06T13:50:51.625 INFO:teuthology.orchestra.run.vm02.stdout:No match for argument: ceph-mgr-cephadm 2026-03-06T13:50:51.626 INFO:teuthology.orchestra.run.vm02.stderr:No packages marked for removal. 2026-03-06T13:50:51.627 INFO:teuthology.orchestra.run.vm02.stdout:Dependencies resolved. 2026-03-06T13:50:51.628 INFO:teuthology.orchestra.run.vm02.stdout:Nothing to do. 2026-03-06T13:50:51.628 INFO:teuthology.orchestra.run.vm02.stdout:Complete! 2026-03-06T13:50:51.713 INFO:teuthology.orchestra.run.vm04.stdout:No match for argument: ceph-mgr-cephadm 2026-03-06T13:50:51.714 INFO:teuthology.orchestra.run.vm04.stderr:No packages marked for removal. 2026-03-06T13:50:51.716 INFO:teuthology.orchestra.run.vm04.stdout:Dependencies resolved. 2026-03-06T13:50:51.716 INFO:teuthology.orchestra.run.vm04.stdout:Nothing to do. 2026-03-06T13:50:51.716 INFO:teuthology.orchestra.run.vm04.stdout:Complete! 2026-03-06T13:50:51.747 INFO:teuthology.orchestra.run.vm06.stdout:Dependencies resolved. 2026-03-06T13:50:51.748 INFO:teuthology.orchestra.run.vm06.stdout:================================================================================ 2026-03-06T13:50:51.748 INFO:teuthology.orchestra.run.vm06.stdout: Package Arch Version Repository Size 2026-03-06T13:50:51.748 INFO:teuthology.orchestra.run.vm06.stdout:================================================================================ 2026-03-06T13:50:51.748 INFO:teuthology.orchestra.run.vm06.stdout:Removing: 2026-03-06T13:50:51.748 INFO:teuthology.orchestra.run.vm06.stdout: ceph-fuse x86_64 2:19.2.3-47.gc24117fd552.el9.clyso @ceph 3.6 M 2026-03-06T13:50:51.748 INFO:teuthology.orchestra.run.vm06.stdout:Removing unused dependencies: 2026-03-06T13:50:51.748 INFO:teuthology.orchestra.run.vm06.stdout: fuse x86_64 2.9.9-17.el9 @baseos 214 k 2026-03-06T13:50:51.748 INFO:teuthology.orchestra.run.vm06.stdout: 2026-03-06T13:50:51.748 INFO:teuthology.orchestra.run.vm06.stdout:Transaction Summary 2026-03-06T13:50:51.748 INFO:teuthology.orchestra.run.vm06.stdout:================================================================================ 2026-03-06T13:50:51.748 INFO:teuthology.orchestra.run.vm06.stdout:Remove 2 Packages 2026-03-06T13:50:51.748 INFO:teuthology.orchestra.run.vm06.stdout: 2026-03-06T13:50:51.748 INFO:teuthology.orchestra.run.vm06.stdout:Freed space: 3.8 M 2026-03-06T13:50:51.748 INFO:teuthology.orchestra.run.vm06.stdout:Running transaction check 2026-03-06T13:50:51.750 INFO:teuthology.orchestra.run.vm06.stdout:Transaction check succeeded. 2026-03-06T13:50:51.750 INFO:teuthology.orchestra.run.vm06.stdout:Running transaction test 2026-03-06T13:50:51.764 INFO:teuthology.orchestra.run.vm06.stdout:Transaction test succeeded. 2026-03-06T13:50:51.764 INFO:teuthology.orchestra.run.vm06.stdout:Running transaction 2026-03-06T13:50:51.792 INFO:teuthology.orchestra.run.vm06.stdout: Preparing : 1/1 2026-03-06T13:50:51.795 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : ceph-fuse-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 1/2 2026-03-06T13:50:51.808 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : fuse-2.9.9-17.el9.x86_64 2/2 2026-03-06T13:50:51.808 INFO:teuthology.orchestra.run.vm02.stdout:Dependencies resolved. 2026-03-06T13:50:51.809 INFO:teuthology.orchestra.run.vm02.stdout:================================================================================ 2026-03-06T13:50:51.809 INFO:teuthology.orchestra.run.vm02.stdout: Package Arch Version Repository Size 2026-03-06T13:50:51.809 INFO:teuthology.orchestra.run.vm02.stdout:================================================================================ 2026-03-06T13:50:51.809 INFO:teuthology.orchestra.run.vm02.stdout:Removing: 2026-03-06T13:50:51.809 INFO:teuthology.orchestra.run.vm02.stdout: ceph-fuse x86_64 2:19.2.3-47.gc24117fd552.el9.clyso @ceph 3.6 M 2026-03-06T13:50:51.809 INFO:teuthology.orchestra.run.vm02.stdout:Removing unused dependencies: 2026-03-06T13:50:51.809 INFO:teuthology.orchestra.run.vm02.stdout: fuse x86_64 2.9.9-17.el9 @baseos 214 k 2026-03-06T13:50:51.809 INFO:teuthology.orchestra.run.vm02.stdout: 2026-03-06T13:50:51.809 INFO:teuthology.orchestra.run.vm02.stdout:Transaction Summary 2026-03-06T13:50:51.809 INFO:teuthology.orchestra.run.vm02.stdout:================================================================================ 2026-03-06T13:50:51.809 INFO:teuthology.orchestra.run.vm02.stdout:Remove 2 Packages 2026-03-06T13:50:51.809 INFO:teuthology.orchestra.run.vm02.stdout: 2026-03-06T13:50:51.809 INFO:teuthology.orchestra.run.vm02.stdout:Freed space: 3.8 M 2026-03-06T13:50:51.809 INFO:teuthology.orchestra.run.vm02.stdout:Running transaction check 2026-03-06T13:50:51.811 INFO:teuthology.orchestra.run.vm02.stdout:Transaction check succeeded. 2026-03-06T13:50:51.811 INFO:teuthology.orchestra.run.vm02.stdout:Running transaction test 2026-03-06T13:50:51.824 INFO:teuthology.orchestra.run.vm02.stdout:Transaction test succeeded. 2026-03-06T13:50:51.824 INFO:teuthology.orchestra.run.vm02.stdout:Running transaction 2026-03-06T13:50:51.851 INFO:teuthology.orchestra.run.vm02.stdout: Preparing : 1/1 2026-03-06T13:50:51.855 INFO:teuthology.orchestra.run.vm02.stdout: Erasing : ceph-fuse-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 1/2 2026-03-06T13:50:51.869 INFO:teuthology.orchestra.run.vm02.stdout: Erasing : fuse-2.9.9-17.el9.x86_64 2/2 2026-03-06T13:50:51.873 INFO:teuthology.orchestra.run.vm06.stdout: Running scriptlet: fuse-2.9.9-17.el9.x86_64 2/2 2026-03-06T13:50:51.873 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : ceph-fuse-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 1/2 2026-03-06T13:50:51.899 INFO:teuthology.orchestra.run.vm04.stdout:Dependencies resolved. 2026-03-06T13:50:51.899 INFO:teuthology.orchestra.run.vm04.stdout:================================================================================ 2026-03-06T13:50:51.899 INFO:teuthology.orchestra.run.vm04.stdout: Package Arch Version Repository Size 2026-03-06T13:50:51.899 INFO:teuthology.orchestra.run.vm04.stdout:================================================================================ 2026-03-06T13:50:51.899 INFO:teuthology.orchestra.run.vm04.stdout:Removing: 2026-03-06T13:50:51.899 INFO:teuthology.orchestra.run.vm04.stdout: ceph-fuse x86_64 2:19.2.3-47.gc24117fd552.el9.clyso @ceph 3.6 M 2026-03-06T13:50:51.899 INFO:teuthology.orchestra.run.vm04.stdout:Removing unused dependencies: 2026-03-06T13:50:51.899 INFO:teuthology.orchestra.run.vm04.stdout: fuse x86_64 2.9.9-17.el9 @baseos 214 k 2026-03-06T13:50:51.899 INFO:teuthology.orchestra.run.vm04.stdout: 2026-03-06T13:50:51.899 INFO:teuthology.orchestra.run.vm04.stdout:Transaction Summary 2026-03-06T13:50:51.899 INFO:teuthology.orchestra.run.vm04.stdout:================================================================================ 2026-03-06T13:50:51.899 INFO:teuthology.orchestra.run.vm04.stdout:Remove 2 Packages 2026-03-06T13:50:51.899 INFO:teuthology.orchestra.run.vm04.stdout: 2026-03-06T13:50:51.899 INFO:teuthology.orchestra.run.vm04.stdout:Freed space: 3.8 M 2026-03-06T13:50:51.899 INFO:teuthology.orchestra.run.vm04.stdout:Running transaction check 2026-03-06T13:50:51.902 INFO:teuthology.orchestra.run.vm04.stdout:Transaction check succeeded. 2026-03-06T13:50:51.902 INFO:teuthology.orchestra.run.vm04.stdout:Running transaction test 2026-03-06T13:50:51.913 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : fuse-2.9.9-17.el9.x86_64 2/2 2026-03-06T13:50:51.913 INFO:teuthology.orchestra.run.vm06.stdout: 2026-03-06T13:50:51.914 INFO:teuthology.orchestra.run.vm06.stdout:Removed: 2026-03-06T13:50:51.914 INFO:teuthology.orchestra.run.vm06.stdout: ceph-fuse-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 fuse-2.9.9-17.el9.x86_64 2026-03-06T13:50:51.914 INFO:teuthology.orchestra.run.vm06.stdout: 2026-03-06T13:50:51.914 INFO:teuthology.orchestra.run.vm06.stdout:Complete! 2026-03-06T13:50:51.914 INFO:teuthology.orchestra.run.vm04.stdout:Transaction test succeeded. 2026-03-06T13:50:51.914 INFO:teuthology.orchestra.run.vm04.stdout:Running transaction 2026-03-06T13:50:51.935 INFO:teuthology.orchestra.run.vm02.stdout: Running scriptlet: fuse-2.9.9-17.el9.x86_64 2/2 2026-03-06T13:50:51.935 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : ceph-fuse-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 1/2 2026-03-06T13:50:51.943 INFO:teuthology.orchestra.run.vm04.stdout: Preparing : 1/1 2026-03-06T13:50:51.948 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : ceph-fuse-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 1/2 2026-03-06T13:50:51.963 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : fuse-2.9.9-17.el9.x86_64 2/2 2026-03-06T13:50:51.977 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : fuse-2.9.9-17.el9.x86_64 2/2 2026-03-06T13:50:51.977 INFO:teuthology.orchestra.run.vm02.stdout: 2026-03-06T13:50:51.977 INFO:teuthology.orchestra.run.vm02.stdout:Removed: 2026-03-06T13:50:51.977 INFO:teuthology.orchestra.run.vm02.stdout: ceph-fuse-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 fuse-2.9.9-17.el9.x86_64 2026-03-06T13:50:51.977 INFO:teuthology.orchestra.run.vm02.stdout: 2026-03-06T13:50:51.977 INFO:teuthology.orchestra.run.vm02.stdout:Complete! 2026-03-06T13:50:52.028 INFO:teuthology.orchestra.run.vm04.stdout: Running scriptlet: fuse-2.9.9-17.el9.x86_64 2/2 2026-03-06T13:50:52.028 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : ceph-fuse-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 1/2 2026-03-06T13:50:52.068 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : fuse-2.9.9-17.el9.x86_64 2/2 2026-03-06T13:50:52.068 INFO:teuthology.orchestra.run.vm04.stdout: 2026-03-06T13:50:52.068 INFO:teuthology.orchestra.run.vm04.stdout:Removed: 2026-03-06T13:50:52.068 INFO:teuthology.orchestra.run.vm04.stdout: ceph-fuse-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 fuse-2.9.9-17.el9.x86_64 2026-03-06T13:50:52.068 INFO:teuthology.orchestra.run.vm04.stdout: 2026-03-06T13:50:52.068 INFO:teuthology.orchestra.run.vm04.stdout:Complete! 2026-03-06T13:50:52.106 INFO:teuthology.orchestra.run.vm06.stdout:No match for argument: ceph-volume 2026-03-06T13:50:52.107 INFO:teuthology.orchestra.run.vm06.stderr:No packages marked for removal. 2026-03-06T13:50:52.109 INFO:teuthology.orchestra.run.vm06.stdout:Dependencies resolved. 2026-03-06T13:50:52.109 INFO:teuthology.orchestra.run.vm06.stdout:Nothing to do. 2026-03-06T13:50:52.109 INFO:teuthology.orchestra.run.vm06.stdout:Complete! 2026-03-06T13:50:52.168 INFO:teuthology.orchestra.run.vm02.stdout:No match for argument: ceph-volume 2026-03-06T13:50:52.168 INFO:teuthology.orchestra.run.vm02.stderr:No packages marked for removal. 2026-03-06T13:50:52.170 INFO:teuthology.orchestra.run.vm02.stdout:Dependencies resolved. 2026-03-06T13:50:52.171 INFO:teuthology.orchestra.run.vm02.stdout:Nothing to do. 2026-03-06T13:50:52.171 INFO:teuthology.orchestra.run.vm02.stdout:Complete! 2026-03-06T13:50:52.268 INFO:teuthology.orchestra.run.vm04.stdout:No match for argument: ceph-volume 2026-03-06T13:50:52.268 INFO:teuthology.orchestra.run.vm04.stderr:No packages marked for removal. 2026-03-06T13:50:52.270 INFO:teuthology.orchestra.run.vm04.stdout:Dependencies resolved. 2026-03-06T13:50:52.270 INFO:teuthology.orchestra.run.vm04.stdout:Nothing to do. 2026-03-06T13:50:52.270 INFO:teuthology.orchestra.run.vm04.stdout:Complete! 2026-03-06T13:50:52.292 INFO:teuthology.orchestra.run.vm06.stdout:Dependencies resolved. 2026-03-06T13:50:52.293 INFO:teuthology.orchestra.run.vm06.stdout:================================================================================ 2026-03-06T13:50:52.293 INFO:teuthology.orchestra.run.vm06.stdout: Package Arch Version Repo Size 2026-03-06T13:50:52.293 INFO:teuthology.orchestra.run.vm06.stdout:================================================================================ 2026-03-06T13:50:52.293 INFO:teuthology.orchestra.run.vm06.stdout:Removing: 2026-03-06T13:50:52.293 INFO:teuthology.orchestra.run.vm06.stdout: librados-devel x86_64 2:19.2.3-47.gc24117fd552.el9.clyso @ceph 456 k 2026-03-06T13:50:52.293 INFO:teuthology.orchestra.run.vm06.stdout:Removing dependent packages: 2026-03-06T13:50:52.293 INFO:teuthology.orchestra.run.vm06.stdout: libcephfs-devel x86_64 2:19.2.3-47.gc24117fd552.el9.clyso @ceph 153 k 2026-03-06T13:50:52.293 INFO:teuthology.orchestra.run.vm06.stdout: 2026-03-06T13:50:52.293 INFO:teuthology.orchestra.run.vm06.stdout:Transaction Summary 2026-03-06T13:50:52.293 INFO:teuthology.orchestra.run.vm06.stdout:================================================================================ 2026-03-06T13:50:52.293 INFO:teuthology.orchestra.run.vm06.stdout:Remove 2 Packages 2026-03-06T13:50:52.293 INFO:teuthology.orchestra.run.vm06.stdout: 2026-03-06T13:50:52.293 INFO:teuthology.orchestra.run.vm06.stdout:Freed space: 610 k 2026-03-06T13:50:52.293 INFO:teuthology.orchestra.run.vm06.stdout:Running transaction check 2026-03-06T13:50:52.295 INFO:teuthology.orchestra.run.vm06.stdout:Transaction check succeeded. 2026-03-06T13:50:52.295 INFO:teuthology.orchestra.run.vm06.stdout:Running transaction test 2026-03-06T13:50:52.305 INFO:teuthology.orchestra.run.vm06.stdout:Transaction test succeeded. 2026-03-06T13:50:52.305 INFO:teuthology.orchestra.run.vm06.stdout:Running transaction 2026-03-06T13:50:52.329 INFO:teuthology.orchestra.run.vm06.stdout: Preparing : 1/1 2026-03-06T13:50:52.331 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : libcephfs-devel-2:19.2.3-47.gc24117fd552.el9.clyso.x 1/2 2026-03-06T13:50:52.344 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : librados-devel-2:19.2.3-47.gc24117fd552.el9.clyso.x8 2/2 2026-03-06T13:50:52.350 INFO:teuthology.orchestra.run.vm02.stdout:Dependencies resolved. 2026-03-06T13:50:52.351 INFO:teuthology.orchestra.run.vm02.stdout:================================================================================ 2026-03-06T13:50:52.351 INFO:teuthology.orchestra.run.vm02.stdout: Package Arch Version Repo Size 2026-03-06T13:50:52.351 INFO:teuthology.orchestra.run.vm02.stdout:================================================================================ 2026-03-06T13:50:52.351 INFO:teuthology.orchestra.run.vm02.stdout:Removing: 2026-03-06T13:50:52.351 INFO:teuthology.orchestra.run.vm02.stdout: librados-devel x86_64 2:19.2.3-47.gc24117fd552.el9.clyso @ceph 456 k 2026-03-06T13:50:52.351 INFO:teuthology.orchestra.run.vm02.stdout:Removing dependent packages: 2026-03-06T13:50:52.351 INFO:teuthology.orchestra.run.vm02.stdout: libcephfs-devel x86_64 2:19.2.3-47.gc24117fd552.el9.clyso @ceph 153 k 2026-03-06T13:50:52.351 INFO:teuthology.orchestra.run.vm02.stdout: 2026-03-06T13:50:52.351 INFO:teuthology.orchestra.run.vm02.stdout:Transaction Summary 2026-03-06T13:50:52.351 INFO:teuthology.orchestra.run.vm02.stdout:================================================================================ 2026-03-06T13:50:52.351 INFO:teuthology.orchestra.run.vm02.stdout:Remove 2 Packages 2026-03-06T13:50:52.351 INFO:teuthology.orchestra.run.vm02.stdout: 2026-03-06T13:50:52.351 INFO:teuthology.orchestra.run.vm02.stdout:Freed space: 610 k 2026-03-06T13:50:52.351 INFO:teuthology.orchestra.run.vm02.stdout:Running transaction check 2026-03-06T13:50:52.353 INFO:teuthology.orchestra.run.vm02.stdout:Transaction check succeeded. 2026-03-06T13:50:52.353 INFO:teuthology.orchestra.run.vm02.stdout:Running transaction test 2026-03-06T13:50:52.362 INFO:teuthology.orchestra.run.vm02.stdout:Transaction test succeeded. 2026-03-06T13:50:52.362 INFO:teuthology.orchestra.run.vm02.stdout:Running transaction 2026-03-06T13:50:52.386 INFO:teuthology.orchestra.run.vm02.stdout: Preparing : 1/1 2026-03-06T13:50:52.388 INFO:teuthology.orchestra.run.vm02.stdout: Erasing : libcephfs-devel-2:19.2.3-47.gc24117fd552.el9.clyso.x 1/2 2026-03-06T13:50:52.402 INFO:teuthology.orchestra.run.vm02.stdout: Erasing : librados-devel-2:19.2.3-47.gc24117fd552.el9.clyso.x8 2/2 2026-03-06T13:50:52.404 INFO:teuthology.orchestra.run.vm06.stdout: Running scriptlet: librados-devel-2:19.2.3-47.gc24117fd552.el9.clyso.x8 2/2 2026-03-06T13:50:52.404 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : libcephfs-devel-2:19.2.3-47.gc24117fd552.el9.clyso.x 1/2 2026-03-06T13:50:52.447 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : librados-devel-2:19.2.3-47.gc24117fd552.el9.clyso.x8 2/2 2026-03-06T13:50:52.447 INFO:teuthology.orchestra.run.vm06.stdout: 2026-03-06T13:50:52.447 INFO:teuthology.orchestra.run.vm06.stdout:Removed: 2026-03-06T13:50:52.447 INFO:teuthology.orchestra.run.vm06.stdout: libcephfs-devel-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 2026-03-06T13:50:52.447 INFO:teuthology.orchestra.run.vm06.stdout: librados-devel-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 2026-03-06T13:50:52.447 INFO:teuthology.orchestra.run.vm06.stdout: 2026-03-06T13:50:52.447 INFO:teuthology.orchestra.run.vm06.stdout:Complete! 2026-03-06T13:50:52.449 INFO:teuthology.orchestra.run.vm04.stdout:Dependencies resolved. 2026-03-06T13:50:52.450 INFO:teuthology.orchestra.run.vm04.stdout:================================================================================ 2026-03-06T13:50:52.450 INFO:teuthology.orchestra.run.vm04.stdout: Package Arch Version Repo Size 2026-03-06T13:50:52.450 INFO:teuthology.orchestra.run.vm04.stdout:================================================================================ 2026-03-06T13:50:52.450 INFO:teuthology.orchestra.run.vm04.stdout:Removing: 2026-03-06T13:50:52.450 INFO:teuthology.orchestra.run.vm04.stdout: librados-devel x86_64 2:19.2.3-47.gc24117fd552.el9.clyso @ceph 456 k 2026-03-06T13:50:52.450 INFO:teuthology.orchestra.run.vm04.stdout:Removing dependent packages: 2026-03-06T13:50:52.450 INFO:teuthology.orchestra.run.vm04.stdout: libcephfs-devel x86_64 2:19.2.3-47.gc24117fd552.el9.clyso @ceph 153 k 2026-03-06T13:50:52.450 INFO:teuthology.orchestra.run.vm04.stdout: 2026-03-06T13:50:52.450 INFO:teuthology.orchestra.run.vm04.stdout:Transaction Summary 2026-03-06T13:50:52.450 INFO:teuthology.orchestra.run.vm04.stdout:================================================================================ 2026-03-06T13:50:52.450 INFO:teuthology.orchestra.run.vm04.stdout:Remove 2 Packages 2026-03-06T13:50:52.450 INFO:teuthology.orchestra.run.vm04.stdout: 2026-03-06T13:50:52.450 INFO:teuthology.orchestra.run.vm04.stdout:Freed space: 610 k 2026-03-06T13:50:52.450 INFO:teuthology.orchestra.run.vm04.stdout:Running transaction check 2026-03-06T13:50:52.452 INFO:teuthology.orchestra.run.vm04.stdout:Transaction check succeeded. 2026-03-06T13:50:52.452 INFO:teuthology.orchestra.run.vm04.stdout:Running transaction test 2026-03-06T13:50:52.463 INFO:teuthology.orchestra.run.vm04.stdout:Transaction test succeeded. 2026-03-06T13:50:52.463 INFO:teuthology.orchestra.run.vm04.stdout:Running transaction 2026-03-06T13:50:52.466 INFO:teuthology.orchestra.run.vm02.stdout: Running scriptlet: librados-devel-2:19.2.3-47.gc24117fd552.el9.clyso.x8 2/2 2026-03-06T13:50:52.466 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : libcephfs-devel-2:19.2.3-47.gc24117fd552.el9.clyso.x 1/2 2026-03-06T13:50:52.491 INFO:teuthology.orchestra.run.vm04.stdout: Preparing : 1/1 2026-03-06T13:50:52.493 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : libcephfs-devel-2:19.2.3-47.gc24117fd552.el9.clyso.x 1/2 2026-03-06T13:50:52.507 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : librados-devel-2:19.2.3-47.gc24117fd552.el9.clyso.x8 2/2 2026-03-06T13:50:52.512 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : librados-devel-2:19.2.3-47.gc24117fd552.el9.clyso.x8 2/2 2026-03-06T13:50:52.512 INFO:teuthology.orchestra.run.vm02.stdout: 2026-03-06T13:50:52.512 INFO:teuthology.orchestra.run.vm02.stdout:Removed: 2026-03-06T13:50:52.512 INFO:teuthology.orchestra.run.vm02.stdout: libcephfs-devel-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 2026-03-06T13:50:52.512 INFO:teuthology.orchestra.run.vm02.stdout: librados-devel-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 2026-03-06T13:50:52.512 INFO:teuthology.orchestra.run.vm02.stdout: 2026-03-06T13:50:52.512 INFO:teuthology.orchestra.run.vm02.stdout:Complete! 2026-03-06T13:50:52.563 INFO:teuthology.orchestra.run.vm04.stdout: Running scriptlet: librados-devel-2:19.2.3-47.gc24117fd552.el9.clyso.x8 2/2 2026-03-06T13:50:52.563 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : libcephfs-devel-2:19.2.3-47.gc24117fd552.el9.clyso.x 1/2 2026-03-06T13:50:52.608 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : librados-devel-2:19.2.3-47.gc24117fd552.el9.clyso.x8 2/2 2026-03-06T13:50:52.608 INFO:teuthology.orchestra.run.vm04.stdout: 2026-03-06T13:50:52.608 INFO:teuthology.orchestra.run.vm04.stdout:Removed: 2026-03-06T13:50:52.608 INFO:teuthology.orchestra.run.vm04.stdout: libcephfs-devel-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 2026-03-06T13:50:52.608 INFO:teuthology.orchestra.run.vm04.stdout: librados-devel-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 2026-03-06T13:50:52.608 INFO:teuthology.orchestra.run.vm04.stdout: 2026-03-06T13:50:52.608 INFO:teuthology.orchestra.run.vm04.stdout:Complete! 2026-03-06T13:50:52.640 INFO:teuthology.orchestra.run.vm06.stdout:Dependencies resolved. 2026-03-06T13:50:52.640 INFO:teuthology.orchestra.run.vm06.stdout:================================================================================ 2026-03-06T13:50:52.640 INFO:teuthology.orchestra.run.vm06.stdout: Package Arch Version Repo Size 2026-03-06T13:50:52.640 INFO:teuthology.orchestra.run.vm06.stdout:================================================================================ 2026-03-06T13:50:52.640 INFO:teuthology.orchestra.run.vm06.stdout:Removing: 2026-03-06T13:50:52.641 INFO:teuthology.orchestra.run.vm06.stdout: libcephfs2 x86_64 2:19.2.3-47.gc24117fd552.el9.clyso @ceph 3.0 M 2026-03-06T13:50:52.641 INFO:teuthology.orchestra.run.vm06.stdout:Removing dependent packages: 2026-03-06T13:50:52.641 INFO:teuthology.orchestra.run.vm06.stdout: python3-cephfs x86_64 2:19.2.3-47.gc24117fd552.el9.clyso @ceph 514 k 2026-03-06T13:50:52.641 INFO:teuthology.orchestra.run.vm06.stdout:Removing unused dependencies: 2026-03-06T13:50:52.641 INFO:teuthology.orchestra.run.vm06.stdout: python3-ceph-argparse x86_64 2:19.2.3-47.gc24117fd552.el9.clyso @ceph 187 k 2026-03-06T13:50:52.641 INFO:teuthology.orchestra.run.vm06.stdout: 2026-03-06T13:50:52.641 INFO:teuthology.orchestra.run.vm06.stdout:Transaction Summary 2026-03-06T13:50:52.641 INFO:teuthology.orchestra.run.vm06.stdout:================================================================================ 2026-03-06T13:50:52.641 INFO:teuthology.orchestra.run.vm06.stdout:Remove 3 Packages 2026-03-06T13:50:52.641 INFO:teuthology.orchestra.run.vm06.stdout: 2026-03-06T13:50:52.641 INFO:teuthology.orchestra.run.vm06.stdout:Freed space: 3.7 M 2026-03-06T13:50:52.641 INFO:teuthology.orchestra.run.vm06.stdout:Running transaction check 2026-03-06T13:50:52.643 INFO:teuthology.orchestra.run.vm06.stdout:Transaction check succeeded. 2026-03-06T13:50:52.643 INFO:teuthology.orchestra.run.vm06.stdout:Running transaction test 2026-03-06T13:50:52.659 INFO:teuthology.orchestra.run.vm06.stdout:Transaction test succeeded. 2026-03-06T13:50:52.659 INFO:teuthology.orchestra.run.vm06.stdout:Running transaction 2026-03-06T13:50:52.687 INFO:teuthology.orchestra.run.vm06.stdout: Preparing : 1/1 2026-03-06T13:50:52.690 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : python3-cephfs-2:19.2.3-47.gc24117fd552.el9.clyso.x8 1/3 2026-03-06T13:50:52.691 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : python3-ceph-argparse-2:19.2.3-47.gc24117fd552.el9.c 2/3 2026-03-06T13:50:52.691 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : libcephfs2-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 3/3 2026-03-06T13:50:52.697 INFO:teuthology.orchestra.run.vm02.stdout:Dependencies resolved. 2026-03-06T13:50:52.697 INFO:teuthology.orchestra.run.vm02.stdout:================================================================================ 2026-03-06T13:50:52.698 INFO:teuthology.orchestra.run.vm02.stdout: Package Arch Version Repo Size 2026-03-06T13:50:52.698 INFO:teuthology.orchestra.run.vm02.stdout:================================================================================ 2026-03-06T13:50:52.698 INFO:teuthology.orchestra.run.vm02.stdout:Removing: 2026-03-06T13:50:52.698 INFO:teuthology.orchestra.run.vm02.stdout: libcephfs2 x86_64 2:19.2.3-47.gc24117fd552.el9.clyso @ceph 3.0 M 2026-03-06T13:50:52.698 INFO:teuthology.orchestra.run.vm02.stdout:Removing dependent packages: 2026-03-06T13:50:52.698 INFO:teuthology.orchestra.run.vm02.stdout: python3-cephfs x86_64 2:19.2.3-47.gc24117fd552.el9.clyso @ceph 514 k 2026-03-06T13:50:52.698 INFO:teuthology.orchestra.run.vm02.stdout:Removing unused dependencies: 2026-03-06T13:50:52.698 INFO:teuthology.orchestra.run.vm02.stdout: python3-ceph-argparse x86_64 2:19.2.3-47.gc24117fd552.el9.clyso @ceph 187 k 2026-03-06T13:50:52.698 INFO:teuthology.orchestra.run.vm02.stdout: 2026-03-06T13:50:52.698 INFO:teuthology.orchestra.run.vm02.stdout:Transaction Summary 2026-03-06T13:50:52.698 INFO:teuthology.orchestra.run.vm02.stdout:================================================================================ 2026-03-06T13:50:52.698 INFO:teuthology.orchestra.run.vm02.stdout:Remove 3 Packages 2026-03-06T13:50:52.698 INFO:teuthology.orchestra.run.vm02.stdout: 2026-03-06T13:50:52.698 INFO:teuthology.orchestra.run.vm02.stdout:Freed space: 3.7 M 2026-03-06T13:50:52.698 INFO:teuthology.orchestra.run.vm02.stdout:Running transaction check 2026-03-06T13:50:52.700 INFO:teuthology.orchestra.run.vm02.stdout:Transaction check succeeded. 2026-03-06T13:50:52.700 INFO:teuthology.orchestra.run.vm02.stdout:Running transaction test 2026-03-06T13:50:52.715 INFO:teuthology.orchestra.run.vm02.stdout:Transaction test succeeded. 2026-03-06T13:50:52.715 INFO:teuthology.orchestra.run.vm02.stdout:Running transaction 2026-03-06T13:50:52.743 INFO:teuthology.orchestra.run.vm02.stdout: Preparing : 1/1 2026-03-06T13:50:52.746 INFO:teuthology.orchestra.run.vm02.stdout: Erasing : python3-cephfs-2:19.2.3-47.gc24117fd552.el9.clyso.x8 1/3 2026-03-06T13:50:52.749 INFO:teuthology.orchestra.run.vm02.stdout: Erasing : python3-ceph-argparse-2:19.2.3-47.gc24117fd552.el9.c 2/3 2026-03-06T13:50:52.749 INFO:teuthology.orchestra.run.vm02.stdout: Erasing : libcephfs2-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 3/3 2026-03-06T13:50:52.751 INFO:teuthology.orchestra.run.vm06.stdout: Running scriptlet: libcephfs2-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 3/3 2026-03-06T13:50:52.751 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : libcephfs2-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 1/3 2026-03-06T13:50:52.751 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-ceph-argparse-2:19.2.3-47.gc24117fd552.el9.c 2/3 2026-03-06T13:50:52.790 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-cephfs-2:19.2.3-47.gc24117fd552.el9.clyso.x8 3/3 2026-03-06T13:50:52.790 INFO:teuthology.orchestra.run.vm06.stdout: 2026-03-06T13:50:52.790 INFO:teuthology.orchestra.run.vm06.stdout:Removed: 2026-03-06T13:50:52.790 INFO:teuthology.orchestra.run.vm06.stdout: libcephfs2-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 2026-03-06T13:50:52.790 INFO:teuthology.orchestra.run.vm06.stdout: python3-ceph-argparse-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 2026-03-06T13:50:52.791 INFO:teuthology.orchestra.run.vm06.stdout: python3-cephfs-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 2026-03-06T13:50:52.791 INFO:teuthology.orchestra.run.vm06.stdout: 2026-03-06T13:50:52.791 INFO:teuthology.orchestra.run.vm06.stdout:Complete! 2026-03-06T13:50:52.801 INFO:teuthology.orchestra.run.vm04.stdout:Dependencies resolved. 2026-03-06T13:50:52.801 INFO:teuthology.orchestra.run.vm04.stdout:================================================================================ 2026-03-06T13:50:52.801 INFO:teuthology.orchestra.run.vm04.stdout: Package Arch Version Repo Size 2026-03-06T13:50:52.801 INFO:teuthology.orchestra.run.vm04.stdout:================================================================================ 2026-03-06T13:50:52.801 INFO:teuthology.orchestra.run.vm04.stdout:Removing: 2026-03-06T13:50:52.801 INFO:teuthology.orchestra.run.vm04.stdout: libcephfs2 x86_64 2:19.2.3-47.gc24117fd552.el9.clyso @ceph 3.0 M 2026-03-06T13:50:52.801 INFO:teuthology.orchestra.run.vm04.stdout:Removing dependent packages: 2026-03-06T13:50:52.802 INFO:teuthology.orchestra.run.vm04.stdout: python3-cephfs x86_64 2:19.2.3-47.gc24117fd552.el9.clyso @ceph 514 k 2026-03-06T13:50:52.802 INFO:teuthology.orchestra.run.vm04.stdout:Removing unused dependencies: 2026-03-06T13:50:52.802 INFO:teuthology.orchestra.run.vm04.stdout: python3-ceph-argparse x86_64 2:19.2.3-47.gc24117fd552.el9.clyso @ceph 187 k 2026-03-06T13:50:52.802 INFO:teuthology.orchestra.run.vm04.stdout: 2026-03-06T13:50:52.802 INFO:teuthology.orchestra.run.vm04.stdout:Transaction Summary 2026-03-06T13:50:52.802 INFO:teuthology.orchestra.run.vm04.stdout:================================================================================ 2026-03-06T13:50:52.802 INFO:teuthology.orchestra.run.vm04.stdout:Remove 3 Packages 2026-03-06T13:50:52.802 INFO:teuthology.orchestra.run.vm04.stdout: 2026-03-06T13:50:52.802 INFO:teuthology.orchestra.run.vm04.stdout:Freed space: 3.7 M 2026-03-06T13:50:52.802 INFO:teuthology.orchestra.run.vm04.stdout:Running transaction check 2026-03-06T13:50:52.803 INFO:teuthology.orchestra.run.vm04.stdout:Transaction check succeeded. 2026-03-06T13:50:52.803 INFO:teuthology.orchestra.run.vm04.stdout:Running transaction test 2026-03-06T13:50:52.811 INFO:teuthology.orchestra.run.vm02.stdout: Running scriptlet: libcephfs2-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 3/3 2026-03-06T13:50:52.811 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : libcephfs2-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 1/3 2026-03-06T13:50:52.811 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : python3-ceph-argparse-2:19.2.3-47.gc24117fd552.el9.c 2/3 2026-03-06T13:50:52.819 INFO:teuthology.orchestra.run.vm04.stdout:Transaction test succeeded. 2026-03-06T13:50:52.819 INFO:teuthology.orchestra.run.vm04.stdout:Running transaction 2026-03-06T13:50:52.846 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : python3-cephfs-2:19.2.3-47.gc24117fd552.el9.clyso.x8 3/3 2026-03-06T13:50:52.846 INFO:teuthology.orchestra.run.vm02.stdout: 2026-03-06T13:50:52.846 INFO:teuthology.orchestra.run.vm02.stdout:Removed: 2026-03-06T13:50:52.846 INFO:teuthology.orchestra.run.vm02.stdout: libcephfs2-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 2026-03-06T13:50:52.846 INFO:teuthology.orchestra.run.vm02.stdout: python3-ceph-argparse-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 2026-03-06T13:50:52.846 INFO:teuthology.orchestra.run.vm02.stdout: python3-cephfs-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 2026-03-06T13:50:52.846 INFO:teuthology.orchestra.run.vm02.stdout: 2026-03-06T13:50:52.846 INFO:teuthology.orchestra.run.vm02.stdout:Complete! 2026-03-06T13:50:52.848 INFO:teuthology.orchestra.run.vm04.stdout: Preparing : 1/1 2026-03-06T13:50:52.850 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : python3-cephfs-2:19.2.3-47.gc24117fd552.el9.clyso.x8 1/3 2026-03-06T13:50:52.851 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : python3-ceph-argparse-2:19.2.3-47.gc24117fd552.el9.c 2/3 2026-03-06T13:50:52.851 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : libcephfs2-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 3/3 2026-03-06T13:50:52.914 INFO:teuthology.orchestra.run.vm04.stdout: Running scriptlet: libcephfs2-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 3/3 2026-03-06T13:50:52.914 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : libcephfs2-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 1/3 2026-03-06T13:50:52.914 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-ceph-argparse-2:19.2.3-47.gc24117fd552.el9.c 2/3 2026-03-06T13:50:52.949 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-cephfs-2:19.2.3-47.gc24117fd552.el9.clyso.x8 3/3 2026-03-06T13:50:52.949 INFO:teuthology.orchestra.run.vm04.stdout: 2026-03-06T13:50:52.949 INFO:teuthology.orchestra.run.vm04.stdout:Removed: 2026-03-06T13:50:52.949 INFO:teuthology.orchestra.run.vm04.stdout: libcephfs2-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 2026-03-06T13:50:52.949 INFO:teuthology.orchestra.run.vm04.stdout: python3-ceph-argparse-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 2026-03-06T13:50:52.949 INFO:teuthology.orchestra.run.vm04.stdout: python3-cephfs-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 2026-03-06T13:50:52.949 INFO:teuthology.orchestra.run.vm04.stdout: 2026-03-06T13:50:52.949 INFO:teuthology.orchestra.run.vm04.stdout:Complete! 2026-03-06T13:50:52.970 INFO:teuthology.orchestra.run.vm06.stdout:No match for argument: libcephfs-devel 2026-03-06T13:50:52.970 INFO:teuthology.orchestra.run.vm06.stderr:No packages marked for removal. 2026-03-06T13:50:52.972 INFO:teuthology.orchestra.run.vm06.stdout:Dependencies resolved. 2026-03-06T13:50:52.973 INFO:teuthology.orchestra.run.vm06.stdout:Nothing to do. 2026-03-06T13:50:52.973 INFO:teuthology.orchestra.run.vm06.stdout:Complete! 2026-03-06T13:50:53.019 INFO:teuthology.orchestra.run.vm02.stdout:No match for argument: libcephfs-devel 2026-03-06T13:50:53.019 INFO:teuthology.orchestra.run.vm02.stderr:No packages marked for removal. 2026-03-06T13:50:53.021 INFO:teuthology.orchestra.run.vm02.stdout:Dependencies resolved. 2026-03-06T13:50:53.022 INFO:teuthology.orchestra.run.vm02.stdout:Nothing to do. 2026-03-06T13:50:53.022 INFO:teuthology.orchestra.run.vm02.stdout:Complete! 2026-03-06T13:50:53.112 INFO:teuthology.orchestra.run.vm04.stdout:No match for argument: libcephfs-devel 2026-03-06T13:50:53.112 INFO:teuthology.orchestra.run.vm04.stderr:No packages marked for removal. 2026-03-06T13:50:53.115 INFO:teuthology.orchestra.run.vm04.stdout:Dependencies resolved. 2026-03-06T13:50:53.115 INFO:teuthology.orchestra.run.vm04.stdout:Nothing to do. 2026-03-06T13:50:53.115 INFO:teuthology.orchestra.run.vm04.stdout:Complete! 2026-03-06T13:50:53.167 INFO:teuthology.orchestra.run.vm06.stdout:Dependencies resolved. 2026-03-06T13:50:53.169 INFO:teuthology.orchestra.run.vm06.stdout:================================================================================ 2026-03-06T13:50:53.169 INFO:teuthology.orchestra.run.vm06.stdout: Package Arch Version Repository Size 2026-03-06T13:50:53.169 INFO:teuthology.orchestra.run.vm06.stdout:================================================================================ 2026-03-06T13:50:53.169 INFO:teuthology.orchestra.run.vm06.stdout:Removing: 2026-03-06T13:50:53.169 INFO:teuthology.orchestra.run.vm06.stdout: librados2 x86_64 2:19.2.3-47.gc24117fd552.el9.clyso @ceph 13 M 2026-03-06T13:50:53.169 INFO:teuthology.orchestra.run.vm06.stdout:Removing dependent packages: 2026-03-06T13:50:53.169 INFO:teuthology.orchestra.run.vm06.stdout: python3-rados x86_64 2:19.2.3-47.gc24117fd552.el9.clyso @ceph 1.1 M 2026-03-06T13:50:53.169 INFO:teuthology.orchestra.run.vm06.stdout: python3-rbd x86_64 2:19.2.3-47.gc24117fd552.el9.clyso @ceph 1.1 M 2026-03-06T13:50:53.169 INFO:teuthology.orchestra.run.vm06.stdout: python3-rgw x86_64 2:19.2.3-47.gc24117fd552.el9.clyso @ceph 265 k 2026-03-06T13:50:53.169 INFO:teuthology.orchestra.run.vm06.stdout: qemu-kvm-block-rbd x86_64 17:10.1.0-15.el9 @appstream 37 k 2026-03-06T13:50:53.169 INFO:teuthology.orchestra.run.vm06.stdout: rbd-fuse x86_64 2:19.2.3-47.gc24117fd552.el9.clyso @ceph 227 k 2026-03-06T13:50:53.169 INFO:teuthology.orchestra.run.vm06.stdout: rbd-nbd x86_64 2:19.2.3-47.gc24117fd552.el9.clyso @ceph 494 k 2026-03-06T13:50:53.169 INFO:teuthology.orchestra.run.vm06.stdout:Removing unused dependencies: 2026-03-06T13:50:53.169 INFO:teuthology.orchestra.run.vm06.stdout: boost-program-options 2026-03-06T13:50:53.169 INFO:teuthology.orchestra.run.vm06.stdout: x86_64 1.75.0-13.el9 @appstream 276 k 2026-03-06T13:50:53.169 INFO:teuthology.orchestra.run.vm06.stdout: libarrow x86_64 9.0.0-15.el9 @epel 18 M 2026-03-06T13:50:53.169 INFO:teuthology.orchestra.run.vm06.stdout: libarrow-doc noarch 9.0.0-15.el9 @epel 122 k 2026-03-06T13:50:53.169 INFO:teuthology.orchestra.run.vm06.stdout: libnbd x86_64 1.20.3-4.el9 @appstream 453 k 2026-03-06T13:50:53.169 INFO:teuthology.orchestra.run.vm06.stdout: libpmemobj x86_64 1.12.1-1.el9 @appstream 383 k 2026-03-06T13:50:53.169 INFO:teuthology.orchestra.run.vm06.stdout: librabbitmq x86_64 0.11.0-7.el9 @appstream 102 k 2026-03-06T13:50:53.169 INFO:teuthology.orchestra.run.vm06.stdout: librbd1 x86_64 2:19.2.3-47.gc24117fd552.el9.clyso @ceph 13 M 2026-03-06T13:50:53.169 INFO:teuthology.orchestra.run.vm06.stdout: librdkafka x86_64 1.6.1-102.el9 @appstream 2.0 M 2026-03-06T13:50:53.169 INFO:teuthology.orchestra.run.vm06.stdout: librgw2 x86_64 2:19.2.3-47.gc24117fd552.el9.clyso @ceph 19 M 2026-03-06T13:50:53.169 INFO:teuthology.orchestra.run.vm06.stdout: lttng-ust x86_64 2.12.0-6.el9 @appstream 1.0 M 2026-03-06T13:50:53.169 INFO:teuthology.orchestra.run.vm06.stdout: parquet-libs x86_64 9.0.0-15.el9 @epel 2.8 M 2026-03-06T13:50:53.169 INFO:teuthology.orchestra.run.vm06.stdout: re2 x86_64 1:20211101-20.el9 @epel 472 k 2026-03-06T13:50:53.169 INFO:teuthology.orchestra.run.vm06.stdout: thrift x86_64 0.15.0-4.el9 @epel 4.8 M 2026-03-06T13:50:53.170 INFO:teuthology.orchestra.run.vm06.stdout: 2026-03-06T13:50:53.170 INFO:teuthology.orchestra.run.vm06.stdout:Transaction Summary 2026-03-06T13:50:53.170 INFO:teuthology.orchestra.run.vm06.stdout:================================================================================ 2026-03-06T13:50:53.170 INFO:teuthology.orchestra.run.vm06.stdout:Remove 20 Packages 2026-03-06T13:50:53.170 INFO:teuthology.orchestra.run.vm06.stdout: 2026-03-06T13:50:53.170 INFO:teuthology.orchestra.run.vm06.stdout:Freed space: 79 M 2026-03-06T13:50:53.170 INFO:teuthology.orchestra.run.vm06.stdout:Running transaction check 2026-03-06T13:50:53.174 INFO:teuthology.orchestra.run.vm06.stdout:Transaction check succeeded. 2026-03-06T13:50:53.174 INFO:teuthology.orchestra.run.vm06.stdout:Running transaction test 2026-03-06T13:50:53.196 INFO:teuthology.orchestra.run.vm06.stdout:Transaction test succeeded. 2026-03-06T13:50:53.196 INFO:teuthology.orchestra.run.vm06.stdout:Running transaction 2026-03-06T13:50:53.197 INFO:teuthology.orchestra.run.vm02.stdout:Dependencies resolved. 2026-03-06T13:50:53.198 INFO:teuthology.orchestra.run.vm02.stdout:================================================================================ 2026-03-06T13:50:53.198 INFO:teuthology.orchestra.run.vm02.stdout: Package Arch Version Repository Size 2026-03-06T13:50:53.198 INFO:teuthology.orchestra.run.vm02.stdout:================================================================================ 2026-03-06T13:50:53.199 INFO:teuthology.orchestra.run.vm02.stdout:Removing: 2026-03-06T13:50:53.199 INFO:teuthology.orchestra.run.vm02.stdout: librados2 x86_64 2:19.2.3-47.gc24117fd552.el9.clyso @ceph 13 M 2026-03-06T13:50:53.199 INFO:teuthology.orchestra.run.vm02.stdout:Removing dependent packages: 2026-03-06T13:50:53.199 INFO:teuthology.orchestra.run.vm02.stdout: python3-rados x86_64 2:19.2.3-47.gc24117fd552.el9.clyso @ceph 1.1 M 2026-03-06T13:50:53.199 INFO:teuthology.orchestra.run.vm02.stdout: python3-rbd x86_64 2:19.2.3-47.gc24117fd552.el9.clyso @ceph 1.1 M 2026-03-06T13:50:53.199 INFO:teuthology.orchestra.run.vm02.stdout: python3-rgw x86_64 2:19.2.3-47.gc24117fd552.el9.clyso @ceph 265 k 2026-03-06T13:50:53.199 INFO:teuthology.orchestra.run.vm02.stdout: qemu-kvm-block-rbd x86_64 17:10.1.0-15.el9 @appstream 37 k 2026-03-06T13:50:53.199 INFO:teuthology.orchestra.run.vm02.stdout: rbd-fuse x86_64 2:19.2.3-47.gc24117fd552.el9.clyso @ceph 227 k 2026-03-06T13:50:53.199 INFO:teuthology.orchestra.run.vm02.stdout: rbd-nbd x86_64 2:19.2.3-47.gc24117fd552.el9.clyso @ceph 494 k 2026-03-06T13:50:53.199 INFO:teuthology.orchestra.run.vm02.stdout:Removing unused dependencies: 2026-03-06T13:50:53.199 INFO:teuthology.orchestra.run.vm02.stdout: boost-program-options 2026-03-06T13:50:53.199 INFO:teuthology.orchestra.run.vm02.stdout: x86_64 1.75.0-13.el9 @appstream 276 k 2026-03-06T13:50:53.199 INFO:teuthology.orchestra.run.vm02.stdout: libarrow x86_64 9.0.0-15.el9 @epel 18 M 2026-03-06T13:50:53.199 INFO:teuthology.orchestra.run.vm02.stdout: libarrow-doc noarch 9.0.0-15.el9 @epel 122 k 2026-03-06T13:50:53.199 INFO:teuthology.orchestra.run.vm02.stdout: libnbd x86_64 1.20.3-4.el9 @appstream 453 k 2026-03-06T13:50:53.199 INFO:teuthology.orchestra.run.vm02.stdout: libpmemobj x86_64 1.12.1-1.el9 @appstream 383 k 2026-03-06T13:50:53.199 INFO:teuthology.orchestra.run.vm02.stdout: librabbitmq x86_64 0.11.0-7.el9 @appstream 102 k 2026-03-06T13:50:53.199 INFO:teuthology.orchestra.run.vm02.stdout: librbd1 x86_64 2:19.2.3-47.gc24117fd552.el9.clyso @ceph 13 M 2026-03-06T13:50:53.199 INFO:teuthology.orchestra.run.vm02.stdout: librdkafka x86_64 1.6.1-102.el9 @appstream 2.0 M 2026-03-06T13:50:53.199 INFO:teuthology.orchestra.run.vm02.stdout: librgw2 x86_64 2:19.2.3-47.gc24117fd552.el9.clyso @ceph 19 M 2026-03-06T13:50:53.199 INFO:teuthology.orchestra.run.vm02.stdout: lttng-ust x86_64 2.12.0-6.el9 @appstream 1.0 M 2026-03-06T13:50:53.199 INFO:teuthology.orchestra.run.vm02.stdout: parquet-libs x86_64 9.0.0-15.el9 @epel 2.8 M 2026-03-06T13:50:53.199 INFO:teuthology.orchestra.run.vm02.stdout: re2 x86_64 1:20211101-20.el9 @epel 472 k 2026-03-06T13:50:53.199 INFO:teuthology.orchestra.run.vm02.stdout: thrift x86_64 0.15.0-4.el9 @epel 4.8 M 2026-03-06T13:50:53.199 INFO:teuthology.orchestra.run.vm02.stdout: 2026-03-06T13:50:53.199 INFO:teuthology.orchestra.run.vm02.stdout:Transaction Summary 2026-03-06T13:50:53.199 INFO:teuthology.orchestra.run.vm02.stdout:================================================================================ 2026-03-06T13:50:53.199 INFO:teuthology.orchestra.run.vm02.stdout:Remove 20 Packages 2026-03-06T13:50:53.199 INFO:teuthology.orchestra.run.vm02.stdout: 2026-03-06T13:50:53.199 INFO:teuthology.orchestra.run.vm02.stdout:Freed space: 79 M 2026-03-06T13:50:53.199 INFO:teuthology.orchestra.run.vm02.stdout:Running transaction check 2026-03-06T13:50:53.203 INFO:teuthology.orchestra.run.vm02.stdout:Transaction check succeeded. 2026-03-06T13:50:53.203 INFO:teuthology.orchestra.run.vm02.stdout:Running transaction test 2026-03-06T13:50:53.224 INFO:teuthology.orchestra.run.vm02.stdout:Transaction test succeeded. 2026-03-06T13:50:53.224 INFO:teuthology.orchestra.run.vm02.stdout:Running transaction 2026-03-06T13:50:53.238 INFO:teuthology.orchestra.run.vm06.stdout: Preparing : 1/1 2026-03-06T13:50:53.241 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : rbd-nbd-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 1/20 2026-03-06T13:50:53.243 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : rbd-fuse-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 2/20 2026-03-06T13:50:53.246 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : python3-rgw-2:19.2.3-47.gc24117fd552.el9.clyso.x86 3/20 2026-03-06T13:50:53.246 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : librgw2-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 4/20 2026-03-06T13:50:53.259 INFO:teuthology.orchestra.run.vm06.stdout: Running scriptlet: librgw2-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 4/20 2026-03-06T13:50:53.261 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : parquet-libs-9.0.0-15.el9.x86_64 5/20 2026-03-06T13:50:53.263 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : python3-rbd-2:19.2.3-47.gc24117fd552.el9.clyso.x86 6/20 2026-03-06T13:50:53.264 INFO:teuthology.orchestra.run.vm02.stdout: Preparing : 1/1 2026-03-06T13:50:53.264 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : python3-rados-2:19.2.3-47.gc24117fd552.el9.clyso.x 7/20 2026-03-06T13:50:53.266 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : qemu-kvm-block-rbd-17:10.1.0-15.el9.x86_64 8/20 2026-03-06T13:50:53.267 INFO:teuthology.orchestra.run.vm02.stdout: Erasing : rbd-nbd-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 1/20 2026-03-06T13:50:53.268 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : libarrow-doc-9.0.0-15.el9.noarch 9/20 2026-03-06T13:50:53.268 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : librbd1-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 10/20 2026-03-06T13:50:53.269 INFO:teuthology.orchestra.run.vm02.stdout: Erasing : rbd-fuse-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 2/20 2026-03-06T13:50:53.272 INFO:teuthology.orchestra.run.vm02.stdout: Erasing : python3-rgw-2:19.2.3-47.gc24117fd552.el9.clyso.x86 3/20 2026-03-06T13:50:53.272 INFO:teuthology.orchestra.run.vm02.stdout: Erasing : librgw2-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 4/20 2026-03-06T13:50:53.282 INFO:teuthology.orchestra.run.vm06.stdout: Running scriptlet: librbd1-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 10/20 2026-03-06T13:50:53.283 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : librados2-2:19.2.3-47.gc24117fd552.el9.clyso.x86_6 11/20 2026-03-06T13:50:53.283 INFO:teuthology.orchestra.run.vm06.stdout:warning: file /etc/ceph: remove failed: No such file or directory 2026-03-06T13:50:53.283 INFO:teuthology.orchestra.run.vm06.stdout: 2026-03-06T13:50:53.285 INFO:teuthology.orchestra.run.vm02.stdout: Running scriptlet: librgw2-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 4/20 2026-03-06T13:50:53.286 INFO:teuthology.orchestra.run.vm02.stdout: Erasing : parquet-libs-9.0.0-15.el9.x86_64 5/20 2026-03-06T13:50:53.288 INFO:teuthology.orchestra.run.vm02.stdout: Erasing : python3-rbd-2:19.2.3-47.gc24117fd552.el9.clyso.x86 6/20 2026-03-06T13:50:53.290 INFO:teuthology.orchestra.run.vm02.stdout: Erasing : python3-rados-2:19.2.3-47.gc24117fd552.el9.clyso.x 7/20 2026-03-06T13:50:53.291 INFO:teuthology.orchestra.run.vm02.stdout: Erasing : qemu-kvm-block-rbd-17:10.1.0-15.el9.x86_64 8/20 2026-03-06T13:50:53.294 INFO:teuthology.orchestra.run.vm02.stdout: Erasing : libarrow-doc-9.0.0-15.el9.noarch 9/20 2026-03-06T13:50:53.294 INFO:teuthology.orchestra.run.vm02.stdout: Erasing : librbd1-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 10/20 2026-03-06T13:50:53.296 INFO:teuthology.orchestra.run.vm06.stdout: Running scriptlet: librados2-2:19.2.3-47.gc24117fd552.el9.clyso.x86_6 11/20 2026-03-06T13:50:53.296 INFO:teuthology.orchestra.run.vm04.stdout:Dependencies resolved. 2026-03-06T13:50:53.297 INFO:teuthology.orchestra.run.vm04.stdout:================================================================================ 2026-03-06T13:50:53.297 INFO:teuthology.orchestra.run.vm04.stdout: Package Arch Version Repository Size 2026-03-06T13:50:53.297 INFO:teuthology.orchestra.run.vm04.stdout:================================================================================ 2026-03-06T13:50:53.297 INFO:teuthology.orchestra.run.vm04.stdout:Removing: 2026-03-06T13:50:53.297 INFO:teuthology.orchestra.run.vm04.stdout: librados2 x86_64 2:19.2.3-47.gc24117fd552.el9.clyso @ceph 13 M 2026-03-06T13:50:53.297 INFO:teuthology.orchestra.run.vm04.stdout:Removing dependent packages: 2026-03-06T13:50:53.297 INFO:teuthology.orchestra.run.vm04.stdout: python3-rados x86_64 2:19.2.3-47.gc24117fd552.el9.clyso @ceph 1.1 M 2026-03-06T13:50:53.297 INFO:teuthology.orchestra.run.vm04.stdout: python3-rbd x86_64 2:19.2.3-47.gc24117fd552.el9.clyso @ceph 1.1 M 2026-03-06T13:50:53.297 INFO:teuthology.orchestra.run.vm04.stdout: python3-rgw x86_64 2:19.2.3-47.gc24117fd552.el9.clyso @ceph 265 k 2026-03-06T13:50:53.298 INFO:teuthology.orchestra.run.vm04.stdout: qemu-kvm-block-rbd x86_64 17:10.1.0-15.el9 @appstream 37 k 2026-03-06T13:50:53.298 INFO:teuthology.orchestra.run.vm04.stdout: rbd-fuse x86_64 2:19.2.3-47.gc24117fd552.el9.clyso @ceph 227 k 2026-03-06T13:50:53.298 INFO:teuthology.orchestra.run.vm04.stdout: rbd-nbd x86_64 2:19.2.3-47.gc24117fd552.el9.clyso @ceph 494 k 2026-03-06T13:50:53.298 INFO:teuthology.orchestra.run.vm04.stdout:Removing unused dependencies: 2026-03-06T13:50:53.298 INFO:teuthology.orchestra.run.vm04.stdout: boost-program-options 2026-03-06T13:50:53.298 INFO:teuthology.orchestra.run.vm04.stdout: x86_64 1.75.0-13.el9 @appstream 276 k 2026-03-06T13:50:53.298 INFO:teuthology.orchestra.run.vm04.stdout: libarrow x86_64 9.0.0-15.el9 @epel 18 M 2026-03-06T13:50:53.298 INFO:teuthology.orchestra.run.vm04.stdout: libarrow-doc noarch 9.0.0-15.el9 @epel 122 k 2026-03-06T13:50:53.298 INFO:teuthology.orchestra.run.vm04.stdout: libnbd x86_64 1.20.3-4.el9 @appstream 453 k 2026-03-06T13:50:53.298 INFO:teuthology.orchestra.run.vm04.stdout: libpmemobj x86_64 1.12.1-1.el9 @appstream 383 k 2026-03-06T13:50:53.298 INFO:teuthology.orchestra.run.vm04.stdout: librabbitmq x86_64 0.11.0-7.el9 @appstream 102 k 2026-03-06T13:50:53.298 INFO:teuthology.orchestra.run.vm04.stdout: librbd1 x86_64 2:19.2.3-47.gc24117fd552.el9.clyso @ceph 13 M 2026-03-06T13:50:53.298 INFO:teuthology.orchestra.run.vm04.stdout: librdkafka x86_64 1.6.1-102.el9 @appstream 2.0 M 2026-03-06T13:50:53.298 INFO:teuthology.orchestra.run.vm04.stdout: librgw2 x86_64 2:19.2.3-47.gc24117fd552.el9.clyso @ceph 19 M 2026-03-06T13:50:53.298 INFO:teuthology.orchestra.run.vm04.stdout: lttng-ust x86_64 2.12.0-6.el9 @appstream 1.0 M 2026-03-06T13:50:53.298 INFO:teuthology.orchestra.run.vm04.stdout: parquet-libs x86_64 9.0.0-15.el9 @epel 2.8 M 2026-03-06T13:50:53.298 INFO:teuthology.orchestra.run.vm04.stdout: re2 x86_64 1:20211101-20.el9 @epel 472 k 2026-03-06T13:50:53.298 INFO:teuthology.orchestra.run.vm04.stdout: thrift x86_64 0.15.0-4.el9 @epel 4.8 M 2026-03-06T13:50:53.298 INFO:teuthology.orchestra.run.vm04.stdout: 2026-03-06T13:50:53.298 INFO:teuthology.orchestra.run.vm04.stdout:Transaction Summary 2026-03-06T13:50:53.298 INFO:teuthology.orchestra.run.vm04.stdout:================================================================================ 2026-03-06T13:50:53.298 INFO:teuthology.orchestra.run.vm04.stdout:Remove 20 Packages 2026-03-06T13:50:53.298 INFO:teuthology.orchestra.run.vm04.stdout: 2026-03-06T13:50:53.298 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : libarrow-9.0.0-15.el9.x86_64 12/20 2026-03-06T13:50:53.298 INFO:teuthology.orchestra.run.vm04.stdout:Freed space: 79 M 2026-03-06T13:50:53.298 INFO:teuthology.orchestra.run.vm04.stdout:Running transaction check 2026-03-06T13:50:53.301 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : re2-1:20211101-20.el9.x86_64 13/20 2026-03-06T13:50:53.302 INFO:teuthology.orchestra.run.vm04.stdout:Transaction check succeeded. 2026-03-06T13:50:53.302 INFO:teuthology.orchestra.run.vm04.stdout:Running transaction test 2026-03-06T13:50:53.304 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : lttng-ust-2.12.0-6.el9.x86_64 14/20 2026-03-06T13:50:53.307 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : thrift-0.15.0-4.el9.x86_64 15/20 2026-03-06T13:50:53.307 INFO:teuthology.orchestra.run.vm02.stdout: Running scriptlet: librbd1-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 10/20 2026-03-06T13:50:53.308 INFO:teuthology.orchestra.run.vm02.stdout: Erasing : librados2-2:19.2.3-47.gc24117fd552.el9.clyso.x86_6 11/20 2026-03-06T13:50:53.308 INFO:teuthology.orchestra.run.vm02.stdout:warning: file /etc/ceph: remove failed: No such file or directory 2026-03-06T13:50:53.308 INFO:teuthology.orchestra.run.vm02.stdout: 2026-03-06T13:50:53.310 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : libnbd-1.20.3-4.el9.x86_64 16/20 2026-03-06T13:50:53.312 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : libpmemobj-1.12.1-1.el9.x86_64 17/20 2026-03-06T13:50:53.314 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : boost-program-options-1.75.0-13.el9.x86_64 18/20 2026-03-06T13:50:53.316 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : librabbitmq-0.11.0-7.el9.x86_64 19/20 2026-03-06T13:50:53.322 INFO:teuthology.orchestra.run.vm02.stdout: Running scriptlet: librados2-2:19.2.3-47.gc24117fd552.el9.clyso.x86_6 11/20 2026-03-06T13:50:53.324 INFO:teuthology.orchestra.run.vm02.stdout: Erasing : libarrow-9.0.0-15.el9.x86_64 12/20 2026-03-06T13:50:53.325 INFO:teuthology.orchestra.run.vm04.stdout:Transaction test succeeded. 2026-03-06T13:50:53.325 INFO:teuthology.orchestra.run.vm04.stdout:Running transaction 2026-03-06T13:50:53.327 INFO:teuthology.orchestra.run.vm02.stdout: Erasing : re2-1:20211101-20.el9.x86_64 13/20 2026-03-06T13:50:53.330 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : librdkafka-1.6.1-102.el9.x86_64 20/20 2026-03-06T13:50:53.331 INFO:teuthology.orchestra.run.vm02.stdout: Erasing : lttng-ust-2.12.0-6.el9.x86_64 14/20 2026-03-06T13:50:53.333 INFO:teuthology.orchestra.run.vm02.stdout: Erasing : thrift-0.15.0-4.el9.x86_64 15/20 2026-03-06T13:50:53.336 INFO:teuthology.orchestra.run.vm02.stdout: Erasing : libnbd-1.20.3-4.el9.x86_64 16/20 2026-03-06T13:50:53.338 INFO:teuthology.orchestra.run.vm02.stdout: Erasing : libpmemobj-1.12.1-1.el9.x86_64 17/20 2026-03-06T13:50:53.340 INFO:teuthology.orchestra.run.vm02.stdout: Erasing : boost-program-options-1.75.0-13.el9.x86_64 18/20 2026-03-06T13:50:53.342 INFO:teuthology.orchestra.run.vm02.stdout: Erasing : librabbitmq-0.11.0-7.el9.x86_64 19/20 2026-03-06T13:50:53.356 INFO:teuthology.orchestra.run.vm02.stdout: Erasing : librdkafka-1.6.1-102.el9.x86_64 20/20 2026-03-06T13:50:53.364 INFO:teuthology.orchestra.run.vm04.stdout: Preparing : 1/1 2026-03-06T13:50:53.367 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : rbd-nbd-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 1/20 2026-03-06T13:50:53.369 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : rbd-fuse-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 2/20 2026-03-06T13:50:53.372 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : python3-rgw-2:19.2.3-47.gc24117fd552.el9.clyso.x86 3/20 2026-03-06T13:50:53.372 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : librgw2-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 4/20 2026-03-06T13:50:53.385 INFO:teuthology.orchestra.run.vm04.stdout: Running scriptlet: librgw2-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 4/20 2026-03-06T13:50:53.387 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : parquet-libs-9.0.0-15.el9.x86_64 5/20 2026-03-06T13:50:53.388 INFO:teuthology.orchestra.run.vm06.stdout: Running scriptlet: librdkafka-1.6.1-102.el9.x86_64 20/20 2026-03-06T13:50:53.388 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : boost-program-options-1.75.0-13.el9.x86_64 1/20 2026-03-06T13:50:53.388 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : libarrow-9.0.0-15.el9.x86_64 2/20 2026-03-06T13:50:53.388 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : libarrow-doc-9.0.0-15.el9.noarch 3/20 2026-03-06T13:50:53.388 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : libnbd-1.20.3-4.el9.x86_64 4/20 2026-03-06T13:50:53.388 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : libpmemobj-1.12.1-1.el9.x86_64 5/20 2026-03-06T13:50:53.388 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : librabbitmq-0.11.0-7.el9.x86_64 6/20 2026-03-06T13:50:53.388 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : librados2-2:19.2.3-47.gc24117fd552.el9.clyso.x86_6 7/20 2026-03-06T13:50:53.389 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : librbd1-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 8/20 2026-03-06T13:50:53.389 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : librdkafka-1.6.1-102.el9.x86_64 9/20 2026-03-06T13:50:53.389 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : librgw2-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 10/20 2026-03-06T13:50:53.389 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : lttng-ust-2.12.0-6.el9.x86_64 11/20 2026-03-06T13:50:53.389 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : python3-rbd-2:19.2.3-47.gc24117fd552.el9.clyso.x86 6/20 2026-03-06T13:50:53.389 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : parquet-libs-9.0.0-15.el9.x86_64 12/20 2026-03-06T13:50:53.389 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-rados-2:19.2.3-47.gc24117fd552.el9.clyso.x 13/20 2026-03-06T13:50:53.389 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-rbd-2:19.2.3-47.gc24117fd552.el9.clyso.x86 14/20 2026-03-06T13:50:53.389 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-rgw-2:19.2.3-47.gc24117fd552.el9.clyso.x86 15/20 2026-03-06T13:50:53.389 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : qemu-kvm-block-rbd-17:10.1.0-15.el9.x86_64 16/20 2026-03-06T13:50:53.389 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : rbd-fuse-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 17/20 2026-03-06T13:50:53.389 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : rbd-nbd-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 18/20 2026-03-06T13:50:53.389 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : re2-1:20211101-20.el9.x86_64 19/20 2026-03-06T13:50:53.390 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : python3-rados-2:19.2.3-47.gc24117fd552.el9.clyso.x 7/20 2026-03-06T13:50:53.392 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : qemu-kvm-block-rbd-17:10.1.0-15.el9.x86_64 8/20 2026-03-06T13:50:53.394 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : libarrow-doc-9.0.0-15.el9.noarch 9/20 2026-03-06T13:50:53.395 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : librbd1-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 10/20 2026-03-06T13:50:53.409 INFO:teuthology.orchestra.run.vm04.stdout: Running scriptlet: librbd1-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 10/20 2026-03-06T13:50:53.409 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : librados2-2:19.2.3-47.gc24117fd552.el9.clyso.x86_6 11/20 2026-03-06T13:50:53.409 INFO:teuthology.orchestra.run.vm04.stdout:warning: file /etc/ceph: remove failed: No such file or directory 2026-03-06T13:50:53.409 INFO:teuthology.orchestra.run.vm04.stdout: 2026-03-06T13:50:53.418 INFO:teuthology.orchestra.run.vm02.stdout: Running scriptlet: librdkafka-1.6.1-102.el9.x86_64 20/20 2026-03-06T13:50:53.418 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : boost-program-options-1.75.0-13.el9.x86_64 1/20 2026-03-06T13:50:53.418 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : libarrow-9.0.0-15.el9.x86_64 2/20 2026-03-06T13:50:53.418 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : libarrow-doc-9.0.0-15.el9.noarch 3/20 2026-03-06T13:50:53.418 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : libnbd-1.20.3-4.el9.x86_64 4/20 2026-03-06T13:50:53.418 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : libpmemobj-1.12.1-1.el9.x86_64 5/20 2026-03-06T13:50:53.418 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : librabbitmq-0.11.0-7.el9.x86_64 6/20 2026-03-06T13:50:53.418 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : librados2-2:19.2.3-47.gc24117fd552.el9.clyso.x86_6 7/20 2026-03-06T13:50:53.418 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : librbd1-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 8/20 2026-03-06T13:50:53.418 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : librdkafka-1.6.1-102.el9.x86_64 9/20 2026-03-06T13:50:53.418 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : librgw2-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 10/20 2026-03-06T13:50:53.418 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : lttng-ust-2.12.0-6.el9.x86_64 11/20 2026-03-06T13:50:53.418 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : parquet-libs-9.0.0-15.el9.x86_64 12/20 2026-03-06T13:50:53.418 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : python3-rados-2:19.2.3-47.gc24117fd552.el9.clyso.x 13/20 2026-03-06T13:50:53.418 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : python3-rbd-2:19.2.3-47.gc24117fd552.el9.clyso.x86 14/20 2026-03-06T13:50:53.418 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : python3-rgw-2:19.2.3-47.gc24117fd552.el9.clyso.x86 15/20 2026-03-06T13:50:53.418 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : qemu-kvm-block-rbd-17:10.1.0-15.el9.x86_64 16/20 2026-03-06T13:50:53.419 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : rbd-fuse-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 17/20 2026-03-06T13:50:53.419 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : rbd-nbd-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 18/20 2026-03-06T13:50:53.419 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : re2-1:20211101-20.el9.x86_64 19/20 2026-03-06T13:50:53.424 INFO:teuthology.orchestra.run.vm04.stdout: Running scriptlet: librados2-2:19.2.3-47.gc24117fd552.el9.clyso.x86_6 11/20 2026-03-06T13:50:53.426 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : libarrow-9.0.0-15.el9.x86_64 12/20 2026-03-06T13:50:53.430 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : re2-1:20211101-20.el9.x86_64 13/20 2026-03-06T13:50:53.433 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : lttng-ust-2.12.0-6.el9.x86_64 14/20 2026-03-06T13:50:53.436 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : thrift-0.15.0-4.el9.x86_64 20/20 2026-03-06T13:50:53.436 INFO:teuthology.orchestra.run.vm06.stdout: 2026-03-06T13:50:53.436 INFO:teuthology.orchestra.run.vm06.stdout:Removed: 2026-03-06T13:50:53.436 INFO:teuthology.orchestra.run.vm06.stdout: boost-program-options-1.75.0-13.el9.x86_64 2026-03-06T13:50:53.436 INFO:teuthology.orchestra.run.vm06.stdout: libarrow-9.0.0-15.el9.x86_64 2026-03-06T13:50:53.436 INFO:teuthology.orchestra.run.vm06.stdout: libarrow-doc-9.0.0-15.el9.noarch 2026-03-06T13:50:53.436 INFO:teuthology.orchestra.run.vm06.stdout: libnbd-1.20.3-4.el9.x86_64 2026-03-06T13:50:53.436 INFO:teuthology.orchestra.run.vm06.stdout: libpmemobj-1.12.1-1.el9.x86_64 2026-03-06T13:50:53.436 INFO:teuthology.orchestra.run.vm06.stdout: librabbitmq-0.11.0-7.el9.x86_64 2026-03-06T13:50:53.436 INFO:teuthology.orchestra.run.vm06.stdout: librados2-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 2026-03-06T13:50:53.436 INFO:teuthology.orchestra.run.vm06.stdout: librbd1-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 2026-03-06T13:50:53.436 INFO:teuthology.orchestra.run.vm06.stdout: librdkafka-1.6.1-102.el9.x86_64 2026-03-06T13:50:53.436 INFO:teuthology.orchestra.run.vm06.stdout: librgw2-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 2026-03-06T13:50:53.436 INFO:teuthology.orchestra.run.vm06.stdout: lttng-ust-2.12.0-6.el9.x86_64 2026-03-06T13:50:53.436 INFO:teuthology.orchestra.run.vm06.stdout: parquet-libs-9.0.0-15.el9.x86_64 2026-03-06T13:50:53.436 INFO:teuthology.orchestra.run.vm06.stdout: python3-rados-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 2026-03-06T13:50:53.436 INFO:teuthology.orchestra.run.vm06.stdout: python3-rbd-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 2026-03-06T13:50:53.436 INFO:teuthology.orchestra.run.vm06.stdout: python3-rgw-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 2026-03-06T13:50:53.436 INFO:teuthology.orchestra.run.vm06.stdout: qemu-kvm-block-rbd-17:10.1.0-15.el9.x86_64 2026-03-06T13:50:53.436 INFO:teuthology.orchestra.run.vm06.stdout: rbd-fuse-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 2026-03-06T13:50:53.436 INFO:teuthology.orchestra.run.vm06.stdout: rbd-nbd-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 2026-03-06T13:50:53.436 INFO:teuthology.orchestra.run.vm06.stdout: re2-1:20211101-20.el9.x86_64 2026-03-06T13:50:53.436 INFO:teuthology.orchestra.run.vm06.stdout: thrift-0.15.0-4.el9.x86_64 2026-03-06T13:50:53.436 INFO:teuthology.orchestra.run.vm06.stdout: 2026-03-06T13:50:53.436 INFO:teuthology.orchestra.run.vm06.stdout:Complete! 2026-03-06T13:50:53.436 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : thrift-0.15.0-4.el9.x86_64 15/20 2026-03-06T13:50:53.439 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : libnbd-1.20.3-4.el9.x86_64 16/20 2026-03-06T13:50:53.441 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : libpmemobj-1.12.1-1.el9.x86_64 17/20 2026-03-06T13:50:53.443 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : boost-program-options-1.75.0-13.el9.x86_64 18/20 2026-03-06T13:50:53.445 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : librabbitmq-0.11.0-7.el9.x86_64 19/20 2026-03-06T13:50:53.458 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : librdkafka-1.6.1-102.el9.x86_64 20/20 2026-03-06T13:50:53.465 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : thrift-0.15.0-4.el9.x86_64 20/20 2026-03-06T13:50:53.465 INFO:teuthology.orchestra.run.vm02.stdout: 2026-03-06T13:50:53.465 INFO:teuthology.orchestra.run.vm02.stdout:Removed: 2026-03-06T13:50:53.465 INFO:teuthology.orchestra.run.vm02.stdout: boost-program-options-1.75.0-13.el9.x86_64 2026-03-06T13:50:53.465 INFO:teuthology.orchestra.run.vm02.stdout: libarrow-9.0.0-15.el9.x86_64 2026-03-06T13:50:53.465 INFO:teuthology.orchestra.run.vm02.stdout: libarrow-doc-9.0.0-15.el9.noarch 2026-03-06T13:50:53.465 INFO:teuthology.orchestra.run.vm02.stdout: libnbd-1.20.3-4.el9.x86_64 2026-03-06T13:50:53.465 INFO:teuthology.orchestra.run.vm02.stdout: libpmemobj-1.12.1-1.el9.x86_64 2026-03-06T13:50:53.465 INFO:teuthology.orchestra.run.vm02.stdout: librabbitmq-0.11.0-7.el9.x86_64 2026-03-06T13:50:53.465 INFO:teuthology.orchestra.run.vm02.stdout: librados2-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 2026-03-06T13:50:53.465 INFO:teuthology.orchestra.run.vm02.stdout: librbd1-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 2026-03-06T13:50:53.465 INFO:teuthology.orchestra.run.vm02.stdout: librdkafka-1.6.1-102.el9.x86_64 2026-03-06T13:50:53.465 INFO:teuthology.orchestra.run.vm02.stdout: librgw2-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 2026-03-06T13:50:53.465 INFO:teuthology.orchestra.run.vm02.stdout: lttng-ust-2.12.0-6.el9.x86_64 2026-03-06T13:50:53.465 INFO:teuthology.orchestra.run.vm02.stdout: parquet-libs-9.0.0-15.el9.x86_64 2026-03-06T13:50:53.465 INFO:teuthology.orchestra.run.vm02.stdout: python3-rados-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 2026-03-06T13:50:53.465 INFO:teuthology.orchestra.run.vm02.stdout: python3-rbd-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 2026-03-06T13:50:53.465 INFO:teuthology.orchestra.run.vm02.stdout: python3-rgw-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 2026-03-06T13:50:53.465 INFO:teuthology.orchestra.run.vm02.stdout: qemu-kvm-block-rbd-17:10.1.0-15.el9.x86_64 2026-03-06T13:50:53.465 INFO:teuthology.orchestra.run.vm02.stdout: rbd-fuse-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 2026-03-06T13:50:53.465 INFO:teuthology.orchestra.run.vm02.stdout: rbd-nbd-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 2026-03-06T13:50:53.465 INFO:teuthology.orchestra.run.vm02.stdout: re2-1:20211101-20.el9.x86_64 2026-03-06T13:50:53.465 INFO:teuthology.orchestra.run.vm02.stdout: thrift-0.15.0-4.el9.x86_64 2026-03-06T13:50:53.465 INFO:teuthology.orchestra.run.vm02.stdout: 2026-03-06T13:50:53.465 INFO:teuthology.orchestra.run.vm02.stdout:Complete! 2026-03-06T13:50:53.517 INFO:teuthology.orchestra.run.vm04.stdout: Running scriptlet: librdkafka-1.6.1-102.el9.x86_64 20/20 2026-03-06T13:50:53.517 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : boost-program-options-1.75.0-13.el9.x86_64 1/20 2026-03-06T13:50:53.517 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : libarrow-9.0.0-15.el9.x86_64 2/20 2026-03-06T13:50:53.517 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : libarrow-doc-9.0.0-15.el9.noarch 3/20 2026-03-06T13:50:53.517 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : libnbd-1.20.3-4.el9.x86_64 4/20 2026-03-06T13:50:53.517 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : libpmemobj-1.12.1-1.el9.x86_64 5/20 2026-03-06T13:50:53.517 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : librabbitmq-0.11.0-7.el9.x86_64 6/20 2026-03-06T13:50:53.517 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : librados2-2:19.2.3-47.gc24117fd552.el9.clyso.x86_6 7/20 2026-03-06T13:50:53.517 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : librbd1-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 8/20 2026-03-06T13:50:53.517 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : librdkafka-1.6.1-102.el9.x86_64 9/20 2026-03-06T13:50:53.517 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : librgw2-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 10/20 2026-03-06T13:50:53.517 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : lttng-ust-2.12.0-6.el9.x86_64 11/20 2026-03-06T13:50:53.517 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : parquet-libs-9.0.0-15.el9.x86_64 12/20 2026-03-06T13:50:53.517 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-rados-2:19.2.3-47.gc24117fd552.el9.clyso.x 13/20 2026-03-06T13:50:53.517 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-rbd-2:19.2.3-47.gc24117fd552.el9.clyso.x86 14/20 2026-03-06T13:50:53.517 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-rgw-2:19.2.3-47.gc24117fd552.el9.clyso.x86 15/20 2026-03-06T13:50:53.517 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : qemu-kvm-block-rbd-17:10.1.0-15.el9.x86_64 16/20 2026-03-06T13:50:53.517 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : rbd-fuse-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 17/20 2026-03-06T13:50:53.517 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : rbd-nbd-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 18/20 2026-03-06T13:50:53.517 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : re2-1:20211101-20.el9.x86_64 19/20 2026-03-06T13:50:53.562 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : thrift-0.15.0-4.el9.x86_64 20/20 2026-03-06T13:50:53.562 INFO:teuthology.orchestra.run.vm04.stdout: 2026-03-06T13:50:53.562 INFO:teuthology.orchestra.run.vm04.stdout:Removed: 2026-03-06T13:50:53.562 INFO:teuthology.orchestra.run.vm04.stdout: boost-program-options-1.75.0-13.el9.x86_64 2026-03-06T13:50:53.562 INFO:teuthology.orchestra.run.vm04.stdout: libarrow-9.0.0-15.el9.x86_64 2026-03-06T13:50:53.562 INFO:teuthology.orchestra.run.vm04.stdout: libarrow-doc-9.0.0-15.el9.noarch 2026-03-06T13:50:53.562 INFO:teuthology.orchestra.run.vm04.stdout: libnbd-1.20.3-4.el9.x86_64 2026-03-06T13:50:53.562 INFO:teuthology.orchestra.run.vm04.stdout: libpmemobj-1.12.1-1.el9.x86_64 2026-03-06T13:50:53.562 INFO:teuthology.orchestra.run.vm04.stdout: librabbitmq-0.11.0-7.el9.x86_64 2026-03-06T13:50:53.562 INFO:teuthology.orchestra.run.vm04.stdout: librados2-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 2026-03-06T13:50:53.562 INFO:teuthology.orchestra.run.vm04.stdout: librbd1-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 2026-03-06T13:50:53.562 INFO:teuthology.orchestra.run.vm04.stdout: librdkafka-1.6.1-102.el9.x86_64 2026-03-06T13:50:53.562 INFO:teuthology.orchestra.run.vm04.stdout: librgw2-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 2026-03-06T13:50:53.562 INFO:teuthology.orchestra.run.vm04.stdout: lttng-ust-2.12.0-6.el9.x86_64 2026-03-06T13:50:53.562 INFO:teuthology.orchestra.run.vm04.stdout: parquet-libs-9.0.0-15.el9.x86_64 2026-03-06T13:50:53.562 INFO:teuthology.orchestra.run.vm04.stdout: python3-rados-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 2026-03-06T13:50:53.562 INFO:teuthology.orchestra.run.vm04.stdout: python3-rbd-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 2026-03-06T13:50:53.562 INFO:teuthology.orchestra.run.vm04.stdout: python3-rgw-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 2026-03-06T13:50:53.562 INFO:teuthology.orchestra.run.vm04.stdout: qemu-kvm-block-rbd-17:10.1.0-15.el9.x86_64 2026-03-06T13:50:53.562 INFO:teuthology.orchestra.run.vm04.stdout: rbd-fuse-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 2026-03-06T13:50:53.562 INFO:teuthology.orchestra.run.vm04.stdout: rbd-nbd-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 2026-03-06T13:50:53.562 INFO:teuthology.orchestra.run.vm04.stdout: re2-1:20211101-20.el9.x86_64 2026-03-06T13:50:53.562 INFO:teuthology.orchestra.run.vm04.stdout: thrift-0.15.0-4.el9.x86_64 2026-03-06T13:50:53.562 INFO:teuthology.orchestra.run.vm04.stdout: 2026-03-06T13:50:53.563 INFO:teuthology.orchestra.run.vm04.stdout:Complete! 2026-03-06T13:50:53.647 INFO:teuthology.orchestra.run.vm06.stdout:No match for argument: librbd1 2026-03-06T13:50:53.648 INFO:teuthology.orchestra.run.vm06.stderr:No packages marked for removal. 2026-03-06T13:50:53.651 INFO:teuthology.orchestra.run.vm06.stdout:Dependencies resolved. 2026-03-06T13:50:53.652 INFO:teuthology.orchestra.run.vm06.stdout:Nothing to do. 2026-03-06T13:50:53.652 INFO:teuthology.orchestra.run.vm06.stdout:Complete! 2026-03-06T13:50:53.670 INFO:teuthology.orchestra.run.vm02.stdout:No match for argument: librbd1 2026-03-06T13:50:53.670 INFO:teuthology.orchestra.run.vm02.stderr:No packages marked for removal. 2026-03-06T13:50:53.672 INFO:teuthology.orchestra.run.vm02.stdout:Dependencies resolved. 2026-03-06T13:50:53.673 INFO:teuthology.orchestra.run.vm02.stdout:Nothing to do. 2026-03-06T13:50:53.673 INFO:teuthology.orchestra.run.vm02.stdout:Complete! 2026-03-06T13:50:53.785 INFO:teuthology.orchestra.run.vm04.stdout:No match for argument: librbd1 2026-03-06T13:50:53.785 INFO:teuthology.orchestra.run.vm04.stderr:No packages marked for removal. 2026-03-06T13:50:53.788 INFO:teuthology.orchestra.run.vm04.stdout:Dependencies resolved. 2026-03-06T13:50:53.788 INFO:teuthology.orchestra.run.vm04.stdout:Nothing to do. 2026-03-06T13:50:53.789 INFO:teuthology.orchestra.run.vm04.stdout:Complete! 2026-03-06T13:50:53.843 INFO:teuthology.orchestra.run.vm06.stdout:No match for argument: python3-rados 2026-03-06T13:50:53.844 INFO:teuthology.orchestra.run.vm06.stderr:No packages marked for removal. 2026-03-06T13:50:53.846 INFO:teuthology.orchestra.run.vm06.stdout:Dependencies resolved. 2026-03-06T13:50:53.846 INFO:teuthology.orchestra.run.vm06.stdout:Nothing to do. 2026-03-06T13:50:53.847 INFO:teuthology.orchestra.run.vm06.stdout:Complete! 2026-03-06T13:50:53.858 INFO:teuthology.orchestra.run.vm02.stdout:No match for argument: python3-rados 2026-03-06T13:50:53.858 INFO:teuthology.orchestra.run.vm02.stderr:No packages marked for removal. 2026-03-06T13:50:53.861 INFO:teuthology.orchestra.run.vm02.stdout:Dependencies resolved. 2026-03-06T13:50:53.861 INFO:teuthology.orchestra.run.vm02.stdout:Nothing to do. 2026-03-06T13:50:53.861 INFO:teuthology.orchestra.run.vm02.stdout:Complete! 2026-03-06T13:50:53.972 INFO:teuthology.orchestra.run.vm04.stdout:No match for argument: python3-rados 2026-03-06T13:50:53.972 INFO:teuthology.orchestra.run.vm04.stderr:No packages marked for removal. 2026-03-06T13:50:53.974 INFO:teuthology.orchestra.run.vm04.stdout:Dependencies resolved. 2026-03-06T13:50:53.975 INFO:teuthology.orchestra.run.vm04.stdout:Nothing to do. 2026-03-06T13:50:53.975 INFO:teuthology.orchestra.run.vm04.stdout:Complete! 2026-03-06T13:50:54.012 INFO:teuthology.orchestra.run.vm06.stdout:No match for argument: python3-rgw 2026-03-06T13:50:54.012 INFO:teuthology.orchestra.run.vm06.stderr:No packages marked for removal. 2026-03-06T13:50:54.014 INFO:teuthology.orchestra.run.vm06.stdout:Dependencies resolved. 2026-03-06T13:50:54.015 INFO:teuthology.orchestra.run.vm06.stdout:Nothing to do. 2026-03-06T13:50:54.015 INFO:teuthology.orchestra.run.vm06.stdout:Complete! 2026-03-06T13:50:54.027 INFO:teuthology.orchestra.run.vm02.stdout:No match for argument: python3-rgw 2026-03-06T13:50:54.027 INFO:teuthology.orchestra.run.vm02.stderr:No packages marked for removal. 2026-03-06T13:50:54.030 INFO:teuthology.orchestra.run.vm02.stdout:Dependencies resolved. 2026-03-06T13:50:54.030 INFO:teuthology.orchestra.run.vm02.stdout:Nothing to do. 2026-03-06T13:50:54.030 INFO:teuthology.orchestra.run.vm02.stdout:Complete! 2026-03-06T13:50:54.138 INFO:teuthology.orchestra.run.vm04.stdout:No match for argument: python3-rgw 2026-03-06T13:50:54.138 INFO:teuthology.orchestra.run.vm04.stderr:No packages marked for removal. 2026-03-06T13:50:54.141 INFO:teuthology.orchestra.run.vm04.stdout:Dependencies resolved. 2026-03-06T13:50:54.141 INFO:teuthology.orchestra.run.vm04.stdout:Nothing to do. 2026-03-06T13:50:54.141 INFO:teuthology.orchestra.run.vm04.stdout:Complete! 2026-03-06T13:50:54.176 INFO:teuthology.orchestra.run.vm06.stdout:No match for argument: python3-cephfs 2026-03-06T13:50:54.177 INFO:teuthology.orchestra.run.vm06.stderr:No packages marked for removal. 2026-03-06T13:50:54.179 INFO:teuthology.orchestra.run.vm06.stdout:Dependencies resolved. 2026-03-06T13:50:54.179 INFO:teuthology.orchestra.run.vm06.stdout:Nothing to do. 2026-03-06T13:50:54.179 INFO:teuthology.orchestra.run.vm06.stdout:Complete! 2026-03-06T13:50:54.191 INFO:teuthology.orchestra.run.vm02.stdout:No match for argument: python3-cephfs 2026-03-06T13:50:54.192 INFO:teuthology.orchestra.run.vm02.stderr:No packages marked for removal. 2026-03-06T13:50:54.194 INFO:teuthology.orchestra.run.vm02.stdout:Dependencies resolved. 2026-03-06T13:50:54.194 INFO:teuthology.orchestra.run.vm02.stdout:Nothing to do. 2026-03-06T13:50:54.194 INFO:teuthology.orchestra.run.vm02.stdout:Complete! 2026-03-06T13:50:54.309 INFO:teuthology.orchestra.run.vm04.stdout:No match for argument: python3-cephfs 2026-03-06T13:50:54.309 INFO:teuthology.orchestra.run.vm04.stderr:No packages marked for removal. 2026-03-06T13:50:54.311 INFO:teuthology.orchestra.run.vm04.stdout:Dependencies resolved. 2026-03-06T13:50:54.312 INFO:teuthology.orchestra.run.vm04.stdout:Nothing to do. 2026-03-06T13:50:54.312 INFO:teuthology.orchestra.run.vm04.stdout:Complete! 2026-03-06T13:50:54.344 INFO:teuthology.orchestra.run.vm06.stdout:No match for argument: python3-rbd 2026-03-06T13:50:54.344 INFO:teuthology.orchestra.run.vm06.stderr:No packages marked for removal. 2026-03-06T13:50:54.346 INFO:teuthology.orchestra.run.vm06.stdout:Dependencies resolved. 2026-03-06T13:50:54.347 INFO:teuthology.orchestra.run.vm06.stdout:Nothing to do. 2026-03-06T13:50:54.347 INFO:teuthology.orchestra.run.vm06.stdout:Complete! 2026-03-06T13:50:54.357 INFO:teuthology.orchestra.run.vm02.stdout:No match for argument: python3-rbd 2026-03-06T13:50:54.357 INFO:teuthology.orchestra.run.vm02.stderr:No packages marked for removal. 2026-03-06T13:50:54.359 INFO:teuthology.orchestra.run.vm02.stdout:Dependencies resolved. 2026-03-06T13:50:54.360 INFO:teuthology.orchestra.run.vm02.stdout:Nothing to do. 2026-03-06T13:50:54.360 INFO:teuthology.orchestra.run.vm02.stdout:Complete! 2026-03-06T13:50:54.487 INFO:teuthology.orchestra.run.vm04.stdout:No match for argument: python3-rbd 2026-03-06T13:50:54.488 INFO:teuthology.orchestra.run.vm04.stderr:No packages marked for removal. 2026-03-06T13:50:54.490 INFO:teuthology.orchestra.run.vm04.stdout:Dependencies resolved. 2026-03-06T13:50:54.491 INFO:teuthology.orchestra.run.vm04.stdout:Nothing to do. 2026-03-06T13:50:54.491 INFO:teuthology.orchestra.run.vm04.stdout:Complete! 2026-03-06T13:50:54.518 INFO:teuthology.orchestra.run.vm06.stdout:No match for argument: rbd-fuse 2026-03-06T13:50:54.518 INFO:teuthology.orchestra.run.vm06.stderr:No packages marked for removal. 2026-03-06T13:50:54.520 INFO:teuthology.orchestra.run.vm06.stdout:Dependencies resolved. 2026-03-06T13:50:54.520 INFO:teuthology.orchestra.run.vm06.stdout:Nothing to do. 2026-03-06T13:50:54.520 INFO:teuthology.orchestra.run.vm06.stdout:Complete! 2026-03-06T13:50:54.527 INFO:teuthology.orchestra.run.vm02.stdout:No match for argument: rbd-fuse 2026-03-06T13:50:54.527 INFO:teuthology.orchestra.run.vm02.stderr:No packages marked for removal. 2026-03-06T13:50:54.529 INFO:teuthology.orchestra.run.vm02.stdout:Dependencies resolved. 2026-03-06T13:50:54.530 INFO:teuthology.orchestra.run.vm02.stdout:Nothing to do. 2026-03-06T13:50:54.530 INFO:teuthology.orchestra.run.vm02.stdout:Complete! 2026-03-06T13:50:54.654 INFO:teuthology.orchestra.run.vm04.stdout:No match for argument: rbd-fuse 2026-03-06T13:50:54.654 INFO:teuthology.orchestra.run.vm04.stderr:No packages marked for removal. 2026-03-06T13:50:54.656 INFO:teuthology.orchestra.run.vm04.stdout:Dependencies resolved. 2026-03-06T13:50:54.657 INFO:teuthology.orchestra.run.vm04.stdout:Nothing to do. 2026-03-06T13:50:54.657 INFO:teuthology.orchestra.run.vm04.stdout:Complete! 2026-03-06T13:50:54.682 INFO:teuthology.orchestra.run.vm06.stdout:No match for argument: rbd-mirror 2026-03-06T13:50:54.682 INFO:teuthology.orchestra.run.vm06.stderr:No packages marked for removal. 2026-03-06T13:50:54.684 INFO:teuthology.orchestra.run.vm06.stdout:Dependencies resolved. 2026-03-06T13:50:54.685 INFO:teuthology.orchestra.run.vm06.stdout:Nothing to do. 2026-03-06T13:50:54.685 INFO:teuthology.orchestra.run.vm06.stdout:Complete! 2026-03-06T13:50:54.691 INFO:teuthology.orchestra.run.vm02.stdout:No match for argument: rbd-mirror 2026-03-06T13:50:54.691 INFO:teuthology.orchestra.run.vm02.stderr:No packages marked for removal. 2026-03-06T13:50:54.694 INFO:teuthology.orchestra.run.vm02.stdout:Dependencies resolved. 2026-03-06T13:50:54.694 INFO:teuthology.orchestra.run.vm02.stdout:Nothing to do. 2026-03-06T13:50:54.694 INFO:teuthology.orchestra.run.vm02.stdout:Complete! 2026-03-06T13:50:54.821 INFO:teuthology.orchestra.run.vm04.stdout:No match for argument: rbd-mirror 2026-03-06T13:50:54.821 INFO:teuthology.orchestra.run.vm04.stderr:No packages marked for removal. 2026-03-06T13:50:54.823 INFO:teuthology.orchestra.run.vm04.stdout:Dependencies resolved. 2026-03-06T13:50:54.824 INFO:teuthology.orchestra.run.vm04.stdout:Nothing to do. 2026-03-06T13:50:54.824 INFO:teuthology.orchestra.run.vm04.stdout:Complete! 2026-03-06T13:50:54.846 INFO:teuthology.orchestra.run.vm06.stdout:No match for argument: rbd-nbd 2026-03-06T13:50:54.846 INFO:teuthology.orchestra.run.vm06.stderr:No packages marked for removal. 2026-03-06T13:50:54.848 INFO:teuthology.orchestra.run.vm06.stdout:Dependencies resolved. 2026-03-06T13:50:54.849 INFO:teuthology.orchestra.run.vm06.stdout:Nothing to do. 2026-03-06T13:50:54.849 INFO:teuthology.orchestra.run.vm06.stdout:Complete! 2026-03-06T13:50:54.856 INFO:teuthology.orchestra.run.vm02.stdout:No match for argument: rbd-nbd 2026-03-06T13:50:54.857 INFO:teuthology.orchestra.run.vm02.stderr:No packages marked for removal. 2026-03-06T13:50:54.859 INFO:teuthology.orchestra.run.vm02.stdout:Dependencies resolved. 2026-03-06T13:50:54.860 INFO:teuthology.orchestra.run.vm02.stdout:Nothing to do. 2026-03-06T13:50:54.860 INFO:teuthology.orchestra.run.vm02.stdout:Complete! 2026-03-06T13:50:54.872 DEBUG:teuthology.orchestra.run.vm06:> sudo yum clean all 2026-03-06T13:50:54.883 DEBUG:teuthology.orchestra.run.vm02:> sudo yum clean all 2026-03-06T13:50:54.986 INFO:teuthology.orchestra.run.vm04.stdout:No match for argument: rbd-nbd 2026-03-06T13:50:54.986 INFO:teuthology.orchestra.run.vm04.stderr:No packages marked for removal. 2026-03-06T13:50:54.988 INFO:teuthology.orchestra.run.vm04.stdout:Dependencies resolved. 2026-03-06T13:50:54.989 INFO:teuthology.orchestra.run.vm04.stdout:Nothing to do. 2026-03-06T13:50:54.989 INFO:teuthology.orchestra.run.vm04.stdout:Complete! 2026-03-06T13:50:55.004 INFO:teuthology.orchestra.run.vm06.stdout:56 files removed 2026-03-06T13:50:55.010 DEBUG:teuthology.orchestra.run.vm04:> sudo yum clean all 2026-03-06T13:50:55.011 INFO:teuthology.orchestra.run.vm02.stdout:56 files removed 2026-03-06T13:50:55.027 DEBUG:teuthology.orchestra.run.vm06:> sudo rm /etc/yum.repos.d/ceph-source.repo 2026-03-06T13:50:55.035 DEBUG:teuthology.orchestra.run.vm02:> sudo rm /etc/yum.repos.d/ceph-source.repo 2026-03-06T13:50:55.050 DEBUG:teuthology.orchestra.run.vm06:> sudo rm /etc/yum.repos.d/ceph-noarch.repo 2026-03-06T13:50:55.058 DEBUG:teuthology.orchestra.run.vm02:> sudo rm /etc/yum.repos.d/ceph-noarch.repo 2026-03-06T13:50:55.114 DEBUG:teuthology.orchestra.run.vm06:> sudo rm /etc/yum.repos.d/ceph.repo 2026-03-06T13:50:55.124 DEBUG:teuthology.orchestra.run.vm02:> sudo rm /etc/yum.repos.d/ceph.repo 2026-03-06T13:50:55.135 INFO:teuthology.orchestra.run.vm04.stdout:56 files removed 2026-03-06T13:50:55.155 DEBUG:teuthology.orchestra.run.vm04:> sudo rm /etc/yum.repos.d/ceph-source.repo 2026-03-06T13:50:55.177 DEBUG:teuthology.orchestra.run.vm04:> sudo rm /etc/yum.repos.d/ceph-noarch.repo 2026-03-06T13:50:55.179 DEBUG:teuthology.orchestra.run.vm06:> sudo yum clean expire-cache 2026-03-06T13:50:55.189 DEBUG:teuthology.orchestra.run.vm02:> sudo yum clean expire-cache 2026-03-06T13:50:55.241 DEBUG:teuthology.orchestra.run.vm04:> sudo rm /etc/yum.repos.d/ceph.repo 2026-03-06T13:50:55.305 DEBUG:teuthology.orchestra.run.vm04:> sudo yum clean expire-cache 2026-03-06T13:50:55.326 INFO:teuthology.orchestra.run.vm06.stdout:Cache was expired 2026-03-06T13:50:55.326 INFO:teuthology.orchestra.run.vm06.stdout:0 files removed 2026-03-06T13:50:55.339 INFO:teuthology.orchestra.run.vm02.stdout:Cache was expired 2026-03-06T13:50:55.339 INFO:teuthology.orchestra.run.vm02.stdout:0 files removed 2026-03-06T13:50:55.344 DEBUG:teuthology.parallel:result is None 2026-03-06T13:50:55.359 DEBUG:teuthology.parallel:result is None 2026-03-06T13:50:55.449 INFO:teuthology.orchestra.run.vm04.stdout:Cache was expired 2026-03-06T13:50:55.449 INFO:teuthology.orchestra.run.vm04.stdout:0 files removed 2026-03-06T13:50:55.467 DEBUG:teuthology.parallel:result is None 2026-03-06T13:50:55.467 INFO:teuthology.task.install:Removing ceph sources lists on ubuntu@vm02.local 2026-03-06T13:50:55.467 INFO:teuthology.task.install:Removing ceph sources lists on ubuntu@vm04.local 2026-03-06T13:50:55.467 INFO:teuthology.task.install:Removing ceph sources lists on ubuntu@vm06.local 2026-03-06T13:50:55.467 DEBUG:teuthology.orchestra.run.vm02:> sudo rm -f /etc/yum.repos.d/ceph.repo 2026-03-06T13:50:55.467 DEBUG:teuthology.orchestra.run.vm04:> sudo rm -f /etc/yum.repos.d/ceph.repo 2026-03-06T13:50:55.467 DEBUG:teuthology.orchestra.run.vm06:> sudo rm -f /etc/yum.repos.d/ceph.repo 2026-03-06T13:50:55.491 DEBUG:teuthology.orchestra.run.vm04:> sudo mv -f /etc/yum/pluginconf.d/priorities.conf.orig /etc/yum/pluginconf.d/priorities.conf 2026-03-06T13:50:55.492 DEBUG:teuthology.orchestra.run.vm06:> sudo mv -f /etc/yum/pluginconf.d/priorities.conf.orig /etc/yum/pluginconf.d/priorities.conf 2026-03-06T13:50:55.495 DEBUG:teuthology.orchestra.run.vm02:> sudo mv -f /etc/yum/pluginconf.d/priorities.conf.orig /etc/yum/pluginconf.d/priorities.conf 2026-03-06T13:50:55.555 INFO:teuthology.orchestra.run.vm04.stderr:mv: cannot stat '/etc/yum/pluginconf.d/priorities.conf.orig': No such file or directory 2026-03-06T13:50:55.555 INFO:teuthology.orchestra.run.vm06.stderr:mv: cannot stat '/etc/yum/pluginconf.d/priorities.conf.orig': No such file or directory 2026-03-06T13:50:55.556 DEBUG:teuthology.orchestra.run:got remote process result: 1 2026-03-06T13:50:55.556 DEBUG:teuthology.parallel:result is None 2026-03-06T13:50:55.556 DEBUG:teuthology.orchestra.run:got remote process result: 1 2026-03-06T13:50:55.556 DEBUG:teuthology.parallel:result is None 2026-03-06T13:50:55.561 INFO:teuthology.orchestra.run.vm02.stderr:mv: cannot stat '/etc/yum/pluginconf.d/priorities.conf.orig': No such file or directory 2026-03-06T13:50:55.562 DEBUG:teuthology.orchestra.run:got remote process result: 1 2026-03-06T13:50:55.562 DEBUG:teuthology.parallel:result is None 2026-03-06T13:50:55.562 DEBUG:teuthology.run_tasks:Unwinding manager clock 2026-03-06T13:50:55.565 INFO:teuthology.task.clock:Checking final clock skew... 2026-03-06T13:50:55.565 DEBUG:teuthology.orchestra.run.vm02:> PATH=/usr/bin:/usr/sbin ntpq -p || PATH=/usr/bin:/usr/sbin chronyc sources || true 2026-03-06T13:50:55.604 DEBUG:teuthology.orchestra.run.vm04:> PATH=/usr/bin:/usr/sbin ntpq -p || PATH=/usr/bin:/usr/sbin chronyc sources || true 2026-03-06T13:50:55.606 DEBUG:teuthology.orchestra.run.vm06:> PATH=/usr/bin:/usr/sbin ntpq -p || PATH=/usr/bin:/usr/sbin chronyc sources || true 2026-03-06T13:50:55.617 INFO:teuthology.orchestra.run.vm02.stderr:bash: line 1: ntpq: command not found 2026-03-06T13:50:55.619 INFO:teuthology.orchestra.run.vm04.stderr:bash: line 1: ntpq: command not found 2026-03-06T13:50:55.620 INFO:teuthology.orchestra.run.vm06.stderr:bash: line 1: ntpq: command not found 2026-03-06T13:50:55.621 INFO:teuthology.orchestra.run.vm02.stdout:MS Name/IP address Stratum Poll Reach LastRx Last sample 2026-03-06T13:50:55.621 INFO:teuthology.orchestra.run.vm02.stdout:=============================================================================== 2026-03-06T13:50:55.621 INFO:teuthology.orchestra.run.vm02.stdout:^- ntp5.kernfusion.at 2 6 377 8 +66us[ +66us] +/- 17ms 2026-03-06T13:50:55.621 INFO:teuthology.orchestra.run.vm02.stdout:^- mail.fu110.de 2 6 377 9 -178us[ -178us] +/- 20ms 2026-03-06T13:50:55.621 INFO:teuthology.orchestra.run.vm02.stdout:^- 79.133.44.142 1 6 377 8 +138us[ +138us] +/- 10ms 2026-03-06T13:50:55.621 INFO:teuthology.orchestra.run.vm02.stdout:^* static.222.16.42.77.clie> 2 6 377 10 -12us[ -18us] +/- 2458us 2026-03-06T13:50:55.622 INFO:teuthology.orchestra.run.vm04.stdout:MS Name/IP address Stratum Poll Reach LastRx Last sample 2026-03-06T13:50:55.622 INFO:teuthology.orchestra.run.vm04.stdout:=============================================================================== 2026-03-06T13:50:55.622 INFO:teuthology.orchestra.run.vm04.stdout:^* static.222.16.42.77.clie> 2 6 377 8 -2301ns[-4060ns] +/- 2453us 2026-03-06T13:50:55.622 INFO:teuthology.orchestra.run.vm04.stdout:^- 79.133.44.142 1 6 377 6 +118us[ +118us] +/- 10ms 2026-03-06T13:50:55.622 INFO:teuthology.orchestra.run.vm04.stdout:^- ntp5.kernfusion.at 2 6 377 8 +117us[ +115us] +/- 17ms 2026-03-06T13:50:55.622 INFO:teuthology.orchestra.run.vm04.stdout:^- mail.fu110.de 2 6 377 9 -189us[ -191us] +/- 20ms 2026-03-06T13:50:55.623 INFO:teuthology.orchestra.run.vm06.stdout:MS Name/IP address Stratum Poll Reach LastRx Last sample 2026-03-06T13:50:55.623 INFO:teuthology.orchestra.run.vm06.stdout:=============================================================================== 2026-03-06T13:50:55.623 INFO:teuthology.orchestra.run.vm06.stdout:^* static.222.16.42.77.clie> 2 6 377 9 -1266ns[-2350ns] +/- 2429us 2026-03-06T13:50:55.623 INFO:teuthology.orchestra.run.vm06.stdout:^- ntp5.kernfusion.at 2 6 377 8 -49us[ -49us] +/- 17ms 2026-03-06T13:50:55.623 INFO:teuthology.orchestra.run.vm06.stdout:^- mail.fu110.de 2 6 377 7 -204us[ -204us] +/- 20ms 2026-03-06T13:50:55.623 INFO:teuthology.orchestra.run.vm06.stdout:^- 79.133.44.142 1 6 377 9 +145us[ +145us] +/- 10ms 2026-03-06T13:50:55.624 DEBUG:teuthology.run_tasks:Unwinding manager ansible.cephlab 2026-03-06T13:50:55.626 INFO:teuthology.task.ansible:Skipping ansible cleanup... 2026-03-06T13:50:55.626 DEBUG:teuthology.run_tasks:Unwinding manager selinux 2026-03-06T13:50:55.628 DEBUG:teuthology.run_tasks:Unwinding manager pcp 2026-03-06T13:50:55.630 DEBUG:teuthology.run_tasks:Unwinding manager internal.timer 2026-03-06T13:50:55.632 INFO:teuthology.task.internal:Duration was 626.495823 seconds 2026-03-06T13:50:55.632 DEBUG:teuthology.run_tasks:Unwinding manager internal.syslog 2026-03-06T13:50:55.634 INFO:teuthology.task.internal.syslog:Shutting down syslog monitoring... 2026-03-06T13:50:55.634 DEBUG:teuthology.orchestra.run.vm02:> sudo rm -f -- /etc/rsyslog.d/80-cephtest.conf && sudo service rsyslog restart 2026-03-06T13:50:55.664 DEBUG:teuthology.orchestra.run.vm04:> sudo rm -f -- /etc/rsyslog.d/80-cephtest.conf && sudo service rsyslog restart 2026-03-06T13:50:55.665 DEBUG:teuthology.orchestra.run.vm06:> sudo rm -f -- /etc/rsyslog.d/80-cephtest.conf && sudo service rsyslog restart 2026-03-06T13:50:55.701 INFO:teuthology.orchestra.run.vm02.stderr:Redirecting to /bin/systemctl restart rsyslog.service 2026-03-06T13:50:55.702 INFO:teuthology.orchestra.run.vm06.stderr:Redirecting to /bin/systemctl restart rsyslog.service 2026-03-06T13:50:55.703 INFO:teuthology.orchestra.run.vm04.stderr:Redirecting to /bin/systemctl restart rsyslog.service 2026-03-06T13:50:56.142 INFO:teuthology.task.internal.syslog:Checking logs for errors... 2026-03-06T13:50:56.142 DEBUG:teuthology.task.internal.syslog:Checking ubuntu@vm02.local 2026-03-06T13:50:56.142 DEBUG:teuthology.orchestra.run.vm02:> grep -E --binary-files=text '\bBUG\b|\bINFO\b|\bDEADLOCK\b' /home/ubuntu/cephtest/archive/syslog/kern.log | grep -v 'task .* blocked for more than .* seconds' | grep -v 'lockdep is turned off' | grep -v 'trying to register non-static key' | grep -v 'DEBUG: fsize' | grep -v CRON | grep -v 'BUG: bad unlock balance detected' | grep -v 'inconsistent lock state' | grep -v '*** DEADLOCK ***' | grep -v 'INFO: possible irq lock inversion dependency detected' | grep -v 'INFO: NMI handler (perf_event_nmi_handler) took too long to run' | grep -v 'INFO: recovery required on readonly' | grep -v 'ceph-create-keys: INFO' | grep -v INFO:ceph-create-keys | grep -v 'Loaded datasource DataSourceOpenStack' | grep -v 'container-storage-setup: INFO: Volume group backing root filesystem could not be determined' | grep -E -v '\bsalt-master\b|\bsalt-minion\b|\bsalt-api\b' | grep -v ceph-crash | grep -E -v '\btcmu-runner\b.*\bINFO\b' | head -n 1 2026-03-06T13:50:56.204 DEBUG:teuthology.task.internal.syslog:Checking ubuntu@vm04.local 2026-03-06T13:50:56.205 DEBUG:teuthology.orchestra.run.vm04:> grep -E --binary-files=text '\bBUG\b|\bINFO\b|\bDEADLOCK\b' /home/ubuntu/cephtest/archive/syslog/kern.log | grep -v 'task .* blocked for more than .* seconds' | grep -v 'lockdep is turned off' | grep -v 'trying to register non-static key' | grep -v 'DEBUG: fsize' | grep -v CRON | grep -v 'BUG: bad unlock balance detected' | grep -v 'inconsistent lock state' | grep -v '*** DEADLOCK ***' | grep -v 'INFO: possible irq lock inversion dependency detected' | grep -v 'INFO: NMI handler (perf_event_nmi_handler) took too long to run' | grep -v 'INFO: recovery required on readonly' | grep -v 'ceph-create-keys: INFO' | grep -v INFO:ceph-create-keys | grep -v 'Loaded datasource DataSourceOpenStack' | grep -v 'container-storage-setup: INFO: Volume group backing root filesystem could not be determined' | grep -E -v '\bsalt-master\b|\bsalt-minion\b|\bsalt-api\b' | grep -v ceph-crash | grep -E -v '\btcmu-runner\b.*\bINFO\b' | head -n 1 2026-03-06T13:50:56.230 DEBUG:teuthology.task.internal.syslog:Checking ubuntu@vm06.local 2026-03-06T13:50:56.230 DEBUG:teuthology.orchestra.run.vm06:> grep -E --binary-files=text '\bBUG\b|\bINFO\b|\bDEADLOCK\b' /home/ubuntu/cephtest/archive/syslog/kern.log | grep -v 'task .* blocked for more than .* seconds' | grep -v 'lockdep is turned off' | grep -v 'trying to register non-static key' | grep -v 'DEBUG: fsize' | grep -v CRON | grep -v 'BUG: bad unlock balance detected' | grep -v 'inconsistent lock state' | grep -v '*** DEADLOCK ***' | grep -v 'INFO: possible irq lock inversion dependency detected' | grep -v 'INFO: NMI handler (perf_event_nmi_handler) took too long to run' | grep -v 'INFO: recovery required on readonly' | grep -v 'ceph-create-keys: INFO' | grep -v INFO:ceph-create-keys | grep -v 'Loaded datasource DataSourceOpenStack' | grep -v 'container-storage-setup: INFO: Volume group backing root filesystem could not be determined' | grep -E -v '\bsalt-master\b|\bsalt-minion\b|\bsalt-api\b' | grep -v ceph-crash | grep -E -v '\btcmu-runner\b.*\bINFO\b' | head -n 1 2026-03-06T13:50:56.253 INFO:teuthology.task.internal.syslog:Gathering journactl... 2026-03-06T13:50:56.253 DEBUG:teuthology.orchestra.run.vm02:> sudo journalctl > /home/ubuntu/cephtest/archive/syslog/journalctl.log 2026-03-06T13:50:56.255 DEBUG:teuthology.orchestra.run.vm04:> sudo journalctl > /home/ubuntu/cephtest/archive/syslog/journalctl.log 2026-03-06T13:50:56.273 DEBUG:teuthology.orchestra.run.vm06:> sudo journalctl > /home/ubuntu/cephtest/archive/syslog/journalctl.log 2026-03-06T13:50:56.744 INFO:teuthology.task.internal.syslog:Compressing syslogs... 2026-03-06T13:50:56.744 DEBUG:teuthology.orchestra.run.vm02:> find /home/ubuntu/cephtest/archive/syslog -name '*.log' -print0 | sudo xargs -0 --max-args=1 --max-procs=0 --verbose --no-run-if-empty -- gzip -5 --verbose -- 2026-03-06T13:50:56.746 DEBUG:teuthology.orchestra.run.vm04:> find /home/ubuntu/cephtest/archive/syslog -name '*.log' -print0 | sudo xargs -0 --max-args=1 --max-procs=0 --verbose --no-run-if-empty -- gzip -5 --verbose -- 2026-03-06T13:50:56.747 DEBUG:teuthology.orchestra.run.vm06:> find /home/ubuntu/cephtest/archive/syslog -name '*.log' -print0 | sudo xargs -0 --max-args=1 --max-procs=0 --verbose --no-run-if-empty -- gzip -5 --verbose -- 2026-03-06T13:50:56.769 INFO:teuthology.orchestra.run.vm02.stderr:gzip -5 --verbose -- /home/ubuntu/cephtest/archive/syslog/kern.log 2026-03-06T13:50:56.769 INFO:teuthology.orchestra.run.vm02.stderr:gzip -5 --verbose -- /home/ubuntu/cephtest/archive/syslog/misc.log 2026-03-06T13:50:56.769 INFO:teuthology.orchestra.run.vm02.stderr:/home/ubuntu/cephtest/archive/syslog/kern.log: 0.0% -- replaced with /home/ubuntu/cephtest/archive/syslog/kern.log.gz 2026-03-06T13:50:56.769 INFO:teuthology.orchestra.run.vm04.stderr:gzip -5 --verbose -- /home/ubuntu/cephtest/archive/syslog/kern.log 2026-03-06T13:50:56.770 INFO:teuthology.orchestra.run.vm04.stderr:gzip -5 --verbose -- /home/ubuntu/cephtest/archive/syslog/misc.log 2026-03-06T13:50:56.770 INFO:teuthology.orchestra.run.vm04.stderr:gzip -5 --verbose -- /home/ubuntu/cephtest/archive/syslog/journalctl.log 2026-03-06T13:50:56.770 INFO:teuthology.orchestra.run.vm04.stderr:/home/ubuntu/cephtest/archive/syslog/kern.log: 0.0% -- replaced with /home/ubuntu/cephtest/archive/syslog/kern.log.gz 2026-03-06T13:50:56.770 INFO:teuthology.orchestra.run.vm02.stderr:gzip -5 --verbose -- /home/ubuntu/cephtest/archive/syslog/journalctl.log 2026-03-06T13:50:56.770 INFO:teuthology.orchestra.run.vm02.stderr:/home/ubuntu/cephtest/archive/syslog/misc.log: 0.0% -- replaced with /home/ubuntu/cephtest/archive/syslog/misc.log.gz 2026-03-06T13:50:56.770 INFO:teuthology.orchestra.run.vm04.stderr:/home/ubuntu/cephtest/archive/syslog/misc.log: 0.0% -- replaced with /home/ubuntu/cephtest/archive/syslog/misc.log.gz 2026-03-06T13:50:56.774 INFO:teuthology.orchestra.run.vm06.stderr:gzip -5 --verbose -- /home/ubuntu/cephtest/archive/syslog/kern.log 2026-03-06T13:50:56.774 INFO:teuthology.orchestra.run.vm06.stderr:gzip -5 --verbose -- /home/ubuntu/cephtest/archive/syslog/misc.log 2026-03-06T13:50:56.774 INFO:teuthology.orchestra.run.vm06.stderr:/home/ubuntu/cephtest/archive/syslog/kern.log: gzip -5 --verbose -- /home/ubuntu/cephtest/archive/syslog/journalctl.log 2026-03-06T13:50:56.774 INFO:teuthology.orchestra.run.vm06.stderr: 0.0% -- replaced with /home/ubuntu/cephtest/archive/syslog/kern.log.gz 2026-03-06T13:50:56.774 INFO:teuthology.orchestra.run.vm06.stderr:/home/ubuntu/cephtest/archive/syslog/misc.log: 0.0% -- replaced with /home/ubuntu/cephtest/archive/syslog/misc.log.gz 2026-03-06T13:50:56.871 INFO:teuthology.orchestra.run.vm02.stderr:/home/ubuntu/cephtest/archive/syslog/journalctl.log: 98.0% -- replaced with /home/ubuntu/cephtest/archive/syslog/journalctl.log.gz 2026-03-06T13:50:56.896 INFO:teuthology.orchestra.run.vm04.stderr:/home/ubuntu/cephtest/archive/syslog/journalctl.log: 98.3% -- replaced with /home/ubuntu/cephtest/archive/syslog/journalctl.log.gz 2026-03-06T13:50:56.930 INFO:teuthology.orchestra.run.vm06.stderr:/home/ubuntu/cephtest/archive/syslog/journalctl.log: 98.4% -- replaced with /home/ubuntu/cephtest/archive/syslog/journalctl.log.gz 2026-03-06T13:50:56.932 DEBUG:teuthology.run_tasks:Unwinding manager internal.sudo 2026-03-06T13:50:56.934 INFO:teuthology.task.internal:Restoring /etc/sudoers... 2026-03-06T13:50:56.934 DEBUG:teuthology.orchestra.run.vm02:> sudo mv -f /etc/sudoers.orig.teuthology /etc/sudoers 2026-03-06T13:50:56.960 DEBUG:teuthology.orchestra.run.vm04:> sudo mv -f /etc/sudoers.orig.teuthology /etc/sudoers 2026-03-06T13:50:56.982 DEBUG:teuthology.orchestra.run.vm06:> sudo mv -f /etc/sudoers.orig.teuthology /etc/sudoers 2026-03-06T13:50:57.005 DEBUG:teuthology.run_tasks:Unwinding manager internal.coredump 2026-03-06T13:50:57.008 DEBUG:teuthology.orchestra.run.vm02:> sudo sysctl -w kernel.core_pattern=core && sudo bash -c 'for f in `find /home/ubuntu/cephtest/archive/coredump -type f`; do file $f | grep -q systemd-sysusers && rm $f || true ; done' && rmdir --ignore-fail-on-non-empty -- /home/ubuntu/cephtest/archive/coredump 2026-03-06T13:50:57.009 DEBUG:teuthology.orchestra.run.vm04:> sudo sysctl -w kernel.core_pattern=core && sudo bash -c 'for f in `find /home/ubuntu/cephtest/archive/coredump -type f`; do file $f | grep -q systemd-sysusers && rm $f || true ; done' && rmdir --ignore-fail-on-non-empty -- /home/ubuntu/cephtest/archive/coredump 2026-03-06T13:50:57.024 DEBUG:teuthology.orchestra.run.vm06:> sudo sysctl -w kernel.core_pattern=core && sudo bash -c 'for f in `find /home/ubuntu/cephtest/archive/coredump -type f`; do file $f | grep -q systemd-sysusers && rm $f || true ; done' && rmdir --ignore-fail-on-non-empty -- /home/ubuntu/cephtest/archive/coredump 2026-03-06T13:50:57.033 INFO:teuthology.orchestra.run.vm02.stdout:kernel.core_pattern = core 2026-03-06T13:50:57.051 INFO:teuthology.orchestra.run.vm04.stdout:kernel.core_pattern = core 2026-03-06T13:50:57.071 INFO:teuthology.orchestra.run.vm06.stdout:kernel.core_pattern = core 2026-03-06T13:50:57.085 DEBUG:teuthology.orchestra.run.vm02:> test -e /home/ubuntu/cephtest/archive/coredump 2026-03-06T13:50:57.104 DEBUG:teuthology.orchestra.run:got remote process result: 1 2026-03-06T13:50:57.104 DEBUG:teuthology.orchestra.run.vm04:> test -e /home/ubuntu/cephtest/archive/coredump 2026-03-06T13:50:57.118 DEBUG:teuthology.orchestra.run:got remote process result: 1 2026-03-06T13:50:57.118 DEBUG:teuthology.orchestra.run.vm06:> test -e /home/ubuntu/cephtest/archive/coredump 2026-03-06T13:50:57.139 DEBUG:teuthology.orchestra.run:got remote process result: 1 2026-03-06T13:50:57.139 DEBUG:teuthology.run_tasks:Unwinding manager internal.archive 2026-03-06T13:50:57.142 INFO:teuthology.task.internal:Transferring archived files... 2026-03-06T13:50:57.143 DEBUG:teuthology.misc:Transferring archived files from vm02:/home/ubuntu/cephtest/archive to /archive/irq0-2026-03-06_13:20:18-orch:cephadm:workunits-cobaltcore-storage-v19.2.3-fasttrack-3-none-default-vps/274/remote/vm02 2026-03-06T13:50:57.143 DEBUG:teuthology.orchestra.run.vm02:> sudo tar c -f - -C /home/ubuntu/cephtest/archive -- . 2026-03-06T13:50:57.169 DEBUG:teuthology.misc:Transferring archived files from vm04:/home/ubuntu/cephtest/archive to /archive/irq0-2026-03-06_13:20:18-orch:cephadm:workunits-cobaltcore-storage-v19.2.3-fasttrack-3-none-default-vps/274/remote/vm04 2026-03-06T13:50:57.169 DEBUG:teuthology.orchestra.run.vm04:> sudo tar c -f - -C /home/ubuntu/cephtest/archive -- . 2026-03-06T13:50:57.197 DEBUG:teuthology.misc:Transferring archived files from vm06:/home/ubuntu/cephtest/archive to /archive/irq0-2026-03-06_13:20:18-orch:cephadm:workunits-cobaltcore-storage-v19.2.3-fasttrack-3-none-default-vps/274/remote/vm06 2026-03-06T13:50:57.198 DEBUG:teuthology.orchestra.run.vm06:> sudo tar c -f - -C /home/ubuntu/cephtest/archive -- . 2026-03-06T13:50:57.224 INFO:teuthology.task.internal:Removing archive directory... 2026-03-06T13:50:57.224 DEBUG:teuthology.orchestra.run.vm02:> rm -rf -- /home/ubuntu/cephtest/archive 2026-03-06T13:50:57.226 DEBUG:teuthology.orchestra.run.vm04:> rm -rf -- /home/ubuntu/cephtest/archive 2026-03-06T13:50:57.240 DEBUG:teuthology.orchestra.run.vm06:> rm -rf -- /home/ubuntu/cephtest/archive 2026-03-06T13:50:57.278 DEBUG:teuthology.run_tasks:Unwinding manager internal.archive_upload 2026-03-06T13:50:57.281 INFO:teuthology.task.internal:Not uploading archives. 2026-03-06T13:50:57.281 DEBUG:teuthology.run_tasks:Unwinding manager internal.base 2026-03-06T13:50:57.283 INFO:teuthology.task.internal:Tidying up after the test... 2026-03-06T13:50:57.283 DEBUG:teuthology.orchestra.run.vm02:> find /home/ubuntu/cephtest -ls ; rmdir -- /home/ubuntu/cephtest 2026-03-06T13:50:57.284 DEBUG:teuthology.orchestra.run.vm04:> find /home/ubuntu/cephtest -ls ; rmdir -- /home/ubuntu/cephtest 2026-03-06T13:50:57.298 DEBUG:teuthology.orchestra.run.vm06:> find /home/ubuntu/cephtest -ls ; rmdir -- /home/ubuntu/cephtest 2026-03-06T13:50:57.298 INFO:teuthology.orchestra.run.vm02.stdout: 8532138 0 drwxr-xr-x 2 ubuntu ubuntu 6 Mar 6 13:50 /home/ubuntu/cephtest 2026-03-06T13:50:57.313 INFO:teuthology.orchestra.run.vm04.stdout: 8532146 0 drwxr-xr-x 2 ubuntu ubuntu 6 Mar 6 13:50 /home/ubuntu/cephtest 2026-03-06T13:50:57.333 INFO:teuthology.orchestra.run.vm06.stdout: 8532144 0 drwxr-xr-x 2 ubuntu ubuntu 6 Mar 6 13:50 /home/ubuntu/cephtest 2026-03-06T13:50:57.334 DEBUG:teuthology.run_tasks:Unwinding manager console_log 2026-03-06T13:50:57.339 INFO:teuthology.run:Summary data: description: orch:cephadm:workunits/{0-distro/centos_9.stream_runc agent/on mon_election/connectivity task/test_set_mon_crush_locations} duration: 626.4958233833313 flavor: default owner: irq0 success: true 2026-03-06T13:50:57.339 DEBUG:teuthology.report:Pushing job info to http://localhost:8080 2026-03-06T13:50:57.363 INFO:teuthology.run:pass