2026-03-09T20:46:47.983 INFO:root:teuthology version: 1.2.4.dev6+g1c580df7a 2026-03-09T20:46:47.992 DEBUG:teuthology.report:Pushing job info to http://localhost:8080 2026-03-09T20:46:48.014 INFO:teuthology.run:Config: archive_path: /archive/kyr-2026-03-09_11:23:05-orch-squid-none-default-vps/653 branch: squid description: orch/cephadm/workunits/{0-distro/centos_9.stream_runc agent/off mon_election/classic task/test_set_mon_crush_locations} email: null first_in_suite: false flavor: default job_id: '653' last_in_suite: false machine_type: vps name: kyr-2026-03-09_11:23:05-orch-squid-none-default-vps no_nested_subset: false os_type: centos os_version: 9.stream overrides: admin_socket: branch: squid ansible.cephlab: branch: main skip_tags: nagios,monitoring-scripts,hostname,pubkeys,zap,sudoers,kerberos,ntp-client,resolvconf,cpan,nfs vars: timezone: UTC ceph: conf: global: mon election default strategy: 1 mgr: debug mgr: 20 debug ms: 1 mgr/cephadm/use_agent: false mon: debug mon: 20 debug ms: 1 debug paxos: 20 osd: debug ms: 1 debug osd: 20 osd mclock iops capacity threshold hdd: 49000 flavor: default log-ignorelist: - \(MDS_ALL_DOWN\) - \(MDS_UP_LESS_THAN_MAX\) - MON_DOWN - POOL_APP_NOT_ENABLED - mon down - mons down - out of quorum - CEPHADM_FAILED_DAEMON log-only-match: - CEPHADM_ sha1: e911bdebe5c8faa3800735d1568fcdca65db60df ceph-deploy: conf: client: log file: /var/log/ceph/ceph-$name.$pid.log mon: {} install: ceph: flavor: default sha1: e911bdebe5c8faa3800735d1568fcdca65db60df extra_system_packages: deb: - python3-xmltodict - python3-jmespath rpm: - bzip2 - perl-Test-Harness - python3-xmltodict - python3-jmespath selinux: allowlist: - scontext=system_u:system_r:logrotate_t:s0 workunit: branch: tt-squid sha1: 569c3e99c9b32a51b4eaf08731c728f4513ed589 owner: kyr priority: 1000 repo: https://github.com/ceph/ceph.git roles: - - host.a - osd.0 - mon.a - mgr.a - - host.b - osd.1 - mon.b - mgr.b - - host.c - osd.2 - mon.c seed: 3443 sha1: e911bdebe5c8faa3800735d1568fcdca65db60df sleep_before_teardown: 0 subset: 1/64 suite: orch suite_branch: tt-squid suite_path: /home/teuthos/src/github.com_kshtsk_ceph_569c3e99c9b32a51b4eaf08731c728f4513ed589/qa suite_relpath: qa suite_repo: https://github.com/kshtsk/ceph.git suite_sha1: 569c3e99c9b32a51b4eaf08731c728f4513ed589 targets: vm01.local: ecdsa-sha2-nistp256 AAAAE2VjZHNhLXNoYTItbmlzdHAyNTYAAAAIbmlzdHAyNTYAAABBBMGU1FsHEdemluZlIHIMj1fdQl/eFuC2O6qwMOrXUv4AZ0zrDixJLPYkymMxOkiewKeW3P8hqjIGROfgkV2Bhzs= vm06.local: ecdsa-sha2-nistp256 AAAAE2VjZHNhLXNoYTItbmlzdHAyNTYAAAAIbmlzdHAyNTYAAABBBAmXGjyBboZ6z2nJd8VOL5j6/++qJHuyHHanZH/PcAjoGLu28RDEpTP/fEr7E9O4ZfnZ0DjvHb4OeNoR09yh4CA= vm08.local: ecdsa-sha2-nistp256 AAAAE2VjZHNhLXNoYTItbmlzdHAyNTYAAAAIbmlzdHAyNTYAAABBBPVrqxwXGDuBfAcnv9cxVjD593KORijs8A3TrgQ1cP0EQUlX3eB2tRlX/+loN1SsRA5abp/49zTHvpkooa6Nq/Q= tasks: - pexec: all: - sudo dnf remove nvme-cli -y - sudo dnf install runc nvmetcli nvme-cli -y - sudo sed -i 's/^#runtime = "crun"/runtime = "runc"/g' /usr/share/containers/containers.conf - sudo sed -i 's/runtime = "crun"/#runtime = "crun"/g' /usr/share/containers/containers.conf - install: null - cephadm: null - cephadm.apply: specs: - placement: count: 3 service_id: foo service_type: mon spec: crush_locations: host.a: - datacenter=a host.b: - datacenter=b - rack=2 host.c: - datacenter=a - rack=3 - cephadm.shell: host.a: - "set -ex\n# since we don't know the real hostnames before the test, the next\n\ # bit is in order to replace the fake hostnames \"host.a/b/c\" with\n# the actual\ \ names cephadm knows the host by within the mon spec\nceph orch host ls --format\ \ json | jq -r '.[] | .hostname' > realnames\necho $'host.a\\nhost.b\\nhost.c'\ \ > fakenames\necho $'a\\nb\\nc' > mon_ids\necho $'{datacenter=a}\\n{datacenter=b,rack=2}\\\ n{datacenter=a,rack=3}' > crush_locs\nceph orch ls --service-name mon --export\ \ > mon.yaml\nMONSPEC=`cat mon.yaml`\necho \"$MONSPEC\"\nwhile read realname\ \ <&3 && read fakename <&4; do\n MONSPEC=\"${MONSPEC//$fakename/$realname}\"\ \ndone 3 mon.yaml\ncat mon.yaml\n\ # now the spec should have the real hostnames, so let's re-apply\nceph orch\ \ apply -i mon.yaml\nsleep 90\nceph orch ps --refresh\nceph orch ls --service-name\ \ mon --export > mon.yaml; ceph orch apply -i mon.yaml\nsleep 90\nceph mon dump\n\ ceph mon dump --format json\n# verify all the crush locations got set from \"\ ceph mon dump\" output\nwhile read monid <&3 && read crushloc <&4; do\n ceph\ \ mon dump --format json | jq --arg monid \"$monid\" --arg crushloc \"$crushloc\"\ \ -e '.mons | .[] | select(.name == $monid) | .crush_location == $crushloc'\n\ done 3, func=.kill_console_loggers at 0x7ff2b9822170>, signals=[15]) 2026-03-09T20:46:48.786 INFO:teuthology.run_tasks:Running task internal.connect... 2026-03-09T20:46:48.787 INFO:teuthology.task.internal:Opening connections... 2026-03-09T20:46:48.787 DEBUG:teuthology.task.internal:connecting to ubuntu@vm01.local 2026-03-09T20:46:48.787 DEBUG:teuthology.orchestra.connection:{'hostname': 'vm01.local', 'username': 'ubuntu', 'timeout': 60} 2026-03-09T20:46:48.844 DEBUG:teuthology.task.internal:connecting to ubuntu@vm06.local 2026-03-09T20:46:48.845 DEBUG:teuthology.orchestra.connection:{'hostname': 'vm06.local', 'username': 'ubuntu', 'timeout': 60} 2026-03-09T20:46:48.904 DEBUG:teuthology.task.internal:connecting to ubuntu@vm08.local 2026-03-09T20:46:48.905 DEBUG:teuthology.orchestra.connection:{'hostname': 'vm08.local', 'username': 'ubuntu', 'timeout': 60} 2026-03-09T20:46:48.964 INFO:teuthology.run_tasks:Running task internal.push_inventory... 2026-03-09T20:46:48.966 DEBUG:teuthology.orchestra.run.vm01:> uname -m 2026-03-09T20:46:48.981 INFO:teuthology.orchestra.run.vm01.stdout:x86_64 2026-03-09T20:46:48.981 DEBUG:teuthology.orchestra.run.vm01:> cat /etc/os-release 2026-03-09T20:46:49.035 INFO:teuthology.orchestra.run.vm01.stdout:NAME="CentOS Stream" 2026-03-09T20:46:49.035 INFO:teuthology.orchestra.run.vm01.stdout:VERSION="9" 2026-03-09T20:46:49.035 INFO:teuthology.orchestra.run.vm01.stdout:ID="centos" 2026-03-09T20:46:49.035 INFO:teuthology.orchestra.run.vm01.stdout:ID_LIKE="rhel fedora" 2026-03-09T20:46:49.035 INFO:teuthology.orchestra.run.vm01.stdout:VERSION_ID="9" 2026-03-09T20:46:49.035 INFO:teuthology.orchestra.run.vm01.stdout:PLATFORM_ID="platform:el9" 2026-03-09T20:46:49.035 INFO:teuthology.orchestra.run.vm01.stdout:PRETTY_NAME="CentOS Stream 9" 2026-03-09T20:46:49.035 INFO:teuthology.orchestra.run.vm01.stdout:ANSI_COLOR="0;31" 2026-03-09T20:46:49.035 INFO:teuthology.orchestra.run.vm01.stdout:LOGO="fedora-logo-icon" 2026-03-09T20:46:49.035 INFO:teuthology.orchestra.run.vm01.stdout:CPE_NAME="cpe:/o:centos:centos:9" 2026-03-09T20:46:49.035 INFO:teuthology.orchestra.run.vm01.stdout:HOME_URL="https://centos.org/" 2026-03-09T20:46:49.035 INFO:teuthology.orchestra.run.vm01.stdout:BUG_REPORT_URL="https://issues.redhat.com/" 2026-03-09T20:46:49.035 INFO:teuthology.orchestra.run.vm01.stdout:REDHAT_SUPPORT_PRODUCT="Red Hat Enterprise Linux 9" 2026-03-09T20:46:49.035 INFO:teuthology.orchestra.run.vm01.stdout:REDHAT_SUPPORT_PRODUCT_VERSION="CentOS Stream" 2026-03-09T20:46:49.036 INFO:teuthology.lock.ops:Updating vm01.local on lock server 2026-03-09T20:46:49.041 DEBUG:teuthology.orchestra.run.vm06:> uname -m 2026-03-09T20:46:49.058 INFO:teuthology.orchestra.run.vm06.stdout:x86_64 2026-03-09T20:46:49.058 DEBUG:teuthology.orchestra.run.vm06:> cat /etc/os-release 2026-03-09T20:46:49.113 INFO:teuthology.orchestra.run.vm06.stdout:NAME="CentOS Stream" 2026-03-09T20:46:49.113 INFO:teuthology.orchestra.run.vm06.stdout:VERSION="9" 2026-03-09T20:46:49.113 INFO:teuthology.orchestra.run.vm06.stdout:ID="centos" 2026-03-09T20:46:49.113 INFO:teuthology.orchestra.run.vm06.stdout:ID_LIKE="rhel fedora" 2026-03-09T20:46:49.113 INFO:teuthology.orchestra.run.vm06.stdout:VERSION_ID="9" 2026-03-09T20:46:49.113 INFO:teuthology.orchestra.run.vm06.stdout:PLATFORM_ID="platform:el9" 2026-03-09T20:46:49.113 INFO:teuthology.orchestra.run.vm06.stdout:PRETTY_NAME="CentOS Stream 9" 2026-03-09T20:46:49.113 INFO:teuthology.orchestra.run.vm06.stdout:ANSI_COLOR="0;31" 2026-03-09T20:46:49.113 INFO:teuthology.orchestra.run.vm06.stdout:LOGO="fedora-logo-icon" 2026-03-09T20:46:49.113 INFO:teuthology.orchestra.run.vm06.stdout:CPE_NAME="cpe:/o:centos:centos:9" 2026-03-09T20:46:49.113 INFO:teuthology.orchestra.run.vm06.stdout:HOME_URL="https://centos.org/" 2026-03-09T20:46:49.113 INFO:teuthology.orchestra.run.vm06.stdout:BUG_REPORT_URL="https://issues.redhat.com/" 2026-03-09T20:46:49.113 INFO:teuthology.orchestra.run.vm06.stdout:REDHAT_SUPPORT_PRODUCT="Red Hat Enterprise Linux 9" 2026-03-09T20:46:49.113 INFO:teuthology.orchestra.run.vm06.stdout:REDHAT_SUPPORT_PRODUCT_VERSION="CentOS Stream" 2026-03-09T20:46:49.113 INFO:teuthology.lock.ops:Updating vm06.local on lock server 2026-03-09T20:46:49.119 DEBUG:teuthology.orchestra.run.vm08:> uname -m 2026-03-09T20:46:49.134 INFO:teuthology.orchestra.run.vm08.stdout:x86_64 2026-03-09T20:46:49.134 DEBUG:teuthology.orchestra.run.vm08:> cat /etc/os-release 2026-03-09T20:46:49.190 INFO:teuthology.orchestra.run.vm08.stdout:NAME="CentOS Stream" 2026-03-09T20:46:49.190 INFO:teuthology.orchestra.run.vm08.stdout:VERSION="9" 2026-03-09T20:46:49.190 INFO:teuthology.orchestra.run.vm08.stdout:ID="centos" 2026-03-09T20:46:49.190 INFO:teuthology.orchestra.run.vm08.stdout:ID_LIKE="rhel fedora" 2026-03-09T20:46:49.190 INFO:teuthology.orchestra.run.vm08.stdout:VERSION_ID="9" 2026-03-09T20:46:49.190 INFO:teuthology.orchestra.run.vm08.stdout:PLATFORM_ID="platform:el9" 2026-03-09T20:46:49.190 INFO:teuthology.orchestra.run.vm08.stdout:PRETTY_NAME="CentOS Stream 9" 2026-03-09T20:46:49.190 INFO:teuthology.orchestra.run.vm08.stdout:ANSI_COLOR="0;31" 2026-03-09T20:46:49.190 INFO:teuthology.orchestra.run.vm08.stdout:LOGO="fedora-logo-icon" 2026-03-09T20:46:49.190 INFO:teuthology.orchestra.run.vm08.stdout:CPE_NAME="cpe:/o:centos:centos:9" 2026-03-09T20:46:49.190 INFO:teuthology.orchestra.run.vm08.stdout:HOME_URL="https://centos.org/" 2026-03-09T20:46:49.190 INFO:teuthology.orchestra.run.vm08.stdout:BUG_REPORT_URL="https://issues.redhat.com/" 2026-03-09T20:46:49.190 INFO:teuthology.orchestra.run.vm08.stdout:REDHAT_SUPPORT_PRODUCT="Red Hat Enterprise Linux 9" 2026-03-09T20:46:49.190 INFO:teuthology.orchestra.run.vm08.stdout:REDHAT_SUPPORT_PRODUCT_VERSION="CentOS Stream" 2026-03-09T20:46:49.191 INFO:teuthology.lock.ops:Updating vm08.local on lock server 2026-03-09T20:46:49.195 INFO:teuthology.run_tasks:Running task internal.serialize_remote_roles... 2026-03-09T20:46:49.197 INFO:teuthology.run_tasks:Running task internal.check_conflict... 2026-03-09T20:46:49.199 INFO:teuthology.task.internal:Checking for old test directory... 2026-03-09T20:46:49.199 DEBUG:teuthology.orchestra.run.vm01:> test '!' -e /home/ubuntu/cephtest 2026-03-09T20:46:49.201 DEBUG:teuthology.orchestra.run.vm06:> test '!' -e /home/ubuntu/cephtest 2026-03-09T20:46:49.203 DEBUG:teuthology.orchestra.run.vm08:> test '!' -e /home/ubuntu/cephtest 2026-03-09T20:46:49.246 INFO:teuthology.run_tasks:Running task internal.check_ceph_data... 2026-03-09T20:46:49.247 INFO:teuthology.task.internal:Checking for non-empty /var/lib/ceph... 2026-03-09T20:46:49.247 DEBUG:teuthology.orchestra.run.vm01:> test -z $(ls -A /var/lib/ceph) 2026-03-09T20:46:49.256 DEBUG:teuthology.orchestra.run.vm06:> test -z $(ls -A /var/lib/ceph) 2026-03-09T20:46:49.259 DEBUG:teuthology.orchestra.run.vm08:> test -z $(ls -A /var/lib/ceph) 2026-03-09T20:46:49.269 INFO:teuthology.orchestra.run.vm01.stderr:ls: cannot access '/var/lib/ceph': No such file or directory 2026-03-09T20:46:49.272 INFO:teuthology.orchestra.run.vm06.stderr:ls: cannot access '/var/lib/ceph': No such file or directory 2026-03-09T20:46:49.306 INFO:teuthology.orchestra.run.vm08.stderr:ls: cannot access '/var/lib/ceph': No such file or directory 2026-03-09T20:46:49.306 INFO:teuthology.run_tasks:Running task internal.vm_setup... 2026-03-09T20:46:49.314 DEBUG:teuthology.orchestra.run.vm01:> test -e /ceph-qa-ready 2026-03-09T20:46:49.327 DEBUG:teuthology.orchestra.run:got remote process result: 1 2026-03-09T20:46:49.512 DEBUG:teuthology.orchestra.run.vm06:> test -e /ceph-qa-ready 2026-03-09T20:46:49.528 DEBUG:teuthology.orchestra.run:got remote process result: 1 2026-03-09T20:46:49.731 DEBUG:teuthology.orchestra.run.vm08:> test -e /ceph-qa-ready 2026-03-09T20:46:49.746 DEBUG:teuthology.orchestra.run:got remote process result: 1 2026-03-09T20:46:49.943 INFO:teuthology.run_tasks:Running task internal.base... 2026-03-09T20:46:49.944 INFO:teuthology.task.internal:Creating test directory... 2026-03-09T20:46:49.944 DEBUG:teuthology.orchestra.run.vm01:> mkdir -p -m0755 -- /home/ubuntu/cephtest 2026-03-09T20:46:49.946 DEBUG:teuthology.orchestra.run.vm06:> mkdir -p -m0755 -- /home/ubuntu/cephtest 2026-03-09T20:46:49.948 DEBUG:teuthology.orchestra.run.vm08:> mkdir -p -m0755 -- /home/ubuntu/cephtest 2026-03-09T20:46:49.966 INFO:teuthology.run_tasks:Running task internal.archive_upload... 2026-03-09T20:46:49.967 INFO:teuthology.run_tasks:Running task internal.archive... 2026-03-09T20:46:49.968 INFO:teuthology.task.internal:Creating archive directory... 2026-03-09T20:46:49.968 DEBUG:teuthology.orchestra.run.vm01:> install -d -m0755 -- /home/ubuntu/cephtest/archive 2026-03-09T20:46:50.006 DEBUG:teuthology.orchestra.run.vm06:> install -d -m0755 -- /home/ubuntu/cephtest/archive 2026-03-09T20:46:50.008 DEBUG:teuthology.orchestra.run.vm08:> install -d -m0755 -- /home/ubuntu/cephtest/archive 2026-03-09T20:46:50.027 INFO:teuthology.run_tasks:Running task internal.coredump... 2026-03-09T20:46:50.029 INFO:teuthology.task.internal:Enabling coredump saving... 2026-03-09T20:46:50.029 DEBUG:teuthology.orchestra.run.vm01:> test -f /run/.containerenv -o -f /.dockerenv 2026-03-09T20:46:50.076 DEBUG:teuthology.orchestra.run:got remote process result: 1 2026-03-09T20:46:50.077 DEBUG:teuthology.orchestra.run.vm06:> test -f /run/.containerenv -o -f /.dockerenv 2026-03-09T20:46:50.090 DEBUG:teuthology.orchestra.run:got remote process result: 1 2026-03-09T20:46:50.090 DEBUG:teuthology.orchestra.run.vm08:> test -f /run/.containerenv -o -f /.dockerenv 2026-03-09T20:46:50.106 DEBUG:teuthology.orchestra.run:got remote process result: 1 2026-03-09T20:46:50.106 DEBUG:teuthology.orchestra.run.vm01:> install -d -m0755 -- /home/ubuntu/cephtest/archive/coredump && sudo sysctl -w kernel.core_pattern=/home/ubuntu/cephtest/archive/coredump/%t.%p.core && echo kernel.core_pattern=/home/ubuntu/cephtest/archive/coredump/%t.%p.core | sudo tee -a /etc/sysctl.conf 2026-03-09T20:46:50.119 DEBUG:teuthology.orchestra.run.vm06:> install -d -m0755 -- /home/ubuntu/cephtest/archive/coredump && sudo sysctl -w kernel.core_pattern=/home/ubuntu/cephtest/archive/coredump/%t.%p.core && echo kernel.core_pattern=/home/ubuntu/cephtest/archive/coredump/%t.%p.core | sudo tee -a /etc/sysctl.conf 2026-03-09T20:46:50.132 DEBUG:teuthology.orchestra.run.vm08:> install -d -m0755 -- /home/ubuntu/cephtest/archive/coredump && sudo sysctl -w kernel.core_pattern=/home/ubuntu/cephtest/archive/coredump/%t.%p.core && echo kernel.core_pattern=/home/ubuntu/cephtest/archive/coredump/%t.%p.core | sudo tee -a /etc/sysctl.conf 2026-03-09T20:46:50.142 INFO:teuthology.orchestra.run.vm01.stdout:kernel.core_pattern = /home/ubuntu/cephtest/archive/coredump/%t.%p.core 2026-03-09T20:46:50.151 INFO:teuthology.orchestra.run.vm01.stdout:kernel.core_pattern=/home/ubuntu/cephtest/archive/coredump/%t.%p.core 2026-03-09T20:46:50.158 INFO:teuthology.orchestra.run.vm06.stdout:kernel.core_pattern = /home/ubuntu/cephtest/archive/coredump/%t.%p.core 2026-03-09T20:46:50.169 INFO:teuthology.orchestra.run.vm06.stdout:kernel.core_pattern=/home/ubuntu/cephtest/archive/coredump/%t.%p.core 2026-03-09T20:46:50.172 INFO:teuthology.orchestra.run.vm08.stdout:kernel.core_pattern = /home/ubuntu/cephtest/archive/coredump/%t.%p.core 2026-03-09T20:46:50.181 INFO:teuthology.orchestra.run.vm08.stdout:kernel.core_pattern=/home/ubuntu/cephtest/archive/coredump/%t.%p.core 2026-03-09T20:46:50.182 INFO:teuthology.run_tasks:Running task internal.sudo... 2026-03-09T20:46:50.184 INFO:teuthology.task.internal:Configuring sudo... 2026-03-09T20:46:50.184 DEBUG:teuthology.orchestra.run.vm01:> sudo sed -i.orig.teuthology -e 's/^\([^#]*\) \(requiretty\)/\1 !\2/g' -e 's/^\([^#]*\) !\(visiblepw\)/\1 \2/g' /etc/sudoers 2026-03-09T20:46:50.195 DEBUG:teuthology.orchestra.run.vm06:> sudo sed -i.orig.teuthology -e 's/^\([^#]*\) \(requiretty\)/\1 !\2/g' -e 's/^\([^#]*\) !\(visiblepw\)/\1 \2/g' /etc/sudoers 2026-03-09T20:46:50.212 DEBUG:teuthology.orchestra.run.vm08:> sudo sed -i.orig.teuthology -e 's/^\([^#]*\) \(requiretty\)/\1 !\2/g' -e 's/^\([^#]*\) !\(visiblepw\)/\1 \2/g' /etc/sudoers 2026-03-09T20:46:50.247 INFO:teuthology.run_tasks:Running task internal.syslog... 2026-03-09T20:46:50.249 INFO:teuthology.task.internal.syslog:Starting syslog monitoring... 2026-03-09T20:46:50.249 DEBUG:teuthology.orchestra.run.vm01:> mkdir -p -m0755 -- /home/ubuntu/cephtest/archive/syslog 2026-03-09T20:46:50.258 DEBUG:teuthology.orchestra.run.vm06:> mkdir -p -m0755 -- /home/ubuntu/cephtest/archive/syslog 2026-03-09T20:46:50.277 DEBUG:teuthology.orchestra.run.vm08:> mkdir -p -m0755 -- /home/ubuntu/cephtest/archive/syslog 2026-03-09T20:46:50.303 DEBUG:teuthology.orchestra.run.vm01:> install -m 666 /dev/null /home/ubuntu/cephtest/archive/syslog/kern.log 2026-03-09T20:46:50.335 DEBUG:teuthology.orchestra.run.vm01:> install -m 666 /dev/null /home/ubuntu/cephtest/archive/syslog/misc.log 2026-03-09T20:46:50.394 DEBUG:teuthology.orchestra.run.vm01:> set -ex 2026-03-09T20:46:50.394 DEBUG:teuthology.orchestra.run.vm01:> sudo dd of=/etc/rsyslog.d/80-cephtest.conf 2026-03-09T20:46:50.452 DEBUG:teuthology.orchestra.run.vm06:> install -m 666 /dev/null /home/ubuntu/cephtest/archive/syslog/kern.log 2026-03-09T20:46:50.476 DEBUG:teuthology.orchestra.run.vm06:> install -m 666 /dev/null /home/ubuntu/cephtest/archive/syslog/misc.log 2026-03-09T20:46:50.532 DEBUG:teuthology.orchestra.run.vm06:> set -ex 2026-03-09T20:46:50.532 DEBUG:teuthology.orchestra.run.vm06:> sudo dd of=/etc/rsyslog.d/80-cephtest.conf 2026-03-09T20:46:50.589 DEBUG:teuthology.orchestra.run.vm08:> install -m 666 /dev/null /home/ubuntu/cephtest/archive/syslog/kern.log 2026-03-09T20:46:50.612 DEBUG:teuthology.orchestra.run.vm08:> install -m 666 /dev/null /home/ubuntu/cephtest/archive/syslog/misc.log 2026-03-09T20:46:50.668 DEBUG:teuthology.orchestra.run.vm08:> set -ex 2026-03-09T20:46:50.669 DEBUG:teuthology.orchestra.run.vm08:> sudo dd of=/etc/rsyslog.d/80-cephtest.conf 2026-03-09T20:46:50.730 DEBUG:teuthology.orchestra.run.vm01:> sudo service rsyslog restart 2026-03-09T20:46:50.732 DEBUG:teuthology.orchestra.run.vm06:> sudo service rsyslog restart 2026-03-09T20:46:50.734 DEBUG:teuthology.orchestra.run.vm08:> sudo service rsyslog restart 2026-03-09T20:46:50.759 INFO:teuthology.orchestra.run.vm01.stderr:Redirecting to /bin/systemctl restart rsyslog.service 2026-03-09T20:46:50.763 INFO:teuthology.orchestra.run.vm06.stderr:Redirecting to /bin/systemctl restart rsyslog.service 2026-03-09T20:46:50.802 INFO:teuthology.orchestra.run.vm08.stderr:Redirecting to /bin/systemctl restart rsyslog.service 2026-03-09T20:46:51.234 INFO:teuthology.run_tasks:Running task internal.timer... 2026-03-09T20:46:51.236 INFO:teuthology.task.internal:Starting timer... 2026-03-09T20:46:51.236 INFO:teuthology.run_tasks:Running task pcp... 2026-03-09T20:46:51.239 INFO:teuthology.run_tasks:Running task selinux... 2026-03-09T20:46:51.241 DEBUG:teuthology.task:Applying overrides for task selinux: {'allowlist': ['scontext=system_u:system_r:logrotate_t:s0']} 2026-03-09T20:46:51.241 INFO:teuthology.task.selinux:Excluding vm01: VMs are not yet supported 2026-03-09T20:46:51.241 INFO:teuthology.task.selinux:Excluding vm06: VMs are not yet supported 2026-03-09T20:46:51.241 INFO:teuthology.task.selinux:Excluding vm08: VMs are not yet supported 2026-03-09T20:46:51.241 DEBUG:teuthology.task.selinux:Getting current SELinux state 2026-03-09T20:46:51.241 DEBUG:teuthology.task.selinux:Existing SELinux modes: {} 2026-03-09T20:46:51.241 INFO:teuthology.task.selinux:Putting SELinux into permissive mode 2026-03-09T20:46:51.241 INFO:teuthology.run_tasks:Running task ansible.cephlab... 2026-03-09T20:46:51.243 DEBUG:teuthology.task:Applying overrides for task ansible.cephlab: {'branch': 'main', 'skip_tags': 'nagios,monitoring-scripts,hostname,pubkeys,zap,sudoers,kerberos,ntp-client,resolvconf,cpan,nfs', 'vars': {'timezone': 'UTC'}} 2026-03-09T20:46:51.243 DEBUG:teuthology.repo_utils:Setting repo remote to https://github.com/ceph/ceph-cm-ansible.git 2026-03-09T20:46:51.245 INFO:teuthology.repo_utils:Fetching github.com_ceph_ceph-cm-ansible_main from origin 2026-03-09T20:46:54.417 DEBUG:teuthology.repo_utils:Resetting repo at /home/teuthos/src/github.com_ceph_ceph-cm-ansible_main to origin/main 2026-03-09T20:46:54.423 INFO:teuthology.task.ansible:Playbook: [{'import_playbook': 'ansible_managed.yml'}, {'import_playbook': 'teuthology.yml'}, {'hosts': 'testnodes', 'tasks': [{'set_fact': {'ran_from_cephlab_playbook': True}}]}, {'import_playbook': 'testnodes.yml'}, {'import_playbook': 'container-host.yml'}, {'import_playbook': 'cobbler.yml'}, {'import_playbook': 'paddles.yml'}, {'import_playbook': 'pulpito.yml'}, {'hosts': 'testnodes', 'become': True, 'tasks': [{'name': 'Touch /ceph-qa-ready', 'file': {'path': '/ceph-qa-ready', 'state': 'touch'}, 'when': 'ran_from_cephlab_playbook|bool'}]}] 2026-03-09T20:46:54.423 DEBUG:teuthology.task.ansible:Running ansible-playbook -v --extra-vars '{"ansible_ssh_user": "ubuntu", "timezone": "UTC"}' -i /tmp/teuth_ansible_inventoryqahqudcf --limit vm01.local,vm06.local,vm08.local /home/teuthos/src/github.com_ceph_ceph-cm-ansible_main/cephlab.yml --skip-tags nagios,monitoring-scripts,hostname,pubkeys,zap,sudoers,kerberos,ntp-client,resolvconf,cpan,nfs 2026-03-09T20:49:07.964 DEBUG:teuthology.task.ansible:Reconnecting to [Remote(name='ubuntu@vm01.local'), Remote(name='ubuntu@vm06.local'), Remote(name='ubuntu@vm08.local')] 2026-03-09T20:49:07.964 INFO:teuthology.orchestra.remote:Trying to reconnect to host 'ubuntu@vm01.local' 2026-03-09T20:49:07.965 DEBUG:teuthology.orchestra.connection:{'hostname': 'vm01.local', 'username': 'ubuntu', 'timeout': 60} 2026-03-09T20:49:08.029 DEBUG:teuthology.orchestra.run.vm01:> true 2026-03-09T20:49:08.129 INFO:teuthology.orchestra.remote:Successfully reconnected to host 'ubuntu@vm01.local' 2026-03-09T20:49:08.129 INFO:teuthology.orchestra.remote:Trying to reconnect to host 'ubuntu@vm06.local' 2026-03-09T20:49:08.129 DEBUG:teuthology.orchestra.connection:{'hostname': 'vm06.local', 'username': 'ubuntu', 'timeout': 60} 2026-03-09T20:49:08.197 DEBUG:teuthology.orchestra.run.vm06:> true 2026-03-09T20:49:08.274 INFO:teuthology.orchestra.remote:Successfully reconnected to host 'ubuntu@vm06.local' 2026-03-09T20:49:08.274 INFO:teuthology.orchestra.remote:Trying to reconnect to host 'ubuntu@vm08.local' 2026-03-09T20:49:08.275 DEBUG:teuthology.orchestra.connection:{'hostname': 'vm08.local', 'username': 'ubuntu', 'timeout': 60} 2026-03-09T20:49:08.341 DEBUG:teuthology.orchestra.run.vm08:> true 2026-03-09T20:49:08.423 INFO:teuthology.orchestra.remote:Successfully reconnected to host 'ubuntu@vm08.local' 2026-03-09T20:49:08.423 INFO:teuthology.run_tasks:Running task clock... 2026-03-09T20:49:08.426 INFO:teuthology.task.clock:Syncing clocks and checking initial clock skew... 2026-03-09T20:49:08.426 INFO:teuthology.orchestra.run:Running command with timeout 360 2026-03-09T20:49:08.426 DEBUG:teuthology.orchestra.run.vm01:> sudo systemctl stop ntp.service || sudo systemctl stop ntpd.service || sudo systemctl stop chronyd.service ; sudo ntpd -gq || sudo chronyc makestep ; sudo systemctl start ntp.service || sudo systemctl start ntpd.service || sudo systemctl start chronyd.service ; PATH=/usr/bin:/usr/sbin ntpq -p || PATH=/usr/bin:/usr/sbin chronyc sources || true 2026-03-09T20:49:08.428 INFO:teuthology.orchestra.run:Running command with timeout 360 2026-03-09T20:49:08.428 DEBUG:teuthology.orchestra.run.vm06:> sudo systemctl stop ntp.service || sudo systemctl stop ntpd.service || sudo systemctl stop chronyd.service ; sudo ntpd -gq || sudo chronyc makestep ; sudo systemctl start ntp.service || sudo systemctl start ntpd.service || sudo systemctl start chronyd.service ; PATH=/usr/bin:/usr/sbin ntpq -p || PATH=/usr/bin:/usr/sbin chronyc sources || true 2026-03-09T20:49:08.432 INFO:teuthology.orchestra.run:Running command with timeout 360 2026-03-09T20:49:08.432 DEBUG:teuthology.orchestra.run.vm08:> sudo systemctl stop ntp.service || sudo systemctl stop ntpd.service || sudo systemctl stop chronyd.service ; sudo ntpd -gq || sudo chronyc makestep ; sudo systemctl start ntp.service || sudo systemctl start ntpd.service || sudo systemctl start chronyd.service ; PATH=/usr/bin:/usr/sbin ntpq -p || PATH=/usr/bin:/usr/sbin chronyc sources || true 2026-03-09T20:49:08.472 INFO:teuthology.orchestra.run.vm06.stderr:Failed to stop ntp.service: Unit ntp.service not loaded. 2026-03-09T20:49:08.482 INFO:teuthology.orchestra.run.vm01.stderr:Failed to stop ntp.service: Unit ntp.service not loaded. 2026-03-09T20:49:08.485 INFO:teuthology.orchestra.run.vm06.stderr:Failed to stop ntpd.service: Unit ntpd.service not loaded. 2026-03-09T20:49:08.498 INFO:teuthology.orchestra.run.vm01.stderr:Failed to stop ntpd.service: Unit ntpd.service not loaded. 2026-03-09T20:49:08.508 INFO:teuthology.orchestra.run.vm08.stderr:Failed to stop ntp.service: Unit ntp.service not loaded. 2026-03-09T20:49:08.509 INFO:teuthology.orchestra.run.vm06.stderr:sudo: ntpd: command not found 2026-03-09T20:49:08.520 INFO:teuthology.orchestra.run.vm06.stdout:506 Cannot talk to daemon 2026-03-09T20:49:08.524 INFO:teuthology.orchestra.run.vm08.stderr:Failed to stop ntpd.service: Unit ntpd.service not loaded. 2026-03-09T20:49:08.532 INFO:teuthology.orchestra.run.vm01.stderr:sudo: ntpd: command not found 2026-03-09T20:49:08.533 INFO:teuthology.orchestra.run.vm06.stderr:Failed to start ntp.service: Unit ntp.service not found. 2026-03-09T20:49:08.543 INFO:teuthology.orchestra.run.vm01.stdout:506 Cannot talk to daemon 2026-03-09T20:49:08.546 INFO:teuthology.orchestra.run.vm06.stderr:Failed to start ntpd.service: Unit ntpd.service not found. 2026-03-09T20:49:08.554 INFO:teuthology.orchestra.run.vm08.stderr:sudo: ntpd: command not found 2026-03-09T20:49:08.557 INFO:teuthology.orchestra.run.vm01.stderr:Failed to start ntp.service: Unit ntp.service not found. 2026-03-09T20:49:08.567 INFO:teuthology.orchestra.run.vm08.stdout:506 Cannot talk to daemon 2026-03-09T20:49:08.575 INFO:teuthology.orchestra.run.vm01.stderr:Failed to start ntpd.service: Unit ntpd.service not found. 2026-03-09T20:49:08.584 INFO:teuthology.orchestra.run.vm08.stderr:Failed to start ntp.service: Unit ntp.service not found. 2026-03-09T20:49:08.595 INFO:teuthology.orchestra.run.vm06.stderr:bash: line 1: ntpq: command not found 2026-03-09T20:49:08.600 INFO:teuthology.orchestra.run.vm08.stderr:Failed to start ntpd.service: Unit ntpd.service not found. 2026-03-09T20:49:08.601 INFO:teuthology.orchestra.run.vm06.stdout:MS Name/IP address Stratum Poll Reach LastRx Last sample 2026-03-09T20:49:08.601 INFO:teuthology.orchestra.run.vm06.stdout:=============================================================================== 2026-03-09T20:49:08.601 INFO:teuthology.orchestra.run.vm06.stdout:^? fa.gnudb.org 0 6 0 - +0ns[ +0ns] +/- 0ns 2026-03-09T20:49:08.601 INFO:teuthology.orchestra.run.vm06.stdout:^? ntp2.uni-ulm.de 0 6 0 - +0ns[ +0ns] +/- 0ns 2026-03-09T20:49:08.601 INFO:teuthology.orchestra.run.vm06.stdout:^? static.119.109.140.128.c> 0 6 0 - +0ns[ +0ns] +/- 0ns 2026-03-09T20:49:08.601 INFO:teuthology.orchestra.run.vm06.stdout:^? x1.ncomputers.org 0 6 0 - +0ns[ +0ns] +/- 0ns 2026-03-09T20:49:08.627 INFO:teuthology.orchestra.run.vm01.stderr:bash: line 1: ntpq: command not found 2026-03-09T20:49:08.629 INFO:teuthology.orchestra.run.vm01.stdout:MS Name/IP address Stratum Poll Reach LastRx Last sample 2026-03-09T20:49:08.629 INFO:teuthology.orchestra.run.vm01.stdout:=============================================================================== 2026-03-09T20:49:08.629 INFO:teuthology.orchestra.run.vm01.stdout:^? ntp2.uni-ulm.de 0 6 0 - +0ns[ +0ns] +/- 0ns 2026-03-09T20:49:08.629 INFO:teuthology.orchestra.run.vm01.stdout:^? static.119.109.140.128.c> 0 6 0 - +0ns[ +0ns] +/- 0ns 2026-03-09T20:49:08.629 INFO:teuthology.orchestra.run.vm01.stdout:^? x1.ncomputers.org 0 6 0 - +0ns[ +0ns] +/- 0ns 2026-03-09T20:49:08.629 INFO:teuthology.orchestra.run.vm01.stdout:^? fa.gnudb.org 0 6 0 - +0ns[ +0ns] +/- 0ns 2026-03-09T20:49:08.652 INFO:teuthology.orchestra.run.vm08.stderr:bash: line 1: ntpq: command not found 2026-03-09T20:49:08.655 INFO:teuthology.orchestra.run.vm08.stdout:MS Name/IP address Stratum Poll Reach LastRx Last sample 2026-03-09T20:49:08.655 INFO:teuthology.orchestra.run.vm08.stdout:=============================================================================== 2026-03-09T20:49:08.655 INFO:teuthology.orchestra.run.vm08.stdout:^? static.119.109.140.128.c> 0 6 0 - +0ns[ +0ns] +/- 0ns 2026-03-09T20:49:08.655 INFO:teuthology.orchestra.run.vm08.stdout:^? x1.ncomputers.org 0 6 0 - +0ns[ +0ns] +/- 0ns 2026-03-09T20:49:08.655 INFO:teuthology.orchestra.run.vm08.stdout:^? fa.gnudb.org 0 6 0 - +0ns[ +0ns] +/- 0ns 2026-03-09T20:49:08.655 INFO:teuthology.orchestra.run.vm08.stdout:^? ntp2.uni-ulm.de 0 6 0 - +0ns[ +0ns] +/- 0ns 2026-03-09T20:49:08.655 INFO:teuthology.run_tasks:Running task pexec... 2026-03-09T20:49:08.659 INFO:teuthology.task.pexec:Executing custom commands... 2026-03-09T20:49:08.659 DEBUG:teuthology.orchestra.run.vm01:> TESTDIR=/home/ubuntu/cephtest bash -s 2026-03-09T20:49:08.659 DEBUG:teuthology.orchestra.run.vm06:> TESTDIR=/home/ubuntu/cephtest bash -s 2026-03-09T20:49:08.659 DEBUG:teuthology.orchestra.run.vm08:> TESTDIR=/home/ubuntu/cephtest bash -s 2026-03-09T20:49:08.661 DEBUG:teuthology.task.pexec:ubuntu@vm06.local< sudo dnf remove nvme-cli -y 2026-03-09T20:49:08.661 DEBUG:teuthology.task.pexec:ubuntu@vm06.local< sudo dnf install runc nvmetcli nvme-cli -y 2026-03-09T20:49:08.661 DEBUG:teuthology.task.pexec:ubuntu@vm06.local< sudo sed -i 's/^#runtime = "crun"/runtime = "runc"/g' /usr/share/containers/containers.conf 2026-03-09T20:49:08.661 DEBUG:teuthology.task.pexec:ubuntu@vm06.local< sudo sed -i 's/runtime = "crun"/#runtime = "crun"/g' /usr/share/containers/containers.conf 2026-03-09T20:49:08.661 INFO:teuthology.task.pexec:Running commands on host ubuntu@vm06.local 2026-03-09T20:49:08.661 INFO:teuthology.task.pexec:sudo dnf remove nvme-cli -y 2026-03-09T20:49:08.661 INFO:teuthology.task.pexec:sudo dnf install runc nvmetcli nvme-cli -y 2026-03-09T20:49:08.661 INFO:teuthology.task.pexec:sudo sed -i 's/^#runtime = "crun"/runtime = "runc"/g' /usr/share/containers/containers.conf 2026-03-09T20:49:08.661 INFO:teuthology.task.pexec:sudo sed -i 's/runtime = "crun"/#runtime = "crun"/g' /usr/share/containers/containers.conf 2026-03-09T20:49:08.662 DEBUG:teuthology.task.pexec:ubuntu@vm01.local< sudo dnf remove nvme-cli -y 2026-03-09T20:49:08.663 DEBUG:teuthology.task.pexec:ubuntu@vm01.local< sudo dnf install runc nvmetcli nvme-cli -y 2026-03-09T20:49:08.663 DEBUG:teuthology.task.pexec:ubuntu@vm01.local< sudo sed -i 's/^#runtime = "crun"/runtime = "runc"/g' /usr/share/containers/containers.conf 2026-03-09T20:49:08.663 DEBUG:teuthology.task.pexec:ubuntu@vm01.local< sudo sed -i 's/runtime = "crun"/#runtime = "crun"/g' /usr/share/containers/containers.conf 2026-03-09T20:49:08.663 INFO:teuthology.task.pexec:Running commands on host ubuntu@vm01.local 2026-03-09T20:49:08.663 INFO:teuthology.task.pexec:sudo dnf remove nvme-cli -y 2026-03-09T20:49:08.663 INFO:teuthology.task.pexec:sudo dnf install runc nvmetcli nvme-cli -y 2026-03-09T20:49:08.663 INFO:teuthology.task.pexec:sudo sed -i 's/^#runtime = "crun"/runtime = "runc"/g' /usr/share/containers/containers.conf 2026-03-09T20:49:08.663 INFO:teuthology.task.pexec:sudo sed -i 's/runtime = "crun"/#runtime = "crun"/g' /usr/share/containers/containers.conf 2026-03-09T20:49:08.698 DEBUG:teuthology.task.pexec:ubuntu@vm08.local< sudo dnf remove nvme-cli -y 2026-03-09T20:49:08.698 DEBUG:teuthology.task.pexec:ubuntu@vm08.local< sudo dnf install runc nvmetcli nvme-cli -y 2026-03-09T20:49:08.698 DEBUG:teuthology.task.pexec:ubuntu@vm08.local< sudo sed -i 's/^#runtime = "crun"/runtime = "runc"/g' /usr/share/containers/containers.conf 2026-03-09T20:49:08.698 DEBUG:teuthology.task.pexec:ubuntu@vm08.local< sudo sed -i 's/runtime = "crun"/#runtime = "crun"/g' /usr/share/containers/containers.conf 2026-03-09T20:49:08.699 INFO:teuthology.task.pexec:Running commands on host ubuntu@vm08.local 2026-03-09T20:49:08.699 INFO:teuthology.task.pexec:sudo dnf remove nvme-cli -y 2026-03-09T20:49:08.699 INFO:teuthology.task.pexec:sudo dnf install runc nvmetcli nvme-cli -y 2026-03-09T20:49:08.699 INFO:teuthology.task.pexec:sudo sed -i 's/^#runtime = "crun"/runtime = "runc"/g' /usr/share/containers/containers.conf 2026-03-09T20:49:08.699 INFO:teuthology.task.pexec:sudo sed -i 's/runtime = "crun"/#runtime = "crun"/g' /usr/share/containers/containers.conf 2026-03-09T20:49:08.873 INFO:teuthology.orchestra.run.vm06.stdout:No match for argument: nvme-cli 2026-03-09T20:49:08.874 INFO:teuthology.orchestra.run.vm06.stderr:No packages marked for removal. 2026-03-09T20:49:08.877 INFO:teuthology.orchestra.run.vm06.stdout:Dependencies resolved. 2026-03-09T20:49:08.877 INFO:teuthology.orchestra.run.vm06.stdout:Nothing to do. 2026-03-09T20:49:08.877 INFO:teuthology.orchestra.run.vm06.stdout:Complete! 2026-03-09T20:49:08.897 INFO:teuthology.orchestra.run.vm01.stdout:No match for argument: nvme-cli 2026-03-09T20:49:08.898 INFO:teuthology.orchestra.run.vm01.stderr:No packages marked for removal. 2026-03-09T20:49:08.903 INFO:teuthology.orchestra.run.vm01.stdout:Dependencies resolved. 2026-03-09T20:49:08.904 INFO:teuthology.orchestra.run.vm01.stdout:Nothing to do. 2026-03-09T20:49:08.904 INFO:teuthology.orchestra.run.vm01.stdout:Complete! 2026-03-09T20:49:08.913 INFO:teuthology.orchestra.run.vm08.stdout:No match for argument: nvme-cli 2026-03-09T20:49:08.913 INFO:teuthology.orchestra.run.vm08.stderr:No packages marked for removal. 2026-03-09T20:49:08.916 INFO:teuthology.orchestra.run.vm08.stdout:Dependencies resolved. 2026-03-09T20:49:08.917 INFO:teuthology.orchestra.run.vm08.stdout:Nothing to do. 2026-03-09T20:49:08.917 INFO:teuthology.orchestra.run.vm08.stdout:Complete! 2026-03-09T20:49:09.284 INFO:teuthology.orchestra.run.vm06.stdout:Last metadata expiration check: 0:01:33 ago on Mon 09 Mar 2026 08:47:36 PM UTC. 2026-03-09T20:49:09.357 INFO:teuthology.orchestra.run.vm08.stdout:Last metadata expiration check: 0:01:34 ago on Mon 09 Mar 2026 08:47:35 PM UTC. 2026-03-09T20:49:09.383 INFO:teuthology.orchestra.run.vm06.stdout:Dependencies resolved. 2026-03-09T20:49:09.383 INFO:teuthology.orchestra.run.vm06.stdout:================================================================================ 2026-03-09T20:49:09.383 INFO:teuthology.orchestra.run.vm06.stdout: Package Arch Version Repository Size 2026-03-09T20:49:09.383 INFO:teuthology.orchestra.run.vm06.stdout:================================================================================ 2026-03-09T20:49:09.383 INFO:teuthology.orchestra.run.vm06.stdout:Installing: 2026-03-09T20:49:09.383 INFO:teuthology.orchestra.run.vm06.stdout: nvme-cli x86_64 2.16-1.el9 baseos 1.2 M 2026-03-09T20:49:09.383 INFO:teuthology.orchestra.run.vm06.stdout: nvmetcli noarch 0.8-3.el9 baseos 44 k 2026-03-09T20:49:09.383 INFO:teuthology.orchestra.run.vm06.stdout: runc x86_64 4:1.4.0-2.el9 appstream 4.0 M 2026-03-09T20:49:09.383 INFO:teuthology.orchestra.run.vm06.stdout:Installing dependencies: 2026-03-09T20:49:09.383 INFO:teuthology.orchestra.run.vm06.stdout: python3-configshell noarch 1:1.1.30-1.el9 baseos 72 k 2026-03-09T20:49:09.383 INFO:teuthology.orchestra.run.vm06.stdout: python3-kmod x86_64 0.9-32.el9 baseos 84 k 2026-03-09T20:49:09.383 INFO:teuthology.orchestra.run.vm06.stdout: python3-pyparsing noarch 2.4.7-9.el9 baseos 150 k 2026-03-09T20:49:09.384 INFO:teuthology.orchestra.run.vm06.stdout: python3-urwid x86_64 2.1.2-4.el9 baseos 837 k 2026-03-09T20:49:09.384 INFO:teuthology.orchestra.run.vm06.stdout: 2026-03-09T20:49:09.384 INFO:teuthology.orchestra.run.vm06.stdout:Transaction Summary 2026-03-09T20:49:09.384 INFO:teuthology.orchestra.run.vm06.stdout:================================================================================ 2026-03-09T20:49:09.384 INFO:teuthology.orchestra.run.vm06.stdout:Install 7 Packages 2026-03-09T20:49:09.384 INFO:teuthology.orchestra.run.vm06.stdout: 2026-03-09T20:49:09.384 INFO:teuthology.orchestra.run.vm06.stdout:Total download size: 6.3 M 2026-03-09T20:49:09.384 INFO:teuthology.orchestra.run.vm06.stdout:Installed size: 24 M 2026-03-09T20:49:09.384 INFO:teuthology.orchestra.run.vm06.stdout:Downloading Packages: 2026-03-09T20:49:09.462 INFO:teuthology.orchestra.run.vm08.stdout:Dependencies resolved. 2026-03-09T20:49:09.462 INFO:teuthology.orchestra.run.vm08.stdout:================================================================================ 2026-03-09T20:49:09.462 INFO:teuthology.orchestra.run.vm08.stdout: Package Arch Version Repository Size 2026-03-09T20:49:09.462 INFO:teuthology.orchestra.run.vm08.stdout:================================================================================ 2026-03-09T20:49:09.462 INFO:teuthology.orchestra.run.vm08.stdout:Installing: 2026-03-09T20:49:09.462 INFO:teuthology.orchestra.run.vm08.stdout: nvme-cli x86_64 2.16-1.el9 baseos 1.2 M 2026-03-09T20:49:09.462 INFO:teuthology.orchestra.run.vm08.stdout: nvmetcli noarch 0.8-3.el9 baseos 44 k 2026-03-09T20:49:09.462 INFO:teuthology.orchestra.run.vm08.stdout: runc x86_64 4:1.4.0-2.el9 appstream 4.0 M 2026-03-09T20:49:09.462 INFO:teuthology.orchestra.run.vm08.stdout:Installing dependencies: 2026-03-09T20:49:09.462 INFO:teuthology.orchestra.run.vm08.stdout: python3-configshell noarch 1:1.1.30-1.el9 baseos 72 k 2026-03-09T20:49:09.463 INFO:teuthology.orchestra.run.vm08.stdout: python3-kmod x86_64 0.9-32.el9 baseos 84 k 2026-03-09T20:49:09.463 INFO:teuthology.orchestra.run.vm08.stdout: python3-pyparsing noarch 2.4.7-9.el9 baseos 150 k 2026-03-09T20:49:09.463 INFO:teuthology.orchestra.run.vm08.stdout: python3-urwid x86_64 2.1.2-4.el9 baseos 837 k 2026-03-09T20:49:09.463 INFO:teuthology.orchestra.run.vm08.stdout: 2026-03-09T20:49:09.463 INFO:teuthology.orchestra.run.vm08.stdout:Transaction Summary 2026-03-09T20:49:09.463 INFO:teuthology.orchestra.run.vm08.stdout:================================================================================ 2026-03-09T20:49:09.463 INFO:teuthology.orchestra.run.vm08.stdout:Install 7 Packages 2026-03-09T20:49:09.463 INFO:teuthology.orchestra.run.vm08.stdout: 2026-03-09T20:49:09.463 INFO:teuthology.orchestra.run.vm08.stdout:Total download size: 6.3 M 2026-03-09T20:49:09.463 INFO:teuthology.orchestra.run.vm08.stdout:Installed size: 24 M 2026-03-09T20:49:09.463 INFO:teuthology.orchestra.run.vm08.stdout:Downloading Packages: 2026-03-09T20:49:09.478 INFO:teuthology.orchestra.run.vm01.stdout:Last metadata expiration check: 0:01:11 ago on Mon 09 Mar 2026 08:47:58 PM UTC. 2026-03-09T20:49:09.629 INFO:teuthology.orchestra.run.vm01.stdout:Dependencies resolved. 2026-03-09T20:49:09.629 INFO:teuthology.orchestra.run.vm01.stdout:================================================================================ 2026-03-09T20:49:09.629 INFO:teuthology.orchestra.run.vm01.stdout: Package Arch Version Repository Size 2026-03-09T20:49:09.629 INFO:teuthology.orchestra.run.vm01.stdout:================================================================================ 2026-03-09T20:49:09.629 INFO:teuthology.orchestra.run.vm01.stdout:Installing: 2026-03-09T20:49:09.629 INFO:teuthology.orchestra.run.vm01.stdout: nvme-cli x86_64 2.16-1.el9 baseos 1.2 M 2026-03-09T20:49:09.629 INFO:teuthology.orchestra.run.vm01.stdout: nvmetcli noarch 0.8-3.el9 baseos 44 k 2026-03-09T20:49:09.629 INFO:teuthology.orchestra.run.vm01.stdout: runc x86_64 4:1.4.0-2.el9 appstream 4.0 M 2026-03-09T20:49:09.629 INFO:teuthology.orchestra.run.vm01.stdout:Installing dependencies: 2026-03-09T20:49:09.629 INFO:teuthology.orchestra.run.vm01.stdout: python3-configshell noarch 1:1.1.30-1.el9 baseos 72 k 2026-03-09T20:49:09.629 INFO:teuthology.orchestra.run.vm01.stdout: python3-kmod x86_64 0.9-32.el9 baseos 84 k 2026-03-09T20:49:09.629 INFO:teuthology.orchestra.run.vm01.stdout: python3-pyparsing noarch 2.4.7-9.el9 baseos 150 k 2026-03-09T20:49:09.629 INFO:teuthology.orchestra.run.vm01.stdout: python3-urwid x86_64 2.1.2-4.el9 baseos 837 k 2026-03-09T20:49:09.629 INFO:teuthology.orchestra.run.vm01.stdout: 2026-03-09T20:49:09.629 INFO:teuthology.orchestra.run.vm01.stdout:Transaction Summary 2026-03-09T20:49:09.629 INFO:teuthology.orchestra.run.vm01.stdout:================================================================================ 2026-03-09T20:49:09.629 INFO:teuthology.orchestra.run.vm01.stdout:Install 7 Packages 2026-03-09T20:49:09.629 INFO:teuthology.orchestra.run.vm01.stdout: 2026-03-09T20:49:09.633 INFO:teuthology.orchestra.run.vm01.stdout:Total download size: 6.3 M 2026-03-09T20:49:09.633 INFO:teuthology.orchestra.run.vm01.stdout:Installed size: 24 M 2026-03-09T20:49:09.634 INFO:teuthology.orchestra.run.vm01.stdout:Downloading Packages: 2026-03-09T20:49:10.269 INFO:teuthology.orchestra.run.vm06.stdout:(1/7): nvmetcli-0.8-3.el9.noarch.rpm 201 kB/s | 44 kB 00:00 2026-03-09T20:49:10.270 INFO:teuthology.orchestra.run.vm06.stdout:(2/7): python3-configshell-1.1.30-1.el9.noarch. 328 kB/s | 72 kB 00:00 2026-03-09T20:49:10.335 INFO:teuthology.orchestra.run.vm06.stdout:(3/7): python3-kmod-0.9-32.el9.x86_64.rpm 1.3 MB/s | 84 kB 00:00 2026-03-09T20:49:10.370 INFO:teuthology.orchestra.run.vm06.stdout:(4/7): python3-pyparsing-2.4.7-9.el9.noarch.rpm 1.5 MB/s | 150 kB 00:00 2026-03-09T20:49:10.372 INFO:teuthology.orchestra.run.vm08.stdout:(1/7): nvmetcli-0.8-3.el9.noarch.rpm 123 kB/s | 44 kB 00:00 2026-03-09T20:49:10.396 INFO:teuthology.orchestra.run.vm01.stdout:(1/7): nvmetcli-0.8-3.el9.noarch.rpm 387 kB/s | 44 kB 00:00 2026-03-09T20:49:10.397 INFO:teuthology.orchestra.run.vm01.stdout:(2/7): python3-configshell-1.1.30-1.el9.noarch. 629 kB/s | 72 kB 00:00 2026-03-09T20:49:10.402 INFO:teuthology.orchestra.run.vm06.stdout:(5/7): nvme-cli-2.16-1.el9.x86_64.rpm 3.3 MB/s | 1.2 MB 00:00 2026-03-09T20:49:10.405 INFO:teuthology.orchestra.run.vm08.stdout:(2/7): python3-configshell-1.1.30-1.el9.noarch. 185 kB/s | 72 kB 00:00 2026-03-09T20:49:10.440 INFO:teuthology.orchestra.run.vm06.stdout:(6/7): python3-urwid-2.1.2-4.el9.x86_64.rpm 7.8 MB/s | 837 kB 00:00 2026-03-09T20:49:10.453 INFO:teuthology.orchestra.run.vm01.stdout:(3/7): python3-kmod-0.9-32.el9.x86_64.rpm 1.5 MB/s | 84 kB 00:00 2026-03-09T20:49:10.482 INFO:teuthology.orchestra.run.vm01.stdout:(4/7): python3-pyparsing-2.4.7-9.el9.noarch.rpm 1.7 MB/s | 150 kB 00:00 2026-03-09T20:49:10.511 INFO:teuthology.orchestra.run.vm01.stdout:(5/7): nvme-cli-2.16-1.el9.x86_64.rpm 5.0 MB/s | 1.2 MB 00:00 2026-03-09T20:49:10.549 INFO:teuthology.orchestra.run.vm08.stdout:(3/7): python3-kmod-0.9-32.el9.x86_64.rpm 477 kB/s | 84 kB 00:00 2026-03-09T20:49:10.550 INFO:teuthology.orchestra.run.vm08.stdout:(4/7): python3-pyparsing-2.4.7-9.el9.noarch.rpm 1.0 MB/s | 150 kB 00:00 2026-03-09T20:49:10.570 INFO:teuthology.orchestra.run.vm01.stdout:(6/7): python3-urwid-2.1.2-4.el9.x86_64.rpm 6.9 MB/s | 837 kB 00:00 2026-03-09T20:49:10.605 INFO:teuthology.orchestra.run.vm08.stdout:(5/7): nvme-cli-2.16-1.el9.x86_64.rpm 2.0 MB/s | 1.2 MB 00:00 2026-03-09T20:49:10.700 INFO:teuthology.orchestra.run.vm08.stdout:(6/7): python3-urwid-2.1.2-4.el9.x86_64.rpm 5.4 MB/s | 837 kB 00:00 2026-03-09T20:49:10.905 INFO:teuthology.orchestra.run.vm06.stdout:(7/7): runc-1.4.0-2.el9.x86_64.rpm 7.4 MB/s | 4.0 MB 00:00 2026-03-09T20:49:10.905 INFO:teuthology.orchestra.run.vm06.stdout:-------------------------------------------------------------------------------- 2026-03-09T20:49:10.905 INFO:teuthology.orchestra.run.vm06.stdout:Total 4.1 MB/s | 6.3 MB 00:01 2026-03-09T20:49:10.976 INFO:teuthology.orchestra.run.vm06.stdout:Running transaction check 2026-03-09T20:49:10.984 INFO:teuthology.orchestra.run.vm06.stdout:Transaction check succeeded. 2026-03-09T20:49:10.984 INFO:teuthology.orchestra.run.vm06.stdout:Running transaction test 2026-03-09T20:49:11.059 INFO:teuthology.orchestra.run.vm06.stdout:Transaction test succeeded. 2026-03-09T20:49:11.059 INFO:teuthology.orchestra.run.vm06.stdout:Running transaction 2026-03-09T20:49:11.182 INFO:teuthology.orchestra.run.vm08.stdout:(7/7): runc-1.4.0-2.el9.x86_64.rpm 6.3 MB/s | 4.0 MB 00:00 2026-03-09T20:49:11.184 INFO:teuthology.orchestra.run.vm08.stdout:-------------------------------------------------------------------------------- 2026-03-09T20:49:11.185 INFO:teuthology.orchestra.run.vm08.stdout:Total 3.6 MB/s | 6.3 MB 00:01 2026-03-09T20:49:11.273 INFO:teuthology.orchestra.run.vm08.stdout:Running transaction check 2026-03-09T20:49:11.284 INFO:teuthology.orchestra.run.vm08.stdout:Transaction check succeeded. 2026-03-09T20:49:11.284 INFO:teuthology.orchestra.run.vm08.stdout:Running transaction test 2026-03-09T20:49:11.311 INFO:teuthology.orchestra.run.vm06.stdout: Preparing : 1/1 2026-03-09T20:49:11.323 INFO:teuthology.orchestra.run.vm06.stdout: Installing : python3-urwid-2.1.2-4.el9.x86_64 1/7 2026-03-09T20:49:11.337 INFO:teuthology.orchestra.run.vm06.stdout: Installing : python3-pyparsing-2.4.7-9.el9.noarch 2/7 2026-03-09T20:49:11.347 INFO:teuthology.orchestra.run.vm06.stdout: Installing : python3-configshell-1:1.1.30-1.el9.noarch 3/7 2026-03-09T20:49:11.359 INFO:teuthology.orchestra.run.vm08.stdout:Transaction test succeeded. 2026-03-09T20:49:11.359 INFO:teuthology.orchestra.run.vm08.stdout:Running transaction 2026-03-09T20:49:11.360 INFO:teuthology.orchestra.run.vm06.stdout: Installing : python3-kmod-0.9-32.el9.x86_64 4/7 2026-03-09T20:49:11.362 INFO:teuthology.orchestra.run.vm06.stdout: Installing : nvmetcli-0.8-3.el9.noarch 5/7 2026-03-09T20:49:11.425 INFO:teuthology.orchestra.run.vm06.stdout: Running scriptlet: nvmetcli-0.8-3.el9.noarch 5/7 2026-03-09T20:49:11.455 INFO:teuthology.orchestra.run.vm01.stdout:(7/7): runc-1.4.0-2.el9.x86_64.rpm 4.1 MB/s | 4.0 MB 00:00 2026-03-09T20:49:11.455 INFO:teuthology.orchestra.run.vm01.stdout:-------------------------------------------------------------------------------- 2026-03-09T20:49:11.455 INFO:teuthology.orchestra.run.vm01.stdout:Total 3.4 MB/s | 6.3 MB 00:01 2026-03-09T20:49:11.536 INFO:teuthology.orchestra.run.vm08.stdout: Preparing : 1/1 2026-03-09T20:49:11.547 INFO:teuthology.orchestra.run.vm01.stdout:Running transaction check 2026-03-09T20:49:11.550 INFO:teuthology.orchestra.run.vm08.stdout: Installing : python3-urwid-2.1.2-4.el9.x86_64 1/7 2026-03-09T20:49:11.555 INFO:teuthology.orchestra.run.vm01.stdout:Transaction check succeeded. 2026-03-09T20:49:11.555 INFO:teuthology.orchestra.run.vm01.stdout:Running transaction test 2026-03-09T20:49:11.561 INFO:teuthology.orchestra.run.vm08.stdout: Installing : python3-pyparsing-2.4.7-9.el9.noarch 2/7 2026-03-09T20:49:11.568 INFO:teuthology.orchestra.run.vm08.stdout: Installing : python3-configshell-1:1.1.30-1.el9.noarch 3/7 2026-03-09T20:49:11.577 INFO:teuthology.orchestra.run.vm08.stdout: Installing : python3-kmod-0.9-32.el9.x86_64 4/7 2026-03-09T20:49:11.578 INFO:teuthology.orchestra.run.vm08.stdout: Installing : nvmetcli-0.8-3.el9.noarch 5/7 2026-03-09T20:49:11.579 INFO:teuthology.orchestra.run.vm06.stdout: Installing : runc-4:1.4.0-2.el9.x86_64 6/7 2026-03-09T20:49:11.585 INFO:teuthology.orchestra.run.vm06.stdout: Installing : nvme-cli-2.16-1.el9.x86_64 7/7 2026-03-09T20:49:11.631 INFO:teuthology.orchestra.run.vm08.stdout: Running scriptlet: nvmetcli-0.8-3.el9.noarch 5/7 2026-03-09T20:49:11.641 INFO:teuthology.orchestra.run.vm01.stdout:Transaction test succeeded. 2026-03-09T20:49:11.641 INFO:teuthology.orchestra.run.vm01.stdout:Running transaction 2026-03-09T20:49:11.775 INFO:teuthology.orchestra.run.vm08.stdout: Installing : runc-4:1.4.0-2.el9.x86_64 6/7 2026-03-09T20:49:11.780 INFO:teuthology.orchestra.run.vm08.stdout: Installing : nvme-cli-2.16-1.el9.x86_64 7/7 2026-03-09T20:49:11.857 INFO:teuthology.orchestra.run.vm01.stdout: Preparing : 1/1 2026-03-09T20:49:11.872 INFO:teuthology.orchestra.run.vm01.stdout: Installing : python3-urwid-2.1.2-4.el9.x86_64 1/7 2026-03-09T20:49:11.887 INFO:teuthology.orchestra.run.vm01.stdout: Installing : python3-pyparsing-2.4.7-9.el9.noarch 2/7 2026-03-09T20:49:11.895 INFO:teuthology.orchestra.run.vm01.stdout: Installing : python3-configshell-1:1.1.30-1.el9.noarch 3/7 2026-03-09T20:49:11.905 INFO:teuthology.orchestra.run.vm01.stdout: Installing : python3-kmod-0.9-32.el9.x86_64 4/7 2026-03-09T20:49:11.910 INFO:teuthology.orchestra.run.vm01.stdout: Installing : nvmetcli-0.8-3.el9.noarch 5/7 2026-03-09T20:49:11.979 INFO:teuthology.orchestra.run.vm01.stdout: Running scriptlet: nvmetcli-0.8-3.el9.noarch 5/7 2026-03-09T20:49:12.006 INFO:teuthology.orchestra.run.vm06.stdout: Running scriptlet: nvme-cli-2.16-1.el9.x86_64 7/7 2026-03-09T20:49:12.007 INFO:teuthology.orchestra.run.vm06.stdout:Created symlink /etc/systemd/system/default.target.wants/nvmefc-boot-connections.service → /usr/lib/systemd/system/nvmefc-boot-connections.service. 2026-03-09T20:49:12.007 INFO:teuthology.orchestra.run.vm06.stdout: 2026-03-09T20:49:12.189 INFO:teuthology.orchestra.run.vm08.stdout: Running scriptlet: nvme-cli-2.16-1.el9.x86_64 7/7 2026-03-09T20:49:12.189 INFO:teuthology.orchestra.run.vm08.stdout:Created symlink /etc/systemd/system/default.target.wants/nvmefc-boot-connections.service → /usr/lib/systemd/system/nvmefc-boot-connections.service. 2026-03-09T20:49:12.189 INFO:teuthology.orchestra.run.vm08.stdout: 2026-03-09T20:49:12.193 INFO:teuthology.orchestra.run.vm01.stdout: Installing : runc-4:1.4.0-2.el9.x86_64 6/7 2026-03-09T20:49:12.198 INFO:teuthology.orchestra.run.vm01.stdout: Installing : nvme-cli-2.16-1.el9.x86_64 7/7 2026-03-09T20:49:12.667 INFO:teuthology.orchestra.run.vm01.stdout: Running scriptlet: nvme-cli-2.16-1.el9.x86_64 7/7 2026-03-09T20:49:12.667 INFO:teuthology.orchestra.run.vm01.stdout:Created symlink /etc/systemd/system/default.target.wants/nvmefc-boot-connections.service → /usr/lib/systemd/system/nvmefc-boot-connections.service. 2026-03-09T20:49:12.667 INFO:teuthology.orchestra.run.vm01.stdout: 2026-03-09T20:49:12.970 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : nvme-cli-2.16-1.el9.x86_64 1/7 2026-03-09T20:49:12.970 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : nvmetcli-0.8-3.el9.noarch 2/7 2026-03-09T20:49:12.970 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-configshell-1:1.1.30-1.el9.noarch 3/7 2026-03-09T20:49:12.970 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-kmod-0.9-32.el9.x86_64 4/7 2026-03-09T20:49:12.970 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-pyparsing-2.4.7-9.el9.noarch 5/7 2026-03-09T20:49:12.970 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-urwid-2.1.2-4.el9.x86_64 6/7 2026-03-09T20:49:13.025 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : nvme-cli-2.16-1.el9.x86_64 1/7 2026-03-09T20:49:13.025 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : nvmetcli-0.8-3.el9.noarch 2/7 2026-03-09T20:49:13.025 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : python3-configshell-1:1.1.30-1.el9.noarch 3/7 2026-03-09T20:49:13.025 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : python3-kmod-0.9-32.el9.x86_64 4/7 2026-03-09T20:49:13.025 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : python3-pyparsing-2.4.7-9.el9.noarch 5/7 2026-03-09T20:49:13.026 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : python3-urwid-2.1.2-4.el9.x86_64 6/7 2026-03-09T20:49:13.065 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : runc-4:1.4.0-2.el9.x86_64 7/7 2026-03-09T20:49:13.065 INFO:teuthology.orchestra.run.vm06.stdout: 2026-03-09T20:49:13.065 INFO:teuthology.orchestra.run.vm06.stdout:Installed: 2026-03-09T20:49:13.065 INFO:teuthology.orchestra.run.vm06.stdout: nvme-cli-2.16-1.el9.x86_64 nvmetcli-0.8-3.el9.noarch 2026-03-09T20:49:13.065 INFO:teuthology.orchestra.run.vm06.stdout: python3-configshell-1:1.1.30-1.el9.noarch python3-kmod-0.9-32.el9.x86_64 2026-03-09T20:49:13.065 INFO:teuthology.orchestra.run.vm06.stdout: python3-pyparsing-2.4.7-9.el9.noarch python3-urwid-2.1.2-4.el9.x86_64 2026-03-09T20:49:13.065 INFO:teuthology.orchestra.run.vm06.stdout: runc-4:1.4.0-2.el9.x86_64 2026-03-09T20:49:13.065 INFO:teuthology.orchestra.run.vm06.stdout: 2026-03-09T20:49:13.065 INFO:teuthology.orchestra.run.vm06.stdout:Complete! 2026-03-09T20:49:13.144 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : runc-4:1.4.0-2.el9.x86_64 7/7 2026-03-09T20:49:13.144 INFO:teuthology.orchestra.run.vm08.stdout: 2026-03-09T20:49:13.144 INFO:teuthology.orchestra.run.vm08.stdout:Installed: 2026-03-09T20:49:13.144 INFO:teuthology.orchestra.run.vm08.stdout: nvme-cli-2.16-1.el9.x86_64 nvmetcli-0.8-3.el9.noarch 2026-03-09T20:49:13.144 INFO:teuthology.orchestra.run.vm08.stdout: python3-configshell-1:1.1.30-1.el9.noarch python3-kmod-0.9-32.el9.x86_64 2026-03-09T20:49:13.144 INFO:teuthology.orchestra.run.vm08.stdout: python3-pyparsing-2.4.7-9.el9.noarch python3-urwid-2.1.2-4.el9.x86_64 2026-03-09T20:49:13.144 INFO:teuthology.orchestra.run.vm08.stdout: runc-4:1.4.0-2.el9.x86_64 2026-03-09T20:49:13.144 INFO:teuthology.orchestra.run.vm08.stdout: 2026-03-09T20:49:13.144 INFO:teuthology.orchestra.run.vm08.stdout:Complete! 2026-03-09T20:49:13.176 DEBUG:teuthology.parallel:result is None 2026-03-09T20:49:13.247 DEBUG:teuthology.parallel:result is None 2026-03-09T20:49:13.564 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : nvme-cli-2.16-1.el9.x86_64 1/7 2026-03-09T20:49:13.564 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : nvmetcli-0.8-3.el9.noarch 2/7 2026-03-09T20:49:13.565 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-configshell-1:1.1.30-1.el9.noarch 3/7 2026-03-09T20:49:13.565 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-kmod-0.9-32.el9.x86_64 4/7 2026-03-09T20:49:13.565 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-pyparsing-2.4.7-9.el9.noarch 5/7 2026-03-09T20:49:13.565 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-urwid-2.1.2-4.el9.x86_64 6/7 2026-03-09T20:49:13.665 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : runc-4:1.4.0-2.el9.x86_64 7/7 2026-03-09T20:49:13.665 INFO:teuthology.orchestra.run.vm01.stdout: 2026-03-09T20:49:13.665 INFO:teuthology.orchestra.run.vm01.stdout:Installed: 2026-03-09T20:49:13.665 INFO:teuthology.orchestra.run.vm01.stdout: nvme-cli-2.16-1.el9.x86_64 nvmetcli-0.8-3.el9.noarch 2026-03-09T20:49:13.665 INFO:teuthology.orchestra.run.vm01.stdout: python3-configshell-1:1.1.30-1.el9.noarch python3-kmod-0.9-32.el9.x86_64 2026-03-09T20:49:13.665 INFO:teuthology.orchestra.run.vm01.stdout: python3-pyparsing-2.4.7-9.el9.noarch python3-urwid-2.1.2-4.el9.x86_64 2026-03-09T20:49:13.665 INFO:teuthology.orchestra.run.vm01.stdout: runc-4:1.4.0-2.el9.x86_64 2026-03-09T20:49:13.665 INFO:teuthology.orchestra.run.vm01.stdout: 2026-03-09T20:49:13.665 INFO:teuthology.orchestra.run.vm01.stdout:Complete! 2026-03-09T20:49:13.760 DEBUG:teuthology.parallel:result is None 2026-03-09T20:49:13.760 INFO:teuthology.run_tasks:Running task install... 2026-03-09T20:49:13.763 DEBUG:teuthology.task.install:project ceph 2026-03-09T20:49:13.763 DEBUG:teuthology.task.install:INSTALL overrides: {'ceph': {'flavor': 'default', 'sha1': 'e911bdebe5c8faa3800735d1568fcdca65db60df'}, 'extra_system_packages': {'deb': ['python3-xmltodict', 'python3-jmespath'], 'rpm': ['bzip2', 'perl-Test-Harness', 'python3-xmltodict', 'python3-jmespath']}} 2026-03-09T20:49:13.763 DEBUG:teuthology.task.install:config {'flavor': 'default', 'sha1': 'e911bdebe5c8faa3800735d1568fcdca65db60df', 'extra_system_packages': {'deb': ['python3-xmltodict', 'python3-jmespath'], 'rpm': ['bzip2', 'perl-Test-Harness', 'python3-xmltodict', 'python3-jmespath']}} 2026-03-09T20:49:13.763 INFO:teuthology.task.install:Using flavor: default 2026-03-09T20:49:13.767 DEBUG:teuthology.task.install:Package list is: {'deb': ['ceph', 'cephadm', 'ceph-mds', 'ceph-mgr', 'ceph-common', 'ceph-fuse', 'ceph-test', 'ceph-volume', 'radosgw', 'python3-rados', 'python3-rgw', 'python3-cephfs', 'python3-rbd', 'libcephfs2', 'libcephfs-dev', 'librados2', 'librbd1', 'rbd-fuse'], 'rpm': ['ceph-radosgw', 'ceph-test', 'ceph', 'ceph-base', 'cephadm', 'ceph-immutable-object-cache', 'ceph-mgr', 'ceph-mgr-dashboard', 'ceph-mgr-diskprediction-local', 'ceph-mgr-rook', 'ceph-mgr-cephadm', 'ceph-fuse', 'ceph-volume', 'librados-devel', 'libcephfs2', 'libcephfs-devel', 'librados2', 'librbd1', 'python3-rados', 'python3-rgw', 'python3-cephfs', 'python3-rbd', 'rbd-fuse', 'rbd-mirror', 'rbd-nbd']} 2026-03-09T20:49:13.767 INFO:teuthology.task.install:extra packages: [] 2026-03-09T20:49:13.767 DEBUG:teuthology.task.install.rpm:_update_package_list_and_install: config is {'branch': None, 'cleanup': None, 'debuginfo': None, 'downgrade_packages': [], 'exclude_packages': [], 'extra_packages': [], 'extra_system_packages': {'deb': ['python3-xmltodict', 'python3-jmespath'], 'rpm': ['bzip2', 'perl-Test-Harness', 'python3-xmltodict', 'python3-jmespath']}, 'extras': None, 'enable_coprs': [], 'flavor': 'default', 'install_ceph_packages': True, 'packages': {}, 'project': 'ceph', 'repos_only': False, 'sha1': 'e911bdebe5c8faa3800735d1568fcdca65db60df', 'tag': None, 'wait_for_package': False} 2026-03-09T20:49:13.767 DEBUG:teuthology.packaging:Querying https://shaman.ceph.com/api/search?status=ready&project=ceph&flavor=default&distros=centos%2F9%2Fx86_64&sha1=e911bdebe5c8faa3800735d1568fcdca65db60df 2026-03-09T20:49:13.768 DEBUG:teuthology.task.install.rpm:_update_package_list_and_install: config is {'branch': None, 'cleanup': None, 'debuginfo': None, 'downgrade_packages': [], 'exclude_packages': [], 'extra_packages': [], 'extra_system_packages': {'deb': ['python3-xmltodict', 'python3-jmespath'], 'rpm': ['bzip2', 'perl-Test-Harness', 'python3-xmltodict', 'python3-jmespath']}, 'extras': None, 'enable_coprs': [], 'flavor': 'default', 'install_ceph_packages': True, 'packages': {}, 'project': 'ceph', 'repos_only': False, 'sha1': 'e911bdebe5c8faa3800735d1568fcdca65db60df', 'tag': None, 'wait_for_package': False} 2026-03-09T20:49:13.768 DEBUG:teuthology.packaging:Querying https://shaman.ceph.com/api/search?status=ready&project=ceph&flavor=default&distros=centos%2F9%2Fx86_64&sha1=e911bdebe5c8faa3800735d1568fcdca65db60df 2026-03-09T20:49:13.768 DEBUG:teuthology.task.install.rpm:_update_package_list_and_install: config is {'branch': None, 'cleanup': None, 'debuginfo': None, 'downgrade_packages': [], 'exclude_packages': [], 'extra_packages': [], 'extra_system_packages': {'deb': ['python3-xmltodict', 'python3-jmespath'], 'rpm': ['bzip2', 'perl-Test-Harness', 'python3-xmltodict', 'python3-jmespath']}, 'extras': None, 'enable_coprs': [], 'flavor': 'default', 'install_ceph_packages': True, 'packages': {}, 'project': 'ceph', 'repos_only': False, 'sha1': 'e911bdebe5c8faa3800735d1568fcdca65db60df', 'tag': None, 'wait_for_package': False} 2026-03-09T20:49:13.768 DEBUG:teuthology.packaging:Querying https://shaman.ceph.com/api/search?status=ready&project=ceph&flavor=default&distros=centos%2F9%2Fx86_64&sha1=e911bdebe5c8faa3800735d1568fcdca65db60df 2026-03-09T20:49:14.494 INFO:teuthology.task.install.rpm:Pulling from https://3.chacra.ceph.com/r/ceph/squid/e911bdebe5c8faa3800735d1568fcdca65db60df/centos/9/flavors/default/ 2026-03-09T20:49:14.494 INFO:teuthology.task.install.rpm:Package version is 19.2.3-678.ge911bdeb 2026-03-09T20:49:14.528 INFO:teuthology.task.install.rpm:Pulling from https://3.chacra.ceph.com/r/ceph/squid/e911bdebe5c8faa3800735d1568fcdca65db60df/centos/9/flavors/default/ 2026-03-09T20:49:14.528 INFO:teuthology.task.install.rpm:Package version is 19.2.3-678.ge911bdeb 2026-03-09T20:49:14.529 INFO:teuthology.task.install.rpm:Pulling from https://3.chacra.ceph.com/r/ceph/squid/e911bdebe5c8faa3800735d1568fcdca65db60df/centos/9/flavors/default/ 2026-03-09T20:49:14.529 INFO:teuthology.task.install.rpm:Package version is 19.2.3-678.ge911bdeb 2026-03-09T20:49:15.015 INFO:teuthology.packaging:Writing yum repo: [ceph] name=ceph packages for $basearch baseurl=https://3.chacra.ceph.com/r/ceph/squid/e911bdebe5c8faa3800735d1568fcdca65db60df/centos/9/flavors/default/$basearch enabled=1 gpgcheck=0 type=rpm-md [ceph-noarch] name=ceph noarch packages baseurl=https://3.chacra.ceph.com/r/ceph/squid/e911bdebe5c8faa3800735d1568fcdca65db60df/centos/9/flavors/default/noarch enabled=1 gpgcheck=0 type=rpm-md [ceph-source] name=ceph source packages baseurl=https://3.chacra.ceph.com/r/ceph/squid/e911bdebe5c8faa3800735d1568fcdca65db60df/centos/9/flavors/default/SRPMS enabled=1 gpgcheck=0 type=rpm-md 2026-03-09T20:49:15.016 DEBUG:teuthology.orchestra.run.vm08:> set -ex 2026-03-09T20:49:15.016 DEBUG:teuthology.orchestra.run.vm08:> sudo dd of=/etc/yum.repos.d/ceph.repo 2026-03-09T20:49:15.025 INFO:teuthology.packaging:Writing yum repo: [ceph] name=ceph packages for $basearch baseurl=https://3.chacra.ceph.com/r/ceph/squid/e911bdebe5c8faa3800735d1568fcdca65db60df/centos/9/flavors/default/$basearch enabled=1 gpgcheck=0 type=rpm-md [ceph-noarch] name=ceph noarch packages baseurl=https://3.chacra.ceph.com/r/ceph/squid/e911bdebe5c8faa3800735d1568fcdca65db60df/centos/9/flavors/default/noarch enabled=1 gpgcheck=0 type=rpm-md [ceph-source] name=ceph source packages baseurl=https://3.chacra.ceph.com/r/ceph/squid/e911bdebe5c8faa3800735d1568fcdca65db60df/centos/9/flavors/default/SRPMS enabled=1 gpgcheck=0 type=rpm-md 2026-03-09T20:49:15.026 DEBUG:teuthology.orchestra.run.vm06:> set -ex 2026-03-09T20:49:15.026 DEBUG:teuthology.orchestra.run.vm06:> sudo dd of=/etc/yum.repos.d/ceph.repo 2026-03-09T20:49:15.054 INFO:teuthology.packaging:Writing yum repo: [ceph] name=ceph packages for $basearch baseurl=https://3.chacra.ceph.com/r/ceph/squid/e911bdebe5c8faa3800735d1568fcdca65db60df/centos/9/flavors/default/$basearch enabled=1 gpgcheck=0 type=rpm-md [ceph-noarch] name=ceph noarch packages baseurl=https://3.chacra.ceph.com/r/ceph/squid/e911bdebe5c8faa3800735d1568fcdca65db60df/centos/9/flavors/default/noarch enabled=1 gpgcheck=0 type=rpm-md [ceph-source] name=ceph source packages baseurl=https://3.chacra.ceph.com/r/ceph/squid/e911bdebe5c8faa3800735d1568fcdca65db60df/centos/9/flavors/default/SRPMS enabled=1 gpgcheck=0 type=rpm-md 2026-03-09T20:49:15.054 DEBUG:teuthology.orchestra.run.vm01:> set -ex 2026-03-09T20:49:15.054 DEBUG:teuthology.orchestra.run.vm01:> sudo dd of=/etc/yum.repos.d/ceph.repo 2026-03-09T20:49:15.055 INFO:teuthology.task.install.rpm:Installing packages: ceph-radosgw, ceph-test, ceph, ceph-base, cephadm, ceph-immutable-object-cache, ceph-mgr, ceph-mgr-dashboard, ceph-mgr-diskprediction-local, ceph-mgr-rook, ceph-mgr-cephadm, ceph-fuse, ceph-volume, librados-devel, libcephfs2, libcephfs-devel, librados2, librbd1, python3-rados, python3-rgw, python3-cephfs, python3-rbd, rbd-fuse, rbd-mirror, rbd-nbd, bzip2, perl-Test-Harness, python3-xmltodict, python3-jmespath on remote rpm x86_64 2026-03-09T20:49:15.055 DEBUG:teuthology.orchestra.run.vm06:> if test -f /etc/yum.repos.d/ceph.repo ; then sudo sed -i -e ':a;N;$!ba;s/enabled=1\ngpg/enabled=1\npriority=1\ngpg/g' -e 's;ref/[a-zA-Z0-9_-]*/;sha1/e911bdebe5c8faa3800735d1568fcdca65db60df/;g' /etc/yum.repos.d/ceph.repo ; fi 2026-03-09T20:49:15.057 INFO:teuthology.task.install.rpm:Installing packages: ceph-radosgw, ceph-test, ceph, ceph-base, cephadm, ceph-immutable-object-cache, ceph-mgr, ceph-mgr-dashboard, ceph-mgr-diskprediction-local, ceph-mgr-rook, ceph-mgr-cephadm, ceph-fuse, ceph-volume, librados-devel, libcephfs2, libcephfs-devel, librados2, librbd1, python3-rados, python3-rgw, python3-cephfs, python3-rbd, rbd-fuse, rbd-mirror, rbd-nbd, bzip2, perl-Test-Harness, python3-xmltodict, python3-jmespath on remote rpm x86_64 2026-03-09T20:49:15.057 DEBUG:teuthology.orchestra.run.vm08:> if test -f /etc/yum.repos.d/ceph.repo ; then sudo sed -i -e ':a;N;$!ba;s/enabled=1\ngpg/enabled=1\npriority=1\ngpg/g' -e 's;ref/[a-zA-Z0-9_-]*/;sha1/e911bdebe5c8faa3800735d1568fcdca65db60df/;g' /etc/yum.repos.d/ceph.repo ; fi 2026-03-09T20:49:15.098 INFO:teuthology.task.install.rpm:Installing packages: ceph-radosgw, ceph-test, ceph, ceph-base, cephadm, ceph-immutable-object-cache, ceph-mgr, ceph-mgr-dashboard, ceph-mgr-diskprediction-local, ceph-mgr-rook, ceph-mgr-cephadm, ceph-fuse, ceph-volume, librados-devel, libcephfs2, libcephfs-devel, librados2, librbd1, python3-rados, python3-rgw, python3-cephfs, python3-rbd, rbd-fuse, rbd-mirror, rbd-nbd, bzip2, perl-Test-Harness, python3-xmltodict, python3-jmespath on remote rpm x86_64 2026-03-09T20:49:15.098 DEBUG:teuthology.orchestra.run.vm01:> if test -f /etc/yum.repos.d/ceph.repo ; then sudo sed -i -e ':a;N;$!ba;s/enabled=1\ngpg/enabled=1\npriority=1\ngpg/g' -e 's;ref/[a-zA-Z0-9_-]*/;sha1/e911bdebe5c8faa3800735d1568fcdca65db60df/;g' /etc/yum.repos.d/ceph.repo ; fi 2026-03-09T20:49:15.137 DEBUG:teuthology.orchestra.run.vm06:> sudo touch -a /etc/yum/pluginconf.d/priorities.conf ; test -e /etc/yum/pluginconf.d/priorities.conf.orig || sudo cp -af /etc/yum/pluginconf.d/priorities.conf /etc/yum/pluginconf.d/priorities.conf.orig 2026-03-09T20:49:15.149 DEBUG:teuthology.orchestra.run.vm08:> sudo touch -a /etc/yum/pluginconf.d/priorities.conf ; test -e /etc/yum/pluginconf.d/priorities.conf.orig || sudo cp -af /etc/yum/pluginconf.d/priorities.conf /etc/yum/pluginconf.d/priorities.conf.orig 2026-03-09T20:49:15.182 DEBUG:teuthology.orchestra.run.vm01:> sudo touch -a /etc/yum/pluginconf.d/priorities.conf ; test -e /etc/yum/pluginconf.d/priorities.conf.orig || sudo cp -af /etc/yum/pluginconf.d/priorities.conf /etc/yum/pluginconf.d/priorities.conf.orig 2026-03-09T20:49:15.193 DEBUG:teuthology.orchestra.run.vm08:> grep check_obsoletes /etc/yum/pluginconf.d/priorities.conf && sudo sed -i 's/check_obsoletes.*0/check_obsoletes = 1/g' /etc/yum/pluginconf.d/priorities.conf || echo 'check_obsoletes = 1' | sudo tee -a /etc/yum/pluginconf.d/priorities.conf 2026-03-09T20:49:15.217 DEBUG:teuthology.orchestra.run.vm06:> grep check_obsoletes /etc/yum/pluginconf.d/priorities.conf && sudo sed -i 's/check_obsoletes.*0/check_obsoletes = 1/g' /etc/yum/pluginconf.d/priorities.conf || echo 'check_obsoletes = 1' | sudo tee -a /etc/yum/pluginconf.d/priorities.conf 2026-03-09T20:49:15.228 INFO:teuthology.orchestra.run.vm08.stdout:check_obsoletes = 1 2026-03-09T20:49:15.230 DEBUG:teuthology.orchestra.run.vm08:> sudo yum clean all 2026-03-09T20:49:15.270 DEBUG:teuthology.orchestra.run.vm01:> grep check_obsoletes /etc/yum/pluginconf.d/priorities.conf && sudo sed -i 's/check_obsoletes.*0/check_obsoletes = 1/g' /etc/yum/pluginconf.d/priorities.conf || echo 'check_obsoletes = 1' | sudo tee -a /etc/yum/pluginconf.d/priorities.conf 2026-03-09T20:49:15.292 INFO:teuthology.orchestra.run.vm06.stdout:check_obsoletes = 1 2026-03-09T20:49:15.293 DEBUG:teuthology.orchestra.run.vm06:> sudo yum clean all 2026-03-09T20:49:15.314 INFO:teuthology.orchestra.run.vm01.stdout:check_obsoletes = 1 2026-03-09T20:49:15.316 DEBUG:teuthology.orchestra.run.vm01:> sudo yum clean all 2026-03-09T20:49:15.429 INFO:teuthology.orchestra.run.vm08.stdout:41 files removed 2026-03-09T20:49:15.454 DEBUG:teuthology.orchestra.run.vm08:> sudo yum -y install ceph-radosgw ceph-test ceph ceph-base cephadm ceph-immutable-object-cache ceph-mgr ceph-mgr-dashboard ceph-mgr-diskprediction-local ceph-mgr-rook ceph-mgr-cephadm ceph-fuse ceph-volume librados-devel libcephfs2 libcephfs-devel librados2 librbd1 python3-rados python3-rgw python3-cephfs python3-rbd rbd-fuse rbd-mirror rbd-nbd bzip2 perl-Test-Harness python3-xmltodict python3-jmespath 2026-03-09T20:49:15.491 INFO:teuthology.orchestra.run.vm06.stdout:41 files removed 2026-03-09T20:49:15.507 INFO:teuthology.orchestra.run.vm01.stdout:41 files removed 2026-03-09T20:49:15.514 DEBUG:teuthology.orchestra.run.vm06:> sudo yum -y install ceph-radosgw ceph-test ceph ceph-base cephadm ceph-immutable-object-cache ceph-mgr ceph-mgr-dashboard ceph-mgr-diskprediction-local ceph-mgr-rook ceph-mgr-cephadm ceph-fuse ceph-volume librados-devel libcephfs2 libcephfs-devel librados2 librbd1 python3-rados python3-rgw python3-cephfs python3-rbd rbd-fuse rbd-mirror rbd-nbd bzip2 perl-Test-Harness python3-xmltodict python3-jmespath 2026-03-09T20:49:15.532 DEBUG:teuthology.orchestra.run.vm01:> sudo yum -y install ceph-radosgw ceph-test ceph ceph-base cephadm ceph-immutable-object-cache ceph-mgr ceph-mgr-dashboard ceph-mgr-diskprediction-local ceph-mgr-rook ceph-mgr-cephadm ceph-fuse ceph-volume librados-devel libcephfs2 libcephfs-devel librados2 librbd1 python3-rados python3-rgw python3-cephfs python3-rbd rbd-fuse rbd-mirror rbd-nbd bzip2 perl-Test-Harness python3-xmltodict python3-jmespath 2026-03-09T20:49:16.806 INFO:teuthology.orchestra.run.vm08.stdout:ceph packages for x86_64 72 kB/s | 84 kB 00:01 2026-03-09T20:49:16.874 INFO:teuthology.orchestra.run.vm06.stdout:ceph packages for x86_64 72 kB/s | 84 kB 00:01 2026-03-09T20:49:16.923 INFO:teuthology.orchestra.run.vm01.stdout:ceph packages for x86_64 72 kB/s | 84 kB 00:01 2026-03-09T20:49:17.769 INFO:teuthology.orchestra.run.vm08.stdout:ceph noarch packages 12 kB/s | 12 kB 00:00 2026-03-09T20:49:17.858 INFO:teuthology.orchestra.run.vm06.stdout:ceph noarch packages 12 kB/s | 12 kB 00:00 2026-03-09T20:49:17.895 INFO:teuthology.orchestra.run.vm01.stdout:ceph noarch packages 12 kB/s | 12 kB 00:00 2026-03-09T20:49:18.749 INFO:teuthology.orchestra.run.vm08.stdout:ceph source packages 2.0 kB/s | 1.9 kB 00:00 2026-03-09T20:49:18.818 INFO:teuthology.orchestra.run.vm06.stdout:ceph source packages 2.0 kB/s | 1.9 kB 00:00 2026-03-09T20:49:18.856 INFO:teuthology.orchestra.run.vm01.stdout:ceph source packages 2.0 kB/s | 1.9 kB 00:00 2026-03-09T20:49:19.683 INFO:teuthology.orchestra.run.vm06.stdout:CentOS Stream 9 - BaseOS 10 MB/s | 8.9 MB 00:00 2026-03-09T20:49:19.858 INFO:teuthology.orchestra.run.vm08.stdout:CentOS Stream 9 - BaseOS 8.2 MB/s | 8.9 MB 00:01 2026-03-09T20:49:21.957 INFO:teuthology.orchestra.run.vm06.stdout:CentOS Stream 9 - AppStream 18 MB/s | 27 MB 00:01 2026-03-09T20:49:22.502 INFO:teuthology.orchestra.run.vm01.stdout:CentOS Stream 9 - BaseOS 2.5 MB/s | 8.9 MB 00:03 2026-03-09T20:49:22.784 INFO:teuthology.orchestra.run.vm08.stdout:CentOS Stream 9 - AppStream 12 MB/s | 27 MB 00:02 2026-03-09T20:49:26.601 INFO:teuthology.orchestra.run.vm06.stdout:CentOS Stream 9 - CRB 4.9 MB/s | 8.0 MB 00:01 2026-03-09T20:49:26.611 INFO:teuthology.orchestra.run.vm01.stdout:CentOS Stream 9 - AppStream 8.5 MB/s | 27 MB 00:03 2026-03-09T20:49:26.966 INFO:teuthology.orchestra.run.vm08.stdout:CentOS Stream 9 - CRB 7.2 MB/s | 8.0 MB 00:01 2026-03-09T20:49:27.972 INFO:teuthology.orchestra.run.vm06.stdout:CentOS Stream 9 - Extras packages 47 kB/s | 20 kB 00:00 2026-03-09T20:49:28.536 INFO:teuthology.orchestra.run.vm08.stdout:CentOS Stream 9 - Extras packages 30 kB/s | 20 kB 00:00 2026-03-09T20:49:28.839 INFO:teuthology.orchestra.run.vm06.stdout:Extra Packages for Enterprise Linux 26 MB/s | 20 MB 00:00 2026-03-09T20:49:30.025 INFO:teuthology.orchestra.run.vm08.stdout:Extra Packages for Enterprise Linux 15 MB/s | 20 MB 00:01 2026-03-09T20:49:31.108 INFO:teuthology.orchestra.run.vm01.stdout:CentOS Stream 9 - CRB 6.5 MB/s | 8.0 MB 00:01 2026-03-09T20:49:32.439 INFO:teuthology.orchestra.run.vm01.stdout:CentOS Stream 9 - Extras packages 46 kB/s | 20 kB 00:00 2026-03-09T20:49:33.351 INFO:teuthology.orchestra.run.vm01.stdout:Extra Packages for Enterprise Linux 25 MB/s | 20 MB 00:00 2026-03-09T20:49:34.254 INFO:teuthology.orchestra.run.vm06.stdout:lab-extras 65 kB/s | 50 kB 00:00 2026-03-09T20:49:35.435 INFO:teuthology.orchestra.run.vm08.stdout:lab-extras 61 kB/s | 50 kB 00:00 2026-03-09T20:49:35.747 INFO:teuthology.orchestra.run.vm06.stdout:Package librados2-2:16.2.4-5.el9.x86_64 is already installed. 2026-03-09T20:49:35.747 INFO:teuthology.orchestra.run.vm06.stdout:Package librbd1-2:16.2.4-5.el9.x86_64 is already installed. 2026-03-09T20:49:35.752 INFO:teuthology.orchestra.run.vm06.stdout:Package bzip2-1.0.8-11.el9.x86_64 is already installed. 2026-03-09T20:49:35.752 INFO:teuthology.orchestra.run.vm06.stdout:Package perl-Test-Harness-1:3.42-461.el9.noarch is already installed. 2026-03-09T20:49:35.781 INFO:teuthology.orchestra.run.vm06.stdout:Dependencies resolved. 2026-03-09T20:49:35.786 INFO:teuthology.orchestra.run.vm06.stdout:====================================================================================== 2026-03-09T20:49:35.786 INFO:teuthology.orchestra.run.vm06.stdout: Package Arch Version Repository Size 2026-03-09T20:49:35.786 INFO:teuthology.orchestra.run.vm06.stdout:====================================================================================== 2026-03-09T20:49:35.786 INFO:teuthology.orchestra.run.vm06.stdout:Installing: 2026-03-09T20:49:35.786 INFO:teuthology.orchestra.run.vm06.stdout: ceph x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 6.5 k 2026-03-09T20:49:35.786 INFO:teuthology.orchestra.run.vm06.stdout: ceph-base x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 5.5 M 2026-03-09T20:49:35.786 INFO:teuthology.orchestra.run.vm06.stdout: ceph-fuse x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 1.2 M 2026-03-09T20:49:35.786 INFO:teuthology.orchestra.run.vm06.stdout: ceph-immutable-object-cache x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 145 k 2026-03-09T20:49:35.786 INFO:teuthology.orchestra.run.vm06.stdout: ceph-mgr x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 1.1 M 2026-03-09T20:49:35.786 INFO:teuthology.orchestra.run.vm06.stdout: ceph-mgr-cephadm noarch 2:19.2.3-678.ge911bdeb.el9 ceph-noarch 150 k 2026-03-09T20:49:35.786 INFO:teuthology.orchestra.run.vm06.stdout: ceph-mgr-dashboard noarch 2:19.2.3-678.ge911bdeb.el9 ceph-noarch 3.8 M 2026-03-09T20:49:35.786 INFO:teuthology.orchestra.run.vm06.stdout: ceph-mgr-diskprediction-local noarch 2:19.2.3-678.ge911bdeb.el9 ceph-noarch 7.4 M 2026-03-09T20:49:35.786 INFO:teuthology.orchestra.run.vm06.stdout: ceph-mgr-rook noarch 2:19.2.3-678.ge911bdeb.el9 ceph-noarch 49 k 2026-03-09T20:49:35.786 INFO:teuthology.orchestra.run.vm06.stdout: ceph-radosgw x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 11 M 2026-03-09T20:49:35.786 INFO:teuthology.orchestra.run.vm06.stdout: ceph-test x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 50 M 2026-03-09T20:49:35.786 INFO:teuthology.orchestra.run.vm06.stdout: ceph-volume noarch 2:19.2.3-678.ge911bdeb.el9 ceph-noarch 299 k 2026-03-09T20:49:35.786 INFO:teuthology.orchestra.run.vm06.stdout: cephadm noarch 2:19.2.3-678.ge911bdeb.el9 ceph-noarch 769 k 2026-03-09T20:49:35.786 INFO:teuthology.orchestra.run.vm06.stdout: libcephfs-devel x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 34 k 2026-03-09T20:49:35.786 INFO:teuthology.orchestra.run.vm06.stdout: libcephfs2 x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 1.0 M 2026-03-09T20:49:35.786 INFO:teuthology.orchestra.run.vm06.stdout: librados-devel x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 127 k 2026-03-09T20:49:35.786 INFO:teuthology.orchestra.run.vm06.stdout: python3-cephfs x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 165 k 2026-03-09T20:49:35.786 INFO:teuthology.orchestra.run.vm06.stdout: python3-jmespath noarch 1.0.1-1.el9 appstream 48 k 2026-03-09T20:49:35.786 INFO:teuthology.orchestra.run.vm06.stdout: python3-rados x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 323 k 2026-03-09T20:49:35.786 INFO:teuthology.orchestra.run.vm06.stdout: python3-rbd x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 303 k 2026-03-09T20:49:35.786 INFO:teuthology.orchestra.run.vm06.stdout: python3-rgw x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 100 k 2026-03-09T20:49:35.786 INFO:teuthology.orchestra.run.vm06.stdout: python3-xmltodict noarch 0.12.0-15.el9 epel 22 k 2026-03-09T20:49:35.786 INFO:teuthology.orchestra.run.vm06.stdout: rbd-fuse x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 85 k 2026-03-09T20:49:35.786 INFO:teuthology.orchestra.run.vm06.stdout: rbd-mirror x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 3.1 M 2026-03-09T20:49:35.786 INFO:teuthology.orchestra.run.vm06.stdout: rbd-nbd x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 171 k 2026-03-09T20:49:35.786 INFO:teuthology.orchestra.run.vm06.stdout:Upgrading: 2026-03-09T20:49:35.786 INFO:teuthology.orchestra.run.vm06.stdout: librados2 x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 3.4 M 2026-03-09T20:49:35.786 INFO:teuthology.orchestra.run.vm06.stdout: librbd1 x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 3.2 M 2026-03-09T20:49:35.786 INFO:teuthology.orchestra.run.vm06.stdout:Installing dependencies: 2026-03-09T20:49:35.786 INFO:teuthology.orchestra.run.vm06.stdout: abseil-cpp x86_64 20211102.0-4.el9 epel 551 k 2026-03-09T20:49:35.786 INFO:teuthology.orchestra.run.vm06.stdout: boost-program-options x86_64 1.75.0-13.el9 appstream 104 k 2026-03-09T20:49:35.786 INFO:teuthology.orchestra.run.vm06.stdout: ceph-common x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 22 M 2026-03-09T20:49:35.786 INFO:teuthology.orchestra.run.vm06.stdout: ceph-grafana-dashboards noarch 2:19.2.3-678.ge911bdeb.el9 ceph-noarch 31 k 2026-03-09T20:49:35.786 INFO:teuthology.orchestra.run.vm06.stdout: ceph-mds x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 2.4 M 2026-03-09T20:49:35.786 INFO:teuthology.orchestra.run.vm06.stdout: ceph-mgr-modules-core noarch 2:19.2.3-678.ge911bdeb.el9 ceph-noarch 253 k 2026-03-09T20:49:35.787 INFO:teuthology.orchestra.run.vm06.stdout: ceph-mon x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 4.7 M 2026-03-09T20:49:35.787 INFO:teuthology.orchestra.run.vm06.stdout: ceph-osd x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 17 M 2026-03-09T20:49:35.787 INFO:teuthology.orchestra.run.vm06.stdout: ceph-prometheus-alerts noarch 2:19.2.3-678.ge911bdeb.el9 ceph-noarch 17 k 2026-03-09T20:49:35.787 INFO:teuthology.orchestra.run.vm06.stdout: ceph-selinux x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 25 k 2026-03-09T20:49:35.787 INFO:teuthology.orchestra.run.vm06.stdout: cryptsetup x86_64 2.8.1-3.el9 baseos 351 k 2026-03-09T20:49:35.787 INFO:teuthology.orchestra.run.vm06.stdout: flexiblas x86_64 3.0.4-9.el9 appstream 30 k 2026-03-09T20:49:35.787 INFO:teuthology.orchestra.run.vm06.stdout: flexiblas-netlib x86_64 3.0.4-9.el9 appstream 3.0 M 2026-03-09T20:49:35.787 INFO:teuthology.orchestra.run.vm06.stdout: flexiblas-openblas-openmp x86_64 3.0.4-9.el9 appstream 15 k 2026-03-09T20:49:35.787 INFO:teuthology.orchestra.run.vm06.stdout: gperftools-libs x86_64 2.9.1-3.el9 epel 308 k 2026-03-09T20:49:35.787 INFO:teuthology.orchestra.run.vm06.stdout: grpc-data noarch 1.46.7-10.el9 epel 19 k 2026-03-09T20:49:35.787 INFO:teuthology.orchestra.run.vm06.stdout: ledmon-libs x86_64 1.1.0-3.el9 baseos 40 k 2026-03-09T20:49:35.787 INFO:teuthology.orchestra.run.vm06.stdout: libarrow x86_64 9.0.0-15.el9 epel 4.4 M 2026-03-09T20:49:35.787 INFO:teuthology.orchestra.run.vm06.stdout: libarrow-doc noarch 9.0.0-15.el9 epel 25 k 2026-03-09T20:49:35.787 INFO:teuthology.orchestra.run.vm06.stdout: libcephsqlite x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 163 k 2026-03-09T20:49:35.787 INFO:teuthology.orchestra.run.vm06.stdout: libconfig x86_64 1.7.2-9.el9 baseos 72 k 2026-03-09T20:49:35.787 INFO:teuthology.orchestra.run.vm06.stdout: libgfortran x86_64 11.5.0-14.el9 baseos 794 k 2026-03-09T20:49:35.787 INFO:teuthology.orchestra.run.vm06.stdout: libnbd x86_64 1.20.3-4.el9 appstream 164 k 2026-03-09T20:49:35.787 INFO:teuthology.orchestra.run.vm06.stdout: liboath x86_64 2.6.12-1.el9 epel 49 k 2026-03-09T20:49:35.787 INFO:teuthology.orchestra.run.vm06.stdout: libpmemobj x86_64 1.12.1-1.el9 appstream 160 k 2026-03-09T20:49:35.787 INFO:teuthology.orchestra.run.vm06.stdout: libquadmath x86_64 11.5.0-14.el9 baseos 184 k 2026-03-09T20:49:35.787 INFO:teuthology.orchestra.run.vm06.stdout: librabbitmq x86_64 0.11.0-7.el9 appstream 45 k 2026-03-09T20:49:35.787 INFO:teuthology.orchestra.run.vm06.stdout: libradosstriper1 x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 503 k 2026-03-09T20:49:35.787 INFO:teuthology.orchestra.run.vm06.stdout: librdkafka x86_64 1.6.1-102.el9 appstream 662 k 2026-03-09T20:49:35.787 INFO:teuthology.orchestra.run.vm06.stdout: librgw2 x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 5.4 M 2026-03-09T20:49:35.787 INFO:teuthology.orchestra.run.vm06.stdout: libstoragemgmt x86_64 1.10.1-1.el9 appstream 246 k 2026-03-09T20:49:35.787 INFO:teuthology.orchestra.run.vm06.stdout: libunwind x86_64 1.6.2-1.el9 epel 67 k 2026-03-09T20:49:35.787 INFO:teuthology.orchestra.run.vm06.stdout: libxslt x86_64 1.1.34-12.el9 appstream 233 k 2026-03-09T20:49:35.787 INFO:teuthology.orchestra.run.vm06.stdout: lttng-ust x86_64 2.12.0-6.el9 appstream 292 k 2026-03-09T20:49:35.787 INFO:teuthology.orchestra.run.vm06.stdout: lua x86_64 5.4.4-4.el9 appstream 188 k 2026-03-09T20:49:35.787 INFO:teuthology.orchestra.run.vm06.stdout: lua-devel x86_64 5.4.4-4.el9 crb 22 k 2026-03-09T20:49:35.787 INFO:teuthology.orchestra.run.vm06.stdout: luarocks noarch 3.9.2-5.el9 epel 151 k 2026-03-09T20:49:35.787 INFO:teuthology.orchestra.run.vm06.stdout: mailcap noarch 2.1.49-5.el9 baseos 33 k 2026-03-09T20:49:35.787 INFO:teuthology.orchestra.run.vm06.stdout: openblas x86_64 0.3.29-1.el9 appstream 42 k 2026-03-09T20:49:35.787 INFO:teuthology.orchestra.run.vm06.stdout: openblas-openmp x86_64 0.3.29-1.el9 appstream 5.3 M 2026-03-09T20:49:35.787 INFO:teuthology.orchestra.run.vm06.stdout: parquet-libs x86_64 9.0.0-15.el9 epel 838 k 2026-03-09T20:49:35.787 INFO:teuthology.orchestra.run.vm06.stdout: pciutils x86_64 3.7.0-7.el9 baseos 93 k 2026-03-09T20:49:35.787 INFO:teuthology.orchestra.run.vm06.stdout: protobuf x86_64 3.14.0-17.el9 appstream 1.0 M 2026-03-09T20:49:35.787 INFO:teuthology.orchestra.run.vm06.stdout: protobuf-compiler x86_64 3.14.0-17.el9 crb 862 k 2026-03-09T20:49:35.787 INFO:teuthology.orchestra.run.vm06.stdout: python3-asyncssh noarch 2.13.2-5.el9 epel 548 k 2026-03-09T20:49:35.787 INFO:teuthology.orchestra.run.vm06.stdout: python3-autocommand noarch 2.2.2-8.el9 epel 29 k 2026-03-09T20:49:35.787 INFO:teuthology.orchestra.run.vm06.stdout: python3-babel noarch 2.9.1-2.el9 appstream 6.0 M 2026-03-09T20:49:35.787 INFO:teuthology.orchestra.run.vm06.stdout: python3-backports-tarfile noarch 1.2.0-1.el9 epel 60 k 2026-03-09T20:49:35.787 INFO:teuthology.orchestra.run.vm06.stdout: python3-bcrypt x86_64 3.2.2-1.el9 epel 43 k 2026-03-09T20:49:35.787 INFO:teuthology.orchestra.run.vm06.stdout: python3-cachetools noarch 4.2.4-1.el9 epel 32 k 2026-03-09T20:49:35.787 INFO:teuthology.orchestra.run.vm06.stdout: python3-ceph-argparse x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 45 k 2026-03-09T20:49:35.787 INFO:teuthology.orchestra.run.vm06.stdout: python3-ceph-common x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 142 k 2026-03-09T20:49:35.787 INFO:teuthology.orchestra.run.vm06.stdout: python3-certifi noarch 2023.05.07-4.el9 epel 14 k 2026-03-09T20:49:35.787 INFO:teuthology.orchestra.run.vm06.stdout: python3-cffi x86_64 1.14.5-5.el9 baseos 253 k 2026-03-09T20:49:35.787 INFO:teuthology.orchestra.run.vm06.stdout: python3-cheroot noarch 10.0.1-4.el9 epel 173 k 2026-03-09T20:49:35.787 INFO:teuthology.orchestra.run.vm06.stdout: python3-cherrypy noarch 18.6.1-2.el9 epel 358 k 2026-03-09T20:49:35.787 INFO:teuthology.orchestra.run.vm06.stdout: python3-cryptography x86_64 36.0.1-5.el9 baseos 1.2 M 2026-03-09T20:49:35.787 INFO:teuthology.orchestra.run.vm06.stdout: python3-devel x86_64 3.9.25-3.el9 appstream 244 k 2026-03-09T20:49:35.787 INFO:teuthology.orchestra.run.vm06.stdout: python3-google-auth noarch 1:2.45.0-1.el9 epel 254 k 2026-03-09T20:49:35.787 INFO:teuthology.orchestra.run.vm06.stdout: python3-grpcio x86_64 1.46.7-10.el9 epel 2.0 M 2026-03-09T20:49:35.787 INFO:teuthology.orchestra.run.vm06.stdout: python3-grpcio-tools x86_64 1.46.7-10.el9 epel 144 k 2026-03-09T20:49:35.787 INFO:teuthology.orchestra.run.vm06.stdout: python3-jaraco noarch 8.2.1-3.el9 epel 11 k 2026-03-09T20:49:35.787 INFO:teuthology.orchestra.run.vm06.stdout: python3-jaraco-classes noarch 3.2.1-5.el9 epel 18 k 2026-03-09T20:49:35.787 INFO:teuthology.orchestra.run.vm06.stdout: python3-jaraco-collections noarch 3.0.0-8.el9 epel 23 k 2026-03-09T20:49:35.787 INFO:teuthology.orchestra.run.vm06.stdout: python3-jaraco-context noarch 6.0.1-3.el9 epel 20 k 2026-03-09T20:49:35.787 INFO:teuthology.orchestra.run.vm06.stdout: python3-jaraco-functools noarch 3.5.0-2.el9 epel 19 k 2026-03-09T20:49:35.787 INFO:teuthology.orchestra.run.vm06.stdout: python3-jaraco-text noarch 4.0.0-2.el9 epel 26 k 2026-03-09T20:49:35.788 INFO:teuthology.orchestra.run.vm06.stdout: python3-jinja2 noarch 2.11.3-8.el9 appstream 249 k 2026-03-09T20:49:35.788 INFO:teuthology.orchestra.run.vm06.stdout: python3-kubernetes noarch 1:26.1.0-3.el9 epel 1.0 M 2026-03-09T20:49:35.788 INFO:teuthology.orchestra.run.vm06.stdout: python3-libstoragemgmt x86_64 1.10.1-1.el9 appstream 177 k 2026-03-09T20:49:35.788 INFO:teuthology.orchestra.run.vm06.stdout: python3-logutils noarch 0.3.5-21.el9 epel 46 k 2026-03-09T20:49:35.788 INFO:teuthology.orchestra.run.vm06.stdout: python3-mako noarch 1.1.4-6.el9 appstream 172 k 2026-03-09T20:49:35.788 INFO:teuthology.orchestra.run.vm06.stdout: python3-markupsafe x86_64 1.1.1-12.el9 appstream 35 k 2026-03-09T20:49:35.788 INFO:teuthology.orchestra.run.vm06.stdout: python3-more-itertools noarch 8.12.0-2.el9 epel 79 k 2026-03-09T20:49:35.788 INFO:teuthology.orchestra.run.vm06.stdout: python3-natsort noarch 7.1.1-5.el9 epel 58 k 2026-03-09T20:49:35.788 INFO:teuthology.orchestra.run.vm06.stdout: python3-numpy x86_64 1:1.23.5-2.el9 appstream 6.1 M 2026-03-09T20:49:35.788 INFO:teuthology.orchestra.run.vm06.stdout: python3-numpy-f2py x86_64 1:1.23.5-2.el9 appstream 442 k 2026-03-09T20:49:35.788 INFO:teuthology.orchestra.run.vm06.stdout: python3-packaging noarch 20.9-5.el9 appstream 77 k 2026-03-09T20:49:35.788 INFO:teuthology.orchestra.run.vm06.stdout: python3-pecan noarch 1.4.2-3.el9 epel 272 k 2026-03-09T20:49:35.788 INFO:teuthology.orchestra.run.vm06.stdout: python3-ply noarch 3.11-14.el9 baseos 106 k 2026-03-09T20:49:35.788 INFO:teuthology.orchestra.run.vm06.stdout: python3-portend noarch 3.1.0-2.el9 epel 16 k 2026-03-09T20:49:35.788 INFO:teuthology.orchestra.run.vm06.stdout: python3-protobuf noarch 3.14.0-17.el9 appstream 267 k 2026-03-09T20:49:35.788 INFO:teuthology.orchestra.run.vm06.stdout: python3-pyOpenSSL noarch 21.0.0-1.el9 epel 90 k 2026-03-09T20:49:35.788 INFO:teuthology.orchestra.run.vm06.stdout: python3-pyasn1 noarch 0.4.8-7.el9 appstream 157 k 2026-03-09T20:49:35.788 INFO:teuthology.orchestra.run.vm06.stdout: python3-pyasn1-modules noarch 0.4.8-7.el9 appstream 277 k 2026-03-09T20:49:35.788 INFO:teuthology.orchestra.run.vm06.stdout: python3-pycparser noarch 2.20-6.el9 baseos 135 k 2026-03-09T20:49:35.788 INFO:teuthology.orchestra.run.vm06.stdout: python3-repoze-lru noarch 0.7-16.el9 epel 31 k 2026-03-09T20:49:35.788 INFO:teuthology.orchestra.run.vm06.stdout: python3-requests noarch 2.25.1-10.el9 baseos 126 k 2026-03-09T20:49:35.788 INFO:teuthology.orchestra.run.vm06.stdout: python3-requests-oauthlib noarch 1.3.0-12.el9 appstream 54 k 2026-03-09T20:49:35.788 INFO:teuthology.orchestra.run.vm06.stdout: python3-routes noarch 2.5.1-5.el9 epel 188 k 2026-03-09T20:49:35.788 INFO:teuthology.orchestra.run.vm06.stdout: python3-rsa noarch 4.9-2.el9 epel 59 k 2026-03-09T20:49:35.788 INFO:teuthology.orchestra.run.vm06.stdout: python3-scipy x86_64 1.9.3-2.el9 appstream 19 M 2026-03-09T20:49:35.788 INFO:teuthology.orchestra.run.vm06.stdout: python3-tempora noarch 5.0.0-2.el9 epel 36 k 2026-03-09T20:49:35.788 INFO:teuthology.orchestra.run.vm06.stdout: python3-toml noarch 0.10.2-6.el9 appstream 42 k 2026-03-09T20:49:35.788 INFO:teuthology.orchestra.run.vm06.stdout: python3-typing-extensions noarch 4.15.0-1.el9 epel 86 k 2026-03-09T20:49:35.788 INFO:teuthology.orchestra.run.vm06.stdout: python3-urllib3 noarch 1.26.5-7.el9 baseos 218 k 2026-03-09T20:49:35.788 INFO:teuthology.orchestra.run.vm06.stdout: python3-webob noarch 1.8.8-2.el9 epel 230 k 2026-03-09T20:49:35.788 INFO:teuthology.orchestra.run.vm06.stdout: python3-websocket-client noarch 1.2.3-2.el9 epel 90 k 2026-03-09T20:49:35.788 INFO:teuthology.orchestra.run.vm06.stdout: python3-werkzeug noarch 2.0.3-3.el9.1 epel 427 k 2026-03-09T20:49:35.788 INFO:teuthology.orchestra.run.vm06.stdout: python3-zc-lockfile noarch 2.0-10.el9 epel 20 k 2026-03-09T20:49:35.788 INFO:teuthology.orchestra.run.vm06.stdout: qatlib x86_64 25.08.0-2.el9 appstream 240 k 2026-03-09T20:49:35.788 INFO:teuthology.orchestra.run.vm06.stdout: qatzip-libs x86_64 1.3.1-1.el9 appstream 66 k 2026-03-09T20:49:35.788 INFO:teuthology.orchestra.run.vm06.stdout: re2 x86_64 1:20211101-20.el9 epel 191 k 2026-03-09T20:49:35.788 INFO:teuthology.orchestra.run.vm06.stdout: socat x86_64 1.7.4.1-8.el9 appstream 303 k 2026-03-09T20:49:35.788 INFO:teuthology.orchestra.run.vm06.stdout: thrift x86_64 0.15.0-4.el9 epel 1.6 M 2026-03-09T20:49:35.788 INFO:teuthology.orchestra.run.vm06.stdout: unzip x86_64 6.0-59.el9 baseos 182 k 2026-03-09T20:49:35.788 INFO:teuthology.orchestra.run.vm06.stdout: xmlstarlet x86_64 1.6.1-20.el9 appstream 64 k 2026-03-09T20:49:35.788 INFO:teuthology.orchestra.run.vm06.stdout: zip x86_64 3.0-35.el9 baseos 266 k 2026-03-09T20:49:35.788 INFO:teuthology.orchestra.run.vm06.stdout:Installing weak dependencies: 2026-03-09T20:49:35.788 INFO:teuthology.orchestra.run.vm06.stdout: qatlib-service x86_64 25.08.0-2.el9 appstream 37 k 2026-03-09T20:49:35.788 INFO:teuthology.orchestra.run.vm06.stdout: 2026-03-09T20:49:35.788 INFO:teuthology.orchestra.run.vm06.stdout:Transaction Summary 2026-03-09T20:49:35.788 INFO:teuthology.orchestra.run.vm06.stdout:====================================================================================== 2026-03-09T20:49:35.788 INFO:teuthology.orchestra.run.vm06.stdout:Install 134 Packages 2026-03-09T20:49:35.788 INFO:teuthology.orchestra.run.vm06.stdout:Upgrade 2 Packages 2026-03-09T20:49:35.788 INFO:teuthology.orchestra.run.vm06.stdout: 2026-03-09T20:49:35.788 INFO:teuthology.orchestra.run.vm06.stdout:Total download size: 210 M 2026-03-09T20:49:35.788 INFO:teuthology.orchestra.run.vm06.stdout:Downloading Packages: 2026-03-09T20:49:36.936 INFO:teuthology.orchestra.run.vm08.stdout:Package librados2-2:16.2.4-5.el9.x86_64 is already installed. 2026-03-09T20:49:36.936 INFO:teuthology.orchestra.run.vm08.stdout:Package librbd1-2:16.2.4-5.el9.x86_64 is already installed. 2026-03-09T20:49:36.941 INFO:teuthology.orchestra.run.vm08.stdout:Package bzip2-1.0.8-11.el9.x86_64 is already installed. 2026-03-09T20:49:36.941 INFO:teuthology.orchestra.run.vm08.stdout:Package perl-Test-Harness-1:3.42-461.el9.noarch is already installed. 2026-03-09T20:49:36.979 INFO:teuthology.orchestra.run.vm08.stdout:Dependencies resolved. 2026-03-09T20:49:36.984 INFO:teuthology.orchestra.run.vm08.stdout:====================================================================================== 2026-03-09T20:49:36.984 INFO:teuthology.orchestra.run.vm08.stdout: Package Arch Version Repository Size 2026-03-09T20:49:36.984 INFO:teuthology.orchestra.run.vm08.stdout:====================================================================================== 2026-03-09T20:49:36.984 INFO:teuthology.orchestra.run.vm08.stdout:Installing: 2026-03-09T20:49:36.984 INFO:teuthology.orchestra.run.vm08.stdout: ceph x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 6.5 k 2026-03-09T20:49:36.984 INFO:teuthology.orchestra.run.vm08.stdout: ceph-base x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 5.5 M 2026-03-09T20:49:36.984 INFO:teuthology.orchestra.run.vm08.stdout: ceph-fuse x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 1.2 M 2026-03-09T20:49:36.984 INFO:teuthology.orchestra.run.vm08.stdout: ceph-immutable-object-cache x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 145 k 2026-03-09T20:49:36.984 INFO:teuthology.orchestra.run.vm08.stdout: ceph-mgr x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 1.1 M 2026-03-09T20:49:36.984 INFO:teuthology.orchestra.run.vm08.stdout: ceph-mgr-cephadm noarch 2:19.2.3-678.ge911bdeb.el9 ceph-noarch 150 k 2026-03-09T20:49:36.984 INFO:teuthology.orchestra.run.vm08.stdout: ceph-mgr-dashboard noarch 2:19.2.3-678.ge911bdeb.el9 ceph-noarch 3.8 M 2026-03-09T20:49:36.984 INFO:teuthology.orchestra.run.vm08.stdout: ceph-mgr-diskprediction-local noarch 2:19.2.3-678.ge911bdeb.el9 ceph-noarch 7.4 M 2026-03-09T20:49:36.984 INFO:teuthology.orchestra.run.vm08.stdout: ceph-mgr-rook noarch 2:19.2.3-678.ge911bdeb.el9 ceph-noarch 49 k 2026-03-09T20:49:36.984 INFO:teuthology.orchestra.run.vm08.stdout: ceph-radosgw x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 11 M 2026-03-09T20:49:36.984 INFO:teuthology.orchestra.run.vm08.stdout: ceph-test x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 50 M 2026-03-09T20:49:36.984 INFO:teuthology.orchestra.run.vm08.stdout: ceph-volume noarch 2:19.2.3-678.ge911bdeb.el9 ceph-noarch 299 k 2026-03-09T20:49:36.984 INFO:teuthology.orchestra.run.vm08.stdout: cephadm noarch 2:19.2.3-678.ge911bdeb.el9 ceph-noarch 769 k 2026-03-09T20:49:36.984 INFO:teuthology.orchestra.run.vm08.stdout: libcephfs-devel x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 34 k 2026-03-09T20:49:36.984 INFO:teuthology.orchestra.run.vm08.stdout: libcephfs2 x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 1.0 M 2026-03-09T20:49:36.984 INFO:teuthology.orchestra.run.vm08.stdout: librados-devel x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 127 k 2026-03-09T20:49:36.984 INFO:teuthology.orchestra.run.vm08.stdout: python3-cephfs x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 165 k 2026-03-09T20:49:36.984 INFO:teuthology.orchestra.run.vm08.stdout: python3-jmespath noarch 1.0.1-1.el9 appstream 48 k 2026-03-09T20:49:36.985 INFO:teuthology.orchestra.run.vm08.stdout: python3-rados x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 323 k 2026-03-09T20:49:36.985 INFO:teuthology.orchestra.run.vm08.stdout: python3-rbd x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 303 k 2026-03-09T20:49:36.985 INFO:teuthology.orchestra.run.vm08.stdout: python3-rgw x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 100 k 2026-03-09T20:49:36.985 INFO:teuthology.orchestra.run.vm08.stdout: python3-xmltodict noarch 0.12.0-15.el9 epel 22 k 2026-03-09T20:49:36.985 INFO:teuthology.orchestra.run.vm08.stdout: rbd-fuse x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 85 k 2026-03-09T20:49:36.985 INFO:teuthology.orchestra.run.vm08.stdout: rbd-mirror x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 3.1 M 2026-03-09T20:49:36.985 INFO:teuthology.orchestra.run.vm08.stdout: rbd-nbd x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 171 k 2026-03-09T20:49:36.985 INFO:teuthology.orchestra.run.vm08.stdout:Upgrading: 2026-03-09T20:49:36.985 INFO:teuthology.orchestra.run.vm08.stdout: librados2 x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 3.4 M 2026-03-09T20:49:36.985 INFO:teuthology.orchestra.run.vm08.stdout: librbd1 x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 3.2 M 2026-03-09T20:49:36.985 INFO:teuthology.orchestra.run.vm08.stdout:Installing dependencies: 2026-03-09T20:49:36.985 INFO:teuthology.orchestra.run.vm08.stdout: abseil-cpp x86_64 20211102.0-4.el9 epel 551 k 2026-03-09T20:49:36.985 INFO:teuthology.orchestra.run.vm08.stdout: boost-program-options x86_64 1.75.0-13.el9 appstream 104 k 2026-03-09T20:49:36.985 INFO:teuthology.orchestra.run.vm08.stdout: ceph-common x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 22 M 2026-03-09T20:49:36.985 INFO:teuthology.orchestra.run.vm08.stdout: ceph-grafana-dashboards noarch 2:19.2.3-678.ge911bdeb.el9 ceph-noarch 31 k 2026-03-09T20:49:36.985 INFO:teuthology.orchestra.run.vm08.stdout: ceph-mds x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 2.4 M 2026-03-09T20:49:36.985 INFO:teuthology.orchestra.run.vm08.stdout: ceph-mgr-modules-core noarch 2:19.2.3-678.ge911bdeb.el9 ceph-noarch 253 k 2026-03-09T20:49:36.985 INFO:teuthology.orchestra.run.vm08.stdout: ceph-mon x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 4.7 M 2026-03-09T20:49:36.985 INFO:teuthology.orchestra.run.vm08.stdout: ceph-osd x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 17 M 2026-03-09T20:49:36.985 INFO:teuthology.orchestra.run.vm08.stdout: ceph-prometheus-alerts noarch 2:19.2.3-678.ge911bdeb.el9 ceph-noarch 17 k 2026-03-09T20:49:36.985 INFO:teuthology.orchestra.run.vm08.stdout: ceph-selinux x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 25 k 2026-03-09T20:49:36.985 INFO:teuthology.orchestra.run.vm08.stdout: cryptsetup x86_64 2.8.1-3.el9 baseos 351 k 2026-03-09T20:49:36.985 INFO:teuthology.orchestra.run.vm08.stdout: flexiblas x86_64 3.0.4-9.el9 appstream 30 k 2026-03-09T20:49:36.985 INFO:teuthology.orchestra.run.vm08.stdout: flexiblas-netlib x86_64 3.0.4-9.el9 appstream 3.0 M 2026-03-09T20:49:36.985 INFO:teuthology.orchestra.run.vm08.stdout: flexiblas-openblas-openmp x86_64 3.0.4-9.el9 appstream 15 k 2026-03-09T20:49:36.985 INFO:teuthology.orchestra.run.vm08.stdout: gperftools-libs x86_64 2.9.1-3.el9 epel 308 k 2026-03-09T20:49:36.985 INFO:teuthology.orchestra.run.vm08.stdout: grpc-data noarch 1.46.7-10.el9 epel 19 k 2026-03-09T20:49:36.985 INFO:teuthology.orchestra.run.vm08.stdout: ledmon-libs x86_64 1.1.0-3.el9 baseos 40 k 2026-03-09T20:49:36.985 INFO:teuthology.orchestra.run.vm08.stdout: libarrow x86_64 9.0.0-15.el9 epel 4.4 M 2026-03-09T20:49:36.985 INFO:teuthology.orchestra.run.vm08.stdout: libarrow-doc noarch 9.0.0-15.el9 epel 25 k 2026-03-09T20:49:36.985 INFO:teuthology.orchestra.run.vm08.stdout: libcephsqlite x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 163 k 2026-03-09T20:49:36.985 INFO:teuthology.orchestra.run.vm08.stdout: libconfig x86_64 1.7.2-9.el9 baseos 72 k 2026-03-09T20:49:36.985 INFO:teuthology.orchestra.run.vm08.stdout: libgfortran x86_64 11.5.0-14.el9 baseos 794 k 2026-03-09T20:49:36.985 INFO:teuthology.orchestra.run.vm08.stdout: libnbd x86_64 1.20.3-4.el9 appstream 164 k 2026-03-09T20:49:36.985 INFO:teuthology.orchestra.run.vm08.stdout: liboath x86_64 2.6.12-1.el9 epel 49 k 2026-03-09T20:49:36.985 INFO:teuthology.orchestra.run.vm08.stdout: libpmemobj x86_64 1.12.1-1.el9 appstream 160 k 2026-03-09T20:49:36.985 INFO:teuthology.orchestra.run.vm08.stdout: libquadmath x86_64 11.5.0-14.el9 baseos 184 k 2026-03-09T20:49:36.985 INFO:teuthology.orchestra.run.vm08.stdout: librabbitmq x86_64 0.11.0-7.el9 appstream 45 k 2026-03-09T20:49:36.985 INFO:teuthology.orchestra.run.vm08.stdout: libradosstriper1 x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 503 k 2026-03-09T20:49:36.985 INFO:teuthology.orchestra.run.vm08.stdout: librdkafka x86_64 1.6.1-102.el9 appstream 662 k 2026-03-09T20:49:36.985 INFO:teuthology.orchestra.run.vm08.stdout: librgw2 x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 5.4 M 2026-03-09T20:49:36.986 INFO:teuthology.orchestra.run.vm08.stdout: libstoragemgmt x86_64 1.10.1-1.el9 appstream 246 k 2026-03-09T20:49:36.986 INFO:teuthology.orchestra.run.vm08.stdout: libunwind x86_64 1.6.2-1.el9 epel 67 k 2026-03-09T20:49:36.986 INFO:teuthology.orchestra.run.vm08.stdout: libxslt x86_64 1.1.34-12.el9 appstream 233 k 2026-03-09T20:49:36.986 INFO:teuthology.orchestra.run.vm08.stdout: lttng-ust x86_64 2.12.0-6.el9 appstream 292 k 2026-03-09T20:49:36.986 INFO:teuthology.orchestra.run.vm08.stdout: lua x86_64 5.4.4-4.el9 appstream 188 k 2026-03-09T20:49:36.986 INFO:teuthology.orchestra.run.vm08.stdout: lua-devel x86_64 5.4.4-4.el9 crb 22 k 2026-03-09T20:49:36.986 INFO:teuthology.orchestra.run.vm08.stdout: luarocks noarch 3.9.2-5.el9 epel 151 k 2026-03-09T20:49:36.986 INFO:teuthology.orchestra.run.vm08.stdout: mailcap noarch 2.1.49-5.el9 baseos 33 k 2026-03-09T20:49:36.986 INFO:teuthology.orchestra.run.vm08.stdout: openblas x86_64 0.3.29-1.el9 appstream 42 k 2026-03-09T20:49:36.986 INFO:teuthology.orchestra.run.vm08.stdout: openblas-openmp x86_64 0.3.29-1.el9 appstream 5.3 M 2026-03-09T20:49:36.986 INFO:teuthology.orchestra.run.vm08.stdout: parquet-libs x86_64 9.0.0-15.el9 epel 838 k 2026-03-09T20:49:36.986 INFO:teuthology.orchestra.run.vm08.stdout: pciutils x86_64 3.7.0-7.el9 baseos 93 k 2026-03-09T20:49:36.986 INFO:teuthology.orchestra.run.vm08.stdout: protobuf x86_64 3.14.0-17.el9 appstream 1.0 M 2026-03-09T20:49:36.986 INFO:teuthology.orchestra.run.vm08.stdout: protobuf-compiler x86_64 3.14.0-17.el9 crb 862 k 2026-03-09T20:49:36.986 INFO:teuthology.orchestra.run.vm08.stdout: python3-asyncssh noarch 2.13.2-5.el9 epel 548 k 2026-03-09T20:49:36.986 INFO:teuthology.orchestra.run.vm08.stdout: python3-autocommand noarch 2.2.2-8.el9 epel 29 k 2026-03-09T20:49:36.986 INFO:teuthology.orchestra.run.vm08.stdout: python3-babel noarch 2.9.1-2.el9 appstream 6.0 M 2026-03-09T20:49:36.986 INFO:teuthology.orchestra.run.vm08.stdout: python3-backports-tarfile noarch 1.2.0-1.el9 epel 60 k 2026-03-09T20:49:36.986 INFO:teuthology.orchestra.run.vm08.stdout: python3-bcrypt x86_64 3.2.2-1.el9 epel 43 k 2026-03-09T20:49:36.986 INFO:teuthology.orchestra.run.vm08.stdout: python3-cachetools noarch 4.2.4-1.el9 epel 32 k 2026-03-09T20:49:36.986 INFO:teuthology.orchestra.run.vm08.stdout: python3-ceph-argparse x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 45 k 2026-03-09T20:49:36.986 INFO:teuthology.orchestra.run.vm08.stdout: python3-ceph-common x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 142 k 2026-03-09T20:49:36.986 INFO:teuthology.orchestra.run.vm08.stdout: python3-certifi noarch 2023.05.07-4.el9 epel 14 k 2026-03-09T20:49:36.986 INFO:teuthology.orchestra.run.vm08.stdout: python3-cffi x86_64 1.14.5-5.el9 baseos 253 k 2026-03-09T20:49:36.986 INFO:teuthology.orchestra.run.vm08.stdout: python3-cheroot noarch 10.0.1-4.el9 epel 173 k 2026-03-09T20:49:36.986 INFO:teuthology.orchestra.run.vm08.stdout: python3-cherrypy noarch 18.6.1-2.el9 epel 358 k 2026-03-09T20:49:36.986 INFO:teuthology.orchestra.run.vm08.stdout: python3-cryptography x86_64 36.0.1-5.el9 baseos 1.2 M 2026-03-09T20:49:36.986 INFO:teuthology.orchestra.run.vm08.stdout: python3-devel x86_64 3.9.25-3.el9 appstream 244 k 2026-03-09T20:49:36.986 INFO:teuthology.orchestra.run.vm08.stdout: python3-google-auth noarch 1:2.45.0-1.el9 epel 254 k 2026-03-09T20:49:36.986 INFO:teuthology.orchestra.run.vm08.stdout: python3-grpcio x86_64 1.46.7-10.el9 epel 2.0 M 2026-03-09T20:49:36.986 INFO:teuthology.orchestra.run.vm08.stdout: python3-grpcio-tools x86_64 1.46.7-10.el9 epel 144 k 2026-03-09T20:49:36.986 INFO:teuthology.orchestra.run.vm08.stdout: python3-jaraco noarch 8.2.1-3.el9 epel 11 k 2026-03-09T20:49:36.986 INFO:teuthology.orchestra.run.vm08.stdout: python3-jaraco-classes noarch 3.2.1-5.el9 epel 18 k 2026-03-09T20:49:36.986 INFO:teuthology.orchestra.run.vm08.stdout: python3-jaraco-collections noarch 3.0.0-8.el9 epel 23 k 2026-03-09T20:49:36.986 INFO:teuthology.orchestra.run.vm08.stdout: python3-jaraco-context noarch 6.0.1-3.el9 epel 20 k 2026-03-09T20:49:36.986 INFO:teuthology.orchestra.run.vm08.stdout: python3-jaraco-functools noarch 3.5.0-2.el9 epel 19 k 2026-03-09T20:49:36.986 INFO:teuthology.orchestra.run.vm08.stdout: python3-jaraco-text noarch 4.0.0-2.el9 epel 26 k 2026-03-09T20:49:36.986 INFO:teuthology.orchestra.run.vm08.stdout: python3-jinja2 noarch 2.11.3-8.el9 appstream 249 k 2026-03-09T20:49:36.986 INFO:teuthology.orchestra.run.vm08.stdout: python3-kubernetes noarch 1:26.1.0-3.el9 epel 1.0 M 2026-03-09T20:49:36.986 INFO:teuthology.orchestra.run.vm08.stdout: python3-libstoragemgmt x86_64 1.10.1-1.el9 appstream 177 k 2026-03-09T20:49:36.986 INFO:teuthology.orchestra.run.vm08.stdout: python3-logutils noarch 0.3.5-21.el9 epel 46 k 2026-03-09T20:49:36.986 INFO:teuthology.orchestra.run.vm08.stdout: python3-mako noarch 1.1.4-6.el9 appstream 172 k 2026-03-09T20:49:36.986 INFO:teuthology.orchestra.run.vm08.stdout: python3-markupsafe x86_64 1.1.1-12.el9 appstream 35 k 2026-03-09T20:49:36.987 INFO:teuthology.orchestra.run.vm08.stdout: python3-more-itertools noarch 8.12.0-2.el9 epel 79 k 2026-03-09T20:49:36.987 INFO:teuthology.orchestra.run.vm08.stdout: python3-natsort noarch 7.1.1-5.el9 epel 58 k 2026-03-09T20:49:36.987 INFO:teuthology.orchestra.run.vm08.stdout: python3-numpy x86_64 1:1.23.5-2.el9 appstream 6.1 M 2026-03-09T20:49:36.987 INFO:teuthology.orchestra.run.vm08.stdout: python3-numpy-f2py x86_64 1:1.23.5-2.el9 appstream 442 k 2026-03-09T20:49:36.987 INFO:teuthology.orchestra.run.vm08.stdout: python3-packaging noarch 20.9-5.el9 appstream 77 k 2026-03-09T20:49:36.987 INFO:teuthology.orchestra.run.vm08.stdout: python3-pecan noarch 1.4.2-3.el9 epel 272 k 2026-03-09T20:49:36.987 INFO:teuthology.orchestra.run.vm08.stdout: python3-ply noarch 3.11-14.el9 baseos 106 k 2026-03-09T20:49:36.987 INFO:teuthology.orchestra.run.vm08.stdout: python3-portend noarch 3.1.0-2.el9 epel 16 k 2026-03-09T20:49:36.987 INFO:teuthology.orchestra.run.vm08.stdout: python3-protobuf noarch 3.14.0-17.el9 appstream 267 k 2026-03-09T20:49:36.987 INFO:teuthology.orchestra.run.vm08.stdout: python3-pyOpenSSL noarch 21.0.0-1.el9 epel 90 k 2026-03-09T20:49:36.987 INFO:teuthology.orchestra.run.vm08.stdout: python3-pyasn1 noarch 0.4.8-7.el9 appstream 157 k 2026-03-09T20:49:36.987 INFO:teuthology.orchestra.run.vm08.stdout: python3-pyasn1-modules noarch 0.4.8-7.el9 appstream 277 k 2026-03-09T20:49:36.987 INFO:teuthology.orchestra.run.vm08.stdout: python3-pycparser noarch 2.20-6.el9 baseos 135 k 2026-03-09T20:49:36.987 INFO:teuthology.orchestra.run.vm08.stdout: python3-repoze-lru noarch 0.7-16.el9 epel 31 k 2026-03-09T20:49:36.987 INFO:teuthology.orchestra.run.vm08.stdout: python3-requests noarch 2.25.1-10.el9 baseos 126 k 2026-03-09T20:49:36.987 INFO:teuthology.orchestra.run.vm08.stdout: python3-requests-oauthlib noarch 1.3.0-12.el9 appstream 54 k 2026-03-09T20:49:36.987 INFO:teuthology.orchestra.run.vm08.stdout: python3-routes noarch 2.5.1-5.el9 epel 188 k 2026-03-09T20:49:36.987 INFO:teuthology.orchestra.run.vm08.stdout: python3-rsa noarch 4.9-2.el9 epel 59 k 2026-03-09T20:49:36.987 INFO:teuthology.orchestra.run.vm08.stdout: python3-scipy x86_64 1.9.3-2.el9 appstream 19 M 2026-03-09T20:49:36.987 INFO:teuthology.orchestra.run.vm08.stdout: python3-tempora noarch 5.0.0-2.el9 epel 36 k 2026-03-09T20:49:36.987 INFO:teuthology.orchestra.run.vm08.stdout: python3-toml noarch 0.10.2-6.el9 appstream 42 k 2026-03-09T20:49:36.987 INFO:teuthology.orchestra.run.vm08.stdout: python3-typing-extensions noarch 4.15.0-1.el9 epel 86 k 2026-03-09T20:49:36.987 INFO:teuthology.orchestra.run.vm08.stdout: python3-urllib3 noarch 1.26.5-7.el9 baseos 218 k 2026-03-09T20:49:36.987 INFO:teuthology.orchestra.run.vm08.stdout: python3-webob noarch 1.8.8-2.el9 epel 230 k 2026-03-09T20:49:36.987 INFO:teuthology.orchestra.run.vm08.stdout: python3-websocket-client noarch 1.2.3-2.el9 epel 90 k 2026-03-09T20:49:36.987 INFO:teuthology.orchestra.run.vm08.stdout: python3-werkzeug noarch 2.0.3-3.el9.1 epel 427 k 2026-03-09T20:49:36.987 INFO:teuthology.orchestra.run.vm08.stdout: python3-zc-lockfile noarch 2.0-10.el9 epel 20 k 2026-03-09T20:49:36.987 INFO:teuthology.orchestra.run.vm08.stdout: qatlib x86_64 25.08.0-2.el9 appstream 240 k 2026-03-09T20:49:36.987 INFO:teuthology.orchestra.run.vm08.stdout: qatzip-libs x86_64 1.3.1-1.el9 appstream 66 k 2026-03-09T20:49:36.987 INFO:teuthology.orchestra.run.vm08.stdout: re2 x86_64 1:20211101-20.el9 epel 191 k 2026-03-09T20:49:36.987 INFO:teuthology.orchestra.run.vm08.stdout: socat x86_64 1.7.4.1-8.el9 appstream 303 k 2026-03-09T20:49:36.987 INFO:teuthology.orchestra.run.vm08.stdout: thrift x86_64 0.15.0-4.el9 epel 1.6 M 2026-03-09T20:49:36.987 INFO:teuthology.orchestra.run.vm08.stdout: unzip x86_64 6.0-59.el9 baseos 182 k 2026-03-09T20:49:36.987 INFO:teuthology.orchestra.run.vm08.stdout: xmlstarlet x86_64 1.6.1-20.el9 appstream 64 k 2026-03-09T20:49:36.987 INFO:teuthology.orchestra.run.vm08.stdout: zip x86_64 3.0-35.el9 baseos 266 k 2026-03-09T20:49:36.987 INFO:teuthology.orchestra.run.vm08.stdout:Installing weak dependencies: 2026-03-09T20:49:36.987 INFO:teuthology.orchestra.run.vm08.stdout: qatlib-service x86_64 25.08.0-2.el9 appstream 37 k 2026-03-09T20:49:36.987 INFO:teuthology.orchestra.run.vm08.stdout: 2026-03-09T20:49:36.987 INFO:teuthology.orchestra.run.vm08.stdout:Transaction Summary 2026-03-09T20:49:36.987 INFO:teuthology.orchestra.run.vm08.stdout:====================================================================================== 2026-03-09T20:49:36.987 INFO:teuthology.orchestra.run.vm08.stdout:Install 134 Packages 2026-03-09T20:49:36.988 INFO:teuthology.orchestra.run.vm08.stdout:Upgrade 2 Packages 2026-03-09T20:49:36.988 INFO:teuthology.orchestra.run.vm08.stdout: 2026-03-09T20:49:36.988 INFO:teuthology.orchestra.run.vm08.stdout:Total download size: 210 M 2026-03-09T20:49:36.988 INFO:teuthology.orchestra.run.vm08.stdout:Downloading Packages: 2026-03-09T20:49:37.908 INFO:teuthology.orchestra.run.vm06.stdout:(1/136): ceph-19.2.3-678.ge911bdeb.el9.x86_64.r 12 kB/s | 6.5 kB 00:00 2026-03-09T20:49:38.383 INFO:teuthology.orchestra.run.vm08.stdout:(1/136): ceph-19.2.3-678.ge911bdeb.el9.x86_64.r 14 kB/s | 6.5 kB 00:00 2026-03-09T20:49:38.639 INFO:teuthology.orchestra.run.vm01.stdout:lab-extras 65 kB/s | 50 kB 00:00 2026-03-09T20:49:38.863 INFO:teuthology.orchestra.run.vm06.stdout:(2/136): ceph-fuse-19.2.3-678.ge911bdeb.el9.x86 1.2 MB/s | 1.2 MB 00:00 2026-03-09T20:49:38.990 INFO:teuthology.orchestra.run.vm06.stdout:(3/136): ceph-base-19.2.3-678.ge911bdeb.el9.x86 3.4 MB/s | 5.5 MB 00:01 2026-03-09T20:49:38.992 INFO:teuthology.orchestra.run.vm06.stdout:(4/136): ceph-immutable-object-cache-19.2.3-678 1.1 MB/s | 145 kB 00:00 2026-03-09T20:49:39.220 INFO:teuthology.orchestra.run.vm08.stdout:(2/136): ceph-fuse-19.2.3-678.ge911bdeb.el9.x86 1.4 MB/s | 1.2 MB 00:00 2026-03-09T20:49:39.260 INFO:teuthology.orchestra.run.vm06.stdout:(5/136): ceph-mds-19.2.3-678.ge911bdeb.el9.x86_ 9.0 MB/s | 2.4 MB 00:00 2026-03-09T20:49:39.344 INFO:teuthology.orchestra.run.vm08.stdout:(3/136): ceph-immutable-object-cache-19.2.3-678 1.2 MB/s | 145 kB 00:00 2026-03-09T20:49:39.392 INFO:teuthology.orchestra.run.vm06.stdout:(6/136): ceph-mgr-19.2.3-678.ge911bdeb.el9.x86_ 2.7 MB/s | 1.1 MB 00:00 2026-03-09T20:49:39.576 INFO:teuthology.orchestra.run.vm08.stdout:(4/136): ceph-base-19.2.3-678.ge911bdeb.el9.x86 3.3 MB/s | 5.5 MB 00:01 2026-03-09T20:49:39.657 INFO:teuthology.orchestra.run.vm06.stdout:(7/136): ceph-mon-19.2.3-678.ge911bdeb.el9.x86_ 12 MB/s | 4.7 MB 00:00 2026-03-09T20:49:39.719 INFO:teuthology.orchestra.run.vm08.stdout:(5/136): ceph-mgr-19.2.3-678.ge911bdeb.el9.x86_ 7.5 MB/s | 1.1 MB 00:00 2026-03-09T20:49:39.749 INFO:teuthology.orchestra.run.vm08.stdout:(6/136): ceph-mds-19.2.3-678.ge911bdeb.el9.x86_ 6.0 MB/s | 2.4 MB 00:00 2026-03-09T20:49:40.053 INFO:teuthology.orchestra.run.vm06.stdout:(8/136): ceph-common-19.2.3-678.ge911bdeb.el9.x 8.1 MB/s | 22 MB 00:02 2026-03-09T20:49:40.083 INFO:teuthology.orchestra.run.vm01.stdout:Package librados2-2:16.2.4-5.el9.x86_64 is already installed. 2026-03-09T20:49:40.083 INFO:teuthology.orchestra.run.vm01.stdout:Package librbd1-2:16.2.4-5.el9.x86_64 is already installed. 2026-03-09T20:49:40.087 INFO:teuthology.orchestra.run.vm01.stdout:Package bzip2-1.0.8-11.el9.x86_64 is already installed. 2026-03-09T20:49:40.088 INFO:teuthology.orchestra.run.vm01.stdout:Package perl-Test-Harness-1:3.42-461.el9.noarch is already installed. 2026-03-09T20:49:40.116 INFO:teuthology.orchestra.run.vm01.stdout:Dependencies resolved. 2026-03-09T20:49:40.120 INFO:teuthology.orchestra.run.vm01.stdout:====================================================================================== 2026-03-09T20:49:40.120 INFO:teuthology.orchestra.run.vm01.stdout: Package Arch Version Repository Size 2026-03-09T20:49:40.120 INFO:teuthology.orchestra.run.vm01.stdout:====================================================================================== 2026-03-09T20:49:40.121 INFO:teuthology.orchestra.run.vm01.stdout:Installing: 2026-03-09T20:49:40.121 INFO:teuthology.orchestra.run.vm01.stdout: ceph x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 6.5 k 2026-03-09T20:49:40.121 INFO:teuthology.orchestra.run.vm01.stdout: ceph-base x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 5.5 M 2026-03-09T20:49:40.121 INFO:teuthology.orchestra.run.vm01.stdout: ceph-fuse x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 1.2 M 2026-03-09T20:49:40.121 INFO:teuthology.orchestra.run.vm01.stdout: ceph-immutable-object-cache x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 145 k 2026-03-09T20:49:40.121 INFO:teuthology.orchestra.run.vm01.stdout: ceph-mgr x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 1.1 M 2026-03-09T20:49:40.121 INFO:teuthology.orchestra.run.vm01.stdout: ceph-mgr-cephadm noarch 2:19.2.3-678.ge911bdeb.el9 ceph-noarch 150 k 2026-03-09T20:49:40.121 INFO:teuthology.orchestra.run.vm01.stdout: ceph-mgr-dashboard noarch 2:19.2.3-678.ge911bdeb.el9 ceph-noarch 3.8 M 2026-03-09T20:49:40.121 INFO:teuthology.orchestra.run.vm01.stdout: ceph-mgr-diskprediction-local noarch 2:19.2.3-678.ge911bdeb.el9 ceph-noarch 7.4 M 2026-03-09T20:49:40.121 INFO:teuthology.orchestra.run.vm01.stdout: ceph-mgr-rook noarch 2:19.2.3-678.ge911bdeb.el9 ceph-noarch 49 k 2026-03-09T20:49:40.121 INFO:teuthology.orchestra.run.vm01.stdout: ceph-radosgw x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 11 M 2026-03-09T20:49:40.121 INFO:teuthology.orchestra.run.vm01.stdout: ceph-test x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 50 M 2026-03-09T20:49:40.121 INFO:teuthology.orchestra.run.vm01.stdout: ceph-volume noarch 2:19.2.3-678.ge911bdeb.el9 ceph-noarch 299 k 2026-03-09T20:49:40.121 INFO:teuthology.orchestra.run.vm01.stdout: cephadm noarch 2:19.2.3-678.ge911bdeb.el9 ceph-noarch 769 k 2026-03-09T20:49:40.121 INFO:teuthology.orchestra.run.vm01.stdout: libcephfs-devel x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 34 k 2026-03-09T20:49:40.121 INFO:teuthology.orchestra.run.vm01.stdout: libcephfs2 x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 1.0 M 2026-03-09T20:49:40.121 INFO:teuthology.orchestra.run.vm01.stdout: librados-devel x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 127 k 2026-03-09T20:49:40.121 INFO:teuthology.orchestra.run.vm01.stdout: python3-cephfs x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 165 k 2026-03-09T20:49:40.121 INFO:teuthology.orchestra.run.vm01.stdout: python3-jmespath noarch 1.0.1-1.el9 appstream 48 k 2026-03-09T20:49:40.121 INFO:teuthology.orchestra.run.vm01.stdout: python3-rados x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 323 k 2026-03-09T20:49:40.121 INFO:teuthology.orchestra.run.vm01.stdout: python3-rbd x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 303 k 2026-03-09T20:49:40.121 INFO:teuthology.orchestra.run.vm01.stdout: python3-rgw x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 100 k 2026-03-09T20:49:40.121 INFO:teuthology.orchestra.run.vm01.stdout: python3-xmltodict noarch 0.12.0-15.el9 epel 22 k 2026-03-09T20:49:40.121 INFO:teuthology.orchestra.run.vm01.stdout: rbd-fuse x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 85 k 2026-03-09T20:49:40.121 INFO:teuthology.orchestra.run.vm01.stdout: rbd-mirror x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 3.1 M 2026-03-09T20:49:40.121 INFO:teuthology.orchestra.run.vm01.stdout: rbd-nbd x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 171 k 2026-03-09T20:49:40.121 INFO:teuthology.orchestra.run.vm01.stdout:Upgrading: 2026-03-09T20:49:40.121 INFO:teuthology.orchestra.run.vm01.stdout: librados2 x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 3.4 M 2026-03-09T20:49:40.121 INFO:teuthology.orchestra.run.vm01.stdout: librbd1 x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 3.2 M 2026-03-09T20:49:40.121 INFO:teuthology.orchestra.run.vm01.stdout:Installing dependencies: 2026-03-09T20:49:40.121 INFO:teuthology.orchestra.run.vm01.stdout: abseil-cpp x86_64 20211102.0-4.el9 epel 551 k 2026-03-09T20:49:40.121 INFO:teuthology.orchestra.run.vm01.stdout: boost-program-options x86_64 1.75.0-13.el9 appstream 104 k 2026-03-09T20:49:40.121 INFO:teuthology.orchestra.run.vm01.stdout: ceph-common x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 22 M 2026-03-09T20:49:40.121 INFO:teuthology.orchestra.run.vm01.stdout: ceph-grafana-dashboards noarch 2:19.2.3-678.ge911bdeb.el9 ceph-noarch 31 k 2026-03-09T20:49:40.121 INFO:teuthology.orchestra.run.vm01.stdout: ceph-mds x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 2.4 M 2026-03-09T20:49:40.121 INFO:teuthology.orchestra.run.vm01.stdout: ceph-mgr-modules-core noarch 2:19.2.3-678.ge911bdeb.el9 ceph-noarch 253 k 2026-03-09T20:49:40.121 INFO:teuthology.orchestra.run.vm01.stdout: ceph-mon x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 4.7 M 2026-03-09T20:49:40.121 INFO:teuthology.orchestra.run.vm01.stdout: ceph-osd x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 17 M 2026-03-09T20:49:40.121 INFO:teuthology.orchestra.run.vm01.stdout: ceph-prometheus-alerts noarch 2:19.2.3-678.ge911bdeb.el9 ceph-noarch 17 k 2026-03-09T20:49:40.121 INFO:teuthology.orchestra.run.vm01.stdout: ceph-selinux x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 25 k 2026-03-09T20:49:40.121 INFO:teuthology.orchestra.run.vm01.stdout: cryptsetup x86_64 2.8.1-3.el9 baseos 351 k 2026-03-09T20:49:40.121 INFO:teuthology.orchestra.run.vm01.stdout: flexiblas x86_64 3.0.4-9.el9 appstream 30 k 2026-03-09T20:49:40.121 INFO:teuthology.orchestra.run.vm01.stdout: flexiblas-netlib x86_64 3.0.4-9.el9 appstream 3.0 M 2026-03-09T20:49:40.121 INFO:teuthology.orchestra.run.vm01.stdout: flexiblas-openblas-openmp x86_64 3.0.4-9.el9 appstream 15 k 2026-03-09T20:49:40.121 INFO:teuthology.orchestra.run.vm01.stdout: gperftools-libs x86_64 2.9.1-3.el9 epel 308 k 2026-03-09T20:49:40.121 INFO:teuthology.orchestra.run.vm01.stdout: grpc-data noarch 1.46.7-10.el9 epel 19 k 2026-03-09T20:49:40.121 INFO:teuthology.orchestra.run.vm01.stdout: ledmon-libs x86_64 1.1.0-3.el9 baseos 40 k 2026-03-09T20:49:40.121 INFO:teuthology.orchestra.run.vm01.stdout: libarrow x86_64 9.0.0-15.el9 epel 4.4 M 2026-03-09T20:49:40.122 INFO:teuthology.orchestra.run.vm01.stdout: libarrow-doc noarch 9.0.0-15.el9 epel 25 k 2026-03-09T20:49:40.122 INFO:teuthology.orchestra.run.vm01.stdout: libcephsqlite x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 163 k 2026-03-09T20:49:40.122 INFO:teuthology.orchestra.run.vm01.stdout: libconfig x86_64 1.7.2-9.el9 baseos 72 k 2026-03-09T20:49:40.122 INFO:teuthology.orchestra.run.vm01.stdout: libgfortran x86_64 11.5.0-14.el9 baseos 794 k 2026-03-09T20:49:40.122 INFO:teuthology.orchestra.run.vm01.stdout: libnbd x86_64 1.20.3-4.el9 appstream 164 k 2026-03-09T20:49:40.122 INFO:teuthology.orchestra.run.vm01.stdout: liboath x86_64 2.6.12-1.el9 epel 49 k 2026-03-09T20:49:40.122 INFO:teuthology.orchestra.run.vm01.stdout: libpmemobj x86_64 1.12.1-1.el9 appstream 160 k 2026-03-09T20:49:40.122 INFO:teuthology.orchestra.run.vm01.stdout: libquadmath x86_64 11.5.0-14.el9 baseos 184 k 2026-03-09T20:49:40.122 INFO:teuthology.orchestra.run.vm01.stdout: librabbitmq x86_64 0.11.0-7.el9 appstream 45 k 2026-03-09T20:49:40.122 INFO:teuthology.orchestra.run.vm01.stdout: libradosstriper1 x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 503 k 2026-03-09T20:49:40.122 INFO:teuthology.orchestra.run.vm01.stdout: librdkafka x86_64 1.6.1-102.el9 appstream 662 k 2026-03-09T20:49:40.122 INFO:teuthology.orchestra.run.vm01.stdout: librgw2 x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 5.4 M 2026-03-09T20:49:40.122 INFO:teuthology.orchestra.run.vm01.stdout: libstoragemgmt x86_64 1.10.1-1.el9 appstream 246 k 2026-03-09T20:49:40.122 INFO:teuthology.orchestra.run.vm01.stdout: libunwind x86_64 1.6.2-1.el9 epel 67 k 2026-03-09T20:49:40.122 INFO:teuthology.orchestra.run.vm01.stdout: libxslt x86_64 1.1.34-12.el9 appstream 233 k 2026-03-09T20:49:40.122 INFO:teuthology.orchestra.run.vm01.stdout: lttng-ust x86_64 2.12.0-6.el9 appstream 292 k 2026-03-09T20:49:40.122 INFO:teuthology.orchestra.run.vm01.stdout: lua x86_64 5.4.4-4.el9 appstream 188 k 2026-03-09T20:49:40.122 INFO:teuthology.orchestra.run.vm01.stdout: lua-devel x86_64 5.4.4-4.el9 crb 22 k 2026-03-09T20:49:40.122 INFO:teuthology.orchestra.run.vm01.stdout: luarocks noarch 3.9.2-5.el9 epel 151 k 2026-03-09T20:49:40.122 INFO:teuthology.orchestra.run.vm01.stdout: mailcap noarch 2.1.49-5.el9 baseos 33 k 2026-03-09T20:49:40.122 INFO:teuthology.orchestra.run.vm01.stdout: openblas x86_64 0.3.29-1.el9 appstream 42 k 2026-03-09T20:49:40.122 INFO:teuthology.orchestra.run.vm01.stdout: openblas-openmp x86_64 0.3.29-1.el9 appstream 5.3 M 2026-03-09T20:49:40.122 INFO:teuthology.orchestra.run.vm01.stdout: parquet-libs x86_64 9.0.0-15.el9 epel 838 k 2026-03-09T20:49:40.122 INFO:teuthology.orchestra.run.vm01.stdout: pciutils x86_64 3.7.0-7.el9 baseos 93 k 2026-03-09T20:49:40.122 INFO:teuthology.orchestra.run.vm01.stdout: protobuf x86_64 3.14.0-17.el9 appstream 1.0 M 2026-03-09T20:49:40.122 INFO:teuthology.orchestra.run.vm01.stdout: protobuf-compiler x86_64 3.14.0-17.el9 crb 862 k 2026-03-09T20:49:40.122 INFO:teuthology.orchestra.run.vm01.stdout: python3-asyncssh noarch 2.13.2-5.el9 epel 548 k 2026-03-09T20:49:40.122 INFO:teuthology.orchestra.run.vm01.stdout: python3-autocommand noarch 2.2.2-8.el9 epel 29 k 2026-03-09T20:49:40.122 INFO:teuthology.orchestra.run.vm01.stdout: python3-babel noarch 2.9.1-2.el9 appstream 6.0 M 2026-03-09T20:49:40.122 INFO:teuthology.orchestra.run.vm01.stdout: python3-backports-tarfile noarch 1.2.0-1.el9 epel 60 k 2026-03-09T20:49:40.122 INFO:teuthology.orchestra.run.vm01.stdout: python3-bcrypt x86_64 3.2.2-1.el9 epel 43 k 2026-03-09T20:49:40.122 INFO:teuthology.orchestra.run.vm01.stdout: python3-cachetools noarch 4.2.4-1.el9 epel 32 k 2026-03-09T20:49:40.122 INFO:teuthology.orchestra.run.vm01.stdout: python3-ceph-argparse x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 45 k 2026-03-09T20:49:40.122 INFO:teuthology.orchestra.run.vm01.stdout: python3-ceph-common x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 142 k 2026-03-09T20:49:40.122 INFO:teuthology.orchestra.run.vm01.stdout: python3-certifi noarch 2023.05.07-4.el9 epel 14 k 2026-03-09T20:49:40.122 INFO:teuthology.orchestra.run.vm01.stdout: python3-cffi x86_64 1.14.5-5.el9 baseos 253 k 2026-03-09T20:49:40.122 INFO:teuthology.orchestra.run.vm01.stdout: python3-cheroot noarch 10.0.1-4.el9 epel 173 k 2026-03-09T20:49:40.122 INFO:teuthology.orchestra.run.vm01.stdout: python3-cherrypy noarch 18.6.1-2.el9 epel 358 k 2026-03-09T20:49:40.122 INFO:teuthology.orchestra.run.vm01.stdout: python3-cryptography x86_64 36.0.1-5.el9 baseos 1.2 M 2026-03-09T20:49:40.122 INFO:teuthology.orchestra.run.vm01.stdout: python3-devel x86_64 3.9.25-3.el9 appstream 244 k 2026-03-09T20:49:40.122 INFO:teuthology.orchestra.run.vm01.stdout: python3-google-auth noarch 1:2.45.0-1.el9 epel 254 k 2026-03-09T20:49:40.122 INFO:teuthology.orchestra.run.vm01.stdout: python3-grpcio x86_64 1.46.7-10.el9 epel 2.0 M 2026-03-09T20:49:40.122 INFO:teuthology.orchestra.run.vm01.stdout: python3-grpcio-tools x86_64 1.46.7-10.el9 epel 144 k 2026-03-09T20:49:40.122 INFO:teuthology.orchestra.run.vm01.stdout: python3-jaraco noarch 8.2.1-3.el9 epel 11 k 2026-03-09T20:49:40.122 INFO:teuthology.orchestra.run.vm01.stdout: python3-jaraco-classes noarch 3.2.1-5.el9 epel 18 k 2026-03-09T20:49:40.122 INFO:teuthology.orchestra.run.vm01.stdout: python3-jaraco-collections noarch 3.0.0-8.el9 epel 23 k 2026-03-09T20:49:40.122 INFO:teuthology.orchestra.run.vm01.stdout: python3-jaraco-context noarch 6.0.1-3.el9 epel 20 k 2026-03-09T20:49:40.122 INFO:teuthology.orchestra.run.vm01.stdout: python3-jaraco-functools noarch 3.5.0-2.el9 epel 19 k 2026-03-09T20:49:40.122 INFO:teuthology.orchestra.run.vm01.stdout: python3-jaraco-text noarch 4.0.0-2.el9 epel 26 k 2026-03-09T20:49:40.123 INFO:teuthology.orchestra.run.vm01.stdout: python3-jinja2 noarch 2.11.3-8.el9 appstream 249 k 2026-03-09T20:49:40.123 INFO:teuthology.orchestra.run.vm01.stdout: python3-kubernetes noarch 1:26.1.0-3.el9 epel 1.0 M 2026-03-09T20:49:40.123 INFO:teuthology.orchestra.run.vm01.stdout: python3-libstoragemgmt x86_64 1.10.1-1.el9 appstream 177 k 2026-03-09T20:49:40.123 INFO:teuthology.orchestra.run.vm01.stdout: python3-logutils noarch 0.3.5-21.el9 epel 46 k 2026-03-09T20:49:40.123 INFO:teuthology.orchestra.run.vm01.stdout: python3-mako noarch 1.1.4-6.el9 appstream 172 k 2026-03-09T20:49:40.123 INFO:teuthology.orchestra.run.vm01.stdout: python3-markupsafe x86_64 1.1.1-12.el9 appstream 35 k 2026-03-09T20:49:40.123 INFO:teuthology.orchestra.run.vm01.stdout: python3-more-itertools noarch 8.12.0-2.el9 epel 79 k 2026-03-09T20:49:40.123 INFO:teuthology.orchestra.run.vm01.stdout: python3-natsort noarch 7.1.1-5.el9 epel 58 k 2026-03-09T20:49:40.123 INFO:teuthology.orchestra.run.vm01.stdout: python3-numpy x86_64 1:1.23.5-2.el9 appstream 6.1 M 2026-03-09T20:49:40.123 INFO:teuthology.orchestra.run.vm01.stdout: python3-numpy-f2py x86_64 1:1.23.5-2.el9 appstream 442 k 2026-03-09T20:49:40.123 INFO:teuthology.orchestra.run.vm01.stdout: python3-packaging noarch 20.9-5.el9 appstream 77 k 2026-03-09T20:49:40.123 INFO:teuthology.orchestra.run.vm01.stdout: python3-pecan noarch 1.4.2-3.el9 epel 272 k 2026-03-09T20:49:40.123 INFO:teuthology.orchestra.run.vm01.stdout: python3-ply noarch 3.11-14.el9 baseos 106 k 2026-03-09T20:49:40.123 INFO:teuthology.orchestra.run.vm01.stdout: python3-portend noarch 3.1.0-2.el9 epel 16 k 2026-03-09T20:49:40.123 INFO:teuthology.orchestra.run.vm01.stdout: python3-protobuf noarch 3.14.0-17.el9 appstream 267 k 2026-03-09T20:49:40.123 INFO:teuthology.orchestra.run.vm01.stdout: python3-pyOpenSSL noarch 21.0.0-1.el9 epel 90 k 2026-03-09T20:49:40.123 INFO:teuthology.orchestra.run.vm01.stdout: python3-pyasn1 noarch 0.4.8-7.el9 appstream 157 k 2026-03-09T20:49:40.123 INFO:teuthology.orchestra.run.vm01.stdout: python3-pyasn1-modules noarch 0.4.8-7.el9 appstream 277 k 2026-03-09T20:49:40.123 INFO:teuthology.orchestra.run.vm01.stdout: python3-pycparser noarch 2.20-6.el9 baseos 135 k 2026-03-09T20:49:40.123 INFO:teuthology.orchestra.run.vm01.stdout: python3-repoze-lru noarch 0.7-16.el9 epel 31 k 2026-03-09T20:49:40.123 INFO:teuthology.orchestra.run.vm01.stdout: python3-requests noarch 2.25.1-10.el9 baseos 126 k 2026-03-09T20:49:40.123 INFO:teuthology.orchestra.run.vm01.stdout: python3-requests-oauthlib noarch 1.3.0-12.el9 appstream 54 k 2026-03-09T20:49:40.123 INFO:teuthology.orchestra.run.vm01.stdout: python3-routes noarch 2.5.1-5.el9 epel 188 k 2026-03-09T20:49:40.123 INFO:teuthology.orchestra.run.vm01.stdout: python3-rsa noarch 4.9-2.el9 epel 59 k 2026-03-09T20:49:40.123 INFO:teuthology.orchestra.run.vm01.stdout: python3-scipy x86_64 1.9.3-2.el9 appstream 19 M 2026-03-09T20:49:40.123 INFO:teuthology.orchestra.run.vm01.stdout: python3-tempora noarch 5.0.0-2.el9 epel 36 k 2026-03-09T20:49:40.123 INFO:teuthology.orchestra.run.vm01.stdout: python3-toml noarch 0.10.2-6.el9 appstream 42 k 2026-03-09T20:49:40.123 INFO:teuthology.orchestra.run.vm01.stdout: python3-typing-extensions noarch 4.15.0-1.el9 epel 86 k 2026-03-09T20:49:40.123 INFO:teuthology.orchestra.run.vm01.stdout: python3-urllib3 noarch 1.26.5-7.el9 baseos 218 k 2026-03-09T20:49:40.123 INFO:teuthology.orchestra.run.vm01.stdout: python3-webob noarch 1.8.8-2.el9 epel 230 k 2026-03-09T20:49:40.123 INFO:teuthology.orchestra.run.vm01.stdout: python3-websocket-client noarch 1.2.3-2.el9 epel 90 k 2026-03-09T20:49:40.123 INFO:teuthology.orchestra.run.vm01.stdout: python3-werkzeug noarch 2.0.3-3.el9.1 epel 427 k 2026-03-09T20:49:40.123 INFO:teuthology.orchestra.run.vm01.stdout: python3-zc-lockfile noarch 2.0-10.el9 epel 20 k 2026-03-09T20:49:40.123 INFO:teuthology.orchestra.run.vm01.stdout: qatlib x86_64 25.08.0-2.el9 appstream 240 k 2026-03-09T20:49:40.123 INFO:teuthology.orchestra.run.vm01.stdout: qatzip-libs x86_64 1.3.1-1.el9 appstream 66 k 2026-03-09T20:49:40.123 INFO:teuthology.orchestra.run.vm01.stdout: re2 x86_64 1:20211101-20.el9 epel 191 k 2026-03-09T20:49:40.123 INFO:teuthology.orchestra.run.vm01.stdout: socat x86_64 1.7.4.1-8.el9 appstream 303 k 2026-03-09T20:49:40.123 INFO:teuthology.orchestra.run.vm01.stdout: thrift x86_64 0.15.0-4.el9 epel 1.6 M 2026-03-09T20:49:40.123 INFO:teuthology.orchestra.run.vm01.stdout: unzip x86_64 6.0-59.el9 baseos 182 k 2026-03-09T20:49:40.123 INFO:teuthology.orchestra.run.vm01.stdout: xmlstarlet x86_64 1.6.1-20.el9 appstream 64 k 2026-03-09T20:49:40.123 INFO:teuthology.orchestra.run.vm01.stdout: zip x86_64 3.0-35.el9 baseos 266 k 2026-03-09T20:49:40.123 INFO:teuthology.orchestra.run.vm01.stdout:Installing weak dependencies: 2026-03-09T20:49:40.123 INFO:teuthology.orchestra.run.vm01.stdout: qatlib-service x86_64 25.08.0-2.el9 appstream 37 k 2026-03-09T20:49:40.123 INFO:teuthology.orchestra.run.vm01.stdout: 2026-03-09T20:49:40.123 INFO:teuthology.orchestra.run.vm01.stdout:Transaction Summary 2026-03-09T20:49:40.123 INFO:teuthology.orchestra.run.vm01.stdout:====================================================================================== 2026-03-09T20:49:40.123 INFO:teuthology.orchestra.run.vm01.stdout:Install 134 Packages 2026-03-09T20:49:40.123 INFO:teuthology.orchestra.run.vm01.stdout:Upgrade 2 Packages 2026-03-09T20:49:40.123 INFO:teuthology.orchestra.run.vm01.stdout: 2026-03-09T20:49:40.124 INFO:teuthology.orchestra.run.vm01.stdout:Total download size: 210 M 2026-03-09T20:49:40.124 INFO:teuthology.orchestra.run.vm01.stdout:Downloading Packages: 2026-03-09T20:49:40.182 INFO:teuthology.orchestra.run.vm06.stdout:(9/136): ceph-selinux-19.2.3-678.ge911bdeb.el9. 196 kB/s | 25 kB 00:00 2026-03-09T20:49:40.191 INFO:teuthology.orchestra.run.vm08.stdout:(7/136): ceph-mon-19.2.3-678.ge911bdeb.el9.x86_ 10 MB/s | 4.7 MB 00:00 2026-03-09T20:49:40.637 INFO:teuthology.orchestra.run.vm06.stdout:(10/136): ceph-radosgw-19.2.3-678.ge911bdeb.el9 11 MB/s | 11 MB 00:00 2026-03-09T20:49:40.764 INFO:teuthology.orchestra.run.vm06.stdout:(11/136): libcephfs-devel-19.2.3-678.ge911bdeb. 265 kB/s | 34 kB 00:00 2026-03-09T20:49:40.875 INFO:teuthology.orchestra.run.vm08.stdout:(8/136): ceph-common-19.2.3-678.ge911bdeb.el9.x 7.3 MB/s | 22 MB 00:02 2026-03-09T20:49:40.992 INFO:teuthology.orchestra.run.vm08.stdout:(9/136): ceph-selinux-19.2.3-678.ge911bdeb.el9. 215 kB/s | 25 kB 00:00 2026-03-09T20:49:41.061 INFO:teuthology.orchestra.run.vm06.stdout:(12/136): libcephfs2-19.2.3-678.ge911bdeb.el9.x 3.3 MB/s | 1.0 MB 00:00 2026-03-09T20:49:41.189 INFO:teuthology.orchestra.run.vm06.stdout:(13/136): libcephsqlite-19.2.3-678.ge911bdeb.el 1.3 MB/s | 163 kB 00:00 2026-03-09T20:49:41.311 INFO:teuthology.orchestra.run.vm06.stdout:(14/136): librados-devel-19.2.3-678.ge911bdeb.e 1.0 MB/s | 127 kB 00:00 2026-03-09T20:49:41.443 INFO:teuthology.orchestra.run.vm06.stdout:(15/136): libradosstriper1-19.2.3-678.ge911bdeb 3.7 MB/s | 503 kB 00:00 2026-03-09T20:49:41.511 INFO:teuthology.orchestra.run.vm08.stdout:(10/136): ceph-radosgw-19.2.3-678.ge911bdeb.el9 8.1 MB/s | 11 MB 00:01 2026-03-09T20:49:41.567 INFO:teuthology.orchestra.run.vm08.stdout:(11/136): ceph-osd-19.2.3-678.ge911bdeb.el9.x86 9.4 MB/s | 17 MB 00:01 2026-03-09T20:49:41.631 INFO:teuthology.orchestra.run.vm08.stdout:(12/136): libcephfs-devel-19.2.3-678.ge911bdeb. 281 kB/s | 34 kB 00:00 2026-03-09T20:49:41.705 INFO:teuthology.orchestra.run.vm08.stdout:(13/136): libcephfs2-19.2.3-678.ge911bdeb.el9.x 7.1 MB/s | 1.0 MB 00:00 2026-03-09T20:49:41.753 INFO:teuthology.orchestra.run.vm08.stdout:(14/136): libcephsqlite-19.2.3-678.ge911bdeb.el 1.3 MB/s | 163 kB 00:00 2026-03-09T20:49:41.827 INFO:teuthology.orchestra.run.vm08.stdout:(15/136): librados-devel-19.2.3-678.ge911bdeb.e 1.0 MB/s | 127 kB 00:00 2026-03-09T20:49:41.878 INFO:teuthology.orchestra.run.vm08.stdout:(16/136): libradosstriper1-19.2.3-678.ge911bdeb 3.9 MB/s | 503 kB 00:00 2026-03-09T20:49:41.999 INFO:teuthology.orchestra.run.vm08.stdout:(17/136): python3-ceph-argparse-19.2.3-678.ge91 376 kB/s | 45 kB 00:00 2026-03-09T20:49:42.122 INFO:teuthology.orchestra.run.vm08.stdout:(18/136): python3-ceph-common-19.2.3-678.ge911b 1.1 MB/s | 142 kB 00:00 2026-03-09T20:49:42.175 INFO:teuthology.orchestra.run.vm06.stdout:(16/136): librgw2-19.2.3-678.ge911bdeb.el9.x86_ 7.4 MB/s | 5.4 MB 00:00 2026-03-09T20:49:42.229 INFO:teuthology.orchestra.run.vm01.stdout:(1/136): ceph-19.2.3-678.ge911bdeb.el9.x86_64.r 12 kB/s | 6.5 kB 00:00 2026-03-09T20:49:42.243 INFO:teuthology.orchestra.run.vm08.stdout:(19/136): python3-cephfs-19.2.3-678.ge911bdeb.e 1.3 MB/s | 165 kB 00:00 2026-03-09T20:49:42.297 INFO:teuthology.orchestra.run.vm06.stdout:(17/136): python3-ceph-argparse-19.2.3-678.ge91 372 kB/s | 45 kB 00:00 2026-03-09T20:49:42.368 INFO:teuthology.orchestra.run.vm08.stdout:(20/136): python3-rados-19.2.3-678.ge911bdeb.el 2.5 MB/s | 323 kB 00:00 2026-03-09T20:49:42.420 INFO:teuthology.orchestra.run.vm06.stdout:(18/136): python3-ceph-common-19.2.3-678.ge911b 1.1 MB/s | 142 kB 00:00 2026-03-09T20:49:42.437 INFO:teuthology.orchestra.run.vm08.stdout:(21/136): librgw2-19.2.3-678.ge911bdeb.el9.x86_ 8.9 MB/s | 5.4 MB 00:00 2026-03-09T20:49:42.491 INFO:teuthology.orchestra.run.vm08.stdout:(22/136): python3-rbd-19.2.3-678.ge911bdeb.el9. 2.4 MB/s | 303 kB 00:00 2026-03-09T20:49:42.542 INFO:teuthology.orchestra.run.vm06.stdout:(19/136): python3-cephfs-19.2.3-678.ge911bdeb.e 1.3 MB/s | 165 kB 00:00 2026-03-09T20:49:42.561 INFO:teuthology.orchestra.run.vm08.stdout:(23/136): python3-rgw-19.2.3-678.ge911bdeb.el9. 808 kB/s | 100 kB 00:00 2026-03-09T20:49:42.611 INFO:teuthology.orchestra.run.vm08.stdout:(24/136): rbd-fuse-19.2.3-678.ge911bdeb.el9.x86 709 kB/s | 85 kB 00:00 2026-03-09T20:49:42.666 INFO:teuthology.orchestra.run.vm06.stdout:(20/136): python3-rados-19.2.3-678.ge911bdeb.el 2.6 MB/s | 323 kB 00:00 2026-03-09T20:49:42.765 INFO:teuthology.orchestra.run.vm08.stdout:(25/136): rbd-nbd-19.2.3-678.ge911bdeb.el9.x86_ 1.1 MB/s | 171 kB 00:00 2026-03-09T20:49:42.790 INFO:teuthology.orchestra.run.vm06.stdout:(21/136): python3-rbd-19.2.3-678.ge911bdeb.el9. 2.4 MB/s | 303 kB 00:00 2026-03-09T20:49:42.887 INFO:teuthology.orchestra.run.vm08.stdout:(26/136): ceph-grafana-dashboards-19.2.3-678.ge 257 kB/s | 31 kB 00:00 2026-03-09T20:49:42.918 INFO:teuthology.orchestra.run.vm06.stdout:(22/136): python3-rgw-19.2.3-678.ge911bdeb.el9. 783 kB/s | 100 kB 00:00 2026-03-09T20:49:42.934 INFO:teuthology.orchestra.run.vm08.stdout:(27/136): rbd-mirror-19.2.3-678.ge911bdeb.el9.x 8.4 MB/s | 3.1 MB 00:00 2026-03-09T20:49:43.010 INFO:teuthology.orchestra.run.vm08.stdout:(28/136): ceph-mgr-cephadm-19.2.3-678.ge911bdeb 1.2 MB/s | 150 kB 00:00 2026-03-09T20:49:43.053 INFO:teuthology.orchestra.run.vm06.stdout:(23/136): rbd-fuse-19.2.3-678.ge911bdeb.el9.x86 631 kB/s | 85 kB 00:00 2026-03-09T20:49:43.095 INFO:teuthology.orchestra.run.vm01.stdout:(2/136): ceph-fuse-19.2.3-678.ge911bdeb.el9.x86 1.3 MB/s | 1.2 MB 00:00 2026-03-09T20:49:43.270 INFO:teuthology.orchestra.run.vm01.stdout:(3/136): ceph-immutable-object-cache-19.2.3-678 833 kB/s | 145 kB 00:00 2026-03-09T20:49:43.330 INFO:teuthology.orchestra.run.vm01.stdout:(4/136): ceph-base-19.2.3-678.ge911bdeb.el9.x86 3.4 MB/s | 5.5 MB 00:01 2026-03-09T20:49:43.467 INFO:teuthology.orchestra.run.vm01.stdout:(5/136): ceph-mgr-19.2.3-678.ge911bdeb.el9.x86_ 7.9 MB/s | 1.1 MB 00:00 2026-03-09T20:49:43.539 INFO:teuthology.orchestra.run.vm08.stdout:(29/136): ceph-mgr-dashboard-19.2.3-678.ge911bd 6.3 MB/s | 3.8 MB 00:00 2026-03-09T20:49:43.686 INFO:teuthology.orchestra.run.vm06.stdout:(24/136): ceph-test-19.2.3-678.ge911bdeb.el9.x8 14 MB/s | 50 MB 00:03 2026-03-09T20:49:43.687 INFO:teuthology.orchestra.run.vm08.stdout:(30/136): ceph-mgr-modules-core-19.2.3-678.ge91 1.7 MB/s | 253 kB 00:00 2026-03-09T20:49:43.701 INFO:teuthology.orchestra.run.vm01.stdout:(6/136): ceph-mds-19.2.3-678.ge911bdeb.el9.x86_ 5.6 MB/s | 2.4 MB 00:00 2026-03-09T20:49:43.808 INFO:teuthology.orchestra.run.vm08.stdout:(31/136): ceph-mgr-rook-19.2.3-678.ge911bdeb.el 406 kB/s | 49 kB 00:00 2026-03-09T20:49:43.817 INFO:teuthology.orchestra.run.vm06.stdout:(25/136): rbd-mirror-19.2.3-678.ge911bdeb.el9.x 4.1 MB/s | 3.1 MB 00:00 2026-03-09T20:49:43.818 INFO:teuthology.orchestra.run.vm06.stdout:(26/136): rbd-nbd-19.2.3-678.ge911bdeb.el9.x86_ 1.3 MB/s | 171 kB 00:00 2026-03-09T20:49:43.868 INFO:teuthology.orchestra.run.vm01.stdout:(7/136): ceph-mon-19.2.3-678.ge911bdeb.el9.x86_ 12 MB/s | 4.7 MB 00:00 2026-03-09T20:49:43.909 INFO:teuthology.orchestra.run.vm08.stdout:(32/136): ceph-mgr-diskprediction-local-19.2.3- 8.2 MB/s | 7.4 MB 00:00 2026-03-09T20:49:43.928 INFO:teuthology.orchestra.run.vm08.stdout:(33/136): ceph-prometheus-alerts-19.2.3-678.ge9 141 kB/s | 17 kB 00:00 2026-03-09T20:49:43.940 INFO:teuthology.orchestra.run.vm06.stdout:(27/136): ceph-mgr-cephadm-19.2.3-678.ge911bdeb 1.2 MB/s | 150 kB 00:00 2026-03-09T20:49:43.941 INFO:teuthology.orchestra.run.vm06.stdout:(28/136): ceph-grafana-dashboards-19.2.3-678.ge 250 kB/s | 31 kB 00:00 2026-03-09T20:49:44.043 INFO:teuthology.orchestra.run.vm08.stdout:(34/136): ceph-volume-19.2.3-678.ge911bdeb.el9. 2.2 MB/s | 299 kB 00:00 2026-03-09T20:49:44.072 INFO:teuthology.orchestra.run.vm08.stdout:(35/136): cephadm-19.2.3-678.ge911bdeb.el9.noar 5.2 MB/s | 769 kB 00:00 2026-03-09T20:49:44.157 INFO:teuthology.orchestra.run.vm08.stdout:(36/136): ledmon-libs-1.1.0-3.el9.x86_64.rpm 479 kB/s | 40 kB 00:00 2026-03-09T20:49:44.215 INFO:teuthology.orchestra.run.vm08.stdout:(37/136): libconfig-1.7.2-9.el9.x86_64.rpm 1.2 MB/s | 72 kB 00:00 2026-03-09T20:49:44.250 INFO:teuthology.orchestra.run.vm08.stdout:(38/136): cryptsetup-2.8.1-3.el9.x86_64.rpm 1.6 MB/s | 351 kB 00:00 2026-03-09T20:49:44.282 INFO:teuthology.orchestra.run.vm08.stdout:(39/136): libquadmath-11.5.0-14.el9.x86_64.rpm 5.8 MB/s | 184 kB 00:00 2026-03-09T20:49:44.308 INFO:teuthology.orchestra.run.vm08.stdout:(40/136): libgfortran-11.5.0-14.el9.x86_64.rpm 8.4 MB/s | 794 kB 00:00 2026-03-09T20:49:44.311 INFO:teuthology.orchestra.run.vm08.stdout:(41/136): mailcap-2.1.49-5.el9.noarch.rpm 1.1 MB/s | 33 kB 00:00 2026-03-09T20:49:44.338 INFO:teuthology.orchestra.run.vm08.stdout:(42/136): pciutils-3.7.0-7.el9.x86_64.rpm 3.0 MB/s | 93 kB 00:00 2026-03-09T20:49:44.344 INFO:teuthology.orchestra.run.vm08.stdout:(43/136): python3-cffi-1.14.5-5.el9.x86_64.rpm 7.8 MB/s | 253 kB 00:00 2026-03-09T20:49:44.378 INFO:teuthology.orchestra.run.vm08.stdout:(44/136): python3-ply-3.11-14.el9.noarch.rpm 3.1 MB/s | 106 kB 00:00 2026-03-09T20:49:44.409 INFO:teuthology.orchestra.run.vm08.stdout:(45/136): python3-pycparser-2.20-6.el9.noarch.r 4.3 MB/s | 135 kB 00:00 2026-03-09T20:49:44.429 INFO:teuthology.orchestra.run.vm08.stdout:(46/136): python3-cryptography-36.0.1-5.el9.x86 14 MB/s | 1.2 MB 00:00 2026-03-09T20:49:44.440 INFO:teuthology.orchestra.run.vm08.stdout:(47/136): python3-requests-2.25.1-10.el9.noarch 4.1 MB/s | 126 kB 00:00 2026-03-09T20:49:44.461 INFO:teuthology.orchestra.run.vm08.stdout:(48/136): python3-urllib3-1.26.5-7.el9.noarch.r 6.7 MB/s | 218 kB 00:00 2026-03-09T20:49:44.471 INFO:teuthology.orchestra.run.vm08.stdout:(49/136): unzip-6.0-59.el9.x86_64.rpm 5.8 MB/s | 182 kB 00:00 2026-03-09T20:49:44.494 INFO:teuthology.orchestra.run.vm08.stdout:(50/136): zip-3.0-35.el9.x86_64.rpm 7.9 MB/s | 266 kB 00:00 2026-03-09T20:49:44.598 INFO:teuthology.orchestra.run.vm08.stdout:(51/136): flexiblas-3.0.4-9.el9.x86_64.rpm 286 kB/s | 30 kB 00:00 2026-03-09T20:49:44.614 INFO:teuthology.orchestra.run.vm06.stdout:(29/136): ceph-mgr-dashboard-19.2.3-678.ge911bd 5.6 MB/s | 3.8 MB 00:00 2026-03-09T20:49:44.639 INFO:teuthology.orchestra.run.vm08.stdout:(52/136): boost-program-options-1.75.0-13.el9.x 622 kB/s | 104 kB 00:00 2026-03-09T20:49:44.690 INFO:teuthology.orchestra.run.vm08.stdout:(53/136): flexiblas-openblas-openmp-3.0.4-9.el9 294 kB/s | 15 kB 00:00 2026-03-09T20:49:44.690 INFO:teuthology.orchestra.run.vm01.stdout:(8/136): ceph-radosgw-19.2.3-678.ge911bdeb.el9. 13 MB/s | 11 MB 00:00 2026-03-09T20:49:44.766 INFO:teuthology.orchestra.run.vm08.stdout:(54/136): libnbd-1.20.3-4.el9.x86_64.rpm 2.1 MB/s | 164 kB 00:00 2026-03-09T20:49:44.766 INFO:teuthology.orchestra.run.vm06.stdout:(30/136): ceph-osd-19.2.3-678.ge911bdeb.el9.x86 3.2 MB/s | 17 MB 00:05 2026-03-09T20:49:44.771 INFO:teuthology.orchestra.run.vm06.stdout:(31/136): ceph-mgr-modules-core-19.2.3-678.ge91 1.6 MB/s | 253 kB 00:00 2026-03-09T20:49:44.809 INFO:teuthology.orchestra.run.vm08.stdout:(55/136): flexiblas-netlib-3.0.4-9.el9.x86_64.r 14 MB/s | 3.0 MB 00:00 2026-03-09T20:49:44.810 INFO:teuthology.orchestra.run.vm01.stdout:(9/136): ceph-selinux-19.2.3-678.ge911bdeb.el9. 210 kB/s | 25 kB 00:00 2026-03-09T20:49:44.819 INFO:teuthology.orchestra.run.vm08.stdout:(56/136): libpmemobj-1.12.1-1.el9.x86_64.rpm 3.0 MB/s | 160 kB 00:00 2026-03-09T20:49:44.838 INFO:teuthology.orchestra.run.vm08.stdout:(57/136): librabbitmq-0.11.0-7.el9.x86_64.rpm 1.5 MB/s | 45 kB 00:00 2026-03-09T20:49:44.883 INFO:teuthology.orchestra.run.vm08.stdout:(58/136): libstoragemgmt-1.10.1-1.el9.x86_64.rp 5.4 MB/s | 246 kB 00:00 2026-03-09T20:49:44.914 INFO:teuthology.orchestra.run.vm08.stdout:(59/136): libxslt-1.1.34-12.el9.x86_64.rpm 7.3 MB/s | 233 kB 00:00 2026-03-09T20:49:44.958 INFO:teuthology.orchestra.run.vm08.stdout:(60/136): librdkafka-1.6.1-102.el9.x86_64.rpm 4.6 MB/s | 662 kB 00:00 2026-03-09T20:49:44.963 INFO:teuthology.orchestra.run.vm08.stdout:(61/136): lttng-ust-2.12.0-6.el9.x86_64.rpm 5.9 MB/s | 292 kB 00:00 2026-03-09T20:49:45.013 INFO:teuthology.orchestra.run.vm08.stdout:(62/136): lua-5.4.4-4.el9.x86_64.rpm 3.4 MB/s | 188 kB 00:00 2026-03-09T20:49:45.013 INFO:teuthology.orchestra.run.vm01.stdout:(10/136): ceph-common-19.2.3-678.ge911bdeb.el9. 6.6 MB/s | 22 MB 00:03 2026-03-09T20:49:45.014 INFO:teuthology.orchestra.run.vm08.stdout:(63/136): openblas-0.3.29-1.el9.x86_64.rpm 833 kB/s | 42 kB 00:00 2026-03-09T20:49:45.135 INFO:teuthology.orchestra.run.vm01.stdout:(11/136): libcephfs-devel-19.2.3-678.ge911bdeb. 277 kB/s | 34 kB 00:00 2026-03-09T20:49:45.185 INFO:teuthology.orchestra.run.vm08.stdout:(64/136): protobuf-3.14.0-17.el9.x86_64.rpm 5.9 MB/s | 1.0 MB 00:00 2026-03-09T20:49:45.276 INFO:teuthology.orchestra.run.vm01.stdout:(12/136): libcephfs2-19.2.3-678.ge911bdeb.el9.x 7.0 MB/s | 1.0 MB 00:00 2026-03-09T20:49:45.294 INFO:teuthology.orchestra.run.vm06.stdout:(32/136): ceph-prometheus-alerts-19.2.3-678.ge9 32 kB/s | 17 kB 00:00 2026-03-09T20:49:45.397 INFO:teuthology.orchestra.run.vm01.stdout:(13/136): libcephsqlite-19.2.3-678.ge911bdeb.el 1.3 MB/s | 163 kB 00:00 2026-03-09T20:49:45.416 INFO:teuthology.orchestra.run.vm06.stdout:(33/136): ceph-volume-19.2.3-678.ge911bdeb.el9. 2.4 MB/s | 299 kB 00:00 2026-03-09T20:49:45.439 INFO:teuthology.orchestra.run.vm08.stdout:(65/136): python3-babel-2.9.1-2.el9.noarch.rpm 23 MB/s | 6.0 MB 00:00 2026-03-09T20:49:45.471 INFO:teuthology.orchestra.run.vm08.stdout:(66/136): python3-devel-3.9.25-3.el9.x86_64.rpm 7.6 MB/s | 244 kB 00:00 2026-03-09T20:49:45.490 INFO:teuthology.orchestra.run.vm08.stdout:(67/136): openblas-openmp-0.3.29-1.el9.x86_64.r 11 MB/s | 5.3 MB 00:00 2026-03-09T20:49:45.509 INFO:teuthology.orchestra.run.vm06.stdout:(34/136): ceph-mgr-rook-19.2.3-678.ge911bdeb.el 66 kB/s | 49 kB 00:00 2026-03-09T20:49:45.525 INFO:teuthology.orchestra.run.vm01.stdout:(14/136): ceph-osd-19.2.3-678.ge911bdeb.el9.x86 9.4 MB/s | 17 MB 00:01 2026-03-09T20:49:45.527 INFO:teuthology.orchestra.run.vm01.stdout:(15/136): librados-devel-19.2.3-678.ge911bdeb.e 979 kB/s | 127 kB 00:00 2026-03-09T20:49:45.528 INFO:teuthology.orchestra.run.vm08.stdout:(68/136): python3-jmespath-1.0.1-1.el9.noarch.r 1.2 MB/s | 48 kB 00:00 2026-03-09T20:49:45.530 INFO:teuthology.orchestra.run.vm08.stdout:(69/136): python3-jinja2-2.11.3-8.el9.noarch.rp 4.1 MB/s | 249 kB 00:00 2026-03-09T20:49:45.582 INFO:teuthology.orchestra.run.vm08.stdout:(70/136): python3-libstoragemgmt-1.10.1-1.el9.x 3.2 MB/s | 177 kB 00:00 2026-03-09T20:49:45.600 INFO:teuthology.orchestra.run.vm08.stdout:(71/136): python3-mako-1.1.4-6.el9.noarch.rpm 2.4 MB/s | 172 kB 00:00 2026-03-09T20:49:45.622 INFO:teuthology.orchestra.run.vm08.stdout:(72/136): python3-markupsafe-1.1.1-12.el9.x86_6 872 kB/s | 35 kB 00:00 2026-03-09T20:49:45.644 INFO:teuthology.orchestra.run.vm06.stdout:(35/136): cryptsetup-2.8.1-3.el9.x86_64.rpm 2.6 MB/s | 351 kB 00:00 2026-03-09T20:49:45.657 INFO:teuthology.orchestra.run.vm01.stdout:(16/136): libradosstriper1-19.2.3-678.ge911bdeb 3.7 MB/s | 503 kB 00:00 2026-03-09T20:49:45.689 INFO:teuthology.orchestra.run.vm06.stdout:(36/136): cephadm-19.2.3-678.ge911bdeb.el9.noar 2.8 MB/s | 769 kB 00:00 2026-03-09T20:49:45.690 INFO:teuthology.orchestra.run.vm08.stdout:(73/136): python3-numpy-f2py-1.23.5-2.el9.x86_6 6.4 MB/s | 442 kB 00:00 2026-03-09T20:49:45.691 INFO:teuthology.orchestra.run.vm06.stdout:(37/136): ledmon-libs-1.1.0-3.el9.x86_64.rpm 876 kB/s | 40 kB 00:00 2026-03-09T20:49:45.724 INFO:teuthology.orchestra.run.vm06.stdout:(38/136): libconfig-1.7.2-9.el9.x86_64.rpm 2.0 MB/s | 72 kB 00:00 2026-03-09T20:49:45.729 INFO:teuthology.orchestra.run.vm08.stdout:(74/136): python3-packaging-20.9-5.el9.noarch.r 2.0 MB/s | 77 kB 00:00 2026-03-09T20:49:45.751 INFO:teuthology.orchestra.run.vm06.stdout:(39/136): libquadmath-11.5.0-14.el9.x86_64.rpm 6.8 MB/s | 184 kB 00:00 2026-03-09T20:49:45.756 INFO:teuthology.orchestra.run.vm06.stdout:(40/136): mailcap-2.1.49-5.el9.noarch.rpm 6.8 MB/s | 33 kB 00:00 2026-03-09T20:49:45.772 INFO:teuthology.orchestra.run.vm06.stdout:(41/136): pciutils-3.7.0-7.el9.x86_64.rpm 5.9 MB/s | 93 kB 00:00 2026-03-09T20:49:45.778 INFO:teuthology.orchestra.run.vm01.stdout:(17/136): python3-ceph-argparse-19.2.3-678.ge91 373 kB/s | 45 kB 00:00 2026-03-09T20:49:45.787 INFO:teuthology.orchestra.run.vm08.stdout:(75/136): python3-protobuf-3.14.0-17.el9.noarch 4.5 MB/s | 267 kB 00:00 2026-03-09T20:49:45.850 INFO:teuthology.orchestra.run.vm08.stdout:(76/136): python3-pyasn1-0.4.8-7.el9.noarch.rpm 2.5 MB/s | 157 kB 00:00 2026-03-09T20:49:45.857 INFO:teuthology.orchestra.run.vm06.stdout:(42/136): libgfortran-11.5.0-14.el9.x86_64.rpm 4.7 MB/s | 794 kB 00:00 2026-03-09T20:49:45.884 INFO:teuthology.orchestra.run.vm06.stdout:(43/136): python3-cffi-1.14.5-5.el9.x86_64.rpm 2.2 MB/s | 253 kB 00:00 2026-03-09T20:49:45.894 INFO:teuthology.orchestra.run.vm08.stdout:(77/136): python3-pyasn1-modules-0.4.8-7.el9.no 6.1 MB/s | 277 kB 00:00 2026-03-09T20:49:45.903 INFO:teuthology.orchestra.run.vm01.stdout:(18/136): python3-ceph-common-19.2.3-678.ge911b 1.1 MB/s | 142 kB 00:00 2026-03-09T20:49:45.911 INFO:teuthology.orchestra.run.vm06.stdout:(44/136): python3-ply-3.11-14.el9.noarch.rpm 4.0 MB/s | 106 kB 00:00 2026-03-09T20:49:45.929 INFO:teuthology.orchestra.run.vm08.stdout:(78/136): python3-requests-oauthlib-1.3.0-12.el 1.5 MB/s | 54 kB 00:00 2026-03-09T20:49:45.966 INFO:teuthology.orchestra.run.vm08.stdout:(79/136): python3-numpy-1.23.5-2.el9.x86_64.rpm 17 MB/s | 6.1 MB 00:00 2026-03-09T20:49:45.966 INFO:teuthology.orchestra.run.vm06.stdout:(45/136): ceph-mgr-diskprediction-local-19.2.3- 3.6 MB/s | 7.4 MB 00:02 2026-03-09T20:49:45.971 INFO:teuthology.orchestra.run.vm06.stdout:(46/136): python3-cryptography-36.0.1-5.el9.x86 11 MB/s | 1.2 MB 00:00 2026-03-09T20:49:45.979 INFO:teuthology.orchestra.run.vm06.stdout:(47/136): python3-urllib3-1.26.5-7.el9.noarch.r 28 MB/s | 218 kB 00:00 2026-03-09T20:49:45.986 INFO:teuthology.orchestra.run.vm06.stdout:(48/136): python3-requests-2.25.1-10.el9.noarch 6.4 MB/s | 126 kB 00:00 2026-03-09T20:49:46.006 INFO:teuthology.orchestra.run.vm08.stdout:(80/136): python3-toml-0.10.2-6.el9.noarch.rpm 1.0 MB/s | 42 kB 00:00 2026-03-09T20:49:46.026 INFO:teuthology.orchestra.run.vm01.stdout:(19/136): python3-cephfs-19.2.3-678.ge911bdeb.e 1.3 MB/s | 165 kB 00:00 2026-03-09T20:49:46.049 INFO:teuthology.orchestra.run.vm06.stdout:(49/136): python3-pycparser-2.20-6.el9.noarch.r 980 kB/s | 135 kB 00:00 2026-03-09T20:49:46.059 INFO:teuthology.orchestra.run.vm01.stdout:(20/136): librgw2-19.2.3-678.ge911bdeb.el9.x86_ 10 MB/s | 5.4 MB 00:00 2026-03-09T20:49:46.060 INFO:teuthology.orchestra.run.vm08.stdout:(81/136): qatlib-25.08.0-2.el9.x86_64.rpm 4.4 MB/s | 240 kB 00:00 2026-03-09T20:49:46.095 INFO:teuthology.orchestra.run.vm08.stdout:(82/136): qatlib-service-25.08.0-2.el9.x86_64.r 1.1 MB/s | 37 kB 00:00 2026-03-09T20:49:46.114 INFO:teuthology.orchestra.run.vm06.stdout:(50/136): unzip-6.0-59.el9.x86_64.rpm 1.3 MB/s | 182 kB 00:00 2026-03-09T20:49:46.119 INFO:teuthology.orchestra.run.vm06.stdout:(51/136): zip-3.0-35.el9.x86_64.rpm 2.0 MB/s | 266 kB 00:00 2026-03-09T20:49:46.139 INFO:teuthology.orchestra.run.vm08.stdout:(83/136): qatzip-libs-1.3.1-1.el9.x86_64.rpm 1.5 MB/s | 66 kB 00:00 2026-03-09T20:49:46.148 INFO:teuthology.orchestra.run.vm01.stdout:(21/136): python3-rados-19.2.3-678.ge911bdeb.el 2.6 MB/s | 323 kB 00:00 2026-03-09T20:49:46.172 INFO:teuthology.orchestra.run.vm08.stdout:(84/136): socat-1.7.4.1-8.el9.x86_64.rpm 9.2 MB/s | 303 kB 00:00 2026-03-09T20:49:46.182 INFO:teuthology.orchestra.run.vm01.stdout:(22/136): python3-rbd-19.2.3-678.ge911bdeb.el9. 2.4 MB/s | 303 kB 00:00 2026-03-09T20:49:46.207 INFO:teuthology.orchestra.run.vm08.stdout:(85/136): xmlstarlet-1.6.1-20.el9.x86_64.rpm 1.8 MB/s | 64 kB 00:00 2026-03-09T20:49:46.226 INFO:teuthology.orchestra.run.vm08.stdout:(86/136): lua-devel-5.4.4-4.el9.x86_64.rpm 1.2 MB/s | 22 kB 00:00 2026-03-09T20:49:46.268 INFO:teuthology.orchestra.run.vm01.stdout:(23/136): python3-rgw-19.2.3-678.ge911bdeb.el9. 833 kB/s | 100 kB 00:00 2026-03-09T20:49:46.302 INFO:teuthology.orchestra.run.vm01.stdout:(24/136): rbd-fuse-19.2.3-678.ge911bdeb.el9.x86 709 kB/s | 85 kB 00:00 2026-03-09T20:49:46.412 INFO:teuthology.orchestra.run.vm06.stdout:(52/136): flexiblas-3.0.4-9.el9.x86_64.rpm 99 kB/s | 30 kB 00:00 2026-03-09T20:49:46.413 INFO:teuthology.orchestra.run.vm08.stdout:(87/136): protobuf-compiler-3.14.0-17.el9.x86_6 4.5 MB/s | 862 kB 00:00 2026-03-09T20:49:46.443 INFO:teuthology.orchestra.run.vm08.stdout:(88/136): abseil-cpp-20211102.0-4.el9.x86_64.rp 18 MB/s | 551 kB 00:00 2026-03-09T20:49:46.444 INFO:teuthology.orchestra.run.vm01.stdout:(25/136): rbd-nbd-19.2.3-678.ge911bdeb.el9.x86_ 1.2 MB/s | 171 kB 00:00 2026-03-09T20:49:46.452 INFO:teuthology.orchestra.run.vm08.stdout:(89/136): gperftools-libs-2.9.1-3.el9.x86_64.rp 36 MB/s | 308 kB 00:00 2026-03-09T20:49:46.455 INFO:teuthology.orchestra.run.vm08.stdout:(90/136): grpc-data-1.46.7-10.el9.noarch.rpm 8.2 MB/s | 19 kB 00:00 2026-03-09T20:49:46.512 INFO:teuthology.orchestra.run.vm06.stdout:(53/136): boost-program-options-1.75.0-13.el9.x 225 kB/s | 104 kB 00:00 2026-03-09T20:49:46.518 INFO:teuthology.orchestra.run.vm06.stdout:(54/136): flexiblas-openblas-openmp-3.0.4-9.el9 140 kB/s | 15 kB 00:00 2026-03-09T20:49:46.539 INFO:teuthology.orchestra.run.vm08.stdout:(91/136): libarrow-9.0.0-15.el9.x86_64.rpm 52 MB/s | 4.4 MB 00:00 2026-03-09T20:49:46.549 INFO:teuthology.orchestra.run.vm08.stdout:(92/136): libarrow-doc-9.0.0-15.el9.noarch.rpm 2.7 MB/s | 25 kB 00:00 2026-03-09T20:49:46.552 INFO:teuthology.orchestra.run.vm08.stdout:(93/136): liboath-2.6.12-1.el9.x86_64.rpm 18 MB/s | 49 kB 00:00 2026-03-09T20:49:46.555 INFO:teuthology.orchestra.run.vm08.stdout:(94/136): libunwind-1.6.2-1.el9.x86_64.rpm 20 MB/s | 67 kB 00:00 2026-03-09T20:49:46.560 INFO:teuthology.orchestra.run.vm08.stdout:(95/136): luarocks-3.9.2-5.el9.noarch.rpm 36 MB/s | 151 kB 00:00 2026-03-09T20:49:46.565 INFO:teuthology.orchestra.run.vm01.stdout:(26/136): ceph-grafana-dashboards-19.2.3-678.ge 259 kB/s | 31 kB 00:00 2026-03-09T20:49:46.579 INFO:teuthology.orchestra.run.vm08.stdout:(96/136): parquet-libs-9.0.0-15.el9.x86_64.rpm 45 MB/s | 838 kB 00:00 2026-03-09T20:49:46.588 INFO:teuthology.orchestra.run.vm08.stdout:(97/136): python3-asyncssh-2.13.2-5.el9.noarch. 58 MB/s | 548 kB 00:00 2026-03-09T20:49:46.592 INFO:teuthology.orchestra.run.vm08.stdout:(98/136): python3-autocommand-2.2.2-8.el9.noarc 9.7 MB/s | 29 kB 00:00 2026-03-09T20:49:46.595 INFO:teuthology.orchestra.run.vm08.stdout:(99/136): python3-backports-tarfile-1.2.0-1.el9 17 MB/s | 60 kB 00:00 2026-03-09T20:49:46.601 INFO:teuthology.orchestra.run.vm08.stdout:(100/136): python3-bcrypt-3.2.2-1.el9.x86_64.rp 9.6 MB/s | 43 kB 00:00 2026-03-09T20:49:46.605 INFO:teuthology.orchestra.run.vm08.stdout:(101/136): python3-cachetools-4.2.4-1.el9.noarc 7.9 MB/s | 32 kB 00:00 2026-03-09T20:49:46.608 INFO:teuthology.orchestra.run.vm08.stdout:(102/136): python3-certifi-2023.05.07-4.el9.noa 5.7 MB/s | 14 kB 00:00 2026-03-09T20:49:46.612 INFO:teuthology.orchestra.run.vm08.stdout:(103/136): python3-cheroot-10.0.1-4.el9.noarch. 42 MB/s | 173 kB 00:00 2026-03-09T20:49:46.619 INFO:teuthology.orchestra.run.vm08.stdout:(104/136): python3-cherrypy-18.6.1-2.el9.noarch 54 MB/s | 358 kB 00:00 2026-03-09T20:49:46.625 INFO:teuthology.orchestra.run.vm08.stdout:(105/136): python3-google-auth-2.45.0-1.el9.noa 45 MB/s | 254 kB 00:00 2026-03-09T20:49:46.666 INFO:teuthology.orchestra.run.vm08.stdout:(106/136): python3-grpcio-1.46.7-10.el9.x86_64. 50 MB/s | 2.0 MB 00:00 2026-03-09T20:49:46.686 INFO:teuthology.orchestra.run.vm01.stdout:(27/136): ceph-mgr-cephadm-19.2.3-678.ge911bdeb 1.2 MB/s | 150 kB 00:00 2026-03-09T20:49:46.686 INFO:teuthology.orchestra.run.vm06.stdout:(55/136): libnbd-1.20.3-4.el9.x86_64.rpm 942 kB/s | 164 kB 00:00 2026-03-09T20:49:46.815 INFO:teuthology.orchestra.run.vm06.stdout:(56/136): libpmemobj-1.12.1-1.el9.x86_64.rpm 539 kB/s | 160 kB 00:00 2026-03-09T20:49:46.816 INFO:teuthology.orchestra.run.vm01.stdout:(28/136): rbd-mirror-19.2.3-678.ge911bdeb.el9.x 5.7 MB/s | 3.1 MB 00:00 2026-03-09T20:49:46.816 INFO:teuthology.orchestra.run.vm08.stdout:(107/136): ceph-test-19.2.3-678.ge911bdeb.el9.x 8.5 MB/s | 50 MB 00:05 2026-03-09T20:49:46.816 INFO:teuthology.orchestra.run.vm06.stdout:(57/136): librabbitmq-0.11.0-7.el9.x86_64.rpm 349 kB/s | 45 kB 00:00 2026-03-09T20:49:46.818 INFO:teuthology.orchestra.run.vm08.stdout:(108/136): python3-grpcio-tools-1.46.7-10.el9.x 954 kB/s | 144 kB 00:00 2026-03-09T20:49:46.824 INFO:teuthology.orchestra.run.vm08.stdout:(109/136): python3-jaraco-classes-3.2.1-5.el9.n 3.2 MB/s | 18 kB 00:00 2026-03-09T20:49:46.824 INFO:teuthology.orchestra.run.vm08.stdout:(110/136): python3-jaraco-8.2.1-3.el9.noarch.rp 1.3 MB/s | 11 kB 00:00 2026-03-09T20:49:46.827 INFO:teuthology.orchestra.run.vm08.stdout:(111/136): python3-jaraco-collections-3.0.0-8.e 8.5 MB/s | 23 kB 00:00 2026-03-09T20:49:46.828 INFO:teuthology.orchestra.run.vm08.stdout:(112/136): python3-jaraco-context-6.0.1-3.el9.n 6.6 MB/s | 20 kB 00:00 2026-03-09T20:49:46.830 INFO:teuthology.orchestra.run.vm08.stdout:(113/136): python3-jaraco-functools-3.5.0-2.el9 5.6 MB/s | 19 kB 00:00 2026-03-09T20:49:46.831 INFO:teuthology.orchestra.run.vm08.stdout:(114/136): python3-jaraco-text-4.0.0-2.el9.noar 7.8 MB/s | 26 kB 00:00 2026-03-09T20:49:46.839 INFO:teuthology.orchestra.run.vm08.stdout:(115/136): python3-logutils-0.3.5-21.el9.noarch 6.0 MB/s | 46 kB 00:00 2026-03-09T20:49:46.849 INFO:teuthology.orchestra.run.vm08.stdout:(116/136): python3-kubernetes-26.1.0-3.el9.noar 59 MB/s | 1.0 MB 00:00 2026-03-09T20:49:46.850 INFO:teuthology.orchestra.run.vm08.stdout:(117/136): python3-more-itertools-8.12.0-2.el9. 7.3 MB/s | 79 kB 00:00 2026-03-09T20:49:46.854 INFO:teuthology.orchestra.run.vm08.stdout:(118/136): python3-natsort-7.1.1-5.el9.noarch.r 11 MB/s | 58 kB 00:00 2026-03-09T20:49:46.859 INFO:teuthology.orchestra.run.vm08.stdout:(119/136): python3-portend-3.1.0-2.el9.noarch.r 3.6 MB/s | 16 kB 00:00 2026-03-09T20:49:46.862 INFO:teuthology.orchestra.run.vm08.stdout:(120/136): python3-pecan-1.4.2-3.el9.noarch.rpm 22 MB/s | 272 kB 00:00 2026-03-09T20:49:46.864 INFO:teuthology.orchestra.run.vm08.stdout:(121/136): python3-pyOpenSSL-21.0.0-1.el9.noarc 17 MB/s | 90 kB 00:00 2026-03-09T20:49:46.866 INFO:teuthology.orchestra.run.vm08.stdout:(122/136): python3-repoze-lru-0.7-16.el9.noarch 8.4 MB/s | 31 kB 00:00 2026-03-09T20:49:46.869 INFO:teuthology.orchestra.run.vm08.stdout:(123/136): python3-routes-2.5.1-5.el9.noarch.rp 44 MB/s | 188 kB 00:00 2026-03-09T20:49:46.869 INFO:teuthology.orchestra.run.vm08.stdout:(124/136): python3-rsa-4.9-2.el9.noarch.rpm 19 MB/s | 59 kB 00:00 2026-03-09T20:49:46.872 INFO:teuthology.orchestra.run.vm08.stdout:(125/136): python3-tempora-5.0.0-2.el9.noarch.r 11 MB/s | 36 kB 00:00 2026-03-09T20:49:46.873 INFO:teuthology.orchestra.run.vm08.stdout:(126/136): python3-typing-extensions-4.15.0-1.e 21 MB/s | 86 kB 00:00 2026-03-09T20:49:46.878 INFO:teuthology.orchestra.run.vm08.stdout:(127/136): python3-webob-1.8.8-2.el9.noarch.rpm 37 MB/s | 230 kB 00:00 2026-03-09T20:49:46.881 INFO:teuthology.orchestra.run.vm08.stdout:(128/136): python3-websocket-client-1.2.3-2.el9 12 MB/s | 90 kB 00:00 2026-03-09T20:49:46.887 INFO:teuthology.orchestra.run.vm08.stdout:(129/136): python3-werkzeug-2.0.3-3.el9.1.noarc 52 MB/s | 427 kB 00:00 2026-03-09T20:49:46.889 INFO:teuthology.orchestra.run.vm08.stdout:(130/136): python3-xmltodict-0.12.0-15.el9.noar 3.0 MB/s | 22 kB 00:00 2026-03-09T20:49:46.893 INFO:teuthology.orchestra.run.vm08.stdout:(131/136): python3-zc-lockfile-2.0-10.el9.noarc 3.4 MB/s | 20 kB 00:00 2026-03-09T20:49:46.896 INFO:teuthology.orchestra.run.vm08.stdout:(132/136): re2-20211101-20.el9.x86_64.rpm 28 MB/s | 191 kB 00:00 2026-03-09T20:49:46.911 INFO:teuthology.orchestra.run.vm06.stdout:(58/136): flexiblas-netlib-3.0.4-9.el9.x86_64.r 3.8 MB/s | 3.0 MB 00:00 2026-03-09T20:49:46.935 INFO:teuthology.orchestra.run.vm08.stdout:(133/136): thrift-0.15.0-4.el9.x86_64.rpm 38 MB/s | 1.6 MB 00:00 2026-03-09T20:49:46.967 INFO:teuthology.orchestra.run.vm06.stdout:(59/136): libstoragemgmt-1.10.1-1.el9.x86_64.rp 1.6 MB/s | 246 kB 00:00 2026-03-09T20:49:47.020 INFO:teuthology.orchestra.run.vm06.stdout:(60/136): libxslt-1.1.34-12.el9.x86_64.rpm 2.1 MB/s | 233 kB 00:00 2026-03-09T20:49:47.048 INFO:teuthology.orchestra.run.vm06.stdout:(61/136): librdkafka-1.6.1-102.el9.x86_64.rpm 2.8 MB/s | 662 kB 00:00 2026-03-09T20:49:47.097 INFO:teuthology.orchestra.run.vm06.stdout:(62/136): lttng-ust-2.12.0-6.el9.x86_64.rpm 2.2 MB/s | 292 kB 00:00 2026-03-09T20:49:47.120 INFO:teuthology.orchestra.run.vm06.stdout:(63/136): lua-5.4.4-4.el9.x86_64.rpm 1.8 MB/s | 188 kB 00:00 2026-03-09T20:49:47.144 INFO:teuthology.orchestra.run.vm06.stdout:(64/136): openblas-0.3.29-1.el9.x86_64.rpm 443 kB/s | 42 kB 00:00 2026-03-09T20:49:47.196 INFO:teuthology.orchestra.run.vm01.stdout:(29/136): ceph-mgr-dashboard-19.2.3-678.ge911bd 7.5 MB/s | 3.8 MB 00:00 2026-03-09T20:49:47.290 INFO:teuthology.orchestra.run.vm06.stdout:(65/136): protobuf-3.14.0-17.el9.x86_64.rpm 5.9 MB/s | 1.0 MB 00:00 2026-03-09T20:49:47.320 INFO:teuthology.orchestra.run.vm01.stdout:(30/136): ceph-mgr-modules-core-19.2.3-678.ge91 2.0 MB/s | 253 kB 00:00 2026-03-09T20:49:47.405 INFO:teuthology.orchestra.run.vm06.stdout:(66/136): python3-devel-3.9.25-3.el9.x86_64.rpm 2.1 MB/s | 244 kB 00:00 2026-03-09T20:49:47.445 INFO:teuthology.orchestra.run.vm01.stdout:(31/136): ceph-mgr-rook-19.2.3-678.ge911bdeb.el 397 kB/s | 49 kB 00:00 2026-03-09T20:49:47.505 INFO:teuthology.orchestra.run.vm06.stdout:(67/136): python3-jinja2-2.11.3-8.el9.noarch.rp 2.4 MB/s | 249 kB 00:00 2026-03-09T20:49:47.566 INFO:teuthology.orchestra.run.vm01.stdout:(32/136): ceph-prometheus-alerts-19.2.3-678.ge9 138 kB/s | 17 kB 00:00 2026-03-09T20:49:47.567 INFO:teuthology.orchestra.run.vm06.stdout:(68/136): python3-babel-2.9.1-2.el9.noarch.rpm 14 MB/s | 6.0 MB 00:00 2026-03-09T20:49:47.606 INFO:teuthology.orchestra.run.vm06.stdout:(69/136): python3-jmespath-1.0.1-1.el9.noarch.r 473 kB/s | 48 kB 00:00 2026-03-09T20:49:47.659 INFO:teuthology.orchestra.run.vm06.stdout:(70/136): python3-libstoragemgmt-1.10.1-1.el9.x 1.9 MB/s | 177 kB 00:00 2026-03-09T20:49:47.686 INFO:teuthology.orchestra.run.vm06.stdout:(71/136): python3-mako-1.1.4-6.el9.noarch.rpm 2.1 MB/s | 172 kB 00:00 2026-03-09T20:49:47.690 INFO:teuthology.orchestra.run.vm01.stdout:(33/136): ceph-volume-19.2.3-678.ge911bdeb.el9. 2.4 MB/s | 299 kB 00:00 2026-03-09T20:49:47.743 INFO:teuthology.orchestra.run.vm06.stdout:(72/136): python3-markupsafe-1.1.1-12.el9.x86_6 419 kB/s | 35 kB 00:00 2026-03-09T20:49:47.810 INFO:teuthology.orchestra.run.vm01.stdout:(34/136): ceph-mgr-diskprediction-local-19.2.3- 7.4 MB/s | 7.4 MB 00:00 2026-03-09T20:49:47.904 INFO:teuthology.orchestra.run.vm06.stdout:(73/136): python3-numpy-f2py-1.23.5-2.el9.x86_6 2.7 MB/s | 442 kB 00:00 2026-03-09T20:49:47.989 INFO:teuthology.orchestra.run.vm08.stdout:(134/136): python3-scipy-1.9.3-2.el9.x86_64.rpm 9.4 MB/s | 19 MB 00:02 2026-03-09T20:49:47.992 INFO:teuthology.orchestra.run.vm06.stdout:(74/136): python3-packaging-20.9-5.el9.noarch.r 879 kB/s | 77 kB 00:00 2026-03-09T20:49:48.002 INFO:teuthology.orchestra.run.vm08.stdout:(135/136): librados2-19.2.3-678.ge911bdeb.el9.x 3.1 MB/s | 3.4 MB 00:01 2026-03-09T20:49:48.015 INFO:teuthology.orchestra.run.vm08.stdout:(136/136): librbd1-19.2.3-678.ge911bdeb.el9.x86 2.9 MB/s | 3.2 MB 00:01 2026-03-09T20:49:48.018 INFO:teuthology.orchestra.run.vm08.stdout:-------------------------------------------------------------------------------- 2026-03-09T20:49:48.018 INFO:teuthology.orchestra.run.vm08.stdout:Total 19 MB/s | 210 MB 00:11 2026-03-09T20:49:48.025 INFO:teuthology.orchestra.run.vm06.stdout:(75/136): openblas-openmp-0.3.29-1.el9.x86_64.r 5.7 MB/s | 5.3 MB 00:00 2026-03-09T20:49:48.034 INFO:teuthology.orchestra.run.vm01.stdout:(35/136): cryptsetup-2.8.1-3.el9.x86_64.rpm 1.5 MB/s | 351 kB 00:00 2026-03-09T20:49:48.083 INFO:teuthology.orchestra.run.vm01.stdout:(36/136): ledmon-libs-1.1.0-3.el9.x86_64.rpm 833 kB/s | 40 kB 00:00 2026-03-09T20:49:48.101 INFO:teuthology.orchestra.run.vm06.stdout:(76/136): python3-protobuf-3.14.0-17.el9.noarch 2.4 MB/s | 267 kB 00:00 2026-03-09T20:49:48.125 INFO:teuthology.orchestra.run.vm01.stdout:(37/136): libconfig-1.7.2-9.el9.x86_64.rpm 1.7 MB/s | 72 kB 00:00 2026-03-09T20:49:48.125 INFO:teuthology.orchestra.run.vm06.stdout:(77/136): python3-numpy-1.23.5-2.el9.x86_64.rpm 14 MB/s | 6.1 MB 00:00 2026-03-09T20:49:48.127 INFO:teuthology.orchestra.run.vm06.stdout:(78/136): python3-pyasn1-0.4.8-7.el9.noarch.rpm 1.5 MB/s | 157 kB 00:00 2026-03-09T20:49:48.204 INFO:teuthology.orchestra.run.vm06.stdout:(79/136): python3-pyasn1-modules-0.4.8-7.el9.no 2.6 MB/s | 277 kB 00:00 2026-03-09T20:49:48.204 INFO:teuthology.orchestra.run.vm06.stdout:(80/136): python3-requests-oauthlib-1.3.0-12.el 679 kB/s | 54 kB 00:00 2026-03-09T20:49:48.207 INFO:teuthology.orchestra.run.vm01.stdout:(38/136): libgfortran-11.5.0-14.el9.x86_64.rpm 9.5 MB/s | 794 kB 00:00 2026-03-09T20:49:48.256 INFO:teuthology.orchestra.run.vm01.stdout:(39/136): libquadmath-11.5.0-14.el9.x86_64.rpm 3.7 MB/s | 184 kB 00:00 2026-03-09T20:49:48.280 INFO:teuthology.orchestra.run.vm06.stdout:(81/136): python3-toml-0.10.2-6.el9.noarch.rpm 547 kB/s | 42 kB 00:00 2026-03-09T20:49:48.290 INFO:teuthology.orchestra.run.vm01.stdout:(40/136): mailcap-2.1.49-5.el9.noarch.rpm 1.0 MB/s | 33 kB 00:00 2026-03-09T20:49:48.311 INFO:teuthology.orchestra.run.vm06.stdout:(82/136): qatlib-25.08.0-2.el9.x86_64.rpm 2.2 MB/s | 240 kB 00:00 2026-03-09T20:49:48.332 INFO:teuthology.orchestra.run.vm01.stdout:(41/136): pciutils-3.7.0-7.el9.x86_64.rpm 2.2 MB/s | 93 kB 00:00 2026-03-09T20:49:48.361 INFO:teuthology.orchestra.run.vm06.stdout:(83/136): qatlib-service-25.08.0-2.el9.x86_64.r 460 kB/s | 37 kB 00:00 2026-03-09T20:49:48.380 INFO:teuthology.orchestra.run.vm01.stdout:(42/136): python3-cffi-1.14.5-5.el9.x86_64.rpm 5.2 MB/s | 253 kB 00:00 2026-03-09T20:49:48.403 INFO:teuthology.orchestra.run.vm06.stdout:(84/136): qatzip-libs-1.3.1-1.el9.x86_64.rpm 725 kB/s | 66 kB 00:00 2026-03-09T20:49:48.485 INFO:teuthology.orchestra.run.vm06.stdout:(85/136): xmlstarlet-1.6.1-20.el9.x86_64.rpm 783 kB/s | 64 kB 00:00 2026-03-09T20:49:48.492 INFO:teuthology.orchestra.run.vm06.stdout:(86/136): socat-1.7.4.1-8.el9.x86_64.rpm 2.3 MB/s | 303 kB 00:00 2026-03-09T20:49:48.524 INFO:teuthology.orchestra.run.vm01.stdout:(43/136): python3-cryptography-36.0.1-5.el9.x86 8.7 MB/s | 1.2 MB 00:00 2026-03-09T20:49:48.568 INFO:teuthology.orchestra.run.vm01.stdout:(44/136): python3-ply-3.11-14.el9.noarch.rpm 2.3 MB/s | 106 kB 00:00 2026-03-09T20:49:48.612 INFO:teuthology.orchestra.run.vm01.stdout:(45/136): python3-pycparser-2.20-6.el9.noarch.r 3.1 MB/s | 135 kB 00:00 2026-03-09T20:49:48.613 INFO:teuthology.orchestra.run.vm06.stdout:(87/136): lua-devel-5.4.4-4.el9.x86_64.rpm 173 kB/s | 22 kB 00:00 2026-03-09T20:49:48.634 INFO:teuthology.orchestra.run.vm06.stdout:(88/136): abseil-cpp-20211102.0-4.el9.x86_64.rp 27 MB/s | 551 kB 00:00 2026-03-09T20:49:48.641 INFO:teuthology.orchestra.run.vm06.stdout:(89/136): gperftools-libs-2.9.1-3.el9.x86_64.rp 48 MB/s | 308 kB 00:00 2026-03-09T20:49:48.643 INFO:teuthology.orchestra.run.vm06.stdout:(90/136): grpc-data-1.46.7-10.el9.noarch.rpm 8.4 MB/s | 19 kB 00:00 2026-03-09T20:49:48.649 INFO:teuthology.orchestra.run.vm08.stdout:Running transaction check 2026-03-09T20:49:48.654 INFO:teuthology.orchestra.run.vm01.stdout:(46/136): python3-requests-2.25.1-10.el9.noarch 2.9 MB/s | 126 kB 00:00 2026-03-09T20:49:48.704 INFO:teuthology.orchestra.run.vm08.stdout:Transaction check succeeded. 2026-03-09T20:49:48.704 INFO:teuthology.orchestra.run.vm08.stdout:Running transaction test 2026-03-09T20:49:48.723 INFO:teuthology.orchestra.run.vm06.stdout:(91/136): libarrow-9.0.0-15.el9.x86_64.rpm 56 MB/s | 4.4 MB 00:00 2026-03-09T20:49:48.726 INFO:teuthology.orchestra.run.vm06.stdout:(92/136): libarrow-doc-9.0.0-15.el9.noarch.rpm 10 MB/s | 25 kB 00:00 2026-03-09T20:49:48.729 INFO:teuthology.orchestra.run.vm06.stdout:(93/136): liboath-2.6.12-1.el9.x86_64.rpm 18 MB/s | 49 kB 00:00 2026-03-09T20:49:48.732 INFO:teuthology.orchestra.run.vm06.stdout:(94/136): libunwind-1.6.2-1.el9.x86_64.rpm 23 MB/s | 67 kB 00:00 2026-03-09T20:49:48.735 INFO:teuthology.orchestra.run.vm06.stdout:(95/136): protobuf-compiler-3.14.0-17.el9.x86_6 3.5 MB/s | 862 kB 00:00 2026-03-09T20:49:48.737 INFO:teuthology.orchestra.run.vm06.stdout:(96/136): luarocks-3.9.2-5.el9.noarch.rpm 31 MB/s | 151 kB 00:00 2026-03-09T20:49:48.751 INFO:teuthology.orchestra.run.vm06.stdout:(97/136): python3-asyncssh-2.13.2-5.el9.noarch. 39 MB/s | 548 kB 00:00 2026-03-09T20:49:48.755 INFO:teuthology.orchestra.run.vm06.stdout:(98/136): python3-autocommand-2.2.2-8.el9.noarc 8.3 MB/s | 29 kB 00:00 2026-03-09T20:49:48.758 INFO:teuthology.orchestra.run.vm06.stdout:(99/136): python3-backports-tarfile-1.2.0-1.el9 23 MB/s | 60 kB 00:00 2026-03-09T20:49:48.760 INFO:teuthology.orchestra.run.vm06.stdout:(100/136): parquet-libs-9.0.0-15.el9.x86_64.rpm 33 MB/s | 838 kB 00:00 2026-03-09T20:49:48.761 INFO:teuthology.orchestra.run.vm06.stdout:(101/136): python3-bcrypt-3.2.2-1.el9.x86_64.rp 12 MB/s | 43 kB 00:00 2026-03-09T20:49:48.762 INFO:teuthology.orchestra.run.vm06.stdout:(102/136): python3-cachetools-4.2.4-1.el9.noarc 14 MB/s | 32 kB 00:00 2026-03-09T20:49:48.763 INFO:teuthology.orchestra.run.vm06.stdout:(103/136): python3-certifi-2023.05.07-4.el9.noa 7.2 MB/s | 14 kB 00:00 2026-03-09T20:49:48.768 INFO:teuthology.orchestra.run.vm06.stdout:(104/136): python3-cheroot-10.0.1-4.el9.noarch. 32 MB/s | 173 kB 00:00 2026-03-09T20:49:48.771 INFO:teuthology.orchestra.run.vm06.stdout:(105/136): python3-cherrypy-18.6.1-2.el9.noarch 48 MB/s | 358 kB 00:00 2026-03-09T20:49:48.773 INFO:teuthology.orchestra.run.vm06.stdout:(106/136): python3-google-auth-2.45.0-1.el9.noa 52 MB/s | 254 kB 00:00 2026-03-09T20:49:48.778 INFO:teuthology.orchestra.run.vm06.stdout:(107/136): python3-grpcio-tools-1.46.7-10.el9.x 29 MB/s | 144 kB 00:00 2026-03-09T20:49:48.781 INFO:teuthology.orchestra.run.vm06.stdout:(108/136): python3-jaraco-8.2.1-3.el9.noarch.rp 3.4 MB/s | 11 kB 00:00 2026-03-09T20:49:48.785 INFO:teuthology.orchestra.run.vm06.stdout:(109/136): python3-jaraco-classes-3.2.1-5.el9.n 5.5 MB/s | 18 kB 00:00 2026-03-09T20:49:48.788 INFO:teuthology.orchestra.run.vm06.stdout:(110/136): python3-jaraco-collections-3.0.0-8.e 7.2 MB/s | 23 kB 00:00 2026-03-09T20:49:48.792 INFO:teuthology.orchestra.run.vm06.stdout:(111/136): python3-jaraco-context-6.0.1-3.el9.n 6.4 MB/s | 20 kB 00:00 2026-03-09T20:49:48.795 INFO:teuthology.orchestra.run.vm06.stdout:(112/136): python3-jaraco-functools-3.5.0-2.el9 6.1 MB/s | 19 kB 00:00 2026-03-09T20:49:48.804 INFO:teuthology.orchestra.run.vm06.stdout:(113/136): python3-grpcio-1.46.7-10.el9.x86_64. 61 MB/s | 2.0 MB 00:00 2026-03-09T20:49:48.805 INFO:teuthology.orchestra.run.vm06.stdout:(114/136): python3-jaraco-text-4.0.0-2.el9.noar 2.8 MB/s | 26 kB 00:00 2026-03-09T20:49:48.805 INFO:teuthology.orchestra.run.vm01.stdout:(47/136): cephadm-19.2.3-678.ge911bdeb.el9.noar 690 kB/s | 769 kB 00:01 2026-03-09T20:49:48.808 INFO:teuthology.orchestra.run.vm06.stdout:(115/136): python3-logutils-0.3.5-21.el9.noarch 15 MB/s | 46 kB 00:00 2026-03-09T20:49:48.812 INFO:teuthology.orchestra.run.vm06.stdout:(116/136): python3-more-itertools-8.12.0-2.el9. 21 MB/s | 79 kB 00:00 2026-03-09T20:49:48.815 INFO:teuthology.orchestra.run.vm06.stdout:(117/136): python3-natsort-7.1.1-5.el9.noarch.r 18 MB/s | 58 kB 00:00 2026-03-09T20:49:48.823 INFO:teuthology.orchestra.run.vm06.stdout:(118/136): python3-kubernetes-26.1.0-3.el9.noar 56 MB/s | 1.0 MB 00:00 2026-03-09T20:49:48.825 INFO:teuthology.orchestra.run.vm06.stdout:(119/136): python3-pecan-1.4.2-3.el9.noarch.rpm 29 MB/s | 272 kB 00:00 2026-03-09T20:49:48.825 INFO:teuthology.orchestra.run.vm06.stdout:(120/136): python3-portend-3.1.0-2.el9.noarch.r 8.0 MB/s | 16 kB 00:00 2026-03-09T20:49:48.828 INFO:teuthology.orchestra.run.vm06.stdout:(121/136): python3-pyOpenSSL-21.0.0-1.el9.noarc 27 MB/s | 90 kB 00:00 2026-03-09T20:49:48.830 INFO:teuthology.orchestra.run.vm06.stdout:(122/136): python3-repoze-lru-0.7-16.el9.noarch 7.5 MB/s | 31 kB 00:00 2026-03-09T20:49:48.833 INFO:teuthology.orchestra.run.vm06.stdout:(123/136): python3-rsa-4.9-2.el9.noarch.rpm 16 MB/s | 59 kB 00:00 2026-03-09T20:49:48.835 INFO:teuthology.orchestra.run.vm06.stdout:(124/136): python3-routes-2.5.1-5.el9.noarch.rp 30 MB/s | 188 kB 00:00 2026-03-09T20:49:48.836 INFO:teuthology.orchestra.run.vm06.stdout:(125/136): python3-tempora-5.0.0-2.el9.noarch.r 14 MB/s | 36 kB 00:00 2026-03-09T20:49:48.838 INFO:teuthology.orchestra.run.vm06.stdout:(126/136): python3-typing-extensions-4.15.0-1.e 29 MB/s | 86 kB 00:00 2026-03-09T20:49:48.842 INFO:teuthology.orchestra.run.vm06.stdout:(127/136): python3-websocket-client-1.2.3-2.el9 24 MB/s | 90 kB 00:00 2026-03-09T20:49:48.843 INFO:teuthology.orchestra.run.vm06.stdout:(128/136): python3-webob-1.8.8-2.el9.noarch.rpm 35 MB/s | 230 kB 00:00 2026-03-09T20:49:48.847 INFO:teuthology.orchestra.run.vm06.stdout:(129/136): python3-xmltodict-0.12.0-15.el9.noar 5.9 MB/s | 22 kB 00:00 2026-03-09T20:49:48.851 INFO:teuthology.orchestra.run.vm06.stdout:(130/136): python3-werkzeug-2.0.3-3.el9.1.noarc 48 MB/s | 427 kB 00:00 2026-03-09T20:49:48.851 INFO:teuthology.orchestra.run.vm06.stdout:(131/136): python3-zc-lockfile-2.0-10.el9.noarc 4.4 MB/s | 20 kB 00:00 2026-03-09T20:49:48.856 INFO:teuthology.orchestra.run.vm06.stdout:(132/136): re2-20211101-20.el9.x86_64.rpm 41 MB/s | 191 kB 00:00 2026-03-09T20:49:48.879 INFO:teuthology.orchestra.run.vm06.stdout:(133/136): thrift-0.15.0-4.el9.x86_64.rpm 58 MB/s | 1.6 MB 00:00 2026-03-09T20:49:48.945 INFO:teuthology.orchestra.run.vm01.stdout:(48/136): python3-urllib3-1.26.5-7.el9.noarch.r 750 kB/s | 218 kB 00:00 2026-03-09T20:49:48.979 INFO:teuthology.orchestra.run.vm01.stdout:(49/136): unzip-6.0-59.el9.x86_64.rpm 1.0 MB/s | 182 kB 00:00 2026-03-09T20:49:49.005 INFO:teuthology.orchestra.run.vm01.stdout:(50/136): zip-3.0-35.el9.x86_64.rpm 4.3 MB/s | 266 kB 00:00 2026-03-09T20:49:49.042 INFO:teuthology.orchestra.run.vm06.stdout:(134/136): python3-scipy-1.9.3-2.el9.x86_64.rpm 21 MB/s | 19 MB 00:00 2026-03-09T20:49:49.124 INFO:teuthology.orchestra.run.vm01.stdout:(51/136): flexiblas-3.0.4-9.el9.x86_64.rpm 250 kB/s | 30 kB 00:00 2026-03-09T20:49:49.203 INFO:teuthology.orchestra.run.vm01.stdout:(52/136): boost-program-options-1.75.0-13.el9.x 464 kB/s | 104 kB 00:00 2026-03-09T20:49:49.271 INFO:teuthology.orchestra.run.vm01.stdout:(53/136): flexiblas-openblas-openmp-3.0.4-9.el9 219 kB/s | 15 kB 00:00 2026-03-09T20:49:49.460 INFO:teuthology.orchestra.run.vm01.stdout:(54/136): flexiblas-netlib-3.0.4-9.el9.x86_64.r 8.9 MB/s | 3.0 MB 00:00 2026-03-09T20:49:49.578 INFO:teuthology.orchestra.run.vm01.stdout:(55/136): libpmemobj-1.12.1-1.el9.x86_64.rpm 1.3 MB/s | 160 kB 00:00 2026-03-09T20:49:49.587 INFO:teuthology.orchestra.run.vm08.stdout:Transaction test succeeded. 2026-03-09T20:49:49.587 INFO:teuthology.orchestra.run.vm08.stdout:Running transaction 2026-03-09T20:49:49.613 INFO:teuthology.orchestra.run.vm01.stdout:(56/136): libnbd-1.20.3-4.el9.x86_64.rpm 479 kB/s | 164 kB 00:00 2026-03-09T20:49:49.640 INFO:teuthology.orchestra.run.vm01.stdout:(57/136): librabbitmq-0.11.0-7.el9.x86_64.rpm 736 kB/s | 45 kB 00:00 2026-03-09T20:49:49.743 INFO:teuthology.orchestra.run.vm01.stdout:(58/136): libstoragemgmt-1.10.1-1.el9.x86_64.rp 2.3 MB/s | 246 kB 00:00 2026-03-09T20:49:50.096 INFO:teuthology.orchestra.run.vm06.stdout:(135/136): librbd1-19.2.3-678.ge911bdeb.el9.x86 2.6 MB/s | 3.2 MB 00:01 2026-03-09T20:49:50.162 INFO:teuthology.orchestra.run.vm01.stdout:(59/136): librdkafka-1.6.1-102.el9.x86_64.rpm 1.2 MB/s | 662 kB 00:00 2026-03-09T20:49:50.166 INFO:teuthology.orchestra.run.vm01.stdout:(60/136): libxslt-1.1.34-12.el9.x86_64.rpm 551 kB/s | 233 kB 00:00 2026-03-09T20:49:50.286 INFO:teuthology.orchestra.run.vm01.stdout:(61/136): lua-5.4.4-4.el9.x86_64.rpm 1.5 MB/s | 188 kB 00:00 2026-03-09T20:49:50.293 INFO:teuthology.orchestra.run.vm06.stdout:(136/136): librados2-19.2.3-678.ge911bdeb.el9.x 2.4 MB/s | 3.4 MB 00:01 2026-03-09T20:49:50.298 INFO:teuthology.orchestra.run.vm06.stdout:-------------------------------------------------------------------------------- 2026-03-09T20:49:50.298 INFO:teuthology.orchestra.run.vm06.stdout:Total 14 MB/s | 210 MB 00:14 2026-03-09T20:49:50.343 INFO:teuthology.orchestra.run.vm01.stdout:(62/136): openblas-0.3.29-1.el9.x86_64.rpm 731 kB/s | 42 kB 00:00 2026-03-09T20:49:50.410 INFO:teuthology.orchestra.run.vm01.stdout:(63/136): lttng-ust-2.12.0-6.el9.x86_64.rpm 1.2 MB/s | 292 kB 00:00 2026-03-09T20:49:50.546 INFO:teuthology.orchestra.run.vm08.stdout: Preparing : 1/1 2026-03-09T20:49:50.563 INFO:teuthology.orchestra.run.vm08.stdout: Installing : python3-more-itertools-8.12.0-2.el9.noarch 1/138 2026-03-09T20:49:50.578 INFO:teuthology.orchestra.run.vm08.stdout: Installing : thrift-0.15.0-4.el9.x86_64 2/138 2026-03-09T20:49:50.761 INFO:teuthology.orchestra.run.vm08.stdout: Installing : lttng-ust-2.12.0-6.el9.x86_64 3/138 2026-03-09T20:49:50.763 INFO:teuthology.orchestra.run.vm08.stdout: Upgrading : librados2-2:19.2.3-678.ge911bdeb.el9.x86_64 4/138 2026-03-09T20:49:50.829 INFO:teuthology.orchestra.run.vm08.stdout: Running scriptlet: librados2-2:19.2.3-678.ge911bdeb.el9.x86_64 4/138 2026-03-09T20:49:50.837 INFO:teuthology.orchestra.run.vm08.stdout: Installing : libcephfs2-2:19.2.3-678.ge911bdeb.el9.x86_64 5/138 2026-03-09T20:49:50.872 INFO:teuthology.orchestra.run.vm08.stdout: Running scriptlet: libcephfs2-2:19.2.3-678.ge911bdeb.el9.x86_64 5/138 2026-03-09T20:49:50.883 INFO:teuthology.orchestra.run.vm08.stdout: Installing : python3-rados-2:19.2.3-678.ge911bdeb.el9.x86_64 6/138 2026-03-09T20:49:50.887 INFO:teuthology.orchestra.run.vm08.stdout: Installing : librdkafka-1.6.1-102.el9.x86_64 7/138 2026-03-09T20:49:50.891 INFO:teuthology.orchestra.run.vm08.stdout: Installing : librabbitmq-0.11.0-7.el9.x86_64 8/138 2026-03-09T20:49:50.897 INFO:teuthology.orchestra.run.vm08.stdout: Installing : python3-jaraco-8.2.1-3.el9.noarch 9/138 2026-03-09T20:49:50.909 INFO:teuthology.orchestra.run.vm08.stdout: Installing : libnbd-1.20.3-4.el9.x86_64 10/138 2026-03-09T20:49:50.911 INFO:teuthology.orchestra.run.vm08.stdout: Installing : libcephsqlite-2:19.2.3-678.ge911bdeb.el9.x86_64 11/138 2026-03-09T20:49:50.952 INFO:teuthology.orchestra.run.vm08.stdout: Running scriptlet: libcephsqlite-2:19.2.3-678.ge911bdeb.el9.x86_64 11/138 2026-03-09T20:49:50.955 INFO:teuthology.orchestra.run.vm08.stdout: Installing : libradosstriper1-2:19.2.3-678.ge911bdeb.el9.x86_ 12/138 2026-03-09T20:49:50.972 INFO:teuthology.orchestra.run.vm06.stdout:Running transaction check 2026-03-09T20:49:50.973 INFO:teuthology.orchestra.run.vm08.stdout: Running scriptlet: libradosstriper1-2:19.2.3-678.ge911bdeb.el9.x86_ 12/138 2026-03-09T20:49:51.013 INFO:teuthology.orchestra.run.vm08.stdout: Installing : re2-1:20211101-20.el9.x86_64 13/138 2026-03-09T20:49:51.023 INFO:teuthology.orchestra.run.vm06.stdout:Transaction check succeeded. 2026-03-09T20:49:51.023 INFO:teuthology.orchestra.run.vm06.stdout:Running transaction test 2026-03-09T20:49:51.076 INFO:teuthology.orchestra.run.vm08.stdout: Installing : libarrow-9.0.0-15.el9.x86_64 14/138 2026-03-09T20:49:51.082 INFO:teuthology.orchestra.run.vm08.stdout: Installing : python3-werkzeug-2.0.3-3.el9.1.noarch 15/138 2026-03-09T20:49:51.110 INFO:teuthology.orchestra.run.vm08.stdout: Installing : liboath-2.6.12-1.el9.x86_64 16/138 2026-03-09T20:49:51.127 INFO:teuthology.orchestra.run.vm08.stdout: Installing : python3-pyasn1-0.4.8-7.el9.noarch 17/138 2026-03-09T20:49:51.136 INFO:teuthology.orchestra.run.vm08.stdout: Installing : python3-packaging-20.9-5.el9.noarch 18/138 2026-03-09T20:49:51.147 INFO:teuthology.orchestra.run.vm08.stdout: Installing : python3-markupsafe-1.1.1-12.el9.x86_64 19/138 2026-03-09T20:49:51.155 INFO:teuthology.orchestra.run.vm08.stdout: Installing : protobuf-3.14.0-17.el9.x86_64 20/138 2026-03-09T20:49:51.159 INFO:teuthology.orchestra.run.vm08.stdout: Installing : lua-5.4.4-4.el9.x86_64 21/138 2026-03-09T20:49:51.169 INFO:teuthology.orchestra.run.vm08.stdout: Installing : flexiblas-3.0.4-9.el9.x86_64 22/138 2026-03-09T20:49:51.204 INFO:teuthology.orchestra.run.vm08.stdout: Installing : unzip-6.0-59.el9.x86_64 23/138 2026-03-09T20:49:51.223 INFO:teuthology.orchestra.run.vm08.stdout: Installing : python3-urllib3-1.26.5-7.el9.noarch 24/138 2026-03-09T20:49:51.227 INFO:teuthology.orchestra.run.vm08.stdout: Installing : python3-requests-2.25.1-10.el9.noarch 25/138 2026-03-09T20:49:51.235 INFO:teuthology.orchestra.run.vm08.stdout: Installing : libquadmath-11.5.0-14.el9.x86_64 26/138 2026-03-09T20:49:51.237 INFO:teuthology.orchestra.run.vm08.stdout: Installing : libgfortran-11.5.0-14.el9.x86_64 27/138 2026-03-09T20:49:51.271 INFO:teuthology.orchestra.run.vm08.stdout: Installing : ledmon-libs-1.1.0-3.el9.x86_64 28/138 2026-03-09T20:49:51.278 INFO:teuthology.orchestra.run.vm08.stdout: Installing : python3-ceph-common-2:19.2.3-678.ge911bdeb.el9.x 29/138 2026-03-09T20:49:51.288 INFO:teuthology.orchestra.run.vm08.stdout: Installing : python3-ceph-argparse-2:19.2.3-678.ge911bdeb.el9 30/138 2026-03-09T20:49:51.303 INFO:teuthology.orchestra.run.vm08.stdout: Installing : python3-cephfs-2:19.2.3-678.ge911bdeb.el9.x86_64 31/138 2026-03-09T20:49:51.312 INFO:teuthology.orchestra.run.vm08.stdout: Installing : python3-requests-oauthlib-1.3.0-12.el9.noarch 32/138 2026-03-09T20:49:51.343 INFO:teuthology.orchestra.run.vm08.stdout: Installing : zip-3.0-35.el9.x86_64 33/138 2026-03-09T20:49:51.349 INFO:teuthology.orchestra.run.vm08.stdout: Installing : luarocks-3.9.2-5.el9.noarch 34/138 2026-03-09T20:49:51.368 INFO:teuthology.orchestra.run.vm08.stdout: Installing : lua-devel-5.4.4-4.el9.x86_64 35/138 2026-03-09T20:49:51.472 INFO:teuthology.orchestra.run.vm08.stdout: Installing : protobuf-compiler-3.14.0-17.el9.x86_64 36/138 2026-03-09T20:49:51.540 INFO:teuthology.orchestra.run.vm08.stdout: Installing : python3-mako-1.1.4-6.el9.noarch 37/138 2026-03-09T20:49:51.558 INFO:teuthology.orchestra.run.vm08.stdout: Installing : python3-pyasn1-modules-0.4.8-7.el9.noarch 38/138 2026-03-09T20:49:51.567 INFO:teuthology.orchestra.run.vm08.stdout: Installing : python3-rsa-4.9-2.el9.noarch 39/138 2026-03-09T20:49:51.578 INFO:teuthology.orchestra.run.vm08.stdout: Installing : python3-jaraco-classes-3.2.1-5.el9.noarch 40/138 2026-03-09T20:49:51.585 INFO:teuthology.orchestra.run.vm08.stdout: Installing : librados-devel-2:19.2.3-678.ge911bdeb.el9.x86_64 41/138 2026-03-09T20:49:51.590 INFO:teuthology.orchestra.run.vm08.stdout: Installing : python3-zc-lockfile-2.0-10.el9.noarch 42/138 2026-03-09T20:49:51.609 INFO:teuthology.orchestra.run.vm08.stdout: Installing : python3-xmltodict-0.12.0-15.el9.noarch 43/138 2026-03-09T20:49:51.637 INFO:teuthology.orchestra.run.vm08.stdout: Installing : python3-websocket-client-1.2.3-2.el9.noarch 44/138 2026-03-09T20:49:51.646 INFO:teuthology.orchestra.run.vm08.stdout: Installing : python3-webob-1.8.8-2.el9.noarch 45/138 2026-03-09T20:49:51.655 INFO:teuthology.orchestra.run.vm08.stdout: Installing : python3-typing-extensions-4.15.0-1.el9.noarch 46/138 2026-03-09T20:49:51.672 INFO:teuthology.orchestra.run.vm08.stdout: Installing : python3-repoze-lru-0.7-16.el9.noarch 47/138 2026-03-09T20:49:51.730 INFO:teuthology.orchestra.run.vm08.stdout: Installing : python3-routes-2.5.1-5.el9.noarch 48/138 2026-03-09T20:49:51.745 INFO:teuthology.orchestra.run.vm01.stdout:(64/136): ceph-test-19.2.3-678.ge911bdeb.el9.x8 7.2 MB/s | 50 MB 00:06 2026-03-09T20:49:51.756 INFO:teuthology.orchestra.run.vm08.stdout: Installing : python3-natsort-7.1.1-5.el9.noarch 49/138 2026-03-09T20:49:51.822 INFO:teuthology.orchestra.run.vm08.stdout: Installing : python3-logutils-0.3.5-21.el9.noarch 50/138 2026-03-09T20:49:51.831 INFO:teuthology.orchestra.run.vm08.stdout: Installing : python3-pecan-1.4.2-3.el9.noarch 51/138 2026-03-09T20:49:51.843 INFO:teuthology.orchestra.run.vm08.stdout: Installing : python3-certifi-2023.05.07-4.el9.noarch 52/138 2026-03-09T20:49:51.893 INFO:teuthology.orchestra.run.vm06.stdout:Transaction test succeeded. 2026-03-09T20:49:51.893 INFO:teuthology.orchestra.run.vm06.stdout:Running transaction 2026-03-09T20:49:51.894 INFO:teuthology.orchestra.run.vm08.stdout: Installing : python3-cachetools-4.2.4-1.el9.noarch 53/138 2026-03-09T20:49:52.312 INFO:teuthology.orchestra.run.vm01.stdout:(65/136): python3-babel-2.9.1-2.el9.noarch.rpm 11 MB/s | 6.0 MB 00:00 2026-03-09T20:49:52.320 INFO:teuthology.orchestra.run.vm08.stdout: Installing : python3-google-auth-1:2.45.0-1.el9.noarch 54/138 2026-03-09T20:49:52.338 INFO:teuthology.orchestra.run.vm08.stdout: Installing : python3-kubernetes-1:26.1.0-3.el9.noarch 55/138 2026-03-09T20:49:52.345 INFO:teuthology.orchestra.run.vm08.stdout: Installing : python3-backports-tarfile-1.2.0-1.el9.noarch 56/138 2026-03-09T20:49:52.353 INFO:teuthology.orchestra.run.vm08.stdout: Installing : python3-jaraco-context-6.0.1-3.el9.noarch 57/138 2026-03-09T20:49:52.361 INFO:teuthology.orchestra.run.vm08.stdout: Installing : python3-autocommand-2.2.2-8.el9.noarch 58/138 2026-03-09T20:49:52.372 INFO:teuthology.orchestra.run.vm08.stdout: Installing : libunwind-1.6.2-1.el9.x86_64 59/138 2026-03-09T20:49:52.377 INFO:teuthology.orchestra.run.vm08.stdout: Installing : gperftools-libs-2.9.1-3.el9.x86_64 60/138 2026-03-09T20:49:52.380 INFO:teuthology.orchestra.run.vm08.stdout: Installing : libarrow-doc-9.0.0-15.el9.noarch 61/138 2026-03-09T20:49:52.416 INFO:teuthology.orchestra.run.vm08.stdout: Installing : grpc-data-1.46.7-10.el9.noarch 62/138 2026-03-09T20:49:52.477 INFO:teuthology.orchestra.run.vm08.stdout: Installing : abseil-cpp-20211102.0-4.el9.x86_64 63/138 2026-03-09T20:49:52.493 INFO:teuthology.orchestra.run.vm08.stdout: Installing : python3-grpcio-1.46.7-10.el9.x86_64 64/138 2026-03-09T20:49:52.509 INFO:teuthology.orchestra.run.vm08.stdout: Installing : socat-1.7.4.1-8.el9.x86_64 65/138 2026-03-09T20:49:52.514 INFO:teuthology.orchestra.run.vm08.stdout: Installing : python3-toml-0.10.2-6.el9.noarch 66/138 2026-03-09T20:49:52.516 INFO:teuthology.orchestra.run.vm01.stdout:(66/136): python3-devel-3.9.25-3.el9.x86_64.rpm 1.2 MB/s | 244 kB 00:00 2026-03-09T20:49:52.523 INFO:teuthology.orchestra.run.vm08.stdout: Installing : python3-jaraco-functools-3.5.0-2.el9.noarch 67/138 2026-03-09T20:49:52.528 INFO:teuthology.orchestra.run.vm08.stdout: Installing : python3-jaraco-text-4.0.0-2.el9.noarch 68/138 2026-03-09T20:49:52.538 INFO:teuthology.orchestra.run.vm08.stdout: Installing : python3-jaraco-collections-3.0.0-8.el9.noarch 69/138 2026-03-09T20:49:52.544 INFO:teuthology.orchestra.run.vm08.stdout: Installing : python3-tempora-5.0.0-2.el9.noarch 70/138 2026-03-09T20:49:52.579 INFO:teuthology.orchestra.run.vm08.stdout: Installing : python3-portend-3.1.0-2.el9.noarch 71/138 2026-03-09T20:49:52.594 INFO:teuthology.orchestra.run.vm01.stdout:(67/136): python3-jinja2-2.11.3-8.el9.noarch.rp 3.1 MB/s | 249 kB 00:00 2026-03-09T20:49:52.594 INFO:teuthology.orchestra.run.vm08.stdout: Installing : python3-protobuf-3.14.0-17.el9.noarch 72/138 2026-03-09T20:49:52.640 INFO:teuthology.orchestra.run.vm08.stdout: Installing : python3-grpcio-tools-1.46.7-10.el9.x86_64 73/138 2026-03-09T20:49:52.836 INFO:teuthology.orchestra.run.vm06.stdout: Preparing : 1/1 2026-03-09T20:49:52.852 INFO:teuthology.orchestra.run.vm06.stdout: Installing : python3-more-itertools-8.12.0-2.el9.noarch 1/138 2026-03-09T20:49:52.872 INFO:teuthology.orchestra.run.vm06.stdout: Installing : thrift-0.15.0-4.el9.x86_64 2/138 2026-03-09T20:49:52.928 INFO:teuthology.orchestra.run.vm08.stdout: Installing : python3-devel-3.9.25-3.el9.x86_64 74/138 2026-03-09T20:49:52.962 INFO:teuthology.orchestra.run.vm08.stdout: Installing : python3-babel-2.9.1-2.el9.noarch 75/138 2026-03-09T20:49:52.970 INFO:teuthology.orchestra.run.vm08.stdout: Installing : python3-jinja2-2.11.3-8.el9.noarch 76/138 2026-03-09T20:49:53.038 INFO:teuthology.orchestra.run.vm08.stdout: Installing : openblas-0.3.29-1.el9.x86_64 77/138 2026-03-09T20:49:53.042 INFO:teuthology.orchestra.run.vm08.stdout: Installing : openblas-openmp-0.3.29-1.el9.x86_64 78/138 2026-03-09T20:49:53.055 INFO:teuthology.orchestra.run.vm06.stdout: Installing : lttng-ust-2.12.0-6.el9.x86_64 3/138 2026-03-09T20:49:53.057 INFO:teuthology.orchestra.run.vm06.stdout: Upgrading : librados2-2:19.2.3-678.ge911bdeb.el9.x86_64 4/138 2026-03-09T20:49:53.075 INFO:teuthology.orchestra.run.vm08.stdout: Installing : flexiblas-openblas-openmp-3.0.4-9.el9.x86_64 79/138 2026-03-09T20:49:53.115 INFO:teuthology.orchestra.run.vm01.stdout:(68/136): python3-jmespath-1.0.1-1.el9.noarch.r 91 kB/s | 48 kB 00:00 2026-03-09T20:49:53.122 INFO:teuthology.orchestra.run.vm06.stdout: Running scriptlet: librados2-2:19.2.3-678.ge911bdeb.el9.x86_64 4/138 2026-03-09T20:49:53.124 INFO:teuthology.orchestra.run.vm06.stdout: Installing : libcephfs2-2:19.2.3-678.ge911bdeb.el9.x86_64 5/138 2026-03-09T20:49:53.157 INFO:teuthology.orchestra.run.vm06.stdout: Running scriptlet: libcephfs2-2:19.2.3-678.ge911bdeb.el9.x86_64 5/138 2026-03-09T20:49:53.168 INFO:teuthology.orchestra.run.vm06.stdout: Installing : python3-rados-2:19.2.3-678.ge911bdeb.el9.x86_64 6/138 2026-03-09T20:49:53.173 INFO:teuthology.orchestra.run.vm06.stdout: Installing : librdkafka-1.6.1-102.el9.x86_64 7/138 2026-03-09T20:49:53.175 INFO:teuthology.orchestra.run.vm06.stdout: Installing : librabbitmq-0.11.0-7.el9.x86_64 8/138 2026-03-09T20:49:53.181 INFO:teuthology.orchestra.run.vm06.stdout: Installing : python3-jaraco-8.2.1-3.el9.noarch 9/138 2026-03-09T20:49:53.192 INFO:teuthology.orchestra.run.vm06.stdout: Installing : libnbd-1.20.3-4.el9.x86_64 10/138 2026-03-09T20:49:53.193 INFO:teuthology.orchestra.run.vm06.stdout: Installing : libcephsqlite-2:19.2.3-678.ge911bdeb.el9.x86_64 11/138 2026-03-09T20:49:53.236 INFO:teuthology.orchestra.run.vm06.stdout: Running scriptlet: libcephsqlite-2:19.2.3-678.ge911bdeb.el9.x86_64 11/138 2026-03-09T20:49:53.238 INFO:teuthology.orchestra.run.vm06.stdout: Installing : libradosstriper1-2:19.2.3-678.ge911bdeb.el9.x86_ 12/138 2026-03-09T20:49:53.260 INFO:teuthology.orchestra.run.vm06.stdout: Running scriptlet: libradosstriper1-2:19.2.3-678.ge911bdeb.el9.x86_ 12/138 2026-03-09T20:49:53.288 INFO:teuthology.orchestra.run.vm01.stdout:(69/136): python3-libstoragemgmt-1.10.1-1.el9.x 1.0 MB/s | 177 kB 00:00 2026-03-09T20:49:53.299 INFO:teuthology.orchestra.run.vm06.stdout: Installing : re2-1:20211101-20.el9.x86_64 13/138 2026-03-09T20:49:53.341 INFO:teuthology.orchestra.run.vm06.stdout: Installing : libarrow-9.0.0-15.el9.x86_64 14/138 2026-03-09T20:49:53.346 INFO:teuthology.orchestra.run.vm06.stdout: Installing : python3-werkzeug-2.0.3-3.el9.1.noarch 15/138 2026-03-09T20:49:53.375 INFO:teuthology.orchestra.run.vm06.stdout: Installing : liboath-2.6.12-1.el9.x86_64 16/138 2026-03-09T20:49:53.392 INFO:teuthology.orchestra.run.vm06.stdout: Installing : python3-pyasn1-0.4.8-7.el9.noarch 17/138 2026-03-09T20:49:53.402 INFO:teuthology.orchestra.run.vm06.stdout: Installing : python3-packaging-20.9-5.el9.noarch 18/138 2026-03-09T20:49:53.413 INFO:teuthology.orchestra.run.vm06.stdout: Installing : python3-markupsafe-1.1.1-12.el9.x86_64 19/138 2026-03-09T20:49:53.421 INFO:teuthology.orchestra.run.vm06.stdout: Installing : protobuf-3.14.0-17.el9.x86_64 20/138 2026-03-09T20:49:53.425 INFO:teuthology.orchestra.run.vm06.stdout: Installing : lua-5.4.4-4.el9.x86_64 21/138 2026-03-09T20:49:53.433 INFO:teuthology.orchestra.run.vm06.stdout: Installing : flexiblas-3.0.4-9.el9.x86_64 22/138 2026-03-09T20:49:53.467 INFO:teuthology.orchestra.run.vm06.stdout: Installing : unzip-6.0-59.el9.x86_64 23/138 2026-03-09T20:49:53.485 INFO:teuthology.orchestra.run.vm06.stdout: Installing : python3-urllib3-1.26.5-7.el9.noarch 24/138 2026-03-09T20:49:53.490 INFO:teuthology.orchestra.run.vm06.stdout: Installing : python3-requests-2.25.1-10.el9.noarch 25/138 2026-03-09T20:49:53.498 INFO:teuthology.orchestra.run.vm06.stdout: Installing : libquadmath-11.5.0-14.el9.x86_64 26/138 2026-03-09T20:49:53.501 INFO:teuthology.orchestra.run.vm06.stdout: Installing : libgfortran-11.5.0-14.el9.x86_64 27/138 2026-03-09T20:49:53.513 INFO:teuthology.orchestra.run.vm08.stdout: Installing : flexiblas-netlib-3.0.4-9.el9.x86_64 80/138 2026-03-09T20:49:53.535 INFO:teuthology.orchestra.run.vm06.stdout: Installing : ledmon-libs-1.1.0-3.el9.x86_64 28/138 2026-03-09T20:49:53.543 INFO:teuthology.orchestra.run.vm06.stdout: Installing : python3-ceph-common-2:19.2.3-678.ge911bdeb.el9.x 29/138 2026-03-09T20:49:53.555 INFO:teuthology.orchestra.run.vm06.stdout: Installing : python3-ceph-argparse-2:19.2.3-678.ge911bdeb.el9 30/138 2026-03-09T20:49:53.577 INFO:teuthology.orchestra.run.vm06.stdout: Installing : python3-cephfs-2:19.2.3-678.ge911bdeb.el9.x86_64 31/138 2026-03-09T20:49:53.607 INFO:teuthology.orchestra.run.vm08.stdout: Installing : python3-numpy-1:1.23.5-2.el9.x86_64 81/138 2026-03-09T20:49:53.637 INFO:teuthology.orchestra.run.vm06.stdout: Installing : python3-requests-oauthlib-1.3.0-12.el9.noarch 32/138 2026-03-09T20:49:53.687 INFO:teuthology.orchestra.run.vm06.stdout: Installing : zip-3.0-35.el9.x86_64 33/138 2026-03-09T20:49:53.831 INFO:teuthology.orchestra.run.vm06.stdout: Installing : luarocks-3.9.2-5.el9.noarch 34/138 2026-03-09T20:49:53.849 INFO:teuthology.orchestra.run.vm06.stdout: Installing : lua-devel-5.4.4-4.el9.x86_64 35/138 2026-03-09T20:49:53.911 INFO:teuthology.orchestra.run.vm01.stdout:(70/136): protobuf-3.14.0-17.el9.x86_64.rpm 294 kB/s | 1.0 MB 00:03 2026-03-09T20:49:53.935 INFO:teuthology.orchestra.run.vm01.stdout:(71/136): python3-mako-1.1.4-6.el9.noarch.rpm 266 kB/s | 172 kB 00:00 2026-03-09T20:49:53.945 INFO:teuthology.orchestra.run.vm06.stdout: Installing : protobuf-compiler-3.14.0-17.el9.x86_64 36/138 2026-03-09T20:49:54.061 INFO:teuthology.orchestra.run.vm01.stdout:(72/136): python3-markupsafe-1.1.1-12.el9.x86_6 232 kB/s | 35 kB 00:00 2026-03-09T20:49:54.155 INFO:teuthology.orchestra.run.vm06.stdout: Installing : python3-mako-1.1.4-6.el9.noarch 37/138 2026-03-09T20:49:54.158 INFO:teuthology.orchestra.run.vm01.stdout:(73/136): openblas-openmp-0.3.29-1.el9.x86_64.r 1.4 MB/s | 5.3 MB 00:03 2026-03-09T20:49:54.216 INFO:teuthology.orchestra.run.vm06.stdout: Installing : python3-pyasn1-modules-0.4.8-7.el9.noarch 38/138 2026-03-09T20:49:54.224 INFO:teuthology.orchestra.run.vm06.stdout: Installing : python3-rsa-4.9-2.el9.noarch 39/138 2026-03-09T20:49:54.234 INFO:teuthology.orchestra.run.vm06.stdout: Installing : python3-jaraco-classes-3.2.1-5.el9.noarch 40/138 2026-03-09T20:49:54.241 INFO:teuthology.orchestra.run.vm06.stdout: Installing : librados-devel-2:19.2.3-678.ge911bdeb.el9.x86_64 41/138 2026-03-09T20:49:54.246 INFO:teuthology.orchestra.run.vm06.stdout: Installing : python3-zc-lockfile-2.0-10.el9.noarch 42/138 2026-03-09T20:49:54.265 INFO:teuthology.orchestra.run.vm06.stdout: Installing : python3-xmltodict-0.12.0-15.el9.noarch 43/138 2026-03-09T20:49:54.284 INFO:teuthology.orchestra.run.vm01.stdout:(74/136): python3-packaging-20.9-5.el9.noarch.r 616 kB/s | 77 kB 00:00 2026-03-09T20:49:54.292 INFO:teuthology.orchestra.run.vm06.stdout: Installing : python3-websocket-client-1.2.3-2.el9.noarch 44/138 2026-03-09T20:49:54.300 INFO:teuthology.orchestra.run.vm06.stdout: Installing : python3-webob-1.8.8-2.el9.noarch 45/138 2026-03-09T20:49:54.307 INFO:teuthology.orchestra.run.vm06.stdout: Installing : python3-typing-extensions-4.15.0-1.el9.noarch 46/138 2026-03-09T20:49:54.324 INFO:teuthology.orchestra.run.vm06.stdout: Installing : python3-repoze-lru-0.7-16.el9.noarch 47/138 2026-03-09T20:49:54.338 INFO:teuthology.orchestra.run.vm06.stdout: Installing : python3-routes-2.5.1-5.el9.noarch 48/138 2026-03-09T20:49:54.352 INFO:teuthology.orchestra.run.vm06.stdout: Installing : python3-natsort-7.1.1-5.el9.noarch 49/138 2026-03-09T20:49:54.419 INFO:teuthology.orchestra.run.vm06.stdout: Installing : python3-logutils-0.3.5-21.el9.noarch 50/138 2026-03-09T20:49:54.430 INFO:teuthology.orchestra.run.vm06.stdout: Installing : python3-pecan-1.4.2-3.el9.noarch 51/138 2026-03-09T20:49:54.441 INFO:teuthology.orchestra.run.vm06.stdout: Installing : python3-certifi-2023.05.07-4.el9.noarch 52/138 2026-03-09T20:49:54.493 INFO:teuthology.orchestra.run.vm06.stdout: Installing : python3-cachetools-4.2.4-1.el9.noarch 53/138 2026-03-09T20:49:54.549 INFO:teuthology.orchestra.run.vm08.stdout: Installing : python3-numpy-f2py-1:1.23.5-2.el9.x86_64 82/138 2026-03-09T20:49:54.581 INFO:teuthology.orchestra.run.vm08.stdout: Installing : python3-scipy-1.9.3-2.el9.x86_64 83/138 2026-03-09T20:49:54.590 INFO:teuthology.orchestra.run.vm08.stdout: Installing : libxslt-1.1.34-12.el9.x86_64 84/138 2026-03-09T20:49:54.597 INFO:teuthology.orchestra.run.vm08.stdout: Installing : xmlstarlet-1.6.1-20.el9.x86_64 85/138 2026-03-09T20:49:54.768 INFO:teuthology.orchestra.run.vm08.stdout: Installing : libpmemobj-1.12.1-1.el9.x86_64 86/138 2026-03-09T20:49:54.772 INFO:teuthology.orchestra.run.vm08.stdout: Upgrading : librbd1-2:19.2.3-678.ge911bdeb.el9.x86_64 87/138 2026-03-09T20:49:54.810 INFO:teuthology.orchestra.run.vm08.stdout: Running scriptlet: librbd1-2:19.2.3-678.ge911bdeb.el9.x86_64 87/138 2026-03-09T20:49:54.814 INFO:teuthology.orchestra.run.vm08.stdout: Installing : python3-rbd-2:19.2.3-678.ge911bdeb.el9.x86_64 88/138 2026-03-09T20:49:54.823 INFO:teuthology.orchestra.run.vm08.stdout: Installing : boost-program-options-1.75.0-13.el9.x86_64 89/138 2026-03-09T20:49:54.893 INFO:teuthology.orchestra.run.vm06.stdout: Installing : python3-google-auth-1:2.45.0-1.el9.noarch 54/138 2026-03-09T20:49:54.938 INFO:teuthology.orchestra.run.vm06.stdout: Installing : python3-kubernetes-1:26.1.0-3.el9.noarch 55/138 2026-03-09T20:49:54.945 INFO:teuthology.orchestra.run.vm06.stdout: Installing : python3-backports-tarfile-1.2.0-1.el9.noarch 56/138 2026-03-09T20:49:54.954 INFO:teuthology.orchestra.run.vm06.stdout: Installing : python3-jaraco-context-6.0.1-3.el9.noarch 57/138 2026-03-09T20:49:54.961 INFO:teuthology.orchestra.run.vm06.stdout: Installing : python3-autocommand-2.2.2-8.el9.noarch 58/138 2026-03-09T20:49:54.969 INFO:teuthology.orchestra.run.vm06.stdout: Installing : libunwind-1.6.2-1.el9.x86_64 59/138 2026-03-09T20:49:54.973 INFO:teuthology.orchestra.run.vm06.stdout: Installing : gperftools-libs-2.9.1-3.el9.x86_64 60/138 2026-03-09T20:49:54.975 INFO:teuthology.orchestra.run.vm06.stdout: Installing : libarrow-doc-9.0.0-15.el9.noarch 61/138 2026-03-09T20:49:55.008 INFO:teuthology.orchestra.run.vm06.stdout: Installing : grpc-data-1.46.7-10.el9.noarch 62/138 2026-03-09T20:49:55.064 INFO:teuthology.orchestra.run.vm06.stdout: Installing : abseil-cpp-20211102.0-4.el9.x86_64 63/138 2026-03-09T20:49:55.095 INFO:teuthology.orchestra.run.vm08.stdout: Installing : parquet-libs-9.0.0-15.el9.x86_64 90/138 2026-03-09T20:49:55.130 INFO:teuthology.orchestra.run.vm01.stdout:(75/136): python3-protobuf-3.14.0-17.el9.noarch 316 kB/s | 267 kB 00:00 2026-03-09T20:49:55.263 INFO:teuthology.orchestra.run.vm01.stdout:(76/136): python3-numpy-f2py-1.23.5-2.el9.x86_6 368 kB/s | 442 kB 00:01 2026-03-09T20:49:55.308 INFO:teuthology.orchestra.run.vm06.stdout: Installing : python3-grpcio-1.46.7-10.el9.x86_64 64/138 2026-03-09T20:49:55.313 INFO:teuthology.orchestra.run.vm08.stdout: Installing : librgw2-2:19.2.3-678.ge911bdeb.el9.x86_64 91/138 2026-03-09T20:49:55.344 INFO:teuthology.orchestra.run.vm06.stdout: Installing : socat-1.7.4.1-8.el9.x86_64 65/138 2026-03-09T20:49:55.385 INFO:teuthology.orchestra.run.vm06.stdout: Installing : python3-toml-0.10.2-6.el9.noarch 66/138 2026-03-09T20:49:55.391 INFO:teuthology.orchestra.run.vm08.stdout: Running scriptlet: librgw2-2:19.2.3-678.ge911bdeb.el9.x86_64 91/138 2026-03-09T20:49:55.394 INFO:teuthology.orchestra.run.vm06.stdout: Installing : python3-jaraco-functools-3.5.0-2.el9.noarch 67/138 2026-03-09T20:49:55.394 INFO:teuthology.orchestra.run.vm08.stdout: Installing : python3-rgw-2:19.2.3-678.ge911bdeb.el9.x86_64 92/138 2026-03-09T20:49:55.400 INFO:teuthology.orchestra.run.vm06.stdout: Installing : python3-jaraco-text-4.0.0-2.el9.noarch 68/138 2026-03-09T20:49:55.409 INFO:teuthology.orchestra.run.vm06.stdout: Installing : python3-jaraco-collections-3.0.0-8.el9.noarch 69/138 2026-03-09T20:49:55.415 INFO:teuthology.orchestra.run.vm06.stdout: Installing : python3-tempora-5.0.0-2.el9.noarch 70/138 2026-03-09T20:49:55.459 INFO:teuthology.orchestra.run.vm06.stdout: Installing : python3-portend-3.1.0-2.el9.noarch 71/138 2026-03-09T20:49:55.476 INFO:teuthology.orchestra.run.vm06.stdout: Installing : python3-protobuf-3.14.0-17.el9.noarch 72/138 2026-03-09T20:49:55.532 INFO:teuthology.orchestra.run.vm06.stdout: Installing : python3-grpcio-tools-1.46.7-10.el9.x86_64 73/138 2026-03-09T20:49:55.585 INFO:teuthology.orchestra.run.vm01.stdout:(77/136): python3-numpy-1.23.5-2.el9.x86_64.rpm 3.7 MB/s | 6.1 MB 00:01 2026-03-09T20:49:55.586 INFO:teuthology.orchestra.run.vm01.stdout:(78/136): python3-pyasn1-0.4.8-7.el9.noarch.rpm 345 kB/s | 157 kB 00:00 2026-03-09T20:49:55.587 INFO:teuthology.orchestra.run.vm01.stdout:(79/136): python3-pyasn1-modules-0.4.8-7.el9.no 855 kB/s | 277 kB 00:00 2026-03-09T20:49:55.691 INFO:teuthology.orchestra.run.vm01.stdout:(80/136): python3-toml-0.10.2-6.el9.noarch.rpm 408 kB/s | 42 kB 00:00 2026-03-09T20:49:55.822 INFO:teuthology.orchestra.run.vm06.stdout: Installing : python3-devel-3.9.25-3.el9.x86_64 74/138 2026-03-09T20:49:55.831 INFO:teuthology.orchestra.run.vm01.stdout:(81/136): python3-requests-oauthlib-1.3.0-12.el 219 kB/s | 54 kB 00:00 2026-03-09T20:49:55.857 INFO:teuthology.orchestra.run.vm06.stdout: Installing : python3-babel-2.9.1-2.el9.noarch 75/138 2026-03-09T20:49:55.863 INFO:teuthology.orchestra.run.vm06.stdout: Installing : python3-jinja2-2.11.3-8.el9.noarch 76/138 2026-03-09T20:49:55.907 INFO:teuthology.orchestra.run.vm01.stdout:(82/136): qatlib-service-25.08.0-2.el9.x86_64.r 490 kB/s | 37 kB 00:00 2026-03-09T20:49:55.936 INFO:teuthology.orchestra.run.vm06.stdout: Installing : openblas-0.3.29-1.el9.x86_64 77/138 2026-03-09T20:49:55.939 INFO:teuthology.orchestra.run.vm06.stdout: Installing : openblas-openmp-0.3.29-1.el9.x86_64 78/138 2026-03-09T20:49:55.962 INFO:teuthology.orchestra.run.vm01.stdout:(83/136): qatzip-libs-1.3.1-1.el9.x86_64.rpm 1.2 MB/s | 66 kB 00:00 2026-03-09T20:49:55.970 INFO:teuthology.orchestra.run.vm06.stdout: Installing : flexiblas-openblas-openmp-3.0.4-9.el9.x86_64 79/138 2026-03-09T20:49:56.387 INFO:teuthology.orchestra.run.vm06.stdout: Installing : flexiblas-netlib-3.0.4-9.el9.x86_64 80/138 2026-03-09T20:49:56.488 INFO:teuthology.orchestra.run.vm06.stdout: Installing : python3-numpy-1:1.23.5-2.el9.x86_64 81/138 2026-03-09T20:49:56.611 INFO:teuthology.orchestra.run.vm08.stdout: Running scriptlet: ceph-common-2:19.2.3-678.ge911bdeb.el9.x86_64 93/138 2026-03-09T20:49:56.666 INFO:teuthology.orchestra.run.vm08.stdout: Installing : ceph-common-2:19.2.3-678.ge911bdeb.el9.x86_64 93/138 2026-03-09T20:49:56.696 INFO:teuthology.orchestra.run.vm08.stdout: Running scriptlet: ceph-common-2:19.2.3-678.ge911bdeb.el9.x86_64 93/138 2026-03-09T20:49:56.722 INFO:teuthology.orchestra.run.vm08.stdout: Installing : python3-ply-3.11-14.el9.noarch 94/138 2026-03-09T20:49:56.744 INFO:teuthology.orchestra.run.vm08.stdout: Installing : python3-pycparser-2.20-6.el9.noarch 95/138 2026-03-09T20:49:56.800 INFO:teuthology.orchestra.run.vm01.stdout:(84/136): socat-1.7.4.1-8.el9.x86_64.rpm 362 kB/s | 303 kB 00:00 2026-03-09T20:49:56.844 INFO:teuthology.orchestra.run.vm08.stdout: Installing : python3-cffi-1.14.5-5.el9.x86_64 96/138 2026-03-09T20:49:56.849 INFO:teuthology.orchestra.run.vm01.stdout:(85/136): xmlstarlet-1.6.1-20.el9.x86_64.rpm 1.3 MB/s | 64 kB 00:00 2026-03-09T20:49:56.858 INFO:teuthology.orchestra.run.vm08.stdout: Installing : python3-cryptography-36.0.1-5.el9.x86_64 97/138 2026-03-09T20:49:56.889 INFO:teuthology.orchestra.run.vm08.stdout: Installing : python3-pyOpenSSL-21.0.0-1.el9.noarch 98/138 2026-03-09T20:49:56.927 INFO:teuthology.orchestra.run.vm08.stdout: Installing : python3-cheroot-10.0.1-4.el9.noarch 99/138 2026-03-09T20:49:56.963 INFO:teuthology.orchestra.run.vm01.stdout:(86/136): lua-devel-5.4.4-4.el9.x86_64.rpm 198 kB/s | 22 kB 00:00 2026-03-09T20:49:56.993 INFO:teuthology.orchestra.run.vm08.stdout: Installing : python3-cherrypy-18.6.1-2.el9.noarch 100/138 2026-03-09T20:49:57.004 INFO:teuthology.orchestra.run.vm08.stdout: Installing : python3-asyncssh-2.13.2-5.el9.noarch 101/138 2026-03-09T20:49:57.009 INFO:teuthology.orchestra.run.vm08.stdout: Installing : python3-bcrypt-3.2.2-1.el9.x86_64 102/138 2026-03-09T20:49:57.016 INFO:teuthology.orchestra.run.vm08.stdout: Installing : pciutils-3.7.0-7.el9.x86_64 103/138 2026-03-09T20:49:57.021 INFO:teuthology.orchestra.run.vm08.stdout: Installing : qatlib-25.08.0-2.el9.x86_64 104/138 2026-03-09T20:49:57.023 INFO:teuthology.orchestra.run.vm08.stdout: Installing : qatlib-service-25.08.0-2.el9.x86_64 105/138 2026-03-09T20:49:57.044 INFO:teuthology.orchestra.run.vm08.stdout: Running scriptlet: qatlib-service-25.08.0-2.el9.x86_64 105/138 2026-03-09T20:49:57.099 INFO:teuthology.orchestra.run.vm01.stdout:(87/136): protobuf-compiler-3.14.0-17.el9.x86_6 6.2 MB/s | 862 kB 00:00 2026-03-09T20:49:57.117 INFO:teuthology.orchestra.run.vm01.stdout:(88/136): abseil-cpp-20211102.0-4.el9.x86_64.rp 32 MB/s | 551 kB 00:00 2026-03-09T20:49:57.125 INFO:teuthology.orchestra.run.vm01.stdout:(89/136): gperftools-libs-2.9.1-3.el9.x86_64.rp 38 MB/s | 308 kB 00:00 2026-03-09T20:49:57.128 INFO:teuthology.orchestra.run.vm01.stdout:(90/136): grpc-data-1.46.7-10.el9.noarch.rpm 8.0 MB/s | 19 kB 00:00 2026-03-09T20:49:57.187 INFO:teuthology.orchestra.run.vm01.stdout:(91/136): libarrow-9.0.0-15.el9.x86_64.rpm 75 MB/s | 4.4 MB 00:00 2026-03-09T20:49:57.190 INFO:teuthology.orchestra.run.vm01.stdout:(92/136): libarrow-doc-9.0.0-15.el9.noarch.rpm 9.3 MB/s | 25 kB 00:00 2026-03-09T20:49:57.193 INFO:teuthology.orchestra.run.vm01.stdout:(93/136): liboath-2.6.12-1.el9.x86_64.rpm 17 MB/s | 49 kB 00:00 2026-03-09T20:49:57.196 INFO:teuthology.orchestra.run.vm01.stdout:(94/136): libunwind-1.6.2-1.el9.x86_64.rpm 24 MB/s | 67 kB 00:00 2026-03-09T20:49:57.201 INFO:teuthology.orchestra.run.vm01.stdout:(95/136): luarocks-3.9.2-5.el9.noarch.rpm 31 MB/s | 151 kB 00:00 2026-03-09T20:49:57.214 INFO:teuthology.orchestra.run.vm01.stdout:(96/136): parquet-libs-9.0.0-15.el9.x86_64.rpm 70 MB/s | 838 kB 00:00 2026-03-09T20:49:57.222 INFO:teuthology.orchestra.run.vm01.stdout:(97/136): python3-asyncssh-2.13.2-5.el9.noarch. 67 MB/s | 548 kB 00:00 2026-03-09T20:49:57.224 INFO:teuthology.orchestra.run.vm01.stdout:(98/136): python3-autocommand-2.2.2-8.el9.noarc 14 MB/s | 29 kB 00:00 2026-03-09T20:49:57.227 INFO:teuthology.orchestra.run.vm01.stdout:(99/136): python3-backports-tarfile-1.2.0-1.el9 21 MB/s | 60 kB 00:00 2026-03-09T20:49:57.229 INFO:teuthology.orchestra.run.vm01.stdout:(100/136): python3-bcrypt-3.2.2-1.el9.x86_64.rp 18 MB/s | 43 kB 00:00 2026-03-09T20:49:57.233 INFO:teuthology.orchestra.run.vm01.stdout:(101/136): python3-cachetools-4.2.4-1.el9.noarc 11 MB/s | 32 kB 00:00 2026-03-09T20:49:57.235 INFO:teuthology.orchestra.run.vm01.stdout:(102/136): python3-certifi-2023.05.07-4.el9.noa 5.8 MB/s | 14 kB 00:00 2026-03-09T20:49:57.241 INFO:teuthology.orchestra.run.vm01.stdout:(103/136): python3-cheroot-10.0.1-4.el9.noarch. 32 MB/s | 173 kB 00:00 2026-03-09T20:49:57.248 INFO:teuthology.orchestra.run.vm01.stdout:(104/136): qatlib-25.08.0-2.el9.x86_64.rpm 154 kB/s | 240 kB 00:01 2026-03-09T20:49:57.249 INFO:teuthology.orchestra.run.vm01.stdout:(105/136): python3-cherrypy-18.6.1-2.el9.noarch 45 MB/s | 358 kB 00:00 2026-03-09T20:49:57.255 INFO:teuthology.orchestra.run.vm01.stdout:(106/136): python3-google-auth-2.45.0-1.el9.noa 48 MB/s | 254 kB 00:00 2026-03-09T20:49:57.259 INFO:teuthology.orchestra.run.vm01.stdout:(107/136): python3-grpcio-tools-1.46.7-10.el9.x 36 MB/s | 144 kB 00:00 2026-03-09T20:49:57.261 INFO:teuthology.orchestra.run.vm01.stdout:(108/136): python3-jaraco-8.2.1-3.el9.noarch.rp 4.7 MB/s | 11 kB 00:00 2026-03-09T20:49:57.264 INFO:teuthology.orchestra.run.vm01.stdout:(109/136): python3-jaraco-classes-3.2.1-5.el9.n 8.4 MB/s | 18 kB 00:00 2026-03-09T20:49:57.266 INFO:teuthology.orchestra.run.vm01.stdout:(110/136): python3-jaraco-collections-3.0.0-8.e 9.2 MB/s | 23 kB 00:00 2026-03-09T20:49:57.269 INFO:teuthology.orchestra.run.vm01.stdout:(111/136): python3-jaraco-context-6.0.1-3.el9.n 8.9 MB/s | 20 kB 00:00 2026-03-09T20:49:57.272 INFO:teuthology.orchestra.run.vm01.stdout:(112/136): python3-jaraco-functools-3.5.0-2.el9 5.6 MB/s | 19 kB 00:00 2026-03-09T20:49:57.275 INFO:teuthology.orchestra.run.vm01.stdout:(113/136): python3-jaraco-text-4.0.0-2.el9.noar 11 MB/s | 26 kB 00:00 2026-03-09T20:49:57.291 INFO:teuthology.orchestra.run.vm01.stdout:(114/136): python3-kubernetes-26.1.0-3.el9.noar 65 MB/s | 1.0 MB 00:00 2026-03-09T20:49:57.294 INFO:teuthology.orchestra.run.vm01.stdout:(115/136): python3-logutils-0.3.5-21.el9.noarch 16 MB/s | 46 kB 00:00 2026-03-09T20:49:57.300 INFO:teuthology.orchestra.run.vm01.stdout:(116/136): python3-more-itertools-8.12.0-2.el9. 15 MB/s | 79 kB 00:00 2026-03-09T20:49:57.303 INFO:teuthology.orchestra.run.vm01.stdout:(117/136): python3-natsort-7.1.1-5.el9.noarch.r 19 MB/s | 58 kB 00:00 2026-03-09T20:49:57.309 INFO:teuthology.orchestra.run.vm01.stdout:(118/136): python3-pecan-1.4.2-3.el9.noarch.rpm 48 MB/s | 272 kB 00:00 2026-03-09T20:49:57.313 INFO:teuthology.orchestra.run.vm01.stdout:(119/136): python3-portend-3.1.0-2.el9.noarch.r 4.5 MB/s | 16 kB 00:00 2026-03-09T20:49:57.321 INFO:teuthology.orchestra.run.vm01.stdout:(120/136): python3-grpcio-1.46.7-10.el9.x86_64. 29 MB/s | 2.0 MB 00:00 2026-03-09T20:49:57.325 INFO:teuthology.orchestra.run.vm01.stdout:(121/136): python3-repoze-lru-0.7-16.el9.noarch 9.3 MB/s | 31 kB 00:00 2026-03-09T20:49:57.327 INFO:teuthology.orchestra.run.vm01.stdout:(122/136): python3-pyOpenSSL-21.0.0-1.el9.noarc 6.5 MB/s | 90 kB 00:00 2026-03-09T20:49:57.330 INFO:teuthology.orchestra.run.vm01.stdout:(123/136): python3-routes-2.5.1-5.el9.noarch.rp 38 MB/s | 188 kB 00:00 2026-03-09T20:49:57.331 INFO:teuthology.orchestra.run.vm01.stdout:(124/136): python3-rsa-4.9-2.el9.noarch.rpm 17 MB/s | 59 kB 00:00 2026-03-09T20:49:57.332 INFO:teuthology.orchestra.run.vm01.stdout:(125/136): python3-tempora-5.0.0-2.el9.noarch.r 16 MB/s | 36 kB 00:00 2026-03-09T20:49:57.337 INFO:teuthology.orchestra.run.vm01.stdout:(126/136): python3-typing-extensions-4.15.0-1.e 14 MB/s | 86 kB 00:00 2026-03-09T20:49:57.338 INFO:teuthology.orchestra.run.vm01.stdout:(127/136): python3-webob-1.8.8-2.el9.noarch.rpm 38 MB/s | 230 kB 00:00 2026-03-09T20:49:57.342 INFO:teuthology.orchestra.run.vm01.stdout:(128/136): python3-websocket-client-1.2.3-2.el9 17 MB/s | 90 kB 00:00 2026-03-09T20:49:57.346 INFO:teuthology.orchestra.run.vm01.stdout:(129/136): python3-werkzeug-2.0.3-3.el9.1.noarc 55 MB/s | 427 kB 00:00 2026-03-09T20:49:57.347 INFO:teuthology.orchestra.run.vm01.stdout:(130/136): python3-xmltodict-0.12.0-15.el9.noar 5.0 MB/s | 22 kB 00:00 2026-03-09T20:49:57.349 INFO:teuthology.orchestra.run.vm01.stdout:(131/136): python3-zc-lockfile-2.0-10.el9.noarc 8.8 MB/s | 20 kB 00:00 2026-03-09T20:49:57.356 INFO:teuthology.orchestra.run.vm01.stdout:(132/136): re2-20211101-20.el9.x86_64.rpm 22 MB/s | 191 kB 00:00 2026-03-09T20:49:57.372 INFO:teuthology.orchestra.run.vm01.stdout:(133/136): thrift-0.15.0-4.el9.x86_64.rpm 70 MB/s | 1.6 MB 00:00 2026-03-09T20:49:57.372 INFO:teuthology.orchestra.run.vm06.stdout: Installing : python3-numpy-f2py-1:1.23.5-2.el9.x86_64 82/138 2026-03-09T20:49:57.392 INFO:teuthology.orchestra.run.vm08.stdout: Installing : qatzip-libs-1.3.1-1.el9.x86_64 106/138 2026-03-09T20:49:57.400 INFO:teuthology.orchestra.run.vm08.stdout: Installing : ceph-base-2:19.2.3-678.ge911bdeb.el9.x86_64 107/138 2026-03-09T20:49:57.403 INFO:teuthology.orchestra.run.vm06.stdout: Installing : python3-scipy-1.9.3-2.el9.x86_64 83/138 2026-03-09T20:49:57.411 INFO:teuthology.orchestra.run.vm06.stdout: Installing : libxslt-1.1.34-12.el9.x86_64 84/138 2026-03-09T20:49:57.438 INFO:teuthology.orchestra.run.vm06.stdout: Installing : xmlstarlet-1.6.1-20.el9.x86_64 85/138 2026-03-09T20:49:57.468 INFO:teuthology.orchestra.run.vm08.stdout: Running scriptlet: ceph-base-2:19.2.3-678.ge911bdeb.el9.x86_64 107/138 2026-03-09T20:49:57.495 INFO:teuthology.orchestra.run.vm08.stdout:Created symlink /etc/systemd/system/multi-user.target.wants/ceph.target → /usr/lib/systemd/system/ceph.target. 2026-03-09T20:49:57.495 INFO:teuthology.orchestra.run.vm08.stdout:Created symlink /etc/systemd/system/ceph.target.wants/ceph-crash.service → /usr/lib/systemd/system/ceph-crash.service. 2026-03-09T20:49:57.495 INFO:teuthology.orchestra.run.vm08.stdout: 2026-03-09T20:49:57.538 INFO:teuthology.orchestra.run.vm08.stdout: Installing : ceph-selinux-2:19.2.3-678.ge911bdeb.el9.x86_64 108/138 2026-03-09T20:49:57.694 INFO:teuthology.orchestra.run.vm06.stdout: Installing : libpmemobj-1.12.1-1.el9.x86_64 86/138 2026-03-09T20:49:57.698 INFO:teuthology.orchestra.run.vm06.stdout: Upgrading : librbd1-2:19.2.3-678.ge911bdeb.el9.x86_64 87/138 2026-03-09T20:49:57.738 INFO:teuthology.orchestra.run.vm06.stdout: Running scriptlet: librbd1-2:19.2.3-678.ge911bdeb.el9.x86_64 87/138 2026-03-09T20:49:57.745 INFO:teuthology.orchestra.run.vm06.stdout: Installing : python3-rbd-2:19.2.3-678.ge911bdeb.el9.x86_64 88/138 2026-03-09T20:49:57.755 INFO:teuthology.orchestra.run.vm06.stdout: Installing : boost-program-options-1.75.0-13.el9.x86_64 89/138 2026-03-09T20:49:58.037 INFO:teuthology.orchestra.run.vm06.stdout: Installing : parquet-libs-9.0.0-15.el9.x86_64 90/138 2026-03-09T20:49:58.041 INFO:teuthology.orchestra.run.vm06.stdout: Installing : librgw2-2:19.2.3-678.ge911bdeb.el9.x86_64 91/138 2026-03-09T20:49:58.066 INFO:teuthology.orchestra.run.vm06.stdout: Running scriptlet: librgw2-2:19.2.3-678.ge911bdeb.el9.x86_64 91/138 2026-03-09T20:49:58.069 INFO:teuthology.orchestra.run.vm06.stdout: Installing : python3-rgw-2:19.2.3-678.ge911bdeb.el9.x86_64 92/138 2026-03-09T20:49:58.432 INFO:teuthology.orchestra.run.vm01.stdout:(134/136): librados2-19.2.3-678.ge911bdeb.el9.x 3.2 MB/s | 3.4 MB 00:01 2026-03-09T20:49:58.559 INFO:teuthology.orchestra.run.vm01.stdout:(135/136): librbd1-19.2.3-678.ge911bdeb.el9.x86 2.7 MB/s | 3.2 MB 00:01 2026-03-09T20:49:59.274 INFO:teuthology.orchestra.run.vm06.stdout: Running scriptlet: ceph-common-2:19.2.3-678.ge911bdeb.el9.x86_64 93/138 2026-03-09T20:49:59.340 INFO:teuthology.orchestra.run.vm06.stdout: Installing : ceph-common-2:19.2.3-678.ge911bdeb.el9.x86_64 93/138 2026-03-09T20:49:59.368 INFO:teuthology.orchestra.run.vm06.stdout: Running scriptlet: ceph-common-2:19.2.3-678.ge911bdeb.el9.x86_64 93/138 2026-03-09T20:49:59.387 INFO:teuthology.orchestra.run.vm06.stdout: Installing : python3-ply-3.11-14.el9.noarch 94/138 2026-03-09T20:49:59.410 INFO:teuthology.orchestra.run.vm06.stdout: Installing : python3-pycparser-2.20-6.el9.noarch 95/138 2026-03-09T20:49:59.511 INFO:teuthology.orchestra.run.vm06.stdout: Installing : python3-cffi-1.14.5-5.el9.x86_64 96/138 2026-03-09T20:49:59.526 INFO:teuthology.orchestra.run.vm06.stdout: Installing : python3-cryptography-36.0.1-5.el9.x86_64 97/138 2026-03-09T20:49:59.561 INFO:teuthology.orchestra.run.vm06.stdout: Installing : python3-pyOpenSSL-21.0.0-1.el9.noarch 98/138 2026-03-09T20:49:59.603 INFO:teuthology.orchestra.run.vm06.stdout: Installing : python3-cheroot-10.0.1-4.el9.noarch 99/138 2026-03-09T20:49:59.673 INFO:teuthology.orchestra.run.vm06.stdout: Installing : python3-cherrypy-18.6.1-2.el9.noarch 100/138 2026-03-09T20:49:59.695 INFO:teuthology.orchestra.run.vm06.stdout: Installing : python3-asyncssh-2.13.2-5.el9.noarch 101/138 2026-03-09T20:49:59.701 INFO:teuthology.orchestra.run.vm06.stdout: Installing : python3-bcrypt-3.2.2-1.el9.x86_64 102/138 2026-03-09T20:49:59.707 INFO:teuthology.orchestra.run.vm06.stdout: Installing : pciutils-3.7.0-7.el9.x86_64 103/138 2026-03-09T20:49:59.712 INFO:teuthology.orchestra.run.vm06.stdout: Installing : qatlib-25.08.0-2.el9.x86_64 104/138 2026-03-09T20:49:59.714 INFO:teuthology.orchestra.run.vm06.stdout: Installing : qatlib-service-25.08.0-2.el9.x86_64 105/138 2026-03-09T20:49:59.733 INFO:teuthology.orchestra.run.vm06.stdout: Running scriptlet: qatlib-service-25.08.0-2.el9.x86_64 105/138 2026-03-09T20:50:00.069 INFO:teuthology.orchestra.run.vm06.stdout: Installing : qatzip-libs-1.3.1-1.el9.x86_64 106/138 2026-03-09T20:50:00.076 INFO:teuthology.orchestra.run.vm06.stdout: Installing : ceph-base-2:19.2.3-678.ge911bdeb.el9.x86_64 107/138 2026-03-09T20:50:00.122 INFO:teuthology.orchestra.run.vm06.stdout: Running scriptlet: ceph-base-2:19.2.3-678.ge911bdeb.el9.x86_64 107/138 2026-03-09T20:50:00.122 INFO:teuthology.orchestra.run.vm06.stdout:Created symlink /etc/systemd/system/multi-user.target.wants/ceph.target → /usr/lib/systemd/system/ceph.target. 2026-03-09T20:50:00.122 INFO:teuthology.orchestra.run.vm06.stdout:Created symlink /etc/systemd/system/ceph.target.wants/ceph-crash.service → /usr/lib/systemd/system/ceph-crash.service. 2026-03-09T20:50:00.122 INFO:teuthology.orchestra.run.vm06.stdout: 2026-03-09T20:50:00.128 INFO:teuthology.orchestra.run.vm06.stdout: Installing : ceph-selinux-2:19.2.3-678.ge911bdeb.el9.x86_64 108/138 2026-03-09T20:50:00.525 INFO:teuthology.orchestra.run.vm01.stdout:(136/136): python3-scipy-1.9.3-2.el9.x86_64.rpm 3.9 MB/s | 19 MB 00:04 2026-03-09T20:50:00.528 INFO:teuthology.orchestra.run.vm01.stdout:-------------------------------------------------------------------------------- 2026-03-09T20:50:00.528 INFO:teuthology.orchestra.run.vm01.stdout:Total 10 MB/s | 210 MB 00:20 2026-03-09T20:50:01.169 INFO:teuthology.orchestra.run.vm01.stdout:Running transaction check 2026-03-09T20:50:01.227 INFO:teuthology.orchestra.run.vm01.stdout:Transaction check succeeded. 2026-03-09T20:50:01.227 INFO:teuthology.orchestra.run.vm01.stdout:Running transaction test 2026-03-09T20:50:02.156 INFO:teuthology.orchestra.run.vm01.stdout:Transaction test succeeded. 2026-03-09T20:50:02.157 INFO:teuthology.orchestra.run.vm01.stdout:Running transaction 2026-03-09T20:50:03.181 INFO:teuthology.orchestra.run.vm01.stdout: Preparing : 1/1 2026-03-09T20:50:03.200 INFO:teuthology.orchestra.run.vm01.stdout: Installing : python3-more-itertools-8.12.0-2.el9.noarch 1/138 2026-03-09T20:50:03.218 INFO:teuthology.orchestra.run.vm01.stdout: Installing : thrift-0.15.0-4.el9.x86_64 2/138 2026-03-09T20:50:03.411 INFO:teuthology.orchestra.run.vm01.stdout: Installing : lttng-ust-2.12.0-6.el9.x86_64 3/138 2026-03-09T20:50:03.414 INFO:teuthology.orchestra.run.vm01.stdout: Upgrading : librados2-2:19.2.3-678.ge911bdeb.el9.x86_64 4/138 2026-03-09T20:50:03.599 INFO:teuthology.orchestra.run.vm01.stdout: Running scriptlet: librados2-2:19.2.3-678.ge911bdeb.el9.x86_64 4/138 2026-03-09T20:50:03.657 INFO:teuthology.orchestra.run.vm01.stdout: Installing : libcephfs2-2:19.2.3-678.ge911bdeb.el9.x86_64 5/138 2026-03-09T20:50:03.994 INFO:teuthology.orchestra.run.vm01.stdout: Running scriptlet: libcephfs2-2:19.2.3-678.ge911bdeb.el9.x86_64 5/138 2026-03-09T20:50:04.043 INFO:teuthology.orchestra.run.vm01.stdout: Installing : python3-rados-2:19.2.3-678.ge911bdeb.el9.x86_64 6/138 2026-03-09T20:50:04.067 INFO:teuthology.orchestra.run.vm01.stdout: Installing : librdkafka-1.6.1-102.el9.x86_64 7/138 2026-03-09T20:50:04.144 INFO:teuthology.orchestra.run.vm01.stdout: Installing : librabbitmq-0.11.0-7.el9.x86_64 8/138 2026-03-09T20:50:04.255 INFO:teuthology.orchestra.run.vm01.stdout: Installing : python3-jaraco-8.2.1-3.el9.noarch 9/138 2026-03-09T20:50:04.519 INFO:teuthology.orchestra.run.vm01.stdout: Installing : libnbd-1.20.3-4.el9.x86_64 10/138 2026-03-09T20:50:04.615 INFO:teuthology.orchestra.run.vm01.stdout: Installing : libcephsqlite-2:19.2.3-678.ge911bdeb.el9.x86_64 11/138 2026-03-09T20:50:04.698 INFO:teuthology.orchestra.run.vm01.stdout: Running scriptlet: libcephsqlite-2:19.2.3-678.ge911bdeb.el9.x86_64 11/138 2026-03-09T20:50:04.700 INFO:teuthology.orchestra.run.vm01.stdout: Installing : libradosstriper1-2:19.2.3-678.ge911bdeb.el9.x86_ 12/138 2026-03-09T20:50:04.769 INFO:teuthology.orchestra.run.vm01.stdout: Running scriptlet: libradosstriper1-2:19.2.3-678.ge911bdeb.el9.x86_ 12/138 2026-03-09T20:50:04.887 INFO:teuthology.orchestra.run.vm01.stdout: Installing : re2-1:20211101-20.el9.x86_64 13/138 2026-03-09T20:50:05.008 INFO:teuthology.orchestra.run.vm01.stdout: Installing : libarrow-9.0.0-15.el9.x86_64 14/138 2026-03-09T20:50:05.014 INFO:teuthology.orchestra.run.vm01.stdout: Installing : python3-werkzeug-2.0.3-3.el9.1.noarch 15/138 2026-03-09T20:50:05.046 INFO:teuthology.orchestra.run.vm01.stdout: Installing : liboath-2.6.12-1.el9.x86_64 16/138 2026-03-09T20:50:05.062 INFO:teuthology.orchestra.run.vm01.stdout: Installing : python3-pyasn1-0.4.8-7.el9.noarch 17/138 2026-03-09T20:50:05.071 INFO:teuthology.orchestra.run.vm01.stdout: Installing : python3-packaging-20.9-5.el9.noarch 18/138 2026-03-09T20:50:05.083 INFO:teuthology.orchestra.run.vm01.stdout: Installing : python3-markupsafe-1.1.1-12.el9.x86_64 19/138 2026-03-09T20:50:05.092 INFO:teuthology.orchestra.run.vm01.stdout: Installing : protobuf-3.14.0-17.el9.x86_64 20/138 2026-03-09T20:50:05.096 INFO:teuthology.orchestra.run.vm01.stdout: Installing : lua-5.4.4-4.el9.x86_64 21/138 2026-03-09T20:50:05.104 INFO:teuthology.orchestra.run.vm01.stdout: Installing : flexiblas-3.0.4-9.el9.x86_64 22/138 2026-03-09T20:50:05.138 INFO:teuthology.orchestra.run.vm01.stdout: Installing : unzip-6.0-59.el9.x86_64 23/138 2026-03-09T20:50:05.157 INFO:teuthology.orchestra.run.vm01.stdout: Installing : python3-urllib3-1.26.5-7.el9.noarch 24/138 2026-03-09T20:50:05.164 INFO:teuthology.orchestra.run.vm01.stdout: Installing : python3-requests-2.25.1-10.el9.noarch 25/138 2026-03-09T20:50:05.174 INFO:teuthology.orchestra.run.vm01.stdout: Installing : libquadmath-11.5.0-14.el9.x86_64 26/138 2026-03-09T20:50:05.176 INFO:teuthology.orchestra.run.vm01.stdout: Installing : libgfortran-11.5.0-14.el9.x86_64 27/138 2026-03-09T20:50:05.218 INFO:teuthology.orchestra.run.vm01.stdout: Installing : ledmon-libs-1.1.0-3.el9.x86_64 28/138 2026-03-09T20:50:05.226 INFO:teuthology.orchestra.run.vm01.stdout: Installing : python3-ceph-common-2:19.2.3-678.ge911bdeb.el9.x 29/138 2026-03-09T20:50:05.239 INFO:teuthology.orchestra.run.vm01.stdout: Installing : python3-ceph-argparse-2:19.2.3-678.ge911bdeb.el9 30/138 2026-03-09T20:50:05.258 INFO:teuthology.orchestra.run.vm01.stdout: Installing : python3-cephfs-2:19.2.3-678.ge911bdeb.el9.x86_64 31/138 2026-03-09T20:50:05.267 INFO:teuthology.orchestra.run.vm01.stdout: Installing : python3-requests-oauthlib-1.3.0-12.el9.noarch 32/138 2026-03-09T20:50:05.298 INFO:teuthology.orchestra.run.vm01.stdout: Installing : zip-3.0-35.el9.x86_64 33/138 2026-03-09T20:50:05.304 INFO:teuthology.orchestra.run.vm01.stdout: Installing : luarocks-3.9.2-5.el9.noarch 34/138 2026-03-09T20:50:05.312 INFO:teuthology.orchestra.run.vm01.stdout: Installing : lua-devel-5.4.4-4.el9.x86_64 35/138 2026-03-09T20:50:05.344 INFO:teuthology.orchestra.run.vm01.stdout: Installing : protobuf-compiler-3.14.0-17.el9.x86_64 36/138 2026-03-09T20:50:05.411 INFO:teuthology.orchestra.run.vm01.stdout: Installing : python3-mako-1.1.4-6.el9.noarch 37/138 2026-03-09T20:50:05.430 INFO:teuthology.orchestra.run.vm01.stdout: Installing : python3-pyasn1-modules-0.4.8-7.el9.noarch 38/138 2026-03-09T20:50:05.438 INFO:teuthology.orchestra.run.vm01.stdout: Installing : python3-rsa-4.9-2.el9.noarch 39/138 2026-03-09T20:50:05.448 INFO:teuthology.orchestra.run.vm01.stdout: Installing : python3-jaraco-classes-3.2.1-5.el9.noarch 40/138 2026-03-09T20:50:05.455 INFO:teuthology.orchestra.run.vm01.stdout: Installing : librados-devel-2:19.2.3-678.ge911bdeb.el9.x86_64 41/138 2026-03-09T20:50:05.460 INFO:teuthology.orchestra.run.vm01.stdout: Installing : python3-zc-lockfile-2.0-10.el9.noarch 42/138 2026-03-09T20:50:05.478 INFO:teuthology.orchestra.run.vm01.stdout: Installing : python3-xmltodict-0.12.0-15.el9.noarch 43/138 2026-03-09T20:50:05.497 INFO:teuthology.orchestra.run.vm08.stdout: Running scriptlet: ceph-selinux-2:19.2.3-678.ge911bdeb.el9.x86_64 108/138 2026-03-09T20:50:05.497 INFO:teuthology.orchestra.run.vm08.stdout:skipping the directory /sys 2026-03-09T20:50:05.497 INFO:teuthology.orchestra.run.vm08.stdout:skipping the directory /proc 2026-03-09T20:50:05.497 INFO:teuthology.orchestra.run.vm08.stdout:skipping the directory /mnt 2026-03-09T20:50:05.497 INFO:teuthology.orchestra.run.vm08.stdout:skipping the directory /var/tmp 2026-03-09T20:50:05.497 INFO:teuthology.orchestra.run.vm08.stdout:skipping the directory /home 2026-03-09T20:50:05.497 INFO:teuthology.orchestra.run.vm08.stdout:skipping the directory /root 2026-03-09T20:50:05.497 INFO:teuthology.orchestra.run.vm08.stdout:skipping the directory /tmp 2026-03-09T20:50:05.497 INFO:teuthology.orchestra.run.vm08.stdout: 2026-03-09T20:50:05.509 INFO:teuthology.orchestra.run.vm01.stdout: Installing : python3-websocket-client-1.2.3-2.el9.noarch 44/138 2026-03-09T20:50:05.516 INFO:teuthology.orchestra.run.vm01.stdout: Installing : python3-webob-1.8.8-2.el9.noarch 45/138 2026-03-09T20:50:05.523 INFO:teuthology.orchestra.run.vm01.stdout: Installing : python3-typing-extensions-4.15.0-1.el9.noarch 46/138 2026-03-09T20:50:05.541 INFO:teuthology.orchestra.run.vm01.stdout: Installing : python3-repoze-lru-0.7-16.el9.noarch 47/138 2026-03-09T20:50:05.553 INFO:teuthology.orchestra.run.vm01.stdout: Installing : python3-routes-2.5.1-5.el9.noarch 48/138 2026-03-09T20:50:05.566 INFO:teuthology.orchestra.run.vm01.stdout: Installing : python3-natsort-7.1.1-5.el9.noarch 49/138 2026-03-09T20:50:05.636 INFO:teuthology.orchestra.run.vm08.stdout: Installing : ceph-mds-2:19.2.3-678.ge911bdeb.el9.x86_64 109/138 2026-03-09T20:50:05.637 INFO:teuthology.orchestra.run.vm01.stdout: Installing : python3-logutils-0.3.5-21.el9.noarch 50/138 2026-03-09T20:50:05.648 INFO:teuthology.orchestra.run.vm01.stdout: Installing : python3-pecan-1.4.2-3.el9.noarch 51/138 2026-03-09T20:50:05.663 INFO:teuthology.orchestra.run.vm08.stdout: Running scriptlet: ceph-mds-2:19.2.3-678.ge911bdeb.el9.x86_64 109/138 2026-03-09T20:50:05.668 INFO:teuthology.orchestra.run.vm08.stdout:Glob pattern passed to enable, but globs are not supported for this. 2026-03-09T20:50:05.668 INFO:teuthology.orchestra.run.vm08.stdout:Invalid unit name "ceph-mds@*.service" escaped as "ceph-mds@\x2a.service". 2026-03-09T20:50:05.668 INFO:teuthology.orchestra.run.vm08.stdout:Created symlink /etc/systemd/system/multi-user.target.wants/ceph-mds.target → /usr/lib/systemd/system/ceph-mds.target. 2026-03-09T20:50:05.668 INFO:teuthology.orchestra.run.vm08.stdout:Created symlink /etc/systemd/system/ceph.target.wants/ceph-mds.target → /usr/lib/systemd/system/ceph-mds.target. 2026-03-09T20:50:05.668 INFO:teuthology.orchestra.run.vm08.stdout: 2026-03-09T20:50:05.668 INFO:teuthology.orchestra.run.vm01.stdout: Installing : python3-certifi-2023.05.07-4.el9.noarch 52/138 2026-03-09T20:50:05.775 INFO:teuthology.orchestra.run.vm01.stdout: Installing : python3-cachetools-4.2.4-1.el9.noarch 53/138 2026-03-09T20:50:05.906 INFO:teuthology.orchestra.run.vm08.stdout: Installing : ceph-mon-2:19.2.3-678.ge911bdeb.el9.x86_64 110/138 2026-03-09T20:50:05.934 INFO:teuthology.orchestra.run.vm08.stdout: Running scriptlet: ceph-mon-2:19.2.3-678.ge911bdeb.el9.x86_64 110/138 2026-03-09T20:50:05.934 INFO:teuthology.orchestra.run.vm08.stdout:Glob pattern passed to enable, but globs are not supported for this. 2026-03-09T20:50:05.934 INFO:teuthology.orchestra.run.vm08.stdout:Invalid unit name "ceph-mon@*.service" escaped as "ceph-mon@\x2a.service". 2026-03-09T20:50:05.934 INFO:teuthology.orchestra.run.vm08.stdout:Created symlink /etc/systemd/system/multi-user.target.wants/ceph-mon.target → /usr/lib/systemd/system/ceph-mon.target. 2026-03-09T20:50:05.935 INFO:teuthology.orchestra.run.vm08.stdout:Created symlink /etc/systemd/system/ceph.target.wants/ceph-mon.target → /usr/lib/systemd/system/ceph-mon.target. 2026-03-09T20:50:05.935 INFO:teuthology.orchestra.run.vm08.stdout: 2026-03-09T20:50:05.944 INFO:teuthology.orchestra.run.vm08.stdout: Installing : mailcap-2.1.49-5.el9.noarch 111/138 2026-03-09T20:50:05.948 INFO:teuthology.orchestra.run.vm08.stdout: Installing : libconfig-1.7.2-9.el9.x86_64 112/138 2026-03-09T20:50:05.971 INFO:teuthology.orchestra.run.vm08.stdout: Running scriptlet: libstoragemgmt-1.10.1-1.el9.x86_64 113/138 2026-03-09T20:50:05.971 INFO:teuthology.orchestra.run.vm08.stdout:Creating group 'qat' with GID 994. 2026-03-09T20:50:05.971 INFO:teuthology.orchestra.run.vm08.stdout:Creating group 'libstoragemgmt' with GID 993. 2026-03-09T20:50:05.971 INFO:teuthology.orchestra.run.vm08.stdout:Creating user 'libstoragemgmt' (daemon account for libstoragemgmt) with UID 993 and GID 993. 2026-03-09T20:50:05.971 INFO:teuthology.orchestra.run.vm08.stdout: 2026-03-09T20:50:05.984 INFO:teuthology.orchestra.run.vm08.stdout: Installing : libstoragemgmt-1.10.1-1.el9.x86_64 113/138 2026-03-09T20:50:06.016 INFO:teuthology.orchestra.run.vm08.stdout: Running scriptlet: libstoragemgmt-1.10.1-1.el9.x86_64 113/138 2026-03-09T20:50:06.016 INFO:teuthology.orchestra.run.vm08.stdout:Created symlink /etc/systemd/system/multi-user.target.wants/libstoragemgmt.service → /usr/lib/systemd/system/libstoragemgmt.service. 2026-03-09T20:50:06.016 INFO:teuthology.orchestra.run.vm08.stdout: 2026-03-09T20:50:06.065 INFO:teuthology.orchestra.run.vm08.stdout: Installing : python3-libstoragemgmt-1.10.1-1.el9.x86_64 114/138 2026-03-09T20:50:06.150 INFO:teuthology.orchestra.run.vm08.stdout: Installing : cryptsetup-2.8.1-3.el9.x86_64 115/138 2026-03-09T20:50:06.155 INFO:teuthology.orchestra.run.vm08.stdout: Installing : ceph-volume-2:19.2.3-678.ge911bdeb.el9.noarch 116/138 2026-03-09T20:50:06.175 INFO:teuthology.orchestra.run.vm08.stdout: Running scriptlet: ceph-volume-2:19.2.3-678.ge911bdeb.el9.noarch 116/138 2026-03-09T20:50:06.175 INFO:teuthology.orchestra.run.vm08.stdout:Glob pattern passed to enable, but globs are not supported for this. 2026-03-09T20:50:06.175 INFO:teuthology.orchestra.run.vm08.stdout:Invalid unit name "ceph-volume@*.service" escaped as "ceph-volume@\x2a.service". 2026-03-09T20:50:06.175 INFO:teuthology.orchestra.run.vm08.stdout: 2026-03-09T20:50:06.211 INFO:teuthology.orchestra.run.vm01.stdout: Installing : python3-google-auth-1:2.45.0-1.el9.noarch 54/138 2026-03-09T20:50:06.233 INFO:teuthology.orchestra.run.vm01.stdout: Installing : python3-kubernetes-1:26.1.0-3.el9.noarch 55/138 2026-03-09T20:50:06.247 INFO:teuthology.orchestra.run.vm01.stdout: Installing : python3-backports-tarfile-1.2.0-1.el9.noarch 56/138 2026-03-09T20:50:06.257 INFO:teuthology.orchestra.run.vm01.stdout: Installing : python3-jaraco-context-6.0.1-3.el9.noarch 57/138 2026-03-09T20:50:06.264 INFO:teuthology.orchestra.run.vm01.stdout: Installing : python3-autocommand-2.2.2-8.el9.noarch 58/138 2026-03-09T20:50:06.275 INFO:teuthology.orchestra.run.vm01.stdout: Installing : libunwind-1.6.2-1.el9.x86_64 59/138 2026-03-09T20:50:06.281 INFO:teuthology.orchestra.run.vm01.stdout: Installing : gperftools-libs-2.9.1-3.el9.x86_64 60/138 2026-03-09T20:50:06.284 INFO:teuthology.orchestra.run.vm01.stdout: Installing : libarrow-doc-9.0.0-15.el9.noarch 61/138 2026-03-09T20:50:06.324 INFO:teuthology.orchestra.run.vm01.stdout: Installing : grpc-data-1.46.7-10.el9.noarch 62/138 2026-03-09T20:50:06.380 INFO:teuthology.orchestra.run.vm01.stdout: Installing : abseil-cpp-20211102.0-4.el9.x86_64 63/138 2026-03-09T20:50:06.398 INFO:teuthology.orchestra.run.vm01.stdout: Installing : python3-grpcio-1.46.7-10.el9.x86_64 64/138 2026-03-09T20:50:06.408 INFO:teuthology.orchestra.run.vm01.stdout: Installing : socat-1.7.4.1-8.el9.x86_64 65/138 2026-03-09T20:50:06.414 INFO:teuthology.orchestra.run.vm01.stdout: Installing : python3-toml-0.10.2-6.el9.noarch 66/138 2026-03-09T20:50:06.424 INFO:teuthology.orchestra.run.vm01.stdout: Installing : python3-jaraco-functools-3.5.0-2.el9.noarch 67/138 2026-03-09T20:50:06.430 INFO:teuthology.orchestra.run.vm01.stdout: Installing : python3-jaraco-text-4.0.0-2.el9.noarch 68/138 2026-03-09T20:50:06.526 INFO:teuthology.orchestra.run.vm01.stdout: Installing : python3-jaraco-collections-3.0.0-8.el9.noarch 69/138 2026-03-09T20:50:06.599 INFO:teuthology.orchestra.run.vm01.stdout: Installing : python3-tempora-5.0.0-2.el9.noarch 70/138 2026-03-09T20:50:06.689 INFO:teuthology.orchestra.run.vm01.stdout: Installing : python3-portend-3.1.0-2.el9.noarch 71/138 2026-03-09T20:50:06.829 INFO:teuthology.orchestra.run.vm01.stdout: Installing : python3-protobuf-3.14.0-17.el9.noarch 72/138 2026-03-09T20:50:06.919 INFO:teuthology.orchestra.run.vm01.stdout: Installing : python3-grpcio-tools-1.46.7-10.el9.x86_64 73/138 2026-03-09T20:50:07.220 INFO:teuthology.orchestra.run.vm08.stdout: Installing : ceph-osd-2:19.2.3-678.ge911bdeb.el9.x86_64 117/138 2026-03-09T20:50:07.250 INFO:teuthology.orchestra.run.vm08.stdout: Running scriptlet: ceph-osd-2:19.2.3-678.ge911bdeb.el9.x86_64 117/138 2026-03-09T20:50:07.251 INFO:teuthology.orchestra.run.vm08.stdout:Glob pattern passed to enable, but globs are not supported for this. 2026-03-09T20:50:07.251 INFO:teuthology.orchestra.run.vm08.stdout:Invalid unit name "ceph-osd@*.service" escaped as "ceph-osd@\x2a.service". 2026-03-09T20:50:07.251 INFO:teuthology.orchestra.run.vm08.stdout:Created symlink /etc/systemd/system/multi-user.target.wants/ceph-osd.target → /usr/lib/systemd/system/ceph-osd.target. 2026-03-09T20:50:07.251 INFO:teuthology.orchestra.run.vm08.stdout:Created symlink /etc/systemd/system/ceph.target.wants/ceph-osd.target → /usr/lib/systemd/system/ceph-osd.target. 2026-03-09T20:50:07.251 INFO:teuthology.orchestra.run.vm08.stdout: 2026-03-09T20:50:07.313 INFO:teuthology.orchestra.run.vm01.stdout: Installing : python3-devel-3.9.25-3.el9.x86_64 74/138 2026-03-09T20:50:07.333 INFO:teuthology.orchestra.run.vm08.stdout: Running scriptlet: cephadm-2:19.2.3-678.ge911bdeb.el9.noarch 118/138 2026-03-09T20:50:07.339 INFO:teuthology.orchestra.run.vm08.stdout: Installing : cephadm-2:19.2.3-678.ge911bdeb.el9.noarch 118/138 2026-03-09T20:50:07.347 INFO:teuthology.orchestra.run.vm08.stdout: Installing : ceph-prometheus-alerts-2:19.2.3-678.ge911bdeb.el 119/138 2026-03-09T20:50:07.358 INFO:teuthology.orchestra.run.vm01.stdout: Installing : python3-babel-2.9.1-2.el9.noarch 75/138 2026-03-09T20:50:07.373 INFO:teuthology.orchestra.run.vm01.stdout: Installing : python3-jinja2-2.11.3-8.el9.noarch 76/138 2026-03-09T20:50:07.374 INFO:teuthology.orchestra.run.vm08.stdout: Installing : ceph-grafana-dashboards-2:19.2.3-678.ge911bdeb.e 120/138 2026-03-09T20:50:07.378 INFO:teuthology.orchestra.run.vm08.stdout: Installing : ceph-mgr-cephadm-2:19.2.3-678.ge911bdeb.el9.noar 121/138 2026-03-09T20:50:07.444 INFO:teuthology.orchestra.run.vm01.stdout: Installing : openblas-0.3.29-1.el9.x86_64 77/138 2026-03-09T20:50:07.448 INFO:teuthology.orchestra.run.vm01.stdout: Installing : openblas-openmp-0.3.29-1.el9.x86_64 78/138 2026-03-09T20:50:07.479 INFO:teuthology.orchestra.run.vm01.stdout: Installing : flexiblas-openblas-openmp-3.0.4-9.el9.x86_64 79/138 2026-03-09T20:50:07.819 INFO:teuthology.orchestra.run.vm06.stdout: Running scriptlet: ceph-selinux-2:19.2.3-678.ge911bdeb.el9.x86_64 108/138 2026-03-09T20:50:07.819 INFO:teuthology.orchestra.run.vm06.stdout:skipping the directory /sys 2026-03-09T20:50:07.819 INFO:teuthology.orchestra.run.vm06.stdout:skipping the directory /proc 2026-03-09T20:50:07.819 INFO:teuthology.orchestra.run.vm06.stdout:skipping the directory /mnt 2026-03-09T20:50:07.819 INFO:teuthology.orchestra.run.vm06.stdout:skipping the directory /var/tmp 2026-03-09T20:50:07.819 INFO:teuthology.orchestra.run.vm06.stdout:skipping the directory /home 2026-03-09T20:50:07.819 INFO:teuthology.orchestra.run.vm06.stdout:skipping the directory /root 2026-03-09T20:50:07.819 INFO:teuthology.orchestra.run.vm06.stdout:skipping the directory /tmp 2026-03-09T20:50:07.819 INFO:teuthology.orchestra.run.vm06.stdout: 2026-03-09T20:50:07.915 INFO:teuthology.orchestra.run.vm01.stdout: Installing : flexiblas-netlib-3.0.4-9.el9.x86_64 80/138 2026-03-09T20:50:07.956 INFO:teuthology.orchestra.run.vm06.stdout: Installing : ceph-mds-2:19.2.3-678.ge911bdeb.el9.x86_64 109/138 2026-03-09T20:50:07.981 INFO:teuthology.orchestra.run.vm06.stdout: Running scriptlet: ceph-mds-2:19.2.3-678.ge911bdeb.el9.x86_64 109/138 2026-03-09T20:50:07.981 INFO:teuthology.orchestra.run.vm06.stdout:Glob pattern passed to enable, but globs are not supported for this. 2026-03-09T20:50:07.981 INFO:teuthology.orchestra.run.vm06.stdout:Invalid unit name "ceph-mds@*.service" escaped as "ceph-mds@\x2a.service". 2026-03-09T20:50:07.981 INFO:teuthology.orchestra.run.vm06.stdout:Created symlink /etc/systemd/system/multi-user.target.wants/ceph-mds.target → /usr/lib/systemd/system/ceph-mds.target. 2026-03-09T20:50:07.981 INFO:teuthology.orchestra.run.vm06.stdout:Created symlink /etc/systemd/system/ceph.target.wants/ceph-mds.target → /usr/lib/systemd/system/ceph-mds.target. 2026-03-09T20:50:07.981 INFO:teuthology.orchestra.run.vm06.stdout: 2026-03-09T20:50:07.985 INFO:teuthology.orchestra.run.vm08.stdout: Running scriptlet: ceph-mgr-cephadm-2:19.2.3-678.ge911bdeb.el9.noar 121/138 2026-03-09T20:50:07.993 INFO:teuthology.orchestra.run.vm08.stdout: Installing : ceph-mgr-dashboard-2:19.2.3-678.ge911bdeb.el9.no 122/138 2026-03-09T20:50:08.012 INFO:teuthology.orchestra.run.vm01.stdout: Installing : python3-numpy-1:1.23.5-2.el9.x86_64 81/138 2026-03-09T20:50:08.228 INFO:teuthology.orchestra.run.vm06.stdout: Installing : ceph-mon-2:19.2.3-678.ge911bdeb.el9.x86_64 110/138 2026-03-09T20:50:08.252 INFO:teuthology.orchestra.run.vm06.stdout: Running scriptlet: ceph-mon-2:19.2.3-678.ge911bdeb.el9.x86_64 110/138 2026-03-09T20:50:08.252 INFO:teuthology.orchestra.run.vm06.stdout:Glob pattern passed to enable, but globs are not supported for this. 2026-03-09T20:50:08.252 INFO:teuthology.orchestra.run.vm06.stdout:Invalid unit name "ceph-mon@*.service" escaped as "ceph-mon@\x2a.service". 2026-03-09T20:50:08.252 INFO:teuthology.orchestra.run.vm06.stdout:Created symlink /etc/systemd/system/multi-user.target.wants/ceph-mon.target → /usr/lib/systemd/system/ceph-mon.target. 2026-03-09T20:50:08.252 INFO:teuthology.orchestra.run.vm06.stdout:Created symlink /etc/systemd/system/ceph.target.wants/ceph-mon.target → /usr/lib/systemd/system/ceph-mon.target. 2026-03-09T20:50:08.252 INFO:teuthology.orchestra.run.vm06.stdout: 2026-03-09T20:50:08.299 INFO:teuthology.orchestra.run.vm06.stdout: Installing : mailcap-2.1.49-5.el9.noarch 111/138 2026-03-09T20:50:08.302 INFO:teuthology.orchestra.run.vm06.stdout: Installing : libconfig-1.7.2-9.el9.x86_64 112/138 2026-03-09T20:50:08.322 INFO:teuthology.orchestra.run.vm06.stdout: Running scriptlet: libstoragemgmt-1.10.1-1.el9.x86_64 113/138 2026-03-09T20:50:08.322 INFO:teuthology.orchestra.run.vm06.stdout:Creating group 'qat' with GID 994. 2026-03-09T20:50:08.322 INFO:teuthology.orchestra.run.vm06.stdout:Creating group 'libstoragemgmt' with GID 993. 2026-03-09T20:50:08.322 INFO:teuthology.orchestra.run.vm06.stdout:Creating user 'libstoragemgmt' (daemon account for libstoragemgmt) with UID 993 and GID 993. 2026-03-09T20:50:08.322 INFO:teuthology.orchestra.run.vm06.stdout: 2026-03-09T20:50:08.339 INFO:teuthology.orchestra.run.vm06.stdout: Installing : libstoragemgmt-1.10.1-1.el9.x86_64 113/138 2026-03-09T20:50:08.368 INFO:teuthology.orchestra.run.vm06.stdout: Running scriptlet: libstoragemgmt-1.10.1-1.el9.x86_64 113/138 2026-03-09T20:50:08.368 INFO:teuthology.orchestra.run.vm06.stdout:Created symlink /etc/systemd/system/multi-user.target.wants/libstoragemgmt.service → /usr/lib/systemd/system/libstoragemgmt.service. 2026-03-09T20:50:08.369 INFO:teuthology.orchestra.run.vm06.stdout: 2026-03-09T20:50:08.417 INFO:teuthology.orchestra.run.vm06.stdout: Installing : python3-libstoragemgmt-1.10.1-1.el9.x86_64 114/138 2026-03-09T20:50:08.501 INFO:teuthology.orchestra.run.vm06.stdout: Installing : cryptsetup-2.8.1-3.el9.x86_64 115/138 2026-03-09T20:50:08.506 INFO:teuthology.orchestra.run.vm06.stdout: Installing : ceph-volume-2:19.2.3-678.ge911bdeb.el9.noarch 116/138 2026-03-09T20:50:08.524 INFO:teuthology.orchestra.run.vm06.stdout: Running scriptlet: ceph-volume-2:19.2.3-678.ge911bdeb.el9.noarch 116/138 2026-03-09T20:50:08.524 INFO:teuthology.orchestra.run.vm06.stdout:Glob pattern passed to enable, but globs are not supported for this. 2026-03-09T20:50:08.524 INFO:teuthology.orchestra.run.vm06.stdout:Invalid unit name "ceph-volume@*.service" escaped as "ceph-volume@\x2a.service". 2026-03-09T20:50:08.524 INFO:teuthology.orchestra.run.vm06.stdout: 2026-03-09T20:50:08.601 INFO:teuthology.orchestra.run.vm08.stdout: Running scriptlet: ceph-mgr-dashboard-2:19.2.3-678.ge911bdeb.el9.no 122/138 2026-03-09T20:50:08.603 INFO:teuthology.orchestra.run.vm08.stdout: Installing : ceph-mgr-diskprediction-local-2:19.2.3-678.ge911 123/138 2026-03-09T20:50:08.684 INFO:teuthology.orchestra.run.vm08.stdout: Running scriptlet: ceph-mgr-diskprediction-local-2:19.2.3-678.ge911 123/138 2026-03-09T20:50:08.761 INFO:teuthology.orchestra.run.vm08.stdout: Installing : ceph-mgr-modules-core-2:19.2.3-678.ge911bdeb.el9 124/138 2026-03-09T20:50:08.764 INFO:teuthology.orchestra.run.vm08.stdout: Installing : ceph-mgr-2:19.2.3-678.ge911bdeb.el9.x86_64 125/138 2026-03-09T20:50:08.789 INFO:teuthology.orchestra.run.vm08.stdout: Running scriptlet: ceph-mgr-2:19.2.3-678.ge911bdeb.el9.x86_64 125/138 2026-03-09T20:50:08.789 INFO:teuthology.orchestra.run.vm08.stdout:Glob pattern passed to enable, but globs are not supported for this. 2026-03-09T20:50:08.789 INFO:teuthology.orchestra.run.vm08.stdout:Invalid unit name "ceph-mgr@*.service" escaped as "ceph-mgr@\x2a.service". 2026-03-09T20:50:08.789 INFO:teuthology.orchestra.run.vm08.stdout:Created symlink /etc/systemd/system/multi-user.target.wants/ceph-mgr.target → /usr/lib/systemd/system/ceph-mgr.target. 2026-03-09T20:50:08.789 INFO:teuthology.orchestra.run.vm08.stdout:Created symlink /etc/systemd/system/ceph.target.wants/ceph-mgr.target → /usr/lib/systemd/system/ceph-mgr.target. 2026-03-09T20:50:08.789 INFO:teuthology.orchestra.run.vm08.stdout: 2026-03-09T20:50:08.804 INFO:teuthology.orchestra.run.vm08.stdout: Installing : ceph-mgr-rook-2:19.2.3-678.ge911bdeb.el9.noarch 126/138 2026-03-09T20:50:08.821 INFO:teuthology.orchestra.run.vm08.stdout: Running scriptlet: ceph-mgr-rook-2:19.2.3-678.ge911bdeb.el9.noarch 126/138 2026-03-09T20:50:08.885 INFO:teuthology.orchestra.run.vm01.stdout: Installing : python3-numpy-f2py-1:1.23.5-2.el9.x86_64 82/138 2026-03-09T20:50:08.917 INFO:teuthology.orchestra.run.vm01.stdout: Installing : python3-scipy-1.9.3-2.el9.x86_64 83/138 2026-03-09T20:50:08.924 INFO:teuthology.orchestra.run.vm01.stdout: Installing : libxslt-1.1.34-12.el9.x86_64 84/138 2026-03-09T20:50:08.931 INFO:teuthology.orchestra.run.vm01.stdout: Installing : xmlstarlet-1.6.1-20.el9.x86_64 85/138 2026-03-09T20:50:09.108 INFO:teuthology.orchestra.run.vm01.stdout: Installing : libpmemobj-1.12.1-1.el9.x86_64 86/138 2026-03-09T20:50:09.112 INFO:teuthology.orchestra.run.vm01.stdout: Upgrading : librbd1-2:19.2.3-678.ge911bdeb.el9.x86_64 87/138 2026-03-09T20:50:09.156 INFO:teuthology.orchestra.run.vm01.stdout: Running scriptlet: librbd1-2:19.2.3-678.ge911bdeb.el9.x86_64 87/138 2026-03-09T20:50:09.161 INFO:teuthology.orchestra.run.vm01.stdout: Installing : python3-rbd-2:19.2.3-678.ge911bdeb.el9.x86_64 88/138 2026-03-09T20:50:09.171 INFO:teuthology.orchestra.run.vm01.stdout: Installing : boost-program-options-1.75.0-13.el9.x86_64 89/138 2026-03-09T20:50:09.394 INFO:teuthology.orchestra.run.vm06.stdout: Installing : ceph-osd-2:19.2.3-678.ge911bdeb.el9.x86_64 117/138 2026-03-09T20:50:09.397 INFO:teuthology.orchestra.run.vm08.stdout: Installing : ceph-2:19.2.3-678.ge911bdeb.el9.x86_64 127/138 2026-03-09T20:50:09.406 INFO:teuthology.orchestra.run.vm08.stdout: Installing : ceph-radosgw-2:19.2.3-678.ge911bdeb.el9.x86_64 128/138 2026-03-09T20:50:09.422 INFO:teuthology.orchestra.run.vm06.stdout: Running scriptlet: ceph-osd-2:19.2.3-678.ge911bdeb.el9.x86_64 117/138 2026-03-09T20:50:09.422 INFO:teuthology.orchestra.run.vm06.stdout:Glob pattern passed to enable, but globs are not supported for this. 2026-03-09T20:50:09.422 INFO:teuthology.orchestra.run.vm06.stdout:Invalid unit name "ceph-osd@*.service" escaped as "ceph-osd@\x2a.service". 2026-03-09T20:50:09.422 INFO:teuthology.orchestra.run.vm06.stdout:Created symlink /etc/systemd/system/multi-user.target.wants/ceph-osd.target → /usr/lib/systemd/system/ceph-osd.target. 2026-03-09T20:50:09.422 INFO:teuthology.orchestra.run.vm06.stdout:Created symlink /etc/systemd/system/ceph.target.wants/ceph-osd.target → /usr/lib/systemd/system/ceph-osd.target. 2026-03-09T20:50:09.422 INFO:teuthology.orchestra.run.vm06.stdout: 2026-03-09T20:50:09.436 INFO:teuthology.orchestra.run.vm08.stdout: Running scriptlet: ceph-radosgw-2:19.2.3-678.ge911bdeb.el9.x86_64 128/138 2026-03-09T20:50:09.437 INFO:teuthology.orchestra.run.vm08.stdout:Glob pattern passed to enable, but globs are not supported for this. 2026-03-09T20:50:09.437 INFO:teuthology.orchestra.run.vm08.stdout:Invalid unit name "ceph-radosgw@*.service" escaped as "ceph-radosgw@\x2a.service". 2026-03-09T20:50:09.437 INFO:teuthology.orchestra.run.vm08.stdout:Created symlink /etc/systemd/system/multi-user.target.wants/ceph-radosgw.target → /usr/lib/systemd/system/ceph-radosgw.target. 2026-03-09T20:50:09.437 INFO:teuthology.orchestra.run.vm08.stdout:Created symlink /etc/systemd/system/ceph.target.wants/ceph-radosgw.target → /usr/lib/systemd/system/ceph-radosgw.target. 2026-03-09T20:50:09.437 INFO:teuthology.orchestra.run.vm08.stdout: 2026-03-09T20:50:09.449 INFO:teuthology.orchestra.run.vm08.stdout: Installing : ceph-immutable-object-cache-2:19.2.3-678.ge911bd 129/138 2026-03-09T20:50:09.464 INFO:teuthology.orchestra.run.vm01.stdout: Installing : parquet-libs-9.0.0-15.el9.x86_64 90/138 2026-03-09T20:50:09.478 INFO:teuthology.orchestra.run.vm08.stdout: Running scriptlet: ceph-immutable-object-cache-2:19.2.3-678.ge911bd 129/138 2026-03-09T20:50:09.478 INFO:teuthology.orchestra.run.vm08.stdout:Glob pattern passed to enable, but globs are not supported for this. 2026-03-09T20:50:09.478 INFO:teuthology.orchestra.run.vm08.stdout:Invalid unit name "ceph-immutable-object-cache@*.service" escaped as "ceph-immutable-object-cache@\x2a.service". 2026-03-09T20:50:09.478 INFO:teuthology.orchestra.run.vm08.stdout: 2026-03-09T20:50:09.485 INFO:teuthology.orchestra.run.vm06.stdout: Running scriptlet: cephadm-2:19.2.3-678.ge911bdeb.el9.noarch 118/138 2026-03-09T20:50:09.572 INFO:teuthology.orchestra.run.vm01.stdout: Installing : librgw2-2:19.2.3-678.ge911bdeb.el9.x86_64 91/138 2026-03-09T20:50:09.574 INFO:teuthology.orchestra.run.vm06.stdout: Installing : cephadm-2:19.2.3-678.ge911bdeb.el9.noarch 118/138 2026-03-09T20:50:09.581 INFO:teuthology.orchestra.run.vm06.stdout: Installing : ceph-prometheus-alerts-2:19.2.3-678.ge911bdeb.el 119/138 2026-03-09T20:50:09.595 INFO:teuthology.orchestra.run.vm01.stdout: Running scriptlet: librgw2-2:19.2.3-678.ge911bdeb.el9.x86_64 91/138 2026-03-09T20:50:09.598 INFO:teuthology.orchestra.run.vm01.stdout: Installing : python3-rgw-2:19.2.3-678.ge911bdeb.el9.x86_64 92/138 2026-03-09T20:50:09.608 INFO:teuthology.orchestra.run.vm06.stdout: Installing : ceph-grafana-dashboards-2:19.2.3-678.ge911bdeb.e 120/138 2026-03-09T20:50:09.612 INFO:teuthology.orchestra.run.vm06.stdout: Installing : ceph-mgr-cephadm-2:19.2.3-678.ge911bdeb.el9.noar 121/138 2026-03-09T20:50:09.713 INFO:teuthology.orchestra.run.vm08.stdout: Installing : rbd-mirror-2:19.2.3-678.ge911bdeb.el9.x86_64 130/138 2026-03-09T20:50:09.738 INFO:teuthology.orchestra.run.vm08.stdout: Running scriptlet: rbd-mirror-2:19.2.3-678.ge911bdeb.el9.x86_64 130/138 2026-03-09T20:50:09.739 INFO:teuthology.orchestra.run.vm08.stdout:Glob pattern passed to enable, but globs are not supported for this. 2026-03-09T20:50:09.739 INFO:teuthology.orchestra.run.vm08.stdout:Invalid unit name "ceph-rbd-mirror@*.service" escaped as "ceph-rbd-mirror@\x2a.service". 2026-03-09T20:50:09.739 INFO:teuthology.orchestra.run.vm08.stdout:Created symlink /etc/systemd/system/multi-user.target.wants/ceph-rbd-mirror.target → /usr/lib/systemd/system/ceph-rbd-mirror.target. 2026-03-09T20:50:09.739 INFO:teuthology.orchestra.run.vm08.stdout:Created symlink /etc/systemd/system/ceph.target.wants/ceph-rbd-mirror.target → /usr/lib/systemd/system/ceph-rbd-mirror.target. 2026-03-09T20:50:09.739 INFO:teuthology.orchestra.run.vm08.stdout: 2026-03-09T20:50:10.211 INFO:teuthology.orchestra.run.vm06.stdout: Running scriptlet: ceph-mgr-cephadm-2:19.2.3-678.ge911bdeb.el9.noar 121/138 2026-03-09T20:50:10.218 INFO:teuthology.orchestra.run.vm06.stdout: Installing : ceph-mgr-dashboard-2:19.2.3-678.ge911bdeb.el9.no 122/138 2026-03-09T20:50:10.818 INFO:teuthology.orchestra.run.vm01.stdout: Running scriptlet: ceph-common-2:19.2.3-678.ge911bdeb.el9.x86_64 93/138 2026-03-09T20:50:10.824 INFO:teuthology.orchestra.run.vm01.stdout: Installing : ceph-common-2:19.2.3-678.ge911bdeb.el9.x86_64 93/138 2026-03-09T20:50:10.850 INFO:teuthology.orchestra.run.vm01.stdout: Running scriptlet: ceph-common-2:19.2.3-678.ge911bdeb.el9.x86_64 93/138 2026-03-09T20:50:10.858 INFO:teuthology.orchestra.run.vm06.stdout: Running scriptlet: ceph-mgr-dashboard-2:19.2.3-678.ge911bdeb.el9.no 122/138 2026-03-09T20:50:10.860 INFO:teuthology.orchestra.run.vm06.stdout: Installing : ceph-mgr-diskprediction-local-2:19.2.3-678.ge911 123/138 2026-03-09T20:50:10.872 INFO:teuthology.orchestra.run.vm01.stdout: Installing : python3-ply-3.11-14.el9.noarch 94/138 2026-03-09T20:50:10.894 INFO:teuthology.orchestra.run.vm01.stdout: Installing : python3-pycparser-2.20-6.el9.noarch 95/138 2026-03-09T20:50:10.933 INFO:teuthology.orchestra.run.vm06.stdout: Running scriptlet: ceph-mgr-diskprediction-local-2:19.2.3-678.ge911 123/138 2026-03-09T20:50:10.992 INFO:teuthology.orchestra.run.vm01.stdout: Installing : python3-cffi-1.14.5-5.el9.x86_64 96/138 2026-03-09T20:50:11.010 INFO:teuthology.orchestra.run.vm06.stdout: Installing : ceph-mgr-modules-core-2:19.2.3-678.ge911bdeb.el9 124/138 2026-03-09T20:50:11.013 INFO:teuthology.orchestra.run.vm01.stdout: Installing : python3-cryptography-36.0.1-5.el9.x86_64 97/138 2026-03-09T20:50:11.015 INFO:teuthology.orchestra.run.vm06.stdout: Installing : ceph-mgr-2:19.2.3-678.ge911bdeb.el9.x86_64 125/138 2026-03-09T20:50:11.044 INFO:teuthology.orchestra.run.vm06.stdout: Running scriptlet: ceph-mgr-2:19.2.3-678.ge911bdeb.el9.x86_64 125/138 2026-03-09T20:50:11.044 INFO:teuthology.orchestra.run.vm06.stdout:Glob pattern passed to enable, but globs are not supported for this. 2026-03-09T20:50:11.044 INFO:teuthology.orchestra.run.vm06.stdout:Invalid unit name "ceph-mgr@*.service" escaped as "ceph-mgr@\x2a.service". 2026-03-09T20:50:11.044 INFO:teuthology.orchestra.run.vm06.stdout:Created symlink /etc/systemd/system/multi-user.target.wants/ceph-mgr.target → /usr/lib/systemd/system/ceph-mgr.target. 2026-03-09T20:50:11.044 INFO:teuthology.orchestra.run.vm06.stdout:Created symlink /etc/systemd/system/ceph.target.wants/ceph-mgr.target → /usr/lib/systemd/system/ceph-mgr.target. 2026-03-09T20:50:11.044 INFO:teuthology.orchestra.run.vm06.stdout: 2026-03-09T20:50:11.048 INFO:teuthology.orchestra.run.vm01.stdout: Installing : python3-pyOpenSSL-21.0.0-1.el9.noarch 98/138 2026-03-09T20:50:11.060 INFO:teuthology.orchestra.run.vm06.stdout: Installing : ceph-mgr-rook-2:19.2.3-678.ge911bdeb.el9.noarch 126/138 2026-03-09T20:50:11.073 INFO:teuthology.orchestra.run.vm06.stdout: Running scriptlet: ceph-mgr-rook-2:19.2.3-678.ge911bdeb.el9.noarch 126/138 2026-03-09T20:50:11.092 INFO:teuthology.orchestra.run.vm01.stdout: Installing : python3-cheroot-10.0.1-4.el9.noarch 99/138 2026-03-09T20:50:11.160 INFO:teuthology.orchestra.run.vm01.stdout: Installing : python3-cherrypy-18.6.1-2.el9.noarch 100/138 2026-03-09T20:50:11.176 INFO:teuthology.orchestra.run.vm01.stdout: Installing : python3-asyncssh-2.13.2-5.el9.noarch 101/138 2026-03-09T20:50:11.185 INFO:teuthology.orchestra.run.vm01.stdout: Installing : python3-bcrypt-3.2.2-1.el9.x86_64 102/138 2026-03-09T20:50:11.194 INFO:teuthology.orchestra.run.vm01.stdout: Installing : pciutils-3.7.0-7.el9.x86_64 103/138 2026-03-09T20:50:11.200 INFO:teuthology.orchestra.run.vm01.stdout: Installing : qatlib-25.08.0-2.el9.x86_64 104/138 2026-03-09T20:50:11.202 INFO:teuthology.orchestra.run.vm01.stdout: Installing : qatlib-service-25.08.0-2.el9.x86_64 105/138 2026-03-09T20:50:11.223 INFO:teuthology.orchestra.run.vm01.stdout: Running scriptlet: qatlib-service-25.08.0-2.el9.x86_64 105/138 2026-03-09T20:50:11.556 INFO:teuthology.orchestra.run.vm01.stdout: Installing : qatzip-libs-1.3.1-1.el9.x86_64 106/138 2026-03-09T20:50:11.563 INFO:teuthology.orchestra.run.vm01.stdout: Installing : ceph-base-2:19.2.3-678.ge911bdeb.el9.x86_64 107/138 2026-03-09T20:50:11.602 INFO:teuthology.orchestra.run.vm01.stdout: Running scriptlet: ceph-base-2:19.2.3-678.ge911bdeb.el9.x86_64 107/138 2026-03-09T20:50:11.602 INFO:teuthology.orchestra.run.vm01.stdout:Created symlink /etc/systemd/system/multi-user.target.wants/ceph.target → /usr/lib/systemd/system/ceph.target. 2026-03-09T20:50:11.602 INFO:teuthology.orchestra.run.vm01.stdout:Created symlink /etc/systemd/system/ceph.target.wants/ceph-crash.service → /usr/lib/systemd/system/ceph-crash.service. 2026-03-09T20:50:11.602 INFO:teuthology.orchestra.run.vm01.stdout: 2026-03-09T20:50:11.608 INFO:teuthology.orchestra.run.vm01.stdout: Installing : ceph-selinux-2:19.2.3-678.ge911bdeb.el9.x86_64 108/138 2026-03-09T20:50:11.660 INFO:teuthology.orchestra.run.vm06.stdout: Installing : ceph-2:19.2.3-678.ge911bdeb.el9.x86_64 127/138 2026-03-09T20:50:11.669 INFO:teuthology.orchestra.run.vm06.stdout: Installing : ceph-radosgw-2:19.2.3-678.ge911bdeb.el9.x86_64 128/138 2026-03-09T20:50:11.697 INFO:teuthology.orchestra.run.vm06.stdout: Running scriptlet: ceph-radosgw-2:19.2.3-678.ge911bdeb.el9.x86_64 128/138 2026-03-09T20:50:11.697 INFO:teuthology.orchestra.run.vm06.stdout:Glob pattern passed to enable, but globs are not supported for this. 2026-03-09T20:50:11.697 INFO:teuthology.orchestra.run.vm06.stdout:Invalid unit name "ceph-radosgw@*.service" escaped as "ceph-radosgw@\x2a.service". 2026-03-09T20:50:11.697 INFO:teuthology.orchestra.run.vm06.stdout:Created symlink /etc/systemd/system/multi-user.target.wants/ceph-radosgw.target → /usr/lib/systemd/system/ceph-radosgw.target. 2026-03-09T20:50:11.697 INFO:teuthology.orchestra.run.vm06.stdout:Created symlink /etc/systemd/system/ceph.target.wants/ceph-radosgw.target → /usr/lib/systemd/system/ceph-radosgw.target. 2026-03-09T20:50:11.697 INFO:teuthology.orchestra.run.vm06.stdout: 2026-03-09T20:50:11.710 INFO:teuthology.orchestra.run.vm06.stdout: Installing : ceph-immutable-object-cache-2:19.2.3-678.ge911bd 129/138 2026-03-09T20:50:11.738 INFO:teuthology.orchestra.run.vm06.stdout: Running scriptlet: ceph-immutable-object-cache-2:19.2.3-678.ge911bd 129/138 2026-03-09T20:50:11.738 INFO:teuthology.orchestra.run.vm06.stdout:Glob pattern passed to enable, but globs are not supported for this. 2026-03-09T20:50:11.738 INFO:teuthology.orchestra.run.vm06.stdout:Invalid unit name "ceph-immutable-object-cache@*.service" escaped as "ceph-immutable-object-cache@\x2a.service". 2026-03-09T20:50:11.738 INFO:teuthology.orchestra.run.vm06.stdout: 2026-03-09T20:50:11.912 INFO:teuthology.orchestra.run.vm06.stdout: Installing : rbd-mirror-2:19.2.3-678.ge911bdeb.el9.x86_64 130/138 2026-03-09T20:50:11.942 INFO:teuthology.orchestra.run.vm06.stdout: Running scriptlet: rbd-mirror-2:19.2.3-678.ge911bdeb.el9.x86_64 130/138 2026-03-09T20:50:11.942 INFO:teuthology.orchestra.run.vm06.stdout:Glob pattern passed to enable, but globs are not supported for this. 2026-03-09T20:50:11.942 INFO:teuthology.orchestra.run.vm06.stdout:Invalid unit name "ceph-rbd-mirror@*.service" escaped as "ceph-rbd-mirror@\x2a.service". 2026-03-09T20:50:11.942 INFO:teuthology.orchestra.run.vm06.stdout:Created symlink /etc/systemd/system/multi-user.target.wants/ceph-rbd-mirror.target → /usr/lib/systemd/system/ceph-rbd-mirror.target. 2026-03-09T20:50:11.943 INFO:teuthology.orchestra.run.vm06.stdout:Created symlink /etc/systemd/system/ceph.target.wants/ceph-rbd-mirror.target → /usr/lib/systemd/system/ceph-rbd-mirror.target. 2026-03-09T20:50:11.943 INFO:teuthology.orchestra.run.vm06.stdout: 2026-03-09T20:50:12.920 INFO:teuthology.orchestra.run.vm08.stdout: Installing : ceph-test-2:19.2.3-678.ge911bdeb.el9.x86_64 131/138 2026-03-09T20:50:12.934 INFO:teuthology.orchestra.run.vm08.stdout: Installing : rbd-fuse-2:19.2.3-678.ge911bdeb.el9.x86_64 132/138 2026-03-09T20:50:12.940 INFO:teuthology.orchestra.run.vm08.stdout: Installing : rbd-nbd-2:19.2.3-678.ge911bdeb.el9.x86_64 133/138 2026-03-09T20:50:13.007 INFO:teuthology.orchestra.run.vm08.stdout: Installing : libcephfs-devel-2:19.2.3-678.ge911bdeb.el9.x86_6 134/138 2026-03-09T20:50:13.018 INFO:teuthology.orchestra.run.vm08.stdout: Installing : ceph-fuse-2:19.2.3-678.ge911bdeb.el9.x86_64 135/138 2026-03-09T20:50:13.023 INFO:teuthology.orchestra.run.vm08.stdout: Installing : python3-jmespath-1.0.1-1.el9.noarch 136/138 2026-03-09T20:50:13.023 INFO:teuthology.orchestra.run.vm08.stdout: Cleanup : librbd1-2:16.2.4-5.el9.x86_64 137/138 2026-03-09T20:50:13.042 INFO:teuthology.orchestra.run.vm08.stdout: Running scriptlet: librbd1-2:16.2.4-5.el9.x86_64 137/138 2026-03-09T20:50:13.042 INFO:teuthology.orchestra.run.vm08.stdout: Cleanup : librados2-2:16.2.4-5.el9.x86_64 138/138 2026-03-09T20:50:14.512 INFO:teuthology.orchestra.run.vm08.stdout: Running scriptlet: librados2-2:16.2.4-5.el9.x86_64 138/138 2026-03-09T20:50:14.512 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : ceph-2:19.2.3-678.ge911bdeb.el9.x86_64 1/138 2026-03-09T20:50:14.512 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : ceph-base-2:19.2.3-678.ge911bdeb.el9.x86_64 2/138 2026-03-09T20:50:14.512 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : ceph-common-2:19.2.3-678.ge911bdeb.el9.x86_64 3/138 2026-03-09T20:50:14.512 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : ceph-fuse-2:19.2.3-678.ge911bdeb.el9.x86_64 4/138 2026-03-09T20:50:14.512 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : ceph-immutable-object-cache-2:19.2.3-678.ge911bd 5/138 2026-03-09T20:50:14.512 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : ceph-mds-2:19.2.3-678.ge911bdeb.el9.x86_64 6/138 2026-03-09T20:50:14.512 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : ceph-mgr-2:19.2.3-678.ge911bdeb.el9.x86_64 7/138 2026-03-09T20:50:14.512 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : ceph-mon-2:19.2.3-678.ge911bdeb.el9.x86_64 8/138 2026-03-09T20:50:14.512 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : ceph-osd-2:19.2.3-678.ge911bdeb.el9.x86_64 9/138 2026-03-09T20:50:14.512 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : ceph-radosgw-2:19.2.3-678.ge911bdeb.el9.x86_64 10/138 2026-03-09T20:50:14.513 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : ceph-selinux-2:19.2.3-678.ge911bdeb.el9.x86_64 11/138 2026-03-09T20:50:14.513 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : ceph-test-2:19.2.3-678.ge911bdeb.el9.x86_64 12/138 2026-03-09T20:50:14.513 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : libcephfs-devel-2:19.2.3-678.ge911bdeb.el9.x86_6 13/138 2026-03-09T20:50:14.513 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : libcephfs2-2:19.2.3-678.ge911bdeb.el9.x86_64 14/138 2026-03-09T20:50:14.513 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : libcephsqlite-2:19.2.3-678.ge911bdeb.el9.x86_64 15/138 2026-03-09T20:50:14.513 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : librados-devel-2:19.2.3-678.ge911bdeb.el9.x86_64 16/138 2026-03-09T20:50:14.513 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : libradosstriper1-2:19.2.3-678.ge911bdeb.el9.x86_ 17/138 2026-03-09T20:50:14.513 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : librgw2-2:19.2.3-678.ge911bdeb.el9.x86_64 18/138 2026-03-09T20:50:14.513 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : python3-ceph-argparse-2:19.2.3-678.ge911bdeb.el9 19/138 2026-03-09T20:50:14.513 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : python3-ceph-common-2:19.2.3-678.ge911bdeb.el9.x 20/138 2026-03-09T20:50:14.513 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : python3-cephfs-2:19.2.3-678.ge911bdeb.el9.x86_64 21/138 2026-03-09T20:50:14.513 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : python3-rados-2:19.2.3-678.ge911bdeb.el9.x86_64 22/138 2026-03-09T20:50:14.513 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : python3-rbd-2:19.2.3-678.ge911bdeb.el9.x86_64 23/138 2026-03-09T20:50:14.513 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : python3-rgw-2:19.2.3-678.ge911bdeb.el9.x86_64 24/138 2026-03-09T20:50:14.513 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : rbd-fuse-2:19.2.3-678.ge911bdeb.el9.x86_64 25/138 2026-03-09T20:50:14.513 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : rbd-mirror-2:19.2.3-678.ge911bdeb.el9.x86_64 26/138 2026-03-09T20:50:14.513 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : rbd-nbd-2:19.2.3-678.ge911bdeb.el9.x86_64 27/138 2026-03-09T20:50:14.513 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : ceph-grafana-dashboards-2:19.2.3-678.ge911bdeb.e 28/138 2026-03-09T20:50:14.513 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : ceph-mgr-cephadm-2:19.2.3-678.ge911bdeb.el9.noar 29/138 2026-03-09T20:50:14.513 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : ceph-mgr-dashboard-2:19.2.3-678.ge911bdeb.el9.no 30/138 2026-03-09T20:50:14.513 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : ceph-mgr-diskprediction-local-2:19.2.3-678.ge911 31/138 2026-03-09T20:50:14.513 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : ceph-mgr-modules-core-2:19.2.3-678.ge911bdeb.el9 32/138 2026-03-09T20:50:14.513 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : ceph-mgr-rook-2:19.2.3-678.ge911bdeb.el9.noarch 33/138 2026-03-09T20:50:14.513 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : ceph-prometheus-alerts-2:19.2.3-678.ge911bdeb.el 34/138 2026-03-09T20:50:14.513 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : ceph-volume-2:19.2.3-678.ge911bdeb.el9.noarch 35/138 2026-03-09T20:50:14.515 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : cephadm-2:19.2.3-678.ge911bdeb.el9.noarch 36/138 2026-03-09T20:50:14.515 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : cryptsetup-2.8.1-3.el9.x86_64 37/138 2026-03-09T20:50:14.515 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : ledmon-libs-1.1.0-3.el9.x86_64 38/138 2026-03-09T20:50:14.515 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : libconfig-1.7.2-9.el9.x86_64 39/138 2026-03-09T20:50:14.515 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : libgfortran-11.5.0-14.el9.x86_64 40/138 2026-03-09T20:50:14.515 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : libquadmath-11.5.0-14.el9.x86_64 41/138 2026-03-09T20:50:14.515 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : mailcap-2.1.49-5.el9.noarch 42/138 2026-03-09T20:50:14.515 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : pciutils-3.7.0-7.el9.x86_64 43/138 2026-03-09T20:50:14.515 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : python3-cffi-1.14.5-5.el9.x86_64 44/138 2026-03-09T20:50:14.515 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : python3-cryptography-36.0.1-5.el9.x86_64 45/138 2026-03-09T20:50:14.515 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : python3-ply-3.11-14.el9.noarch 46/138 2026-03-09T20:50:14.515 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : python3-pycparser-2.20-6.el9.noarch 47/138 2026-03-09T20:50:14.515 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : python3-requests-2.25.1-10.el9.noarch 48/138 2026-03-09T20:50:14.515 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : python3-urllib3-1.26.5-7.el9.noarch 49/138 2026-03-09T20:50:14.515 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : unzip-6.0-59.el9.x86_64 50/138 2026-03-09T20:50:14.515 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : zip-3.0-35.el9.x86_64 51/138 2026-03-09T20:50:14.515 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : boost-program-options-1.75.0-13.el9.x86_64 52/138 2026-03-09T20:50:14.515 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : flexiblas-3.0.4-9.el9.x86_64 53/138 2026-03-09T20:50:14.516 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : flexiblas-netlib-3.0.4-9.el9.x86_64 54/138 2026-03-09T20:50:14.516 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : flexiblas-openblas-openmp-3.0.4-9.el9.x86_64 55/138 2026-03-09T20:50:14.516 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : libnbd-1.20.3-4.el9.x86_64 56/138 2026-03-09T20:50:14.516 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : libpmemobj-1.12.1-1.el9.x86_64 57/138 2026-03-09T20:50:14.516 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : librabbitmq-0.11.0-7.el9.x86_64 58/138 2026-03-09T20:50:14.516 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : librdkafka-1.6.1-102.el9.x86_64 59/138 2026-03-09T20:50:14.516 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : libstoragemgmt-1.10.1-1.el9.x86_64 60/138 2026-03-09T20:50:14.516 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : libxslt-1.1.34-12.el9.x86_64 61/138 2026-03-09T20:50:14.516 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : lttng-ust-2.12.0-6.el9.x86_64 62/138 2026-03-09T20:50:14.516 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : lua-5.4.4-4.el9.x86_64 63/138 2026-03-09T20:50:14.516 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : openblas-0.3.29-1.el9.x86_64 64/138 2026-03-09T20:50:14.516 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : openblas-openmp-0.3.29-1.el9.x86_64 65/138 2026-03-09T20:50:14.516 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : protobuf-3.14.0-17.el9.x86_64 66/138 2026-03-09T20:50:14.516 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : python3-babel-2.9.1-2.el9.noarch 67/138 2026-03-09T20:50:14.516 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : python3-devel-3.9.25-3.el9.x86_64 68/138 2026-03-09T20:50:14.516 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : python3-jinja2-2.11.3-8.el9.noarch 69/138 2026-03-09T20:50:14.516 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : python3-jmespath-1.0.1-1.el9.noarch 70/138 2026-03-09T20:50:14.516 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : python3-libstoragemgmt-1.10.1-1.el9.x86_64 71/138 2026-03-09T20:50:14.516 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : python3-mako-1.1.4-6.el9.noarch 72/138 2026-03-09T20:50:14.516 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : python3-markupsafe-1.1.1-12.el9.x86_64 73/138 2026-03-09T20:50:14.516 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : python3-numpy-1:1.23.5-2.el9.x86_64 74/138 2026-03-09T20:50:14.516 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : python3-numpy-f2py-1:1.23.5-2.el9.x86_64 75/138 2026-03-09T20:50:14.516 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : python3-packaging-20.9-5.el9.noarch 76/138 2026-03-09T20:50:14.516 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : python3-protobuf-3.14.0-17.el9.noarch 77/138 2026-03-09T20:50:14.516 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : python3-pyasn1-0.4.8-7.el9.noarch 78/138 2026-03-09T20:50:14.516 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : python3-pyasn1-modules-0.4.8-7.el9.noarch 79/138 2026-03-09T20:50:14.516 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : python3-requests-oauthlib-1.3.0-12.el9.noarch 80/138 2026-03-09T20:50:14.516 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : python3-scipy-1.9.3-2.el9.x86_64 81/138 2026-03-09T20:50:14.516 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : python3-toml-0.10.2-6.el9.noarch 82/138 2026-03-09T20:50:14.516 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : qatlib-25.08.0-2.el9.x86_64 83/138 2026-03-09T20:50:14.516 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : qatlib-service-25.08.0-2.el9.x86_64 84/138 2026-03-09T20:50:14.516 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : qatzip-libs-1.3.1-1.el9.x86_64 85/138 2026-03-09T20:50:14.516 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : socat-1.7.4.1-8.el9.x86_64 86/138 2026-03-09T20:50:14.516 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : xmlstarlet-1.6.1-20.el9.x86_64 87/138 2026-03-09T20:50:14.516 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : lua-devel-5.4.4-4.el9.x86_64 88/138 2026-03-09T20:50:14.516 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : protobuf-compiler-3.14.0-17.el9.x86_64 89/138 2026-03-09T20:50:14.516 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : abseil-cpp-20211102.0-4.el9.x86_64 90/138 2026-03-09T20:50:14.516 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : gperftools-libs-2.9.1-3.el9.x86_64 91/138 2026-03-09T20:50:14.516 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : grpc-data-1.46.7-10.el9.noarch 92/138 2026-03-09T20:50:14.516 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : libarrow-9.0.0-15.el9.x86_64 93/138 2026-03-09T20:50:14.516 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : libarrow-doc-9.0.0-15.el9.noarch 94/138 2026-03-09T20:50:14.516 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : liboath-2.6.12-1.el9.x86_64 95/138 2026-03-09T20:50:14.516 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : libunwind-1.6.2-1.el9.x86_64 96/138 2026-03-09T20:50:14.516 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : luarocks-3.9.2-5.el9.noarch 97/138 2026-03-09T20:50:14.516 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : parquet-libs-9.0.0-15.el9.x86_64 98/138 2026-03-09T20:50:14.516 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : python3-asyncssh-2.13.2-5.el9.noarch 99/138 2026-03-09T20:50:14.516 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : python3-autocommand-2.2.2-8.el9.noarch 100/138 2026-03-09T20:50:14.516 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : python3-backports-tarfile-1.2.0-1.el9.noarch 101/138 2026-03-09T20:50:14.516 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : python3-bcrypt-3.2.2-1.el9.x86_64 102/138 2026-03-09T20:50:14.516 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : python3-cachetools-4.2.4-1.el9.noarch 103/138 2026-03-09T20:50:14.516 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : python3-certifi-2023.05.07-4.el9.noarch 104/138 2026-03-09T20:50:14.516 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : python3-cheroot-10.0.1-4.el9.noarch 105/138 2026-03-09T20:50:14.517 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : python3-cherrypy-18.6.1-2.el9.noarch 106/138 2026-03-09T20:50:14.517 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : python3-google-auth-1:2.45.0-1.el9.noarch 107/138 2026-03-09T20:50:14.517 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : python3-grpcio-1.46.7-10.el9.x86_64 108/138 2026-03-09T20:50:14.517 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : python3-grpcio-tools-1.46.7-10.el9.x86_64 109/138 2026-03-09T20:50:14.517 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : python3-jaraco-8.2.1-3.el9.noarch 110/138 2026-03-09T20:50:14.517 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : python3-jaraco-classes-3.2.1-5.el9.noarch 111/138 2026-03-09T20:50:14.517 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : python3-jaraco-collections-3.0.0-8.el9.noarch 112/138 2026-03-09T20:50:14.517 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : python3-jaraco-context-6.0.1-3.el9.noarch 113/138 2026-03-09T20:50:14.517 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : python3-jaraco-functools-3.5.0-2.el9.noarch 114/138 2026-03-09T20:50:14.517 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : python3-jaraco-text-4.0.0-2.el9.noarch 115/138 2026-03-09T20:50:14.517 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : python3-kubernetes-1:26.1.0-3.el9.noarch 116/138 2026-03-09T20:50:14.517 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : python3-logutils-0.3.5-21.el9.noarch 117/138 2026-03-09T20:50:14.517 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : python3-more-itertools-8.12.0-2.el9.noarch 118/138 2026-03-09T20:50:14.517 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : python3-natsort-7.1.1-5.el9.noarch 119/138 2026-03-09T20:50:14.517 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : python3-pecan-1.4.2-3.el9.noarch 120/138 2026-03-09T20:50:14.517 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : python3-portend-3.1.0-2.el9.noarch 121/138 2026-03-09T20:50:14.517 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : python3-pyOpenSSL-21.0.0-1.el9.noarch 122/138 2026-03-09T20:50:14.517 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : python3-repoze-lru-0.7-16.el9.noarch 123/138 2026-03-09T20:50:14.517 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : python3-routes-2.5.1-5.el9.noarch 124/138 2026-03-09T20:50:14.517 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : python3-rsa-4.9-2.el9.noarch 125/138 2026-03-09T20:50:14.517 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : python3-tempora-5.0.0-2.el9.noarch 126/138 2026-03-09T20:50:14.517 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : python3-typing-extensions-4.15.0-1.el9.noarch 127/138 2026-03-09T20:50:14.517 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : python3-webob-1.8.8-2.el9.noarch 128/138 2026-03-09T20:50:14.517 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : python3-websocket-client-1.2.3-2.el9.noarch 129/138 2026-03-09T20:50:14.517 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : python3-werkzeug-2.0.3-3.el9.1.noarch 130/138 2026-03-09T20:50:14.517 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : python3-xmltodict-0.12.0-15.el9.noarch 131/138 2026-03-09T20:50:14.517 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : python3-zc-lockfile-2.0-10.el9.noarch 132/138 2026-03-09T20:50:14.517 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : re2-1:20211101-20.el9.x86_64 133/138 2026-03-09T20:50:14.517 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : thrift-0.15.0-4.el9.x86_64 134/138 2026-03-09T20:50:14.517 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : librados2-2:19.2.3-678.ge911bdeb.el9.x86_64 135/138 2026-03-09T20:50:14.517 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : librados2-2:16.2.4-5.el9.x86_64 136/138 2026-03-09T20:50:14.517 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : librbd1-2:19.2.3-678.ge911bdeb.el9.x86_64 137/138 2026-03-09T20:50:14.627 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : librbd1-2:16.2.4-5.el9.x86_64 138/138 2026-03-09T20:50:14.627 INFO:teuthology.orchestra.run.vm08.stdout: 2026-03-09T20:50:14.627 INFO:teuthology.orchestra.run.vm08.stdout:Upgraded: 2026-03-09T20:50:14.627 INFO:teuthology.orchestra.run.vm08.stdout: librados2-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T20:50:14.627 INFO:teuthology.orchestra.run.vm08.stdout: librbd1-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T20:50:14.627 INFO:teuthology.orchestra.run.vm08.stdout:Installed: 2026-03-09T20:50:14.627 INFO:teuthology.orchestra.run.vm08.stdout: abseil-cpp-20211102.0-4.el9.x86_64 2026-03-09T20:50:14.627 INFO:teuthology.orchestra.run.vm08.stdout: boost-program-options-1.75.0-13.el9.x86_64 2026-03-09T20:50:14.627 INFO:teuthology.orchestra.run.vm08.stdout: ceph-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T20:50:14.627 INFO:teuthology.orchestra.run.vm08.stdout: ceph-base-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T20:50:14.627 INFO:teuthology.orchestra.run.vm08.stdout: ceph-common-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T20:50:14.627 INFO:teuthology.orchestra.run.vm08.stdout: ceph-fuse-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T20:50:14.627 INFO:teuthology.orchestra.run.vm08.stdout: ceph-grafana-dashboards-2:19.2.3-678.ge911bdeb.el9.noarch 2026-03-09T20:50:14.627 INFO:teuthology.orchestra.run.vm08.stdout: ceph-immutable-object-cache-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T20:50:14.627 INFO:teuthology.orchestra.run.vm08.stdout: ceph-mds-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T20:50:14.627 INFO:teuthology.orchestra.run.vm08.stdout: ceph-mgr-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T20:50:14.627 INFO:teuthology.orchestra.run.vm08.stdout: ceph-mgr-cephadm-2:19.2.3-678.ge911bdeb.el9.noarch 2026-03-09T20:50:14.627 INFO:teuthology.orchestra.run.vm08.stdout: ceph-mgr-dashboard-2:19.2.3-678.ge911bdeb.el9.noarch 2026-03-09T20:50:14.627 INFO:teuthology.orchestra.run.vm08.stdout: ceph-mgr-diskprediction-local-2:19.2.3-678.ge911bdeb.el9.noarch 2026-03-09T20:50:14.627 INFO:teuthology.orchestra.run.vm08.stdout: ceph-mgr-modules-core-2:19.2.3-678.ge911bdeb.el9.noarch 2026-03-09T20:50:14.627 INFO:teuthology.orchestra.run.vm08.stdout: ceph-mgr-rook-2:19.2.3-678.ge911bdeb.el9.noarch 2026-03-09T20:50:14.627 INFO:teuthology.orchestra.run.vm08.stdout: ceph-mon-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T20:50:14.627 INFO:teuthology.orchestra.run.vm08.stdout: ceph-osd-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T20:50:14.627 INFO:teuthology.orchestra.run.vm08.stdout: ceph-prometheus-alerts-2:19.2.3-678.ge911bdeb.el9.noarch 2026-03-09T20:50:14.627 INFO:teuthology.orchestra.run.vm08.stdout: ceph-radosgw-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T20:50:14.627 INFO:teuthology.orchestra.run.vm08.stdout: ceph-selinux-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T20:50:14.627 INFO:teuthology.orchestra.run.vm08.stdout: ceph-test-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T20:50:14.627 INFO:teuthology.orchestra.run.vm08.stdout: ceph-volume-2:19.2.3-678.ge911bdeb.el9.noarch 2026-03-09T20:50:14.627 INFO:teuthology.orchestra.run.vm08.stdout: cephadm-2:19.2.3-678.ge911bdeb.el9.noarch 2026-03-09T20:50:14.627 INFO:teuthology.orchestra.run.vm08.stdout: cryptsetup-2.8.1-3.el9.x86_64 2026-03-09T20:50:14.627 INFO:teuthology.orchestra.run.vm08.stdout: flexiblas-3.0.4-9.el9.x86_64 2026-03-09T20:50:14.627 INFO:teuthology.orchestra.run.vm08.stdout: flexiblas-netlib-3.0.4-9.el9.x86_64 2026-03-09T20:50:14.627 INFO:teuthology.orchestra.run.vm08.stdout: flexiblas-openblas-openmp-3.0.4-9.el9.x86_64 2026-03-09T20:50:14.627 INFO:teuthology.orchestra.run.vm08.stdout: gperftools-libs-2.9.1-3.el9.x86_64 2026-03-09T20:50:14.627 INFO:teuthology.orchestra.run.vm08.stdout: grpc-data-1.46.7-10.el9.noarch 2026-03-09T20:50:14.627 INFO:teuthology.orchestra.run.vm08.stdout: ledmon-libs-1.1.0-3.el9.x86_64 2026-03-09T20:50:14.627 INFO:teuthology.orchestra.run.vm08.stdout: libarrow-9.0.0-15.el9.x86_64 2026-03-09T20:50:14.627 INFO:teuthology.orchestra.run.vm08.stdout: libarrow-doc-9.0.0-15.el9.noarch 2026-03-09T20:50:14.627 INFO:teuthology.orchestra.run.vm08.stdout: libcephfs-devel-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T20:50:14.627 INFO:teuthology.orchestra.run.vm08.stdout: libcephfs2-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T20:50:14.627 INFO:teuthology.orchestra.run.vm08.stdout: libcephsqlite-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T20:50:14.627 INFO:teuthology.orchestra.run.vm08.stdout: libconfig-1.7.2-9.el9.x86_64 2026-03-09T20:50:14.627 INFO:teuthology.orchestra.run.vm08.stdout: libgfortran-11.5.0-14.el9.x86_64 2026-03-09T20:50:14.627 INFO:teuthology.orchestra.run.vm08.stdout: libnbd-1.20.3-4.el9.x86_64 2026-03-09T20:50:14.627 INFO:teuthology.orchestra.run.vm08.stdout: liboath-2.6.12-1.el9.x86_64 2026-03-09T20:50:14.628 INFO:teuthology.orchestra.run.vm08.stdout: libpmemobj-1.12.1-1.el9.x86_64 2026-03-09T20:50:14.628 INFO:teuthology.orchestra.run.vm08.stdout: libquadmath-11.5.0-14.el9.x86_64 2026-03-09T20:50:14.628 INFO:teuthology.orchestra.run.vm08.stdout: librabbitmq-0.11.0-7.el9.x86_64 2026-03-09T20:50:14.628 INFO:teuthology.orchestra.run.vm08.stdout: librados-devel-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T20:50:14.628 INFO:teuthology.orchestra.run.vm08.stdout: libradosstriper1-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T20:50:14.628 INFO:teuthology.orchestra.run.vm08.stdout: librdkafka-1.6.1-102.el9.x86_64 2026-03-09T20:50:14.628 INFO:teuthology.orchestra.run.vm08.stdout: librgw2-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T20:50:14.628 INFO:teuthology.orchestra.run.vm08.stdout: libstoragemgmt-1.10.1-1.el9.x86_64 2026-03-09T20:50:14.628 INFO:teuthology.orchestra.run.vm08.stdout: libunwind-1.6.2-1.el9.x86_64 2026-03-09T20:50:14.628 INFO:teuthology.orchestra.run.vm08.stdout: libxslt-1.1.34-12.el9.x86_64 2026-03-09T20:50:14.628 INFO:teuthology.orchestra.run.vm08.stdout: lttng-ust-2.12.0-6.el9.x86_64 2026-03-09T20:50:14.628 INFO:teuthology.orchestra.run.vm08.stdout: lua-5.4.4-4.el9.x86_64 2026-03-09T20:50:14.628 INFO:teuthology.orchestra.run.vm08.stdout: lua-devel-5.4.4-4.el9.x86_64 2026-03-09T20:50:14.628 INFO:teuthology.orchestra.run.vm08.stdout: luarocks-3.9.2-5.el9.noarch 2026-03-09T20:50:14.628 INFO:teuthology.orchestra.run.vm08.stdout: mailcap-2.1.49-5.el9.noarch 2026-03-09T20:50:14.628 INFO:teuthology.orchestra.run.vm08.stdout: openblas-0.3.29-1.el9.x86_64 2026-03-09T20:50:14.628 INFO:teuthology.orchestra.run.vm08.stdout: openblas-openmp-0.3.29-1.el9.x86_64 2026-03-09T20:50:14.628 INFO:teuthology.orchestra.run.vm08.stdout: parquet-libs-9.0.0-15.el9.x86_64 2026-03-09T20:50:14.628 INFO:teuthology.orchestra.run.vm08.stdout: pciutils-3.7.0-7.el9.x86_64 2026-03-09T20:50:14.628 INFO:teuthology.orchestra.run.vm08.stdout: protobuf-3.14.0-17.el9.x86_64 2026-03-09T20:50:14.628 INFO:teuthology.orchestra.run.vm08.stdout: protobuf-compiler-3.14.0-17.el9.x86_64 2026-03-09T20:50:14.628 INFO:teuthology.orchestra.run.vm08.stdout: python3-asyncssh-2.13.2-5.el9.noarch 2026-03-09T20:50:14.628 INFO:teuthology.orchestra.run.vm08.stdout: python3-autocommand-2.2.2-8.el9.noarch 2026-03-09T20:50:14.628 INFO:teuthology.orchestra.run.vm08.stdout: python3-babel-2.9.1-2.el9.noarch 2026-03-09T20:50:14.628 INFO:teuthology.orchestra.run.vm08.stdout: python3-backports-tarfile-1.2.0-1.el9.noarch 2026-03-09T20:50:14.628 INFO:teuthology.orchestra.run.vm08.stdout: python3-bcrypt-3.2.2-1.el9.x86_64 2026-03-09T20:50:14.628 INFO:teuthology.orchestra.run.vm08.stdout: python3-cachetools-4.2.4-1.el9.noarch 2026-03-09T20:50:14.628 INFO:teuthology.orchestra.run.vm08.stdout: python3-ceph-argparse-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T20:50:14.628 INFO:teuthology.orchestra.run.vm08.stdout: python3-ceph-common-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T20:50:14.628 INFO:teuthology.orchestra.run.vm08.stdout: python3-cephfs-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T20:50:14.628 INFO:teuthology.orchestra.run.vm08.stdout: python3-certifi-2023.05.07-4.el9.noarch 2026-03-09T20:50:14.628 INFO:teuthology.orchestra.run.vm08.stdout: python3-cffi-1.14.5-5.el9.x86_64 2026-03-09T20:50:14.628 INFO:teuthology.orchestra.run.vm08.stdout: python3-cheroot-10.0.1-4.el9.noarch 2026-03-09T20:50:14.628 INFO:teuthology.orchestra.run.vm08.stdout: python3-cherrypy-18.6.1-2.el9.noarch 2026-03-09T20:50:14.628 INFO:teuthology.orchestra.run.vm08.stdout: python3-cryptography-36.0.1-5.el9.x86_64 2026-03-09T20:50:14.628 INFO:teuthology.orchestra.run.vm08.stdout: python3-devel-3.9.25-3.el9.x86_64 2026-03-09T20:50:14.628 INFO:teuthology.orchestra.run.vm08.stdout: python3-google-auth-1:2.45.0-1.el9.noarch 2026-03-09T20:50:14.628 INFO:teuthology.orchestra.run.vm08.stdout: python3-grpcio-1.46.7-10.el9.x86_64 2026-03-09T20:50:14.628 INFO:teuthology.orchestra.run.vm08.stdout: python3-grpcio-tools-1.46.7-10.el9.x86_64 2026-03-09T20:50:14.628 INFO:teuthology.orchestra.run.vm08.stdout: python3-jaraco-8.2.1-3.el9.noarch 2026-03-09T20:50:14.628 INFO:teuthology.orchestra.run.vm08.stdout: python3-jaraco-classes-3.2.1-5.el9.noarch 2026-03-09T20:50:14.628 INFO:teuthology.orchestra.run.vm08.stdout: python3-jaraco-collections-3.0.0-8.el9.noarch 2026-03-09T20:50:14.628 INFO:teuthology.orchestra.run.vm08.stdout: python3-jaraco-context-6.0.1-3.el9.noarch 2026-03-09T20:50:14.628 INFO:teuthology.orchestra.run.vm08.stdout: python3-jaraco-functools-3.5.0-2.el9.noarch 2026-03-09T20:50:14.628 INFO:teuthology.orchestra.run.vm08.stdout: python3-jaraco-text-4.0.0-2.el9.noarch 2026-03-09T20:50:14.628 INFO:teuthology.orchestra.run.vm08.stdout: python3-jinja2-2.11.3-8.el9.noarch 2026-03-09T20:50:14.628 INFO:teuthology.orchestra.run.vm08.stdout: python3-jmespath-1.0.1-1.el9.noarch 2026-03-09T20:50:14.628 INFO:teuthology.orchestra.run.vm08.stdout: python3-kubernetes-1:26.1.0-3.el9.noarch 2026-03-09T20:50:14.628 INFO:teuthology.orchestra.run.vm08.stdout: python3-libstoragemgmt-1.10.1-1.el9.x86_64 2026-03-09T20:50:14.628 INFO:teuthology.orchestra.run.vm08.stdout: python3-logutils-0.3.5-21.el9.noarch 2026-03-09T20:50:14.628 INFO:teuthology.orchestra.run.vm08.stdout: python3-mako-1.1.4-6.el9.noarch 2026-03-09T20:50:14.628 INFO:teuthology.orchestra.run.vm08.stdout: python3-markupsafe-1.1.1-12.el9.x86_64 2026-03-09T20:50:14.628 INFO:teuthology.orchestra.run.vm08.stdout: python3-more-itertools-8.12.0-2.el9.noarch 2026-03-09T20:50:14.628 INFO:teuthology.orchestra.run.vm08.stdout: python3-natsort-7.1.1-5.el9.noarch 2026-03-09T20:50:14.628 INFO:teuthology.orchestra.run.vm08.stdout: python3-numpy-1:1.23.5-2.el9.x86_64 2026-03-09T20:50:14.628 INFO:teuthology.orchestra.run.vm08.stdout: python3-numpy-f2py-1:1.23.5-2.el9.x86_64 2026-03-09T20:50:14.628 INFO:teuthology.orchestra.run.vm08.stdout: python3-packaging-20.9-5.el9.noarch 2026-03-09T20:50:14.628 INFO:teuthology.orchestra.run.vm08.stdout: python3-pecan-1.4.2-3.el9.noarch 2026-03-09T20:50:14.628 INFO:teuthology.orchestra.run.vm08.stdout: python3-ply-3.11-14.el9.noarch 2026-03-09T20:50:14.628 INFO:teuthology.orchestra.run.vm08.stdout: python3-portend-3.1.0-2.el9.noarch 2026-03-09T20:50:14.628 INFO:teuthology.orchestra.run.vm08.stdout: python3-protobuf-3.14.0-17.el9.noarch 2026-03-09T20:50:14.629 INFO:teuthology.orchestra.run.vm08.stdout: python3-pyOpenSSL-21.0.0-1.el9.noarch 2026-03-09T20:50:14.629 INFO:teuthology.orchestra.run.vm08.stdout: python3-pyasn1-0.4.8-7.el9.noarch 2026-03-09T20:50:14.629 INFO:teuthology.orchestra.run.vm08.stdout: python3-pyasn1-modules-0.4.8-7.el9.noarch 2026-03-09T20:50:14.629 INFO:teuthology.orchestra.run.vm08.stdout: python3-pycparser-2.20-6.el9.noarch 2026-03-09T20:50:14.629 INFO:teuthology.orchestra.run.vm08.stdout: python3-rados-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T20:50:14.629 INFO:teuthology.orchestra.run.vm08.stdout: python3-rbd-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T20:50:14.629 INFO:teuthology.orchestra.run.vm08.stdout: python3-repoze-lru-0.7-16.el9.noarch 2026-03-09T20:50:14.629 INFO:teuthology.orchestra.run.vm08.stdout: python3-requests-2.25.1-10.el9.noarch 2026-03-09T20:50:14.629 INFO:teuthology.orchestra.run.vm08.stdout: python3-requests-oauthlib-1.3.0-12.el9.noarch 2026-03-09T20:50:14.629 INFO:teuthology.orchestra.run.vm08.stdout: python3-rgw-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T20:50:14.629 INFO:teuthology.orchestra.run.vm08.stdout: python3-routes-2.5.1-5.el9.noarch 2026-03-09T20:50:14.629 INFO:teuthology.orchestra.run.vm08.stdout: python3-rsa-4.9-2.el9.noarch 2026-03-09T20:50:14.629 INFO:teuthology.orchestra.run.vm08.stdout: python3-scipy-1.9.3-2.el9.x86_64 2026-03-09T20:50:14.629 INFO:teuthology.orchestra.run.vm08.stdout: python3-tempora-5.0.0-2.el9.noarch 2026-03-09T20:50:14.629 INFO:teuthology.orchestra.run.vm08.stdout: python3-toml-0.10.2-6.el9.noarch 2026-03-09T20:50:14.629 INFO:teuthology.orchestra.run.vm08.stdout: python3-typing-extensions-4.15.0-1.el9.noarch 2026-03-09T20:50:14.629 INFO:teuthology.orchestra.run.vm08.stdout: python3-urllib3-1.26.5-7.el9.noarch 2026-03-09T20:50:14.629 INFO:teuthology.orchestra.run.vm08.stdout: python3-webob-1.8.8-2.el9.noarch 2026-03-09T20:50:14.629 INFO:teuthology.orchestra.run.vm08.stdout: python3-websocket-client-1.2.3-2.el9.noarch 2026-03-09T20:50:14.629 INFO:teuthology.orchestra.run.vm08.stdout: python3-werkzeug-2.0.3-3.el9.1.noarch 2026-03-09T20:50:14.629 INFO:teuthology.orchestra.run.vm08.stdout: python3-xmltodict-0.12.0-15.el9.noarch 2026-03-09T20:50:14.629 INFO:teuthology.orchestra.run.vm08.stdout: python3-zc-lockfile-2.0-10.el9.noarch 2026-03-09T20:50:14.629 INFO:teuthology.orchestra.run.vm08.stdout: qatlib-25.08.0-2.el9.x86_64 2026-03-09T20:50:14.629 INFO:teuthology.orchestra.run.vm08.stdout: qatlib-service-25.08.0-2.el9.x86_64 2026-03-09T20:50:14.629 INFO:teuthology.orchestra.run.vm08.stdout: qatzip-libs-1.3.1-1.el9.x86_64 2026-03-09T20:50:14.629 INFO:teuthology.orchestra.run.vm08.stdout: rbd-fuse-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T20:50:14.629 INFO:teuthology.orchestra.run.vm08.stdout: rbd-mirror-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T20:50:14.629 INFO:teuthology.orchestra.run.vm08.stdout: rbd-nbd-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T20:50:14.629 INFO:teuthology.orchestra.run.vm08.stdout: re2-1:20211101-20.el9.x86_64 2026-03-09T20:50:14.629 INFO:teuthology.orchestra.run.vm08.stdout: socat-1.7.4.1-8.el9.x86_64 2026-03-09T20:50:14.629 INFO:teuthology.orchestra.run.vm08.stdout: thrift-0.15.0-4.el9.x86_64 2026-03-09T20:50:14.629 INFO:teuthology.orchestra.run.vm08.stdout: unzip-6.0-59.el9.x86_64 2026-03-09T20:50:14.629 INFO:teuthology.orchestra.run.vm08.stdout: xmlstarlet-1.6.1-20.el9.x86_64 2026-03-09T20:50:14.629 INFO:teuthology.orchestra.run.vm08.stdout: zip-3.0-35.el9.x86_64 2026-03-09T20:50:14.629 INFO:teuthology.orchestra.run.vm08.stdout: 2026-03-09T20:50:14.629 INFO:teuthology.orchestra.run.vm08.stdout:Complete! 2026-03-09T20:50:14.733 DEBUG:teuthology.parallel:result is None 2026-03-09T20:50:15.019 INFO:teuthology.orchestra.run.vm06.stdout: Installing : ceph-test-2:19.2.3-678.ge911bdeb.el9.x86_64 131/138 2026-03-09T20:50:15.032 INFO:teuthology.orchestra.run.vm06.stdout: Installing : rbd-fuse-2:19.2.3-678.ge911bdeb.el9.x86_64 132/138 2026-03-09T20:50:15.039 INFO:teuthology.orchestra.run.vm06.stdout: Installing : rbd-nbd-2:19.2.3-678.ge911bdeb.el9.x86_64 133/138 2026-03-09T20:50:15.107 INFO:teuthology.orchestra.run.vm06.stdout: Installing : libcephfs-devel-2:19.2.3-678.ge911bdeb.el9.x86_6 134/138 2026-03-09T20:50:15.118 INFO:teuthology.orchestra.run.vm06.stdout: Installing : ceph-fuse-2:19.2.3-678.ge911bdeb.el9.x86_64 135/138 2026-03-09T20:50:15.124 INFO:teuthology.orchestra.run.vm06.stdout: Installing : python3-jmespath-1.0.1-1.el9.noarch 136/138 2026-03-09T20:50:15.124 INFO:teuthology.orchestra.run.vm06.stdout: Cleanup : librbd1-2:16.2.4-5.el9.x86_64 137/138 2026-03-09T20:50:15.147 INFO:teuthology.orchestra.run.vm06.stdout: Running scriptlet: librbd1-2:16.2.4-5.el9.x86_64 137/138 2026-03-09T20:50:15.147 INFO:teuthology.orchestra.run.vm06.stdout: Cleanup : librados2-2:16.2.4-5.el9.x86_64 138/138 2026-03-09T20:50:17.116 INFO:teuthology.orchestra.run.vm06.stdout: Running scriptlet: librados2-2:16.2.4-5.el9.x86_64 138/138 2026-03-09T20:50:17.116 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : ceph-2:19.2.3-678.ge911bdeb.el9.x86_64 1/138 2026-03-09T20:50:17.116 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : ceph-base-2:19.2.3-678.ge911bdeb.el9.x86_64 2/138 2026-03-09T20:50:17.116 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : ceph-common-2:19.2.3-678.ge911bdeb.el9.x86_64 3/138 2026-03-09T20:50:17.116 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : ceph-fuse-2:19.2.3-678.ge911bdeb.el9.x86_64 4/138 2026-03-09T20:50:17.116 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : ceph-immutable-object-cache-2:19.2.3-678.ge911bd 5/138 2026-03-09T20:50:17.116 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : ceph-mds-2:19.2.3-678.ge911bdeb.el9.x86_64 6/138 2026-03-09T20:50:17.116 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : ceph-mgr-2:19.2.3-678.ge911bdeb.el9.x86_64 7/138 2026-03-09T20:50:17.116 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : ceph-mon-2:19.2.3-678.ge911bdeb.el9.x86_64 8/138 2026-03-09T20:50:17.116 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : ceph-osd-2:19.2.3-678.ge911bdeb.el9.x86_64 9/138 2026-03-09T20:50:17.116 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : ceph-radosgw-2:19.2.3-678.ge911bdeb.el9.x86_64 10/138 2026-03-09T20:50:17.116 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : ceph-selinux-2:19.2.3-678.ge911bdeb.el9.x86_64 11/138 2026-03-09T20:50:17.116 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : ceph-test-2:19.2.3-678.ge911bdeb.el9.x86_64 12/138 2026-03-09T20:50:17.116 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : libcephfs-devel-2:19.2.3-678.ge911bdeb.el9.x86_6 13/138 2026-03-09T20:50:17.116 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : libcephfs2-2:19.2.3-678.ge911bdeb.el9.x86_64 14/138 2026-03-09T20:50:17.116 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : libcephsqlite-2:19.2.3-678.ge911bdeb.el9.x86_64 15/138 2026-03-09T20:50:17.116 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : librados-devel-2:19.2.3-678.ge911bdeb.el9.x86_64 16/138 2026-03-09T20:50:17.116 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : libradosstriper1-2:19.2.3-678.ge911bdeb.el9.x86_ 17/138 2026-03-09T20:50:17.116 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : librgw2-2:19.2.3-678.ge911bdeb.el9.x86_64 18/138 2026-03-09T20:50:17.116 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-ceph-argparse-2:19.2.3-678.ge911bdeb.el9 19/138 2026-03-09T20:50:17.116 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-ceph-common-2:19.2.3-678.ge911bdeb.el9.x 20/138 2026-03-09T20:50:17.116 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-cephfs-2:19.2.3-678.ge911bdeb.el9.x86_64 21/138 2026-03-09T20:50:17.116 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-rados-2:19.2.3-678.ge911bdeb.el9.x86_64 22/138 2026-03-09T20:50:17.116 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-rbd-2:19.2.3-678.ge911bdeb.el9.x86_64 23/138 2026-03-09T20:50:17.116 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-rgw-2:19.2.3-678.ge911bdeb.el9.x86_64 24/138 2026-03-09T20:50:17.116 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : rbd-fuse-2:19.2.3-678.ge911bdeb.el9.x86_64 25/138 2026-03-09T20:50:17.116 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : rbd-mirror-2:19.2.3-678.ge911bdeb.el9.x86_64 26/138 2026-03-09T20:50:17.116 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : rbd-nbd-2:19.2.3-678.ge911bdeb.el9.x86_64 27/138 2026-03-09T20:50:17.116 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : ceph-grafana-dashboards-2:19.2.3-678.ge911bdeb.e 28/138 2026-03-09T20:50:17.116 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : ceph-mgr-cephadm-2:19.2.3-678.ge911bdeb.el9.noar 29/138 2026-03-09T20:50:17.116 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : ceph-mgr-dashboard-2:19.2.3-678.ge911bdeb.el9.no 30/138 2026-03-09T20:50:17.116 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : ceph-mgr-diskprediction-local-2:19.2.3-678.ge911 31/138 2026-03-09T20:50:17.116 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : ceph-mgr-modules-core-2:19.2.3-678.ge911bdeb.el9 32/138 2026-03-09T20:50:17.116 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : ceph-mgr-rook-2:19.2.3-678.ge911bdeb.el9.noarch 33/138 2026-03-09T20:50:17.116 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : ceph-prometheus-alerts-2:19.2.3-678.ge911bdeb.el 34/138 2026-03-09T20:50:17.116 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : ceph-volume-2:19.2.3-678.ge911bdeb.el9.noarch 35/138 2026-03-09T20:50:17.118 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : cephadm-2:19.2.3-678.ge911bdeb.el9.noarch 36/138 2026-03-09T20:50:17.118 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : cryptsetup-2.8.1-3.el9.x86_64 37/138 2026-03-09T20:50:17.118 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : ledmon-libs-1.1.0-3.el9.x86_64 38/138 2026-03-09T20:50:17.118 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : libconfig-1.7.2-9.el9.x86_64 39/138 2026-03-09T20:50:17.118 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : libgfortran-11.5.0-14.el9.x86_64 40/138 2026-03-09T20:50:17.119 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : libquadmath-11.5.0-14.el9.x86_64 41/138 2026-03-09T20:50:17.119 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : mailcap-2.1.49-5.el9.noarch 42/138 2026-03-09T20:50:17.119 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : pciutils-3.7.0-7.el9.x86_64 43/138 2026-03-09T20:50:17.119 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-cffi-1.14.5-5.el9.x86_64 44/138 2026-03-09T20:50:17.119 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-cryptography-36.0.1-5.el9.x86_64 45/138 2026-03-09T20:50:17.119 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-ply-3.11-14.el9.noarch 46/138 2026-03-09T20:50:17.119 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-pycparser-2.20-6.el9.noarch 47/138 2026-03-09T20:50:17.119 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-requests-2.25.1-10.el9.noarch 48/138 2026-03-09T20:50:17.119 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-urllib3-1.26.5-7.el9.noarch 49/138 2026-03-09T20:50:17.119 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : unzip-6.0-59.el9.x86_64 50/138 2026-03-09T20:50:17.119 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : zip-3.0-35.el9.x86_64 51/138 2026-03-09T20:50:17.119 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : boost-program-options-1.75.0-13.el9.x86_64 52/138 2026-03-09T20:50:17.119 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : flexiblas-3.0.4-9.el9.x86_64 53/138 2026-03-09T20:50:17.119 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : flexiblas-netlib-3.0.4-9.el9.x86_64 54/138 2026-03-09T20:50:17.119 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : flexiblas-openblas-openmp-3.0.4-9.el9.x86_64 55/138 2026-03-09T20:50:17.119 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : libnbd-1.20.3-4.el9.x86_64 56/138 2026-03-09T20:50:17.119 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : libpmemobj-1.12.1-1.el9.x86_64 57/138 2026-03-09T20:50:17.119 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : librabbitmq-0.11.0-7.el9.x86_64 58/138 2026-03-09T20:50:17.119 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : librdkafka-1.6.1-102.el9.x86_64 59/138 2026-03-09T20:50:17.119 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : libstoragemgmt-1.10.1-1.el9.x86_64 60/138 2026-03-09T20:50:17.119 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : libxslt-1.1.34-12.el9.x86_64 61/138 2026-03-09T20:50:17.119 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : lttng-ust-2.12.0-6.el9.x86_64 62/138 2026-03-09T20:50:17.119 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : lua-5.4.4-4.el9.x86_64 63/138 2026-03-09T20:50:17.119 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : openblas-0.3.29-1.el9.x86_64 64/138 2026-03-09T20:50:17.119 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : openblas-openmp-0.3.29-1.el9.x86_64 65/138 2026-03-09T20:50:17.119 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : protobuf-3.14.0-17.el9.x86_64 66/138 2026-03-09T20:50:17.119 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-babel-2.9.1-2.el9.noarch 67/138 2026-03-09T20:50:17.119 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-devel-3.9.25-3.el9.x86_64 68/138 2026-03-09T20:50:17.119 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-jinja2-2.11.3-8.el9.noarch 69/138 2026-03-09T20:50:17.119 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-jmespath-1.0.1-1.el9.noarch 70/138 2026-03-09T20:50:17.119 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-libstoragemgmt-1.10.1-1.el9.x86_64 71/138 2026-03-09T20:50:17.119 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-mako-1.1.4-6.el9.noarch 72/138 2026-03-09T20:50:17.119 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-markupsafe-1.1.1-12.el9.x86_64 73/138 2026-03-09T20:50:17.119 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-numpy-1:1.23.5-2.el9.x86_64 74/138 2026-03-09T20:50:17.119 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-numpy-f2py-1:1.23.5-2.el9.x86_64 75/138 2026-03-09T20:50:17.119 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-packaging-20.9-5.el9.noarch 76/138 2026-03-09T20:50:17.119 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-protobuf-3.14.0-17.el9.noarch 77/138 2026-03-09T20:50:17.119 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-pyasn1-0.4.8-7.el9.noarch 78/138 2026-03-09T20:50:17.119 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-pyasn1-modules-0.4.8-7.el9.noarch 79/138 2026-03-09T20:50:17.119 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-requests-oauthlib-1.3.0-12.el9.noarch 80/138 2026-03-09T20:50:17.119 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-scipy-1.9.3-2.el9.x86_64 81/138 2026-03-09T20:50:17.119 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-toml-0.10.2-6.el9.noarch 82/138 2026-03-09T20:50:17.119 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : qatlib-25.08.0-2.el9.x86_64 83/138 2026-03-09T20:50:17.119 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : qatlib-service-25.08.0-2.el9.x86_64 84/138 2026-03-09T20:50:17.119 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : qatzip-libs-1.3.1-1.el9.x86_64 85/138 2026-03-09T20:50:17.119 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : socat-1.7.4.1-8.el9.x86_64 86/138 2026-03-09T20:50:17.119 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : xmlstarlet-1.6.1-20.el9.x86_64 87/138 2026-03-09T20:50:17.119 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : lua-devel-5.4.4-4.el9.x86_64 88/138 2026-03-09T20:50:17.119 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : protobuf-compiler-3.14.0-17.el9.x86_64 89/138 2026-03-09T20:50:17.119 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : abseil-cpp-20211102.0-4.el9.x86_64 90/138 2026-03-09T20:50:17.119 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : gperftools-libs-2.9.1-3.el9.x86_64 91/138 2026-03-09T20:50:17.119 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : grpc-data-1.46.7-10.el9.noarch 92/138 2026-03-09T20:50:17.119 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : libarrow-9.0.0-15.el9.x86_64 93/138 2026-03-09T20:50:17.119 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : libarrow-doc-9.0.0-15.el9.noarch 94/138 2026-03-09T20:50:17.119 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : liboath-2.6.12-1.el9.x86_64 95/138 2026-03-09T20:50:17.119 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : libunwind-1.6.2-1.el9.x86_64 96/138 2026-03-09T20:50:17.119 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : luarocks-3.9.2-5.el9.noarch 97/138 2026-03-09T20:50:17.119 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : parquet-libs-9.0.0-15.el9.x86_64 98/138 2026-03-09T20:50:17.119 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-asyncssh-2.13.2-5.el9.noarch 99/138 2026-03-09T20:50:17.119 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-autocommand-2.2.2-8.el9.noarch 100/138 2026-03-09T20:50:17.119 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-backports-tarfile-1.2.0-1.el9.noarch 101/138 2026-03-09T20:50:17.120 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-bcrypt-3.2.2-1.el9.x86_64 102/138 2026-03-09T20:50:17.120 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-cachetools-4.2.4-1.el9.noarch 103/138 2026-03-09T20:50:17.120 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-certifi-2023.05.07-4.el9.noarch 104/138 2026-03-09T20:50:17.120 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-cheroot-10.0.1-4.el9.noarch 105/138 2026-03-09T20:50:17.120 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-cherrypy-18.6.1-2.el9.noarch 106/138 2026-03-09T20:50:17.120 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-google-auth-1:2.45.0-1.el9.noarch 107/138 2026-03-09T20:50:17.120 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-grpcio-1.46.7-10.el9.x86_64 108/138 2026-03-09T20:50:17.120 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-grpcio-tools-1.46.7-10.el9.x86_64 109/138 2026-03-09T20:50:17.120 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-jaraco-8.2.1-3.el9.noarch 110/138 2026-03-09T20:50:17.120 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-jaraco-classes-3.2.1-5.el9.noarch 111/138 2026-03-09T20:50:17.120 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-jaraco-collections-3.0.0-8.el9.noarch 112/138 2026-03-09T20:50:17.120 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-jaraco-context-6.0.1-3.el9.noarch 113/138 2026-03-09T20:50:17.120 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-jaraco-functools-3.5.0-2.el9.noarch 114/138 2026-03-09T20:50:17.120 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-jaraco-text-4.0.0-2.el9.noarch 115/138 2026-03-09T20:50:17.120 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-kubernetes-1:26.1.0-3.el9.noarch 116/138 2026-03-09T20:50:17.120 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-logutils-0.3.5-21.el9.noarch 117/138 2026-03-09T20:50:17.120 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-more-itertools-8.12.0-2.el9.noarch 118/138 2026-03-09T20:50:17.120 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-natsort-7.1.1-5.el9.noarch 119/138 2026-03-09T20:50:17.120 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-pecan-1.4.2-3.el9.noarch 120/138 2026-03-09T20:50:17.120 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-portend-3.1.0-2.el9.noarch 121/138 2026-03-09T20:50:17.120 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-pyOpenSSL-21.0.0-1.el9.noarch 122/138 2026-03-09T20:50:17.120 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-repoze-lru-0.7-16.el9.noarch 123/138 2026-03-09T20:50:17.120 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-routes-2.5.1-5.el9.noarch 124/138 2026-03-09T20:50:17.120 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-rsa-4.9-2.el9.noarch 125/138 2026-03-09T20:50:17.120 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-tempora-5.0.0-2.el9.noarch 126/138 2026-03-09T20:50:17.120 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-typing-extensions-4.15.0-1.el9.noarch 127/138 2026-03-09T20:50:17.120 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-webob-1.8.8-2.el9.noarch 128/138 2026-03-09T20:50:17.120 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-websocket-client-1.2.3-2.el9.noarch 129/138 2026-03-09T20:50:17.120 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-werkzeug-2.0.3-3.el9.1.noarch 130/138 2026-03-09T20:50:17.120 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-xmltodict-0.12.0-15.el9.noarch 131/138 2026-03-09T20:50:17.120 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-zc-lockfile-2.0-10.el9.noarch 132/138 2026-03-09T20:50:17.120 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : re2-1:20211101-20.el9.x86_64 133/138 2026-03-09T20:50:17.120 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : thrift-0.15.0-4.el9.x86_64 134/138 2026-03-09T20:50:17.120 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : librados2-2:19.2.3-678.ge911bdeb.el9.x86_64 135/138 2026-03-09T20:50:17.120 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : librados2-2:16.2.4-5.el9.x86_64 136/138 2026-03-09T20:50:17.120 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : librbd1-2:19.2.3-678.ge911bdeb.el9.x86_64 137/138 2026-03-09T20:50:17.263 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : librbd1-2:16.2.4-5.el9.x86_64 138/138 2026-03-09T20:50:17.263 INFO:teuthology.orchestra.run.vm06.stdout: 2026-03-09T20:50:17.263 INFO:teuthology.orchestra.run.vm06.stdout:Upgraded: 2026-03-09T20:50:17.263 INFO:teuthology.orchestra.run.vm06.stdout: librados2-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T20:50:17.263 INFO:teuthology.orchestra.run.vm06.stdout: librbd1-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T20:50:17.263 INFO:teuthology.orchestra.run.vm06.stdout:Installed: 2026-03-09T20:50:17.263 INFO:teuthology.orchestra.run.vm06.stdout: abseil-cpp-20211102.0-4.el9.x86_64 2026-03-09T20:50:17.263 INFO:teuthology.orchestra.run.vm06.stdout: boost-program-options-1.75.0-13.el9.x86_64 2026-03-09T20:50:17.263 INFO:teuthology.orchestra.run.vm06.stdout: ceph-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T20:50:17.263 INFO:teuthology.orchestra.run.vm06.stdout: ceph-base-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T20:50:17.263 INFO:teuthology.orchestra.run.vm06.stdout: ceph-common-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T20:50:17.263 INFO:teuthology.orchestra.run.vm06.stdout: ceph-fuse-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T20:50:17.263 INFO:teuthology.orchestra.run.vm06.stdout: ceph-grafana-dashboards-2:19.2.3-678.ge911bdeb.el9.noarch 2026-03-09T20:50:17.263 INFO:teuthology.orchestra.run.vm06.stdout: ceph-immutable-object-cache-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T20:50:17.263 INFO:teuthology.orchestra.run.vm06.stdout: ceph-mds-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T20:50:17.263 INFO:teuthology.orchestra.run.vm06.stdout: ceph-mgr-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T20:50:17.263 INFO:teuthology.orchestra.run.vm06.stdout: ceph-mgr-cephadm-2:19.2.3-678.ge911bdeb.el9.noarch 2026-03-09T20:50:17.263 INFO:teuthology.orchestra.run.vm06.stdout: ceph-mgr-dashboard-2:19.2.3-678.ge911bdeb.el9.noarch 2026-03-09T20:50:17.264 INFO:teuthology.orchestra.run.vm06.stdout: ceph-mgr-diskprediction-local-2:19.2.3-678.ge911bdeb.el9.noarch 2026-03-09T20:50:17.264 INFO:teuthology.orchestra.run.vm06.stdout: ceph-mgr-modules-core-2:19.2.3-678.ge911bdeb.el9.noarch 2026-03-09T20:50:17.264 INFO:teuthology.orchestra.run.vm06.stdout: ceph-mgr-rook-2:19.2.3-678.ge911bdeb.el9.noarch 2026-03-09T20:50:17.264 INFO:teuthology.orchestra.run.vm06.stdout: ceph-mon-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T20:50:17.264 INFO:teuthology.orchestra.run.vm06.stdout: ceph-osd-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T20:50:17.264 INFO:teuthology.orchestra.run.vm06.stdout: ceph-prometheus-alerts-2:19.2.3-678.ge911bdeb.el9.noarch 2026-03-09T20:50:17.264 INFO:teuthology.orchestra.run.vm06.stdout: ceph-radosgw-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T20:50:17.264 INFO:teuthology.orchestra.run.vm06.stdout: ceph-selinux-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T20:50:17.264 INFO:teuthology.orchestra.run.vm06.stdout: ceph-test-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T20:50:17.264 INFO:teuthology.orchestra.run.vm06.stdout: ceph-volume-2:19.2.3-678.ge911bdeb.el9.noarch 2026-03-09T20:50:17.264 INFO:teuthology.orchestra.run.vm06.stdout: cephadm-2:19.2.3-678.ge911bdeb.el9.noarch 2026-03-09T20:50:17.264 INFO:teuthology.orchestra.run.vm06.stdout: cryptsetup-2.8.1-3.el9.x86_64 2026-03-09T20:50:17.264 INFO:teuthology.orchestra.run.vm06.stdout: flexiblas-3.0.4-9.el9.x86_64 2026-03-09T20:50:17.264 INFO:teuthology.orchestra.run.vm06.stdout: flexiblas-netlib-3.0.4-9.el9.x86_64 2026-03-09T20:50:17.264 INFO:teuthology.orchestra.run.vm06.stdout: flexiblas-openblas-openmp-3.0.4-9.el9.x86_64 2026-03-09T20:50:17.264 INFO:teuthology.orchestra.run.vm06.stdout: gperftools-libs-2.9.1-3.el9.x86_64 2026-03-09T20:50:17.264 INFO:teuthology.orchestra.run.vm06.stdout: grpc-data-1.46.7-10.el9.noarch 2026-03-09T20:50:17.264 INFO:teuthology.orchestra.run.vm06.stdout: ledmon-libs-1.1.0-3.el9.x86_64 2026-03-09T20:50:17.264 INFO:teuthology.orchestra.run.vm06.stdout: libarrow-9.0.0-15.el9.x86_64 2026-03-09T20:50:17.264 INFO:teuthology.orchestra.run.vm06.stdout: libarrow-doc-9.0.0-15.el9.noarch 2026-03-09T20:50:17.264 INFO:teuthology.orchestra.run.vm06.stdout: libcephfs-devel-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T20:50:17.264 INFO:teuthology.orchestra.run.vm06.stdout: libcephfs2-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T20:50:17.264 INFO:teuthology.orchestra.run.vm06.stdout: libcephsqlite-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T20:50:17.264 INFO:teuthology.orchestra.run.vm06.stdout: libconfig-1.7.2-9.el9.x86_64 2026-03-09T20:50:17.264 INFO:teuthology.orchestra.run.vm06.stdout: libgfortran-11.5.0-14.el9.x86_64 2026-03-09T20:50:17.264 INFO:teuthology.orchestra.run.vm06.stdout: libnbd-1.20.3-4.el9.x86_64 2026-03-09T20:50:17.264 INFO:teuthology.orchestra.run.vm06.stdout: liboath-2.6.12-1.el9.x86_64 2026-03-09T20:50:17.264 INFO:teuthology.orchestra.run.vm06.stdout: libpmemobj-1.12.1-1.el9.x86_64 2026-03-09T20:50:17.264 INFO:teuthology.orchestra.run.vm06.stdout: libquadmath-11.5.0-14.el9.x86_64 2026-03-09T20:50:17.264 INFO:teuthology.orchestra.run.vm06.stdout: librabbitmq-0.11.0-7.el9.x86_64 2026-03-09T20:50:17.264 INFO:teuthology.orchestra.run.vm06.stdout: librados-devel-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T20:50:17.264 INFO:teuthology.orchestra.run.vm06.stdout: libradosstriper1-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T20:50:17.264 INFO:teuthology.orchestra.run.vm06.stdout: librdkafka-1.6.1-102.el9.x86_64 2026-03-09T20:50:17.264 INFO:teuthology.orchestra.run.vm06.stdout: librgw2-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T20:50:17.264 INFO:teuthology.orchestra.run.vm06.stdout: libstoragemgmt-1.10.1-1.el9.x86_64 2026-03-09T20:50:17.264 INFO:teuthology.orchestra.run.vm06.stdout: libunwind-1.6.2-1.el9.x86_64 2026-03-09T20:50:17.264 INFO:teuthology.orchestra.run.vm06.stdout: libxslt-1.1.34-12.el9.x86_64 2026-03-09T20:50:17.264 INFO:teuthology.orchestra.run.vm06.stdout: lttng-ust-2.12.0-6.el9.x86_64 2026-03-09T20:50:17.264 INFO:teuthology.orchestra.run.vm06.stdout: lua-5.4.4-4.el9.x86_64 2026-03-09T20:50:17.264 INFO:teuthology.orchestra.run.vm06.stdout: lua-devel-5.4.4-4.el9.x86_64 2026-03-09T20:50:17.264 INFO:teuthology.orchestra.run.vm06.stdout: luarocks-3.9.2-5.el9.noarch 2026-03-09T20:50:17.264 INFO:teuthology.orchestra.run.vm06.stdout: mailcap-2.1.49-5.el9.noarch 2026-03-09T20:50:17.264 INFO:teuthology.orchestra.run.vm06.stdout: openblas-0.3.29-1.el9.x86_64 2026-03-09T20:50:17.264 INFO:teuthology.orchestra.run.vm06.stdout: openblas-openmp-0.3.29-1.el9.x86_64 2026-03-09T20:50:17.264 INFO:teuthology.orchestra.run.vm06.stdout: parquet-libs-9.0.0-15.el9.x86_64 2026-03-09T20:50:17.264 INFO:teuthology.orchestra.run.vm06.stdout: pciutils-3.7.0-7.el9.x86_64 2026-03-09T20:50:17.264 INFO:teuthology.orchestra.run.vm06.stdout: protobuf-3.14.0-17.el9.x86_64 2026-03-09T20:50:17.264 INFO:teuthology.orchestra.run.vm06.stdout: protobuf-compiler-3.14.0-17.el9.x86_64 2026-03-09T20:50:17.264 INFO:teuthology.orchestra.run.vm06.stdout: python3-asyncssh-2.13.2-5.el9.noarch 2026-03-09T20:50:17.264 INFO:teuthology.orchestra.run.vm06.stdout: python3-autocommand-2.2.2-8.el9.noarch 2026-03-09T20:50:17.264 INFO:teuthology.orchestra.run.vm06.stdout: python3-babel-2.9.1-2.el9.noarch 2026-03-09T20:50:17.264 INFO:teuthology.orchestra.run.vm06.stdout: python3-backports-tarfile-1.2.0-1.el9.noarch 2026-03-09T20:50:17.264 INFO:teuthology.orchestra.run.vm06.stdout: python3-bcrypt-3.2.2-1.el9.x86_64 2026-03-09T20:50:17.264 INFO:teuthology.orchestra.run.vm06.stdout: python3-cachetools-4.2.4-1.el9.noarch 2026-03-09T20:50:17.264 INFO:teuthology.orchestra.run.vm06.stdout: python3-ceph-argparse-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T20:50:17.264 INFO:teuthology.orchestra.run.vm06.stdout: python3-ceph-common-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T20:50:17.264 INFO:teuthology.orchestra.run.vm06.stdout: python3-cephfs-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T20:50:17.264 INFO:teuthology.orchestra.run.vm06.stdout: python3-certifi-2023.05.07-4.el9.noarch 2026-03-09T20:50:17.264 INFO:teuthology.orchestra.run.vm06.stdout: python3-cffi-1.14.5-5.el9.x86_64 2026-03-09T20:50:17.264 INFO:teuthology.orchestra.run.vm06.stdout: python3-cheroot-10.0.1-4.el9.noarch 2026-03-09T20:50:17.264 INFO:teuthology.orchestra.run.vm06.stdout: python3-cherrypy-18.6.1-2.el9.noarch 2026-03-09T20:50:17.264 INFO:teuthology.orchestra.run.vm06.stdout: python3-cryptography-36.0.1-5.el9.x86_64 2026-03-09T20:50:17.265 INFO:teuthology.orchestra.run.vm06.stdout: python3-devel-3.9.25-3.el9.x86_64 2026-03-09T20:50:17.265 INFO:teuthology.orchestra.run.vm06.stdout: python3-google-auth-1:2.45.0-1.el9.noarch 2026-03-09T20:50:17.265 INFO:teuthology.orchestra.run.vm06.stdout: python3-grpcio-1.46.7-10.el9.x86_64 2026-03-09T20:50:17.265 INFO:teuthology.orchestra.run.vm06.stdout: python3-grpcio-tools-1.46.7-10.el9.x86_64 2026-03-09T20:50:17.265 INFO:teuthology.orchestra.run.vm06.stdout: python3-jaraco-8.2.1-3.el9.noarch 2026-03-09T20:50:17.265 INFO:teuthology.orchestra.run.vm06.stdout: python3-jaraco-classes-3.2.1-5.el9.noarch 2026-03-09T20:50:17.265 INFO:teuthology.orchestra.run.vm06.stdout: python3-jaraco-collections-3.0.0-8.el9.noarch 2026-03-09T20:50:17.265 INFO:teuthology.orchestra.run.vm06.stdout: python3-jaraco-context-6.0.1-3.el9.noarch 2026-03-09T20:50:17.265 INFO:teuthology.orchestra.run.vm06.stdout: python3-jaraco-functools-3.5.0-2.el9.noarch 2026-03-09T20:50:17.265 INFO:teuthology.orchestra.run.vm06.stdout: python3-jaraco-text-4.0.0-2.el9.noarch 2026-03-09T20:50:17.265 INFO:teuthology.orchestra.run.vm06.stdout: python3-jinja2-2.11.3-8.el9.noarch 2026-03-09T20:50:17.265 INFO:teuthology.orchestra.run.vm06.stdout: python3-jmespath-1.0.1-1.el9.noarch 2026-03-09T20:50:17.265 INFO:teuthology.orchestra.run.vm06.stdout: python3-kubernetes-1:26.1.0-3.el9.noarch 2026-03-09T20:50:17.265 INFO:teuthology.orchestra.run.vm06.stdout: python3-libstoragemgmt-1.10.1-1.el9.x86_64 2026-03-09T20:50:17.265 INFO:teuthology.orchestra.run.vm06.stdout: python3-logutils-0.3.5-21.el9.noarch 2026-03-09T20:50:17.265 INFO:teuthology.orchestra.run.vm06.stdout: python3-mako-1.1.4-6.el9.noarch 2026-03-09T20:50:17.265 INFO:teuthology.orchestra.run.vm06.stdout: python3-markupsafe-1.1.1-12.el9.x86_64 2026-03-09T20:50:17.265 INFO:teuthology.orchestra.run.vm06.stdout: python3-more-itertools-8.12.0-2.el9.noarch 2026-03-09T20:50:17.265 INFO:teuthology.orchestra.run.vm06.stdout: python3-natsort-7.1.1-5.el9.noarch 2026-03-09T20:50:17.265 INFO:teuthology.orchestra.run.vm06.stdout: python3-numpy-1:1.23.5-2.el9.x86_64 2026-03-09T20:50:17.265 INFO:teuthology.orchestra.run.vm06.stdout: python3-numpy-f2py-1:1.23.5-2.el9.x86_64 2026-03-09T20:50:17.265 INFO:teuthology.orchestra.run.vm06.stdout: python3-packaging-20.9-5.el9.noarch 2026-03-09T20:50:17.265 INFO:teuthology.orchestra.run.vm06.stdout: python3-pecan-1.4.2-3.el9.noarch 2026-03-09T20:50:17.265 INFO:teuthology.orchestra.run.vm06.stdout: python3-ply-3.11-14.el9.noarch 2026-03-09T20:50:17.265 INFO:teuthology.orchestra.run.vm06.stdout: python3-portend-3.1.0-2.el9.noarch 2026-03-09T20:50:17.265 INFO:teuthology.orchestra.run.vm06.stdout: python3-protobuf-3.14.0-17.el9.noarch 2026-03-09T20:50:17.265 INFO:teuthology.orchestra.run.vm06.stdout: python3-pyOpenSSL-21.0.0-1.el9.noarch 2026-03-09T20:50:17.265 INFO:teuthology.orchestra.run.vm06.stdout: python3-pyasn1-0.4.8-7.el9.noarch 2026-03-09T20:50:17.265 INFO:teuthology.orchestra.run.vm06.stdout: python3-pyasn1-modules-0.4.8-7.el9.noarch 2026-03-09T20:50:17.265 INFO:teuthology.orchestra.run.vm06.stdout: python3-pycparser-2.20-6.el9.noarch 2026-03-09T20:50:17.265 INFO:teuthology.orchestra.run.vm06.stdout: python3-rados-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T20:50:17.265 INFO:teuthology.orchestra.run.vm06.stdout: python3-rbd-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T20:50:17.265 INFO:teuthology.orchestra.run.vm06.stdout: python3-repoze-lru-0.7-16.el9.noarch 2026-03-09T20:50:17.265 INFO:teuthology.orchestra.run.vm06.stdout: python3-requests-2.25.1-10.el9.noarch 2026-03-09T20:50:17.265 INFO:teuthology.orchestra.run.vm06.stdout: python3-requests-oauthlib-1.3.0-12.el9.noarch 2026-03-09T20:50:17.265 INFO:teuthology.orchestra.run.vm06.stdout: python3-rgw-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T20:50:17.265 INFO:teuthology.orchestra.run.vm06.stdout: python3-routes-2.5.1-5.el9.noarch 2026-03-09T20:50:17.265 INFO:teuthology.orchestra.run.vm06.stdout: python3-rsa-4.9-2.el9.noarch 2026-03-09T20:50:17.265 INFO:teuthology.orchestra.run.vm06.stdout: python3-scipy-1.9.3-2.el9.x86_64 2026-03-09T20:50:17.265 INFO:teuthology.orchestra.run.vm06.stdout: python3-tempora-5.0.0-2.el9.noarch 2026-03-09T20:50:17.265 INFO:teuthology.orchestra.run.vm06.stdout: python3-toml-0.10.2-6.el9.noarch 2026-03-09T20:50:17.265 INFO:teuthology.orchestra.run.vm06.stdout: python3-typing-extensions-4.15.0-1.el9.noarch 2026-03-09T20:50:17.265 INFO:teuthology.orchestra.run.vm06.stdout: python3-urllib3-1.26.5-7.el9.noarch 2026-03-09T20:50:17.265 INFO:teuthology.orchestra.run.vm06.stdout: python3-webob-1.8.8-2.el9.noarch 2026-03-09T20:50:17.265 INFO:teuthology.orchestra.run.vm06.stdout: python3-websocket-client-1.2.3-2.el9.noarch 2026-03-09T20:50:17.265 INFO:teuthology.orchestra.run.vm06.stdout: python3-werkzeug-2.0.3-3.el9.1.noarch 2026-03-09T20:50:17.265 INFO:teuthology.orchestra.run.vm06.stdout: python3-xmltodict-0.12.0-15.el9.noarch 2026-03-09T20:50:17.265 INFO:teuthology.orchestra.run.vm06.stdout: python3-zc-lockfile-2.0-10.el9.noarch 2026-03-09T20:50:17.265 INFO:teuthology.orchestra.run.vm06.stdout: qatlib-25.08.0-2.el9.x86_64 2026-03-09T20:50:17.265 INFO:teuthology.orchestra.run.vm06.stdout: qatlib-service-25.08.0-2.el9.x86_64 2026-03-09T20:50:17.265 INFO:teuthology.orchestra.run.vm06.stdout: qatzip-libs-1.3.1-1.el9.x86_64 2026-03-09T20:50:17.265 INFO:teuthology.orchestra.run.vm06.stdout: rbd-fuse-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T20:50:17.265 INFO:teuthology.orchestra.run.vm06.stdout: rbd-mirror-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T20:50:17.265 INFO:teuthology.orchestra.run.vm06.stdout: rbd-nbd-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T20:50:17.265 INFO:teuthology.orchestra.run.vm06.stdout: re2-1:20211101-20.el9.x86_64 2026-03-09T20:50:17.265 INFO:teuthology.orchestra.run.vm06.stdout: socat-1.7.4.1-8.el9.x86_64 2026-03-09T20:50:17.265 INFO:teuthology.orchestra.run.vm06.stdout: thrift-0.15.0-4.el9.x86_64 2026-03-09T20:50:17.265 INFO:teuthology.orchestra.run.vm06.stdout: unzip-6.0-59.el9.x86_64 2026-03-09T20:50:17.265 INFO:teuthology.orchestra.run.vm06.stdout: xmlstarlet-1.6.1-20.el9.x86_64 2026-03-09T20:50:17.266 INFO:teuthology.orchestra.run.vm06.stdout: zip-3.0-35.el9.x86_64 2026-03-09T20:50:17.266 INFO:teuthology.orchestra.run.vm06.stdout: 2026-03-09T20:50:17.266 INFO:teuthology.orchestra.run.vm06.stdout:Complete! 2026-03-09T20:50:17.368 DEBUG:teuthology.parallel:result is None 2026-03-09T20:50:19.499 INFO:teuthology.orchestra.run.vm01.stdout: Running scriptlet: ceph-selinux-2:19.2.3-678.ge911bdeb.el9.x86_64 108/138 2026-03-09T20:50:19.499 INFO:teuthology.orchestra.run.vm01.stdout:skipping the directory /sys 2026-03-09T20:50:19.499 INFO:teuthology.orchestra.run.vm01.stdout:skipping the directory /proc 2026-03-09T20:50:19.499 INFO:teuthology.orchestra.run.vm01.stdout:skipping the directory /mnt 2026-03-09T20:50:19.499 INFO:teuthology.orchestra.run.vm01.stdout:skipping the directory /var/tmp 2026-03-09T20:50:19.499 INFO:teuthology.orchestra.run.vm01.stdout:skipping the directory /home 2026-03-09T20:50:19.499 INFO:teuthology.orchestra.run.vm01.stdout:skipping the directory /root 2026-03-09T20:50:19.499 INFO:teuthology.orchestra.run.vm01.stdout:skipping the directory /tmp 2026-03-09T20:50:19.499 INFO:teuthology.orchestra.run.vm01.stdout: 2026-03-09T20:50:19.643 INFO:teuthology.orchestra.run.vm01.stdout: Installing : ceph-mds-2:19.2.3-678.ge911bdeb.el9.x86_64 109/138 2026-03-09T20:50:19.674 INFO:teuthology.orchestra.run.vm01.stdout: Running scriptlet: ceph-mds-2:19.2.3-678.ge911bdeb.el9.x86_64 109/138 2026-03-09T20:50:19.674 INFO:teuthology.orchestra.run.vm01.stdout:Glob pattern passed to enable, but globs are not supported for this. 2026-03-09T20:50:19.674 INFO:teuthology.orchestra.run.vm01.stdout:Invalid unit name "ceph-mds@*.service" escaped as "ceph-mds@\x2a.service". 2026-03-09T20:50:19.674 INFO:teuthology.orchestra.run.vm01.stdout:Created symlink /etc/systemd/system/multi-user.target.wants/ceph-mds.target → /usr/lib/systemd/system/ceph-mds.target. 2026-03-09T20:50:19.674 INFO:teuthology.orchestra.run.vm01.stdout:Created symlink /etc/systemd/system/ceph.target.wants/ceph-mds.target → /usr/lib/systemd/system/ceph-mds.target. 2026-03-09T20:50:19.674 INFO:teuthology.orchestra.run.vm01.stdout: 2026-03-09T20:50:19.947 INFO:teuthology.orchestra.run.vm01.stdout: Installing : ceph-mon-2:19.2.3-678.ge911bdeb.el9.x86_64 110/138 2026-03-09T20:50:19.975 INFO:teuthology.orchestra.run.vm01.stdout: Running scriptlet: ceph-mon-2:19.2.3-678.ge911bdeb.el9.x86_64 110/138 2026-03-09T20:50:19.975 INFO:teuthology.orchestra.run.vm01.stdout:Glob pattern passed to enable, but globs are not supported for this. 2026-03-09T20:50:19.975 INFO:teuthology.orchestra.run.vm01.stdout:Invalid unit name "ceph-mon@*.service" escaped as "ceph-mon@\x2a.service". 2026-03-09T20:50:19.975 INFO:teuthology.orchestra.run.vm01.stdout:Created symlink /etc/systemd/system/multi-user.target.wants/ceph-mon.target → /usr/lib/systemd/system/ceph-mon.target. 2026-03-09T20:50:19.975 INFO:teuthology.orchestra.run.vm01.stdout:Created symlink /etc/systemd/system/ceph.target.wants/ceph-mon.target → /usr/lib/systemd/system/ceph-mon.target. 2026-03-09T20:50:19.975 INFO:teuthology.orchestra.run.vm01.stdout: 2026-03-09T20:50:19.984 INFO:teuthology.orchestra.run.vm01.stdout: Installing : mailcap-2.1.49-5.el9.noarch 111/138 2026-03-09T20:50:19.988 INFO:teuthology.orchestra.run.vm01.stdout: Installing : libconfig-1.7.2-9.el9.x86_64 112/138 2026-03-09T20:50:20.005 INFO:teuthology.orchestra.run.vm01.stdout: Running scriptlet: libstoragemgmt-1.10.1-1.el9.x86_64 113/138 2026-03-09T20:50:20.005 INFO:teuthology.orchestra.run.vm01.stdout:Creating group 'qat' with GID 994. 2026-03-09T20:50:20.005 INFO:teuthology.orchestra.run.vm01.stdout:Creating group 'libstoragemgmt' with GID 993. 2026-03-09T20:50:20.005 INFO:teuthology.orchestra.run.vm01.stdout:Creating user 'libstoragemgmt' (daemon account for libstoragemgmt) with UID 993 and GID 993. 2026-03-09T20:50:20.005 INFO:teuthology.orchestra.run.vm01.stdout: 2026-03-09T20:50:20.016 INFO:teuthology.orchestra.run.vm01.stdout: Installing : libstoragemgmt-1.10.1-1.el9.x86_64 113/138 2026-03-09T20:50:20.041 INFO:teuthology.orchestra.run.vm01.stdout: Running scriptlet: libstoragemgmt-1.10.1-1.el9.x86_64 113/138 2026-03-09T20:50:20.041 INFO:teuthology.orchestra.run.vm01.stdout:Created symlink /etc/systemd/system/multi-user.target.wants/libstoragemgmt.service → /usr/lib/systemd/system/libstoragemgmt.service. 2026-03-09T20:50:20.042 INFO:teuthology.orchestra.run.vm01.stdout: 2026-03-09T20:50:20.085 INFO:teuthology.orchestra.run.vm01.stdout: Installing : python3-libstoragemgmt-1.10.1-1.el9.x86_64 114/138 2026-03-09T20:50:20.161 INFO:teuthology.orchestra.run.vm01.stdout: Installing : cryptsetup-2.8.1-3.el9.x86_64 115/138 2026-03-09T20:50:20.167 INFO:teuthology.orchestra.run.vm01.stdout: Installing : ceph-volume-2:19.2.3-678.ge911bdeb.el9.noarch 116/138 2026-03-09T20:50:20.183 INFO:teuthology.orchestra.run.vm01.stdout: Running scriptlet: ceph-volume-2:19.2.3-678.ge911bdeb.el9.noarch 116/138 2026-03-09T20:50:20.183 INFO:teuthology.orchestra.run.vm01.stdout:Glob pattern passed to enable, but globs are not supported for this. 2026-03-09T20:50:20.183 INFO:teuthology.orchestra.run.vm01.stdout:Invalid unit name "ceph-volume@*.service" escaped as "ceph-volume@\x2a.service". 2026-03-09T20:50:20.183 INFO:teuthology.orchestra.run.vm01.stdout: 2026-03-09T20:50:21.127 INFO:teuthology.orchestra.run.vm01.stdout: Installing : ceph-osd-2:19.2.3-678.ge911bdeb.el9.x86_64 117/138 2026-03-09T20:50:21.167 INFO:teuthology.orchestra.run.vm01.stdout: Running scriptlet: ceph-osd-2:19.2.3-678.ge911bdeb.el9.x86_64 117/138 2026-03-09T20:50:21.167 INFO:teuthology.orchestra.run.vm01.stdout:Glob pattern passed to enable, but globs are not supported for this. 2026-03-09T20:50:21.167 INFO:teuthology.orchestra.run.vm01.stdout:Invalid unit name "ceph-osd@*.service" escaped as "ceph-osd@\x2a.service". 2026-03-09T20:50:21.167 INFO:teuthology.orchestra.run.vm01.stdout:Created symlink /etc/systemd/system/multi-user.target.wants/ceph-osd.target → /usr/lib/systemd/system/ceph-osd.target. 2026-03-09T20:50:21.167 INFO:teuthology.orchestra.run.vm01.stdout:Created symlink /etc/systemd/system/ceph.target.wants/ceph-osd.target → /usr/lib/systemd/system/ceph-osd.target. 2026-03-09T20:50:21.167 INFO:teuthology.orchestra.run.vm01.stdout: 2026-03-09T20:50:21.247 INFO:teuthology.orchestra.run.vm01.stdout: Running scriptlet: cephadm-2:19.2.3-678.ge911bdeb.el9.noarch 118/138 2026-03-09T20:50:21.252 INFO:teuthology.orchestra.run.vm01.stdout: Installing : cephadm-2:19.2.3-678.ge911bdeb.el9.noarch 118/138 2026-03-09T20:50:21.259 INFO:teuthology.orchestra.run.vm01.stdout: Installing : ceph-prometheus-alerts-2:19.2.3-678.ge911bdeb.el 119/138 2026-03-09T20:50:21.285 INFO:teuthology.orchestra.run.vm01.stdout: Installing : ceph-grafana-dashboards-2:19.2.3-678.ge911bdeb.e 120/138 2026-03-09T20:50:21.290 INFO:teuthology.orchestra.run.vm01.stdout: Installing : ceph-mgr-cephadm-2:19.2.3-678.ge911bdeb.el9.noar 121/138 2026-03-09T20:50:21.864 INFO:teuthology.orchestra.run.vm01.stdout: Running scriptlet: ceph-mgr-cephadm-2:19.2.3-678.ge911bdeb.el9.noar 121/138 2026-03-09T20:50:21.871 INFO:teuthology.orchestra.run.vm01.stdout: Installing : ceph-mgr-dashboard-2:19.2.3-678.ge911bdeb.el9.no 122/138 2026-03-09T20:50:22.450 INFO:teuthology.orchestra.run.vm01.stdout: Running scriptlet: ceph-mgr-dashboard-2:19.2.3-678.ge911bdeb.el9.no 122/138 2026-03-09T20:50:22.453 INFO:teuthology.orchestra.run.vm01.stdout: Installing : ceph-mgr-diskprediction-local-2:19.2.3-678.ge911 123/138 2026-03-09T20:50:22.524 INFO:teuthology.orchestra.run.vm01.stdout: Running scriptlet: ceph-mgr-diskprediction-local-2:19.2.3-678.ge911 123/138 2026-03-09T20:50:22.663 INFO:teuthology.orchestra.run.vm01.stdout: Installing : ceph-mgr-modules-core-2:19.2.3-678.ge911bdeb.el9 124/138 2026-03-09T20:50:22.668 INFO:teuthology.orchestra.run.vm01.stdout: Installing : ceph-mgr-2:19.2.3-678.ge911bdeb.el9.x86_64 125/138 2026-03-09T20:50:22.693 INFO:teuthology.orchestra.run.vm01.stdout: Running scriptlet: ceph-mgr-2:19.2.3-678.ge911bdeb.el9.x86_64 125/138 2026-03-09T20:50:22.693 INFO:teuthology.orchestra.run.vm01.stdout:Glob pattern passed to enable, but globs are not supported for this. 2026-03-09T20:50:22.693 INFO:teuthology.orchestra.run.vm01.stdout:Invalid unit name "ceph-mgr@*.service" escaped as "ceph-mgr@\x2a.service". 2026-03-09T20:50:22.693 INFO:teuthology.orchestra.run.vm01.stdout:Created symlink /etc/systemd/system/multi-user.target.wants/ceph-mgr.target → /usr/lib/systemd/system/ceph-mgr.target. 2026-03-09T20:50:22.693 INFO:teuthology.orchestra.run.vm01.stdout:Created symlink /etc/systemd/system/ceph.target.wants/ceph-mgr.target → /usr/lib/systemd/system/ceph-mgr.target. 2026-03-09T20:50:22.693 INFO:teuthology.orchestra.run.vm01.stdout: 2026-03-09T20:50:22.710 INFO:teuthology.orchestra.run.vm01.stdout: Installing : ceph-mgr-rook-2:19.2.3-678.ge911bdeb.el9.noarch 126/138 2026-03-09T20:50:22.725 INFO:teuthology.orchestra.run.vm01.stdout: Running scriptlet: ceph-mgr-rook-2:19.2.3-678.ge911bdeb.el9.noarch 126/138 2026-03-09T20:50:23.356 INFO:teuthology.orchestra.run.vm01.stdout: Installing : ceph-2:19.2.3-678.ge911bdeb.el9.x86_64 127/138 2026-03-09T20:50:23.452 INFO:teuthology.orchestra.run.vm01.stdout: Installing : ceph-radosgw-2:19.2.3-678.ge911bdeb.el9.x86_64 128/138 2026-03-09T20:50:23.476 INFO:teuthology.orchestra.run.vm01.stdout: Running scriptlet: ceph-radosgw-2:19.2.3-678.ge911bdeb.el9.x86_64 128/138 2026-03-09T20:50:23.476 INFO:teuthology.orchestra.run.vm01.stdout:Glob pattern passed to enable, but globs are not supported for this. 2026-03-09T20:50:23.476 INFO:teuthology.orchestra.run.vm01.stdout:Invalid unit name "ceph-radosgw@*.service" escaped as "ceph-radosgw@\x2a.service". 2026-03-09T20:50:23.476 INFO:teuthology.orchestra.run.vm01.stdout:Created symlink /etc/systemd/system/multi-user.target.wants/ceph-radosgw.target → /usr/lib/systemd/system/ceph-radosgw.target. 2026-03-09T20:50:23.476 INFO:teuthology.orchestra.run.vm01.stdout:Created symlink /etc/systemd/system/ceph.target.wants/ceph-radosgw.target → /usr/lib/systemd/system/ceph-radosgw.target. 2026-03-09T20:50:23.477 INFO:teuthology.orchestra.run.vm01.stdout: 2026-03-09T20:50:23.489 INFO:teuthology.orchestra.run.vm01.stdout: Installing : ceph-immutable-object-cache-2:19.2.3-678.ge911bd 129/138 2026-03-09T20:50:23.510 INFO:teuthology.orchestra.run.vm01.stdout: Running scriptlet: ceph-immutable-object-cache-2:19.2.3-678.ge911bd 129/138 2026-03-09T20:50:23.510 INFO:teuthology.orchestra.run.vm01.stdout:Glob pattern passed to enable, but globs are not supported for this. 2026-03-09T20:50:23.510 INFO:teuthology.orchestra.run.vm01.stdout:Invalid unit name "ceph-immutable-object-cache@*.service" escaped as "ceph-immutable-object-cache@\x2a.service". 2026-03-09T20:50:23.510 INFO:teuthology.orchestra.run.vm01.stdout: 2026-03-09T20:50:23.713 INFO:teuthology.orchestra.run.vm01.stdout: Installing : rbd-mirror-2:19.2.3-678.ge911bdeb.el9.x86_64 130/138 2026-03-09T20:50:23.739 INFO:teuthology.orchestra.run.vm01.stdout: Running scriptlet: rbd-mirror-2:19.2.3-678.ge911bdeb.el9.x86_64 130/138 2026-03-09T20:50:23.739 INFO:teuthology.orchestra.run.vm01.stdout:Glob pattern passed to enable, but globs are not supported for this. 2026-03-09T20:50:23.739 INFO:teuthology.orchestra.run.vm01.stdout:Invalid unit name "ceph-rbd-mirror@*.service" escaped as "ceph-rbd-mirror@\x2a.service". 2026-03-09T20:50:23.740 INFO:teuthology.orchestra.run.vm01.stdout:Created symlink /etc/systemd/system/multi-user.target.wants/ceph-rbd-mirror.target → /usr/lib/systemd/system/ceph-rbd-mirror.target. 2026-03-09T20:50:23.740 INFO:teuthology.orchestra.run.vm01.stdout:Created symlink /etc/systemd/system/ceph.target.wants/ceph-rbd-mirror.target → /usr/lib/systemd/system/ceph-rbd-mirror.target. 2026-03-09T20:50:23.740 INFO:teuthology.orchestra.run.vm01.stdout: 2026-03-09T20:50:26.740 INFO:teuthology.orchestra.run.vm01.stdout: Installing : ceph-test-2:19.2.3-678.ge911bdeb.el9.x86_64 131/138 2026-03-09T20:50:26.753 INFO:teuthology.orchestra.run.vm01.stdout: Installing : rbd-fuse-2:19.2.3-678.ge911bdeb.el9.x86_64 132/138 2026-03-09T20:50:26.760 INFO:teuthology.orchestra.run.vm01.stdout: Installing : rbd-nbd-2:19.2.3-678.ge911bdeb.el9.x86_64 133/138 2026-03-09T20:50:26.821 INFO:teuthology.orchestra.run.vm01.stdout: Installing : libcephfs-devel-2:19.2.3-678.ge911bdeb.el9.x86_6 134/138 2026-03-09T20:50:26.832 INFO:teuthology.orchestra.run.vm01.stdout: Installing : ceph-fuse-2:19.2.3-678.ge911bdeb.el9.x86_64 135/138 2026-03-09T20:50:26.836 INFO:teuthology.orchestra.run.vm01.stdout: Installing : python3-jmespath-1.0.1-1.el9.noarch 136/138 2026-03-09T20:50:26.836 INFO:teuthology.orchestra.run.vm01.stdout: Cleanup : librbd1-2:16.2.4-5.el9.x86_64 137/138 2026-03-09T20:50:26.856 INFO:teuthology.orchestra.run.vm01.stdout: Running scriptlet: librbd1-2:16.2.4-5.el9.x86_64 137/138 2026-03-09T20:50:26.857 INFO:teuthology.orchestra.run.vm01.stdout: Cleanup : librados2-2:16.2.4-5.el9.x86_64 138/138 2026-03-09T20:50:28.189 INFO:teuthology.orchestra.run.vm01.stdout: Running scriptlet: librados2-2:16.2.4-5.el9.x86_64 138/138 2026-03-09T20:50:28.190 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : ceph-2:19.2.3-678.ge911bdeb.el9.x86_64 1/138 2026-03-09T20:50:28.190 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : ceph-base-2:19.2.3-678.ge911bdeb.el9.x86_64 2/138 2026-03-09T20:50:28.190 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : ceph-common-2:19.2.3-678.ge911bdeb.el9.x86_64 3/138 2026-03-09T20:50:28.190 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : ceph-fuse-2:19.2.3-678.ge911bdeb.el9.x86_64 4/138 2026-03-09T20:50:28.190 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : ceph-immutable-object-cache-2:19.2.3-678.ge911bd 5/138 2026-03-09T20:50:28.190 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : ceph-mds-2:19.2.3-678.ge911bdeb.el9.x86_64 6/138 2026-03-09T20:50:28.190 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : ceph-mgr-2:19.2.3-678.ge911bdeb.el9.x86_64 7/138 2026-03-09T20:50:28.190 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : ceph-mon-2:19.2.3-678.ge911bdeb.el9.x86_64 8/138 2026-03-09T20:50:28.190 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : ceph-osd-2:19.2.3-678.ge911bdeb.el9.x86_64 9/138 2026-03-09T20:50:28.190 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : ceph-radosgw-2:19.2.3-678.ge911bdeb.el9.x86_64 10/138 2026-03-09T20:50:28.190 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : ceph-selinux-2:19.2.3-678.ge911bdeb.el9.x86_64 11/138 2026-03-09T20:50:28.190 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : ceph-test-2:19.2.3-678.ge911bdeb.el9.x86_64 12/138 2026-03-09T20:50:28.190 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : libcephfs-devel-2:19.2.3-678.ge911bdeb.el9.x86_6 13/138 2026-03-09T20:50:28.190 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : libcephfs2-2:19.2.3-678.ge911bdeb.el9.x86_64 14/138 2026-03-09T20:50:28.190 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : libcephsqlite-2:19.2.3-678.ge911bdeb.el9.x86_64 15/138 2026-03-09T20:50:28.190 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : librados-devel-2:19.2.3-678.ge911bdeb.el9.x86_64 16/138 2026-03-09T20:50:28.190 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : libradosstriper1-2:19.2.3-678.ge911bdeb.el9.x86_ 17/138 2026-03-09T20:50:28.190 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : librgw2-2:19.2.3-678.ge911bdeb.el9.x86_64 18/138 2026-03-09T20:50:28.190 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-ceph-argparse-2:19.2.3-678.ge911bdeb.el9 19/138 2026-03-09T20:50:28.190 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-ceph-common-2:19.2.3-678.ge911bdeb.el9.x 20/138 2026-03-09T20:50:28.190 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-cephfs-2:19.2.3-678.ge911bdeb.el9.x86_64 21/138 2026-03-09T20:50:28.190 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-rados-2:19.2.3-678.ge911bdeb.el9.x86_64 22/138 2026-03-09T20:50:28.190 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-rbd-2:19.2.3-678.ge911bdeb.el9.x86_64 23/138 2026-03-09T20:50:28.190 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-rgw-2:19.2.3-678.ge911bdeb.el9.x86_64 24/138 2026-03-09T20:50:28.190 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : rbd-fuse-2:19.2.3-678.ge911bdeb.el9.x86_64 25/138 2026-03-09T20:50:28.191 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : rbd-mirror-2:19.2.3-678.ge911bdeb.el9.x86_64 26/138 2026-03-09T20:50:28.191 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : rbd-nbd-2:19.2.3-678.ge911bdeb.el9.x86_64 27/138 2026-03-09T20:50:28.191 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : ceph-grafana-dashboards-2:19.2.3-678.ge911bdeb.e 28/138 2026-03-09T20:50:28.191 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : ceph-mgr-cephadm-2:19.2.3-678.ge911bdeb.el9.noar 29/138 2026-03-09T20:50:28.191 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : ceph-mgr-dashboard-2:19.2.3-678.ge911bdeb.el9.no 30/138 2026-03-09T20:50:28.191 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : ceph-mgr-diskprediction-local-2:19.2.3-678.ge911 31/138 2026-03-09T20:50:28.191 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : ceph-mgr-modules-core-2:19.2.3-678.ge911bdeb.el9 32/138 2026-03-09T20:50:28.191 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : ceph-mgr-rook-2:19.2.3-678.ge911bdeb.el9.noarch 33/138 2026-03-09T20:50:28.191 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : ceph-prometheus-alerts-2:19.2.3-678.ge911bdeb.el 34/138 2026-03-09T20:50:28.191 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : ceph-volume-2:19.2.3-678.ge911bdeb.el9.noarch 35/138 2026-03-09T20:50:28.191 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : cephadm-2:19.2.3-678.ge911bdeb.el9.noarch 36/138 2026-03-09T20:50:28.191 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : cryptsetup-2.8.1-3.el9.x86_64 37/138 2026-03-09T20:50:28.191 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : ledmon-libs-1.1.0-3.el9.x86_64 38/138 2026-03-09T20:50:28.191 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : libconfig-1.7.2-9.el9.x86_64 39/138 2026-03-09T20:50:28.191 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : libgfortran-11.5.0-14.el9.x86_64 40/138 2026-03-09T20:50:28.191 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : libquadmath-11.5.0-14.el9.x86_64 41/138 2026-03-09T20:50:28.191 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : mailcap-2.1.49-5.el9.noarch 42/138 2026-03-09T20:50:28.191 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : pciutils-3.7.0-7.el9.x86_64 43/138 2026-03-09T20:50:28.191 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-cffi-1.14.5-5.el9.x86_64 44/138 2026-03-09T20:50:28.191 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-cryptography-36.0.1-5.el9.x86_64 45/138 2026-03-09T20:50:28.191 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-ply-3.11-14.el9.noarch 46/138 2026-03-09T20:50:28.191 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-pycparser-2.20-6.el9.noarch 47/138 2026-03-09T20:50:28.191 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-requests-2.25.1-10.el9.noarch 48/138 2026-03-09T20:50:28.191 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-urllib3-1.26.5-7.el9.noarch 49/138 2026-03-09T20:50:28.192 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : unzip-6.0-59.el9.x86_64 50/138 2026-03-09T20:50:28.192 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : zip-3.0-35.el9.x86_64 51/138 2026-03-09T20:50:28.192 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : boost-program-options-1.75.0-13.el9.x86_64 52/138 2026-03-09T20:50:28.192 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : flexiblas-3.0.4-9.el9.x86_64 53/138 2026-03-09T20:50:28.192 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : flexiblas-netlib-3.0.4-9.el9.x86_64 54/138 2026-03-09T20:50:28.192 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : flexiblas-openblas-openmp-3.0.4-9.el9.x86_64 55/138 2026-03-09T20:50:28.192 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : libnbd-1.20.3-4.el9.x86_64 56/138 2026-03-09T20:50:28.192 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : libpmemobj-1.12.1-1.el9.x86_64 57/138 2026-03-09T20:50:28.192 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : librabbitmq-0.11.0-7.el9.x86_64 58/138 2026-03-09T20:50:28.192 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : librdkafka-1.6.1-102.el9.x86_64 59/138 2026-03-09T20:50:28.192 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : libstoragemgmt-1.10.1-1.el9.x86_64 60/138 2026-03-09T20:50:28.192 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : libxslt-1.1.34-12.el9.x86_64 61/138 2026-03-09T20:50:28.192 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : lttng-ust-2.12.0-6.el9.x86_64 62/138 2026-03-09T20:50:28.192 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : lua-5.4.4-4.el9.x86_64 63/138 2026-03-09T20:50:28.192 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : openblas-0.3.29-1.el9.x86_64 64/138 2026-03-09T20:50:28.192 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : openblas-openmp-0.3.29-1.el9.x86_64 65/138 2026-03-09T20:50:28.192 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : protobuf-3.14.0-17.el9.x86_64 66/138 2026-03-09T20:50:28.192 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-babel-2.9.1-2.el9.noarch 67/138 2026-03-09T20:50:28.192 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-devel-3.9.25-3.el9.x86_64 68/138 2026-03-09T20:50:28.192 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-jinja2-2.11.3-8.el9.noarch 69/138 2026-03-09T20:50:28.192 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-jmespath-1.0.1-1.el9.noarch 70/138 2026-03-09T20:50:28.192 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-libstoragemgmt-1.10.1-1.el9.x86_64 71/138 2026-03-09T20:50:28.192 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-mako-1.1.4-6.el9.noarch 72/138 2026-03-09T20:50:28.192 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-markupsafe-1.1.1-12.el9.x86_64 73/138 2026-03-09T20:50:28.192 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-numpy-1:1.23.5-2.el9.x86_64 74/138 2026-03-09T20:50:28.193 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-numpy-f2py-1:1.23.5-2.el9.x86_64 75/138 2026-03-09T20:50:28.193 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-packaging-20.9-5.el9.noarch 76/138 2026-03-09T20:50:28.193 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-protobuf-3.14.0-17.el9.noarch 77/138 2026-03-09T20:50:28.193 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-pyasn1-0.4.8-7.el9.noarch 78/138 2026-03-09T20:50:28.193 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-pyasn1-modules-0.4.8-7.el9.noarch 79/138 2026-03-09T20:50:28.193 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-requests-oauthlib-1.3.0-12.el9.noarch 80/138 2026-03-09T20:50:28.193 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-scipy-1.9.3-2.el9.x86_64 81/138 2026-03-09T20:50:28.193 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-toml-0.10.2-6.el9.noarch 82/138 2026-03-09T20:50:28.193 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : qatlib-25.08.0-2.el9.x86_64 83/138 2026-03-09T20:50:28.193 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : qatlib-service-25.08.0-2.el9.x86_64 84/138 2026-03-09T20:50:28.193 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : qatzip-libs-1.3.1-1.el9.x86_64 85/138 2026-03-09T20:50:28.194 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : socat-1.7.4.1-8.el9.x86_64 86/138 2026-03-09T20:50:28.194 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : xmlstarlet-1.6.1-20.el9.x86_64 87/138 2026-03-09T20:50:28.194 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : lua-devel-5.4.4-4.el9.x86_64 88/138 2026-03-09T20:50:28.194 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : protobuf-compiler-3.14.0-17.el9.x86_64 89/138 2026-03-09T20:50:28.194 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : abseil-cpp-20211102.0-4.el9.x86_64 90/138 2026-03-09T20:50:28.194 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : gperftools-libs-2.9.1-3.el9.x86_64 91/138 2026-03-09T20:50:28.194 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : grpc-data-1.46.7-10.el9.noarch 92/138 2026-03-09T20:50:28.194 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : libarrow-9.0.0-15.el9.x86_64 93/138 2026-03-09T20:50:28.194 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : libarrow-doc-9.0.0-15.el9.noarch 94/138 2026-03-09T20:50:28.194 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : liboath-2.6.12-1.el9.x86_64 95/138 2026-03-09T20:50:28.194 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : libunwind-1.6.2-1.el9.x86_64 96/138 2026-03-09T20:50:28.194 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : luarocks-3.9.2-5.el9.noarch 97/138 2026-03-09T20:50:28.194 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : parquet-libs-9.0.0-15.el9.x86_64 98/138 2026-03-09T20:50:28.194 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-asyncssh-2.13.2-5.el9.noarch 99/138 2026-03-09T20:50:28.194 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-autocommand-2.2.2-8.el9.noarch 100/138 2026-03-09T20:50:28.194 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-backports-tarfile-1.2.0-1.el9.noarch 101/138 2026-03-09T20:50:28.194 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-bcrypt-3.2.2-1.el9.x86_64 102/138 2026-03-09T20:50:28.194 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-cachetools-4.2.4-1.el9.noarch 103/138 2026-03-09T20:50:28.194 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-certifi-2023.05.07-4.el9.noarch 104/138 2026-03-09T20:50:28.194 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-cheroot-10.0.1-4.el9.noarch 105/138 2026-03-09T20:50:28.194 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-cherrypy-18.6.1-2.el9.noarch 106/138 2026-03-09T20:50:28.194 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-google-auth-1:2.45.0-1.el9.noarch 107/138 2026-03-09T20:50:28.194 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-grpcio-1.46.7-10.el9.x86_64 108/138 2026-03-09T20:50:28.194 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-grpcio-tools-1.46.7-10.el9.x86_64 109/138 2026-03-09T20:50:28.194 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-jaraco-8.2.1-3.el9.noarch 110/138 2026-03-09T20:50:28.194 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-jaraco-classes-3.2.1-5.el9.noarch 111/138 2026-03-09T20:50:28.194 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-jaraco-collections-3.0.0-8.el9.noarch 112/138 2026-03-09T20:50:28.194 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-jaraco-context-6.0.1-3.el9.noarch 113/138 2026-03-09T20:50:28.194 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-jaraco-functools-3.5.0-2.el9.noarch 114/138 2026-03-09T20:50:28.194 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-jaraco-text-4.0.0-2.el9.noarch 115/138 2026-03-09T20:50:28.194 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-kubernetes-1:26.1.0-3.el9.noarch 116/138 2026-03-09T20:50:28.194 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-logutils-0.3.5-21.el9.noarch 117/138 2026-03-09T20:50:28.194 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-more-itertools-8.12.0-2.el9.noarch 118/138 2026-03-09T20:50:28.195 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-natsort-7.1.1-5.el9.noarch 119/138 2026-03-09T20:50:28.195 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-pecan-1.4.2-3.el9.noarch 120/138 2026-03-09T20:50:28.195 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-portend-3.1.0-2.el9.noarch 121/138 2026-03-09T20:50:28.195 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-pyOpenSSL-21.0.0-1.el9.noarch 122/138 2026-03-09T20:50:28.195 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-repoze-lru-0.7-16.el9.noarch 123/138 2026-03-09T20:50:28.195 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-routes-2.5.1-5.el9.noarch 124/138 2026-03-09T20:50:28.195 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-rsa-4.9-2.el9.noarch 125/138 2026-03-09T20:50:28.195 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-tempora-5.0.0-2.el9.noarch 126/138 2026-03-09T20:50:28.195 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-typing-extensions-4.15.0-1.el9.noarch 127/138 2026-03-09T20:50:28.195 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-webob-1.8.8-2.el9.noarch 128/138 2026-03-09T20:50:28.195 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-websocket-client-1.2.3-2.el9.noarch 129/138 2026-03-09T20:50:28.195 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-werkzeug-2.0.3-3.el9.1.noarch 130/138 2026-03-09T20:50:28.195 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-xmltodict-0.12.0-15.el9.noarch 131/138 2026-03-09T20:50:28.195 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-zc-lockfile-2.0-10.el9.noarch 132/138 2026-03-09T20:50:28.195 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : re2-1:20211101-20.el9.x86_64 133/138 2026-03-09T20:50:28.195 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : thrift-0.15.0-4.el9.x86_64 134/138 2026-03-09T20:50:28.195 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : librados2-2:19.2.3-678.ge911bdeb.el9.x86_64 135/138 2026-03-09T20:50:28.195 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : librados2-2:16.2.4-5.el9.x86_64 136/138 2026-03-09T20:50:28.195 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : librbd1-2:19.2.3-678.ge911bdeb.el9.x86_64 137/138 2026-03-09T20:50:28.314 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : librbd1-2:16.2.4-5.el9.x86_64 138/138 2026-03-09T20:50:28.314 INFO:teuthology.orchestra.run.vm01.stdout: 2026-03-09T20:50:28.314 INFO:teuthology.orchestra.run.vm01.stdout:Upgraded: 2026-03-09T20:50:28.314 INFO:teuthology.orchestra.run.vm01.stdout: librados2-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T20:50:28.314 INFO:teuthology.orchestra.run.vm01.stdout: librbd1-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T20:50:28.315 INFO:teuthology.orchestra.run.vm01.stdout:Installed: 2026-03-09T20:50:28.315 INFO:teuthology.orchestra.run.vm01.stdout: abseil-cpp-20211102.0-4.el9.x86_64 2026-03-09T20:50:28.315 INFO:teuthology.orchestra.run.vm01.stdout: boost-program-options-1.75.0-13.el9.x86_64 2026-03-09T20:50:28.315 INFO:teuthology.orchestra.run.vm01.stdout: ceph-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T20:50:28.315 INFO:teuthology.orchestra.run.vm01.stdout: ceph-base-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T20:50:28.315 INFO:teuthology.orchestra.run.vm01.stdout: ceph-common-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T20:50:28.315 INFO:teuthology.orchestra.run.vm01.stdout: ceph-fuse-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T20:50:28.315 INFO:teuthology.orchestra.run.vm01.stdout: ceph-grafana-dashboards-2:19.2.3-678.ge911bdeb.el9.noarch 2026-03-09T20:50:28.315 INFO:teuthology.orchestra.run.vm01.stdout: ceph-immutable-object-cache-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T20:50:28.315 INFO:teuthology.orchestra.run.vm01.stdout: ceph-mds-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T20:50:28.315 INFO:teuthology.orchestra.run.vm01.stdout: ceph-mgr-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T20:50:28.315 INFO:teuthology.orchestra.run.vm01.stdout: ceph-mgr-cephadm-2:19.2.3-678.ge911bdeb.el9.noarch 2026-03-09T20:50:28.315 INFO:teuthology.orchestra.run.vm01.stdout: ceph-mgr-dashboard-2:19.2.3-678.ge911bdeb.el9.noarch 2026-03-09T20:50:28.315 INFO:teuthology.orchestra.run.vm01.stdout: ceph-mgr-diskprediction-local-2:19.2.3-678.ge911bdeb.el9.noarch 2026-03-09T20:50:28.315 INFO:teuthology.orchestra.run.vm01.stdout: ceph-mgr-modules-core-2:19.2.3-678.ge911bdeb.el9.noarch 2026-03-09T20:50:28.315 INFO:teuthology.orchestra.run.vm01.stdout: ceph-mgr-rook-2:19.2.3-678.ge911bdeb.el9.noarch 2026-03-09T20:50:28.315 INFO:teuthology.orchestra.run.vm01.stdout: ceph-mon-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T20:50:28.315 INFO:teuthology.orchestra.run.vm01.stdout: ceph-osd-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T20:50:28.315 INFO:teuthology.orchestra.run.vm01.stdout: ceph-prometheus-alerts-2:19.2.3-678.ge911bdeb.el9.noarch 2026-03-09T20:50:28.315 INFO:teuthology.orchestra.run.vm01.stdout: ceph-radosgw-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T20:50:28.315 INFO:teuthology.orchestra.run.vm01.stdout: ceph-selinux-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T20:50:28.315 INFO:teuthology.orchestra.run.vm01.stdout: ceph-test-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T20:50:28.315 INFO:teuthology.orchestra.run.vm01.stdout: ceph-volume-2:19.2.3-678.ge911bdeb.el9.noarch 2026-03-09T20:50:28.315 INFO:teuthology.orchestra.run.vm01.stdout: cephadm-2:19.2.3-678.ge911bdeb.el9.noarch 2026-03-09T20:50:28.315 INFO:teuthology.orchestra.run.vm01.stdout: cryptsetup-2.8.1-3.el9.x86_64 2026-03-09T20:50:28.315 INFO:teuthology.orchestra.run.vm01.stdout: flexiblas-3.0.4-9.el9.x86_64 2026-03-09T20:50:28.315 INFO:teuthology.orchestra.run.vm01.stdout: flexiblas-netlib-3.0.4-9.el9.x86_64 2026-03-09T20:50:28.315 INFO:teuthology.orchestra.run.vm01.stdout: flexiblas-openblas-openmp-3.0.4-9.el9.x86_64 2026-03-09T20:50:28.315 INFO:teuthology.orchestra.run.vm01.stdout: gperftools-libs-2.9.1-3.el9.x86_64 2026-03-09T20:50:28.315 INFO:teuthology.orchestra.run.vm01.stdout: grpc-data-1.46.7-10.el9.noarch 2026-03-09T20:50:28.315 INFO:teuthology.orchestra.run.vm01.stdout: ledmon-libs-1.1.0-3.el9.x86_64 2026-03-09T20:50:28.315 INFO:teuthology.orchestra.run.vm01.stdout: libarrow-9.0.0-15.el9.x86_64 2026-03-09T20:50:28.315 INFO:teuthology.orchestra.run.vm01.stdout: libarrow-doc-9.0.0-15.el9.noarch 2026-03-09T20:50:28.315 INFO:teuthology.orchestra.run.vm01.stdout: libcephfs-devel-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T20:50:28.315 INFO:teuthology.orchestra.run.vm01.stdout: libcephfs2-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T20:50:28.315 INFO:teuthology.orchestra.run.vm01.stdout: libcephsqlite-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T20:50:28.315 INFO:teuthology.orchestra.run.vm01.stdout: libconfig-1.7.2-9.el9.x86_64 2026-03-09T20:50:28.315 INFO:teuthology.orchestra.run.vm01.stdout: libgfortran-11.5.0-14.el9.x86_64 2026-03-09T20:50:28.315 INFO:teuthology.orchestra.run.vm01.stdout: libnbd-1.20.3-4.el9.x86_64 2026-03-09T20:50:28.315 INFO:teuthology.orchestra.run.vm01.stdout: liboath-2.6.12-1.el9.x86_64 2026-03-09T20:50:28.315 INFO:teuthology.orchestra.run.vm01.stdout: libpmemobj-1.12.1-1.el9.x86_64 2026-03-09T20:50:28.315 INFO:teuthology.orchestra.run.vm01.stdout: libquadmath-11.5.0-14.el9.x86_64 2026-03-09T20:50:28.315 INFO:teuthology.orchestra.run.vm01.stdout: librabbitmq-0.11.0-7.el9.x86_64 2026-03-09T20:50:28.315 INFO:teuthology.orchestra.run.vm01.stdout: librados-devel-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T20:50:28.315 INFO:teuthology.orchestra.run.vm01.stdout: libradosstriper1-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T20:50:28.315 INFO:teuthology.orchestra.run.vm01.stdout: librdkafka-1.6.1-102.el9.x86_64 2026-03-09T20:50:28.315 INFO:teuthology.orchestra.run.vm01.stdout: librgw2-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T20:50:28.315 INFO:teuthology.orchestra.run.vm01.stdout: libstoragemgmt-1.10.1-1.el9.x86_64 2026-03-09T20:50:28.315 INFO:teuthology.orchestra.run.vm01.stdout: libunwind-1.6.2-1.el9.x86_64 2026-03-09T20:50:28.315 INFO:teuthology.orchestra.run.vm01.stdout: libxslt-1.1.34-12.el9.x86_64 2026-03-09T20:50:28.315 INFO:teuthology.orchestra.run.vm01.stdout: lttng-ust-2.12.0-6.el9.x86_64 2026-03-09T20:50:28.315 INFO:teuthology.orchestra.run.vm01.stdout: lua-5.4.4-4.el9.x86_64 2026-03-09T20:50:28.315 INFO:teuthology.orchestra.run.vm01.stdout: lua-devel-5.4.4-4.el9.x86_64 2026-03-09T20:50:28.315 INFO:teuthology.orchestra.run.vm01.stdout: luarocks-3.9.2-5.el9.noarch 2026-03-09T20:50:28.315 INFO:teuthology.orchestra.run.vm01.stdout: mailcap-2.1.49-5.el9.noarch 2026-03-09T20:50:28.315 INFO:teuthology.orchestra.run.vm01.stdout: openblas-0.3.29-1.el9.x86_64 2026-03-09T20:50:28.315 INFO:teuthology.orchestra.run.vm01.stdout: openblas-openmp-0.3.29-1.el9.x86_64 2026-03-09T20:50:28.316 INFO:teuthology.orchestra.run.vm01.stdout: parquet-libs-9.0.0-15.el9.x86_64 2026-03-09T20:50:28.316 INFO:teuthology.orchestra.run.vm01.stdout: pciutils-3.7.0-7.el9.x86_64 2026-03-09T20:50:28.316 INFO:teuthology.orchestra.run.vm01.stdout: protobuf-3.14.0-17.el9.x86_64 2026-03-09T20:50:28.316 INFO:teuthology.orchestra.run.vm01.stdout: protobuf-compiler-3.14.0-17.el9.x86_64 2026-03-09T20:50:28.316 INFO:teuthology.orchestra.run.vm01.stdout: python3-asyncssh-2.13.2-5.el9.noarch 2026-03-09T20:50:28.316 INFO:teuthology.orchestra.run.vm01.stdout: python3-autocommand-2.2.2-8.el9.noarch 2026-03-09T20:50:28.316 INFO:teuthology.orchestra.run.vm01.stdout: python3-babel-2.9.1-2.el9.noarch 2026-03-09T20:50:28.316 INFO:teuthology.orchestra.run.vm01.stdout: python3-backports-tarfile-1.2.0-1.el9.noarch 2026-03-09T20:50:28.316 INFO:teuthology.orchestra.run.vm01.stdout: python3-bcrypt-3.2.2-1.el9.x86_64 2026-03-09T20:50:28.316 INFO:teuthology.orchestra.run.vm01.stdout: python3-cachetools-4.2.4-1.el9.noarch 2026-03-09T20:50:28.316 INFO:teuthology.orchestra.run.vm01.stdout: python3-ceph-argparse-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T20:50:28.316 INFO:teuthology.orchestra.run.vm01.stdout: python3-ceph-common-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T20:50:28.316 INFO:teuthology.orchestra.run.vm01.stdout: python3-cephfs-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T20:50:28.316 INFO:teuthology.orchestra.run.vm01.stdout: python3-certifi-2023.05.07-4.el9.noarch 2026-03-09T20:50:28.316 INFO:teuthology.orchestra.run.vm01.stdout: python3-cffi-1.14.5-5.el9.x86_64 2026-03-09T20:50:28.316 INFO:teuthology.orchestra.run.vm01.stdout: python3-cheroot-10.0.1-4.el9.noarch 2026-03-09T20:50:28.316 INFO:teuthology.orchestra.run.vm01.stdout: python3-cherrypy-18.6.1-2.el9.noarch 2026-03-09T20:50:28.316 INFO:teuthology.orchestra.run.vm01.stdout: python3-cryptography-36.0.1-5.el9.x86_64 2026-03-09T20:50:28.316 INFO:teuthology.orchestra.run.vm01.stdout: python3-devel-3.9.25-3.el9.x86_64 2026-03-09T20:50:28.316 INFO:teuthology.orchestra.run.vm01.stdout: python3-google-auth-1:2.45.0-1.el9.noarch 2026-03-09T20:50:28.316 INFO:teuthology.orchestra.run.vm01.stdout: python3-grpcio-1.46.7-10.el9.x86_64 2026-03-09T20:50:28.316 INFO:teuthology.orchestra.run.vm01.stdout: python3-grpcio-tools-1.46.7-10.el9.x86_64 2026-03-09T20:50:28.316 INFO:teuthology.orchestra.run.vm01.stdout: python3-jaraco-8.2.1-3.el9.noarch 2026-03-09T20:50:28.316 INFO:teuthology.orchestra.run.vm01.stdout: python3-jaraco-classes-3.2.1-5.el9.noarch 2026-03-09T20:50:28.316 INFO:teuthology.orchestra.run.vm01.stdout: python3-jaraco-collections-3.0.0-8.el9.noarch 2026-03-09T20:50:28.316 INFO:teuthology.orchestra.run.vm01.stdout: python3-jaraco-context-6.0.1-3.el9.noarch 2026-03-09T20:50:28.316 INFO:teuthology.orchestra.run.vm01.stdout: python3-jaraco-functools-3.5.0-2.el9.noarch 2026-03-09T20:50:28.316 INFO:teuthology.orchestra.run.vm01.stdout: python3-jaraco-text-4.0.0-2.el9.noarch 2026-03-09T20:50:28.316 INFO:teuthology.orchestra.run.vm01.stdout: python3-jinja2-2.11.3-8.el9.noarch 2026-03-09T20:50:28.316 INFO:teuthology.orchestra.run.vm01.stdout: python3-jmespath-1.0.1-1.el9.noarch 2026-03-09T20:50:28.316 INFO:teuthology.orchestra.run.vm01.stdout: python3-kubernetes-1:26.1.0-3.el9.noarch 2026-03-09T20:50:28.316 INFO:teuthology.orchestra.run.vm01.stdout: python3-libstoragemgmt-1.10.1-1.el9.x86_64 2026-03-09T20:50:28.316 INFO:teuthology.orchestra.run.vm01.stdout: python3-logutils-0.3.5-21.el9.noarch 2026-03-09T20:50:28.316 INFO:teuthology.orchestra.run.vm01.stdout: python3-mako-1.1.4-6.el9.noarch 2026-03-09T20:50:28.316 INFO:teuthology.orchestra.run.vm01.stdout: python3-markupsafe-1.1.1-12.el9.x86_64 2026-03-09T20:50:28.316 INFO:teuthology.orchestra.run.vm01.stdout: python3-more-itertools-8.12.0-2.el9.noarch 2026-03-09T20:50:28.316 INFO:teuthology.orchestra.run.vm01.stdout: python3-natsort-7.1.1-5.el9.noarch 2026-03-09T20:50:28.316 INFO:teuthology.orchestra.run.vm01.stdout: python3-numpy-1:1.23.5-2.el9.x86_64 2026-03-09T20:50:28.316 INFO:teuthology.orchestra.run.vm01.stdout: python3-numpy-f2py-1:1.23.5-2.el9.x86_64 2026-03-09T20:50:28.316 INFO:teuthology.orchestra.run.vm01.stdout: python3-packaging-20.9-5.el9.noarch 2026-03-09T20:50:28.316 INFO:teuthology.orchestra.run.vm01.stdout: python3-pecan-1.4.2-3.el9.noarch 2026-03-09T20:50:28.316 INFO:teuthology.orchestra.run.vm01.stdout: python3-ply-3.11-14.el9.noarch 2026-03-09T20:50:28.316 INFO:teuthology.orchestra.run.vm01.stdout: python3-portend-3.1.0-2.el9.noarch 2026-03-09T20:50:28.316 INFO:teuthology.orchestra.run.vm01.stdout: python3-protobuf-3.14.0-17.el9.noarch 2026-03-09T20:50:28.316 INFO:teuthology.orchestra.run.vm01.stdout: python3-pyOpenSSL-21.0.0-1.el9.noarch 2026-03-09T20:50:28.316 INFO:teuthology.orchestra.run.vm01.stdout: python3-pyasn1-0.4.8-7.el9.noarch 2026-03-09T20:50:28.316 INFO:teuthology.orchestra.run.vm01.stdout: python3-pyasn1-modules-0.4.8-7.el9.noarch 2026-03-09T20:50:28.316 INFO:teuthology.orchestra.run.vm01.stdout: python3-pycparser-2.20-6.el9.noarch 2026-03-09T20:50:28.316 INFO:teuthology.orchestra.run.vm01.stdout: python3-rados-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T20:50:28.316 INFO:teuthology.orchestra.run.vm01.stdout: python3-rbd-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T20:50:28.316 INFO:teuthology.orchestra.run.vm01.stdout: python3-repoze-lru-0.7-16.el9.noarch 2026-03-09T20:50:28.316 INFO:teuthology.orchestra.run.vm01.stdout: python3-requests-2.25.1-10.el9.noarch 2026-03-09T20:50:28.316 INFO:teuthology.orchestra.run.vm01.stdout: python3-requests-oauthlib-1.3.0-12.el9.noarch 2026-03-09T20:50:28.316 INFO:teuthology.orchestra.run.vm01.stdout: python3-rgw-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T20:50:28.316 INFO:teuthology.orchestra.run.vm01.stdout: python3-routes-2.5.1-5.el9.noarch 2026-03-09T20:50:28.316 INFO:teuthology.orchestra.run.vm01.stdout: python3-rsa-4.9-2.el9.noarch 2026-03-09T20:50:28.316 INFO:teuthology.orchestra.run.vm01.stdout: python3-scipy-1.9.3-2.el9.x86_64 2026-03-09T20:50:28.316 INFO:teuthology.orchestra.run.vm01.stdout: python3-tempora-5.0.0-2.el9.noarch 2026-03-09T20:50:28.316 INFO:teuthology.orchestra.run.vm01.stdout: python3-toml-0.10.2-6.el9.noarch 2026-03-09T20:50:28.316 INFO:teuthology.orchestra.run.vm01.stdout: python3-typing-extensions-4.15.0-1.el9.noarch 2026-03-09T20:50:28.317 INFO:teuthology.orchestra.run.vm01.stdout: python3-urllib3-1.26.5-7.el9.noarch 2026-03-09T20:50:28.317 INFO:teuthology.orchestra.run.vm01.stdout: python3-webob-1.8.8-2.el9.noarch 2026-03-09T20:50:28.317 INFO:teuthology.orchestra.run.vm01.stdout: python3-websocket-client-1.2.3-2.el9.noarch 2026-03-09T20:50:28.317 INFO:teuthology.orchestra.run.vm01.stdout: python3-werkzeug-2.0.3-3.el9.1.noarch 2026-03-09T20:50:28.317 INFO:teuthology.orchestra.run.vm01.stdout: python3-xmltodict-0.12.0-15.el9.noarch 2026-03-09T20:50:28.317 INFO:teuthology.orchestra.run.vm01.stdout: python3-zc-lockfile-2.0-10.el9.noarch 2026-03-09T20:50:28.317 INFO:teuthology.orchestra.run.vm01.stdout: qatlib-25.08.0-2.el9.x86_64 2026-03-09T20:50:28.317 INFO:teuthology.orchestra.run.vm01.stdout: qatlib-service-25.08.0-2.el9.x86_64 2026-03-09T20:50:28.317 INFO:teuthology.orchestra.run.vm01.stdout: qatzip-libs-1.3.1-1.el9.x86_64 2026-03-09T20:50:28.317 INFO:teuthology.orchestra.run.vm01.stdout: rbd-fuse-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T20:50:28.317 INFO:teuthology.orchestra.run.vm01.stdout: rbd-mirror-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T20:50:28.317 INFO:teuthology.orchestra.run.vm01.stdout: rbd-nbd-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T20:50:28.317 INFO:teuthology.orchestra.run.vm01.stdout: re2-1:20211101-20.el9.x86_64 2026-03-09T20:50:28.317 INFO:teuthology.orchestra.run.vm01.stdout: socat-1.7.4.1-8.el9.x86_64 2026-03-09T20:50:28.317 INFO:teuthology.orchestra.run.vm01.stdout: thrift-0.15.0-4.el9.x86_64 2026-03-09T20:50:28.317 INFO:teuthology.orchestra.run.vm01.stdout: unzip-6.0-59.el9.x86_64 2026-03-09T20:50:28.317 INFO:teuthology.orchestra.run.vm01.stdout: xmlstarlet-1.6.1-20.el9.x86_64 2026-03-09T20:50:28.317 INFO:teuthology.orchestra.run.vm01.stdout: zip-3.0-35.el9.x86_64 2026-03-09T20:50:28.317 INFO:teuthology.orchestra.run.vm01.stdout: 2026-03-09T20:50:28.317 INFO:teuthology.orchestra.run.vm01.stdout:Complete! 2026-03-09T20:50:28.427 DEBUG:teuthology.parallel:result is None 2026-03-09T20:50:28.428 DEBUG:teuthology.packaging:Querying https://shaman.ceph.com/api/search?status=ready&project=ceph&flavor=default&distros=centos%2F9%2Fx86_64&sha1=e911bdebe5c8faa3800735d1568fcdca65db60df 2026-03-09T20:50:29.026 DEBUG:teuthology.orchestra.run.vm01:> rpm -q ceph --qf '%{VERSION}-%{RELEASE}' 2026-03-09T20:50:29.046 INFO:teuthology.orchestra.run.vm01.stdout:19.2.3-678.ge911bdeb.el9 2026-03-09T20:50:29.047 INFO:teuthology.packaging:The installed version of ceph is 19.2.3-678.ge911bdeb.el9 2026-03-09T20:50:29.047 INFO:teuthology.task.install:The correct ceph version 19.2.3-678.ge911bdeb is installed. 2026-03-09T20:50:29.048 DEBUG:teuthology.packaging:Querying https://shaman.ceph.com/api/search?status=ready&project=ceph&flavor=default&distros=centos%2F9%2Fx86_64&sha1=e911bdebe5c8faa3800735d1568fcdca65db60df 2026-03-09T20:50:29.742 DEBUG:teuthology.orchestra.run.vm06:> rpm -q ceph --qf '%{VERSION}-%{RELEASE}' 2026-03-09T20:50:29.764 INFO:teuthology.orchestra.run.vm06.stdout:19.2.3-678.ge911bdeb.el9 2026-03-09T20:50:29.764 INFO:teuthology.packaging:The installed version of ceph is 19.2.3-678.ge911bdeb.el9 2026-03-09T20:50:29.764 INFO:teuthology.task.install:The correct ceph version 19.2.3-678.ge911bdeb is installed. 2026-03-09T20:50:29.765 DEBUG:teuthology.packaging:Querying https://shaman.ceph.com/api/search?status=ready&project=ceph&flavor=default&distros=centos%2F9%2Fx86_64&sha1=e911bdebe5c8faa3800735d1568fcdca65db60df 2026-03-09T20:50:30.385 DEBUG:teuthology.orchestra.run.vm08:> rpm -q ceph --qf '%{VERSION}-%{RELEASE}' 2026-03-09T20:50:30.406 INFO:teuthology.orchestra.run.vm08.stdout:19.2.3-678.ge911bdeb.el9 2026-03-09T20:50:30.407 INFO:teuthology.packaging:The installed version of ceph is 19.2.3-678.ge911bdeb.el9 2026-03-09T20:50:30.407 INFO:teuthology.task.install:The correct ceph version 19.2.3-678.ge911bdeb is installed. 2026-03-09T20:50:30.408 INFO:teuthology.task.install.util:Shipping valgrind.supp... 2026-03-09T20:50:30.408 DEBUG:teuthology.orchestra.run.vm01:> set -ex 2026-03-09T20:50:30.408 DEBUG:teuthology.orchestra.run.vm01:> sudo dd of=/home/ubuntu/cephtest/valgrind.supp 2026-03-09T20:50:30.439 DEBUG:teuthology.orchestra.run.vm06:> set -ex 2026-03-09T20:50:30.439 DEBUG:teuthology.orchestra.run.vm06:> sudo dd of=/home/ubuntu/cephtest/valgrind.supp 2026-03-09T20:50:30.474 DEBUG:teuthology.orchestra.run.vm08:> set -ex 2026-03-09T20:50:30.475 DEBUG:teuthology.orchestra.run.vm08:> sudo dd of=/home/ubuntu/cephtest/valgrind.supp 2026-03-09T20:50:30.509 INFO:teuthology.task.install.util:Shipping 'daemon-helper'... 2026-03-09T20:50:30.509 DEBUG:teuthology.orchestra.run.vm01:> set -ex 2026-03-09T20:50:30.509 DEBUG:teuthology.orchestra.run.vm01:> sudo dd of=/usr/bin/daemon-helper 2026-03-09T20:50:30.534 DEBUG:teuthology.orchestra.run.vm01:> sudo chmod a=rx -- /usr/bin/daemon-helper 2026-03-09T20:50:30.603 DEBUG:teuthology.orchestra.run.vm06:> set -ex 2026-03-09T20:50:30.603 DEBUG:teuthology.orchestra.run.vm06:> sudo dd of=/usr/bin/daemon-helper 2026-03-09T20:50:30.634 DEBUG:teuthology.orchestra.run.vm06:> sudo chmod a=rx -- /usr/bin/daemon-helper 2026-03-09T20:50:30.704 DEBUG:teuthology.orchestra.run.vm08:> set -ex 2026-03-09T20:50:30.704 DEBUG:teuthology.orchestra.run.vm08:> sudo dd of=/usr/bin/daemon-helper 2026-03-09T20:50:30.739 DEBUG:teuthology.orchestra.run.vm08:> sudo chmod a=rx -- /usr/bin/daemon-helper 2026-03-09T20:50:30.810 INFO:teuthology.task.install.util:Shipping 'adjust-ulimits'... 2026-03-09T20:50:30.810 DEBUG:teuthology.orchestra.run.vm01:> set -ex 2026-03-09T20:50:30.810 DEBUG:teuthology.orchestra.run.vm01:> sudo dd of=/usr/bin/adjust-ulimits 2026-03-09T20:50:30.836 DEBUG:teuthology.orchestra.run.vm01:> sudo chmod a=rx -- /usr/bin/adjust-ulimits 2026-03-09T20:50:30.907 DEBUG:teuthology.orchestra.run.vm06:> set -ex 2026-03-09T20:50:30.907 DEBUG:teuthology.orchestra.run.vm06:> sudo dd of=/usr/bin/adjust-ulimits 2026-03-09T20:50:30.931 DEBUG:teuthology.orchestra.run.vm06:> sudo chmod a=rx -- /usr/bin/adjust-ulimits 2026-03-09T20:50:30.999 DEBUG:teuthology.orchestra.run.vm08:> set -ex 2026-03-09T20:50:30.999 DEBUG:teuthology.orchestra.run.vm08:> sudo dd of=/usr/bin/adjust-ulimits 2026-03-09T20:50:31.027 DEBUG:teuthology.orchestra.run.vm08:> sudo chmod a=rx -- /usr/bin/adjust-ulimits 2026-03-09T20:50:31.097 INFO:teuthology.task.install.util:Shipping 'stdin-killer'... 2026-03-09T20:50:31.104 DEBUG:teuthology.orchestra.run.vm01:> set -ex 2026-03-09T20:50:31.104 DEBUG:teuthology.orchestra.run.vm01:> sudo dd of=/usr/bin/stdin-killer 2026-03-09T20:50:31.130 DEBUG:teuthology.orchestra.run.vm01:> sudo chmod a=rx -- /usr/bin/stdin-killer 2026-03-09T20:50:31.197 DEBUG:teuthology.orchestra.run.vm06:> set -ex 2026-03-09T20:50:31.202 DEBUG:teuthology.orchestra.run.vm06:> sudo dd of=/usr/bin/stdin-killer 2026-03-09T20:50:31.231 DEBUG:teuthology.orchestra.run.vm06:> sudo chmod a=rx -- /usr/bin/stdin-killer 2026-03-09T20:50:31.298 DEBUG:teuthology.orchestra.run.vm08:> set -ex 2026-03-09T20:50:31.298 DEBUG:teuthology.orchestra.run.vm08:> sudo dd of=/usr/bin/stdin-killer 2026-03-09T20:50:31.324 DEBUG:teuthology.orchestra.run.vm08:> sudo chmod a=rx -- /usr/bin/stdin-killer 2026-03-09T20:50:31.391 INFO:teuthology.run_tasks:Running task cephadm... 2026-03-09T20:50:31.499 INFO:tasks.cephadm:Config: {'conf': {'global': {'mon election default strategy': 1}, 'mgr': {'debug mgr': 20, 'debug ms': 1, 'mgr/cephadm/use_agent': False}, 'mon': {'debug mon': 20, 'debug ms': 1, 'debug paxos': 20}, 'osd': {'debug ms': 1, 'debug osd': 20, 'osd mclock iops capacity threshold hdd': 49000}}, 'flavor': 'default', 'log-ignorelist': ['\\(MDS_ALL_DOWN\\)', '\\(MDS_UP_LESS_THAN_MAX\\)', 'MON_DOWN', 'POOL_APP_NOT_ENABLED', 'mon down', 'mons down', 'out of quorum', 'CEPHADM_FAILED_DAEMON'], 'log-only-match': ['CEPHADM_'], 'sha1': 'e911bdebe5c8faa3800735d1568fcdca65db60df'} 2026-03-09T20:50:31.499 INFO:tasks.cephadm:Cluster image is quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df 2026-03-09T20:50:31.499 INFO:tasks.cephadm:Cluster fsid is 9cb345a8-1bf9-11f1-a195-0375563c5891 2026-03-09T20:50:31.499 INFO:tasks.cephadm:Choosing monitor IPs and ports... 2026-03-09T20:50:31.499 INFO:tasks.cephadm:Monitor IPs: {'mon.a': '192.168.123.101', 'mon.b': '192.168.123.106', 'mon.c': '192.168.123.108'} 2026-03-09T20:50:31.499 INFO:tasks.cephadm:First mon is mon.a on vm01 2026-03-09T20:50:31.499 INFO:tasks.cephadm:First mgr is a 2026-03-09T20:50:31.499 INFO:tasks.cephadm:Normalizing hostnames... 2026-03-09T20:50:31.499 DEBUG:teuthology.orchestra.run.vm01:> sudo hostname $(hostname -s) 2026-03-09T20:50:31.524 DEBUG:teuthology.orchestra.run.vm06:> sudo hostname $(hostname -s) 2026-03-09T20:50:31.551 DEBUG:teuthology.orchestra.run.vm08:> sudo hostname $(hostname -s) 2026-03-09T20:50:31.595 INFO:tasks.cephadm:Downloading "compiled" cephadm from cachra 2026-03-09T20:50:31.603 DEBUG:teuthology.packaging:Querying https://shaman.ceph.com/api/search?status=ready&project=ceph&flavor=default&distros=centos%2F9%2Fx86_64&sha1=e911bdebe5c8faa3800735d1568fcdca65db60df 2026-03-09T20:50:32.250 INFO:tasks.cephadm:builder_project result: [{'url': 'https://3.chacra.ceph.com/r/ceph/squid/e911bdebe5c8faa3800735d1568fcdca65db60df/centos/9/flavors/default/', 'chacra_url': 'https://3.chacra.ceph.com/repos/ceph/squid/e911bdebe5c8faa3800735d1568fcdca65db60df/centos/9/flavors/default/', 'ref': 'squid', 'sha1': 'e911bdebe5c8faa3800735d1568fcdca65db60df', 'distro': 'centos', 'distro_version': '9', 'distro_codename': None, 'modified': '2026-02-25 18:55:15.146628', 'status': 'ready', 'flavor': 'default', 'project': 'ceph', 'archs': ['source', 'x86_64'], 'extra': {'version': '19.2.3-678-ge911bdeb', 'package_manager_version': '19.2.3-678.ge911bdeb', 'build_url': 'https://jenkins.ceph.com/job/ceph-dev-pipeline/3275/', 'root_build_cause': '', 'node_name': '10.20.192.26+soko16', 'job_name': 'ceph-dev-pipeline'}}] 2026-03-09T20:50:32.939 INFO:tasks.util.chacra:got chacra host 3.chacra.ceph.com, ref squid, sha1 e911bdebe5c8faa3800735d1568fcdca65db60df from https://shaman.ceph.com/api/search/?project=ceph&distros=centos%2F9%2Fx86_64&flavor=default&sha1=e911bdebe5c8faa3800735d1568fcdca65db60df 2026-03-09T20:50:32.941 INFO:tasks.cephadm:Discovered cachra url: https://3.chacra.ceph.com/binaries/ceph/squid/e911bdebe5c8faa3800735d1568fcdca65db60df/centos/9/x86_64/flavors/default/cephadm 2026-03-09T20:50:32.941 INFO:tasks.cephadm:Downloading cephadm from url: https://3.chacra.ceph.com/binaries/ceph/squid/e911bdebe5c8faa3800735d1568fcdca65db60df/centos/9/x86_64/flavors/default/cephadm 2026-03-09T20:50:32.941 DEBUG:teuthology.orchestra.run.vm01:> curl --silent -L https://3.chacra.ceph.com/binaries/ceph/squid/e911bdebe5c8faa3800735d1568fcdca65db60df/centos/9/x86_64/flavors/default/cephadm > /home/ubuntu/cephtest/cephadm && ls -l /home/ubuntu/cephtest/cephadm 2026-03-09T20:50:34.344 INFO:teuthology.orchestra.run.vm01.stdout:-rw-r--r--. 1 ubuntu ubuntu 788355 Mar 9 20:50 /home/ubuntu/cephtest/cephadm 2026-03-09T20:50:34.344 DEBUG:teuthology.orchestra.run.vm06:> curl --silent -L https://3.chacra.ceph.com/binaries/ceph/squid/e911bdebe5c8faa3800735d1568fcdca65db60df/centos/9/x86_64/flavors/default/cephadm > /home/ubuntu/cephtest/cephadm && ls -l /home/ubuntu/cephtest/cephadm 2026-03-09T20:50:35.726 INFO:teuthology.orchestra.run.vm06.stdout:-rw-r--r--. 1 ubuntu ubuntu 788355 Mar 9 20:50 /home/ubuntu/cephtest/cephadm 2026-03-09T20:50:35.726 DEBUG:teuthology.orchestra.run.vm08:> curl --silent -L https://3.chacra.ceph.com/binaries/ceph/squid/e911bdebe5c8faa3800735d1568fcdca65db60df/centos/9/x86_64/flavors/default/cephadm > /home/ubuntu/cephtest/cephadm && ls -l /home/ubuntu/cephtest/cephadm 2026-03-09T20:50:37.106 INFO:teuthology.orchestra.run.vm08.stdout:-rw-r--r--. 1 ubuntu ubuntu 788355 Mar 9 20:50 /home/ubuntu/cephtest/cephadm 2026-03-09T20:50:37.106 DEBUG:teuthology.orchestra.run.vm01:> test -s /home/ubuntu/cephtest/cephadm && test $(stat -c%s /home/ubuntu/cephtest/cephadm) -gt 1000 && chmod +x /home/ubuntu/cephtest/cephadm 2026-03-09T20:50:37.125 DEBUG:teuthology.orchestra.run.vm06:> test -s /home/ubuntu/cephtest/cephadm && test $(stat -c%s /home/ubuntu/cephtest/cephadm) -gt 1000 && chmod +x /home/ubuntu/cephtest/cephadm 2026-03-09T20:50:37.140 DEBUG:teuthology.orchestra.run.vm08:> test -s /home/ubuntu/cephtest/cephadm && test $(stat -c%s /home/ubuntu/cephtest/cephadm) -gt 1000 && chmod +x /home/ubuntu/cephtest/cephadm 2026-03-09T20:50:37.162 INFO:tasks.cephadm:Pulling image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df on all hosts... 2026-03-09T20:50:37.163 DEBUG:teuthology.orchestra.run.vm01:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df pull 2026-03-09T20:50:37.168 DEBUG:teuthology.orchestra.run.vm06:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df pull 2026-03-09T20:50:37.182 DEBUG:teuthology.orchestra.run.vm08:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df pull 2026-03-09T20:50:37.354 INFO:teuthology.orchestra.run.vm01.stderr:Pulling container image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df... 2026-03-09T20:50:37.399 INFO:teuthology.orchestra.run.vm06.stderr:Pulling container image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df... 2026-03-09T20:50:37.401 INFO:teuthology.orchestra.run.vm08.stderr:Pulling container image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df... 2026-03-09T20:51:29.744 INFO:teuthology.orchestra.run.vm06.stdout:{ 2026-03-09T20:51:29.744 INFO:teuthology.orchestra.run.vm06.stdout: "ceph_version": "ceph version 19.2.3-678-ge911bdeb (e911bdebe5c8faa3800735d1568fcdca65db60df) squid (stable)", 2026-03-09T20:51:29.744 INFO:teuthology.orchestra.run.vm06.stdout: "image_id": "654f31e6858eb235bbece362255b685a945f2b6a367e2b88c4930c984fbb214c", 2026-03-09T20:51:29.744 INFO:teuthology.orchestra.run.vm06.stdout: "repo_digests": [ 2026-03-09T20:51:29.744 INFO:teuthology.orchestra.run.vm06.stdout: "quay.ceph.io/ceph-ci/ceph@sha256:8fda260ab1d2d3118a1622f7df75f44f285dfe74e71793626152a711c12bf2cc" 2026-03-09T20:51:29.744 INFO:teuthology.orchestra.run.vm06.stdout: ] 2026-03-09T20:51:29.744 INFO:teuthology.orchestra.run.vm06.stdout:} 2026-03-09T20:51:32.670 INFO:teuthology.orchestra.run.vm08.stdout:{ 2026-03-09T20:51:32.671 INFO:teuthology.orchestra.run.vm08.stdout: "ceph_version": "ceph version 19.2.3-678-ge911bdeb (e911bdebe5c8faa3800735d1568fcdca65db60df) squid (stable)", 2026-03-09T20:51:32.671 INFO:teuthology.orchestra.run.vm08.stdout: "image_id": "654f31e6858eb235bbece362255b685a945f2b6a367e2b88c4930c984fbb214c", 2026-03-09T20:51:32.671 INFO:teuthology.orchestra.run.vm08.stdout: "repo_digests": [ 2026-03-09T20:51:32.671 INFO:teuthology.orchestra.run.vm08.stdout: "quay.ceph.io/ceph-ci/ceph@sha256:8fda260ab1d2d3118a1622f7df75f44f285dfe74e71793626152a711c12bf2cc" 2026-03-09T20:51:32.671 INFO:teuthology.orchestra.run.vm08.stdout: ] 2026-03-09T20:51:32.671 INFO:teuthology.orchestra.run.vm08.stdout:} 2026-03-09T20:51:36.600 INFO:teuthology.orchestra.run.vm01.stdout:{ 2026-03-09T20:51:36.600 INFO:teuthology.orchestra.run.vm01.stdout: "ceph_version": "ceph version 19.2.3-678-ge911bdeb (e911bdebe5c8faa3800735d1568fcdca65db60df) squid (stable)", 2026-03-09T20:51:36.600 INFO:teuthology.orchestra.run.vm01.stdout: "image_id": "654f31e6858eb235bbece362255b685a945f2b6a367e2b88c4930c984fbb214c", 2026-03-09T20:51:36.600 INFO:teuthology.orchestra.run.vm01.stdout: "repo_digests": [ 2026-03-09T20:51:36.600 INFO:teuthology.orchestra.run.vm01.stdout: "quay.ceph.io/ceph-ci/ceph@sha256:8fda260ab1d2d3118a1622f7df75f44f285dfe74e71793626152a711c12bf2cc" 2026-03-09T20:51:36.600 INFO:teuthology.orchestra.run.vm01.stdout: ] 2026-03-09T20:51:36.600 INFO:teuthology.orchestra.run.vm01.stdout:} 2026-03-09T20:51:36.621 DEBUG:teuthology.orchestra.run.vm01:> sudo mkdir -p /etc/ceph 2026-03-09T20:51:36.650 DEBUG:teuthology.orchestra.run.vm06:> sudo mkdir -p /etc/ceph 2026-03-09T20:51:36.678 DEBUG:teuthology.orchestra.run.vm08:> sudo mkdir -p /etc/ceph 2026-03-09T20:51:36.708 DEBUG:teuthology.orchestra.run.vm01:> sudo chmod 777 /etc/ceph 2026-03-09T20:51:36.735 DEBUG:teuthology.orchestra.run.vm06:> sudo chmod 777 /etc/ceph 2026-03-09T20:51:36.765 DEBUG:teuthology.orchestra.run.vm08:> sudo chmod 777 /etc/ceph 2026-03-09T20:51:36.794 INFO:tasks.cephadm:Writing seed config... 2026-03-09T20:51:36.794 INFO:tasks.cephadm: override: [global] mon election default strategy = 1 2026-03-09T20:51:36.795 INFO:tasks.cephadm: override: [mgr] debug mgr = 20 2026-03-09T20:51:36.795 INFO:tasks.cephadm: override: [mgr] debug ms = 1 2026-03-09T20:51:36.795 INFO:tasks.cephadm: override: [mgr] mgr/cephadm/use_agent = False 2026-03-09T20:51:36.795 INFO:tasks.cephadm: override: [mon] debug mon = 20 2026-03-09T20:51:36.795 INFO:tasks.cephadm: override: [mon] debug ms = 1 2026-03-09T20:51:36.795 INFO:tasks.cephadm: override: [mon] debug paxos = 20 2026-03-09T20:51:36.795 INFO:tasks.cephadm: override: [osd] debug ms = 1 2026-03-09T20:51:36.795 INFO:tasks.cephadm: override: [osd] debug osd = 20 2026-03-09T20:51:36.795 INFO:tasks.cephadm: override: [osd] osd mclock iops capacity threshold hdd = 49000 2026-03-09T20:51:36.795 DEBUG:teuthology.orchestra.run.vm01:> set -ex 2026-03-09T20:51:36.795 DEBUG:teuthology.orchestra.run.vm01:> dd of=/home/ubuntu/cephtest/seed.ceph.conf 2026-03-09T20:51:36.810 DEBUG:tasks.cephadm:Final config: [global] # make logging friendly to teuthology log_to_file = true log_to_stderr = false log to journald = false mon cluster log to file = true mon cluster log file level = debug mon clock drift allowed = 1.000 # replicate across OSDs, not hosts osd crush chooseleaf type = 0 #osd pool default size = 2 osd pool default erasure code profile = plugin=jerasure technique=reed_sol_van k=2 m=1 crush-failure-domain=osd # enable some debugging auth debug = true ms die on old message = true ms die on bug = true debug asserts on shutdown = true # adjust warnings mon max pg per osd = 10000# >= luminous mon pg warn max object skew = 0 mon osd allow primary affinity = true mon osd allow pg remap = true mon warn on legacy crush tunables = false mon warn on crush straw calc version zero = false mon warn on no sortbitwise = false mon warn on osd down out interval zero = false mon warn on too few osds = false mon_warn_on_pool_pg_num_not_power_of_two = false # disable pg_autoscaler by default for new pools osd_pool_default_pg_autoscale_mode = off # tests delete pools mon allow pool delete = true fsid = 9cb345a8-1bf9-11f1-a195-0375563c5891 mon election default strategy = 1 [osd] osd scrub load threshold = 5.0 osd scrub max interval = 600 osd mclock profile = high_recovery_ops osd recover clone overlap = true osd recovery max chunk = 1048576 osd deep scrub update digest min age = 30 osd map max advance = 10 osd memory target autotune = true # debugging osd debug shutdown = true osd debug op order = true osd debug verify stray on activate = true osd debug pg log writeout = true osd debug verify cached snaps = true osd debug verify missing on start = true osd debug misdirected ops = true osd op queue = debug_random osd op queue cut off = debug_random osd shutdown pgref assert = true bdev debug aio = true osd sloppy crc = true debug ms = 1 debug osd = 20 osd mclock iops capacity threshold hdd = 49000 [mgr] mon reweight min pgs per osd = 4 mon reweight min bytes per osd = 10 mgr/telemetry/nag = false debug mgr = 20 debug ms = 1 mgr/cephadm/use_agent = False [mon] mon data avail warn = 5 mon mgr mkfs grace = 240 mon reweight min pgs per osd = 4 mon osd reporter subtree level = osd mon osd prime pg temp = true mon reweight min bytes per osd = 10 # rotate auth tickets quickly to exercise renewal paths auth mon ticket ttl = 660# 11m auth service ticket ttl = 240# 4m # don't complain about global id reclaim mon_warn_on_insecure_global_id_reclaim = false mon_warn_on_insecure_global_id_reclaim_allowed = false debug mon = 20 debug ms = 1 debug paxos = 20 [client.rgw] rgw cache enabled = true rgw enable ops log = true rgw enable usage log = true 2026-03-09T20:51:36.811 DEBUG:teuthology.orchestra.run.vm01:mon.a> sudo journalctl -f -n 0 -u ceph-9cb345a8-1bf9-11f1-a195-0375563c5891@mon.a.service 2026-03-09T20:51:36.853 DEBUG:teuthology.orchestra.run.vm01:mgr.a> sudo journalctl -f -n 0 -u ceph-9cb345a8-1bf9-11f1-a195-0375563c5891@mgr.a.service 2026-03-09T20:51:36.895 INFO:tasks.cephadm:Bootstrapping... 2026-03-09T20:51:36.895 DEBUG:teuthology.orchestra.run.vm01:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df -v bootstrap --fsid 9cb345a8-1bf9-11f1-a195-0375563c5891 --config /home/ubuntu/cephtest/seed.ceph.conf --output-config /etc/ceph/ceph.conf --output-keyring /etc/ceph/ceph.client.admin.keyring --output-pub-ssh-key /home/ubuntu/cephtest/ceph.pub --mon-id a --mgr-id a --orphan-initial-daemons --skip-monitoring-stack --mon-ip 192.168.123.101 --skip-admin-label && sudo chmod +r /etc/ceph/ceph.client.admin.keyring 2026-03-09T20:51:37.041 INFO:teuthology.orchestra.run.vm01.stdout:-------------------------------------------------------------------------------- 2026-03-09T20:51:37.041 INFO:teuthology.orchestra.run.vm01.stdout:cephadm ['--image', 'quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df', '-v', 'bootstrap', '--fsid', '9cb345a8-1bf9-11f1-a195-0375563c5891', '--config', '/home/ubuntu/cephtest/seed.ceph.conf', '--output-config', '/etc/ceph/ceph.conf', '--output-keyring', '/etc/ceph/ceph.client.admin.keyring', '--output-pub-ssh-key', '/home/ubuntu/cephtest/ceph.pub', '--mon-id', 'a', '--mgr-id', 'a', '--orphan-initial-daemons', '--skip-monitoring-stack', '--mon-ip', '192.168.123.101', '--skip-admin-label'] 2026-03-09T20:51:37.041 INFO:teuthology.orchestra.run.vm01.stderr:Specifying an fsid for your cluster offers no advantages and may increase the likelihood of fsid conflicts. 2026-03-09T20:51:37.041 INFO:teuthology.orchestra.run.vm01.stdout:Verifying podman|docker is present... 2026-03-09T20:51:37.062 INFO:teuthology.orchestra.run.vm01.stdout:/bin/podman: stdout 5.8.0 2026-03-09T20:51:37.062 INFO:teuthology.orchestra.run.vm01.stdout:Verifying lvm2 is present... 2026-03-09T20:51:37.062 INFO:teuthology.orchestra.run.vm01.stdout:Verifying time synchronization is in place... 2026-03-09T20:51:37.072 INFO:teuthology.orchestra.run.vm01.stdout:Non-zero exit code 1 from systemctl is-enabled chrony.service 2026-03-09T20:51:37.072 INFO:teuthology.orchestra.run.vm01.stdout:systemctl: stderr Failed to get unit file state for chrony.service: No such file or directory 2026-03-09T20:51:37.079 INFO:teuthology.orchestra.run.vm01.stdout:Non-zero exit code 3 from systemctl is-active chrony.service 2026-03-09T20:51:37.079 INFO:teuthology.orchestra.run.vm01.stdout:systemctl: stdout inactive 2026-03-09T20:51:37.086 INFO:teuthology.orchestra.run.vm01.stdout:systemctl: stdout enabled 2026-03-09T20:51:37.092 INFO:teuthology.orchestra.run.vm01.stdout:systemctl: stdout active 2026-03-09T20:51:37.092 INFO:teuthology.orchestra.run.vm01.stdout:Unit chronyd.service is enabled and running 2026-03-09T20:51:37.092 INFO:teuthology.orchestra.run.vm01.stdout:Repeating the final host check... 2026-03-09T20:51:37.113 INFO:teuthology.orchestra.run.vm01.stdout:/bin/podman: stdout 5.8.0 2026-03-09T20:51:37.113 INFO:teuthology.orchestra.run.vm01.stdout:podman (/bin/podman) version 5.8.0 is present 2026-03-09T20:51:37.113 INFO:teuthology.orchestra.run.vm01.stdout:systemctl is present 2026-03-09T20:51:37.113 INFO:teuthology.orchestra.run.vm01.stdout:lvcreate is present 2026-03-09T20:51:37.120 INFO:teuthology.orchestra.run.vm01.stdout:Non-zero exit code 1 from systemctl is-enabled chrony.service 2026-03-09T20:51:37.120 INFO:teuthology.orchestra.run.vm01.stdout:systemctl: stderr Failed to get unit file state for chrony.service: No such file or directory 2026-03-09T20:51:37.127 INFO:teuthology.orchestra.run.vm01.stdout:Non-zero exit code 3 from systemctl is-active chrony.service 2026-03-09T20:51:37.127 INFO:teuthology.orchestra.run.vm01.stdout:systemctl: stdout inactive 2026-03-09T20:51:37.134 INFO:teuthology.orchestra.run.vm01.stdout:systemctl: stdout enabled 2026-03-09T20:51:37.141 INFO:teuthology.orchestra.run.vm01.stdout:systemctl: stdout active 2026-03-09T20:51:37.142 INFO:teuthology.orchestra.run.vm01.stdout:Unit chronyd.service is enabled and running 2026-03-09T20:51:37.142 INFO:teuthology.orchestra.run.vm01.stdout:Host looks OK 2026-03-09T20:51:37.142 INFO:teuthology.orchestra.run.vm01.stdout:Cluster fsid: 9cb345a8-1bf9-11f1-a195-0375563c5891 2026-03-09T20:51:37.142 INFO:teuthology.orchestra.run.vm01.stdout:Acquiring lock 140484603072272 on /run/cephadm/9cb345a8-1bf9-11f1-a195-0375563c5891.lock 2026-03-09T20:51:37.142 INFO:teuthology.orchestra.run.vm01.stdout:Lock 140484603072272 acquired on /run/cephadm/9cb345a8-1bf9-11f1-a195-0375563c5891.lock 2026-03-09T20:51:37.142 INFO:teuthology.orchestra.run.vm01.stdout:Verifying IP 192.168.123.101 port 3300 ... 2026-03-09T20:51:37.143 INFO:teuthology.orchestra.run.vm01.stdout:Verifying IP 192.168.123.101 port 6789 ... 2026-03-09T20:51:37.143 INFO:teuthology.orchestra.run.vm01.stdout:Base mon IP(s) is [192.168.123.101:3300, 192.168.123.101:6789], mon addrv is [v2:192.168.123.101:3300,v1:192.168.123.101:6789] 2026-03-09T20:51:37.146 INFO:teuthology.orchestra.run.vm01.stdout:/sbin/ip: stdout default via 192.168.123.1 dev eth0 proto dhcp src 192.168.123.101 metric 100 2026-03-09T20:51:37.146 INFO:teuthology.orchestra.run.vm01.stdout:/sbin/ip: stdout 192.168.123.0/24 dev eth0 proto kernel scope link src 192.168.123.101 metric 100 2026-03-09T20:51:37.149 INFO:teuthology.orchestra.run.vm01.stdout:/sbin/ip: stdout ::1 dev lo proto kernel metric 256 pref medium 2026-03-09T20:51:37.149 INFO:teuthology.orchestra.run.vm01.stdout:/sbin/ip: stdout fe80::/64 dev eth0 proto kernel metric 1024 pref medium 2026-03-09T20:51:37.151 INFO:teuthology.orchestra.run.vm01.stdout:/sbin/ip: stdout 1: lo: mtu 65536 state UNKNOWN qlen 1000 2026-03-09T20:51:37.151 INFO:teuthology.orchestra.run.vm01.stdout:/sbin/ip: stdout inet6 ::1/128 scope host 2026-03-09T20:51:37.151 INFO:teuthology.orchestra.run.vm01.stdout:/sbin/ip: stdout valid_lft forever preferred_lft forever 2026-03-09T20:51:37.151 INFO:teuthology.orchestra.run.vm01.stdout:/sbin/ip: stdout 2: eth0: mtu 1500 state UP qlen 1000 2026-03-09T20:51:37.151 INFO:teuthology.orchestra.run.vm01.stdout:/sbin/ip: stdout inet6 fe80::5055:ff:fe00:1/64 scope link noprefixroute 2026-03-09T20:51:37.151 INFO:teuthology.orchestra.run.vm01.stdout:/sbin/ip: stdout valid_lft forever preferred_lft forever 2026-03-09T20:51:37.151 INFO:teuthology.orchestra.run.vm01.stdout:Mon IP `192.168.123.101` is in CIDR network `192.168.123.0/24` 2026-03-09T20:51:37.152 INFO:teuthology.orchestra.run.vm01.stdout:Mon IP `192.168.123.101` is in CIDR network `192.168.123.0/24` 2026-03-09T20:51:37.152 INFO:teuthology.orchestra.run.vm01.stdout:Inferred mon public CIDR from local network configuration ['192.168.123.0/24', '192.168.123.0/24'] 2026-03-09T20:51:37.152 INFO:teuthology.orchestra.run.vm01.stdout:Internal network (--cluster-network) has not been provided, OSD replication will default to the public_network 2026-03-09T20:51:37.152 INFO:teuthology.orchestra.run.vm01.stdout:Pulling container image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df... 2026-03-09T20:51:38.620 INFO:teuthology.orchestra.run.vm01.stdout:/bin/podman: stdout 654f31e6858eb235bbece362255b685a945f2b6a367e2b88c4930c984fbb214c 2026-03-09T20:51:38.620 INFO:teuthology.orchestra.run.vm01.stdout:/bin/podman: stderr Trying to pull quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df... 2026-03-09T20:51:38.620 INFO:teuthology.orchestra.run.vm01.stdout:/bin/podman: stderr Getting image source signatures 2026-03-09T20:51:38.620 INFO:teuthology.orchestra.run.vm01.stdout:/bin/podman: stderr Copying blob sha256:1752b8d01aa0dd33bbe0ab24e8316174c94fbdcd5d26252e2680bba0624747a7 2026-03-09T20:51:38.620 INFO:teuthology.orchestra.run.vm01.stdout:/bin/podman: stderr Copying blob sha256:8e380faede39ebd4286247457b408d979ab568aafd8389c42ec304b8cfba4e92 2026-03-09T20:51:38.620 INFO:teuthology.orchestra.run.vm01.stdout:/bin/podman: stderr Copying config sha256:654f31e6858eb235bbece362255b685a945f2b6a367e2b88c4930c984fbb214c 2026-03-09T20:51:38.620 INFO:teuthology.orchestra.run.vm01.stdout:/bin/podman: stderr Writing manifest to image destination 2026-03-09T20:51:39.023 INFO:teuthology.orchestra.run.vm01.stdout:ceph: stdout ceph version 19.2.3-678-ge911bdeb (e911bdebe5c8faa3800735d1568fcdca65db60df) squid (stable) 2026-03-09T20:51:39.023 INFO:teuthology.orchestra.run.vm01.stdout:Ceph version: ceph version 19.2.3-678-ge911bdeb (e911bdebe5c8faa3800735d1568fcdca65db60df) squid (stable) 2026-03-09T20:51:39.023 INFO:teuthology.orchestra.run.vm01.stdout:Extracting ceph user uid/gid from container image... 2026-03-09T20:51:39.813 INFO:teuthology.orchestra.run.vm01.stdout:stat: stdout 167 167 2026-03-09T20:51:39.813 INFO:teuthology.orchestra.run.vm01.stdout:Creating initial keys... 2026-03-09T20:51:40.598 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph-authtool: stdout AQDcMq9pTi1fEBAAUj2oPIr+ltgDphu2iXwY0Q== 2026-03-09T20:51:40.986 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph-authtool: stdout AQDcMq9p46H2KRAA8w7ifvwIPaPocp9ueVFukg== 2026-03-09T20:51:41.211 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph-authtool: stdout AQDdMq9psMo2BRAAO23VRD1QSiGQ5uTATw+dIg== 2026-03-09T20:51:41.211 INFO:teuthology.orchestra.run.vm01.stdout:Creating initial monmap... 2026-03-09T20:51:41.440 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/monmaptool: stdout /usr/bin/monmaptool: monmap file /tmp/monmap 2026-03-09T20:51:41.440 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/monmaptool: stdout setting min_mon_release = quincy 2026-03-09T20:51:41.440 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/monmaptool: stdout /usr/bin/monmaptool: set fsid to 9cb345a8-1bf9-11f1-a195-0375563c5891 2026-03-09T20:51:41.440 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/monmaptool: stdout /usr/bin/monmaptool: writing epoch 0 to /tmp/monmap (1 monitors) 2026-03-09T20:51:41.440 INFO:teuthology.orchestra.run.vm01.stdout:monmaptool for a [v2:192.168.123.101:3300,v1:192.168.123.101:6789] on /usr/bin/monmaptool: monmap file /tmp/monmap 2026-03-09T20:51:41.440 INFO:teuthology.orchestra.run.vm01.stdout:setting min_mon_release = quincy 2026-03-09T20:51:41.440 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/monmaptool: set fsid to 9cb345a8-1bf9-11f1-a195-0375563c5891 2026-03-09T20:51:41.440 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/monmaptool: writing epoch 0 to /tmp/monmap (1 monitors) 2026-03-09T20:51:41.440 INFO:teuthology.orchestra.run.vm01.stdout: 2026-03-09T20:51:41.440 INFO:teuthology.orchestra.run.vm01.stdout:Creating mon... 2026-03-09T20:51:41.867 INFO:teuthology.orchestra.run.vm01.stdout:create mon.a on 2026-03-09T20:51:42.027 INFO:teuthology.orchestra.run.vm01.stdout:systemctl: stderr Removed "/etc/systemd/system/multi-user.target.wants/ceph.target". 2026-03-09T20:51:42.156 INFO:teuthology.orchestra.run.vm01.stdout:systemctl: stderr Created symlink /etc/systemd/system/multi-user.target.wants/ceph.target → /etc/systemd/system/ceph.target. 2026-03-09T20:51:42.291 INFO:teuthology.orchestra.run.vm01.stdout:systemctl: stderr Created symlink /etc/systemd/system/multi-user.target.wants/ceph-9cb345a8-1bf9-11f1-a195-0375563c5891.target → /etc/systemd/system/ceph-9cb345a8-1bf9-11f1-a195-0375563c5891.target. 2026-03-09T20:51:42.291 INFO:teuthology.orchestra.run.vm01.stdout:systemctl: stderr Created symlink /etc/systemd/system/ceph.target.wants/ceph-9cb345a8-1bf9-11f1-a195-0375563c5891.target → /etc/systemd/system/ceph-9cb345a8-1bf9-11f1-a195-0375563c5891.target. 2026-03-09T20:51:42.455 INFO:teuthology.orchestra.run.vm01.stdout:Non-zero exit code 1 from systemctl reset-failed ceph-9cb345a8-1bf9-11f1-a195-0375563c5891@mon.a 2026-03-09T20:51:42.455 INFO:teuthology.orchestra.run.vm01.stdout:systemctl: stderr Failed to reset failed state of unit ceph-9cb345a8-1bf9-11f1-a195-0375563c5891@mon.a.service: Unit ceph-9cb345a8-1bf9-11f1-a195-0375563c5891@mon.a.service not loaded. 2026-03-09T20:51:42.608 INFO:teuthology.orchestra.run.vm01.stdout:systemctl: stderr Created symlink /etc/systemd/system/ceph-9cb345a8-1bf9-11f1-a195-0375563c5891.target.wants/ceph-9cb345a8-1bf9-11f1-a195-0375563c5891@mon.a.service → /etc/systemd/system/ceph-9cb345a8-1bf9-11f1-a195-0375563c5891@.service. 2026-03-09T20:51:43.051 INFO:teuthology.orchestra.run.vm01.stdout:firewalld does not appear to be present 2026-03-09T20:51:43.051 INFO:teuthology.orchestra.run.vm01.stdout:Not possible to enable service . firewalld.service is not available 2026-03-09T20:51:43.051 INFO:teuthology.orchestra.run.vm01.stdout:Waiting for mon to start... 2026-03-09T20:51:43.051 INFO:teuthology.orchestra.run.vm01.stdout:Waiting for mon... 2026-03-09T20:51:43.168 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:51:43 vm01 systemd[1]: Started Ceph mon.a for 9cb345a8-1bf9-11f1-a195-0375563c5891. 2026-03-09T20:51:43.424 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:51:43 vm01 ceph-mon[53013]: mkfs 9cb345a8-1bf9-11f1-a195-0375563c5891 2026-03-09T20:51:43.424 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:51:43 vm01 ceph-mon[53013]: mon.a is new leader, mons a in quorum (ranks 0) 2026-03-09T20:51:43.486 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout cluster: 2026-03-09T20:51:43.486 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout id: 9cb345a8-1bf9-11f1-a195-0375563c5891 2026-03-09T20:51:43.486 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout health: HEALTH_OK 2026-03-09T20:51:43.486 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout 2026-03-09T20:51:43.486 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout services: 2026-03-09T20:51:43.486 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout mon: 1 daemons, quorum a (age 0.164012s) 2026-03-09T20:51:43.486 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout mgr: no daemons active 2026-03-09T20:51:43.486 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout osd: 0 osds: 0 up, 0 in 2026-03-09T20:51:43.487 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout 2026-03-09T20:51:43.487 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout data: 2026-03-09T20:51:43.487 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout pools: 0 pools, 0 pgs 2026-03-09T20:51:43.487 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout objects: 0 objects, 0 B 2026-03-09T20:51:43.487 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout usage: 0 B used, 0 B / 0 B avail 2026-03-09T20:51:43.487 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout pgs: 2026-03-09T20:51:43.487 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout 2026-03-09T20:51:43.487 INFO:teuthology.orchestra.run.vm01.stdout:mon is available 2026-03-09T20:51:43.487 INFO:teuthology.orchestra.run.vm01.stdout:Assimilating anything we can from ceph.conf... 2026-03-09T20:51:43.803 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout 2026-03-09T20:51:43.803 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout [global] 2026-03-09T20:51:43.803 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout fsid = 9cb345a8-1bf9-11f1-a195-0375563c5891 2026-03-09T20:51:43.803 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout mon_cluster_log_file_level = debug 2026-03-09T20:51:43.803 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout mon_host = [v2:192.168.123.101:3300,v1:192.168.123.101:6789] 2026-03-09T20:51:43.803 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout mon_osd_allow_pg_remap = true 2026-03-09T20:51:43.803 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout mon_osd_allow_primary_affinity = true 2026-03-09T20:51:43.803 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout mon_warn_on_no_sortbitwise = false 2026-03-09T20:51:43.803 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout osd_crush_chooseleaf_type = 0 2026-03-09T20:51:43.803 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout 2026-03-09T20:51:43.803 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout [mgr] 2026-03-09T20:51:43.803 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout mgr/cephadm/use_agent = False 2026-03-09T20:51:43.803 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout mgr/telemetry/nag = false 2026-03-09T20:51:43.803 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout 2026-03-09T20:51:43.803 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout [osd] 2026-03-09T20:51:43.804 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout osd_map_max_advance = 10 2026-03-09T20:51:43.804 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout osd_sloppy_crc = true 2026-03-09T20:51:43.804 INFO:teuthology.orchestra.run.vm01.stdout:Generating new minimal ceph.conf... 2026-03-09T20:51:44.305 INFO:teuthology.orchestra.run.vm01.stdout:Restarting the monitor... 2026-03-09T20:51:44.543 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:51:44 vm01 ceph-mon[53013]: mon.a is new leader, mons a in quorum (ranks 0) 2026-03-09T20:51:44.543 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:51:44 vm01 ceph-mon[53013]: monmap epoch 1 2026-03-09T20:51:44.544 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:51:44 vm01 ceph-mon[53013]: fsid 9cb345a8-1bf9-11f1-a195-0375563c5891 2026-03-09T20:51:44.544 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:51:44 vm01 ceph-mon[53013]: last_changed 2026-03-09T20:51:41.312920+0000 2026-03-09T20:51:44.544 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:51:44 vm01 ceph-mon[53013]: created 2026-03-09T20:51:41.312920+0000 2026-03-09T20:51:44.544 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:51:44 vm01 ceph-mon[53013]: min_mon_release 19 (squid) 2026-03-09T20:51:44.544 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:51:44 vm01 ceph-mon[53013]: election_strategy: 1 2026-03-09T20:51:44.544 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:51:44 vm01 ceph-mon[53013]: 0: [v2:192.168.123.101:3300/0,v1:192.168.123.101:6789/0] mon.a 2026-03-09T20:51:44.544 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:51:44 vm01 ceph-mon[53013]: fsmap 2026-03-09T20:51:44.544 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:51:44 vm01 ceph-mon[53013]: osdmap e1: 0 total, 0 up, 0 in 2026-03-09T20:51:44.544 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:51:44 vm01 ceph-mon[53013]: mgrmap e1: no daemons active 2026-03-09T20:51:44.544 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:51:44 vm01 ceph-mon[53013]: from='client.? 192.168.123.101:0/2896904380' entity='client.admin' cmd=[{"prefix": "status"}]: dispatch 2026-03-09T20:51:44.544 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:51:44 vm01 ceph-mon[53013]: from='client.? 192.168.123.101:0/1514369363' entity='client.admin' cmd=[{"prefix": "config assimilate-conf"}]: dispatch 2026-03-09T20:51:44.544 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:51:44 vm01 ceph-mon[53013]: from='client.? 192.168.123.101:0/1514369363' entity='client.admin' cmd='[{"prefix": "config assimilate-conf"}]': finished 2026-03-09T20:51:44.544 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:51:44 vm01 ceph-mon[53013]: from='client.? 192.168.123.101:0/100484673' entity='client.admin' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T20:51:44.544 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:51:44 vm01 systemd[1]: Stopping Ceph mon.a for 9cb345a8-1bf9-11f1-a195-0375563c5891... 2026-03-09T20:51:44.544 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:51:44 vm01 ceph-9cb345a8-1bf9-11f1-a195-0375563c5891-mon-a[52990]: 2026-03-09T20:51:44.388+0000 7f8598124640 -1 received signal: Terminated from /run/podman-init -- /usr/bin/ceph-mon -n mon.a -f --setuser ceph --setgroup ceph --default-log-to-file=false --default-log-to-journald=true --default-log-to-stderr=false --default-mon-cluster-log-to-file=false --default-mon-cluster-log-to-journald=true --default-mon-cluster-log-to-stderr=false (PID: 1) UID: 0 2026-03-09T20:51:44.544 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:51:44 vm01 ceph-9cb345a8-1bf9-11f1-a195-0375563c5891-mon-a[52990]: 2026-03-09T20:51:44.388+0000 7f8598124640 -1 mon.a@0(leader) e1 *** Got Signal Terminated *** 2026-03-09T20:51:44.544 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:51:44 vm01 podman[53288]: 2026-03-09 20:51:44.43205537 +0000 UTC m=+0.059720115 container died 6c5028c1b249ac00045fa8b53fb8cf24846a5d1353ec52c30c8b316004d0dcdd (image=quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df, name=ceph-9cb345a8-1bf9-11f1-a195-0375563c5891-mon-a, GANESHA_REPO_BASEURL=https://buildlogs.centos.org/centos/$releasever-stream/storage/$basearch/nfsganesha-5/, org.label-schema.vendor=CentOS, ceph=True, org.label-schema.name=CentOS Stream 9 Base Image, io.buildah.version=1.41.3, OSD_FLAVOR=default, CEPH_REF=squid, org.opencontainers.image.authors=Ceph Release Team , CEPH_SHA1=e911bdebe5c8faa3800735d1568fcdca65db60df, org.label-schema.schema-version=1.0, CEPH_GIT_REPO=https://github.com/ceph/ceph-ci.git, FROM_IMAGE=quay.io/centos/centos:stream9, org.label-schema.build-date=20260223, org.label-schema.license=GPLv2, org.opencontainers.image.documentation=https://docs.ceph.com/) 2026-03-09T20:51:44.824 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:51:44 vm01 podman[53288]: 2026-03-09 20:51:44.548368859 +0000 UTC m=+0.176033604 container remove 6c5028c1b249ac00045fa8b53fb8cf24846a5d1353ec52c30c8b316004d0dcdd (image=quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df, name=ceph-9cb345a8-1bf9-11f1-a195-0375563c5891-mon-a, org.opencontainers.image.authors=Ceph Release Team , CEPH_GIT_REPO=https://github.com/ceph/ceph-ci.git, CEPH_SHA1=e911bdebe5c8faa3800735d1568fcdca65db60df, GANESHA_REPO_BASEURL=https://buildlogs.centos.org/centos/$releasever-stream/storage/$basearch/nfsganesha-5/, org.opencontainers.image.documentation=https://docs.ceph.com/, OSD_FLAVOR=default, ceph=True, org.label-schema.build-date=20260223, FROM_IMAGE=quay.io/centos/centos:stream9, org.label-schema.schema-version=1.0, io.buildah.version=1.41.3, CEPH_REF=squid, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.license=GPLv2, org.label-schema.vendor=CentOS) 2026-03-09T20:51:44.824 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:51:44 vm01 bash[53288]: ceph-9cb345a8-1bf9-11f1-a195-0375563c5891-mon-a 2026-03-09T20:51:44.824 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:51:44 vm01 systemd[1]: ceph-9cb345a8-1bf9-11f1-a195-0375563c5891@mon.a.service: Deactivated successfully. 2026-03-09T20:51:44.824 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:51:44 vm01 systemd[1]: Stopped Ceph mon.a for 9cb345a8-1bf9-11f1-a195-0375563c5891. 2026-03-09T20:51:44.824 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:51:44 vm01 systemd[1]: Starting Ceph mon.a for 9cb345a8-1bf9-11f1-a195-0375563c5891... 2026-03-09T20:51:44.824 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:51:44 vm01 podman[53368]: 2026-03-09 20:51:44.731852602 +0000 UTC m=+0.019234724 container create 6532f28f79f17d6052d58fbcbb6916057ce4c8090ab12f74503d522d2fc4c18d (image=quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df, name=ceph-9cb345a8-1bf9-11f1-a195-0375563c5891-mon-a, CEPH_SHA1=e911bdebe5c8faa3800735d1568fcdca65db60df, org.label-schema.license=GPLv2, io.buildah.version=1.41.3, org.opencontainers.image.documentation=https://docs.ceph.com/, FROM_IMAGE=quay.io/centos/centos:stream9, org.label-schema.schema-version=1.0, OSD_FLAVOR=default, ceph=True, org.label-schema.name=CentOS Stream 9 Base Image, CEPH_GIT_REPO=https://github.com/ceph/ceph-ci.git, org.opencontainers.image.authors=Ceph Release Team , org.label-schema.build-date=20260223, GANESHA_REPO_BASEURL=https://buildlogs.centos.org/centos/$releasever-stream/storage/$basearch/nfsganesha-5/, org.label-schema.vendor=CentOS, CEPH_REF=squid) 2026-03-09T20:51:44.986 INFO:teuthology.orchestra.run.vm01.stdout:Setting public_network to 192.168.123.0/24 in mon config section 2026-03-09T20:51:45.079 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:51:44 vm01 podman[53368]: 2026-03-09 20:51:44.72414493 +0000 UTC m=+0.011527052 image pull 654f31e6858eb235bbece362255b685a945f2b6a367e2b88c4930c984fbb214c quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df 2026-03-09T20:51:45.079 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:51:44 vm01 podman[53368]: 2026-03-09 20:51:44.858612516 +0000 UTC m=+0.145994638 container init 6532f28f79f17d6052d58fbcbb6916057ce4c8090ab12f74503d522d2fc4c18d (image=quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df, name=ceph-9cb345a8-1bf9-11f1-a195-0375563c5891-mon-a, org.label-schema.vendor=CentOS, OSD_FLAVOR=default, io.buildah.version=1.41.3, ceph=True, GANESHA_REPO_BASEURL=https://buildlogs.centos.org/centos/$releasever-stream/storage/$basearch/nfsganesha-5/, CEPH_GIT_REPO=https://github.com/ceph/ceph-ci.git, CEPH_REF=squid, org.label-schema.build-date=20260223, org.label-schema.license=GPLv2, org.opencontainers.image.documentation=https://docs.ceph.com/, CEPH_SHA1=e911bdebe5c8faa3800735d1568fcdca65db60df, org.opencontainers.image.authors=Ceph Release Team , FROM_IMAGE=quay.io/centos/centos:stream9, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.schema-version=1.0) 2026-03-09T20:51:45.079 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:51:44 vm01 podman[53368]: 2026-03-09 20:51:44.862263467 +0000 UTC m=+0.149645589 container start 6532f28f79f17d6052d58fbcbb6916057ce4c8090ab12f74503d522d2fc4c18d (image=quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df, name=ceph-9cb345a8-1bf9-11f1-a195-0375563c5891-mon-a, OSD_FLAVOR=default, org.label-schema.license=GPLv2, CEPH_GIT_REPO=https://github.com/ceph/ceph-ci.git, io.buildah.version=1.41.3, org.opencontainers.image.documentation=https://docs.ceph.com/, ceph=True, GANESHA_REPO_BASEURL=https://buildlogs.centos.org/centos/$releasever-stream/storage/$basearch/nfsganesha-5/, FROM_IMAGE=quay.io/centos/centos:stream9, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.schema-version=1.0, org.label-schema.vendor=CentOS, org.label-schema.build-date=20260223, CEPH_REF=squid, org.opencontainers.image.authors=Ceph Release Team , CEPH_SHA1=e911bdebe5c8faa3800735d1568fcdca65db60df) 2026-03-09T20:51:45.079 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:51:44 vm01 ceph-mon[53402]: set uid:gid to 167:167 (ceph:ceph) 2026-03-09T20:51:45.079 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:51:44 vm01 ceph-mon[53402]: ceph version 19.2.3-678-ge911bdeb (e911bdebe5c8faa3800735d1568fcdca65db60df) squid (stable), process ceph-mon, pid 6 2026-03-09T20:51:45.079 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:51:44 vm01 ceph-mon[53402]: pidfile_write: ignore empty --pid-file 2026-03-09T20:51:45.079 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:51:44 vm01 ceph-mon[53402]: load: jerasure load: lrc 2026-03-09T20:51:45.079 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:51:44 vm01 ceph-mon[53402]: rocksdb: RocksDB version: 7.9.2 2026-03-09T20:51:45.079 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:51:44 vm01 ceph-mon[53402]: rocksdb: Git sha 0 2026-03-09T20:51:45.079 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:51:44 vm01 ceph-mon[53402]: rocksdb: Compile date 2026-02-25 18:11:04 2026-03-09T20:51:45.079 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:51:44 vm01 ceph-mon[53402]: rocksdb: DB SUMMARY 2026-03-09T20:51:45.079 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:51:44 vm01 ceph-mon[53402]: rocksdb: DB Session ID: WI04U55T9ZGTYHQFHFTJ 2026-03-09T20:51:45.079 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:51:44 vm01 ceph-mon[53402]: rocksdb: CURRENT file: CURRENT 2026-03-09T20:51:45.079 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:51:44 vm01 ceph-mon[53402]: rocksdb: IDENTITY file: IDENTITY 2026-03-09T20:51:45.079 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:51:44 vm01 ceph-mon[53402]: rocksdb: MANIFEST file: MANIFEST-000010 size: 179 Bytes 2026-03-09T20:51:45.079 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:51:44 vm01 ceph-mon[53402]: rocksdb: SST files in /var/lib/ceph/mon/ceph-a/store.db dir, Total Num: 1, files: 000008.sst 2026-03-09T20:51:45.080 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:51:44 vm01 ceph-mon[53402]: rocksdb: Write Ahead Log file in /var/lib/ceph/mon/ceph-a/store.db: 000009.log size: 88017 ; 2026-03-09T20:51:45.080 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:51:44 vm01 ceph-mon[53402]: rocksdb: Options.error_if_exists: 0 2026-03-09T20:51:45.080 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:51:44 vm01 ceph-mon[53402]: rocksdb: Options.create_if_missing: 0 2026-03-09T20:51:45.080 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:51:44 vm01 ceph-mon[53402]: rocksdb: Options.paranoid_checks: 1 2026-03-09T20:51:45.080 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:51:44 vm01 ceph-mon[53402]: rocksdb: Options.flush_verify_memtable_count: 1 2026-03-09T20:51:45.080 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:51:44 vm01 ceph-mon[53402]: rocksdb: Options.track_and_verify_wals_in_manifest: 0 2026-03-09T20:51:45.080 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:51:44 vm01 ceph-mon[53402]: rocksdb: Options.verify_sst_unique_id_in_manifest: 1 2026-03-09T20:51:45.080 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:51:44 vm01 ceph-mon[53402]: rocksdb: Options.env: 0x55b969a67dc0 2026-03-09T20:51:45.080 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:51:44 vm01 ceph-mon[53402]: rocksdb: Options.fs: PosixFileSystem 2026-03-09T20:51:45.080 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:51:44 vm01 ceph-mon[53402]: rocksdb: Options.info_log: 0x55b96b2e5880 2026-03-09T20:51:45.080 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:51:44 vm01 ceph-mon[53402]: rocksdb: Options.max_file_opening_threads: 16 2026-03-09T20:51:45.080 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:51:44 vm01 ceph-mon[53402]: rocksdb: Options.statistics: (nil) 2026-03-09T20:51:45.080 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:51:44 vm01 ceph-mon[53402]: rocksdb: Options.use_fsync: 0 2026-03-09T20:51:45.080 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:51:44 vm01 ceph-mon[53402]: rocksdb: Options.max_log_file_size: 0 2026-03-09T20:51:45.080 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:51:44 vm01 ceph-mon[53402]: rocksdb: Options.max_manifest_file_size: 1073741824 2026-03-09T20:51:45.080 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:51:44 vm01 ceph-mon[53402]: rocksdb: Options.log_file_time_to_roll: 0 2026-03-09T20:51:45.080 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:51:44 vm01 ceph-mon[53402]: rocksdb: Options.keep_log_file_num: 1000 2026-03-09T20:51:45.080 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:51:44 vm01 ceph-mon[53402]: rocksdb: Options.recycle_log_file_num: 0 2026-03-09T20:51:45.080 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:51:44 vm01 ceph-mon[53402]: rocksdb: Options.allow_fallocate: 1 2026-03-09T20:51:45.080 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:51:44 vm01 ceph-mon[53402]: rocksdb: Options.allow_mmap_reads: 0 2026-03-09T20:51:45.080 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:51:44 vm01 ceph-mon[53402]: rocksdb: Options.allow_mmap_writes: 0 2026-03-09T20:51:45.080 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:51:44 vm01 ceph-mon[53402]: rocksdb: Options.use_direct_reads: 0 2026-03-09T20:51:45.080 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:51:44 vm01 ceph-mon[53402]: rocksdb: Options.use_direct_io_for_flush_and_compaction: 0 2026-03-09T20:51:45.080 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:51:44 vm01 ceph-mon[53402]: rocksdb: Options.create_missing_column_families: 0 2026-03-09T20:51:45.080 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:51:44 vm01 ceph-mon[53402]: rocksdb: Options.db_log_dir: 2026-03-09T20:51:45.080 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:51:44 vm01 ceph-mon[53402]: rocksdb: Options.wal_dir: 2026-03-09T20:51:45.080 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:51:44 vm01 ceph-mon[53402]: rocksdb: Options.table_cache_numshardbits: 6 2026-03-09T20:51:45.080 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:51:44 vm01 ceph-mon[53402]: rocksdb: Options.WAL_ttl_seconds: 0 2026-03-09T20:51:45.080 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:51:44 vm01 ceph-mon[53402]: rocksdb: Options.WAL_size_limit_MB: 0 2026-03-09T20:51:45.080 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:51:44 vm01 ceph-mon[53402]: rocksdb: Options.max_write_batch_group_size_bytes: 1048576 2026-03-09T20:51:45.080 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:51:44 vm01 ceph-mon[53402]: rocksdb: Options.manifest_preallocation_size: 4194304 2026-03-09T20:51:45.080 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:51:44 vm01 ceph-mon[53402]: rocksdb: Options.is_fd_close_on_exec: 1 2026-03-09T20:51:45.080 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:51:44 vm01 ceph-mon[53402]: rocksdb: Options.advise_random_on_open: 1 2026-03-09T20:51:45.080 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:51:44 vm01 ceph-mon[53402]: rocksdb: Options.db_write_buffer_size: 0 2026-03-09T20:51:45.080 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:51:44 vm01 ceph-mon[53402]: rocksdb: Options.write_buffer_manager: 0x55b96b2e9900 2026-03-09T20:51:45.080 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:51:44 vm01 ceph-mon[53402]: rocksdb: Options.access_hint_on_compaction_start: 1 2026-03-09T20:51:45.080 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:51:44 vm01 ceph-mon[53402]: rocksdb: Options.random_access_max_buffer_size: 1048576 2026-03-09T20:51:45.080 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:51:44 vm01 ceph-mon[53402]: rocksdb: Options.use_adaptive_mutex: 0 2026-03-09T20:51:45.080 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:51:44 vm01 ceph-mon[53402]: rocksdb: Options.rate_limiter: (nil) 2026-03-09T20:51:45.080 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:51:44 vm01 ceph-mon[53402]: rocksdb: Options.sst_file_manager.rate_bytes_per_sec: 0 2026-03-09T20:51:45.080 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:51:44 vm01 ceph-mon[53402]: rocksdb: Options.wal_recovery_mode: 2 2026-03-09T20:51:45.080 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:51:44 vm01 ceph-mon[53402]: rocksdb: Options.enable_thread_tracking: 0 2026-03-09T20:51:45.080 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:51:44 vm01 ceph-mon[53402]: rocksdb: Options.enable_pipelined_write: 0 2026-03-09T20:51:45.080 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:51:44 vm01 ceph-mon[53402]: rocksdb: Options.unordered_write: 0 2026-03-09T20:51:45.080 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:51:44 vm01 ceph-mon[53402]: rocksdb: Options.allow_concurrent_memtable_write: 1 2026-03-09T20:51:45.080 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:51:44 vm01 ceph-mon[53402]: rocksdb: Options.enable_write_thread_adaptive_yield: 1 2026-03-09T20:51:45.080 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:51:44 vm01 ceph-mon[53402]: rocksdb: Options.write_thread_max_yield_usec: 100 2026-03-09T20:51:45.080 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:51:44 vm01 ceph-mon[53402]: rocksdb: Options.write_thread_slow_yield_usec: 3 2026-03-09T20:51:45.080 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:51:44 vm01 ceph-mon[53402]: rocksdb: Options.row_cache: None 2026-03-09T20:51:45.080 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:51:44 vm01 ceph-mon[53402]: rocksdb: Options.wal_filter: None 2026-03-09T20:51:45.080 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:51:44 vm01 ceph-mon[53402]: rocksdb: Options.avoid_flush_during_recovery: 0 2026-03-09T20:51:45.080 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:51:44 vm01 ceph-mon[53402]: rocksdb: Options.allow_ingest_behind: 0 2026-03-09T20:51:45.080 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:51:44 vm01 ceph-mon[53402]: rocksdb: Options.two_write_queues: 0 2026-03-09T20:51:45.080 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:51:44 vm01 ceph-mon[53402]: rocksdb: Options.manual_wal_flush: 0 2026-03-09T20:51:45.080 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:51:44 vm01 ceph-mon[53402]: rocksdb: Options.wal_compression: 0 2026-03-09T20:51:45.080 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:51:44 vm01 ceph-mon[53402]: rocksdb: Options.atomic_flush: 0 2026-03-09T20:51:45.080 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:51:44 vm01 ceph-mon[53402]: rocksdb: Options.avoid_unnecessary_blocking_io: 0 2026-03-09T20:51:45.080 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:51:44 vm01 ceph-mon[53402]: rocksdb: Options.persist_stats_to_disk: 0 2026-03-09T20:51:45.080 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:51:44 vm01 ceph-mon[53402]: rocksdb: Options.write_dbid_to_manifest: 0 2026-03-09T20:51:45.081 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:51:44 vm01 ceph-mon[53402]: rocksdb: Options.log_readahead_size: 0 2026-03-09T20:51:45.081 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:51:44 vm01 ceph-mon[53402]: rocksdb: Options.file_checksum_gen_factory: Unknown 2026-03-09T20:51:45.081 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:51:44 vm01 ceph-mon[53402]: rocksdb: Options.best_efforts_recovery: 0 2026-03-09T20:51:45.081 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:51:44 vm01 ceph-mon[53402]: rocksdb: Options.max_bgerror_resume_count: 2147483647 2026-03-09T20:51:45.081 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:51:44 vm01 ceph-mon[53402]: rocksdb: Options.bgerror_resume_retry_interval: 1000000 2026-03-09T20:51:45.081 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:51:44 vm01 ceph-mon[53402]: rocksdb: Options.allow_data_in_errors: 0 2026-03-09T20:51:45.081 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:51:44 vm01 ceph-mon[53402]: rocksdb: Options.db_host_id: __hostname__ 2026-03-09T20:51:45.081 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:51:44 vm01 ceph-mon[53402]: rocksdb: Options.enforce_single_del_contracts: true 2026-03-09T20:51:45.081 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:51:44 vm01 ceph-mon[53402]: rocksdb: Options.max_background_jobs: 2 2026-03-09T20:51:45.081 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:51:44 vm01 ceph-mon[53402]: rocksdb: Options.max_background_compactions: -1 2026-03-09T20:51:45.081 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:51:44 vm01 ceph-mon[53402]: rocksdb: Options.max_subcompactions: 1 2026-03-09T20:51:45.081 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:51:44 vm01 ceph-mon[53402]: rocksdb: Options.avoid_flush_during_shutdown: 0 2026-03-09T20:51:45.081 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:51:44 vm01 ceph-mon[53402]: rocksdb: Options.writable_file_max_buffer_size: 1048576 2026-03-09T20:51:45.081 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:51:44 vm01 ceph-mon[53402]: rocksdb: Options.delayed_write_rate : 16777216 2026-03-09T20:51:45.081 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:51:44 vm01 ceph-mon[53402]: rocksdb: Options.max_total_wal_size: 0 2026-03-09T20:51:45.081 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:51:44 vm01 ceph-mon[53402]: rocksdb: Options.delete_obsolete_files_period_micros: 21600000000 2026-03-09T20:51:45.081 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:51:44 vm01 ceph-mon[53402]: rocksdb: Options.stats_dump_period_sec: 600 2026-03-09T20:51:45.081 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:51:44 vm01 ceph-mon[53402]: rocksdb: Options.stats_persist_period_sec: 600 2026-03-09T20:51:45.081 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:51:44 vm01 ceph-mon[53402]: rocksdb: Options.stats_history_buffer_size: 1048576 2026-03-09T20:51:45.081 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:51:44 vm01 ceph-mon[53402]: rocksdb: Options.max_open_files: -1 2026-03-09T20:51:45.081 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:51:44 vm01 ceph-mon[53402]: rocksdb: Options.bytes_per_sync: 0 2026-03-09T20:51:45.081 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:51:44 vm01 ceph-mon[53402]: rocksdb: Options.wal_bytes_per_sync: 0 2026-03-09T20:51:45.081 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:51:44 vm01 ceph-mon[53402]: rocksdb: Options.strict_bytes_per_sync: 0 2026-03-09T20:51:45.081 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:51:44 vm01 ceph-mon[53402]: rocksdb: Options.compaction_readahead_size: 0 2026-03-09T20:51:45.081 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:51:44 vm01 ceph-mon[53402]: rocksdb: Options.max_background_flushes: -1 2026-03-09T20:51:45.081 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:51:44 vm01 ceph-mon[53402]: rocksdb: Compression algorithms supported: 2026-03-09T20:51:45.081 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:51:44 vm01 ceph-mon[53402]: rocksdb: kZSTD supported: 0 2026-03-09T20:51:45.081 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:51:44 vm01 ceph-mon[53402]: rocksdb: kXpressCompression supported: 0 2026-03-09T20:51:45.081 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:51:44 vm01 ceph-mon[53402]: rocksdb: kBZip2Compression supported: 0 2026-03-09T20:51:45.081 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:51:44 vm01 ceph-mon[53402]: rocksdb: kZSTDNotFinalCompression supported: 0 2026-03-09T20:51:45.081 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:51:44 vm01 ceph-mon[53402]: rocksdb: kLZ4Compression supported: 1 2026-03-09T20:51:45.081 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:51:44 vm01 ceph-mon[53402]: rocksdb: kZlibCompression supported: 1 2026-03-09T20:51:45.081 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:51:44 vm01 ceph-mon[53402]: rocksdb: kLZ4HCCompression supported: 1 2026-03-09T20:51:45.081 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:51:44 vm01 ceph-mon[53402]: rocksdb: kSnappyCompression supported: 1 2026-03-09T20:51:45.081 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:51:44 vm01 ceph-mon[53402]: rocksdb: Fast CRC32 supported: Supported on x86 2026-03-09T20:51:45.081 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:51:44 vm01 ceph-mon[53402]: rocksdb: DMutex implementation: pthread_mutex_t 2026-03-09T20:51:45.081 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:51:44 vm01 ceph-mon[53402]: rocksdb: [db/version_set.cc:5527] Recovering from manifest file: /var/lib/ceph/mon/ceph-a/store.db/MANIFEST-000010 2026-03-09T20:51:45.081 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:51:44 vm01 ceph-mon[53402]: rocksdb: [db/column_family.cc:630] --------------- Options for column family [default]: 2026-03-09T20:51:45.081 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:51:44 vm01 ceph-mon[53402]: rocksdb: Options.comparator: leveldb.BytewiseComparator 2026-03-09T20:51:45.081 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:51:44 vm01 ceph-mon[53402]: rocksdb: Options.merge_operator: 2026-03-09T20:51:45.081 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:51:44 vm01 ceph-mon[53402]: rocksdb: Options.compaction_filter: None 2026-03-09T20:51:45.081 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:51:44 vm01 ceph-mon[53402]: rocksdb: Options.compaction_filter_factory: None 2026-03-09T20:51:45.081 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:51:44 vm01 ceph-mon[53402]: rocksdb: Options.sst_partitioner_factory: None 2026-03-09T20:51:45.081 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:51:44 vm01 ceph-mon[53402]: rocksdb: Options.memtable_factory: SkipListFactory 2026-03-09T20:51:45.081 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:51:44 vm01 ceph-mon[53402]: rocksdb: Options.table_factory: BlockBasedTable 2026-03-09T20:51:45.081 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:51:44 vm01 ceph-mon[53402]: rocksdb: table_factory options: flush_block_policy_factory: FlushBlockBySizePolicyFactory (0x55b96b2e4480) 2026-03-09T20:51:45.081 INFO:journalctl@ceph.mon.a.vm01.stdout: cache_index_and_filter_blocks: 1 2026-03-09T20:51:45.081 INFO:journalctl@ceph.mon.a.vm01.stdout: cache_index_and_filter_blocks_with_high_priority: 0 2026-03-09T20:51:45.081 INFO:journalctl@ceph.mon.a.vm01.stdout: pin_l0_filter_and_index_blocks_in_cache: 0 2026-03-09T20:51:45.081 INFO:journalctl@ceph.mon.a.vm01.stdout: pin_top_level_index_and_filter: 1 2026-03-09T20:51:45.081 INFO:journalctl@ceph.mon.a.vm01.stdout: index_type: 0 2026-03-09T20:51:45.081 INFO:journalctl@ceph.mon.a.vm01.stdout: data_block_index_type: 0 2026-03-09T20:51:45.081 INFO:journalctl@ceph.mon.a.vm01.stdout: index_shortening: 1 2026-03-09T20:51:45.081 INFO:journalctl@ceph.mon.a.vm01.stdout: data_block_hash_table_util_ratio: 0.750000 2026-03-09T20:51:45.081 INFO:journalctl@ceph.mon.a.vm01.stdout: checksum: 4 2026-03-09T20:51:45.081 INFO:journalctl@ceph.mon.a.vm01.stdout: no_block_cache: 0 2026-03-09T20:51:45.081 INFO:journalctl@ceph.mon.a.vm01.stdout: block_cache: 0x55b96b309350 2026-03-09T20:51:45.081 INFO:journalctl@ceph.mon.a.vm01.stdout: block_cache_name: BinnedLRUCache 2026-03-09T20:51:45.081 INFO:journalctl@ceph.mon.a.vm01.stdout: block_cache_options: 2026-03-09T20:51:45.081 INFO:journalctl@ceph.mon.a.vm01.stdout: capacity : 536870912 2026-03-09T20:51:45.081 INFO:journalctl@ceph.mon.a.vm01.stdout: num_shard_bits : 4 2026-03-09T20:51:45.081 INFO:journalctl@ceph.mon.a.vm01.stdout: strict_capacity_limit : 0 2026-03-09T20:51:45.082 INFO:journalctl@ceph.mon.a.vm01.stdout: high_pri_pool_ratio: 0.000 2026-03-09T20:51:45.082 INFO:journalctl@ceph.mon.a.vm01.stdout: block_cache_compressed: (nil) 2026-03-09T20:51:45.082 INFO:journalctl@ceph.mon.a.vm01.stdout: persistent_cache: (nil) 2026-03-09T20:51:45.082 INFO:journalctl@ceph.mon.a.vm01.stdout: block_size: 4096 2026-03-09T20:51:45.082 INFO:journalctl@ceph.mon.a.vm01.stdout: block_size_deviation: 10 2026-03-09T20:51:45.082 INFO:journalctl@ceph.mon.a.vm01.stdout: block_restart_interval: 16 2026-03-09T20:51:45.082 INFO:journalctl@ceph.mon.a.vm01.stdout: index_block_restart_interval: 1 2026-03-09T20:51:45.082 INFO:journalctl@ceph.mon.a.vm01.stdout: metadata_block_size: 4096 2026-03-09T20:51:45.082 INFO:journalctl@ceph.mon.a.vm01.stdout: partition_filters: 0 2026-03-09T20:51:45.082 INFO:journalctl@ceph.mon.a.vm01.stdout: use_delta_encoding: 1 2026-03-09T20:51:45.082 INFO:journalctl@ceph.mon.a.vm01.stdout: filter_policy: bloomfilter 2026-03-09T20:51:45.082 INFO:journalctl@ceph.mon.a.vm01.stdout: whole_key_filtering: 1 2026-03-09T20:51:45.082 INFO:journalctl@ceph.mon.a.vm01.stdout: verify_compression: 0 2026-03-09T20:51:45.082 INFO:journalctl@ceph.mon.a.vm01.stdout: read_amp_bytes_per_bit: 0 2026-03-09T20:51:45.082 INFO:journalctl@ceph.mon.a.vm01.stdout: format_version: 5 2026-03-09T20:51:45.082 INFO:journalctl@ceph.mon.a.vm01.stdout: enable_index_compression: 1 2026-03-09T20:51:45.082 INFO:journalctl@ceph.mon.a.vm01.stdout: block_align: 0 2026-03-09T20:51:45.082 INFO:journalctl@ceph.mon.a.vm01.stdout: max_auto_readahead_size: 262144 2026-03-09T20:51:45.082 INFO:journalctl@ceph.mon.a.vm01.stdout: prepopulate_block_cache: 0 2026-03-09T20:51:45.082 INFO:journalctl@ceph.mon.a.vm01.stdout: initial_auto_readahead_size: 8192 2026-03-09T20:51:45.082 INFO:journalctl@ceph.mon.a.vm01.stdout: num_file_reads_for_auto_readahead: 2 2026-03-09T20:51:45.082 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:51:44 vm01 ceph-mon[53402]: rocksdb: Options.write_buffer_size: 33554432 2026-03-09T20:51:45.082 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:51:44 vm01 ceph-mon[53402]: rocksdb: Options.max_write_buffer_number: 2 2026-03-09T20:51:45.082 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:51:44 vm01 ceph-mon[53402]: rocksdb: Options.compression: NoCompression 2026-03-09T20:51:45.082 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:51:44 vm01 ceph-mon[53402]: rocksdb: Options.bottommost_compression: Disabled 2026-03-09T20:51:45.082 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:51:44 vm01 ceph-mon[53402]: rocksdb: Options.prefix_extractor: nullptr 2026-03-09T20:51:45.082 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:51:44 vm01 ceph-mon[53402]: rocksdb: Options.memtable_insert_with_hint_prefix_extractor: nullptr 2026-03-09T20:51:45.082 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:51:44 vm01 ceph-mon[53402]: rocksdb: Options.num_levels: 7 2026-03-09T20:51:45.082 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:51:44 vm01 ceph-mon[53402]: rocksdb: Options.min_write_buffer_number_to_merge: 1 2026-03-09T20:51:45.082 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:51:44 vm01 ceph-mon[53402]: rocksdb: Options.max_write_buffer_number_to_maintain: 0 2026-03-09T20:51:45.082 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:51:44 vm01 ceph-mon[53402]: rocksdb: Options.max_write_buffer_size_to_maintain: 0 2026-03-09T20:51:45.082 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:51:44 vm01 ceph-mon[53402]: rocksdb: Options.bottommost_compression_opts.window_bits: -14 2026-03-09T20:51:45.082 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:51:44 vm01 ceph-mon[53402]: rocksdb: Options.bottommost_compression_opts.level: 32767 2026-03-09T20:51:45.082 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:51:44 vm01 ceph-mon[53402]: rocksdb: Options.bottommost_compression_opts.strategy: 0 2026-03-09T20:51:45.082 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:51:44 vm01 ceph-mon[53402]: rocksdb: Options.bottommost_compression_opts.max_dict_bytes: 0 2026-03-09T20:51:45.082 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:51:44 vm01 ceph-mon[53402]: rocksdb: Options.bottommost_compression_opts.zstd_max_train_bytes: 0 2026-03-09T20:51:45.082 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:51:44 vm01 ceph-mon[53402]: rocksdb: Options.bottommost_compression_opts.parallel_threads: 1 2026-03-09T20:51:45.082 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:51:44 vm01 ceph-mon[53402]: rocksdb: Options.bottommost_compression_opts.enabled: false 2026-03-09T20:51:45.082 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:51:44 vm01 ceph-mon[53402]: rocksdb: Options.bottommost_compression_opts.max_dict_buffer_bytes: 0 2026-03-09T20:51:45.082 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:51:44 vm01 ceph-mon[53402]: rocksdb: Options.bottommost_compression_opts.use_zstd_dict_trainer: true 2026-03-09T20:51:45.082 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:51:44 vm01 ceph-mon[53402]: rocksdb: Options.compression_opts.window_bits: -14 2026-03-09T20:51:45.082 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:51:44 vm01 ceph-mon[53402]: rocksdb: Options.compression_opts.level: 32767 2026-03-09T20:51:45.083 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:51:44 vm01 ceph-mon[53402]: rocksdb: Options.compression_opts.strategy: 0 2026-03-09T20:51:45.083 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:51:44 vm01 ceph-mon[53402]: rocksdb: Options.compression_opts.max_dict_bytes: 0 2026-03-09T20:51:45.083 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:51:44 vm01 ceph-mon[53402]: rocksdb: Options.compression_opts.zstd_max_train_bytes: 0 2026-03-09T20:51:45.083 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:51:44 vm01 ceph-mon[53402]: rocksdb: Options.compression_opts.use_zstd_dict_trainer: true 2026-03-09T20:51:45.083 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:51:44 vm01 ceph-mon[53402]: rocksdb: Options.compression_opts.parallel_threads: 1 2026-03-09T20:51:45.083 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:51:44 vm01 ceph-mon[53402]: rocksdb: Options.compression_opts.enabled: false 2026-03-09T20:51:45.083 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:51:44 vm01 ceph-mon[53402]: rocksdb: Options.compression_opts.max_dict_buffer_bytes: 0 2026-03-09T20:51:45.083 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:51:44 vm01 ceph-mon[53402]: rocksdb: Options.level0_file_num_compaction_trigger: 4 2026-03-09T20:51:45.083 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:51:44 vm01 ceph-mon[53402]: rocksdb: Options.level0_slowdown_writes_trigger: 20 2026-03-09T20:51:45.083 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:51:44 vm01 ceph-mon[53402]: rocksdb: Options.level0_stop_writes_trigger: 36 2026-03-09T20:51:45.083 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:51:44 vm01 ceph-mon[53402]: rocksdb: Options.target_file_size_base: 67108864 2026-03-09T20:51:45.083 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:51:44 vm01 ceph-mon[53402]: rocksdb: Options.target_file_size_multiplier: 1 2026-03-09T20:51:45.083 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:51:44 vm01 ceph-mon[53402]: rocksdb: Options.max_bytes_for_level_base: 268435456 2026-03-09T20:51:45.083 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:51:44 vm01 ceph-mon[53402]: rocksdb: Options.level_compaction_dynamic_level_bytes: 1 2026-03-09T20:51:45.083 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:51:44 vm01 ceph-mon[53402]: rocksdb: Options.max_bytes_for_level_multiplier: 10.000000 2026-03-09T20:51:45.083 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:51:44 vm01 ceph-mon[53402]: rocksdb: Options.max_bytes_for_level_multiplier_addtl[0]: 1 2026-03-09T20:51:45.083 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:51:44 vm01 ceph-mon[53402]: rocksdb: Options.max_bytes_for_level_multiplier_addtl[1]: 1 2026-03-09T20:51:45.083 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:51:44 vm01 ceph-mon[53402]: rocksdb: Options.max_bytes_for_level_multiplier_addtl[2]: 1 2026-03-09T20:51:45.083 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:51:44 vm01 ceph-mon[53402]: rocksdb: Options.max_bytes_for_level_multiplier_addtl[3]: 1 2026-03-09T20:51:45.083 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:51:44 vm01 ceph-mon[53402]: rocksdb: Options.max_bytes_for_level_multiplier_addtl[4]: 1 2026-03-09T20:51:45.083 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:51:44 vm01 ceph-mon[53402]: rocksdb: Options.max_bytes_for_level_multiplier_addtl[5]: 1 2026-03-09T20:51:45.083 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:51:44 vm01 ceph-mon[53402]: rocksdb: Options.max_bytes_for_level_multiplier_addtl[6]: 1 2026-03-09T20:51:45.083 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:51:44 vm01 ceph-mon[53402]: rocksdb: Options.max_sequential_skip_in_iterations: 8 2026-03-09T20:51:45.083 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:51:44 vm01 ceph-mon[53402]: rocksdb: Options.max_compaction_bytes: 1677721600 2026-03-09T20:51:45.083 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:51:44 vm01 ceph-mon[53402]: rocksdb: Options.ignore_max_compaction_bytes_for_input: true 2026-03-09T20:51:45.083 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:51:44 vm01 ceph-mon[53402]: rocksdb: Options.arena_block_size: 1048576 2026-03-09T20:51:45.083 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:51:44 vm01 ceph-mon[53402]: rocksdb: Options.soft_pending_compaction_bytes_limit: 68719476736 2026-03-09T20:51:45.083 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:51:44 vm01 ceph-mon[53402]: rocksdb: Options.hard_pending_compaction_bytes_limit: 274877906944 2026-03-09T20:51:45.083 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:51:44 vm01 ceph-mon[53402]: rocksdb: Options.disable_auto_compactions: 0 2026-03-09T20:51:45.083 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:51:44 vm01 ceph-mon[53402]: rocksdb: Options.compaction_style: kCompactionStyleLevel 2026-03-09T20:51:45.083 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:51:44 vm01 ceph-mon[53402]: rocksdb: Options.compaction_pri: kMinOverlappingRatio 2026-03-09T20:51:45.083 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:51:44 vm01 ceph-mon[53402]: rocksdb: Options.compaction_options_universal.size_ratio: 1 2026-03-09T20:51:45.083 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:51:44 vm01 ceph-mon[53402]: rocksdb: Options.compaction_options_universal.min_merge_width: 2 2026-03-09T20:51:45.083 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:51:44 vm01 ceph-mon[53402]: rocksdb: Options.compaction_options_universal.max_merge_width: 4294967295 2026-03-09T20:51:45.083 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:51:44 vm01 ceph-mon[53402]: rocksdb: Options.compaction_options_universal.max_size_amplification_percent: 200 2026-03-09T20:51:45.083 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:51:44 vm01 ceph-mon[53402]: rocksdb: Options.compaction_options_universal.compression_size_percent: -1 2026-03-09T20:51:45.083 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:51:44 vm01 ceph-mon[53402]: rocksdb: Options.compaction_options_universal.stop_style: kCompactionStopStyleTotalSize 2026-03-09T20:51:45.083 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:51:44 vm01 ceph-mon[53402]: rocksdb: Options.compaction_options_fifo.max_table_files_size: 1073741824 2026-03-09T20:51:45.083 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:51:44 vm01 ceph-mon[53402]: rocksdb: Options.compaction_options_fifo.allow_compaction: 0 2026-03-09T20:51:45.083 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:51:44 vm01 ceph-mon[53402]: rocksdb: Options.table_properties_collectors: CompactOnDeletionCollector (Sliding window size = 32768 Deletion trigger = 16384 Deletion ratio = 0); 2026-03-09T20:51:45.083 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:51:44 vm01 ceph-mon[53402]: rocksdb: Options.inplace_update_support: 0 2026-03-09T20:51:45.083 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:51:44 vm01 ceph-mon[53402]: rocksdb: Options.inplace_update_num_locks: 10000 2026-03-09T20:51:45.083 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:51:44 vm01 ceph-mon[53402]: rocksdb: Options.memtable_prefix_bloom_size_ratio: 0.000000 2026-03-09T20:51:45.083 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:51:44 vm01 ceph-mon[53402]: rocksdb: Options.memtable_whole_key_filtering: 0 2026-03-09T20:51:45.083 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:51:44 vm01 ceph-mon[53402]: rocksdb: Options.memtable_huge_page_size: 0 2026-03-09T20:51:45.083 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:51:44 vm01 ceph-mon[53402]: rocksdb: Options.bloom_locality: 0 2026-03-09T20:51:45.083 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:51:44 vm01 ceph-mon[53402]: rocksdb: Options.max_successive_merges: 0 2026-03-09T20:51:45.083 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:51:44 vm01 ceph-mon[53402]: rocksdb: Options.optimize_filters_for_hits: 0 2026-03-09T20:51:45.083 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:51:44 vm01 ceph-mon[53402]: rocksdb: Options.paranoid_file_checks: 0 2026-03-09T20:51:45.083 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:51:44 vm01 ceph-mon[53402]: rocksdb: Options.force_consistency_checks: 1 2026-03-09T20:51:45.083 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:51:44 vm01 ceph-mon[53402]: rocksdb: Options.report_bg_io_stats: 0 2026-03-09T20:51:45.083 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:51:44 vm01 ceph-mon[53402]: rocksdb: Options.ttl: 2592000 2026-03-09T20:51:45.084 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:51:44 vm01 ceph-mon[53402]: rocksdb: Options.periodic_compaction_seconds: 0 2026-03-09T20:51:45.084 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:51:44 vm01 ceph-mon[53402]: rocksdb: Options.preclude_last_level_data_seconds: 0 2026-03-09T20:51:45.084 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:51:44 vm01 ceph-mon[53402]: rocksdb: Options.preserve_internal_time_seconds: 0 2026-03-09T20:51:45.084 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:51:44 vm01 ceph-mon[53402]: rocksdb: Options.enable_blob_files: false 2026-03-09T20:51:45.084 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:51:44 vm01 ceph-mon[53402]: rocksdb: Options.min_blob_size: 0 2026-03-09T20:51:45.084 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:51:44 vm01 ceph-mon[53402]: rocksdb: Options.blob_file_size: 268435456 2026-03-09T20:51:45.084 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:51:44 vm01 ceph-mon[53402]: rocksdb: Options.blob_compression_type: NoCompression 2026-03-09T20:51:45.084 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:51:44 vm01 ceph-mon[53402]: rocksdb: Options.enable_blob_garbage_collection: false 2026-03-09T20:51:45.084 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:51:44 vm01 ceph-mon[53402]: rocksdb: Options.blob_garbage_collection_age_cutoff: 0.250000 2026-03-09T20:51:45.084 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:51:44 vm01 ceph-mon[53402]: rocksdb: Options.blob_garbage_collection_force_threshold: 1.000000 2026-03-09T20:51:45.084 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:51:44 vm01 ceph-mon[53402]: rocksdb: Options.blob_compaction_readahead_size: 0 2026-03-09T20:51:45.084 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:51:44 vm01 ceph-mon[53402]: rocksdb: Options.blob_file_starting_level: 0 2026-03-09T20:51:45.084 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:51:44 vm01 ceph-mon[53402]: rocksdb: Options.experimental_mempurge_threshold: 0.000000 2026-03-09T20:51:45.084 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:51:44 vm01 ceph-mon[53402]: rocksdb: [db/version_set.cc:5566] Recovered from manifest file:/var/lib/ceph/mon/ceph-a/store.db/MANIFEST-000010 succeeded,manifest_file_number is 10, next_file_number is 12, last_sequence is 5, log_number is 5,prev_log_number is 0,max_column_family is 0,min_log_number_to_keep is 5 2026-03-09T20:51:45.084 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:51:44 vm01 ceph-mon[53402]: rocksdb: [db/version_set.cc:5581] Column family [default] (ID 0), log number is 5 2026-03-09T20:51:45.084 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:51:44 vm01 ceph-mon[53402]: rocksdb: [db/db_impl/db_impl_open.cc:539] DB ID: 91566189-23a8-41b2-987e-56821ccfb635 2026-03-09T20:51:45.084 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:51:44 vm01 ceph-mon[53402]: rocksdb: EVENT_LOG_v1 {"time_micros": 1773089504882965, "job": 1, "event": "recovery_started", "wal_files": [9]} 2026-03-09T20:51:45.084 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:51:44 vm01 ceph-mon[53402]: rocksdb: [db/db_impl/db_impl_open.cc:1043] Recovering log #9 mode 2 2026-03-09T20:51:45.084 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:51:44 vm01 ceph-mon[53402]: rocksdb: EVENT_LOG_v1 {"time_micros": 1773089504974978, "cf_name": "default", "job": 1, "event": "table_file_creation", "file_number": 13, "file_size": 84995, "file_checksum": "", "file_checksum_func_name": "Unknown", "smallest_seqno": 8, "largest_seqno": 246, "table_properties": {"data_size": 83147, "index_size": 237, "index_partitions": 0, "top_level_index_size": 0, "index_key_is_user_key": 1, "index_value_is_delta_encoded": 1, "filter_size": 581, "raw_key_size": 10161, "raw_average_key_size": 47, "raw_value_size": 77316, "raw_average_value_size": 362, "num_data_blocks": 11, "num_entries": 213, "num_filter_entries": 213, "num_deletions": 3, "num_merge_operands": 0, "num_range_deletions": 0, "format_version": 0, "fixed_key_len": 0, "filter_policy": "bloomfilter", "column_family_name": "default", "column_family_id": 0, "comparator": "leveldb.BytewiseComparator", "merge_operator": "", "prefix_extractor_name": "nullptr", "property_collectors": "[CompactOnDeletionCollector]", "compression": "NoCompression", "compression_options": "window_bits=-14; level=32767; strategy=0; max_dict_bytes=0; zstd_max_train_bytes=0; enabled=0; max_dict_buffer_bytes=0; use_zstd_dict_trainer=1; ", "creation_time": 1773089504, "oldest_key_time": 0, "file_creation_time": 0, "slow_compression_estimated_data_size": 0, "fast_compression_estimated_data_size": 0, "db_id": "91566189-23a8-41b2-987e-56821ccfb635", "db_session_id": "WI04U55T9ZGTYHQFHFTJ", "orig_file_number": 13, "seqno_to_time_mapping": "N/A"}} 2026-03-09T20:51:45.084 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:51:44 vm01 ceph-mon[53402]: rocksdb: EVENT_LOG_v1 {"time_micros": 1773089504975088, "job": 1, "event": "recovery_finished"} 2026-03-09T20:51:45.084 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:51:44 vm01 ceph-mon[53402]: rocksdb: [db/version_set.cc:5047] Creating manifest 15 2026-03-09T20:51:45.084 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:51:44 vm01 bash[53368]: 6532f28f79f17d6052d58fbcbb6916057ce4c8090ab12f74503d522d2fc4c18d 2026-03-09T20:51:45.084 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:51:44 vm01 systemd[1]: Started Ceph mon.a for 9cb345a8-1bf9-11f1-a195-0375563c5891. 2026-03-09T20:51:45.424 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:51:45 vm01 ceph-mon[53402]: rocksdb: [file/delete_scheduler.cc:74] Deleted file /var/lib/ceph/mon/ceph-a/store.db/000009.log immediately, rate_bytes_per_sec 0, total_trash_size 0 max_trash_db_ratio 0.250000 2026-03-09T20:51:45.424 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:51:45 vm01 ceph-mon[53402]: rocksdb: [db/db_impl/db_impl_open.cc:1987] SstFileManager instance 0x55b96b30ae00 2026-03-09T20:51:45.424 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:51:45 vm01 ceph-mon[53402]: rocksdb: DB pointer 0x55b96b414000 2026-03-09T20:51:45.424 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:51:45 vm01 ceph-mon[53402]: starting mon.a rank 0 at public addrs [v2:192.168.123.101:3300/0,v1:192.168.123.101:6789/0] at bind addrs [v2:192.168.123.101:3300/0,v1:192.168.123.101:6789/0] mon_data /var/lib/ceph/mon/ceph-a fsid 9cb345a8-1bf9-11f1-a195-0375563c5891 2026-03-09T20:51:45.424 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:51:45 vm01 ceph-mon[53402]: rocksdb: [db/db_impl/db_impl.cc:1109] ------- DUMPING STATS ------- 2026-03-09T20:51:45.424 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:51:45 vm01 ceph-mon[53402]: rocksdb: [db/db_impl/db_impl.cc:1111] 2026-03-09T20:51:45.424 INFO:journalctl@ceph.mon.a.vm01.stdout: ** DB Stats ** 2026-03-09T20:51:45.424 INFO:journalctl@ceph.mon.a.vm01.stdout: Uptime(secs): 0.2 total, 0.2 interval 2026-03-09T20:51:45.424 INFO:journalctl@ceph.mon.a.vm01.stdout: Cumulative writes: 0 writes, 0 keys, 0 commit groups, 0.0 writes per commit group, ingest: 0.00 GB, 0.00 MB/s 2026-03-09T20:51:45.424 INFO:journalctl@ceph.mon.a.vm01.stdout: Cumulative WAL: 0 writes, 0 syncs, 0.00 writes per sync, written: 0.00 GB, 0.00 MB/s 2026-03-09T20:51:45.424 INFO:journalctl@ceph.mon.a.vm01.stdout: Cumulative stall: 00:00:0.000 H:M:S, 0.0 percent 2026-03-09T20:51:45.424 INFO:journalctl@ceph.mon.a.vm01.stdout: Interval writes: 0 writes, 0 keys, 0 commit groups, 0.0 writes per commit group, ingest: 0.00 MB, 0.00 MB/s 2026-03-09T20:51:45.424 INFO:journalctl@ceph.mon.a.vm01.stdout: Interval WAL: 0 writes, 0 syncs, 0.00 writes per sync, written: 0.00 GB, 0.00 MB/s 2026-03-09T20:51:45.424 INFO:journalctl@ceph.mon.a.vm01.stdout: Interval stall: 00:00:0.000 H:M:S, 0.0 percent 2026-03-09T20:51:45.424 INFO:journalctl@ceph.mon.a.vm01.stdout: 2026-03-09T20:51:45.424 INFO:journalctl@ceph.mon.a.vm01.stdout: ** Compaction Stats [default] ** 2026-03-09T20:51:45.424 INFO:journalctl@ceph.mon.a.vm01.stdout: Level Files Size Score Read(GB) Rn(GB) Rnp1(GB) Write(GB) Wnew(GB) Moved(GB) W-Amp Rd(MB/s) Wr(MB/s) Comp(sec) CompMergeCPU(sec) Comp(cnt) Avg(sec) KeyIn KeyDrop Rblob(GB) Wblob(GB) 2026-03-09T20:51:45.424 INFO:journalctl@ceph.mon.a.vm01.stdout: ------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------ 2026-03-09T20:51:45.424 INFO:journalctl@ceph.mon.a.vm01.stdout: L0 2/0 84.86 KB 0.5 0.0 0.0 0.0 0.0 0.0 0.0 1.0 0.0 0.9 0.09 0.00 1 0.092 0 0 0.0 0.0 2026-03-09T20:51:45.424 INFO:journalctl@ceph.mon.a.vm01.stdout: Sum 2/0 84.86 KB 0.0 0.0 0.0 0.0 0.0 0.0 0.0 1.0 0.0 0.9 0.09 0.00 1 0.092 0 0 0.0 0.0 2026-03-09T20:51:45.424 INFO:journalctl@ceph.mon.a.vm01.stdout: Int 0/0 0.00 KB 0.0 0.0 0.0 0.0 0.0 0.0 0.0 1.0 0.0 0.9 0.09 0.00 1 0.092 0 0 0.0 0.0 2026-03-09T20:51:45.424 INFO:journalctl@ceph.mon.a.vm01.stdout: 2026-03-09T20:51:45.424 INFO:journalctl@ceph.mon.a.vm01.stdout: ** Compaction Stats [default] ** 2026-03-09T20:51:45.424 INFO:journalctl@ceph.mon.a.vm01.stdout: Priority Files Size Score Read(GB) Rn(GB) Rnp1(GB) Write(GB) Wnew(GB) Moved(GB) W-Amp Rd(MB/s) Wr(MB/s) Comp(sec) CompMergeCPU(sec) Comp(cnt) Avg(sec) KeyIn KeyDrop Rblob(GB) Wblob(GB) 2026-03-09T20:51:45.424 INFO:journalctl@ceph.mon.a.vm01.stdout: --------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------- 2026-03-09T20:51:45.424 INFO:journalctl@ceph.mon.a.vm01.stdout: User 0/0 0.00 KB 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0.9 0.09 0.00 1 0.092 0 0 0.0 0.0 2026-03-09T20:51:45.424 INFO:journalctl@ceph.mon.a.vm01.stdout: 2026-03-09T20:51:45.424 INFO:journalctl@ceph.mon.a.vm01.stdout: Blob file count: 0, total size: 0.0 GB, garbage size: 0.0 GB, space amp: 0.0 2026-03-09T20:51:45.424 INFO:journalctl@ceph.mon.a.vm01.stdout: 2026-03-09T20:51:45.424 INFO:journalctl@ceph.mon.a.vm01.stdout: Uptime(secs): 0.2 total, 0.2 interval 2026-03-09T20:51:45.424 INFO:journalctl@ceph.mon.a.vm01.stdout: Flush(GB): cumulative 0.000, interval 0.000 2026-03-09T20:51:45.424 INFO:journalctl@ceph.mon.a.vm01.stdout: AddFile(GB): cumulative 0.000, interval 0.000 2026-03-09T20:51:45.424 INFO:journalctl@ceph.mon.a.vm01.stdout: AddFile(Total Files): cumulative 0, interval 0 2026-03-09T20:51:45.424 INFO:journalctl@ceph.mon.a.vm01.stdout: AddFile(L0 Files): cumulative 0, interval 0 2026-03-09T20:51:45.424 INFO:journalctl@ceph.mon.a.vm01.stdout: AddFile(Keys): cumulative 0, interval 0 2026-03-09T20:51:45.425 INFO:journalctl@ceph.mon.a.vm01.stdout: Cumulative compaction: 0.00 GB write, 0.41 MB/s write, 0.00 GB read, 0.00 MB/s read, 0.1 seconds 2026-03-09T20:51:45.425 INFO:journalctl@ceph.mon.a.vm01.stdout: Interval compaction: 0.00 GB write, 0.41 MB/s write, 0.00 GB read, 0.00 MB/s read, 0.1 seconds 2026-03-09T20:51:45.425 INFO:journalctl@ceph.mon.a.vm01.stdout: Stalls(count): 0 level0_slowdown, 0 level0_slowdown_with_compaction, 0 level0_numfiles, 0 level0_numfiles_with_compaction, 0 stop for pending_compaction_bytes, 0 slowdown for pending_compaction_bytes, 0 memtable_compaction, 0 memtable_slowdown, interval 0 total count 2026-03-09T20:51:45.425 INFO:journalctl@ceph.mon.a.vm01.stdout: Block cache BinnedLRUCache@0x55b96b309350#6 capacity: 512.00 MB usage: 26.30 KB table_size: 0 occupancy: 18446744073709551615 collections: 1 last_copies: 0 last_secs: 1.2e-05 secs_since: 0 2026-03-09T20:51:45.425 INFO:journalctl@ceph.mon.a.vm01.stdout: Block cache entry stats(count,size,portion): DataBlock(3,25.11 KB,0.00478923%) FilterBlock(2,0.77 KB,0.000146031%) IndexBlock(2,0.42 KB,8.04663e-05%) Misc(1,0.00 KB,0%) 2026-03-09T20:51:45.425 INFO:journalctl@ceph.mon.a.vm01.stdout: 2026-03-09T20:51:45.425 INFO:journalctl@ceph.mon.a.vm01.stdout: ** File Read Latency Histogram By Level [default] ** 2026-03-09T20:51:45.425 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:51:45 vm01 ceph-mon[53402]: mon.a@-1(???) e1 preinit fsid 9cb345a8-1bf9-11f1-a195-0375563c5891 2026-03-09T20:51:45.425 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:51:45 vm01 ceph-mon[53402]: mon.a@-1(???).mds e1 new map 2026-03-09T20:51:45.425 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:51:45 vm01 ceph-mon[53402]: mon.a@-1(???).mds e1 print_map 2026-03-09T20:51:45.425 INFO:journalctl@ceph.mon.a.vm01.stdout: e1 2026-03-09T20:51:45.425 INFO:journalctl@ceph.mon.a.vm01.stdout: btime 2026-03-09T20:51:43:189291+0000 2026-03-09T20:51:45.425 INFO:journalctl@ceph.mon.a.vm01.stdout: enable_multiple, ever_enabled_multiple: 1,1 2026-03-09T20:51:45.425 INFO:journalctl@ceph.mon.a.vm01.stdout: default compat: compat={},rocompat={},incompat={1=base v0.20,2=client writeable ranges,3=default file layouts on dirs,4=dir inode in separate object,5=mds uses versioned encoding,6=dirfrag is stored in omap,8=no anchor table,9=file layout v2,10=snaprealm v2,11=minor log segments,12=quiesce subvolumes} 2026-03-09T20:51:45.425 INFO:journalctl@ceph.mon.a.vm01.stdout: legacy client fscid: -1 2026-03-09T20:51:45.425 INFO:journalctl@ceph.mon.a.vm01.stdout: 2026-03-09T20:51:45.425 INFO:journalctl@ceph.mon.a.vm01.stdout: No filesystems configured 2026-03-09T20:51:45.425 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:51:45 vm01 ceph-mon[53402]: mon.a@-1(???).osd e1 crush map has features 3314932999778484224, adjusting msgr requires 2026-03-09T20:51:45.425 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:51:45 vm01 ceph-mon[53402]: mon.a@-1(???).osd e1 crush map has features 288514050185494528, adjusting msgr requires 2026-03-09T20:51:45.425 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:51:45 vm01 ceph-mon[53402]: mon.a@-1(???).osd e1 crush map has features 288514050185494528, adjusting msgr requires 2026-03-09T20:51:45.425 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:51:45 vm01 ceph-mon[53402]: mon.a@-1(???).osd e1 crush map has features 288514050185494528, adjusting msgr requires 2026-03-09T20:51:45.425 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:51:45 vm01 ceph-mon[53402]: mon.a@-1(???).paxosservice(auth 1..2) refresh upgraded, format 0 -> 3 2026-03-09T20:51:45.924 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:51:45 vm01 ceph-mon[53402]: mon.a is new leader, mons a in quorum (ranks 0) 2026-03-09T20:51:45.924 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:51:45 vm01 ceph-mon[53402]: monmap epoch 1 2026-03-09T20:51:45.924 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:51:45 vm01 ceph-mon[53402]: fsid 9cb345a8-1bf9-11f1-a195-0375563c5891 2026-03-09T20:51:45.924 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:51:45 vm01 ceph-mon[53402]: last_changed 2026-03-09T20:51:41.312920+0000 2026-03-09T20:51:45.924 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:51:45 vm01 ceph-mon[53402]: created 2026-03-09T20:51:41.312920+0000 2026-03-09T20:51:45.924 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:51:45 vm01 ceph-mon[53402]: min_mon_release 19 (squid) 2026-03-09T20:51:45.924 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:51:45 vm01 ceph-mon[53402]: election_strategy: 1 2026-03-09T20:51:45.924 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:51:45 vm01 ceph-mon[53402]: 0: [v2:192.168.123.101:3300/0,v1:192.168.123.101:6789/0] mon.a 2026-03-09T20:51:45.924 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:51:45 vm01 ceph-mon[53402]: fsmap 2026-03-09T20:51:45.924 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:51:45 vm01 ceph-mon[53402]: osdmap e1: 0 total, 0 up, 0 in 2026-03-09T20:51:45.924 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:51:45 vm01 ceph-mon[53402]: mgrmap e1: no daemons active 2026-03-09T20:51:46.548 INFO:teuthology.orchestra.run.vm01.stdout:Wrote config to /etc/ceph/ceph.conf 2026-03-09T20:51:46.550 INFO:teuthology.orchestra.run.vm01.stdout:Wrote keyring to /etc/ceph/ceph.client.admin.keyring 2026-03-09T20:51:46.550 INFO:teuthology.orchestra.run.vm01.stdout:Creating mgr... 2026-03-09T20:51:46.550 INFO:teuthology.orchestra.run.vm01.stdout:Verifying port 0.0.0.0:9283 ... 2026-03-09T20:51:46.550 INFO:teuthology.orchestra.run.vm01.stdout:Verifying port 0.0.0.0:8765 ... 2026-03-09T20:51:46.699 INFO:teuthology.orchestra.run.vm01.stdout:Non-zero exit code 1 from systemctl reset-failed ceph-9cb345a8-1bf9-11f1-a195-0375563c5891@mgr.a 2026-03-09T20:51:46.699 INFO:teuthology.orchestra.run.vm01.stdout:systemctl: stderr Failed to reset failed state of unit ceph-9cb345a8-1bf9-11f1-a195-0375563c5891@mgr.a.service: Unit ceph-9cb345a8-1bf9-11f1-a195-0375563c5891@mgr.a.service not loaded. 2026-03-09T20:51:46.826 INFO:teuthology.orchestra.run.vm01.stdout:systemctl: stderr Created symlink /etc/systemd/system/ceph-9cb345a8-1bf9-11f1-a195-0375563c5891.target.wants/ceph-9cb345a8-1bf9-11f1-a195-0375563c5891@mgr.a.service → /etc/systemd/system/ceph-9cb345a8-1bf9-11f1-a195-0375563c5891@.service. 2026-03-09T20:51:46.968 INFO:journalctl@ceph.mgr.a.vm01.stdout:Mar 09 20:51:46 vm01 podman[53628]: 2026-03-09 20:51:46.937405715 +0000 UTC m=+0.018161342 container create 9fc252e619ba3254f69778a8ecbaac0c12aa7a6d41ba3880147c41cb162f5d15 (image=quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df, name=ceph-9cb345a8-1bf9-11f1-a195-0375563c5891-mgr-a, CEPH_GIT_REPO=https://github.com/ceph/ceph-ci.git, org.label-schema.license=GPLv2, io.buildah.version=1.41.3, CEPH_SHA1=e911bdebe5c8faa3800735d1568fcdca65db60df, FROM_IMAGE=quay.io/centos/centos:stream9, org.label-schema.vendor=CentOS, GANESHA_REPO_BASEURL=https://buildlogs.centos.org/centos/$releasever-stream/storage/$basearch/nfsganesha-5/, org.label-schema.build-date=20260223, OSD_FLAVOR=default, org.opencontainers.image.documentation=https://docs.ceph.com/, org.label-schema.schema-version=1.0, CEPH_REF=squid, ceph=True, org.opencontainers.image.authors=Ceph Release Team , org.label-schema.name=CentOS Stream 9 Base Image) 2026-03-09T20:51:46.996 INFO:teuthology.orchestra.run.vm01.stdout:firewalld does not appear to be present 2026-03-09T20:51:46.996 INFO:teuthology.orchestra.run.vm01.stdout:Not possible to enable service . firewalld.service is not available 2026-03-09T20:51:46.996 INFO:teuthology.orchestra.run.vm01.stdout:firewalld does not appear to be present 2026-03-09T20:51:46.996 INFO:teuthology.orchestra.run.vm01.stdout:Not possible to open ports <[9283, 8765]>. firewalld.service is not available 2026-03-09T20:51:46.996 INFO:teuthology.orchestra.run.vm01.stdout:Waiting for mgr to start... 2026-03-09T20:51:46.996 INFO:teuthology.orchestra.run.vm01.stdout:Waiting for mgr... 2026-03-09T20:51:47.361 INFO:journalctl@ceph.mgr.a.vm01.stdout:Mar 09 20:51:46 vm01 podman[53628]: 2026-03-09 20:51:46.982933565 +0000 UTC m=+0.063689192 container init 9fc252e619ba3254f69778a8ecbaac0c12aa7a6d41ba3880147c41cb162f5d15 (image=quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df, name=ceph-9cb345a8-1bf9-11f1-a195-0375563c5891-mgr-a, org.label-schema.license=GPLv2, org.opencontainers.image.authors=Ceph Release Team , CEPH_SHA1=e911bdebe5c8faa3800735d1568fcdca65db60df, org.label-schema.vendor=CentOS, org.label-schema.name=CentOS Stream 9 Base Image, FROM_IMAGE=quay.io/centos/centos:stream9, io.buildah.version=1.41.3, CEPH_GIT_REPO=https://github.com/ceph/ceph-ci.git, OSD_FLAVOR=default, org.opencontainers.image.documentation=https://docs.ceph.com/, org.label-schema.schema-version=1.0, CEPH_REF=squid, ceph=True, GANESHA_REPO_BASEURL=https://buildlogs.centos.org/centos/$releasever-stream/storage/$basearch/nfsganesha-5/, org.label-schema.build-date=20260223) 2026-03-09T20:51:47.361 INFO:journalctl@ceph.mgr.a.vm01.stdout:Mar 09 20:51:46 vm01 podman[53628]: 2026-03-09 20:51:46.987832385 +0000 UTC m=+0.068588012 container start 9fc252e619ba3254f69778a8ecbaac0c12aa7a6d41ba3880147c41cb162f5d15 (image=quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df, name=ceph-9cb345a8-1bf9-11f1-a195-0375563c5891-mgr-a, org.label-schema.schema-version=1.0, org.label-schema.vendor=CentOS, ceph=True, CEPH_GIT_REPO=https://github.com/ceph/ceph-ci.git, GANESHA_REPO_BASEURL=https://buildlogs.centos.org/centos/$releasever-stream/storage/$basearch/nfsganesha-5/, org.label-schema.build-date=20260223, org.opencontainers.image.authors=Ceph Release Team , org.label-schema.name=CentOS Stream 9 Base Image, io.buildah.version=1.41.3, org.label-schema.license=GPLv2, CEPH_REF=squid, CEPH_SHA1=e911bdebe5c8faa3800735d1568fcdca65db60df, OSD_FLAVOR=default, FROM_IMAGE=quay.io/centos/centos:stream9, org.opencontainers.image.documentation=https://docs.ceph.com/) 2026-03-09T20:51:47.361 INFO:journalctl@ceph.mgr.a.vm01.stdout:Mar 09 20:51:46 vm01 bash[53628]: 9fc252e619ba3254f69778a8ecbaac0c12aa7a6d41ba3880147c41cb162f5d15 2026-03-09T20:51:47.361 INFO:journalctl@ceph.mgr.a.vm01.stdout:Mar 09 20:51:46 vm01 podman[53628]: 2026-03-09 20:51:46.929572558 +0000 UTC m=+0.010328194 image pull 654f31e6858eb235bbece362255b685a945f2b6a367e2b88c4930c984fbb214c quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df 2026-03-09T20:51:47.362 INFO:journalctl@ceph.mgr.a.vm01.stdout:Mar 09 20:51:46 vm01 systemd[1]: Started Ceph mgr.a for 9cb345a8-1bf9-11f1-a195-0375563c5891. 2026-03-09T20:51:47.362 INFO:journalctl@ceph.mgr.a.vm01.stdout:Mar 09 20:51:47 vm01 ceph-9cb345a8-1bf9-11f1-a195-0375563c5891-mgr-a[53639]: 2026-03-09T20:51:47.085+0000 7f7a638ed140 -1 mgr[py] Module alerts has missing NOTIFY_TYPES member 2026-03-09T20:51:47.362 INFO:journalctl@ceph.mgr.a.vm01.stdout:Mar 09 20:51:47 vm01 ceph-9cb345a8-1bf9-11f1-a195-0375563c5891-mgr-a[53639]: 2026-03-09T20:51:47.126+0000 7f7a638ed140 -1 mgr[py] Module balancer has missing NOTIFY_TYPES member 2026-03-09T20:51:47.600 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout 2026-03-09T20:51:47.600 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout { 2026-03-09T20:51:47.600 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "fsid": "9cb345a8-1bf9-11f1-a195-0375563c5891", 2026-03-09T20:51:47.600 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "health": { 2026-03-09T20:51:47.600 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "status": "HEALTH_OK", 2026-03-09T20:51:47.600 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "checks": {}, 2026-03-09T20:51:47.601 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "mutes": [] 2026-03-09T20:51:47.601 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout }, 2026-03-09T20:51:47.601 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "election_epoch": 5, 2026-03-09T20:51:47.601 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "quorum": [ 2026-03-09T20:51:47.601 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout 0 2026-03-09T20:51:47.601 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout ], 2026-03-09T20:51:47.601 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "quorum_names": [ 2026-03-09T20:51:47.601 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "a" 2026-03-09T20:51:47.601 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout ], 2026-03-09T20:51:47.601 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "quorum_age": 2, 2026-03-09T20:51:47.601 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "monmap": { 2026-03-09T20:51:47.601 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "epoch": 1, 2026-03-09T20:51:47.601 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "min_mon_release_name": "squid", 2026-03-09T20:51:47.601 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "num_mons": 1 2026-03-09T20:51:47.601 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout }, 2026-03-09T20:51:47.601 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "osdmap": { 2026-03-09T20:51:47.601 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "epoch": 1, 2026-03-09T20:51:47.601 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "num_osds": 0, 2026-03-09T20:51:47.601 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "num_up_osds": 0, 2026-03-09T20:51:47.601 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "osd_up_since": 0, 2026-03-09T20:51:47.601 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "num_in_osds": 0, 2026-03-09T20:51:47.601 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "osd_in_since": 0, 2026-03-09T20:51:47.601 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "num_remapped_pgs": 0 2026-03-09T20:51:47.601 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout }, 2026-03-09T20:51:47.601 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "pgmap": { 2026-03-09T20:51:47.601 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "pgs_by_state": [], 2026-03-09T20:51:47.601 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "num_pgs": 0, 2026-03-09T20:51:47.601 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "num_pools": 0, 2026-03-09T20:51:47.601 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "num_objects": 0, 2026-03-09T20:51:47.601 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "data_bytes": 0, 2026-03-09T20:51:47.601 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "bytes_used": 0, 2026-03-09T20:51:47.601 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "bytes_avail": 0, 2026-03-09T20:51:47.601 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "bytes_total": 0 2026-03-09T20:51:47.601 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout }, 2026-03-09T20:51:47.601 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "fsmap": { 2026-03-09T20:51:47.601 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "epoch": 1, 2026-03-09T20:51:47.601 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "btime": "2026-03-09T20:51:43:189291+0000", 2026-03-09T20:51:47.601 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "by_rank": [], 2026-03-09T20:51:47.601 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "up:standby": 0 2026-03-09T20:51:47.601 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout }, 2026-03-09T20:51:47.601 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "mgrmap": { 2026-03-09T20:51:47.602 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "available": false, 2026-03-09T20:51:47.602 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "num_standbys": 0, 2026-03-09T20:51:47.602 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "modules": [ 2026-03-09T20:51:47.602 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "iostat", 2026-03-09T20:51:47.602 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "nfs", 2026-03-09T20:51:47.602 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "restful" 2026-03-09T20:51:47.602 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout ], 2026-03-09T20:51:47.602 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "services": {} 2026-03-09T20:51:47.602 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout }, 2026-03-09T20:51:47.602 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "servicemap": { 2026-03-09T20:51:47.602 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "epoch": 1, 2026-03-09T20:51:47.602 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "modified": "2026-03-09T20:51:43.190306+0000", 2026-03-09T20:51:47.602 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "services": {} 2026-03-09T20:51:47.602 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout }, 2026-03-09T20:51:47.602 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "progress_events": {} 2026-03-09T20:51:47.602 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout } 2026-03-09T20:51:47.602 INFO:teuthology.orchestra.run.vm01.stdout:mgr not available, waiting (1/15)... 2026-03-09T20:51:47.674 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:51:47 vm01 ceph-mon[53402]: from='client.? 192.168.123.101:0/3478986751' entity='client.admin' 2026-03-09T20:51:47.674 INFO:journalctl@ceph.mgr.a.vm01.stdout:Mar 09 20:51:47 vm01 ceph-9cb345a8-1bf9-11f1-a195-0375563c5891-mgr-a[53639]: 2026-03-09T20:51:47.564+0000 7f7a638ed140 -1 mgr[py] Module crash has missing NOTIFY_TYPES member 2026-03-09T20:51:48.174 INFO:journalctl@ceph.mgr.a.vm01.stdout:Mar 09 20:51:47 vm01 ceph-9cb345a8-1bf9-11f1-a195-0375563c5891-mgr-a[53639]: 2026-03-09T20:51:47.885+0000 7f7a638ed140 -1 mgr[py] Module devicehealth has missing NOTIFY_TYPES member 2026-03-09T20:51:48.174 INFO:journalctl@ceph.mgr.a.vm01.stdout:Mar 09 20:51:47 vm01 ceph-9cb345a8-1bf9-11f1-a195-0375563c5891-mgr-a[53639]: /lib64/python3.9/site-packages/scipy/__init__.py:73: UserWarning: NumPy was imported from a Python sub-interpreter but NumPy does not properly support sub-interpreters. This will likely work for most users but might cause hard to track down issues or subtle bugs. A common user of the rare sub-interpreter feature is wsgi which also allows single-interpreter mode. 2026-03-09T20:51:48.174 INFO:journalctl@ceph.mgr.a.vm01.stdout:Mar 09 20:51:47 vm01 ceph-9cb345a8-1bf9-11f1-a195-0375563c5891-mgr-a[53639]: Improvements in the case of bugs are welcome, but is not on the NumPy roadmap, and full support may require significant effort to achieve. 2026-03-09T20:51:48.174 INFO:journalctl@ceph.mgr.a.vm01.stdout:Mar 09 20:51:47 vm01 ceph-9cb345a8-1bf9-11f1-a195-0375563c5891-mgr-a[53639]: from numpy import show_config as show_numpy_config 2026-03-09T20:51:48.174 INFO:journalctl@ceph.mgr.a.vm01.stdout:Mar 09 20:51:47 vm01 ceph-9cb345a8-1bf9-11f1-a195-0375563c5891-mgr-a[53639]: 2026-03-09T20:51:47.970+0000 7f7a638ed140 -1 mgr[py] Module diskprediction_local has missing NOTIFY_TYPES member 2026-03-09T20:51:48.174 INFO:journalctl@ceph.mgr.a.vm01.stdout:Mar 09 20:51:48 vm01 ceph-9cb345a8-1bf9-11f1-a195-0375563c5891-mgr-a[53639]: 2026-03-09T20:51:48.006+0000 7f7a638ed140 -1 mgr[py] Module influx has missing NOTIFY_TYPES member 2026-03-09T20:51:48.174 INFO:journalctl@ceph.mgr.a.vm01.stdout:Mar 09 20:51:48 vm01 ceph-9cb345a8-1bf9-11f1-a195-0375563c5891-mgr-a[53639]: 2026-03-09T20:51:48.075+0000 7f7a638ed140 -1 mgr[py] Module iostat has missing NOTIFY_TYPES member 2026-03-09T20:51:48.674 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:51:48 vm01 ceph-mon[53402]: from='client.? 192.168.123.101:0/1368915480' entity='client.admin' cmd=[{"prefix": "status", "format": "json-pretty"}]: dispatch 2026-03-09T20:51:48.674 INFO:journalctl@ceph.mgr.a.vm01.stdout:Mar 09 20:51:48 vm01 ceph-9cb345a8-1bf9-11f1-a195-0375563c5891-mgr-a[53639]: 2026-03-09T20:51:48.577+0000 7f7a638ed140 -1 mgr[py] Module nfs has missing NOTIFY_TYPES member 2026-03-09T20:51:49.039 INFO:journalctl@ceph.mgr.a.vm01.stdout:Mar 09 20:51:48 vm01 ceph-9cb345a8-1bf9-11f1-a195-0375563c5891-mgr-a[53639]: 2026-03-09T20:51:48.699+0000 7f7a638ed140 -1 mgr[py] Module orchestrator has missing NOTIFY_TYPES member 2026-03-09T20:51:49.039 INFO:journalctl@ceph.mgr.a.vm01.stdout:Mar 09 20:51:48 vm01 ceph-9cb345a8-1bf9-11f1-a195-0375563c5891-mgr-a[53639]: 2026-03-09T20:51:48.739+0000 7f7a638ed140 -1 mgr[py] Module osd_perf_query has missing NOTIFY_TYPES member 2026-03-09T20:51:49.039 INFO:journalctl@ceph.mgr.a.vm01.stdout:Mar 09 20:51:48 vm01 ceph-9cb345a8-1bf9-11f1-a195-0375563c5891-mgr-a[53639]: 2026-03-09T20:51:48.779+0000 7f7a638ed140 -1 mgr[py] Module osd_support has missing NOTIFY_TYPES member 2026-03-09T20:51:49.039 INFO:journalctl@ceph.mgr.a.vm01.stdout:Mar 09 20:51:48 vm01 ceph-9cb345a8-1bf9-11f1-a195-0375563c5891-mgr-a[53639]: 2026-03-09T20:51:48.821+0000 7f7a638ed140 -1 mgr[py] Module pg_autoscaler has missing NOTIFY_TYPES member 2026-03-09T20:51:49.039 INFO:journalctl@ceph.mgr.a.vm01.stdout:Mar 09 20:51:48 vm01 ceph-9cb345a8-1bf9-11f1-a195-0375563c5891-mgr-a[53639]: 2026-03-09T20:51:48.858+0000 7f7a638ed140 -1 mgr[py] Module progress has missing NOTIFY_TYPES member 2026-03-09T20:51:49.039 INFO:journalctl@ceph.mgr.a.vm01.stdout:Mar 09 20:51:49 vm01 ceph-9cb345a8-1bf9-11f1-a195-0375563c5891-mgr-a[53639]: 2026-03-09T20:51:49.037+0000 7f7a638ed140 -1 mgr[py] Module prometheus has missing NOTIFY_TYPES member 2026-03-09T20:51:49.304 INFO:journalctl@ceph.mgr.a.vm01.stdout:Mar 09 20:51:49 vm01 ceph-9cb345a8-1bf9-11f1-a195-0375563c5891-mgr-a[53639]: 2026-03-09T20:51:49.087+0000 7f7a638ed140 -1 mgr[py] Module rbd_support has missing NOTIFY_TYPES member 2026-03-09T20:51:49.304 INFO:journalctl@ceph.mgr.a.vm01.stdout:Mar 09 20:51:49 vm01 ceph-9cb345a8-1bf9-11f1-a195-0375563c5891-mgr-a[53639]: 2026-03-09T20:51:49.302+0000 7f7a638ed140 -1 mgr[py] Module rgw has missing NOTIFY_TYPES member 2026-03-09T20:51:49.850 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:51:49 vm01 ceph-mon[53402]: from='client.? 192.168.123.101:0/3266560349' entity='client.admin' cmd=[{"prefix": "status", "format": "json-pretty"}]: dispatch 2026-03-09T20:51:49.851 INFO:journalctl@ceph.mgr.a.vm01.stdout:Mar 09 20:51:49 vm01 ceph-9cb345a8-1bf9-11f1-a195-0375563c5891-mgr-a[53639]: 2026-03-09T20:51:49.579+0000 7f7a638ed140 -1 mgr[py] Module rook has missing NOTIFY_TYPES member 2026-03-09T20:51:49.851 INFO:journalctl@ceph.mgr.a.vm01.stdout:Mar 09 20:51:49 vm01 ceph-9cb345a8-1bf9-11f1-a195-0375563c5891-mgr-a[53639]: 2026-03-09T20:51:49.616+0000 7f7a638ed140 -1 mgr[py] Module selftest has missing NOTIFY_TYPES member 2026-03-09T20:51:49.851 INFO:journalctl@ceph.mgr.a.vm01.stdout:Mar 09 20:51:49 vm01 ceph-9cb345a8-1bf9-11f1-a195-0375563c5891-mgr-a[53639]: 2026-03-09T20:51:49.663+0000 7f7a638ed140 -1 mgr[py] Module snap_schedule has missing NOTIFY_TYPES member 2026-03-09T20:51:49.851 INFO:journalctl@ceph.mgr.a.vm01.stdout:Mar 09 20:51:49 vm01 ceph-9cb345a8-1bf9-11f1-a195-0375563c5891-mgr-a[53639]: 2026-03-09T20:51:49.752+0000 7f7a638ed140 -1 mgr[py] Module status has missing NOTIFY_TYPES member 2026-03-09T20:51:49.851 INFO:journalctl@ceph.mgr.a.vm01.stdout:Mar 09 20:51:49 vm01 ceph-9cb345a8-1bf9-11f1-a195-0375563c5891-mgr-a[53639]: 2026-03-09T20:51:49.799+0000 7f7a638ed140 -1 mgr[py] Module telegraf has missing NOTIFY_TYPES member 2026-03-09T20:51:49.942 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout 2026-03-09T20:51:49.942 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout { 2026-03-09T20:51:49.943 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "fsid": "9cb345a8-1bf9-11f1-a195-0375563c5891", 2026-03-09T20:51:49.943 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "health": { 2026-03-09T20:51:49.943 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "status": "HEALTH_OK", 2026-03-09T20:51:49.943 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "checks": {}, 2026-03-09T20:51:49.943 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "mutes": [] 2026-03-09T20:51:49.943 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout }, 2026-03-09T20:51:49.943 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "election_epoch": 5, 2026-03-09T20:51:49.943 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "quorum": [ 2026-03-09T20:51:49.943 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout 0 2026-03-09T20:51:49.943 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout ], 2026-03-09T20:51:49.943 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "quorum_names": [ 2026-03-09T20:51:49.943 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "a" 2026-03-09T20:51:49.943 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout ], 2026-03-09T20:51:49.943 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "quorum_age": 4, 2026-03-09T20:51:49.943 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "monmap": { 2026-03-09T20:51:49.943 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "epoch": 1, 2026-03-09T20:51:49.943 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "min_mon_release_name": "squid", 2026-03-09T20:51:49.944 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "num_mons": 1 2026-03-09T20:51:49.944 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout }, 2026-03-09T20:51:49.944 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "osdmap": { 2026-03-09T20:51:49.944 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "epoch": 1, 2026-03-09T20:51:49.944 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "num_osds": 0, 2026-03-09T20:51:49.944 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "num_up_osds": 0, 2026-03-09T20:51:49.944 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "osd_up_since": 0, 2026-03-09T20:51:49.944 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "num_in_osds": 0, 2026-03-09T20:51:49.944 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "osd_in_since": 0, 2026-03-09T20:51:49.944 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "num_remapped_pgs": 0 2026-03-09T20:51:49.944 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout }, 2026-03-09T20:51:49.944 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "pgmap": { 2026-03-09T20:51:49.944 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "pgs_by_state": [], 2026-03-09T20:51:49.944 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "num_pgs": 0, 2026-03-09T20:51:49.944 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "num_pools": 0, 2026-03-09T20:51:49.944 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "num_objects": 0, 2026-03-09T20:51:49.944 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "data_bytes": 0, 2026-03-09T20:51:49.944 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "bytes_used": 0, 2026-03-09T20:51:49.944 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "bytes_avail": 0, 2026-03-09T20:51:49.944 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "bytes_total": 0 2026-03-09T20:51:49.944 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout }, 2026-03-09T20:51:49.944 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "fsmap": { 2026-03-09T20:51:49.944 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "epoch": 1, 2026-03-09T20:51:49.944 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "btime": "2026-03-09T20:51:43:189291+0000", 2026-03-09T20:51:49.944 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "by_rank": [], 2026-03-09T20:51:49.944 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "up:standby": 0 2026-03-09T20:51:49.944 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout }, 2026-03-09T20:51:49.944 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "mgrmap": { 2026-03-09T20:51:49.944 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "available": false, 2026-03-09T20:51:49.944 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "num_standbys": 0, 2026-03-09T20:51:49.944 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "modules": [ 2026-03-09T20:51:49.944 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "iostat", 2026-03-09T20:51:49.944 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "nfs", 2026-03-09T20:51:49.944 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "restful" 2026-03-09T20:51:49.944 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout ], 2026-03-09T20:51:49.944 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "services": {} 2026-03-09T20:51:49.944 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout }, 2026-03-09T20:51:49.944 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "servicemap": { 2026-03-09T20:51:49.944 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "epoch": 1, 2026-03-09T20:51:49.944 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "modified": "2026-03-09T20:51:43.190306+0000", 2026-03-09T20:51:49.944 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "services": {} 2026-03-09T20:51:49.944 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout }, 2026-03-09T20:51:49.944 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "progress_events": {} 2026-03-09T20:51:49.944 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout } 2026-03-09T20:51:49.944 INFO:teuthology.orchestra.run.vm01.stdout:mgr not available, waiting (2/15)... 2026-03-09T20:51:50.139 INFO:journalctl@ceph.mgr.a.vm01.stdout:Mar 09 20:51:49 vm01 ceph-9cb345a8-1bf9-11f1-a195-0375563c5891-mgr-a[53639]: 2026-03-09T20:51:49.887+0000 7f7a638ed140 -1 mgr[py] Module telemetry has missing NOTIFY_TYPES member 2026-03-09T20:51:50.139 INFO:journalctl@ceph.mgr.a.vm01.stdout:Mar 09 20:51:50 vm01 ceph-9cb345a8-1bf9-11f1-a195-0375563c5891-mgr-a[53639]: 2026-03-09T20:51:50.003+0000 7f7a638ed140 -1 mgr[py] Module test_orchestrator has missing NOTIFY_TYPES member 2026-03-09T20:51:50.139 INFO:journalctl@ceph.mgr.a.vm01.stdout:Mar 09 20:51:50 vm01 ceph-9cb345a8-1bf9-11f1-a195-0375563c5891-mgr-a[53639]: 2026-03-09T20:51:50.137+0000 7f7a638ed140 -1 mgr[py] Module volumes has missing NOTIFY_TYPES member 2026-03-09T20:51:50.424 INFO:journalctl@ceph.mgr.a.vm01.stdout:Mar 09 20:51:50 vm01 ceph-9cb345a8-1bf9-11f1-a195-0375563c5891-mgr-a[53639]: 2026-03-09T20:51:50.174+0000 7f7a638ed140 -1 mgr[py] Module zabbix has missing NOTIFY_TYPES member 2026-03-09T20:51:51.174 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:51:50 vm01 ceph-mon[53402]: Activating manager daemon a 2026-03-09T20:51:51.174 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:51:50 vm01 ceph-mon[53402]: mgrmap e2: a(active, starting, since 0.00396147s) 2026-03-09T20:51:51.174 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:51:50 vm01 ceph-mon[53402]: from='mgr.14100 192.168.123.101:0/1115159257' entity='mgr.a' cmd=[{"prefix": "mds metadata"}]: dispatch 2026-03-09T20:51:51.174 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:51:50 vm01 ceph-mon[53402]: from='mgr.14100 192.168.123.101:0/1115159257' entity='mgr.a' cmd=[{"prefix": "osd metadata"}]: dispatch 2026-03-09T20:51:51.174 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:51:50 vm01 ceph-mon[53402]: from='mgr.14100 192.168.123.101:0/1115159257' entity='mgr.a' cmd=[{"prefix": "mon metadata"}]: dispatch 2026-03-09T20:51:51.174 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:51:50 vm01 ceph-mon[53402]: from='mgr.14100 192.168.123.101:0/1115159257' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "a"}]: dispatch 2026-03-09T20:51:51.174 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:51:50 vm01 ceph-mon[53402]: from='mgr.14100 192.168.123.101:0/1115159257' entity='mgr.a' cmd=[{"prefix": "mgr metadata", "who": "a", "id": "a"}]: dispatch 2026-03-09T20:51:51.174 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:51:50 vm01 ceph-mon[53402]: Manager daemon a is now available 2026-03-09T20:51:51.174 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:51:50 vm01 ceph-mon[53402]: from='mgr.14100 192.168.123.101:0/1115159257' entity='mgr.a' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/a/mirror_snapshot_schedule"}]: dispatch 2026-03-09T20:51:51.175 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:51:50 vm01 ceph-mon[53402]: from='mgr.14100 192.168.123.101:0/1115159257' entity='mgr.a' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/a/trash_purge_schedule"}]: dispatch 2026-03-09T20:51:51.175 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:51:50 vm01 ceph-mon[53402]: from='mgr.14100 192.168.123.101:0/1115159257' entity='mgr.a' 2026-03-09T20:51:51.175 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:51:50 vm01 ceph-mon[53402]: from='mgr.14100 192.168.123.101:0/1115159257' entity='mgr.a' 2026-03-09T20:51:51.175 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:51:50 vm01 ceph-mon[53402]: from='mgr.14100 192.168.123.101:0/1115159257' entity='mgr.a' 2026-03-09T20:51:52.341 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:51:52 vm01 ceph-mon[53402]: mgrmap e3: a(active, since 1.00801s) 2026-03-09T20:51:52.515 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout 2026-03-09T20:51:52.515 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout { 2026-03-09T20:51:52.515 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "fsid": "9cb345a8-1bf9-11f1-a195-0375563c5891", 2026-03-09T20:51:52.515 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "health": { 2026-03-09T20:51:52.515 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "status": "HEALTH_OK", 2026-03-09T20:51:52.515 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "checks": {}, 2026-03-09T20:51:52.515 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "mutes": [] 2026-03-09T20:51:52.515 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout }, 2026-03-09T20:51:52.515 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "election_epoch": 5, 2026-03-09T20:51:52.515 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "quorum": [ 2026-03-09T20:51:52.515 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout 0 2026-03-09T20:51:52.515 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout ], 2026-03-09T20:51:52.515 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "quorum_names": [ 2026-03-09T20:51:52.515 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "a" 2026-03-09T20:51:52.515 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout ], 2026-03-09T20:51:52.515 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "quorum_age": 7, 2026-03-09T20:51:52.515 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "monmap": { 2026-03-09T20:51:52.515 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "epoch": 1, 2026-03-09T20:51:52.515 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "min_mon_release_name": "squid", 2026-03-09T20:51:52.515 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "num_mons": 1 2026-03-09T20:51:52.515 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout }, 2026-03-09T20:51:52.515 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "osdmap": { 2026-03-09T20:51:52.515 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "epoch": 1, 2026-03-09T20:51:52.515 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "num_osds": 0, 2026-03-09T20:51:52.515 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "num_up_osds": 0, 2026-03-09T20:51:52.515 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "osd_up_since": 0, 2026-03-09T20:51:52.515 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "num_in_osds": 0, 2026-03-09T20:51:52.515 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "osd_in_since": 0, 2026-03-09T20:51:52.515 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "num_remapped_pgs": 0 2026-03-09T20:51:52.515 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout }, 2026-03-09T20:51:52.515 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "pgmap": { 2026-03-09T20:51:52.515 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "pgs_by_state": [], 2026-03-09T20:51:52.515 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "num_pgs": 0, 2026-03-09T20:51:52.515 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "num_pools": 0, 2026-03-09T20:51:52.515 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "num_objects": 0, 2026-03-09T20:51:52.515 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "data_bytes": 0, 2026-03-09T20:51:52.515 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "bytes_used": 0, 2026-03-09T20:51:52.515 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "bytes_avail": 0, 2026-03-09T20:51:52.516 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "bytes_total": 0 2026-03-09T20:51:52.516 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout }, 2026-03-09T20:51:52.516 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "fsmap": { 2026-03-09T20:51:52.516 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "epoch": 1, 2026-03-09T20:51:52.516 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "btime": "2026-03-09T20:51:43:189291+0000", 2026-03-09T20:51:52.516 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "by_rank": [], 2026-03-09T20:51:52.516 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "up:standby": 0 2026-03-09T20:51:52.516 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout }, 2026-03-09T20:51:52.516 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "mgrmap": { 2026-03-09T20:51:52.516 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "available": true, 2026-03-09T20:51:52.516 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "num_standbys": 0, 2026-03-09T20:51:52.516 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "modules": [ 2026-03-09T20:51:52.516 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "iostat", 2026-03-09T20:51:52.516 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "nfs", 2026-03-09T20:51:52.516 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "restful" 2026-03-09T20:51:52.516 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout ], 2026-03-09T20:51:52.516 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "services": {} 2026-03-09T20:51:52.516 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout }, 2026-03-09T20:51:52.516 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "servicemap": { 2026-03-09T20:51:52.516 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "epoch": 1, 2026-03-09T20:51:52.516 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "modified": "2026-03-09T20:51:43.190306+0000", 2026-03-09T20:51:52.516 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "services": {} 2026-03-09T20:51:52.516 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout }, 2026-03-09T20:51:52.516 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "progress_events": {} 2026-03-09T20:51:52.516 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout } 2026-03-09T20:51:52.516 INFO:teuthology.orchestra.run.vm01.stdout:mgr is available 2026-03-09T20:51:52.913 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout 2026-03-09T20:51:52.913 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout [global] 2026-03-09T20:51:52.913 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout fsid = 9cb345a8-1bf9-11f1-a195-0375563c5891 2026-03-09T20:51:52.913 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout mon_cluster_log_file_level = debug 2026-03-09T20:51:52.913 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout mon_host = [v2:192.168.123.101:3300,v1:192.168.123.101:6789] 2026-03-09T20:51:52.914 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout mon_osd_allow_pg_remap = true 2026-03-09T20:51:52.914 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout mon_osd_allow_primary_affinity = true 2026-03-09T20:51:52.914 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout mon_warn_on_no_sortbitwise = false 2026-03-09T20:51:52.914 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout osd_crush_chooseleaf_type = 0 2026-03-09T20:51:52.914 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout 2026-03-09T20:51:52.914 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout [mgr] 2026-03-09T20:51:52.914 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout mgr/telemetry/nag = false 2026-03-09T20:51:52.914 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout 2026-03-09T20:51:52.914 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout [osd] 2026-03-09T20:51:52.914 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout osd_map_max_advance = 10 2026-03-09T20:51:52.914 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout osd_sloppy_crc = true 2026-03-09T20:51:52.914 INFO:teuthology.orchestra.run.vm01.stdout:Enabling cephadm module... 2026-03-09T20:51:53.674 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:51:53 vm01 ceph-mon[53402]: mgrmap e4: a(active, since 2s) 2026-03-09T20:51:53.674 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:51:53 vm01 ceph-mon[53402]: from='client.? 192.168.123.101:0/183043839' entity='client.admin' cmd=[{"prefix": "status", "format": "json-pretty"}]: dispatch 2026-03-09T20:51:53.674 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:51:53 vm01 ceph-mon[53402]: from='client.? 192.168.123.101:0/127714295' entity='client.admin' cmd=[{"prefix": "config assimilate-conf"}]: dispatch 2026-03-09T20:51:53.674 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:51:53 vm01 ceph-mon[53402]: from='client.? 192.168.123.101:0/127714295' entity='client.admin' cmd='[{"prefix": "config assimilate-conf"}]': finished 2026-03-09T20:51:53.674 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:51:53 vm01 ceph-mon[53402]: from='client.? 192.168.123.101:0/1793140337' entity='client.admin' cmd=[{"prefix": "mgr module enable", "module": "cephadm"}]: dispatch 2026-03-09T20:51:54.174 INFO:journalctl@ceph.mgr.a.vm01.stdout:Mar 09 20:51:53 vm01 ceph-9cb345a8-1bf9-11f1-a195-0375563c5891-mgr-a[53639]: ignoring --setuser ceph since I am not root 2026-03-09T20:51:54.174 INFO:journalctl@ceph.mgr.a.vm01.stdout:Mar 09 20:51:53 vm01 ceph-9cb345a8-1bf9-11f1-a195-0375563c5891-mgr-a[53639]: ignoring --setgroup ceph since I am not root 2026-03-09T20:51:54.174 INFO:journalctl@ceph.mgr.a.vm01.stdout:Mar 09 20:51:53 vm01 ceph-9cb345a8-1bf9-11f1-a195-0375563c5891-mgr-a[53639]: 2026-03-09T20:51:53.873+0000 7f6b219c2140 -1 mgr[py] Module alerts has missing NOTIFY_TYPES member 2026-03-09T20:51:54.174 INFO:journalctl@ceph.mgr.a.vm01.stdout:Mar 09 20:51:53 vm01 ceph-9cb345a8-1bf9-11f1-a195-0375563c5891-mgr-a[53639]: 2026-03-09T20:51:53.930+0000 7f6b219c2140 -1 mgr[py] Module balancer has missing NOTIFY_TYPES member 2026-03-09T20:51:54.323 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout { 2026-03-09T20:51:54.323 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "epoch": 5, 2026-03-09T20:51:54.323 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "available": true, 2026-03-09T20:51:54.323 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "active_name": "a", 2026-03-09T20:51:54.323 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "num_standby": 0 2026-03-09T20:51:54.323 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout } 2026-03-09T20:51:54.323 INFO:teuthology.orchestra.run.vm01.stdout:Waiting for the mgr to restart... 2026-03-09T20:51:54.323 INFO:teuthology.orchestra.run.vm01.stdout:Waiting for mgr epoch 5... 2026-03-09T20:51:54.439 INFO:journalctl@ceph.mgr.a.vm01.stdout:Mar 09 20:51:54 vm01 ceph-9cb345a8-1bf9-11f1-a195-0375563c5891-mgr-a[53639]: 2026-03-09T20:51:54.351+0000 7f6b219c2140 -1 mgr[py] Module crash has missing NOTIFY_TYPES member 2026-03-09T20:51:54.757 INFO:journalctl@ceph.mgr.a.vm01.stdout:Mar 09 20:51:54 vm01 ceph-9cb345a8-1bf9-11f1-a195-0375563c5891-mgr-a[53639]: 2026-03-09T20:51:54.686+0000 7f6b219c2140 -1 mgr[py] Module devicehealth has missing NOTIFY_TYPES member 2026-03-09T20:51:55.174 INFO:journalctl@ceph.mgr.a.vm01.stdout:Mar 09 20:51:54 vm01 ceph-9cb345a8-1bf9-11f1-a195-0375563c5891-mgr-a[53639]: /lib64/python3.9/site-packages/scipy/__init__.py:73: UserWarning: NumPy was imported from a Python sub-interpreter but NumPy does not properly support sub-interpreters. This will likely work for most users but might cause hard to track down issues or subtle bugs. A common user of the rare sub-interpreter feature is wsgi which also allows single-interpreter mode. 2026-03-09T20:51:55.174 INFO:journalctl@ceph.mgr.a.vm01.stdout:Mar 09 20:51:54 vm01 ceph-9cb345a8-1bf9-11f1-a195-0375563c5891-mgr-a[53639]: Improvements in the case of bugs are welcome, but is not on the NumPy roadmap, and full support may require significant effort to achieve. 2026-03-09T20:51:55.174 INFO:journalctl@ceph.mgr.a.vm01.stdout:Mar 09 20:51:54 vm01 ceph-9cb345a8-1bf9-11f1-a195-0375563c5891-mgr-a[53639]: from numpy import show_config as show_numpy_config 2026-03-09T20:51:55.174 INFO:journalctl@ceph.mgr.a.vm01.stdout:Mar 09 20:51:54 vm01 ceph-9cb345a8-1bf9-11f1-a195-0375563c5891-mgr-a[53639]: 2026-03-09T20:51:54.772+0000 7f6b219c2140 -1 mgr[py] Module diskprediction_local has missing NOTIFY_TYPES member 2026-03-09T20:51:55.174 INFO:journalctl@ceph.mgr.a.vm01.stdout:Mar 09 20:51:54 vm01 ceph-9cb345a8-1bf9-11f1-a195-0375563c5891-mgr-a[53639]: 2026-03-09T20:51:54.812+0000 7f6b219c2140 -1 mgr[py] Module influx has missing NOTIFY_TYPES member 2026-03-09T20:51:55.174 INFO:journalctl@ceph.mgr.a.vm01.stdout:Mar 09 20:51:54 vm01 ceph-9cb345a8-1bf9-11f1-a195-0375563c5891-mgr-a[53639]: 2026-03-09T20:51:54.881+0000 7f6b219c2140 -1 mgr[py] Module iostat has missing NOTIFY_TYPES member 2026-03-09T20:51:55.174 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:51:54 vm01 ceph-mon[53402]: from='client.? 192.168.123.101:0/1793140337' entity='client.admin' cmd='[{"prefix": "mgr module enable", "module": "cephadm"}]': finished 2026-03-09T20:51:55.174 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:51:54 vm01 ceph-mon[53402]: mgrmap e5: a(active, since 3s) 2026-03-09T20:51:55.174 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:51:54 vm01 ceph-mon[53402]: from='client.? 192.168.123.101:0/3553558713' entity='client.admin' cmd=[{"prefix": "mgr stat"}]: dispatch 2026-03-09T20:51:55.674 INFO:journalctl@ceph.mgr.a.vm01.stdout:Mar 09 20:51:55 vm01 ceph-9cb345a8-1bf9-11f1-a195-0375563c5891-mgr-a[53639]: 2026-03-09T20:51:55.349+0000 7f6b219c2140 -1 mgr[py] Module nfs has missing NOTIFY_TYPES member 2026-03-09T20:51:55.674 INFO:journalctl@ceph.mgr.a.vm01.stdout:Mar 09 20:51:55 vm01 ceph-9cb345a8-1bf9-11f1-a195-0375563c5891-mgr-a[53639]: 2026-03-09T20:51:55.452+0000 7f6b219c2140 -1 mgr[py] Module orchestrator has missing NOTIFY_TYPES member 2026-03-09T20:51:55.674 INFO:journalctl@ceph.mgr.a.vm01.stdout:Mar 09 20:51:55 vm01 ceph-9cb345a8-1bf9-11f1-a195-0375563c5891-mgr-a[53639]: 2026-03-09T20:51:55.489+0000 7f6b219c2140 -1 mgr[py] Module osd_perf_query has missing NOTIFY_TYPES member 2026-03-09T20:51:55.674 INFO:journalctl@ceph.mgr.a.vm01.stdout:Mar 09 20:51:55 vm01 ceph-9cb345a8-1bf9-11f1-a195-0375563c5891-mgr-a[53639]: 2026-03-09T20:51:55.521+0000 7f6b219c2140 -1 mgr[py] Module osd_support has missing NOTIFY_TYPES member 2026-03-09T20:51:55.674 INFO:journalctl@ceph.mgr.a.vm01.stdout:Mar 09 20:51:55 vm01 ceph-9cb345a8-1bf9-11f1-a195-0375563c5891-mgr-a[53639]: 2026-03-09T20:51:55.561+0000 7f6b219c2140 -1 mgr[py] Module pg_autoscaler has missing NOTIFY_TYPES member 2026-03-09T20:51:55.674 INFO:journalctl@ceph.mgr.a.vm01.stdout:Mar 09 20:51:55 vm01 ceph-9cb345a8-1bf9-11f1-a195-0375563c5891-mgr-a[53639]: 2026-03-09T20:51:55.596+0000 7f6b219c2140 -1 mgr[py] Module progress has missing NOTIFY_TYPES member 2026-03-09T20:51:56.014 INFO:journalctl@ceph.mgr.a.vm01.stdout:Mar 09 20:51:55 vm01 ceph-9cb345a8-1bf9-11f1-a195-0375563c5891-mgr-a[53639]: 2026-03-09T20:51:55.757+0000 7f6b219c2140 -1 mgr[py] Module prometheus has missing NOTIFY_TYPES member 2026-03-09T20:51:56.014 INFO:journalctl@ceph.mgr.a.vm01.stdout:Mar 09 20:51:55 vm01 ceph-9cb345a8-1bf9-11f1-a195-0375563c5891-mgr-a[53639]: 2026-03-09T20:51:55.805+0000 7f6b219c2140 -1 mgr[py] Module rbd_support has missing NOTIFY_TYPES member 2026-03-09T20:51:56.277 INFO:journalctl@ceph.mgr.a.vm01.stdout:Mar 09 20:51:56 vm01 ceph-9cb345a8-1bf9-11f1-a195-0375563c5891-mgr-a[53639]: 2026-03-09T20:51:56.012+0000 7f6b219c2140 -1 mgr[py] Module rgw has missing NOTIFY_TYPES member 2026-03-09T20:51:56.533 INFO:journalctl@ceph.mgr.a.vm01.stdout:Mar 09 20:51:56 vm01 ceph-9cb345a8-1bf9-11f1-a195-0375563c5891-mgr-a[53639]: 2026-03-09T20:51:56.276+0000 7f6b219c2140 -1 mgr[py] Module rook has missing NOTIFY_TYPES member 2026-03-09T20:51:56.533 INFO:journalctl@ceph.mgr.a.vm01.stdout:Mar 09 20:51:56 vm01 ceph-9cb345a8-1bf9-11f1-a195-0375563c5891-mgr-a[53639]: 2026-03-09T20:51:56.311+0000 7f6b219c2140 -1 mgr[py] Module selftest has missing NOTIFY_TYPES member 2026-03-09T20:51:56.533 INFO:journalctl@ceph.mgr.a.vm01.stdout:Mar 09 20:51:56 vm01 ceph-9cb345a8-1bf9-11f1-a195-0375563c5891-mgr-a[53639]: 2026-03-09T20:51:56.350+0000 7f6b219c2140 -1 mgr[py] Module snap_schedule has missing NOTIFY_TYPES member 2026-03-09T20:51:56.533 INFO:journalctl@ceph.mgr.a.vm01.stdout:Mar 09 20:51:56 vm01 ceph-9cb345a8-1bf9-11f1-a195-0375563c5891-mgr-a[53639]: 2026-03-09T20:51:56.422+0000 7f6b219c2140 -1 mgr[py] Module status has missing NOTIFY_TYPES member 2026-03-09T20:51:56.533 INFO:journalctl@ceph.mgr.a.vm01.stdout:Mar 09 20:51:56 vm01 ceph-9cb345a8-1bf9-11f1-a195-0375563c5891-mgr-a[53639]: 2026-03-09T20:51:56.456+0000 7f6b219c2140 -1 mgr[py] Module telegraf has missing NOTIFY_TYPES member 2026-03-09T20:51:56.803 INFO:journalctl@ceph.mgr.a.vm01.stdout:Mar 09 20:51:56 vm01 ceph-9cb345a8-1bf9-11f1-a195-0375563c5891-mgr-a[53639]: 2026-03-09T20:51:56.531+0000 7f6b219c2140 -1 mgr[py] Module telemetry has missing NOTIFY_TYPES member 2026-03-09T20:51:56.803 INFO:journalctl@ceph.mgr.a.vm01.stdout:Mar 09 20:51:56 vm01 ceph-9cb345a8-1bf9-11f1-a195-0375563c5891-mgr-a[53639]: 2026-03-09T20:51:56.637+0000 7f6b219c2140 -1 mgr[py] Module test_orchestrator has missing NOTIFY_TYPES member 2026-03-09T20:51:56.803 INFO:journalctl@ceph.mgr.a.vm01.stdout:Mar 09 20:51:56 vm01 ceph-9cb345a8-1bf9-11f1-a195-0375563c5891-mgr-a[53639]: 2026-03-09T20:51:56.764+0000 7f6b219c2140 -1 mgr[py] Module volumes has missing NOTIFY_TYPES member 2026-03-09T20:51:57.174 INFO:journalctl@ceph.mgr.a.vm01.stdout:Mar 09 20:51:56 vm01 ceph-9cb345a8-1bf9-11f1-a195-0375563c5891-mgr-a[53639]: 2026-03-09T20:51:56.802+0000 7f6b219c2140 -1 mgr[py] Module zabbix has missing NOTIFY_TYPES member 2026-03-09T20:51:57.174 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:51:56 vm01 ceph-mon[53402]: Active manager daemon a restarted 2026-03-09T20:51:57.174 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:51:56 vm01 ceph-mon[53402]: Activating manager daemon a 2026-03-09T20:51:57.174 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:51:56 vm01 ceph-mon[53402]: osdmap e2: 0 total, 0 up, 0 in 2026-03-09T20:51:57.174 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:51:56 vm01 ceph-mon[53402]: mgrmap e6: a(active, starting, since 0.00589906s) 2026-03-09T20:51:57.174 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:51:56 vm01 ceph-mon[53402]: from='mgr.14118 192.168.123.101:0/3188408464' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "a"}]: dispatch 2026-03-09T20:51:57.174 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:51:56 vm01 ceph-mon[53402]: from='mgr.14118 192.168.123.101:0/3188408464' entity='mgr.a' cmd=[{"prefix": "mgr metadata", "who": "a", "id": "a"}]: dispatch 2026-03-09T20:51:57.174 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:51:56 vm01 ceph-mon[53402]: from='mgr.14118 192.168.123.101:0/3188408464' entity='mgr.a' cmd=[{"prefix": "mds metadata"}]: dispatch 2026-03-09T20:51:57.174 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:51:56 vm01 ceph-mon[53402]: from='mgr.14118 192.168.123.101:0/3188408464' entity='mgr.a' cmd=[{"prefix": "osd metadata"}]: dispatch 2026-03-09T20:51:57.174 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:51:56 vm01 ceph-mon[53402]: from='mgr.14118 192.168.123.101:0/3188408464' entity='mgr.a' cmd=[{"prefix": "mon metadata"}]: dispatch 2026-03-09T20:51:57.174 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:51:56 vm01 ceph-mon[53402]: Manager daemon a is now available 2026-03-09T20:51:57.174 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:51:56 vm01 ceph-mon[53402]: from='mgr.14118 192.168.123.101:0/3188408464' entity='mgr.a' 2026-03-09T20:51:57.174 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:51:56 vm01 ceph-mon[53402]: from='mgr.14118 192.168.123.101:0/3188408464' entity='mgr.a' 2026-03-09T20:51:57.174 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:51:56 vm01 ceph-mon[53402]: from='mgr.14118 192.168.123.101:0/3188408464' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T20:51:57.174 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:51:56 vm01 ceph-mon[53402]: from='mgr.14118 192.168.123.101:0/3188408464' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T20:51:57.174 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:51:56 vm01 ceph-mon[53402]: from='mgr.14118 192.168.123.101:0/3188408464' entity='mgr.a' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/a/mirror_snapshot_schedule"}]: dispatch 2026-03-09T20:51:57.966 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout { 2026-03-09T20:51:57.966 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "mgrmap_epoch": 7, 2026-03-09T20:51:57.966 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "initialized": true 2026-03-09T20:51:57.966 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout } 2026-03-09T20:51:57.966 INFO:teuthology.orchestra.run.vm01.stdout:mgr epoch 5 is available 2026-03-09T20:51:57.966 INFO:teuthology.orchestra.run.vm01.stdout:Setting orchestrator backend to cephadm... 2026-03-09T20:51:58.074 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:51:57 vm01 ceph-mon[53402]: Found migration_current of "None". Setting to last migration. 2026-03-09T20:51:58.074 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:51:57 vm01 ceph-mon[53402]: from='mgr.14118 192.168.123.101:0/3188408464' entity='mgr.a' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/a/trash_purge_schedule"}]: dispatch 2026-03-09T20:51:58.074 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:51:57 vm01 ceph-mon[53402]: from='mgr.14118 192.168.123.101:0/3188408464' entity='mgr.a' 2026-03-09T20:51:58.074 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:51:57 vm01 ceph-mon[53402]: from='mgr.14118 192.168.123.101:0/3188408464' entity='mgr.a' 2026-03-09T20:51:58.074 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:51:57 vm01 ceph-mon[53402]: mgrmap e7: a(active, since 1.00808s) 2026-03-09T20:51:58.757 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout value unchanged 2026-03-09T20:51:58.757 INFO:teuthology.orchestra.run.vm01.stdout:Generating ssh key... 2026-03-09T20:51:59.174 INFO:journalctl@ceph.mgr.a.vm01.stdout:Mar 09 20:51:59 vm01 ceph-9cb345a8-1bf9-11f1-a195-0375563c5891-mgr-a[53639]: Generating public/private ed25519 key pair. 2026-03-09T20:51:59.174 INFO:journalctl@ceph.mgr.a.vm01.stdout:Mar 09 20:51:59 vm01 ceph-9cb345a8-1bf9-11f1-a195-0375563c5891-mgr-a[53639]: Your identification has been saved in /tmp/tmpcps51wht/key 2026-03-09T20:51:59.174 INFO:journalctl@ceph.mgr.a.vm01.stdout:Mar 09 20:51:59 vm01 ceph-9cb345a8-1bf9-11f1-a195-0375563c5891-mgr-a[53639]: Your public key has been saved in /tmp/tmpcps51wht/key.pub 2026-03-09T20:51:59.174 INFO:journalctl@ceph.mgr.a.vm01.stdout:Mar 09 20:51:59 vm01 ceph-9cb345a8-1bf9-11f1-a195-0375563c5891-mgr-a[53639]: The key fingerprint is: 2026-03-09T20:51:59.174 INFO:journalctl@ceph.mgr.a.vm01.stdout:Mar 09 20:51:59 vm01 ceph-9cb345a8-1bf9-11f1-a195-0375563c5891-mgr-a[53639]: SHA256:MVjhYo1oZa/0A9YbJnM7dWYwv7E5DBww9h2owGD/d1g ceph-9cb345a8-1bf9-11f1-a195-0375563c5891 2026-03-09T20:51:59.174 INFO:journalctl@ceph.mgr.a.vm01.stdout:Mar 09 20:51:59 vm01 ceph-9cb345a8-1bf9-11f1-a195-0375563c5891-mgr-a[53639]: The key's randomart image is: 2026-03-09T20:51:59.175 INFO:journalctl@ceph.mgr.a.vm01.stdout:Mar 09 20:51:59 vm01 ceph-9cb345a8-1bf9-11f1-a195-0375563c5891-mgr-a[53639]: +--[ED25519 256]--+ 2026-03-09T20:51:59.175 INFO:journalctl@ceph.mgr.a.vm01.stdout:Mar 09 20:51:59 vm01 ceph-9cb345a8-1bf9-11f1-a195-0375563c5891-mgr-a[53639]: | ooo *o+.. | 2026-03-09T20:51:59.175 INFO:journalctl@ceph.mgr.a.vm01.stdout:Mar 09 20:51:59 vm01 ceph-9cb345a8-1bf9-11f1-a195-0375563c5891-mgr-a[53639]: | . =oX +.* . | 2026-03-09T20:51:59.175 INFO:journalctl@ceph.mgr.a.vm01.stdout:Mar 09 20:51:59 vm01 ceph-9cb345a8-1bf9-11f1-a195-0375563c5891-mgr-a[53639]: | o %.@.= E | 2026-03-09T20:51:59.175 INFO:journalctl@ceph.mgr.a.vm01.stdout:Mar 09 20:51:59 vm01 ceph-9cb345a8-1bf9-11f1-a195-0375563c5891-mgr-a[53639]: | . + @.B O = | 2026-03-09T20:51:59.175 INFO:journalctl@ceph.mgr.a.vm01.stdout:Mar 09 20:51:59 vm01 ceph-9cb345a8-1bf9-11f1-a195-0375563c5891-mgr-a[53639]: | . S o B | 2026-03-09T20:51:59.175 INFO:journalctl@ceph.mgr.a.vm01.stdout:Mar 09 20:51:59 vm01 ceph-9cb345a8-1bf9-11f1-a195-0375563c5891-mgr-a[53639]: | + . . | 2026-03-09T20:51:59.175 INFO:journalctl@ceph.mgr.a.vm01.stdout:Mar 09 20:51:59 vm01 ceph-9cb345a8-1bf9-11f1-a195-0375563c5891-mgr-a[53639]: | | 2026-03-09T20:51:59.175 INFO:journalctl@ceph.mgr.a.vm01.stdout:Mar 09 20:51:59 vm01 ceph-9cb345a8-1bf9-11f1-a195-0375563c5891-mgr-a[53639]: | | 2026-03-09T20:51:59.175 INFO:journalctl@ceph.mgr.a.vm01.stdout:Mar 09 20:51:59 vm01 ceph-9cb345a8-1bf9-11f1-a195-0375563c5891-mgr-a[53639]: | | 2026-03-09T20:51:59.175 INFO:journalctl@ceph.mgr.a.vm01.stdout:Mar 09 20:51:59 vm01 ceph-9cb345a8-1bf9-11f1-a195-0375563c5891-mgr-a[53639]: +----[SHA256]-----+ 2026-03-09T20:51:59.446 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:51:59 vm01 ceph-mon[53402]: from='client.14122 -' entity='client.admin' cmd=[{"prefix": "get_command_descriptions"}]: dispatch 2026-03-09T20:51:59.446 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:51:59 vm01 ceph-mon[53402]: from='client.14122 -' entity='client.admin' cmd=[{"prefix": "mgr_status"}]: dispatch 2026-03-09T20:51:59.446 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:51:59 vm01 ceph-mon[53402]: from='client.14130 -' entity='client.admin' cmd=[{"prefix": "orch set backend", "module_name": "cephadm", "target": ["mon-mgr", ""]}]: dispatch 2026-03-09T20:51:59.446 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:51:59 vm01 ceph-mon[53402]: [09/Mar/2026:20:51:58] ENGINE Bus STARTING 2026-03-09T20:51:59.446 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:51:59 vm01 ceph-mon[53402]: from='mgr.14118 192.168.123.101:0/3188408464' entity='mgr.a' 2026-03-09T20:51:59.446 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:51:59 vm01 ceph-mon[53402]: from='mgr.14118 192.168.123.101:0/3188408464' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T20:51:59.446 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:51:59 vm01 ceph-mon[53402]: [09/Mar/2026:20:51:58] ENGINE Serving on https://192.168.123.101:7150 2026-03-09T20:51:59.446 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:51:59 vm01 ceph-mon[53402]: [09/Mar/2026:20:51:58] ENGINE Client ('192.168.123.101', 44204) lost — peer dropped the TLS connection suddenly, during handshake: (6, 'TLS/SSL connection has been closed (EOF) (_ssl.c:1147)') 2026-03-09T20:51:59.446 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:51:59 vm01 ceph-mon[53402]: [09/Mar/2026:20:51:58] ENGINE Serving on http://192.168.123.101:8765 2026-03-09T20:51:59.446 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:51:59 vm01 ceph-mon[53402]: [09/Mar/2026:20:51:58] ENGINE Bus STARTED 2026-03-09T20:51:59.446 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:51:59 vm01 ceph-mon[53402]: from='mgr.14118 192.168.123.101:0/3188408464' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T20:51:59.446 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:51:59 vm01 ceph-mon[53402]: from='client.14132 -' entity='client.admin' cmd=[{"prefix": "cephadm set-user", "user": "root", "target": ["mon-mgr", ""]}]: dispatch 2026-03-09T20:51:59.446 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:51:59 vm01 ceph-mon[53402]: from='mgr.14118 192.168.123.101:0/3188408464' entity='mgr.a' 2026-03-09T20:51:59.446 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:51:59 vm01 ceph-mon[53402]: from='mgr.14118 192.168.123.101:0/3188408464' entity='mgr.a' 2026-03-09T20:51:59.596 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout ssh-ed25519 AAAAC3NzaC1lZDI1NTE5AAAAIIqPEA2Kc+F2uZ8qMyPBvV6OAFu2Rygy2M/FhERcxuCl ceph-9cb345a8-1bf9-11f1-a195-0375563c5891 2026-03-09T20:51:59.596 INFO:teuthology.orchestra.run.vm01.stdout:Wrote public SSH key to /home/ubuntu/cephtest/ceph.pub 2026-03-09T20:51:59.596 INFO:teuthology.orchestra.run.vm01.stdout:Adding key to root@localhost authorized_keys... 2026-03-09T20:51:59.596 INFO:teuthology.orchestra.run.vm01.stdout:Adding host vm01... 2026-03-09T20:52:00.255 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:52:00 vm01 ceph-mon[53402]: from='client.14134 -' entity='client.admin' cmd=[{"prefix": "cephadm generate-key", "target": ["mon-mgr", ""]}]: dispatch 2026-03-09T20:52:00.255 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:52:00 vm01 ceph-mon[53402]: Generating ssh key... 2026-03-09T20:52:00.255 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:52:00 vm01 ceph-mon[53402]: mgrmap e8: a(active, since 2s) 2026-03-09T20:52:00.255 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:52:00 vm01 ceph-mon[53402]: from='client.14136 -' entity='client.admin' cmd=[{"prefix": "cephadm get-pub-key", "target": ["mon-mgr", ""]}]: dispatch 2026-03-09T20:52:01.424 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:52:01 vm01 ceph-mon[53402]: from='client.14138 -' entity='client.admin' cmd=[{"prefix": "orch host add", "hostname": "vm01", "addr": "192.168.123.101", "target": ["mon-mgr", ""]}]: dispatch 2026-03-09T20:52:01.424 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:52:01 vm01 ceph-mon[53402]: Deploying cephadm binary to vm01 2026-03-09T20:52:01.585 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout Added host 'vm01' with addr '192.168.123.101' 2026-03-09T20:52:01.585 INFO:teuthology.orchestra.run.vm01.stdout:Deploying unmanaged mon service... 2026-03-09T20:52:01.977 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout Scheduled mon update... 2026-03-09T20:52:01.977 INFO:teuthology.orchestra.run.vm01.stdout:Deploying unmanaged mgr service... 2026-03-09T20:52:02.352 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout Scheduled mgr update... 2026-03-09T20:52:02.591 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:52:02 vm01 ceph-mon[53402]: from='mgr.14118 192.168.123.101:0/3188408464' entity='mgr.a' 2026-03-09T20:52:02.591 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:52:02 vm01 ceph-mon[53402]: Added host vm01 2026-03-09T20:52:02.591 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:52:02 vm01 ceph-mon[53402]: from='mgr.14118 192.168.123.101:0/3188408464' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T20:52:02.591 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:52:02 vm01 ceph-mon[53402]: from='mgr.14118 192.168.123.101:0/3188408464' entity='mgr.a' 2026-03-09T20:52:02.591 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:52:02 vm01 ceph-mon[53402]: from='mgr.14118 192.168.123.101:0/3188408464' entity='mgr.a' 2026-03-09T20:52:03.123 INFO:teuthology.orchestra.run.vm01.stdout:Enabling the dashboard module... 2026-03-09T20:52:03.875 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:52:03 vm01 ceph-mon[53402]: from='client.14140 -' entity='client.admin' cmd=[{"prefix": "orch apply", "service_type": "mon", "unmanaged": true, "target": ["mon-mgr", ""]}]: dispatch 2026-03-09T20:52:03.876 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:52:03 vm01 ceph-mon[53402]: Saving service mon spec with placement count:5 2026-03-09T20:52:03.876 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:52:03 vm01 ceph-mon[53402]: from='client.14142 -' entity='client.admin' cmd=[{"prefix": "orch apply", "service_type": "mgr", "unmanaged": true, "target": ["mon-mgr", ""]}]: dispatch 2026-03-09T20:52:03.876 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:52:03 vm01 ceph-mon[53402]: Saving service mgr spec with placement count:2 2026-03-09T20:52:03.876 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:52:03 vm01 ceph-mon[53402]: from='client.? 192.168.123.101:0/247036883' entity='client.admin' 2026-03-09T20:52:03.876 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:52:03 vm01 ceph-mon[53402]: from='client.? 192.168.123.101:0/559943007' entity='client.admin' 2026-03-09T20:52:03.876 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:52:03 vm01 ceph-mon[53402]: from='mgr.14118 192.168.123.101:0/3188408464' entity='mgr.a' 2026-03-09T20:52:03.876 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:52:03 vm01 ceph-mon[53402]: from='client.? 192.168.123.101:0/1627249645' entity='client.admin' cmd=[{"prefix": "mgr module enable", "module": "dashboard"}]: dispatch 2026-03-09T20:52:03.876 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:52:03 vm01 ceph-mon[53402]: from='mgr.14118 192.168.123.101:0/3188408464' entity='mgr.a' 2026-03-09T20:52:04.623 INFO:journalctl@ceph.mgr.a.vm01.stdout:Mar 09 20:52:04 vm01 ceph-9cb345a8-1bf9-11f1-a195-0375563c5891-mgr-a[53639]: ignoring --setuser ceph since I am not root 2026-03-09T20:52:04.623 INFO:journalctl@ceph.mgr.a.vm01.stdout:Mar 09 20:52:04 vm01 ceph-9cb345a8-1bf9-11f1-a195-0375563c5891-mgr-a[53639]: ignoring --setgroup ceph since I am not root 2026-03-09T20:52:04.623 INFO:journalctl@ceph.mgr.a.vm01.stdout:Mar 09 20:52:04 vm01 ceph-9cb345a8-1bf9-11f1-a195-0375563c5891-mgr-a[53639]: 2026-03-09T20:52:04.466+0000 7f0c0f461140 -1 mgr[py] Module alerts has missing NOTIFY_TYPES member 2026-03-09T20:52:04.623 INFO:journalctl@ceph.mgr.a.vm01.stdout:Mar 09 20:52:04 vm01 ceph-9cb345a8-1bf9-11f1-a195-0375563c5891-mgr-a[53639]: 2026-03-09T20:52:04.507+0000 7f0c0f461140 -1 mgr[py] Module balancer has missing NOTIFY_TYPES member 2026-03-09T20:52:05.132 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:52:04 vm01 ceph-mon[53402]: from='mgr.14118 192.168.123.101:0/3188408464' entity='mgr.a' 2026-03-09T20:52:05.132 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:52:04 vm01 ceph-mon[53402]: from='client.? 192.168.123.101:0/1627249645' entity='client.admin' cmd='[{"prefix": "mgr module enable", "module": "dashboard"}]': finished 2026-03-09T20:52:05.132 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:52:04 vm01 ceph-mon[53402]: mgrmap e9: a(active, since 7s) 2026-03-09T20:52:05.132 INFO:journalctl@ceph.mgr.a.vm01.stdout:Mar 09 20:52:04 vm01 ceph-9cb345a8-1bf9-11f1-a195-0375563c5891-mgr-a[53639]: 2026-03-09T20:52:04.879+0000 7f0c0f461140 -1 mgr[py] Module crash has missing NOTIFY_TYPES member 2026-03-09T20:52:05.275 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout { 2026-03-09T20:52:05.275 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "epoch": 9, 2026-03-09T20:52:05.275 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "available": true, 2026-03-09T20:52:05.275 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "active_name": "a", 2026-03-09T20:52:05.275 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "num_standby": 0 2026-03-09T20:52:05.275 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout } 2026-03-09T20:52:05.276 INFO:teuthology.orchestra.run.vm01.stdout:Waiting for the mgr to restart... 2026-03-09T20:52:05.276 INFO:teuthology.orchestra.run.vm01.stdout:Waiting for mgr epoch 9... 2026-03-09T20:52:05.382 INFO:journalctl@ceph.mgr.a.vm01.stdout:Mar 09 20:52:05 vm01 ceph-9cb345a8-1bf9-11f1-a195-0375563c5891-mgr-a[53639]: 2026-03-09T20:52:05.207+0000 7f0c0f461140 -1 mgr[py] Module devicehealth has missing NOTIFY_TYPES member 2026-03-09T20:52:05.382 INFO:journalctl@ceph.mgr.a.vm01.stdout:Mar 09 20:52:05 vm01 ceph-9cb345a8-1bf9-11f1-a195-0375563c5891-mgr-a[53639]: /lib64/python3.9/site-packages/scipy/__init__.py:73: UserWarning: NumPy was imported from a Python sub-interpreter but NumPy does not properly support sub-interpreters. This will likely work for most users but might cause hard to track down issues or subtle bugs. A common user of the rare sub-interpreter feature is wsgi which also allows single-interpreter mode. 2026-03-09T20:52:05.382 INFO:journalctl@ceph.mgr.a.vm01.stdout:Mar 09 20:52:05 vm01 ceph-9cb345a8-1bf9-11f1-a195-0375563c5891-mgr-a[53639]: Improvements in the case of bugs are welcome, but is not on the NumPy roadmap, and full support may require significant effort to achieve. 2026-03-09T20:52:05.382 INFO:journalctl@ceph.mgr.a.vm01.stdout:Mar 09 20:52:05 vm01 ceph-9cb345a8-1bf9-11f1-a195-0375563c5891-mgr-a[53639]: from numpy import show_config as show_numpy_config 2026-03-09T20:52:05.382 INFO:journalctl@ceph.mgr.a.vm01.stdout:Mar 09 20:52:05 vm01 ceph-9cb345a8-1bf9-11f1-a195-0375563c5891-mgr-a[53639]: 2026-03-09T20:52:05.299+0000 7f0c0f461140 -1 mgr[py] Module diskprediction_local has missing NOTIFY_TYPES member 2026-03-09T20:52:05.382 INFO:journalctl@ceph.mgr.a.vm01.stdout:Mar 09 20:52:05 vm01 ceph-9cb345a8-1bf9-11f1-a195-0375563c5891-mgr-a[53639]: 2026-03-09T20:52:05.338+0000 7f0c0f461140 -1 mgr[py] Module influx has missing NOTIFY_TYPES member 2026-03-09T20:52:05.674 INFO:journalctl@ceph.mgr.a.vm01.stdout:Mar 09 20:52:05 vm01 ceph-9cb345a8-1bf9-11f1-a195-0375563c5891-mgr-a[53639]: 2026-03-09T20:52:05.415+0000 7f0c0f461140 -1 mgr[py] Module iostat has missing NOTIFY_TYPES member 2026-03-09T20:52:06.160 INFO:journalctl@ceph.mgr.a.vm01.stdout:Mar 09 20:52:05 vm01 ceph-9cb345a8-1bf9-11f1-a195-0375563c5891-mgr-a[53639]: 2026-03-09T20:52:05.894+0000 7f0c0f461140 -1 mgr[py] Module nfs has missing NOTIFY_TYPES member 2026-03-09T20:52:06.160 INFO:journalctl@ceph.mgr.a.vm01.stdout:Mar 09 20:52:06 vm01 ceph-9cb345a8-1bf9-11f1-a195-0375563c5891-mgr-a[53639]: 2026-03-09T20:52:06.005+0000 7f0c0f461140 -1 mgr[py] Module orchestrator has missing NOTIFY_TYPES member 2026-03-09T20:52:06.160 INFO:journalctl@ceph.mgr.a.vm01.stdout:Mar 09 20:52:06 vm01 ceph-9cb345a8-1bf9-11f1-a195-0375563c5891-mgr-a[53639]: 2026-03-09T20:52:06.046+0000 7f0c0f461140 -1 mgr[py] Module osd_perf_query has missing NOTIFY_TYPES member 2026-03-09T20:52:06.160 INFO:journalctl@ceph.mgr.a.vm01.stdout:Mar 09 20:52:06 vm01 ceph-9cb345a8-1bf9-11f1-a195-0375563c5891-mgr-a[53639]: 2026-03-09T20:52:06.082+0000 7f0c0f461140 -1 mgr[py] Module osd_support has missing NOTIFY_TYPES member 2026-03-09T20:52:06.160 INFO:journalctl@ceph.mgr.a.vm01.stdout:Mar 09 20:52:06 vm01 ceph-9cb345a8-1bf9-11f1-a195-0375563c5891-mgr-a[53639]: 2026-03-09T20:52:06.121+0000 7f0c0f461140 -1 mgr[py] Module pg_autoscaler has missing NOTIFY_TYPES member 2026-03-09T20:52:06.160 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:52:05 vm01 ceph-mon[53402]: from='client.? 192.168.123.101:0/3674177876' entity='client.admin' cmd=[{"prefix": "mgr stat"}]: dispatch 2026-03-09T20:52:06.423 INFO:journalctl@ceph.mgr.a.vm01.stdout:Mar 09 20:52:06 vm01 ceph-9cb345a8-1bf9-11f1-a195-0375563c5891-mgr-a[53639]: 2026-03-09T20:52:06.158+0000 7f0c0f461140 -1 mgr[py] Module progress has missing NOTIFY_TYPES member 2026-03-09T20:52:06.424 INFO:journalctl@ceph.mgr.a.vm01.stdout:Mar 09 20:52:06 vm01 ceph-9cb345a8-1bf9-11f1-a195-0375563c5891-mgr-a[53639]: 2026-03-09T20:52:06.331+0000 7f0c0f461140 -1 mgr[py] Module prometheus has missing NOTIFY_TYPES member 2026-03-09T20:52:06.424 INFO:journalctl@ceph.mgr.a.vm01.stdout:Mar 09 20:52:06 vm01 ceph-9cb345a8-1bf9-11f1-a195-0375563c5891-mgr-a[53639]: 2026-03-09T20:52:06.379+0000 7f0c0f461140 -1 mgr[py] Module rbd_support has missing NOTIFY_TYPES member 2026-03-09T20:52:06.859 INFO:journalctl@ceph.mgr.a.vm01.stdout:Mar 09 20:52:06 vm01 ceph-9cb345a8-1bf9-11f1-a195-0375563c5891-mgr-a[53639]: 2026-03-09T20:52:06.590+0000 7f0c0f461140 -1 mgr[py] Module rgw has missing NOTIFY_TYPES member 2026-03-09T20:52:07.117 INFO:journalctl@ceph.mgr.a.vm01.stdout:Mar 09 20:52:06 vm01 ceph-9cb345a8-1bf9-11f1-a195-0375563c5891-mgr-a[53639]: 2026-03-09T20:52:06.858+0000 7f0c0f461140 -1 mgr[py] Module rook has missing NOTIFY_TYPES member 2026-03-09T20:52:07.117 INFO:journalctl@ceph.mgr.a.vm01.stdout:Mar 09 20:52:06 vm01 ceph-9cb345a8-1bf9-11f1-a195-0375563c5891-mgr-a[53639]: 2026-03-09T20:52:06.891+0000 7f0c0f461140 -1 mgr[py] Module selftest has missing NOTIFY_TYPES member 2026-03-09T20:52:07.117 INFO:journalctl@ceph.mgr.a.vm01.stdout:Mar 09 20:52:06 vm01 ceph-9cb345a8-1bf9-11f1-a195-0375563c5891-mgr-a[53639]: 2026-03-09T20:52:06.931+0000 7f0c0f461140 -1 mgr[py] Module snap_schedule has missing NOTIFY_TYPES member 2026-03-09T20:52:07.117 INFO:journalctl@ceph.mgr.a.vm01.stdout:Mar 09 20:52:07 vm01 ceph-9cb345a8-1bf9-11f1-a195-0375563c5891-mgr-a[53639]: 2026-03-09T20:52:07.004+0000 7f0c0f461140 -1 mgr[py] Module status has missing NOTIFY_TYPES member 2026-03-09T20:52:07.117 INFO:journalctl@ceph.mgr.a.vm01.stdout:Mar 09 20:52:07 vm01 ceph-9cb345a8-1bf9-11f1-a195-0375563c5891-mgr-a[53639]: 2026-03-09T20:52:07.038+0000 7f0c0f461140 -1 mgr[py] Module telegraf has missing NOTIFY_TYPES member 2026-03-09T20:52:07.385 INFO:journalctl@ceph.mgr.a.vm01.stdout:Mar 09 20:52:07 vm01 ceph-9cb345a8-1bf9-11f1-a195-0375563c5891-mgr-a[53639]: 2026-03-09T20:52:07.115+0000 7f0c0f461140 -1 mgr[py] Module telemetry has missing NOTIFY_TYPES member 2026-03-09T20:52:07.385 INFO:journalctl@ceph.mgr.a.vm01.stdout:Mar 09 20:52:07 vm01 ceph-9cb345a8-1bf9-11f1-a195-0375563c5891-mgr-a[53639]: 2026-03-09T20:52:07.222+0000 7f0c0f461140 -1 mgr[py] Module test_orchestrator has missing NOTIFY_TYPES member 2026-03-09T20:52:07.385 INFO:journalctl@ceph.mgr.a.vm01.stdout:Mar 09 20:52:07 vm01 ceph-9cb345a8-1bf9-11f1-a195-0375563c5891-mgr-a[53639]: 2026-03-09T20:52:07.350+0000 7f0c0f461140 -1 mgr[py] Module volumes has missing NOTIFY_TYPES member 2026-03-09T20:52:07.673 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:52:07 vm01 ceph-mon[53402]: Active manager daemon a restarted 2026-03-09T20:52:07.673 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:52:07 vm01 ceph-mon[53402]: Activating manager daemon a 2026-03-09T20:52:07.674 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:52:07 vm01 ceph-mon[53402]: osdmap e3: 0 total, 0 up, 0 in 2026-03-09T20:52:07.674 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:52:07 vm01 ceph-mon[53402]: mgrmap e10: a(active, starting, since 0.00634276s) 2026-03-09T20:52:07.674 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:52:07 vm01 ceph-mon[53402]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "a"}]: dispatch 2026-03-09T20:52:07.674 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:52:07 vm01 ceph-mon[53402]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' cmd=[{"prefix": "mgr metadata", "who": "a", "id": "a"}]: dispatch 2026-03-09T20:52:07.674 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:52:07 vm01 ceph-mon[53402]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' cmd=[{"prefix": "mds metadata"}]: dispatch 2026-03-09T20:52:07.674 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:52:07 vm01 ceph-mon[53402]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' cmd=[{"prefix": "osd metadata"}]: dispatch 2026-03-09T20:52:07.674 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:52:07 vm01 ceph-mon[53402]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' cmd=[{"prefix": "mon metadata"}]: dispatch 2026-03-09T20:52:07.674 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:52:07 vm01 ceph-mon[53402]: Manager daemon a is now available 2026-03-09T20:52:07.674 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:52:07 vm01 ceph-mon[53402]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T20:52:07.674 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:52:07 vm01 ceph-mon[53402]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/a/mirror_snapshot_schedule"}]: dispatch 2026-03-09T20:52:07.674 INFO:journalctl@ceph.mgr.a.vm01.stdout:Mar 09 20:52:07 vm01 ceph-9cb345a8-1bf9-11f1-a195-0375563c5891-mgr-a[53639]: 2026-03-09T20:52:07.384+0000 7f0c0f461140 -1 mgr[py] Module zabbix has missing NOTIFY_TYPES member 2026-03-09T20:52:08.586 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout { 2026-03-09T20:52:08.586 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "mgrmap_epoch": 11, 2026-03-09T20:52:08.586 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "initialized": true 2026-03-09T20:52:08.586 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout } 2026-03-09T20:52:08.586 INFO:teuthology.orchestra.run.vm01.stdout:mgr epoch 9 is available 2026-03-09T20:52:08.586 INFO:teuthology.orchestra.run.vm01.stdout:Generating a dashboard self-signed certificate... 2026-03-09T20:52:08.684 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:52:08 vm01 ceph-mon[53402]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/a/trash_purge_schedule"}]: dispatch 2026-03-09T20:52:08.684 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:52:08 vm01 ceph-mon[53402]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' 2026-03-09T20:52:08.684 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:52:08 vm01 ceph-mon[53402]: [09/Mar/2026:20:52:08] ENGINE Bus STARTING 2026-03-09T20:52:08.684 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:52:08 vm01 ceph-mon[53402]: mgrmap e11: a(active, since 1.01907s) 2026-03-09T20:52:08.684 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:52:08 vm01 ceph-mon[53402]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' 2026-03-09T20:52:09.054 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout Self-signed certificate created 2026-03-09T20:52:09.054 INFO:teuthology.orchestra.run.vm01.stdout:Creating initial admin user... 2026-03-09T20:52:09.565 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout {"username": "admin", "password": "$2b$12$AnrMx2kOf558eMu176kbAuURZEweaqaZyGrTsoVVFDvUJsbIELlwy", "roles": ["administrator"], "name": null, "email": null, "lastUpdate": 1773089529, "enabled": true, "pwdExpirationDate": null, "pwdUpdateRequired": true} 2026-03-09T20:52:09.565 INFO:teuthology.orchestra.run.vm01.stdout:Fetching dashboard port number... 2026-03-09T20:52:09.945 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout 8443 2026-03-09T20:52:09.945 INFO:teuthology.orchestra.run.vm01.stdout:firewalld does not appear to be present 2026-03-09T20:52:09.945 INFO:teuthology.orchestra.run.vm01.stdout:Not possible to open ports <[8443]>. firewalld.service is not available 2026-03-09T20:52:09.947 INFO:teuthology.orchestra.run.vm01.stdout:Ceph Dashboard is now available at: 2026-03-09T20:52:09.947 INFO:teuthology.orchestra.run.vm01.stdout: 2026-03-09T20:52:09.947 INFO:teuthology.orchestra.run.vm01.stdout: URL: https://vm01.local:8443/ 2026-03-09T20:52:09.947 INFO:teuthology.orchestra.run.vm01.stdout: User: admin 2026-03-09T20:52:09.947 INFO:teuthology.orchestra.run.vm01.stdout: Password: njxnm5e9er 2026-03-09T20:52:09.947 INFO:teuthology.orchestra.run.vm01.stdout: 2026-03-09T20:52:09.947 INFO:teuthology.orchestra.run.vm01.stdout:Saving cluster configuration to /var/lib/ceph/9cb345a8-1bf9-11f1-a195-0375563c5891/config directory 2026-03-09T20:52:10.064 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:52:09 vm01 ceph-mon[53402]: from='client.14154 -' entity='client.admin' cmd=[{"prefix": "get_command_descriptions"}]: dispatch 2026-03-09T20:52:10.064 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:52:09 vm01 ceph-mon[53402]: from='client.14154 -' entity='client.admin' cmd=[{"prefix": "mgr_status"}]: dispatch 2026-03-09T20:52:10.064 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:52:09 vm01 ceph-mon[53402]: [09/Mar/2026:20:52:08] ENGINE Serving on https://192.168.123.101:7150 2026-03-09T20:52:10.064 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:52:09 vm01 ceph-mon[53402]: [09/Mar/2026:20:52:08] ENGINE Client ('192.168.123.101', 55504) lost — peer dropped the TLS connection suddenly, during handshake: (6, 'TLS/SSL connection has been closed (EOF) (_ssl.c:1147)') 2026-03-09T20:52:10.064 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:52:09 vm01 ceph-mon[53402]: [09/Mar/2026:20:52:08] ENGINE Serving on http://192.168.123.101:8765 2026-03-09T20:52:10.064 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:52:09 vm01 ceph-mon[53402]: [09/Mar/2026:20:52:08] ENGINE Bus STARTED 2026-03-09T20:52:10.064 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:52:09 vm01 ceph-mon[53402]: from='client.14162 -' entity='client.admin' cmd=[{"prefix": "dashboard create-self-signed-cert", "target": ["mon-mgr", ""]}]: dispatch 2026-03-09T20:52:10.064 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:52:09 vm01 ceph-mon[53402]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' 2026-03-09T20:52:10.064 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:52:09 vm01 ceph-mon[53402]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' 2026-03-09T20:52:10.064 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:52:09 vm01 ceph-mon[53402]: from='client.14164 -' entity='client.admin' cmd=[{"prefix": "dashboard ac-user-create", "username": "admin", "rolename": "administrator", "force_password": true, "pwd_update_required": true, "target": ["mon-mgr", ""]}]: dispatch 2026-03-09T20:52:10.064 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:52:09 vm01 ceph-mon[53402]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' 2026-03-09T20:52:10.064 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:52:09 vm01 ceph-mon[53402]: from='client.? 192.168.123.101:0/2122775969' entity='client.admin' cmd=[{"prefix": "config get", "who": "mgr", "key": "mgr/dashboard/ssl_server_port"}]: dispatch 2026-03-09T20:52:10.380 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stderr set mgr/dashboard/cluster/status 2026-03-09T20:52:10.381 INFO:teuthology.orchestra.run.vm01.stdout:You can access the Ceph CLI as following in case of multi-cluster or non-default config: 2026-03-09T20:52:10.381 INFO:teuthology.orchestra.run.vm01.stdout: 2026-03-09T20:52:10.381 INFO:teuthology.orchestra.run.vm01.stdout: sudo /home/ubuntu/cephtest/cephadm shell --fsid 9cb345a8-1bf9-11f1-a195-0375563c5891 -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring 2026-03-09T20:52:10.381 INFO:teuthology.orchestra.run.vm01.stdout: 2026-03-09T20:52:10.381 INFO:teuthology.orchestra.run.vm01.stdout:Or, if you are only running a single cluster on this host: 2026-03-09T20:52:10.381 INFO:teuthology.orchestra.run.vm01.stdout: 2026-03-09T20:52:10.381 INFO:teuthology.orchestra.run.vm01.stdout: sudo /home/ubuntu/cephtest/cephadm shell 2026-03-09T20:52:10.381 INFO:teuthology.orchestra.run.vm01.stdout: 2026-03-09T20:52:10.381 INFO:teuthology.orchestra.run.vm01.stdout:Please consider enabling telemetry to help improve Ceph: 2026-03-09T20:52:10.381 INFO:teuthology.orchestra.run.vm01.stdout: 2026-03-09T20:52:10.381 INFO:teuthology.orchestra.run.vm01.stdout: ceph telemetry on 2026-03-09T20:52:10.381 INFO:teuthology.orchestra.run.vm01.stdout: 2026-03-09T20:52:10.381 INFO:teuthology.orchestra.run.vm01.stdout:For more information see: 2026-03-09T20:52:10.381 INFO:teuthology.orchestra.run.vm01.stdout: 2026-03-09T20:52:10.381 INFO:teuthology.orchestra.run.vm01.stdout: https://docs.ceph.com/en/latest/mgr/telemetry/ 2026-03-09T20:52:10.381 INFO:teuthology.orchestra.run.vm01.stdout: 2026-03-09T20:52:10.381 INFO:teuthology.orchestra.run.vm01.stdout:Bootstrap complete. 2026-03-09T20:52:10.413 INFO:tasks.cephadm:Fetching config... 2026-03-09T20:52:10.413 DEBUG:teuthology.orchestra.run.vm01:> set -ex 2026-03-09T20:52:10.413 DEBUG:teuthology.orchestra.run.vm01:> dd if=/etc/ceph/ceph.conf of=/dev/stdout 2026-03-09T20:52:10.438 INFO:tasks.cephadm:Fetching client.admin keyring... 2026-03-09T20:52:10.438 DEBUG:teuthology.orchestra.run.vm01:> set -ex 2026-03-09T20:52:10.438 DEBUG:teuthology.orchestra.run.vm01:> dd if=/etc/ceph/ceph.client.admin.keyring of=/dev/stdout 2026-03-09T20:52:10.519 INFO:tasks.cephadm:Fetching mon keyring... 2026-03-09T20:52:10.519 DEBUG:teuthology.orchestra.run.vm01:> set -ex 2026-03-09T20:52:10.519 DEBUG:teuthology.orchestra.run.vm01:> sudo dd if=/var/lib/ceph/9cb345a8-1bf9-11f1-a195-0375563c5891/mon.a/keyring of=/dev/stdout 2026-03-09T20:52:10.596 INFO:tasks.cephadm:Fetching pub ssh key... 2026-03-09T20:52:10.596 DEBUG:teuthology.orchestra.run.vm01:> set -ex 2026-03-09T20:52:10.596 DEBUG:teuthology.orchestra.run.vm01:> dd if=/home/ubuntu/cephtest/ceph.pub of=/dev/stdout 2026-03-09T20:52:10.652 INFO:tasks.cephadm:Installing pub ssh key for root users... 2026-03-09T20:52:10.652 DEBUG:teuthology.orchestra.run.vm01:> sudo install -d -m 0700 /root/.ssh && echo 'ssh-ed25519 AAAAC3NzaC1lZDI1NTE5AAAAIIqPEA2Kc+F2uZ8qMyPBvV6OAFu2Rygy2M/FhERcxuCl ceph-9cb345a8-1bf9-11f1-a195-0375563c5891' | sudo tee -a /root/.ssh/authorized_keys && sudo chmod 0600 /root/.ssh/authorized_keys 2026-03-09T20:52:10.752 INFO:teuthology.orchestra.run.vm01.stdout:ssh-ed25519 AAAAC3NzaC1lZDI1NTE5AAAAIIqPEA2Kc+F2uZ8qMyPBvV6OAFu2Rygy2M/FhERcxuCl ceph-9cb345a8-1bf9-11f1-a195-0375563c5891 2026-03-09T20:52:10.774 DEBUG:teuthology.orchestra.run.vm06:> sudo install -d -m 0700 /root/.ssh && echo 'ssh-ed25519 AAAAC3NzaC1lZDI1NTE5AAAAIIqPEA2Kc+F2uZ8qMyPBvV6OAFu2Rygy2M/FhERcxuCl ceph-9cb345a8-1bf9-11f1-a195-0375563c5891' | sudo tee -a /root/.ssh/authorized_keys && sudo chmod 0600 /root/.ssh/authorized_keys 2026-03-09T20:52:10.812 INFO:teuthology.orchestra.run.vm06.stdout:ssh-ed25519 AAAAC3NzaC1lZDI1NTE5AAAAIIqPEA2Kc+F2uZ8qMyPBvV6OAFu2Rygy2M/FhERcxuCl ceph-9cb345a8-1bf9-11f1-a195-0375563c5891 2026-03-09T20:52:10.826 DEBUG:teuthology.orchestra.run.vm08:> sudo install -d -m 0700 /root/.ssh && echo 'ssh-ed25519 AAAAC3NzaC1lZDI1NTE5AAAAIIqPEA2Kc+F2uZ8qMyPBvV6OAFu2Rygy2M/FhERcxuCl ceph-9cb345a8-1bf9-11f1-a195-0375563c5891' | sudo tee -a /root/.ssh/authorized_keys && sudo chmod 0600 /root/.ssh/authorized_keys 2026-03-09T20:52:10.864 INFO:teuthology.orchestra.run.vm08.stdout:ssh-ed25519 AAAAC3NzaC1lZDI1NTE5AAAAIIqPEA2Kc+F2uZ8qMyPBvV6OAFu2Rygy2M/FhERcxuCl ceph-9cb345a8-1bf9-11f1-a195-0375563c5891 2026-03-09T20:52:10.876 DEBUG:teuthology.orchestra.run.vm01:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 9cb345a8-1bf9-11f1-a195-0375563c5891 -- ceph config set mgr mgr/cephadm/allow_ptrace true 2026-03-09T20:52:11.072 INFO:teuthology.orchestra.run.vm01.stderr:Inferring config /var/lib/ceph/9cb345a8-1bf9-11f1-a195-0375563c5891/mon.a/config 2026-03-09T20:52:11.106 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:52:10 vm01 ceph-mon[53402]: mgrmap e12: a(active, since 2s) 2026-03-09T20:52:11.106 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:52:10 vm01 ceph-mon[53402]: from='client.? 192.168.123.101:0/1503767849' entity='client.admin' 2026-03-09T20:52:11.527 INFO:tasks.cephadm:Distributing conf and client.admin keyring to all hosts + 0755 2026-03-09T20:52:11.527 DEBUG:teuthology.orchestra.run.vm01:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 9cb345a8-1bf9-11f1-a195-0375563c5891 -- ceph orch client-keyring set client.admin '*' --mode 0755 2026-03-09T20:52:11.710 INFO:teuthology.orchestra.run.vm01.stderr:Inferring config /var/lib/ceph/9cb345a8-1bf9-11f1-a195-0375563c5891/mon.a/config 2026-03-09T20:52:12.115 INFO:tasks.cephadm:Writing (initial) conf and keyring to vm06 2026-03-09T20:52:12.115 DEBUG:teuthology.orchestra.run.vm06:> set -ex 2026-03-09T20:52:12.115 DEBUG:teuthology.orchestra.run.vm06:> dd of=/etc/ceph/ceph.conf 2026-03-09T20:52:12.132 DEBUG:teuthology.orchestra.run.vm06:> set -ex 2026-03-09T20:52:12.132 DEBUG:teuthology.orchestra.run.vm06:> dd of=/etc/ceph/ceph.client.admin.keyring 2026-03-09T20:52:12.190 INFO:tasks.cephadm:Adding host vm06 to orchestrator... 2026-03-09T20:52:12.190 DEBUG:teuthology.orchestra.run.vm01:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 9cb345a8-1bf9-11f1-a195-0375563c5891 -- ceph orch host add vm06 2026-03-09T20:52:12.341 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:52:12 vm01 ceph-mon[53402]: from='client.? 192.168.123.101:0/2339663519' entity='client.admin' 2026-03-09T20:52:12.341 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:52:12 vm01 ceph-mon[53402]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' 2026-03-09T20:52:12.341 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:52:12 vm01 ceph-mon[53402]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' 2026-03-09T20:52:12.341 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:52:12 vm01 ceph-mon[53402]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' cmd=[{"prefix": "config rm", "who": "osd/host:vm01", "name": "osd_memory_target"}]: dispatch 2026-03-09T20:52:12.341 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:52:12 vm01 ceph-mon[53402]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' 2026-03-09T20:52:12.341 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:52:12 vm01 ceph-mon[53402]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T20:52:12.341 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:52:12 vm01 ceph-mon[53402]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' 2026-03-09T20:52:12.341 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:52:12 vm01 ceph-mon[53402]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' 2026-03-09T20:52:12.341 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:52:12 vm01 ceph-mon[53402]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T20:52:12.341 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:52:12 vm01 ceph-mon[53402]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T20:52:12.341 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:52:12 vm01 ceph-mon[53402]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T20:52:12.374 INFO:teuthology.orchestra.run.vm01.stderr:Inferring config /var/lib/ceph/9cb345a8-1bf9-11f1-a195-0375563c5891/mon.a/config 2026-03-09T20:52:13.674 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:52:13 vm01 ceph-mon[53402]: from='client.14172 -' entity='client.admin' cmd=[{"prefix": "orch client-keyring set", "entity": "client.admin", "placement": "*", "mode": "0755", "target": ["mon-mgr", ""]}]: dispatch 2026-03-09T20:52:13.674 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:52:13 vm01 ceph-mon[53402]: Updating vm01:/etc/ceph/ceph.conf 2026-03-09T20:52:13.674 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:52:13 vm01 ceph-mon[53402]: Updating vm01:/var/lib/ceph/9cb345a8-1bf9-11f1-a195-0375563c5891/config/ceph.conf 2026-03-09T20:52:13.674 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:52:13 vm01 ceph-mon[53402]: Updating vm01:/etc/ceph/ceph.client.admin.keyring 2026-03-09T20:52:13.674 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:52:13 vm01 ceph-mon[53402]: Updating vm01:/var/lib/ceph/9cb345a8-1bf9-11f1-a195-0375563c5891/config/ceph.client.admin.keyring 2026-03-09T20:52:13.674 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:52:13 vm01 ceph-mon[53402]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' 2026-03-09T20:52:13.674 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:52:13 vm01 ceph-mon[53402]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' 2026-03-09T20:52:13.674 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:52:13 vm01 ceph-mon[53402]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' 2026-03-09T20:52:14.140 INFO:teuthology.orchestra.run.vm01.stdout:Added host 'vm06' with addr '192.168.123.106' 2026-03-09T20:52:14.296 DEBUG:teuthology.orchestra.run.vm01:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 9cb345a8-1bf9-11f1-a195-0375563c5891 -- ceph orch host ls --format=json 2026-03-09T20:52:14.424 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:52:14 vm01 ceph-mon[53402]: from='client.14174 -' entity='client.admin' cmd=[{"prefix": "orch host add", "hostname": "vm06", "target": ["mon-mgr", ""]}]: dispatch 2026-03-09T20:52:14.424 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:52:14 vm01 ceph-mon[53402]: Deploying cephadm binary to vm06 2026-03-09T20:52:14.424 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:52:14 vm01 ceph-mon[53402]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' 2026-03-09T20:52:14.424 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:52:14 vm01 ceph-mon[53402]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T20:52:14.424 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:52:14 vm01 ceph-mon[53402]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' 2026-03-09T20:52:14.482 INFO:teuthology.orchestra.run.vm01.stderr:Inferring config /var/lib/ceph/9cb345a8-1bf9-11f1-a195-0375563c5891/mon.a/config 2026-03-09T20:52:14.718 INFO:teuthology.orchestra.run.vm01.stdout: 2026-03-09T20:52:14.718 INFO:teuthology.orchestra.run.vm01.stdout:[{"addr": "192.168.123.101", "hostname": "vm01", "labels": [], "status": ""}, {"addr": "192.168.123.106", "hostname": "vm06", "labels": [], "status": ""}] 2026-03-09T20:52:14.883 INFO:tasks.cephadm:Writing (initial) conf and keyring to vm08 2026-03-09T20:52:14.883 DEBUG:teuthology.orchestra.run.vm08:> set -ex 2026-03-09T20:52:14.883 DEBUG:teuthology.orchestra.run.vm08:> dd of=/etc/ceph/ceph.conf 2026-03-09T20:52:14.897 DEBUG:teuthology.orchestra.run.vm08:> set -ex 2026-03-09T20:52:14.897 DEBUG:teuthology.orchestra.run.vm08:> dd of=/etc/ceph/ceph.client.admin.keyring 2026-03-09T20:52:14.952 INFO:tasks.cephadm:Adding host vm08 to orchestrator... 2026-03-09T20:52:14.952 DEBUG:teuthology.orchestra.run.vm01:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 9cb345a8-1bf9-11f1-a195-0375563c5891 -- ceph orch host add vm08 2026-03-09T20:52:15.112 INFO:teuthology.orchestra.run.vm01.stderr:Inferring config /var/lib/ceph/9cb345a8-1bf9-11f1-a195-0375563c5891/mon.a/config 2026-03-09T20:52:15.424 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:52:15 vm01 ceph-mon[53402]: Added host vm06 2026-03-09T20:52:15.424 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:52:15 vm01 ceph-mon[53402]: mgrmap e13: a(active, since 6s) 2026-03-09T20:52:15.424 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:52:15 vm01 ceph-mon[53402]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' 2026-03-09T20:52:16.674 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:52:16 vm01 ceph-mon[53402]: from='client.14176 -' entity='client.admin' cmd=[{"prefix": "orch host ls", "target": ["mon-mgr", ""], "format": "json"}]: dispatch 2026-03-09T20:52:16.674 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:52:16 vm01 ceph-mon[53402]: from='client.14178 -' entity='client.admin' cmd=[{"prefix": "orch host add", "hostname": "vm08", "target": ["mon-mgr", ""]}]: dispatch 2026-03-09T20:52:16.674 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:52:16 vm01 ceph-mon[53402]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' 2026-03-09T20:52:16.871 INFO:teuthology.orchestra.run.vm01.stdout:Added host 'vm08' with addr '192.168.123.108' 2026-03-09T20:52:17.031 DEBUG:teuthology.orchestra.run.vm01:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 9cb345a8-1bf9-11f1-a195-0375563c5891 -- ceph orch host ls --format=json 2026-03-09T20:52:17.215 INFO:teuthology.orchestra.run.vm01.stderr:Inferring config /var/lib/ceph/9cb345a8-1bf9-11f1-a195-0375563c5891/mon.a/config 2026-03-09T20:52:17.484 INFO:teuthology.orchestra.run.vm01.stdout: 2026-03-09T20:52:17.484 INFO:teuthology.orchestra.run.vm01.stdout:[{"addr": "192.168.123.101", "hostname": "vm01", "labels": [], "status": ""}, {"addr": "192.168.123.106", "hostname": "vm06", "labels": [], "status": ""}, {"addr": "192.168.123.108", "hostname": "vm08", "labels": [], "status": ""}] 2026-03-09T20:52:17.608 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:52:17 vm01 ceph-mon[53402]: Deploying cephadm binary to vm08 2026-03-09T20:52:17.608 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:52:17 vm01 ceph-mon[53402]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' 2026-03-09T20:52:17.608 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:52:17 vm01 ceph-mon[53402]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' 2026-03-09T20:52:17.608 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:52:17 vm01 ceph-mon[53402]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' 2026-03-09T20:52:17.608 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:52:17 vm01 ceph-mon[53402]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' 2026-03-09T20:52:17.608 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:52:17 vm01 ceph-mon[53402]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' 2026-03-09T20:52:17.608 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:52:17 vm01 ceph-mon[53402]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' cmd=[{"prefix": "config rm", "who": "osd/host:vm06", "name": "osd_memory_target"}]: dispatch 2026-03-09T20:52:17.608 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:52:17 vm01 ceph-mon[53402]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T20:52:17.608 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:52:17 vm01 ceph-mon[53402]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T20:52:17.636 INFO:tasks.cephadm:Setting crush tunables to default 2026-03-09T20:52:17.636 DEBUG:teuthology.orchestra.run.vm01:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 9cb345a8-1bf9-11f1-a195-0375563c5891 -- ceph osd crush tunables default 2026-03-09T20:52:17.804 INFO:teuthology.orchestra.run.vm01.stderr:Inferring config /var/lib/ceph/9cb345a8-1bf9-11f1-a195-0375563c5891/mon.a/config 2026-03-09T20:52:18.674 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:52:18 vm01 ceph-mon[53402]: Added host vm08 2026-03-09T20:52:18.674 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:52:18 vm01 ceph-mon[53402]: Updating vm06:/etc/ceph/ceph.conf 2026-03-09T20:52:18.674 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:52:18 vm01 ceph-mon[53402]: Updating vm06:/var/lib/ceph/9cb345a8-1bf9-11f1-a195-0375563c5891/config/ceph.conf 2026-03-09T20:52:18.674 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:52:18 vm01 ceph-mon[53402]: Updating vm06:/etc/ceph/ceph.client.admin.keyring 2026-03-09T20:52:18.674 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:52:18 vm01 ceph-mon[53402]: Updating vm06:/var/lib/ceph/9cb345a8-1bf9-11f1-a195-0375563c5891/config/ceph.client.admin.keyring 2026-03-09T20:52:18.674 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:52:18 vm01 ceph-mon[53402]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' 2026-03-09T20:52:18.674 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:52:18 vm01 ceph-mon[53402]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' 2026-03-09T20:52:18.674 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:52:18 vm01 ceph-mon[53402]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' 2026-03-09T20:52:18.674 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:52:18 vm01 ceph-mon[53402]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T20:52:18.674 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:52:18 vm01 ceph-mon[53402]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' 2026-03-09T20:52:18.674 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:52:18 vm01 ceph-mon[53402]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' 2026-03-09T20:52:18.674 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:52:18 vm01 ceph-mon[53402]: from='client.? 192.168.123.101:0/4290406823' entity='client.admin' cmd=[{"prefix": "osd crush tunables", "profile": "default"}]: dispatch 2026-03-09T20:52:18.883 INFO:teuthology.orchestra.run.vm01.stderr:adjusted tunables profile to default 2026-03-09T20:52:19.039 INFO:tasks.cephadm:Adding mon.a on vm01 2026-03-09T20:52:19.039 INFO:tasks.cephadm:Adding mon.b on vm06 2026-03-09T20:52:19.039 INFO:tasks.cephadm:Adding mon.c on vm08 2026-03-09T20:52:19.040 DEBUG:teuthology.orchestra.run.vm08:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 9cb345a8-1bf9-11f1-a195-0375563c5891 -- ceph orch apply mon '3;vm01:192.168.123.101=a;vm06:192.168.123.106=b;vm08:192.168.123.108=c' 2026-03-09T20:52:19.209 INFO:teuthology.orchestra.run.vm08.stderr:Inferring config /etc/ceph/ceph.conf 2026-03-09T20:52:19.258 INFO:teuthology.orchestra.run.vm08.stderr:Inferring config /etc/ceph/ceph.conf 2026-03-09T20:52:19.533 INFO:teuthology.orchestra.run.vm08.stdout:Scheduled mon update... 2026-03-09T20:52:19.674 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:52:19 vm01 ceph-mon[53402]: from='client.14180 -' entity='client.admin' cmd=[{"prefix": "orch host ls", "target": ["mon-mgr", ""], "format": "json"}]: dispatch 2026-03-09T20:52:19.674 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:52:19 vm01 ceph-mon[53402]: from='client.? 192.168.123.101:0/4290406823' entity='client.admin' cmd='[{"prefix": "osd crush tunables", "profile": "default"}]': finished 2026-03-09T20:52:19.674 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:52:19 vm01 ceph-mon[53402]: osdmap e4: 0 total, 0 up, 0 in 2026-03-09T20:52:19.674 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:52:19 vm01 ceph-mon[53402]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' 2026-03-09T20:52:19.697 DEBUG:teuthology.orchestra.run.vm06:mon.b> sudo journalctl -f -n 0 -u ceph-9cb345a8-1bf9-11f1-a195-0375563c5891@mon.b.service 2026-03-09T20:52:19.699 DEBUG:teuthology.orchestra.run.vm08:mon.c> sudo journalctl -f -n 0 -u ceph-9cb345a8-1bf9-11f1-a195-0375563c5891@mon.c.service 2026-03-09T20:52:19.700 INFO:tasks.cephadm:Waiting for 3 mons in monmap... 2026-03-09T20:52:19.700 DEBUG:teuthology.orchestra.run.vm08:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 9cb345a8-1bf9-11f1-a195-0375563c5891 -- ceph mon dump -f json 2026-03-09T20:52:19.920 INFO:teuthology.orchestra.run.vm08.stderr:Inferring config /etc/ceph/ceph.conf 2026-03-09T20:52:19.962 INFO:teuthology.orchestra.run.vm08.stderr:Inferring config /etc/ceph/ceph.conf 2026-03-09T20:52:20.239 INFO:teuthology.orchestra.run.vm08.stdout: 2026-03-09T20:52:20.239 INFO:teuthology.orchestra.run.vm08.stdout:{"epoch":1,"fsid":"9cb345a8-1bf9-11f1-a195-0375563c5891","modified":"2026-03-09T20:51:41.312920Z","created":"2026-03-09T20:51:41.312920Z","min_mon_release":19,"min_mon_release_name":"squid","election_strategy":1,"disallowed_leaders":"","stretch_mode":false,"tiebreaker_mon":"","removed_ranks":"","features":{"persistent":["kraken","luminous","mimic","osdmap-prune","nautilus","octopus","pacific","elector-pinging","quincy","reef","squid"],"optional":[]},"mons":[{"rank":0,"name":"a","public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.101:3300","nonce":0},{"type":"v1","addr":"192.168.123.101:6789","nonce":0}]},"addr":"192.168.123.101:6789/0","public_addr":"192.168.123.101:6789/0","priority":0,"weight":0,"crush_location":"{}"}],"quorum":[0]} 2026-03-09T20:52:20.239 INFO:teuthology.orchestra.run.vm08.stderr:dumped monmap epoch 1 2026-03-09T20:52:20.923 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:52:20 vm01 ceph-mon[53402]: from='client.14184 -' entity='client.admin' cmd=[{"prefix": "orch apply", "service_type": "mon", "placement": "3;vm01:192.168.123.101=a;vm06:192.168.123.106=b;vm08:192.168.123.108=c", "target": ["mon-mgr", ""]}]: dispatch 2026-03-09T20:52:20.924 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:52:20 vm01 ceph-mon[53402]: Saving service mon spec with placement vm01:192.168.123.101=a;vm06:192.168.123.106=b;vm08:192.168.123.108=c;count:3 2026-03-09T20:52:20.924 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:52:20 vm01 ceph-mon[53402]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' 2026-03-09T20:52:20.924 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:52:20 vm01 ceph-mon[53402]: from='client.? 192.168.123.108:0/1564587413' entity='client.admin' cmd=[{"prefix": "mon dump", "format": "json"}]: dispatch 2026-03-09T20:52:21.387 INFO:tasks.cephadm:Waiting for 3 mons in monmap... 2026-03-09T20:52:21.397 DEBUG:teuthology.orchestra.run.vm08:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 9cb345a8-1bf9-11f1-a195-0375563c5891 -- ceph mon dump -f json 2026-03-09T20:52:21.780 INFO:teuthology.orchestra.run.vm08.stderr:Inferring config /var/lib/ceph/9cb345a8-1bf9-11f1-a195-0375563c5891/config/ceph.conf 2026-03-09T20:52:21.924 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:52:21 vm01 ceph-mon[53402]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' 2026-03-09T20:52:21.924 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:52:21 vm01 ceph-mon[53402]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' 2026-03-09T20:52:21.924 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:52:21 vm01 ceph-mon[53402]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' 2026-03-09T20:52:21.924 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:52:21 vm01 ceph-mon[53402]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' 2026-03-09T20:52:21.924 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:52:21 vm01 ceph-mon[53402]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' cmd=[{"prefix": "config rm", "who": "osd/host:vm08", "name": "osd_memory_target"}]: dispatch 2026-03-09T20:52:21.924 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:52:21 vm01 ceph-mon[53402]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T20:52:21.924 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:52:21 vm01 ceph-mon[53402]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T20:52:21.924 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:52:21 vm01 ceph-mon[53402]: Updating vm08:/etc/ceph/ceph.conf 2026-03-09T20:52:21.924 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:52:21 vm01 ceph-mon[53402]: Updating vm08:/var/lib/ceph/9cb345a8-1bf9-11f1-a195-0375563c5891/config/ceph.conf 2026-03-09T20:52:21.924 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:52:21 vm01 ceph-mon[53402]: Updating vm08:/etc/ceph/ceph.client.admin.keyring 2026-03-09T20:52:21.924 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:52:21 vm01 ceph-mon[53402]: Updating vm08:/var/lib/ceph/9cb345a8-1bf9-11f1-a195-0375563c5891/config/ceph.client.admin.keyring 2026-03-09T20:52:21.924 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:52:21 vm01 ceph-mon[53402]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' 2026-03-09T20:52:21.924 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:52:21 vm01 ceph-mon[53402]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' 2026-03-09T20:52:21.924 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:52:21 vm01 ceph-mon[53402]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' 2026-03-09T20:52:21.924 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:52:21 vm01 ceph-mon[53402]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "mon."}]: dispatch 2026-03-09T20:52:21.924 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:52:21 vm01 ceph-mon[53402]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T20:52:21.924 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:52:21 vm01 ceph-mon[53402]: Deploying daemon mon.c on vm08 2026-03-09T20:52:22.066 INFO:teuthology.orchestra.run.vm08.stdout: 2026-03-09T20:52:22.066 INFO:teuthology.orchestra.run.vm08.stdout:{"epoch":1,"fsid":"9cb345a8-1bf9-11f1-a195-0375563c5891","modified":"2026-03-09T20:51:41.312920Z","created":"2026-03-09T20:51:41.312920Z","min_mon_release":19,"min_mon_release_name":"squid","election_strategy":1,"disallowed_leaders":"","stretch_mode":false,"tiebreaker_mon":"","removed_ranks":"","features":{"persistent":["kraken","luminous","mimic","osdmap-prune","nautilus","octopus","pacific","elector-pinging","quincy","reef","squid"],"optional":[]},"mons":[{"rank":0,"name":"a","public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.101:3300","nonce":0},{"type":"v1","addr":"192.168.123.101:6789","nonce":0}]},"addr":"192.168.123.101:6789/0","public_addr":"192.168.123.101:6789/0","priority":0,"weight":0,"crush_location":"{}"}],"quorum":[0]} 2026-03-09T20:52:22.066 INFO:teuthology.orchestra.run.vm08.stderr:dumped monmap epoch 1 2026-03-09T20:52:22.924 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:52:22 vm01 ceph-mon[53402]: from='client.? 192.168.123.108:0/648178151' entity='client.admin' cmd=[{"prefix": "mon dump", "format": "json"}]: dispatch 2026-03-09T20:52:23.207 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:52:23 vm08 ceph-mon[54611]: mon.c@-1(synchronizing).mgr e13 mkfs or daemon transitioned to available, loading commands 2026-03-09T20:52:23.229 INFO:tasks.cephadm:Waiting for 3 mons in monmap... 2026-03-09T20:52:23.229 DEBUG:teuthology.orchestra.run.vm08:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 9cb345a8-1bf9-11f1-a195-0375563c5891 -- ceph mon dump -f json 2026-03-09T20:52:23.397 INFO:teuthology.orchestra.run.vm08.stderr:Inferring config /var/lib/ceph/9cb345a8-1bf9-11f1-a195-0375563c5891/mon.c/config 2026-03-09T20:52:24.796 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 09 20:52:24 vm06 ceph-mon[52939]: mon.b@-1(synchronizing).paxosservice(auth 1..3) refresh upgraded, format 0 -> 3 2026-03-09T20:52:28.362 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:52:28 vm08 ceph-mon[54611]: Deploying daemon mon.b on vm06 2026-03-09T20:52:28.362 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:52:28 vm08 ceph-mon[54611]: mon.a calling monitor election 2026-03-09T20:52:28.362 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:52:28 vm08 ceph-mon[54611]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "a"}]: dispatch 2026-03-09T20:52:28.362 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:52:28 vm08 ceph-mon[54611]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "c"}]: dispatch 2026-03-09T20:52:28.362 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:52:28 vm08 ceph-mon[54611]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "c"}]: dispatch 2026-03-09T20:52:28.362 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:52:28 vm08 ceph-mon[54611]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-09T20:52:28.362 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:52:28 vm08 ceph-mon[54611]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "c"}]: dispatch 2026-03-09T20:52:28.362 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:52:28 vm08 ceph-mon[54611]: mon.c calling monitor election 2026-03-09T20:52:28.362 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:52:28 vm08 ceph-mon[54611]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-09T20:52:28.362 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:52:28 vm08 ceph-mon[54611]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "c"}]: dispatch 2026-03-09T20:52:28.362 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:52:28 vm08 ceph-mon[54611]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-09T20:52:28.362 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:52:28 vm08 ceph-mon[54611]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "c"}]: dispatch 2026-03-09T20:52:28.362 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:52:28 vm08 ceph-mon[54611]: pgmap v4: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-09T20:52:28.362 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:52:28 vm08 ceph-mon[54611]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-09T20:52:28.362 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:52:28 vm08 ceph-mon[54611]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "c"}]: dispatch 2026-03-09T20:52:28.362 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:52:28 vm08 ceph-mon[54611]: mon.a is new leader, mons a,c in quorum (ranks 0,1) 2026-03-09T20:52:28.362 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:52:28 vm08 ceph-mon[54611]: monmap epoch 2 2026-03-09T20:52:28.362 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:52:28 vm08 ceph-mon[54611]: fsid 9cb345a8-1bf9-11f1-a195-0375563c5891 2026-03-09T20:52:28.362 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:52:28 vm08 ceph-mon[54611]: last_changed 2026-03-09T20:52:23.012475+0000 2026-03-09T20:52:28.362 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:52:28 vm08 ceph-mon[54611]: created 2026-03-09T20:51:41.312920+0000 2026-03-09T20:52:28.362 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:52:28 vm08 ceph-mon[54611]: min_mon_release 19 (squid) 2026-03-09T20:52:28.362 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:52:28 vm08 ceph-mon[54611]: election_strategy: 1 2026-03-09T20:52:28.362 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:52:28 vm08 ceph-mon[54611]: 0: [v2:192.168.123.101:3300/0,v1:192.168.123.101:6789/0] mon.a 2026-03-09T20:52:28.362 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:52:28 vm08 ceph-mon[54611]: 1: [v2:192.168.123.108:3300/0,v1:192.168.123.108:6789/0] mon.c 2026-03-09T20:52:28.362 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:52:28 vm08 ceph-mon[54611]: fsmap 2026-03-09T20:52:28.362 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:52:28 vm08 ceph-mon[54611]: osdmap e4: 0 total, 0 up, 0 in 2026-03-09T20:52:28.362 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:52:28 vm08 ceph-mon[54611]: mgrmap e13: a(active, since 20s) 2026-03-09T20:52:28.362 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:52:28 vm08 ceph-mon[54611]: overall HEALTH_OK 2026-03-09T20:52:28.362 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:52:28 vm08 ceph-mon[54611]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' 2026-03-09T20:52:28.362 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:52:28 vm08 ceph-mon[54611]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' 2026-03-09T20:52:28.362 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:52:28 vm08 ceph-mon[54611]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' 2026-03-09T20:52:28.362 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:52:28 vm08 ceph-mon[54611]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' 2026-03-09T20:52:28.424 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:52:28 vm01 ceph-mon[53402]: Deploying daemon mon.b on vm06 2026-03-09T20:52:28.424 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:52:28 vm01 ceph-mon[53402]: mon.a calling monitor election 2026-03-09T20:52:28.424 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:52:28 vm01 ceph-mon[53402]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "a"}]: dispatch 2026-03-09T20:52:28.424 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:52:28 vm01 ceph-mon[53402]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "c"}]: dispatch 2026-03-09T20:52:28.424 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:52:28 vm01 ceph-mon[53402]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "c"}]: dispatch 2026-03-09T20:52:28.424 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:52:28 vm01 ceph-mon[53402]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-09T20:52:28.424 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:52:28 vm01 ceph-mon[53402]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "c"}]: dispatch 2026-03-09T20:52:28.424 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:52:28 vm01 ceph-mon[53402]: mon.c calling monitor election 2026-03-09T20:52:28.424 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:52:28 vm01 ceph-mon[53402]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-09T20:52:28.424 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:52:28 vm01 ceph-mon[53402]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "c"}]: dispatch 2026-03-09T20:52:28.424 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:52:28 vm01 ceph-mon[53402]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-09T20:52:28.424 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:52:28 vm01 ceph-mon[53402]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "c"}]: dispatch 2026-03-09T20:52:28.424 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:52:28 vm01 ceph-mon[53402]: pgmap v4: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-09T20:52:28.424 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:52:28 vm01 ceph-mon[53402]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-09T20:52:28.424 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:52:28 vm01 ceph-mon[53402]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "c"}]: dispatch 2026-03-09T20:52:28.424 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:52:28 vm01 ceph-mon[53402]: mon.a is new leader, mons a,c in quorum (ranks 0,1) 2026-03-09T20:52:28.424 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:52:28 vm01 ceph-mon[53402]: monmap epoch 2 2026-03-09T20:52:28.424 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:52:28 vm01 ceph-mon[53402]: fsid 9cb345a8-1bf9-11f1-a195-0375563c5891 2026-03-09T20:52:28.425 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:52:28 vm01 ceph-mon[53402]: last_changed 2026-03-09T20:52:23.012475+0000 2026-03-09T20:52:28.425 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:52:28 vm01 ceph-mon[53402]: created 2026-03-09T20:51:41.312920+0000 2026-03-09T20:52:28.425 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:52:28 vm01 ceph-mon[53402]: min_mon_release 19 (squid) 2026-03-09T20:52:28.425 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:52:28 vm01 ceph-mon[53402]: election_strategy: 1 2026-03-09T20:52:28.425 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:52:28 vm01 ceph-mon[53402]: 0: [v2:192.168.123.101:3300/0,v1:192.168.123.101:6789/0] mon.a 2026-03-09T20:52:28.425 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:52:28 vm01 ceph-mon[53402]: 1: [v2:192.168.123.108:3300/0,v1:192.168.123.108:6789/0] mon.c 2026-03-09T20:52:28.425 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:52:28 vm01 ceph-mon[53402]: fsmap 2026-03-09T20:52:28.425 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:52:28 vm01 ceph-mon[53402]: osdmap e4: 0 total, 0 up, 0 in 2026-03-09T20:52:28.425 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:52:28 vm01 ceph-mon[53402]: mgrmap e13: a(active, since 20s) 2026-03-09T20:52:28.425 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:52:28 vm01 ceph-mon[53402]: overall HEALTH_OK 2026-03-09T20:52:28.425 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:52:28 vm01 ceph-mon[53402]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' 2026-03-09T20:52:28.425 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:52:28 vm01 ceph-mon[53402]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' 2026-03-09T20:52:28.425 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:52:28 vm01 ceph-mon[53402]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' 2026-03-09T20:52:28.425 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:52:28 vm01 ceph-mon[53402]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' 2026-03-09T20:52:29.423 INFO:journalctl@ceph.mgr.a.vm01.stdout:Mar 09 20:52:29 vm01 ceph-9cb345a8-1bf9-11f1-a195-0375563c5891-mgr-a[53639]: 2026-03-09T20:52:29.012+0000 7f0bdb7c1640 -1 mgr.server handle_report got status from non-daemon mon.c 2026-03-09T20:52:31.046 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 09 20:52:30 vm06 ceph-mon[52939]: Deploying daemon mon.b on vm06 2026-03-09T20:52:31.046 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 09 20:52:30 vm06 ceph-mon[52939]: mon.a calling monitor election 2026-03-09T20:52:31.046 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 09 20:52:30 vm06 ceph-mon[52939]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "a"}]: dispatch 2026-03-09T20:52:31.046 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 09 20:52:30 vm06 ceph-mon[52939]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "c"}]: dispatch 2026-03-09T20:52:31.046 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 09 20:52:30 vm06 ceph-mon[52939]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "c"}]: dispatch 2026-03-09T20:52:31.046 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 09 20:52:30 vm06 ceph-mon[52939]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-09T20:52:31.046 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 09 20:52:30 vm06 ceph-mon[52939]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "c"}]: dispatch 2026-03-09T20:52:31.046 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 09 20:52:30 vm06 ceph-mon[52939]: mon.c calling monitor election 2026-03-09T20:52:31.046 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 09 20:52:30 vm06 ceph-mon[52939]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-09T20:52:31.046 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 09 20:52:30 vm06 ceph-mon[52939]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "c"}]: dispatch 2026-03-09T20:52:31.046 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 09 20:52:30 vm06 ceph-mon[52939]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-09T20:52:31.046 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 09 20:52:30 vm06 ceph-mon[52939]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "c"}]: dispatch 2026-03-09T20:52:31.046 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 09 20:52:30 vm06 ceph-mon[52939]: pgmap v4: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-09T20:52:31.046 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 09 20:52:30 vm06 ceph-mon[52939]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-09T20:52:31.046 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 09 20:52:30 vm06 ceph-mon[52939]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "c"}]: dispatch 2026-03-09T20:52:31.046 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 09 20:52:30 vm06 ceph-mon[52939]: mon.a is new leader, mons a,c in quorum (ranks 0,1) 2026-03-09T20:52:31.046 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 09 20:52:30 vm06 ceph-mon[52939]: monmap epoch 2 2026-03-09T20:52:31.046 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 09 20:52:30 vm06 ceph-mon[52939]: fsid 9cb345a8-1bf9-11f1-a195-0375563c5891 2026-03-09T20:52:31.046 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 09 20:52:30 vm06 ceph-mon[52939]: last_changed 2026-03-09T20:52:23.012475+0000 2026-03-09T20:52:31.046 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 09 20:52:30 vm06 ceph-mon[52939]: created 2026-03-09T20:51:41.312920+0000 2026-03-09T20:52:31.046 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 09 20:52:30 vm06 ceph-mon[52939]: min_mon_release 19 (squid) 2026-03-09T20:52:31.046 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 09 20:52:30 vm06 ceph-mon[52939]: election_strategy: 1 2026-03-09T20:52:31.046 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 09 20:52:30 vm06 ceph-mon[52939]: 0: [v2:192.168.123.101:3300/0,v1:192.168.123.101:6789/0] mon.a 2026-03-09T20:52:31.046 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 09 20:52:30 vm06 ceph-mon[52939]: 1: [v2:192.168.123.108:3300/0,v1:192.168.123.108:6789/0] mon.c 2026-03-09T20:52:31.046 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 09 20:52:30 vm06 ceph-mon[52939]: fsmap 2026-03-09T20:52:31.046 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 09 20:52:30 vm06 ceph-mon[52939]: osdmap e4: 0 total, 0 up, 0 in 2026-03-09T20:52:31.046 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 09 20:52:30 vm06 ceph-mon[52939]: mgrmap e13: a(active, since 20s) 2026-03-09T20:52:31.046 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 09 20:52:30 vm06 ceph-mon[52939]: overall HEALTH_OK 2026-03-09T20:52:31.046 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 09 20:52:30 vm06 ceph-mon[52939]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' 2026-03-09T20:52:31.046 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 09 20:52:30 vm06 ceph-mon[52939]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' 2026-03-09T20:52:31.046 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 09 20:52:30 vm06 ceph-mon[52939]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' 2026-03-09T20:52:31.046 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 09 20:52:30 vm06 ceph-mon[52939]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' 2026-03-09T20:52:33.890 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:52:33 vm01 ceph-mon[53402]: Updating vm01:/etc/ceph/ceph.conf 2026-03-09T20:52:33.890 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:52:33 vm01 ceph-mon[53402]: Updating vm06:/etc/ceph/ceph.conf 2026-03-09T20:52:33.890 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:52:33 vm01 ceph-mon[53402]: Updating vm08:/etc/ceph/ceph.conf 2026-03-09T20:52:33.890 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:52:33 vm01 ceph-mon[53402]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "a"}]: dispatch 2026-03-09T20:52:33.890 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:52:33 vm01 ceph-mon[53402]: mon.a calling monitor election 2026-03-09T20:52:33.890 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:52:33 vm01 ceph-mon[53402]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-09T20:52:33.890 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:52:33 vm01 ceph-mon[53402]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "c"}]: dispatch 2026-03-09T20:52:33.890 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:52:33 vm01 ceph-mon[53402]: mon.c calling monitor election 2026-03-09T20:52:33.890 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:52:33 vm01 ceph-mon[53402]: Updating vm08:/var/lib/ceph/9cb345a8-1bf9-11f1-a195-0375563c5891/config/ceph.conf 2026-03-09T20:52:33.890 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:52:33 vm01 ceph-mon[53402]: Updating vm06:/var/lib/ceph/9cb345a8-1bf9-11f1-a195-0375563c5891/config/ceph.conf 2026-03-09T20:52:33.890 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:52:33 vm01 ceph-mon[53402]: Updating vm01:/var/lib/ceph/9cb345a8-1bf9-11f1-a195-0375563c5891/config/ceph.conf 2026-03-09T20:52:33.890 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:52:33 vm01 ceph-mon[53402]: pgmap v5: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-09T20:52:33.890 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:52:33 vm01 ceph-mon[53402]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-09T20:52:33.890 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:52:33 vm01 ceph-mon[53402]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-09T20:52:33.890 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:52:33 vm01 ceph-mon[53402]: pgmap v6: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-09T20:52:33.890 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:52:33 vm01 ceph-mon[53402]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-09T20:52:33.890 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:52:33 vm01 ceph-mon[53402]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-09T20:52:33.890 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:52:33 vm01 ceph-mon[53402]: pgmap v7: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-09T20:52:33.890 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:52:33 vm01 ceph-mon[53402]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-09T20:52:33.890 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:52:33 vm01 ceph-mon[53402]: mon.a is new leader, mons a,c in quorum (ranks 0,1) 2026-03-09T20:52:33.890 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:52:33 vm01 ceph-mon[53402]: monmap epoch 3 2026-03-09T20:52:33.890 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:52:33 vm01 ceph-mon[53402]: fsid 9cb345a8-1bf9-11f1-a195-0375563c5891 2026-03-09T20:52:33.890 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:52:33 vm01 ceph-mon[53402]: last_changed 2026-03-09T20:52:28.571447+0000 2026-03-09T20:52:33.890 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:52:33 vm01 ceph-mon[53402]: created 2026-03-09T20:51:41.312920+0000 2026-03-09T20:52:33.890 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:52:33 vm01 ceph-mon[53402]: min_mon_release 19 (squid) 2026-03-09T20:52:33.890 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:52:33 vm01 ceph-mon[53402]: election_strategy: 1 2026-03-09T20:52:33.890 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:52:33 vm01 ceph-mon[53402]: 0: [v2:192.168.123.101:3300/0,v1:192.168.123.101:6789/0] mon.a 2026-03-09T20:52:33.890 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:52:33 vm01 ceph-mon[53402]: 1: [v2:192.168.123.108:3300/0,v1:192.168.123.108:6789/0] mon.c 2026-03-09T20:52:33.890 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:52:33 vm01 ceph-mon[53402]: 2: [v2:192.168.123.106:3300/0,v1:192.168.123.106:6789/0] mon.b 2026-03-09T20:52:33.890 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:52:33 vm01 ceph-mon[53402]: fsmap 2026-03-09T20:52:33.890 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:52:33 vm01 ceph-mon[53402]: osdmap e4: 0 total, 0 up, 0 in 2026-03-09T20:52:33.890 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:52:33 vm01 ceph-mon[53402]: mgrmap e13: a(active, since 26s) 2026-03-09T20:52:33.890 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:52:33 vm01 ceph-mon[53402]: overall HEALTH_OK 2026-03-09T20:52:33.890 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:52:33 vm01 ceph-mon[53402]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' 2026-03-09T20:52:33.890 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:52:33 vm01 ceph-mon[53402]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' 2026-03-09T20:52:33.890 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:52:33 vm01 ceph-mon[53402]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' 2026-03-09T20:52:33.890 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:52:33 vm01 ceph-mon[53402]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' 2026-03-09T20:52:33.891 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:52:33 vm01 ceph-mon[53402]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' 2026-03-09T20:52:33.891 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:52:33 vm01 ceph-mon[53402]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' 2026-03-09T20:52:33.891 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:52:33 vm01 ceph-mon[53402]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' 2026-03-09T20:52:33.891 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:52:33 vm01 ceph-mon[53402]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' 2026-03-09T20:52:33.957 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:52:33 vm08 ceph-mon[54611]: Updating vm01:/etc/ceph/ceph.conf 2026-03-09T20:52:33.957 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:52:33 vm08 ceph-mon[54611]: Updating vm06:/etc/ceph/ceph.conf 2026-03-09T20:52:33.957 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:52:33 vm08 ceph-mon[54611]: Updating vm08:/etc/ceph/ceph.conf 2026-03-09T20:52:33.957 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:52:33 vm08 ceph-mon[54611]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "a"}]: dispatch 2026-03-09T20:52:33.957 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:52:33 vm08 ceph-mon[54611]: mon.a calling monitor election 2026-03-09T20:52:33.957 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:52:33 vm08 ceph-mon[54611]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-09T20:52:33.957 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:52:33 vm08 ceph-mon[54611]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "c"}]: dispatch 2026-03-09T20:52:33.957 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:52:33 vm08 ceph-mon[54611]: mon.c calling monitor election 2026-03-09T20:52:33.957 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:52:33 vm08 ceph-mon[54611]: Updating vm08:/var/lib/ceph/9cb345a8-1bf9-11f1-a195-0375563c5891/config/ceph.conf 2026-03-09T20:52:33.957 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:52:33 vm08 ceph-mon[54611]: Updating vm06:/var/lib/ceph/9cb345a8-1bf9-11f1-a195-0375563c5891/config/ceph.conf 2026-03-09T20:52:33.957 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:52:33 vm08 ceph-mon[54611]: Updating vm01:/var/lib/ceph/9cb345a8-1bf9-11f1-a195-0375563c5891/config/ceph.conf 2026-03-09T20:52:33.957 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:52:33 vm08 ceph-mon[54611]: pgmap v5: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-09T20:52:33.957 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:52:33 vm08 ceph-mon[54611]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-09T20:52:33.957 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:52:33 vm08 ceph-mon[54611]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-09T20:52:33.957 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:52:33 vm08 ceph-mon[54611]: pgmap v6: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-09T20:52:33.957 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:52:33 vm08 ceph-mon[54611]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-09T20:52:33.957 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:52:33 vm08 ceph-mon[54611]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-09T20:52:33.957 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:52:33 vm08 ceph-mon[54611]: pgmap v7: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-09T20:52:33.957 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:52:33 vm08 ceph-mon[54611]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-09T20:52:33.957 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:52:33 vm08 ceph-mon[54611]: mon.a is new leader, mons a,c in quorum (ranks 0,1) 2026-03-09T20:52:33.957 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:52:33 vm08 ceph-mon[54611]: monmap epoch 3 2026-03-09T20:52:33.957 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:52:33 vm08 ceph-mon[54611]: fsid 9cb345a8-1bf9-11f1-a195-0375563c5891 2026-03-09T20:52:33.957 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:52:33 vm08 ceph-mon[54611]: last_changed 2026-03-09T20:52:28.571447+0000 2026-03-09T20:52:33.957 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:52:33 vm08 ceph-mon[54611]: created 2026-03-09T20:51:41.312920+0000 2026-03-09T20:52:33.957 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:52:33 vm08 ceph-mon[54611]: min_mon_release 19 (squid) 2026-03-09T20:52:33.957 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:52:33 vm08 ceph-mon[54611]: election_strategy: 1 2026-03-09T20:52:33.957 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:52:33 vm08 ceph-mon[54611]: 0: [v2:192.168.123.101:3300/0,v1:192.168.123.101:6789/0] mon.a 2026-03-09T20:52:33.957 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:52:33 vm08 ceph-mon[54611]: 1: [v2:192.168.123.108:3300/0,v1:192.168.123.108:6789/0] mon.c 2026-03-09T20:52:33.957 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:52:33 vm08 ceph-mon[54611]: 2: [v2:192.168.123.106:3300/0,v1:192.168.123.106:6789/0] mon.b 2026-03-09T20:52:33.957 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:52:33 vm08 ceph-mon[54611]: fsmap 2026-03-09T20:52:33.957 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:52:33 vm08 ceph-mon[54611]: osdmap e4: 0 total, 0 up, 0 in 2026-03-09T20:52:33.958 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:52:33 vm08 ceph-mon[54611]: mgrmap e13: a(active, since 26s) 2026-03-09T20:52:33.958 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:52:33 vm08 ceph-mon[54611]: overall HEALTH_OK 2026-03-09T20:52:33.958 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:52:33 vm08 ceph-mon[54611]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' 2026-03-09T20:52:33.958 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:52:33 vm08 ceph-mon[54611]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' 2026-03-09T20:52:33.958 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:52:33 vm08 ceph-mon[54611]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' 2026-03-09T20:52:33.958 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:52:33 vm08 ceph-mon[54611]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' 2026-03-09T20:52:33.958 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:52:33 vm08 ceph-mon[54611]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' 2026-03-09T20:52:33.958 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:52:33 vm08 ceph-mon[54611]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' 2026-03-09T20:52:33.958 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:52:33 vm08 ceph-mon[54611]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' 2026-03-09T20:52:33.958 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:52:33 vm08 ceph-mon[54611]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' 2026-03-09T20:52:34.067 INFO:teuthology.orchestra.run.vm08.stdout: 2026-03-09T20:52:34.067 INFO:teuthology.orchestra.run.vm08.stdout:{"epoch":3,"fsid":"9cb345a8-1bf9-11f1-a195-0375563c5891","modified":"2026-03-09T20:52:28.571447Z","created":"2026-03-09T20:51:41.312920Z","min_mon_release":19,"min_mon_release_name":"squid","election_strategy":1,"disallowed_leaders":"","stretch_mode":false,"tiebreaker_mon":"","removed_ranks":"","features":{"persistent":["kraken","luminous","mimic","osdmap-prune","nautilus","octopus","pacific","elector-pinging","quincy","reef","squid"],"optional":[]},"mons":[{"rank":0,"name":"a","public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.101:3300","nonce":0},{"type":"v1","addr":"192.168.123.101:6789","nonce":0}]},"addr":"192.168.123.101:6789/0","public_addr":"192.168.123.101:6789/0","priority":0,"weight":0,"crush_location":"{}"},{"rank":1,"name":"c","public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.108:3300","nonce":0},{"type":"v1","addr":"192.168.123.108:6789","nonce":0}]},"addr":"192.168.123.108:6789/0","public_addr":"192.168.123.108:6789/0","priority":0,"weight":0,"crush_location":"{}"},{"rank":2,"name":"b","public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.106:3300","nonce":0},{"type":"v1","addr":"192.168.123.106:6789","nonce":0}]},"addr":"192.168.123.106:6789/0","public_addr":"192.168.123.106:6789/0","priority":0,"weight":0,"crush_location":"{}"}],"quorum":[0,1]} 2026-03-09T20:52:34.067 INFO:teuthology.orchestra.run.vm08.stderr:dumped monmap epoch 3 2026-03-09T20:52:34.237 INFO:tasks.cephadm:Generating final ceph.conf file... 2026-03-09T20:52:34.237 DEBUG:teuthology.orchestra.run.vm01:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 9cb345a8-1bf9-11f1-a195-0375563c5891 -- ceph config generate-minimal-conf 2026-03-09T20:52:34.397 INFO:teuthology.orchestra.run.vm01.stderr:Inferring config /var/lib/ceph/9cb345a8-1bf9-11f1-a195-0375563c5891/mon.a/config 2026-03-09T20:52:34.626 INFO:teuthology.orchestra.run.vm01.stdout:# minimal ceph.conf for 9cb345a8-1bf9-11f1-a195-0375563c5891 2026-03-09T20:52:34.626 INFO:teuthology.orchestra.run.vm01.stdout:[global] 2026-03-09T20:52:34.626 INFO:teuthology.orchestra.run.vm01.stdout: fsid = 9cb345a8-1bf9-11f1-a195-0375563c5891 2026-03-09T20:52:34.626 INFO:teuthology.orchestra.run.vm01.stdout: mon_host = [v2:192.168.123.101:3300/0,v1:192.168.123.101:6789/0] [v2:192.168.123.106:3300/0,v1:192.168.123.106:6789/0] [v2:192.168.123.108:3300/0,v1:192.168.123.108:6789/0] 2026-03-09T20:52:34.770 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:52:34 vm01 ceph-mon[53402]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' 2026-03-09T20:52:34.770 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:52:34 vm01 ceph-mon[53402]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' 2026-03-09T20:52:34.770 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:52:34 vm01 ceph-mon[53402]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' 2026-03-09T20:52:34.770 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:52:34 vm01 ceph-mon[53402]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' 2026-03-09T20:52:34.770 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:52:34 vm01 ceph-mon[53402]: Reconfiguring mon.a (unknown last config time)... 2026-03-09T20:52:34.770 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:52:34 vm01 ceph-mon[53402]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "mon."}]: dispatch 2026-03-09T20:52:34.770 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:52:34 vm01 ceph-mon[53402]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' cmd=[{"prefix": "config get", "who": "mon", "key": "public_network"}]: dispatch 2026-03-09T20:52:34.770 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:52:34 vm01 ceph-mon[53402]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T20:52:34.770 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:52:34 vm01 ceph-mon[53402]: Reconfiguring daemon mon.a on vm01 2026-03-09T20:52:34.770 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:52:34 vm01 ceph-mon[53402]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' 2026-03-09T20:52:34.770 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:52:34 vm01 ceph-mon[53402]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' 2026-03-09T20:52:34.770 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:52:34 vm01 ceph-mon[53402]: Reconfiguring mon.b (monmap changed)... 2026-03-09T20:52:34.770 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:52:34 vm01 ceph-mon[53402]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "mon."}]: dispatch 2026-03-09T20:52:34.770 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:52:34 vm01 ceph-mon[53402]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' cmd=[{"prefix": "config get", "who": "mon", "key": "public_network"}]: dispatch 2026-03-09T20:52:34.770 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:52:34 vm01 ceph-mon[53402]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T20:52:34.770 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:52:34 vm01 ceph-mon[53402]: Reconfiguring daemon mon.b on vm06 2026-03-09T20:52:34.770 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:52:34 vm01 ceph-mon[53402]: from='client.? 192.168.123.108:0/2579873569' entity='client.admin' cmd=[{"prefix": "mon dump", "format": "json"}]: dispatch 2026-03-09T20:52:34.770 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:52:34 vm01 ceph-mon[53402]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' 2026-03-09T20:52:34.770 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:52:34 vm01 ceph-mon[53402]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' 2026-03-09T20:52:34.770 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:52:34 vm01 ceph-mon[53402]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "mon."}]: dispatch 2026-03-09T20:52:34.770 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:52:34 vm01 ceph-mon[53402]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' cmd=[{"prefix": "config get", "who": "mon", "key": "public_network"}]: dispatch 2026-03-09T20:52:34.770 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:52:34 vm01 ceph-mon[53402]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T20:52:34.770 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:52:34 vm01 ceph-mon[53402]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-09T20:52:34.770 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:52:34 vm01 ceph-mon[53402]: from='client.? 192.168.123.101:0/4095839030' entity='client.admin' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T20:52:34.798 INFO:tasks.cephadm:Distributing (final) config and client.admin keyring... 2026-03-09T20:52:34.798 DEBUG:teuthology.orchestra.run.vm01:> set -ex 2026-03-09T20:52:34.798 DEBUG:teuthology.orchestra.run.vm01:> sudo dd of=/etc/ceph/ceph.conf 2026-03-09T20:52:34.831 DEBUG:teuthology.orchestra.run.vm01:> set -ex 2026-03-09T20:52:34.831 DEBUG:teuthology.orchestra.run.vm01:> sudo dd of=/etc/ceph/ceph.client.admin.keyring 2026-03-09T20:52:34.910 DEBUG:teuthology.orchestra.run.vm06:> set -ex 2026-03-09T20:52:34.910 DEBUG:teuthology.orchestra.run.vm06:> sudo dd of=/etc/ceph/ceph.conf 2026-03-09T20:52:34.937 DEBUG:teuthology.orchestra.run.vm06:> set -ex 2026-03-09T20:52:34.937 DEBUG:teuthology.orchestra.run.vm06:> sudo dd of=/etc/ceph/ceph.client.admin.keyring 2026-03-09T20:52:34.957 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:52:34 vm08 ceph-mon[54611]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' 2026-03-09T20:52:34.957 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:52:34 vm08 ceph-mon[54611]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' 2026-03-09T20:52:34.957 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:52:34 vm08 ceph-mon[54611]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' 2026-03-09T20:52:34.957 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:52:34 vm08 ceph-mon[54611]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' 2026-03-09T20:52:34.957 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:52:34 vm08 ceph-mon[54611]: Reconfiguring mon.a (unknown last config time)... 2026-03-09T20:52:34.957 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:52:34 vm08 ceph-mon[54611]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "mon."}]: dispatch 2026-03-09T20:52:34.957 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:52:34 vm08 ceph-mon[54611]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' cmd=[{"prefix": "config get", "who": "mon", "key": "public_network"}]: dispatch 2026-03-09T20:52:34.957 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:52:34 vm08 ceph-mon[54611]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T20:52:34.957 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:52:34 vm08 ceph-mon[54611]: Reconfiguring daemon mon.a on vm01 2026-03-09T20:52:34.957 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:52:34 vm08 ceph-mon[54611]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' 2026-03-09T20:52:34.957 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:52:34 vm08 ceph-mon[54611]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' 2026-03-09T20:52:34.957 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:52:34 vm08 ceph-mon[54611]: Reconfiguring mon.b (monmap changed)... 2026-03-09T20:52:34.957 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:52:34 vm08 ceph-mon[54611]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "mon."}]: dispatch 2026-03-09T20:52:34.957 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:52:34 vm08 ceph-mon[54611]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' cmd=[{"prefix": "config get", "who": "mon", "key": "public_network"}]: dispatch 2026-03-09T20:52:34.957 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:52:34 vm08 ceph-mon[54611]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T20:52:34.957 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:52:34 vm08 ceph-mon[54611]: Reconfiguring daemon mon.b on vm06 2026-03-09T20:52:34.957 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:52:34 vm08 ceph-mon[54611]: from='client.? 192.168.123.108:0/2579873569' entity='client.admin' cmd=[{"prefix": "mon dump", "format": "json"}]: dispatch 2026-03-09T20:52:34.957 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:52:34 vm08 ceph-mon[54611]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' 2026-03-09T20:52:34.957 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:52:34 vm08 ceph-mon[54611]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' 2026-03-09T20:52:34.957 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:52:34 vm08 ceph-mon[54611]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "mon."}]: dispatch 2026-03-09T20:52:34.957 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:52:34 vm08 ceph-mon[54611]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' cmd=[{"prefix": "config get", "who": "mon", "key": "public_network"}]: dispatch 2026-03-09T20:52:34.957 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:52:34 vm08 ceph-mon[54611]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T20:52:34.957 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:52:34 vm08 ceph-mon[54611]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-09T20:52:34.957 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:52:34 vm08 ceph-mon[54611]: from='client.? 192.168.123.101:0/4095839030' entity='client.admin' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T20:52:35.008 DEBUG:teuthology.orchestra.run.vm08:> set -ex 2026-03-09T20:52:35.008 DEBUG:teuthology.orchestra.run.vm08:> sudo dd of=/etc/ceph/ceph.conf 2026-03-09T20:52:35.036 DEBUG:teuthology.orchestra.run.vm08:> set -ex 2026-03-09T20:52:35.036 DEBUG:teuthology.orchestra.run.vm08:> sudo dd of=/etc/ceph/ceph.client.admin.keyring 2026-03-09T20:52:35.106 INFO:tasks.cephadm:Adding mgr.a on vm01 2026-03-09T20:52:35.106 INFO:tasks.cephadm:Adding mgr.b on vm06 2026-03-09T20:52:35.106 DEBUG:teuthology.orchestra.run.vm08:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 9cb345a8-1bf9-11f1-a195-0375563c5891 -- ceph orch apply mgr '2;vm01=a;vm06=b' 2026-03-09T20:52:35.299 INFO:teuthology.orchestra.run.vm08.stderr:Inferring config /var/lib/ceph/9cb345a8-1bf9-11f1-a195-0375563c5891/mon.c/config 2026-03-09T20:52:35.546 INFO:teuthology.orchestra.run.vm08.stdout:Scheduled mgr update... 2026-03-09T20:52:35.725 DEBUG:teuthology.orchestra.run.vm06:mgr.b> sudo journalctl -f -n 0 -u ceph-9cb345a8-1bf9-11f1-a195-0375563c5891@mgr.b.service 2026-03-09T20:52:35.727 INFO:tasks.cephadm:Deploying OSDs... 2026-03-09T20:52:35.727 DEBUG:teuthology.orchestra.run.vm01:> set -ex 2026-03-09T20:52:35.727 DEBUG:teuthology.orchestra.run.vm01:> dd if=/scratch_devs of=/dev/stdout 2026-03-09T20:52:35.747 DEBUG:teuthology.orchestra.run:got remote process result: 1 2026-03-09T20:52:35.747 DEBUG:teuthology.orchestra.run.vm01:> ls /dev/[sv]d? 2026-03-09T20:52:35.807 INFO:teuthology.orchestra.run.vm01.stdout:/dev/vda 2026-03-09T20:52:35.807 INFO:teuthology.orchestra.run.vm01.stdout:/dev/vdb 2026-03-09T20:52:35.807 INFO:teuthology.orchestra.run.vm01.stdout:/dev/vdc 2026-03-09T20:52:35.807 INFO:teuthology.orchestra.run.vm01.stdout:/dev/vdd 2026-03-09T20:52:35.807 INFO:teuthology.orchestra.run.vm01.stdout:/dev/vde 2026-03-09T20:52:35.807 WARNING:teuthology.misc:Removing root device: /dev/vda from device list 2026-03-09T20:52:35.807 DEBUG:teuthology.misc:devs=['/dev/vdb', '/dev/vdc', '/dev/vdd', '/dev/vde'] 2026-03-09T20:52:35.807 DEBUG:teuthology.orchestra.run.vm01:> stat /dev/vdb 2026-03-09T20:52:35.838 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 09 20:52:35 vm06 ceph-mon[52939]: Updating vm01:/etc/ceph/ceph.conf 2026-03-09T20:52:35.838 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 09 20:52:35 vm06 ceph-mon[52939]: Updating vm06:/etc/ceph/ceph.conf 2026-03-09T20:52:35.838 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 09 20:52:35 vm06 ceph-mon[52939]: Updating vm08:/etc/ceph/ceph.conf 2026-03-09T20:52:35.838 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 09 20:52:35 vm06 ceph-mon[52939]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "a"}]: dispatch 2026-03-09T20:52:35.838 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 09 20:52:35 vm06 ceph-mon[52939]: mon.a calling monitor election 2026-03-09T20:52:35.838 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 09 20:52:35 vm06 ceph-mon[52939]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-09T20:52:35.838 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 09 20:52:35 vm06 ceph-mon[52939]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "c"}]: dispatch 2026-03-09T20:52:35.838 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 09 20:52:35 vm06 ceph-mon[52939]: mon.c calling monitor election 2026-03-09T20:52:35.838 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 09 20:52:35 vm06 ceph-mon[52939]: Updating vm08:/var/lib/ceph/9cb345a8-1bf9-11f1-a195-0375563c5891/config/ceph.conf 2026-03-09T20:52:35.838 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 09 20:52:35 vm06 ceph-mon[52939]: Updating vm06:/var/lib/ceph/9cb345a8-1bf9-11f1-a195-0375563c5891/config/ceph.conf 2026-03-09T20:52:35.838 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 09 20:52:35 vm06 ceph-mon[52939]: Updating vm01:/var/lib/ceph/9cb345a8-1bf9-11f1-a195-0375563c5891/config/ceph.conf 2026-03-09T20:52:35.838 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 09 20:52:35 vm06 ceph-mon[52939]: pgmap v5: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-09T20:52:35.838 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 09 20:52:35 vm06 ceph-mon[52939]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-09T20:52:35.838 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 09 20:52:35 vm06 ceph-mon[52939]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-09T20:52:35.838 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 09 20:52:35 vm06 ceph-mon[52939]: pgmap v6: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-09T20:52:35.838 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 09 20:52:35 vm06 ceph-mon[52939]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-09T20:52:35.838 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 09 20:52:35 vm06 ceph-mon[52939]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-09T20:52:35.838 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 09 20:52:35 vm06 ceph-mon[52939]: pgmap v7: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-09T20:52:35.839 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 09 20:52:35 vm06 ceph-mon[52939]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-09T20:52:35.839 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 09 20:52:35 vm06 ceph-mon[52939]: mon.a is new leader, mons a,c in quorum (ranks 0,1) 2026-03-09T20:52:35.839 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 09 20:52:35 vm06 ceph-mon[52939]: monmap epoch 3 2026-03-09T20:52:35.839 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 09 20:52:35 vm06 ceph-mon[52939]: fsid 9cb345a8-1bf9-11f1-a195-0375563c5891 2026-03-09T20:52:35.839 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 09 20:52:35 vm06 ceph-mon[52939]: last_changed 2026-03-09T20:52:28.571447+0000 2026-03-09T20:52:35.839 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 09 20:52:35 vm06 ceph-mon[52939]: created 2026-03-09T20:51:41.312920+0000 2026-03-09T20:52:35.839 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 09 20:52:35 vm06 ceph-mon[52939]: min_mon_release 19 (squid) 2026-03-09T20:52:35.839 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 09 20:52:35 vm06 ceph-mon[52939]: election_strategy: 1 2026-03-09T20:52:35.839 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 09 20:52:35 vm06 ceph-mon[52939]: 0: [v2:192.168.123.101:3300/0,v1:192.168.123.101:6789/0] mon.a 2026-03-09T20:52:35.839 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 09 20:52:35 vm06 ceph-mon[52939]: 1: [v2:192.168.123.108:3300/0,v1:192.168.123.108:6789/0] mon.c 2026-03-09T20:52:35.839 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 09 20:52:35 vm06 ceph-mon[52939]: 2: [v2:192.168.123.106:3300/0,v1:192.168.123.106:6789/0] mon.b 2026-03-09T20:52:35.839 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 09 20:52:35 vm06 ceph-mon[52939]: fsmap 2026-03-09T20:52:35.839 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 09 20:52:35 vm06 ceph-mon[52939]: osdmap e4: 0 total, 0 up, 0 in 2026-03-09T20:52:35.839 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 09 20:52:35 vm06 ceph-mon[52939]: mgrmap e13: a(active, since 26s) 2026-03-09T20:52:35.839 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 09 20:52:35 vm06 ceph-mon[52939]: overall HEALTH_OK 2026-03-09T20:52:35.839 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 09 20:52:35 vm06 ceph-mon[52939]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' 2026-03-09T20:52:35.839 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 09 20:52:35 vm06 ceph-mon[52939]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' 2026-03-09T20:52:35.839 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 09 20:52:35 vm06 ceph-mon[52939]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' 2026-03-09T20:52:35.839 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 09 20:52:35 vm06 ceph-mon[52939]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' 2026-03-09T20:52:35.839 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 09 20:52:35 vm06 ceph-mon[52939]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' 2026-03-09T20:52:35.839 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 09 20:52:35 vm06 ceph-mon[52939]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' 2026-03-09T20:52:35.839 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 09 20:52:35 vm06 ceph-mon[52939]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' 2026-03-09T20:52:35.839 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 09 20:52:35 vm06 ceph-mon[52939]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' 2026-03-09T20:52:35.839 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 09 20:52:35 vm06 ceph-mon[52939]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' 2026-03-09T20:52:35.839 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 09 20:52:35 vm06 ceph-mon[52939]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' 2026-03-09T20:52:35.839 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 09 20:52:35 vm06 ceph-mon[52939]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' 2026-03-09T20:52:35.839 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 09 20:52:35 vm06 ceph-mon[52939]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' 2026-03-09T20:52:35.839 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 09 20:52:35 vm06 ceph-mon[52939]: Reconfiguring mon.a (unknown last config time)... 2026-03-09T20:52:35.839 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 09 20:52:35 vm06 ceph-mon[52939]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "mon."}]: dispatch 2026-03-09T20:52:35.839 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 09 20:52:35 vm06 ceph-mon[52939]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' cmd=[{"prefix": "config get", "who": "mon", "key": "public_network"}]: dispatch 2026-03-09T20:52:35.839 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 09 20:52:35 vm06 ceph-mon[52939]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T20:52:35.839 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 09 20:52:35 vm06 ceph-mon[52939]: Reconfiguring daemon mon.a on vm01 2026-03-09T20:52:35.839 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 09 20:52:35 vm06 ceph-mon[52939]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' 2026-03-09T20:52:35.839 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 09 20:52:35 vm06 ceph-mon[52939]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' 2026-03-09T20:52:35.839 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 09 20:52:35 vm06 ceph-mon[52939]: Reconfiguring mon.b (monmap changed)... 2026-03-09T20:52:35.839 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 09 20:52:35 vm06 ceph-mon[52939]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "mon."}]: dispatch 2026-03-09T20:52:35.839 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 09 20:52:35 vm06 ceph-mon[52939]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' cmd=[{"prefix": "config get", "who": "mon", "key": "public_network"}]: dispatch 2026-03-09T20:52:35.839 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 09 20:52:35 vm06 ceph-mon[52939]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T20:52:35.839 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 09 20:52:35 vm06 ceph-mon[52939]: Reconfiguring daemon mon.b on vm06 2026-03-09T20:52:35.839 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 09 20:52:35 vm06 ceph-mon[52939]: from='client.? 192.168.123.108:0/2579873569' entity='client.admin' cmd=[{"prefix": "mon dump", "format": "json"}]: dispatch 2026-03-09T20:52:35.839 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 09 20:52:35 vm06 ceph-mon[52939]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' 2026-03-09T20:52:35.839 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 09 20:52:35 vm06 ceph-mon[52939]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' 2026-03-09T20:52:35.839 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 09 20:52:35 vm06 ceph-mon[52939]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "mon."}]: dispatch 2026-03-09T20:52:35.839 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 09 20:52:35 vm06 ceph-mon[52939]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' cmd=[{"prefix": "config get", "who": "mon", "key": "public_network"}]: dispatch 2026-03-09T20:52:35.839 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 09 20:52:35 vm06 ceph-mon[52939]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T20:52:35.839 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 09 20:52:35 vm06 ceph-mon[52939]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-09T20:52:35.839 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 09 20:52:35 vm06 ceph-mon[52939]: from='client.? 192.168.123.101:0/4095839030' entity='client.admin' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T20:52:35.865 INFO:teuthology.orchestra.run.vm01.stdout: File: /dev/vdb 2026-03-09T20:52:35.865 INFO:teuthology.orchestra.run.vm01.stdout: Size: 0 Blocks: 0 IO Block: 512 block special file 2026-03-09T20:52:35.866 INFO:teuthology.orchestra.run.vm01.stdout:Device: 6h/6d Inode: 223 Links: 1 Device type: fc,10 2026-03-09T20:52:35.866 INFO:teuthology.orchestra.run.vm01.stdout:Access: (0660/brw-rw----) Uid: ( 0/ root) Gid: ( 6/ disk) 2026-03-09T20:52:35.866 INFO:teuthology.orchestra.run.vm01.stdout:Context: system_u:object_r:fixed_disk_device_t:s0 2026-03-09T20:52:35.866 INFO:teuthology.orchestra.run.vm01.stdout:Access: 2026-03-09 20:52:11.348368339 +0000 2026-03-09T20:52:35.866 INFO:teuthology.orchestra.run.vm01.stdout:Modify: 2026-03-09 20:49:12.758525867 +0000 2026-03-09T20:52:35.866 INFO:teuthology.orchestra.run.vm01.stdout:Change: 2026-03-09 20:49:12.758525867 +0000 2026-03-09T20:52:35.866 INFO:teuthology.orchestra.run.vm01.stdout: Birth: 2026-03-09 20:45:34.259000000 +0000 2026-03-09T20:52:35.866 DEBUG:teuthology.orchestra.run.vm01:> sudo dd if=/dev/vdb of=/dev/null count=1 2026-03-09T20:52:35.931 INFO:teuthology.orchestra.run.vm01.stderr:1+0 records in 2026-03-09T20:52:35.931 INFO:teuthology.orchestra.run.vm01.stderr:1+0 records out 2026-03-09T20:52:35.931 INFO:teuthology.orchestra.run.vm01.stderr:512 bytes copied, 0.000171711 s, 3.0 MB/s 2026-03-09T20:52:35.933 DEBUG:teuthology.orchestra.run.vm01:> ! mount | grep -v devtmpfs | grep -q /dev/vdb 2026-03-09T20:52:35.991 DEBUG:teuthology.orchestra.run.vm01:> stat /dev/vdc 2026-03-09T20:52:36.050 INFO:teuthology.orchestra.run.vm01.stdout: File: /dev/vdc 2026-03-09T20:52:36.050 INFO:teuthology.orchestra.run.vm01.stdout: Size: 0 Blocks: 0 IO Block: 512 block special file 2026-03-09T20:52:36.050 INFO:teuthology.orchestra.run.vm01.stdout:Device: 6h/6d Inode: 224 Links: 1 Device type: fc,20 2026-03-09T20:52:36.050 INFO:teuthology.orchestra.run.vm01.stdout:Access: (0660/brw-rw----) Uid: ( 0/ root) Gid: ( 6/ disk) 2026-03-09T20:52:36.050 INFO:teuthology.orchestra.run.vm01.stdout:Context: system_u:object_r:fixed_disk_device_t:s0 2026-03-09T20:52:36.050 INFO:teuthology.orchestra.run.vm01.stdout:Access: 2026-03-09 20:52:11.393368479 +0000 2026-03-09T20:52:36.050 INFO:teuthology.orchestra.run.vm01.stdout:Modify: 2026-03-09 20:49:12.765525875 +0000 2026-03-09T20:52:36.050 INFO:teuthology.orchestra.run.vm01.stdout:Change: 2026-03-09 20:49:12.765525875 +0000 2026-03-09T20:52:36.050 INFO:teuthology.orchestra.run.vm01.stdout: Birth: 2026-03-09 20:45:34.262000000 +0000 2026-03-09T20:52:36.051 DEBUG:teuthology.orchestra.run.vm01:> sudo dd if=/dev/vdc of=/dev/null count=1 2026-03-09T20:52:36.114 INFO:teuthology.orchestra.run.vm01.stderr:1+0 records in 2026-03-09T20:52:36.114 INFO:teuthology.orchestra.run.vm01.stderr:1+0 records out 2026-03-09T20:52:36.114 INFO:teuthology.orchestra.run.vm01.stderr:512 bytes copied, 0.000184055 s, 2.8 MB/s 2026-03-09T20:52:36.115 DEBUG:teuthology.orchestra.run.vm01:> ! mount | grep -v devtmpfs | grep -q /dev/vdc 2026-03-09T20:52:36.172 DEBUG:teuthology.orchestra.run.vm01:> stat /dev/vdd 2026-03-09T20:52:36.228 INFO:teuthology.orchestra.run.vm01.stdout: File: /dev/vdd 2026-03-09T20:52:36.228 INFO:teuthology.orchestra.run.vm01.stdout: Size: 0 Blocks: 0 IO Block: 512 block special file 2026-03-09T20:52:36.228 INFO:teuthology.orchestra.run.vm01.stdout:Device: 6h/6d Inode: 225 Links: 1 Device type: fc,30 2026-03-09T20:52:36.228 INFO:teuthology.orchestra.run.vm01.stdout:Access: (0660/brw-rw----) Uid: ( 0/ root) Gid: ( 6/ disk) 2026-03-09T20:52:36.228 INFO:teuthology.orchestra.run.vm01.stdout:Context: system_u:object_r:fixed_disk_device_t:s0 2026-03-09T20:52:36.228 INFO:teuthology.orchestra.run.vm01.stdout:Access: 2026-03-09 20:52:11.417368553 +0000 2026-03-09T20:52:36.228 INFO:teuthology.orchestra.run.vm01.stdout:Modify: 2026-03-09 20:49:12.738525844 +0000 2026-03-09T20:52:36.228 INFO:teuthology.orchestra.run.vm01.stdout:Change: 2026-03-09 20:49:12.738525844 +0000 2026-03-09T20:52:36.228 INFO:teuthology.orchestra.run.vm01.stdout: Birth: 2026-03-09 20:45:34.265000000 +0000 2026-03-09T20:52:36.228 DEBUG:teuthology.orchestra.run.vm01:> sudo dd if=/dev/vdd of=/dev/null count=1 2026-03-09T20:52:36.292 INFO:teuthology.orchestra.run.vm01.stderr:1+0 records in 2026-03-09T20:52:36.292 INFO:teuthology.orchestra.run.vm01.stderr:1+0 records out 2026-03-09T20:52:36.292 INFO:teuthology.orchestra.run.vm01.stderr:512 bytes copied, 0.000177532 s, 2.9 MB/s 2026-03-09T20:52:36.293 DEBUG:teuthology.orchestra.run.vm01:> ! mount | grep -v devtmpfs | grep -q /dev/vdd 2026-03-09T20:52:36.349 DEBUG:teuthology.orchestra.run.vm01:> stat /dev/vde 2026-03-09T20:52:36.389 INFO:journalctl@ceph.mgr.b.vm06.stdout:Mar 09 20:52:36 vm06 podman[54249]: 2026-03-09 20:52:36.359086293 +0000 UTC m=+0.015998123 container create 355207d5317a091a502516e081024127f73f66af33082f8711fff45a57f9714d (image=quay.ceph.io/ceph-ci/ceph@sha256:8fda260ab1d2d3118a1622f7df75f44f285dfe74e71793626152a711c12bf2cc, name=ceph-9cb345a8-1bf9-11f1-a195-0375563c5891-mgr-b, OSD_FLAVOR=default, ceph=True, CEPH_REF=squid, org.label-schema.vendor=CentOS, CEPH_GIT_REPO=https://github.com/ceph/ceph-ci.git, io.buildah.version=1.41.3, FROM_IMAGE=quay.io/centos/centos:stream9, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.schema-version=1.0, GANESHA_REPO_BASEURL=https://buildlogs.centos.org/centos/$releasever-stream/storage/$basearch/nfsganesha-5/, org.label-schema.license=GPLv2, org.opencontainers.image.documentation=https://docs.ceph.com/, org.opencontainers.image.authors=Ceph Release Team , org.label-schema.build-date=20260223, CEPH_SHA1=e911bdebe5c8faa3800735d1568fcdca65db60df) 2026-03-09T20:52:36.405 INFO:teuthology.orchestra.run.vm01.stdout: File: /dev/vde 2026-03-09T20:52:36.405 INFO:teuthology.orchestra.run.vm01.stdout: Size: 0 Blocks: 0 IO Block: 512 block special file 2026-03-09T20:52:36.405 INFO:teuthology.orchestra.run.vm01.stdout:Device: 6h/6d Inode: 226 Links: 1 Device type: fc,40 2026-03-09T20:52:36.405 INFO:teuthology.orchestra.run.vm01.stdout:Access: (0660/brw-rw----) Uid: ( 0/ root) Gid: ( 6/ disk) 2026-03-09T20:52:36.405 INFO:teuthology.orchestra.run.vm01.stdout:Context: system_u:object_r:fixed_disk_device_t:s0 2026-03-09T20:52:36.405 INFO:teuthology.orchestra.run.vm01.stdout:Access: 2026-03-09 20:52:11.440368625 +0000 2026-03-09T20:52:36.406 INFO:teuthology.orchestra.run.vm01.stdout:Modify: 2026-03-09 20:49:12.780525893 +0000 2026-03-09T20:52:36.406 INFO:teuthology.orchestra.run.vm01.stdout:Change: 2026-03-09 20:49:12.780525893 +0000 2026-03-09T20:52:36.406 INFO:teuthology.orchestra.run.vm01.stdout: Birth: 2026-03-09 20:45:34.269000000 +0000 2026-03-09T20:52:36.406 DEBUG:teuthology.orchestra.run.vm01:> sudo dd if=/dev/vde of=/dev/null count=1 2026-03-09T20:52:36.476 INFO:teuthology.orchestra.run.vm01.stderr:1+0 records in 2026-03-09T20:52:36.476 INFO:teuthology.orchestra.run.vm01.stderr:1+0 records out 2026-03-09T20:52:36.476 INFO:teuthology.orchestra.run.vm01.stderr:512 bytes copied, 0.000154329 s, 3.3 MB/s 2026-03-09T20:52:36.477 DEBUG:teuthology.orchestra.run.vm01:> ! mount | grep -v devtmpfs | grep -q /dev/vde 2026-03-09T20:52:36.536 DEBUG:teuthology.orchestra.run.vm06:> set -ex 2026-03-09T20:52:36.536 DEBUG:teuthology.orchestra.run.vm06:> dd if=/scratch_devs of=/dev/stdout 2026-03-09T20:52:36.554 DEBUG:teuthology.orchestra.run:got remote process result: 1 2026-03-09T20:52:36.555 DEBUG:teuthology.orchestra.run.vm06:> ls /dev/[sv]d? 2026-03-09T20:52:36.564 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:52:36 vm01 ceph-mon[53402]: mon.b calling monitor election 2026-03-09T20:52:36.564 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:52:36 vm01 ceph-mon[53402]: from='client.14205 -' entity='client.admin' cmd=[{"prefix": "orch apply", "service_type": "mgr", "placement": "2;vm01=a;vm06=b", "target": ["mon-mgr", ""]}]: dispatch 2026-03-09T20:52:36.564 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:52:36 vm01 ceph-mon[53402]: Saving service mgr spec with placement vm01=a;vm06=b;count:2 2026-03-09T20:52:36.564 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:52:36 vm01 ceph-mon[53402]: Deploying daemon mgr.b on vm06 2026-03-09T20:52:36.564 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:52:36 vm01 ceph-mon[53402]: mon.b calling monitor election 2026-03-09T20:52:36.564 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:52:36 vm01 ceph-mon[53402]: mon.c calling monitor election 2026-03-09T20:52:36.564 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:52:36 vm01 ceph-mon[53402]: mon.a calling monitor election 2026-03-09T20:52:36.564 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:52:36 vm01 ceph-mon[53402]: mon.a is new leader, mons a,c,b in quorum (ranks 0,1,2) 2026-03-09T20:52:36.564 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:52:36 vm01 ceph-mon[53402]: monmap epoch 3 2026-03-09T20:52:36.564 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:52:36 vm01 ceph-mon[53402]: fsid 9cb345a8-1bf9-11f1-a195-0375563c5891 2026-03-09T20:52:36.564 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:52:36 vm01 ceph-mon[53402]: last_changed 2026-03-09T20:52:28.571447+0000 2026-03-09T20:52:36.564 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:52:36 vm01 ceph-mon[53402]: created 2026-03-09T20:51:41.312920+0000 2026-03-09T20:52:36.564 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:52:36 vm01 ceph-mon[53402]: min_mon_release 19 (squid) 2026-03-09T20:52:36.564 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:52:36 vm01 ceph-mon[53402]: election_strategy: 1 2026-03-09T20:52:36.564 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:52:36 vm01 ceph-mon[53402]: 0: [v2:192.168.123.101:3300/0,v1:192.168.123.101:6789/0] mon.a 2026-03-09T20:52:36.564 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:52:36 vm01 ceph-mon[53402]: 1: [v2:192.168.123.108:3300/0,v1:192.168.123.108:6789/0] mon.c 2026-03-09T20:52:36.564 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:52:36 vm01 ceph-mon[53402]: 2: [v2:192.168.123.106:3300/0,v1:192.168.123.106:6789/0] mon.b 2026-03-09T20:52:36.564 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:52:36 vm01 ceph-mon[53402]: fsmap 2026-03-09T20:52:36.564 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:52:36 vm01 ceph-mon[53402]: osdmap e4: 0 total, 0 up, 0 in 2026-03-09T20:52:36.564 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:52:36 vm01 ceph-mon[53402]: mgrmap e13: a(active, since 28s) 2026-03-09T20:52:36.564 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:52:36 vm01 ceph-mon[53402]: overall HEALTH_OK 2026-03-09T20:52:36.564 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:52:36 vm01 ceph-mon[53402]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' 2026-03-09T20:52:36.564 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:52:36 vm01 ceph-mon[53402]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' 2026-03-09T20:52:36.564 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:52:36 vm01 ceph-mon[53402]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' 2026-03-09T20:52:36.564 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:52:36 vm01 ceph-mon[53402]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' 2026-03-09T20:52:36.564 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:52:36 vm01 ceph-mon[53402]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T20:52:36.676 INFO:teuthology.orchestra.run.vm06.stdout:/dev/vda 2026-03-09T20:52:36.676 INFO:teuthology.orchestra.run.vm06.stdout:/dev/vdb 2026-03-09T20:52:36.676 INFO:teuthology.orchestra.run.vm06.stdout:/dev/vdc 2026-03-09T20:52:36.676 INFO:teuthology.orchestra.run.vm06.stdout:/dev/vdd 2026-03-09T20:52:36.676 INFO:teuthology.orchestra.run.vm06.stdout:/dev/vde 2026-03-09T20:52:36.676 WARNING:teuthology.misc:Removing root device: /dev/vda from device list 2026-03-09T20:52:36.677 DEBUG:teuthology.misc:devs=['/dev/vdb', '/dev/vdc', '/dev/vdd', '/dev/vde'] 2026-03-09T20:52:36.677 DEBUG:teuthology.orchestra.run.vm06:> stat /dev/vdb 2026-03-09T20:52:36.777 INFO:teuthology.orchestra.run.vm06.stdout: File: /dev/vdb 2026-03-09T20:52:36.777 INFO:teuthology.orchestra.run.vm06.stdout: Size: 0 Blocks: 0 IO Block: 512 block special file 2026-03-09T20:52:36.777 INFO:teuthology.orchestra.run.vm06.stdout:Device: 6h/6d Inode: 218 Links: 1 Device type: fc,10 2026-03-09T20:52:36.777 INFO:teuthology.orchestra.run.vm06.stdout:Access: (0660/brw-rw----) Uid: ( 0/ root) Gid: ( 6/ disk) 2026-03-09T20:52:36.777 INFO:teuthology.orchestra.run.vm06.stdout:Context: system_u:object_r:fixed_disk_device_t:s0 2026-03-09T20:52:36.777 INFO:teuthology.orchestra.run.vm06.stdout:Access: 2026-03-09 20:52:16.657568894 +0000 2026-03-09T20:52:36.777 INFO:teuthology.orchestra.run.vm06.stdout:Modify: 2026-03-09 20:49:12.064437628 +0000 2026-03-09T20:52:36.777 INFO:teuthology.orchestra.run.vm06.stdout:Change: 2026-03-09 20:49:12.064437628 +0000 2026-03-09T20:52:36.777 INFO:teuthology.orchestra.run.vm06.stdout: Birth: 2026-03-09 20:45:03.255000000 +0000 2026-03-09T20:52:36.777 DEBUG:teuthology.orchestra.run.vm06:> sudo dd if=/dev/vdb of=/dev/null count=1 2026-03-09T20:52:36.797 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 09 20:52:36 vm06 ceph-mon[52939]: mon.b calling monitor election 2026-03-09T20:52:36.797 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 09 20:52:36 vm06 ceph-mon[52939]: from='client.14205 -' entity='client.admin' cmd=[{"prefix": "orch apply", "service_type": "mgr", "placement": "2;vm01=a;vm06=b", "target": ["mon-mgr", ""]}]: dispatch 2026-03-09T20:52:36.797 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 09 20:52:36 vm06 ceph-mon[52939]: Saving service mgr spec with placement vm01=a;vm06=b;count:2 2026-03-09T20:52:36.797 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 09 20:52:36 vm06 ceph-mon[52939]: Deploying daemon mgr.b on vm06 2026-03-09T20:52:36.797 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 09 20:52:36 vm06 ceph-mon[52939]: mon.b calling monitor election 2026-03-09T20:52:36.797 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 09 20:52:36 vm06 ceph-mon[52939]: mon.c calling monitor election 2026-03-09T20:52:36.797 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 09 20:52:36 vm06 ceph-mon[52939]: mon.a calling monitor election 2026-03-09T20:52:36.797 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 09 20:52:36 vm06 ceph-mon[52939]: mon.a is new leader, mons a,c,b in quorum (ranks 0,1,2) 2026-03-09T20:52:36.797 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 09 20:52:36 vm06 ceph-mon[52939]: monmap epoch 3 2026-03-09T20:52:36.797 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 09 20:52:36 vm06 ceph-mon[52939]: fsid 9cb345a8-1bf9-11f1-a195-0375563c5891 2026-03-09T20:52:36.797 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 09 20:52:36 vm06 ceph-mon[52939]: last_changed 2026-03-09T20:52:28.571447+0000 2026-03-09T20:52:36.797 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 09 20:52:36 vm06 ceph-mon[52939]: created 2026-03-09T20:51:41.312920+0000 2026-03-09T20:52:36.797 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 09 20:52:36 vm06 ceph-mon[52939]: min_mon_release 19 (squid) 2026-03-09T20:52:36.797 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 09 20:52:36 vm06 ceph-mon[52939]: election_strategy: 1 2026-03-09T20:52:36.797 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 09 20:52:36 vm06 ceph-mon[52939]: 0: [v2:192.168.123.101:3300/0,v1:192.168.123.101:6789/0] mon.a 2026-03-09T20:52:36.797 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 09 20:52:36 vm06 ceph-mon[52939]: 1: [v2:192.168.123.108:3300/0,v1:192.168.123.108:6789/0] mon.c 2026-03-09T20:52:36.797 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 09 20:52:36 vm06 ceph-mon[52939]: 2: [v2:192.168.123.106:3300/0,v1:192.168.123.106:6789/0] mon.b 2026-03-09T20:52:36.797 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 09 20:52:36 vm06 ceph-mon[52939]: fsmap 2026-03-09T20:52:36.797 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 09 20:52:36 vm06 ceph-mon[52939]: osdmap e4: 0 total, 0 up, 0 in 2026-03-09T20:52:36.797 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 09 20:52:36 vm06 ceph-mon[52939]: mgrmap e13: a(active, since 28s) 2026-03-09T20:52:36.797 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 09 20:52:36 vm06 ceph-mon[52939]: overall HEALTH_OK 2026-03-09T20:52:36.797 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 09 20:52:36 vm06 ceph-mon[52939]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' 2026-03-09T20:52:36.797 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 09 20:52:36 vm06 ceph-mon[52939]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' 2026-03-09T20:52:36.797 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 09 20:52:36 vm06 ceph-mon[52939]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' 2026-03-09T20:52:36.797 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 09 20:52:36 vm06 ceph-mon[52939]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' 2026-03-09T20:52:36.797 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 09 20:52:36 vm06 ceph-mon[52939]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T20:52:36.797 INFO:journalctl@ceph.mgr.b.vm06.stdout:Mar 09 20:52:36 vm06 podman[54249]: 2026-03-09 20:52:36.405647374 +0000 UTC m=+0.062559213 container init 355207d5317a091a502516e081024127f73f66af33082f8711fff45a57f9714d (image=quay.ceph.io/ceph-ci/ceph@sha256:8fda260ab1d2d3118a1622f7df75f44f285dfe74e71793626152a711c12bf2cc, name=ceph-9cb345a8-1bf9-11f1-a195-0375563c5891-mgr-b, ceph=True, org.label-schema.name=CentOS Stream 9 Base Image, FROM_IMAGE=quay.io/centos/centos:stream9, org.opencontainers.image.authors=Ceph Release Team , GANESHA_REPO_BASEURL=https://buildlogs.centos.org/centos/$releasever-stream/storage/$basearch/nfsganesha-5/, io.buildah.version=1.41.3, org.label-schema.schema-version=1.0, CEPH_REF=squid, CEPH_GIT_REPO=https://github.com/ceph/ceph-ci.git, org.label-schema.build-date=20260223, OSD_FLAVOR=default, org.label-schema.vendor=CentOS, org.opencontainers.image.documentation=https://docs.ceph.com/, org.label-schema.license=GPLv2, CEPH_SHA1=e911bdebe5c8faa3800735d1568fcdca65db60df) 2026-03-09T20:52:36.797 INFO:journalctl@ceph.mgr.b.vm06.stdout:Mar 09 20:52:36 vm06 podman[54249]: 2026-03-09 20:52:36.410539813 +0000 UTC m=+0.067451643 container start 355207d5317a091a502516e081024127f73f66af33082f8711fff45a57f9714d (image=quay.ceph.io/ceph-ci/ceph@sha256:8fda260ab1d2d3118a1622f7df75f44f285dfe74e71793626152a711c12bf2cc, name=ceph-9cb345a8-1bf9-11f1-a195-0375563c5891-mgr-b, org.label-schema.license=GPLv2, CEPH_REF=squid, org.opencontainers.image.documentation=https://docs.ceph.com/, org.label-schema.vendor=CentOS, org.opencontainers.image.authors=Ceph Release Team , CEPH_GIT_REPO=https://github.com/ceph/ceph-ci.git, io.buildah.version=1.41.3, FROM_IMAGE=quay.io/centos/centos:stream9, OSD_FLAVOR=default, ceph=True, org.label-schema.build-date=20260223, GANESHA_REPO_BASEURL=https://buildlogs.centos.org/centos/$releasever-stream/storage/$basearch/nfsganesha-5/, org.label-schema.name=CentOS Stream 9 Base Image, CEPH_SHA1=e911bdebe5c8faa3800735d1568fcdca65db60df, org.label-schema.schema-version=1.0) 2026-03-09T20:52:36.797 INFO:journalctl@ceph.mgr.b.vm06.stdout:Mar 09 20:52:36 vm06 bash[54249]: 355207d5317a091a502516e081024127f73f66af33082f8711fff45a57f9714d 2026-03-09T20:52:36.797 INFO:journalctl@ceph.mgr.b.vm06.stdout:Mar 09 20:52:36 vm06 podman[54249]: 2026-03-09 20:52:36.352668538 +0000 UTC m=+0.009580379 image pull 654f31e6858eb235bbece362255b685a945f2b6a367e2b88c4930c984fbb214c quay.ceph.io/ceph-ci/ceph@sha256:8fda260ab1d2d3118a1622f7df75f44f285dfe74e71793626152a711c12bf2cc 2026-03-09T20:52:36.797 INFO:journalctl@ceph.mgr.b.vm06.stdout:Mar 09 20:52:36 vm06 systemd[1]: Started Ceph mgr.b for 9cb345a8-1bf9-11f1-a195-0375563c5891. 2026-03-09T20:52:36.797 INFO:journalctl@ceph.mgr.b.vm06.stdout:Mar 09 20:52:36 vm06 ceph-9cb345a8-1bf9-11f1-a195-0375563c5891-mgr-b[54259]: 2026-03-09T20:52:36.507+0000 7f173a491140 -1 mgr[py] Module alerts has missing NOTIFY_TYPES member 2026-03-09T20:52:36.797 INFO:journalctl@ceph.mgr.b.vm06.stdout:Mar 09 20:52:36 vm06 ceph-9cb345a8-1bf9-11f1-a195-0375563c5891-mgr-b[54259]: 2026-03-09T20:52:36.564+0000 7f173a491140 -1 mgr[py] Module balancer has missing NOTIFY_TYPES member 2026-03-09T20:52:36.836 INFO:teuthology.orchestra.run.vm06.stderr:1+0 records in 2026-03-09T20:52:36.836 INFO:teuthology.orchestra.run.vm06.stderr:1+0 records out 2026-03-09T20:52:36.836 INFO:teuthology.orchestra.run.vm06.stderr:512 bytes copied, 0.000134242 s, 3.8 MB/s 2026-03-09T20:52:36.838 DEBUG:teuthology.orchestra.run.vm06:> ! mount | grep -v devtmpfs | grep -q /dev/vdb 2026-03-09T20:52:36.923 DEBUG:teuthology.orchestra.run.vm06:> stat /dev/vdc 2026-03-09T20:52:36.956 INFO:teuthology.orchestra.run.vm06.stdout: File: /dev/vdc 2026-03-09T20:52:36.957 INFO:teuthology.orchestra.run.vm06.stdout: Size: 0 Blocks: 0 IO Block: 512 block special file 2026-03-09T20:52:36.957 INFO:teuthology.orchestra.run.vm06.stdout:Device: 6h/6d Inode: 254 Links: 1 Device type: fc,20 2026-03-09T20:52:36.957 INFO:teuthology.orchestra.run.vm06.stdout:Access: (0660/brw-rw----) Uid: ( 0/ root) Gid: ( 6/ disk) 2026-03-09T20:52:36.957 INFO:teuthology.orchestra.run.vm06.stdout:Context: system_u:object_r:fixed_disk_device_t:s0 2026-03-09T20:52:36.957 INFO:teuthology.orchestra.run.vm06.stdout:Access: 2026-03-09 20:52:16.686568979 +0000 2026-03-09T20:52:36.957 INFO:teuthology.orchestra.run.vm06.stdout:Modify: 2026-03-09 20:49:12.090437649 +0000 2026-03-09T20:52:36.957 INFO:teuthology.orchestra.run.vm06.stdout:Change: 2026-03-09 20:49:12.090437649 +0000 2026-03-09T20:52:36.957 INFO:teuthology.orchestra.run.vm06.stdout: Birth: 2026-03-09 20:45:03.282000000 +0000 2026-03-09T20:52:36.957 DEBUG:teuthology.orchestra.run.vm06:> sudo dd if=/dev/vdc of=/dev/null count=1 2026-03-09T20:52:36.957 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:52:36 vm08 ceph-mon[54611]: mon.b calling monitor election 2026-03-09T20:52:36.957 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:52:36 vm08 ceph-mon[54611]: from='client.14205 -' entity='client.admin' cmd=[{"prefix": "orch apply", "service_type": "mgr", "placement": "2;vm01=a;vm06=b", "target": ["mon-mgr", ""]}]: dispatch 2026-03-09T20:52:36.958 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:52:36 vm08 ceph-mon[54611]: Saving service mgr spec with placement vm01=a;vm06=b;count:2 2026-03-09T20:52:36.958 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:52:36 vm08 ceph-mon[54611]: Deploying daemon mgr.b on vm06 2026-03-09T20:52:36.958 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:52:36 vm08 ceph-mon[54611]: mon.b calling monitor election 2026-03-09T20:52:36.958 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:52:36 vm08 ceph-mon[54611]: mon.c calling monitor election 2026-03-09T20:52:36.958 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:52:36 vm08 ceph-mon[54611]: mon.a calling monitor election 2026-03-09T20:52:36.958 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:52:36 vm08 ceph-mon[54611]: mon.a is new leader, mons a,c,b in quorum (ranks 0,1,2) 2026-03-09T20:52:36.958 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:52:36 vm08 ceph-mon[54611]: monmap epoch 3 2026-03-09T20:52:36.958 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:52:36 vm08 ceph-mon[54611]: fsid 9cb345a8-1bf9-11f1-a195-0375563c5891 2026-03-09T20:52:36.958 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:52:36 vm08 ceph-mon[54611]: last_changed 2026-03-09T20:52:28.571447+0000 2026-03-09T20:52:36.958 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:52:36 vm08 ceph-mon[54611]: created 2026-03-09T20:51:41.312920+0000 2026-03-09T20:52:36.958 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:52:36 vm08 ceph-mon[54611]: min_mon_release 19 (squid) 2026-03-09T20:52:36.958 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:52:36 vm08 ceph-mon[54611]: election_strategy: 1 2026-03-09T20:52:36.958 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:52:36 vm08 ceph-mon[54611]: 0: [v2:192.168.123.101:3300/0,v1:192.168.123.101:6789/0] mon.a 2026-03-09T20:52:36.958 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:52:36 vm08 ceph-mon[54611]: 1: [v2:192.168.123.108:3300/0,v1:192.168.123.108:6789/0] mon.c 2026-03-09T20:52:36.958 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:52:36 vm08 ceph-mon[54611]: 2: [v2:192.168.123.106:3300/0,v1:192.168.123.106:6789/0] mon.b 2026-03-09T20:52:36.958 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:52:36 vm08 ceph-mon[54611]: fsmap 2026-03-09T20:52:36.958 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:52:36 vm08 ceph-mon[54611]: osdmap e4: 0 total, 0 up, 0 in 2026-03-09T20:52:36.958 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:52:36 vm08 ceph-mon[54611]: mgrmap e13: a(active, since 28s) 2026-03-09T20:52:36.958 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:52:36 vm08 ceph-mon[54611]: overall HEALTH_OK 2026-03-09T20:52:36.958 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:52:36 vm08 ceph-mon[54611]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' 2026-03-09T20:52:36.958 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:52:36 vm08 ceph-mon[54611]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' 2026-03-09T20:52:36.958 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:52:36 vm08 ceph-mon[54611]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' 2026-03-09T20:52:36.958 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:52:36 vm08 ceph-mon[54611]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' 2026-03-09T20:52:36.958 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:52:36 vm08 ceph-mon[54611]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T20:52:37.054 INFO:teuthology.orchestra.run.vm06.stderr:1+0 records in 2026-03-09T20:52:37.054 INFO:teuthology.orchestra.run.vm06.stderr:1+0 records out 2026-03-09T20:52:37.054 INFO:teuthology.orchestra.run.vm06.stderr:512 bytes copied, 0.000722172 s, 709 kB/s 2026-03-09T20:52:37.056 DEBUG:teuthology.orchestra.run.vm06:> ! mount | grep -v devtmpfs | grep -q /dev/vdc 2026-03-09T20:52:37.110 DEBUG:teuthology.orchestra.run.vm06:> stat /dev/vdd 2026-03-09T20:52:37.134 INFO:teuthology.orchestra.run.vm06.stdout: File: /dev/vdd 2026-03-09T20:52:37.134 INFO:teuthology.orchestra.run.vm06.stdout: Size: 0 Blocks: 0 IO Block: 512 block special file 2026-03-09T20:52:37.134 INFO:teuthology.orchestra.run.vm06.stdout:Device: 6h/6d Inode: 256 Links: 1 Device type: fc,30 2026-03-09T20:52:37.134 INFO:teuthology.orchestra.run.vm06.stdout:Access: (0660/brw-rw----) Uid: ( 0/ root) Gid: ( 6/ disk) 2026-03-09T20:52:37.134 INFO:teuthology.orchestra.run.vm06.stdout:Context: system_u:object_r:fixed_disk_device_t:s0 2026-03-09T20:52:37.134 INFO:teuthology.orchestra.run.vm06.stdout:Access: 2026-03-09 20:52:16.715569063 +0000 2026-03-09T20:52:37.134 INFO:teuthology.orchestra.run.vm06.stdout:Modify: 2026-03-09 20:49:12.073437635 +0000 2026-03-09T20:52:37.134 INFO:teuthology.orchestra.run.vm06.stdout:Change: 2026-03-09 20:49:12.073437635 +0000 2026-03-09T20:52:37.134 INFO:teuthology.orchestra.run.vm06.stdout: Birth: 2026-03-09 20:45:03.286000000 +0000 2026-03-09T20:52:37.135 DEBUG:teuthology.orchestra.run.vm06:> sudo dd if=/dev/vdd of=/dev/null count=1 2026-03-09T20:52:37.199 INFO:journalctl@ceph.mgr.b.vm06.stdout:Mar 09 20:52:37 vm06 ceph-9cb345a8-1bf9-11f1-a195-0375563c5891-mgr-b[54259]: 2026-03-09T20:52:37.024+0000 7f173a491140 -1 mgr[py] Module crash has missing NOTIFY_TYPES member 2026-03-09T20:52:37.201 INFO:teuthology.orchestra.run.vm06.stderr:1+0 records in 2026-03-09T20:52:37.201 INFO:teuthology.orchestra.run.vm06.stderr:1+0 records out 2026-03-09T20:52:37.201 INFO:teuthology.orchestra.run.vm06.stderr:512 bytes copied, 0.00016551 s, 3.1 MB/s 2026-03-09T20:52:37.202 DEBUG:teuthology.orchestra.run.vm06:> ! mount | grep -v devtmpfs | grep -q /dev/vdd 2026-03-09T20:52:37.263 DEBUG:teuthology.orchestra.run.vm06:> stat /dev/vde 2026-03-09T20:52:37.320 INFO:teuthology.orchestra.run.vm06.stdout: File: /dev/vde 2026-03-09T20:52:37.320 INFO:teuthology.orchestra.run.vm06.stdout: Size: 0 Blocks: 0 IO Block: 512 block special file 2026-03-09T20:52:37.320 INFO:teuthology.orchestra.run.vm06.stdout:Device: 6h/6d Inode: 257 Links: 1 Device type: fc,40 2026-03-09T20:52:37.320 INFO:teuthology.orchestra.run.vm06.stdout:Access: (0660/brw-rw----) Uid: ( 0/ root) Gid: ( 6/ disk) 2026-03-09T20:52:37.320 INFO:teuthology.orchestra.run.vm06.stdout:Context: system_u:object_r:fixed_disk_device_t:s0 2026-03-09T20:52:37.320 INFO:teuthology.orchestra.run.vm06.stdout:Access: 2026-03-09 20:52:16.744569148 +0000 2026-03-09T20:52:37.320 INFO:teuthology.orchestra.run.vm06.stdout:Modify: 2026-03-09 20:49:12.059437624 +0000 2026-03-09T20:52:37.320 INFO:teuthology.orchestra.run.vm06.stdout:Change: 2026-03-09 20:49:12.059437624 +0000 2026-03-09T20:52:37.320 INFO:teuthology.orchestra.run.vm06.stdout: Birth: 2026-03-09 20:45:03.330000000 +0000 2026-03-09T20:52:37.321 DEBUG:teuthology.orchestra.run.vm06:> sudo dd if=/dev/vde of=/dev/null count=1 2026-03-09T20:52:37.384 INFO:teuthology.orchestra.run.vm06.stderr:1+0 records in 2026-03-09T20:52:37.384 INFO:teuthology.orchestra.run.vm06.stderr:1+0 records out 2026-03-09T20:52:37.384 INFO:teuthology.orchestra.run.vm06.stderr:512 bytes copied, 0.00021369 s, 2.4 MB/s 2026-03-09T20:52:37.385 DEBUG:teuthology.orchestra.run.vm06:> ! mount | grep -v devtmpfs | grep -q /dev/vde 2026-03-09T20:52:37.444 DEBUG:teuthology.orchestra.run.vm08:> set -ex 2026-03-09T20:52:37.444 DEBUG:teuthology.orchestra.run.vm08:> dd if=/scratch_devs of=/dev/stdout 2026-03-09T20:52:37.459 DEBUG:teuthology.orchestra.run:got remote process result: 1 2026-03-09T20:52:37.459 DEBUG:teuthology.orchestra.run.vm08:> ls /dev/[sv]d? 2026-03-09T20:52:37.469 INFO:journalctl@ceph.mgr.b.vm06.stdout:Mar 09 20:52:37 vm06 ceph-9cb345a8-1bf9-11f1-a195-0375563c5891-mgr-b[54259]: 2026-03-09T20:52:37.378+0000 7f173a491140 -1 mgr[py] Module devicehealth has missing NOTIFY_TYPES member 2026-03-09T20:52:37.515 INFO:teuthology.orchestra.run.vm08.stdout:/dev/vda 2026-03-09T20:52:37.515 INFO:teuthology.orchestra.run.vm08.stdout:/dev/vdb 2026-03-09T20:52:37.515 INFO:teuthology.orchestra.run.vm08.stdout:/dev/vdc 2026-03-09T20:52:37.515 INFO:teuthology.orchestra.run.vm08.stdout:/dev/vdd 2026-03-09T20:52:37.515 INFO:teuthology.orchestra.run.vm08.stdout:/dev/vde 2026-03-09T20:52:37.515 WARNING:teuthology.misc:Removing root device: /dev/vda from device list 2026-03-09T20:52:37.515 DEBUG:teuthology.misc:devs=['/dev/vdb', '/dev/vdc', '/dev/vdd', '/dev/vde'] 2026-03-09T20:52:37.515 DEBUG:teuthology.orchestra.run.vm08:> stat /dev/vdb 2026-03-09T20:52:37.574 INFO:teuthology.orchestra.run.vm08.stdout: File: /dev/vdb 2026-03-09T20:52:37.574 INFO:teuthology.orchestra.run.vm08.stdout: Size: 0 Blocks: 0 IO Block: 512 block special file 2026-03-09T20:52:37.574 INFO:teuthology.orchestra.run.vm08.stdout:Device: 6h/6d Inode: 254 Links: 1 Device type: fc,10 2026-03-09T20:52:37.574 INFO:teuthology.orchestra.run.vm08.stdout:Access: (0660/brw-rw----) Uid: ( 0/ root) Gid: ( 6/ disk) 2026-03-09T20:52:37.574 INFO:teuthology.orchestra.run.vm08.stdout:Context: system_u:object_r:fixed_disk_device_t:s0 2026-03-09T20:52:37.574 INFO:teuthology.orchestra.run.vm08.stdout:Access: 2026-03-09 20:52:20.301032167 +0000 2026-03-09T20:52:37.574 INFO:teuthology.orchestra.run.vm08.stdout:Modify: 2026-03-09 20:49:12.247637895 +0000 2026-03-09T20:52:37.574 INFO:teuthology.orchestra.run.vm08.stdout:Change: 2026-03-09 20:49:12.247637895 +0000 2026-03-09T20:52:37.574 INFO:teuthology.orchestra.run.vm08.stdout: Birth: 2026-03-09 20:46:05.258000000 +0000 2026-03-09T20:52:37.574 DEBUG:teuthology.orchestra.run.vm08:> sudo dd if=/dev/vdb of=/dev/null count=1 2026-03-09T20:52:37.637 INFO:teuthology.orchestra.run.vm08.stderr:1+0 records in 2026-03-09T20:52:37.638 INFO:teuthology.orchestra.run.vm08.stderr:1+0 records out 2026-03-09T20:52:37.638 INFO:teuthology.orchestra.run.vm08.stderr:512 bytes copied, 8.559e-05 s, 6.0 MB/s 2026-03-09T20:52:37.639 DEBUG:teuthology.orchestra.run.vm08:> ! mount | grep -v devtmpfs | grep -q /dev/vdb 2026-03-09T20:52:37.674 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:52:37 vm01 ceph-mon[53402]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-09T20:52:37.674 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:52:37 vm01 ceph-mon[53402]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' 2026-03-09T20:52:37.674 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:52:37 vm01 ceph-mon[53402]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' 2026-03-09T20:52:37.674 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:52:37 vm01 ceph-mon[53402]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T20:52:37.674 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:52:37 vm01 ceph-mon[53402]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T20:52:37.674 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:52:37 vm01 ceph-mon[53402]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' 2026-03-09T20:52:37.674 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:52:37 vm01 ceph-mon[53402]: Reconfiguring mgr.a (unknown last config time)... 2026-03-09T20:52:37.674 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:52:37 vm01 ceph-mon[53402]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' cmd=[{"prefix": "auth get-or-create", "entity": "mgr.a", "caps": ["mon", "profile mgr", "osd", "allow *", "mds", "allow *"]}]: dispatch 2026-03-09T20:52:37.674 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:52:37 vm01 ceph-mon[53402]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' cmd=[{"prefix": "mgr services"}]: dispatch 2026-03-09T20:52:37.674 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:52:37 vm01 ceph-mon[53402]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T20:52:37.674 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:52:37 vm01 ceph-mon[53402]: Reconfiguring daemon mgr.a on vm01 2026-03-09T20:52:37.674 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:52:37 vm01 ceph-mon[53402]: pgmap v9: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-09T20:52:37.674 INFO:journalctl@ceph.mgr.a.vm01.stdout:Mar 09 20:52:37 vm01 ceph-9cb345a8-1bf9-11f1-a195-0375563c5891-mgr-a[53639]: 2026-03-09T20:52:37.571+0000 7f0bdb7c1640 -1 mgr.server handle_report got status from non-daemon mon.b 2026-03-09T20:52:37.693 DEBUG:teuthology.orchestra.run.vm08:> stat /dev/vdc 2026-03-09T20:52:37.751 INFO:teuthology.orchestra.run.vm08.stdout: File: /dev/vdc 2026-03-09T20:52:37.751 INFO:teuthology.orchestra.run.vm08.stdout: Size: 0 Blocks: 0 IO Block: 512 block special file 2026-03-09T20:52:37.752 INFO:teuthology.orchestra.run.vm08.stdout:Device: 6h/6d Inode: 255 Links: 1 Device type: fc,20 2026-03-09T20:52:37.752 INFO:teuthology.orchestra.run.vm08.stdout:Access: (0660/brw-rw----) Uid: ( 0/ root) Gid: ( 6/ disk) 2026-03-09T20:52:37.752 INFO:teuthology.orchestra.run.vm08.stdout:Context: system_u:object_r:fixed_disk_device_t:s0 2026-03-09T20:52:37.752 INFO:teuthology.orchestra.run.vm08.stdout:Access: 2026-03-09 20:52:20.328032238 +0000 2026-03-09T20:52:37.752 INFO:teuthology.orchestra.run.vm08.stdout:Modify: 2026-03-09 20:49:12.244637891 +0000 2026-03-09T20:52:37.752 INFO:teuthology.orchestra.run.vm08.stdout:Change: 2026-03-09 20:49:12.244637891 +0000 2026-03-09T20:52:37.752 INFO:teuthology.orchestra.run.vm08.stdout: Birth: 2026-03-09 20:46:05.262000000 +0000 2026-03-09T20:52:37.752 DEBUG:teuthology.orchestra.run.vm08:> sudo dd if=/dev/vdc of=/dev/null count=1 2026-03-09T20:52:37.795 INFO:journalctl@ceph.mgr.b.vm06.stdout:Mar 09 20:52:37 vm06 ceph-9cb345a8-1bf9-11f1-a195-0375563c5891-mgr-b[54259]: /lib64/python3.9/site-packages/scipy/__init__.py:73: UserWarning: NumPy was imported from a Python sub-interpreter but NumPy does not properly support sub-interpreters. This will likely work for most users but might cause hard to track down issues or subtle bugs. A common user of the rare sub-interpreter feature is wsgi which also allows single-interpreter mode. 2026-03-09T20:52:37.795 INFO:journalctl@ceph.mgr.b.vm06.stdout:Mar 09 20:52:37 vm06 ceph-9cb345a8-1bf9-11f1-a195-0375563c5891-mgr-b[54259]: Improvements in the case of bugs are welcome, but is not on the NumPy roadmap, and full support may require significant effort to achieve. 2026-03-09T20:52:37.795 INFO:journalctl@ceph.mgr.b.vm06.stdout:Mar 09 20:52:37 vm06 ceph-9cb345a8-1bf9-11f1-a195-0375563c5891-mgr-b[54259]: from numpy import show_config as show_numpy_config 2026-03-09T20:52:37.796 INFO:journalctl@ceph.mgr.b.vm06.stdout:Mar 09 20:52:37 vm06 ceph-9cb345a8-1bf9-11f1-a195-0375563c5891-mgr-b[54259]: 2026-03-09T20:52:37.472+0000 7f173a491140 -1 mgr[py] Module diskprediction_local has missing NOTIFY_TYPES member 2026-03-09T20:52:37.796 INFO:journalctl@ceph.mgr.b.vm06.stdout:Mar 09 20:52:37 vm06 ceph-9cb345a8-1bf9-11f1-a195-0375563c5891-mgr-b[54259]: 2026-03-09T20:52:37.506+0000 7f173a491140 -1 mgr[py] Module influx has missing NOTIFY_TYPES member 2026-03-09T20:52:37.796 INFO:journalctl@ceph.mgr.b.vm06.stdout:Mar 09 20:52:37 vm06 ceph-9cb345a8-1bf9-11f1-a195-0375563c5891-mgr-b[54259]: 2026-03-09T20:52:37.577+0000 7f173a491140 -1 mgr[py] Module iostat has missing NOTIFY_TYPES member 2026-03-09T20:52:37.796 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 09 20:52:37 vm06 ceph-mon[52939]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-09T20:52:37.796 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 09 20:52:37 vm06 ceph-mon[52939]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' 2026-03-09T20:52:37.796 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 09 20:52:37 vm06 ceph-mon[52939]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' 2026-03-09T20:52:37.796 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 09 20:52:37 vm06 ceph-mon[52939]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T20:52:37.796 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 09 20:52:37 vm06 ceph-mon[52939]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T20:52:37.796 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 09 20:52:37 vm06 ceph-mon[52939]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' 2026-03-09T20:52:37.796 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 09 20:52:37 vm06 ceph-mon[52939]: Reconfiguring mgr.a (unknown last config time)... 2026-03-09T20:52:37.796 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 09 20:52:37 vm06 ceph-mon[52939]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' cmd=[{"prefix": "auth get-or-create", "entity": "mgr.a", "caps": ["mon", "profile mgr", "osd", "allow *", "mds", "allow *"]}]: dispatch 2026-03-09T20:52:37.796 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 09 20:52:37 vm06 ceph-mon[52939]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' cmd=[{"prefix": "mgr services"}]: dispatch 2026-03-09T20:52:37.796 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 09 20:52:37 vm06 ceph-mon[52939]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T20:52:37.796 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 09 20:52:37 vm06 ceph-mon[52939]: Reconfiguring daemon mgr.a on vm01 2026-03-09T20:52:37.796 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 09 20:52:37 vm06 ceph-mon[52939]: pgmap v9: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-09T20:52:37.817 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:52:37 vm08 ceph-mon[54611]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-09T20:52:37.817 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:52:37 vm08 ceph-mon[54611]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' 2026-03-09T20:52:37.817 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:52:37 vm08 ceph-mon[54611]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' 2026-03-09T20:52:37.817 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:52:37 vm08 ceph-mon[54611]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T20:52:37.817 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:52:37 vm08 ceph-mon[54611]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T20:52:37.817 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:52:37 vm08 ceph-mon[54611]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' 2026-03-09T20:52:37.817 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:52:37 vm08 ceph-mon[54611]: Reconfiguring mgr.a (unknown last config time)... 2026-03-09T20:52:37.817 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:52:37 vm08 ceph-mon[54611]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' cmd=[{"prefix": "auth get-or-create", "entity": "mgr.a", "caps": ["mon", "profile mgr", "osd", "allow *", "mds", "allow *"]}]: dispatch 2026-03-09T20:52:37.817 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:52:37 vm08 ceph-mon[54611]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' cmd=[{"prefix": "mgr services"}]: dispatch 2026-03-09T20:52:37.817 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:52:37 vm08 ceph-mon[54611]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T20:52:37.817 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:52:37 vm08 ceph-mon[54611]: Reconfiguring daemon mgr.a on vm01 2026-03-09T20:52:37.817 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:52:37 vm08 ceph-mon[54611]: pgmap v9: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-09T20:52:37.819 INFO:teuthology.orchestra.run.vm08.stderr:1+0 records in 2026-03-09T20:52:37.819 INFO:teuthology.orchestra.run.vm08.stderr:1+0 records out 2026-03-09T20:52:37.819 INFO:teuthology.orchestra.run.vm08.stderr:512 bytes copied, 0.000116838 s, 4.4 MB/s 2026-03-09T20:52:37.820 DEBUG:teuthology.orchestra.run.vm08:> ! mount | grep -v devtmpfs | grep -q /dev/vdc 2026-03-09T20:52:37.878 DEBUG:teuthology.orchestra.run.vm08:> stat /dev/vdd 2026-03-09T20:52:37.935 INFO:teuthology.orchestra.run.vm08.stdout: File: /dev/vdd 2026-03-09T20:52:37.935 INFO:teuthology.orchestra.run.vm08.stdout: Size: 0 Blocks: 0 IO Block: 512 block special file 2026-03-09T20:52:37.935 INFO:teuthology.orchestra.run.vm08.stdout:Device: 6h/6d Inode: 256 Links: 1 Device type: fc,30 2026-03-09T20:52:37.935 INFO:teuthology.orchestra.run.vm08.stdout:Access: (0660/brw-rw----) Uid: ( 0/ root) Gid: ( 6/ disk) 2026-03-09T20:52:37.935 INFO:teuthology.orchestra.run.vm08.stdout:Context: system_u:object_r:fixed_disk_device_t:s0 2026-03-09T20:52:37.935 INFO:teuthology.orchestra.run.vm08.stdout:Access: 2026-03-09 20:52:20.355032310 +0000 2026-03-09T20:52:37.935 INFO:teuthology.orchestra.run.vm08.stdout:Modify: 2026-03-09 20:49:12.242637888 +0000 2026-03-09T20:52:37.935 INFO:teuthology.orchestra.run.vm08.stdout:Change: 2026-03-09 20:49:12.242637888 +0000 2026-03-09T20:52:37.935 INFO:teuthology.orchestra.run.vm08.stdout: Birth: 2026-03-09 20:46:05.265000000 +0000 2026-03-09T20:52:37.935 DEBUG:teuthology.orchestra.run.vm08:> sudo dd if=/dev/vdd of=/dev/null count=1 2026-03-09T20:52:37.998 INFO:teuthology.orchestra.run.vm08.stderr:1+0 records in 2026-03-09T20:52:37.998 INFO:teuthology.orchestra.run.vm08.stderr:1+0 records out 2026-03-09T20:52:37.998 INFO:teuthology.orchestra.run.vm08.stderr:512 bytes copied, 9.0459e-05 s, 5.7 MB/s 2026-03-09T20:52:37.999 DEBUG:teuthology.orchestra.run.vm08:> ! mount | grep -v devtmpfs | grep -q /dev/vdd 2026-03-09T20:52:38.055 DEBUG:teuthology.orchestra.run.vm08:> stat /dev/vde 2026-03-09T20:52:38.113 INFO:teuthology.orchestra.run.vm08.stdout: File: /dev/vde 2026-03-09T20:52:38.113 INFO:teuthology.orchestra.run.vm08.stdout: Size: 0 Blocks: 0 IO Block: 512 block special file 2026-03-09T20:52:38.113 INFO:teuthology.orchestra.run.vm08.stdout:Device: 6h/6d Inode: 257 Links: 1 Device type: fc,40 2026-03-09T20:52:38.113 INFO:teuthology.orchestra.run.vm08.stdout:Access: (0660/brw-rw----) Uid: ( 0/ root) Gid: ( 6/ disk) 2026-03-09T20:52:38.113 INFO:teuthology.orchestra.run.vm08.stdout:Context: system_u:object_r:fixed_disk_device_t:s0 2026-03-09T20:52:38.113 INFO:teuthology.orchestra.run.vm08.stdout:Access: 2026-03-09 20:52:20.395032415 +0000 2026-03-09T20:52:38.113 INFO:teuthology.orchestra.run.vm08.stdout:Modify: 2026-03-09 20:49:12.248637897 +0000 2026-03-09T20:52:38.113 INFO:teuthology.orchestra.run.vm08.stdout:Change: 2026-03-09 20:49:12.248637897 +0000 2026-03-09T20:52:38.113 INFO:teuthology.orchestra.run.vm08.stdout: Birth: 2026-03-09 20:46:05.272000000 +0000 2026-03-09T20:52:38.113 DEBUG:teuthology.orchestra.run.vm08:> sudo dd if=/dev/vde of=/dev/null count=1 2026-03-09T20:52:38.175 INFO:teuthology.orchestra.run.vm08.stderr:1+0 records in 2026-03-09T20:52:38.175 INFO:teuthology.orchestra.run.vm08.stderr:1+0 records out 2026-03-09T20:52:38.175 INFO:teuthology.orchestra.run.vm08.stderr:512 bytes copied, 0.000113613 s, 4.5 MB/s 2026-03-09T20:52:38.176 DEBUG:teuthology.orchestra.run.vm08:> ! mount | grep -v devtmpfs | grep -q /dev/vde 2026-03-09T20:52:38.232 INFO:tasks.cephadm:Deploying osd.0 on vm01 with /dev/vde... 2026-03-09T20:52:38.232 DEBUG:teuthology.orchestra.run.vm01:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df ceph-volume -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 9cb345a8-1bf9-11f1-a195-0375563c5891 -- lvm zap /dev/vde 2026-03-09T20:52:38.329 INFO:journalctl@ceph.mgr.b.vm06.stdout:Mar 09 20:52:38 vm06 ceph-9cb345a8-1bf9-11f1-a195-0375563c5891-mgr-b[54259]: 2026-03-09T20:52:38.073+0000 7f173a491140 -1 mgr[py] Module nfs has missing NOTIFY_TYPES member 2026-03-09T20:52:38.329 INFO:journalctl@ceph.mgr.b.vm06.stdout:Mar 09 20:52:38 vm06 ceph-9cb345a8-1bf9-11f1-a195-0375563c5891-mgr-b[54259]: 2026-03-09T20:52:38.181+0000 7f173a491140 -1 mgr[py] Module orchestrator has missing NOTIFY_TYPES member 2026-03-09T20:52:38.329 INFO:journalctl@ceph.mgr.b.vm06.stdout:Mar 09 20:52:38 vm06 ceph-9cb345a8-1bf9-11f1-a195-0375563c5891-mgr-b[54259]: 2026-03-09T20:52:38.219+0000 7f173a491140 -1 mgr[py] Module osd_perf_query has missing NOTIFY_TYPES member 2026-03-09T20:52:38.329 INFO:journalctl@ceph.mgr.b.vm06.stdout:Mar 09 20:52:38 vm06 ceph-9cb345a8-1bf9-11f1-a195-0375563c5891-mgr-b[54259]: 2026-03-09T20:52:38.253+0000 7f173a491140 -1 mgr[py] Module osd_support has missing NOTIFY_TYPES member 2026-03-09T20:52:38.329 INFO:journalctl@ceph.mgr.b.vm06.stdout:Mar 09 20:52:38 vm06 ceph-9cb345a8-1bf9-11f1-a195-0375563c5891-mgr-b[54259]: 2026-03-09T20:52:38.292+0000 7f173a491140 -1 mgr[py] Module pg_autoscaler has missing NOTIFY_TYPES member 2026-03-09T20:52:38.396 INFO:teuthology.orchestra.run.vm01.stderr:Inferring config /var/lib/ceph/9cb345a8-1bf9-11f1-a195-0375563c5891/mon.a/config 2026-03-09T20:52:38.589 INFO:journalctl@ceph.mgr.b.vm06.stdout:Mar 09 20:52:38 vm06 ceph-9cb345a8-1bf9-11f1-a195-0375563c5891-mgr-b[54259]: 2026-03-09T20:52:38.327+0000 7f173a491140 -1 mgr[py] Module progress has missing NOTIFY_TYPES member 2026-03-09T20:52:38.590 INFO:journalctl@ceph.mgr.b.vm06.stdout:Mar 09 20:52:38 vm06 ceph-9cb345a8-1bf9-11f1-a195-0375563c5891-mgr-b[54259]: 2026-03-09T20:52:38.492+0000 7f173a491140 -1 mgr[py] Module prometheus has missing NOTIFY_TYPES member 2026-03-09T20:52:38.590 INFO:journalctl@ceph.mgr.b.vm06.stdout:Mar 09 20:52:38 vm06 ceph-9cb345a8-1bf9-11f1-a195-0375563c5891-mgr-b[54259]: 2026-03-09T20:52:38.545+0000 7f173a491140 -1 mgr[py] Module rbd_support has missing NOTIFY_TYPES member 2026-03-09T20:52:38.908 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:52:38 vm01 ceph-mon[53402]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' 2026-03-09T20:52:38.908 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:52:38 vm01 ceph-mon[53402]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' 2026-03-09T20:52:38.908 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:52:38 vm01 ceph-mon[53402]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T20:52:38.908 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:52:38 vm01 ceph-mon[53402]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T20:52:38.908 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:52:38 vm01 ceph-mon[53402]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T20:52:38.908 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:52:38 vm01 ceph-mon[53402]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' 2026-03-09T20:52:38.957 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:52:38 vm08 ceph-mon[54611]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' 2026-03-09T20:52:38.957 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:52:38 vm08 ceph-mon[54611]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' 2026-03-09T20:52:38.957 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:52:38 vm08 ceph-mon[54611]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T20:52:38.957 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:52:38 vm08 ceph-mon[54611]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T20:52:38.957 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:52:38 vm08 ceph-mon[54611]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T20:52:38.957 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:52:38 vm08 ceph-mon[54611]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' 2026-03-09T20:52:39.045 INFO:journalctl@ceph.mgr.b.vm06.stdout:Mar 09 20:52:38 vm06 ceph-9cb345a8-1bf9-11f1-a195-0375563c5891-mgr-b[54259]: 2026-03-09T20:52:38.790+0000 7f173a491140 -1 mgr[py] Module rgw has missing NOTIFY_TYPES member 2026-03-09T20:52:39.046 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 09 20:52:38 vm06 ceph-mon[52939]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' 2026-03-09T20:52:39.046 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 09 20:52:38 vm06 ceph-mon[52939]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' 2026-03-09T20:52:39.046 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 09 20:52:38 vm06 ceph-mon[52939]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T20:52:39.046 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 09 20:52:38 vm06 ceph-mon[52939]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T20:52:39.046 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 09 20:52:38 vm06 ceph-mon[52939]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T20:52:39.046 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 09 20:52:38 vm06 ceph-mon[52939]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' 2026-03-09T20:52:39.348 INFO:journalctl@ceph.mgr.b.vm06.stdout:Mar 09 20:52:39 vm06 ceph-9cb345a8-1bf9-11f1-a195-0375563c5891-mgr-b[54259]: 2026-03-09T20:52:39.064+0000 7f173a491140 -1 mgr[py] Module rook has missing NOTIFY_TYPES member 2026-03-09T20:52:39.348 INFO:journalctl@ceph.mgr.b.vm06.stdout:Mar 09 20:52:39 vm06 ceph-9cb345a8-1bf9-11f1-a195-0375563c5891-mgr-b[54259]: 2026-03-09T20:52:39.102+0000 7f173a491140 -1 mgr[py] Module selftest has missing NOTIFY_TYPES member 2026-03-09T20:52:39.348 INFO:journalctl@ceph.mgr.b.vm06.stdout:Mar 09 20:52:39 vm06 ceph-9cb345a8-1bf9-11f1-a195-0375563c5891-mgr-b[54259]: 2026-03-09T20:52:39.144+0000 7f173a491140 -1 mgr[py] Module snap_schedule has missing NOTIFY_TYPES member 2026-03-09T20:52:39.348 INFO:journalctl@ceph.mgr.b.vm06.stdout:Mar 09 20:52:39 vm06 ceph-9cb345a8-1bf9-11f1-a195-0375563c5891-mgr-b[54259]: 2026-03-09T20:52:39.224+0000 7f173a491140 -1 mgr[py] Module status has missing NOTIFY_TYPES member 2026-03-09T20:52:39.348 INFO:journalctl@ceph.mgr.b.vm06.stdout:Mar 09 20:52:39 vm06 ceph-9cb345a8-1bf9-11f1-a195-0375563c5891-mgr-b[54259]: 2026-03-09T20:52:39.263+0000 7f173a491140 -1 mgr[py] Module telegraf has missing NOTIFY_TYPES member 2026-03-09T20:52:39.411 INFO:teuthology.orchestra.run.vm01.stdout: 2026-03-09T20:52:39.429 DEBUG:teuthology.orchestra.run.vm01:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 9cb345a8-1bf9-11f1-a195-0375563c5891 -- ceph orch daemon add osd vm01:/dev/vde 2026-03-09T20:52:39.605 INFO:journalctl@ceph.mgr.b.vm06.stdout:Mar 09 20:52:39 vm06 ceph-9cb345a8-1bf9-11f1-a195-0375563c5891-mgr-b[54259]: 2026-03-09T20:52:39.346+0000 7f173a491140 -1 mgr[py] Module telemetry has missing NOTIFY_TYPES member 2026-03-09T20:52:39.605 INFO:journalctl@ceph.mgr.b.vm06.stdout:Mar 09 20:52:39 vm06 ceph-9cb345a8-1bf9-11f1-a195-0375563c5891-mgr-b[54259]: 2026-03-09T20:52:39.462+0000 7f173a491140 -1 mgr[py] Module test_orchestrator has missing NOTIFY_TYPES member 2026-03-09T20:52:39.609 INFO:teuthology.orchestra.run.vm01.stderr:Inferring config /var/lib/ceph/9cb345a8-1bf9-11f1-a195-0375563c5891/mon.a/config 2026-03-09T20:52:39.617 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:52:39 vm01 ceph-mon[53402]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' 2026-03-09T20:52:39.617 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:52:39 vm01 ceph-mon[53402]: pgmap v10: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-09T20:52:39.957 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:52:39 vm08 ceph-mon[54611]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' 2026-03-09T20:52:39.957 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:52:39 vm08 ceph-mon[54611]: pgmap v10: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-09T20:52:40.046 INFO:journalctl@ceph.mgr.b.vm06.stdout:Mar 09 20:52:39 vm06 ceph-9cb345a8-1bf9-11f1-a195-0375563c5891-mgr-b[54259]: 2026-03-09T20:52:39.603+0000 7f173a491140 -1 mgr[py] Module volumes has missing NOTIFY_TYPES member 2026-03-09T20:52:40.046 INFO:journalctl@ceph.mgr.b.vm06.stdout:Mar 09 20:52:39 vm06 ceph-9cb345a8-1bf9-11f1-a195-0375563c5891-mgr-b[54259]: 2026-03-09T20:52:39.646+0000 7f173a491140 -1 mgr[py] Module zabbix has missing NOTIFY_TYPES member 2026-03-09T20:52:40.046 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 09 20:52:39 vm06 ceph-mon[52939]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' 2026-03-09T20:52:40.046 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 09 20:52:39 vm06 ceph-mon[52939]: pgmap v10: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-09T20:52:40.834 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:52:40 vm01 ceph-mon[53402]: Standby manager daemon b started 2026-03-09T20:52:40.834 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:52:40 vm01 ceph-mon[53402]: from='mgr.? 192.168.123.106:0/256216871' entity='mgr.b' cmd=[{"prefix": "config-key get", "key": "mgr/dashboard/b/crt"}]: dispatch 2026-03-09T20:52:40.834 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:52:40 vm01 ceph-mon[53402]: from='mgr.? 192.168.123.106:0/256216871' entity='mgr.b' cmd=[{"prefix": "config-key get", "key": "mgr/dashboard/crt"}]: dispatch 2026-03-09T20:52:40.834 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:52:40 vm01 ceph-mon[53402]: from='mgr.? 192.168.123.106:0/256216871' entity='mgr.b' cmd=[{"prefix": "config-key get", "key": "mgr/dashboard/b/key"}]: dispatch 2026-03-09T20:52:40.834 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:52:40 vm01 ceph-mon[53402]: from='mgr.? 192.168.123.106:0/256216871' entity='mgr.b' cmd=[{"prefix": "config-key get", "key": "mgr/dashboard/key"}]: dispatch 2026-03-09T20:52:40.834 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:52:40 vm01 ceph-mon[53402]: from='client.24100 -' entity='client.admin' cmd=[{"prefix": "orch daemon add osd", "svc_arg": "vm01:/dev/vde", "target": ["mon-mgr", ""]}]: dispatch 2026-03-09T20:52:40.834 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:52:40 vm01 ceph-mon[53402]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' cmd=[{"prefix": "osd tree", "states": ["destroyed"], "format": "json"}]: dispatch 2026-03-09T20:52:40.834 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:52:40 vm01 ceph-mon[53402]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.bootstrap-osd"}]: dispatch 2026-03-09T20:52:40.834 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:52:40 vm01 ceph-mon[53402]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T20:52:40.957 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:52:40 vm08 ceph-mon[54611]: Standby manager daemon b started 2026-03-09T20:52:40.957 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:52:40 vm08 ceph-mon[54611]: from='mgr.? 192.168.123.106:0/256216871' entity='mgr.b' cmd=[{"prefix": "config-key get", "key": "mgr/dashboard/b/crt"}]: dispatch 2026-03-09T20:52:40.957 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:52:40 vm08 ceph-mon[54611]: from='mgr.? 192.168.123.106:0/256216871' entity='mgr.b' cmd=[{"prefix": "config-key get", "key": "mgr/dashboard/crt"}]: dispatch 2026-03-09T20:52:40.957 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:52:40 vm08 ceph-mon[54611]: from='mgr.? 192.168.123.106:0/256216871' entity='mgr.b' cmd=[{"prefix": "config-key get", "key": "mgr/dashboard/b/key"}]: dispatch 2026-03-09T20:52:40.957 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:52:40 vm08 ceph-mon[54611]: from='mgr.? 192.168.123.106:0/256216871' entity='mgr.b' cmd=[{"prefix": "config-key get", "key": "mgr/dashboard/key"}]: dispatch 2026-03-09T20:52:40.957 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:52:40 vm08 ceph-mon[54611]: from='client.24100 -' entity='client.admin' cmd=[{"prefix": "orch daemon add osd", "svc_arg": "vm01:/dev/vde", "target": ["mon-mgr", ""]}]: dispatch 2026-03-09T20:52:40.957 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:52:40 vm08 ceph-mon[54611]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' cmd=[{"prefix": "osd tree", "states": ["destroyed"], "format": "json"}]: dispatch 2026-03-09T20:52:40.957 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:52:40 vm08 ceph-mon[54611]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.bootstrap-osd"}]: dispatch 2026-03-09T20:52:40.957 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:52:40 vm08 ceph-mon[54611]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T20:52:41.046 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 09 20:52:40 vm06 ceph-mon[52939]: Standby manager daemon b started 2026-03-09T20:52:41.046 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 09 20:52:40 vm06 ceph-mon[52939]: from='mgr.? 192.168.123.106:0/256216871' entity='mgr.b' cmd=[{"prefix": "config-key get", "key": "mgr/dashboard/b/crt"}]: dispatch 2026-03-09T20:52:41.046 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 09 20:52:40 vm06 ceph-mon[52939]: from='mgr.? 192.168.123.106:0/256216871' entity='mgr.b' cmd=[{"prefix": "config-key get", "key": "mgr/dashboard/crt"}]: dispatch 2026-03-09T20:52:41.046 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 09 20:52:40 vm06 ceph-mon[52939]: from='mgr.? 192.168.123.106:0/256216871' entity='mgr.b' cmd=[{"prefix": "config-key get", "key": "mgr/dashboard/b/key"}]: dispatch 2026-03-09T20:52:41.046 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 09 20:52:40 vm06 ceph-mon[52939]: from='mgr.? 192.168.123.106:0/256216871' entity='mgr.b' cmd=[{"prefix": "config-key get", "key": "mgr/dashboard/key"}]: dispatch 2026-03-09T20:52:41.046 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 09 20:52:40 vm06 ceph-mon[52939]: from='client.24100 -' entity='client.admin' cmd=[{"prefix": "orch daemon add osd", "svc_arg": "vm01:/dev/vde", "target": ["mon-mgr", ""]}]: dispatch 2026-03-09T20:52:41.046 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 09 20:52:40 vm06 ceph-mon[52939]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' cmd=[{"prefix": "osd tree", "states": ["destroyed"], "format": "json"}]: dispatch 2026-03-09T20:52:41.046 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 09 20:52:40 vm06 ceph-mon[52939]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.bootstrap-osd"}]: dispatch 2026-03-09T20:52:41.046 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 09 20:52:40 vm06 ceph-mon[52939]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T20:52:41.924 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:52:41 vm01 ceph-mon[53402]: mgrmap e14: a(active, since 33s), standbys: b 2026-03-09T20:52:41.924 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:52:41 vm01 ceph-mon[53402]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' cmd=[{"prefix": "mgr metadata", "who": "b", "id": "b"}]: dispatch 2026-03-09T20:52:41.924 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:52:41 vm01 ceph-mon[53402]: from='client.? 192.168.123.101:0/3512101599' entity='client.bootstrap-osd' cmd=[{"prefix": "osd new", "uuid": "58efc62d-2ef3-4005-88e4-4cf11f5576fe"}]: dispatch 2026-03-09T20:52:41.924 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:52:41 vm01 ceph-mon[53402]: from='client.? 192.168.123.101:0/3512101599' entity='client.bootstrap-osd' cmd='[{"prefix": "osd new", "uuid": "58efc62d-2ef3-4005-88e4-4cf11f5576fe"}]': finished 2026-03-09T20:52:41.924 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:52:41 vm01 ceph-mon[53402]: osdmap e5: 1 total, 0 up, 1 in 2026-03-09T20:52:41.924 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:52:41 vm01 ceph-mon[53402]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 0}]: dispatch 2026-03-09T20:52:41.924 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:52:41 vm01 ceph-mon[53402]: from='client.? 192.168.123.101:0/4012395384' entity='client.bootstrap-osd' cmd=[{"prefix": "mon getmap"}]: dispatch 2026-03-09T20:52:41.924 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:52:41 vm01 ceph-mon[53402]: pgmap v12: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-09T20:52:41.957 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:52:41 vm08 ceph-mon[54611]: mgrmap e14: a(active, since 33s), standbys: b 2026-03-09T20:52:41.957 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:52:41 vm08 ceph-mon[54611]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' cmd=[{"prefix": "mgr metadata", "who": "b", "id": "b"}]: dispatch 2026-03-09T20:52:41.957 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:52:41 vm08 ceph-mon[54611]: from='client.? 192.168.123.101:0/3512101599' entity='client.bootstrap-osd' cmd=[{"prefix": "osd new", "uuid": "58efc62d-2ef3-4005-88e4-4cf11f5576fe"}]: dispatch 2026-03-09T20:52:41.957 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:52:41 vm08 ceph-mon[54611]: from='client.? 192.168.123.101:0/3512101599' entity='client.bootstrap-osd' cmd='[{"prefix": "osd new", "uuid": "58efc62d-2ef3-4005-88e4-4cf11f5576fe"}]': finished 2026-03-09T20:52:41.957 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:52:41 vm08 ceph-mon[54611]: osdmap e5: 1 total, 0 up, 1 in 2026-03-09T20:52:41.957 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:52:41 vm08 ceph-mon[54611]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 0}]: dispatch 2026-03-09T20:52:41.957 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:52:41 vm08 ceph-mon[54611]: from='client.? 192.168.123.101:0/4012395384' entity='client.bootstrap-osd' cmd=[{"prefix": "mon getmap"}]: dispatch 2026-03-09T20:52:41.957 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:52:41 vm08 ceph-mon[54611]: pgmap v12: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-09T20:52:42.045 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 09 20:52:41 vm06 ceph-mon[52939]: mgrmap e14: a(active, since 33s), standbys: b 2026-03-09T20:52:42.045 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 09 20:52:41 vm06 ceph-mon[52939]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' cmd=[{"prefix": "mgr metadata", "who": "b", "id": "b"}]: dispatch 2026-03-09T20:52:42.045 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 09 20:52:41 vm06 ceph-mon[52939]: from='client.? 192.168.123.101:0/3512101599' entity='client.bootstrap-osd' cmd=[{"prefix": "osd new", "uuid": "58efc62d-2ef3-4005-88e4-4cf11f5576fe"}]: dispatch 2026-03-09T20:52:42.046 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 09 20:52:41 vm06 ceph-mon[52939]: from='client.? 192.168.123.101:0/3512101599' entity='client.bootstrap-osd' cmd='[{"prefix": "osd new", "uuid": "58efc62d-2ef3-4005-88e4-4cf11f5576fe"}]': finished 2026-03-09T20:52:42.046 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 09 20:52:41 vm06 ceph-mon[52939]: osdmap e5: 1 total, 0 up, 1 in 2026-03-09T20:52:42.046 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 09 20:52:41 vm06 ceph-mon[52939]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 0}]: dispatch 2026-03-09T20:52:42.046 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 09 20:52:41 vm06 ceph-mon[52939]: from='client.? 192.168.123.101:0/4012395384' entity='client.bootstrap-osd' cmd=[{"prefix": "mon getmap"}]: dispatch 2026-03-09T20:52:42.046 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 09 20:52:41 vm06 ceph-mon[52939]: pgmap v12: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-09T20:52:43.796 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 09 20:52:43 vm06 ceph-mon[52939]: pgmap v13: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-09T20:52:43.924 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:52:43 vm01 ceph-mon[53402]: pgmap v13: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-09T20:52:43.957 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:52:43 vm08 ceph-mon[54611]: pgmap v13: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-09T20:52:45.517 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:52:45 vm01 ceph-mon[53402]: pgmap v14: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-09T20:52:45.796 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 09 20:52:45 vm06 ceph-mon[52939]: pgmap v14: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-09T20:52:45.957 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:52:45 vm08 ceph-mon[54611]: pgmap v14: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-09T20:52:46.748 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:52:46 vm01 ceph-mon[53402]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "osd.0"}]: dispatch 2026-03-09T20:52:46.748 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:52:46 vm01 ceph-mon[53402]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T20:52:46.748 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:52:46 vm01 ceph-mon[53402]: Deploying daemon osd.0 on vm01 2026-03-09T20:52:46.795 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 09 20:52:46 vm06 ceph-mon[52939]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "osd.0"}]: dispatch 2026-03-09T20:52:46.796 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 09 20:52:46 vm06 ceph-mon[52939]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T20:52:46.796 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 09 20:52:46 vm06 ceph-mon[52939]: Deploying daemon osd.0 on vm01 2026-03-09T20:52:46.957 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:52:46 vm08 ceph-mon[54611]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "osd.0"}]: dispatch 2026-03-09T20:52:46.957 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:52:46 vm08 ceph-mon[54611]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T20:52:46.957 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:52:46 vm08 ceph-mon[54611]: Deploying daemon osd.0 on vm01 2026-03-09T20:52:47.795 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 09 20:52:47 vm06 ceph-mon[52939]: pgmap v15: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-09T20:52:47.822 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:52:47 vm01 ceph-mon[53402]: pgmap v15: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-09T20:52:47.957 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:52:47 vm08 ceph-mon[54611]: pgmap v15: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-09T20:52:48.674 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:52:48 vm01 ceph-mon[53402]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T20:52:48.674 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:52:48 vm01 ceph-mon[53402]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' 2026-03-09T20:52:48.674 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:52:48 vm01 ceph-mon[53402]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' 2026-03-09T20:52:48.795 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 09 20:52:48 vm06 ceph-mon[52939]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T20:52:48.796 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 09 20:52:48 vm06 ceph-mon[52939]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' 2026-03-09T20:52:48.796 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 09 20:52:48 vm06 ceph-mon[52939]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' 2026-03-09T20:52:48.957 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:52:48 vm08 ceph-mon[54611]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T20:52:48.957 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:52:48 vm08 ceph-mon[54611]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' 2026-03-09T20:52:48.957 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:52:48 vm08 ceph-mon[54611]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' 2026-03-09T20:52:49.546 INFO:teuthology.orchestra.run.vm01.stdout:Created osd(s) 0 on host 'vm01' 2026-03-09T20:52:49.706 DEBUG:teuthology.orchestra.run.vm01:osd.0> sudo journalctl -f -n 0 -u ceph-9cb345a8-1bf9-11f1-a195-0375563c5891@osd.0.service 2026-03-09T20:52:49.708 INFO:tasks.cephadm:Deploying osd.1 on vm06 with /dev/vde... 2026-03-09T20:52:49.708 DEBUG:teuthology.orchestra.run.vm06:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df ceph-volume -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 9cb345a8-1bf9-11f1-a195-0375563c5891 -- lvm zap /dev/vde 2026-03-09T20:52:49.933 INFO:teuthology.orchestra.run.vm06.stderr:Inferring config /var/lib/ceph/9cb345a8-1bf9-11f1-a195-0375563c5891/mon.b/config 2026-03-09T20:52:49.980 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 09 20:52:49 vm06 ceph-mon[52939]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' 2026-03-09T20:52:49.980 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 09 20:52:49 vm06 ceph-mon[52939]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' 2026-03-09T20:52:49.980 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 09 20:52:49 vm06 ceph-mon[52939]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T20:52:49.980 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 09 20:52:49 vm06 ceph-mon[52939]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T20:52:49.980 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 09 20:52:49 vm06 ceph-mon[52939]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' 2026-03-09T20:52:49.980 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 09 20:52:49 vm06 ceph-mon[52939]: pgmap v16: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-09T20:52:49.980 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 09 20:52:49 vm06 ceph-mon[52939]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T20:52:49.980 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 09 20:52:49 vm06 ceph-mon[52939]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' 2026-03-09T20:52:49.980 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 09 20:52:49 vm06 ceph-mon[52939]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' 2026-03-09T20:52:49.989 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:52:49 vm01 ceph-mon[53402]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' 2026-03-09T20:52:49.989 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:52:49 vm01 ceph-mon[53402]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' 2026-03-09T20:52:49.989 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:52:49 vm01 ceph-mon[53402]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T20:52:49.989 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:52:49 vm01 ceph-mon[53402]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T20:52:49.989 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:52:49 vm01 ceph-mon[53402]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' 2026-03-09T20:52:49.989 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:52:49 vm01 ceph-mon[53402]: pgmap v16: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-09T20:52:49.989 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:52:49 vm01 ceph-mon[53402]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T20:52:49.989 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:52:49 vm01 ceph-mon[53402]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' 2026-03-09T20:52:49.989 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:52:49 vm01 ceph-mon[53402]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' 2026-03-09T20:52:49.990 INFO:journalctl@ceph.osd.0.vm01.stdout:Mar 09 20:52:49 vm01 ceph-9cb345a8-1bf9-11f1-a195-0375563c5891-osd-0[64225]: 2026-03-09T20:52:49.819+0000 7fd60b125740 -1 osd.0 0 log_to_monitors true 2026-03-09T20:52:50.207 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:52:49 vm08 ceph-mon[54611]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' 2026-03-09T20:52:50.207 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:52:49 vm08 ceph-mon[54611]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' 2026-03-09T20:52:50.207 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:52:49 vm08 ceph-mon[54611]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T20:52:50.207 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:52:49 vm08 ceph-mon[54611]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T20:52:50.207 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:52:49 vm08 ceph-mon[54611]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' 2026-03-09T20:52:50.207 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:52:49 vm08 ceph-mon[54611]: pgmap v16: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-09T20:52:50.207 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:52:49 vm08 ceph-mon[54611]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T20:52:50.207 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:52:49 vm08 ceph-mon[54611]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' 2026-03-09T20:52:50.207 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:52:49 vm08 ceph-mon[54611]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' 2026-03-09T20:52:50.924 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:52:50 vm01 ceph-mon[53402]: from='osd.0 [v2:192.168.123.101:6802/3447435909,v1:192.168.123.101:6803/3447435909]' entity='osd.0' cmd=[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["0"]}]: dispatch 2026-03-09T20:52:50.924 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:52:50 vm01 ceph-mon[53402]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' 2026-03-09T20:52:50.924 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:52:50 vm01 ceph-mon[53402]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' 2026-03-09T20:52:50.924 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:52:50 vm01 ceph-mon[53402]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' cmd=[{"prefix": "config rm", "who": "osd.0", "name": "osd_memory_target"}]: dispatch 2026-03-09T20:52:50.924 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:52:50 vm01 ceph-mon[53402]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T20:52:50.924 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:52:50 vm01 ceph-mon[53402]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T20:52:50.924 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:52:50 vm01 ceph-mon[53402]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' 2026-03-09T20:52:51.064 INFO:teuthology.orchestra.run.vm06.stdout: 2026-03-09T20:52:51.086 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 09 20:52:50 vm06 ceph-mon[52939]: from='osd.0 [v2:192.168.123.101:6802/3447435909,v1:192.168.123.101:6803/3447435909]' entity='osd.0' cmd=[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["0"]}]: dispatch 2026-03-09T20:52:51.086 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 09 20:52:50 vm06 ceph-mon[52939]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' 2026-03-09T20:52:51.086 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 09 20:52:50 vm06 ceph-mon[52939]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' 2026-03-09T20:52:51.086 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 09 20:52:50 vm06 ceph-mon[52939]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' cmd=[{"prefix": "config rm", "who": "osd.0", "name": "osd_memory_target"}]: dispatch 2026-03-09T20:52:51.086 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 09 20:52:50 vm06 ceph-mon[52939]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T20:52:51.086 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 09 20:52:50 vm06 ceph-mon[52939]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T20:52:51.086 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 09 20:52:50 vm06 ceph-mon[52939]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' 2026-03-09T20:52:51.086 DEBUG:teuthology.orchestra.run.vm06:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 9cb345a8-1bf9-11f1-a195-0375563c5891 -- ceph orch daemon add osd vm06:/dev/vde 2026-03-09T20:52:51.207 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:52:50 vm08 ceph-mon[54611]: from='osd.0 [v2:192.168.123.101:6802/3447435909,v1:192.168.123.101:6803/3447435909]' entity='osd.0' cmd=[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["0"]}]: dispatch 2026-03-09T20:52:51.207 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:52:50 vm08 ceph-mon[54611]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' 2026-03-09T20:52:51.207 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:52:50 vm08 ceph-mon[54611]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' 2026-03-09T20:52:51.207 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:52:50 vm08 ceph-mon[54611]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' cmd=[{"prefix": "config rm", "who": "osd.0", "name": "osd_memory_target"}]: dispatch 2026-03-09T20:52:51.207 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:52:50 vm08 ceph-mon[54611]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T20:52:51.207 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:52:50 vm08 ceph-mon[54611]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T20:52:51.207 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:52:50 vm08 ceph-mon[54611]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' 2026-03-09T20:52:51.267 INFO:teuthology.orchestra.run.vm06.stderr:Inferring config /var/lib/ceph/9cb345a8-1bf9-11f1-a195-0375563c5891/mon.b/config 2026-03-09T20:52:52.091 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 09 20:52:51 vm06 ceph-mon[52939]: Detected new or changed devices on vm01 2026-03-09T20:52:52.091 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 09 20:52:51 vm06 ceph-mon[52939]: Adjusting osd_memory_target on vm01 to 257.0M 2026-03-09T20:52:52.091 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 09 20:52:51 vm06 ceph-mon[52939]: Unable to set osd_memory_target on vm01 to 269530726: error parsing value: Value '269530726' is below minimum 939524096 2026-03-09T20:52:52.091 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 09 20:52:51 vm06 ceph-mon[52939]: from='osd.0 [v2:192.168.123.101:6802/3447435909,v1:192.168.123.101:6803/3447435909]' entity='osd.0' cmd='[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["0"]}]': finished 2026-03-09T20:52:52.091 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 09 20:52:51 vm06 ceph-mon[52939]: osdmap e6: 1 total, 0 up, 1 in 2026-03-09T20:52:52.091 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 09 20:52:51 vm06 ceph-mon[52939]: from='osd.0 [v2:192.168.123.101:6802/3447435909,v1:192.168.123.101:6803/3447435909]' entity='osd.0' cmd=[{"prefix": "osd crush create-or-move", "id": 0, "weight":0.0195, "args": ["host=vm01", "root=default"]}]: dispatch 2026-03-09T20:52:52.091 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 09 20:52:51 vm06 ceph-mon[52939]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 0}]: dispatch 2026-03-09T20:52:52.091 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 09 20:52:51 vm06 ceph-mon[52939]: pgmap v18: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-09T20:52:52.091 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 09 20:52:51 vm06 ceph-mon[52939]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' cmd=[{"prefix": "osd tree", "states": ["destroyed"], "format": "json"}]: dispatch 2026-03-09T20:52:52.091 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 09 20:52:51 vm06 ceph-mon[52939]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.bootstrap-osd"}]: dispatch 2026-03-09T20:52:52.091 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 09 20:52:51 vm06 ceph-mon[52939]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T20:52:52.173 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:52:51 vm01 ceph-mon[53402]: Detected new or changed devices on vm01 2026-03-09T20:52:52.174 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:52:51 vm01 ceph-mon[53402]: Adjusting osd_memory_target on vm01 to 257.0M 2026-03-09T20:52:52.174 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:52:51 vm01 ceph-mon[53402]: Unable to set osd_memory_target on vm01 to 269530726: error parsing value: Value '269530726' is below minimum 939524096 2026-03-09T20:52:52.174 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:52:51 vm01 ceph-mon[53402]: from='osd.0 [v2:192.168.123.101:6802/3447435909,v1:192.168.123.101:6803/3447435909]' entity='osd.0' cmd='[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["0"]}]': finished 2026-03-09T20:52:52.174 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:52:51 vm01 ceph-mon[53402]: osdmap e6: 1 total, 0 up, 1 in 2026-03-09T20:52:52.174 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:52:51 vm01 ceph-mon[53402]: from='osd.0 [v2:192.168.123.101:6802/3447435909,v1:192.168.123.101:6803/3447435909]' entity='osd.0' cmd=[{"prefix": "osd crush create-or-move", "id": 0, "weight":0.0195, "args": ["host=vm01", "root=default"]}]: dispatch 2026-03-09T20:52:52.174 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:52:51 vm01 ceph-mon[53402]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 0}]: dispatch 2026-03-09T20:52:52.174 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:52:51 vm01 ceph-mon[53402]: pgmap v18: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-09T20:52:52.174 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:52:51 vm01 ceph-mon[53402]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' cmd=[{"prefix": "osd tree", "states": ["destroyed"], "format": "json"}]: dispatch 2026-03-09T20:52:52.174 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:52:51 vm01 ceph-mon[53402]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.bootstrap-osd"}]: dispatch 2026-03-09T20:52:52.174 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:52:51 vm01 ceph-mon[53402]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T20:52:52.174 INFO:journalctl@ceph.osd.0.vm01.stdout:Mar 09 20:52:51 vm01 ceph-9cb345a8-1bf9-11f1-a195-0375563c5891-osd-0[64225]: 2026-03-09T20:52:51.832+0000 7fd6070a6640 -1 osd.0 0 waiting for initial osdmap 2026-03-09T20:52:52.174 INFO:journalctl@ceph.osd.0.vm01.stdout:Mar 09 20:52:51 vm01 ceph-9cb345a8-1bf9-11f1-a195-0375563c5891-osd-0[64225]: 2026-03-09T20:52:51.837+0000 7fd602ed0640 -1 osd.0 7 set_numa_affinity unable to identify public interface '' numa node: (2) No such file or directory 2026-03-09T20:52:52.207 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:52:51 vm08 ceph-mon[54611]: Detected new or changed devices on vm01 2026-03-09T20:52:52.207 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:52:51 vm08 ceph-mon[54611]: Adjusting osd_memory_target on vm01 to 257.0M 2026-03-09T20:52:52.207 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:52:51 vm08 ceph-mon[54611]: Unable to set osd_memory_target on vm01 to 269530726: error parsing value: Value '269530726' is below minimum 939524096 2026-03-09T20:52:52.207 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:52:51 vm08 ceph-mon[54611]: from='osd.0 [v2:192.168.123.101:6802/3447435909,v1:192.168.123.101:6803/3447435909]' entity='osd.0' cmd='[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["0"]}]': finished 2026-03-09T20:52:52.207 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:52:51 vm08 ceph-mon[54611]: osdmap e6: 1 total, 0 up, 1 in 2026-03-09T20:52:52.207 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:52:51 vm08 ceph-mon[54611]: from='osd.0 [v2:192.168.123.101:6802/3447435909,v1:192.168.123.101:6803/3447435909]' entity='osd.0' cmd=[{"prefix": "osd crush create-or-move", "id": 0, "weight":0.0195, "args": ["host=vm01", "root=default"]}]: dispatch 2026-03-09T20:52:52.207 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:52:51 vm08 ceph-mon[54611]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 0}]: dispatch 2026-03-09T20:52:52.207 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:52:51 vm08 ceph-mon[54611]: pgmap v18: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-09T20:52:52.207 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:52:51 vm08 ceph-mon[54611]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' cmd=[{"prefix": "osd tree", "states": ["destroyed"], "format": "json"}]: dispatch 2026-03-09T20:52:52.207 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:52:51 vm08 ceph-mon[54611]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.bootstrap-osd"}]: dispatch 2026-03-09T20:52:52.207 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:52:51 vm08 ceph-mon[54611]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T20:52:53.046 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 09 20:52:52 vm06 ceph-mon[52939]: purged_snaps scrub starts 2026-03-09T20:52:53.046 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 09 20:52:52 vm06 ceph-mon[52939]: purged_snaps scrub ok 2026-03-09T20:52:53.046 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 09 20:52:52 vm06 ceph-mon[52939]: from='client.24125 -' entity='client.admin' cmd=[{"prefix": "orch daemon add osd", "svc_arg": "vm06:/dev/vde", "target": ["mon-mgr", ""]}]: dispatch 2026-03-09T20:52:53.046 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 09 20:52:52 vm06 ceph-mon[52939]: from='osd.0 [v2:192.168.123.101:6802/3447435909,v1:192.168.123.101:6803/3447435909]' entity='osd.0' cmd='[{"prefix": "osd crush create-or-move", "id": 0, "weight":0.0195, "args": ["host=vm01", "root=default"]}]': finished 2026-03-09T20:52:53.046 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 09 20:52:52 vm06 ceph-mon[52939]: osdmap e7: 1 total, 0 up, 1 in 2026-03-09T20:52:53.046 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 09 20:52:52 vm06 ceph-mon[52939]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 0}]: dispatch 2026-03-09T20:52:53.046 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 09 20:52:52 vm06 ceph-mon[52939]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 0}]: dispatch 2026-03-09T20:52:53.046 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 09 20:52:52 vm06 ceph-mon[52939]: from='client.? 192.168.123.106:0/1785544513' entity='client.bootstrap-osd' cmd=[{"prefix": "osd new", "uuid": "38f01073-940b-498e-91d3-7a23e98bcb53"}]: dispatch 2026-03-09T20:52:53.046 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 09 20:52:52 vm06 ceph-mon[52939]: from='client.? ' entity='client.bootstrap-osd' cmd=[{"prefix": "osd new", "uuid": "38f01073-940b-498e-91d3-7a23e98bcb53"}]: dispatch 2026-03-09T20:52:53.046 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 09 20:52:52 vm06 ceph-mon[52939]: from='client.? ' entity='client.bootstrap-osd' cmd='[{"prefix": "osd new", "uuid": "38f01073-940b-498e-91d3-7a23e98bcb53"}]': finished 2026-03-09T20:52:53.046 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 09 20:52:52 vm06 ceph-mon[52939]: osd.0 [v2:192.168.123.101:6802/3447435909,v1:192.168.123.101:6803/3447435909] boot 2026-03-09T20:52:53.046 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 09 20:52:52 vm06 ceph-mon[52939]: osdmap e8: 2 total, 1 up, 2 in 2026-03-09T20:52:53.046 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 09 20:52:52 vm06 ceph-mon[52939]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 0}]: dispatch 2026-03-09T20:52:53.046 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 09 20:52:52 vm06 ceph-mon[52939]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 1}]: dispatch 2026-03-09T20:52:53.173 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:52:52 vm01 ceph-mon[53402]: purged_snaps scrub starts 2026-03-09T20:52:53.173 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:52:52 vm01 ceph-mon[53402]: purged_snaps scrub ok 2026-03-09T20:52:53.173 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:52:52 vm01 ceph-mon[53402]: from='client.24125 -' entity='client.admin' cmd=[{"prefix": "orch daemon add osd", "svc_arg": "vm06:/dev/vde", "target": ["mon-mgr", ""]}]: dispatch 2026-03-09T20:52:53.173 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:52:52 vm01 ceph-mon[53402]: from='osd.0 [v2:192.168.123.101:6802/3447435909,v1:192.168.123.101:6803/3447435909]' entity='osd.0' cmd='[{"prefix": "osd crush create-or-move", "id": 0, "weight":0.0195, "args": ["host=vm01", "root=default"]}]': finished 2026-03-09T20:52:53.174 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:52:52 vm01 ceph-mon[53402]: osdmap e7: 1 total, 0 up, 1 in 2026-03-09T20:52:53.174 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:52:52 vm01 ceph-mon[53402]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 0}]: dispatch 2026-03-09T20:52:53.174 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:52:52 vm01 ceph-mon[53402]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 0}]: dispatch 2026-03-09T20:52:53.174 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:52:52 vm01 ceph-mon[53402]: from='client.? 192.168.123.106:0/1785544513' entity='client.bootstrap-osd' cmd=[{"prefix": "osd new", "uuid": "38f01073-940b-498e-91d3-7a23e98bcb53"}]: dispatch 2026-03-09T20:52:53.174 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:52:52 vm01 ceph-mon[53402]: from='client.? ' entity='client.bootstrap-osd' cmd=[{"prefix": "osd new", "uuid": "38f01073-940b-498e-91d3-7a23e98bcb53"}]: dispatch 2026-03-09T20:52:53.174 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:52:52 vm01 ceph-mon[53402]: from='client.? ' entity='client.bootstrap-osd' cmd='[{"prefix": "osd new", "uuid": "38f01073-940b-498e-91d3-7a23e98bcb53"}]': finished 2026-03-09T20:52:53.174 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:52:52 vm01 ceph-mon[53402]: osd.0 [v2:192.168.123.101:6802/3447435909,v1:192.168.123.101:6803/3447435909] boot 2026-03-09T20:52:53.174 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:52:52 vm01 ceph-mon[53402]: osdmap e8: 2 total, 1 up, 2 in 2026-03-09T20:52:53.174 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:52:52 vm01 ceph-mon[53402]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 0}]: dispatch 2026-03-09T20:52:53.174 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:52:52 vm01 ceph-mon[53402]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 1}]: dispatch 2026-03-09T20:52:53.207 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:52:52 vm08 ceph-mon[54611]: purged_snaps scrub starts 2026-03-09T20:52:53.207 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:52:52 vm08 ceph-mon[54611]: purged_snaps scrub ok 2026-03-09T20:52:53.207 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:52:52 vm08 ceph-mon[54611]: from='client.24125 -' entity='client.admin' cmd=[{"prefix": "orch daemon add osd", "svc_arg": "vm06:/dev/vde", "target": ["mon-mgr", ""]}]: dispatch 2026-03-09T20:52:53.207 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:52:52 vm08 ceph-mon[54611]: from='osd.0 [v2:192.168.123.101:6802/3447435909,v1:192.168.123.101:6803/3447435909]' entity='osd.0' cmd='[{"prefix": "osd crush create-or-move", "id": 0, "weight":0.0195, "args": ["host=vm01", "root=default"]}]': finished 2026-03-09T20:52:53.207 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:52:52 vm08 ceph-mon[54611]: osdmap e7: 1 total, 0 up, 1 in 2026-03-09T20:52:53.207 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:52:52 vm08 ceph-mon[54611]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 0}]: dispatch 2026-03-09T20:52:53.207 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:52:52 vm08 ceph-mon[54611]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 0}]: dispatch 2026-03-09T20:52:53.207 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:52:52 vm08 ceph-mon[54611]: from='client.? 192.168.123.106:0/1785544513' entity='client.bootstrap-osd' cmd=[{"prefix": "osd new", "uuid": "38f01073-940b-498e-91d3-7a23e98bcb53"}]: dispatch 2026-03-09T20:52:53.207 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:52:52 vm08 ceph-mon[54611]: from='client.? ' entity='client.bootstrap-osd' cmd=[{"prefix": "osd new", "uuid": "38f01073-940b-498e-91d3-7a23e98bcb53"}]: dispatch 2026-03-09T20:52:53.207 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:52:52 vm08 ceph-mon[54611]: from='client.? ' entity='client.bootstrap-osd' cmd='[{"prefix": "osd new", "uuid": "38f01073-940b-498e-91d3-7a23e98bcb53"}]': finished 2026-03-09T20:52:53.207 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:52:52 vm08 ceph-mon[54611]: osd.0 [v2:192.168.123.101:6802/3447435909,v1:192.168.123.101:6803/3447435909] boot 2026-03-09T20:52:53.207 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:52:52 vm08 ceph-mon[54611]: osdmap e8: 2 total, 1 up, 2 in 2026-03-09T20:52:53.207 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:52:52 vm08 ceph-mon[54611]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 0}]: dispatch 2026-03-09T20:52:53.207 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:52:52 vm08 ceph-mon[54611]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 1}]: dispatch 2026-03-09T20:52:54.173 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:52:53 vm01 ceph-mon[53402]: from='client.? 192.168.123.106:0/2529297670' entity='client.bootstrap-osd' cmd=[{"prefix": "mon getmap"}]: dispatch 2026-03-09T20:52:54.174 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:52:53 vm01 ceph-mon[53402]: pgmap v21: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-09T20:52:54.174 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:52:53 vm01 ceph-mon[53402]: osdmap e9: 2 total, 1 up, 2 in 2026-03-09T20:52:54.174 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:52:53 vm01 ceph-mon[53402]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 1}]: dispatch 2026-03-09T20:52:54.207 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:52:53 vm08 ceph-mon[54611]: from='client.? 192.168.123.106:0/2529297670' entity='client.bootstrap-osd' cmd=[{"prefix": "mon getmap"}]: dispatch 2026-03-09T20:52:54.207 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:52:53 vm08 ceph-mon[54611]: pgmap v21: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-09T20:52:54.207 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:52:53 vm08 ceph-mon[54611]: osdmap e9: 2 total, 1 up, 2 in 2026-03-09T20:52:54.207 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:52:53 vm08 ceph-mon[54611]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 1}]: dispatch 2026-03-09T20:52:54.295 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 09 20:52:53 vm06 ceph-mon[52939]: from='client.? 192.168.123.106:0/2529297670' entity='client.bootstrap-osd' cmd=[{"prefix": "mon getmap"}]: dispatch 2026-03-09T20:52:54.296 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 09 20:52:53 vm06 ceph-mon[52939]: pgmap v21: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-09T20:52:54.296 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 09 20:52:53 vm06 ceph-mon[52939]: osdmap e9: 2 total, 1 up, 2 in 2026-03-09T20:52:54.296 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 09 20:52:53 vm06 ceph-mon[52939]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 1}]: dispatch 2026-03-09T20:52:55.795 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 09 20:52:55 vm06 ceph-mon[52939]: pgmap v23: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-09T20:52:55.923 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:52:55 vm01 ceph-mon[53402]: pgmap v23: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-09T20:52:55.957 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:52:55 vm08 ceph-mon[54611]: pgmap v23: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-09T20:52:57.534 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 09 20:52:57 vm06 ceph-mon[52939]: pgmap v24: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-09T20:52:57.923 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:52:57 vm01 ceph-mon[53402]: pgmap v24: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-09T20:52:57.957 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:52:57 vm08 ceph-mon[54611]: pgmap v24: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-09T20:52:58.680 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 09 20:52:58 vm06 ceph-mon[52939]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "osd.1"}]: dispatch 2026-03-09T20:52:58.681 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 09 20:52:58 vm06 ceph-mon[52939]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T20:52:58.681 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 09 20:52:58 vm06 ceph-mon[52939]: Deploying daemon osd.1 on vm06 2026-03-09T20:52:58.923 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:52:58 vm01 ceph-mon[53402]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "osd.1"}]: dispatch 2026-03-09T20:52:58.923 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:52:58 vm01 ceph-mon[53402]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T20:52:58.924 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:52:58 vm01 ceph-mon[53402]: Deploying daemon osd.1 on vm06 2026-03-09T20:52:58.957 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:52:58 vm08 ceph-mon[54611]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "osd.1"}]: dispatch 2026-03-09T20:52:58.957 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:52:58 vm08 ceph-mon[54611]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T20:52:58.957 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:52:58 vm08 ceph-mon[54611]: Deploying daemon osd.1 on vm06 2026-03-09T20:52:59.650 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 09 20:52:59 vm06 ceph-mon[52939]: pgmap v25: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-09T20:52:59.923 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:52:59 vm01 ceph-mon[53402]: pgmap v25: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-09T20:52:59.957 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:52:59 vm08 ceph-mon[54611]: pgmap v25: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-09T20:53:00.795 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 09 20:53:00 vm06 ceph-mon[52939]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T20:53:00.796 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 09 20:53:00 vm06 ceph-mon[52939]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' 2026-03-09T20:53:00.796 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 09 20:53:00 vm06 ceph-mon[52939]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' 2026-03-09T20:53:00.923 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:53:00 vm01 ceph-mon[53402]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T20:53:00.923 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:53:00 vm01 ceph-mon[53402]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' 2026-03-09T20:53:00.923 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:53:00 vm01 ceph-mon[53402]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' 2026-03-09T20:53:00.957 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:53:00 vm08 ceph-mon[54611]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T20:53:00.957 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:53:00 vm08 ceph-mon[54611]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' 2026-03-09T20:53:00.957 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:53:00 vm08 ceph-mon[54611]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' 2026-03-09T20:53:01.239 INFO:teuthology.orchestra.run.vm06.stdout:Created osd(s) 1 on host 'vm06' 2026-03-09T20:53:01.404 DEBUG:teuthology.orchestra.run.vm06:osd.1> sudo journalctl -f -n 0 -u ceph-9cb345a8-1bf9-11f1-a195-0375563c5891@osd.1.service 2026-03-09T20:53:01.406 INFO:tasks.cephadm:Deploying osd.2 on vm08 with /dev/vde... 2026-03-09T20:53:01.406 DEBUG:teuthology.orchestra.run.vm08:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df ceph-volume -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 9cb345a8-1bf9-11f1-a195-0375563c5891 -- lvm zap /dev/vde 2026-03-09T20:53:01.581 INFO:teuthology.orchestra.run.vm08.stderr:Inferring config /var/lib/ceph/9cb345a8-1bf9-11f1-a195-0375563c5891/mon.c/config 2026-03-09T20:53:01.693 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 09 20:53:01 vm06 ceph-mon[52939]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' 2026-03-09T20:53:01.693 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 09 20:53:01 vm06 ceph-mon[52939]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' 2026-03-09T20:53:01.693 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 09 20:53:01 vm06 ceph-mon[52939]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T20:53:01.693 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 09 20:53:01 vm06 ceph-mon[52939]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T20:53:01.693 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 09 20:53:01 vm06 ceph-mon[52939]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' 2026-03-09T20:53:01.693 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 09 20:53:01 vm06 ceph-mon[52939]: from='osd.1 [v2:192.168.123.106:6800/3453001158,v1:192.168.123.106:6801/3453001158]' entity='osd.1' cmd=[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["1"]}]: dispatch 2026-03-09T20:53:01.693 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 09 20:53:01 vm06 ceph-mon[52939]: from='osd.1 ' entity='osd.1' cmd=[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["1"]}]: dispatch 2026-03-09T20:53:01.693 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 09 20:53:01 vm06 ceph-mon[52939]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T20:53:01.693 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 09 20:53:01 vm06 ceph-mon[52939]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' 2026-03-09T20:53:01.693 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 09 20:53:01 vm06 ceph-mon[52939]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' 2026-03-09T20:53:01.693 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 09 20:53:01 vm06 ceph-mon[52939]: pgmap v26: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-09T20:53:01.702 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:53:01 vm08 ceph-mon[54611]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' 2026-03-09T20:53:01.702 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:53:01 vm08 ceph-mon[54611]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' 2026-03-09T20:53:01.702 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:53:01 vm08 ceph-mon[54611]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T20:53:01.702 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:53:01 vm08 ceph-mon[54611]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T20:53:01.702 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:53:01 vm08 ceph-mon[54611]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' 2026-03-09T20:53:01.702 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:53:01 vm08 ceph-mon[54611]: from='osd.1 [v2:192.168.123.106:6800/3453001158,v1:192.168.123.106:6801/3453001158]' entity='osd.1' cmd=[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["1"]}]: dispatch 2026-03-09T20:53:01.702 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:53:01 vm08 ceph-mon[54611]: from='osd.1 ' entity='osd.1' cmd=[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["1"]}]: dispatch 2026-03-09T20:53:01.702 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:53:01 vm08 ceph-mon[54611]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T20:53:01.702 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:53:01 vm08 ceph-mon[54611]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' 2026-03-09T20:53:01.702 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:53:01 vm08 ceph-mon[54611]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' 2026-03-09T20:53:01.702 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:53:01 vm08 ceph-mon[54611]: pgmap v26: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-09T20:53:01.924 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:53:01 vm01 ceph-mon[53402]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' 2026-03-09T20:53:01.924 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:53:01 vm01 ceph-mon[53402]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' 2026-03-09T20:53:01.924 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:53:01 vm01 ceph-mon[53402]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T20:53:01.924 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:53:01 vm01 ceph-mon[53402]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T20:53:01.924 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:53:01 vm01 ceph-mon[53402]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' 2026-03-09T20:53:01.924 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:53:01 vm01 ceph-mon[53402]: from='osd.1 [v2:192.168.123.106:6800/3453001158,v1:192.168.123.106:6801/3453001158]' entity='osd.1' cmd=[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["1"]}]: dispatch 2026-03-09T20:53:01.924 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:53:01 vm01 ceph-mon[53402]: from='osd.1 ' entity='osd.1' cmd=[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["1"]}]: dispatch 2026-03-09T20:53:01.924 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:53:01 vm01 ceph-mon[53402]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T20:53:01.924 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:53:01 vm01 ceph-mon[53402]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' 2026-03-09T20:53:01.924 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:53:01 vm01 ceph-mon[53402]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' 2026-03-09T20:53:01.924 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:53:01 vm01 ceph-mon[53402]: pgmap v26: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-09T20:53:02.612 INFO:teuthology.orchestra.run.vm08.stdout: 2026-03-09T20:53:02.633 DEBUG:teuthology.orchestra.run.vm08:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 9cb345a8-1bf9-11f1-a195-0375563c5891 -- ceph orch daemon add osd vm08:/dev/vde 2026-03-09T20:53:02.659 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:53:02 vm08 ceph-mon[54611]: from='osd.1 ' entity='osd.1' cmd='[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["1"]}]': finished 2026-03-09T20:53:02.659 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:53:02 vm08 ceph-mon[54611]: osdmap e10: 2 total, 1 up, 2 in 2026-03-09T20:53:02.659 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:53:02 vm08 ceph-mon[54611]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 1}]: dispatch 2026-03-09T20:53:02.659 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:53:02 vm08 ceph-mon[54611]: from='osd.1 [v2:192.168.123.106:6800/3453001158,v1:192.168.123.106:6801/3453001158]' entity='osd.1' cmd=[{"prefix": "osd crush create-or-move", "id": 1, "weight":0.0195, "args": ["host=vm06", "root=default"]}]: dispatch 2026-03-09T20:53:02.659 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:53:02 vm08 ceph-mon[54611]: from='osd.1 ' entity='osd.1' cmd=[{"prefix": "osd crush create-or-move", "id": 1, "weight":0.0195, "args": ["host=vm06", "root=default"]}]: dispatch 2026-03-09T20:53:02.659 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:53:02 vm08 ceph-mon[54611]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' 2026-03-09T20:53:02.659 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:53:02 vm08 ceph-mon[54611]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' 2026-03-09T20:53:02.659 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:53:02 vm08 ceph-mon[54611]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' cmd=[{"prefix": "config rm", "who": "osd.1", "name": "osd_memory_target"}]: dispatch 2026-03-09T20:53:02.659 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:53:02 vm08 ceph-mon[54611]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T20:53:02.659 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:53:02 vm08 ceph-mon[54611]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T20:53:02.659 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:53:02 vm08 ceph-mon[54611]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' 2026-03-09T20:53:02.809 INFO:teuthology.orchestra.run.vm08.stderr:Inferring config /var/lib/ceph/9cb345a8-1bf9-11f1-a195-0375563c5891/mon.c/config 2026-03-09T20:53:02.924 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:53:02 vm01 ceph-mon[53402]: from='osd.1 ' entity='osd.1' cmd='[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["1"]}]': finished 2026-03-09T20:53:02.924 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:53:02 vm01 ceph-mon[53402]: osdmap e10: 2 total, 1 up, 2 in 2026-03-09T20:53:02.924 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:53:02 vm01 ceph-mon[53402]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 1}]: dispatch 2026-03-09T20:53:02.924 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:53:02 vm01 ceph-mon[53402]: from='osd.1 [v2:192.168.123.106:6800/3453001158,v1:192.168.123.106:6801/3453001158]' entity='osd.1' cmd=[{"prefix": "osd crush create-or-move", "id": 1, "weight":0.0195, "args": ["host=vm06", "root=default"]}]: dispatch 2026-03-09T20:53:02.924 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:53:02 vm01 ceph-mon[53402]: from='osd.1 ' entity='osd.1' cmd=[{"prefix": "osd crush create-or-move", "id": 1, "weight":0.0195, "args": ["host=vm06", "root=default"]}]: dispatch 2026-03-09T20:53:02.924 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:53:02 vm01 ceph-mon[53402]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' 2026-03-09T20:53:02.924 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:53:02 vm01 ceph-mon[53402]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' 2026-03-09T20:53:02.924 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:53:02 vm01 ceph-mon[53402]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' cmd=[{"prefix": "config rm", "who": "osd.1", "name": "osd_memory_target"}]: dispatch 2026-03-09T20:53:02.924 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:53:02 vm01 ceph-mon[53402]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T20:53:02.924 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:53:02 vm01 ceph-mon[53402]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T20:53:02.924 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:53:02 vm01 ceph-mon[53402]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' 2026-03-09T20:53:03.046 INFO:journalctl@ceph.osd.1.vm06.stdout:Mar 09 20:53:02 vm06 ceph-9cb345a8-1bf9-11f1-a195-0375563c5891-osd-1[57192]: 2026-03-09T20:53:02.599+0000 7f3ef5641640 -1 osd.1 0 waiting for initial osdmap 2026-03-09T20:53:03.046 INFO:journalctl@ceph.osd.1.vm06.stdout:Mar 09 20:53:02 vm06 ceph-9cb345a8-1bf9-11f1-a195-0375563c5891-osd-1[57192]: 2026-03-09T20:53:02.603+0000 7f3ef146b640 -1 osd.1 11 set_numa_affinity unable to identify public interface '' numa node: (2) No such file or directory 2026-03-09T20:53:03.046 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 09 20:53:02 vm06 ceph-mon[52939]: from='osd.1 ' entity='osd.1' cmd='[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["1"]}]': finished 2026-03-09T20:53:03.046 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 09 20:53:02 vm06 ceph-mon[52939]: osdmap e10: 2 total, 1 up, 2 in 2026-03-09T20:53:03.046 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 09 20:53:02 vm06 ceph-mon[52939]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 1}]: dispatch 2026-03-09T20:53:03.046 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 09 20:53:02 vm06 ceph-mon[52939]: from='osd.1 [v2:192.168.123.106:6800/3453001158,v1:192.168.123.106:6801/3453001158]' entity='osd.1' cmd=[{"prefix": "osd crush create-or-move", "id": 1, "weight":0.0195, "args": ["host=vm06", "root=default"]}]: dispatch 2026-03-09T20:53:03.046 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 09 20:53:02 vm06 ceph-mon[52939]: from='osd.1 ' entity='osd.1' cmd=[{"prefix": "osd crush create-or-move", "id": 1, "weight":0.0195, "args": ["host=vm06", "root=default"]}]: dispatch 2026-03-09T20:53:03.046 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 09 20:53:02 vm06 ceph-mon[52939]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' 2026-03-09T20:53:03.046 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 09 20:53:02 vm06 ceph-mon[52939]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' 2026-03-09T20:53:03.046 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 09 20:53:02 vm06 ceph-mon[52939]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' cmd=[{"prefix": "config rm", "who": "osd.1", "name": "osd_memory_target"}]: dispatch 2026-03-09T20:53:03.046 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 09 20:53:02 vm06 ceph-mon[52939]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T20:53:03.046 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 09 20:53:02 vm06 ceph-mon[52939]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T20:53:03.046 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 09 20:53:02 vm06 ceph-mon[52939]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' 2026-03-09T20:53:03.778 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:53:03 vm08 ceph-mon[54611]: Detected new or changed devices on vm06 2026-03-09T20:53:03.778 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:53:03 vm08 ceph-mon[54611]: Adjusting osd_memory_target on vm06 to 257.0M 2026-03-09T20:53:03.778 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:53:03 vm08 ceph-mon[54611]: Unable to set osd_memory_target on vm06 to 269536460: error parsing value: Value '269536460' is below minimum 939524096 2026-03-09T20:53:03.778 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:53:03 vm08 ceph-mon[54611]: from='osd.1 ' entity='osd.1' cmd='[{"prefix": "osd crush create-or-move", "id": 1, "weight":0.0195, "args": ["host=vm06", "root=default"]}]': finished 2026-03-09T20:53:03.778 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:53:03 vm08 ceph-mon[54611]: osdmap e11: 2 total, 1 up, 2 in 2026-03-09T20:53:03.778 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:53:03 vm08 ceph-mon[54611]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 1}]: dispatch 2026-03-09T20:53:03.778 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:53:03 vm08 ceph-mon[54611]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 1}]: dispatch 2026-03-09T20:53:03.778 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:53:03 vm08 ceph-mon[54611]: from='client.14271 -' entity='client.admin' cmd=[{"prefix": "orch daemon add osd", "svc_arg": "vm08:/dev/vde", "target": ["mon-mgr", ""]}]: dispatch 2026-03-09T20:53:03.778 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:53:03 vm08 ceph-mon[54611]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' cmd=[{"prefix": "osd tree", "states": ["destroyed"], "format": "json"}]: dispatch 2026-03-09T20:53:03.778 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:53:03 vm08 ceph-mon[54611]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.bootstrap-osd"}]: dispatch 2026-03-09T20:53:03.778 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:53:03 vm08 ceph-mon[54611]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T20:53:03.778 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:53:03 vm08 ceph-mon[54611]: pgmap v29: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-09T20:53:03.924 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:53:03 vm01 ceph-mon[53402]: Detected new or changed devices on vm06 2026-03-09T20:53:03.924 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:53:03 vm01 ceph-mon[53402]: Adjusting osd_memory_target on vm06 to 257.0M 2026-03-09T20:53:03.924 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:53:03 vm01 ceph-mon[53402]: Unable to set osd_memory_target on vm06 to 269536460: error parsing value: Value '269536460' is below minimum 939524096 2026-03-09T20:53:03.924 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:53:03 vm01 ceph-mon[53402]: from='osd.1 ' entity='osd.1' cmd='[{"prefix": "osd crush create-or-move", "id": 1, "weight":0.0195, "args": ["host=vm06", "root=default"]}]': finished 2026-03-09T20:53:03.924 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:53:03 vm01 ceph-mon[53402]: osdmap e11: 2 total, 1 up, 2 in 2026-03-09T20:53:03.924 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:53:03 vm01 ceph-mon[53402]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 1}]: dispatch 2026-03-09T20:53:03.924 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:53:03 vm01 ceph-mon[53402]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 1}]: dispatch 2026-03-09T20:53:03.924 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:53:03 vm01 ceph-mon[53402]: from='client.14271 -' entity='client.admin' cmd=[{"prefix": "orch daemon add osd", "svc_arg": "vm08:/dev/vde", "target": ["mon-mgr", ""]}]: dispatch 2026-03-09T20:53:03.924 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:53:03 vm01 ceph-mon[53402]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' cmd=[{"prefix": "osd tree", "states": ["destroyed"], "format": "json"}]: dispatch 2026-03-09T20:53:03.924 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:53:03 vm01 ceph-mon[53402]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.bootstrap-osd"}]: dispatch 2026-03-09T20:53:03.924 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:53:03 vm01 ceph-mon[53402]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T20:53:03.924 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:53:03 vm01 ceph-mon[53402]: pgmap v29: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-09T20:53:04.046 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 09 20:53:03 vm06 ceph-mon[52939]: Detected new or changed devices on vm06 2026-03-09T20:53:04.046 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 09 20:53:03 vm06 ceph-mon[52939]: Adjusting osd_memory_target on vm06 to 257.0M 2026-03-09T20:53:04.046 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 09 20:53:03 vm06 ceph-mon[52939]: Unable to set osd_memory_target on vm06 to 269536460: error parsing value: Value '269536460' is below minimum 939524096 2026-03-09T20:53:04.046 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 09 20:53:03 vm06 ceph-mon[52939]: from='osd.1 ' entity='osd.1' cmd='[{"prefix": "osd crush create-or-move", "id": 1, "weight":0.0195, "args": ["host=vm06", "root=default"]}]': finished 2026-03-09T20:53:04.046 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 09 20:53:03 vm06 ceph-mon[52939]: osdmap e11: 2 total, 1 up, 2 in 2026-03-09T20:53:04.046 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 09 20:53:03 vm06 ceph-mon[52939]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 1}]: dispatch 2026-03-09T20:53:04.046 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 09 20:53:03 vm06 ceph-mon[52939]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 1}]: dispatch 2026-03-09T20:53:04.046 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 09 20:53:03 vm06 ceph-mon[52939]: from='client.14271 -' entity='client.admin' cmd=[{"prefix": "orch daemon add osd", "svc_arg": "vm08:/dev/vde", "target": ["mon-mgr", ""]}]: dispatch 2026-03-09T20:53:04.046 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 09 20:53:03 vm06 ceph-mon[52939]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' cmd=[{"prefix": "osd tree", "states": ["destroyed"], "format": "json"}]: dispatch 2026-03-09T20:53:04.046 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 09 20:53:03 vm06 ceph-mon[52939]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.bootstrap-osd"}]: dispatch 2026-03-09T20:53:04.046 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 09 20:53:03 vm06 ceph-mon[52939]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T20:53:04.046 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 09 20:53:03 vm06 ceph-mon[52939]: pgmap v29: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-09T20:53:04.707 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:53:04 vm08 ceph-mon[54611]: purged_snaps scrub starts 2026-03-09T20:53:04.707 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:53:04 vm08 ceph-mon[54611]: purged_snaps scrub ok 2026-03-09T20:53:04.707 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:53:04 vm08 ceph-mon[54611]: osd.1 [v2:192.168.123.106:6800/3453001158,v1:192.168.123.106:6801/3453001158] boot 2026-03-09T20:53:04.707 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:53:04 vm08 ceph-mon[54611]: osdmap e12: 2 total, 2 up, 2 in 2026-03-09T20:53:04.707 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:53:04 vm08 ceph-mon[54611]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 1}]: dispatch 2026-03-09T20:53:04.707 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:53:04 vm08 ceph-mon[54611]: from='client.? 192.168.123.108:0/2899344003' entity='client.bootstrap-osd' cmd=[{"prefix": "osd new", "uuid": "13684d61-15e6-4d3c-8355-a5e681450c79"}]: dispatch 2026-03-09T20:53:04.707 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:53:04 vm08 ceph-mon[54611]: from='client.? ' entity='client.bootstrap-osd' cmd=[{"prefix": "osd new", "uuid": "13684d61-15e6-4d3c-8355-a5e681450c79"}]: dispatch 2026-03-09T20:53:04.707 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:53:04 vm08 ceph-mon[54611]: from='client.? ' entity='client.bootstrap-osd' cmd='[{"prefix": "osd new", "uuid": "13684d61-15e6-4d3c-8355-a5e681450c79"}]': finished 2026-03-09T20:53:04.707 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:53:04 vm08 ceph-mon[54611]: osdmap e13: 3 total, 2 up, 3 in 2026-03-09T20:53:04.707 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:53:04 vm08 ceph-mon[54611]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 2}]: dispatch 2026-03-09T20:53:04.707 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:53:04 vm08 ceph-mon[54611]: from='client.? 192.168.123.108:0/3748053026' entity='client.bootstrap-osd' cmd=[{"prefix": "mon getmap"}]: dispatch 2026-03-09T20:53:04.923 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:53:04 vm01 ceph-mon[53402]: purged_snaps scrub starts 2026-03-09T20:53:04.924 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:53:04 vm01 ceph-mon[53402]: purged_snaps scrub ok 2026-03-09T20:53:04.924 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:53:04 vm01 ceph-mon[53402]: osd.1 [v2:192.168.123.106:6800/3453001158,v1:192.168.123.106:6801/3453001158] boot 2026-03-09T20:53:04.924 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:53:04 vm01 ceph-mon[53402]: osdmap e12: 2 total, 2 up, 2 in 2026-03-09T20:53:04.924 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:53:04 vm01 ceph-mon[53402]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 1}]: dispatch 2026-03-09T20:53:04.924 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:53:04 vm01 ceph-mon[53402]: from='client.? 192.168.123.108:0/2899344003' entity='client.bootstrap-osd' cmd=[{"prefix": "osd new", "uuid": "13684d61-15e6-4d3c-8355-a5e681450c79"}]: dispatch 2026-03-09T20:53:04.924 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:53:04 vm01 ceph-mon[53402]: from='client.? ' entity='client.bootstrap-osd' cmd=[{"prefix": "osd new", "uuid": "13684d61-15e6-4d3c-8355-a5e681450c79"}]: dispatch 2026-03-09T20:53:04.924 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:53:04 vm01 ceph-mon[53402]: from='client.? ' entity='client.bootstrap-osd' cmd='[{"prefix": "osd new", "uuid": "13684d61-15e6-4d3c-8355-a5e681450c79"}]': finished 2026-03-09T20:53:04.924 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:53:04 vm01 ceph-mon[53402]: osdmap e13: 3 total, 2 up, 3 in 2026-03-09T20:53:04.924 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:53:04 vm01 ceph-mon[53402]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 2}]: dispatch 2026-03-09T20:53:04.924 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:53:04 vm01 ceph-mon[53402]: from='client.? 192.168.123.108:0/3748053026' entity='client.bootstrap-osd' cmd=[{"prefix": "mon getmap"}]: dispatch 2026-03-09T20:53:05.046 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 09 20:53:04 vm06 ceph-mon[52939]: purged_snaps scrub starts 2026-03-09T20:53:05.046 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 09 20:53:04 vm06 ceph-mon[52939]: purged_snaps scrub ok 2026-03-09T20:53:05.046 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 09 20:53:04 vm06 ceph-mon[52939]: osd.1 [v2:192.168.123.106:6800/3453001158,v1:192.168.123.106:6801/3453001158] boot 2026-03-09T20:53:05.046 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 09 20:53:04 vm06 ceph-mon[52939]: osdmap e12: 2 total, 2 up, 2 in 2026-03-09T20:53:05.046 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 09 20:53:04 vm06 ceph-mon[52939]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 1}]: dispatch 2026-03-09T20:53:05.046 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 09 20:53:04 vm06 ceph-mon[52939]: from='client.? 192.168.123.108:0/2899344003' entity='client.bootstrap-osd' cmd=[{"prefix": "osd new", "uuid": "13684d61-15e6-4d3c-8355-a5e681450c79"}]: dispatch 2026-03-09T20:53:05.046 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 09 20:53:04 vm06 ceph-mon[52939]: from='client.? ' entity='client.bootstrap-osd' cmd=[{"prefix": "osd new", "uuid": "13684d61-15e6-4d3c-8355-a5e681450c79"}]: dispatch 2026-03-09T20:53:05.046 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 09 20:53:04 vm06 ceph-mon[52939]: from='client.? ' entity='client.bootstrap-osd' cmd='[{"prefix": "osd new", "uuid": "13684d61-15e6-4d3c-8355-a5e681450c79"}]': finished 2026-03-09T20:53:05.046 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 09 20:53:04 vm06 ceph-mon[52939]: osdmap e13: 3 total, 2 up, 3 in 2026-03-09T20:53:05.046 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 09 20:53:04 vm06 ceph-mon[52939]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 2}]: dispatch 2026-03-09T20:53:05.046 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 09 20:53:04 vm06 ceph-mon[52939]: from='client.? 192.168.123.108:0/3748053026' entity='client.bootstrap-osd' cmd=[{"prefix": "mon getmap"}]: dispatch 2026-03-09T20:53:06.296 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 09 20:53:05 vm06 ceph-mon[52939]: osdmap e14: 3 total, 2 up, 3 in 2026-03-09T20:53:06.296 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 09 20:53:05 vm06 ceph-mon[52939]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 2}]: dispatch 2026-03-09T20:53:06.296 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 09 20:53:05 vm06 ceph-mon[52939]: pgmap v33: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-09T20:53:06.424 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:53:05 vm01 ceph-mon[53402]: osdmap e14: 3 total, 2 up, 3 in 2026-03-09T20:53:06.424 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:53:05 vm01 ceph-mon[53402]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 2}]: dispatch 2026-03-09T20:53:06.424 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:53:05 vm01 ceph-mon[53402]: pgmap v33: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-09T20:53:06.457 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:53:05 vm08 ceph-mon[54611]: osdmap e14: 3 total, 2 up, 3 in 2026-03-09T20:53:06.457 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:53:05 vm08 ceph-mon[54611]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 2}]: dispatch 2026-03-09T20:53:06.457 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:53:05 vm08 ceph-mon[54611]: pgmap v33: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-09T20:53:07.678 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:53:07 vm08 ceph-mon[54611]: pgmap v34: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-09T20:53:07.796 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 09 20:53:07 vm06 ceph-mon[52939]: pgmap v34: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-09T20:53:07.924 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:53:07 vm01 ceph-mon[53402]: pgmap v34: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-09T20:53:08.522 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:53:08 vm08 ceph-mon[54611]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "osd.2"}]: dispatch 2026-03-09T20:53:08.522 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:53:08 vm08 ceph-mon[54611]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T20:53:08.522 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:53:08 vm08 ceph-mon[54611]: Deploying daemon osd.2 on vm08 2026-03-09T20:53:08.796 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 09 20:53:08 vm06 ceph-mon[52939]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "osd.2"}]: dispatch 2026-03-09T20:53:08.796 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 09 20:53:08 vm06 ceph-mon[52939]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T20:53:08.796 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 09 20:53:08 vm06 ceph-mon[52939]: Deploying daemon osd.2 on vm08 2026-03-09T20:53:08.924 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:53:08 vm01 ceph-mon[53402]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "osd.2"}]: dispatch 2026-03-09T20:53:08.924 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:53:08 vm01 ceph-mon[53402]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T20:53:08.924 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:53:08 vm01 ceph-mon[53402]: Deploying daemon osd.2 on vm08 2026-03-09T20:53:09.708 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:53:09 vm08 ceph-mon[54611]: pgmap v35: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-09T20:53:09.795 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 09 20:53:09 vm06 ceph-mon[52939]: pgmap v35: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-09T20:53:09.923 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:53:09 vm01 ceph-mon[53402]: pgmap v35: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-09T20:53:10.914 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:53:10 vm08 ceph-mon[54611]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T20:53:10.914 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:53:10 vm08 ceph-mon[54611]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' 2026-03-09T20:53:10.914 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:53:10 vm08 ceph-mon[54611]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' 2026-03-09T20:53:11.045 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 09 20:53:10 vm06 ceph-mon[52939]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T20:53:11.046 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 09 20:53:10 vm06 ceph-mon[52939]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' 2026-03-09T20:53:11.046 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 09 20:53:10 vm06 ceph-mon[52939]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' 2026-03-09T20:53:11.173 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:53:10 vm01 ceph-mon[53402]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T20:53:11.174 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:53:10 vm01 ceph-mon[53402]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' 2026-03-09T20:53:11.174 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:53:10 vm01 ceph-mon[53402]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' 2026-03-09T20:53:12.014 INFO:teuthology.orchestra.run.vm08.stdout:Created osd(s) 2 on host 'vm08' 2026-03-09T20:53:12.660 DEBUG:teuthology.orchestra.run.vm08:osd.2> sudo journalctl -f -n 0 -u ceph-9cb345a8-1bf9-11f1-a195-0375563c5891@osd.2.service 2026-03-09T20:53:12.701 INFO:tasks.cephadm:Waiting for 3 OSDs to come up... 2026-03-09T20:53:12.701 DEBUG:teuthology.orchestra.run.vm01:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 9cb345a8-1bf9-11f1-a195-0375563c5891 -- ceph osd stat -f json 2026-03-09T20:53:12.881 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:53:12 vm08 ceph-mon[54611]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' 2026-03-09T20:53:12.881 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:53:12 vm08 ceph-mon[54611]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' 2026-03-09T20:53:12.881 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:53:12 vm08 ceph-mon[54611]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T20:53:12.881 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:53:12 vm08 ceph-mon[54611]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T20:53:12.881 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:53:12 vm08 ceph-mon[54611]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' 2026-03-09T20:53:12.881 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:53:12 vm08 ceph-mon[54611]: pgmap v36: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-09T20:53:12.881 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:53:12 vm08 ceph-mon[54611]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T20:53:12.881 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:53:12 vm08 ceph-mon[54611]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' 2026-03-09T20:53:12.881 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:53:12 vm08 ceph-mon[54611]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' 2026-03-09T20:53:12.899 INFO:teuthology.orchestra.run.vm01.stderr:Inferring config /var/lib/ceph/9cb345a8-1bf9-11f1-a195-0375563c5891/mon.a/config 2026-03-09T20:53:12.923 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:53:12 vm01 ceph-mon[53402]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' 2026-03-09T20:53:12.923 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:53:12 vm01 ceph-mon[53402]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' 2026-03-09T20:53:12.923 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:53:12 vm01 ceph-mon[53402]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T20:53:12.923 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:53:12 vm01 ceph-mon[53402]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T20:53:12.923 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:53:12 vm01 ceph-mon[53402]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' 2026-03-09T20:53:12.923 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:53:12 vm01 ceph-mon[53402]: pgmap v36: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-09T20:53:12.924 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:53:12 vm01 ceph-mon[53402]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T20:53:12.924 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:53:12 vm01 ceph-mon[53402]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' 2026-03-09T20:53:12.924 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:53:12 vm01 ceph-mon[53402]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' 2026-03-09T20:53:13.046 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 09 20:53:12 vm06 ceph-mon[52939]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' 2026-03-09T20:53:13.046 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 09 20:53:12 vm06 ceph-mon[52939]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' 2026-03-09T20:53:13.046 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 09 20:53:12 vm06 ceph-mon[52939]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T20:53:13.046 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 09 20:53:12 vm06 ceph-mon[52939]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T20:53:13.046 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 09 20:53:12 vm06 ceph-mon[52939]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' 2026-03-09T20:53:13.046 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 09 20:53:12 vm06 ceph-mon[52939]: pgmap v36: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-09T20:53:13.046 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 09 20:53:12 vm06 ceph-mon[52939]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T20:53:13.046 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 09 20:53:12 vm06 ceph-mon[52939]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' 2026-03-09T20:53:13.046 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 09 20:53:12 vm06 ceph-mon[52939]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' 2026-03-09T20:53:13.244 INFO:teuthology.orchestra.run.vm01.stdout: 2026-03-09T20:53:13.438 INFO:teuthology.orchestra.run.vm01.stdout:{"epoch":14,"num_osds":3,"num_up_osds":2,"osd_up_since":1773089583,"num_in_osds":3,"osd_in_since":1773089583,"num_remapped_pgs":0} 2026-03-09T20:53:14.046 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 09 20:53:13 vm06 ceph-mon[52939]: from='osd.2 [v2:192.168.123.108:6800/4083304280,v1:192.168.123.108:6801/4083304280]' entity='osd.2' cmd=[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["2"]}]: dispatch 2026-03-09T20:53:14.046 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 09 20:53:13 vm06 ceph-mon[52939]: from='osd.2 ' entity='osd.2' cmd=[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["2"]}]: dispatch 2026-03-09T20:53:14.046 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 09 20:53:13 vm06 ceph-mon[52939]: from='client.? 192.168.123.101:0/2444259423' entity='client.admin' cmd=[{"prefix": "osd stat", "format": "json"}]: dispatch 2026-03-09T20:53:14.046 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 09 20:53:13 vm06 ceph-mon[52939]: pgmap v37: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-09T20:53:14.173 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:53:13 vm01 ceph-mon[53402]: from='osd.2 [v2:192.168.123.108:6800/4083304280,v1:192.168.123.108:6801/4083304280]' entity='osd.2' cmd=[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["2"]}]: dispatch 2026-03-09T20:53:14.173 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:53:13 vm01 ceph-mon[53402]: from='osd.2 ' entity='osd.2' cmd=[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["2"]}]: dispatch 2026-03-09T20:53:14.173 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:53:13 vm01 ceph-mon[53402]: from='client.? 192.168.123.101:0/2444259423' entity='client.admin' cmd=[{"prefix": "osd stat", "format": "json"}]: dispatch 2026-03-09T20:53:14.174 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:53:13 vm01 ceph-mon[53402]: pgmap v37: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-09T20:53:14.207 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:53:13 vm08 ceph-mon[54611]: from='osd.2 [v2:192.168.123.108:6800/4083304280,v1:192.168.123.108:6801/4083304280]' entity='osd.2' cmd=[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["2"]}]: dispatch 2026-03-09T20:53:14.207 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:53:13 vm08 ceph-mon[54611]: from='osd.2 ' entity='osd.2' cmd=[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["2"]}]: dispatch 2026-03-09T20:53:14.207 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:53:13 vm08 ceph-mon[54611]: from='client.? 192.168.123.101:0/2444259423' entity='client.admin' cmd=[{"prefix": "osd stat", "format": "json"}]: dispatch 2026-03-09T20:53:14.207 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:53:13 vm08 ceph-mon[54611]: pgmap v37: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-09T20:53:14.439 DEBUG:teuthology.orchestra.run.vm01:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 9cb345a8-1bf9-11f1-a195-0375563c5891 -- ceph osd stat -f json 2026-03-09T20:53:14.623 INFO:teuthology.orchestra.run.vm01.stderr:Inferring config /var/lib/ceph/9cb345a8-1bf9-11f1-a195-0375563c5891/mon.a/config 2026-03-09T20:53:14.888 INFO:teuthology.orchestra.run.vm01.stdout: 2026-03-09T20:53:15.038 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:53:14 vm01 ceph-mon[53402]: from='osd.2 ' entity='osd.2' cmd='[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["2"]}]': finished 2026-03-09T20:53:15.038 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:53:14 vm01 ceph-mon[53402]: from='osd.2 [v2:192.168.123.108:6800/4083304280,v1:192.168.123.108:6801/4083304280]' entity='osd.2' cmd=[{"prefix": "osd crush create-or-move", "id": 2, "weight":0.0195, "args": ["host=vm08", "root=default"]}]: dispatch 2026-03-09T20:53:15.038 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:53:14 vm01 ceph-mon[53402]: osdmap e15: 3 total, 2 up, 3 in 2026-03-09T20:53:15.038 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:53:14 vm01 ceph-mon[53402]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 2}]: dispatch 2026-03-09T20:53:15.038 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:53:14 vm01 ceph-mon[53402]: from='osd.2 ' entity='osd.2' cmd=[{"prefix": "osd crush create-or-move", "id": 2, "weight":0.0195, "args": ["host=vm08", "root=default"]}]: dispatch 2026-03-09T20:53:15.038 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:53:14 vm01 ceph-mon[53402]: Detected new or changed devices on vm08 2026-03-09T20:53:15.038 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:53:14 vm01 ceph-mon[53402]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' 2026-03-09T20:53:15.038 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:53:14 vm01 ceph-mon[53402]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' 2026-03-09T20:53:15.038 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:53:14 vm01 ceph-mon[53402]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' cmd=[{"prefix": "config rm", "who": "osd.2", "name": "osd_memory_target"}]: dispatch 2026-03-09T20:53:15.038 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:53:14 vm01 ceph-mon[53402]: Adjusting osd_memory_target on vm08 to 4353M 2026-03-09T20:53:15.038 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:53:14 vm01 ceph-mon[53402]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' 2026-03-09T20:53:15.038 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:53:14 vm01 ceph-mon[53402]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T20:53:15.038 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:53:14 vm01 ceph-mon[53402]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T20:53:15.038 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:53:14 vm01 ceph-mon[53402]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' 2026-03-09T20:53:15.046 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 09 20:53:14 vm06 ceph-mon[52939]: from='osd.2 ' entity='osd.2' cmd='[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["2"]}]': finished 2026-03-09T20:53:15.046 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 09 20:53:14 vm06 ceph-mon[52939]: from='osd.2 [v2:192.168.123.108:6800/4083304280,v1:192.168.123.108:6801/4083304280]' entity='osd.2' cmd=[{"prefix": "osd crush create-or-move", "id": 2, "weight":0.0195, "args": ["host=vm08", "root=default"]}]: dispatch 2026-03-09T20:53:15.046 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 09 20:53:14 vm06 ceph-mon[52939]: osdmap e15: 3 total, 2 up, 3 in 2026-03-09T20:53:15.046 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 09 20:53:14 vm06 ceph-mon[52939]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 2}]: dispatch 2026-03-09T20:53:15.046 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 09 20:53:14 vm06 ceph-mon[52939]: from='osd.2 ' entity='osd.2' cmd=[{"prefix": "osd crush create-or-move", "id": 2, "weight":0.0195, "args": ["host=vm08", "root=default"]}]: dispatch 2026-03-09T20:53:15.046 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 09 20:53:14 vm06 ceph-mon[52939]: Detected new or changed devices on vm08 2026-03-09T20:53:15.046 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 09 20:53:14 vm06 ceph-mon[52939]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' 2026-03-09T20:53:15.046 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 09 20:53:14 vm06 ceph-mon[52939]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' 2026-03-09T20:53:15.046 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 09 20:53:14 vm06 ceph-mon[52939]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' cmd=[{"prefix": "config rm", "who": "osd.2", "name": "osd_memory_target"}]: dispatch 2026-03-09T20:53:15.046 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 09 20:53:14 vm06 ceph-mon[52939]: Adjusting osd_memory_target on vm08 to 4353M 2026-03-09T20:53:15.046 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 09 20:53:14 vm06 ceph-mon[52939]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' 2026-03-09T20:53:15.046 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 09 20:53:14 vm06 ceph-mon[52939]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T20:53:15.046 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 09 20:53:14 vm06 ceph-mon[52939]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T20:53:15.046 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 09 20:53:14 vm06 ceph-mon[52939]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' 2026-03-09T20:53:15.068 INFO:teuthology.orchestra.run.vm01.stdout:{"epoch":16,"num_osds":3,"num_up_osds":2,"osd_up_since":1773089583,"num_in_osds":3,"osd_in_since":1773089583,"num_remapped_pgs":0} 2026-03-09T20:53:15.207 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:53:14 vm08 ceph-mon[54611]: from='osd.2 ' entity='osd.2' cmd='[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["2"]}]': finished 2026-03-09T20:53:15.207 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:53:14 vm08 ceph-mon[54611]: from='osd.2 [v2:192.168.123.108:6800/4083304280,v1:192.168.123.108:6801/4083304280]' entity='osd.2' cmd=[{"prefix": "osd crush create-or-move", "id": 2, "weight":0.0195, "args": ["host=vm08", "root=default"]}]: dispatch 2026-03-09T20:53:15.207 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:53:14 vm08 ceph-mon[54611]: osdmap e15: 3 total, 2 up, 3 in 2026-03-09T20:53:15.207 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:53:14 vm08 ceph-mon[54611]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 2}]: dispatch 2026-03-09T20:53:15.207 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:53:14 vm08 ceph-mon[54611]: from='osd.2 ' entity='osd.2' cmd=[{"prefix": "osd crush create-or-move", "id": 2, "weight":0.0195, "args": ["host=vm08", "root=default"]}]: dispatch 2026-03-09T20:53:15.207 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:53:14 vm08 ceph-mon[54611]: Detected new or changed devices on vm08 2026-03-09T20:53:15.207 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:53:14 vm08 ceph-mon[54611]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' 2026-03-09T20:53:15.207 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:53:14 vm08 ceph-mon[54611]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' 2026-03-09T20:53:15.207 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:53:14 vm08 ceph-mon[54611]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' cmd=[{"prefix": "config rm", "who": "osd.2", "name": "osd_memory_target"}]: dispatch 2026-03-09T20:53:15.207 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:53:14 vm08 ceph-mon[54611]: Adjusting osd_memory_target on vm08 to 4353M 2026-03-09T20:53:15.207 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:53:14 vm08 ceph-mon[54611]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' 2026-03-09T20:53:15.207 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:53:14 vm08 ceph-mon[54611]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T20:53:15.207 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:53:14 vm08 ceph-mon[54611]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T20:53:15.207 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:53:14 vm08 ceph-mon[54611]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' 2026-03-09T20:53:15.208 INFO:journalctl@ceph.osd.2.vm08.stdout:Mar 09 20:53:14 vm08 ceph-9cb345a8-1bf9-11f1-a195-0375563c5891-osd-2[58655]: 2026-03-09T20:53:14.773+0000 7fc58328c640 -1 osd.2 0 waiting for initial osdmap 2026-03-09T20:53:15.208 INFO:journalctl@ceph.osd.2.vm08.stdout:Mar 09 20:53:14 vm08 ceph-9cb345a8-1bf9-11f1-a195-0375563c5891-osd-2[58655]: 2026-03-09T20:53:14.780+0000 7fc57e8a3640 -1 osd.2 16 set_numa_affinity unable to identify public interface '' numa node: (2) No such file or directory 2026-03-09T20:53:16.046 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 09 20:53:15 vm06 ceph-mon[52939]: purged_snaps scrub starts 2026-03-09T20:53:16.046 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 09 20:53:15 vm06 ceph-mon[52939]: purged_snaps scrub ok 2026-03-09T20:53:16.046 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 09 20:53:15 vm06 ceph-mon[52939]: from='osd.2 ' entity='osd.2' cmd='[{"prefix": "osd crush create-or-move", "id": 2, "weight":0.0195, "args": ["host=vm08", "root=default"]}]': finished 2026-03-09T20:53:16.046 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 09 20:53:15 vm06 ceph-mon[52939]: osdmap e16: 3 total, 2 up, 3 in 2026-03-09T20:53:16.046 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 09 20:53:15 vm06 ceph-mon[52939]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 2}]: dispatch 2026-03-09T20:53:16.046 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 09 20:53:15 vm06 ceph-mon[52939]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 2}]: dispatch 2026-03-09T20:53:16.046 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 09 20:53:15 vm06 ceph-mon[52939]: from='client.? 192.168.123.101:0/525415226' entity='client.admin' cmd=[{"prefix": "osd stat", "format": "json"}]: dispatch 2026-03-09T20:53:16.046 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 09 20:53:15 vm06 ceph-mon[52939]: pgmap v40: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-09T20:53:16.069 DEBUG:teuthology.orchestra.run.vm01:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 9cb345a8-1bf9-11f1-a195-0375563c5891 -- ceph osd stat -f json 2026-03-09T20:53:16.092 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:53:15 vm01 ceph-mon[53402]: purged_snaps scrub starts 2026-03-09T20:53:16.093 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:53:15 vm01 ceph-mon[53402]: purged_snaps scrub ok 2026-03-09T20:53:16.093 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:53:15 vm01 ceph-mon[53402]: from='osd.2 ' entity='osd.2' cmd='[{"prefix": "osd crush create-or-move", "id": 2, "weight":0.0195, "args": ["host=vm08", "root=default"]}]': finished 2026-03-09T20:53:16.093 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:53:15 vm01 ceph-mon[53402]: osdmap e16: 3 total, 2 up, 3 in 2026-03-09T20:53:16.093 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:53:15 vm01 ceph-mon[53402]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 2}]: dispatch 2026-03-09T20:53:16.093 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:53:15 vm01 ceph-mon[53402]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 2}]: dispatch 2026-03-09T20:53:16.093 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:53:15 vm01 ceph-mon[53402]: from='client.? 192.168.123.101:0/525415226' entity='client.admin' cmd=[{"prefix": "osd stat", "format": "json"}]: dispatch 2026-03-09T20:53:16.093 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:53:15 vm01 ceph-mon[53402]: pgmap v40: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-09T20:53:16.207 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:53:15 vm08 ceph-mon[54611]: purged_snaps scrub starts 2026-03-09T20:53:16.207 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:53:15 vm08 ceph-mon[54611]: purged_snaps scrub ok 2026-03-09T20:53:16.207 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:53:15 vm08 ceph-mon[54611]: from='osd.2 ' entity='osd.2' cmd='[{"prefix": "osd crush create-or-move", "id": 2, "weight":0.0195, "args": ["host=vm08", "root=default"]}]': finished 2026-03-09T20:53:16.207 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:53:15 vm08 ceph-mon[54611]: osdmap e16: 3 total, 2 up, 3 in 2026-03-09T20:53:16.207 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:53:15 vm08 ceph-mon[54611]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 2}]: dispatch 2026-03-09T20:53:16.207 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:53:15 vm08 ceph-mon[54611]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 2}]: dispatch 2026-03-09T20:53:16.207 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:53:15 vm08 ceph-mon[54611]: from='client.? 192.168.123.101:0/525415226' entity='client.admin' cmd=[{"prefix": "osd stat", "format": "json"}]: dispatch 2026-03-09T20:53:16.207 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:53:15 vm08 ceph-mon[54611]: pgmap v40: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-09T20:53:16.233 INFO:teuthology.orchestra.run.vm01.stderr:Inferring config /var/lib/ceph/9cb345a8-1bf9-11f1-a195-0375563c5891/mon.a/config 2026-03-09T20:53:16.562 INFO:teuthology.orchestra.run.vm01.stdout: 2026-03-09T20:53:16.754 INFO:teuthology.orchestra.run.vm01.stdout:{"epoch":17,"num_osds":3,"num_up_osds":3,"osd_up_since":1773089595,"num_in_osds":3,"osd_in_since":1773089583,"num_remapped_pgs":0} 2026-03-09T20:53:16.754 DEBUG:teuthology.orchestra.run.vm01:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell --fsid 9cb345a8-1bf9-11f1-a195-0375563c5891 -- ceph osd dump --format=json 2026-03-09T20:53:16.924 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:53:16 vm01 ceph-mon[53402]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 2}]: dispatch 2026-03-09T20:53:16.924 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:53:16 vm01 ceph-mon[53402]: osd.2 [v2:192.168.123.108:6800/4083304280,v1:192.168.123.108:6801/4083304280] boot 2026-03-09T20:53:16.924 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:53:16 vm01 ceph-mon[53402]: osdmap e17: 3 total, 3 up, 3 in 2026-03-09T20:53:16.924 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:53:16 vm01 ceph-mon[53402]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 2}]: dispatch 2026-03-09T20:53:16.924 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:53:16 vm01 ceph-mon[53402]: from='client.? 192.168.123.101:0/2333947852' entity='client.admin' cmd=[{"prefix": "osd stat", "format": "json"}]: dispatch 2026-03-09T20:53:16.979 INFO:teuthology.orchestra.run.vm01.stderr:Inferring config /var/lib/ceph/9cb345a8-1bf9-11f1-a195-0375563c5891/mon.a/config 2026-03-09T20:53:17.045 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 09 20:53:16 vm06 ceph-mon[52939]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 2}]: dispatch 2026-03-09T20:53:17.046 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 09 20:53:16 vm06 ceph-mon[52939]: osd.2 [v2:192.168.123.108:6800/4083304280,v1:192.168.123.108:6801/4083304280] boot 2026-03-09T20:53:17.046 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 09 20:53:16 vm06 ceph-mon[52939]: osdmap e17: 3 total, 3 up, 3 in 2026-03-09T20:53:17.046 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 09 20:53:16 vm06 ceph-mon[52939]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 2}]: dispatch 2026-03-09T20:53:17.046 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 09 20:53:16 vm06 ceph-mon[52939]: from='client.? 192.168.123.101:0/2333947852' entity='client.admin' cmd=[{"prefix": "osd stat", "format": "json"}]: dispatch 2026-03-09T20:53:17.207 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:53:16 vm08 ceph-mon[54611]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 2}]: dispatch 2026-03-09T20:53:17.207 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:53:16 vm08 ceph-mon[54611]: osd.2 [v2:192.168.123.108:6800/4083304280,v1:192.168.123.108:6801/4083304280] boot 2026-03-09T20:53:17.207 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:53:16 vm08 ceph-mon[54611]: osdmap e17: 3 total, 3 up, 3 in 2026-03-09T20:53:17.207 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:53:16 vm08 ceph-mon[54611]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 2}]: dispatch 2026-03-09T20:53:17.207 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:53:16 vm08 ceph-mon[54611]: from='client.? 192.168.123.101:0/2333947852' entity='client.admin' cmd=[{"prefix": "osd stat", "format": "json"}]: dispatch 2026-03-09T20:53:17.228 INFO:teuthology.orchestra.run.vm01.stdout: 2026-03-09T20:53:17.228 INFO:teuthology.orchestra.run.vm01.stdout:{"epoch":18,"fsid":"9cb345a8-1bf9-11f1-a195-0375563c5891","created":"2026-03-09T20:51:43.189739+0000","modified":"2026-03-09T20:53:16.775040+0000","last_up_change":"2026-03-09T20:53:15.768033+0000","last_in_change":"2026-03-09T20:53:03.945618+0000","flags":"sortbitwise,recovery_deletes,purged_snapdirs,pglog_hardlimit","flags_num":5799936,"flags_set":["pglog_hardlimit","purged_snapdirs","recovery_deletes","sortbitwise"],"crush_version":8,"full_ratio":0.94999998807907104,"backfillfull_ratio":0.89999997615814209,"nearfull_ratio":0.85000002384185791,"cluster_snapshot":"","pool_max":0,"max_osd":3,"require_min_compat_client":"luminous","min_compat_client":"jewel","require_osd_release":"squid","allow_crimson":false,"pools":[],"osds":[{"osd":0,"uuid":"58efc62d-2ef3-4005-88e4-4cf11f5576fe","up":1,"in":1,"weight":1,"primary_affinity":1,"last_clean_begin":0,"last_clean_end":0,"up_from":8,"up_thru":0,"down_at":0,"lost_at":0,"public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.101:6802","nonce":3447435909},{"type":"v1","addr":"192.168.123.101:6803","nonce":3447435909}]},"cluster_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.101:6804","nonce":3447435909},{"type":"v1","addr":"192.168.123.101:6805","nonce":3447435909}]},"heartbeat_back_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.101:6808","nonce":3447435909},{"type":"v1","addr":"192.168.123.101:6809","nonce":3447435909}]},"heartbeat_front_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.101:6806","nonce":3447435909},{"type":"v1","addr":"192.168.123.101:6807","nonce":3447435909}]},"public_addr":"192.168.123.101:6803/3447435909","cluster_addr":"192.168.123.101:6805/3447435909","heartbeat_back_addr":"192.168.123.101:6809/3447435909","heartbeat_front_addr":"192.168.123.101:6807/3447435909","state":["exists","up"]},{"osd":1,"uuid":"38f01073-940b-498e-91d3-7a23e98bcb53","up":1,"in":1,"weight":1,"primary_affinity":1,"last_clean_begin":0,"last_clean_end":0,"up_from":12,"up_thru":0,"down_at":0,"lost_at":0,"public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.106:6800","nonce":3453001158},{"type":"v1","addr":"192.168.123.106:6801","nonce":3453001158}]},"cluster_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.106:6802","nonce":3453001158},{"type":"v1","addr":"192.168.123.106:6803","nonce":3453001158}]},"heartbeat_back_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.106:6806","nonce":3453001158},{"type":"v1","addr":"192.168.123.106:6807","nonce":3453001158}]},"heartbeat_front_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.106:6804","nonce":3453001158},{"type":"v1","addr":"192.168.123.106:6805","nonce":3453001158}]},"public_addr":"192.168.123.106:6801/3453001158","cluster_addr":"192.168.123.106:6803/3453001158","heartbeat_back_addr":"192.168.123.106:6807/3453001158","heartbeat_front_addr":"192.168.123.106:6805/3453001158","state":["exists","up"]},{"osd":2,"uuid":"13684d61-15e6-4d3c-8355-a5e681450c79","up":1,"in":1,"weight":1,"primary_affinity":1,"last_clean_begin":0,"last_clean_end":0,"up_from":17,"up_thru":0,"down_at":0,"lost_at":0,"public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.108:6800","nonce":4083304280},{"type":"v1","addr":"192.168.123.108:6801","nonce":4083304280}]},"cluster_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.108:6802","nonce":4083304280},{"type":"v1","addr":"192.168.123.108:6803","nonce":4083304280}]},"heartbeat_back_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.108:6806","nonce":4083304280},{"type":"v1","addr":"192.168.123.108:6807","nonce":4083304280}]},"heartbeat_front_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.108:6804","nonce":4083304280},{"type":"v1","addr":"192.168.123.108:6805","nonce":4083304280}]},"public_addr":"192.168.123.108:6801/4083304280","cluster_addr":"192.168.123.108:6803/4083304280","heartbeat_back_addr":"192.168.123.108:6807/4083304280","heartbeat_front_addr":"192.168.123.108:6805/4083304280","state":["exists","up"]}],"osd_xinfo":[{"osd":0,"down_stamp":"0.000000","laggy_probability":0,"laggy_interval":0,"features":4540701547738038271,"old_weight":0,"last_purged_snaps_scrub":"2026-03-09T20:52:50.809851+0000","dead_epoch":0},{"osd":1,"down_stamp":"0.000000","laggy_probability":0,"laggy_interval":0,"features":4540701547738038271,"old_weight":0,"last_purged_snaps_scrub":"2026-03-09T20:53:02.121506+0000","dead_epoch":0},{"osd":2,"down_stamp":"0.000000","laggy_probability":0,"laggy_interval":0,"features":4540701547738038271,"old_weight":0,"last_purged_snaps_scrub":"2026-03-09T20:53:13.662870+0000","dead_epoch":0}],"pg_upmap":[],"pg_upmap_items":[],"pg_upmap_primaries":[],"pg_temp":[],"primary_temp":[],"blocklist":{"192.168.123.101:0/1300460699":"2026-03-10T20:52:07.386219+0000","192.168.123.101:6801/119456608":"2026-03-10T20:52:07.386219+0000","192.168.123.101:0/2446502297":"2026-03-10T20:51:56.804238+0000","192.168.123.101:0/1626717301":"2026-03-10T20:52:07.386219+0000","192.168.123.101:0/4056848261":"2026-03-10T20:51:56.804238+0000","192.168.123.101:6800/119456608":"2026-03-10T20:52:07.386219+0000","192.168.123.101:6801/668143714":"2026-03-10T20:51:56.804238+0000","192.168.123.101:6800/668143714":"2026-03-10T20:51:56.804238+0000","192.168.123.101:0/1165500474":"2026-03-10T20:52:07.386219+0000","192.168.123.101:0/4063840047":"2026-03-10T20:51:56.804238+0000"},"range_blocklist":{},"erasure_code_profiles":{"default":{"crush-failure-domain":"osd","k":"2","m":"1","plugin":"jerasure","technique":"reed_sol_van"}},"removed_snaps_queue":[],"new_removed_snaps":[],"new_purged_snaps":[],"crush_node_flags":{},"device_class_flags":{},"stretch_mode":{"stretch_mode_enabled":false,"stretch_bucket_count":0,"degraded_stretch_mode":0,"recovering_stretch_mode":0,"stretch_mode_bucket":0}} 2026-03-09T20:53:17.405 INFO:tasks.cephadm.ceph_manager.ceph:[] 2026-03-09T20:53:17.405 INFO:tasks.cephadm:Setting up client nodes... 2026-03-09T20:53:17.405 INFO:tasks.ceph:Waiting until ceph daemons up and pgs clean... 2026-03-09T20:53:17.405 INFO:tasks.cephadm.ceph_manager.ceph:waiting for mgr available 2026-03-09T20:53:17.406 DEBUG:teuthology.orchestra.run.vm01:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell --fsid 9cb345a8-1bf9-11f1-a195-0375563c5891 -- ceph mgr dump --format=json 2026-03-09T20:53:17.584 INFO:teuthology.orchestra.run.vm01.stderr:Inferring config /var/lib/ceph/9cb345a8-1bf9-11f1-a195-0375563c5891/mon.a/config 2026-03-09T20:53:17.854 INFO:teuthology.orchestra.run.vm01.stdout: 2026-03-09T20:53:17.858 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:53:17 vm01 ceph-mon[53402]: osdmap e18: 3 total, 3 up, 3 in 2026-03-09T20:53:17.858 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:53:17 vm01 ceph-mon[53402]: from='client.? 192.168.123.101:0/3140589146' entity='client.admin' cmd=[{"prefix": "osd dump", "format": "json"}]: dispatch 2026-03-09T20:53:17.858 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:53:17 vm01 ceph-mon[53402]: pgmap v43: 0 pgs: ; 0 B data, 79 MiB used, 60 GiB / 60 GiB avail 2026-03-09T20:53:17.858 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:53:17 vm01 ceph-mon[53402]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' cmd=[{"prefix": "osd pool create", "format": "json", "pool": ".mgr", "pg_num": 1, "pg_num_min": 1, "pg_num_max": 32, "yes_i_really_mean_it": true}]: dispatch 2026-03-09T20:53:18.012 INFO:teuthology.orchestra.run.vm01.stdout:{"epoch":14,"flags":0,"active_gid":14150,"active_name":"a","active_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.101:6800","nonce":2709248488},{"type":"v1","addr":"192.168.123.101:6801","nonce":2709248488}]},"active_addr":"192.168.123.101:6801/2709248488","active_change":"2026-03-09T20:52:07.386327+0000","active_mgr_features":4540701547738038271,"available":true,"standbys":[{"gid":14211,"name":"b","mgr_features":4540701547738038271,"available_modules":[{"name":"alerts","can_run":true,"error_string":"","module_options":{"interval":{"name":"interval","type":"secs","level":"advanced","flags":1,"default_value":"60","min":"","max":"","enum_allowed":[],"desc":"How frequently to reexamine health status","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"smtp_destination":{"name":"smtp_destination","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"Email address to send alerts to","long_desc":"","tags":[],"see_also":[]},"smtp_from_name":{"name":"smtp_from_name","type":"str","level":"advanced","flags":1,"default_value":"Ceph","min":"","max":"","enum_allowed":[],"desc":"Email From: name","long_desc":"","tags":[],"see_also":[]},"smtp_host":{"name":"smtp_host","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"SMTP server","long_desc":"","tags":[],"see_also":[]},"smtp_password":{"name":"smtp_password","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"Password to authenticate with","long_desc":"","tags":[],"see_also":[]},"smtp_port":{"name":"smtp_port","type":"int","level":"advanced","flags":1,"default_value":"465","min":"","max":"","enum_allowed":[],"desc":"SMTP port","long_desc":"","tags":[],"see_also":[]},"smtp_sender":{"name":"smtp_sender","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"SMTP envelope sender","long_desc":"","tags":[],"see_also":[]},"smtp_ssl":{"name":"smtp_ssl","type":"bool","level":"advanced","flags":1,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"Use SSL to connect to SMTP server","long_desc":"","tags":[],"see_also":[]},"smtp_user":{"name":"smtp_user","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"User to authenticate as","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"balancer","can_run":true,"error_string":"","module_options":{"active":{"name":"active","type":"bool","level":"advanced","flags":1,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"automatically balance PGs across cluster","long_desc":"","tags":[],"see_also":[]},"begin_time":{"name":"begin_time","type":"str","level":"advanced","flags":1,"default_value":"0000","min":"","max":"","enum_allowed":[],"desc":"beginning time of day to automatically balance","long_desc":"This is a time of day in the format HHMM.","tags":[],"see_also":[]},"begin_weekday":{"name":"begin_weekday","type":"uint","level":"advanced","flags":1,"default_value":"0","min":"0","max":"6","enum_allowed":[],"desc":"Restrict automatic balancing to this day of the week or later","long_desc":"0 = Sunday, 1 = Monday, etc.","tags":[],"see_also":[]},"crush_compat_max_iterations":{"name":"crush_compat_max_iterations","type":"uint","level":"advanced","flags":1,"default_value":"25","min":"1","max":"250","enum_allowed":[],"desc":"maximum number of iterations to attempt optimization","long_desc":"","tags":[],"see_also":[]},"crush_compat_metrics":{"name":"crush_compat_metrics","type":"str","level":"advanced","flags":1,"default_value":"pgs,objects,bytes","min":"","max":"","enum_allowed":[],"desc":"metrics with which to calculate OSD utilization","long_desc":"Value is a list of one or more of \"pgs\", \"objects\", or \"bytes\", and indicates which metrics to use to balance utilization.","tags":[],"see_also":[]},"crush_compat_step":{"name":"crush_compat_step","type":"float","level":"advanced","flags":1,"default_value":"0.5","min":"0.001","max":"0.999","enum_allowed":[],"desc":"aggressiveness of optimization","long_desc":".99 is very aggressive, .01 is less aggressive","tags":[],"see_also":[]},"end_time":{"name":"end_time","type":"str","level":"advanced","flags":1,"default_value":"2359","min":"","max":"","enum_allowed":[],"desc":"ending time of day to automatically balance","long_desc":"This is a time of day in the format HHMM.","tags":[],"see_also":[]},"end_weekday":{"name":"end_weekday","type":"uint","level":"advanced","flags":1,"default_value":"0","min":"0","max":"6","enum_allowed":[],"desc":"Restrict automatic balancing to days of the week earlier than this","long_desc":"0 = Sunday, 1 = Monday, etc.","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"min_score":{"name":"min_score","type":"float","level":"advanced","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"minimum score, below which no optimization is attempted","long_desc":"","tags":[],"see_also":[]},"mode":{"name":"mode","type":"str","level":"advanced","flags":1,"default_value":"upmap","min":"","max":"","enum_allowed":["crush-compat","none","read","upmap","upmap-read"],"desc":"Balancer mode","long_desc":"","tags":[],"see_also":[]},"pool_ids":{"name":"pool_ids","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"pools which the automatic balancing will be limited to","long_desc":"","tags":[],"see_also":[]},"sleep_interval":{"name":"sleep_interval","type":"secs","level":"advanced","flags":1,"default_value":"60","min":"","max":"","enum_allowed":[],"desc":"how frequently to wake up and attempt optimization","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"update_pg_upmap_activity":{"name":"update_pg_upmap_activity","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Updates pg_upmap activity stats to be used in `balancer status detail`","long_desc":"","tags":[],"see_also":[]},"upmap_max_deviation":{"name":"upmap_max_deviation","type":"int","level":"advanced","flags":1,"default_value":"5","min":"1","max":"","enum_allowed":[],"desc":"deviation below which no optimization is attempted","long_desc":"If the number of PGs are within this count then no optimization is attempted","tags":[],"see_also":[]},"upmap_max_optimizations":{"name":"upmap_max_optimizations","type":"uint","level":"advanced","flags":1,"default_value":"10","min":"","max":"","enum_allowed":[],"desc":"maximum upmap optimizations to make per attempt","long_desc":"","tags":[],"see_also":[]}}},{"name":"cephadm","can_run":true,"error_string":"","module_options":{"agent_down_multiplier":{"name":"agent_down_multiplier","type":"float","level":"advanced","flags":0,"default_value":"3.0","min":"","max":"","enum_allowed":[],"desc":"Multiplied by agent refresh rate to calculate how long agent must not report before being marked down","long_desc":"","tags":[],"see_also":[]},"agent_refresh_rate":{"name":"agent_refresh_rate","type":"secs","level":"advanced","flags":0,"default_value":"20","min":"","max":"","enum_allowed":[],"desc":"How often agent on each host will try to gather and send metadata","long_desc":"","tags":[],"see_also":[]},"agent_starting_port":{"name":"agent_starting_port","type":"int","level":"advanced","flags":0,"default_value":"4721","min":"","max":"","enum_allowed":[],"desc":"First port agent will try to bind to (will also try up to next 1000 subsequent ports if blocked)","long_desc":"","tags":[],"see_also":[]},"allow_ptrace":{"name":"allow_ptrace","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"allow SYS_PTRACE capability on ceph containers","long_desc":"The SYS_PTRACE capability is needed to attach to a process with gdb or strace. Enabling this options can allow debugging daemons that encounter problems at runtime.","tags":[],"see_also":[]},"autotune_interval":{"name":"autotune_interval","type":"secs","level":"advanced","flags":0,"default_value":"600","min":"","max":"","enum_allowed":[],"desc":"how frequently to autotune daemon memory","long_desc":"","tags":[],"see_also":[]},"autotune_memory_target_ratio":{"name":"autotune_memory_target_ratio","type":"float","level":"advanced","flags":0,"default_value":"0.7","min":"","max":"","enum_allowed":[],"desc":"ratio of total system memory to divide amongst autotuned daemons","long_desc":"","tags":[],"see_also":[]},"cephadm_log_destination":{"name":"cephadm_log_destination","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":["file","file,syslog","syslog"],"desc":"Destination for cephadm command's persistent logging","long_desc":"","tags":[],"see_also":[]},"cgroups_split":{"name":"cgroups_split","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"Pass --cgroups=split when cephadm creates containers (currently podman only)","long_desc":"","tags":[],"see_also":[]},"config_checks_enabled":{"name":"config_checks_enabled","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Enable or disable the cephadm configuration analysis","long_desc":"","tags":[],"see_also":[]},"config_dashboard":{"name":"config_dashboard","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"manage configs like API endpoints in Dashboard.","long_desc":"","tags":[],"see_also":[]},"container_image_alertmanager":{"name":"container_image_alertmanager","type":"str","level":"advanced","flags":0,"default_value":"quay.io/prometheus/alertmanager:v0.25.0","min":"","max":"","enum_allowed":[],"desc":"Prometheus container image","long_desc":"","tags":[],"see_also":[]},"container_image_base":{"name":"container_image_base","type":"str","level":"advanced","flags":1,"default_value":"quay.io/ceph/ceph","min":"","max":"","enum_allowed":[],"desc":"Container image name, without the tag","long_desc":"","tags":[],"see_also":[]},"container_image_elasticsearch":{"name":"container_image_elasticsearch","type":"str","level":"advanced","flags":0,"default_value":"quay.io/omrizeneva/elasticsearch:6.8.23","min":"","max":"","enum_allowed":[],"desc":"elasticsearch container image","long_desc":"","tags":[],"see_also":[]},"container_image_grafana":{"name":"container_image_grafana","type":"str","level":"advanced","flags":0,"default_value":"quay.io/ceph/grafana:10.4.0","min":"","max":"","enum_allowed":[],"desc":"Prometheus container image","long_desc":"","tags":[],"see_also":[]},"container_image_haproxy":{"name":"container_image_haproxy","type":"str","level":"advanced","flags":0,"default_value":"quay.io/ceph/haproxy:2.3","min":"","max":"","enum_allowed":[],"desc":"HAproxy container image","long_desc":"","tags":[],"see_also":[]},"container_image_jaeger_agent":{"name":"container_image_jaeger_agent","type":"str","level":"advanced","flags":0,"default_value":"quay.io/jaegertracing/jaeger-agent:1.29","min":"","max":"","enum_allowed":[],"desc":"Jaeger agent container image","long_desc":"","tags":[],"see_also":[]},"container_image_jaeger_collector":{"name":"container_image_jaeger_collector","type":"str","level":"advanced","flags":0,"default_value":"quay.io/jaegertracing/jaeger-collector:1.29","min":"","max":"","enum_allowed":[],"desc":"Jaeger collector container image","long_desc":"","tags":[],"see_also":[]},"container_image_jaeger_query":{"name":"container_image_jaeger_query","type":"str","level":"advanced","flags":0,"default_value":"quay.io/jaegertracing/jaeger-query:1.29","min":"","max":"","enum_allowed":[],"desc":"Jaeger query container image","long_desc":"","tags":[],"see_also":[]},"container_image_keepalived":{"name":"container_image_keepalived","type":"str","level":"advanced","flags":0,"default_value":"quay.io/ceph/keepalived:2.2.4","min":"","max":"","enum_allowed":[],"desc":"Keepalived container image","long_desc":"","tags":[],"see_also":[]},"container_image_loki":{"name":"container_image_loki","type":"str","level":"advanced","flags":0,"default_value":"quay.io/ceph/loki:3.0.0","min":"","max":"","enum_allowed":[],"desc":"Loki container image","long_desc":"","tags":[],"see_also":[]},"container_image_node_exporter":{"name":"container_image_node_exporter","type":"str","level":"advanced","flags":0,"default_value":"quay.io/prometheus/node-exporter:v1.7.0","min":"","max":"","enum_allowed":[],"desc":"Prometheus container image","long_desc":"","tags":[],"see_also":[]},"container_image_nvmeof":{"name":"container_image_nvmeof","type":"str","level":"advanced","flags":0,"default_value":"quay.io/ceph/nvmeof:1.2.5","min":"","max":"","enum_allowed":[],"desc":"Nvme-of container image","long_desc":"","tags":[],"see_also":[]},"container_image_prometheus":{"name":"container_image_prometheus","type":"str","level":"advanced","flags":0,"default_value":"quay.io/prometheus/prometheus:v2.51.0","min":"","max":"","enum_allowed":[],"desc":"Prometheus container image","long_desc":"","tags":[],"see_also":[]},"container_image_promtail":{"name":"container_image_promtail","type":"str","level":"advanced","flags":0,"default_value":"quay.io/ceph/promtail:3.0.0","min":"","max":"","enum_allowed":[],"desc":"Promtail container image","long_desc":"","tags":[],"see_also":[]},"container_image_samba":{"name":"container_image_samba","type":"str","level":"advanced","flags":0,"default_value":"quay.io/samba.org/samba-server:devbuilds-centos-amd64","min":"","max":"","enum_allowed":[],"desc":"Samba/SMB container image","long_desc":"","tags":[],"see_also":[]},"container_image_snmp_gateway":{"name":"container_image_snmp_gateway","type":"str","level":"advanced","flags":0,"default_value":"quay.io/ceph/snmp-notifier:v1.2.1","min":"","max":"","enum_allowed":[],"desc":"SNMP Gateway container image","long_desc":"","tags":[],"see_also":[]},"container_init":{"name":"container_init","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"Run podman/docker with `--init`","long_desc":"","tags":[],"see_also":[]},"daemon_cache_timeout":{"name":"daemon_cache_timeout","type":"secs","level":"advanced","flags":0,"default_value":"600","min":"","max":"","enum_allowed":[],"desc":"seconds to cache service (daemon) inventory","long_desc":"","tags":[],"see_also":[]},"default_cephadm_command_timeout":{"name":"default_cephadm_command_timeout","type":"int","level":"advanced","flags":0,"default_value":"900","min":"","max":"","enum_allowed":[],"desc":"Default timeout applied to cephadm commands run directly on the host (in seconds)","long_desc":"","tags":[],"see_also":[]},"default_registry":{"name":"default_registry","type":"str","level":"advanced","flags":0,"default_value":"quay.io","min":"","max":"","enum_allowed":[],"desc":"Search-registry to which we should normalize unqualified image names. This is not the default registry","long_desc":"","tags":[],"see_also":[]},"device_cache_timeout":{"name":"device_cache_timeout","type":"secs","level":"advanced","flags":0,"default_value":"1800","min":"","max":"","enum_allowed":[],"desc":"seconds to cache device inventory","long_desc":"","tags":[],"see_also":[]},"device_enhanced_scan":{"name":"device_enhanced_scan","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Use libstoragemgmt during device scans","long_desc":"","tags":[],"see_also":[]},"facts_cache_timeout":{"name":"facts_cache_timeout","type":"secs","level":"advanced","flags":0,"default_value":"60","min":"","max":"","enum_allowed":[],"desc":"seconds to cache host facts data","long_desc":"","tags":[],"see_also":[]},"grafana_dashboards_path":{"name":"grafana_dashboards_path","type":"str","level":"advanced","flags":0,"default_value":"/etc/grafana/dashboards/ceph-dashboard/","min":"","max":"","enum_allowed":[],"desc":"location of dashboards to include in grafana deployments","long_desc":"","tags":[],"see_also":[]},"host_check_interval":{"name":"host_check_interval","type":"secs","level":"advanced","flags":0,"default_value":"600","min":"","max":"","enum_allowed":[],"desc":"how frequently to perform a host check","long_desc":"","tags":[],"see_also":[]},"hw_monitoring":{"name":"hw_monitoring","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Deploy hw monitoring daemon on every host.","long_desc":"","tags":[],"see_also":[]},"inventory_list_all":{"name":"inventory_list_all","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Whether ceph-volume inventory should report more devices (mostly mappers (LVs / mpaths), partitions...)","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_refresh_metadata":{"name":"log_refresh_metadata","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Log all refresh metadata. Includes daemon, device, and host info collected regularly. Only has effect if logging at debug level","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"log to the \"cephadm\" cluster log channel\"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"manage_etc_ceph_ceph_conf":{"name":"manage_etc_ceph_ceph_conf","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Manage and own /etc/ceph/ceph.conf on the hosts.","long_desc":"","tags":[],"see_also":[]},"manage_etc_ceph_ceph_conf_hosts":{"name":"manage_etc_ceph_ceph_conf_hosts","type":"str","level":"advanced","flags":0,"default_value":"*","min":"","max":"","enum_allowed":[],"desc":"PlacementSpec describing on which hosts to manage /etc/ceph/ceph.conf","long_desc":"","tags":[],"see_also":[]},"max_count_per_host":{"name":"max_count_per_host","type":"int","level":"advanced","flags":0,"default_value":"10","min":"","max":"","enum_allowed":[],"desc":"max number of daemons per service per host","long_desc":"","tags":[],"see_also":[]},"max_osd_draining_count":{"name":"max_osd_draining_count","type":"int","level":"advanced","flags":0,"default_value":"10","min":"","max":"","enum_allowed":[],"desc":"max number of osds that will be drained simultaneously when osds are removed","long_desc":"","tags":[],"see_also":[]},"migration_current":{"name":"migration_current","type":"int","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"internal - do not modify","long_desc":"","tags":[],"see_also":[]},"mode":{"name":"mode","type":"str","level":"advanced","flags":0,"default_value":"root","min":"","max":"","enum_allowed":["cephadm-package","root"],"desc":"mode for remote execution of cephadm","long_desc":"","tags":[],"see_also":[]},"oob_default_addr":{"name":"oob_default_addr","type":"str","level":"advanced","flags":0,"default_value":"169.254.1.1","min":"","max":"","enum_allowed":[],"desc":"Default address for RedFish API (oob management).","long_desc":"","tags":[],"see_also":[]},"prometheus_alerts_path":{"name":"prometheus_alerts_path","type":"str","level":"advanced","flags":0,"default_value":"/etc/prometheus/ceph/ceph_default_alerts.yml","min":"","max":"","enum_allowed":[],"desc":"location of alerts to include in prometheus deployments","long_desc":"","tags":[],"see_also":[]},"registry_insecure":{"name":"registry_insecure","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Registry is to be considered insecure (no TLS available). Only for development purposes.","long_desc":"","tags":[],"see_also":[]},"registry_password":{"name":"registry_password","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"Custom repository password. Only used for logging into a registry.","long_desc":"","tags":[],"see_also":[]},"registry_url":{"name":"registry_url","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"Registry url for login purposes. This is not the default registry","long_desc":"","tags":[],"see_also":[]},"registry_username":{"name":"registry_username","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"Custom repository username. Only used for logging into a registry.","long_desc":"","tags":[],"see_also":[]},"secure_monitoring_stack":{"name":"secure_monitoring_stack","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Enable TLS security for all the monitoring stack daemons","long_desc":"","tags":[],"see_also":[]},"service_discovery_port":{"name":"service_discovery_port","type":"int","level":"advanced","flags":0,"default_value":"8765","min":"","max":"","enum_allowed":[],"desc":"cephadm service discovery port","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"ssh_config_file":{"name":"ssh_config_file","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"customized SSH config file to connect to managed hosts","long_desc":"","tags":[],"see_also":[]},"ssh_keepalive_count_max":{"name":"ssh_keepalive_count_max","type":"int","level":"advanced","flags":0,"default_value":"3","min":"","max":"","enum_allowed":[],"desc":"How many times ssh connections can fail liveness checks before the host is marked offline","long_desc":"","tags":[],"see_also":[]},"ssh_keepalive_interval":{"name":"ssh_keepalive_interval","type":"int","level":"advanced","flags":0,"default_value":"7","min":"","max":"","enum_allowed":[],"desc":"How often ssh connections are checked for liveness","long_desc":"","tags":[],"see_also":[]},"use_agent":{"name":"use_agent","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Use cephadm agent on each host to gather and send metadata","long_desc":"","tags":[],"see_also":[]},"use_repo_digest":{"name":"use_repo_digest","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"Automatically convert image tags to image digest. Make sure all daemons use the same image","long_desc":"","tags":[],"see_also":[]},"warn_on_failed_host_check":{"name":"warn_on_failed_host_check","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"raise a health warning if the host check fails","long_desc":"","tags":[],"see_also":[]},"warn_on_stray_daemons":{"name":"warn_on_stray_daemons","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"raise a health warning if daemons are detected that are not managed by cephadm","long_desc":"","tags":[],"see_also":[]},"warn_on_stray_hosts":{"name":"warn_on_stray_hosts","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"raise a health warning if daemons are detected on a host that is not managed by cephadm","long_desc":"","tags":[],"see_also":[]}}},{"name":"crash","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"retain_interval":{"name":"retain_interval","type":"secs","level":"advanced","flags":1,"default_value":"31536000","min":"","max":"","enum_allowed":[],"desc":"how long to retain crashes before pruning them","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"warn_recent_interval":{"name":"warn_recent_interval","type":"secs","level":"advanced","flags":1,"default_value":"1209600","min":"","max":"","enum_allowed":[],"desc":"time interval in which to warn about recent crashes","long_desc":"","tags":[],"see_also":[]}}},{"name":"dashboard","can_run":true,"error_string":"","module_options":{"ACCOUNT_LOCKOUT_ATTEMPTS":{"name":"ACCOUNT_LOCKOUT_ATTEMPTS","type":"int","level":"advanced","flags":0,"default_value":"10","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"ALERTMANAGER_API_HOST":{"name":"ALERTMANAGER_API_HOST","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"ALERTMANAGER_API_SSL_VERIFY":{"name":"ALERTMANAGER_API_SSL_VERIFY","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"AUDIT_API_ENABLED":{"name":"AUDIT_API_ENABLED","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"AUDIT_API_LOG_PAYLOAD":{"name":"AUDIT_API_LOG_PAYLOAD","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"ENABLE_BROWSABLE_API":{"name":"ENABLE_BROWSABLE_API","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"FEATURE_TOGGLE_CEPHFS":{"name":"FEATURE_TOGGLE_CEPHFS","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"FEATURE_TOGGLE_DASHBOARD":{"name":"FEATURE_TOGGLE_DASHBOARD","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"FEATURE_TOGGLE_ISCSI":{"name":"FEATURE_TOGGLE_ISCSI","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"FEATURE_TOGGLE_MIRRORING":{"name":"FEATURE_TOGGLE_MIRRORING","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"FEATURE_TOGGLE_NFS":{"name":"FEATURE_TOGGLE_NFS","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"FEATURE_TOGGLE_RBD":{"name":"FEATURE_TOGGLE_RBD","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"FEATURE_TOGGLE_RGW":{"name":"FEATURE_TOGGLE_RGW","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"GANESHA_CLUSTERS_RADOS_POOL_NAMESPACE":{"name":"GANESHA_CLUSTERS_RADOS_POOL_NAMESPACE","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"GRAFANA_API_PASSWORD":{"name":"GRAFANA_API_PASSWORD","type":"str","level":"advanced","flags":0,"default_value":"admin","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"GRAFANA_API_SSL_VERIFY":{"name":"GRAFANA_API_SSL_VERIFY","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"GRAFANA_API_URL":{"name":"GRAFANA_API_URL","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"GRAFANA_API_USERNAME":{"name":"GRAFANA_API_USERNAME","type":"str","level":"advanced","flags":0,"default_value":"admin","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"GRAFANA_FRONTEND_API_URL":{"name":"GRAFANA_FRONTEND_API_URL","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"GRAFANA_UPDATE_DASHBOARDS":{"name":"GRAFANA_UPDATE_DASHBOARDS","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"ISCSI_API_SSL_VERIFICATION":{"name":"ISCSI_API_SSL_VERIFICATION","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"ISSUE_TRACKER_API_KEY":{"name":"ISSUE_TRACKER_API_KEY","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PROMETHEUS_API_HOST":{"name":"PROMETHEUS_API_HOST","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PROMETHEUS_API_SSL_VERIFY":{"name":"PROMETHEUS_API_SSL_VERIFY","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PWD_POLICY_CHECK_COMPLEXITY_ENABLED":{"name":"PWD_POLICY_CHECK_COMPLEXITY_ENABLED","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PWD_POLICY_CHECK_EXCLUSION_LIST_ENABLED":{"name":"PWD_POLICY_CHECK_EXCLUSION_LIST_ENABLED","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PWD_POLICY_CHECK_LENGTH_ENABLED":{"name":"PWD_POLICY_CHECK_LENGTH_ENABLED","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PWD_POLICY_CHECK_OLDPWD_ENABLED":{"name":"PWD_POLICY_CHECK_OLDPWD_ENABLED","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PWD_POLICY_CHECK_REPETITIVE_CHARS_ENABLED":{"name":"PWD_POLICY_CHECK_REPETITIVE_CHARS_ENABLED","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PWD_POLICY_CHECK_SEQUENTIAL_CHARS_ENABLED":{"name":"PWD_POLICY_CHECK_SEQUENTIAL_CHARS_ENABLED","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PWD_POLICY_CHECK_USERNAME_ENABLED":{"name":"PWD_POLICY_CHECK_USERNAME_ENABLED","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PWD_POLICY_ENABLED":{"name":"PWD_POLICY_ENABLED","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PWD_POLICY_EXCLUSION_LIST":{"name":"PWD_POLICY_EXCLUSION_LIST","type":"str","level":"advanced","flags":0,"default_value":"osd,host,dashboard,pool,block,nfs,ceph,monitors,gateway,logs,crush,maps","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PWD_POLICY_MIN_COMPLEXITY":{"name":"PWD_POLICY_MIN_COMPLEXITY","type":"int","level":"advanced","flags":0,"default_value":"10","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PWD_POLICY_MIN_LENGTH":{"name":"PWD_POLICY_MIN_LENGTH","type":"int","level":"advanced","flags":0,"default_value":"8","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"REST_REQUESTS_TIMEOUT":{"name":"REST_REQUESTS_TIMEOUT","type":"int","level":"advanced","flags":0,"default_value":"45","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"RGW_API_ACCESS_KEY":{"name":"RGW_API_ACCESS_KEY","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"RGW_API_ADMIN_RESOURCE":{"name":"RGW_API_ADMIN_RESOURCE","type":"str","level":"advanced","flags":0,"default_value":"admin","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"RGW_API_SECRET_KEY":{"name":"RGW_API_SECRET_KEY","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"RGW_API_SSL_VERIFY":{"name":"RGW_API_SSL_VERIFY","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"USER_PWD_EXPIRATION_SPAN":{"name":"USER_PWD_EXPIRATION_SPAN","type":"int","level":"advanced","flags":0,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"USER_PWD_EXPIRATION_WARNING_1":{"name":"USER_PWD_EXPIRATION_WARNING_1","type":"int","level":"advanced","flags":0,"default_value":"10","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"USER_PWD_EXPIRATION_WARNING_2":{"name":"USER_PWD_EXPIRATION_WARNING_2","type":"int","level":"advanced","flags":0,"default_value":"5","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"cross_origin_url":{"name":"cross_origin_url","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"crt_file":{"name":"crt_file","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"debug":{"name":"debug","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Enable/disable debug options","long_desc":"","tags":[],"see_also":[]},"jwt_token_ttl":{"name":"jwt_token_ttl","type":"int","level":"advanced","flags":0,"default_value":"28800","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"key_file":{"name":"key_file","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"motd":{"name":"motd","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"The message of the day","long_desc":"","tags":[],"see_also":[]},"redirect_resolve_ip_addr":{"name":"redirect_resolve_ip_addr","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"server_addr":{"name":"server_addr","type":"str","level":"advanced","flags":0,"default_value":"::","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"server_port":{"name":"server_port","type":"int","level":"advanced","flags":0,"default_value":"8080","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"ssl":{"name":"ssl","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"ssl_server_port":{"name":"ssl_server_port","type":"int","level":"advanced","flags":0,"default_value":"8443","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"standby_behaviour":{"name":"standby_behaviour","type":"str","level":"advanced","flags":0,"default_value":"redirect","min":"","max":"","enum_allowed":["error","redirect"],"desc":"","long_desc":"","tags":[],"see_also":[]},"standby_error_status_code":{"name":"standby_error_status_code","type":"int","level":"advanced","flags":0,"default_value":"500","min":"400","max":"599","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"url_prefix":{"name":"url_prefix","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"devicehealth","can_run":true,"error_string":"","module_options":{"enable_monitoring":{"name":"enable_monitoring","type":"bool","level":"advanced","flags":1,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"monitor device health metrics","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"mark_out_threshold":{"name":"mark_out_threshold","type":"secs","level":"advanced","flags":1,"default_value":"2419200","min":"","max":"","enum_allowed":[],"desc":"automatically mark OSD if it may fail before this long","long_desc":"","tags":[],"see_also":[]},"pool_name":{"name":"pool_name","type":"str","level":"advanced","flags":1,"default_value":"device_health_metrics","min":"","max":"","enum_allowed":[],"desc":"name of pool in which to store device health metrics","long_desc":"","tags":[],"see_also":[]},"retention_period":{"name":"retention_period","type":"secs","level":"advanced","flags":1,"default_value":"15552000","min":"","max":"","enum_allowed":[],"desc":"how long to retain device health metrics","long_desc":"","tags":[],"see_also":[]},"scrape_frequency":{"name":"scrape_frequency","type":"secs","level":"advanced","flags":1,"default_value":"86400","min":"","max":"","enum_allowed":[],"desc":"how frequently to scrape device health metrics","long_desc":"","tags":[],"see_also":[]},"self_heal":{"name":"self_heal","type":"bool","level":"advanced","flags":1,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"preemptively heal cluster around devices that may fail","long_desc":"","tags":[],"see_also":[]},"sleep_interval":{"name":"sleep_interval","type":"secs","level":"advanced","flags":1,"default_value":"600","min":"","max":"","enum_allowed":[],"desc":"how frequently to wake up and check device health","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"warn_threshold":{"name":"warn_threshold","type":"secs","level":"advanced","flags":1,"default_value":"7257600","min":"","max":"","enum_allowed":[],"desc":"raise health warning if OSD may fail before this long","long_desc":"","tags":[],"see_also":[]}}},{"name":"diskprediction_local","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"predict_interval":{"name":"predict_interval","type":"str","level":"advanced","flags":0,"default_value":"86400","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"predictor_model":{"name":"predictor_model","type":"str","level":"advanced","flags":0,"default_value":"prophetstor","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sleep_interval":{"name":"sleep_interval","type":"str","level":"advanced","flags":0,"default_value":"600","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"influx","can_run":false,"error_string":"influxdb python module not found","module_options":{"batch_size":{"name":"batch_size","type":"int","level":"advanced","flags":0,"default_value":"5000","min":"","max":"","enum_allowed":[],"desc":"How big batches of data points should be when sending to InfluxDB.","long_desc":"","tags":[],"see_also":[]},"database":{"name":"database","type":"str","level":"advanced","flags":0,"default_value":"ceph","min":"","max":"","enum_allowed":[],"desc":"InfluxDB database name. You will need to create this database and grant write privileges to the configured username or the username must have admin privileges to create it.","long_desc":"","tags":[],"see_also":[]},"hostname":{"name":"hostname","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"InfluxDB server hostname","long_desc":"","tags":[],"see_also":[]},"interval":{"name":"interval","type":"secs","level":"advanced","flags":0,"default_value":"30","min":"5","max":"","enum_allowed":[],"desc":"Time between reports to InfluxDB. Default 30 seconds.","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"password":{"name":"password","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"password of InfluxDB server user","long_desc":"","tags":[],"see_also":[]},"port":{"name":"port","type":"int","level":"advanced","flags":0,"default_value":"8086","min":"","max":"","enum_allowed":[],"desc":"InfluxDB server port","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"ssl":{"name":"ssl","type":"str","level":"advanced","flags":0,"default_value":"false","min":"","max":"","enum_allowed":[],"desc":"Use https connection for InfluxDB server. Use \"true\" or \"false\".","long_desc":"","tags":[],"see_also":[]},"threads":{"name":"threads","type":"int","level":"advanced","flags":0,"default_value":"5","min":"1","max":"32","enum_allowed":[],"desc":"How many worker threads should be spawned for sending data to InfluxDB.","long_desc":"","tags":[],"see_also":[]},"username":{"name":"username","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"username of InfluxDB server user","long_desc":"","tags":[],"see_also":[]},"verify_ssl":{"name":"verify_ssl","type":"str","level":"advanced","flags":0,"default_value":"true","min":"","max":"","enum_allowed":[],"desc":"Verify https cert for InfluxDB server. Use \"true\" or \"false\".","long_desc":"","tags":[],"see_also":[]}}},{"name":"insights","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"iostat","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"k8sevents","can_run":true,"error_string":"","module_options":{"ceph_event_retention_days":{"name":"ceph_event_retention_days","type":"int","level":"advanced","flags":0,"default_value":"7","min":"","max":"","enum_allowed":[],"desc":"Days to hold ceph event information within local cache","long_desc":"","tags":[],"see_also":[]},"config_check_secs":{"name":"config_check_secs","type":"int","level":"advanced","flags":0,"default_value":"10","min":"10","max":"","enum_allowed":[],"desc":"interval (secs) to check for cluster configuration changes","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"localpool","can_run":true,"error_string":"","module_options":{"failure_domain":{"name":"failure_domain","type":"str","level":"advanced","flags":1,"default_value":"host","min":"","max":"","enum_allowed":[],"desc":"failure domain for any created local pool","long_desc":"what failure domain we should separate data replicas across.","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"min_size":{"name":"min_size","type":"int","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"default min_size for any created local pool","long_desc":"value to set min_size to (unchanged from Ceph's default if this option is not set)","tags":[],"see_also":[]},"num_rep":{"name":"num_rep","type":"int","level":"advanced","flags":1,"default_value":"3","min":"","max":"","enum_allowed":[],"desc":"default replica count for any created local pool","long_desc":"","tags":[],"see_also":[]},"pg_num":{"name":"pg_num","type":"int","level":"advanced","flags":1,"default_value":"128","min":"","max":"","enum_allowed":[],"desc":"default pg_num for any created local pool","long_desc":"","tags":[],"see_also":[]},"prefix":{"name":"prefix","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"name prefix for any created local pool","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"subtree":{"name":"subtree","type":"str","level":"advanced","flags":1,"default_value":"rack","min":"","max":"","enum_allowed":[],"desc":"CRUSH level for which to create a local pool","long_desc":"which CRUSH subtree type the module should create a pool for.","tags":[],"see_also":[]}}},{"name":"mds_autoscaler","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"mirroring","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"nfs","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"orchestrator","can_run":true,"error_string":"","module_options":{"fail_fs":{"name":"fail_fs","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Fail filesystem for rapid multi-rank mds upgrade","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"orchestrator":{"name":"orchestrator","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["cephadm","rook","test_orchestrator"],"desc":"Orchestrator backend","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"osd_perf_query","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"osd_support","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"pg_autoscaler","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sleep_interval":{"name":"sleep_interval","type":"secs","level":"advanced","flags":0,"default_value":"60","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"threshold":{"name":"threshold","type":"float","level":"advanced","flags":0,"default_value":"3.0","min":"1.0","max":"","enum_allowed":[],"desc":"scaling threshold","long_desc":"The factor by which the `NEW PG_NUM` must vary from the current`PG_NUM` before being accepted. Cannot be less than 1.0","tags":[],"see_also":[]}}},{"name":"progress","can_run":true,"error_string":"","module_options":{"allow_pg_recovery_event":{"name":"allow_pg_recovery_event","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"allow the module to show pg recovery progress","long_desc":"","tags":[],"see_also":[]},"enabled":{"name":"enabled","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"max_completed_events":{"name":"max_completed_events","type":"int","level":"advanced","flags":1,"default_value":"50","min":"","max":"","enum_allowed":[],"desc":"number of past completed events to remember","long_desc":"","tags":[],"see_also":[]},"sleep_interval":{"name":"sleep_interval","type":"secs","level":"advanced","flags":1,"default_value":"5","min":"","max":"","enum_allowed":[],"desc":"how long the module is going to sleep","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"prometheus","can_run":true,"error_string":"","module_options":{"cache":{"name":"cache","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"exclude_perf_counters":{"name":"exclude_perf_counters","type":"bool","level":"advanced","flags":1,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"Do not include perf-counters in the metrics output","long_desc":"Gathering perf-counters from a single Prometheus exporter can degrade ceph-mgr performance, especially in large clusters. Instead, Ceph-exporter daemons are now used by default for perf-counter gathering. This should only be disabled when no ceph-exporters are deployed.","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"rbd_stats_pools":{"name":"rbd_stats_pools","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"rbd_stats_pools_refresh_interval":{"name":"rbd_stats_pools_refresh_interval","type":"int","level":"advanced","flags":0,"default_value":"300","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"scrape_interval":{"name":"scrape_interval","type":"float","level":"advanced","flags":0,"default_value":"15.0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"server_addr":{"name":"server_addr","type":"str","level":"advanced","flags":0,"default_value":"::","min":"","max":"","enum_allowed":[],"desc":"the IPv4 or IPv6 address on which the module listens for HTTP requests","long_desc":"","tags":[],"see_also":[]},"server_port":{"name":"server_port","type":"int","level":"advanced","flags":1,"default_value":"9283","min":"","max":"","enum_allowed":[],"desc":"the port on which the module listens for HTTP requests","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"stale_cache_strategy":{"name":"stale_cache_strategy","type":"str","level":"advanced","flags":0,"default_value":"log","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"standby_behaviour":{"name":"standby_behaviour","type":"str","level":"advanced","flags":1,"default_value":"default","min":"","max":"","enum_allowed":["default","error"],"desc":"","long_desc":"","tags":[],"see_also":[]},"standby_error_status_code":{"name":"standby_error_status_code","type":"int","level":"advanced","flags":1,"default_value":"500","min":"400","max":"599","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"rbd_support","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"max_concurrent_snap_create":{"name":"max_concurrent_snap_create","type":"int","level":"advanced","flags":0,"default_value":"10","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"mirror_snapshot_schedule":{"name":"mirror_snapshot_schedule","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"trash_purge_schedule":{"name":"trash_purge_schedule","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"restful","can_run":true,"error_string":"","module_options":{"enable_auth":{"name":"enable_auth","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"key_file":{"name":"key_file","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"max_requests":{"name":"max_requests","type":"int","level":"advanced","flags":0,"default_value":"500","min":"","max":"","enum_allowed":[],"desc":"Maximum number of requests to keep in memory. When new request comes in, the oldest request will be removed if the number of requests exceeds the max request number. if un-finished request is removed, error message will be logged in the ceph-mgr log.","long_desc":"","tags":[],"see_also":[]},"server_addr":{"name":"server_addr","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"server_port":{"name":"server_port","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"rgw","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"secondary_zone_period_retry_limit":{"name":"secondary_zone_period_retry_limit","type":"int","level":"advanced","flags":0,"default_value":"5","min":"","max":"","enum_allowed":[],"desc":"RGW module period update retry limit for secondary site","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"rook","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"storage_class":{"name":"storage_class","type":"str","level":"advanced","flags":0,"default_value":"local","min":"","max":"","enum_allowed":[],"desc":"storage class name for LSO-discovered PVs","long_desc":"","tags":[],"see_also":[]}}},{"name":"selftest","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"roption1":{"name":"roption1","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"roption2":{"name":"roption2","type":"str","level":"advanced","flags":0,"default_value":"xyz","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"rwoption1":{"name":"rwoption1","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"rwoption2":{"name":"rwoption2","type":"int","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"rwoption3":{"name":"rwoption3","type":"float","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"rwoption4":{"name":"rwoption4","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"rwoption5":{"name":"rwoption5","type":"bool","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"rwoption6":{"name":"rwoption6","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"rwoption7":{"name":"rwoption7","type":"int","level":"advanced","flags":0,"default_value":"","min":"1","max":"42","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"testkey":{"name":"testkey","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"testlkey":{"name":"testlkey","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"testnewline":{"name":"testnewline","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"snap_schedule","can_run":true,"error_string":"","module_options":{"allow_m_granularity":{"name":"allow_m_granularity","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"allow minute scheduled snapshots","long_desc":"","tags":[],"see_also":[]},"dump_on_update":{"name":"dump_on_update","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"dump database to debug log on update","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"stats","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"status","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"telegraf","can_run":true,"error_string":"","module_options":{"address":{"name":"address","type":"str","level":"advanced","flags":0,"default_value":"unixgram:///tmp/telegraf.sock","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"interval":{"name":"interval","type":"secs","level":"advanced","flags":0,"default_value":"15","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"telemetry","can_run":true,"error_string":"","module_options":{"channel_basic":{"name":"channel_basic","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"Share basic cluster information (size, version)","long_desc":"","tags":[],"see_also":[]},"channel_crash":{"name":"channel_crash","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"Share metadata about Ceph daemon crashes (version, stack straces, etc)","long_desc":"","tags":[],"see_also":[]},"channel_device":{"name":"channel_device","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"Share device health metrics (e.g., SMART data, minus potentially identifying info like serial numbers)","long_desc":"","tags":[],"see_also":[]},"channel_ident":{"name":"channel_ident","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Share a user-provided description and/or contact email for the cluster","long_desc":"","tags":[],"see_also":[]},"channel_perf":{"name":"channel_perf","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Share various performance metrics of a cluster","long_desc":"","tags":[],"see_also":[]},"contact":{"name":"contact","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"description":{"name":"description","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"device_url":{"name":"device_url","type":"str","level":"advanced","flags":0,"default_value":"https://telemetry.ceph.com/device","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"enabled":{"name":"enabled","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"interval":{"name":"interval","type":"int","level":"advanced","flags":0,"default_value":"24","min":"8","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"last_opt_revision":{"name":"last_opt_revision","type":"int","level":"advanced","flags":0,"default_value":"1","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"leaderboard":{"name":"leaderboard","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"leaderboard_description":{"name":"leaderboard_description","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"organization":{"name":"organization","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"proxy":{"name":"proxy","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"url":{"name":"url","type":"str","level":"advanced","flags":0,"default_value":"https://telemetry.ceph.com/report","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"test_orchestrator","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"volumes","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"max_concurrent_clones":{"name":"max_concurrent_clones","type":"int","level":"advanced","flags":0,"default_value":"4","min":"","max":"","enum_allowed":[],"desc":"Number of asynchronous cloner threads","long_desc":"","tags":[],"see_also":[]},"periodic_async_work":{"name":"periodic_async_work","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Periodically check for async work","long_desc":"","tags":[],"see_also":[]},"snapshot_clone_delay":{"name":"snapshot_clone_delay","type":"int","level":"advanced","flags":0,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"Delay clone begin operation by snapshot_clone_delay seconds","long_desc":"","tags":[],"see_also":[]},"snapshot_clone_no_wait":{"name":"snapshot_clone_no_wait","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"Reject subvolume clone request when cloner threads are busy","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"zabbix","can_run":true,"error_string":"","module_options":{"discovery_interval":{"name":"discovery_interval","type":"uint","level":"advanced","flags":0,"default_value":"100","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"identifier":{"name":"identifier","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"interval":{"name":"interval","type":"secs","level":"advanced","flags":0,"default_value":"60","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"zabbix_host":{"name":"zabbix_host","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"zabbix_port":{"name":"zabbix_port","type":"int","level":"advanced","flags":0,"default_value":"10051","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"zabbix_sender":{"name":"zabbix_sender","type":"str","level":"advanced","flags":0,"default_value":"/usr/bin/zabbix_sender","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}}]}],"modules":["cephadm","dashboard","iostat","nfs","restful"],"available_modules":[{"name":"alerts","can_run":true,"error_string":"","module_options":{"interval":{"name":"interval","type":"secs","level":"advanced","flags":1,"default_value":"60","min":"","max":"","enum_allowed":[],"desc":"How frequently to reexamine health status","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"smtp_destination":{"name":"smtp_destination","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"Email address to send alerts to","long_desc":"","tags":[],"see_also":[]},"smtp_from_name":{"name":"smtp_from_name","type":"str","level":"advanced","flags":1,"default_value":"Ceph","min":"","max":"","enum_allowed":[],"desc":"Email From: name","long_desc":"","tags":[],"see_also":[]},"smtp_host":{"name":"smtp_host","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"SMTP server","long_desc":"","tags":[],"see_also":[]},"smtp_password":{"name":"smtp_password","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"Password to authenticate with","long_desc":"","tags":[],"see_also":[]},"smtp_port":{"name":"smtp_port","type":"int","level":"advanced","flags":1,"default_value":"465","min":"","max":"","enum_allowed":[],"desc":"SMTP port","long_desc":"","tags":[],"see_also":[]},"smtp_sender":{"name":"smtp_sender","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"SMTP envelope sender","long_desc":"","tags":[],"see_also":[]},"smtp_ssl":{"name":"smtp_ssl","type":"bool","level":"advanced","flags":1,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"Use SSL to connect to SMTP server","long_desc":"","tags":[],"see_also":[]},"smtp_user":{"name":"smtp_user","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"User to authenticate as","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"balancer","can_run":true,"error_string":"","module_options":{"active":{"name":"active","type":"bool","level":"advanced","flags":1,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"automatically balance PGs across cluster","long_desc":"","tags":[],"see_also":[]},"begin_time":{"name":"begin_time","type":"str","level":"advanced","flags":1,"default_value":"0000","min":"","max":"","enum_allowed":[],"desc":"beginning time of day to automatically balance","long_desc":"This is a time of day in the format HHMM.","tags":[],"see_also":[]},"begin_weekday":{"name":"begin_weekday","type":"uint","level":"advanced","flags":1,"default_value":"0","min":"0","max":"6","enum_allowed":[],"desc":"Restrict automatic balancing to this day of the week or later","long_desc":"0 = Sunday, 1 = Monday, etc.","tags":[],"see_also":[]},"crush_compat_max_iterations":{"name":"crush_compat_max_iterations","type":"uint","level":"advanced","flags":1,"default_value":"25","min":"1","max":"250","enum_allowed":[],"desc":"maximum number of iterations to attempt optimization","long_desc":"","tags":[],"see_also":[]},"crush_compat_metrics":{"name":"crush_compat_metrics","type":"str","level":"advanced","flags":1,"default_value":"pgs,objects,bytes","min":"","max":"","enum_allowed":[],"desc":"metrics with which to calculate OSD utilization","long_desc":"Value is a list of one or more of \"pgs\", \"objects\", or \"bytes\", and indicates which metrics to use to balance utilization.","tags":[],"see_also":[]},"crush_compat_step":{"name":"crush_compat_step","type":"float","level":"advanced","flags":1,"default_value":"0.5","min":"0.001","max":"0.999","enum_allowed":[],"desc":"aggressiveness of optimization","long_desc":".99 is very aggressive, .01 is less aggressive","tags":[],"see_also":[]},"end_time":{"name":"end_time","type":"str","level":"advanced","flags":1,"default_value":"2359","min":"","max":"","enum_allowed":[],"desc":"ending time of day to automatically balance","long_desc":"This is a time of day in the format HHMM.","tags":[],"see_also":[]},"end_weekday":{"name":"end_weekday","type":"uint","level":"advanced","flags":1,"default_value":"0","min":"0","max":"6","enum_allowed":[],"desc":"Restrict automatic balancing to days of the week earlier than this","long_desc":"0 = Sunday, 1 = Monday, etc.","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"min_score":{"name":"min_score","type":"float","level":"advanced","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"minimum score, below which no optimization is attempted","long_desc":"","tags":[],"see_also":[]},"mode":{"name":"mode","type":"str","level":"advanced","flags":1,"default_value":"upmap","min":"","max":"","enum_allowed":["crush-compat","none","read","upmap","upmap-read"],"desc":"Balancer mode","long_desc":"","tags":[],"see_also":[]},"pool_ids":{"name":"pool_ids","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"pools which the automatic balancing will be limited to","long_desc":"","tags":[],"see_also":[]},"sleep_interval":{"name":"sleep_interval","type":"secs","level":"advanced","flags":1,"default_value":"60","min":"","max":"","enum_allowed":[],"desc":"how frequently to wake up and attempt optimization","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"update_pg_upmap_activity":{"name":"update_pg_upmap_activity","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Updates pg_upmap activity stats to be used in `balancer status detail`","long_desc":"","tags":[],"see_also":[]},"upmap_max_deviation":{"name":"upmap_max_deviation","type":"int","level":"advanced","flags":1,"default_value":"5","min":"1","max":"","enum_allowed":[],"desc":"deviation below which no optimization is attempted","long_desc":"If the number of PGs are within this count then no optimization is attempted","tags":[],"see_also":[]},"upmap_max_optimizations":{"name":"upmap_max_optimizations","type":"uint","level":"advanced","flags":1,"default_value":"10","min":"","max":"","enum_allowed":[],"desc":"maximum upmap optimizations to make per attempt","long_desc":"","tags":[],"see_also":[]}}},{"name":"cephadm","can_run":true,"error_string":"","module_options":{"agent_down_multiplier":{"name":"agent_down_multiplier","type":"float","level":"advanced","flags":0,"default_value":"3.0","min":"","max":"","enum_allowed":[],"desc":"Multiplied by agent refresh rate to calculate how long agent must not report before being marked down","long_desc":"","tags":[],"see_also":[]},"agent_refresh_rate":{"name":"agent_refresh_rate","type":"secs","level":"advanced","flags":0,"default_value":"20","min":"","max":"","enum_allowed":[],"desc":"How often agent on each host will try to gather and send metadata","long_desc":"","tags":[],"see_also":[]},"agent_starting_port":{"name":"agent_starting_port","type":"int","level":"advanced","flags":0,"default_value":"4721","min":"","max":"","enum_allowed":[],"desc":"First port agent will try to bind to (will also try up to next 1000 subsequent ports if blocked)","long_desc":"","tags":[],"see_also":[]},"allow_ptrace":{"name":"allow_ptrace","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"allow SYS_PTRACE capability on ceph containers","long_desc":"The SYS_PTRACE capability is needed to attach to a process with gdb or strace. Enabling this options can allow debugging daemons that encounter problems at runtime.","tags":[],"see_also":[]},"autotune_interval":{"name":"autotune_interval","type":"secs","level":"advanced","flags":0,"default_value":"600","min":"","max":"","enum_allowed":[],"desc":"how frequently to autotune daemon memory","long_desc":"","tags":[],"see_also":[]},"autotune_memory_target_ratio":{"name":"autotune_memory_target_ratio","type":"float","level":"advanced","flags":0,"default_value":"0.7","min":"","max":"","enum_allowed":[],"desc":"ratio of total system memory to divide amongst autotuned daemons","long_desc":"","tags":[],"see_also":[]},"cephadm_log_destination":{"name":"cephadm_log_destination","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":["file","file,syslog","syslog"],"desc":"Destination for cephadm command's persistent logging","long_desc":"","tags":[],"see_also":[]},"cgroups_split":{"name":"cgroups_split","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"Pass --cgroups=split when cephadm creates containers (currently podman only)","long_desc":"","tags":[],"see_also":[]},"config_checks_enabled":{"name":"config_checks_enabled","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Enable or disable the cephadm configuration analysis","long_desc":"","tags":[],"see_also":[]},"config_dashboard":{"name":"config_dashboard","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"manage configs like API endpoints in Dashboard.","long_desc":"","tags":[],"see_also":[]},"container_image_alertmanager":{"name":"container_image_alertmanager","type":"str","level":"advanced","flags":0,"default_value":"quay.io/prometheus/alertmanager:v0.25.0","min":"","max":"","enum_allowed":[],"desc":"Prometheus container image","long_desc":"","tags":[],"see_also":[]},"container_image_base":{"name":"container_image_base","type":"str","level":"advanced","flags":1,"default_value":"quay.io/ceph/ceph","min":"","max":"","enum_allowed":[],"desc":"Container image name, without the tag","long_desc":"","tags":[],"see_also":[]},"container_image_elasticsearch":{"name":"container_image_elasticsearch","type":"str","level":"advanced","flags":0,"default_value":"quay.io/omrizeneva/elasticsearch:6.8.23","min":"","max":"","enum_allowed":[],"desc":"elasticsearch container image","long_desc":"","tags":[],"see_also":[]},"container_image_grafana":{"name":"container_image_grafana","type":"str","level":"advanced","flags":0,"default_value":"quay.io/ceph/grafana:10.4.0","min":"","max":"","enum_allowed":[],"desc":"Prometheus container image","long_desc":"","tags":[],"see_also":[]},"container_image_haproxy":{"name":"container_image_haproxy","type":"str","level":"advanced","flags":0,"default_value":"quay.io/ceph/haproxy:2.3","min":"","max":"","enum_allowed":[],"desc":"HAproxy container image","long_desc":"","tags":[],"see_also":[]},"container_image_jaeger_agent":{"name":"container_image_jaeger_agent","type":"str","level":"advanced","flags":0,"default_value":"quay.io/jaegertracing/jaeger-agent:1.29","min":"","max":"","enum_allowed":[],"desc":"Jaeger agent container image","long_desc":"","tags":[],"see_also":[]},"container_image_jaeger_collector":{"name":"container_image_jaeger_collector","type":"str","level":"advanced","flags":0,"default_value":"quay.io/jaegertracing/jaeger-collector:1.29","min":"","max":"","enum_allowed":[],"desc":"Jaeger collector container image","long_desc":"","tags":[],"see_also":[]},"container_image_jaeger_query":{"name":"container_image_jaeger_query","type":"str","level":"advanced","flags":0,"default_value":"quay.io/jaegertracing/jaeger-query:1.29","min":"","max":"","enum_allowed":[],"desc":"Jaeger query container image","long_desc":"","tags":[],"see_also":[]},"container_image_keepalived":{"name":"container_image_keepalived","type":"str","level":"advanced","flags":0,"default_value":"quay.io/ceph/keepalived:2.2.4","min":"","max":"","enum_allowed":[],"desc":"Keepalived container image","long_desc":"","tags":[],"see_also":[]},"container_image_loki":{"name":"container_image_loki","type":"str","level":"advanced","flags":0,"default_value":"quay.io/ceph/loki:3.0.0","min":"","max":"","enum_allowed":[],"desc":"Loki container image","long_desc":"","tags":[],"see_also":[]},"container_image_node_exporter":{"name":"container_image_node_exporter","type":"str","level":"advanced","flags":0,"default_value":"quay.io/prometheus/node-exporter:v1.7.0","min":"","max":"","enum_allowed":[],"desc":"Prometheus container image","long_desc":"","tags":[],"see_also":[]},"container_image_nvmeof":{"name":"container_image_nvmeof","type":"str","level":"advanced","flags":0,"default_value":"quay.io/ceph/nvmeof:1.2.5","min":"","max":"","enum_allowed":[],"desc":"Nvme-of container image","long_desc":"","tags":[],"see_also":[]},"container_image_prometheus":{"name":"container_image_prometheus","type":"str","level":"advanced","flags":0,"default_value":"quay.io/prometheus/prometheus:v2.51.0","min":"","max":"","enum_allowed":[],"desc":"Prometheus container image","long_desc":"","tags":[],"see_also":[]},"container_image_promtail":{"name":"container_image_promtail","type":"str","level":"advanced","flags":0,"default_value":"quay.io/ceph/promtail:3.0.0","min":"","max":"","enum_allowed":[],"desc":"Promtail container image","long_desc":"","tags":[],"see_also":[]},"container_image_samba":{"name":"container_image_samba","type":"str","level":"advanced","flags":0,"default_value":"quay.io/samba.org/samba-server:devbuilds-centos-amd64","min":"","max":"","enum_allowed":[],"desc":"Samba/SMB container image","long_desc":"","tags":[],"see_also":[]},"container_image_snmp_gateway":{"name":"container_image_snmp_gateway","type":"str","level":"advanced","flags":0,"default_value":"quay.io/ceph/snmp-notifier:v1.2.1","min":"","max":"","enum_allowed":[],"desc":"SNMP Gateway container image","long_desc":"","tags":[],"see_also":[]},"container_init":{"name":"container_init","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"Run podman/docker with `--init`","long_desc":"","tags":[],"see_also":[]},"daemon_cache_timeout":{"name":"daemon_cache_timeout","type":"secs","level":"advanced","flags":0,"default_value":"600","min":"","max":"","enum_allowed":[],"desc":"seconds to cache service (daemon) inventory","long_desc":"","tags":[],"see_also":[]},"default_cephadm_command_timeout":{"name":"default_cephadm_command_timeout","type":"int","level":"advanced","flags":0,"default_value":"900","min":"","max":"","enum_allowed":[],"desc":"Default timeout applied to cephadm commands run directly on the host (in seconds)","long_desc":"","tags":[],"see_also":[]},"default_registry":{"name":"default_registry","type":"str","level":"advanced","flags":0,"default_value":"quay.io","min":"","max":"","enum_allowed":[],"desc":"Search-registry to which we should normalize unqualified image names. This is not the default registry","long_desc":"","tags":[],"see_also":[]},"device_cache_timeout":{"name":"device_cache_timeout","type":"secs","level":"advanced","flags":0,"default_value":"1800","min":"","max":"","enum_allowed":[],"desc":"seconds to cache device inventory","long_desc":"","tags":[],"see_also":[]},"device_enhanced_scan":{"name":"device_enhanced_scan","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Use libstoragemgmt during device scans","long_desc":"","tags":[],"see_also":[]},"facts_cache_timeout":{"name":"facts_cache_timeout","type":"secs","level":"advanced","flags":0,"default_value":"60","min":"","max":"","enum_allowed":[],"desc":"seconds to cache host facts data","long_desc":"","tags":[],"see_also":[]},"grafana_dashboards_path":{"name":"grafana_dashboards_path","type":"str","level":"advanced","flags":0,"default_value":"/etc/grafana/dashboards/ceph-dashboard/","min":"","max":"","enum_allowed":[],"desc":"location of dashboards to include in grafana deployments","long_desc":"","tags":[],"see_also":[]},"host_check_interval":{"name":"host_check_interval","type":"secs","level":"advanced","flags":0,"default_value":"600","min":"","max":"","enum_allowed":[],"desc":"how frequently to perform a host check","long_desc":"","tags":[],"see_also":[]},"hw_monitoring":{"name":"hw_monitoring","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Deploy hw monitoring daemon on every host.","long_desc":"","tags":[],"see_also":[]},"inventory_list_all":{"name":"inventory_list_all","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Whether ceph-volume inventory should report more devices (mostly mappers (LVs / mpaths), partitions...)","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_refresh_metadata":{"name":"log_refresh_metadata","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Log all refresh metadata. Includes daemon, device, and host info collected regularly. Only has effect if logging at debug level","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"log to the \"cephadm\" cluster log channel\"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"manage_etc_ceph_ceph_conf":{"name":"manage_etc_ceph_ceph_conf","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Manage and own /etc/ceph/ceph.conf on the hosts.","long_desc":"","tags":[],"see_also":[]},"manage_etc_ceph_ceph_conf_hosts":{"name":"manage_etc_ceph_ceph_conf_hosts","type":"str","level":"advanced","flags":0,"default_value":"*","min":"","max":"","enum_allowed":[],"desc":"PlacementSpec describing on which hosts to manage /etc/ceph/ceph.conf","long_desc":"","tags":[],"see_also":[]},"max_count_per_host":{"name":"max_count_per_host","type":"int","level":"advanced","flags":0,"default_value":"10","min":"","max":"","enum_allowed":[],"desc":"max number of daemons per service per host","long_desc":"","tags":[],"see_also":[]},"max_osd_draining_count":{"name":"max_osd_draining_count","type":"int","level":"advanced","flags":0,"default_value":"10","min":"","max":"","enum_allowed":[],"desc":"max number of osds that will be drained simultaneously when osds are removed","long_desc":"","tags":[],"see_also":[]},"migration_current":{"name":"migration_current","type":"int","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"internal - do not modify","long_desc":"","tags":[],"see_also":[]},"mode":{"name":"mode","type":"str","level":"advanced","flags":0,"default_value":"root","min":"","max":"","enum_allowed":["cephadm-package","root"],"desc":"mode for remote execution of cephadm","long_desc":"","tags":[],"see_also":[]},"oob_default_addr":{"name":"oob_default_addr","type":"str","level":"advanced","flags":0,"default_value":"169.254.1.1","min":"","max":"","enum_allowed":[],"desc":"Default address for RedFish API (oob management).","long_desc":"","tags":[],"see_also":[]},"prometheus_alerts_path":{"name":"prometheus_alerts_path","type":"str","level":"advanced","flags":0,"default_value":"/etc/prometheus/ceph/ceph_default_alerts.yml","min":"","max":"","enum_allowed":[],"desc":"location of alerts to include in prometheus deployments","long_desc":"","tags":[],"see_also":[]},"registry_insecure":{"name":"registry_insecure","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Registry is to be considered insecure (no TLS available). Only for development purposes.","long_desc":"","tags":[],"see_also":[]},"registry_password":{"name":"registry_password","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"Custom repository password. Only used for logging into a registry.","long_desc":"","tags":[],"see_also":[]},"registry_url":{"name":"registry_url","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"Registry url for login purposes. This is not the default registry","long_desc":"","tags":[],"see_also":[]},"registry_username":{"name":"registry_username","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"Custom repository username. Only used for logging into a registry.","long_desc":"","tags":[],"see_also":[]},"secure_monitoring_stack":{"name":"secure_monitoring_stack","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Enable TLS security for all the monitoring stack daemons","long_desc":"","tags":[],"see_also":[]},"service_discovery_port":{"name":"service_discovery_port","type":"int","level":"advanced","flags":0,"default_value":"8765","min":"","max":"","enum_allowed":[],"desc":"cephadm service discovery port","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"ssh_config_file":{"name":"ssh_config_file","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"customized SSH config file to connect to managed hosts","long_desc":"","tags":[],"see_also":[]},"ssh_keepalive_count_max":{"name":"ssh_keepalive_count_max","type":"int","level":"advanced","flags":0,"default_value":"3","min":"","max":"","enum_allowed":[],"desc":"How many times ssh connections can fail liveness checks before the host is marked offline","long_desc":"","tags":[],"see_also":[]},"ssh_keepalive_interval":{"name":"ssh_keepalive_interval","type":"int","level":"advanced","flags":0,"default_value":"7","min":"","max":"","enum_allowed":[],"desc":"How often ssh connections are checked for liveness","long_desc":"","tags":[],"see_also":[]},"use_agent":{"name":"use_agent","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Use cephadm agent on each host to gather and send metadata","long_desc":"","tags":[],"see_also":[]},"use_repo_digest":{"name":"use_repo_digest","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"Automatically convert image tags to image digest. Make sure all daemons use the same image","long_desc":"","tags":[],"see_also":[]},"warn_on_failed_host_check":{"name":"warn_on_failed_host_check","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"raise a health warning if the host check fails","long_desc":"","tags":[],"see_also":[]},"warn_on_stray_daemons":{"name":"warn_on_stray_daemons","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"raise a health warning if daemons are detected that are not managed by cephadm","long_desc":"","tags":[],"see_also":[]},"warn_on_stray_hosts":{"name":"warn_on_stray_hosts","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"raise a health warning if daemons are detected on a host that is not managed by cephadm","long_desc":"","tags":[],"see_also":[]}}},{"name":"crash","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"retain_interval":{"name":"retain_interval","type":"secs","level":"advanced","flags":1,"default_value":"31536000","min":"","max":"","enum_allowed":[],"desc":"how long to retain crashes before pruning them","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"warn_recent_interval":{"name":"warn_recent_interval","type":"secs","level":"advanced","flags":1,"default_value":"1209600","min":"","max":"","enum_allowed":[],"desc":"time interval in which to warn about recent crashes","long_desc":"","tags":[],"see_also":[]}}},{"name":"dashboard","can_run":true,"error_string":"","module_options":{"ACCOUNT_LOCKOUT_ATTEMPTS":{"name":"ACCOUNT_LOCKOUT_ATTEMPTS","type":"int","level":"advanced","flags":0,"default_value":"10","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"ALERTMANAGER_API_HOST":{"name":"ALERTMANAGER_API_HOST","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"ALERTMANAGER_API_SSL_VERIFY":{"name":"ALERTMANAGER_API_SSL_VERIFY","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"AUDIT_API_ENABLED":{"name":"AUDIT_API_ENABLED","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"AUDIT_API_LOG_PAYLOAD":{"name":"AUDIT_API_LOG_PAYLOAD","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"ENABLE_BROWSABLE_API":{"name":"ENABLE_BROWSABLE_API","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"FEATURE_TOGGLE_CEPHFS":{"name":"FEATURE_TOGGLE_CEPHFS","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"FEATURE_TOGGLE_DASHBOARD":{"name":"FEATURE_TOGGLE_DASHBOARD","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"FEATURE_TOGGLE_ISCSI":{"name":"FEATURE_TOGGLE_ISCSI","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"FEATURE_TOGGLE_MIRRORING":{"name":"FEATURE_TOGGLE_MIRRORING","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"FEATURE_TOGGLE_NFS":{"name":"FEATURE_TOGGLE_NFS","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"FEATURE_TOGGLE_RBD":{"name":"FEATURE_TOGGLE_RBD","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"FEATURE_TOGGLE_RGW":{"name":"FEATURE_TOGGLE_RGW","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"GANESHA_CLUSTERS_RADOS_POOL_NAMESPACE":{"name":"GANESHA_CLUSTERS_RADOS_POOL_NAMESPACE","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"GRAFANA_API_PASSWORD":{"name":"GRAFANA_API_PASSWORD","type":"str","level":"advanced","flags":0,"default_value":"admin","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"GRAFANA_API_SSL_VERIFY":{"name":"GRAFANA_API_SSL_VERIFY","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"GRAFANA_API_URL":{"name":"GRAFANA_API_URL","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"GRAFANA_API_USERNAME":{"name":"GRAFANA_API_USERNAME","type":"str","level":"advanced","flags":0,"default_value":"admin","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"GRAFANA_FRONTEND_API_URL":{"name":"GRAFANA_FRONTEND_API_URL","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"GRAFANA_UPDATE_DASHBOARDS":{"name":"GRAFANA_UPDATE_DASHBOARDS","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"ISCSI_API_SSL_VERIFICATION":{"name":"ISCSI_API_SSL_VERIFICATION","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"ISSUE_TRACKER_API_KEY":{"name":"ISSUE_TRACKER_API_KEY","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PROMETHEUS_API_HOST":{"name":"PROMETHEUS_API_HOST","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PROMETHEUS_API_SSL_VERIFY":{"name":"PROMETHEUS_API_SSL_VERIFY","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PWD_POLICY_CHECK_COMPLEXITY_ENABLED":{"name":"PWD_POLICY_CHECK_COMPLEXITY_ENABLED","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PWD_POLICY_CHECK_EXCLUSION_LIST_ENABLED":{"name":"PWD_POLICY_CHECK_EXCLUSION_LIST_ENABLED","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PWD_POLICY_CHECK_LENGTH_ENABLED":{"name":"PWD_POLICY_CHECK_LENGTH_ENABLED","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PWD_POLICY_CHECK_OLDPWD_ENABLED":{"name":"PWD_POLICY_CHECK_OLDPWD_ENABLED","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PWD_POLICY_CHECK_REPETITIVE_CHARS_ENABLED":{"name":"PWD_POLICY_CHECK_REPETITIVE_CHARS_ENABLED","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PWD_POLICY_CHECK_SEQUENTIAL_CHARS_ENABLED":{"name":"PWD_POLICY_CHECK_SEQUENTIAL_CHARS_ENABLED","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PWD_POLICY_CHECK_USERNAME_ENABLED":{"name":"PWD_POLICY_CHECK_USERNAME_ENABLED","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PWD_POLICY_ENABLED":{"name":"PWD_POLICY_ENABLED","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PWD_POLICY_EXCLUSION_LIST":{"name":"PWD_POLICY_EXCLUSION_LIST","type":"str","level":"advanced","flags":0,"default_value":"osd,host,dashboard,pool,block,nfs,ceph,monitors,gateway,logs,crush,maps","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PWD_POLICY_MIN_COMPLEXITY":{"name":"PWD_POLICY_MIN_COMPLEXITY","type":"int","level":"advanced","flags":0,"default_value":"10","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PWD_POLICY_MIN_LENGTH":{"name":"PWD_POLICY_MIN_LENGTH","type":"int","level":"advanced","flags":0,"default_value":"8","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"REST_REQUESTS_TIMEOUT":{"name":"REST_REQUESTS_TIMEOUT","type":"int","level":"advanced","flags":0,"default_value":"45","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"RGW_API_ACCESS_KEY":{"name":"RGW_API_ACCESS_KEY","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"RGW_API_ADMIN_RESOURCE":{"name":"RGW_API_ADMIN_RESOURCE","type":"str","level":"advanced","flags":0,"default_value":"admin","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"RGW_API_SECRET_KEY":{"name":"RGW_API_SECRET_KEY","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"RGW_API_SSL_VERIFY":{"name":"RGW_API_SSL_VERIFY","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"USER_PWD_EXPIRATION_SPAN":{"name":"USER_PWD_EXPIRATION_SPAN","type":"int","level":"advanced","flags":0,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"USER_PWD_EXPIRATION_WARNING_1":{"name":"USER_PWD_EXPIRATION_WARNING_1","type":"int","level":"advanced","flags":0,"default_value":"10","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"USER_PWD_EXPIRATION_WARNING_2":{"name":"USER_PWD_EXPIRATION_WARNING_2","type":"int","level":"advanced","flags":0,"default_value":"5","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"cross_origin_url":{"name":"cross_origin_url","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"crt_file":{"name":"crt_file","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"debug":{"name":"debug","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Enable/disable debug options","long_desc":"","tags":[],"see_also":[]},"jwt_token_ttl":{"name":"jwt_token_ttl","type":"int","level":"advanced","flags":0,"default_value":"28800","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"key_file":{"name":"key_file","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"motd":{"name":"motd","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"The message of the day","long_desc":"","tags":[],"see_also":[]},"redirect_resolve_ip_addr":{"name":"redirect_resolve_ip_addr","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"server_addr":{"name":"server_addr","type":"str","level":"advanced","flags":0,"default_value":"::","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"server_port":{"name":"server_port","type":"int","level":"advanced","flags":0,"default_value":"8080","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"ssl":{"name":"ssl","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"ssl_server_port":{"name":"ssl_server_port","type":"int","level":"advanced","flags":0,"default_value":"8443","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"standby_behaviour":{"name":"standby_behaviour","type":"str","level":"advanced","flags":0,"default_value":"redirect","min":"","max":"","enum_allowed":["error","redirect"],"desc":"","long_desc":"","tags":[],"see_also":[]},"standby_error_status_code":{"name":"standby_error_status_code","type":"int","level":"advanced","flags":0,"default_value":"500","min":"400","max":"599","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"url_prefix":{"name":"url_prefix","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"devicehealth","can_run":true,"error_string":"","module_options":{"enable_monitoring":{"name":"enable_monitoring","type":"bool","level":"advanced","flags":1,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"monitor device health metrics","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"mark_out_threshold":{"name":"mark_out_threshold","type":"secs","level":"advanced","flags":1,"default_value":"2419200","min":"","max":"","enum_allowed":[],"desc":"automatically mark OSD if it may fail before this long","long_desc":"","tags":[],"see_also":[]},"pool_name":{"name":"pool_name","type":"str","level":"advanced","flags":1,"default_value":"device_health_metrics","min":"","max":"","enum_allowed":[],"desc":"name of pool in which to store device health metrics","long_desc":"","tags":[],"see_also":[]},"retention_period":{"name":"retention_period","type":"secs","level":"advanced","flags":1,"default_value":"15552000","min":"","max":"","enum_allowed":[],"desc":"how long to retain device health metrics","long_desc":"","tags":[],"see_also":[]},"scrape_frequency":{"name":"scrape_frequency","type":"secs","level":"advanced","flags":1,"default_value":"86400","min":"","max":"","enum_allowed":[],"desc":"how frequently to scrape device health metrics","long_desc":"","tags":[],"see_also":[]},"self_heal":{"name":"self_heal","type":"bool","level":"advanced","flags":1,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"preemptively heal cluster around devices that may fail","long_desc":"","tags":[],"see_also":[]},"sleep_interval":{"name":"sleep_interval","type":"secs","level":"advanced","flags":1,"default_value":"600","min":"","max":"","enum_allowed":[],"desc":"how frequently to wake up and check device health","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"warn_threshold":{"name":"warn_threshold","type":"secs","level":"advanced","flags":1,"default_value":"7257600","min":"","max":"","enum_allowed":[],"desc":"raise health warning if OSD may fail before this long","long_desc":"","tags":[],"see_also":[]}}},{"name":"diskprediction_local","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"predict_interval":{"name":"predict_interval","type":"str","level":"advanced","flags":0,"default_value":"86400","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"predictor_model":{"name":"predictor_model","type":"str","level":"advanced","flags":0,"default_value":"prophetstor","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sleep_interval":{"name":"sleep_interval","type":"str","level":"advanced","flags":0,"default_value":"600","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"influx","can_run":false,"error_string":"influxdb python module not found","module_options":{"batch_size":{"name":"batch_size","type":"int","level":"advanced","flags":0,"default_value":"5000","min":"","max":"","enum_allowed":[],"desc":"How big batches of data points should be when sending to InfluxDB.","long_desc":"","tags":[],"see_also":[]},"database":{"name":"database","type":"str","level":"advanced","flags":0,"default_value":"ceph","min":"","max":"","enum_allowed":[],"desc":"InfluxDB database name. You will need to create this database and grant write privileges to the configured username or the username must have admin privileges to create it.","long_desc":"","tags":[],"see_also":[]},"hostname":{"name":"hostname","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"InfluxDB server hostname","long_desc":"","tags":[],"see_also":[]},"interval":{"name":"interval","type":"secs","level":"advanced","flags":0,"default_value":"30","min":"5","max":"","enum_allowed":[],"desc":"Time between reports to InfluxDB. Default 30 seconds.","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"password":{"name":"password","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"password of InfluxDB server user","long_desc":"","tags":[],"see_also":[]},"port":{"name":"port","type":"int","level":"advanced","flags":0,"default_value":"8086","min":"","max":"","enum_allowed":[],"desc":"InfluxDB server port","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"ssl":{"name":"ssl","type":"str","level":"advanced","flags":0,"default_value":"false","min":"","max":"","enum_allowed":[],"desc":"Use https connection for InfluxDB server. Use \"true\" or \"false\".","long_desc":"","tags":[],"see_also":[]},"threads":{"name":"threads","type":"int","level":"advanced","flags":0,"default_value":"5","min":"1","max":"32","enum_allowed":[],"desc":"How many worker threads should be spawned for sending data to InfluxDB.","long_desc":"","tags":[],"see_also":[]},"username":{"name":"username","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"username of InfluxDB server user","long_desc":"","tags":[],"see_also":[]},"verify_ssl":{"name":"verify_ssl","type":"str","level":"advanced","flags":0,"default_value":"true","min":"","max":"","enum_allowed":[],"desc":"Verify https cert for InfluxDB server. Use \"true\" or \"false\".","long_desc":"","tags":[],"see_also":[]}}},{"name":"insights","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"iostat","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"k8sevents","can_run":true,"error_string":"","module_options":{"ceph_event_retention_days":{"name":"ceph_event_retention_days","type":"int","level":"advanced","flags":0,"default_value":"7","min":"","max":"","enum_allowed":[],"desc":"Days to hold ceph event information within local cache","long_desc":"","tags":[],"see_also":[]},"config_check_secs":{"name":"config_check_secs","type":"int","level":"advanced","flags":0,"default_value":"10","min":"10","max":"","enum_allowed":[],"desc":"interval (secs) to check for cluster configuration changes","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"localpool","can_run":true,"error_string":"","module_options":{"failure_domain":{"name":"failure_domain","type":"str","level":"advanced","flags":1,"default_value":"host","min":"","max":"","enum_allowed":[],"desc":"failure domain for any created local pool","long_desc":"what failure domain we should separate data replicas across.","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"min_size":{"name":"min_size","type":"int","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"default min_size for any created local pool","long_desc":"value to set min_size to (unchanged from Ceph's default if this option is not set)","tags":[],"see_also":[]},"num_rep":{"name":"num_rep","type":"int","level":"advanced","flags":1,"default_value":"3","min":"","max":"","enum_allowed":[],"desc":"default replica count for any created local pool","long_desc":"","tags":[],"see_also":[]},"pg_num":{"name":"pg_num","type":"int","level":"advanced","flags":1,"default_value":"128","min":"","max":"","enum_allowed":[],"desc":"default pg_num for any created local pool","long_desc":"","tags":[],"see_also":[]},"prefix":{"name":"prefix","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"name prefix for any created local pool","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"subtree":{"name":"subtree","type":"str","level":"advanced","flags":1,"default_value":"rack","min":"","max":"","enum_allowed":[],"desc":"CRUSH level for which to create a local pool","long_desc":"which CRUSH subtree type the module should create a pool for.","tags":[],"see_also":[]}}},{"name":"mds_autoscaler","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"mirroring","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"nfs","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"orchestrator","can_run":true,"error_string":"","module_options":{"fail_fs":{"name":"fail_fs","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Fail filesystem for rapid multi-rank mds upgrade","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"orchestrator":{"name":"orchestrator","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["cephadm","rook","test_orchestrator"],"desc":"Orchestrator backend","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"osd_perf_query","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"osd_support","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"pg_autoscaler","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sleep_interval":{"name":"sleep_interval","type":"secs","level":"advanced","flags":0,"default_value":"60","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"threshold":{"name":"threshold","type":"float","level":"advanced","flags":0,"default_value":"3.0","min":"1.0","max":"","enum_allowed":[],"desc":"scaling threshold","long_desc":"The factor by which the `NEW PG_NUM` must vary from the current`PG_NUM` before being accepted. Cannot be less than 1.0","tags":[],"see_also":[]}}},{"name":"progress","can_run":true,"error_string":"","module_options":{"allow_pg_recovery_event":{"name":"allow_pg_recovery_event","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"allow the module to show pg recovery progress","long_desc":"","tags":[],"see_also":[]},"enabled":{"name":"enabled","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"max_completed_events":{"name":"max_completed_events","type":"int","level":"advanced","flags":1,"default_value":"50","min":"","max":"","enum_allowed":[],"desc":"number of past completed events to remember","long_desc":"","tags":[],"see_also":[]},"sleep_interval":{"name":"sleep_interval","type":"secs","level":"advanced","flags":1,"default_value":"5","min":"","max":"","enum_allowed":[],"desc":"how long the module is going to sleep","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"prometheus","can_run":true,"error_string":"","module_options":{"cache":{"name":"cache","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"exclude_perf_counters":{"name":"exclude_perf_counters","type":"bool","level":"advanced","flags":1,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"Do not include perf-counters in the metrics output","long_desc":"Gathering perf-counters from a single Prometheus exporter can degrade ceph-mgr performance, especially in large clusters. Instead, Ceph-exporter daemons are now used by default for perf-counter gathering. This should only be disabled when no ceph-exporters are deployed.","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"rbd_stats_pools":{"name":"rbd_stats_pools","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"rbd_stats_pools_refresh_interval":{"name":"rbd_stats_pools_refresh_interval","type":"int","level":"advanced","flags":0,"default_value":"300","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"scrape_interval":{"name":"scrape_interval","type":"float","level":"advanced","flags":0,"default_value":"15.0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"server_addr":{"name":"server_addr","type":"str","level":"advanced","flags":0,"default_value":"::","min":"","max":"","enum_allowed":[],"desc":"the IPv4 or IPv6 address on which the module listens for HTTP requests","long_desc":"","tags":[],"see_also":[]},"server_port":{"name":"server_port","type":"int","level":"advanced","flags":1,"default_value":"9283","min":"","max":"","enum_allowed":[],"desc":"the port on which the module listens for HTTP requests","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"stale_cache_strategy":{"name":"stale_cache_strategy","type":"str","level":"advanced","flags":0,"default_value":"log","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"standby_behaviour":{"name":"standby_behaviour","type":"str","level":"advanced","flags":1,"default_value":"default","min":"","max":"","enum_allowed":["default","error"],"desc":"","long_desc":"","tags":[],"see_also":[]},"standby_error_status_code":{"name":"standby_error_status_code","type":"int","level":"advanced","flags":1,"default_value":"500","min":"400","max":"599","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"rbd_support","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"max_concurrent_snap_create":{"name":"max_concurrent_snap_create","type":"int","level":"advanced","flags":0,"default_value":"10","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"mirror_snapshot_schedule":{"name":"mirror_snapshot_schedule","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"trash_purge_schedule":{"name":"trash_purge_schedule","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"restful","can_run":true,"error_string":"","module_options":{"enable_auth":{"name":"enable_auth","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"key_file":{"name":"key_file","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"max_requests":{"name":"max_requests","type":"int","level":"advanced","flags":0,"default_value":"500","min":"","max":"","enum_allowed":[],"desc":"Maximum number of requests to keep in memory. When new request comes in, the oldest request will be removed if the number of requests exceeds the max request number. if un-finished request is removed, error message will be logged in the ceph-mgr log.","long_desc":"","tags":[],"see_also":[]},"server_addr":{"name":"server_addr","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"server_port":{"name":"server_port","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"rgw","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"secondary_zone_period_retry_limit":{"name":"secondary_zone_period_retry_limit","type":"int","level":"advanced","flags":0,"default_value":"5","min":"","max":"","enum_allowed":[],"desc":"RGW module period update retry limit for secondary site","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"rook","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"storage_class":{"name":"storage_class","type":"str","level":"advanced","flags":0,"default_value":"local","min":"","max":"","enum_allowed":[],"desc":"storage class name for LSO-discovered PVs","long_desc":"","tags":[],"see_also":[]}}},{"name":"selftest","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"roption1":{"name":"roption1","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"roption2":{"name":"roption2","type":"str","level":"advanced","flags":0,"default_value":"xyz","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"rwoption1":{"name":"rwoption1","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"rwoption2":{"name":"rwoption2","type":"int","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"rwoption3":{"name":"rwoption3","type":"float","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"rwoption4":{"name":"rwoption4","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"rwoption5":{"name":"rwoption5","type":"bool","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"rwoption6":{"name":"rwoption6","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"rwoption7":{"name":"rwoption7","type":"int","level":"advanced","flags":0,"default_value":"","min":"1","max":"42","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"testkey":{"name":"testkey","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"testlkey":{"name":"testlkey","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"testnewline":{"name":"testnewline","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"snap_schedule","can_run":true,"error_string":"","module_options":{"allow_m_granularity":{"name":"allow_m_granularity","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"allow minute scheduled snapshots","long_desc":"","tags":[],"see_also":[]},"dump_on_update":{"name":"dump_on_update","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"dump database to debug log on update","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"stats","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"status","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"telegraf","can_run":true,"error_string":"","module_options":{"address":{"name":"address","type":"str","level":"advanced","flags":0,"default_value":"unixgram:///tmp/telegraf.sock","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"interval":{"name":"interval","type":"secs","level":"advanced","flags":0,"default_value":"15","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"telemetry","can_run":true,"error_string":"","module_options":{"channel_basic":{"name":"channel_basic","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"Share basic cluster information (size, version)","long_desc":"","tags":[],"see_also":[]},"channel_crash":{"name":"channel_crash","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"Share metadata about Ceph daemon crashes (version, stack straces, etc)","long_desc":"","tags":[],"see_also":[]},"channel_device":{"name":"channel_device","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"Share device health metrics (e.g., SMART data, minus potentially identifying info like serial numbers)","long_desc":"","tags":[],"see_also":[]},"channel_ident":{"name":"channel_ident","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Share a user-provided description and/or contact email for the cluster","long_desc":"","tags":[],"see_also":[]},"channel_perf":{"name":"channel_perf","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Share various performance metrics of a cluster","long_desc":"","tags":[],"see_also":[]},"contact":{"name":"contact","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"description":{"name":"description","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"device_url":{"name":"device_url","type":"str","level":"advanced","flags":0,"default_value":"https://telemetry.ceph.com/device","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"enabled":{"name":"enabled","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"interval":{"name":"interval","type":"int","level":"advanced","flags":0,"default_value":"24","min":"8","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"last_opt_revision":{"name":"last_opt_revision","type":"int","level":"advanced","flags":0,"default_value":"1","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"leaderboard":{"name":"leaderboard","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"leaderboard_description":{"name":"leaderboard_description","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"organization":{"name":"organization","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"proxy":{"name":"proxy","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"url":{"name":"url","type":"str","level":"advanced","flags":0,"default_value":"https://telemetry.ceph.com/report","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"test_orchestrator","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"volumes","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"max_concurrent_clones":{"name":"max_concurrent_clones","type":"int","level":"advanced","flags":0,"default_value":"4","min":"","max":"","enum_allowed":[],"desc":"Number of asynchronous cloner threads","long_desc":"","tags":[],"see_also":[]},"periodic_async_work":{"name":"periodic_async_work","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Periodically check for async work","long_desc":"","tags":[],"see_also":[]},"snapshot_clone_delay":{"name":"snapshot_clone_delay","type":"int","level":"advanced","flags":0,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"Delay clone begin operation by snapshot_clone_delay seconds","long_desc":"","tags":[],"see_also":[]},"snapshot_clone_no_wait":{"name":"snapshot_clone_no_wait","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"Reject subvolume clone request when cloner threads are busy","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"zabbix","can_run":true,"error_string":"","module_options":{"discovery_interval":{"name":"discovery_interval","type":"uint","level":"advanced","flags":0,"default_value":"100","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"identifier":{"name":"identifier","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"interval":{"name":"interval","type":"secs","level":"advanced","flags":0,"default_value":"60","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"zabbix_host":{"name":"zabbix_host","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"zabbix_port":{"name":"zabbix_port","type":"int","level":"advanced","flags":0,"default_value":"10051","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"zabbix_sender":{"name":"zabbix_sender","type":"str","level":"advanced","flags":0,"default_value":"/usr/bin/zabbix_sender","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}}],"services":{"dashboard":"https://192.168.123.101:8443/"},"always_on_modules":{"octopus":["balancer","crash","devicehealth","orchestrator","pg_autoscaler","progress","rbd_support","status","telemetry","volumes"],"pacific":["balancer","crash","devicehealth","orchestrator","pg_autoscaler","progress","rbd_support","status","telemetry","volumes"],"quincy":["balancer","crash","devicehealth","orchestrator","pg_autoscaler","progress","rbd_support","status","telemetry","volumes"],"reef":["balancer","crash","devicehealth","orchestrator","pg_autoscaler","progress","rbd_support","status","telemetry","volumes"],"squid":["balancer","crash","devicehealth","orchestrator","pg_autoscaler","progress","rbd_support","status","telemetry","volumes"]},"force_disabled_modules":{},"last_failure_osd_epoch":3,"active_clients":[{"name":"libcephsqlite","addrvec":[{"type":"v2","addr":"192.168.123.101:0","nonce":385116697}]},{"name":"rbd_support","addrvec":[{"type":"v2","addr":"192.168.123.101:0","nonce":1183377538}]},{"name":"volumes","addrvec":[{"type":"v2","addr":"192.168.123.101:0","nonce":2806126584}]}]} 2026-03-09T20:53:18.014 INFO:tasks.cephadm.ceph_manager.ceph:mgr available! 2026-03-09T20:53:18.014 INFO:tasks.cephadm.ceph_manager.ceph:waiting for all up 2026-03-09T20:53:18.014 DEBUG:teuthology.orchestra.run.vm01:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell --fsid 9cb345a8-1bf9-11f1-a195-0375563c5891 -- ceph osd dump --format=json 2026-03-09T20:53:18.045 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 09 20:53:17 vm06 ceph-mon[52939]: osdmap e18: 3 total, 3 up, 3 in 2026-03-09T20:53:18.046 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 09 20:53:17 vm06 ceph-mon[52939]: from='client.? 192.168.123.101:0/3140589146' entity='client.admin' cmd=[{"prefix": "osd dump", "format": "json"}]: dispatch 2026-03-09T20:53:18.046 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 09 20:53:17 vm06 ceph-mon[52939]: pgmap v43: 0 pgs: ; 0 B data, 79 MiB used, 60 GiB / 60 GiB avail 2026-03-09T20:53:18.046 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 09 20:53:17 vm06 ceph-mon[52939]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' cmd=[{"prefix": "osd pool create", "format": "json", "pool": ".mgr", "pg_num": 1, "pg_num_min": 1, "pg_num_max": 32, "yes_i_really_mean_it": true}]: dispatch 2026-03-09T20:53:18.207 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:53:17 vm08 ceph-mon[54611]: osdmap e18: 3 total, 3 up, 3 in 2026-03-09T20:53:18.207 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:53:17 vm08 ceph-mon[54611]: from='client.? 192.168.123.101:0/3140589146' entity='client.admin' cmd=[{"prefix": "osd dump", "format": "json"}]: dispatch 2026-03-09T20:53:18.207 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:53:17 vm08 ceph-mon[54611]: pgmap v43: 0 pgs: ; 0 B data, 79 MiB used, 60 GiB / 60 GiB avail 2026-03-09T20:53:18.207 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:53:17 vm08 ceph-mon[54611]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' cmd=[{"prefix": "osd pool create", "format": "json", "pool": ".mgr", "pg_num": 1, "pg_num_min": 1, "pg_num_max": 32, "yes_i_really_mean_it": true}]: dispatch 2026-03-09T20:53:18.210 INFO:teuthology.orchestra.run.vm01.stderr:Inferring config /var/lib/ceph/9cb345a8-1bf9-11f1-a195-0375563c5891/mon.a/config 2026-03-09T20:53:18.452 INFO:teuthology.orchestra.run.vm01.stdout: 2026-03-09T20:53:18.452 INFO:teuthology.orchestra.run.vm01.stdout:{"epoch":19,"fsid":"9cb345a8-1bf9-11f1-a195-0375563c5891","created":"2026-03-09T20:51:43.189739+0000","modified":"2026-03-09T20:53:17.782831+0000","last_up_change":"2026-03-09T20:53:15.768033+0000","last_in_change":"2026-03-09T20:53:03.945618+0000","flags":"sortbitwise,recovery_deletes,purged_snapdirs,pglog_hardlimit","flags_num":5799936,"flags_set":["pglog_hardlimit","purged_snapdirs","recovery_deletes","sortbitwise"],"crush_version":8,"full_ratio":0.94999998807907104,"backfillfull_ratio":0.89999997615814209,"nearfull_ratio":0.85000002384185791,"cluster_snapshot":"","pool_max":1,"max_osd":3,"require_min_compat_client":"luminous","min_compat_client":"jewel","require_osd_release":"squid","allow_crimson":false,"pools":[{"pool":1,"pool_name":".mgr","create_time":"2026-03-09T20:53:17.443588+0000","flags":32769,"flags_names":"hashpspool,creating","type":1,"size":3,"min_size":2,"crush_rule":0,"peering_crush_bucket_count":0,"peering_crush_bucket_target":0,"peering_crush_bucket_barrier":0,"peering_crush_bucket_mandatory_member":2147483647,"is_stretch_pool":false,"object_hash":2,"pg_autoscale_mode":"off","pg_num":1,"pg_placement_num":1,"pg_placement_num_target":1,"pg_num_target":1,"pg_num_pending":1,"last_pg_merge_meta":{"source_pgid":"0.0","ready_epoch":0,"last_epoch_started":0,"last_epoch_clean":0,"source_version":"0'0","target_version":"0'0"},"last_change":"19","last_force_op_resend":"0","last_force_op_resend_prenautilus":"0","last_force_op_resend_preluminous":"0","auid":0,"snap_mode":"selfmanaged","snap_seq":0,"snap_epoch":0,"pool_snaps":[],"removed_snaps":"[]","quota_max_bytes":0,"quota_max_objects":0,"tiers":[],"tier_of":-1,"read_tier":-1,"write_tier":-1,"cache_mode":"none","target_max_bytes":0,"target_max_objects":0,"cache_target_dirty_ratio_micro":400000,"cache_target_dirty_high_ratio_micro":600000,"cache_target_full_ratio_micro":800000,"cache_min_flush_age":0,"cache_min_evict_age":0,"erasure_code_profile":"","hit_set_params":{"type":"none"},"hit_set_period":0,"hit_set_count":0,"use_gmt_hitset":true,"min_read_recency_for_promote":0,"min_write_recency_for_promote":0,"hit_set_grade_decay_rate":0,"hit_set_search_last_n":0,"grade_table":[],"stripe_width":0,"expected_num_objects":0,"fast_read":false,"options":{"pg_num_max":32,"pg_num_min":1},"application_metadata":{},"read_balance":{"score_type":"Fair distribution","score_acting":3,"score_stable":3,"optimal_score":1,"raw_score_acting":3,"raw_score_stable":3,"primary_affinity_weighted":1,"average_primary_affinity":1,"average_primary_affinity_weighted":1}}],"osds":[{"osd":0,"uuid":"58efc62d-2ef3-4005-88e4-4cf11f5576fe","up":1,"in":1,"weight":1,"primary_affinity":1,"last_clean_begin":0,"last_clean_end":0,"up_from":8,"up_thru":0,"down_at":0,"lost_at":0,"public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.101:6802","nonce":3447435909},{"type":"v1","addr":"192.168.123.101:6803","nonce":3447435909}]},"cluster_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.101:6804","nonce":3447435909},{"type":"v1","addr":"192.168.123.101:6805","nonce":3447435909}]},"heartbeat_back_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.101:6808","nonce":3447435909},{"type":"v1","addr":"192.168.123.101:6809","nonce":3447435909}]},"heartbeat_front_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.101:6806","nonce":3447435909},{"type":"v1","addr":"192.168.123.101:6807","nonce":3447435909}]},"public_addr":"192.168.123.101:6803/3447435909","cluster_addr":"192.168.123.101:6805/3447435909","heartbeat_back_addr":"192.168.123.101:6809/3447435909","heartbeat_front_addr":"192.168.123.101:6807/3447435909","state":["exists","up"]},{"osd":1,"uuid":"38f01073-940b-498e-91d3-7a23e98bcb53","up":1,"in":1,"weight":1,"primary_affinity":1,"last_clean_begin":0,"last_clean_end":0,"up_from":12,"up_thru":0,"down_at":0,"lost_at":0,"public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.106:6800","nonce":3453001158},{"type":"v1","addr":"192.168.123.106:6801","nonce":3453001158}]},"cluster_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.106:6802","nonce":3453001158},{"type":"v1","addr":"192.168.123.106:6803","nonce":3453001158}]},"heartbeat_back_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.106:6806","nonce":3453001158},{"type":"v1","addr":"192.168.123.106:6807","nonce":3453001158}]},"heartbeat_front_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.106:6804","nonce":3453001158},{"type":"v1","addr":"192.168.123.106:6805","nonce":3453001158}]},"public_addr":"192.168.123.106:6801/3453001158","cluster_addr":"192.168.123.106:6803/3453001158","heartbeat_back_addr":"192.168.123.106:6807/3453001158","heartbeat_front_addr":"192.168.123.106:6805/3453001158","state":["exists","up"]},{"osd":2,"uuid":"13684d61-15e6-4d3c-8355-a5e681450c79","up":1,"in":1,"weight":1,"primary_affinity":1,"last_clean_begin":0,"last_clean_end":0,"up_from":17,"up_thru":0,"down_at":0,"lost_at":0,"public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.108:6800","nonce":4083304280},{"type":"v1","addr":"192.168.123.108:6801","nonce":4083304280}]},"cluster_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.108:6802","nonce":4083304280},{"type":"v1","addr":"192.168.123.108:6803","nonce":4083304280}]},"heartbeat_back_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.108:6806","nonce":4083304280},{"type":"v1","addr":"192.168.123.108:6807","nonce":4083304280}]},"heartbeat_front_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.108:6804","nonce":4083304280},{"type":"v1","addr":"192.168.123.108:6805","nonce":4083304280}]},"public_addr":"192.168.123.108:6801/4083304280","cluster_addr":"192.168.123.108:6803/4083304280","heartbeat_back_addr":"192.168.123.108:6807/4083304280","heartbeat_front_addr":"192.168.123.108:6805/4083304280","state":["exists","up"]}],"osd_xinfo":[{"osd":0,"down_stamp":"0.000000","laggy_probability":0,"laggy_interval":0,"features":4540701547738038271,"old_weight":0,"last_purged_snaps_scrub":"2026-03-09T20:52:50.809851+0000","dead_epoch":0},{"osd":1,"down_stamp":"0.000000","laggy_probability":0,"laggy_interval":0,"features":4540701547738038271,"old_weight":0,"last_purged_snaps_scrub":"2026-03-09T20:53:02.121506+0000","dead_epoch":0},{"osd":2,"down_stamp":"0.000000","laggy_probability":0,"laggy_interval":0,"features":4540701547738038271,"old_weight":0,"last_purged_snaps_scrub":"2026-03-09T20:53:13.662870+0000","dead_epoch":0}],"pg_upmap":[],"pg_upmap_items":[],"pg_upmap_primaries":[],"pg_temp":[],"primary_temp":[],"blocklist":{"192.168.123.101:0/1300460699":"2026-03-10T20:52:07.386219+0000","192.168.123.101:6801/119456608":"2026-03-10T20:52:07.386219+0000","192.168.123.101:0/2446502297":"2026-03-10T20:51:56.804238+0000","192.168.123.101:0/1626717301":"2026-03-10T20:52:07.386219+0000","192.168.123.101:0/4056848261":"2026-03-10T20:51:56.804238+0000","192.168.123.101:6800/119456608":"2026-03-10T20:52:07.386219+0000","192.168.123.101:6801/668143714":"2026-03-10T20:51:56.804238+0000","192.168.123.101:6800/668143714":"2026-03-10T20:51:56.804238+0000","192.168.123.101:0/1165500474":"2026-03-10T20:52:07.386219+0000","192.168.123.101:0/4063840047":"2026-03-10T20:51:56.804238+0000"},"range_blocklist":{},"erasure_code_profiles":{"default":{"crush-failure-domain":"osd","k":"2","m":"1","plugin":"jerasure","technique":"reed_sol_van"}},"removed_snaps_queue":[],"new_removed_snaps":[],"new_purged_snaps":[],"crush_node_flags":{},"device_class_flags":{},"stretch_mode":{"stretch_mode_enabled":false,"stretch_bucket_count":0,"degraded_stretch_mode":0,"recovering_stretch_mode":0,"stretch_mode_bucket":0}} 2026-03-09T20:53:18.635 INFO:tasks.cephadm.ceph_manager.ceph:all up! 2026-03-09T20:53:18.635 DEBUG:teuthology.orchestra.run.vm01:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell --fsid 9cb345a8-1bf9-11f1-a195-0375563c5891 -- ceph osd dump --format=json 2026-03-09T20:53:18.810 INFO:teuthology.orchestra.run.vm01.stderr:Inferring config /var/lib/ceph/9cb345a8-1bf9-11f1-a195-0375563c5891/mon.a/config 2026-03-09T20:53:19.106 INFO:teuthology.orchestra.run.vm01.stdout: 2026-03-09T20:53:19.106 INFO:teuthology.orchestra.run.vm01.stdout:{"epoch":20,"fsid":"9cb345a8-1bf9-11f1-a195-0375563c5891","created":"2026-03-09T20:51:43.189739+0000","modified":"2026-03-09T20:53:18.788591+0000","last_up_change":"2026-03-09T20:53:15.768033+0000","last_in_change":"2026-03-09T20:53:03.945618+0000","flags":"sortbitwise,recovery_deletes,purged_snapdirs,pglog_hardlimit","flags_num":5799936,"flags_set":["pglog_hardlimit","purged_snapdirs","recovery_deletes","sortbitwise"],"crush_version":8,"full_ratio":0.94999998807907104,"backfillfull_ratio":0.89999997615814209,"nearfull_ratio":0.85000002384185791,"cluster_snapshot":"","pool_max":1,"max_osd":3,"require_min_compat_client":"luminous","min_compat_client":"jewel","require_osd_release":"squid","allow_crimson":false,"pools":[{"pool":1,"pool_name":".mgr","create_time":"2026-03-09T20:53:17.443588+0000","flags":32769,"flags_names":"hashpspool,creating","type":1,"size":3,"min_size":2,"crush_rule":0,"peering_crush_bucket_count":0,"peering_crush_bucket_target":0,"peering_crush_bucket_barrier":0,"peering_crush_bucket_mandatory_member":2147483647,"is_stretch_pool":false,"object_hash":2,"pg_autoscale_mode":"off","pg_num":1,"pg_placement_num":1,"pg_placement_num_target":1,"pg_num_target":1,"pg_num_pending":1,"last_pg_merge_meta":{"source_pgid":"0.0","ready_epoch":0,"last_epoch_started":0,"last_epoch_clean":0,"source_version":"0'0","target_version":"0'0"},"last_change":"20","last_force_op_resend":"0","last_force_op_resend_prenautilus":"0","last_force_op_resend_preluminous":"0","auid":0,"snap_mode":"selfmanaged","snap_seq":0,"snap_epoch":0,"pool_snaps":[],"removed_snaps":"[]","quota_max_bytes":0,"quota_max_objects":0,"tiers":[],"tier_of":-1,"read_tier":-1,"write_tier":-1,"cache_mode":"none","target_max_bytes":0,"target_max_objects":0,"cache_target_dirty_ratio_micro":400000,"cache_target_dirty_high_ratio_micro":600000,"cache_target_full_ratio_micro":800000,"cache_min_flush_age":0,"cache_min_evict_age":0,"erasure_code_profile":"","hit_set_params":{"type":"none"},"hit_set_period":0,"hit_set_count":0,"use_gmt_hitset":true,"min_read_recency_for_promote":0,"min_write_recency_for_promote":0,"hit_set_grade_decay_rate":0,"hit_set_search_last_n":0,"grade_table":[],"stripe_width":0,"expected_num_objects":0,"fast_read":false,"options":{"pg_num_max":32,"pg_num_min":1},"application_metadata":{"mgr":{}},"read_balance":{"score_type":"Fair distribution","score_acting":3,"score_stable":3,"optimal_score":1,"raw_score_acting":3,"raw_score_stable":3,"primary_affinity_weighted":1,"average_primary_affinity":1,"average_primary_affinity_weighted":1}}],"osds":[{"osd":0,"uuid":"58efc62d-2ef3-4005-88e4-4cf11f5576fe","up":1,"in":1,"weight":1,"primary_affinity":1,"last_clean_begin":0,"last_clean_end":0,"up_from":8,"up_thru":0,"down_at":0,"lost_at":0,"public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.101:6802","nonce":3447435909},{"type":"v1","addr":"192.168.123.101:6803","nonce":3447435909}]},"cluster_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.101:6804","nonce":3447435909},{"type":"v1","addr":"192.168.123.101:6805","nonce":3447435909}]},"heartbeat_back_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.101:6808","nonce":3447435909},{"type":"v1","addr":"192.168.123.101:6809","nonce":3447435909}]},"heartbeat_front_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.101:6806","nonce":3447435909},{"type":"v1","addr":"192.168.123.101:6807","nonce":3447435909}]},"public_addr":"192.168.123.101:6803/3447435909","cluster_addr":"192.168.123.101:6805/3447435909","heartbeat_back_addr":"192.168.123.101:6809/3447435909","heartbeat_front_addr":"192.168.123.101:6807/3447435909","state":["exists","up"]},{"osd":1,"uuid":"38f01073-940b-498e-91d3-7a23e98bcb53","up":1,"in":1,"weight":1,"primary_affinity":1,"last_clean_begin":0,"last_clean_end":0,"up_from":12,"up_thru":19,"down_at":0,"lost_at":0,"public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.106:6800","nonce":3453001158},{"type":"v1","addr":"192.168.123.106:6801","nonce":3453001158}]},"cluster_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.106:6802","nonce":3453001158},{"type":"v1","addr":"192.168.123.106:6803","nonce":3453001158}]},"heartbeat_back_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.106:6806","nonce":3453001158},{"type":"v1","addr":"192.168.123.106:6807","nonce":3453001158}]},"heartbeat_front_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.106:6804","nonce":3453001158},{"type":"v1","addr":"192.168.123.106:6805","nonce":3453001158}]},"public_addr":"192.168.123.106:6801/3453001158","cluster_addr":"192.168.123.106:6803/3453001158","heartbeat_back_addr":"192.168.123.106:6807/3453001158","heartbeat_front_addr":"192.168.123.106:6805/3453001158","state":["exists","up"]},{"osd":2,"uuid":"13684d61-15e6-4d3c-8355-a5e681450c79","up":1,"in":1,"weight":1,"primary_affinity":1,"last_clean_begin":0,"last_clean_end":0,"up_from":17,"up_thru":0,"down_at":0,"lost_at":0,"public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.108:6800","nonce":4083304280},{"type":"v1","addr":"192.168.123.108:6801","nonce":4083304280}]},"cluster_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.108:6802","nonce":4083304280},{"type":"v1","addr":"192.168.123.108:6803","nonce":4083304280}]},"heartbeat_back_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.108:6806","nonce":4083304280},{"type":"v1","addr":"192.168.123.108:6807","nonce":4083304280}]},"heartbeat_front_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.108:6804","nonce":4083304280},{"type":"v1","addr":"192.168.123.108:6805","nonce":4083304280}]},"public_addr":"192.168.123.108:6801/4083304280","cluster_addr":"192.168.123.108:6803/4083304280","heartbeat_back_addr":"192.168.123.108:6807/4083304280","heartbeat_front_addr":"192.168.123.108:6805/4083304280","state":["exists","up"]}],"osd_xinfo":[{"osd":0,"down_stamp":"0.000000","laggy_probability":0,"laggy_interval":0,"features":4540701547738038271,"old_weight":0,"last_purged_snaps_scrub":"2026-03-09T20:52:50.809851+0000","dead_epoch":0},{"osd":1,"down_stamp":"0.000000","laggy_probability":0,"laggy_interval":0,"features":4540701547738038271,"old_weight":0,"last_purged_snaps_scrub":"2026-03-09T20:53:02.121506+0000","dead_epoch":0},{"osd":2,"down_stamp":"0.000000","laggy_probability":0,"laggy_interval":0,"features":4540701547738038271,"old_weight":0,"last_purged_snaps_scrub":"2026-03-09T20:53:13.662870+0000","dead_epoch":0}],"pg_upmap":[],"pg_upmap_items":[],"pg_upmap_primaries":[],"pg_temp":[],"primary_temp":[],"blocklist":{"192.168.123.101:0/1300460699":"2026-03-10T20:52:07.386219+0000","192.168.123.101:6801/119456608":"2026-03-10T20:52:07.386219+0000","192.168.123.101:0/2446502297":"2026-03-10T20:51:56.804238+0000","192.168.123.101:0/1626717301":"2026-03-10T20:52:07.386219+0000","192.168.123.101:0/4056848261":"2026-03-10T20:51:56.804238+0000","192.168.123.101:6800/119456608":"2026-03-10T20:52:07.386219+0000","192.168.123.101:6801/668143714":"2026-03-10T20:51:56.804238+0000","192.168.123.101:6800/668143714":"2026-03-10T20:51:56.804238+0000","192.168.123.101:0/1165500474":"2026-03-10T20:52:07.386219+0000","192.168.123.101:0/4063840047":"2026-03-10T20:51:56.804238+0000"},"range_blocklist":{},"erasure_code_profiles":{"default":{"crush-failure-domain":"osd","k":"2","m":"1","plugin":"jerasure","technique":"reed_sol_van"}},"removed_snaps_queue":[],"new_removed_snaps":[],"new_purged_snaps":[],"crush_node_flags":{},"device_class_flags":{},"stretch_mode":{"stretch_mode_enabled":false,"stretch_bucket_count":0,"degraded_stretch_mode":0,"recovering_stretch_mode":0,"stretch_mode_bucket":0}} 2026-03-09T20:53:19.117 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:53:18 vm01 ceph-mon[53402]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' cmd='[{"prefix": "osd pool create", "format": "json", "pool": ".mgr", "pg_num": 1, "pg_num_min": 1, "pg_num_max": 32, "yes_i_really_mean_it": true}]': finished 2026-03-09T20:53:19.117 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:53:18 vm01 ceph-mon[53402]: osdmap e19: 3 total, 3 up, 3 in 2026-03-09T20:53:19.117 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:53:18 vm01 ceph-mon[53402]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' cmd=[{"prefix": "osd pool application enable", "format": "json", "pool": ".mgr", "app": "mgr", "yes_i_really_mean_it": true}]: dispatch 2026-03-09T20:53:19.117 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:53:18 vm01 ceph-mon[53402]: from='client.? 192.168.123.101:0/1181953921' entity='client.admin' cmd=[{"prefix": "mgr dump", "format": "json"}]: dispatch 2026-03-09T20:53:19.117 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:53:18 vm01 ceph-mon[53402]: from='client.? 192.168.123.101:0/3469612210' entity='client.admin' cmd=[{"prefix": "osd dump", "format": "json"}]: dispatch 2026-03-09T20:53:19.207 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:53:18 vm08 ceph-mon[54611]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' cmd='[{"prefix": "osd pool create", "format": "json", "pool": ".mgr", "pg_num": 1, "pg_num_min": 1, "pg_num_max": 32, "yes_i_really_mean_it": true}]': finished 2026-03-09T20:53:19.207 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:53:18 vm08 ceph-mon[54611]: osdmap e19: 3 total, 3 up, 3 in 2026-03-09T20:53:19.207 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:53:18 vm08 ceph-mon[54611]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' cmd=[{"prefix": "osd pool application enable", "format": "json", "pool": ".mgr", "app": "mgr", "yes_i_really_mean_it": true}]: dispatch 2026-03-09T20:53:19.207 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:53:18 vm08 ceph-mon[54611]: from='client.? 192.168.123.101:0/1181953921' entity='client.admin' cmd=[{"prefix": "mgr dump", "format": "json"}]: dispatch 2026-03-09T20:53:19.207 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:53:18 vm08 ceph-mon[54611]: from='client.? 192.168.123.101:0/3469612210' entity='client.admin' cmd=[{"prefix": "osd dump", "format": "json"}]: dispatch 2026-03-09T20:53:19.279 DEBUG:teuthology.orchestra.run.vm01:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell --fsid 9cb345a8-1bf9-11f1-a195-0375563c5891 -- ceph tell osd.0 flush_pg_stats 2026-03-09T20:53:19.279 DEBUG:teuthology.orchestra.run.vm01:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell --fsid 9cb345a8-1bf9-11f1-a195-0375563c5891 -- ceph tell osd.1 flush_pg_stats 2026-03-09T20:53:19.279 DEBUG:teuthology.orchestra.run.vm01:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell --fsid 9cb345a8-1bf9-11f1-a195-0375563c5891 -- ceph tell osd.2 flush_pg_stats 2026-03-09T20:53:19.295 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 09 20:53:18 vm06 ceph-mon[52939]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' cmd='[{"prefix": "osd pool create", "format": "json", "pool": ".mgr", "pg_num": 1, "pg_num_min": 1, "pg_num_max": 32, "yes_i_really_mean_it": true}]': finished 2026-03-09T20:53:19.295 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 09 20:53:18 vm06 ceph-mon[52939]: osdmap e19: 3 total, 3 up, 3 in 2026-03-09T20:53:19.295 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 09 20:53:18 vm06 ceph-mon[52939]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' cmd=[{"prefix": "osd pool application enable", "format": "json", "pool": ".mgr", "app": "mgr", "yes_i_really_mean_it": true}]: dispatch 2026-03-09T20:53:19.295 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 09 20:53:18 vm06 ceph-mon[52939]: from='client.? 192.168.123.101:0/1181953921' entity='client.admin' cmd=[{"prefix": "mgr dump", "format": "json"}]: dispatch 2026-03-09T20:53:19.295 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 09 20:53:18 vm06 ceph-mon[52939]: from='client.? 192.168.123.101:0/3469612210' entity='client.admin' cmd=[{"prefix": "osd dump", "format": "json"}]: dispatch 2026-03-09T20:53:19.524 INFO:teuthology.orchestra.run.vm01.stderr:Inferring config /var/lib/ceph/9cb345a8-1bf9-11f1-a195-0375563c5891/mon.a/config 2026-03-09T20:53:19.628 INFO:teuthology.orchestra.run.vm01.stderr:Inferring config /var/lib/ceph/9cb345a8-1bf9-11f1-a195-0375563c5891/mon.a/config 2026-03-09T20:53:19.633 INFO:teuthology.orchestra.run.vm01.stderr:Inferring config /var/lib/ceph/9cb345a8-1bf9-11f1-a195-0375563c5891/mon.a/config 2026-03-09T20:53:19.865 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:53:19 vm01 sudo[68855]: ceph : PWD=/ ; USER=root ; COMMAND=/usr/sbin/smartctl -x --json=o /dev/vda 2026-03-09T20:53:19.865 INFO:journalctl@ceph.osd.0.vm01.stdout:Mar 09 20:53:19 vm01 sudo[68828]: ceph : PWD=/ ; USER=root ; COMMAND=/usr/sbin/smartctl -x --json=o /dev/vde 2026-03-09T20:53:19.865 INFO:journalctl@ceph.osd.0.vm01.stdout:Mar 09 20:53:19 vm01 sudo[68828]: pam_systemd(sudo:session): Failed to connect to system bus: No such file or directory 2026-03-09T20:53:19.865 INFO:journalctl@ceph.osd.0.vm01.stdout:Mar 09 20:53:19 vm01 sudo[68828]: pam_unix(sudo:session): session opened for user root(uid=0) by (uid=167) 2026-03-09T20:53:19.865 INFO:journalctl@ceph.osd.0.vm01.stdout:Mar 09 20:53:19 vm01 sudo[68828]: pam_unix(sudo:session): session closed for user root 2026-03-09T20:53:19.925 INFO:teuthology.orchestra.run.vm01.stdout:34359738375 2026-03-09T20:53:19.925 DEBUG:teuthology.orchestra.run.vm01:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell --fsid 9cb345a8-1bf9-11f1-a195-0375563c5891 -- ceph osd last-stat-seq osd.0 2026-03-09T20:53:20.117 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:53:19 vm01 sudo[68855]: pam_systemd(sudo:session): Failed to connect to system bus: No such file or directory 2026-03-09T20:53:20.117 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:53:19 vm01 sudo[68855]: pam_unix(sudo:session): session opened for user root(uid=0) by (uid=167) 2026-03-09T20:53:20.117 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:53:19 vm01 sudo[68855]: pam_unix(sudo:session): session closed for user root 2026-03-09T20:53:20.117 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:53:19 vm01 ceph-mon[53402]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' cmd='[{"prefix": "osd pool application enable", "format": "json", "pool": ".mgr", "app": "mgr", "yes_i_really_mean_it": true}]': finished 2026-03-09T20:53:20.117 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:53:19 vm01 ceph-mon[53402]: osdmap e20: 3 total, 3 up, 3 in 2026-03-09T20:53:20.117 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:53:19 vm01 ceph-mon[53402]: from='client.? 192.168.123.101:0/1475126905' entity='client.admin' cmd=[{"prefix": "osd dump", "format": "json"}]: dispatch 2026-03-09T20:53:20.117 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:53:19 vm01 ceph-mon[53402]: pgmap v46: 1 pgs: 1 unknown; 0 B data, 79 MiB used, 60 GiB / 60 GiB avail 2026-03-09T20:53:20.117 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:53:19 vm01 ceph-mon[53402]: from='admin socket' entity='admin socket' cmd='smart' args=[json]: dispatch 2026-03-09T20:53:20.117 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:53:19 vm01 ceph-mon[53402]: from='admin socket' entity='admin socket' cmd=smart args=[json]: finished 2026-03-09T20:53:20.117 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:53:19 vm01 ceph-mon[53402]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "a"}]: dispatch 2026-03-09T20:53:20.117 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:53:19 vm01 ceph-mon[53402]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-09T20:53:20.117 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:53:19 vm01 ceph-mon[53402]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "c"}]: dispatch 2026-03-09T20:53:20.146 INFO:teuthology.orchestra.run.vm01.stdout:51539607557 2026-03-09T20:53:20.146 DEBUG:teuthology.orchestra.run.vm01:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell --fsid 9cb345a8-1bf9-11f1-a195-0375563c5891 -- ceph osd last-stat-seq osd.1 2026-03-09T20:53:20.151 INFO:teuthology.orchestra.run.vm01.stdout:73014444035 2026-03-09T20:53:20.151 DEBUG:teuthology.orchestra.run.vm01:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell --fsid 9cb345a8-1bf9-11f1-a195-0375563c5891 -- ceph osd last-stat-seq osd.2 2026-03-09T20:53:20.171 INFO:teuthology.orchestra.run.vm01.stderr:Inferring config /var/lib/ceph/9cb345a8-1bf9-11f1-a195-0375563c5891/mon.a/config 2026-03-09T20:53:20.207 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:53:19 vm08 ceph-mon[54611]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' cmd='[{"prefix": "osd pool application enable", "format": "json", "pool": ".mgr", "app": "mgr", "yes_i_really_mean_it": true}]': finished 2026-03-09T20:53:20.207 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:53:19 vm08 ceph-mon[54611]: osdmap e20: 3 total, 3 up, 3 in 2026-03-09T20:53:20.207 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:53:19 vm08 ceph-mon[54611]: from='client.? 192.168.123.101:0/1475126905' entity='client.admin' cmd=[{"prefix": "osd dump", "format": "json"}]: dispatch 2026-03-09T20:53:20.207 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:53:19 vm08 ceph-mon[54611]: pgmap v46: 1 pgs: 1 unknown; 0 B data, 79 MiB used, 60 GiB / 60 GiB avail 2026-03-09T20:53:20.207 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:53:19 vm08 ceph-mon[54611]: from='admin socket' entity='admin socket' cmd='smart' args=[json]: dispatch 2026-03-09T20:53:20.207 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:53:19 vm08 ceph-mon[54611]: from='admin socket' entity='admin socket' cmd=smart args=[json]: finished 2026-03-09T20:53:20.207 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:53:19 vm08 ceph-mon[54611]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "a"}]: dispatch 2026-03-09T20:53:20.207 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:53:19 vm08 ceph-mon[54611]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-09T20:53:20.207 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:53:19 vm08 ceph-mon[54611]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "c"}]: dispatch 2026-03-09T20:53:20.207 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:53:19 vm08 sudo[61680]: ceph : PWD=/ ; USER=root ; COMMAND=/usr/sbin/smartctl -x --json=o /dev/vda 2026-03-09T20:53:20.207 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:53:19 vm08 sudo[61680]: pam_systemd(sudo:session): Failed to connect to system bus: No such file or directory 2026-03-09T20:53:20.207 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:53:19 vm08 sudo[61680]: pam_unix(sudo:session): session opened for user root(uid=0) by (uid=167) 2026-03-09T20:53:20.207 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:53:19 vm08 sudo[61680]: pam_unix(sudo:session): session closed for user root 2026-03-09T20:53:20.207 INFO:journalctl@ceph.osd.2.vm08.stdout:Mar 09 20:53:19 vm08 sudo[61676]: ceph : PWD=/ ; USER=root ; COMMAND=/usr/sbin/smartctl -x --json=o /dev/vde 2026-03-09T20:53:20.207 INFO:journalctl@ceph.osd.2.vm08.stdout:Mar 09 20:53:19 vm08 sudo[61676]: pam_systemd(sudo:session): Failed to connect to system bus: No such file or directory 2026-03-09T20:53:20.207 INFO:journalctl@ceph.osd.2.vm08.stdout:Mar 09 20:53:19 vm08 sudo[61676]: pam_unix(sudo:session): session opened for user root(uid=0) by (uid=167) 2026-03-09T20:53:20.207 INFO:journalctl@ceph.osd.2.vm08.stdout:Mar 09 20:53:19 vm08 sudo[61676]: pam_unix(sudo:session): session closed for user root 2026-03-09T20:53:20.296 INFO:journalctl@ceph.osd.1.vm06.stdout:Mar 09 20:53:19 vm06 sudo[60295]: ceph : PWD=/ ; USER=root ; COMMAND=/usr/sbin/smartctl -x --json=o /dev/vde 2026-03-09T20:53:20.296 INFO:journalctl@ceph.osd.1.vm06.stdout:Mar 09 20:53:19 vm06 sudo[60295]: pam_systemd(sudo:session): Failed to connect to system bus: No such file or directory 2026-03-09T20:53:20.296 INFO:journalctl@ceph.osd.1.vm06.stdout:Mar 09 20:53:19 vm06 sudo[60295]: pam_unix(sudo:session): session opened for user root(uid=0) by (uid=167) 2026-03-09T20:53:20.296 INFO:journalctl@ceph.osd.1.vm06.stdout:Mar 09 20:53:19 vm06 sudo[60295]: pam_unix(sudo:session): session closed for user root 2026-03-09T20:53:20.296 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 09 20:53:19 vm06 ceph-mon[52939]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' cmd='[{"prefix": "osd pool application enable", "format": "json", "pool": ".mgr", "app": "mgr", "yes_i_really_mean_it": true}]': finished 2026-03-09T20:53:20.296 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 09 20:53:19 vm06 ceph-mon[52939]: osdmap e20: 3 total, 3 up, 3 in 2026-03-09T20:53:20.296 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 09 20:53:19 vm06 ceph-mon[52939]: from='client.? 192.168.123.101:0/1475126905' entity='client.admin' cmd=[{"prefix": "osd dump", "format": "json"}]: dispatch 2026-03-09T20:53:20.296 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 09 20:53:19 vm06 ceph-mon[52939]: pgmap v46: 1 pgs: 1 unknown; 0 B data, 79 MiB used, 60 GiB / 60 GiB avail 2026-03-09T20:53:20.296 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 09 20:53:19 vm06 ceph-mon[52939]: from='admin socket' entity='admin socket' cmd='smart' args=[json]: dispatch 2026-03-09T20:53:20.296 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 09 20:53:19 vm06 ceph-mon[52939]: from='admin socket' entity='admin socket' cmd=smart args=[json]: finished 2026-03-09T20:53:20.296 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 09 20:53:19 vm06 ceph-mon[52939]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "a"}]: dispatch 2026-03-09T20:53:20.296 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 09 20:53:19 vm06 ceph-mon[52939]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-09T20:53:20.296 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 09 20:53:19 vm06 ceph-mon[52939]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "c"}]: dispatch 2026-03-09T20:53:20.296 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 09 20:53:19 vm06 sudo[60299]: ceph : PWD=/ ; USER=root ; COMMAND=/usr/sbin/smartctl -x --json=o /dev/vda 2026-03-09T20:53:20.296 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 09 20:53:19 vm06 sudo[60299]: pam_systemd(sudo:session): Failed to connect to system bus: No such file or directory 2026-03-09T20:53:20.296 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 09 20:53:19 vm06 sudo[60299]: pam_unix(sudo:session): session opened for user root(uid=0) by (uid=167) 2026-03-09T20:53:20.296 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 09 20:53:19 vm06 sudo[60299]: pam_unix(sudo:session): session closed for user root 2026-03-09T20:53:20.420 INFO:teuthology.orchestra.run.vm01.stderr:Inferring config /var/lib/ceph/9cb345a8-1bf9-11f1-a195-0375563c5891/mon.a/config 2026-03-09T20:53:20.497 INFO:teuthology.orchestra.run.vm01.stdout:34359738374 2026-03-09T20:53:20.526 INFO:teuthology.orchestra.run.vm01.stderr:Inferring config /var/lib/ceph/9cb345a8-1bf9-11f1-a195-0375563c5891/mon.a/config 2026-03-09T20:53:20.691 INFO:tasks.cephadm.ceph_manager.ceph:need seq 34359738375 got 34359738374 for osd.0 2026-03-09T20:53:20.797 INFO:teuthology.orchestra.run.vm01.stdout:73014444034 2026-03-09T20:53:20.798 INFO:teuthology.orchestra.run.vm01.stdout:51539607555 2026-03-09T20:53:21.100 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:53:21 vm01 ceph-mon[53402]: from='admin socket' entity='admin socket' cmd='smart' args=[json]: dispatch 2026-03-09T20:53:21.100 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:53:21 vm01 ceph-mon[53402]: osdmap e21: 3 total, 3 up, 3 in 2026-03-09T20:53:21.100 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:53:21 vm01 ceph-mon[53402]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "a"}]: dispatch 2026-03-09T20:53:21.100 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:53:21 vm01 ceph-mon[53402]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-09T20:53:21.136 INFO:tasks.cephadm.ceph_manager.ceph:need seq 73014444035 got 73014444034 for osd.2 2026-03-09T20:53:21.189 INFO:tasks.cephadm.ceph_manager.ceph:need seq 51539607557 got 51539607555 for osd.1 2026-03-09T20:53:21.424 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:53:21 vm01 ceph-mon[53402]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "c"}]: dispatch 2026-03-09T20:53:21.424 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:53:21 vm01 ceph-mon[53402]: from='admin socket' entity='admin socket' cmd=smart args=[json]: finished 2026-03-09T20:53:21.424 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:53:21 vm01 ceph-mon[53402]: from='admin socket' entity='admin socket' cmd='smart' args=[json]: dispatch 2026-03-09T20:53:21.424 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:53:21 vm01 ceph-mon[53402]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "a"}]: dispatch 2026-03-09T20:53:21.424 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:53:21 vm01 ceph-mon[53402]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-09T20:53:21.424 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:53:21 vm01 ceph-mon[53402]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "c"}]: dispatch 2026-03-09T20:53:21.424 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:53:21 vm01 ceph-mon[53402]: from='admin socket' entity='admin socket' cmd=smart args=[json]: finished 2026-03-09T20:53:21.424 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:53:21 vm01 ceph-mon[53402]: from='client.? 192.168.123.101:0/1046277700' entity='client.admin' cmd=[{"prefix": "osd last-stat-seq", "id": 0}]: dispatch 2026-03-09T20:53:21.424 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:53:21 vm01 ceph-mon[53402]: from='client.? 192.168.123.101:0/1920888798' entity='client.admin' cmd=[{"prefix": "osd last-stat-seq", "id": 2}]: dispatch 2026-03-09T20:53:21.424 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:53:21 vm01 ceph-mon[53402]: from='client.? 192.168.123.101:0/4050147034' entity='client.admin' cmd=[{"prefix": "osd last-stat-seq", "id": 1}]: dispatch 2026-03-09T20:53:21.457 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:53:21 vm08 ceph-mon[54611]: from='admin socket' entity='admin socket' cmd='smart' args=[json]: dispatch 2026-03-09T20:53:21.457 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:53:21 vm08 ceph-mon[54611]: osdmap e21: 3 total, 3 up, 3 in 2026-03-09T20:53:21.457 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:53:21 vm08 ceph-mon[54611]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "a"}]: dispatch 2026-03-09T20:53:21.457 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:53:21 vm08 ceph-mon[54611]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-09T20:53:21.457 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:53:21 vm08 ceph-mon[54611]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "c"}]: dispatch 2026-03-09T20:53:21.457 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:53:21 vm08 ceph-mon[54611]: from='admin socket' entity='admin socket' cmd=smart args=[json]: finished 2026-03-09T20:53:21.457 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:53:21 vm08 ceph-mon[54611]: from='admin socket' entity='admin socket' cmd='smart' args=[json]: dispatch 2026-03-09T20:53:21.457 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:53:21 vm08 ceph-mon[54611]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "a"}]: dispatch 2026-03-09T20:53:21.457 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:53:21 vm08 ceph-mon[54611]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-09T20:53:21.457 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:53:21 vm08 ceph-mon[54611]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "c"}]: dispatch 2026-03-09T20:53:21.458 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:53:21 vm08 ceph-mon[54611]: from='admin socket' entity='admin socket' cmd=smart args=[json]: finished 2026-03-09T20:53:21.458 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:53:21 vm08 ceph-mon[54611]: from='client.? 192.168.123.101:0/1046277700' entity='client.admin' cmd=[{"prefix": "osd last-stat-seq", "id": 0}]: dispatch 2026-03-09T20:53:21.458 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:53:21 vm08 ceph-mon[54611]: from='client.? 192.168.123.101:0/1920888798' entity='client.admin' cmd=[{"prefix": "osd last-stat-seq", "id": 2}]: dispatch 2026-03-09T20:53:21.458 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:53:21 vm08 ceph-mon[54611]: from='client.? 192.168.123.101:0/4050147034' entity='client.admin' cmd=[{"prefix": "osd last-stat-seq", "id": 1}]: dispatch 2026-03-09T20:53:21.546 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 09 20:53:21 vm06 ceph-mon[52939]: from='admin socket' entity='admin socket' cmd='smart' args=[json]: dispatch 2026-03-09T20:53:21.546 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 09 20:53:21 vm06 ceph-mon[52939]: osdmap e21: 3 total, 3 up, 3 in 2026-03-09T20:53:21.546 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 09 20:53:21 vm06 ceph-mon[52939]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "a"}]: dispatch 2026-03-09T20:53:21.546 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 09 20:53:21 vm06 ceph-mon[52939]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-09T20:53:21.546 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 09 20:53:21 vm06 ceph-mon[52939]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "c"}]: dispatch 2026-03-09T20:53:21.546 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 09 20:53:21 vm06 ceph-mon[52939]: from='admin socket' entity='admin socket' cmd=smart args=[json]: finished 2026-03-09T20:53:21.546 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 09 20:53:21 vm06 ceph-mon[52939]: from='admin socket' entity='admin socket' cmd='smart' args=[json]: dispatch 2026-03-09T20:53:21.546 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 09 20:53:21 vm06 ceph-mon[52939]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "a"}]: dispatch 2026-03-09T20:53:21.546 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 09 20:53:21 vm06 ceph-mon[52939]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-09T20:53:21.546 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 09 20:53:21 vm06 ceph-mon[52939]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "c"}]: dispatch 2026-03-09T20:53:21.546 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 09 20:53:21 vm06 ceph-mon[52939]: from='admin socket' entity='admin socket' cmd=smart args=[json]: finished 2026-03-09T20:53:21.546 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 09 20:53:21 vm06 ceph-mon[52939]: from='client.? 192.168.123.101:0/1046277700' entity='client.admin' cmd=[{"prefix": "osd last-stat-seq", "id": 0}]: dispatch 2026-03-09T20:53:21.546 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 09 20:53:21 vm06 ceph-mon[52939]: from='client.? 192.168.123.101:0/1920888798' entity='client.admin' cmd=[{"prefix": "osd last-stat-seq", "id": 2}]: dispatch 2026-03-09T20:53:21.546 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 09 20:53:21 vm06 ceph-mon[52939]: from='client.? 192.168.123.101:0/4050147034' entity='client.admin' cmd=[{"prefix": "osd last-stat-seq", "id": 1}]: dispatch 2026-03-09T20:53:21.692 DEBUG:teuthology.orchestra.run.vm01:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell --fsid 9cb345a8-1bf9-11f1-a195-0375563c5891 -- ceph osd last-stat-seq osd.0 2026-03-09T20:53:21.867 INFO:teuthology.orchestra.run.vm01.stderr:Inferring config /var/lib/ceph/9cb345a8-1bf9-11f1-a195-0375563c5891/mon.a/config 2026-03-09T20:53:22.137 DEBUG:teuthology.orchestra.run.vm01:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell --fsid 9cb345a8-1bf9-11f1-a195-0375563c5891 -- ceph osd last-stat-seq osd.2 2026-03-09T20:53:22.190 DEBUG:teuthology.orchestra.run.vm01:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell --fsid 9cb345a8-1bf9-11f1-a195-0375563c5891 -- ceph osd last-stat-seq osd.1 2026-03-09T20:53:22.366 INFO:teuthology.orchestra.run.vm01.stdout:34359738374 2026-03-09T20:53:22.579 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:53:22 vm01 ceph-mon[53402]: pgmap v48: 1 pgs: 1 active+clean; 320 KiB data, 79 MiB used, 60 GiB / 60 GiB avail 2026-03-09T20:53:22.661 INFO:teuthology.orchestra.run.vm01.stderr:Inferring config /var/lib/ceph/9cb345a8-1bf9-11f1-a195-0375563c5891/mon.a/config 2026-03-09T20:53:22.701 INFO:tasks.cephadm.ceph_manager.ceph:need seq 34359738375 got 34359738374 for osd.0 2026-03-09T20:53:22.703 INFO:teuthology.orchestra.run.vm01.stderr:Inferring config /var/lib/ceph/9cb345a8-1bf9-11f1-a195-0375563c5891/mon.a/config 2026-03-09T20:53:22.957 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:53:22 vm08 ceph-mon[54611]: pgmap v48: 1 pgs: 1 active+clean; 320 KiB data, 79 MiB used, 60 GiB / 60 GiB avail 2026-03-09T20:53:23.045 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 09 20:53:22 vm06 ceph-mon[52939]: pgmap v48: 1 pgs: 1 active+clean; 320 KiB data, 79 MiB used, 60 GiB / 60 GiB avail 2026-03-09T20:53:23.117 INFO:teuthology.orchestra.run.vm01.stdout:73014444035 2026-03-09T20:53:23.246 INFO:teuthology.orchestra.run.vm01.stdout:51539607557 2026-03-09T20:53:23.378 INFO:tasks.cephadm.ceph_manager.ceph:need seq 73014444035 got 73014444035 for osd.2 2026-03-09T20:53:23.379 DEBUG:teuthology.parallel:result is None 2026-03-09T20:53:23.448 INFO:tasks.cephadm.ceph_manager.ceph:need seq 51539607557 got 51539607557 for osd.1 2026-03-09T20:53:23.448 DEBUG:teuthology.parallel:result is None 2026-03-09T20:53:23.673 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:53:23 vm01 ceph-mon[53402]: from='client.? 192.168.123.101:0/3597052812' entity='client.admin' cmd=[{"prefix": "osd last-stat-seq", "id": 0}]: dispatch 2026-03-09T20:53:23.673 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:53:23 vm01 ceph-mon[53402]: mgrmap e15: a(active, since 75s), standbys: b 2026-03-09T20:53:23.673 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:53:23 vm01 ceph-mon[53402]: from='client.? 192.168.123.101:0/3931554977' entity='client.admin' cmd=[{"prefix": "osd last-stat-seq", "id": 2}]: dispatch 2026-03-09T20:53:23.673 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:53:23 vm01 ceph-mon[53402]: from='client.? 192.168.123.101:0/3555577039' entity='client.admin' cmd=[{"prefix": "osd last-stat-seq", "id": 1}]: dispatch 2026-03-09T20:53:23.673 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:53:23 vm01 ceph-mon[53402]: pgmap v49: 1 pgs: 1 active+clean; 320 KiB data, 80 MiB used, 60 GiB / 60 GiB avail 2026-03-09T20:53:23.702 DEBUG:teuthology.orchestra.run.vm01:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell --fsid 9cb345a8-1bf9-11f1-a195-0375563c5891 -- ceph osd last-stat-seq osd.0 2026-03-09T20:53:23.878 INFO:teuthology.orchestra.run.vm01.stderr:Inferring config /var/lib/ceph/9cb345a8-1bf9-11f1-a195-0375563c5891/mon.a/config 2026-03-09T20:53:23.957 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:53:23 vm08 ceph-mon[54611]: from='client.? 192.168.123.101:0/3597052812' entity='client.admin' cmd=[{"prefix": "osd last-stat-seq", "id": 0}]: dispatch 2026-03-09T20:53:23.957 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:53:23 vm08 ceph-mon[54611]: mgrmap e15: a(active, since 75s), standbys: b 2026-03-09T20:53:23.957 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:53:23 vm08 ceph-mon[54611]: from='client.? 192.168.123.101:0/3931554977' entity='client.admin' cmd=[{"prefix": "osd last-stat-seq", "id": 2}]: dispatch 2026-03-09T20:53:23.957 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:53:23 vm08 ceph-mon[54611]: from='client.? 192.168.123.101:0/3555577039' entity='client.admin' cmd=[{"prefix": "osd last-stat-seq", "id": 1}]: dispatch 2026-03-09T20:53:23.957 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:53:23 vm08 ceph-mon[54611]: pgmap v49: 1 pgs: 1 active+clean; 320 KiB data, 80 MiB used, 60 GiB / 60 GiB avail 2026-03-09T20:53:24.045 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 09 20:53:23 vm06 ceph-mon[52939]: from='client.? 192.168.123.101:0/3597052812' entity='client.admin' cmd=[{"prefix": "osd last-stat-seq", "id": 0}]: dispatch 2026-03-09T20:53:24.045 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 09 20:53:23 vm06 ceph-mon[52939]: mgrmap e15: a(active, since 75s), standbys: b 2026-03-09T20:53:24.045 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 09 20:53:23 vm06 ceph-mon[52939]: from='client.? 192.168.123.101:0/3931554977' entity='client.admin' cmd=[{"prefix": "osd last-stat-seq", "id": 2}]: dispatch 2026-03-09T20:53:24.045 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 09 20:53:23 vm06 ceph-mon[52939]: from='client.? 192.168.123.101:0/3555577039' entity='client.admin' cmd=[{"prefix": "osd last-stat-seq", "id": 1}]: dispatch 2026-03-09T20:53:24.045 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 09 20:53:23 vm06 ceph-mon[52939]: pgmap v49: 1 pgs: 1 active+clean; 320 KiB data, 80 MiB used, 60 GiB / 60 GiB avail 2026-03-09T20:53:24.122 INFO:teuthology.orchestra.run.vm01.stdout:34359738376 2026-03-09T20:53:24.293 INFO:tasks.cephadm.ceph_manager.ceph:need seq 34359738375 got 34359738376 for osd.0 2026-03-09T20:53:24.293 DEBUG:teuthology.parallel:result is None 2026-03-09T20:53:24.293 INFO:tasks.cephadm.ceph_manager.ceph:waiting for clean 2026-03-09T20:53:24.293 DEBUG:teuthology.orchestra.run.vm01:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell --fsid 9cb345a8-1bf9-11f1-a195-0375563c5891 -- ceph pg dump --format=json 2026-03-09T20:53:24.461 INFO:teuthology.orchestra.run.vm01.stderr:Inferring config /var/lib/ceph/9cb345a8-1bf9-11f1-a195-0375563c5891/mon.a/config 2026-03-09T20:53:24.675 INFO:teuthology.orchestra.run.vm01.stdout: 2026-03-09T20:53:24.675 INFO:teuthology.orchestra.run.vm01.stderr:dumped all 2026-03-09T20:53:24.798 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:53:24 vm01 ceph-mon[53402]: from='client.? 192.168.123.101:0/3668378898' entity='client.admin' cmd=[{"prefix": "osd last-stat-seq", "id": 0}]: dispatch 2026-03-09T20:53:24.821 INFO:teuthology.orchestra.run.vm01.stdout:{"pg_ready":true,"pg_map":{"version":49,"stamp":"2026-03-09T20:53:23.403350+0000","last_osdmap_epoch":0,"last_pg_scan":0,"pg_stats_sum":{"stat_sum":{"num_bytes":328192,"num_objects":2,"num_object_clones":0,"num_object_copies":6,"num_objects_missing_on_primary":0,"num_objects_missing":0,"num_objects_degraded":0,"num_objects_misplaced":0,"num_objects_unfound":0,"num_objects_dirty":2,"num_whiteouts":0,"num_read":36,"num_read_kb":28,"num_write":41,"num_write_kb":322,"num_scrub_errors":0,"num_shallow_scrub_errors":0,"num_deep_scrub_errors":0,"num_objects_recovered":0,"num_bytes_recovered":0,"num_keys_recovered":0,"num_objects_omap":0,"num_objects_hit_set_archive":0,"num_bytes_hit_set_archive":0,"num_flush":0,"num_flush_kb":0,"num_evict":0,"num_evict_kb":0,"num_promote":0,"num_flush_mode_high":0,"num_flush_mode_low":0,"num_evict_mode_some":0,"num_evict_mode_full":0,"num_objects_pinned":0,"num_legacy_snapsets":0,"num_large_omap_objects":0,"num_objects_manifest":0,"num_omap_bytes":0,"num_omap_keys":0,"num_objects_repaired":0},"store_stats":{"total":0,"available":0,"internally_reserved":0,"allocated":0,"data_stored":0,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":0},"log_size":19,"ondisk_log_size":19,"up":3,"acting":3,"num_store_stats":0},"osd_stats_sum":{"up_from":0,"seq":0,"num_pgs":2,"num_osds":3,"num_per_pool_osds":3,"num_per_pool_omap_osds":3,"kb":62902272,"kb_used":82200,"kb_used_data":1424,"kb_used_omap":4,"kb_used_meta":80443,"kb_avail":62820072,"statfs":{"total":64411926528,"available":64327753728,"internally_reserved":0,"allocated":1458176,"data_stored":1081997,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":4770,"internal_metadata":82373982},"hb_peers":[],"snap_trim_queue_len":0,"num_snap_trimming":0,"num_shards_repaired":0,"op_queue_age_hist":{"histogram":[],"upper_bound":1},"perf_stat":{"commit_latency_ms":2,"apply_latency_ms":2,"commit_latency_ns":2000000,"apply_latency_ns":2000000},"alerts":[],"network_ping_times":[]},"pg_stats_delta":{"stat_sum":{"num_bytes":0,"num_objects":0,"num_object_clones":0,"num_object_copies":0,"num_objects_missing_on_primary":0,"num_objects_missing":0,"num_objects_degraded":0,"num_objects_misplaced":0,"num_objects_unfound":0,"num_objects_dirty":0,"num_whiteouts":0,"num_read":0,"num_read_kb":0,"num_write":0,"num_write_kb":0,"num_scrub_errors":0,"num_shallow_scrub_errors":0,"num_deep_scrub_errors":0,"num_objects_recovered":0,"num_bytes_recovered":0,"num_keys_recovered":0,"num_objects_omap":0,"num_objects_hit_set_archive":0,"num_bytes_hit_set_archive":0,"num_flush":0,"num_flush_kb":0,"num_evict":0,"num_evict_kb":0,"num_promote":0,"num_flush_mode_high":0,"num_flush_mode_low":0,"num_evict_mode_some":0,"num_evict_mode_full":0,"num_objects_pinned":0,"num_legacy_snapsets":0,"num_large_omap_objects":0,"num_objects_manifest":0,"num_omap_bytes":0,"num_omap_keys":0,"num_objects_repaired":0},"store_stats":{"total":0,"available":0,"internally_reserved":0,"allocated":0,"data_stored":0,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":0},"log_size":0,"ondisk_log_size":0,"up":0,"acting":0,"num_store_stats":0,"stamp_delta":"2.000334"},"pg_stats":[{"pgid":"1.0","version":"20'19","reported_seq":38,"reported_epoch":20,"state":"active+clean","last_fresh":"2026-03-09T20:53:19.750797+0000","last_change":"2026-03-09T20:53:19.701593+0000","last_active":"2026-03-09T20:53:19.750797+0000","last_peered":"2026-03-09T20:53:19.750797+0000","last_clean":"2026-03-09T20:53:19.750797+0000","last_became_active":"2026-03-09T20:53:19.701365+0000","last_became_peered":"2026-03-09T20:53:19.701365+0000","last_unstale":"2026-03-09T20:53:19.750797+0000","last_undegraded":"2026-03-09T20:53:19.750797+0000","last_fullsized":"2026-03-09T20:53:19.750797+0000","mapping_epoch":19,"log_start":"0'0","ondisk_log_start":"0'0","created":19,"last_epoch_clean":20,"parent":"0.0","parent_split_bits":0,"last_scrub":"0'0","last_scrub_stamp":"2026-03-09T20:53:17.782831+0000","last_deep_scrub":"0'0","last_deep_scrub_stamp":"2026-03-09T20:53:17.782831+0000","last_clean_scrub_stamp":"2026-03-09T20:53:17.782831+0000","objects_scrubbed":0,"log_size":19,"log_dups_size":0,"ondisk_log_size":19,"stats_invalid":false,"dirty_stats_invalid":false,"omap_stats_invalid":false,"hitset_stats_invalid":false,"hitset_bytes_stats_invalid":false,"pin_stats_invalid":false,"manifest_stats_invalid":false,"snaptrimq_len":0,"last_scrub_duration":0,"scrub_schedule":"periodic scrub scheduled @ 2026-03-11T04:24:56.341492+0000","scrub_duration":0,"objects_trimmed":0,"snaptrim_duration":0,"stat_sum":{"num_bytes":328192,"num_objects":2,"num_object_clones":0,"num_object_copies":6,"num_objects_missing_on_primary":0,"num_objects_missing":0,"num_objects_degraded":0,"num_objects_misplaced":0,"num_objects_unfound":0,"num_objects_dirty":2,"num_whiteouts":0,"num_read":36,"num_read_kb":28,"num_write":41,"num_write_kb":322,"num_scrub_errors":0,"num_shallow_scrub_errors":0,"num_deep_scrub_errors":0,"num_objects_recovered":0,"num_bytes_recovered":0,"num_keys_recovered":0,"num_objects_omap":0,"num_objects_hit_set_archive":0,"num_bytes_hit_set_archive":0,"num_flush":0,"num_flush_kb":0,"num_evict":0,"num_evict_kb":0,"num_promote":0,"num_flush_mode_high":0,"num_flush_mode_low":0,"num_evict_mode_some":0,"num_evict_mode_full":0,"num_objects_pinned":0,"num_legacy_snapsets":0,"num_large_omap_objects":0,"num_objects_manifest":0,"num_omap_bytes":0,"num_omap_keys":0,"num_objects_repaired":0},"up":[1,2,0],"acting":[1,2,0],"avail_no_missing":[],"object_location_counts":[],"blocked_by":[],"up_primary":1,"acting_primary":1,"purged_snaps":[]}],"pool_stats":[{"poolid":1,"num_pg":1,"stat_sum":{"num_bytes":328192,"num_objects":2,"num_object_clones":0,"num_object_copies":6,"num_objects_missing_on_primary":0,"num_objects_missing":0,"num_objects_degraded":0,"num_objects_misplaced":0,"num_objects_unfound":0,"num_objects_dirty":2,"num_whiteouts":0,"num_read":36,"num_read_kb":28,"num_write":41,"num_write_kb":322,"num_scrub_errors":0,"num_shallow_scrub_errors":0,"num_deep_scrub_errors":0,"num_objects_recovered":0,"num_bytes_recovered":0,"num_keys_recovered":0,"num_objects_omap":0,"num_objects_hit_set_archive":0,"num_bytes_hit_set_archive":0,"num_flush":0,"num_flush_kb":0,"num_evict":0,"num_evict_kb":0,"num_promote":0,"num_flush_mode_high":0,"num_flush_mode_low":0,"num_evict_mode_some":0,"num_evict_mode_full":0,"num_objects_pinned":0,"num_legacy_snapsets":0,"num_large_omap_objects":0,"num_objects_manifest":0,"num_omap_bytes":0,"num_omap_keys":0,"num_objects_repaired":0},"store_stats":{"total":0,"available":0,"internally_reserved":0,"allocated":1257472,"data_stored":1246752,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":0},"log_size":19,"ondisk_log_size":19,"up":3,"acting":3,"num_store_stats":3}],"osd_stats":[{"osd":2,"up_from":17,"seq":73014444036,"num_pgs":1,"num_osds":1,"num_per_pool_osds":1,"num_per_pool_omap_osds":1,"kb":20967424,"kb_used":27596,"kb_used_data":628,"kb_used_omap":1,"kb_used_meta":26814,"kb_avail":20939828,"statfs":{"total":21470642176,"available":21442383872,"internally_reserved":0,"allocated":643072,"data_stored":515018,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":1590,"internal_metadata":27457994},"hb_peers":[0,1],"snap_trim_queue_len":0,"num_snap_trimming":0,"num_shards_repaired":0,"op_queue_age_hist":{"histogram":[],"upper_bound":1},"perf_stat":{"commit_latency_ms":1,"apply_latency_ms":1,"commit_latency_ns":1000000,"apply_latency_ns":1000000},"alerts":[]},{"osd":1,"up_from":12,"seq":51539607557,"num_pgs":0,"num_osds":1,"num_per_pool_osds":1,"num_per_pool_omap_osds":1,"kb":20967424,"kb_used":27012,"kb_used_data":168,"kb_used_omap":1,"kb_used_meta":26814,"kb_avail":20940412,"statfs":{"total":21470642176,"available":21442981888,"internally_reserved":0,"allocated":172032,"data_stored":51961,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":1590,"internal_metadata":27457994},"hb_peers":[0,2],"snap_trim_queue_len":0,"num_snap_trimming":0,"num_shards_repaired":0,"op_queue_age_hist":{"histogram":[],"upper_bound":1},"perf_stat":{"commit_latency_ms":0,"apply_latency_ms":0,"commit_latency_ns":0,"apply_latency_ns":0},"alerts":[]},{"osd":0,"up_from":8,"seq":34359738376,"num_pgs":1,"num_osds":1,"num_per_pool_osds":1,"num_per_pool_omap_osds":1,"kb":20967424,"kb_used":27592,"kb_used_data":628,"kb_used_omap":1,"kb_used_meta":26814,"kb_avail":20939832,"statfs":{"total":21470642176,"available":21442387968,"internally_reserved":0,"allocated":643072,"data_stored":515018,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":1590,"internal_metadata":27457994},"hb_peers":[1,2],"snap_trim_queue_len":0,"num_snap_trimming":0,"num_shards_repaired":0,"op_queue_age_hist":{"histogram":[],"upper_bound":1},"perf_stat":{"commit_latency_ms":1,"apply_latency_ms":1,"commit_latency_ns":1000000,"apply_latency_ns":1000000},"alerts":[]}],"pool_statfs":[{"poolid":1,"osd":0,"total":0,"available":0,"internally_reserved":0,"allocated":462848,"data_stored":459280,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":0},{"poolid":1,"osd":1,"total":0,"available":0,"internally_reserved":0,"allocated":331776,"data_stored":328192,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":0},{"poolid":1,"osd":2,"total":0,"available":0,"internally_reserved":0,"allocated":462848,"data_stored":459280,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":0}]}} 2026-03-09T20:53:24.821 DEBUG:teuthology.orchestra.run.vm01:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell --fsid 9cb345a8-1bf9-11f1-a195-0375563c5891 -- ceph pg dump --format=json 2026-03-09T20:53:24.957 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:53:24 vm08 ceph-mon[54611]: from='client.? 192.168.123.101:0/3668378898' entity='client.admin' cmd=[{"prefix": "osd last-stat-seq", "id": 0}]: dispatch 2026-03-09T20:53:24.971 INFO:teuthology.orchestra.run.vm01.stderr:Inferring config /var/lib/ceph/9cb345a8-1bf9-11f1-a195-0375563c5891/mon.a/config 2026-03-09T20:53:25.045 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 09 20:53:24 vm06 ceph-mon[52939]: from='client.? 192.168.123.101:0/3668378898' entity='client.admin' cmd=[{"prefix": "osd last-stat-seq", "id": 0}]: dispatch 2026-03-09T20:53:25.214 INFO:teuthology.orchestra.run.vm01.stdout: 2026-03-09T20:53:25.214 INFO:teuthology.orchestra.run.vm01.stderr:dumped all 2026-03-09T20:53:25.381 INFO:teuthology.orchestra.run.vm01.stdout:{"pg_ready":true,"pg_map":{"version":49,"stamp":"2026-03-09T20:53:23.403350+0000","last_osdmap_epoch":0,"last_pg_scan":0,"pg_stats_sum":{"stat_sum":{"num_bytes":328192,"num_objects":2,"num_object_clones":0,"num_object_copies":6,"num_objects_missing_on_primary":0,"num_objects_missing":0,"num_objects_degraded":0,"num_objects_misplaced":0,"num_objects_unfound":0,"num_objects_dirty":2,"num_whiteouts":0,"num_read":36,"num_read_kb":28,"num_write":41,"num_write_kb":322,"num_scrub_errors":0,"num_shallow_scrub_errors":0,"num_deep_scrub_errors":0,"num_objects_recovered":0,"num_bytes_recovered":0,"num_keys_recovered":0,"num_objects_omap":0,"num_objects_hit_set_archive":0,"num_bytes_hit_set_archive":0,"num_flush":0,"num_flush_kb":0,"num_evict":0,"num_evict_kb":0,"num_promote":0,"num_flush_mode_high":0,"num_flush_mode_low":0,"num_evict_mode_some":0,"num_evict_mode_full":0,"num_objects_pinned":0,"num_legacy_snapsets":0,"num_large_omap_objects":0,"num_objects_manifest":0,"num_omap_bytes":0,"num_omap_keys":0,"num_objects_repaired":0},"store_stats":{"total":0,"available":0,"internally_reserved":0,"allocated":0,"data_stored":0,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":0},"log_size":19,"ondisk_log_size":19,"up":3,"acting":3,"num_store_stats":0},"osd_stats_sum":{"up_from":0,"seq":0,"num_pgs":2,"num_osds":3,"num_per_pool_osds":3,"num_per_pool_omap_osds":3,"kb":62902272,"kb_used":82200,"kb_used_data":1424,"kb_used_omap":4,"kb_used_meta":80443,"kb_avail":62820072,"statfs":{"total":64411926528,"available":64327753728,"internally_reserved":0,"allocated":1458176,"data_stored":1081997,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":4770,"internal_metadata":82373982},"hb_peers":[],"snap_trim_queue_len":0,"num_snap_trimming":0,"num_shards_repaired":0,"op_queue_age_hist":{"histogram":[],"upper_bound":1},"perf_stat":{"commit_latency_ms":2,"apply_latency_ms":2,"commit_latency_ns":2000000,"apply_latency_ns":2000000},"alerts":[],"network_ping_times":[]},"pg_stats_delta":{"stat_sum":{"num_bytes":0,"num_objects":0,"num_object_clones":0,"num_object_copies":0,"num_objects_missing_on_primary":0,"num_objects_missing":0,"num_objects_degraded":0,"num_objects_misplaced":0,"num_objects_unfound":0,"num_objects_dirty":0,"num_whiteouts":0,"num_read":0,"num_read_kb":0,"num_write":0,"num_write_kb":0,"num_scrub_errors":0,"num_shallow_scrub_errors":0,"num_deep_scrub_errors":0,"num_objects_recovered":0,"num_bytes_recovered":0,"num_keys_recovered":0,"num_objects_omap":0,"num_objects_hit_set_archive":0,"num_bytes_hit_set_archive":0,"num_flush":0,"num_flush_kb":0,"num_evict":0,"num_evict_kb":0,"num_promote":0,"num_flush_mode_high":0,"num_flush_mode_low":0,"num_evict_mode_some":0,"num_evict_mode_full":0,"num_objects_pinned":0,"num_legacy_snapsets":0,"num_large_omap_objects":0,"num_objects_manifest":0,"num_omap_bytes":0,"num_omap_keys":0,"num_objects_repaired":0},"store_stats":{"total":0,"available":0,"internally_reserved":0,"allocated":0,"data_stored":0,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":0},"log_size":0,"ondisk_log_size":0,"up":0,"acting":0,"num_store_stats":0,"stamp_delta":"2.000334"},"pg_stats":[{"pgid":"1.0","version":"20'19","reported_seq":38,"reported_epoch":20,"state":"active+clean","last_fresh":"2026-03-09T20:53:19.750797+0000","last_change":"2026-03-09T20:53:19.701593+0000","last_active":"2026-03-09T20:53:19.750797+0000","last_peered":"2026-03-09T20:53:19.750797+0000","last_clean":"2026-03-09T20:53:19.750797+0000","last_became_active":"2026-03-09T20:53:19.701365+0000","last_became_peered":"2026-03-09T20:53:19.701365+0000","last_unstale":"2026-03-09T20:53:19.750797+0000","last_undegraded":"2026-03-09T20:53:19.750797+0000","last_fullsized":"2026-03-09T20:53:19.750797+0000","mapping_epoch":19,"log_start":"0'0","ondisk_log_start":"0'0","created":19,"last_epoch_clean":20,"parent":"0.0","parent_split_bits":0,"last_scrub":"0'0","last_scrub_stamp":"2026-03-09T20:53:17.782831+0000","last_deep_scrub":"0'0","last_deep_scrub_stamp":"2026-03-09T20:53:17.782831+0000","last_clean_scrub_stamp":"2026-03-09T20:53:17.782831+0000","objects_scrubbed":0,"log_size":19,"log_dups_size":0,"ondisk_log_size":19,"stats_invalid":false,"dirty_stats_invalid":false,"omap_stats_invalid":false,"hitset_stats_invalid":false,"hitset_bytes_stats_invalid":false,"pin_stats_invalid":false,"manifest_stats_invalid":false,"snaptrimq_len":0,"last_scrub_duration":0,"scrub_schedule":"periodic scrub scheduled @ 2026-03-11T04:24:56.341492+0000","scrub_duration":0,"objects_trimmed":0,"snaptrim_duration":0,"stat_sum":{"num_bytes":328192,"num_objects":2,"num_object_clones":0,"num_object_copies":6,"num_objects_missing_on_primary":0,"num_objects_missing":0,"num_objects_degraded":0,"num_objects_misplaced":0,"num_objects_unfound":0,"num_objects_dirty":2,"num_whiteouts":0,"num_read":36,"num_read_kb":28,"num_write":41,"num_write_kb":322,"num_scrub_errors":0,"num_shallow_scrub_errors":0,"num_deep_scrub_errors":0,"num_objects_recovered":0,"num_bytes_recovered":0,"num_keys_recovered":0,"num_objects_omap":0,"num_objects_hit_set_archive":0,"num_bytes_hit_set_archive":0,"num_flush":0,"num_flush_kb":0,"num_evict":0,"num_evict_kb":0,"num_promote":0,"num_flush_mode_high":0,"num_flush_mode_low":0,"num_evict_mode_some":0,"num_evict_mode_full":0,"num_objects_pinned":0,"num_legacy_snapsets":0,"num_large_omap_objects":0,"num_objects_manifest":0,"num_omap_bytes":0,"num_omap_keys":0,"num_objects_repaired":0},"up":[1,2,0],"acting":[1,2,0],"avail_no_missing":[],"object_location_counts":[],"blocked_by":[],"up_primary":1,"acting_primary":1,"purged_snaps":[]}],"pool_stats":[{"poolid":1,"num_pg":1,"stat_sum":{"num_bytes":328192,"num_objects":2,"num_object_clones":0,"num_object_copies":6,"num_objects_missing_on_primary":0,"num_objects_missing":0,"num_objects_degraded":0,"num_objects_misplaced":0,"num_objects_unfound":0,"num_objects_dirty":2,"num_whiteouts":0,"num_read":36,"num_read_kb":28,"num_write":41,"num_write_kb":322,"num_scrub_errors":0,"num_shallow_scrub_errors":0,"num_deep_scrub_errors":0,"num_objects_recovered":0,"num_bytes_recovered":0,"num_keys_recovered":0,"num_objects_omap":0,"num_objects_hit_set_archive":0,"num_bytes_hit_set_archive":0,"num_flush":0,"num_flush_kb":0,"num_evict":0,"num_evict_kb":0,"num_promote":0,"num_flush_mode_high":0,"num_flush_mode_low":0,"num_evict_mode_some":0,"num_evict_mode_full":0,"num_objects_pinned":0,"num_legacy_snapsets":0,"num_large_omap_objects":0,"num_objects_manifest":0,"num_omap_bytes":0,"num_omap_keys":0,"num_objects_repaired":0},"store_stats":{"total":0,"available":0,"internally_reserved":0,"allocated":1257472,"data_stored":1246752,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":0},"log_size":19,"ondisk_log_size":19,"up":3,"acting":3,"num_store_stats":3}],"osd_stats":[{"osd":2,"up_from":17,"seq":73014444036,"num_pgs":1,"num_osds":1,"num_per_pool_osds":1,"num_per_pool_omap_osds":1,"kb":20967424,"kb_used":27596,"kb_used_data":628,"kb_used_omap":1,"kb_used_meta":26814,"kb_avail":20939828,"statfs":{"total":21470642176,"available":21442383872,"internally_reserved":0,"allocated":643072,"data_stored":515018,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":1590,"internal_metadata":27457994},"hb_peers":[0,1],"snap_trim_queue_len":0,"num_snap_trimming":0,"num_shards_repaired":0,"op_queue_age_hist":{"histogram":[],"upper_bound":1},"perf_stat":{"commit_latency_ms":1,"apply_latency_ms":1,"commit_latency_ns":1000000,"apply_latency_ns":1000000},"alerts":[]},{"osd":1,"up_from":12,"seq":51539607557,"num_pgs":0,"num_osds":1,"num_per_pool_osds":1,"num_per_pool_omap_osds":1,"kb":20967424,"kb_used":27012,"kb_used_data":168,"kb_used_omap":1,"kb_used_meta":26814,"kb_avail":20940412,"statfs":{"total":21470642176,"available":21442981888,"internally_reserved":0,"allocated":172032,"data_stored":51961,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":1590,"internal_metadata":27457994},"hb_peers":[0,2],"snap_trim_queue_len":0,"num_snap_trimming":0,"num_shards_repaired":0,"op_queue_age_hist":{"histogram":[],"upper_bound":1},"perf_stat":{"commit_latency_ms":0,"apply_latency_ms":0,"commit_latency_ns":0,"apply_latency_ns":0},"alerts":[]},{"osd":0,"up_from":8,"seq":34359738376,"num_pgs":1,"num_osds":1,"num_per_pool_osds":1,"num_per_pool_omap_osds":1,"kb":20967424,"kb_used":27592,"kb_used_data":628,"kb_used_omap":1,"kb_used_meta":26814,"kb_avail":20939832,"statfs":{"total":21470642176,"available":21442387968,"internally_reserved":0,"allocated":643072,"data_stored":515018,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":1590,"internal_metadata":27457994},"hb_peers":[1,2],"snap_trim_queue_len":0,"num_snap_trimming":0,"num_shards_repaired":0,"op_queue_age_hist":{"histogram":[],"upper_bound":1},"perf_stat":{"commit_latency_ms":1,"apply_latency_ms":1,"commit_latency_ns":1000000,"apply_latency_ns":1000000},"alerts":[]}],"pool_statfs":[{"poolid":1,"osd":0,"total":0,"available":0,"internally_reserved":0,"allocated":462848,"data_stored":459280,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":0},{"poolid":1,"osd":1,"total":0,"available":0,"internally_reserved":0,"allocated":331776,"data_stored":328192,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":0},{"poolid":1,"osd":2,"total":0,"available":0,"internally_reserved":0,"allocated":462848,"data_stored":459280,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":0}]}} 2026-03-09T20:53:25.381 INFO:tasks.cephadm.ceph_manager.ceph:clean! 2026-03-09T20:53:25.381 INFO:tasks.ceph:Waiting until ceph cluster ceph is healthy... 2026-03-09T20:53:25.381 INFO:tasks.cephadm.ceph_manager.ceph:wait_until_healthy 2026-03-09T20:53:25.382 DEBUG:teuthology.orchestra.run.vm01:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell --fsid 9cb345a8-1bf9-11f1-a195-0375563c5891 -- ceph health --format=json 2026-03-09T20:53:25.551 INFO:teuthology.orchestra.run.vm01.stderr:Inferring config /var/lib/ceph/9cb345a8-1bf9-11f1-a195-0375563c5891/mon.a/config 2026-03-09T20:53:25.805 INFO:teuthology.orchestra.run.vm01.stdout: 2026-03-09T20:53:25.805 INFO:teuthology.orchestra.run.vm01.stdout:{"status":"HEALTH_OK","checks":{},"mutes":[]} 2026-03-09T20:53:25.923 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:53:25 vm01 ceph-mon[53402]: from='client.14400 -' entity='client.admin' cmd=[{"prefix": "pg dump", "target": ["mon-mgr", ""], "format": "json"}]: dispatch 2026-03-09T20:53:25.923 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:53:25 vm01 ceph-mon[53402]: from='client.24245 -' entity='client.admin' cmd=[{"prefix": "pg dump", "target": ["mon-mgr", ""], "format": "json"}]: dispatch 2026-03-09T20:53:25.923 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:53:25 vm01 ceph-mon[53402]: pgmap v50: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail; 2.2 KiB/s rd, 65 KiB/s wr, 6 op/s 2026-03-09T20:53:25.955 INFO:tasks.cephadm.ceph_manager.ceph:wait_until_healthy done 2026-03-09T20:53:25.955 INFO:tasks.cephadm:Setup complete, yielding 2026-03-09T20:53:25.955 INFO:teuthology.run_tasks:Running task cephadm.apply... 2026-03-09T20:53:25.957 INFO:tasks.cephadm:Applying spec(s): placement: count: 3 service_id: foo service_type: mon spec: crush_locations: host.a: - datacenter=a host.b: - datacenter=b - rack=2 host.c: - datacenter=a - rack=3 2026-03-09T20:53:25.958 DEBUG:teuthology.orchestra.run.vm01:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 9cb345a8-1bf9-11f1-a195-0375563c5891 -- ceph orch apply -i - 2026-03-09T20:53:25.958 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:53:25 vm08 ceph-mon[54611]: from='client.14400 -' entity='client.admin' cmd=[{"prefix": "pg dump", "target": ["mon-mgr", ""], "format": "json"}]: dispatch 2026-03-09T20:53:25.958 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:53:25 vm08 ceph-mon[54611]: from='client.24245 -' entity='client.admin' cmd=[{"prefix": "pg dump", "target": ["mon-mgr", ""], "format": "json"}]: dispatch 2026-03-09T20:53:25.958 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:53:25 vm08 ceph-mon[54611]: pgmap v50: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail; 2.2 KiB/s rd, 65 KiB/s wr, 6 op/s 2026-03-09T20:53:26.046 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 09 20:53:25 vm06 ceph-mon[52939]: from='client.14400 -' entity='client.admin' cmd=[{"prefix": "pg dump", "target": ["mon-mgr", ""], "format": "json"}]: dispatch 2026-03-09T20:53:26.046 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 09 20:53:25 vm06 ceph-mon[52939]: from='client.24245 -' entity='client.admin' cmd=[{"prefix": "pg dump", "target": ["mon-mgr", ""], "format": "json"}]: dispatch 2026-03-09T20:53:26.046 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 09 20:53:25 vm06 ceph-mon[52939]: pgmap v50: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail; 2.2 KiB/s rd, 65 KiB/s wr, 6 op/s 2026-03-09T20:53:26.127 INFO:teuthology.orchestra.run.vm01.stderr:Inferring config /var/lib/ceph/9cb345a8-1bf9-11f1-a195-0375563c5891/mon.a/config 2026-03-09T20:53:26.375 INFO:teuthology.orchestra.run.vm01.stdout:Scheduled mon update... 2026-03-09T20:53:26.539 INFO:teuthology.run_tasks:Running task cephadm.shell... 2026-03-09T20:53:26.542 INFO:tasks.cephadm:Running commands on role host.a host ubuntu@vm01.local 2026-03-09T20:53:26.542 DEBUG:teuthology.orchestra.run.vm01:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 9cb345a8-1bf9-11f1-a195-0375563c5891 -- bash -c 'set -ex 2026-03-09T20:53:26.542 DEBUG:teuthology.orchestra.run.vm01:> # since we don'"'"'t know the real hostnames before the test, the next 2026-03-09T20:53:26.542 DEBUG:teuthology.orchestra.run.vm01:> # bit is in order to replace the fake hostnames "host.a/b/c" with 2026-03-09T20:53:26.542 DEBUG:teuthology.orchestra.run.vm01:> # the actual names cephadm knows the host by within the mon spec 2026-03-09T20:53:26.542 DEBUG:teuthology.orchestra.run.vm01:> ceph orch host ls --format json | jq -r '"'"'.[] | .hostname'"'"' > realnames 2026-03-09T20:53:26.542 DEBUG:teuthology.orchestra.run.vm01:> echo $'"'"'host.a\nhost.b\nhost.c'"'"' > fakenames 2026-03-09T20:53:26.542 DEBUG:teuthology.orchestra.run.vm01:> echo $'"'"'a\nb\nc'"'"' > mon_ids 2026-03-09T20:53:26.542 DEBUG:teuthology.orchestra.run.vm01:> echo $'"'"'{datacenter=a}\n{datacenter=b,rack=2}\n{datacenter=a,rack=3}'"'"' > crush_locs 2026-03-09T20:53:26.542 DEBUG:teuthology.orchestra.run.vm01:> ceph orch ls --service-name mon --export > mon.yaml 2026-03-09T20:53:26.542 DEBUG:teuthology.orchestra.run.vm01:> MONSPEC=`cat mon.yaml` 2026-03-09T20:53:26.542 DEBUG:teuthology.orchestra.run.vm01:> echo "$MONSPEC" 2026-03-09T20:53:26.542 DEBUG:teuthology.orchestra.run.vm01:> while read realname <&3 && read fakename <&4; do 2026-03-09T20:53:26.542 DEBUG:teuthology.orchestra.run.vm01:> MONSPEC="${MONSPEC//$fakename/$realname}" 2026-03-09T20:53:26.542 DEBUG:teuthology.orchestra.run.vm01:> done 3 echo "$MONSPEC" > mon.yaml 2026-03-09T20:53:26.542 DEBUG:teuthology.orchestra.run.vm01:> cat mon.yaml 2026-03-09T20:53:26.542 DEBUG:teuthology.orchestra.run.vm01:> # now the spec should have the real hostnames, so let'"'"'s re-apply 2026-03-09T20:53:26.542 DEBUG:teuthology.orchestra.run.vm01:> ceph orch apply -i mon.yaml 2026-03-09T20:53:26.542 DEBUG:teuthology.orchestra.run.vm01:> sleep 90 2026-03-09T20:53:26.542 DEBUG:teuthology.orchestra.run.vm01:> ceph orch ps --refresh 2026-03-09T20:53:26.542 DEBUG:teuthology.orchestra.run.vm01:> ceph orch ls --service-name mon --export > mon.yaml; ceph orch apply -i mon.yaml 2026-03-09T20:53:26.542 DEBUG:teuthology.orchestra.run.vm01:> sleep 90 2026-03-09T20:53:26.542 DEBUG:teuthology.orchestra.run.vm01:> ceph mon dump 2026-03-09T20:53:26.542 DEBUG:teuthology.orchestra.run.vm01:> ceph mon dump --format json 2026-03-09T20:53:26.542 DEBUG:teuthology.orchestra.run.vm01:> # verify all the crush locations got set from "ceph mon dump" output 2026-03-09T20:53:26.542 DEBUG:teuthology.orchestra.run.vm01:> while read monid <&3 && read crushloc <&4; do 2026-03-09T20:53:26.542 DEBUG:teuthology.orchestra.run.vm01:> ceph mon dump --format json | jq --arg monid "$monid" --arg crushloc "$crushloc" -e '"'"'.mons | .[] | select(.name == $monid) | .crush_location == $crushloc'"'"' 2026-03-09T20:53:26.542 DEBUG:teuthology.orchestra.run.vm01:> done 3 ' 2026-03-09T20:53:26.750 INFO:teuthology.orchestra.run.vm01.stderr:Inferring config /var/lib/ceph/9cb345a8-1bf9-11f1-a195-0375563c5891/mon.a/config 2026-03-09T20:53:26.775 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:53:26 vm01 ceph-mon[53402]: from='client.? 192.168.123.101:0/3372928475' entity='client.admin' cmd=[{"prefix": "health", "format": "json"}]: dispatch 2026-03-09T20:53:26.775 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:53:26 vm01 ceph-mon[53402]: from='client.24257 -' entity='client.admin' cmd=[{"prefix": "orch apply", "target": ["mon-mgr", ""]}]: dispatch 2026-03-09T20:53:26.775 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:53:26 vm01 ceph-mon[53402]: Saving service mon spec with placement count:3 2026-03-09T20:53:26.775 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:53:26 vm01 ceph-mon[53402]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' 2026-03-09T20:53:26.775 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:53:26 vm01 ceph-mon[53402]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T20:53:26.844 INFO:teuthology.orchestra.run.vm01.stderr:+ ceph orch host ls --format json 2026-03-09T20:53:26.849 INFO:teuthology.orchestra.run.vm01.stderr:+ jq -r '.[] | .hostname' 2026-03-09T20:53:26.957 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:53:26 vm08 ceph-mon[54611]: from='client.? 192.168.123.101:0/3372928475' entity='client.admin' cmd=[{"prefix": "health", "format": "json"}]: dispatch 2026-03-09T20:53:26.957 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:53:26 vm08 ceph-mon[54611]: from='client.24257 -' entity='client.admin' cmd=[{"prefix": "orch apply", "target": ["mon-mgr", ""]}]: dispatch 2026-03-09T20:53:26.957 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:53:26 vm08 ceph-mon[54611]: Saving service mon spec with placement count:3 2026-03-09T20:53:26.957 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:53:26 vm08 ceph-mon[54611]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' 2026-03-09T20:53:26.957 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:53:26 vm08 ceph-mon[54611]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T20:53:27.020 INFO:teuthology.orchestra.run.vm01.stderr:+ echo 'host.a 2026-03-09T20:53:27.020 INFO:teuthology.orchestra.run.vm01.stderr:host.b 2026-03-09T20:53:27.020 INFO:teuthology.orchestra.run.vm01.stderr:host.c' 2026-03-09T20:53:27.020 INFO:teuthology.orchestra.run.vm01.stderr:+ echo 'a 2026-03-09T20:53:27.020 INFO:teuthology.orchestra.run.vm01.stderr:b 2026-03-09T20:53:27.020 INFO:teuthology.orchestra.run.vm01.stderr:c' 2026-03-09T20:53:27.020 INFO:teuthology.orchestra.run.vm01.stderr:+ echo '{datacenter=a} 2026-03-09T20:53:27.021 INFO:teuthology.orchestra.run.vm01.stderr:{datacenter=b,rack=2} 2026-03-09T20:53:27.021 INFO:teuthology.orchestra.run.vm01.stderr:{datacenter=a,rack=3}' 2026-03-09T20:53:27.021 INFO:teuthology.orchestra.run.vm01.stderr:+ ceph orch ls --service-name mon --export 2026-03-09T20:53:27.045 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 09 20:53:26 vm06 ceph-mon[52939]: from='client.? 192.168.123.101:0/3372928475' entity='client.admin' cmd=[{"prefix": "health", "format": "json"}]: dispatch 2026-03-09T20:53:27.046 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 09 20:53:26 vm06 ceph-mon[52939]: from='client.24257 -' entity='client.admin' cmd=[{"prefix": "orch apply", "target": ["mon-mgr", ""]}]: dispatch 2026-03-09T20:53:27.046 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 09 20:53:26 vm06 ceph-mon[52939]: Saving service mon spec with placement count:3 2026-03-09T20:53:27.046 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 09 20:53:26 vm06 ceph-mon[52939]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' 2026-03-09T20:53:27.046 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 09 20:53:26 vm06 ceph-mon[52939]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T20:53:27.191 INFO:teuthology.orchestra.run.vm01.stderr:++ cat mon.yaml 2026-03-09T20:53:27.192 INFO:teuthology.orchestra.run.vm01.stderr:+ MONSPEC='service_type: mon 2026-03-09T20:53:27.192 INFO:teuthology.orchestra.run.vm01.stderr:service_name: mon 2026-03-09T20:53:27.192 INFO:teuthology.orchestra.run.vm01.stderr:placement: 2026-03-09T20:53:27.192 INFO:teuthology.orchestra.run.vm01.stderr: count: 3 2026-03-09T20:53:27.192 INFO:teuthology.orchestra.run.vm01.stderr:spec: 2026-03-09T20:53:27.192 INFO:teuthology.orchestra.run.vm01.stderr: crush_locations: 2026-03-09T20:53:27.192 INFO:teuthology.orchestra.run.vm01.stderr: host.a: 2026-03-09T20:53:27.192 INFO:teuthology.orchestra.run.vm01.stderr: - datacenter=a 2026-03-09T20:53:27.193 INFO:teuthology.orchestra.run.vm01.stderr: host.b: 2026-03-09T20:53:27.193 INFO:teuthology.orchestra.run.vm01.stderr: - datacenter=b 2026-03-09T20:53:27.193 INFO:teuthology.orchestra.run.vm01.stderr: - rack=2 2026-03-09T20:53:27.193 INFO:teuthology.orchestra.run.vm01.stderr: host.c: 2026-03-09T20:53:27.193 INFO:teuthology.orchestra.run.vm01.stderr: - datacenter=a 2026-03-09T20:53:27.193 INFO:teuthology.orchestra.run.vm01.stderr: - rack=3' 2026-03-09T20:53:27.193 INFO:teuthology.orchestra.run.vm01.stdout:service_type: mon 2026-03-09T20:53:27.193 INFO:teuthology.orchestra.run.vm01.stdout:service_name: mon 2026-03-09T20:53:27.193 INFO:teuthology.orchestra.run.vm01.stdout:placement: 2026-03-09T20:53:27.193 INFO:teuthology.orchestra.run.vm01.stdout: count: 3 2026-03-09T20:53:27.193 INFO:teuthology.orchestra.run.vm01.stdout:spec: 2026-03-09T20:53:27.193 INFO:teuthology.orchestra.run.vm01.stdout: crush_locations: 2026-03-09T20:53:27.193 INFO:teuthology.orchestra.run.vm01.stdout: host.a: 2026-03-09T20:53:27.193 INFO:teuthology.orchestra.run.vm01.stdout: - datacenter=a 2026-03-09T20:53:27.193 INFO:teuthology.orchestra.run.vm01.stdout: host.b: 2026-03-09T20:53:27.193 INFO:teuthology.orchestra.run.vm01.stdout: - datacenter=b 2026-03-09T20:53:27.193 INFO:teuthology.orchestra.run.vm01.stdout: - rack=2 2026-03-09T20:53:27.193 INFO:teuthology.orchestra.run.vm01.stdout: host.c: 2026-03-09T20:53:27.193 INFO:teuthology.orchestra.run.vm01.stdout: - datacenter=a 2026-03-09T20:53:27.193 INFO:teuthology.orchestra.run.vm01.stderr:+ echo 'service_type: mon 2026-03-09T20:53:27.193 INFO:teuthology.orchestra.run.vm01.stderr:service_name: mon 2026-03-09T20:53:27.193 INFO:teuthology.orchestra.run.vm01.stderr:placement: 2026-03-09T20:53:27.193 INFO:teuthology.orchestra.run.vm01.stderr: count: 3 2026-03-09T20:53:27.193 INFO:teuthology.orchestra.run.vm01.stderr:spec: 2026-03-09T20:53:27.193 INFO:teuthology.orchestra.run.vm01.stderr: crush_locations: 2026-03-09T20:53:27.193 INFO:teuthology.orchestra.run.vm01.stderr: host.a: 2026-03-09T20:53:27.193 INFO:teuthology.orchestra.run.vm01.stderr: - datacenter=a 2026-03-09T20:53:27.193 INFO:teuthology.orchestra.run.vm01.stderr: host.b: 2026-03-09T20:53:27.193 INFO:teuthology.orchestra.run.vm01.stderr: - datacenter=b 2026-03-09T20:53:27.193 INFO:teuthology.orchestra.run.vm01.stderr: - rack=2 2026-03-09T20:53:27.193 INFO:teuthology.orchestra.run.vm01.stderr: host.c: 2026-03-09T20:53:27.193 INFO:teuthology.orchestra.run.vm01.stderr: - datacenter=a 2026-03-09T20:53:27.193 INFO:teuthology.orchestra.run.vm01.stderr: - rack=3' 2026-03-09T20:53:27.193 INFO:teuthology.orchestra.run.vm01.stdout: - rack=3 2026-03-09T20:53:27.193 INFO:teuthology.orchestra.run.vm01.stderr:+ read realname 2026-03-09T20:53:27.193 INFO:teuthology.orchestra.run.vm01.stderr:+ read fakename 2026-03-09T20:53:27.193 INFO:teuthology.orchestra.run.vm01.stderr:+ MONSPEC='service_type: mon 2026-03-09T20:53:27.193 INFO:teuthology.orchestra.run.vm01.stderr:service_name: mon 2026-03-09T20:53:27.193 INFO:teuthology.orchestra.run.vm01.stderr:placement: 2026-03-09T20:53:27.193 INFO:teuthology.orchestra.run.vm01.stderr: count: 3 2026-03-09T20:53:27.193 INFO:teuthology.orchestra.run.vm01.stderr:spec: 2026-03-09T20:53:27.193 INFO:teuthology.orchestra.run.vm01.stderr: crush_locations: 2026-03-09T20:53:27.194 INFO:teuthology.orchestra.run.vm01.stderr: vm01: 2026-03-09T20:53:27.194 INFO:teuthology.orchestra.run.vm01.stderr: - datacenter=a 2026-03-09T20:53:27.194 INFO:teuthology.orchestra.run.vm01.stderr: host.b: 2026-03-09T20:53:27.194 INFO:teuthology.orchestra.run.vm01.stderr: - datacenter=b 2026-03-09T20:53:27.194 INFO:teuthology.orchestra.run.vm01.stderr: - rack=2 2026-03-09T20:53:27.194 INFO:teuthology.orchestra.run.vm01.stderr: host.c: 2026-03-09T20:53:27.194 INFO:teuthology.orchestra.run.vm01.stderr: - datacenter=a 2026-03-09T20:53:27.194 INFO:teuthology.orchestra.run.vm01.stderr: - rack=3' 2026-03-09T20:53:27.194 INFO:teuthology.orchestra.run.vm01.stderr:+ read realname 2026-03-09T20:53:27.194 INFO:teuthology.orchestra.run.vm01.stderr:+ read fakename 2026-03-09T20:53:27.194 INFO:teuthology.orchestra.run.vm01.stderr:+ MONSPEC='service_type: mon 2026-03-09T20:53:27.194 INFO:teuthology.orchestra.run.vm01.stderr:service_name: mon 2026-03-09T20:53:27.194 INFO:teuthology.orchestra.run.vm01.stderr:placement: 2026-03-09T20:53:27.194 INFO:teuthology.orchestra.run.vm01.stderr: count: 3 2026-03-09T20:53:27.194 INFO:teuthology.orchestra.run.vm01.stderr:spec: 2026-03-09T20:53:27.194 INFO:teuthology.orchestra.run.vm01.stderr: crush_locations: 2026-03-09T20:53:27.194 INFO:teuthology.orchestra.run.vm01.stderr: vm01: 2026-03-09T20:53:27.194 INFO:teuthology.orchestra.run.vm01.stderr: - datacenter=a 2026-03-09T20:53:27.194 INFO:teuthology.orchestra.run.vm01.stderr: vm06: 2026-03-09T20:53:27.194 INFO:teuthology.orchestra.run.vm01.stderr: - datacenter=b 2026-03-09T20:53:27.194 INFO:teuthology.orchestra.run.vm01.stderr: - rack=2 2026-03-09T20:53:27.194 INFO:teuthology.orchestra.run.vm01.stderr: host.c: 2026-03-09T20:53:27.194 INFO:teuthology.orchestra.run.vm01.stderr: - datacenter=a 2026-03-09T20:53:27.194 INFO:teuthology.orchestra.run.vm01.stderr: - rack=3' 2026-03-09T20:53:27.194 INFO:teuthology.orchestra.run.vm01.stderr:+ read realname 2026-03-09T20:53:27.194 INFO:teuthology.orchestra.run.vm01.stderr:+ read fakename 2026-03-09T20:53:27.194 INFO:teuthology.orchestra.run.vm01.stderr:+ MONSPEC='service_type: mon 2026-03-09T20:53:27.194 INFO:teuthology.orchestra.run.vm01.stderr:service_name: mon 2026-03-09T20:53:27.194 INFO:teuthology.orchestra.run.vm01.stderr:placement: 2026-03-09T20:53:27.194 INFO:teuthology.orchestra.run.vm01.stderr: count: 3 2026-03-09T20:53:27.194 INFO:teuthology.orchestra.run.vm01.stderr:spec: 2026-03-09T20:53:27.194 INFO:teuthology.orchestra.run.vm01.stderr: crush_locations: 2026-03-09T20:53:27.194 INFO:teuthology.orchestra.run.vm01.stderr: vm01: 2026-03-09T20:53:27.194 INFO:teuthology.orchestra.run.vm01.stderr: - datacenter=a 2026-03-09T20:53:27.194 INFO:teuthology.orchestra.run.vm01.stderr: vm06: 2026-03-09T20:53:27.194 INFO:teuthology.orchestra.run.vm01.stderr: - datacenter=b 2026-03-09T20:53:27.194 INFO:teuthology.orchestra.run.vm01.stderr: - rack=2 2026-03-09T20:53:27.194 INFO:teuthology.orchestra.run.vm01.stderr: vm08: 2026-03-09T20:53:27.194 INFO:teuthology.orchestra.run.vm01.stderr: - datacenter=a 2026-03-09T20:53:27.194 INFO:teuthology.orchestra.run.vm01.stderr: - rack=3' 2026-03-09T20:53:27.194 INFO:teuthology.orchestra.run.vm01.stderr:+ read realname 2026-03-09T20:53:27.194 INFO:teuthology.orchestra.run.vm01.stderr:+ echo 'service_type: mon 2026-03-09T20:53:27.194 INFO:teuthology.orchestra.run.vm01.stderr:service_name: mon 2026-03-09T20:53:27.194 INFO:teuthology.orchestra.run.vm01.stderr:placement: 2026-03-09T20:53:27.194 INFO:teuthology.orchestra.run.vm01.stderr: count: 3 2026-03-09T20:53:27.194 INFO:teuthology.orchestra.run.vm01.stderr:spec: 2026-03-09T20:53:27.194 INFO:teuthology.orchestra.run.vm01.stderr: crush_locations: 2026-03-09T20:53:27.194 INFO:teuthology.orchestra.run.vm01.stderr: vm01: 2026-03-09T20:53:27.194 INFO:teuthology.orchestra.run.vm01.stderr: - datacenter=a 2026-03-09T20:53:27.194 INFO:teuthology.orchestra.run.vm01.stderr: vm06: 2026-03-09T20:53:27.194 INFO:teuthology.orchestra.run.vm01.stderr: - datacenter=b 2026-03-09T20:53:27.195 INFO:teuthology.orchestra.run.vm01.stderr: - rack=2 2026-03-09T20:53:27.195 INFO:teuthology.orchestra.run.vm01.stderr: vm08: 2026-03-09T20:53:27.195 INFO:teuthology.orchestra.run.vm01.stderr: - datacenter=a 2026-03-09T20:53:27.195 INFO:teuthology.orchestra.run.vm01.stderr: - rack=3' 2026-03-09T20:53:27.195 INFO:teuthology.orchestra.run.vm01.stderr:+ cat mon.yaml 2026-03-09T20:53:27.195 INFO:teuthology.orchestra.run.vm01.stdout:service_type: mon 2026-03-09T20:53:27.195 INFO:teuthology.orchestra.run.vm01.stdout:service_name: mon 2026-03-09T20:53:27.195 INFO:teuthology.orchestra.run.vm01.stdout:placement: 2026-03-09T20:53:27.195 INFO:teuthology.orchestra.run.vm01.stdout: count: 3 2026-03-09T20:53:27.195 INFO:teuthology.orchestra.run.vm01.stdout:spec: 2026-03-09T20:53:27.195 INFO:teuthology.orchestra.run.vm01.stdout: crush_locations: 2026-03-09T20:53:27.195 INFO:teuthology.orchestra.run.vm01.stdout: vm01: 2026-03-09T20:53:27.195 INFO:teuthology.orchestra.run.vm01.stdout: - datacenter=a 2026-03-09T20:53:27.195 INFO:teuthology.orchestra.run.vm01.stdout: vm06: 2026-03-09T20:53:27.195 INFO:teuthology.orchestra.run.vm01.stdout: - datacenter=b 2026-03-09T20:53:27.195 INFO:teuthology.orchestra.run.vm01.stdout: - rack=2 2026-03-09T20:53:27.195 INFO:teuthology.orchestra.run.vm01.stdout: vm08: 2026-03-09T20:53:27.195 INFO:teuthology.orchestra.run.vm01.stdout: - datacenter=a 2026-03-09T20:53:27.195 INFO:teuthology.orchestra.run.vm01.stdout: - rack=3 2026-03-09T20:53:27.195 INFO:teuthology.orchestra.run.vm01.stderr:+ ceph orch apply -i mon.yaml 2026-03-09T20:53:27.360 INFO:teuthology.orchestra.run.vm01.stdout:Scheduled mon update... 2026-03-09T20:53:27.381 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 90 2026-03-09T20:53:32.731 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:53:32 vm01 ceph-mon[53402]: from='client.14424 -' entity='client.admin' cmd=[{"prefix": "orch host ls", "target": ["mon-mgr", ""], "format": "json"}]: dispatch 2026-03-09T20:53:32.731 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:53:32 vm01 ceph-mon[53402]: from='client.14430 -' entity='client.admin' cmd=[{"prefix": "orch ls", "service_name": "mon", "export": true, "target": ["mon-mgr", ""]}]: dispatch 2026-03-09T20:53:32.731 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:53:32 vm01 ceph-mon[53402]: from='client.14436 -' entity='client.admin' cmd=[{"prefix": "orch apply", "target": ["mon-mgr", ""]}]: dispatch 2026-03-09T20:53:32.731 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:53:32 vm01 ceph-mon[53402]: Saving service mon spec with placement count:3 2026-03-09T20:53:32.731 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:53:32 vm01 ceph-mon[53402]: Setting crush location for mon a to {datacenter=a} 2026-03-09T20:53:32.731 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:53:32 vm01 ceph-mon[53402]: Setting crush location for mon b to {datacenter=b,rack=2} 2026-03-09T20:53:32.731 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:53:32 vm01 ceph-mon[53402]: Setting crush location for mon c to {datacenter=a,rack=3} 2026-03-09T20:53:32.731 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:53:32 vm01 ceph-mon[53402]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' cmd='[{"prefix": "mon set_location", "name": "c", "args": ["datacenter=a", "rack=3"]}]': finished 2026-03-09T20:53:32.732 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:53:32 vm01 ceph-mon[53402]: pgmap v51: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail; 1.5 KiB/s rd, 44 KiB/s wr, 4 op/s 2026-03-09T20:53:32.732 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:53:32 vm01 ceph-mon[53402]: mon.b calling monitor election 2026-03-09T20:53:32.732 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:53:32 vm01 ceph-mon[53402]: mon.a calling monitor election 2026-03-09T20:53:32.732 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:53:32 vm01 ceph-mon[53402]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "a"}]: dispatch 2026-03-09T20:53:32.732 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:53:32 vm01 ceph-mon[53402]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-09T20:53:32.732 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:53:32 vm01 ceph-mon[53402]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "c"}]: dispatch 2026-03-09T20:53:32.732 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:53:32 vm01 ceph-mon[53402]: mon.c calling monitor election 2026-03-09T20:53:32.732 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:53:32 vm01 ceph-mon[53402]: pgmap v52: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail; 1.1 KiB/s rd, 33 KiB/s wr, 3 op/s 2026-03-09T20:53:32.732 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:53:32 vm01 ceph-mon[53402]: pgmap v53: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail; 921 B/s rd, 26 KiB/s wr, 2 op/s 2026-03-09T20:53:32.732 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:53:32 vm01 ceph-mon[53402]: mon.a is new leader, mons a,b in quorum (ranks 0,2) 2026-03-09T20:53:32.732 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:53:32 vm01 ceph-mon[53402]: overall HEALTH_OK 2026-03-09T20:53:32.732 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:53:32 vm01 ceph-mon[53402]: mon.a calling monitor election 2026-03-09T20:53:32.732 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:53:32 vm01 ceph-mon[53402]: mon.a is new leader, mons a,c,b in quorum (ranks 0,1,2) 2026-03-09T20:53:32.732 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:53:32 vm01 ceph-mon[53402]: monmap epoch 6 2026-03-09T20:53:32.732 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:53:32 vm01 ceph-mon[53402]: fsid 9cb345a8-1bf9-11f1-a195-0375563c5891 2026-03-09T20:53:32.732 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:53:32 vm01 ceph-mon[53402]: last_changed 2026-03-09T20:53:27.400224+0000 2026-03-09T20:53:32.732 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:53:32 vm01 ceph-mon[53402]: created 2026-03-09T20:51:41.312920+0000 2026-03-09T20:53:32.732 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:53:32 vm01 ceph-mon[53402]: min_mon_release 19 (squid) 2026-03-09T20:53:32.732 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:53:32 vm01 ceph-mon[53402]: election_strategy: 1 2026-03-09T20:53:32.732 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:53:32 vm01 ceph-mon[53402]: 0: [v2:192.168.123.101:3300/0,v1:192.168.123.101:6789/0] mon.a; crush_location {datacenter=a} 2026-03-09T20:53:32.732 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:53:32 vm01 ceph-mon[53402]: 1: [v2:192.168.123.108:3300/0,v1:192.168.123.108:6789/0] mon.c; crush_location {datacenter=a,rack=3} 2026-03-09T20:53:32.732 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:53:32 vm01 ceph-mon[53402]: 2: [v2:192.168.123.106:3300/0,v1:192.168.123.106:6789/0] mon.b; crush_location {datacenter=b,rack=2} 2026-03-09T20:53:32.732 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:53:32 vm01 ceph-mon[53402]: fsmap 2026-03-09T20:53:32.732 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:53:32 vm01 ceph-mon[53402]: osdmap e21: 3 total, 3 up, 3 in 2026-03-09T20:53:32.732 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:53:32 vm01 ceph-mon[53402]: mgrmap e15: a(active, since 85s), standbys: b 2026-03-09T20:53:32.732 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:53:32 vm01 ceph-mon[53402]: overall HEALTH_OK 2026-03-09T20:53:32.732 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:53:32 vm01 ceph-mon[53402]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' 2026-03-09T20:53:32.796 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 09 20:53:32 vm06 ceph-mon[52939]: from='client.14424 -' entity='client.admin' cmd=[{"prefix": "orch host ls", "target": ["mon-mgr", ""], "format": "json"}]: dispatch 2026-03-09T20:53:32.796 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 09 20:53:32 vm06 ceph-mon[52939]: from='client.14430 -' entity='client.admin' cmd=[{"prefix": "orch ls", "service_name": "mon", "export": true, "target": ["mon-mgr", ""]}]: dispatch 2026-03-09T20:53:32.796 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 09 20:53:32 vm06 ceph-mon[52939]: from='client.14436 -' entity='client.admin' cmd=[{"prefix": "orch apply", "target": ["mon-mgr", ""]}]: dispatch 2026-03-09T20:53:32.796 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 09 20:53:32 vm06 ceph-mon[52939]: Saving service mon spec with placement count:3 2026-03-09T20:53:32.796 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 09 20:53:32 vm06 ceph-mon[52939]: Setting crush location for mon a to {datacenter=a} 2026-03-09T20:53:32.796 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 09 20:53:32 vm06 ceph-mon[52939]: Setting crush location for mon b to {datacenter=b,rack=2} 2026-03-09T20:53:32.796 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 09 20:53:32 vm06 ceph-mon[52939]: Setting crush location for mon c to {datacenter=a,rack=3} 2026-03-09T20:53:32.796 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 09 20:53:32 vm06 ceph-mon[52939]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' cmd='[{"prefix": "mon set_location", "name": "c", "args": ["datacenter=a", "rack=3"]}]': finished 2026-03-09T20:53:32.796 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 09 20:53:32 vm06 ceph-mon[52939]: pgmap v51: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail; 1.5 KiB/s rd, 44 KiB/s wr, 4 op/s 2026-03-09T20:53:32.796 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 09 20:53:32 vm06 ceph-mon[52939]: mon.b calling monitor election 2026-03-09T20:53:32.796 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 09 20:53:32 vm06 ceph-mon[52939]: mon.a calling monitor election 2026-03-09T20:53:32.796 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 09 20:53:32 vm06 ceph-mon[52939]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "a"}]: dispatch 2026-03-09T20:53:32.796 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 09 20:53:32 vm06 ceph-mon[52939]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-09T20:53:32.796 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 09 20:53:32 vm06 ceph-mon[52939]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "c"}]: dispatch 2026-03-09T20:53:32.796 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 09 20:53:32 vm06 ceph-mon[52939]: mon.c calling monitor election 2026-03-09T20:53:32.796 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 09 20:53:32 vm06 ceph-mon[52939]: pgmap v52: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail; 1.1 KiB/s rd, 33 KiB/s wr, 3 op/s 2026-03-09T20:53:32.796 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 09 20:53:32 vm06 ceph-mon[52939]: pgmap v53: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail; 921 B/s rd, 26 KiB/s wr, 2 op/s 2026-03-09T20:53:32.796 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 09 20:53:32 vm06 ceph-mon[52939]: mon.a is new leader, mons a,b in quorum (ranks 0,2) 2026-03-09T20:53:32.796 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 09 20:53:32 vm06 ceph-mon[52939]: overall HEALTH_OK 2026-03-09T20:53:32.796 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 09 20:53:32 vm06 ceph-mon[52939]: mon.a calling monitor election 2026-03-09T20:53:32.796 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 09 20:53:32 vm06 ceph-mon[52939]: mon.a is new leader, mons a,c,b in quorum (ranks 0,1,2) 2026-03-09T20:53:32.796 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 09 20:53:32 vm06 ceph-mon[52939]: monmap epoch 6 2026-03-09T20:53:32.796 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 09 20:53:32 vm06 ceph-mon[52939]: fsid 9cb345a8-1bf9-11f1-a195-0375563c5891 2026-03-09T20:53:32.796 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 09 20:53:32 vm06 ceph-mon[52939]: last_changed 2026-03-09T20:53:27.400224+0000 2026-03-09T20:53:32.796 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 09 20:53:32 vm06 ceph-mon[52939]: created 2026-03-09T20:51:41.312920+0000 2026-03-09T20:53:32.796 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 09 20:53:32 vm06 ceph-mon[52939]: min_mon_release 19 (squid) 2026-03-09T20:53:32.796 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 09 20:53:32 vm06 ceph-mon[52939]: election_strategy: 1 2026-03-09T20:53:32.796 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 09 20:53:32 vm06 ceph-mon[52939]: 0: [v2:192.168.123.101:3300/0,v1:192.168.123.101:6789/0] mon.a; crush_location {datacenter=a} 2026-03-09T20:53:32.796 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 09 20:53:32 vm06 ceph-mon[52939]: 1: [v2:192.168.123.108:3300/0,v1:192.168.123.108:6789/0] mon.c; crush_location {datacenter=a,rack=3} 2026-03-09T20:53:32.796 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 09 20:53:32 vm06 ceph-mon[52939]: 2: [v2:192.168.123.106:3300/0,v1:192.168.123.106:6789/0] mon.b; crush_location {datacenter=b,rack=2} 2026-03-09T20:53:32.796 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 09 20:53:32 vm06 ceph-mon[52939]: fsmap 2026-03-09T20:53:32.796 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 09 20:53:32 vm06 ceph-mon[52939]: osdmap e21: 3 total, 3 up, 3 in 2026-03-09T20:53:32.796 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 09 20:53:32 vm06 ceph-mon[52939]: mgrmap e15: a(active, since 85s), standbys: b 2026-03-09T20:53:32.796 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 09 20:53:32 vm06 ceph-mon[52939]: overall HEALTH_OK 2026-03-09T20:53:32.796 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 09 20:53:32 vm06 ceph-mon[52939]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' 2026-03-09T20:53:32.957 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:53:32 vm08 ceph-mon[54611]: from='client.14424 -' entity='client.admin' cmd=[{"prefix": "orch host ls", "target": ["mon-mgr", ""], "format": "json"}]: dispatch 2026-03-09T20:53:32.957 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:53:32 vm08 ceph-mon[54611]: from='client.14430 -' entity='client.admin' cmd=[{"prefix": "orch ls", "service_name": "mon", "export": true, "target": ["mon-mgr", ""]}]: dispatch 2026-03-09T20:53:32.957 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:53:32 vm08 ceph-mon[54611]: from='client.14436 -' entity='client.admin' cmd=[{"prefix": "orch apply", "target": ["mon-mgr", ""]}]: dispatch 2026-03-09T20:53:32.957 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:53:32 vm08 ceph-mon[54611]: Saving service mon spec with placement count:3 2026-03-09T20:53:32.957 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:53:32 vm08 ceph-mon[54611]: Setting crush location for mon a to {datacenter=a} 2026-03-09T20:53:32.957 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:53:32 vm08 ceph-mon[54611]: Setting crush location for mon b to {datacenter=b,rack=2} 2026-03-09T20:53:32.957 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:53:32 vm08 ceph-mon[54611]: Setting crush location for mon c to {datacenter=a,rack=3} 2026-03-09T20:53:32.957 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:53:32 vm08 ceph-mon[54611]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' cmd='[{"prefix": "mon set_location", "name": "c", "args": ["datacenter=a", "rack=3"]}]': finished 2026-03-09T20:53:32.957 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:53:32 vm08 ceph-mon[54611]: pgmap v51: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail; 1.5 KiB/s rd, 44 KiB/s wr, 4 op/s 2026-03-09T20:53:32.958 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:53:32 vm08 ceph-mon[54611]: mon.b calling monitor election 2026-03-09T20:53:32.958 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:53:32 vm08 ceph-mon[54611]: mon.a calling monitor election 2026-03-09T20:53:32.958 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:53:32 vm08 ceph-mon[54611]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "a"}]: dispatch 2026-03-09T20:53:32.958 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:53:32 vm08 ceph-mon[54611]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-09T20:53:32.958 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:53:32 vm08 ceph-mon[54611]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "c"}]: dispatch 2026-03-09T20:53:32.958 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:53:32 vm08 ceph-mon[54611]: mon.c calling monitor election 2026-03-09T20:53:32.958 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:53:32 vm08 ceph-mon[54611]: pgmap v52: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail; 1.1 KiB/s rd, 33 KiB/s wr, 3 op/s 2026-03-09T20:53:32.958 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:53:32 vm08 ceph-mon[54611]: pgmap v53: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail; 921 B/s rd, 26 KiB/s wr, 2 op/s 2026-03-09T20:53:32.958 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:53:32 vm08 ceph-mon[54611]: mon.a is new leader, mons a,b in quorum (ranks 0,2) 2026-03-09T20:53:32.958 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:53:32 vm08 ceph-mon[54611]: overall HEALTH_OK 2026-03-09T20:53:32.958 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:53:32 vm08 ceph-mon[54611]: mon.a calling monitor election 2026-03-09T20:53:32.958 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:53:32 vm08 ceph-mon[54611]: mon.a is new leader, mons a,c,b in quorum (ranks 0,1,2) 2026-03-09T20:53:32.958 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:53:32 vm08 ceph-mon[54611]: monmap epoch 6 2026-03-09T20:53:32.958 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:53:32 vm08 ceph-mon[54611]: fsid 9cb345a8-1bf9-11f1-a195-0375563c5891 2026-03-09T20:53:32.958 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:53:32 vm08 ceph-mon[54611]: last_changed 2026-03-09T20:53:27.400224+0000 2026-03-09T20:53:32.958 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:53:32 vm08 ceph-mon[54611]: created 2026-03-09T20:51:41.312920+0000 2026-03-09T20:53:32.958 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:53:32 vm08 ceph-mon[54611]: min_mon_release 19 (squid) 2026-03-09T20:53:32.958 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:53:32 vm08 ceph-mon[54611]: election_strategy: 1 2026-03-09T20:53:32.958 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:53:32 vm08 ceph-mon[54611]: 0: [v2:192.168.123.101:3300/0,v1:192.168.123.101:6789/0] mon.a; crush_location {datacenter=a} 2026-03-09T20:53:32.958 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:53:32 vm08 ceph-mon[54611]: 1: [v2:192.168.123.108:3300/0,v1:192.168.123.108:6789/0] mon.c; crush_location {datacenter=a,rack=3} 2026-03-09T20:53:32.958 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:53:32 vm08 ceph-mon[54611]: 2: [v2:192.168.123.106:3300/0,v1:192.168.123.106:6789/0] mon.b; crush_location {datacenter=b,rack=2} 2026-03-09T20:53:32.958 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:53:32 vm08 ceph-mon[54611]: fsmap 2026-03-09T20:53:32.958 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:53:32 vm08 ceph-mon[54611]: osdmap e21: 3 total, 3 up, 3 in 2026-03-09T20:53:32.958 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:53:32 vm08 ceph-mon[54611]: mgrmap e15: a(active, since 85s), standbys: b 2026-03-09T20:53:32.958 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:53:32 vm08 ceph-mon[54611]: overall HEALTH_OK 2026-03-09T20:53:32.958 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:53:32 vm08 ceph-mon[54611]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' 2026-03-09T20:53:33.796 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 09 20:53:33 vm06 ceph-mon[52939]: Reconfiguring mon.a (monmap changed)... 2026-03-09T20:53:33.796 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 09 20:53:33 vm06 ceph-mon[52939]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "mon."}]: dispatch 2026-03-09T20:53:33.796 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 09 20:53:33 vm06 ceph-mon[52939]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' cmd=[{"prefix": "config get", "who": "mon", "key": "public_network"}]: dispatch 2026-03-09T20:53:33.796 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 09 20:53:33 vm06 ceph-mon[52939]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T20:53:33.796 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 09 20:53:33 vm06 ceph-mon[52939]: Reconfiguring daemon mon.a on vm01 2026-03-09T20:53:33.796 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 09 20:53:33 vm06 ceph-mon[52939]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' 2026-03-09T20:53:33.796 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 09 20:53:33 vm06 ceph-mon[52939]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' 2026-03-09T20:53:33.796 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 09 20:53:33 vm06 ceph-mon[52939]: Reconfiguring mgr.a (monmap changed)... 2026-03-09T20:53:33.796 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 09 20:53:33 vm06 ceph-mon[52939]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' cmd=[{"prefix": "auth get-or-create", "entity": "mgr.a", "caps": ["mon", "profile mgr", "osd", "allow *", "mds", "allow *"]}]: dispatch 2026-03-09T20:53:33.796 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 09 20:53:33 vm06 ceph-mon[52939]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' cmd=[{"prefix": "mgr services"}]: dispatch 2026-03-09T20:53:33.796 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 09 20:53:33 vm06 ceph-mon[52939]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T20:53:33.796 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 09 20:53:33 vm06 ceph-mon[52939]: Reconfiguring daemon mgr.a on vm01 2026-03-09T20:53:33.796 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 09 20:53:33 vm06 ceph-mon[52939]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' 2026-03-09T20:53:33.796 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 09 20:53:33 vm06 ceph-mon[52939]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' 2026-03-09T20:53:33.796 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 09 20:53:33 vm06 ceph-mon[52939]: Reconfiguring osd.0 (monmap changed)... 2026-03-09T20:53:33.796 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 09 20:53:33 vm06 ceph-mon[52939]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "osd.0"}]: dispatch 2026-03-09T20:53:33.796 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 09 20:53:33 vm06 ceph-mon[52939]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T20:53:33.796 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 09 20:53:33 vm06 ceph-mon[52939]: Reconfiguring daemon osd.0 on vm01 2026-03-09T20:53:33.796 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 09 20:53:33 vm06 ceph-mon[52939]: pgmap v54: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail; 767 B/s rd, 22 KiB/s wr, 2 op/s 2026-03-09T20:53:33.923 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:53:33 vm01 ceph-mon[53402]: Reconfiguring mon.a (monmap changed)... 2026-03-09T20:53:33.923 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:53:33 vm01 ceph-mon[53402]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "mon."}]: dispatch 2026-03-09T20:53:33.923 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:53:33 vm01 ceph-mon[53402]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' cmd=[{"prefix": "config get", "who": "mon", "key": "public_network"}]: dispatch 2026-03-09T20:53:33.923 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:53:33 vm01 ceph-mon[53402]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T20:53:33.923 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:53:33 vm01 ceph-mon[53402]: Reconfiguring daemon mon.a on vm01 2026-03-09T20:53:33.923 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:53:33 vm01 ceph-mon[53402]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' 2026-03-09T20:53:33.923 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:53:33 vm01 ceph-mon[53402]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' 2026-03-09T20:53:33.923 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:53:33 vm01 ceph-mon[53402]: Reconfiguring mgr.a (monmap changed)... 2026-03-09T20:53:33.923 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:53:33 vm01 ceph-mon[53402]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' cmd=[{"prefix": "auth get-or-create", "entity": "mgr.a", "caps": ["mon", "profile mgr", "osd", "allow *", "mds", "allow *"]}]: dispatch 2026-03-09T20:53:33.923 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:53:33 vm01 ceph-mon[53402]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' cmd=[{"prefix": "mgr services"}]: dispatch 2026-03-09T20:53:33.924 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:53:33 vm01 ceph-mon[53402]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T20:53:33.924 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:53:33 vm01 ceph-mon[53402]: Reconfiguring daemon mgr.a on vm01 2026-03-09T20:53:33.924 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:53:33 vm01 ceph-mon[53402]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' 2026-03-09T20:53:33.924 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:53:33 vm01 ceph-mon[53402]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' 2026-03-09T20:53:33.924 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:53:33 vm01 ceph-mon[53402]: Reconfiguring osd.0 (monmap changed)... 2026-03-09T20:53:33.924 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:53:33 vm01 ceph-mon[53402]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "osd.0"}]: dispatch 2026-03-09T20:53:33.924 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:53:33 vm01 ceph-mon[53402]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T20:53:33.924 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:53:33 vm01 ceph-mon[53402]: Reconfiguring daemon osd.0 on vm01 2026-03-09T20:53:33.924 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:53:33 vm01 ceph-mon[53402]: pgmap v54: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail; 767 B/s rd, 22 KiB/s wr, 2 op/s 2026-03-09T20:53:33.957 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:53:33 vm08 ceph-mon[54611]: Reconfiguring mon.a (monmap changed)... 2026-03-09T20:53:33.957 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:53:33 vm08 ceph-mon[54611]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "mon."}]: dispatch 2026-03-09T20:53:33.957 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:53:33 vm08 ceph-mon[54611]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' cmd=[{"prefix": "config get", "who": "mon", "key": "public_network"}]: dispatch 2026-03-09T20:53:33.957 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:53:33 vm08 ceph-mon[54611]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T20:53:33.957 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:53:33 vm08 ceph-mon[54611]: Reconfiguring daemon mon.a on vm01 2026-03-09T20:53:33.957 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:53:33 vm08 ceph-mon[54611]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' 2026-03-09T20:53:33.957 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:53:33 vm08 ceph-mon[54611]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' 2026-03-09T20:53:33.957 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:53:33 vm08 ceph-mon[54611]: Reconfiguring mgr.a (monmap changed)... 2026-03-09T20:53:33.957 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:53:33 vm08 ceph-mon[54611]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' cmd=[{"prefix": "auth get-or-create", "entity": "mgr.a", "caps": ["mon", "profile mgr", "osd", "allow *", "mds", "allow *"]}]: dispatch 2026-03-09T20:53:33.957 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:53:33 vm08 ceph-mon[54611]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' cmd=[{"prefix": "mgr services"}]: dispatch 2026-03-09T20:53:33.957 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:53:33 vm08 ceph-mon[54611]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T20:53:33.957 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:53:33 vm08 ceph-mon[54611]: Reconfiguring daemon mgr.a on vm01 2026-03-09T20:53:33.957 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:53:33 vm08 ceph-mon[54611]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' 2026-03-09T20:53:33.957 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:53:33 vm08 ceph-mon[54611]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' 2026-03-09T20:53:33.957 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:53:33 vm08 ceph-mon[54611]: Reconfiguring osd.0 (monmap changed)... 2026-03-09T20:53:33.957 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:53:33 vm08 ceph-mon[54611]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "osd.0"}]: dispatch 2026-03-09T20:53:33.957 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:53:33 vm08 ceph-mon[54611]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T20:53:33.957 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:53:33 vm08 ceph-mon[54611]: Reconfiguring daemon osd.0 on vm01 2026-03-09T20:53:33.957 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:53:33 vm08 ceph-mon[54611]: pgmap v54: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail; 767 B/s rd, 22 KiB/s wr, 2 op/s 2026-03-09T20:53:34.923 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:53:34 vm01 ceph-mon[53402]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' 2026-03-09T20:53:34.924 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:53:34 vm01 ceph-mon[53402]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' 2026-03-09T20:53:34.924 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:53:34 vm01 ceph-mon[53402]: Reconfiguring mon.b (monmap changed)... 2026-03-09T20:53:34.924 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:53:34 vm01 ceph-mon[53402]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "mon."}]: dispatch 2026-03-09T20:53:34.924 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:53:34 vm01 ceph-mon[53402]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' cmd=[{"prefix": "config get", "who": "mon", "key": "public_network"}]: dispatch 2026-03-09T20:53:34.924 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:53:34 vm01 ceph-mon[53402]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T20:53:34.924 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:53:34 vm01 ceph-mon[53402]: Reconfiguring daemon mon.b on vm06 2026-03-09T20:53:34.924 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:53:34 vm01 ceph-mon[53402]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' 2026-03-09T20:53:34.924 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:53:34 vm01 ceph-mon[53402]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' 2026-03-09T20:53:34.924 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:53:34 vm01 ceph-mon[53402]: Reconfiguring mgr.b (monmap changed)... 2026-03-09T20:53:34.924 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:53:34 vm01 ceph-mon[53402]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' cmd=[{"prefix": "auth get-or-create", "entity": "mgr.b", "caps": ["mon", "profile mgr", "osd", "allow *", "mds", "allow *"]}]: dispatch 2026-03-09T20:53:34.924 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:53:34 vm01 ceph-mon[53402]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' cmd=[{"prefix": "mgr services"}]: dispatch 2026-03-09T20:53:34.924 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:53:34 vm01 ceph-mon[53402]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T20:53:34.924 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:53:34 vm01 ceph-mon[53402]: Reconfiguring daemon mgr.b on vm06 2026-03-09T20:53:34.924 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:53:34 vm01 ceph-mon[53402]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' 2026-03-09T20:53:34.924 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:53:34 vm01 ceph-mon[53402]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' 2026-03-09T20:53:34.924 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:53:34 vm01 ceph-mon[53402]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "osd.1"}]: dispatch 2026-03-09T20:53:34.924 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:53:34 vm01 ceph-mon[53402]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T20:53:34.957 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:53:34 vm08 ceph-mon[54611]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' 2026-03-09T20:53:34.957 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:53:34 vm08 ceph-mon[54611]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' 2026-03-09T20:53:34.957 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:53:34 vm08 ceph-mon[54611]: Reconfiguring mon.b (monmap changed)... 2026-03-09T20:53:34.957 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:53:34 vm08 ceph-mon[54611]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "mon."}]: dispatch 2026-03-09T20:53:34.957 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:53:34 vm08 ceph-mon[54611]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' cmd=[{"prefix": "config get", "who": "mon", "key": "public_network"}]: dispatch 2026-03-09T20:53:34.957 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:53:34 vm08 ceph-mon[54611]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T20:53:34.957 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:53:34 vm08 ceph-mon[54611]: Reconfiguring daemon mon.b on vm06 2026-03-09T20:53:34.957 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:53:34 vm08 ceph-mon[54611]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' 2026-03-09T20:53:34.957 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:53:34 vm08 ceph-mon[54611]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' 2026-03-09T20:53:34.957 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:53:34 vm08 ceph-mon[54611]: Reconfiguring mgr.b (monmap changed)... 2026-03-09T20:53:34.957 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:53:34 vm08 ceph-mon[54611]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' cmd=[{"prefix": "auth get-or-create", "entity": "mgr.b", "caps": ["mon", "profile mgr", "osd", "allow *", "mds", "allow *"]}]: dispatch 2026-03-09T20:53:34.957 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:53:34 vm08 ceph-mon[54611]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' cmd=[{"prefix": "mgr services"}]: dispatch 2026-03-09T20:53:34.957 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:53:34 vm08 ceph-mon[54611]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T20:53:34.957 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:53:34 vm08 ceph-mon[54611]: Reconfiguring daemon mgr.b on vm06 2026-03-09T20:53:34.957 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:53:34 vm08 ceph-mon[54611]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' 2026-03-09T20:53:34.957 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:53:34 vm08 ceph-mon[54611]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' 2026-03-09T20:53:34.957 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:53:34 vm08 ceph-mon[54611]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "osd.1"}]: dispatch 2026-03-09T20:53:34.957 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:53:34 vm08 ceph-mon[54611]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T20:53:35.046 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 09 20:53:34 vm06 ceph-mon[52939]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' 2026-03-09T20:53:35.046 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 09 20:53:34 vm06 ceph-mon[52939]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' 2026-03-09T20:53:35.046 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 09 20:53:34 vm06 ceph-mon[52939]: Reconfiguring mon.b (monmap changed)... 2026-03-09T20:53:35.046 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 09 20:53:34 vm06 ceph-mon[52939]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "mon."}]: dispatch 2026-03-09T20:53:35.046 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 09 20:53:34 vm06 ceph-mon[52939]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' cmd=[{"prefix": "config get", "who": "mon", "key": "public_network"}]: dispatch 2026-03-09T20:53:35.046 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 09 20:53:34 vm06 ceph-mon[52939]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T20:53:35.046 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 09 20:53:34 vm06 ceph-mon[52939]: Reconfiguring daemon mon.b on vm06 2026-03-09T20:53:35.046 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 09 20:53:34 vm06 ceph-mon[52939]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' 2026-03-09T20:53:35.046 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 09 20:53:34 vm06 ceph-mon[52939]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' 2026-03-09T20:53:35.046 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 09 20:53:34 vm06 ceph-mon[52939]: Reconfiguring mgr.b (monmap changed)... 2026-03-09T20:53:35.046 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 09 20:53:34 vm06 ceph-mon[52939]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' cmd=[{"prefix": "auth get-or-create", "entity": "mgr.b", "caps": ["mon", "profile mgr", "osd", "allow *", "mds", "allow *"]}]: dispatch 2026-03-09T20:53:35.046 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 09 20:53:34 vm06 ceph-mon[52939]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' cmd=[{"prefix": "mgr services"}]: dispatch 2026-03-09T20:53:35.046 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 09 20:53:34 vm06 ceph-mon[52939]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T20:53:35.046 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 09 20:53:34 vm06 ceph-mon[52939]: Reconfiguring daemon mgr.b on vm06 2026-03-09T20:53:35.046 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 09 20:53:34 vm06 ceph-mon[52939]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' 2026-03-09T20:53:35.046 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 09 20:53:34 vm06 ceph-mon[52939]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' 2026-03-09T20:53:35.046 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 09 20:53:34 vm06 ceph-mon[52939]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "osd.1"}]: dispatch 2026-03-09T20:53:35.046 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 09 20:53:34 vm06 ceph-mon[52939]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T20:53:35.894 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:53:35 vm08 ceph-mon[54611]: Reconfiguring osd.1 (monmap changed)... 2026-03-09T20:53:35.894 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:53:35 vm08 ceph-mon[54611]: Reconfiguring daemon osd.1 on vm06 2026-03-09T20:53:35.894 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:53:35 vm08 ceph-mon[54611]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' 2026-03-09T20:53:35.894 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:53:35 vm08 ceph-mon[54611]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' 2026-03-09T20:53:35.894 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:53:35 vm08 ceph-mon[54611]: Reconfiguring mon.c (monmap changed)... 2026-03-09T20:53:35.894 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:53:35 vm08 ceph-mon[54611]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "mon."}]: dispatch 2026-03-09T20:53:35.895 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:53:35 vm08 ceph-mon[54611]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' cmd=[{"prefix": "config get", "who": "mon", "key": "public_network"}]: dispatch 2026-03-09T20:53:35.895 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:53:35 vm08 ceph-mon[54611]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T20:53:35.895 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:53:35 vm08 ceph-mon[54611]: Reconfiguring daemon mon.c on vm08 2026-03-09T20:53:35.895 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:53:35 vm08 ceph-mon[54611]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' 2026-03-09T20:53:35.895 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:53:35 vm08 ceph-mon[54611]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' 2026-03-09T20:53:35.895 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:53:35 vm08 ceph-mon[54611]: Reconfiguring osd.2 (monmap changed)... 2026-03-09T20:53:35.895 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:53:35 vm08 ceph-mon[54611]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "osd.2"}]: dispatch 2026-03-09T20:53:35.895 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:53:35 vm08 ceph-mon[54611]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T20:53:35.895 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:53:35 vm08 ceph-mon[54611]: Reconfiguring daemon osd.2 on vm08 2026-03-09T20:53:35.895 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:53:35 vm08 ceph-mon[54611]: pgmap v55: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail; 767 B/s rd, 22 KiB/s wr, 2 op/s 2026-03-09T20:53:35.895 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:53:35 vm08 ceph-mon[54611]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' 2026-03-09T20:53:35.895 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:53:35 vm08 ceph-mon[54611]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' 2026-03-09T20:53:36.173 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:53:35 vm01 ceph-mon[53402]: Reconfiguring osd.1 (monmap changed)... 2026-03-09T20:53:36.174 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:53:35 vm01 ceph-mon[53402]: Reconfiguring daemon osd.1 on vm06 2026-03-09T20:53:36.174 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:53:35 vm01 ceph-mon[53402]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' 2026-03-09T20:53:36.174 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:53:35 vm01 ceph-mon[53402]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' 2026-03-09T20:53:36.174 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:53:35 vm01 ceph-mon[53402]: Reconfiguring mon.c (monmap changed)... 2026-03-09T20:53:36.174 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:53:35 vm01 ceph-mon[53402]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "mon."}]: dispatch 2026-03-09T20:53:36.174 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:53:35 vm01 ceph-mon[53402]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' cmd=[{"prefix": "config get", "who": "mon", "key": "public_network"}]: dispatch 2026-03-09T20:53:36.174 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:53:35 vm01 ceph-mon[53402]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T20:53:36.174 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:53:35 vm01 ceph-mon[53402]: Reconfiguring daemon mon.c on vm08 2026-03-09T20:53:36.174 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:53:35 vm01 ceph-mon[53402]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' 2026-03-09T20:53:36.174 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:53:35 vm01 ceph-mon[53402]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' 2026-03-09T20:53:36.174 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:53:35 vm01 ceph-mon[53402]: Reconfiguring osd.2 (monmap changed)... 2026-03-09T20:53:36.174 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:53:35 vm01 ceph-mon[53402]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "osd.2"}]: dispatch 2026-03-09T20:53:36.174 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:53:35 vm01 ceph-mon[53402]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T20:53:36.174 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:53:35 vm01 ceph-mon[53402]: Reconfiguring daemon osd.2 on vm08 2026-03-09T20:53:36.174 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:53:35 vm01 ceph-mon[53402]: pgmap v55: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail; 767 B/s rd, 22 KiB/s wr, 2 op/s 2026-03-09T20:53:36.174 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:53:35 vm01 ceph-mon[53402]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' 2026-03-09T20:53:36.174 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:53:35 vm01 ceph-mon[53402]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' 2026-03-09T20:53:36.295 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 09 20:53:35 vm06 ceph-mon[52939]: Reconfiguring osd.1 (monmap changed)... 2026-03-09T20:53:36.296 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 09 20:53:35 vm06 ceph-mon[52939]: Reconfiguring daemon osd.1 on vm06 2026-03-09T20:53:36.296 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 09 20:53:35 vm06 ceph-mon[52939]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' 2026-03-09T20:53:36.296 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 09 20:53:35 vm06 ceph-mon[52939]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' 2026-03-09T20:53:36.296 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 09 20:53:35 vm06 ceph-mon[52939]: Reconfiguring mon.c (monmap changed)... 2026-03-09T20:53:36.296 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 09 20:53:35 vm06 ceph-mon[52939]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "mon."}]: dispatch 2026-03-09T20:53:36.296 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 09 20:53:35 vm06 ceph-mon[52939]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' cmd=[{"prefix": "config get", "who": "mon", "key": "public_network"}]: dispatch 2026-03-09T20:53:36.296 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 09 20:53:35 vm06 ceph-mon[52939]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T20:53:36.296 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 09 20:53:35 vm06 ceph-mon[52939]: Reconfiguring daemon mon.c on vm08 2026-03-09T20:53:36.296 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 09 20:53:35 vm06 ceph-mon[52939]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' 2026-03-09T20:53:36.296 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 09 20:53:35 vm06 ceph-mon[52939]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' 2026-03-09T20:53:36.296 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 09 20:53:35 vm06 ceph-mon[52939]: Reconfiguring osd.2 (monmap changed)... 2026-03-09T20:53:36.296 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 09 20:53:35 vm06 ceph-mon[52939]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "osd.2"}]: dispatch 2026-03-09T20:53:36.296 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 09 20:53:35 vm06 ceph-mon[52939]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T20:53:36.296 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 09 20:53:35 vm06 ceph-mon[52939]: Reconfiguring daemon osd.2 on vm08 2026-03-09T20:53:36.296 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 09 20:53:35 vm06 ceph-mon[52939]: pgmap v55: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail; 767 B/s rd, 22 KiB/s wr, 2 op/s 2026-03-09T20:53:36.296 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 09 20:53:35 vm06 ceph-mon[52939]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' 2026-03-09T20:53:36.296 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 09 20:53:35 vm06 ceph-mon[52939]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' 2026-03-09T20:53:37.795 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 09 20:53:37 vm06 ceph-mon[52939]: pgmap v56: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-09T20:53:37.923 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:53:37 vm01 ceph-mon[53402]: pgmap v56: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-09T20:53:37.957 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:53:37 vm08 ceph-mon[54611]: pgmap v56: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-09T20:53:39.796 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 09 20:53:39 vm06 ceph-mon[52939]: pgmap v57: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-09T20:53:39.923 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:53:39 vm01 ceph-mon[53402]: pgmap v57: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-09T20:53:39.957 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:53:39 vm08 ceph-mon[54611]: pgmap v57: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-09T20:53:41.795 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 09 20:53:41 vm06 ceph-mon[52939]: pgmap v58: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-09T20:53:41.923 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:53:41 vm01 ceph-mon[53402]: pgmap v58: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-09T20:53:41.957 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:53:41 vm08 ceph-mon[54611]: pgmap v58: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-09T20:53:43.795 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 09 20:53:43 vm06 ceph-mon[52939]: pgmap v59: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-09T20:53:43.923 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:53:43 vm01 ceph-mon[53402]: pgmap v59: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-09T20:53:43.957 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:53:43 vm08 ceph-mon[54611]: pgmap v59: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-09T20:53:45.795 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 09 20:53:45 vm06 ceph-mon[52939]: pgmap v60: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-09T20:53:45.923 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:53:45 vm01 ceph-mon[53402]: pgmap v60: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-09T20:53:45.957 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:53:45 vm08 ceph-mon[54611]: pgmap v60: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-09T20:53:47.795 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 09 20:53:47 vm06 ceph-mon[52939]: pgmap v61: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-09T20:53:47.923 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:53:47 vm01 ceph-mon[53402]: pgmap v61: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-09T20:53:47.957 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:53:47 vm08 ceph-mon[54611]: pgmap v61: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-09T20:53:49.795 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 09 20:53:49 vm06 ceph-mon[52939]: pgmap v62: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-09T20:53:49.923 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:53:49 vm01 ceph-mon[53402]: pgmap v62: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-09T20:53:49.957 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:53:49 vm08 ceph-mon[54611]: pgmap v62: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-09T20:53:51.795 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 09 20:53:51 vm06 ceph-mon[52939]: pgmap v63: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-09T20:53:51.923 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:53:51 vm01 ceph-mon[53402]: pgmap v63: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-09T20:53:51.957 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:53:51 vm08 ceph-mon[54611]: pgmap v63: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-09T20:53:53.795 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 09 20:53:53 vm06 ceph-mon[52939]: pgmap v64: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-09T20:53:53.923 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:53:53 vm01 ceph-mon[53402]: pgmap v64: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-09T20:53:53.957 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:53:53 vm08 ceph-mon[54611]: pgmap v64: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-09T20:53:55.795 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 09 20:53:55 vm06 ceph-mon[52939]: pgmap v65: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-09T20:53:55.923 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:53:55 vm01 ceph-mon[53402]: pgmap v65: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-09T20:53:55.957 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:53:55 vm08 ceph-mon[54611]: pgmap v65: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-09T20:53:57.795 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 09 20:53:57 vm06 ceph-mon[52939]: pgmap v66: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-09T20:53:57.923 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:53:57 vm01 ceph-mon[53402]: pgmap v66: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-09T20:53:57.957 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:53:57 vm08 ceph-mon[54611]: pgmap v66: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-09T20:53:59.795 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 09 20:53:59 vm06 ceph-mon[52939]: pgmap v67: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-09T20:53:59.923 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:53:59 vm01 ceph-mon[53402]: pgmap v67: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-09T20:53:59.957 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:53:59 vm08 ceph-mon[54611]: pgmap v67: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-09T20:54:01.796 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 09 20:54:01 vm06 ceph-mon[52939]: pgmap v68: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-09T20:54:01.923 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:54:01 vm01 ceph-mon[53402]: pgmap v68: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-09T20:54:01.957 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:54:01 vm08 ceph-mon[54611]: pgmap v68: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-09T20:54:03.796 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 09 20:54:03 vm06 ceph-mon[52939]: pgmap v69: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-09T20:54:03.923 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:54:03 vm01 ceph-mon[53402]: pgmap v69: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-09T20:54:03.957 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:54:03 vm08 ceph-mon[54611]: pgmap v69: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-09T20:54:05.795 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 09 20:54:05 vm06 ceph-mon[52939]: pgmap v70: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-09T20:54:05.923 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:54:05 vm01 ceph-mon[53402]: pgmap v70: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-09T20:54:05.957 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:54:05 vm08 ceph-mon[54611]: pgmap v70: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-09T20:54:07.795 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 09 20:54:07 vm06 ceph-mon[52939]: pgmap v71: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-09T20:54:07.923 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:54:07 vm01 ceph-mon[53402]: pgmap v71: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-09T20:54:07.957 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:54:07 vm08 ceph-mon[54611]: pgmap v71: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-09T20:54:09.795 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 09 20:54:09 vm06 ceph-mon[52939]: pgmap v72: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-09T20:54:09.923 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:54:09 vm01 ceph-mon[53402]: pgmap v72: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-09T20:54:09.957 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:54:09 vm08 ceph-mon[54611]: pgmap v72: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-09T20:54:11.795 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 09 20:54:11 vm06 ceph-mon[52939]: pgmap v73: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-09T20:54:11.923 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:54:11 vm01 ceph-mon[53402]: pgmap v73: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-09T20:54:11.957 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:54:11 vm08 ceph-mon[54611]: pgmap v73: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-09T20:54:13.795 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 09 20:54:13 vm06 ceph-mon[52939]: pgmap v74: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-09T20:54:13.923 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:54:13 vm01 ceph-mon[53402]: pgmap v74: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-09T20:54:13.957 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:54:13 vm08 ceph-mon[54611]: pgmap v74: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-09T20:54:15.795 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 09 20:54:15 vm06 ceph-mon[52939]: pgmap v75: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-09T20:54:15.923 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:54:15 vm01 ceph-mon[53402]: pgmap v75: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-09T20:54:15.957 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:54:15 vm08 ceph-mon[54611]: pgmap v75: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-09T20:54:17.795 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 09 20:54:17 vm06 ceph-mon[52939]: pgmap v76: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-09T20:54:17.923 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:54:17 vm01 ceph-mon[53402]: pgmap v76: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-09T20:54:17.957 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:54:17 vm08 ceph-mon[54611]: pgmap v76: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-09T20:54:19.795 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 09 20:54:19 vm06 ceph-mon[52939]: pgmap v77: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-09T20:54:19.923 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:54:19 vm01 ceph-mon[53402]: pgmap v77: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-09T20:54:19.957 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:54:19 vm08 ceph-mon[54611]: pgmap v77: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-09T20:54:21.795 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 09 20:54:21 vm06 ceph-mon[52939]: pgmap v78: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-09T20:54:21.923 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:54:21 vm01 ceph-mon[53402]: pgmap v78: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-09T20:54:21.957 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:54:21 vm08 ceph-mon[54611]: pgmap v78: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-09T20:54:23.795 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 09 20:54:23 vm06 ceph-mon[52939]: pgmap v79: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-09T20:54:23.923 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:54:23 vm01 ceph-mon[53402]: pgmap v79: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-09T20:54:23.957 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:54:23 vm08 ceph-mon[54611]: pgmap v79: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-09T20:54:25.795 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 09 20:54:25 vm06 ceph-mon[52939]: pgmap v80: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-09T20:54:25.923 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:54:25 vm01 ceph-mon[53402]: pgmap v80: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-09T20:54:25.957 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:54:25 vm08 ceph-mon[54611]: pgmap v80: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-09T20:54:27.795 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 09 20:54:27 vm06 ceph-mon[52939]: pgmap v81: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-09T20:54:27.923 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:54:27 vm01 ceph-mon[53402]: pgmap v81: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-09T20:54:27.957 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:54:27 vm08 ceph-mon[54611]: pgmap v81: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-09T20:54:29.795 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 09 20:54:29 vm06 ceph-mon[52939]: pgmap v82: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-09T20:54:29.923 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:54:29 vm01 ceph-mon[53402]: pgmap v82: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-09T20:54:29.957 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:54:29 vm08 ceph-mon[54611]: pgmap v82: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-09T20:54:31.795 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 09 20:54:31 vm06 ceph-mon[52939]: pgmap v83: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-09T20:54:31.923 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:54:31 vm01 ceph-mon[53402]: pgmap v83: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-09T20:54:31.957 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:54:31 vm08 ceph-mon[54611]: pgmap v83: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-09T20:54:33.795 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 09 20:54:33 vm06 ceph-mon[52939]: pgmap v84: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-09T20:54:33.923 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:54:33 vm01 ceph-mon[53402]: pgmap v84: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-09T20:54:33.957 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:54:33 vm08 ceph-mon[54611]: pgmap v84: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-09T20:54:35.795 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 09 20:54:35 vm06 ceph-mon[52939]: pgmap v85: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-09T20:54:35.923 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:54:35 vm01 ceph-mon[53402]: pgmap v85: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-09T20:54:35.957 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:54:35 vm08 ceph-mon[54611]: pgmap v85: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-09T20:54:36.795 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 09 20:54:36 vm06 ceph-mon[52939]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T20:54:36.795 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 09 20:54:36 vm06 ceph-mon[52939]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T20:54:36.795 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 09 20:54:36 vm06 ceph-mon[52939]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T20:54:36.795 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 09 20:54:36 vm06 ceph-mon[52939]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' 2026-03-09T20:54:36.923 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:54:36 vm01 ceph-mon[53402]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T20:54:36.923 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:54:36 vm01 ceph-mon[53402]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T20:54:36.923 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:54:36 vm01 ceph-mon[53402]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T20:54:36.923 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:54:36 vm01 ceph-mon[53402]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' 2026-03-09T20:54:36.957 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:54:36 vm08 ceph-mon[54611]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T20:54:36.957 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:54:36 vm08 ceph-mon[54611]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T20:54:36.957 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:54:36 vm08 ceph-mon[54611]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T20:54:36.957 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:54:36 vm08 ceph-mon[54611]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' 2026-03-09T20:54:37.795 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 09 20:54:37 vm06 ceph-mon[52939]: pgmap v86: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-09T20:54:37.923 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:54:37 vm01 ceph-mon[53402]: pgmap v86: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-09T20:54:37.957 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:54:37 vm08 ceph-mon[54611]: pgmap v86: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-09T20:54:39.795 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 09 20:54:39 vm06 ceph-mon[52939]: pgmap v87: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-09T20:54:39.923 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:54:39 vm01 ceph-mon[53402]: pgmap v87: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-09T20:54:39.957 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:54:39 vm08 ceph-mon[54611]: pgmap v87: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-09T20:54:41.795 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 09 20:54:41 vm06 ceph-mon[52939]: pgmap v88: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-09T20:54:41.923 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:54:41 vm01 ceph-mon[53402]: pgmap v88: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-09T20:54:41.957 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:54:41 vm08 ceph-mon[54611]: pgmap v88: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-09T20:54:43.795 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 09 20:54:43 vm06 ceph-mon[52939]: pgmap v89: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-09T20:54:43.923 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:54:43 vm01 ceph-mon[53402]: pgmap v89: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-09T20:54:43.957 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:54:43 vm08 ceph-mon[54611]: pgmap v89: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-09T20:54:45.795 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 09 20:54:45 vm06 ceph-mon[52939]: pgmap v90: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-09T20:54:45.923 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:54:45 vm01 ceph-mon[53402]: pgmap v90: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-09T20:54:45.957 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:54:45 vm08 ceph-mon[54611]: pgmap v90: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-09T20:54:47.795 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 09 20:54:47 vm06 ceph-mon[52939]: pgmap v91: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-09T20:54:47.923 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:54:47 vm01 ceph-mon[53402]: pgmap v91: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-09T20:54:47.957 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:54:47 vm08 ceph-mon[54611]: pgmap v91: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-09T20:54:49.795 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 09 20:54:49 vm06 ceph-mon[52939]: pgmap v92: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-09T20:54:49.923 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:54:49 vm01 ceph-mon[53402]: pgmap v92: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-09T20:54:49.957 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:54:49 vm08 ceph-mon[54611]: pgmap v92: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-09T20:54:51.795 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 09 20:54:51 vm06 ceph-mon[52939]: pgmap v93: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-09T20:54:51.923 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:54:51 vm01 ceph-mon[53402]: pgmap v93: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-09T20:54:51.957 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:54:51 vm08 ceph-mon[54611]: pgmap v93: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-09T20:54:53.795 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 09 20:54:53 vm06 ceph-mon[52939]: pgmap v94: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-09T20:54:53.923 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:54:53 vm01 ceph-mon[53402]: pgmap v94: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-09T20:54:53.957 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:54:53 vm08 ceph-mon[54611]: pgmap v94: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-09T20:54:55.795 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 09 20:54:55 vm06 ceph-mon[52939]: pgmap v95: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-09T20:54:55.923 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:54:55 vm01 ceph-mon[53402]: pgmap v95: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-09T20:54:55.957 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:54:55 vm08 ceph-mon[54611]: pgmap v95: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-09T20:54:57.385 INFO:teuthology.orchestra.run.vm01.stderr:+ ceph orch ps --refresh 2026-03-09T20:54:57.550 INFO:teuthology.orchestra.run.vm01.stdout:NAME HOST PORTS STATUS REFRESHED AGE MEM USE MEM LIM VERSION IMAGE ID CONTAINER ID 2026-03-09T20:54:57.550 INFO:teuthology.orchestra.run.vm01.stdout:mgr.a vm01 *:9283,8765 running (3m) 2m ago 3m 539M - 19.2.3-678-ge911bdeb 654f31e6858e 9fc252e619ba 2026-03-09T20:54:57.550 INFO:teuthology.orchestra.run.vm01.stdout:mgr.b vm06 *:8443,8765 running (2m) 116s ago 2m 486M - 19.2.3-678-ge911bdeb 654f31e6858e 355207d5317a 2026-03-09T20:54:57.550 INFO:teuthology.orchestra.run.vm01.stdout:mon.a vm01 running (3m) 2m ago 3m 41.1M 2048M 19.2.3-678-ge911bdeb 654f31e6858e 6532f28f79f1 2026-03-09T20:54:57.550 INFO:teuthology.orchestra.run.vm01.stdout:mon.b vm06 running (2m) 116s ago 2m 34.9M 2048M 19.2.3-678-ge911bdeb 654f31e6858e d29f2c50f910 2026-03-09T20:54:57.550 INFO:teuthology.orchestra.run.vm01.stdout:mon.c vm08 running (2m) 106s ago 2m 40.5M 2048M 19.2.3-678-ge911bdeb 654f31e6858e f3c5e5978ae6 2026-03-09T20:54:57.550 INFO:teuthology.orchestra.run.vm01.stdout:osd.0 vm01 running (2m) 2m ago 2m 15.4M 4096M 19.2.3-678-ge911bdeb 654f31e6858e 0d379550ff05 2026-03-09T20:54:57.550 INFO:teuthology.orchestra.run.vm01.stdout:osd.1 vm06 running (117s) 116s ago 117s 12.3M 4096M 19.2.3-678-ge911bdeb 654f31e6858e 1efb5185c5db 2026-03-09T20:54:57.550 INFO:teuthology.orchestra.run.vm01.stdout:osd.2 vm08 running (106s) 106s ago 106s 12.5M 4353M 19.2.3-678-ge911bdeb 654f31e6858e b061149d1825 2026-03-09T20:54:57.563 INFO:teuthology.orchestra.run.vm01.stderr:+ ceph orch ls --service-name mon --export 2026-03-09T20:54:57.673 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:54:57 vm01 ceph-mon[53402]: pgmap v96: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-09T20:54:57.760 INFO:teuthology.orchestra.run.vm01.stderr:+ ceph orch apply -i mon.yaml 2026-03-09T20:54:57.795 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 09 20:54:57 vm06 ceph-mon[52939]: pgmap v96: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-09T20:54:57.849 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:54:57 vm08 ceph-mon[54611]: pgmap v96: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-09T20:54:57.947 INFO:teuthology.orchestra.run.vm01.stdout:Scheduled mon update... 2026-03-09T20:54:57.962 INFO:teuthology.orchestra.run.vm01.stderr:+ sleep 90 2026-03-09T20:54:58.795 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 09 20:54:58 vm06 ceph-mon[52939]: from='client.14442 -' entity='client.admin' cmd=[{"prefix": "orch ps", "refresh": true, "target": ["mon-mgr", ""]}]: dispatch 2026-03-09T20:54:58.796 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 09 20:54:58 vm06 ceph-mon[52939]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T20:54:58.796 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 09 20:54:58 vm06 ceph-mon[52939]: from='client.24272 -' entity='client.admin' cmd=[{"prefix": "orch ls", "service_name": "mon", "export": true, "target": ["mon-mgr", ""]}]: dispatch 2026-03-09T20:54:58.796 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 09 20:54:58 vm06 ceph-mon[52939]: from='client.14454 -' entity='client.admin' cmd=[{"prefix": "orch apply", "target": ["mon-mgr", ""]}]: dispatch 2026-03-09T20:54:58.796 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 09 20:54:58 vm06 ceph-mon[52939]: Saving service mon spec with placement count:3 2026-03-09T20:54:58.796 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 09 20:54:58 vm06 ceph-mon[52939]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' 2026-03-09T20:54:58.796 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 09 20:54:58 vm06 ceph-mon[52939]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' 2026-03-09T20:54:58.796 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 09 20:54:58 vm06 ceph-mon[52939]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' 2026-03-09T20:54:58.796 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 09 20:54:58 vm06 ceph-mon[52939]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' 2026-03-09T20:54:58.796 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 09 20:54:58 vm06 ceph-mon[52939]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' 2026-03-09T20:54:58.796 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 09 20:54:58 vm06 ceph-mon[52939]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' 2026-03-09T20:54:58.796 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 09 20:54:58 vm06 ceph-mon[52939]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' 2026-03-09T20:54:58.796 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 09 20:54:58 vm06 ceph-mon[52939]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T20:54:58.796 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 09 20:54:58 vm06 ceph-mon[52939]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T20:54:58.796 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 09 20:54:58 vm06 ceph-mon[52939]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' 2026-03-09T20:54:58.796 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 09 20:54:58 vm06 ceph-mon[52939]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' cmd=[{"prefix": "quorum_status"}]: dispatch 2026-03-09T20:54:58.796 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 09 20:54:58 vm06 ceph-mon[52939]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' 2026-03-09T20:54:58.796 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 09 20:54:58 vm06 ceph-mon[52939]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T20:54:58.796 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 09 20:54:58 vm06 ceph-mon[52939]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T20:54:58.796 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 09 20:54:58 vm06 ceph-mon[52939]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T20:54:58.796 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 09 20:54:58 vm06 ceph-mon[52939]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' 2026-03-09T20:54:58.923 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:54:58 vm01 ceph-mon[53402]: from='client.14442 -' entity='client.admin' cmd=[{"prefix": "orch ps", "refresh": true, "target": ["mon-mgr", ""]}]: dispatch 2026-03-09T20:54:58.923 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:54:58 vm01 ceph-mon[53402]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T20:54:58.923 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:54:58 vm01 ceph-mon[53402]: from='client.24272 -' entity='client.admin' cmd=[{"prefix": "orch ls", "service_name": "mon", "export": true, "target": ["mon-mgr", ""]}]: dispatch 2026-03-09T20:54:58.923 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:54:58 vm01 ceph-mon[53402]: from='client.14454 -' entity='client.admin' cmd=[{"prefix": "orch apply", "target": ["mon-mgr", ""]}]: dispatch 2026-03-09T20:54:58.923 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:54:58 vm01 ceph-mon[53402]: Saving service mon spec with placement count:3 2026-03-09T20:54:58.923 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:54:58 vm01 ceph-mon[53402]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' 2026-03-09T20:54:58.923 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:54:58 vm01 ceph-mon[53402]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' 2026-03-09T20:54:58.923 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:54:58 vm01 ceph-mon[53402]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' 2026-03-09T20:54:58.923 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:54:58 vm01 ceph-mon[53402]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' 2026-03-09T20:54:58.924 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:54:58 vm01 ceph-mon[53402]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' 2026-03-09T20:54:58.924 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:54:58 vm01 ceph-mon[53402]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' 2026-03-09T20:54:58.924 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:54:58 vm01 ceph-mon[53402]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' 2026-03-09T20:54:58.924 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:54:58 vm01 ceph-mon[53402]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T20:54:58.924 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:54:58 vm01 ceph-mon[53402]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T20:54:58.924 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:54:58 vm01 ceph-mon[53402]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' 2026-03-09T20:54:58.924 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:54:58 vm01 ceph-mon[53402]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' cmd=[{"prefix": "quorum_status"}]: dispatch 2026-03-09T20:54:58.924 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:54:58 vm01 ceph-mon[53402]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' 2026-03-09T20:54:58.924 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:54:58 vm01 ceph-mon[53402]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T20:54:58.924 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:54:58 vm01 ceph-mon[53402]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T20:54:58.924 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:54:58 vm01 ceph-mon[53402]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T20:54:58.924 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:54:58 vm01 ceph-mon[53402]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' 2026-03-09T20:54:58.957 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:54:58 vm08 ceph-mon[54611]: from='client.14442 -' entity='client.admin' cmd=[{"prefix": "orch ps", "refresh": true, "target": ["mon-mgr", ""]}]: dispatch 2026-03-09T20:54:58.957 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:54:58 vm08 ceph-mon[54611]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T20:54:58.957 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:54:58 vm08 ceph-mon[54611]: from='client.24272 -' entity='client.admin' cmd=[{"prefix": "orch ls", "service_name": "mon", "export": true, "target": ["mon-mgr", ""]}]: dispatch 2026-03-09T20:54:58.957 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:54:58 vm08 ceph-mon[54611]: from='client.14454 -' entity='client.admin' cmd=[{"prefix": "orch apply", "target": ["mon-mgr", ""]}]: dispatch 2026-03-09T20:54:58.957 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:54:58 vm08 ceph-mon[54611]: Saving service mon spec with placement count:3 2026-03-09T20:54:58.957 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:54:58 vm08 ceph-mon[54611]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' 2026-03-09T20:54:58.957 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:54:58 vm08 ceph-mon[54611]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' 2026-03-09T20:54:58.957 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:54:58 vm08 ceph-mon[54611]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' 2026-03-09T20:54:58.957 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:54:58 vm08 ceph-mon[54611]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' 2026-03-09T20:54:58.957 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:54:58 vm08 ceph-mon[54611]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' 2026-03-09T20:54:58.957 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:54:58 vm08 ceph-mon[54611]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' 2026-03-09T20:54:58.957 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:54:58 vm08 ceph-mon[54611]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' 2026-03-09T20:54:58.957 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:54:58 vm08 ceph-mon[54611]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T20:54:58.957 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:54:58 vm08 ceph-mon[54611]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T20:54:58.957 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:54:58 vm08 ceph-mon[54611]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' 2026-03-09T20:54:58.957 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:54:58 vm08 ceph-mon[54611]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' cmd=[{"prefix": "quorum_status"}]: dispatch 2026-03-09T20:54:58.957 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:54:58 vm08 ceph-mon[54611]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' 2026-03-09T20:54:58.957 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:54:58 vm08 ceph-mon[54611]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T20:54:58.957 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:54:58 vm08 ceph-mon[54611]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T20:54:58.957 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:54:58 vm08 ceph-mon[54611]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T20:54:58.957 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:54:58 vm08 ceph-mon[54611]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' 2026-03-09T20:54:59.796 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 09 20:54:59 vm06 ceph-mon[52939]: pgmap v97: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-09T20:54:59.923 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:54:59 vm01 ceph-mon[53402]: pgmap v97: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-09T20:54:59.957 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:54:59 vm08 ceph-mon[54611]: pgmap v97: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-09T20:55:01.795 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 09 20:55:01 vm06 ceph-mon[52939]: pgmap v98: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-09T20:55:01.923 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:55:01 vm01 ceph-mon[53402]: pgmap v98: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-09T20:55:01.957 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:55:01 vm08 ceph-mon[54611]: pgmap v98: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-09T20:55:03.795 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 09 20:55:03 vm06 ceph-mon[52939]: pgmap v99: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-09T20:55:03.923 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:55:03 vm01 ceph-mon[53402]: pgmap v99: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-09T20:55:03.957 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:55:03 vm08 ceph-mon[54611]: pgmap v99: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-09T20:55:05.795 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 09 20:55:05 vm06 ceph-mon[52939]: pgmap v100: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-09T20:55:05.923 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:55:05 vm01 ceph-mon[53402]: pgmap v100: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-09T20:55:05.957 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:55:05 vm08 ceph-mon[54611]: pgmap v100: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-09T20:55:07.795 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 09 20:55:07 vm06 ceph-mon[52939]: pgmap v101: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-09T20:55:07.923 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:55:07 vm01 ceph-mon[53402]: pgmap v101: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-09T20:55:07.957 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:55:07 vm08 ceph-mon[54611]: pgmap v101: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-09T20:55:09.796 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 09 20:55:09 vm06 ceph-mon[52939]: pgmap v102: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-09T20:55:09.923 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:55:09 vm01 ceph-mon[53402]: pgmap v102: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-09T20:55:09.957 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:55:09 vm08 ceph-mon[54611]: pgmap v102: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-09T20:55:11.795 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 09 20:55:11 vm06 ceph-mon[52939]: pgmap v103: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-09T20:55:11.923 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:55:11 vm01 ceph-mon[53402]: pgmap v103: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-09T20:55:11.957 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:55:11 vm08 ceph-mon[54611]: pgmap v103: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-09T20:55:13.795 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 09 20:55:13 vm06 ceph-mon[52939]: pgmap v104: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-09T20:55:13.923 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:55:13 vm01 ceph-mon[53402]: pgmap v104: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-09T20:55:13.957 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:55:13 vm08 ceph-mon[54611]: pgmap v104: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-09T20:55:15.795 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 09 20:55:15 vm06 ceph-mon[52939]: pgmap v105: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-09T20:55:15.923 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:55:15 vm01 ceph-mon[53402]: pgmap v105: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-09T20:55:15.957 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:55:15 vm08 ceph-mon[54611]: pgmap v105: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-09T20:55:17.795 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 09 20:55:17 vm06 ceph-mon[52939]: pgmap v106: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-09T20:55:17.923 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:55:17 vm01 ceph-mon[53402]: pgmap v106: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-09T20:55:17.957 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:55:17 vm08 ceph-mon[54611]: pgmap v106: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-09T20:55:19.795 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 09 20:55:19 vm06 ceph-mon[52939]: pgmap v107: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-09T20:55:19.923 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:55:19 vm01 ceph-mon[53402]: pgmap v107: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-09T20:55:19.957 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:55:19 vm08 ceph-mon[54611]: pgmap v107: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-09T20:55:21.795 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 09 20:55:21 vm06 ceph-mon[52939]: pgmap v108: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-09T20:55:21.923 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:55:21 vm01 ceph-mon[53402]: pgmap v108: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-09T20:55:21.957 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:55:21 vm08 ceph-mon[54611]: pgmap v108: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-09T20:55:23.795 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 09 20:55:23 vm06 ceph-mon[52939]: pgmap v109: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-09T20:55:23.923 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:55:23 vm01 ceph-mon[53402]: pgmap v109: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-09T20:55:23.957 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:55:23 vm08 ceph-mon[54611]: pgmap v109: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-09T20:55:25.795 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 09 20:55:25 vm06 ceph-mon[52939]: pgmap v110: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-09T20:55:25.923 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:55:25 vm01 ceph-mon[53402]: pgmap v110: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-09T20:55:25.957 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:55:25 vm08 ceph-mon[54611]: pgmap v110: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-09T20:55:27.795 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 09 20:55:27 vm06 ceph-mon[52939]: pgmap v111: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-09T20:55:27.923 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:55:27 vm01 ceph-mon[53402]: pgmap v111: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-09T20:55:27.957 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:55:27 vm08 ceph-mon[54611]: pgmap v111: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-09T20:55:29.795 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 09 20:55:29 vm06 ceph-mon[52939]: pgmap v112: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-09T20:55:29.923 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:55:29 vm01 ceph-mon[53402]: pgmap v112: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-09T20:55:29.957 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:55:29 vm08 ceph-mon[54611]: pgmap v112: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-09T20:55:31.795 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 09 20:55:31 vm06 ceph-mon[52939]: pgmap v113: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-09T20:55:31.923 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:55:31 vm01 ceph-mon[53402]: pgmap v113: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-09T20:55:31.957 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:55:31 vm08 ceph-mon[54611]: pgmap v113: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-09T20:55:33.795 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 09 20:55:33 vm06 ceph-mon[52939]: pgmap v114: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-09T20:55:33.923 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:55:33 vm01 ceph-mon[53402]: pgmap v114: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-09T20:55:33.957 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:55:33 vm08 ceph-mon[54611]: pgmap v114: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-09T20:55:35.795 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 09 20:55:35 vm06 ceph-mon[52939]: pgmap v115: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-09T20:55:35.923 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:55:35 vm01 ceph-mon[53402]: pgmap v115: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-09T20:55:35.957 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:55:35 vm08 ceph-mon[54611]: pgmap v115: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-09T20:55:37.795 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 09 20:55:37 vm06 ceph-mon[52939]: pgmap v116: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-09T20:55:37.923 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:55:37 vm01 ceph-mon[53402]: pgmap v116: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-09T20:55:37.957 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:55:37 vm08 ceph-mon[54611]: pgmap v116: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-09T20:55:39.795 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 09 20:55:39 vm06 ceph-mon[52939]: pgmap v117: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-09T20:55:39.923 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:55:39 vm01 ceph-mon[53402]: pgmap v117: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-09T20:55:39.957 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:55:39 vm08 ceph-mon[54611]: pgmap v117: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-09T20:55:41.795 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 09 20:55:41 vm06 ceph-mon[52939]: pgmap v118: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-09T20:55:41.923 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:55:41 vm01 ceph-mon[53402]: pgmap v118: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-09T20:55:41.957 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:55:41 vm08 ceph-mon[54611]: pgmap v118: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-09T20:55:43.795 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 09 20:55:43 vm06 ceph-mon[52939]: pgmap v119: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-09T20:55:43.923 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:55:43 vm01 ceph-mon[53402]: pgmap v119: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-09T20:55:43.957 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:55:43 vm08 ceph-mon[54611]: pgmap v119: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-09T20:55:45.796 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 09 20:55:45 vm06 ceph-mon[52939]: pgmap v120: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-09T20:55:45.923 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:55:45 vm01 ceph-mon[53402]: pgmap v120: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-09T20:55:45.957 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:55:45 vm08 ceph-mon[54611]: pgmap v120: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-09T20:55:47.795 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 09 20:55:47 vm06 ceph-mon[52939]: pgmap v121: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-09T20:55:47.923 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:55:47 vm01 ceph-mon[53402]: pgmap v121: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-09T20:55:47.957 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:55:47 vm08 ceph-mon[54611]: pgmap v121: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-09T20:55:49.795 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 09 20:55:49 vm06 ceph-mon[52939]: pgmap v122: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-09T20:55:49.923 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:55:49 vm01 ceph-mon[53402]: pgmap v122: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-09T20:55:49.957 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:55:49 vm08 ceph-mon[54611]: pgmap v122: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-09T20:55:51.795 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 09 20:55:51 vm06 ceph-mon[52939]: pgmap v123: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-09T20:55:51.923 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:55:51 vm01 ceph-mon[53402]: pgmap v123: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-09T20:55:51.957 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:55:51 vm08 ceph-mon[54611]: pgmap v123: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-09T20:55:53.795 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 09 20:55:53 vm06 ceph-mon[52939]: pgmap v124: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-09T20:55:53.923 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:55:53 vm01 ceph-mon[53402]: pgmap v124: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-09T20:55:53.957 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:55:53 vm08 ceph-mon[54611]: pgmap v124: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-09T20:55:55.795 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 09 20:55:55 vm06 ceph-mon[52939]: pgmap v125: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-09T20:55:55.923 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:55:55 vm01 ceph-mon[53402]: pgmap v125: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-09T20:55:55.957 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:55:55 vm08 ceph-mon[54611]: pgmap v125: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-09T20:55:57.795 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 09 20:55:57 vm06 ceph-mon[52939]: pgmap v126: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-09T20:55:57.923 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:55:57 vm01 ceph-mon[53402]: pgmap v126: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-09T20:55:57.957 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:55:57 vm08 ceph-mon[54611]: pgmap v126: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-09T20:55:58.795 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 09 20:55:58 vm06 ceph-mon[52939]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T20:55:58.795 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 09 20:55:58 vm06 ceph-mon[52939]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T20:55:58.795 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 09 20:55:58 vm06 ceph-mon[52939]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T20:55:58.795 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 09 20:55:58 vm06 ceph-mon[52939]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' 2026-03-09T20:55:58.923 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:55:58 vm01 ceph-mon[53402]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T20:55:58.923 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:55:58 vm01 ceph-mon[53402]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T20:55:58.923 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:55:58 vm01 ceph-mon[53402]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T20:55:58.923 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:55:58 vm01 ceph-mon[53402]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' 2026-03-09T20:55:58.957 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:55:58 vm08 ceph-mon[54611]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T20:55:58.957 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:55:58 vm08 ceph-mon[54611]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T20:55:58.957 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:55:58 vm08 ceph-mon[54611]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T20:55:58.957 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:55:58 vm08 ceph-mon[54611]: from='mgr.14150 192.168.123.101:0/1274504658' entity='mgr.a' 2026-03-09T20:55:59.795 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 09 20:55:59 vm06 ceph-mon[52939]: pgmap v127: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-09T20:55:59.923 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:55:59 vm01 ceph-mon[53402]: pgmap v127: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-09T20:55:59.957 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:55:59 vm08 ceph-mon[54611]: pgmap v127: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-09T20:56:01.795 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 09 20:56:01 vm06 ceph-mon[52939]: pgmap v128: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-09T20:56:01.923 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:56:01 vm01 ceph-mon[53402]: pgmap v128: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-09T20:56:01.957 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:56:01 vm08 ceph-mon[54611]: pgmap v128: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-09T20:56:03.795 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 09 20:56:03 vm06 ceph-mon[52939]: pgmap v129: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-09T20:56:03.923 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:56:03 vm01 ceph-mon[53402]: pgmap v129: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-09T20:56:03.957 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:56:03 vm08 ceph-mon[54611]: pgmap v129: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-09T20:56:05.795 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 09 20:56:05 vm06 ceph-mon[52939]: pgmap v130: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-09T20:56:05.923 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:56:05 vm01 ceph-mon[53402]: pgmap v130: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-09T20:56:05.957 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:56:05 vm08 ceph-mon[54611]: pgmap v130: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-09T20:56:07.795 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 09 20:56:07 vm06 ceph-mon[52939]: pgmap v131: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-09T20:56:07.923 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:56:07 vm01 ceph-mon[53402]: pgmap v131: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-09T20:56:07.957 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:56:07 vm08 ceph-mon[54611]: pgmap v131: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-09T20:56:09.795 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 09 20:56:09 vm06 ceph-mon[52939]: pgmap v132: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-09T20:56:09.923 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:56:09 vm01 ceph-mon[53402]: pgmap v132: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-09T20:56:09.957 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:56:09 vm08 ceph-mon[54611]: pgmap v132: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-09T20:56:11.795 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 09 20:56:11 vm06 ceph-mon[52939]: pgmap v133: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-09T20:56:11.923 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:56:11 vm01 ceph-mon[53402]: pgmap v133: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-09T20:56:11.957 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:56:11 vm08 ceph-mon[54611]: pgmap v133: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-09T20:56:13.795 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 09 20:56:13 vm06 ceph-mon[52939]: pgmap v134: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-09T20:56:13.923 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:56:13 vm01 ceph-mon[53402]: pgmap v134: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-09T20:56:13.957 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:56:13 vm08 ceph-mon[54611]: pgmap v134: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-09T20:56:15.795 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 09 20:56:15 vm06 ceph-mon[52939]: pgmap v135: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-09T20:56:15.923 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:56:15 vm01 ceph-mon[53402]: pgmap v135: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-09T20:56:15.957 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:56:15 vm08 ceph-mon[54611]: pgmap v135: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-09T20:56:17.795 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 09 20:56:17 vm06 ceph-mon[52939]: pgmap v136: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-09T20:56:17.923 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:56:17 vm01 ceph-mon[53402]: pgmap v136: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-09T20:56:17.957 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:56:17 vm08 ceph-mon[54611]: pgmap v136: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-09T20:56:19.795 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 09 20:56:19 vm06 ceph-mon[52939]: pgmap v137: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-09T20:56:19.923 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:56:19 vm01 ceph-mon[53402]: pgmap v137: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-09T20:56:19.957 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:56:19 vm08 ceph-mon[54611]: pgmap v137: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-09T20:56:21.795 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 09 20:56:21 vm06 ceph-mon[52939]: pgmap v138: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-09T20:56:21.923 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:56:21 vm01 ceph-mon[53402]: pgmap v138: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-09T20:56:21.957 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:56:21 vm08 ceph-mon[54611]: pgmap v138: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-09T20:56:23.795 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 09 20:56:23 vm06 ceph-mon[52939]: pgmap v139: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-09T20:56:23.923 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:56:23 vm01 ceph-mon[53402]: pgmap v139: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-09T20:56:23.957 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:56:23 vm08 ceph-mon[54611]: pgmap v139: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-09T20:56:25.795 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 09 20:56:25 vm06 ceph-mon[52939]: pgmap v140: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-09T20:56:25.923 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:56:25 vm01 ceph-mon[53402]: pgmap v140: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-09T20:56:25.957 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:56:25 vm08 ceph-mon[54611]: pgmap v140: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-09T20:56:27.796 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 09 20:56:27 vm06 ceph-mon[52939]: pgmap v141: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-09T20:56:27.923 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:56:27 vm01 ceph-mon[53402]: pgmap v141: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-09T20:56:27.957 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:56:27 vm08 ceph-mon[54611]: pgmap v141: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-09T20:56:27.964 INFO:teuthology.orchestra.run.vm01.stderr:+ ceph mon dump 2026-03-09T20:56:28.168 INFO:teuthology.orchestra.run.vm01.stdout:epoch 6 2026-03-09T20:56:28.169 INFO:teuthology.orchestra.run.vm01.stdout:fsid 9cb345a8-1bf9-11f1-a195-0375563c5891 2026-03-09T20:56:28.169 INFO:teuthology.orchestra.run.vm01.stdout:last_changed 2026-03-09T20:53:27.400224+0000 2026-03-09T20:56:28.169 INFO:teuthology.orchestra.run.vm01.stdout:created 2026-03-09T20:51:41.312920+0000 2026-03-09T20:56:28.169 INFO:teuthology.orchestra.run.vm01.stdout:min_mon_release 19 (squid) 2026-03-09T20:56:28.169 INFO:teuthology.orchestra.run.vm01.stdout:election_strategy: 1 2026-03-09T20:56:28.169 INFO:teuthology.orchestra.run.vm01.stdout:0: [v2:192.168.123.101:3300/0,v1:192.168.123.101:6789/0] mon.a; crush_location {datacenter=a} 2026-03-09T20:56:28.169 INFO:teuthology.orchestra.run.vm01.stdout:1: [v2:192.168.123.108:3300/0,v1:192.168.123.108:6789/0] mon.c; crush_location {datacenter=a,rack=3} 2026-03-09T20:56:28.169 INFO:teuthology.orchestra.run.vm01.stdout:2: [v2:192.168.123.106:3300/0,v1:192.168.123.106:6789/0] mon.b; crush_location {datacenter=b,rack=2} 2026-03-09T20:56:28.169 INFO:teuthology.orchestra.run.vm01.stderr:dumped monmap epoch 6 2026-03-09T20:56:28.177 INFO:teuthology.orchestra.run.vm01.stderr:+ ceph mon dump --format json 2026-03-09T20:56:28.376 INFO:teuthology.orchestra.run.vm01.stdout: 2026-03-09T20:56:28.376 INFO:teuthology.orchestra.run.vm01.stdout:{"epoch":6,"fsid":"9cb345a8-1bf9-11f1-a195-0375563c5891","modified":"2026-03-09T20:53:27.400224Z","created":"2026-03-09T20:51:41.312920Z","min_mon_release":19,"min_mon_release_name":"squid","election_strategy":1,"disallowed_leaders":"","stretch_mode":false,"tiebreaker_mon":"","removed_ranks":"","features":{"persistent":["kraken","luminous","mimic","osdmap-prune","nautilus","octopus","pacific","elector-pinging","quincy","reef","squid"],"optional":[]},"mons":[{"rank":0,"name":"a","public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.101:3300","nonce":0},{"type":"v1","addr":"192.168.123.101:6789","nonce":0}]},"addr":"192.168.123.101:6789/0","public_addr":"192.168.123.101:6789/0","priority":0,"weight":0,"crush_location":"{datacenter=a}"},{"rank":1,"name":"c","public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.108:3300","nonce":0},{"type":"v1","addr":"192.168.123.108:6789","nonce":0}]},"addr":"192.168.123.108:6789/0","public_addr":"192.168.123.108:6789/0","priority":0,"weight":0,"crush_location":"{datacenter=a,rack=3}"},{"rank":2,"name":"b","public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.106:3300","nonce":0},{"type":"v1","addr":"192.168.123.106:6789","nonce":0}]},"addr":"192.168.123.106:6789/0","public_addr":"192.168.123.106:6789/0","priority":0,"weight":0,"crush_location":"{datacenter=b,rack=2}"}],"quorum":[0,1,2]} 2026-03-09T20:56:28.376 INFO:teuthology.orchestra.run.vm01.stderr:dumped monmap epoch 6 2026-03-09T20:56:28.385 INFO:teuthology.orchestra.run.vm01.stderr:+ read monid 2026-03-09T20:56:28.385 INFO:teuthology.orchestra.run.vm01.stderr:+ read crushloc 2026-03-09T20:56:28.385 INFO:teuthology.orchestra.run.vm01.stderr:+ jq --arg monid a --arg crushloc '{datacenter=a}' -e '.mons | .[] | select(.name == $monid) | .crush_location == $crushloc' 2026-03-09T20:56:28.385 INFO:teuthology.orchestra.run.vm01.stderr:+ ceph mon dump --format json 2026-03-09T20:56:28.569 INFO:teuthology.orchestra.run.vm01.stderr:dumped monmap epoch 6 2026-03-09T20:56:28.579 INFO:teuthology.orchestra.run.vm01.stdout:true 2026-03-09T20:56:28.579 INFO:teuthology.orchestra.run.vm01.stderr:+ read monid 2026-03-09T20:56:28.579 INFO:teuthology.orchestra.run.vm01.stderr:+ read crushloc 2026-03-09T20:56:28.580 INFO:teuthology.orchestra.run.vm01.stderr:+ jq --arg monid b --arg crushloc '{datacenter=b,rack=2}' -e '.mons | .[] | select(.name == $monid) | .crush_location == $crushloc' 2026-03-09T20:56:28.580 INFO:teuthology.orchestra.run.vm01.stderr:+ ceph mon dump --format json 2026-03-09T20:56:28.673 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:56:28 vm01 ceph-mon[53402]: from='client.? 192.168.123.101:0/3018511921' entity='client.admin' cmd=[{"prefix": "mon dump"}]: dispatch 2026-03-09T20:56:28.673 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:56:28 vm01 ceph-mon[53402]: from='client.? 192.168.123.101:0/792168454' entity='client.admin' cmd=[{"prefix": "mon dump", "format": "json"}]: dispatch 2026-03-09T20:56:28.789 INFO:teuthology.orchestra.run.vm01.stderr:dumped monmap epoch 6 2026-03-09T20:56:28.795 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 09 20:56:28 vm06 ceph-mon[52939]: from='client.? 192.168.123.101:0/3018511921' entity='client.admin' cmd=[{"prefix": "mon dump"}]: dispatch 2026-03-09T20:56:28.795 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 09 20:56:28 vm06 ceph-mon[52939]: from='client.? 192.168.123.101:0/792168454' entity='client.admin' cmd=[{"prefix": "mon dump", "format": "json"}]: dispatch 2026-03-09T20:56:28.799 INFO:teuthology.orchestra.run.vm01.stdout:true 2026-03-09T20:56:28.799 INFO:teuthology.orchestra.run.vm01.stderr:+ read monid 2026-03-09T20:56:28.799 INFO:teuthology.orchestra.run.vm01.stderr:+ read crushloc 2026-03-09T20:56:28.800 INFO:teuthology.orchestra.run.vm01.stderr:+ ceph mon dump --format json 2026-03-09T20:56:28.800 INFO:teuthology.orchestra.run.vm01.stderr:+ jq --arg monid c --arg crushloc '{datacenter=a,rack=3}' -e '.mons | .[] | select(.name == $monid) | .crush_location == $crushloc' 2026-03-09T20:56:28.957 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:56:28 vm08 ceph-mon[54611]: from='client.? 192.168.123.101:0/3018511921' entity='client.admin' cmd=[{"prefix": "mon dump"}]: dispatch 2026-03-09T20:56:28.957 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:56:28 vm08 ceph-mon[54611]: from='client.? 192.168.123.101:0/792168454' entity='client.admin' cmd=[{"prefix": "mon dump", "format": "json"}]: dispatch 2026-03-09T20:56:28.988 INFO:teuthology.orchestra.run.vm01.stderr:dumped monmap epoch 6 2026-03-09T20:56:28.997 INFO:teuthology.orchestra.run.vm01.stdout:true 2026-03-09T20:56:28.998 INFO:teuthology.orchestra.run.vm01.stderr:+ read monid 2026-03-09T20:56:29.159 DEBUG:teuthology.run_tasks:Unwinding manager cephadm 2026-03-09T20:56:29.161 INFO:tasks.cephadm:Teardown begin 2026-03-09T20:56:29.161 DEBUG:teuthology.orchestra.run.vm01:> sudo rm -f /etc/ceph/ceph.conf /etc/ceph/ceph.client.admin.keyring 2026-03-09T20:56:29.191 DEBUG:teuthology.orchestra.run.vm06:> sudo rm -f /etc/ceph/ceph.conf /etc/ceph/ceph.client.admin.keyring 2026-03-09T20:56:29.224 DEBUG:teuthology.orchestra.run.vm08:> sudo rm -f /etc/ceph/ceph.conf /etc/ceph/ceph.client.admin.keyring 2026-03-09T20:56:29.251 INFO:tasks.cephadm:Disabling cephadm mgr module 2026-03-09T20:56:29.251 DEBUG:teuthology.orchestra.run.vm01:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 9cb345a8-1bf9-11f1-a195-0375563c5891 -- ceph mgr module disable cephadm 2026-03-09T20:56:29.433 INFO:teuthology.orchestra.run.vm01.stderr:Inferring config /var/lib/ceph/9cb345a8-1bf9-11f1-a195-0375563c5891/mon.a/config 2026-03-09T20:56:29.453 INFO:teuthology.orchestra.run.vm01.stderr:Error: statfs /etc/ceph/ceph.client.admin.keyring: no such file or directory 2026-03-09T20:56:29.476 DEBUG:teuthology.orchestra.run:got remote process result: 125 2026-03-09T20:56:29.477 INFO:tasks.cephadm:Cleaning up testdir ceph.* files... 2026-03-09T20:56:29.477 DEBUG:teuthology.orchestra.run.vm01:> rm -f /home/ubuntu/cephtest/seed.ceph.conf /home/ubuntu/cephtest/ceph.pub 2026-03-09T20:56:29.495 DEBUG:teuthology.orchestra.run.vm06:> rm -f /home/ubuntu/cephtest/seed.ceph.conf /home/ubuntu/cephtest/ceph.pub 2026-03-09T20:56:29.502 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:56:29 vm08 ceph-mon[54611]: from='client.? 192.168.123.101:0/1327028988' entity='client.admin' cmd=[{"prefix": "mon dump", "format": "json"}]: dispatch 2026-03-09T20:56:29.502 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:56:29 vm08 ceph-mon[54611]: from='client.? 192.168.123.101:0/131853574' entity='client.admin' cmd=[{"prefix": "mon dump", "format": "json"}]: dispatch 2026-03-09T20:56:29.502 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:56:29 vm08 ceph-mon[54611]: from='client.? 192.168.123.101:0/2204246205' entity='client.admin' cmd=[{"prefix": "mon dump", "format": "json"}]: dispatch 2026-03-09T20:56:29.502 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:56:29 vm08 ceph-mon[54611]: pgmap v142: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-09T20:56:29.503 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 09 20:56:29 vm06 ceph-mon[52939]: from='client.? 192.168.123.101:0/1327028988' entity='client.admin' cmd=[{"prefix": "mon dump", "format": "json"}]: dispatch 2026-03-09T20:56:29.503 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 09 20:56:29 vm06 ceph-mon[52939]: from='client.? 192.168.123.101:0/131853574' entity='client.admin' cmd=[{"prefix": "mon dump", "format": "json"}]: dispatch 2026-03-09T20:56:29.503 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 09 20:56:29 vm06 ceph-mon[52939]: from='client.? 192.168.123.101:0/2204246205' entity='client.admin' cmd=[{"prefix": "mon dump", "format": "json"}]: dispatch 2026-03-09T20:56:29.503 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 09 20:56:29 vm06 ceph-mon[52939]: pgmap v142: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-09T20:56:29.516 DEBUG:teuthology.orchestra.run.vm08:> rm -f /home/ubuntu/cephtest/seed.ceph.conf /home/ubuntu/cephtest/ceph.pub 2026-03-09T20:56:29.535 INFO:tasks.cephadm:Stopping all daemons... 2026-03-09T20:56:29.535 INFO:tasks.cephadm.mon.a:Stopping mon.a... 2026-03-09T20:56:29.536 DEBUG:teuthology.orchestra.run.vm01:> sudo systemctl stop ceph-9cb345a8-1bf9-11f1-a195-0375563c5891@mon.a 2026-03-09T20:56:29.873 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:56:29 vm01 ceph-mon[53402]: from='client.? 192.168.123.101:0/1327028988' entity='client.admin' cmd=[{"prefix": "mon dump", "format": "json"}]: dispatch 2026-03-09T20:56:29.873 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:56:29 vm01 ceph-mon[53402]: from='client.? 192.168.123.101:0/131853574' entity='client.admin' cmd=[{"prefix": "mon dump", "format": "json"}]: dispatch 2026-03-09T20:56:29.873 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:56:29 vm01 ceph-mon[53402]: from='client.? 192.168.123.101:0/2204246205' entity='client.admin' cmd=[{"prefix": "mon dump", "format": "json"}]: dispatch 2026-03-09T20:56:29.873 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:56:29 vm01 ceph-mon[53402]: pgmap v142: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-09T20:56:29.873 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:56:29 vm01 systemd[1]: Stopping Ceph mon.a for 9cb345a8-1bf9-11f1-a195-0375563c5891... 2026-03-09T20:56:29.873 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:56:29 vm01 ceph-9cb345a8-1bf9-11f1-a195-0375563c5891-mon-a[53378]: 2026-03-09T20:56:29.656+0000 7fb80ecad640 -1 received signal: Terminated from /run/podman-init -- /usr/bin/ceph-mon -n mon.a -f --setuser ceph --setgroup ceph --default-log-to-file=false --default-log-to-journald=true --default-log-to-stderr=false --default-mon-cluster-log-to-file=false --default-mon-cluster-log-to-journald=true --default-mon-cluster-log-to-stderr=false (PID: 1) UID: 0 2026-03-09T20:56:29.873 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:56:29 vm01 ceph-9cb345a8-1bf9-11f1-a195-0375563c5891-mon-a[53378]: 2026-03-09T20:56:29.656+0000 7fb80ecad640 -1 mon.a@0(leader) e6 *** Got Signal Terminated *** 2026-03-09T20:56:29.873 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 20:56:29 vm01 podman[71547]: 2026-03-09 20:56:29.872168523 +0000 UTC m=+0.235355239 container died 6532f28f79f17d6052d58fbcbb6916057ce4c8090ab12f74503d522d2fc4c18d (image=quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df, name=ceph-9cb345a8-1bf9-11f1-a195-0375563c5891-mon-a, CEPH_REF=squid, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.license=GPLv2, FROM_IMAGE=quay.io/centos/centos:stream9, OSD_FLAVOR=default, CEPH_GIT_REPO=https://github.com/ceph/ceph-ci.git, org.opencontainers.image.documentation=https://docs.ceph.com/, org.label-schema.schema-version=1.0, io.buildah.version=1.41.3, ceph=True, GANESHA_REPO_BASEURL=https://buildlogs.centos.org/centos/$releasever-stream/storage/$basearch/nfsganesha-5/, org.label-schema.build-date=20260223, org.label-schema.vendor=CentOS, org.opencontainers.image.authors=Ceph Release Team , CEPH_SHA1=e911bdebe5c8faa3800735d1568fcdca65db60df) 2026-03-09T20:56:30.055 DEBUG:teuthology.orchestra.run.vm01:> sudo pkill -f 'journalctl -f -n 0 -u ceph-9cb345a8-1bf9-11f1-a195-0375563c5891@mon.a.service' 2026-03-09T20:56:30.093 DEBUG:teuthology.orchestra.run:got remote process result: None 2026-03-09T20:56:30.093 INFO:tasks.cephadm.mon.a:Stopped mon.a 2026-03-09T20:56:30.093 INFO:tasks.cephadm.mon.c:Stopping mon.b... 2026-03-09T20:56:30.093 DEBUG:teuthology.orchestra.run.vm06:> sudo systemctl stop ceph-9cb345a8-1bf9-11f1-a195-0375563c5891@mon.b 2026-03-09T20:56:30.446 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 09 20:56:30 vm06 systemd[1]: Stopping Ceph mon.b for 9cb345a8-1bf9-11f1-a195-0375563c5891... 2026-03-09T20:56:30.446 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 09 20:56:30 vm06 ceph-9cb345a8-1bf9-11f1-a195-0375563c5891-mon-b[52916]: 2026-03-09T20:56:30.202+0000 7f4bc4b10640 -1 received signal: Terminated from /run/podman-init -- /usr/bin/ceph-mon -n mon.b -f --setuser ceph --setgroup ceph --default-log-to-file=false --default-log-to-journald=true --default-log-to-stderr=false --default-mon-cluster-log-to-file=false --default-mon-cluster-log-to-journald=true --default-mon-cluster-log-to-stderr=false (PID: 1) UID: 0 2026-03-09T20:56:30.446 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 09 20:56:30 vm06 ceph-9cb345a8-1bf9-11f1-a195-0375563c5891-mon-b[52916]: 2026-03-09T20:56:30.202+0000 7f4bc4b10640 -1 mon.b@2(peon) e6 *** Got Signal Terminated *** 2026-03-09T20:56:30.446 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 09 20:56:30 vm06 podman[61261]: 2026-03-09 20:56:30.285809901 +0000 UTC m=+0.099462068 container died d29f2c50f9108b148c97ff12926e75b06fe479cbb001eebd82c147a078554edc (image=quay.ceph.io/ceph-ci/ceph@sha256:8fda260ab1d2d3118a1622f7df75f44f285dfe74e71793626152a711c12bf2cc, name=ceph-9cb345a8-1bf9-11f1-a195-0375563c5891-mon-b, ceph=True, org.label-schema.schema-version=1.0, CEPH_REF=squid, org.label-schema.build-date=20260223, org.label-schema.license=GPLv2, FROM_IMAGE=quay.io/centos/centos:stream9, CEPH_GIT_REPO=https://github.com/ceph/ceph-ci.git, org.opencontainers.image.documentation=https://docs.ceph.com/, OSD_FLAVOR=default, org.label-schema.vendor=CentOS, GANESHA_REPO_BASEURL=https://buildlogs.centos.org/centos/$releasever-stream/storage/$basearch/nfsganesha-5/, CEPH_SHA1=e911bdebe5c8faa3800735d1568fcdca65db60df, io.buildah.version=1.41.3, org.label-schema.name=CentOS Stream 9 Base Image, org.opencontainers.image.authors=Ceph Release Team ) 2026-03-09T20:56:30.514 DEBUG:teuthology.orchestra.run.vm06:> sudo pkill -f 'journalctl -f -n 0 -u ceph-9cb345a8-1bf9-11f1-a195-0375563c5891@mon.b.service' 2026-03-09T20:56:30.549 DEBUG:teuthology.orchestra.run:got remote process result: None 2026-03-09T20:56:30.549 INFO:tasks.cephadm.mon.c:Stopped mon.b 2026-03-09T20:56:30.549 INFO:tasks.cephadm.mon.c:Stopping mon.c... 2026-03-09T20:56:30.550 DEBUG:teuthology.orchestra.run.vm08:> sudo systemctl stop ceph-9cb345a8-1bf9-11f1-a195-0375563c5891@mon.c 2026-03-09T20:56:30.873 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:56:30 vm08 systemd[1]: Stopping Ceph mon.c for 9cb345a8-1bf9-11f1-a195-0375563c5891... 2026-03-09T20:56:30.873 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:56:30 vm08 ceph-9cb345a8-1bf9-11f1-a195-0375563c5891-mon-c[54586]: 2026-03-09T20:56:30.644+0000 7f0cae370640 -1 received signal: Terminated from /run/podman-init -- /usr/bin/ceph-mon -n mon.c -f --setuser ceph --setgroup ceph --default-log-to-file=false --default-log-to-journald=true --default-log-to-stderr=false --default-mon-cluster-log-to-file=false --default-mon-cluster-log-to-journald=true --default-mon-cluster-log-to-stderr=false (PID: 1) UID: 0 2026-03-09T20:56:30.873 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:56:30 vm08 ceph-9cb345a8-1bf9-11f1-a195-0375563c5891-mon-c[54586]: 2026-03-09T20:56:30.644+0000 7f0cae370640 -1 mon.c@1(peon) e6 *** Got Signal Terminated *** 2026-03-09T20:56:31.061 DEBUG:teuthology.orchestra.run.vm08:> sudo pkill -f 'journalctl -f -n 0 -u ceph-9cb345a8-1bf9-11f1-a195-0375563c5891@mon.c.service' 2026-03-09T20:56:31.094 DEBUG:teuthology.orchestra.run:got remote process result: None 2026-03-09T20:56:31.094 INFO:tasks.cephadm.mon.c:Stopped mon.c 2026-03-09T20:56:31.094 INFO:tasks.cephadm.mgr.a:Stopping mgr.a... 2026-03-09T20:56:31.095 DEBUG:teuthology.orchestra.run.vm01:> sudo systemctl stop ceph-9cb345a8-1bf9-11f1-a195-0375563c5891@mgr.a 2026-03-09T20:56:31.370 INFO:journalctl@ceph.mgr.a.vm01.stdout:Mar 09 20:56:31 vm01 systemd[1]: Stopping Ceph mgr.a for 9cb345a8-1bf9-11f1-a195-0375563c5891... 2026-03-09T20:56:31.370 INFO:journalctl@ceph.mgr.a.vm01.stdout:Mar 09 20:56:31 vm01 podman[71661]: 2026-03-09 20:56:31.246266069 +0000 UTC m=+0.056724296 container died 9fc252e619ba3254f69778a8ecbaac0c12aa7a6d41ba3880147c41cb162f5d15 (image=quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df, name=ceph-9cb345a8-1bf9-11f1-a195-0375563c5891-mgr-a, CEPH_GIT_REPO=https://github.com/ceph/ceph-ci.git, org.opencontainers.image.authors=Ceph Release Team , org.label-schema.name=CentOS Stream 9 Base Image, OSD_FLAVOR=default, io.buildah.version=1.41.3, FROM_IMAGE=quay.io/centos/centos:stream9, org.label-schema.build-date=20260223, org.label-schema.license=GPLv2, org.label-schema.schema-version=1.0, CEPH_SHA1=e911bdebe5c8faa3800735d1568fcdca65db60df, GANESHA_REPO_BASEURL=https://buildlogs.centos.org/centos/$releasever-stream/storage/$basearch/nfsganesha-5/, org.label-schema.vendor=CentOS, ceph=True, CEPH_REF=squid, org.opencontainers.image.documentation=https://docs.ceph.com/) 2026-03-09T20:56:31.433 DEBUG:teuthology.orchestra.run.vm01:> sudo pkill -f 'journalctl -f -n 0 -u ceph-9cb345a8-1bf9-11f1-a195-0375563c5891@mgr.a.service' 2026-03-09T20:56:31.463 DEBUG:teuthology.orchestra.run:got remote process result: None 2026-03-09T20:56:31.463 INFO:tasks.cephadm.mgr.a:Stopped mgr.a 2026-03-09T20:56:31.463 INFO:tasks.cephadm.mgr.b:Stopping mgr.b... 2026-03-09T20:56:31.463 DEBUG:teuthology.orchestra.run.vm06:> sudo systemctl stop ceph-9cb345a8-1bf9-11f1-a195-0375563c5891@mgr.b 2026-03-09T20:56:31.502 INFO:journalctl@ceph.mgr.b.vm06.stdout:Mar 09 20:56:31 vm06 systemd[1]: Stopping Ceph mgr.b for 9cb345a8-1bf9-11f1-a195-0375563c5891... 2026-03-09T20:56:31.789 INFO:journalctl@ceph.mgr.b.vm06.stdout:Mar 09 20:56:31 vm06 podman[61375]: 2026-03-09 20:56:31.61273878 +0000 UTC m=+0.053764682 container died 355207d5317a091a502516e081024127f73f66af33082f8711fff45a57f9714d (image=quay.ceph.io/ceph-ci/ceph@sha256:8fda260ab1d2d3118a1622f7df75f44f285dfe74e71793626152a711c12bf2cc, name=ceph-9cb345a8-1bf9-11f1-a195-0375563c5891-mgr-b, GANESHA_REPO_BASEURL=https://buildlogs.centos.org/centos/$releasever-stream/storage/$basearch/nfsganesha-5/, OSD_FLAVOR=default, org.label-schema.build-date=20260223, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.license=GPLv2, org.label-schema.schema-version=1.0, CEPH_GIT_REPO=https://github.com/ceph/ceph-ci.git, org.opencontainers.image.authors=Ceph Release Team , org.opencontainers.image.documentation=https://docs.ceph.com/, FROM_IMAGE=quay.io/centos/centos:stream9, org.label-schema.vendor=CentOS, ceph=True, io.buildah.version=1.41.3, CEPH_REF=squid, CEPH_SHA1=e911bdebe5c8faa3800735d1568fcdca65db60df) 2026-03-09T20:56:31.790 INFO:journalctl@ceph.mgr.b.vm06.stdout:Mar 09 20:56:31 vm06 podman[61375]: 2026-03-09 20:56:31.738866571 +0000 UTC m=+0.179892473 container remove 355207d5317a091a502516e081024127f73f66af33082f8711fff45a57f9714d (image=quay.ceph.io/ceph-ci/ceph@sha256:8fda260ab1d2d3118a1622f7df75f44f285dfe74e71793626152a711c12bf2cc, name=ceph-9cb345a8-1bf9-11f1-a195-0375563c5891-mgr-b, CEPH_SHA1=e911bdebe5c8faa3800735d1568fcdca65db60df, io.buildah.version=1.41.3, CEPH_REF=squid, org.label-schema.build-date=20260223, org.label-schema.vendor=CentOS, OSD_FLAVOR=default, org.opencontainers.image.authors=Ceph Release Team , CEPH_GIT_REPO=https://github.com/ceph/ceph-ci.git, org.label-schema.name=CentOS Stream 9 Base Image, org.opencontainers.image.documentation=https://docs.ceph.com/, ceph=True, org.label-schema.schema-version=1.0, FROM_IMAGE=quay.io/centos/centos:stream9, GANESHA_REPO_BASEURL=https://buildlogs.centos.org/centos/$releasever-stream/storage/$basearch/nfsganesha-5/, org.label-schema.license=GPLv2) 2026-03-09T20:56:31.790 INFO:journalctl@ceph.mgr.b.vm06.stdout:Mar 09 20:56:31 vm06 bash[61375]: ceph-9cb345a8-1bf9-11f1-a195-0375563c5891-mgr-b 2026-03-09T20:56:31.790 INFO:journalctl@ceph.mgr.b.vm06.stdout:Mar 09 20:56:31 vm06 systemd[1]: ceph-9cb345a8-1bf9-11f1-a195-0375563c5891@mgr.b.service: Main process exited, code=exited, status=143/n/a 2026-03-09T20:56:31.797 DEBUG:teuthology.orchestra.run.vm06:> sudo pkill -f 'journalctl -f -n 0 -u ceph-9cb345a8-1bf9-11f1-a195-0375563c5891@mgr.b.service' 2026-03-09T20:56:31.869 DEBUG:teuthology.orchestra.run:got remote process result: None 2026-03-09T20:56:31.869 INFO:tasks.cephadm.mgr.b:Stopped mgr.b 2026-03-09T20:56:31.869 INFO:tasks.cephadm.osd.0:Stopping osd.0... 2026-03-09T20:56:31.869 DEBUG:teuthology.orchestra.run.vm01:> sudo systemctl stop ceph-9cb345a8-1bf9-11f1-a195-0375563c5891@osd.0 2026-03-09T20:56:32.173 INFO:journalctl@ceph.osd.0.vm01.stdout:Mar 09 20:56:31 vm01 systemd[1]: Stopping Ceph osd.0 for 9cb345a8-1bf9-11f1-a195-0375563c5891... 2026-03-09T20:56:32.173 INFO:journalctl@ceph.osd.0.vm01.stdout:Mar 09 20:56:31 vm01 ceph-9cb345a8-1bf9-11f1-a195-0375563c5891-osd-0[64225]: 2026-03-09T20:56:31.979+0000 7fd6080ba640 -1 received signal: Terminated from /run/podman-init -- /usr/bin/ceph-osd -n osd.0 -f --setuser ceph --setgroup ceph --default-log-to-file=false --default-log-to-journald=true --default-log-to-stderr=false (PID: 1) UID: 0 2026-03-09T20:56:32.173 INFO:journalctl@ceph.osd.0.vm01.stdout:Mar 09 20:56:31 vm01 ceph-9cb345a8-1bf9-11f1-a195-0375563c5891-osd-0[64225]: 2026-03-09T20:56:31.979+0000 7fd6080ba640 -1 osd.0 21 *** Got signal Terminated *** 2026-03-09T20:56:32.173 INFO:journalctl@ceph.osd.0.vm01.stdout:Mar 09 20:56:31 vm01 ceph-9cb345a8-1bf9-11f1-a195-0375563c5891-osd-0[64225]: 2026-03-09T20:56:31.979+0000 7fd6080ba640 -1 osd.0 21 *** Immediate shutdown (osd_fast_shutdown=true) *** 2026-03-09T20:56:37.293 INFO:journalctl@ceph.osd.0.vm01.stdout:Mar 09 20:56:37 vm01 podman[71773]: 2026-03-09 20:56:37.00200148 +0000 UTC m=+5.040688065 container died 0d379550ff05eac54b7473f1e90caca92d79fca5e5eb472bf36858ea4548859d (image=quay.ceph.io/ceph-ci/ceph@sha256:8fda260ab1d2d3118a1622f7df75f44f285dfe74e71793626152a711c12bf2cc, name=ceph-9cb345a8-1bf9-11f1-a195-0375563c5891-osd-0, ceph=True, OSD_FLAVOR=default, FROM_IMAGE=quay.io/centos/centos:stream9, org.label-schema.license=GPLv2, GANESHA_REPO_BASEURL=https://buildlogs.centos.org/centos/$releasever-stream/storage/$basearch/nfsganesha-5/, io.buildah.version=1.41.3, org.label-schema.name=CentOS Stream 9 Base Image, CEPH_REF=squid, org.label-schema.vendor=CentOS, CEPH_GIT_REPO=https://github.com/ceph/ceph-ci.git, org.label-schema.build-date=20260223, CEPH_SHA1=e911bdebe5c8faa3800735d1568fcdca65db60df, org.opencontainers.image.authors=Ceph Release Team , org.opencontainers.image.documentation=https://docs.ceph.com/, org.label-schema.schema-version=1.0) 2026-03-09T20:56:37.294 INFO:journalctl@ceph.osd.0.vm01.stdout:Mar 09 20:56:37 vm01 podman[71773]: 2026-03-09 20:56:37.133940027 +0000 UTC m=+5.172626612 container remove 0d379550ff05eac54b7473f1e90caca92d79fca5e5eb472bf36858ea4548859d (image=quay.ceph.io/ceph-ci/ceph@sha256:8fda260ab1d2d3118a1622f7df75f44f285dfe74e71793626152a711c12bf2cc, name=ceph-9cb345a8-1bf9-11f1-a195-0375563c5891-osd-0, CEPH_GIT_REPO=https://github.com/ceph/ceph-ci.git, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.schema-version=1.0, ceph=True, GANESHA_REPO_BASEURL=https://buildlogs.centos.org/centos/$releasever-stream/storage/$basearch/nfsganesha-5/, OSD_FLAVOR=default, io.buildah.version=1.41.3, org.label-schema.build-date=20260223, CEPH_SHA1=e911bdebe5c8faa3800735d1568fcdca65db60df, FROM_IMAGE=quay.io/centos/centos:stream9, org.opencontainers.image.documentation=https://docs.ceph.com/, org.label-schema.vendor=CentOS, org.label-schema.license=GPLv2, CEPH_REF=squid, org.opencontainers.image.authors=Ceph Release Team ) 2026-03-09T20:56:37.294 INFO:journalctl@ceph.osd.0.vm01.stdout:Mar 09 20:56:37 vm01 bash[71773]: ceph-9cb345a8-1bf9-11f1-a195-0375563c5891-osd-0 2026-03-09T20:56:37.618 INFO:journalctl@ceph.osd.0.vm01.stdout:Mar 09 20:56:37 vm01 podman[71850]: 2026-03-09 20:56:37.293554289 +0000 UTC m=+0.019188586 container create 1e159c2caeadd414d1d026f98f4b4747c6d85d2f8bc5c92d1432310e62b853e9 (image=quay.ceph.io/ceph-ci/ceph@sha256:8fda260ab1d2d3118a1622f7df75f44f285dfe74e71793626152a711c12bf2cc, name=ceph-9cb345a8-1bf9-11f1-a195-0375563c5891-osd-0-deactivate, CEPH_REF=squid, ceph=True, CEPH_GIT_REPO=https://github.com/ceph/ceph-ci.git, GANESHA_REPO_BASEURL=https://buildlogs.centos.org/centos/$releasever-stream/storage/$basearch/nfsganesha-5/, OSD_FLAVOR=default, io.buildah.version=1.41.3, org.label-schema.vendor=CentOS, org.label-schema.schema-version=1.0, org.label-schema.name=CentOS Stream 9 Base Image, org.opencontainers.image.documentation=https://docs.ceph.com/, org.label-schema.build-date=20260223, org.opencontainers.image.authors=Ceph Release Team , org.label-schema.license=GPLv2, CEPH_SHA1=e911bdebe5c8faa3800735d1568fcdca65db60df, FROM_IMAGE=quay.io/centos/centos:stream9) 2026-03-09T20:56:37.619 INFO:journalctl@ceph.osd.0.vm01.stdout:Mar 09 20:56:37 vm01 podman[71850]: 2026-03-09 20:56:37.345285472 +0000 UTC m=+0.070919769 container init 1e159c2caeadd414d1d026f98f4b4747c6d85d2f8bc5c92d1432310e62b853e9 (image=quay.ceph.io/ceph-ci/ceph@sha256:8fda260ab1d2d3118a1622f7df75f44f285dfe74e71793626152a711c12bf2cc, name=ceph-9cb345a8-1bf9-11f1-a195-0375563c5891-osd-0-deactivate, CEPH_SHA1=e911bdebe5c8faa3800735d1568fcdca65db60df, GANESHA_REPO_BASEURL=https://buildlogs.centos.org/centos/$releasever-stream/storage/$basearch/nfsganesha-5/, org.label-schema.build-date=20260223, FROM_IMAGE=quay.io/centos/centos:stream9, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.license=GPLv2, org.opencontainers.image.documentation=https://docs.ceph.com/, io.buildah.version=1.41.3, CEPH_GIT_REPO=https://github.com/ceph/ceph-ci.git, OSD_FLAVOR=default, org.label-schema.vendor=CentOS, org.label-schema.schema-version=1.0, CEPH_REF=squid, org.opencontainers.image.authors=Ceph Release Team , ceph=True) 2026-03-09T20:56:37.619 INFO:journalctl@ceph.osd.0.vm01.stdout:Mar 09 20:56:37 vm01 podman[71850]: 2026-03-09 20:56:37.353824153 +0000 UTC m=+0.079458450 container start 1e159c2caeadd414d1d026f98f4b4747c6d85d2f8bc5c92d1432310e62b853e9 (image=quay.ceph.io/ceph-ci/ceph@sha256:8fda260ab1d2d3118a1622f7df75f44f285dfe74e71793626152a711c12bf2cc, name=ceph-9cb345a8-1bf9-11f1-a195-0375563c5891-osd-0-deactivate, org.label-schema.license=GPLv2, CEPH_SHA1=e911bdebe5c8faa3800735d1568fcdca65db60df, GANESHA_REPO_BASEURL=https://buildlogs.centos.org/centos/$releasever-stream/storage/$basearch/nfsganesha-5/, OSD_FLAVOR=default, CEPH_GIT_REPO=https://github.com/ceph/ceph-ci.git, org.label-schema.vendor=CentOS, org.label-schema.schema-version=1.0, org.label-schema.build-date=20260223, org.opencontainers.image.documentation=https://docs.ceph.com/, FROM_IMAGE=quay.io/centos/centos:stream9, CEPH_REF=squid, ceph=True, io.buildah.version=1.41.3, org.opencontainers.image.authors=Ceph Release Team , org.label-schema.name=CentOS Stream 9 Base Image) 2026-03-09T20:56:37.619 INFO:journalctl@ceph.osd.0.vm01.stdout:Mar 09 20:56:37 vm01 podman[71850]: 2026-03-09 20:56:37.358972795 +0000 UTC m=+0.084607092 container attach 1e159c2caeadd414d1d026f98f4b4747c6d85d2f8bc5c92d1432310e62b853e9 (image=quay.ceph.io/ceph-ci/ceph@sha256:8fda260ab1d2d3118a1622f7df75f44f285dfe74e71793626152a711c12bf2cc, name=ceph-9cb345a8-1bf9-11f1-a195-0375563c5891-osd-0-deactivate, org.label-schema.schema-version=1.0, CEPH_REF=squid, org.label-schema.license=GPLv2, ceph=True, OSD_FLAVOR=default, org.label-schema.build-date=20260223, org.label-schema.vendor=CentOS, org.label-schema.name=CentOS Stream 9 Base Image, org.opencontainers.image.documentation=https://docs.ceph.com/, CEPH_SHA1=e911bdebe5c8faa3800735d1568fcdca65db60df, FROM_IMAGE=quay.io/centos/centos:stream9, org.opencontainers.image.authors=Ceph Release Team , CEPH_GIT_REPO=https://github.com/ceph/ceph-ci.git, GANESHA_REPO_BASEURL=https://buildlogs.centos.org/centos/$releasever-stream/storage/$basearch/nfsganesha-5/, io.buildah.version=1.41.3) 2026-03-09T20:56:37.619 INFO:journalctl@ceph.osd.0.vm01.stdout:Mar 09 20:56:37 vm01 podman[71850]: 2026-03-09 20:56:37.285858406 +0000 UTC m=+0.011492714 image pull 654f31e6858eb235bbece362255b685a945f2b6a367e2b88c4930c984fbb214c quay.ceph.io/ceph-ci/ceph@sha256:8fda260ab1d2d3118a1622f7df75f44f285dfe74e71793626152a711c12bf2cc 2026-03-09T20:56:37.619 INFO:journalctl@ceph.osd.0.vm01.stdout:Mar 09 20:56:37 vm01 conmon[71860]: conmon 1e159c2caeadd414d1d0 : Failed to open cgroups file: /sys/fs/cgroup/machine.slice/libpod-1e159c2caeadd414d1d026f98f4b4747c6d85d2f8bc5c92d1432310e62b853e9.scope/memory.events 2026-03-09T20:56:37.619 INFO:journalctl@ceph.osd.0.vm01.stdout:Mar 09 20:56:37 vm01 podman[71850]: 2026-03-09 20:56:37.490811042 +0000 UTC m=+0.216445339 container died 1e159c2caeadd414d1d026f98f4b4747c6d85d2f8bc5c92d1432310e62b853e9 (image=quay.ceph.io/ceph-ci/ceph@sha256:8fda260ab1d2d3118a1622f7df75f44f285dfe74e71793626152a711c12bf2cc, name=ceph-9cb345a8-1bf9-11f1-a195-0375563c5891-osd-0-deactivate, OSD_FLAVOR=default, org.label-schema.build-date=20260223, io.buildah.version=1.41.3, org.label-schema.schema-version=1.0, ceph=True, CEPH_REF=squid, org.label-schema.name=CentOS Stream 9 Base Image, CEPH_SHA1=e911bdebe5c8faa3800735d1568fcdca65db60df, GANESHA_REPO_BASEURL=https://buildlogs.centos.org/centos/$releasever-stream/storage/$basearch/nfsganesha-5/, org.opencontainers.image.documentation=https://docs.ceph.com/, FROM_IMAGE=quay.io/centos/centos:stream9, org.label-schema.vendor=CentOS, org.opencontainers.image.authors=Ceph Release Team , org.label-schema.license=GPLv2, CEPH_GIT_REPO=https://github.com/ceph/ceph-ci.git) 2026-03-09T20:56:37.636 DEBUG:teuthology.orchestra.run.vm01:> sudo pkill -f 'journalctl -f -n 0 -u ceph-9cb345a8-1bf9-11f1-a195-0375563c5891@osd.0.service' 2026-03-09T20:56:37.673 DEBUG:teuthology.orchestra.run:got remote process result: None 2026-03-09T20:56:37.673 INFO:tasks.cephadm.osd.0:Stopped osd.0 2026-03-09T20:56:37.673 INFO:tasks.cephadm.osd.1:Stopping osd.1... 2026-03-09T20:56:37.673 DEBUG:teuthology.orchestra.run.vm06:> sudo systemctl stop ceph-9cb345a8-1bf9-11f1-a195-0375563c5891@osd.1 2026-03-09T20:56:38.045 INFO:journalctl@ceph.osd.1.vm06.stdout:Mar 09 20:56:37 vm06 systemd[1]: Stopping Ceph osd.1 for 9cb345a8-1bf9-11f1-a195-0375563c5891... 2026-03-09T20:56:38.046 INFO:journalctl@ceph.osd.1.vm06.stdout:Mar 09 20:56:37 vm06 ceph-9cb345a8-1bf9-11f1-a195-0375563c5891-osd-1[57192]: 2026-03-09T20:56:37.781+0000 7f3ef6655640 -1 received signal: Terminated from /run/podman-init -- /usr/bin/ceph-osd -n osd.1 -f --setuser ceph --setgroup ceph --default-log-to-file=false --default-log-to-journald=true --default-log-to-stderr=false (PID: 1) UID: 0 2026-03-09T20:56:38.046 INFO:journalctl@ceph.osd.1.vm06.stdout:Mar 09 20:56:37 vm06 ceph-9cb345a8-1bf9-11f1-a195-0375563c5891-osd-1[57192]: 2026-03-09T20:56:37.781+0000 7f3ef6655640 -1 osd.1 21 *** Got signal Terminated *** 2026-03-09T20:56:38.046 INFO:journalctl@ceph.osd.1.vm06.stdout:Mar 09 20:56:37 vm06 ceph-9cb345a8-1bf9-11f1-a195-0375563c5891-osd-1[57192]: 2026-03-09T20:56:37.781+0000 7f3ef6655640 -1 osd.1 21 *** Immediate shutdown (osd_fast_shutdown=true) *** 2026-03-09T20:56:43.116 INFO:journalctl@ceph.osd.1.vm06.stdout:Mar 09 20:56:42 vm06 podman[61487]: 2026-03-09 20:56:42.815717714 +0000 UTC m=+5.049691085 container died 1efb5185c5db9397867a06fdf739315855f7cc03b4734ea18eeae99f57e4dddf (image=quay.ceph.io/ceph-ci/ceph@sha256:8fda260ab1d2d3118a1622f7df75f44f285dfe74e71793626152a711c12bf2cc, name=ceph-9cb345a8-1bf9-11f1-a195-0375563c5891-osd-1, CEPH_REF=squid, OSD_FLAVOR=default, org.label-schema.schema-version=1.0, FROM_IMAGE=quay.io/centos/centos:stream9, GANESHA_REPO_BASEURL=https://buildlogs.centos.org/centos/$releasever-stream/storage/$basearch/nfsganesha-5/, org.label-schema.build-date=20260223, org.label-schema.name=CentOS Stream 9 Base Image, ceph=True, org.label-schema.license=GPLv2, CEPH_SHA1=e911bdebe5c8faa3800735d1568fcdca65db60df, org.opencontainers.image.authors=Ceph Release Team , org.label-schema.vendor=CentOS, CEPH_GIT_REPO=https://github.com/ceph/ceph-ci.git, org.opencontainers.image.documentation=https://docs.ceph.com/, io.buildah.version=1.41.3) 2026-03-09T20:56:43.116 INFO:journalctl@ceph.osd.1.vm06.stdout:Mar 09 20:56:42 vm06 podman[61487]: 2026-03-09 20:56:42.94955059 +0000 UTC m=+5.183523961 container remove 1efb5185c5db9397867a06fdf739315855f7cc03b4734ea18eeae99f57e4dddf (image=quay.ceph.io/ceph-ci/ceph@sha256:8fda260ab1d2d3118a1622f7df75f44f285dfe74e71793626152a711c12bf2cc, name=ceph-9cb345a8-1bf9-11f1-a195-0375563c5891-osd-1, org.label-schema.license=GPLv2, OSD_FLAVOR=default, ceph=True, org.label-schema.schema-version=1.0, CEPH_REF=squid, org.opencontainers.image.authors=Ceph Release Team , org.label-schema.name=CentOS Stream 9 Base Image, org.opencontainers.image.documentation=https://docs.ceph.com/, io.buildah.version=1.41.3, CEPH_GIT_REPO=https://github.com/ceph/ceph-ci.git, org.label-schema.vendor=CentOS, FROM_IMAGE=quay.io/centos/centos:stream9, CEPH_SHA1=e911bdebe5c8faa3800735d1568fcdca65db60df, GANESHA_REPO_BASEURL=https://buildlogs.centos.org/centos/$releasever-stream/storage/$basearch/nfsganesha-5/, org.label-schema.build-date=20260223) 2026-03-09T20:56:43.116 INFO:journalctl@ceph.osd.1.vm06.stdout:Mar 09 20:56:42 vm06 bash[61487]: ceph-9cb345a8-1bf9-11f1-a195-0375563c5891-osd-1 2026-03-09T20:56:43.438 INFO:journalctl@ceph.osd.1.vm06.stdout:Mar 09 20:56:43 vm06 podman[61569]: 2026-03-09 20:56:43.115436419 +0000 UTC m=+0.029570232 container create c6ba29df1899403a7d765541951ac62b5d792af406e55c4e284880aa711f5e3d (image=quay.ceph.io/ceph-ci/ceph@sha256:8fda260ab1d2d3118a1622f7df75f44f285dfe74e71793626152a711c12bf2cc, name=ceph-9cb345a8-1bf9-11f1-a195-0375563c5891-osd-1-deactivate, ceph=True, OSD_FLAVOR=default, io.buildah.version=1.41.3, org.label-schema.vendor=CentOS, CEPH_GIT_REPO=https://github.com/ceph/ceph-ci.git, org.opencontainers.image.documentation=https://docs.ceph.com/, org.label-schema.license=GPLv2, org.label-schema.build-date=20260223, CEPH_SHA1=e911bdebe5c8faa3800735d1568fcdca65db60df, org.label-schema.name=CentOS Stream 9 Base Image, CEPH_REF=squid, FROM_IMAGE=quay.io/centos/centos:stream9, org.opencontainers.image.authors=Ceph Release Team , GANESHA_REPO_BASEURL=https://buildlogs.centos.org/centos/$releasever-stream/storage/$basearch/nfsganesha-5/, org.label-schema.schema-version=1.0) 2026-03-09T20:56:43.438 INFO:journalctl@ceph.osd.1.vm06.stdout:Mar 09 20:56:43 vm06 podman[61569]: 2026-03-09 20:56:43.175399593 +0000 UTC m=+0.089533397 container init c6ba29df1899403a7d765541951ac62b5d792af406e55c4e284880aa711f5e3d (image=quay.ceph.io/ceph-ci/ceph@sha256:8fda260ab1d2d3118a1622f7df75f44f285dfe74e71793626152a711c12bf2cc, name=ceph-9cb345a8-1bf9-11f1-a195-0375563c5891-osd-1-deactivate, ceph=True, GANESHA_REPO_BASEURL=https://buildlogs.centos.org/centos/$releasever-stream/storage/$basearch/nfsganesha-5/, org.label-schema.schema-version=1.0, OSD_FLAVOR=default, CEPH_GIT_REPO=https://github.com/ceph/ceph-ci.git, org.label-schema.build-date=20260223, FROM_IMAGE=quay.io/centos/centos:stream9, org.opencontainers.image.authors=Ceph Release Team , org.label-schema.vendor=CentOS, org.label-schema.name=CentOS Stream 9 Base Image, org.opencontainers.image.documentation=https://docs.ceph.com/, CEPH_REF=squid, org.label-schema.license=GPLv2, CEPH_SHA1=e911bdebe5c8faa3800735d1568fcdca65db60df, io.buildah.version=1.41.3) 2026-03-09T20:56:43.438 INFO:journalctl@ceph.osd.1.vm06.stdout:Mar 09 20:56:43 vm06 podman[61569]: 2026-03-09 20:56:43.181660616 +0000 UTC m=+0.095794410 container start c6ba29df1899403a7d765541951ac62b5d792af406e55c4e284880aa711f5e3d (image=quay.ceph.io/ceph-ci/ceph@sha256:8fda260ab1d2d3118a1622f7df75f44f285dfe74e71793626152a711c12bf2cc, name=ceph-9cb345a8-1bf9-11f1-a195-0375563c5891-osd-1-deactivate, CEPH_REF=squid, ceph=True, CEPH_GIT_REPO=https://github.com/ceph/ceph-ci.git, org.label-schema.name=CentOS Stream 9 Base Image, org.opencontainers.image.authors=Ceph Release Team , org.label-schema.license=GPLv2, GANESHA_REPO_BASEURL=https://buildlogs.centos.org/centos/$releasever-stream/storage/$basearch/nfsganesha-5/, org.label-schema.schema-version=1.0, FROM_IMAGE=quay.io/centos/centos:stream9, org.label-schema.vendor=CentOS, org.label-schema.build-date=20260223, io.buildah.version=1.41.3, org.opencontainers.image.documentation=https://docs.ceph.com/, OSD_FLAVOR=default, CEPH_SHA1=e911bdebe5c8faa3800735d1568fcdca65db60df) 2026-03-09T20:56:43.438 INFO:journalctl@ceph.osd.1.vm06.stdout:Mar 09 20:56:43 vm06 podman[61569]: 2026-03-09 20:56:43.18278949 +0000 UTC m=+0.096923294 container attach c6ba29df1899403a7d765541951ac62b5d792af406e55c4e284880aa711f5e3d (image=quay.ceph.io/ceph-ci/ceph@sha256:8fda260ab1d2d3118a1622f7df75f44f285dfe74e71793626152a711c12bf2cc, name=ceph-9cb345a8-1bf9-11f1-a195-0375563c5891-osd-1-deactivate, FROM_IMAGE=quay.io/centos/centos:stream9, org.opencontainers.image.authors=Ceph Release Team , org.label-schema.license=GPLv2, CEPH_GIT_REPO=https://github.com/ceph/ceph-ci.git, CEPH_SHA1=e911bdebe5c8faa3800735d1568fcdca65db60df, org.label-schema.vendor=CentOS, org.label-schema.name=CentOS Stream 9 Base Image, org.opencontainers.image.documentation=https://docs.ceph.com/, ceph=True, GANESHA_REPO_BASEURL=https://buildlogs.centos.org/centos/$releasever-stream/storage/$basearch/nfsganesha-5/, OSD_FLAVOR=default, io.buildah.version=1.41.3, CEPH_REF=squid, org.label-schema.schema-version=1.0, org.label-schema.build-date=20260223) 2026-03-09T20:56:43.438 INFO:journalctl@ceph.osd.1.vm06.stdout:Mar 09 20:56:43 vm06 podman[61569]: 2026-03-09 20:56:43.095912605 +0000 UTC m=+0.010046409 image pull 654f31e6858eb235bbece362255b685a945f2b6a367e2b88c4930c984fbb214c quay.ceph.io/ceph-ci/ceph@sha256:8fda260ab1d2d3118a1622f7df75f44f285dfe74e71793626152a711c12bf2cc 2026-03-09T20:56:43.438 INFO:journalctl@ceph.osd.1.vm06.stdout:Mar 09 20:56:43 vm06 podman[61569]: 2026-03-09 20:56:43.319517908 +0000 UTC m=+0.233651712 container died c6ba29df1899403a7d765541951ac62b5d792af406e55c4e284880aa711f5e3d (image=quay.ceph.io/ceph-ci/ceph@sha256:8fda260ab1d2d3118a1622f7df75f44f285dfe74e71793626152a711c12bf2cc, name=ceph-9cb345a8-1bf9-11f1-a195-0375563c5891-osd-1-deactivate, io.buildah.version=1.41.3, CEPH_SHA1=e911bdebe5c8faa3800735d1568fcdca65db60df, org.opencontainers.image.authors=Ceph Release Team , org.label-schema.schema-version=1.0, org.label-schema.vendor=CentOS, CEPH_GIT_REPO=https://github.com/ceph/ceph-ci.git, org.label-schema.name=CentOS Stream 9 Base Image, org.opencontainers.image.documentation=https://docs.ceph.com/, CEPH_REF=squid, FROM_IMAGE=quay.io/centos/centos:stream9, org.label-schema.build-date=20260223, ceph=True, org.label-schema.license=GPLv2, GANESHA_REPO_BASEURL=https://buildlogs.centos.org/centos/$releasever-stream/storage/$basearch/nfsganesha-5/, OSD_FLAVOR=default) 2026-03-09T20:56:43.450 DEBUG:teuthology.orchestra.run.vm06:> sudo pkill -f 'journalctl -f -n 0 -u ceph-9cb345a8-1bf9-11f1-a195-0375563c5891@osd.1.service' 2026-03-09T20:56:43.486 DEBUG:teuthology.orchestra.run:got remote process result: None 2026-03-09T20:56:43.486 INFO:tasks.cephadm.osd.1:Stopped osd.1 2026-03-09T20:56:43.486 INFO:tasks.cephadm.osd.2:Stopping osd.2... 2026-03-09T20:56:43.486 DEBUG:teuthology.orchestra.run.vm08:> sudo systemctl stop ceph-9cb345a8-1bf9-11f1-a195-0375563c5891@osd.2 2026-03-09T20:56:43.957 INFO:journalctl@ceph.osd.2.vm08.stdout:Mar 09 20:56:43 vm08 systemd[1]: Stopping Ceph osd.2 for 9cb345a8-1bf9-11f1-a195-0375563c5891... 2026-03-09T20:56:43.957 INFO:journalctl@ceph.osd.2.vm08.stdout:Mar 09 20:56:43 vm08 ceph-9cb345a8-1bf9-11f1-a195-0375563c5891-osd-2[58655]: 2026-03-09T20:56:43.600+0000 7fc583a8d640 -1 received signal: Terminated from /run/podman-init -- /usr/bin/ceph-osd -n osd.2 -f --setuser ceph --setgroup ceph --default-log-to-file=false --default-log-to-journald=true --default-log-to-stderr=false (PID: 1) UID: 0 2026-03-09T20:56:43.957 INFO:journalctl@ceph.osd.2.vm08.stdout:Mar 09 20:56:43 vm08 ceph-9cb345a8-1bf9-11f1-a195-0375563c5891-osd-2[58655]: 2026-03-09T20:56:43.600+0000 7fc583a8d640 -1 osd.2 21 *** Got signal Terminated *** 2026-03-09T20:56:43.957 INFO:journalctl@ceph.osd.2.vm08.stdout:Mar 09 20:56:43 vm08 ceph-9cb345a8-1bf9-11f1-a195-0375563c5891-osd-2[58655]: 2026-03-09T20:56:43.600+0000 7fc583a8d640 -1 osd.2 21 *** Immediate shutdown (osd_fast_shutdown=true) *** 2026-03-09T20:56:48.899 INFO:journalctl@ceph.osd.2.vm08.stdout:Mar 09 20:56:48 vm08 podman[62627]: 2026-03-09 20:56:48.621842627 +0000 UTC m=+5.038488095 container died b061149d182532e63637686488950a2cf939b0d7f172359d2fcd93389b2718eb (image=quay.ceph.io/ceph-ci/ceph@sha256:8fda260ab1d2d3118a1622f7df75f44f285dfe74e71793626152a711c12bf2cc, name=ceph-9cb345a8-1bf9-11f1-a195-0375563c5891-osd-2, FROM_IMAGE=quay.io/centos/centos:stream9, ceph=True, org.label-schema.license=GPLv2, CEPH_SHA1=e911bdebe5c8faa3800735d1568fcdca65db60df, io.buildah.version=1.41.3, org.label-schema.schema-version=1.0, org.label-schema.name=CentOS Stream 9 Base Image, org.opencontainers.image.authors=Ceph Release Team , GANESHA_REPO_BASEURL=https://buildlogs.centos.org/centos/$releasever-stream/storage/$basearch/nfsganesha-5/, org.opencontainers.image.documentation=https://docs.ceph.com/, CEPH_REF=squid, org.label-schema.vendor=CentOS, CEPH_GIT_REPO=https://github.com/ceph/ceph-ci.git, OSD_FLAVOR=default, org.label-schema.build-date=20260223) 2026-03-09T20:56:48.899 INFO:journalctl@ceph.osd.2.vm08.stdout:Mar 09 20:56:48 vm08 podman[62627]: 2026-03-09 20:56:48.74658488 +0000 UTC m=+5.163230339 container remove b061149d182532e63637686488950a2cf939b0d7f172359d2fcd93389b2718eb (image=quay.ceph.io/ceph-ci/ceph@sha256:8fda260ab1d2d3118a1622f7df75f44f285dfe74e71793626152a711c12bf2cc, name=ceph-9cb345a8-1bf9-11f1-a195-0375563c5891-osd-2, org.label-schema.license=GPLv2, io.buildah.version=1.41.3, CEPH_REF=squid, org.opencontainers.image.authors=Ceph Release Team , CEPH_SHA1=e911bdebe5c8faa3800735d1568fcdca65db60df, org.label-schema.schema-version=1.0, org.label-schema.vendor=CentOS, FROM_IMAGE=quay.io/centos/centos:stream9, org.label-schema.build-date=20260223, CEPH_GIT_REPO=https://github.com/ceph/ceph-ci.git, GANESHA_REPO_BASEURL=https://buildlogs.centos.org/centos/$releasever-stream/storage/$basearch/nfsganesha-5/, OSD_FLAVOR=default, ceph=True, org.label-schema.name=CentOS Stream 9 Base Image, org.opencontainers.image.documentation=https://docs.ceph.com/) 2026-03-09T20:56:48.899 INFO:journalctl@ceph.osd.2.vm08.stdout:Mar 09 20:56:48 vm08 bash[62627]: ceph-9cb345a8-1bf9-11f1-a195-0375563c5891-osd-2 2026-03-09T20:56:49.207 INFO:journalctl@ceph.osd.2.vm08.stdout:Mar 09 20:56:48 vm08 podman[62705]: 2026-03-09 20:56:48.898959686 +0000 UTC m=+0.017796243 container create e476c3b493988b5fb25e1b36b37cdbb769678c8ecb2612e93e92601848bfb0a8 (image=quay.ceph.io/ceph-ci/ceph@sha256:8fda260ab1d2d3118a1622f7df75f44f285dfe74e71793626152a711c12bf2cc, name=ceph-9cb345a8-1bf9-11f1-a195-0375563c5891-osd-2-deactivate, org.opencontainers.image.authors=Ceph Release Team , CEPH_SHA1=e911bdebe5c8faa3800735d1568fcdca65db60df, OSD_FLAVOR=default, org.label-schema.license=GPLv2, org.label-schema.build-date=20260223, io.buildah.version=1.41.3, GANESHA_REPO_BASEURL=https://buildlogs.centos.org/centos/$releasever-stream/storage/$basearch/nfsganesha-5/, FROM_IMAGE=quay.io/centos/centos:stream9, ceph=True, org.label-schema.name=CentOS Stream 9 Base Image, CEPH_GIT_REPO=https://github.com/ceph/ceph-ci.git, org.opencontainers.image.documentation=https://docs.ceph.com/, org.label-schema.schema-version=1.0, org.label-schema.vendor=CentOS, CEPH_REF=squid) 2026-03-09T20:56:49.207 INFO:journalctl@ceph.osd.2.vm08.stdout:Mar 09 20:56:48 vm08 podman[62705]: 2026-03-09 20:56:48.937689521 +0000 UTC m=+0.056526078 container init e476c3b493988b5fb25e1b36b37cdbb769678c8ecb2612e93e92601848bfb0a8 (image=quay.ceph.io/ceph-ci/ceph@sha256:8fda260ab1d2d3118a1622f7df75f44f285dfe74e71793626152a711c12bf2cc, name=ceph-9cb345a8-1bf9-11f1-a195-0375563c5891-osd-2-deactivate, CEPH_SHA1=e911bdebe5c8faa3800735d1568fcdca65db60df, io.buildah.version=1.41.3, org.label-schema.name=CentOS Stream 9 Base Image, CEPH_REF=squid, OSD_FLAVOR=default, org.label-schema.license=GPLv2, GANESHA_REPO_BASEURL=https://buildlogs.centos.org/centos/$releasever-stream/storage/$basearch/nfsganesha-5/, org.label-schema.schema-version=1.0, FROM_IMAGE=quay.io/centos/centos:stream9, org.label-schema.vendor=CentOS, org.opencontainers.image.documentation=https://docs.ceph.com/, org.label-schema.build-date=20260223, org.opencontainers.image.authors=Ceph Release Team , ceph=True, CEPH_GIT_REPO=https://github.com/ceph/ceph-ci.git) 2026-03-09T20:56:49.207 INFO:journalctl@ceph.osd.2.vm08.stdout:Mar 09 20:56:48 vm08 podman[62705]: 2026-03-09 20:56:48.942594232 +0000 UTC m=+0.061430789 container start e476c3b493988b5fb25e1b36b37cdbb769678c8ecb2612e93e92601848bfb0a8 (image=quay.ceph.io/ceph-ci/ceph@sha256:8fda260ab1d2d3118a1622f7df75f44f285dfe74e71793626152a711c12bf2cc, name=ceph-9cb345a8-1bf9-11f1-a195-0375563c5891-osd-2-deactivate, org.opencontainers.image.documentation=https://docs.ceph.com/, org.label-schema.build-date=20260223, OSD_FLAVOR=default, org.opencontainers.image.authors=Ceph Release Team , org.label-schema.vendor=CentOS, org.label-schema.license=GPLv2, GANESHA_REPO_BASEURL=https://buildlogs.centos.org/centos/$releasever-stream/storage/$basearch/nfsganesha-5/, FROM_IMAGE=quay.io/centos/centos:stream9, ceph=True, CEPH_GIT_REPO=https://github.com/ceph/ceph-ci.git, io.buildah.version=1.41.3, org.label-schema.schema-version=1.0, org.label-schema.name=CentOS Stream 9 Base Image, CEPH_REF=squid, CEPH_SHA1=e911bdebe5c8faa3800735d1568fcdca65db60df) 2026-03-09T20:56:49.207 INFO:journalctl@ceph.osd.2.vm08.stdout:Mar 09 20:56:48 vm08 podman[62705]: 2026-03-09 20:56:48.94431361 +0000 UTC m=+0.063150167 container attach e476c3b493988b5fb25e1b36b37cdbb769678c8ecb2612e93e92601848bfb0a8 (image=quay.ceph.io/ceph-ci/ceph@sha256:8fda260ab1d2d3118a1622f7df75f44f285dfe74e71793626152a711c12bf2cc, name=ceph-9cb345a8-1bf9-11f1-a195-0375563c5891-osd-2-deactivate, OSD_FLAVOR=default, org.label-schema.license=GPLv2, ceph=True, CEPH_GIT_REPO=https://github.com/ceph/ceph-ci.git, FROM_IMAGE=quay.io/centos/centos:stream9, org.label-schema.name=CentOS Stream 9 Base Image, CEPH_SHA1=e911bdebe5c8faa3800735d1568fcdca65db60df, org.opencontainers.image.documentation=https://docs.ceph.com/, io.buildah.version=1.41.3, org.label-schema.schema-version=1.0, org.opencontainers.image.authors=Ceph Release Team , GANESHA_REPO_BASEURL=https://buildlogs.centos.org/centos/$releasever-stream/storage/$basearch/nfsganesha-5/, org.label-schema.vendor=CentOS, CEPH_REF=squid, org.label-schema.build-date=20260223) 2026-03-09T20:56:49.207 INFO:journalctl@ceph.osd.2.vm08.stdout:Mar 09 20:56:48 vm08 podman[62705]: 2026-03-09 20:56:48.89258297 +0000 UTC m=+0.011419537 image pull 654f31e6858eb235bbece362255b685a945f2b6a367e2b88c4930c984fbb214c quay.ceph.io/ceph-ci/ceph@sha256:8fda260ab1d2d3118a1622f7df75f44f285dfe74e71793626152a711c12bf2cc 2026-03-09T20:56:49.207 INFO:journalctl@ceph.osd.2.vm08.stdout:Mar 09 20:56:49 vm08 podman[62705]: 2026-03-09 20:56:49.072372772 +0000 UTC m=+0.191209329 container died e476c3b493988b5fb25e1b36b37cdbb769678c8ecb2612e93e92601848bfb0a8 (image=quay.ceph.io/ceph-ci/ceph@sha256:8fda260ab1d2d3118a1622f7df75f44f285dfe74e71793626152a711c12bf2cc, name=ceph-9cb345a8-1bf9-11f1-a195-0375563c5891-osd-2-deactivate, OSD_FLAVOR=default, io.buildah.version=1.41.3, GANESHA_REPO_BASEURL=https://buildlogs.centos.org/centos/$releasever-stream/storage/$basearch/nfsganesha-5/, FROM_IMAGE=quay.io/centos/centos:stream9, org.label-schema.name=CentOS Stream 9 Base Image, CEPH_REF=squid, org.label-schema.schema-version=1.0, org.label-schema.license=GPLv2, org.opencontainers.image.authors=Ceph Release Team , org.opencontainers.image.documentation=https://docs.ceph.com/, org.label-schema.build-date=20260223, ceph=True, org.label-schema.vendor=CentOS, CEPH_SHA1=e911bdebe5c8faa3800735d1568fcdca65db60df, CEPH_GIT_REPO=https://github.com/ceph/ceph-ci.git) 2026-03-09T20:56:49.255 DEBUG:teuthology.orchestra.run.vm08:> sudo pkill -f 'journalctl -f -n 0 -u ceph-9cb345a8-1bf9-11f1-a195-0375563c5891@osd.2.service' 2026-03-09T20:56:49.285 DEBUG:teuthology.orchestra.run:got remote process result: None 2026-03-09T20:56:49.285 INFO:tasks.cephadm.osd.2:Stopped osd.2 2026-03-09T20:56:49.285 DEBUG:teuthology.orchestra.run.vm01:> sudo /home/ubuntu/cephtest/cephadm rm-cluster --fsid 9cb345a8-1bf9-11f1-a195-0375563c5891 --force --keep-logs 2026-03-09T20:56:49.422 INFO:teuthology.orchestra.run.vm01.stdout:Deleting cluster with fsid: 9cb345a8-1bf9-11f1-a195-0375563c5891 2026-03-09T20:56:50.404 DEBUG:teuthology.orchestra.run.vm06:> sudo /home/ubuntu/cephtest/cephadm rm-cluster --fsid 9cb345a8-1bf9-11f1-a195-0375563c5891 --force --keep-logs 2026-03-09T20:56:50.535 INFO:teuthology.orchestra.run.vm06.stdout:Deleting cluster with fsid: 9cb345a8-1bf9-11f1-a195-0375563c5891 2026-03-09T20:56:51.531 DEBUG:teuthology.orchestra.run.vm08:> sudo /home/ubuntu/cephtest/cephadm rm-cluster --fsid 9cb345a8-1bf9-11f1-a195-0375563c5891 --force --keep-logs 2026-03-09T20:56:51.673 INFO:teuthology.orchestra.run.vm08.stdout:Deleting cluster with fsid: 9cb345a8-1bf9-11f1-a195-0375563c5891 2026-03-09T20:56:52.472 DEBUG:teuthology.orchestra.run.vm01:> sudo rm -f /etc/ceph/ceph.conf /etc/ceph/ceph.client.admin.keyring 2026-03-09T20:56:52.499 DEBUG:teuthology.orchestra.run.vm06:> sudo rm -f /etc/ceph/ceph.conf /etc/ceph/ceph.client.admin.keyring 2026-03-09T20:56:52.526 DEBUG:teuthology.orchestra.run.vm08:> sudo rm -f /etc/ceph/ceph.conf /etc/ceph/ceph.client.admin.keyring 2026-03-09T20:56:52.554 INFO:tasks.cephadm:Archiving crash dumps... 2026-03-09T20:56:52.554 DEBUG:teuthology.misc:Transferring archived files from vm01:/var/lib/ceph/9cb345a8-1bf9-11f1-a195-0375563c5891/crash to /archive/kyr-2026-03-09_11:23:05-orch-squid-none-default-vps/653/remote/vm01/crash 2026-03-09T20:56:52.554 DEBUG:teuthology.orchestra.run.vm01:> sudo tar c -f - -C /var/lib/ceph/9cb345a8-1bf9-11f1-a195-0375563c5891/crash -- . 2026-03-09T20:56:52.579 INFO:teuthology.orchestra.run.vm01.stderr:tar: /var/lib/ceph/9cb345a8-1bf9-11f1-a195-0375563c5891/crash: Cannot open: No such file or directory 2026-03-09T20:56:52.579 INFO:teuthology.orchestra.run.vm01.stderr:tar: Error is not recoverable: exiting now 2026-03-09T20:56:52.580 DEBUG:teuthology.misc:Transferring archived files from vm06:/var/lib/ceph/9cb345a8-1bf9-11f1-a195-0375563c5891/crash to /archive/kyr-2026-03-09_11:23:05-orch-squid-none-default-vps/653/remote/vm06/crash 2026-03-09T20:56:52.580 DEBUG:teuthology.orchestra.run.vm06:> sudo tar c -f - -C /var/lib/ceph/9cb345a8-1bf9-11f1-a195-0375563c5891/crash -- . 2026-03-09T20:56:52.608 INFO:teuthology.orchestra.run.vm06.stderr:tar: /var/lib/ceph/9cb345a8-1bf9-11f1-a195-0375563c5891/crash: Cannot open: No such file or directory 2026-03-09T20:56:52.608 INFO:teuthology.orchestra.run.vm06.stderr:tar: Error is not recoverable: exiting now 2026-03-09T20:56:52.609 DEBUG:teuthology.misc:Transferring archived files from vm08:/var/lib/ceph/9cb345a8-1bf9-11f1-a195-0375563c5891/crash to /archive/kyr-2026-03-09_11:23:05-orch-squid-none-default-vps/653/remote/vm08/crash 2026-03-09T20:56:52.609 DEBUG:teuthology.orchestra.run.vm08:> sudo tar c -f - -C /var/lib/ceph/9cb345a8-1bf9-11f1-a195-0375563c5891/crash -- . 2026-03-09T20:56:52.634 INFO:teuthology.orchestra.run.vm08.stderr:tar: /var/lib/ceph/9cb345a8-1bf9-11f1-a195-0375563c5891/crash: Cannot open: No such file or directory 2026-03-09T20:56:52.634 INFO:teuthology.orchestra.run.vm08.stderr:tar: Error is not recoverable: exiting now 2026-03-09T20:56:52.635 INFO:tasks.cephadm:Checking cluster log for badness... 2026-03-09T20:56:52.636 DEBUG:teuthology.orchestra.run.vm01:> sudo egrep '\[ERR\]|\[WRN\]|\[SEC\]' /var/log/ceph/9cb345a8-1bf9-11f1-a195-0375563c5891/ceph.log | egrep CEPHADM_ | egrep -v '\(MDS_ALL_DOWN\)' | egrep -v '\(MDS_UP_LESS_THAN_MAX\)' | egrep -v MON_DOWN | egrep -v POOL_APP_NOT_ENABLED | egrep -v 'mon down' | egrep -v 'mons down' | egrep -v 'out of quorum' | egrep -v CEPHADM_FAILED_DAEMON | head -n 1 2026-03-09T20:56:52.667 INFO:tasks.cephadm:Compressing logs... 2026-03-09T20:56:52.667 DEBUG:teuthology.orchestra.run.vm01:> time sudo find /var/log/ceph /var/log/rbd-target-api -name '*.log' -print0 | sudo xargs --max-args=1 --max-procs=0 --verbose -0 --no-run-if-empty -- gzip -5 --verbose -- 2026-03-09T20:56:52.709 DEBUG:teuthology.orchestra.run.vm06:> time sudo find /var/log/ceph /var/log/rbd-target-api -name '*.log' -print0 | sudo xargs --max-args=1 --max-procs=0 --verbose -0 --no-run-if-empty -- gzip -5 --verbose -- 2026-03-09T20:56:52.710 DEBUG:teuthology.orchestra.run.vm08:> time sudo find /var/log/ceph /var/log/rbd-target-api -name '*.log' -print0 | sudo xargs --max-args=1 --max-procs=0 --verbose -0 --no-run-if-empty -- gzip -5 --verbose -- 2026-03-09T20:56:52.732 INFO:teuthology.orchestra.run.vm01.stderr:find: gzip -5 --verbose -- /var/log/ceph/cephadm.log 2026-03-09T20:56:52.732 INFO:teuthology.orchestra.run.vm01.stderr:‘/var/log/rbd-target-api’: No such file or directory 2026-03-09T20:56:52.733 INFO:teuthology.orchestra.run.vm01.stderr:gzip -5 --verbose -- /var/log/ceph/9cb345a8-1bf9-11f1-a195-0375563c5891/ceph-mon.a.log 2026-03-09T20:56:52.733 INFO:teuthology.orchestra.run.vm01.stderr:gzip -5 --verbose -- /var/log/ceph/9cb345a8-1bf9-11f1-a195-0375563c5891/ceph.log 2026-03-09T20:56:52.734 INFO:teuthology.orchestra.run.vm01.stderr:/var/log/ceph/cephadm.log: /var/log/ceph/9cb345a8-1bf9-11f1-a195-0375563c5891/ceph-mon.a.log: 89.7% -- replaced with /var/log/ceph/cephadm.log.gz 2026-03-09T20:56:52.735 INFO:teuthology.orchestra.run.vm01.stderr:gzip -5 --verbose -- /var/log/ceph/9cb345a8-1bf9-11f1-a195-0375563c5891/ceph.audit.log 2026-03-09T20:56:52.735 INFO:teuthology.orchestra.run.vm01.stderr:/var/log/ceph/9cb345a8-1bf9-11f1-a195-0375563c5891/ceph.log: 87.0% -- replaced with /var/log/ceph/9cb345a8-1bf9-11f1-a195-0375563c5891/ceph.log.gz 2026-03-09T20:56:52.735 INFO:teuthology.orchestra.run.vm01.stderr:gzip -5 --verbose -- /var/log/ceph/9cb345a8-1bf9-11f1-a195-0375563c5891/ceph-mgr.a.log 2026-03-09T20:56:52.737 INFO:teuthology.orchestra.run.vm01.stderr:/var/log/ceph/9cb345a8-1bf9-11f1-a195-0375563c5891/ceph.audit.log: 89.7% -- replaced with /var/log/ceph/9cb345a8-1bf9-11f1-a195-0375563c5891/ceph.audit.log.gz 2026-03-09T20:56:52.737 INFO:teuthology.orchestra.run.vm06.stderr:find: gzip -5 --verbose -- /var/log/ceph/cephadm.log 2026-03-09T20:56:52.737 INFO:teuthology.orchestra.run.vm06.stderr:‘/var/log/rbd-target-api’: No such file or directory 2026-03-09T20:56:52.737 INFO:teuthology.orchestra.run.vm01.stderr:gzip -5 --verbose -- /var/log/ceph/9cb345a8-1bf9-11f1-a195-0375563c5891/ceph.cephadm.log 2026-03-09T20:56:52.738 INFO:teuthology.orchestra.run.vm06.stderr:/var/log/ceph/cephadm.log: 88.5% -- replaced with /var/log/ceph/cephadm.log.gz 2026-03-09T20:56:52.738 INFO:teuthology.orchestra.run.vm06.stderr:gzip -5 --verbose -- /var/log/ceph/9cb345a8-1bf9-11f1-a195-0375563c5891/ceph-volume.log 2026-03-09T20:56:52.738 INFO:teuthology.orchestra.run.vm06.stderr:gzip -5 --verbose -- /var/log/ceph/9cb345a8-1bf9-11f1-a195-0375563c5891/ceph-mon.b.log 2026-03-09T20:56:52.740 INFO:teuthology.orchestra.run.vm08.stderr:find: ‘/var/log/rbd-target-api’: No such file or directory 2026-03-09T20:56:52.740 INFO:teuthology.orchestra.run.vm06.stderr:/var/log/ceph/9cb345a8-1bf9-11f1-a195-0375563c5891/ceph-volume.log: 94.6% -- replaced with /var/log/ceph/9cb345a8-1bf9-11f1-a195-0375563c5891/ceph-volume.log.gz 2026-03-09T20:56:52.740 INFO:teuthology.orchestra.run.vm06.stderr:gzip -5 --verbose -- /var/log/ceph/9cb345a8-1bf9-11f1-a195-0375563c5891/ceph.cephadm.log 2026-03-09T20:56:52.740 INFO:teuthology.orchestra.run.vm08.stderr:gzip -5 --verbose -- /var/log/ceph/cephadm.log 2026-03-09T20:56:52.741 INFO:teuthology.orchestra.run.vm06.stderr:gzip -5 --verbose -- /var/log/ceph/9cb345a8-1bf9-11f1-a195-0375563c5891/ceph.log 2026-03-09T20:56:52.741 INFO:teuthology.orchestra.run.vm08.stderr:gzip -5 --verbose -- /var/log/ceph/9cb345a8-1bf9-11f1-a195-0375563c5891/ceph-volume.log 2026-03-09T20:56:52.741 INFO:teuthology.orchestra.run.vm06.stderr:/var/log/ceph/9cb345a8-1bf9-11f1-a195-0375563c5891/ceph.cephadm.log: /var/log/ceph/9cb345a8-1bf9-11f1-a195-0375563c5891/ceph-mon.b.log: 80.7% -- replaced with /var/log/ceph/9cb345a8-1bf9-11f1-a195-0375563c5891/ceph.cephadm.log.gz 2026-03-09T20:56:52.741 INFO:teuthology.orchestra.run.vm08.stderr:gzip -5 --verbose -- /var/log/ceph/9cb345a8-1bf9-11f1-a195-0375563c5891/ceph-mon.c.log 2026-03-09T20:56:52.741 INFO:teuthology.orchestra.run.vm06.stderr:gzip -5 --verbose -- /var/log/ceph/9cb345a8-1bf9-11f1-a195-0375563c5891/ceph.audit.log 2026-03-09T20:56:52.741 INFO:teuthology.orchestra.run.vm01.stderr:/var/log/ceph/9cb345a8-1bf9-11f1-a195-0375563c5891/ceph-mgr.a.log: gzip -5 --verbose -- /var/log/ceph/9cb345a8-1bf9-11f1-a195-0375563c5891/ceph-volume.log 2026-03-09T20:56:52.742 INFO:teuthology.orchestra.run.vm08.stderr:/var/log/ceph/cephadm.log: /var/log/ceph/9cb345a8-1bf9-11f1-a195-0375563c5891/ceph-volume.log: 88.8% -- replaced with /var/log/ceph/cephadm.log.gz 2026-03-09T20:56:52.742 INFO:teuthology.orchestra.run.vm06.stderr:/var/log/ceph/9cb345a8-1bf9-11f1-a195-0375563c5891/ceph.log: 87.0% -- replaced with /var/log/ceph/9cb345a8-1bf9-11f1-a195-0375563c5891/ceph.log.gz 2026-03-09T20:56:52.742 INFO:teuthology.orchestra.run.vm06.stderr:gzip -5 --verbose -- /var/log/ceph/9cb345a8-1bf9-11f1-a195-0375563c5891/ceph-mgr.b.log 2026-03-09T20:56:52.743 INFO:teuthology.orchestra.run.vm01.stderr:/var/log/ceph/9cb345a8-1bf9-11f1-a195-0375563c5891/ceph.cephadm.log: 81.3% -- replaced with /var/log/ceph/9cb345a8-1bf9-11f1-a195-0375563c5891/ceph.cephadm.log.gz 2026-03-09T20:56:52.744 INFO:teuthology.orchestra.run.vm08.stderr:gzip -5 --verbose -- /var/log/ceph/9cb345a8-1bf9-11f1-a195-0375563c5891/ceph.cephadm.log 2026-03-09T20:56:52.744 INFO:teuthology.orchestra.run.vm08.stderr: 94.8% -- replaced with /var/log/ceph/9cb345a8-1bf9-11f1-a195-0375563c5891/ceph-volume.log.gz 2026-03-09T20:56:52.744 INFO:teuthology.orchestra.run.vm08.stderr:/var/log/ceph/9cb345a8-1bf9-11f1-a195-0375563c5891/ceph-mon.c.log: gzip -5 --verbose -- /var/log/ceph/9cb345a8-1bf9-11f1-a195-0375563c5891/ceph.log 2026-03-09T20:56:52.744 INFO:teuthology.orchestra.run.vm06.stderr:/var/log/ceph/9cb345a8-1bf9-11f1-a195-0375563c5891/ceph.audit.log: 90.0% -- replaced with /var/log/ceph/9cb345a8-1bf9-11f1-a195-0375563c5891/ceph.audit.log.gz 2026-03-09T20:56:52.744 INFO:teuthology.orchestra.run.vm06.stderr:gzip -5 --verbose -- /var/log/ceph/9cb345a8-1bf9-11f1-a195-0375563c5891/ceph-osd.1.log 2026-03-09T20:56:52.744 INFO:teuthology.orchestra.run.vm08.stderr:/var/log/ceph/9cb345a8-1bf9-11f1-a195-0375563c5891/ceph.cephadm.log: 80.7% -- replaced with /var/log/ceph/9cb345a8-1bf9-11f1-a195-0375563c5891/ceph.cephadm.log.gz 2026-03-09T20:56:52.744 INFO:teuthology.orchestra.run.vm08.stderr:gzip -5 --verbose -- /var/log/ceph/9cb345a8-1bf9-11f1-a195-0375563c5891/ceph.audit.log 2026-03-09T20:56:52.745 INFO:teuthology.orchestra.run.vm08.stderr:/var/log/ceph/9cb345a8-1bf9-11f1-a195-0375563c5891/ceph.log: 87.0% -- replaced with /var/log/ceph/9cb345a8-1bf9-11f1-a195-0375563c5891/ceph.log.gz 2026-03-09T20:56:52.745 INFO:teuthology.orchestra.run.vm08.stderr:gzip -5 --verbose -- /var/log/ceph/9cb345a8-1bf9-11f1-a195-0375563c5891/ceph-osd.2.log 2026-03-09T20:56:52.746 INFO:teuthology.orchestra.run.vm06.stderr:/var/log/ceph/9cb345a8-1bf9-11f1-a195-0375563c5891/ceph-mgr.b.log: 90.9% -- replaced with /var/log/ceph/9cb345a8-1bf9-11f1-a195-0375563c5891/ceph-mgr.b.log.gz 2026-03-09T20:56:52.746 INFO:teuthology.orchestra.run.vm08.stderr:/var/log/ceph/9cb345a8-1bf9-11f1-a195-0375563c5891/ceph.audit.log: 90.0% -- replaced with /var/log/ceph/9cb345a8-1bf9-11f1-a195-0375563c5891/ceph.audit.log.gz 2026-03-09T20:56:52.748 INFO:teuthology.orchestra.run.vm01.stderr:gzip -5 --verbose -- /var/log/ceph/9cb345a8-1bf9-11f1-a195-0375563c5891/ceph-osd.0.log 2026-03-09T20:56:52.752 INFO:teuthology.orchestra.run.vm01.stderr:/var/log/ceph/9cb345a8-1bf9-11f1-a195-0375563c5891/ceph-volume.log: 94.7% -- replaced with /var/log/ceph/9cb345a8-1bf9-11f1-a195-0375563c5891/ceph-volume.log.gz 2026-03-09T20:56:52.762 INFO:teuthology.orchestra.run.vm08.stderr:/var/log/ceph/9cb345a8-1bf9-11f1-a195-0375563c5891/ceph-osd.2.log: 93.4% -- replaced with /var/log/ceph/9cb345a8-1bf9-11f1-a195-0375563c5891/ceph-osd.2.log.gz 2026-03-09T20:56:52.769 INFO:teuthology.orchestra.run.vm06.stderr:/var/log/ceph/9cb345a8-1bf9-11f1-a195-0375563c5891/ceph-osd.1.log: 93.4% -- replaced with /var/log/ceph/9cb345a8-1bf9-11f1-a195-0375563c5891/ceph-osd.1.log.gz 2026-03-09T20:56:52.777 INFO:teuthology.orchestra.run.vm06.stderr: 93.1% -- replaced with /var/log/ceph/9cb345a8-1bf9-11f1-a195-0375563c5891/ceph-mon.b.log.gz 2026-03-09T20:56:52.779 INFO:teuthology.orchestra.run.vm08.stderr: 92.8% -- replaced with /var/log/ceph/9cb345a8-1bf9-11f1-a195-0375563c5891/ceph-mon.c.log.gz 2026-03-09T20:56:52.780 INFO:teuthology.orchestra.run.vm08.stderr: 2026-03-09T20:56:52.780 INFO:teuthology.orchestra.run.vm08.stderr:real 0m0.051s 2026-03-09T20:56:52.780 INFO:teuthology.orchestra.run.vm08.stderr:user 0m0.055s 2026-03-09T20:56:52.780 INFO:teuthology.orchestra.run.vm08.stderr:sys 0m0.023s 2026-03-09T20:56:52.780 INFO:teuthology.orchestra.run.vm06.stderr: 2026-03-09T20:56:52.780 INFO:teuthology.orchestra.run.vm06.stderr:real 0m0.055s 2026-03-09T20:56:52.780 INFO:teuthology.orchestra.run.vm06.stderr:user 0m0.067s 2026-03-09T20:56:52.780 INFO:teuthology.orchestra.run.vm06.stderr:sys 0m0.022s 2026-03-09T20:56:52.788 INFO:teuthology.orchestra.run.vm01.stderr:/var/log/ceph/9cb345a8-1bf9-11f1-a195-0375563c5891/ceph-osd.0.log: 93.2% -- replaced with /var/log/ceph/9cb345a8-1bf9-11f1-a195-0375563c5891/ceph-osd.0.log.gz 2026-03-09T20:56:52.792 INFO:teuthology.orchestra.run.vm01.stderr: 90.7% -- replaced with /var/log/ceph/9cb345a8-1bf9-11f1-a195-0375563c5891/ceph-mgr.a.log.gz 2026-03-09T20:56:52.845 INFO:teuthology.orchestra.run.vm01.stderr: 91.3% -- replaced with /var/log/ceph/9cb345a8-1bf9-11f1-a195-0375563c5891/ceph-mon.a.log.gz 2026-03-09T20:56:52.848 INFO:teuthology.orchestra.run.vm01.stderr: 2026-03-09T20:56:52.848 INFO:teuthology.orchestra.run.vm01.stderr:real 0m0.125s 2026-03-09T20:56:52.848 INFO:teuthology.orchestra.run.vm01.stderr:user 0m0.170s 2026-03-09T20:56:52.848 INFO:teuthology.orchestra.run.vm01.stderr:sys 0m0.019s 2026-03-09T20:56:52.848 INFO:tasks.cephadm:Archiving logs... 2026-03-09T20:56:52.848 DEBUG:teuthology.misc:Transferring archived files from vm01:/var/log/ceph to /archive/kyr-2026-03-09_11:23:05-orch-squid-none-default-vps/653/remote/vm01/log 2026-03-09T20:56:52.848 DEBUG:teuthology.orchestra.run.vm01:> sudo tar c -f - -C /var/log/ceph -- . 2026-03-09T20:56:52.927 DEBUG:teuthology.misc:Transferring archived files from vm06:/var/log/ceph to /archive/kyr-2026-03-09_11:23:05-orch-squid-none-default-vps/653/remote/vm06/log 2026-03-09T20:56:52.928 DEBUG:teuthology.orchestra.run.vm06:> sudo tar c -f - -C /var/log/ceph -- . 2026-03-09T20:56:52.958 DEBUG:teuthology.misc:Transferring archived files from vm08:/var/log/ceph to /archive/kyr-2026-03-09_11:23:05-orch-squid-none-default-vps/653/remote/vm08/log 2026-03-09T20:56:52.958 DEBUG:teuthology.orchestra.run.vm08:> sudo tar c -f - -C /var/log/ceph -- . 2026-03-09T20:56:52.990 INFO:tasks.cephadm:Removing cluster... 2026-03-09T20:56:52.990 DEBUG:teuthology.orchestra.run.vm01:> sudo /home/ubuntu/cephtest/cephadm rm-cluster --fsid 9cb345a8-1bf9-11f1-a195-0375563c5891 --force 2026-03-09T20:56:53.134 INFO:teuthology.orchestra.run.vm01.stdout:Deleting cluster with fsid: 9cb345a8-1bf9-11f1-a195-0375563c5891 2026-03-09T20:56:53.411 DEBUG:teuthology.orchestra.run.vm06:> sudo /home/ubuntu/cephtest/cephadm rm-cluster --fsid 9cb345a8-1bf9-11f1-a195-0375563c5891 --force 2026-03-09T20:56:53.554 INFO:teuthology.orchestra.run.vm06.stdout:Deleting cluster with fsid: 9cb345a8-1bf9-11f1-a195-0375563c5891 2026-03-09T20:56:53.809 DEBUG:teuthology.orchestra.run.vm08:> sudo /home/ubuntu/cephtest/cephadm rm-cluster --fsid 9cb345a8-1bf9-11f1-a195-0375563c5891 --force 2026-03-09T20:56:53.947 INFO:teuthology.orchestra.run.vm08.stdout:Deleting cluster with fsid: 9cb345a8-1bf9-11f1-a195-0375563c5891 2026-03-09T20:56:54.153 INFO:tasks.cephadm:Removing cephadm ... 2026-03-09T20:56:54.153 DEBUG:teuthology.orchestra.run.vm01:> rm -rf /home/ubuntu/cephtest/cephadm 2026-03-09T20:56:54.169 DEBUG:teuthology.orchestra.run.vm06:> rm -rf /home/ubuntu/cephtest/cephadm 2026-03-09T20:56:54.194 DEBUG:teuthology.orchestra.run.vm08:> rm -rf /home/ubuntu/cephtest/cephadm 2026-03-09T20:56:54.217 INFO:tasks.cephadm:Teardown complete 2026-03-09T20:56:54.217 DEBUG:teuthology.run_tasks:Unwinding manager install 2026-03-09T20:56:54.219 INFO:teuthology.task.install.util:Removing shipped files: /home/ubuntu/cephtest/valgrind.supp /usr/bin/daemon-helper /usr/bin/adjust-ulimits /usr/bin/stdin-killer... 2026-03-09T20:56:54.219 DEBUG:teuthology.orchestra.run.vm01:> sudo rm -f -- /home/ubuntu/cephtest/valgrind.supp /usr/bin/daemon-helper /usr/bin/adjust-ulimits /usr/bin/stdin-killer 2026-03-09T20:56:54.221 DEBUG:teuthology.orchestra.run.vm06:> sudo rm -f -- /home/ubuntu/cephtest/valgrind.supp /usr/bin/daemon-helper /usr/bin/adjust-ulimits /usr/bin/stdin-killer 2026-03-09T20:56:54.237 DEBUG:teuthology.orchestra.run.vm08:> sudo rm -f -- /home/ubuntu/cephtest/valgrind.supp /usr/bin/daemon-helper /usr/bin/adjust-ulimits /usr/bin/stdin-killer 2026-03-09T20:56:54.298 INFO:teuthology.task.install.rpm:Removing packages: ceph-radosgw, ceph-test, ceph, ceph-base, cephadm, ceph-immutable-object-cache, ceph-mgr, ceph-mgr-dashboard, ceph-mgr-diskprediction-local, ceph-mgr-rook, ceph-mgr-cephadm, ceph-fuse, ceph-volume, librados-devel, libcephfs2, libcephfs-devel, librados2, librbd1, python3-rados, python3-rgw, python3-cephfs, python3-rbd, rbd-fuse, rbd-mirror, rbd-nbd on rpm system. 2026-03-09T20:56:54.298 DEBUG:teuthology.orchestra.run.vm01:> 2026-03-09T20:56:54.298 DEBUG:teuthology.orchestra.run.vm01:> for d in ceph-radosgw ceph-test ceph ceph-base cephadm ceph-immutable-object-cache ceph-mgr ceph-mgr-dashboard ceph-mgr-diskprediction-local ceph-mgr-rook ceph-mgr-cephadm ceph-fuse ceph-volume librados-devel libcephfs2 libcephfs-devel librados2 librbd1 python3-rados python3-rgw python3-cephfs python3-rbd rbd-fuse rbd-mirror rbd-nbd ; do 2026-03-09T20:56:54.298 DEBUG:teuthology.orchestra.run.vm01:> sudo yum -y remove $d || true 2026-03-09T20:56:54.298 DEBUG:teuthology.orchestra.run.vm01:> done 2026-03-09T20:56:54.302 INFO:teuthology.task.install.rpm:Removing packages: ceph-radosgw, ceph-test, ceph, ceph-base, cephadm, ceph-immutable-object-cache, ceph-mgr, ceph-mgr-dashboard, ceph-mgr-diskprediction-local, ceph-mgr-rook, ceph-mgr-cephadm, ceph-fuse, ceph-volume, librados-devel, libcephfs2, libcephfs-devel, librados2, librbd1, python3-rados, python3-rgw, python3-cephfs, python3-rbd, rbd-fuse, rbd-mirror, rbd-nbd on rpm system. 2026-03-09T20:56:54.302 DEBUG:teuthology.orchestra.run.vm06:> 2026-03-09T20:56:54.302 DEBUG:teuthology.orchestra.run.vm06:> for d in ceph-radosgw ceph-test ceph ceph-base cephadm ceph-immutable-object-cache ceph-mgr ceph-mgr-dashboard ceph-mgr-diskprediction-local ceph-mgr-rook ceph-mgr-cephadm ceph-fuse ceph-volume librados-devel libcephfs2 libcephfs-devel librados2 librbd1 python3-rados python3-rgw python3-cephfs python3-rbd rbd-fuse rbd-mirror rbd-nbd ; do 2026-03-09T20:56:54.302 DEBUG:teuthology.orchestra.run.vm06:> sudo yum -y remove $d || true 2026-03-09T20:56:54.303 DEBUG:teuthology.orchestra.run.vm06:> done 2026-03-09T20:56:54.307 INFO:teuthology.task.install.rpm:Removing packages: ceph-radosgw, ceph-test, ceph, ceph-base, cephadm, ceph-immutable-object-cache, ceph-mgr, ceph-mgr-dashboard, ceph-mgr-diskprediction-local, ceph-mgr-rook, ceph-mgr-cephadm, ceph-fuse, ceph-volume, librados-devel, libcephfs2, libcephfs-devel, librados2, librbd1, python3-rados, python3-rgw, python3-cephfs, python3-rbd, rbd-fuse, rbd-mirror, rbd-nbd on rpm system. 2026-03-09T20:56:54.308 DEBUG:teuthology.orchestra.run.vm08:> 2026-03-09T20:56:54.308 DEBUG:teuthology.orchestra.run.vm08:> for d in ceph-radosgw ceph-test ceph ceph-base cephadm ceph-immutable-object-cache ceph-mgr ceph-mgr-dashboard ceph-mgr-diskprediction-local ceph-mgr-rook ceph-mgr-cephadm ceph-fuse ceph-volume librados-devel libcephfs2 libcephfs-devel librados2 librbd1 python3-rados python3-rgw python3-cephfs python3-rbd rbd-fuse rbd-mirror rbd-nbd ; do 2026-03-09T20:56:54.308 DEBUG:teuthology.orchestra.run.vm08:> sudo yum -y remove $d || true 2026-03-09T20:56:54.308 DEBUG:teuthology.orchestra.run.vm08:> done 2026-03-09T20:56:54.494 INFO:teuthology.orchestra.run.vm01.stdout:Dependencies resolved. 2026-03-09T20:56:54.494 INFO:teuthology.orchestra.run.vm01.stdout:================================================================================ 2026-03-09T20:56:54.494 INFO:teuthology.orchestra.run.vm01.stdout: Package Arch Version Repository Size 2026-03-09T20:56:54.494 INFO:teuthology.orchestra.run.vm01.stdout:================================================================================ 2026-03-09T20:56:54.494 INFO:teuthology.orchestra.run.vm01.stdout:Removing: 2026-03-09T20:56:54.494 INFO:teuthology.orchestra.run.vm01.stdout: ceph-radosgw x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 39 M 2026-03-09T20:56:54.494 INFO:teuthology.orchestra.run.vm01.stdout:Removing unused dependencies: 2026-03-09T20:56:54.494 INFO:teuthology.orchestra.run.vm01.stdout: mailcap noarch 2.1.49-5.el9 @baseos 78 k 2026-03-09T20:56:54.494 INFO:teuthology.orchestra.run.vm01.stdout: 2026-03-09T20:56:54.494 INFO:teuthology.orchestra.run.vm01.stdout:Transaction Summary 2026-03-09T20:56:54.494 INFO:teuthology.orchestra.run.vm01.stdout:================================================================================ 2026-03-09T20:56:54.494 INFO:teuthology.orchestra.run.vm01.stdout:Remove 2 Packages 2026-03-09T20:56:54.494 INFO:teuthology.orchestra.run.vm01.stdout: 2026-03-09T20:56:54.495 INFO:teuthology.orchestra.run.vm01.stdout:Freed space: 39 M 2026-03-09T20:56:54.495 INFO:teuthology.orchestra.run.vm01.stdout:Running transaction check 2026-03-09T20:56:54.496 INFO:teuthology.orchestra.run.vm06.stdout:Dependencies resolved. 2026-03-09T20:56:54.496 INFO:teuthology.orchestra.run.vm06.stdout:================================================================================ 2026-03-09T20:56:54.497 INFO:teuthology.orchestra.run.vm06.stdout: Package Arch Version Repository Size 2026-03-09T20:56:54.497 INFO:teuthology.orchestra.run.vm06.stdout:================================================================================ 2026-03-09T20:56:54.497 INFO:teuthology.orchestra.run.vm06.stdout:Removing: 2026-03-09T20:56:54.497 INFO:teuthology.orchestra.run.vm06.stdout: ceph-radosgw x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 39 M 2026-03-09T20:56:54.497 INFO:teuthology.orchestra.run.vm06.stdout:Removing unused dependencies: 2026-03-09T20:56:54.497 INFO:teuthology.orchestra.run.vm06.stdout: mailcap noarch 2.1.49-5.el9 @baseos 78 k 2026-03-09T20:56:54.497 INFO:teuthology.orchestra.run.vm06.stdout: 2026-03-09T20:56:54.497 INFO:teuthology.orchestra.run.vm06.stdout:Transaction Summary 2026-03-09T20:56:54.497 INFO:teuthology.orchestra.run.vm06.stdout:================================================================================ 2026-03-09T20:56:54.497 INFO:teuthology.orchestra.run.vm06.stdout:Remove 2 Packages 2026-03-09T20:56:54.497 INFO:teuthology.orchestra.run.vm06.stdout: 2026-03-09T20:56:54.497 INFO:teuthology.orchestra.run.vm06.stdout:Freed space: 39 M 2026-03-09T20:56:54.497 INFO:teuthology.orchestra.run.vm06.stdout:Running transaction check 2026-03-09T20:56:54.497 INFO:teuthology.orchestra.run.vm01.stdout:Transaction check succeeded. 2026-03-09T20:56:54.497 INFO:teuthology.orchestra.run.vm01.stdout:Running transaction test 2026-03-09T20:56:54.499 INFO:teuthology.orchestra.run.vm06.stdout:Transaction check succeeded. 2026-03-09T20:56:54.499 INFO:teuthology.orchestra.run.vm06.stdout:Running transaction test 2026-03-09T20:56:54.511 INFO:teuthology.orchestra.run.vm01.stdout:Transaction test succeeded. 2026-03-09T20:56:54.512 INFO:teuthology.orchestra.run.vm01.stdout:Running transaction 2026-03-09T20:56:54.514 INFO:teuthology.orchestra.run.vm06.stdout:Transaction test succeeded. 2026-03-09T20:56:54.515 INFO:teuthology.orchestra.run.vm06.stdout:Running transaction 2026-03-09T20:56:54.541 INFO:teuthology.orchestra.run.vm08.stdout:Dependencies resolved. 2026-03-09T20:56:54.542 INFO:teuthology.orchestra.run.vm08.stdout:================================================================================ 2026-03-09T20:56:54.542 INFO:teuthology.orchestra.run.vm08.stdout: Package Arch Version Repository Size 2026-03-09T20:56:54.542 INFO:teuthology.orchestra.run.vm08.stdout:================================================================================ 2026-03-09T20:56:54.542 INFO:teuthology.orchestra.run.vm08.stdout:Removing: 2026-03-09T20:56:54.542 INFO:teuthology.orchestra.run.vm08.stdout: ceph-radosgw x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 39 M 2026-03-09T20:56:54.542 INFO:teuthology.orchestra.run.vm08.stdout:Removing unused dependencies: 2026-03-09T20:56:54.542 INFO:teuthology.orchestra.run.vm08.stdout: mailcap noarch 2.1.49-5.el9 @baseos 78 k 2026-03-09T20:56:54.542 INFO:teuthology.orchestra.run.vm08.stdout: 2026-03-09T20:56:54.542 INFO:teuthology.orchestra.run.vm08.stdout:Transaction Summary 2026-03-09T20:56:54.542 INFO:teuthology.orchestra.run.vm08.stdout:================================================================================ 2026-03-09T20:56:54.542 INFO:teuthology.orchestra.run.vm08.stdout:Remove 2 Packages 2026-03-09T20:56:54.542 INFO:teuthology.orchestra.run.vm08.stdout: 2026-03-09T20:56:54.542 INFO:teuthology.orchestra.run.vm08.stdout:Freed space: 39 M 2026-03-09T20:56:54.542 INFO:teuthology.orchestra.run.vm08.stdout:Running transaction check 2026-03-09T20:56:54.544 INFO:teuthology.orchestra.run.vm08.stdout:Transaction check succeeded. 2026-03-09T20:56:54.544 INFO:teuthology.orchestra.run.vm08.stdout:Running transaction test 2026-03-09T20:56:54.545 INFO:teuthology.orchestra.run.vm01.stdout: Preparing : 1/1 2026-03-09T20:56:54.546 INFO:teuthology.orchestra.run.vm06.stdout: Preparing : 1/1 2026-03-09T20:56:54.559 INFO:teuthology.orchestra.run.vm08.stdout:Transaction test succeeded. 2026-03-09T20:56:54.559 INFO:teuthology.orchestra.run.vm08.stdout:Running transaction 2026-03-09T20:56:54.569 INFO:teuthology.orchestra.run.vm01.stdout: Running scriptlet: ceph-radosgw-2:19.2.3-678.ge911bdeb.el9.x86_64 1/2 2026-03-09T20:56:54.570 INFO:teuthology.orchestra.run.vm01.stdout:Glob pattern passed to enable, but globs are not supported for this. 2026-03-09T20:56:54.570 INFO:teuthology.orchestra.run.vm01.stdout:Invalid unit name "ceph-radosgw@*.service" escaped as "ceph-radosgw@\x2a.service". 2026-03-09T20:56:54.570 INFO:teuthology.orchestra.run.vm01.stdout:Removed "/etc/systemd/system/multi-user.target.wants/ceph-radosgw.target". 2026-03-09T20:56:54.570 INFO:teuthology.orchestra.run.vm01.stdout:Removed "/etc/systemd/system/ceph.target.wants/ceph-radosgw.target". 2026-03-09T20:56:54.570 INFO:teuthology.orchestra.run.vm01.stdout: 2026-03-09T20:56:54.570 INFO:teuthology.orchestra.run.vm06.stdout: Running scriptlet: ceph-radosgw-2:19.2.3-678.ge911bdeb.el9.x86_64 1/2 2026-03-09T20:56:54.570 INFO:teuthology.orchestra.run.vm06.stdout:Glob pattern passed to enable, but globs are not supported for this. 2026-03-09T20:56:54.570 INFO:teuthology.orchestra.run.vm06.stdout:Invalid unit name "ceph-radosgw@*.service" escaped as "ceph-radosgw@\x2a.service". 2026-03-09T20:56:54.570 INFO:teuthology.orchestra.run.vm06.stdout:Removed "/etc/systemd/system/multi-user.target.wants/ceph-radosgw.target". 2026-03-09T20:56:54.570 INFO:teuthology.orchestra.run.vm06.stdout:Removed "/etc/systemd/system/ceph.target.wants/ceph-radosgw.target". 2026-03-09T20:56:54.570 INFO:teuthology.orchestra.run.vm06.stdout: 2026-03-09T20:56:54.573 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : ceph-radosgw-2:19.2.3-678.ge911bdeb.el9.x86_64 1/2 2026-03-09T20:56:54.573 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : ceph-radosgw-2:19.2.3-678.ge911bdeb.el9.x86_64 1/2 2026-03-09T20:56:54.583 INFO:teuthology.orchestra.run.vm06.stdout: Running scriptlet: ceph-radosgw-2:19.2.3-678.ge911bdeb.el9.x86_64 1/2 2026-03-09T20:56:54.583 INFO:teuthology.orchestra.run.vm01.stdout: Running scriptlet: ceph-radosgw-2:19.2.3-678.ge911bdeb.el9.x86_64 1/2 2026-03-09T20:56:54.590 INFO:teuthology.orchestra.run.vm08.stdout: Preparing : 1/1 2026-03-09T20:56:54.598 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : mailcap-2.1.49-5.el9.noarch 2/2 2026-03-09T20:56:54.599 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : mailcap-2.1.49-5.el9.noarch 2/2 2026-03-09T20:56:54.612 INFO:teuthology.orchestra.run.vm08.stdout: Running scriptlet: ceph-radosgw-2:19.2.3-678.ge911bdeb.el9.x86_64 1/2 2026-03-09T20:56:54.612 INFO:teuthology.orchestra.run.vm08.stdout:Glob pattern passed to enable, but globs are not supported for this. 2026-03-09T20:56:54.612 INFO:teuthology.orchestra.run.vm08.stdout:Invalid unit name "ceph-radosgw@*.service" escaped as "ceph-radosgw@\x2a.service". 2026-03-09T20:56:54.612 INFO:teuthology.orchestra.run.vm08.stdout:Removed "/etc/systemd/system/multi-user.target.wants/ceph-radosgw.target". 2026-03-09T20:56:54.612 INFO:teuthology.orchestra.run.vm08.stdout:Removed "/etc/systemd/system/ceph.target.wants/ceph-radosgw.target". 2026-03-09T20:56:54.612 INFO:teuthology.orchestra.run.vm08.stdout: 2026-03-09T20:56:54.615 INFO:teuthology.orchestra.run.vm08.stdout: Erasing : ceph-radosgw-2:19.2.3-678.ge911bdeb.el9.x86_64 1/2 2026-03-09T20:56:54.625 INFO:teuthology.orchestra.run.vm08.stdout: Running scriptlet: ceph-radosgw-2:19.2.3-678.ge911bdeb.el9.x86_64 1/2 2026-03-09T20:56:54.640 INFO:teuthology.orchestra.run.vm08.stdout: Erasing : mailcap-2.1.49-5.el9.noarch 2/2 2026-03-09T20:56:54.663 INFO:teuthology.orchestra.run.vm06.stdout: Running scriptlet: mailcap-2.1.49-5.el9.noarch 2/2 2026-03-09T20:56:54.663 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : ceph-radosgw-2:19.2.3-678.ge911bdeb.el9.x86_64 1/2 2026-03-09T20:56:54.677 INFO:teuthology.orchestra.run.vm01.stdout: Running scriptlet: mailcap-2.1.49-5.el9.noarch 2/2 2026-03-09T20:56:54.677 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : ceph-radosgw-2:19.2.3-678.ge911bdeb.el9.x86_64 1/2 2026-03-09T20:56:54.724 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : mailcap-2.1.49-5.el9.noarch 2/2 2026-03-09T20:56:54.724 INFO:teuthology.orchestra.run.vm06.stdout: 2026-03-09T20:56:54.724 INFO:teuthology.orchestra.run.vm06.stdout:Removed: 2026-03-09T20:56:54.724 INFO:teuthology.orchestra.run.vm06.stdout: ceph-radosgw-2:19.2.3-678.ge911bdeb.el9.x86_64 mailcap-2.1.49-5.el9.noarch 2026-03-09T20:56:54.724 INFO:teuthology.orchestra.run.vm06.stdout: 2026-03-09T20:56:54.724 INFO:teuthology.orchestra.run.vm06.stdout:Complete! 2026-03-09T20:56:54.734 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : mailcap-2.1.49-5.el9.noarch 2/2 2026-03-09T20:56:54.734 INFO:teuthology.orchestra.run.vm01.stdout: 2026-03-09T20:56:54.735 INFO:teuthology.orchestra.run.vm01.stdout:Removed: 2026-03-09T20:56:54.735 INFO:teuthology.orchestra.run.vm01.stdout: ceph-radosgw-2:19.2.3-678.ge911bdeb.el9.x86_64 mailcap-2.1.49-5.el9.noarch 2026-03-09T20:56:54.735 INFO:teuthology.orchestra.run.vm01.stdout: 2026-03-09T20:56:54.735 INFO:teuthology.orchestra.run.vm01.stdout:Complete! 2026-03-09T20:56:54.743 INFO:teuthology.orchestra.run.vm08.stdout: Running scriptlet: mailcap-2.1.49-5.el9.noarch 2/2 2026-03-09T20:56:54.743 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : ceph-radosgw-2:19.2.3-678.ge911bdeb.el9.x86_64 1/2 2026-03-09T20:56:54.803 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : mailcap-2.1.49-5.el9.noarch 2/2 2026-03-09T20:56:54.803 INFO:teuthology.orchestra.run.vm08.stdout: 2026-03-09T20:56:54.803 INFO:teuthology.orchestra.run.vm08.stdout:Removed: 2026-03-09T20:56:54.803 INFO:teuthology.orchestra.run.vm08.stdout: ceph-radosgw-2:19.2.3-678.ge911bdeb.el9.x86_64 mailcap-2.1.49-5.el9.noarch 2026-03-09T20:56:54.803 INFO:teuthology.orchestra.run.vm08.stdout: 2026-03-09T20:56:54.803 INFO:teuthology.orchestra.run.vm08.stdout:Complete! 2026-03-09T20:56:54.944 INFO:teuthology.orchestra.run.vm06.stdout:Dependencies resolved. 2026-03-09T20:56:54.945 INFO:teuthology.orchestra.run.vm06.stdout:================================================================================ 2026-03-09T20:56:54.945 INFO:teuthology.orchestra.run.vm06.stdout: Package Arch Version Repository Size 2026-03-09T20:56:54.945 INFO:teuthology.orchestra.run.vm06.stdout:================================================================================ 2026-03-09T20:56:54.945 INFO:teuthology.orchestra.run.vm06.stdout:Removing: 2026-03-09T20:56:54.945 INFO:teuthology.orchestra.run.vm06.stdout: ceph-test x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 210 M 2026-03-09T20:56:54.945 INFO:teuthology.orchestra.run.vm06.stdout:Removing unused dependencies: 2026-03-09T20:56:54.945 INFO:teuthology.orchestra.run.vm06.stdout: libxslt x86_64 1.1.34-12.el9 @appstream 743 k 2026-03-09T20:56:54.945 INFO:teuthology.orchestra.run.vm06.stdout: socat x86_64 1.7.4.1-8.el9 @appstream 1.1 M 2026-03-09T20:56:54.945 INFO:teuthology.orchestra.run.vm06.stdout: xmlstarlet x86_64 1.6.1-20.el9 @appstream 195 k 2026-03-09T20:56:54.945 INFO:teuthology.orchestra.run.vm06.stdout: 2026-03-09T20:56:54.945 INFO:teuthology.orchestra.run.vm06.stdout:Transaction Summary 2026-03-09T20:56:54.945 INFO:teuthology.orchestra.run.vm06.stdout:================================================================================ 2026-03-09T20:56:54.945 INFO:teuthology.orchestra.run.vm06.stdout:Remove 4 Packages 2026-03-09T20:56:54.945 INFO:teuthology.orchestra.run.vm06.stdout: 2026-03-09T20:56:54.945 INFO:teuthology.orchestra.run.vm06.stdout:Freed space: 212 M 2026-03-09T20:56:54.945 INFO:teuthology.orchestra.run.vm06.stdout:Running transaction check 2026-03-09T20:56:54.948 INFO:teuthology.orchestra.run.vm01.stdout:Dependencies resolved. 2026-03-09T20:56:54.948 INFO:teuthology.orchestra.run.vm06.stdout:Transaction check succeeded. 2026-03-09T20:56:54.948 INFO:teuthology.orchestra.run.vm06.stdout:Running transaction test 2026-03-09T20:56:54.948 INFO:teuthology.orchestra.run.vm01.stdout:================================================================================ 2026-03-09T20:56:54.948 INFO:teuthology.orchestra.run.vm01.stdout: Package Arch Version Repository Size 2026-03-09T20:56:54.948 INFO:teuthology.orchestra.run.vm01.stdout:================================================================================ 2026-03-09T20:56:54.948 INFO:teuthology.orchestra.run.vm01.stdout:Removing: 2026-03-09T20:56:54.948 INFO:teuthology.orchestra.run.vm01.stdout: ceph-test x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 210 M 2026-03-09T20:56:54.948 INFO:teuthology.orchestra.run.vm01.stdout:Removing unused dependencies: 2026-03-09T20:56:54.948 INFO:teuthology.orchestra.run.vm01.stdout: libxslt x86_64 1.1.34-12.el9 @appstream 743 k 2026-03-09T20:56:54.948 INFO:teuthology.orchestra.run.vm01.stdout: socat x86_64 1.7.4.1-8.el9 @appstream 1.1 M 2026-03-09T20:56:54.948 INFO:teuthology.orchestra.run.vm01.stdout: xmlstarlet x86_64 1.6.1-20.el9 @appstream 195 k 2026-03-09T20:56:54.948 INFO:teuthology.orchestra.run.vm01.stdout: 2026-03-09T20:56:54.948 INFO:teuthology.orchestra.run.vm01.stdout:Transaction Summary 2026-03-09T20:56:54.948 INFO:teuthology.orchestra.run.vm01.stdout:================================================================================ 2026-03-09T20:56:54.948 INFO:teuthology.orchestra.run.vm01.stdout:Remove 4 Packages 2026-03-09T20:56:54.949 INFO:teuthology.orchestra.run.vm01.stdout: 2026-03-09T20:56:54.949 INFO:teuthology.orchestra.run.vm01.stdout:Freed space: 212 M 2026-03-09T20:56:54.949 INFO:teuthology.orchestra.run.vm01.stdout:Running transaction check 2026-03-09T20:56:54.952 INFO:teuthology.orchestra.run.vm01.stdout:Transaction check succeeded. 2026-03-09T20:56:54.952 INFO:teuthology.orchestra.run.vm01.stdout:Running transaction test 2026-03-09T20:56:54.971 INFO:teuthology.orchestra.run.vm06.stdout:Transaction test succeeded. 2026-03-09T20:56:54.971 INFO:teuthology.orchestra.run.vm06.stdout:Running transaction 2026-03-09T20:56:54.975 INFO:teuthology.orchestra.run.vm01.stdout:Transaction test succeeded. 2026-03-09T20:56:54.975 INFO:teuthology.orchestra.run.vm01.stdout:Running transaction 2026-03-09T20:56:55.034 INFO:teuthology.orchestra.run.vm06.stdout: Preparing : 1/1 2026-03-09T20:56:55.041 INFO:teuthology.orchestra.run.vm01.stdout: Preparing : 1/1 2026-03-09T20:56:55.041 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : ceph-test-2:19.2.3-678.ge911bdeb.el9.x86_64 1/4 2026-03-09T20:56:55.044 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : xmlstarlet-1.6.1-20.el9.x86_64 2/4 2026-03-09T20:56:55.044 INFO:teuthology.orchestra.run.vm08.stdout:Dependencies resolved. 2026-03-09T20:56:55.045 INFO:teuthology.orchestra.run.vm08.stdout:================================================================================ 2026-03-09T20:56:55.045 INFO:teuthology.orchestra.run.vm08.stdout: Package Arch Version Repository Size 2026-03-09T20:56:55.045 INFO:teuthology.orchestra.run.vm08.stdout:================================================================================ 2026-03-09T20:56:55.045 INFO:teuthology.orchestra.run.vm08.stdout:Removing: 2026-03-09T20:56:55.045 INFO:teuthology.orchestra.run.vm08.stdout: ceph-test x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 210 M 2026-03-09T20:56:55.045 INFO:teuthology.orchestra.run.vm08.stdout:Removing unused dependencies: 2026-03-09T20:56:55.045 INFO:teuthology.orchestra.run.vm08.stdout: libxslt x86_64 1.1.34-12.el9 @appstream 743 k 2026-03-09T20:56:55.045 INFO:teuthology.orchestra.run.vm08.stdout: socat x86_64 1.7.4.1-8.el9 @appstream 1.1 M 2026-03-09T20:56:55.045 INFO:teuthology.orchestra.run.vm08.stdout: xmlstarlet x86_64 1.6.1-20.el9 @appstream 195 k 2026-03-09T20:56:55.045 INFO:teuthology.orchestra.run.vm08.stdout: 2026-03-09T20:56:55.045 INFO:teuthology.orchestra.run.vm08.stdout:Transaction Summary 2026-03-09T20:56:55.045 INFO:teuthology.orchestra.run.vm08.stdout:================================================================================ 2026-03-09T20:56:55.045 INFO:teuthology.orchestra.run.vm08.stdout:Remove 4 Packages 2026-03-09T20:56:55.045 INFO:teuthology.orchestra.run.vm08.stdout: 2026-03-09T20:56:55.045 INFO:teuthology.orchestra.run.vm08.stdout:Freed space: 212 M 2026-03-09T20:56:55.045 INFO:teuthology.orchestra.run.vm08.stdout:Running transaction check 2026-03-09T20:56:55.048 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : libxslt-1.1.34-12.el9.x86_64 3/4 2026-03-09T20:56:55.049 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : ceph-test-2:19.2.3-678.ge911bdeb.el9.x86_64 1/4 2026-03-09T20:56:55.049 INFO:teuthology.orchestra.run.vm08.stdout:Transaction check succeeded. 2026-03-09T20:56:55.049 INFO:teuthology.orchestra.run.vm08.stdout:Running transaction test 2026-03-09T20:56:55.052 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : xmlstarlet-1.6.1-20.el9.x86_64 2/4 2026-03-09T20:56:55.055 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : libxslt-1.1.34-12.el9.x86_64 3/4 2026-03-09T20:56:55.066 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : socat-1.7.4.1-8.el9.x86_64 4/4 2026-03-09T20:56:55.071 INFO:teuthology.orchestra.run.vm08.stdout:Transaction test succeeded. 2026-03-09T20:56:55.072 INFO:teuthology.orchestra.run.vm08.stdout:Running transaction 2026-03-09T20:56:55.072 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : socat-1.7.4.1-8.el9.x86_64 4/4 2026-03-09T20:56:55.138 INFO:teuthology.orchestra.run.vm08.stdout: Preparing : 1/1 2026-03-09T20:56:55.142 INFO:teuthology.orchestra.run.vm01.stdout: Running scriptlet: socat-1.7.4.1-8.el9.x86_64 4/4 2026-03-09T20:56:55.142 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : ceph-test-2:19.2.3-678.ge911bdeb.el9.x86_64 1/4 2026-03-09T20:56:55.142 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : libxslt-1.1.34-12.el9.x86_64 2/4 2026-03-09T20:56:55.142 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : socat-1.7.4.1-8.el9.x86_64 3/4 2026-03-09T20:56:55.143 INFO:teuthology.orchestra.run.vm06.stdout: Running scriptlet: socat-1.7.4.1-8.el9.x86_64 4/4 2026-03-09T20:56:55.143 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : ceph-test-2:19.2.3-678.ge911bdeb.el9.x86_64 1/4 2026-03-09T20:56:55.143 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : libxslt-1.1.34-12.el9.x86_64 2/4 2026-03-09T20:56:55.143 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : socat-1.7.4.1-8.el9.x86_64 3/4 2026-03-09T20:56:55.147 INFO:teuthology.orchestra.run.vm08.stdout: Erasing : ceph-test-2:19.2.3-678.ge911bdeb.el9.x86_64 1/4 2026-03-09T20:56:55.150 INFO:teuthology.orchestra.run.vm08.stdout: Erasing : xmlstarlet-1.6.1-20.el9.x86_64 2/4 2026-03-09T20:56:55.153 INFO:teuthology.orchestra.run.vm08.stdout: Erasing : libxslt-1.1.34-12.el9.x86_64 3/4 2026-03-09T20:56:55.170 INFO:teuthology.orchestra.run.vm08.stdout: Erasing : socat-1.7.4.1-8.el9.x86_64 4/4 2026-03-09T20:56:55.197 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : xmlstarlet-1.6.1-20.el9.x86_64 4/4 2026-03-09T20:56:55.197 INFO:teuthology.orchestra.run.vm01.stdout: 2026-03-09T20:56:55.197 INFO:teuthology.orchestra.run.vm01.stdout:Removed: 2026-03-09T20:56:55.197 INFO:teuthology.orchestra.run.vm01.stdout: ceph-test-2:19.2.3-678.ge911bdeb.el9.x86_64 libxslt-1.1.34-12.el9.x86_64 2026-03-09T20:56:55.197 INFO:teuthology.orchestra.run.vm01.stdout: socat-1.7.4.1-8.el9.x86_64 xmlstarlet-1.6.1-20.el9.x86_64 2026-03-09T20:56:55.197 INFO:teuthology.orchestra.run.vm01.stdout: 2026-03-09T20:56:55.197 INFO:teuthology.orchestra.run.vm01.stdout:Complete! 2026-03-09T20:56:55.204 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : xmlstarlet-1.6.1-20.el9.x86_64 4/4 2026-03-09T20:56:55.204 INFO:teuthology.orchestra.run.vm06.stdout: 2026-03-09T20:56:55.204 INFO:teuthology.orchestra.run.vm06.stdout:Removed: 2026-03-09T20:56:55.204 INFO:teuthology.orchestra.run.vm06.stdout: ceph-test-2:19.2.3-678.ge911bdeb.el9.x86_64 libxslt-1.1.34-12.el9.x86_64 2026-03-09T20:56:55.204 INFO:teuthology.orchestra.run.vm06.stdout: socat-1.7.4.1-8.el9.x86_64 xmlstarlet-1.6.1-20.el9.x86_64 2026-03-09T20:56:55.204 INFO:teuthology.orchestra.run.vm06.stdout: 2026-03-09T20:56:55.204 INFO:teuthology.orchestra.run.vm06.stdout:Complete! 2026-03-09T20:56:55.243 INFO:teuthology.orchestra.run.vm08.stdout: Running scriptlet: socat-1.7.4.1-8.el9.x86_64 4/4 2026-03-09T20:56:55.243 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : ceph-test-2:19.2.3-678.ge911bdeb.el9.x86_64 1/4 2026-03-09T20:56:55.243 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : libxslt-1.1.34-12.el9.x86_64 2/4 2026-03-09T20:56:55.243 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : socat-1.7.4.1-8.el9.x86_64 3/4 2026-03-09T20:56:55.297 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : xmlstarlet-1.6.1-20.el9.x86_64 4/4 2026-03-09T20:56:55.297 INFO:teuthology.orchestra.run.vm08.stdout: 2026-03-09T20:56:55.297 INFO:teuthology.orchestra.run.vm08.stdout:Removed: 2026-03-09T20:56:55.297 INFO:teuthology.orchestra.run.vm08.stdout: ceph-test-2:19.2.3-678.ge911bdeb.el9.x86_64 libxslt-1.1.34-12.el9.x86_64 2026-03-09T20:56:55.297 INFO:teuthology.orchestra.run.vm08.stdout: socat-1.7.4.1-8.el9.x86_64 xmlstarlet-1.6.1-20.el9.x86_64 2026-03-09T20:56:55.297 INFO:teuthology.orchestra.run.vm08.stdout: 2026-03-09T20:56:55.297 INFO:teuthology.orchestra.run.vm08.stdout:Complete! 2026-03-09T20:56:55.423 INFO:teuthology.orchestra.run.vm01.stdout:Dependencies resolved. 2026-03-09T20:56:55.424 INFO:teuthology.orchestra.run.vm01.stdout:================================================================================ 2026-03-09T20:56:55.424 INFO:teuthology.orchestra.run.vm01.stdout: Package Arch Version Repository Size 2026-03-09T20:56:55.424 INFO:teuthology.orchestra.run.vm01.stdout:================================================================================ 2026-03-09T20:56:55.424 INFO:teuthology.orchestra.run.vm01.stdout:Removing: 2026-03-09T20:56:55.424 INFO:teuthology.orchestra.run.vm01.stdout: ceph x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 0 2026-03-09T20:56:55.424 INFO:teuthology.orchestra.run.vm01.stdout:Removing unused dependencies: 2026-03-09T20:56:55.424 INFO:teuthology.orchestra.run.vm01.stdout: ceph-mds x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 7.5 M 2026-03-09T20:56:55.424 INFO:teuthology.orchestra.run.vm01.stdout: ceph-mon x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 18 M 2026-03-09T20:56:55.424 INFO:teuthology.orchestra.run.vm01.stdout: lua x86_64 5.4.4-4.el9 @appstream 593 k 2026-03-09T20:56:55.424 INFO:teuthology.orchestra.run.vm01.stdout: lua-devel x86_64 5.4.4-4.el9 @crb 49 k 2026-03-09T20:56:55.424 INFO:teuthology.orchestra.run.vm01.stdout: luarocks noarch 3.9.2-5.el9 @epel 692 k 2026-03-09T20:56:55.424 INFO:teuthology.orchestra.run.vm01.stdout: unzip x86_64 6.0-59.el9 @baseos 389 k 2026-03-09T20:56:55.424 INFO:teuthology.orchestra.run.vm01.stdout: zip x86_64 3.0-35.el9 @baseos 724 k 2026-03-09T20:56:55.424 INFO:teuthology.orchestra.run.vm01.stdout: 2026-03-09T20:56:55.425 INFO:teuthology.orchestra.run.vm01.stdout:Transaction Summary 2026-03-09T20:56:55.425 INFO:teuthology.orchestra.run.vm01.stdout:================================================================================ 2026-03-09T20:56:55.425 INFO:teuthology.orchestra.run.vm01.stdout:Remove 8 Packages 2026-03-09T20:56:55.425 INFO:teuthology.orchestra.run.vm01.stdout: 2026-03-09T20:56:55.425 INFO:teuthology.orchestra.run.vm01.stdout:Freed space: 28 M 2026-03-09T20:56:55.425 INFO:teuthology.orchestra.run.vm01.stdout:Running transaction check 2026-03-09T20:56:55.427 INFO:teuthology.orchestra.run.vm01.stdout:Transaction check succeeded. 2026-03-09T20:56:55.427 INFO:teuthology.orchestra.run.vm01.stdout:Running transaction test 2026-03-09T20:56:55.445 INFO:teuthology.orchestra.run.vm06.stdout:Dependencies resolved. 2026-03-09T20:56:55.446 INFO:teuthology.orchestra.run.vm06.stdout:================================================================================ 2026-03-09T20:56:55.446 INFO:teuthology.orchestra.run.vm06.stdout: Package Arch Version Repository Size 2026-03-09T20:56:55.446 INFO:teuthology.orchestra.run.vm06.stdout:================================================================================ 2026-03-09T20:56:55.446 INFO:teuthology.orchestra.run.vm06.stdout:Removing: 2026-03-09T20:56:55.446 INFO:teuthology.orchestra.run.vm06.stdout: ceph x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 0 2026-03-09T20:56:55.446 INFO:teuthology.orchestra.run.vm06.stdout:Removing unused dependencies: 2026-03-09T20:56:55.446 INFO:teuthology.orchestra.run.vm06.stdout: ceph-mds x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 7.5 M 2026-03-09T20:56:55.446 INFO:teuthology.orchestra.run.vm06.stdout: ceph-mon x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 18 M 2026-03-09T20:56:55.446 INFO:teuthology.orchestra.run.vm06.stdout: lua x86_64 5.4.4-4.el9 @appstream 593 k 2026-03-09T20:56:55.446 INFO:teuthology.orchestra.run.vm06.stdout: lua-devel x86_64 5.4.4-4.el9 @crb 49 k 2026-03-09T20:56:55.446 INFO:teuthology.orchestra.run.vm06.stdout: luarocks noarch 3.9.2-5.el9 @epel 692 k 2026-03-09T20:56:55.446 INFO:teuthology.orchestra.run.vm06.stdout: unzip x86_64 6.0-59.el9 @baseos 389 k 2026-03-09T20:56:55.446 INFO:teuthology.orchestra.run.vm06.stdout: zip x86_64 3.0-35.el9 @baseos 724 k 2026-03-09T20:56:55.446 INFO:teuthology.orchestra.run.vm06.stdout: 2026-03-09T20:56:55.446 INFO:teuthology.orchestra.run.vm06.stdout:Transaction Summary 2026-03-09T20:56:55.446 INFO:teuthology.orchestra.run.vm06.stdout:================================================================================ 2026-03-09T20:56:55.446 INFO:teuthology.orchestra.run.vm06.stdout:Remove 8 Packages 2026-03-09T20:56:55.446 INFO:teuthology.orchestra.run.vm06.stdout: 2026-03-09T20:56:55.446 INFO:teuthology.orchestra.run.vm06.stdout:Freed space: 28 M 2026-03-09T20:56:55.446 INFO:teuthology.orchestra.run.vm06.stdout:Running transaction check 2026-03-09T20:56:55.449 INFO:teuthology.orchestra.run.vm06.stdout:Transaction check succeeded. 2026-03-09T20:56:55.449 INFO:teuthology.orchestra.run.vm06.stdout:Running transaction test 2026-03-09T20:56:55.452 INFO:teuthology.orchestra.run.vm01.stdout:Transaction test succeeded. 2026-03-09T20:56:55.452 INFO:teuthology.orchestra.run.vm01.stdout:Running transaction 2026-03-09T20:56:55.474 INFO:teuthology.orchestra.run.vm06.stdout:Transaction test succeeded. 2026-03-09T20:56:55.474 INFO:teuthology.orchestra.run.vm06.stdout:Running transaction 2026-03-09T20:56:55.496 INFO:teuthology.orchestra.run.vm01.stdout: Preparing : 1/1 2026-03-09T20:56:55.501 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : ceph-2:19.2.3-678.ge911bdeb.el9.x86_64 1/8 2026-03-09T20:56:55.504 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : luarocks-3.9.2-5.el9.noarch 2/8 2026-03-09T20:56:55.506 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : lua-devel-5.4.4-4.el9.x86_64 3/8 2026-03-09T20:56:55.509 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : zip-3.0-35.el9.x86_64 4/8 2026-03-09T20:56:55.511 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : unzip-6.0-59.el9.x86_64 5/8 2026-03-09T20:56:55.513 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : lua-5.4.4-4.el9.x86_64 6/8 2026-03-09T20:56:55.517 INFO:teuthology.orchestra.run.vm06.stdout: Preparing : 1/1 2026-03-09T20:56:55.519 INFO:teuthology.orchestra.run.vm08.stdout:Dependencies resolved. 2026-03-09T20:56:55.520 INFO:teuthology.orchestra.run.vm08.stdout:================================================================================ 2026-03-09T20:56:55.520 INFO:teuthology.orchestra.run.vm08.stdout: Package Arch Version Repository Size 2026-03-09T20:56:55.520 INFO:teuthology.orchestra.run.vm08.stdout:================================================================================ 2026-03-09T20:56:55.520 INFO:teuthology.orchestra.run.vm08.stdout:Removing: 2026-03-09T20:56:55.520 INFO:teuthology.orchestra.run.vm08.stdout: ceph x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 0 2026-03-09T20:56:55.520 INFO:teuthology.orchestra.run.vm08.stdout:Removing unused dependencies: 2026-03-09T20:56:55.520 INFO:teuthology.orchestra.run.vm08.stdout: ceph-mds x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 7.5 M 2026-03-09T20:56:55.520 INFO:teuthology.orchestra.run.vm08.stdout: ceph-mon x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 18 M 2026-03-09T20:56:55.520 INFO:teuthology.orchestra.run.vm08.stdout: lua x86_64 5.4.4-4.el9 @appstream 593 k 2026-03-09T20:56:55.520 INFO:teuthology.orchestra.run.vm08.stdout: lua-devel x86_64 5.4.4-4.el9 @crb 49 k 2026-03-09T20:56:55.520 INFO:teuthology.orchestra.run.vm08.stdout: luarocks noarch 3.9.2-5.el9 @epel 692 k 2026-03-09T20:56:55.520 INFO:teuthology.orchestra.run.vm08.stdout: unzip x86_64 6.0-59.el9 @baseos 389 k 2026-03-09T20:56:55.520 INFO:teuthology.orchestra.run.vm08.stdout: zip x86_64 3.0-35.el9 @baseos 724 k 2026-03-09T20:56:55.520 INFO:teuthology.orchestra.run.vm08.stdout: 2026-03-09T20:56:55.520 INFO:teuthology.orchestra.run.vm08.stdout:Transaction Summary 2026-03-09T20:56:55.520 INFO:teuthology.orchestra.run.vm08.stdout:================================================================================ 2026-03-09T20:56:55.520 INFO:teuthology.orchestra.run.vm08.stdout:Remove 8 Packages 2026-03-09T20:56:55.520 INFO:teuthology.orchestra.run.vm08.stdout: 2026-03-09T20:56:55.520 INFO:teuthology.orchestra.run.vm08.stdout:Freed space: 28 M 2026-03-09T20:56:55.520 INFO:teuthology.orchestra.run.vm08.stdout:Running transaction check 2026-03-09T20:56:55.522 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : ceph-2:19.2.3-678.ge911bdeb.el9.x86_64 1/8 2026-03-09T20:56:55.523 INFO:teuthology.orchestra.run.vm08.stdout:Transaction check succeeded. 2026-03-09T20:56:55.523 INFO:teuthology.orchestra.run.vm08.stdout:Running transaction test 2026-03-09T20:56:55.526 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : luarocks-3.9.2-5.el9.noarch 2/8 2026-03-09T20:56:55.529 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : lua-devel-5.4.4-4.el9.x86_64 3/8 2026-03-09T20:56:55.532 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : zip-3.0-35.el9.x86_64 4/8 2026-03-09T20:56:55.534 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : unzip-6.0-59.el9.x86_64 5/8 2026-03-09T20:56:55.536 INFO:teuthology.orchestra.run.vm01.stdout: Running scriptlet: ceph-mds-2:19.2.3-678.ge911bdeb.el9.x86_64 7/8 2026-03-09T20:56:55.536 INFO:teuthology.orchestra.run.vm01.stdout:Glob pattern passed to enable, but globs are not supported for this. 2026-03-09T20:56:55.536 INFO:teuthology.orchestra.run.vm01.stdout:Invalid unit name "ceph-mds@*.service" escaped as "ceph-mds@\x2a.service". 2026-03-09T20:56:55.536 INFO:teuthology.orchestra.run.vm01.stdout:Removed "/etc/systemd/system/multi-user.target.wants/ceph-mds.target". 2026-03-09T20:56:55.536 INFO:teuthology.orchestra.run.vm01.stdout:Removed "/etc/systemd/system/ceph.target.wants/ceph-mds.target". 2026-03-09T20:56:55.536 INFO:teuthology.orchestra.run.vm01.stdout: 2026-03-09T20:56:55.536 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : lua-5.4.4-4.el9.x86_64 6/8 2026-03-09T20:56:55.536 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : ceph-mds-2:19.2.3-678.ge911bdeb.el9.x86_64 7/8 2026-03-09T20:56:55.546 INFO:teuthology.orchestra.run.vm01.stdout: Running scriptlet: ceph-mds-2:19.2.3-678.ge911bdeb.el9.x86_64 7/8 2026-03-09T20:56:55.548 INFO:teuthology.orchestra.run.vm08.stdout:Transaction test succeeded. 2026-03-09T20:56:55.548 INFO:teuthology.orchestra.run.vm08.stdout:Running transaction 2026-03-09T20:56:55.556 INFO:teuthology.orchestra.run.vm06.stdout: Running scriptlet: ceph-mds-2:19.2.3-678.ge911bdeb.el9.x86_64 7/8 2026-03-09T20:56:55.557 INFO:teuthology.orchestra.run.vm06.stdout:Glob pattern passed to enable, but globs are not supported for this. 2026-03-09T20:56:55.557 INFO:teuthology.orchestra.run.vm06.stdout:Invalid unit name "ceph-mds@*.service" escaped as "ceph-mds@\x2a.service". 2026-03-09T20:56:55.557 INFO:teuthology.orchestra.run.vm06.stdout:Removed "/etc/systemd/system/multi-user.target.wants/ceph-mds.target". 2026-03-09T20:56:55.557 INFO:teuthology.orchestra.run.vm06.stdout:Removed "/etc/systemd/system/ceph.target.wants/ceph-mds.target". 2026-03-09T20:56:55.557 INFO:teuthology.orchestra.run.vm06.stdout: 2026-03-09T20:56:55.558 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : ceph-mds-2:19.2.3-678.ge911bdeb.el9.x86_64 7/8 2026-03-09T20:56:55.567 INFO:teuthology.orchestra.run.vm06.stdout: Running scriptlet: ceph-mds-2:19.2.3-678.ge911bdeb.el9.x86_64 7/8 2026-03-09T20:56:55.570 INFO:teuthology.orchestra.run.vm01.stdout: Running scriptlet: ceph-mon-2:19.2.3-678.ge911bdeb.el9.x86_64 8/8 2026-03-09T20:56:55.570 INFO:teuthology.orchestra.run.vm01.stdout:Glob pattern passed to enable, but globs are not supported for this. 2026-03-09T20:56:55.570 INFO:teuthology.orchestra.run.vm01.stdout:Invalid unit name "ceph-mon@*.service" escaped as "ceph-mon@\x2a.service". 2026-03-09T20:56:55.570 INFO:teuthology.orchestra.run.vm01.stdout:Removed "/etc/systemd/system/multi-user.target.wants/ceph-mon.target". 2026-03-09T20:56:55.570 INFO:teuthology.orchestra.run.vm01.stdout:Removed "/etc/systemd/system/ceph.target.wants/ceph-mon.target". 2026-03-09T20:56:55.570 INFO:teuthology.orchestra.run.vm01.stdout: 2026-03-09T20:56:55.572 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : ceph-mon-2:19.2.3-678.ge911bdeb.el9.x86_64 8/8 2026-03-09T20:56:55.589 INFO:teuthology.orchestra.run.vm06.stdout: Running scriptlet: ceph-mon-2:19.2.3-678.ge911bdeb.el9.x86_64 8/8 2026-03-09T20:56:55.589 INFO:teuthology.orchestra.run.vm06.stdout:Glob pattern passed to enable, but globs are not supported for this. 2026-03-09T20:56:55.589 INFO:teuthology.orchestra.run.vm06.stdout:Invalid unit name "ceph-mon@*.service" escaped as "ceph-mon@\x2a.service". 2026-03-09T20:56:55.589 INFO:teuthology.orchestra.run.vm06.stdout:Removed "/etc/systemd/system/multi-user.target.wants/ceph-mon.target". 2026-03-09T20:56:55.589 INFO:teuthology.orchestra.run.vm06.stdout:Removed "/etc/systemd/system/ceph.target.wants/ceph-mon.target". 2026-03-09T20:56:55.589 INFO:teuthology.orchestra.run.vm06.stdout: 2026-03-09T20:56:55.591 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : ceph-mon-2:19.2.3-678.ge911bdeb.el9.x86_64 8/8 2026-03-09T20:56:55.592 INFO:teuthology.orchestra.run.vm08.stdout: Preparing : 1/1 2026-03-09T20:56:55.597 INFO:teuthology.orchestra.run.vm08.stdout: Erasing : ceph-2:19.2.3-678.ge911bdeb.el9.x86_64 1/8 2026-03-09T20:56:55.601 INFO:teuthology.orchestra.run.vm08.stdout: Erasing : luarocks-3.9.2-5.el9.noarch 2/8 2026-03-09T20:56:55.604 INFO:teuthology.orchestra.run.vm08.stdout: Erasing : lua-devel-5.4.4-4.el9.x86_64 3/8 2026-03-09T20:56:55.607 INFO:teuthology.orchestra.run.vm08.stdout: Erasing : zip-3.0-35.el9.x86_64 4/8 2026-03-09T20:56:55.610 INFO:teuthology.orchestra.run.vm08.stdout: Erasing : unzip-6.0-59.el9.x86_64 5/8 2026-03-09T20:56:55.612 INFO:teuthology.orchestra.run.vm08.stdout: Erasing : lua-5.4.4-4.el9.x86_64 6/8 2026-03-09T20:56:55.633 INFO:teuthology.orchestra.run.vm08.stdout: Running scriptlet: ceph-mds-2:19.2.3-678.ge911bdeb.el9.x86_64 7/8 2026-03-09T20:56:55.633 INFO:teuthology.orchestra.run.vm08.stdout:Glob pattern passed to enable, but globs are not supported for this. 2026-03-09T20:56:55.633 INFO:teuthology.orchestra.run.vm08.stdout:Invalid unit name "ceph-mds@*.service" escaped as "ceph-mds@\x2a.service". 2026-03-09T20:56:55.633 INFO:teuthology.orchestra.run.vm08.stdout:Removed "/etc/systemd/system/multi-user.target.wants/ceph-mds.target". 2026-03-09T20:56:55.633 INFO:teuthology.orchestra.run.vm08.stdout:Removed "/etc/systemd/system/ceph.target.wants/ceph-mds.target". 2026-03-09T20:56:55.633 INFO:teuthology.orchestra.run.vm08.stdout: 2026-03-09T20:56:55.634 INFO:teuthology.orchestra.run.vm08.stdout: Erasing : ceph-mds-2:19.2.3-678.ge911bdeb.el9.x86_64 7/8 2026-03-09T20:56:55.642 INFO:teuthology.orchestra.run.vm08.stdout: Running scriptlet: ceph-mds-2:19.2.3-678.ge911bdeb.el9.x86_64 7/8 2026-03-09T20:56:55.664 INFO:teuthology.orchestra.run.vm01.stdout: Running scriptlet: ceph-mon-2:19.2.3-678.ge911bdeb.el9.x86_64 8/8 2026-03-09T20:56:55.664 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : ceph-2:19.2.3-678.ge911bdeb.el9.x86_64 1/8 2026-03-09T20:56:55.664 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : ceph-mds-2:19.2.3-678.ge911bdeb.el9.x86_64 2/8 2026-03-09T20:56:55.664 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : ceph-mon-2:19.2.3-678.ge911bdeb.el9.x86_64 3/8 2026-03-09T20:56:55.664 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : lua-5.4.4-4.el9.x86_64 4/8 2026-03-09T20:56:55.664 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : lua-devel-5.4.4-4.el9.x86_64 5/8 2026-03-09T20:56:55.664 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : luarocks-3.9.2-5.el9.noarch 6/8 2026-03-09T20:56:55.664 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : unzip-6.0-59.el9.x86_64 7/8 2026-03-09T20:56:55.664 INFO:teuthology.orchestra.run.vm08.stdout: Running scriptlet: ceph-mon-2:19.2.3-678.ge911bdeb.el9.x86_64 8/8 2026-03-09T20:56:55.664 INFO:teuthology.orchestra.run.vm08.stdout:Glob pattern passed to enable, but globs are not supported for this. 2026-03-09T20:56:55.664 INFO:teuthology.orchestra.run.vm08.stdout:Invalid unit name "ceph-mon@*.service" escaped as "ceph-mon@\x2a.service". 2026-03-09T20:56:55.664 INFO:teuthology.orchestra.run.vm08.stdout:Removed "/etc/systemd/system/multi-user.target.wants/ceph-mon.target". 2026-03-09T20:56:55.664 INFO:teuthology.orchestra.run.vm08.stdout:Removed "/etc/systemd/system/ceph.target.wants/ceph-mon.target". 2026-03-09T20:56:55.664 INFO:teuthology.orchestra.run.vm08.stdout: 2026-03-09T20:56:55.666 INFO:teuthology.orchestra.run.vm08.stdout: Erasing : ceph-mon-2:19.2.3-678.ge911bdeb.el9.x86_64 8/8 2026-03-09T20:56:55.688 INFO:teuthology.orchestra.run.vm06.stdout: Running scriptlet: ceph-mon-2:19.2.3-678.ge911bdeb.el9.x86_64 8/8 2026-03-09T20:56:55.688 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : ceph-2:19.2.3-678.ge911bdeb.el9.x86_64 1/8 2026-03-09T20:56:55.688 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : ceph-mds-2:19.2.3-678.ge911bdeb.el9.x86_64 2/8 2026-03-09T20:56:55.688 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : ceph-mon-2:19.2.3-678.ge911bdeb.el9.x86_64 3/8 2026-03-09T20:56:55.688 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : lua-5.4.4-4.el9.x86_64 4/8 2026-03-09T20:56:55.688 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : lua-devel-5.4.4-4.el9.x86_64 5/8 2026-03-09T20:56:55.688 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : luarocks-3.9.2-5.el9.noarch 6/8 2026-03-09T20:56:55.688 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : unzip-6.0-59.el9.x86_64 7/8 2026-03-09T20:56:55.718 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : zip-3.0-35.el9.x86_64 8/8 2026-03-09T20:56:55.718 INFO:teuthology.orchestra.run.vm01.stdout: 2026-03-09T20:56:55.718 INFO:teuthology.orchestra.run.vm01.stdout:Removed: 2026-03-09T20:56:55.718 INFO:teuthology.orchestra.run.vm01.stdout: ceph-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T20:56:55.718 INFO:teuthology.orchestra.run.vm01.stdout: ceph-mds-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T20:56:55.718 INFO:teuthology.orchestra.run.vm01.stdout: ceph-mon-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T20:56:55.718 INFO:teuthology.orchestra.run.vm01.stdout: lua-5.4.4-4.el9.x86_64 2026-03-09T20:56:55.718 INFO:teuthology.orchestra.run.vm01.stdout: lua-devel-5.4.4-4.el9.x86_64 2026-03-09T20:56:55.718 INFO:teuthology.orchestra.run.vm01.stdout: luarocks-3.9.2-5.el9.noarch 2026-03-09T20:56:55.718 INFO:teuthology.orchestra.run.vm01.stdout: unzip-6.0-59.el9.x86_64 2026-03-09T20:56:55.718 INFO:teuthology.orchestra.run.vm01.stdout: zip-3.0-35.el9.x86_64 2026-03-09T20:56:55.718 INFO:teuthology.orchestra.run.vm01.stdout: 2026-03-09T20:56:55.718 INFO:teuthology.orchestra.run.vm01.stdout:Complete! 2026-03-09T20:56:55.742 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : zip-3.0-35.el9.x86_64 8/8 2026-03-09T20:56:55.742 INFO:teuthology.orchestra.run.vm06.stdout: 2026-03-09T20:56:55.742 INFO:teuthology.orchestra.run.vm06.stdout:Removed: 2026-03-09T20:56:55.743 INFO:teuthology.orchestra.run.vm06.stdout: ceph-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T20:56:55.743 INFO:teuthology.orchestra.run.vm06.stdout: ceph-mds-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T20:56:55.743 INFO:teuthology.orchestra.run.vm06.stdout: ceph-mon-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T20:56:55.743 INFO:teuthology.orchestra.run.vm06.stdout: lua-5.4.4-4.el9.x86_64 2026-03-09T20:56:55.743 INFO:teuthology.orchestra.run.vm06.stdout: lua-devel-5.4.4-4.el9.x86_64 2026-03-09T20:56:55.743 INFO:teuthology.orchestra.run.vm06.stdout: luarocks-3.9.2-5.el9.noarch 2026-03-09T20:56:55.743 INFO:teuthology.orchestra.run.vm06.stdout: unzip-6.0-59.el9.x86_64 2026-03-09T20:56:55.743 INFO:teuthology.orchestra.run.vm06.stdout: zip-3.0-35.el9.x86_64 2026-03-09T20:56:55.743 INFO:teuthology.orchestra.run.vm06.stdout: 2026-03-09T20:56:55.743 INFO:teuthology.orchestra.run.vm06.stdout:Complete! 2026-03-09T20:56:55.761 INFO:teuthology.orchestra.run.vm08.stdout: Running scriptlet: ceph-mon-2:19.2.3-678.ge911bdeb.el9.x86_64 8/8 2026-03-09T20:56:55.761 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : ceph-2:19.2.3-678.ge911bdeb.el9.x86_64 1/8 2026-03-09T20:56:55.761 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : ceph-mds-2:19.2.3-678.ge911bdeb.el9.x86_64 2/8 2026-03-09T20:56:55.761 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : ceph-mon-2:19.2.3-678.ge911bdeb.el9.x86_64 3/8 2026-03-09T20:56:55.761 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : lua-5.4.4-4.el9.x86_64 4/8 2026-03-09T20:56:55.761 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : lua-devel-5.4.4-4.el9.x86_64 5/8 2026-03-09T20:56:55.761 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : luarocks-3.9.2-5.el9.noarch 6/8 2026-03-09T20:56:55.761 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : unzip-6.0-59.el9.x86_64 7/8 2026-03-09T20:56:55.816 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : zip-3.0-35.el9.x86_64 8/8 2026-03-09T20:56:55.816 INFO:teuthology.orchestra.run.vm08.stdout: 2026-03-09T20:56:55.816 INFO:teuthology.orchestra.run.vm08.stdout:Removed: 2026-03-09T20:56:55.816 INFO:teuthology.orchestra.run.vm08.stdout: ceph-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T20:56:55.816 INFO:teuthology.orchestra.run.vm08.stdout: ceph-mds-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T20:56:55.816 INFO:teuthology.orchestra.run.vm08.stdout: ceph-mon-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T20:56:55.816 INFO:teuthology.orchestra.run.vm08.stdout: lua-5.4.4-4.el9.x86_64 2026-03-09T20:56:55.816 INFO:teuthology.orchestra.run.vm08.stdout: lua-devel-5.4.4-4.el9.x86_64 2026-03-09T20:56:55.816 INFO:teuthology.orchestra.run.vm08.stdout: luarocks-3.9.2-5.el9.noarch 2026-03-09T20:56:55.816 INFO:teuthology.orchestra.run.vm08.stdout: unzip-6.0-59.el9.x86_64 2026-03-09T20:56:55.816 INFO:teuthology.orchestra.run.vm08.stdout: zip-3.0-35.el9.x86_64 2026-03-09T20:56:55.816 INFO:teuthology.orchestra.run.vm08.stdout: 2026-03-09T20:56:55.816 INFO:teuthology.orchestra.run.vm08.stdout:Complete! 2026-03-09T20:56:55.950 INFO:teuthology.orchestra.run.vm01.stdout:Dependencies resolved. 2026-03-09T20:56:55.956 INFO:teuthology.orchestra.run.vm01.stdout:=========================================================================================== 2026-03-09T20:56:55.957 INFO:teuthology.orchestra.run.vm01.stdout: Package Arch Version Repository Size 2026-03-09T20:56:55.957 INFO:teuthology.orchestra.run.vm01.stdout:=========================================================================================== 2026-03-09T20:56:55.957 INFO:teuthology.orchestra.run.vm01.stdout:Removing: 2026-03-09T20:56:55.957 INFO:teuthology.orchestra.run.vm01.stdout: ceph-base x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 23 M 2026-03-09T20:56:55.957 INFO:teuthology.orchestra.run.vm01.stdout:Removing dependent packages: 2026-03-09T20:56:55.957 INFO:teuthology.orchestra.run.vm01.stdout: ceph-immutable-object-cache x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 431 k 2026-03-09T20:56:55.957 INFO:teuthology.orchestra.run.vm01.stdout: ceph-mgr x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 3.4 M 2026-03-09T20:56:55.957 INFO:teuthology.orchestra.run.vm01.stdout: ceph-mgr-cephadm noarch 2:19.2.3-678.ge911bdeb.el9 @ceph-noarch 806 k 2026-03-09T20:56:55.957 INFO:teuthology.orchestra.run.vm01.stdout: ceph-mgr-dashboard noarch 2:19.2.3-678.ge911bdeb.el9 @ceph-noarch 88 M 2026-03-09T20:56:55.957 INFO:teuthology.orchestra.run.vm01.stdout: ceph-mgr-diskprediction-local noarch 2:19.2.3-678.ge911bdeb.el9 @ceph-noarch 66 M 2026-03-09T20:56:55.957 INFO:teuthology.orchestra.run.vm01.stdout: ceph-mgr-rook noarch 2:19.2.3-678.ge911bdeb.el9 @ceph-noarch 563 k 2026-03-09T20:56:55.957 INFO:teuthology.orchestra.run.vm01.stdout: ceph-osd x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 59 M 2026-03-09T20:56:55.957 INFO:teuthology.orchestra.run.vm01.stdout: ceph-volume noarch 2:19.2.3-678.ge911bdeb.el9 @ceph-noarch 1.4 M 2026-03-09T20:56:55.957 INFO:teuthology.orchestra.run.vm01.stdout: rbd-mirror x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 13 M 2026-03-09T20:56:55.957 INFO:teuthology.orchestra.run.vm01.stdout:Removing unused dependencies: 2026-03-09T20:56:55.957 INFO:teuthology.orchestra.run.vm01.stdout: abseil-cpp x86_64 20211102.0-4.el9 @epel 1.9 M 2026-03-09T20:56:55.957 INFO:teuthology.orchestra.run.vm01.stdout: ceph-common x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 85 M 2026-03-09T20:56:55.957 INFO:teuthology.orchestra.run.vm01.stdout: ceph-grafana-dashboards noarch 2:19.2.3-678.ge911bdeb.el9 @ceph-noarch 628 k 2026-03-09T20:56:55.957 INFO:teuthology.orchestra.run.vm01.stdout: ceph-mgr-modules-core noarch 2:19.2.3-678.ge911bdeb.el9 @ceph-noarch 1.5 M 2026-03-09T20:56:55.957 INFO:teuthology.orchestra.run.vm01.stdout: ceph-prometheus-alerts noarch 2:19.2.3-678.ge911bdeb.el9 @ceph-noarch 52 k 2026-03-09T20:56:55.957 INFO:teuthology.orchestra.run.vm01.stdout: ceph-selinux x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 138 k 2026-03-09T20:56:55.957 INFO:teuthology.orchestra.run.vm01.stdout: cryptsetup x86_64 2.8.1-3.el9 @baseos 770 k 2026-03-09T20:56:55.957 INFO:teuthology.orchestra.run.vm01.stdout: flexiblas x86_64 3.0.4-9.el9 @appstream 68 k 2026-03-09T20:56:55.957 INFO:teuthology.orchestra.run.vm01.stdout: flexiblas-netlib x86_64 3.0.4-9.el9 @appstream 11 M 2026-03-09T20:56:55.957 INFO:teuthology.orchestra.run.vm01.stdout: flexiblas-openblas-openmp x86_64 3.0.4-9.el9 @appstream 39 k 2026-03-09T20:56:55.957 INFO:teuthology.orchestra.run.vm01.stdout: gperftools-libs x86_64 2.9.1-3.el9 @epel 1.4 M 2026-03-09T20:56:55.957 INFO:teuthology.orchestra.run.vm01.stdout: grpc-data noarch 1.46.7-10.el9 @epel 13 k 2026-03-09T20:56:55.957 INFO:teuthology.orchestra.run.vm01.stdout: ledmon-libs x86_64 1.1.0-3.el9 @baseos 80 k 2026-03-09T20:56:55.957 INFO:teuthology.orchestra.run.vm01.stdout: libcephsqlite x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 425 k 2026-03-09T20:56:55.957 INFO:teuthology.orchestra.run.vm01.stdout: libconfig x86_64 1.7.2-9.el9 @baseos 220 k 2026-03-09T20:56:55.957 INFO:teuthology.orchestra.run.vm01.stdout: libgfortran x86_64 11.5.0-14.el9 @baseos 2.8 M 2026-03-09T20:56:55.957 INFO:teuthology.orchestra.run.vm01.stdout: liboath x86_64 2.6.12-1.el9 @epel 94 k 2026-03-09T20:56:55.957 INFO:teuthology.orchestra.run.vm01.stdout: libquadmath x86_64 11.5.0-14.el9 @baseos 330 k 2026-03-09T20:56:55.957 INFO:teuthology.orchestra.run.vm01.stdout: libradosstriper1 x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 1.6 M 2026-03-09T20:56:55.957 INFO:teuthology.orchestra.run.vm01.stdout: libstoragemgmt x86_64 1.10.1-1.el9 @appstream 685 k 2026-03-09T20:56:55.957 INFO:teuthology.orchestra.run.vm01.stdout: libunwind x86_64 1.6.2-1.el9 @epel 170 k 2026-03-09T20:56:55.957 INFO:teuthology.orchestra.run.vm01.stdout: openblas x86_64 0.3.29-1.el9 @appstream 112 k 2026-03-09T20:56:55.957 INFO:teuthology.orchestra.run.vm01.stdout: openblas-openmp x86_64 0.3.29-1.el9 @appstream 46 M 2026-03-09T20:56:55.958 INFO:teuthology.orchestra.run.vm01.stdout: pciutils x86_64 3.7.0-7.el9 @baseos 216 k 2026-03-09T20:56:55.958 INFO:teuthology.orchestra.run.vm01.stdout: protobuf x86_64 3.14.0-17.el9 @appstream 3.5 M 2026-03-09T20:56:55.958 INFO:teuthology.orchestra.run.vm01.stdout: protobuf-compiler x86_64 3.14.0-17.el9 @crb 2.9 M 2026-03-09T20:56:55.958 INFO:teuthology.orchestra.run.vm01.stdout: python3-asyncssh noarch 2.13.2-5.el9 @epel 3.9 M 2026-03-09T20:56:55.958 INFO:teuthology.orchestra.run.vm01.stdout: python3-autocommand noarch 2.2.2-8.el9 @epel 82 k 2026-03-09T20:56:55.958 INFO:teuthology.orchestra.run.vm01.stdout: python3-babel noarch 2.9.1-2.el9 @appstream 27 M 2026-03-09T20:56:55.958 INFO:teuthology.orchestra.run.vm01.stdout: python3-backports-tarfile noarch 1.2.0-1.el9 @epel 254 k 2026-03-09T20:56:55.958 INFO:teuthology.orchestra.run.vm01.stdout: python3-bcrypt x86_64 3.2.2-1.el9 @epel 87 k 2026-03-09T20:56:55.958 INFO:teuthology.orchestra.run.vm01.stdout: python3-cachetools noarch 4.2.4-1.el9 @epel 93 k 2026-03-09T20:56:55.958 INFO:teuthology.orchestra.run.vm01.stdout: python3-ceph-common x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 702 k 2026-03-09T20:56:55.958 INFO:teuthology.orchestra.run.vm01.stdout: python3-certifi noarch 2023.05.07-4.el9 @epel 6.3 k 2026-03-09T20:56:55.958 INFO:teuthology.orchestra.run.vm01.stdout: python3-cffi x86_64 1.14.5-5.el9 @baseos 1.0 M 2026-03-09T20:56:55.958 INFO:teuthology.orchestra.run.vm01.stdout: python3-chardet noarch 4.0.0-5.el9 @anaconda 1.4 M 2026-03-09T20:56:55.958 INFO:teuthology.orchestra.run.vm01.stdout: python3-cheroot noarch 10.0.1-4.el9 @epel 682 k 2026-03-09T20:56:55.958 INFO:teuthology.orchestra.run.vm01.stdout: python3-cherrypy noarch 18.6.1-2.el9 @epel 1.1 M 2026-03-09T20:56:55.958 INFO:teuthology.orchestra.run.vm01.stdout: python3-cryptography x86_64 36.0.1-5.el9 @baseos 4.5 M 2026-03-09T20:56:55.958 INFO:teuthology.orchestra.run.vm01.stdout: python3-devel x86_64 3.9.25-3.el9 @appstream 765 k 2026-03-09T20:56:55.958 INFO:teuthology.orchestra.run.vm01.stdout: python3-google-auth noarch 1:2.45.0-1.el9 @epel 1.4 M 2026-03-09T20:56:55.958 INFO:teuthology.orchestra.run.vm01.stdout: python3-grpcio x86_64 1.46.7-10.el9 @epel 6.7 M 2026-03-09T20:56:55.958 INFO:teuthology.orchestra.run.vm01.stdout: python3-grpcio-tools x86_64 1.46.7-10.el9 @epel 418 k 2026-03-09T20:56:55.958 INFO:teuthology.orchestra.run.vm01.stdout: python3-idna noarch 2.10-7.el9.1 @anaconda 513 k 2026-03-09T20:56:55.958 INFO:teuthology.orchestra.run.vm01.stdout: python3-jaraco noarch 8.2.1-3.el9 @epel 3.7 k 2026-03-09T20:56:55.958 INFO:teuthology.orchestra.run.vm01.stdout: python3-jaraco-classes noarch 3.2.1-5.el9 @epel 24 k 2026-03-09T20:56:55.958 INFO:teuthology.orchestra.run.vm01.stdout: python3-jaraco-collections noarch 3.0.0-8.el9 @epel 55 k 2026-03-09T20:56:55.958 INFO:teuthology.orchestra.run.vm01.stdout: python3-jaraco-context noarch 6.0.1-3.el9 @epel 31 k 2026-03-09T20:56:55.958 INFO:teuthology.orchestra.run.vm01.stdout: python3-jaraco-functools noarch 3.5.0-2.el9 @epel 33 k 2026-03-09T20:56:55.958 INFO:teuthology.orchestra.run.vm01.stdout: python3-jaraco-text noarch 4.0.0-2.el9 @epel 51 k 2026-03-09T20:56:55.958 INFO:teuthology.orchestra.run.vm01.stdout: python3-jinja2 noarch 2.11.3-8.el9 @appstream 1.1 M 2026-03-09T20:56:55.958 INFO:teuthology.orchestra.run.vm01.stdout: python3-jsonpatch noarch 1.21-16.el9 @koji-override-0 55 k 2026-03-09T20:56:55.958 INFO:teuthology.orchestra.run.vm01.stdout: python3-jsonpointer noarch 2.0-4.el9 @koji-override-0 34 k 2026-03-09T20:56:55.958 INFO:teuthology.orchestra.run.vm01.stdout: python3-kubernetes noarch 1:26.1.0-3.el9 @epel 21 M 2026-03-09T20:56:55.958 INFO:teuthology.orchestra.run.vm01.stdout: python3-libstoragemgmt x86_64 1.10.1-1.el9 @appstream 832 k 2026-03-09T20:56:55.958 INFO:teuthology.orchestra.run.vm01.stdout: python3-logutils noarch 0.3.5-21.el9 @epel 126 k 2026-03-09T20:56:55.958 INFO:teuthology.orchestra.run.vm01.stdout: python3-mako noarch 1.1.4-6.el9 @appstream 534 k 2026-03-09T20:56:55.958 INFO:teuthology.orchestra.run.vm01.stdout: python3-markupsafe x86_64 1.1.1-12.el9 @appstream 60 k 2026-03-09T20:56:55.958 INFO:teuthology.orchestra.run.vm01.stdout: python3-more-itertools noarch 8.12.0-2.el9 @epel 378 k 2026-03-09T20:56:55.958 INFO:teuthology.orchestra.run.vm01.stdout: python3-natsort noarch 7.1.1-5.el9 @epel 215 k 2026-03-09T20:56:55.958 INFO:teuthology.orchestra.run.vm01.stdout: python3-numpy x86_64 1:1.23.5-2.el9 @appstream 30 M 2026-03-09T20:56:55.958 INFO:teuthology.orchestra.run.vm01.stdout: python3-numpy-f2py x86_64 1:1.23.5-2.el9 @appstream 1.7 M 2026-03-09T20:56:55.958 INFO:teuthology.orchestra.run.vm01.stdout: python3-oauthlib noarch 3.1.1-5.el9 @koji-override-0 888 k 2026-03-09T20:56:55.959 INFO:teuthology.orchestra.run.vm01.stdout: python3-packaging noarch 20.9-5.el9 @appstream 248 k 2026-03-09T20:56:55.959 INFO:teuthology.orchestra.run.vm01.stdout: python3-pecan noarch 1.4.2-3.el9 @epel 1.3 M 2026-03-09T20:56:55.959 INFO:teuthology.orchestra.run.vm01.stdout: python3-ply noarch 3.11-14.el9 @baseos 430 k 2026-03-09T20:56:55.959 INFO:teuthology.orchestra.run.vm01.stdout: python3-portend noarch 3.1.0-2.el9 @epel 20 k 2026-03-09T20:56:55.959 INFO:teuthology.orchestra.run.vm01.stdout: python3-prettytable noarch 0.7.2-27.el9 @koji-override-0 166 k 2026-03-09T20:56:55.959 INFO:teuthology.orchestra.run.vm01.stdout: python3-protobuf noarch 3.14.0-17.el9 @appstream 1.4 M 2026-03-09T20:56:55.959 INFO:teuthology.orchestra.run.vm01.stdout: python3-pyOpenSSL noarch 21.0.0-1.el9 @epel 389 k 2026-03-09T20:56:55.959 INFO:teuthology.orchestra.run.vm01.stdout: python3-pyasn1 noarch 0.4.8-7.el9 @appstream 622 k 2026-03-09T20:56:55.959 INFO:teuthology.orchestra.run.vm01.stdout: python3-pyasn1-modules noarch 0.4.8-7.el9 @appstream 1.0 M 2026-03-09T20:56:55.959 INFO:teuthology.orchestra.run.vm01.stdout: python3-pycparser noarch 2.20-6.el9 @baseos 745 k 2026-03-09T20:56:55.959 INFO:teuthology.orchestra.run.vm01.stdout: python3-pysocks noarch 1.7.1-12.el9 @anaconda 88 k 2026-03-09T20:56:55.959 INFO:teuthology.orchestra.run.vm01.stdout: python3-pytz noarch 2021.1-5.el9 @koji-override-0 176 k 2026-03-09T20:56:55.959 INFO:teuthology.orchestra.run.vm01.stdout: python3-repoze-lru noarch 0.7-16.el9 @epel 83 k 2026-03-09T20:56:55.959 INFO:teuthology.orchestra.run.vm01.stdout: python3-requests noarch 2.25.1-10.el9 @baseos 405 k 2026-03-09T20:56:55.959 INFO:teuthology.orchestra.run.vm01.stdout: python3-requests-oauthlib noarch 1.3.0-12.el9 @appstream 119 k 2026-03-09T20:56:55.959 INFO:teuthology.orchestra.run.vm01.stdout: python3-routes noarch 2.5.1-5.el9 @epel 459 k 2026-03-09T20:56:55.959 INFO:teuthology.orchestra.run.vm01.stdout: python3-rsa noarch 4.9-2.el9 @epel 202 k 2026-03-09T20:56:55.959 INFO:teuthology.orchestra.run.vm01.stdout: python3-scipy x86_64 1.9.3-2.el9 @appstream 76 M 2026-03-09T20:56:55.959 INFO:teuthology.orchestra.run.vm01.stdout: python3-tempora noarch 5.0.0-2.el9 @epel 96 k 2026-03-09T20:56:55.959 INFO:teuthology.orchestra.run.vm01.stdout: python3-toml noarch 0.10.2-6.el9 @appstream 99 k 2026-03-09T20:56:55.959 INFO:teuthology.orchestra.run.vm01.stdout: python3-typing-extensions noarch 4.15.0-1.el9 @epel 447 k 2026-03-09T20:56:55.959 INFO:teuthology.orchestra.run.vm01.stdout: python3-urllib3 noarch 1.26.5-7.el9 @baseos 746 k 2026-03-09T20:56:55.959 INFO:teuthology.orchestra.run.vm01.stdout: python3-webob noarch 1.8.8-2.el9 @epel 1.2 M 2026-03-09T20:56:55.959 INFO:teuthology.orchestra.run.vm01.stdout: python3-websocket-client noarch 1.2.3-2.el9 @epel 319 k 2026-03-09T20:56:55.959 INFO:teuthology.orchestra.run.vm01.stdout: python3-werkzeug noarch 2.0.3-3.el9.1 @epel 1.9 M 2026-03-09T20:56:55.959 INFO:teuthology.orchestra.run.vm01.stdout: python3-zc-lockfile noarch 2.0-10.el9 @epel 35 k 2026-03-09T20:56:55.959 INFO:teuthology.orchestra.run.vm01.stdout: qatlib x86_64 25.08.0-2.el9 @appstream 639 k 2026-03-09T20:56:55.959 INFO:teuthology.orchestra.run.vm01.stdout: qatlib-service x86_64 25.08.0-2.el9 @appstream 69 k 2026-03-09T20:56:55.959 INFO:teuthology.orchestra.run.vm01.stdout: qatzip-libs x86_64 1.3.1-1.el9 @appstream 148 k 2026-03-09T20:56:55.959 INFO:teuthology.orchestra.run.vm01.stdout: 2026-03-09T20:56:55.959 INFO:teuthology.orchestra.run.vm01.stdout:Transaction Summary 2026-03-09T20:56:55.959 INFO:teuthology.orchestra.run.vm01.stdout:=========================================================================================== 2026-03-09T20:56:55.959 INFO:teuthology.orchestra.run.vm01.stdout:Remove 102 Packages 2026-03-09T20:56:55.959 INFO:teuthology.orchestra.run.vm01.stdout: 2026-03-09T20:56:55.960 INFO:teuthology.orchestra.run.vm01.stdout:Freed space: 613 M 2026-03-09T20:56:55.960 INFO:teuthology.orchestra.run.vm01.stdout:Running transaction check 2026-03-09T20:56:55.985 INFO:teuthology.orchestra.run.vm06.stdout:Dependencies resolved. 2026-03-09T20:56:55.986 INFO:teuthology.orchestra.run.vm01.stdout:Transaction check succeeded. 2026-03-09T20:56:55.986 INFO:teuthology.orchestra.run.vm01.stdout:Running transaction test 2026-03-09T20:56:55.993 INFO:teuthology.orchestra.run.vm06.stdout:=========================================================================================== 2026-03-09T20:56:55.993 INFO:teuthology.orchestra.run.vm06.stdout: Package Arch Version Repository Size 2026-03-09T20:56:55.993 INFO:teuthology.orchestra.run.vm06.stdout:=========================================================================================== 2026-03-09T20:56:55.993 INFO:teuthology.orchestra.run.vm06.stdout:Removing: 2026-03-09T20:56:55.993 INFO:teuthology.orchestra.run.vm06.stdout: ceph-base x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 23 M 2026-03-09T20:56:55.993 INFO:teuthology.orchestra.run.vm06.stdout:Removing dependent packages: 2026-03-09T20:56:55.993 INFO:teuthology.orchestra.run.vm06.stdout: ceph-immutable-object-cache x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 431 k 2026-03-09T20:56:55.993 INFO:teuthology.orchestra.run.vm06.stdout: ceph-mgr x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 3.4 M 2026-03-09T20:56:55.993 INFO:teuthology.orchestra.run.vm06.stdout: ceph-mgr-cephadm noarch 2:19.2.3-678.ge911bdeb.el9 @ceph-noarch 806 k 2026-03-09T20:56:55.993 INFO:teuthology.orchestra.run.vm06.stdout: ceph-mgr-dashboard noarch 2:19.2.3-678.ge911bdeb.el9 @ceph-noarch 88 M 2026-03-09T20:56:55.993 INFO:teuthology.orchestra.run.vm06.stdout: ceph-mgr-diskprediction-local noarch 2:19.2.3-678.ge911bdeb.el9 @ceph-noarch 66 M 2026-03-09T20:56:55.993 INFO:teuthology.orchestra.run.vm06.stdout: ceph-mgr-rook noarch 2:19.2.3-678.ge911bdeb.el9 @ceph-noarch 563 k 2026-03-09T20:56:55.993 INFO:teuthology.orchestra.run.vm06.stdout: ceph-osd x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 59 M 2026-03-09T20:56:55.993 INFO:teuthology.orchestra.run.vm06.stdout: ceph-volume noarch 2:19.2.3-678.ge911bdeb.el9 @ceph-noarch 1.4 M 2026-03-09T20:56:55.993 INFO:teuthology.orchestra.run.vm06.stdout: rbd-mirror x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 13 M 2026-03-09T20:56:55.993 INFO:teuthology.orchestra.run.vm06.stdout:Removing unused dependencies: 2026-03-09T20:56:55.993 INFO:teuthology.orchestra.run.vm06.stdout: abseil-cpp x86_64 20211102.0-4.el9 @epel 1.9 M 2026-03-09T20:56:55.993 INFO:teuthology.orchestra.run.vm06.stdout: ceph-common x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 85 M 2026-03-09T20:56:55.993 INFO:teuthology.orchestra.run.vm06.stdout: ceph-grafana-dashboards noarch 2:19.2.3-678.ge911bdeb.el9 @ceph-noarch 628 k 2026-03-09T20:56:55.993 INFO:teuthology.orchestra.run.vm06.stdout: ceph-mgr-modules-core noarch 2:19.2.3-678.ge911bdeb.el9 @ceph-noarch 1.5 M 2026-03-09T20:56:55.993 INFO:teuthology.orchestra.run.vm06.stdout: ceph-prometheus-alerts noarch 2:19.2.3-678.ge911bdeb.el9 @ceph-noarch 52 k 2026-03-09T20:56:55.993 INFO:teuthology.orchestra.run.vm06.stdout: ceph-selinux x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 138 k 2026-03-09T20:56:55.993 INFO:teuthology.orchestra.run.vm06.stdout: cryptsetup x86_64 2.8.1-3.el9 @baseos 770 k 2026-03-09T20:56:55.993 INFO:teuthology.orchestra.run.vm06.stdout: flexiblas x86_64 3.0.4-9.el9 @appstream 68 k 2026-03-09T20:56:55.993 INFO:teuthology.orchestra.run.vm06.stdout: flexiblas-netlib x86_64 3.0.4-9.el9 @appstream 11 M 2026-03-09T20:56:55.993 INFO:teuthology.orchestra.run.vm06.stdout: flexiblas-openblas-openmp x86_64 3.0.4-9.el9 @appstream 39 k 2026-03-09T20:56:55.993 INFO:teuthology.orchestra.run.vm06.stdout: gperftools-libs x86_64 2.9.1-3.el9 @epel 1.4 M 2026-03-09T20:56:55.993 INFO:teuthology.orchestra.run.vm06.stdout: grpc-data noarch 1.46.7-10.el9 @epel 13 k 2026-03-09T20:56:55.993 INFO:teuthology.orchestra.run.vm06.stdout: ledmon-libs x86_64 1.1.0-3.el9 @baseos 80 k 2026-03-09T20:56:55.993 INFO:teuthology.orchestra.run.vm06.stdout: libcephsqlite x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 425 k 2026-03-09T20:56:55.993 INFO:teuthology.orchestra.run.vm06.stdout: libconfig x86_64 1.7.2-9.el9 @baseos 220 k 2026-03-09T20:56:55.993 INFO:teuthology.orchestra.run.vm06.stdout: libgfortran x86_64 11.5.0-14.el9 @baseos 2.8 M 2026-03-09T20:56:55.993 INFO:teuthology.orchestra.run.vm06.stdout: liboath x86_64 2.6.12-1.el9 @epel 94 k 2026-03-09T20:56:55.993 INFO:teuthology.orchestra.run.vm06.stdout: libquadmath x86_64 11.5.0-14.el9 @baseos 330 k 2026-03-09T20:56:55.994 INFO:teuthology.orchestra.run.vm06.stdout: libradosstriper1 x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 1.6 M 2026-03-09T20:56:55.994 INFO:teuthology.orchestra.run.vm06.stdout: libstoragemgmt x86_64 1.10.1-1.el9 @appstream 685 k 2026-03-09T20:56:55.994 INFO:teuthology.orchestra.run.vm06.stdout: libunwind x86_64 1.6.2-1.el9 @epel 170 k 2026-03-09T20:56:55.994 INFO:teuthology.orchestra.run.vm06.stdout: openblas x86_64 0.3.29-1.el9 @appstream 112 k 2026-03-09T20:56:55.994 INFO:teuthology.orchestra.run.vm06.stdout: openblas-openmp x86_64 0.3.29-1.el9 @appstream 46 M 2026-03-09T20:56:55.994 INFO:teuthology.orchestra.run.vm06.stdout: pciutils x86_64 3.7.0-7.el9 @baseos 216 k 2026-03-09T20:56:55.994 INFO:teuthology.orchestra.run.vm06.stdout: protobuf x86_64 3.14.0-17.el9 @appstream 3.5 M 2026-03-09T20:56:55.994 INFO:teuthology.orchestra.run.vm06.stdout: protobuf-compiler x86_64 3.14.0-17.el9 @crb 2.9 M 2026-03-09T20:56:55.994 INFO:teuthology.orchestra.run.vm06.stdout: python3-asyncssh noarch 2.13.2-5.el9 @epel 3.9 M 2026-03-09T20:56:55.994 INFO:teuthology.orchestra.run.vm06.stdout: python3-autocommand noarch 2.2.2-8.el9 @epel 82 k 2026-03-09T20:56:55.994 INFO:teuthology.orchestra.run.vm06.stdout: python3-babel noarch 2.9.1-2.el9 @appstream 27 M 2026-03-09T20:56:55.994 INFO:teuthology.orchestra.run.vm06.stdout: python3-backports-tarfile noarch 1.2.0-1.el9 @epel 254 k 2026-03-09T20:56:55.994 INFO:teuthology.orchestra.run.vm06.stdout: python3-bcrypt x86_64 3.2.2-1.el9 @epel 87 k 2026-03-09T20:56:55.994 INFO:teuthology.orchestra.run.vm06.stdout: python3-cachetools noarch 4.2.4-1.el9 @epel 93 k 2026-03-09T20:56:55.994 INFO:teuthology.orchestra.run.vm06.stdout: python3-ceph-common x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 702 k 2026-03-09T20:56:55.994 INFO:teuthology.orchestra.run.vm06.stdout: python3-certifi noarch 2023.05.07-4.el9 @epel 6.3 k 2026-03-09T20:56:55.994 INFO:teuthology.orchestra.run.vm06.stdout: python3-cffi x86_64 1.14.5-5.el9 @baseos 1.0 M 2026-03-09T20:56:55.994 INFO:teuthology.orchestra.run.vm06.stdout: python3-chardet noarch 4.0.0-5.el9 @anaconda 1.4 M 2026-03-09T20:56:55.994 INFO:teuthology.orchestra.run.vm06.stdout: python3-cheroot noarch 10.0.1-4.el9 @epel 682 k 2026-03-09T20:56:55.994 INFO:teuthology.orchestra.run.vm06.stdout: python3-cherrypy noarch 18.6.1-2.el9 @epel 1.1 M 2026-03-09T20:56:55.994 INFO:teuthology.orchestra.run.vm06.stdout: python3-cryptography x86_64 36.0.1-5.el9 @baseos 4.5 M 2026-03-09T20:56:55.994 INFO:teuthology.orchestra.run.vm06.stdout: python3-devel x86_64 3.9.25-3.el9 @appstream 765 k 2026-03-09T20:56:55.994 INFO:teuthology.orchestra.run.vm06.stdout: python3-google-auth noarch 1:2.45.0-1.el9 @epel 1.4 M 2026-03-09T20:56:55.994 INFO:teuthology.orchestra.run.vm06.stdout: python3-grpcio x86_64 1.46.7-10.el9 @epel 6.7 M 2026-03-09T20:56:55.994 INFO:teuthology.orchestra.run.vm06.stdout: python3-grpcio-tools x86_64 1.46.7-10.el9 @epel 418 k 2026-03-09T20:56:55.994 INFO:teuthology.orchestra.run.vm06.stdout: python3-idna noarch 2.10-7.el9.1 @anaconda 513 k 2026-03-09T20:56:55.994 INFO:teuthology.orchestra.run.vm06.stdout: python3-jaraco noarch 8.2.1-3.el9 @epel 3.7 k 2026-03-09T20:56:55.994 INFO:teuthology.orchestra.run.vm06.stdout: python3-jaraco-classes noarch 3.2.1-5.el9 @epel 24 k 2026-03-09T20:56:55.994 INFO:teuthology.orchestra.run.vm06.stdout: python3-jaraco-collections noarch 3.0.0-8.el9 @epel 55 k 2026-03-09T20:56:55.994 INFO:teuthology.orchestra.run.vm06.stdout: python3-jaraco-context noarch 6.0.1-3.el9 @epel 31 k 2026-03-09T20:56:55.994 INFO:teuthology.orchestra.run.vm06.stdout: python3-jaraco-functools noarch 3.5.0-2.el9 @epel 33 k 2026-03-09T20:56:55.994 INFO:teuthology.orchestra.run.vm06.stdout: python3-jaraco-text noarch 4.0.0-2.el9 @epel 51 k 2026-03-09T20:56:55.994 INFO:teuthology.orchestra.run.vm06.stdout: python3-jinja2 noarch 2.11.3-8.el9 @appstream 1.1 M 2026-03-09T20:56:55.994 INFO:teuthology.orchestra.run.vm06.stdout: python3-jsonpatch noarch 1.21-16.el9 @koji-override-0 55 k 2026-03-09T20:56:55.994 INFO:teuthology.orchestra.run.vm06.stdout: python3-jsonpointer noarch 2.0-4.el9 @koji-override-0 34 k 2026-03-09T20:56:55.994 INFO:teuthology.orchestra.run.vm06.stdout: python3-kubernetes noarch 1:26.1.0-3.el9 @epel 21 M 2026-03-09T20:56:55.994 INFO:teuthology.orchestra.run.vm06.stdout: python3-libstoragemgmt x86_64 1.10.1-1.el9 @appstream 832 k 2026-03-09T20:56:55.994 INFO:teuthology.orchestra.run.vm06.stdout: python3-logutils noarch 0.3.5-21.el9 @epel 126 k 2026-03-09T20:56:55.994 INFO:teuthology.orchestra.run.vm06.stdout: python3-mako noarch 1.1.4-6.el9 @appstream 534 k 2026-03-09T20:56:55.994 INFO:teuthology.orchestra.run.vm06.stdout: python3-markupsafe x86_64 1.1.1-12.el9 @appstream 60 k 2026-03-09T20:56:55.994 INFO:teuthology.orchestra.run.vm06.stdout: python3-more-itertools noarch 8.12.0-2.el9 @epel 378 k 2026-03-09T20:56:55.994 INFO:teuthology.orchestra.run.vm06.stdout: python3-natsort noarch 7.1.1-5.el9 @epel 215 k 2026-03-09T20:56:55.994 INFO:teuthology.orchestra.run.vm06.stdout: python3-numpy x86_64 1:1.23.5-2.el9 @appstream 30 M 2026-03-09T20:56:55.994 INFO:teuthology.orchestra.run.vm06.stdout: python3-numpy-f2py x86_64 1:1.23.5-2.el9 @appstream 1.7 M 2026-03-09T20:56:55.994 INFO:teuthology.orchestra.run.vm06.stdout: python3-oauthlib noarch 3.1.1-5.el9 @koji-override-0 888 k 2026-03-09T20:56:55.994 INFO:teuthology.orchestra.run.vm06.stdout: python3-packaging noarch 20.9-5.el9 @appstream 248 k 2026-03-09T20:56:55.994 INFO:teuthology.orchestra.run.vm06.stdout: python3-pecan noarch 1.4.2-3.el9 @epel 1.3 M 2026-03-09T20:56:55.994 INFO:teuthology.orchestra.run.vm06.stdout: python3-ply noarch 3.11-14.el9 @baseos 430 k 2026-03-09T20:56:55.994 INFO:teuthology.orchestra.run.vm06.stdout: python3-portend noarch 3.1.0-2.el9 @epel 20 k 2026-03-09T20:56:55.994 INFO:teuthology.orchestra.run.vm06.stdout: python3-prettytable noarch 0.7.2-27.el9 @koji-override-0 166 k 2026-03-09T20:56:55.994 INFO:teuthology.orchestra.run.vm06.stdout: python3-protobuf noarch 3.14.0-17.el9 @appstream 1.4 M 2026-03-09T20:56:55.994 INFO:teuthology.orchestra.run.vm06.stdout: python3-pyOpenSSL noarch 21.0.0-1.el9 @epel 389 k 2026-03-09T20:56:55.994 INFO:teuthology.orchestra.run.vm06.stdout: python3-pyasn1 noarch 0.4.8-7.el9 @appstream 622 k 2026-03-09T20:56:55.994 INFO:teuthology.orchestra.run.vm06.stdout: python3-pyasn1-modules noarch 0.4.8-7.el9 @appstream 1.0 M 2026-03-09T20:56:55.994 INFO:teuthology.orchestra.run.vm06.stdout: python3-pycparser noarch 2.20-6.el9 @baseos 745 k 2026-03-09T20:56:55.994 INFO:teuthology.orchestra.run.vm06.stdout: python3-pysocks noarch 1.7.1-12.el9 @anaconda 88 k 2026-03-09T20:56:55.994 INFO:teuthology.orchestra.run.vm06.stdout: python3-pytz noarch 2021.1-5.el9 @koji-override-0 176 k 2026-03-09T20:56:55.994 INFO:teuthology.orchestra.run.vm06.stdout: python3-repoze-lru noarch 0.7-16.el9 @epel 83 k 2026-03-09T20:56:55.994 INFO:teuthology.orchestra.run.vm06.stdout: python3-requests noarch 2.25.1-10.el9 @baseos 405 k 2026-03-09T20:56:55.995 INFO:teuthology.orchestra.run.vm06.stdout: python3-requests-oauthlib noarch 1.3.0-12.el9 @appstream 119 k 2026-03-09T20:56:55.995 INFO:teuthology.orchestra.run.vm06.stdout: python3-routes noarch 2.5.1-5.el9 @epel 459 k 2026-03-09T20:56:55.995 INFO:teuthology.orchestra.run.vm06.stdout: python3-rsa noarch 4.9-2.el9 @epel 202 k 2026-03-09T20:56:55.995 INFO:teuthology.orchestra.run.vm06.stdout: python3-scipy x86_64 1.9.3-2.el9 @appstream 76 M 2026-03-09T20:56:55.995 INFO:teuthology.orchestra.run.vm06.stdout: python3-tempora noarch 5.0.0-2.el9 @epel 96 k 2026-03-09T20:56:55.995 INFO:teuthology.orchestra.run.vm06.stdout: python3-toml noarch 0.10.2-6.el9 @appstream 99 k 2026-03-09T20:56:55.995 INFO:teuthology.orchestra.run.vm06.stdout: python3-typing-extensions noarch 4.15.0-1.el9 @epel 447 k 2026-03-09T20:56:55.995 INFO:teuthology.orchestra.run.vm06.stdout: python3-urllib3 noarch 1.26.5-7.el9 @baseos 746 k 2026-03-09T20:56:55.995 INFO:teuthology.orchestra.run.vm06.stdout: python3-webob noarch 1.8.8-2.el9 @epel 1.2 M 2026-03-09T20:56:55.995 INFO:teuthology.orchestra.run.vm06.stdout: python3-websocket-client noarch 1.2.3-2.el9 @epel 319 k 2026-03-09T20:56:55.995 INFO:teuthology.orchestra.run.vm06.stdout: python3-werkzeug noarch 2.0.3-3.el9.1 @epel 1.9 M 2026-03-09T20:56:55.995 INFO:teuthology.orchestra.run.vm06.stdout: python3-zc-lockfile noarch 2.0-10.el9 @epel 35 k 2026-03-09T20:56:55.995 INFO:teuthology.orchestra.run.vm06.stdout: qatlib x86_64 25.08.0-2.el9 @appstream 639 k 2026-03-09T20:56:55.995 INFO:teuthology.orchestra.run.vm06.stdout: qatlib-service x86_64 25.08.0-2.el9 @appstream 69 k 2026-03-09T20:56:55.995 INFO:teuthology.orchestra.run.vm06.stdout: qatzip-libs x86_64 1.3.1-1.el9 @appstream 148 k 2026-03-09T20:56:55.995 INFO:teuthology.orchestra.run.vm06.stdout: 2026-03-09T20:56:55.995 INFO:teuthology.orchestra.run.vm06.stdout:Transaction Summary 2026-03-09T20:56:55.995 INFO:teuthology.orchestra.run.vm06.stdout:=========================================================================================== 2026-03-09T20:56:55.995 INFO:teuthology.orchestra.run.vm06.stdout:Remove 102 Packages 2026-03-09T20:56:55.995 INFO:teuthology.orchestra.run.vm06.stdout: 2026-03-09T20:56:55.995 INFO:teuthology.orchestra.run.vm06.stdout:Freed space: 613 M 2026-03-09T20:56:55.995 INFO:teuthology.orchestra.run.vm06.stdout:Running transaction check 2026-03-09T20:56:56.023 INFO:teuthology.orchestra.run.vm06.stdout:Transaction check succeeded. 2026-03-09T20:56:56.023 INFO:teuthology.orchestra.run.vm06.stdout:Running transaction test 2026-03-09T20:56:56.036 INFO:teuthology.orchestra.run.vm08.stdout:Dependencies resolved. 2026-03-09T20:56:56.042 INFO:teuthology.orchestra.run.vm08.stdout:=========================================================================================== 2026-03-09T20:56:56.042 INFO:teuthology.orchestra.run.vm08.stdout: Package Arch Version Repository Size 2026-03-09T20:56:56.042 INFO:teuthology.orchestra.run.vm08.stdout:=========================================================================================== 2026-03-09T20:56:56.042 INFO:teuthology.orchestra.run.vm08.stdout:Removing: 2026-03-09T20:56:56.042 INFO:teuthology.orchestra.run.vm08.stdout: ceph-base x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 23 M 2026-03-09T20:56:56.042 INFO:teuthology.orchestra.run.vm08.stdout:Removing dependent packages: 2026-03-09T20:56:56.042 INFO:teuthology.orchestra.run.vm08.stdout: ceph-immutable-object-cache x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 431 k 2026-03-09T20:56:56.042 INFO:teuthology.orchestra.run.vm08.stdout: ceph-mgr x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 3.4 M 2026-03-09T20:56:56.042 INFO:teuthology.orchestra.run.vm08.stdout: ceph-mgr-cephadm noarch 2:19.2.3-678.ge911bdeb.el9 @ceph-noarch 806 k 2026-03-09T20:56:56.042 INFO:teuthology.orchestra.run.vm08.stdout: ceph-mgr-dashboard noarch 2:19.2.3-678.ge911bdeb.el9 @ceph-noarch 88 M 2026-03-09T20:56:56.042 INFO:teuthology.orchestra.run.vm08.stdout: ceph-mgr-diskprediction-local noarch 2:19.2.3-678.ge911bdeb.el9 @ceph-noarch 66 M 2026-03-09T20:56:56.042 INFO:teuthology.orchestra.run.vm08.stdout: ceph-mgr-rook noarch 2:19.2.3-678.ge911bdeb.el9 @ceph-noarch 563 k 2026-03-09T20:56:56.042 INFO:teuthology.orchestra.run.vm08.stdout: ceph-osd x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 59 M 2026-03-09T20:56:56.042 INFO:teuthology.orchestra.run.vm08.stdout: ceph-volume noarch 2:19.2.3-678.ge911bdeb.el9 @ceph-noarch 1.4 M 2026-03-09T20:56:56.042 INFO:teuthology.orchestra.run.vm08.stdout: rbd-mirror x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 13 M 2026-03-09T20:56:56.042 INFO:teuthology.orchestra.run.vm08.stdout:Removing unused dependencies: 2026-03-09T20:56:56.042 INFO:teuthology.orchestra.run.vm08.stdout: abseil-cpp x86_64 20211102.0-4.el9 @epel 1.9 M 2026-03-09T20:56:56.042 INFO:teuthology.orchestra.run.vm08.stdout: ceph-common x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 85 M 2026-03-09T20:56:56.042 INFO:teuthology.orchestra.run.vm08.stdout: ceph-grafana-dashboards noarch 2:19.2.3-678.ge911bdeb.el9 @ceph-noarch 628 k 2026-03-09T20:56:56.042 INFO:teuthology.orchestra.run.vm08.stdout: ceph-mgr-modules-core noarch 2:19.2.3-678.ge911bdeb.el9 @ceph-noarch 1.5 M 2026-03-09T20:56:56.042 INFO:teuthology.orchestra.run.vm08.stdout: ceph-prometheus-alerts noarch 2:19.2.3-678.ge911bdeb.el9 @ceph-noarch 52 k 2026-03-09T20:56:56.042 INFO:teuthology.orchestra.run.vm08.stdout: ceph-selinux x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 138 k 2026-03-09T20:56:56.042 INFO:teuthology.orchestra.run.vm08.stdout: cryptsetup x86_64 2.8.1-3.el9 @baseos 770 k 2026-03-09T20:56:56.043 INFO:teuthology.orchestra.run.vm08.stdout: flexiblas x86_64 3.0.4-9.el9 @appstream 68 k 2026-03-09T20:56:56.043 INFO:teuthology.orchestra.run.vm08.stdout: flexiblas-netlib x86_64 3.0.4-9.el9 @appstream 11 M 2026-03-09T20:56:56.043 INFO:teuthology.orchestra.run.vm08.stdout: flexiblas-openblas-openmp x86_64 3.0.4-9.el9 @appstream 39 k 2026-03-09T20:56:56.043 INFO:teuthology.orchestra.run.vm08.stdout: gperftools-libs x86_64 2.9.1-3.el9 @epel 1.4 M 2026-03-09T20:56:56.043 INFO:teuthology.orchestra.run.vm08.stdout: grpc-data noarch 1.46.7-10.el9 @epel 13 k 2026-03-09T20:56:56.043 INFO:teuthology.orchestra.run.vm08.stdout: ledmon-libs x86_64 1.1.0-3.el9 @baseos 80 k 2026-03-09T20:56:56.043 INFO:teuthology.orchestra.run.vm08.stdout: libcephsqlite x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 425 k 2026-03-09T20:56:56.043 INFO:teuthology.orchestra.run.vm08.stdout: libconfig x86_64 1.7.2-9.el9 @baseos 220 k 2026-03-09T20:56:56.043 INFO:teuthology.orchestra.run.vm08.stdout: libgfortran x86_64 11.5.0-14.el9 @baseos 2.8 M 2026-03-09T20:56:56.043 INFO:teuthology.orchestra.run.vm08.stdout: liboath x86_64 2.6.12-1.el9 @epel 94 k 2026-03-09T20:56:56.043 INFO:teuthology.orchestra.run.vm08.stdout: libquadmath x86_64 11.5.0-14.el9 @baseos 330 k 2026-03-09T20:56:56.043 INFO:teuthology.orchestra.run.vm08.stdout: libradosstriper1 x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 1.6 M 2026-03-09T20:56:56.043 INFO:teuthology.orchestra.run.vm08.stdout: libstoragemgmt x86_64 1.10.1-1.el9 @appstream 685 k 2026-03-09T20:56:56.043 INFO:teuthology.orchestra.run.vm08.stdout: libunwind x86_64 1.6.2-1.el9 @epel 170 k 2026-03-09T20:56:56.043 INFO:teuthology.orchestra.run.vm08.stdout: openblas x86_64 0.3.29-1.el9 @appstream 112 k 2026-03-09T20:56:56.043 INFO:teuthology.orchestra.run.vm08.stdout: openblas-openmp x86_64 0.3.29-1.el9 @appstream 46 M 2026-03-09T20:56:56.043 INFO:teuthology.orchestra.run.vm08.stdout: pciutils x86_64 3.7.0-7.el9 @baseos 216 k 2026-03-09T20:56:56.043 INFO:teuthology.orchestra.run.vm08.stdout: protobuf x86_64 3.14.0-17.el9 @appstream 3.5 M 2026-03-09T20:56:56.043 INFO:teuthology.orchestra.run.vm08.stdout: protobuf-compiler x86_64 3.14.0-17.el9 @crb 2.9 M 2026-03-09T20:56:56.043 INFO:teuthology.orchestra.run.vm08.stdout: python3-asyncssh noarch 2.13.2-5.el9 @epel 3.9 M 2026-03-09T20:56:56.043 INFO:teuthology.orchestra.run.vm08.stdout: python3-autocommand noarch 2.2.2-8.el9 @epel 82 k 2026-03-09T20:56:56.043 INFO:teuthology.orchestra.run.vm08.stdout: python3-babel noarch 2.9.1-2.el9 @appstream 27 M 2026-03-09T20:56:56.043 INFO:teuthology.orchestra.run.vm08.stdout: python3-backports-tarfile noarch 1.2.0-1.el9 @epel 254 k 2026-03-09T20:56:56.043 INFO:teuthology.orchestra.run.vm08.stdout: python3-bcrypt x86_64 3.2.2-1.el9 @epel 87 k 2026-03-09T20:56:56.043 INFO:teuthology.orchestra.run.vm08.stdout: python3-cachetools noarch 4.2.4-1.el9 @epel 93 k 2026-03-09T20:56:56.043 INFO:teuthology.orchestra.run.vm08.stdout: python3-ceph-common x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 702 k 2026-03-09T20:56:56.043 INFO:teuthology.orchestra.run.vm08.stdout: python3-certifi noarch 2023.05.07-4.el9 @epel 6.3 k 2026-03-09T20:56:56.043 INFO:teuthology.orchestra.run.vm08.stdout: python3-cffi x86_64 1.14.5-5.el9 @baseos 1.0 M 2026-03-09T20:56:56.043 INFO:teuthology.orchestra.run.vm08.stdout: python3-chardet noarch 4.0.0-5.el9 @anaconda 1.4 M 2026-03-09T20:56:56.043 INFO:teuthology.orchestra.run.vm08.stdout: python3-cheroot noarch 10.0.1-4.el9 @epel 682 k 2026-03-09T20:56:56.043 INFO:teuthology.orchestra.run.vm08.stdout: python3-cherrypy noarch 18.6.1-2.el9 @epel 1.1 M 2026-03-09T20:56:56.043 INFO:teuthology.orchestra.run.vm08.stdout: python3-cryptography x86_64 36.0.1-5.el9 @baseos 4.5 M 2026-03-09T20:56:56.043 INFO:teuthology.orchestra.run.vm08.stdout: python3-devel x86_64 3.9.25-3.el9 @appstream 765 k 2026-03-09T20:56:56.043 INFO:teuthology.orchestra.run.vm08.stdout: python3-google-auth noarch 1:2.45.0-1.el9 @epel 1.4 M 2026-03-09T20:56:56.043 INFO:teuthology.orchestra.run.vm08.stdout: python3-grpcio x86_64 1.46.7-10.el9 @epel 6.7 M 2026-03-09T20:56:56.043 INFO:teuthology.orchestra.run.vm08.stdout: python3-grpcio-tools x86_64 1.46.7-10.el9 @epel 418 k 2026-03-09T20:56:56.043 INFO:teuthology.orchestra.run.vm08.stdout: python3-idna noarch 2.10-7.el9.1 @anaconda 513 k 2026-03-09T20:56:56.043 INFO:teuthology.orchestra.run.vm08.stdout: python3-jaraco noarch 8.2.1-3.el9 @epel 3.7 k 2026-03-09T20:56:56.043 INFO:teuthology.orchestra.run.vm08.stdout: python3-jaraco-classes noarch 3.2.1-5.el9 @epel 24 k 2026-03-09T20:56:56.043 INFO:teuthology.orchestra.run.vm08.stdout: python3-jaraco-collections noarch 3.0.0-8.el9 @epel 55 k 2026-03-09T20:56:56.043 INFO:teuthology.orchestra.run.vm08.stdout: python3-jaraco-context noarch 6.0.1-3.el9 @epel 31 k 2026-03-09T20:56:56.043 INFO:teuthology.orchestra.run.vm08.stdout: python3-jaraco-functools noarch 3.5.0-2.el9 @epel 33 k 2026-03-09T20:56:56.043 INFO:teuthology.orchestra.run.vm08.stdout: python3-jaraco-text noarch 4.0.0-2.el9 @epel 51 k 2026-03-09T20:56:56.043 INFO:teuthology.orchestra.run.vm08.stdout: python3-jinja2 noarch 2.11.3-8.el9 @appstream 1.1 M 2026-03-09T20:56:56.043 INFO:teuthology.orchestra.run.vm08.stdout: python3-jsonpatch noarch 1.21-16.el9 @koji-override-0 55 k 2026-03-09T20:56:56.043 INFO:teuthology.orchestra.run.vm08.stdout: python3-jsonpointer noarch 2.0-4.el9 @koji-override-0 34 k 2026-03-09T20:56:56.043 INFO:teuthology.orchestra.run.vm08.stdout: python3-kubernetes noarch 1:26.1.0-3.el9 @epel 21 M 2026-03-09T20:56:56.043 INFO:teuthology.orchestra.run.vm08.stdout: python3-libstoragemgmt x86_64 1.10.1-1.el9 @appstream 832 k 2026-03-09T20:56:56.043 INFO:teuthology.orchestra.run.vm08.stdout: python3-logutils noarch 0.3.5-21.el9 @epel 126 k 2026-03-09T20:56:56.043 INFO:teuthology.orchestra.run.vm08.stdout: python3-mako noarch 1.1.4-6.el9 @appstream 534 k 2026-03-09T20:56:56.043 INFO:teuthology.orchestra.run.vm08.stdout: python3-markupsafe x86_64 1.1.1-12.el9 @appstream 60 k 2026-03-09T20:56:56.043 INFO:teuthology.orchestra.run.vm08.stdout: python3-more-itertools noarch 8.12.0-2.el9 @epel 378 k 2026-03-09T20:56:56.043 INFO:teuthology.orchestra.run.vm08.stdout: python3-natsort noarch 7.1.1-5.el9 @epel 215 k 2026-03-09T20:56:56.043 INFO:teuthology.orchestra.run.vm08.stdout: python3-numpy x86_64 1:1.23.5-2.el9 @appstream 30 M 2026-03-09T20:56:56.043 INFO:teuthology.orchestra.run.vm08.stdout: python3-numpy-f2py x86_64 1:1.23.5-2.el9 @appstream 1.7 M 2026-03-09T20:56:56.043 INFO:teuthology.orchestra.run.vm08.stdout: python3-oauthlib noarch 3.1.1-5.el9 @koji-override-0 888 k 2026-03-09T20:56:56.043 INFO:teuthology.orchestra.run.vm08.stdout: python3-packaging noarch 20.9-5.el9 @appstream 248 k 2026-03-09T20:56:56.043 INFO:teuthology.orchestra.run.vm08.stdout: python3-pecan noarch 1.4.2-3.el9 @epel 1.3 M 2026-03-09T20:56:56.043 INFO:teuthology.orchestra.run.vm08.stdout: python3-ply noarch 3.11-14.el9 @baseos 430 k 2026-03-09T20:56:56.043 INFO:teuthology.orchestra.run.vm08.stdout: python3-portend noarch 3.1.0-2.el9 @epel 20 k 2026-03-09T20:56:56.043 INFO:teuthology.orchestra.run.vm08.stdout: python3-prettytable noarch 0.7.2-27.el9 @koji-override-0 166 k 2026-03-09T20:56:56.043 INFO:teuthology.orchestra.run.vm08.stdout: python3-protobuf noarch 3.14.0-17.el9 @appstream 1.4 M 2026-03-09T20:56:56.044 INFO:teuthology.orchestra.run.vm08.stdout: python3-pyOpenSSL noarch 21.0.0-1.el9 @epel 389 k 2026-03-09T20:56:56.044 INFO:teuthology.orchestra.run.vm08.stdout: python3-pyasn1 noarch 0.4.8-7.el9 @appstream 622 k 2026-03-09T20:56:56.044 INFO:teuthology.orchestra.run.vm08.stdout: python3-pyasn1-modules noarch 0.4.8-7.el9 @appstream 1.0 M 2026-03-09T20:56:56.044 INFO:teuthology.orchestra.run.vm08.stdout: python3-pycparser noarch 2.20-6.el9 @baseos 745 k 2026-03-09T20:56:56.044 INFO:teuthology.orchestra.run.vm08.stdout: python3-pysocks noarch 1.7.1-12.el9 @anaconda 88 k 2026-03-09T20:56:56.044 INFO:teuthology.orchestra.run.vm08.stdout: python3-pytz noarch 2021.1-5.el9 @koji-override-0 176 k 2026-03-09T20:56:56.044 INFO:teuthology.orchestra.run.vm08.stdout: python3-repoze-lru noarch 0.7-16.el9 @epel 83 k 2026-03-09T20:56:56.044 INFO:teuthology.orchestra.run.vm08.stdout: python3-requests noarch 2.25.1-10.el9 @baseos 405 k 2026-03-09T20:56:56.044 INFO:teuthology.orchestra.run.vm08.stdout: python3-requests-oauthlib noarch 1.3.0-12.el9 @appstream 119 k 2026-03-09T20:56:56.044 INFO:teuthology.orchestra.run.vm08.stdout: python3-routes noarch 2.5.1-5.el9 @epel 459 k 2026-03-09T20:56:56.044 INFO:teuthology.orchestra.run.vm08.stdout: python3-rsa noarch 4.9-2.el9 @epel 202 k 2026-03-09T20:56:56.044 INFO:teuthology.orchestra.run.vm08.stdout: python3-scipy x86_64 1.9.3-2.el9 @appstream 76 M 2026-03-09T20:56:56.044 INFO:teuthology.orchestra.run.vm08.stdout: python3-tempora noarch 5.0.0-2.el9 @epel 96 k 2026-03-09T20:56:56.044 INFO:teuthology.orchestra.run.vm08.stdout: python3-toml noarch 0.10.2-6.el9 @appstream 99 k 2026-03-09T20:56:56.044 INFO:teuthology.orchestra.run.vm08.stdout: python3-typing-extensions noarch 4.15.0-1.el9 @epel 447 k 2026-03-09T20:56:56.044 INFO:teuthology.orchestra.run.vm08.stdout: python3-urllib3 noarch 1.26.5-7.el9 @baseos 746 k 2026-03-09T20:56:56.044 INFO:teuthology.orchestra.run.vm08.stdout: python3-webob noarch 1.8.8-2.el9 @epel 1.2 M 2026-03-09T20:56:56.044 INFO:teuthology.orchestra.run.vm08.stdout: python3-websocket-client noarch 1.2.3-2.el9 @epel 319 k 2026-03-09T20:56:56.044 INFO:teuthology.orchestra.run.vm08.stdout: python3-werkzeug noarch 2.0.3-3.el9.1 @epel 1.9 M 2026-03-09T20:56:56.044 INFO:teuthology.orchestra.run.vm08.stdout: python3-zc-lockfile noarch 2.0-10.el9 @epel 35 k 2026-03-09T20:56:56.044 INFO:teuthology.orchestra.run.vm08.stdout: qatlib x86_64 25.08.0-2.el9 @appstream 639 k 2026-03-09T20:56:56.044 INFO:teuthology.orchestra.run.vm08.stdout: qatlib-service x86_64 25.08.0-2.el9 @appstream 69 k 2026-03-09T20:56:56.044 INFO:teuthology.orchestra.run.vm08.stdout: qatzip-libs x86_64 1.3.1-1.el9 @appstream 148 k 2026-03-09T20:56:56.044 INFO:teuthology.orchestra.run.vm08.stdout: 2026-03-09T20:56:56.044 INFO:teuthology.orchestra.run.vm08.stdout:Transaction Summary 2026-03-09T20:56:56.044 INFO:teuthology.orchestra.run.vm08.stdout:=========================================================================================== 2026-03-09T20:56:56.044 INFO:teuthology.orchestra.run.vm08.stdout:Remove 102 Packages 2026-03-09T20:56:56.044 INFO:teuthology.orchestra.run.vm08.stdout: 2026-03-09T20:56:56.044 INFO:teuthology.orchestra.run.vm08.stdout:Freed space: 613 M 2026-03-09T20:56:56.044 INFO:teuthology.orchestra.run.vm08.stdout:Running transaction check 2026-03-09T20:56:56.072 INFO:teuthology.orchestra.run.vm08.stdout:Transaction check succeeded. 2026-03-09T20:56:56.072 INFO:teuthology.orchestra.run.vm08.stdout:Running transaction test 2026-03-09T20:56:56.095 INFO:teuthology.orchestra.run.vm01.stdout:Transaction test succeeded. 2026-03-09T20:56:56.095 INFO:teuthology.orchestra.run.vm01.stdout:Running transaction 2026-03-09T20:56:56.140 INFO:teuthology.orchestra.run.vm06.stdout:Transaction test succeeded. 2026-03-09T20:56:56.140 INFO:teuthology.orchestra.run.vm06.stdout:Running transaction 2026-03-09T20:56:56.184 INFO:teuthology.orchestra.run.vm08.stdout:Transaction test succeeded. 2026-03-09T20:56:56.184 INFO:teuthology.orchestra.run.vm08.stdout:Running transaction 2026-03-09T20:56:56.247 INFO:teuthology.orchestra.run.vm01.stdout: Preparing : 1/1 2026-03-09T20:56:56.247 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : ceph-mgr-rook-2:19.2.3-678.ge911bdeb.el9.noarch 1/102 2026-03-09T20:56:56.255 INFO:teuthology.orchestra.run.vm01.stdout: Running scriptlet: ceph-mgr-rook-2:19.2.3-678.ge911bdeb.el9.noarch 1/102 2026-03-09T20:56:56.278 INFO:teuthology.orchestra.run.vm01.stdout: Running scriptlet: ceph-mgr-2:19.2.3-678.ge911bdeb.el9.x86_64 2/102 2026-03-09T20:56:56.278 INFO:teuthology.orchestra.run.vm01.stdout:Glob pattern passed to enable, but globs are not supported for this. 2026-03-09T20:56:56.278 INFO:teuthology.orchestra.run.vm01.stdout:Invalid unit name "ceph-mgr@*.service" escaped as "ceph-mgr@\x2a.service". 2026-03-09T20:56:56.278 INFO:teuthology.orchestra.run.vm01.stdout:Removed "/etc/systemd/system/multi-user.target.wants/ceph-mgr.target". 2026-03-09T20:56:56.278 INFO:teuthology.orchestra.run.vm01.stdout:Removed "/etc/systemd/system/ceph.target.wants/ceph-mgr.target". 2026-03-09T20:56:56.278 INFO:teuthology.orchestra.run.vm01.stdout: 2026-03-09T20:56:56.279 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : ceph-mgr-2:19.2.3-678.ge911bdeb.el9.x86_64 2/102 2026-03-09T20:56:56.287 INFO:teuthology.orchestra.run.vm06.stdout: Preparing : 1/1 2026-03-09T20:56:56.287 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : ceph-mgr-rook-2:19.2.3-678.ge911bdeb.el9.noarch 1/102 2026-03-09T20:56:56.293 INFO:teuthology.orchestra.run.vm01.stdout: Running scriptlet: ceph-mgr-2:19.2.3-678.ge911bdeb.el9.x86_64 2/102 2026-03-09T20:56:56.295 INFO:teuthology.orchestra.run.vm06.stdout: Running scriptlet: ceph-mgr-rook-2:19.2.3-678.ge911bdeb.el9.noarch 1/102 2026-03-09T20:56:56.312 INFO:teuthology.orchestra.run.vm06.stdout: Running scriptlet: ceph-mgr-2:19.2.3-678.ge911bdeb.el9.x86_64 2/102 2026-03-09T20:56:56.312 INFO:teuthology.orchestra.run.vm06.stdout:Glob pattern passed to enable, but globs are not supported for this. 2026-03-09T20:56:56.312 INFO:teuthology.orchestra.run.vm06.stdout:Invalid unit name "ceph-mgr@*.service" escaped as "ceph-mgr@\x2a.service". 2026-03-09T20:56:56.312 INFO:teuthology.orchestra.run.vm06.stdout:Removed "/etc/systemd/system/multi-user.target.wants/ceph-mgr.target". 2026-03-09T20:56:56.312 INFO:teuthology.orchestra.run.vm06.stdout:Removed "/etc/systemd/system/ceph.target.wants/ceph-mgr.target". 2026-03-09T20:56:56.312 INFO:teuthology.orchestra.run.vm06.stdout: 2026-03-09T20:56:56.313 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : ceph-mgr-2:19.2.3-678.ge911bdeb.el9.x86_64 2/102 2026-03-09T20:56:56.320 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : ceph-mgr-modules-core-2:19.2.3-678.ge911bdeb.el9 3/102 2026-03-09T20:56:56.320 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : ceph-mgr-dashboard-2:19.2.3-678.ge911bdeb.el9.no 4/102 2026-03-09T20:56:56.326 INFO:teuthology.orchestra.run.vm06.stdout: Running scriptlet: ceph-mgr-2:19.2.3-678.ge911bdeb.el9.x86_64 2/102 2026-03-09T20:56:56.339 INFO:teuthology.orchestra.run.vm08.stdout: Preparing : 1/1 2026-03-09T20:56:56.339 INFO:teuthology.orchestra.run.vm08.stdout: Erasing : ceph-mgr-rook-2:19.2.3-678.ge911bdeb.el9.noarch 1/102 2026-03-09T20:56:56.348 INFO:teuthology.orchestra.run.vm08.stdout: Running scriptlet: ceph-mgr-rook-2:19.2.3-678.ge911bdeb.el9.noarch 1/102 2026-03-09T20:56:56.350 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : ceph-mgr-modules-core-2:19.2.3-678.ge911bdeb.el9 3/102 2026-03-09T20:56:56.350 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : ceph-mgr-dashboard-2:19.2.3-678.ge911bdeb.el9.no 4/102 2026-03-09T20:56:56.369 INFO:teuthology.orchestra.run.vm08.stdout: Running scriptlet: ceph-mgr-2:19.2.3-678.ge911bdeb.el9.x86_64 2/102 2026-03-09T20:56:56.369 INFO:teuthology.orchestra.run.vm08.stdout:Glob pattern passed to enable, but globs are not supported for this. 2026-03-09T20:56:56.369 INFO:teuthology.orchestra.run.vm08.stdout:Invalid unit name "ceph-mgr@*.service" escaped as "ceph-mgr@\x2a.service". 2026-03-09T20:56:56.369 INFO:teuthology.orchestra.run.vm08.stdout:Removed "/etc/systemd/system/multi-user.target.wants/ceph-mgr.target". 2026-03-09T20:56:56.369 INFO:teuthology.orchestra.run.vm08.stdout:Removed "/etc/systemd/system/ceph.target.wants/ceph-mgr.target". 2026-03-09T20:56:56.369 INFO:teuthology.orchestra.run.vm08.stdout: 2026-03-09T20:56:56.370 INFO:teuthology.orchestra.run.vm08.stdout: Erasing : ceph-mgr-2:19.2.3-678.ge911bdeb.el9.x86_64 2/102 2026-03-09T20:56:56.379 INFO:teuthology.orchestra.run.vm01.stdout: Running scriptlet: ceph-mgr-dashboard-2:19.2.3-678.ge911bdeb.el9.no 4/102 2026-03-09T20:56:56.383 INFO:teuthology.orchestra.run.vm08.stdout: Running scriptlet: ceph-mgr-2:19.2.3-678.ge911bdeb.el9.x86_64 2/102 2026-03-09T20:56:56.390 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : python3-kubernetes-1:26.1.0-3.el9.noarch 5/102 2026-03-09T20:56:56.406 INFO:teuthology.orchestra.run.vm08.stdout: Erasing : ceph-mgr-modules-core-2:19.2.3-678.ge911bdeb.el9 3/102 2026-03-09T20:56:56.406 INFO:teuthology.orchestra.run.vm08.stdout: Erasing : ceph-mgr-dashboard-2:19.2.3-678.ge911bdeb.el9.no 4/102 2026-03-09T20:56:56.407 INFO:teuthology.orchestra.run.vm06.stdout: Running scriptlet: ceph-mgr-dashboard-2:19.2.3-678.ge911bdeb.el9.no 4/102 2026-03-09T20:56:56.460 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : python3-requests-oauthlib-1.3.0-12.el9.noarch 6/102 2026-03-09T20:56:56.460 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : ceph-mgr-cephadm-2:19.2.3-678.ge911bdeb.el9.noar 7/102 2026-03-09T20:56:56.461 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : python3-kubernetes-1:26.1.0-3.el9.noarch 5/102 2026-03-09T20:56:56.466 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : python3-requests-oauthlib-1.3.0-12.el9.noarch 6/102 2026-03-09T20:56:56.467 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : ceph-mgr-cephadm-2:19.2.3-678.ge911bdeb.el9.noar 7/102 2026-03-09T20:56:56.472 INFO:teuthology.orchestra.run.vm01.stdout: Running scriptlet: ceph-mgr-cephadm-2:19.2.3-678.ge911bdeb.el9.noar 7/102 2026-03-09T20:56:56.480 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : python3-cherrypy-18.6.1-2.el9.noarch 8/102 2026-03-09T20:56:56.481 INFO:teuthology.orchestra.run.vm06.stdout: Running scriptlet: ceph-mgr-cephadm-2:19.2.3-678.ge911bdeb.el9.noar 7/102 2026-03-09T20:56:56.484 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : python3-cheroot-10.0.1-4.el9.noarch 9/102 2026-03-09T20:56:56.488 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : python3-cherrypy-18.6.1-2.el9.noarch 8/102 2026-03-09T20:56:56.492 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : python3-cheroot-10.0.1-4.el9.noarch 9/102 2026-03-09T20:56:56.493 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : python3-grpcio-tools-1.46.7-10.el9.x86_64 10/102 2026-03-09T20:56:56.496 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : python3-grpcio-1.46.7-10.el9.x86_64 11/102 2026-03-09T20:56:56.501 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : python3-grpcio-tools-1.46.7-10.el9.x86_64 10/102 2026-03-09T20:56:56.505 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : python3-grpcio-1.46.7-10.el9.x86_64 11/102 2026-03-09T20:56:56.507 INFO:teuthology.orchestra.run.vm08.stdout: Running scriptlet: ceph-mgr-dashboard-2:19.2.3-678.ge911bdeb.el9.no 4/102 2026-03-09T20:56:56.515 INFO:teuthology.orchestra.run.vm08.stdout: Erasing : python3-kubernetes-1:26.1.0-3.el9.noarch 5/102 2026-03-09T20:56:56.516 INFO:teuthology.orchestra.run.vm01.stdout: Running scriptlet: ceph-osd-2:19.2.3-678.ge911bdeb.el9.x86_64 12/102 2026-03-09T20:56:56.517 INFO:teuthology.orchestra.run.vm01.stdout:Glob pattern passed to enable, but globs are not supported for this. 2026-03-09T20:56:56.517 INFO:teuthology.orchestra.run.vm01.stdout:Invalid unit name "ceph-osd@*.service" escaped as "ceph-osd@\x2a.service". 2026-03-09T20:56:56.517 INFO:teuthology.orchestra.run.vm01.stdout:Removed "/etc/systemd/system/multi-user.target.wants/ceph-osd.target". 2026-03-09T20:56:56.517 INFO:teuthology.orchestra.run.vm01.stdout:Removed "/etc/systemd/system/ceph.target.wants/ceph-osd.target". 2026-03-09T20:56:56.517 INFO:teuthology.orchestra.run.vm01.stdout: 2026-03-09T20:56:56.520 INFO:teuthology.orchestra.run.vm08.stdout: Erasing : python3-requests-oauthlib-1.3.0-12.el9.noarch 6/102 2026-03-09T20:56:56.520 INFO:teuthology.orchestra.run.vm08.stdout: Erasing : ceph-mgr-cephadm-2:19.2.3-678.ge911bdeb.el9.noar 7/102 2026-03-09T20:56:56.521 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : ceph-osd-2:19.2.3-678.ge911bdeb.el9.x86_64 12/102 2026-03-09T20:56:56.523 INFO:teuthology.orchestra.run.vm06.stdout: Running scriptlet: ceph-osd-2:19.2.3-678.ge911bdeb.el9.x86_64 12/102 2026-03-09T20:56:56.523 INFO:teuthology.orchestra.run.vm06.stdout:Glob pattern passed to enable, but globs are not supported for this. 2026-03-09T20:56:56.523 INFO:teuthology.orchestra.run.vm06.stdout:Invalid unit name "ceph-osd@*.service" escaped as "ceph-osd@\x2a.service". 2026-03-09T20:56:56.523 INFO:teuthology.orchestra.run.vm06.stdout:Removed "/etc/systemd/system/multi-user.target.wants/ceph-osd.target". 2026-03-09T20:56:56.523 INFO:teuthology.orchestra.run.vm06.stdout:Removed "/etc/systemd/system/ceph.target.wants/ceph-osd.target". 2026-03-09T20:56:56.523 INFO:teuthology.orchestra.run.vm06.stdout: 2026-03-09T20:56:56.529 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : ceph-osd-2:19.2.3-678.ge911bdeb.el9.x86_64 12/102 2026-03-09T20:56:56.530 INFO:teuthology.orchestra.run.vm01.stdout: Running scriptlet: ceph-osd-2:19.2.3-678.ge911bdeb.el9.x86_64 12/102 2026-03-09T20:56:56.530 INFO:teuthology.orchestra.run.vm08.stdout: Running scriptlet: ceph-mgr-cephadm-2:19.2.3-678.ge911bdeb.el9.noar 7/102 2026-03-09T20:56:56.537 INFO:teuthology.orchestra.run.vm08.stdout: Erasing : python3-cherrypy-18.6.1-2.el9.noarch 8/102 2026-03-09T20:56:56.539 INFO:teuthology.orchestra.run.vm06.stdout: Running scriptlet: ceph-osd-2:19.2.3-678.ge911bdeb.el9.x86_64 12/102 2026-03-09T20:56:56.542 INFO:teuthology.orchestra.run.vm08.stdout: Erasing : python3-cheroot-10.0.1-4.el9.noarch 9/102 2026-03-09T20:56:56.547 INFO:teuthology.orchestra.run.vm01.stdout: Running scriptlet: ceph-volume-2:19.2.3-678.ge911bdeb.el9.noarch 13/102 2026-03-09T20:56:56.548 INFO:teuthology.orchestra.run.vm01.stdout:Glob pattern passed to enable, but globs are not supported for this. 2026-03-09T20:56:56.548 INFO:teuthology.orchestra.run.vm01.stdout:Invalid unit name "ceph-volume@*.service" escaped as "ceph-volume@\x2a.service". 2026-03-09T20:56:56.548 INFO:teuthology.orchestra.run.vm01.stdout: 2026-03-09T20:56:56.550 INFO:teuthology.orchestra.run.vm08.stdout: Erasing : python3-grpcio-tools-1.46.7-10.el9.x86_64 10/102 2026-03-09T20:56:56.554 INFO:teuthology.orchestra.run.vm08.stdout: Erasing : python3-grpcio-1.46.7-10.el9.x86_64 11/102 2026-03-09T20:56:56.556 INFO:teuthology.orchestra.run.vm06.stdout: Running scriptlet: ceph-volume-2:19.2.3-678.ge911bdeb.el9.noarch 13/102 2026-03-09T20:56:56.556 INFO:teuthology.orchestra.run.vm06.stdout:Glob pattern passed to enable, but globs are not supported for this. 2026-03-09T20:56:56.556 INFO:teuthology.orchestra.run.vm06.stdout:Invalid unit name "ceph-volume@*.service" escaped as "ceph-volume@\x2a.service". 2026-03-09T20:56:56.556 INFO:teuthology.orchestra.run.vm06.stdout: 2026-03-09T20:56:56.556 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : ceph-volume-2:19.2.3-678.ge911bdeb.el9.noarch 13/102 2026-03-09T20:56:56.564 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : ceph-volume-2:19.2.3-678.ge911bdeb.el9.noarch 13/102 2026-03-09T20:56:56.567 INFO:teuthology.orchestra.run.vm01.stdout: Running scriptlet: ceph-volume-2:19.2.3-678.ge911bdeb.el9.noarch 13/102 2026-03-09T20:56:56.570 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : python3-jaraco-collections-3.0.0-8.el9.noarch 14/102 2026-03-09T20:56:56.575 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : python3-jaraco-text-4.0.0-2.el9.noarch 15/102 2026-03-09T20:56:56.576 INFO:teuthology.orchestra.run.vm06.stdout: Running scriptlet: ceph-volume-2:19.2.3-678.ge911bdeb.el9.noarch 13/102 2026-03-09T20:56:56.577 INFO:teuthology.orchestra.run.vm08.stdout: Running scriptlet: ceph-osd-2:19.2.3-678.ge911bdeb.el9.x86_64 12/102 2026-03-09T20:56:56.577 INFO:teuthology.orchestra.run.vm08.stdout:Glob pattern passed to enable, but globs are not supported for this. 2026-03-09T20:56:56.577 INFO:teuthology.orchestra.run.vm08.stdout:Invalid unit name "ceph-osd@*.service" escaped as "ceph-osd@\x2a.service". 2026-03-09T20:56:56.577 INFO:teuthology.orchestra.run.vm08.stdout:Removed "/etc/systemd/system/multi-user.target.wants/ceph-osd.target". 2026-03-09T20:56:56.577 INFO:teuthology.orchestra.run.vm08.stdout:Removed "/etc/systemd/system/ceph.target.wants/ceph-osd.target". 2026-03-09T20:56:56.577 INFO:teuthology.orchestra.run.vm08.stdout: 2026-03-09T20:56:56.579 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : python3-jaraco-collections-3.0.0-8.el9.noarch 14/102 2026-03-09T20:56:56.580 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : python3-jinja2-2.11.3-8.el9.noarch 16/102 2026-03-09T20:56:56.584 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : python3-jaraco-text-4.0.0-2.el9.noarch 15/102 2026-03-09T20:56:56.584 INFO:teuthology.orchestra.run.vm08.stdout: Erasing : ceph-osd-2:19.2.3-678.ge911bdeb.el9.x86_64 12/102 2026-03-09T20:56:56.589 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : python3-requests-2.25.1-10.el9.noarch 17/102 2026-03-09T20:56:56.660 INFO:teuthology.orchestra.run.vm08.stdout: Running scriptlet: ceph-osd-2:19.2.3-678.ge911bdeb.el9.x86_64 12/102 2026-03-09T20:56:56.674 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : python3-google-auth-1:2.45.0-1.el9.noarch 18/102 2026-03-09T20:56:56.674 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : python3-jinja2-2.11.3-8.el9.noarch 16/102 2026-03-09T20:56:56.677 INFO:teuthology.orchestra.run.vm08.stdout: Running scriptlet: ceph-volume-2:19.2.3-678.ge911bdeb.el9.noarch 13/102 2026-03-09T20:56:56.677 INFO:teuthology.orchestra.run.vm08.stdout:Glob pattern passed to enable, but globs are not supported for this. 2026-03-09T20:56:56.677 INFO:teuthology.orchestra.run.vm08.stdout:Invalid unit name "ceph-volume@*.service" escaped as "ceph-volume@\x2a.service". 2026-03-09T20:56:56.677 INFO:teuthology.orchestra.run.vm08.stdout: 2026-03-09T20:56:56.681 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : python3-pecan-1.4.2-3.el9.noarch 19/102 2026-03-09T20:56:56.685 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : python3-requests-2.25.1-10.el9.noarch 17/102 2026-03-09T20:56:56.685 INFO:teuthology.orchestra.run.vm08.stdout: Erasing : ceph-volume-2:19.2.3-678.ge911bdeb.el9.noarch 13/102 2026-03-09T20:56:56.691 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : python3-rsa-4.9-2.el9.noarch 20/102 2026-03-09T20:56:56.696 INFO:teuthology.orchestra.run.vm08.stdout: Running scriptlet: ceph-volume-2:19.2.3-678.ge911bdeb.el9.noarch 13/102 2026-03-09T20:56:56.697 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : python3-google-auth-1:2.45.0-1.el9.noarch 18/102 2026-03-09T20:56:56.697 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : python3-pyasn1-modules-0.4.8-7.el9.noarch 21/102 2026-03-09T20:56:56.699 INFO:teuthology.orchestra.run.vm08.stdout: Erasing : python3-jaraco-collections-3.0.0-8.el9.noarch 14/102 2026-03-09T20:56:56.703 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : python3-pecan-1.4.2-3.el9.noarch 19/102 2026-03-09T20:56:56.705 INFO:teuthology.orchestra.run.vm08.stdout: Erasing : python3-jaraco-text-4.0.0-2.el9.noarch 15/102 2026-03-09T20:56:56.710 INFO:teuthology.orchestra.run.vm08.stdout: Erasing : python3-jinja2-2.11.3-8.el9.noarch 16/102 2026-03-09T20:56:56.715 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : python3-rsa-4.9-2.el9.noarch 20/102 2026-03-09T20:56:56.720 INFO:teuthology.orchestra.run.vm08.stdout: Erasing : python3-requests-2.25.1-10.el9.noarch 17/102 2026-03-09T20:56:56.721 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : python3-pyasn1-modules-0.4.8-7.el9.noarch 21/102 2026-03-09T20:56:56.729 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : python3-urllib3-1.26.5-7.el9.noarch 22/102 2026-03-09T20:56:56.734 INFO:teuthology.orchestra.run.vm08.stdout: Erasing : python3-google-auth-1:2.45.0-1.el9.noarch 18/102 2026-03-09T20:56:56.737 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : python3-babel-2.9.1-2.el9.noarch 23/102 2026-03-09T20:56:56.740 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : python3-jaraco-classes-3.2.1-5.el9.noarch 24/102 2026-03-09T20:56:56.741 INFO:teuthology.orchestra.run.vm08.stdout: Erasing : python3-pecan-1.4.2-3.el9.noarch 19/102 2026-03-09T20:56:56.749 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : python3-pyOpenSSL-21.0.0-1.el9.noarch 25/102 2026-03-09T20:56:56.752 INFO:teuthology.orchestra.run.vm08.stdout: Erasing : python3-rsa-4.9-2.el9.noarch 20/102 2026-03-09T20:56:56.752 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : python3-urllib3-1.26.5-7.el9.noarch 22/102 2026-03-09T20:56:56.759 INFO:teuthology.orchestra.run.vm08.stdout: Erasing : python3-pyasn1-modules-0.4.8-7.el9.noarch 21/102 2026-03-09T20:56:56.760 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : python3-asyncssh-2.13.2-5.el9.noarch 26/102 2026-03-09T20:56:56.760 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : ceph-mgr-diskprediction-local-2:19.2.3-678.ge911 27/102 2026-03-09T20:56:56.761 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : python3-babel-2.9.1-2.el9.noarch 23/102 2026-03-09T20:56:56.763 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : python3-jaraco-classes-3.2.1-5.el9.noarch 24/102 2026-03-09T20:56:56.768 INFO:teuthology.orchestra.run.vm01.stdout: Running scriptlet: ceph-mgr-diskprediction-local-2:19.2.3-678.ge911 27/102 2026-03-09T20:56:56.772 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : python3-pyOpenSSL-21.0.0-1.el9.noarch 25/102 2026-03-09T20:56:56.784 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : python3-asyncssh-2.13.2-5.el9.noarch 26/102 2026-03-09T20:56:56.784 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : ceph-mgr-diskprediction-local-2:19.2.3-678.ge911 27/102 2026-03-09T20:56:56.792 INFO:teuthology.orchestra.run.vm06.stdout: Running scriptlet: ceph-mgr-diskprediction-local-2:19.2.3-678.ge911 27/102 2026-03-09T20:56:56.794 INFO:teuthology.orchestra.run.vm08.stdout: Erasing : python3-urllib3-1.26.5-7.el9.noarch 22/102 2026-03-09T20:56:56.801 INFO:teuthology.orchestra.run.vm08.stdout: Erasing : python3-babel-2.9.1-2.el9.noarch 23/102 2026-03-09T20:56:56.805 INFO:teuthology.orchestra.run.vm08.stdout: Erasing : python3-jaraco-classes-3.2.1-5.el9.noarch 24/102 2026-03-09T20:56:56.815 INFO:teuthology.orchestra.run.vm08.stdout: Erasing : python3-pyOpenSSL-21.0.0-1.el9.noarch 25/102 2026-03-09T20:56:56.829 INFO:teuthology.orchestra.run.vm08.stdout: Erasing : python3-asyncssh-2.13.2-5.el9.noarch 26/102 2026-03-09T20:56:56.830 INFO:teuthology.orchestra.run.vm08.stdout: Erasing : ceph-mgr-diskprediction-local-2:19.2.3-678.ge911 27/102 2026-03-09T20:56:56.837 INFO:teuthology.orchestra.run.vm08.stdout: Running scriptlet: ceph-mgr-diskprediction-local-2:19.2.3-678.ge911 27/102 2026-03-09T20:56:56.861 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : python3-jsonpatch-1.21-16.el9.noarch 28/102 2026-03-09T20:56:56.878 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : python3-scipy-1.9.3-2.el9.x86_64 29/102 2026-03-09T20:56:56.891 INFO:teuthology.orchestra.run.vm01.stdout: Running scriptlet: libstoragemgmt-1.10.1-1.el9.x86_64 30/102 2026-03-09T20:56:56.891 INFO:teuthology.orchestra.run.vm01.stdout:Removed "/etc/systemd/system/multi-user.target.wants/libstoragemgmt.service". 2026-03-09T20:56:56.891 INFO:teuthology.orchestra.run.vm01.stdout: 2026-03-09T20:56:56.892 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : libstoragemgmt-1.10.1-1.el9.x86_64 30/102 2026-03-09T20:56:56.894 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : python3-jsonpatch-1.21-16.el9.noarch 28/102 2026-03-09T20:56:56.910 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : python3-scipy-1.9.3-2.el9.x86_64 29/102 2026-03-09T20:56:56.920 INFO:teuthology.orchestra.run.vm01.stdout: Running scriptlet: libstoragemgmt-1.10.1-1.el9.x86_64 30/102 2026-03-09T20:56:56.924 INFO:teuthology.orchestra.run.vm06.stdout: Running scriptlet: libstoragemgmt-1.10.1-1.el9.x86_64 30/102 2026-03-09T20:56:56.924 INFO:teuthology.orchestra.run.vm06.stdout:Removed "/etc/systemd/system/multi-user.target.wants/libstoragemgmt.service". 2026-03-09T20:56:56.924 INFO:teuthology.orchestra.run.vm06.stdout: 2026-03-09T20:56:56.925 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : libstoragemgmt-1.10.1-1.el9.x86_64 30/102 2026-03-09T20:56:56.930 INFO:teuthology.orchestra.run.vm08.stdout: Erasing : python3-jsonpatch-1.21-16.el9.noarch 28/102 2026-03-09T20:56:56.936 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : python3-numpy-f2py-1:1.23.5-2.el9.x86_64 31/102 2026-03-09T20:56:56.944 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : python3-cryptography-36.0.1-5.el9.x86_64 32/102 2026-03-09T20:56:56.947 INFO:teuthology.orchestra.run.vm08.stdout: Erasing : python3-scipy-1.9.3-2.el9.x86_64 29/102 2026-03-09T20:56:56.947 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : protobuf-compiler-3.14.0-17.el9.x86_64 33/102 2026-03-09T20:56:56.950 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : python3-bcrypt-3.2.2-1.el9.x86_64 34/102 2026-03-09T20:56:56.955 INFO:teuthology.orchestra.run.vm06.stdout: Running scriptlet: libstoragemgmt-1.10.1-1.el9.x86_64 30/102 2026-03-09T20:56:56.963 INFO:teuthology.orchestra.run.vm08.stdout: Running scriptlet: libstoragemgmt-1.10.1-1.el9.x86_64 30/102 2026-03-09T20:56:56.963 INFO:teuthology.orchestra.run.vm08.stdout:Removed "/etc/systemd/system/multi-user.target.wants/libstoragemgmt.service". 2026-03-09T20:56:56.963 INFO:teuthology.orchestra.run.vm08.stdout: 2026-03-09T20:56:56.964 INFO:teuthology.orchestra.run.vm08.stdout: Erasing : libstoragemgmt-1.10.1-1.el9.x86_64 30/102 2026-03-09T20:56:56.972 INFO:teuthology.orchestra.run.vm01.stdout: Running scriptlet: rbd-mirror-2:19.2.3-678.ge911bdeb.el9.x86_64 35/102 2026-03-09T20:56:56.973 INFO:teuthology.orchestra.run.vm01.stdout:Glob pattern passed to enable, but globs are not supported for this. 2026-03-09T20:56:56.973 INFO:teuthology.orchestra.run.vm01.stdout:Invalid unit name "ceph-rbd-mirror@*.service" escaped as "ceph-rbd-mirror@\x2a.service". 2026-03-09T20:56:56.973 INFO:teuthology.orchestra.run.vm01.stdout:Removed "/etc/systemd/system/multi-user.target.wants/ceph-rbd-mirror.target". 2026-03-09T20:56:56.973 INFO:teuthology.orchestra.run.vm01.stdout:Removed "/etc/systemd/system/ceph.target.wants/ceph-rbd-mirror.target". 2026-03-09T20:56:56.973 INFO:teuthology.orchestra.run.vm01.stdout: 2026-03-09T20:56:56.974 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : rbd-mirror-2:19.2.3-678.ge911bdeb.el9.x86_64 35/102 2026-03-09T20:56:56.976 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : python3-numpy-f2py-1:1.23.5-2.el9.x86_64 31/102 2026-03-09T20:56:56.982 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : python3-cryptography-36.0.1-5.el9.x86_64 32/102 2026-03-09T20:56:56.985 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : protobuf-compiler-3.14.0-17.el9.x86_64 33/102 2026-03-09T20:56:56.987 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : python3-bcrypt-3.2.2-1.el9.x86_64 34/102 2026-03-09T20:56:56.988 INFO:teuthology.orchestra.run.vm01.stdout: Running scriptlet: rbd-mirror-2:19.2.3-678.ge911bdeb.el9.x86_64 35/102 2026-03-09T20:56:56.991 INFO:teuthology.orchestra.run.vm08.stdout: Running scriptlet: libstoragemgmt-1.10.1-1.el9.x86_64 30/102 2026-03-09T20:56:56.992 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : python3-mako-1.1.4-6.el9.noarch 36/102 2026-03-09T20:56:56.995 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : python3-jaraco-context-6.0.1-3.el9.noarch 37/102 2026-03-09T20:56:56.998 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : python3-portend-3.1.0-2.el9.noarch 38/102 2026-03-09T20:56:57.001 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : python3-tempora-5.0.0-2.el9.noarch 39/102 2026-03-09T20:56:57.004 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : python3-jaraco-functools-3.5.0-2.el9.noarch 40/102 2026-03-09T20:56:57.006 INFO:teuthology.orchestra.run.vm08.stdout: Erasing : python3-numpy-f2py-1:1.23.5-2.el9.x86_64 31/102 2026-03-09T20:56:57.007 INFO:teuthology.orchestra.run.vm06.stdout: Running scriptlet: rbd-mirror-2:19.2.3-678.ge911bdeb.el9.x86_64 35/102 2026-03-09T20:56:57.007 INFO:teuthology.orchestra.run.vm06.stdout:Glob pattern passed to enable, but globs are not supported for this. 2026-03-09T20:56:57.007 INFO:teuthology.orchestra.run.vm06.stdout:Invalid unit name "ceph-rbd-mirror@*.service" escaped as "ceph-rbd-mirror@\x2a.service". 2026-03-09T20:56:57.007 INFO:teuthology.orchestra.run.vm06.stdout:Removed "/etc/systemd/system/multi-user.target.wants/ceph-rbd-mirror.target". 2026-03-09T20:56:57.007 INFO:teuthology.orchestra.run.vm06.stdout:Removed "/etc/systemd/system/ceph.target.wants/ceph-rbd-mirror.target". 2026-03-09T20:56:57.007 INFO:teuthology.orchestra.run.vm06.stdout: 2026-03-09T20:56:57.008 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : python3-routes-2.5.1-5.el9.noarch 41/102 2026-03-09T20:56:57.009 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : rbd-mirror-2:19.2.3-678.ge911bdeb.el9.x86_64 35/102 2026-03-09T20:56:57.011 INFO:teuthology.orchestra.run.vm08.stdout: Erasing : python3-cryptography-36.0.1-5.el9.x86_64 32/102 2026-03-09T20:56:57.013 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : python3-cffi-1.14.5-5.el9.x86_64 42/102 2026-03-09T20:56:57.014 INFO:teuthology.orchestra.run.vm08.stdout: Erasing : protobuf-compiler-3.14.0-17.el9.x86_64 33/102 2026-03-09T20:56:57.016 INFO:teuthology.orchestra.run.vm08.stdout: Erasing : python3-bcrypt-3.2.2-1.el9.x86_64 34/102 2026-03-09T20:56:57.022 INFO:teuthology.orchestra.run.vm06.stdout: Running scriptlet: rbd-mirror-2:19.2.3-678.ge911bdeb.el9.x86_64 35/102 2026-03-09T20:56:57.026 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : python3-mako-1.1.4-6.el9.noarch 36/102 2026-03-09T20:56:57.029 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : python3-jaraco-context-6.0.1-3.el9.noarch 37/102 2026-03-09T20:56:57.032 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : python3-portend-3.1.0-2.el9.noarch 38/102 2026-03-09T20:56:57.035 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : python3-tempora-5.0.0-2.el9.noarch 39/102 2026-03-09T20:56:57.036 INFO:teuthology.orchestra.run.vm08.stdout: Running scriptlet: rbd-mirror-2:19.2.3-678.ge911bdeb.el9.x86_64 35/102 2026-03-09T20:56:57.036 INFO:teuthology.orchestra.run.vm08.stdout:Glob pattern passed to enable, but globs are not supported for this. 2026-03-09T20:56:57.036 INFO:teuthology.orchestra.run.vm08.stdout:Invalid unit name "ceph-rbd-mirror@*.service" escaped as "ceph-rbd-mirror@\x2a.service". 2026-03-09T20:56:57.036 INFO:teuthology.orchestra.run.vm08.stdout:Removed "/etc/systemd/system/multi-user.target.wants/ceph-rbd-mirror.target". 2026-03-09T20:56:57.036 INFO:teuthology.orchestra.run.vm08.stdout:Removed "/etc/systemd/system/ceph.target.wants/ceph-rbd-mirror.target". 2026-03-09T20:56:57.036 INFO:teuthology.orchestra.run.vm08.stdout: 2026-03-09T20:56:57.037 INFO:teuthology.orchestra.run.vm08.stdout: Erasing : rbd-mirror-2:19.2.3-678.ge911bdeb.el9.x86_64 35/102 2026-03-09T20:56:57.038 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : python3-jaraco-functools-3.5.0-2.el9.noarch 40/102 2026-03-09T20:56:57.042 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : python3-routes-2.5.1-5.el9.noarch 41/102 2026-03-09T20:56:57.047 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : python3-cffi-1.14.5-5.el9.x86_64 42/102 2026-03-09T20:56:57.049 INFO:teuthology.orchestra.run.vm08.stdout: Running scriptlet: rbd-mirror-2:19.2.3-678.ge911bdeb.el9.x86_64 35/102 2026-03-09T20:56:57.054 INFO:teuthology.orchestra.run.vm08.stdout: Erasing : python3-mako-1.1.4-6.el9.noarch 36/102 2026-03-09T20:56:57.057 INFO:teuthology.orchestra.run.vm08.stdout: Erasing : python3-jaraco-context-6.0.1-3.el9.noarch 37/102 2026-03-09T20:56:57.059 INFO:teuthology.orchestra.run.vm08.stdout: Erasing : python3-portend-3.1.0-2.el9.noarch 38/102 2026-03-09T20:56:57.061 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : python3-pycparser-2.20-6.el9.noarch 43/102 2026-03-09T20:56:57.062 INFO:teuthology.orchestra.run.vm08.stdout: Erasing : python3-tempora-5.0.0-2.el9.noarch 39/102 2026-03-09T20:56:57.065 INFO:teuthology.orchestra.run.vm08.stdout: Erasing : python3-jaraco-functools-3.5.0-2.el9.noarch 40/102 2026-03-09T20:56:57.070 INFO:teuthology.orchestra.run.vm08.stdout: Erasing : python3-routes-2.5.1-5.el9.noarch 41/102 2026-03-09T20:56:57.073 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : python3-numpy-1:1.23.5-2.el9.x86_64 44/102 2026-03-09T20:56:57.075 INFO:teuthology.orchestra.run.vm08.stdout: Erasing : python3-cffi-1.14.5-5.el9.x86_64 42/102 2026-03-09T20:56:57.075 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : flexiblas-netlib-3.0.4-9.el9.x86_64 45/102 2026-03-09T20:56:57.081 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : flexiblas-openblas-openmp-3.0.4-9.el9.x86_64 46/102 2026-03-09T20:56:57.083 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : openblas-openmp-0.3.29-1.el9.x86_64 47/102 2026-03-09T20:56:57.088 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : libgfortran-11.5.0-14.el9.x86_64 48/102 2026-03-09T20:56:57.091 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : python3-libstoragemgmt-1.10.1-1.el9.x86_64 49/102 2026-03-09T20:56:57.096 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : python3-pycparser-2.20-6.el9.noarch 43/102 2026-03-09T20:56:57.108 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : python3-numpy-1:1.23.5-2.el9.x86_64 44/102 2026-03-09T20:56:57.111 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : flexiblas-netlib-3.0.4-9.el9.x86_64 45/102 2026-03-09T20:56:57.114 INFO:teuthology.orchestra.run.vm01.stdout: Running scriptlet: ceph-immutable-object-cache-2:19.2.3-678.ge911bd 50/102 2026-03-09T20:56:57.114 INFO:teuthology.orchestra.run.vm01.stdout:Glob pattern passed to enable, but globs are not supported for this. 2026-03-09T20:56:57.114 INFO:teuthology.orchestra.run.vm01.stdout:Invalid unit name "ceph-immutable-object-cache@*.service" escaped as "ceph-immutable-object-cache@\x2a.service". 2026-03-09T20:56:57.114 INFO:teuthology.orchestra.run.vm01.stdout: 2026-03-09T20:56:57.114 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : ceph-immutable-object-cache-2:19.2.3-678.ge911bd 50/102 2026-03-09T20:56:57.117 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : flexiblas-openblas-openmp-3.0.4-9.el9.x86_64 46/102 2026-03-09T20:56:57.119 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : openblas-openmp-0.3.29-1.el9.x86_64 47/102 2026-03-09T20:56:57.123 INFO:teuthology.orchestra.run.vm01.stdout: Running scriptlet: ceph-immutable-object-cache-2:19.2.3-678.ge911bd 50/102 2026-03-09T20:56:57.123 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : libgfortran-11.5.0-14.el9.x86_64 48/102 2026-03-09T20:56:57.125 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : openblas-0.3.29-1.el9.x86_64 51/102 2026-03-09T20:56:57.127 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : python3-libstoragemgmt-1.10.1-1.el9.x86_64 49/102 2026-03-09T20:56:57.127 INFO:teuthology.orchestra.run.vm08.stdout: Erasing : python3-pycparser-2.20-6.el9.noarch 43/102 2026-03-09T20:56:57.127 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : flexiblas-3.0.4-9.el9.x86_64 52/102 2026-03-09T20:56:57.131 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : python3-ply-3.11-14.el9.noarch 53/102 2026-03-09T20:56:57.133 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : python3-repoze-lru-0.7-16.el9.noarch 54/102 2026-03-09T20:56:57.137 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : python3-jaraco-8.2.1-3.el9.noarch 55/102 2026-03-09T20:56:57.140 INFO:teuthology.orchestra.run.vm08.stdout: Erasing : python3-numpy-1:1.23.5-2.el9.x86_64 44/102 2026-03-09T20:56:57.140 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : python3-more-itertools-8.12.0-2.el9.noarch 56/102 2026-03-09T20:56:57.142 INFO:teuthology.orchestra.run.vm08.stdout: Erasing : flexiblas-netlib-3.0.4-9.el9.x86_64 45/102 2026-03-09T20:56:57.142 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : python3-toml-0.10.2-6.el9.noarch 57/102 2026-03-09T20:56:57.146 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : python3-pytz-2021.1-5.el9.noarch 58/102 2026-03-09T20:56:57.147 INFO:teuthology.orchestra.run.vm08.stdout: Erasing : flexiblas-openblas-openmp-3.0.4-9.el9.x86_64 46/102 2026-03-09T20:56:57.148 INFO:teuthology.orchestra.run.vm06.stdout: Running scriptlet: ceph-immutable-object-cache-2:19.2.3-678.ge911bd 50/102 2026-03-09T20:56:57.148 INFO:teuthology.orchestra.run.vm06.stdout:Glob pattern passed to enable, but globs are not supported for this. 2026-03-09T20:56:57.148 INFO:teuthology.orchestra.run.vm06.stdout:Invalid unit name "ceph-immutable-object-cache@*.service" escaped as "ceph-immutable-object-cache@\x2a.service". 2026-03-09T20:56:57.148 INFO:teuthology.orchestra.run.vm06.stdout: 2026-03-09T20:56:57.149 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : ceph-immutable-object-cache-2:19.2.3-678.ge911bd 50/102 2026-03-09T20:56:57.149 INFO:teuthology.orchestra.run.vm08.stdout: Erasing : openblas-openmp-0.3.29-1.el9.x86_64 47/102 2026-03-09T20:56:57.153 INFO:teuthology.orchestra.run.vm08.stdout: Erasing : libgfortran-11.5.0-14.el9.x86_64 48/102 2026-03-09T20:56:57.153 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : python3-backports-tarfile-1.2.0-1.el9.noarch 59/102 2026-03-09T20:56:57.155 INFO:teuthology.orchestra.run.vm08.stdout: Erasing : python3-libstoragemgmt-1.10.1-1.el9.x86_64 49/102 2026-03-09T20:56:57.157 INFO:teuthology.orchestra.run.vm06.stdout: Running scriptlet: ceph-immutable-object-cache-2:19.2.3-678.ge911bd 50/102 2026-03-09T20:56:57.158 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : python3-devel-3.9.25-3.el9.x86_64 60/102 2026-03-09T20:56:57.159 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : openblas-0.3.29-1.el9.x86_64 51/102 2026-03-09T20:56:57.161 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : python3-jsonpointer-2.0-4.el9.noarch 61/102 2026-03-09T20:56:57.162 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : flexiblas-3.0.4-9.el9.x86_64 52/102 2026-03-09T20:56:57.164 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : python3-typing-extensions-4.15.0-1.el9.noarch 62/102 2026-03-09T20:56:57.164 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : python3-ply-3.11-14.el9.noarch 53/102 2026-03-09T20:56:57.167 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : python3-repoze-lru-0.7-16.el9.noarch 54/102 2026-03-09T20:56:57.168 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : python3-idna-2.10-7.el9.1.noarch 63/102 2026-03-09T20:56:57.170 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : python3-jaraco-8.2.1-3.el9.noarch 55/102 2026-03-09T20:56:57.172 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : python3-more-itertools-8.12.0-2.el9.noarch 56/102 2026-03-09T20:56:57.174 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : python3-pysocks-1.7.1-12.el9.noarch 64/102 2026-03-09T20:56:57.175 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : python3-toml-0.10.2-6.el9.noarch 57/102 2026-03-09T20:56:57.176 INFO:teuthology.orchestra.run.vm08.stdout: Running scriptlet: ceph-immutable-object-cache-2:19.2.3-678.ge911bd 50/102 2026-03-09T20:56:57.176 INFO:teuthology.orchestra.run.vm08.stdout:Glob pattern passed to enable, but globs are not supported for this. 2026-03-09T20:56:57.176 INFO:teuthology.orchestra.run.vm08.stdout:Invalid unit name "ceph-immutable-object-cache@*.service" escaped as "ceph-immutable-object-cache@\x2a.service". 2026-03-09T20:56:57.176 INFO:teuthology.orchestra.run.vm08.stdout: 2026-03-09T20:56:57.176 INFO:teuthology.orchestra.run.vm08.stdout: Erasing : ceph-immutable-object-cache-2:19.2.3-678.ge911bd 50/102 2026-03-09T20:56:57.179 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : python3-pytz-2021.1-5.el9.noarch 58/102 2026-03-09T20:56:57.179 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : python3-pyasn1-0.4.8-7.el9.noarch 65/102 2026-03-09T20:56:57.186 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : python3-logutils-0.3.5-21.el9.noarch 66/102 2026-03-09T20:56:57.187 INFO:teuthology.orchestra.run.vm08.stdout: Running scriptlet: ceph-immutable-object-cache-2:19.2.3-678.ge911bd 50/102 2026-03-09T20:56:57.187 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : python3-backports-tarfile-1.2.0-1.el9.noarch 59/102 2026-03-09T20:56:57.188 INFO:teuthology.orchestra.run.vm08.stdout: Erasing : openblas-0.3.29-1.el9.x86_64 51/102 2026-03-09T20:56:57.190 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : python3-webob-1.8.8-2.el9.noarch 67/102 2026-03-09T20:56:57.191 INFO:teuthology.orchestra.run.vm08.stdout: Erasing : flexiblas-3.0.4-9.el9.x86_64 52/102 2026-03-09T20:56:57.193 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : python3-devel-3.9.25-3.el9.x86_64 60/102 2026-03-09T20:56:57.193 INFO:teuthology.orchestra.run.vm08.stdout: Erasing : python3-ply-3.11-14.el9.noarch 53/102 2026-03-09T20:56:57.195 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : python3-jsonpointer-2.0-4.el9.noarch 61/102 2026-03-09T20:56:57.195 INFO:teuthology.orchestra.run.vm08.stdout: Erasing : python3-repoze-lru-0.7-16.el9.noarch 54/102 2026-03-09T20:56:57.197 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : python3-cachetools-4.2.4-1.el9.noarch 68/102 2026-03-09T20:56:57.198 INFO:teuthology.orchestra.run.vm08.stdout: Erasing : python3-jaraco-8.2.1-3.el9.noarch 55/102 2026-03-09T20:56:57.198 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : python3-typing-extensions-4.15.0-1.el9.noarch 62/102 2026-03-09T20:56:57.200 INFO:teuthology.orchestra.run.vm08.stdout: Erasing : python3-more-itertools-8.12.0-2.el9.noarch 56/102 2026-03-09T20:56:57.200 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : python3-chardet-4.0.0-5.el9.noarch 69/102 2026-03-09T20:56:57.201 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : python3-idna-2.10-7.el9.1.noarch 63/102 2026-03-09T20:56:57.203 INFO:teuthology.orchestra.run.vm08.stdout: Erasing : python3-toml-0.10.2-6.el9.noarch 57/102 2026-03-09T20:56:57.204 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : python3-autocommand-2.2.2-8.el9.noarch 70/102 2026-03-09T20:56:57.206 INFO:teuthology.orchestra.run.vm08.stdout: Erasing : python3-pytz-2021.1-5.el9.noarch 58/102 2026-03-09T20:56:57.206 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : python3-pysocks-1.7.1-12.el9.noarch 64/102 2026-03-09T20:56:57.207 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : python3-packaging-20.9-5.el9.noarch 71/102 2026-03-09T20:56:57.211 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : python3-pyasn1-0.4.8-7.el9.noarch 65/102 2026-03-09T20:56:57.213 INFO:teuthology.orchestra.run.vm08.stdout: Erasing : python3-backports-tarfile-1.2.0-1.el9.noarch 59/102 2026-03-09T20:56:57.213 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : grpc-data-1.46.7-10.el9.noarch 72/102 2026-03-09T20:56:57.216 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : python3-logutils-0.3.5-21.el9.noarch 66/102 2026-03-09T20:56:57.217 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : python3-protobuf-3.14.0-17.el9.noarch 73/102 2026-03-09T20:56:57.218 INFO:teuthology.orchestra.run.vm08.stdout: Erasing : python3-devel-3.9.25-3.el9.x86_64 60/102 2026-03-09T20:56:57.220 INFO:teuthology.orchestra.run.vm08.stdout: Erasing : python3-jsonpointer-2.0-4.el9.noarch 61/102 2026-03-09T20:56:57.221 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : python3-zc-lockfile-2.0-10.el9.noarch 74/102 2026-03-09T20:56:57.222 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : python3-webob-1.8.8-2.el9.noarch 67/102 2026-03-09T20:56:57.223 INFO:teuthology.orchestra.run.vm08.stdout: Erasing : python3-typing-extensions-4.15.0-1.el9.noarch 62/102 2026-03-09T20:56:57.225 INFO:teuthology.orchestra.run.vm08.stdout: Erasing : python3-idna-2.10-7.el9.1.noarch 63/102 2026-03-09T20:56:57.228 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : python3-cachetools-4.2.4-1.el9.noarch 68/102 2026-03-09T20:56:57.230 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : python3-natsort-7.1.1-5.el9.noarch 75/102 2026-03-09T20:56:57.230 INFO:teuthology.orchestra.run.vm08.stdout: Erasing : python3-pysocks-1.7.1-12.el9.noarch 64/102 2026-03-09T20:56:57.231 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : python3-chardet-4.0.0-5.el9.noarch 69/102 2026-03-09T20:56:57.235 INFO:teuthology.orchestra.run.vm08.stdout: Erasing : python3-pyasn1-0.4.8-7.el9.noarch 65/102 2026-03-09T20:56:57.236 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : python3-autocommand-2.2.2-8.el9.noarch 70/102 2026-03-09T20:56:57.237 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : python3-oauthlib-3.1.1-5.el9.noarch 76/102 2026-03-09T20:56:57.239 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : python3-packaging-20.9-5.el9.noarch 71/102 2026-03-09T20:56:57.240 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : python3-websocket-client-1.2.3-2.el9.noarch 77/102 2026-03-09T20:56:57.241 INFO:teuthology.orchestra.run.vm08.stdout: Erasing : python3-logutils-0.3.5-21.el9.noarch 66/102 2026-03-09T20:56:57.243 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : python3-certifi-2023.05.07-4.el9.noarch 78/102 2026-03-09T20:56:57.244 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : ceph-grafana-dashboards-2:19.2.3-678.ge911bdeb.e 79/102 2026-03-09T20:56:57.245 INFO:teuthology.orchestra.run.vm08.stdout: Erasing : python3-webob-1.8.8-2.el9.noarch 67/102 2026-03-09T20:56:57.245 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : grpc-data-1.46.7-10.el9.noarch 72/102 2026-03-09T20:56:57.250 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : python3-protobuf-3.14.0-17.el9.noarch 73/102 2026-03-09T20:56:57.250 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : ceph-prometheus-alerts-2:19.2.3-678.ge911bdeb.el 80/102 2026-03-09T20:56:57.251 INFO:teuthology.orchestra.run.vm08.stdout: Erasing : python3-cachetools-4.2.4-1.el9.noarch 68/102 2026-03-09T20:56:57.254 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : python3-zc-lockfile-2.0-10.el9.noarch 74/102 2026-03-09T20:56:57.254 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : python3-werkzeug-2.0.3-3.el9.1.noarch 81/102 2026-03-09T20:56:57.255 INFO:teuthology.orchestra.run.vm08.stdout: Erasing : python3-chardet-4.0.0-5.el9.noarch 69/102 2026-03-09T20:56:57.259 INFO:teuthology.orchestra.run.vm08.stdout: Erasing : python3-autocommand-2.2.2-8.el9.noarch 70/102 2026-03-09T20:56:57.262 INFO:teuthology.orchestra.run.vm08.stdout: Erasing : python3-packaging-20.9-5.el9.noarch 71/102 2026-03-09T20:56:57.262 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : python3-natsort-7.1.1-5.el9.noarch 75/102 2026-03-09T20:56:57.268 INFO:teuthology.orchestra.run.vm08.stdout: Erasing : grpc-data-1.46.7-10.el9.noarch 72/102 2026-03-09T20:56:57.268 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : python3-oauthlib-3.1.1-5.el9.noarch 76/102 2026-03-09T20:56:57.272 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : python3-websocket-client-1.2.3-2.el9.noarch 77/102 2026-03-09T20:56:57.272 INFO:teuthology.orchestra.run.vm08.stdout: Erasing : python3-protobuf-3.14.0-17.el9.noarch 73/102 2026-03-09T20:56:57.274 INFO:teuthology.orchestra.run.vm01.stdout: Running scriptlet: ceph-base-2:19.2.3-678.ge911bdeb.el9.x86_64 82/102 2026-03-09T20:56:57.274 INFO:teuthology.orchestra.run.vm01.stdout:Removed "/etc/systemd/system/ceph.target.wants/ceph-crash.service". 2026-03-09T20:56:57.274 INFO:teuthology.orchestra.run.vm01.stdout: 2026-03-09T20:56:57.275 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : python3-certifi-2023.05.07-4.el9.noarch 78/102 2026-03-09T20:56:57.276 INFO:teuthology.orchestra.run.vm08.stdout: Erasing : python3-zc-lockfile-2.0-10.el9.noarch 74/102 2026-03-09T20:56:57.277 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : ceph-grafana-dashboards-2:19.2.3-678.ge911bdeb.e 79/102 2026-03-09T20:56:57.281 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : ceph-base-2:19.2.3-678.ge911bdeb.el9.x86_64 82/102 2026-03-09T20:56:57.283 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : ceph-prometheus-alerts-2:19.2.3-678.ge911bdeb.el 80/102 2026-03-09T20:56:57.284 INFO:teuthology.orchestra.run.vm08.stdout: Erasing : python3-natsort-7.1.1-5.el9.noarch 75/102 2026-03-09T20:56:57.288 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : python3-werkzeug-2.0.3-3.el9.1.noarch 81/102 2026-03-09T20:56:57.291 INFO:teuthology.orchestra.run.vm08.stdout: Erasing : python3-oauthlib-3.1.1-5.el9.noarch 76/102 2026-03-09T20:56:57.294 INFO:teuthology.orchestra.run.vm08.stdout: Erasing : python3-websocket-client-1.2.3-2.el9.noarch 77/102 2026-03-09T20:56:57.296 INFO:teuthology.orchestra.run.vm08.stdout: Erasing : python3-certifi-2023.05.07-4.el9.noarch 78/102 2026-03-09T20:56:57.297 INFO:teuthology.orchestra.run.vm08.stdout: Erasing : ceph-grafana-dashboards-2:19.2.3-678.ge911bdeb.e 79/102 2026-03-09T20:56:57.303 INFO:teuthology.orchestra.run.vm08.stdout: Erasing : ceph-prometheus-alerts-2:19.2.3-678.ge911bdeb.el 80/102 2026-03-09T20:56:57.307 INFO:teuthology.orchestra.run.vm08.stdout: Erasing : python3-werkzeug-2.0.3-3.el9.1.noarch 81/102 2026-03-09T20:56:57.308 INFO:teuthology.orchestra.run.vm06.stdout: Running scriptlet: ceph-base-2:19.2.3-678.ge911bdeb.el9.x86_64 82/102 2026-03-09T20:56:57.308 INFO:teuthology.orchestra.run.vm06.stdout:Removed "/etc/systemd/system/ceph.target.wants/ceph-crash.service". 2026-03-09T20:56:57.308 INFO:teuthology.orchestra.run.vm06.stdout: 2026-03-09T20:56:57.311 INFO:teuthology.orchestra.run.vm01.stdout: Running scriptlet: ceph-base-2:19.2.3-678.ge911bdeb.el9.x86_64 82/102 2026-03-09T20:56:57.311 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : ceph-common-2:19.2.3-678.ge911bdeb.el9.x86_64 83/102 2026-03-09T20:56:57.316 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : ceph-base-2:19.2.3-678.ge911bdeb.el9.x86_64 82/102 2026-03-09T20:56:57.324 INFO:teuthology.orchestra.run.vm01.stdout: Running scriptlet: ceph-common-2:19.2.3-678.ge911bdeb.el9.x86_64 83/102 2026-03-09T20:56:57.330 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : qatzip-libs-1.3.1-1.el9.x86_64 84/102 2026-03-09T20:56:57.330 INFO:teuthology.orchestra.run.vm08.stdout: Running scriptlet: ceph-base-2:19.2.3-678.ge911bdeb.el9.x86_64 82/102 2026-03-09T20:56:57.330 INFO:teuthology.orchestra.run.vm08.stdout:Removed "/etc/systemd/system/ceph.target.wants/ceph-crash.service". 2026-03-09T20:56:57.330 INFO:teuthology.orchestra.run.vm08.stdout: 2026-03-09T20:56:57.334 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : python3-ceph-common-2:19.2.3-678.ge911bdeb.el9.x 85/102 2026-03-09T20:56:57.336 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : python3-prettytable-0.7.2-27.el9.noarch 86/102 2026-03-09T20:56:57.336 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : ceph-selinux-2:19.2.3-678.ge911bdeb.el9.x86_64 87/102 2026-03-09T20:56:57.338 INFO:teuthology.orchestra.run.vm08.stdout: Erasing : ceph-base-2:19.2.3-678.ge911bdeb.el9.x86_64 82/102 2026-03-09T20:56:57.347 INFO:teuthology.orchestra.run.vm06.stdout: Running scriptlet: ceph-base-2:19.2.3-678.ge911bdeb.el9.x86_64 82/102 2026-03-09T20:56:57.347 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : ceph-common-2:19.2.3-678.ge911bdeb.el9.x86_64 83/102 2026-03-09T20:56:57.361 INFO:teuthology.orchestra.run.vm06.stdout: Running scriptlet: ceph-common-2:19.2.3-678.ge911bdeb.el9.x86_64 83/102 2026-03-09T20:56:57.365 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : qatzip-libs-1.3.1-1.el9.x86_64 84/102 2026-03-09T20:56:57.367 INFO:teuthology.orchestra.run.vm08.stdout: Running scriptlet: ceph-base-2:19.2.3-678.ge911bdeb.el9.x86_64 82/102 2026-03-09T20:56:57.367 INFO:teuthology.orchestra.run.vm08.stdout: Erasing : ceph-common-2:19.2.3-678.ge911bdeb.el9.x86_64 83/102 2026-03-09T20:56:57.368 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : python3-ceph-common-2:19.2.3-678.ge911bdeb.el9.x 85/102 2026-03-09T20:56:57.371 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : python3-prettytable-0.7.2-27.el9.noarch 86/102 2026-03-09T20:56:57.371 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : ceph-selinux-2:19.2.3-678.ge911bdeb.el9.x86_64 87/102 2026-03-09T20:56:57.379 INFO:teuthology.orchestra.run.vm08.stdout: Running scriptlet: ceph-common-2:19.2.3-678.ge911bdeb.el9.x86_64 83/102 2026-03-09T20:56:57.384 INFO:teuthology.orchestra.run.vm08.stdout: Erasing : qatzip-libs-1.3.1-1.el9.x86_64 84/102 2026-03-09T20:56:57.387 INFO:teuthology.orchestra.run.vm08.stdout: Erasing : python3-ceph-common-2:19.2.3-678.ge911bdeb.el9.x 85/102 2026-03-09T20:56:57.389 INFO:teuthology.orchestra.run.vm08.stdout: Erasing : python3-prettytable-0.7.2-27.el9.noarch 86/102 2026-03-09T20:56:57.389 INFO:teuthology.orchestra.run.vm08.stdout: Erasing : ceph-selinux-2:19.2.3-678.ge911bdeb.el9.x86_64 87/102 2026-03-09T20:57:03.080 INFO:teuthology.orchestra.run.vm08.stdout: Running scriptlet: ceph-selinux-2:19.2.3-678.ge911bdeb.el9.x86_64 87/102 2026-03-09T20:57:03.080 INFO:teuthology.orchestra.run.vm08.stdout:skipping the directory /sys 2026-03-09T20:57:03.080 INFO:teuthology.orchestra.run.vm08.stdout:skipping the directory /proc 2026-03-09T20:57:03.080 INFO:teuthology.orchestra.run.vm08.stdout:skipping the directory /mnt 2026-03-09T20:57:03.080 INFO:teuthology.orchestra.run.vm08.stdout:skipping the directory /var/tmp 2026-03-09T20:57:03.080 INFO:teuthology.orchestra.run.vm08.stdout:skipping the directory /home 2026-03-09T20:57:03.080 INFO:teuthology.orchestra.run.vm08.stdout:skipping the directory /root 2026-03-09T20:57:03.080 INFO:teuthology.orchestra.run.vm08.stdout:skipping the directory /tmp 2026-03-09T20:57:03.080 INFO:teuthology.orchestra.run.vm08.stdout: 2026-03-09T20:57:03.108 INFO:teuthology.orchestra.run.vm08.stdout: Erasing : qatlib-25.08.0-2.el9.x86_64 88/102 2026-03-09T20:57:03.128 INFO:teuthology.orchestra.run.vm08.stdout: Running scriptlet: qatlib-service-25.08.0-2.el9.x86_64 89/102 2026-03-09T20:57:03.128 INFO:teuthology.orchestra.run.vm08.stdout: Erasing : qatlib-service-25.08.0-2.el9.x86_64 89/102 2026-03-09T20:57:03.152 INFO:teuthology.orchestra.run.vm06.stdout: Running scriptlet: ceph-selinux-2:19.2.3-678.ge911bdeb.el9.x86_64 87/102 2026-03-09T20:57:03.152 INFO:teuthology.orchestra.run.vm06.stdout:skipping the directory /sys 2026-03-09T20:57:03.152 INFO:teuthology.orchestra.run.vm06.stdout:skipping the directory /proc 2026-03-09T20:57:03.152 INFO:teuthology.orchestra.run.vm06.stdout:skipping the directory /mnt 2026-03-09T20:57:03.152 INFO:teuthology.orchestra.run.vm06.stdout:skipping the directory /var/tmp 2026-03-09T20:57:03.152 INFO:teuthology.orchestra.run.vm06.stdout:skipping the directory /home 2026-03-09T20:57:03.152 INFO:teuthology.orchestra.run.vm06.stdout:skipping the directory /root 2026-03-09T20:57:03.152 INFO:teuthology.orchestra.run.vm06.stdout:skipping the directory /tmp 2026-03-09T20:57:03.152 INFO:teuthology.orchestra.run.vm06.stdout: 2026-03-09T20:57:03.167 INFO:teuthology.orchestra.run.vm08.stdout: Running scriptlet: qatlib-service-25.08.0-2.el9.x86_64 89/102 2026-03-09T20:57:03.170 INFO:teuthology.orchestra.run.vm08.stdout: Erasing : gperftools-libs-2.9.1-3.el9.x86_64 90/102 2026-03-09T20:57:03.170 INFO:teuthology.orchestra.run.vm01.stdout: Running scriptlet: ceph-selinux-2:19.2.3-678.ge911bdeb.el9.x86_64 87/102 2026-03-09T20:57:03.170 INFO:teuthology.orchestra.run.vm01.stdout:skipping the directory /sys 2026-03-09T20:57:03.170 INFO:teuthology.orchestra.run.vm01.stdout:skipping the directory /proc 2026-03-09T20:57:03.170 INFO:teuthology.orchestra.run.vm01.stdout:skipping the directory /mnt 2026-03-09T20:57:03.170 INFO:teuthology.orchestra.run.vm01.stdout:skipping the directory /var/tmp 2026-03-09T20:57:03.170 INFO:teuthology.orchestra.run.vm01.stdout:skipping the directory /home 2026-03-09T20:57:03.170 INFO:teuthology.orchestra.run.vm01.stdout:skipping the directory /root 2026-03-09T20:57:03.170 INFO:teuthology.orchestra.run.vm01.stdout:skipping the directory /tmp 2026-03-09T20:57:03.170 INFO:teuthology.orchestra.run.vm01.stdout: 2026-03-09T20:57:03.172 INFO:teuthology.orchestra.run.vm08.stdout: Erasing : libunwind-1.6.2-1.el9.x86_64 91/102 2026-03-09T20:57:03.173 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : qatlib-25.08.0-2.el9.x86_64 88/102 2026-03-09T20:57:03.174 INFO:teuthology.orchestra.run.vm08.stdout: Erasing : pciutils-3.7.0-7.el9.x86_64 92/102 2026-03-09T20:57:03.176 INFO:teuthology.orchestra.run.vm08.stdout: Erasing : liboath-2.6.12-1.el9.x86_64 93/102 2026-03-09T20:57:03.176 INFO:teuthology.orchestra.run.vm08.stdout: Erasing : libradosstriper1-2:19.2.3-678.ge911bdeb.el9.x86_ 94/102 2026-03-09T20:57:03.181 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : qatlib-25.08.0-2.el9.x86_64 88/102 2026-03-09T20:57:03.192 INFO:teuthology.orchestra.run.vm06.stdout: Running scriptlet: qatlib-service-25.08.0-2.el9.x86_64 89/102 2026-03-09T20:57:03.192 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : qatlib-service-25.08.0-2.el9.x86_64 89/102 2026-03-09T20:57:03.193 INFO:teuthology.orchestra.run.vm08.stdout: Running scriptlet: libradosstriper1-2:19.2.3-678.ge911bdeb.el9.x86_ 94/102 2026-03-09T20:57:03.196 INFO:teuthology.orchestra.run.vm08.stdout: Erasing : ledmon-libs-1.1.0-3.el9.x86_64 95/102 2026-03-09T20:57:03.197 INFO:teuthology.orchestra.run.vm01.stdout: Running scriptlet: qatlib-service-25.08.0-2.el9.x86_64 89/102 2026-03-09T20:57:03.197 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : qatlib-service-25.08.0-2.el9.x86_64 89/102 2026-03-09T20:57:03.199 INFO:teuthology.orchestra.run.vm08.stdout: Erasing : libquadmath-11.5.0-14.el9.x86_64 96/102 2026-03-09T20:57:03.202 INFO:teuthology.orchestra.run.vm08.stdout: Erasing : python3-markupsafe-1.1.1-12.el9.x86_64 97/102 2026-03-09T20:57:03.202 INFO:teuthology.orchestra.run.vm06.stdout: Running scriptlet: qatlib-service-25.08.0-2.el9.x86_64 89/102 2026-03-09T20:57:03.205 INFO:teuthology.orchestra.run.vm08.stdout: Erasing : protobuf-3.14.0-17.el9.x86_64 98/102 2026-03-09T20:57:03.205 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : gperftools-libs-2.9.1-3.el9.x86_64 90/102 2026-03-09T20:57:03.206 INFO:teuthology.orchestra.run.vm01.stdout: Running scriptlet: qatlib-service-25.08.0-2.el9.x86_64 89/102 2026-03-09T20:57:03.208 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : libunwind-1.6.2-1.el9.x86_64 91/102 2026-03-09T20:57:03.209 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : gperftools-libs-2.9.1-3.el9.x86_64 90/102 2026-03-09T20:57:03.210 INFO:teuthology.orchestra.run.vm08.stdout: Erasing : libconfig-1.7.2-9.el9.x86_64 99/102 2026-03-09T20:57:03.211 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : pciutils-3.7.0-7.el9.x86_64 92/102 2026-03-09T20:57:03.211 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : libunwind-1.6.2-1.el9.x86_64 91/102 2026-03-09T20:57:03.214 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : liboath-2.6.12-1.el9.x86_64 93/102 2026-03-09T20:57:03.214 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : libradosstriper1-2:19.2.3-678.ge911bdeb.el9.x86_ 94/102 2026-03-09T20:57:03.214 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : pciutils-3.7.0-7.el9.x86_64 92/102 2026-03-09T20:57:03.216 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : liboath-2.6.12-1.el9.x86_64 93/102 2026-03-09T20:57:03.216 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : libradosstriper1-2:19.2.3-678.ge911bdeb.el9.x86_ 94/102 2026-03-09T20:57:03.218 INFO:teuthology.orchestra.run.vm08.stdout: Erasing : cryptsetup-2.8.1-3.el9.x86_64 100/102 2026-03-09T20:57:03.222 INFO:teuthology.orchestra.run.vm08.stdout: Erasing : abseil-cpp-20211102.0-4.el9.x86_64 101/102 2026-03-09T20:57:03.222 INFO:teuthology.orchestra.run.vm08.stdout: Erasing : libcephsqlite-2:19.2.3-678.ge911bdeb.el9.x86_64 102/102 2026-03-09T20:57:03.228 INFO:teuthology.orchestra.run.vm06.stdout: Running scriptlet: libradosstriper1-2:19.2.3-678.ge911bdeb.el9.x86_ 94/102 2026-03-09T20:57:03.230 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : ledmon-libs-1.1.0-3.el9.x86_64 95/102 2026-03-09T20:57:03.232 INFO:teuthology.orchestra.run.vm01.stdout: Running scriptlet: libradosstriper1-2:19.2.3-678.ge911bdeb.el9.x86_ 94/102 2026-03-09T20:57:03.233 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : libquadmath-11.5.0-14.el9.x86_64 96/102 2026-03-09T20:57:03.235 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : ledmon-libs-1.1.0-3.el9.x86_64 95/102 2026-03-09T20:57:03.236 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : python3-markupsafe-1.1.1-12.el9.x86_64 97/102 2026-03-09T20:57:03.237 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : libquadmath-11.5.0-14.el9.x86_64 96/102 2026-03-09T20:57:03.239 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : protobuf-3.14.0-17.el9.x86_64 98/102 2026-03-09T20:57:03.241 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : python3-markupsafe-1.1.1-12.el9.x86_64 97/102 2026-03-09T20:57:03.243 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : protobuf-3.14.0-17.el9.x86_64 98/102 2026-03-09T20:57:03.245 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : libconfig-1.7.2-9.el9.x86_64 99/102 2026-03-09T20:57:03.249 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : libconfig-1.7.2-9.el9.x86_64 99/102 2026-03-09T20:57:03.253 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : cryptsetup-2.8.1-3.el9.x86_64 100/102 2026-03-09T20:57:03.257 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : cryptsetup-2.8.1-3.el9.x86_64 100/102 2026-03-09T20:57:03.259 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : abseil-cpp-20211102.0-4.el9.x86_64 101/102 2026-03-09T20:57:03.259 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : libcephsqlite-2:19.2.3-678.ge911bdeb.el9.x86_64 102/102 2026-03-09T20:57:03.262 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : abseil-cpp-20211102.0-4.el9.x86_64 101/102 2026-03-09T20:57:03.262 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : libcephsqlite-2:19.2.3-678.ge911bdeb.el9.x86_64 102/102 2026-03-09T20:57:03.319 INFO:teuthology.orchestra.run.vm08.stdout: Running scriptlet: libcephsqlite-2:19.2.3-678.ge911bdeb.el9.x86_64 102/102 2026-03-09T20:57:03.319 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : abseil-cpp-20211102.0-4.el9.x86_64 1/102 2026-03-09T20:57:03.320 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : ceph-base-2:19.2.3-678.ge911bdeb.el9.x86_64 2/102 2026-03-09T20:57:03.320 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : ceph-common-2:19.2.3-678.ge911bdeb.el9.x86_64 3/102 2026-03-09T20:57:03.320 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : ceph-grafana-dashboards-2:19.2.3-678.ge911bdeb.e 4/102 2026-03-09T20:57:03.320 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : ceph-immutable-object-cache-2:19.2.3-678.ge911bd 5/102 2026-03-09T20:57:03.320 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : ceph-mgr-2:19.2.3-678.ge911bdeb.el9.x86_64 6/102 2026-03-09T20:57:03.320 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : ceph-mgr-cephadm-2:19.2.3-678.ge911bdeb.el9.noar 7/102 2026-03-09T20:57:03.320 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : ceph-mgr-dashboard-2:19.2.3-678.ge911bdeb.el9.no 8/102 2026-03-09T20:57:03.320 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : ceph-mgr-diskprediction-local-2:19.2.3-678.ge911 9/102 2026-03-09T20:57:03.320 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : ceph-mgr-modules-core-2:19.2.3-678.ge911bdeb.el9 10/102 2026-03-09T20:57:03.320 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : ceph-mgr-rook-2:19.2.3-678.ge911bdeb.el9.noarch 11/102 2026-03-09T20:57:03.320 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : ceph-osd-2:19.2.3-678.ge911bdeb.el9.x86_64 12/102 2026-03-09T20:57:03.320 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : ceph-prometheus-alerts-2:19.2.3-678.ge911bdeb.el 13/102 2026-03-09T20:57:03.320 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : ceph-selinux-2:19.2.3-678.ge911bdeb.el9.x86_64 14/102 2026-03-09T20:57:03.320 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : ceph-volume-2:19.2.3-678.ge911bdeb.el9.noarch 15/102 2026-03-09T20:57:03.320 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : cryptsetup-2.8.1-3.el9.x86_64 16/102 2026-03-09T20:57:03.320 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : flexiblas-3.0.4-9.el9.x86_64 17/102 2026-03-09T20:57:03.320 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : flexiblas-netlib-3.0.4-9.el9.x86_64 18/102 2026-03-09T20:57:03.320 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : flexiblas-openblas-openmp-3.0.4-9.el9.x86_64 19/102 2026-03-09T20:57:03.320 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : gperftools-libs-2.9.1-3.el9.x86_64 20/102 2026-03-09T20:57:03.320 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : grpc-data-1.46.7-10.el9.noarch 21/102 2026-03-09T20:57:03.320 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : ledmon-libs-1.1.0-3.el9.x86_64 22/102 2026-03-09T20:57:03.320 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : libcephsqlite-2:19.2.3-678.ge911bdeb.el9.x86_64 23/102 2026-03-09T20:57:03.320 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : libconfig-1.7.2-9.el9.x86_64 24/102 2026-03-09T20:57:03.320 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : libgfortran-11.5.0-14.el9.x86_64 25/102 2026-03-09T20:57:03.320 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : liboath-2.6.12-1.el9.x86_64 26/102 2026-03-09T20:57:03.320 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : libquadmath-11.5.0-14.el9.x86_64 27/102 2026-03-09T20:57:03.320 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : libradosstriper1-2:19.2.3-678.ge911bdeb.el9.x86_ 28/102 2026-03-09T20:57:03.320 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : libstoragemgmt-1.10.1-1.el9.x86_64 29/102 2026-03-09T20:57:03.320 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : libunwind-1.6.2-1.el9.x86_64 30/102 2026-03-09T20:57:03.320 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : openblas-0.3.29-1.el9.x86_64 31/102 2026-03-09T20:57:03.320 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : openblas-openmp-0.3.29-1.el9.x86_64 32/102 2026-03-09T20:57:03.320 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : pciutils-3.7.0-7.el9.x86_64 33/102 2026-03-09T20:57:03.320 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : protobuf-3.14.0-17.el9.x86_64 34/102 2026-03-09T20:57:03.320 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : protobuf-compiler-3.14.0-17.el9.x86_64 35/102 2026-03-09T20:57:03.320 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : python3-asyncssh-2.13.2-5.el9.noarch 36/102 2026-03-09T20:57:03.320 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : python3-autocommand-2.2.2-8.el9.noarch 37/102 2026-03-09T20:57:03.320 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : python3-babel-2.9.1-2.el9.noarch 38/102 2026-03-09T20:57:03.320 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : python3-backports-tarfile-1.2.0-1.el9.noarch 39/102 2026-03-09T20:57:03.320 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : python3-bcrypt-3.2.2-1.el9.x86_64 40/102 2026-03-09T20:57:03.320 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : python3-cachetools-4.2.4-1.el9.noarch 41/102 2026-03-09T20:57:03.320 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : python3-ceph-common-2:19.2.3-678.ge911bdeb.el9.x 42/102 2026-03-09T20:57:03.320 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : python3-certifi-2023.05.07-4.el9.noarch 43/102 2026-03-09T20:57:03.320 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : python3-cffi-1.14.5-5.el9.x86_64 44/102 2026-03-09T20:57:03.320 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : python3-chardet-4.0.0-5.el9.noarch 45/102 2026-03-09T20:57:03.320 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : python3-cheroot-10.0.1-4.el9.noarch 46/102 2026-03-09T20:57:03.320 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : python3-cherrypy-18.6.1-2.el9.noarch 47/102 2026-03-09T20:57:03.320 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : python3-cryptography-36.0.1-5.el9.x86_64 48/102 2026-03-09T20:57:03.320 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : python3-devel-3.9.25-3.el9.x86_64 49/102 2026-03-09T20:57:03.320 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : python3-google-auth-1:2.45.0-1.el9.noarch 50/102 2026-03-09T20:57:03.320 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : python3-grpcio-1.46.7-10.el9.x86_64 51/102 2026-03-09T20:57:03.320 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : python3-grpcio-tools-1.46.7-10.el9.x86_64 52/102 2026-03-09T20:57:03.320 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : python3-idna-2.10-7.el9.1.noarch 53/102 2026-03-09T20:57:03.320 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : python3-jaraco-8.2.1-3.el9.noarch 54/102 2026-03-09T20:57:03.320 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : python3-jaraco-classes-3.2.1-5.el9.noarch 55/102 2026-03-09T20:57:03.320 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : python3-jaraco-collections-3.0.0-8.el9.noarch 56/102 2026-03-09T20:57:03.320 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : python3-jaraco-context-6.0.1-3.el9.noarch 57/102 2026-03-09T20:57:03.320 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : python3-jaraco-functools-3.5.0-2.el9.noarch 58/102 2026-03-09T20:57:03.320 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : python3-jaraco-text-4.0.0-2.el9.noarch 59/102 2026-03-09T20:57:03.320 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : python3-jinja2-2.11.3-8.el9.noarch 60/102 2026-03-09T20:57:03.320 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : python3-jsonpatch-1.21-16.el9.noarch 61/102 2026-03-09T20:57:03.320 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : python3-jsonpointer-2.0-4.el9.noarch 62/102 2026-03-09T20:57:03.320 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : python3-kubernetes-1:26.1.0-3.el9.noarch 63/102 2026-03-09T20:57:03.320 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : python3-libstoragemgmt-1.10.1-1.el9.x86_64 64/102 2026-03-09T20:57:03.321 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : python3-logutils-0.3.5-21.el9.noarch 65/102 2026-03-09T20:57:03.321 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : python3-mako-1.1.4-6.el9.noarch 66/102 2026-03-09T20:57:03.321 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : python3-markupsafe-1.1.1-12.el9.x86_64 67/102 2026-03-09T20:57:03.321 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : python3-more-itertools-8.12.0-2.el9.noarch 68/102 2026-03-09T20:57:03.321 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : python3-natsort-7.1.1-5.el9.noarch 69/102 2026-03-09T20:57:03.321 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : python3-numpy-1:1.23.5-2.el9.x86_64 70/102 2026-03-09T20:57:03.321 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : python3-numpy-f2py-1:1.23.5-2.el9.x86_64 71/102 2026-03-09T20:57:03.321 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : python3-oauthlib-3.1.1-5.el9.noarch 72/102 2026-03-09T20:57:03.321 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : python3-packaging-20.9-5.el9.noarch 73/102 2026-03-09T20:57:03.321 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : python3-pecan-1.4.2-3.el9.noarch 74/102 2026-03-09T20:57:03.321 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : python3-ply-3.11-14.el9.noarch 75/102 2026-03-09T20:57:03.321 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : python3-portend-3.1.0-2.el9.noarch 76/102 2026-03-09T20:57:03.321 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : python3-prettytable-0.7.2-27.el9.noarch 77/102 2026-03-09T20:57:03.321 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : python3-protobuf-3.14.0-17.el9.noarch 78/102 2026-03-09T20:57:03.321 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : python3-pyOpenSSL-21.0.0-1.el9.noarch 79/102 2026-03-09T20:57:03.321 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : python3-pyasn1-0.4.8-7.el9.noarch 80/102 2026-03-09T20:57:03.321 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : python3-pyasn1-modules-0.4.8-7.el9.noarch 81/102 2026-03-09T20:57:03.321 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : python3-pycparser-2.20-6.el9.noarch 82/102 2026-03-09T20:57:03.321 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : python3-pysocks-1.7.1-12.el9.noarch 83/102 2026-03-09T20:57:03.321 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : python3-pytz-2021.1-5.el9.noarch 84/102 2026-03-09T20:57:03.321 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : python3-repoze-lru-0.7-16.el9.noarch 85/102 2026-03-09T20:57:03.321 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : python3-requests-2.25.1-10.el9.noarch 86/102 2026-03-09T20:57:03.321 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : python3-requests-oauthlib-1.3.0-12.el9.noarch 87/102 2026-03-09T20:57:03.321 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : python3-routes-2.5.1-5.el9.noarch 88/102 2026-03-09T20:57:03.321 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : python3-rsa-4.9-2.el9.noarch 89/102 2026-03-09T20:57:03.321 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : python3-scipy-1.9.3-2.el9.x86_64 90/102 2026-03-09T20:57:03.321 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : python3-tempora-5.0.0-2.el9.noarch 91/102 2026-03-09T20:57:03.321 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : python3-toml-0.10.2-6.el9.noarch 92/102 2026-03-09T20:57:03.321 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : python3-typing-extensions-4.15.0-1.el9.noarch 93/102 2026-03-09T20:57:03.321 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : python3-urllib3-1.26.5-7.el9.noarch 94/102 2026-03-09T20:57:03.321 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : python3-webob-1.8.8-2.el9.noarch 95/102 2026-03-09T20:57:03.321 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : python3-websocket-client-1.2.3-2.el9.noarch 96/102 2026-03-09T20:57:03.321 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : python3-werkzeug-2.0.3-3.el9.1.noarch 97/102 2026-03-09T20:57:03.321 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : python3-zc-lockfile-2.0-10.el9.noarch 98/102 2026-03-09T20:57:03.321 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : qatlib-25.08.0-2.el9.x86_64 99/102 2026-03-09T20:57:03.321 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : qatlib-service-25.08.0-2.el9.x86_64 100/102 2026-03-09T20:57:03.321 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : qatzip-libs-1.3.1-1.el9.x86_64 101/102 2026-03-09T20:57:03.363 INFO:teuthology.orchestra.run.vm06.stdout: Running scriptlet: libcephsqlite-2:19.2.3-678.ge911bdeb.el9.x86_64 102/102 2026-03-09T20:57:03.363 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : abseil-cpp-20211102.0-4.el9.x86_64 1/102 2026-03-09T20:57:03.363 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : ceph-base-2:19.2.3-678.ge911bdeb.el9.x86_64 2/102 2026-03-09T20:57:03.363 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : ceph-common-2:19.2.3-678.ge911bdeb.el9.x86_64 3/102 2026-03-09T20:57:03.363 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : ceph-grafana-dashboards-2:19.2.3-678.ge911bdeb.e 4/102 2026-03-09T20:57:03.363 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : ceph-immutable-object-cache-2:19.2.3-678.ge911bd 5/102 2026-03-09T20:57:03.363 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : ceph-mgr-2:19.2.3-678.ge911bdeb.el9.x86_64 6/102 2026-03-09T20:57:03.363 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : ceph-mgr-cephadm-2:19.2.3-678.ge911bdeb.el9.noar 7/102 2026-03-09T20:57:03.363 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : ceph-mgr-dashboard-2:19.2.3-678.ge911bdeb.el9.no 8/102 2026-03-09T20:57:03.363 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : ceph-mgr-diskprediction-local-2:19.2.3-678.ge911 9/102 2026-03-09T20:57:03.363 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : ceph-mgr-modules-core-2:19.2.3-678.ge911bdeb.el9 10/102 2026-03-09T20:57:03.364 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : ceph-mgr-rook-2:19.2.3-678.ge911bdeb.el9.noarch 11/102 2026-03-09T20:57:03.364 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : ceph-osd-2:19.2.3-678.ge911bdeb.el9.x86_64 12/102 2026-03-09T20:57:03.364 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : ceph-prometheus-alerts-2:19.2.3-678.ge911bdeb.el 13/102 2026-03-09T20:57:03.364 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : ceph-selinux-2:19.2.3-678.ge911bdeb.el9.x86_64 14/102 2026-03-09T20:57:03.364 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : ceph-volume-2:19.2.3-678.ge911bdeb.el9.noarch 15/102 2026-03-09T20:57:03.364 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : cryptsetup-2.8.1-3.el9.x86_64 16/102 2026-03-09T20:57:03.364 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : flexiblas-3.0.4-9.el9.x86_64 17/102 2026-03-09T20:57:03.364 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : flexiblas-netlib-3.0.4-9.el9.x86_64 18/102 2026-03-09T20:57:03.364 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : flexiblas-openblas-openmp-3.0.4-9.el9.x86_64 19/102 2026-03-09T20:57:03.364 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : gperftools-libs-2.9.1-3.el9.x86_64 20/102 2026-03-09T20:57:03.364 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : grpc-data-1.46.7-10.el9.noarch 21/102 2026-03-09T20:57:03.364 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : ledmon-libs-1.1.0-3.el9.x86_64 22/102 2026-03-09T20:57:03.364 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : libcephsqlite-2:19.2.3-678.ge911bdeb.el9.x86_64 23/102 2026-03-09T20:57:03.364 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : libconfig-1.7.2-9.el9.x86_64 24/102 2026-03-09T20:57:03.364 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : libgfortran-11.5.0-14.el9.x86_64 25/102 2026-03-09T20:57:03.364 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : liboath-2.6.12-1.el9.x86_64 26/102 2026-03-09T20:57:03.364 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : libquadmath-11.5.0-14.el9.x86_64 27/102 2026-03-09T20:57:03.364 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : libradosstriper1-2:19.2.3-678.ge911bdeb.el9.x86_ 28/102 2026-03-09T20:57:03.365 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : libstoragemgmt-1.10.1-1.el9.x86_64 29/102 2026-03-09T20:57:03.365 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : libunwind-1.6.2-1.el9.x86_64 30/102 2026-03-09T20:57:03.365 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : openblas-0.3.29-1.el9.x86_64 31/102 2026-03-09T20:57:03.365 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : openblas-openmp-0.3.29-1.el9.x86_64 32/102 2026-03-09T20:57:03.365 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : pciutils-3.7.0-7.el9.x86_64 33/102 2026-03-09T20:57:03.365 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : protobuf-3.14.0-17.el9.x86_64 34/102 2026-03-09T20:57:03.365 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : protobuf-compiler-3.14.0-17.el9.x86_64 35/102 2026-03-09T20:57:03.365 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-asyncssh-2.13.2-5.el9.noarch 36/102 2026-03-09T20:57:03.365 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-autocommand-2.2.2-8.el9.noarch 37/102 2026-03-09T20:57:03.365 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-babel-2.9.1-2.el9.noarch 38/102 2026-03-09T20:57:03.365 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-backports-tarfile-1.2.0-1.el9.noarch 39/102 2026-03-09T20:57:03.365 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-bcrypt-3.2.2-1.el9.x86_64 40/102 2026-03-09T20:57:03.365 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-cachetools-4.2.4-1.el9.noarch 41/102 2026-03-09T20:57:03.365 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-ceph-common-2:19.2.3-678.ge911bdeb.el9.x 42/102 2026-03-09T20:57:03.365 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-certifi-2023.05.07-4.el9.noarch 43/102 2026-03-09T20:57:03.365 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-cffi-1.14.5-5.el9.x86_64 44/102 2026-03-09T20:57:03.365 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-chardet-4.0.0-5.el9.noarch 45/102 2026-03-09T20:57:03.365 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-cheroot-10.0.1-4.el9.noarch 46/102 2026-03-09T20:57:03.365 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-cherrypy-18.6.1-2.el9.noarch 47/102 2026-03-09T20:57:03.365 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-cryptography-36.0.1-5.el9.x86_64 48/102 2026-03-09T20:57:03.365 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-devel-3.9.25-3.el9.x86_64 49/102 2026-03-09T20:57:03.365 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-google-auth-1:2.45.0-1.el9.noarch 50/102 2026-03-09T20:57:03.365 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-grpcio-1.46.7-10.el9.x86_64 51/102 2026-03-09T20:57:03.366 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-grpcio-tools-1.46.7-10.el9.x86_64 52/102 2026-03-09T20:57:03.366 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-idna-2.10-7.el9.1.noarch 53/102 2026-03-09T20:57:03.366 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-jaraco-8.2.1-3.el9.noarch 54/102 2026-03-09T20:57:03.366 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-jaraco-classes-3.2.1-5.el9.noarch 55/102 2026-03-09T20:57:03.366 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-jaraco-collections-3.0.0-8.el9.noarch 56/102 2026-03-09T20:57:03.366 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-jaraco-context-6.0.1-3.el9.noarch 57/102 2026-03-09T20:57:03.366 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-jaraco-functools-3.5.0-2.el9.noarch 58/102 2026-03-09T20:57:03.366 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-jaraco-text-4.0.0-2.el9.noarch 59/102 2026-03-09T20:57:03.366 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-jinja2-2.11.3-8.el9.noarch 60/102 2026-03-09T20:57:03.366 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-jsonpatch-1.21-16.el9.noarch 61/102 2026-03-09T20:57:03.366 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-jsonpointer-2.0-4.el9.noarch 62/102 2026-03-09T20:57:03.366 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-kubernetes-1:26.1.0-3.el9.noarch 63/102 2026-03-09T20:57:03.366 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-libstoragemgmt-1.10.1-1.el9.x86_64 64/102 2026-03-09T20:57:03.366 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-logutils-0.3.5-21.el9.noarch 65/102 2026-03-09T20:57:03.366 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-mako-1.1.4-6.el9.noarch 66/102 2026-03-09T20:57:03.366 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-markupsafe-1.1.1-12.el9.x86_64 67/102 2026-03-09T20:57:03.366 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-more-itertools-8.12.0-2.el9.noarch 68/102 2026-03-09T20:57:03.366 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-natsort-7.1.1-5.el9.noarch 69/102 2026-03-09T20:57:03.366 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-numpy-1:1.23.5-2.el9.x86_64 70/102 2026-03-09T20:57:03.366 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-numpy-f2py-1:1.23.5-2.el9.x86_64 71/102 2026-03-09T20:57:03.366 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-oauthlib-3.1.1-5.el9.noarch 72/102 2026-03-09T20:57:03.366 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-packaging-20.9-5.el9.noarch 73/102 2026-03-09T20:57:03.366 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-pecan-1.4.2-3.el9.noarch 74/102 2026-03-09T20:57:03.366 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-ply-3.11-14.el9.noarch 75/102 2026-03-09T20:57:03.366 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-portend-3.1.0-2.el9.noarch 76/102 2026-03-09T20:57:03.367 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-prettytable-0.7.2-27.el9.noarch 77/102 2026-03-09T20:57:03.367 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-protobuf-3.14.0-17.el9.noarch 78/102 2026-03-09T20:57:03.367 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-pyOpenSSL-21.0.0-1.el9.noarch 79/102 2026-03-09T20:57:03.367 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-pyasn1-0.4.8-7.el9.noarch 80/102 2026-03-09T20:57:03.367 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-pyasn1-modules-0.4.8-7.el9.noarch 81/102 2026-03-09T20:57:03.367 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-pycparser-2.20-6.el9.noarch 82/102 2026-03-09T20:57:03.367 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-pysocks-1.7.1-12.el9.noarch 83/102 2026-03-09T20:57:03.367 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-pytz-2021.1-5.el9.noarch 84/102 2026-03-09T20:57:03.367 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-repoze-lru-0.7-16.el9.noarch 85/102 2026-03-09T20:57:03.367 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-requests-2.25.1-10.el9.noarch 86/102 2026-03-09T20:57:03.367 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-requests-oauthlib-1.3.0-12.el9.noarch 87/102 2026-03-09T20:57:03.367 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-routes-2.5.1-5.el9.noarch 88/102 2026-03-09T20:57:03.367 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-rsa-4.9-2.el9.noarch 89/102 2026-03-09T20:57:03.367 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-scipy-1.9.3-2.el9.x86_64 90/102 2026-03-09T20:57:03.367 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-tempora-5.0.0-2.el9.noarch 91/102 2026-03-09T20:57:03.367 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-toml-0.10.2-6.el9.noarch 92/102 2026-03-09T20:57:03.367 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-typing-extensions-4.15.0-1.el9.noarch 93/102 2026-03-09T20:57:03.367 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-urllib3-1.26.5-7.el9.noarch 94/102 2026-03-09T20:57:03.367 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-webob-1.8.8-2.el9.noarch 95/102 2026-03-09T20:57:03.367 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-websocket-client-1.2.3-2.el9.noarch 96/102 2026-03-09T20:57:03.367 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-werkzeug-2.0.3-3.el9.1.noarch 97/102 2026-03-09T20:57:03.367 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-zc-lockfile-2.0-10.el9.noarch 98/102 2026-03-09T20:57:03.367 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : qatlib-25.08.0-2.el9.x86_64 99/102 2026-03-09T20:57:03.367 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : qatlib-service-25.08.0-2.el9.x86_64 100/102 2026-03-09T20:57:03.367 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : qatzip-libs-1.3.1-1.el9.x86_64 101/102 2026-03-09T20:57:03.373 INFO:teuthology.orchestra.run.vm01.stdout: Running scriptlet: libcephsqlite-2:19.2.3-678.ge911bdeb.el9.x86_64 102/102 2026-03-09T20:57:03.373 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : abseil-cpp-20211102.0-4.el9.x86_64 1/102 2026-03-09T20:57:03.373 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : ceph-base-2:19.2.3-678.ge911bdeb.el9.x86_64 2/102 2026-03-09T20:57:03.373 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : ceph-common-2:19.2.3-678.ge911bdeb.el9.x86_64 3/102 2026-03-09T20:57:03.374 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : ceph-grafana-dashboards-2:19.2.3-678.ge911bdeb.e 4/102 2026-03-09T20:57:03.374 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : ceph-immutable-object-cache-2:19.2.3-678.ge911bd 5/102 2026-03-09T20:57:03.374 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : ceph-mgr-2:19.2.3-678.ge911bdeb.el9.x86_64 6/102 2026-03-09T20:57:03.374 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : ceph-mgr-cephadm-2:19.2.3-678.ge911bdeb.el9.noar 7/102 2026-03-09T20:57:03.374 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : ceph-mgr-dashboard-2:19.2.3-678.ge911bdeb.el9.no 8/102 2026-03-09T20:57:03.374 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : ceph-mgr-diskprediction-local-2:19.2.3-678.ge911 9/102 2026-03-09T20:57:03.375 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : ceph-mgr-modules-core-2:19.2.3-678.ge911bdeb.el9 10/102 2026-03-09T20:57:03.375 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : ceph-mgr-rook-2:19.2.3-678.ge911bdeb.el9.noarch 11/102 2026-03-09T20:57:03.375 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : ceph-osd-2:19.2.3-678.ge911bdeb.el9.x86_64 12/102 2026-03-09T20:57:03.375 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : ceph-prometheus-alerts-2:19.2.3-678.ge911bdeb.el 13/102 2026-03-09T20:57:03.375 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : ceph-selinux-2:19.2.3-678.ge911bdeb.el9.x86_64 14/102 2026-03-09T20:57:03.375 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : ceph-volume-2:19.2.3-678.ge911bdeb.el9.noarch 15/102 2026-03-09T20:57:03.375 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : cryptsetup-2.8.1-3.el9.x86_64 16/102 2026-03-09T20:57:03.375 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : flexiblas-3.0.4-9.el9.x86_64 17/102 2026-03-09T20:57:03.375 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : flexiblas-netlib-3.0.4-9.el9.x86_64 18/102 2026-03-09T20:57:03.375 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : flexiblas-openblas-openmp-3.0.4-9.el9.x86_64 19/102 2026-03-09T20:57:03.375 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : gperftools-libs-2.9.1-3.el9.x86_64 20/102 2026-03-09T20:57:03.375 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : grpc-data-1.46.7-10.el9.noarch 21/102 2026-03-09T20:57:03.375 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : ledmon-libs-1.1.0-3.el9.x86_64 22/102 2026-03-09T20:57:03.375 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : libcephsqlite-2:19.2.3-678.ge911bdeb.el9.x86_64 23/102 2026-03-09T20:57:03.375 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : libconfig-1.7.2-9.el9.x86_64 24/102 2026-03-09T20:57:03.375 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : libgfortran-11.5.0-14.el9.x86_64 25/102 2026-03-09T20:57:03.375 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : liboath-2.6.12-1.el9.x86_64 26/102 2026-03-09T20:57:03.376 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : libquadmath-11.5.0-14.el9.x86_64 27/102 2026-03-09T20:57:03.376 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : libradosstriper1-2:19.2.3-678.ge911bdeb.el9.x86_ 28/102 2026-03-09T20:57:03.376 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : libstoragemgmt-1.10.1-1.el9.x86_64 29/102 2026-03-09T20:57:03.376 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : libunwind-1.6.2-1.el9.x86_64 30/102 2026-03-09T20:57:03.376 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : openblas-0.3.29-1.el9.x86_64 31/102 2026-03-09T20:57:03.376 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : openblas-openmp-0.3.29-1.el9.x86_64 32/102 2026-03-09T20:57:03.376 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : pciutils-3.7.0-7.el9.x86_64 33/102 2026-03-09T20:57:03.376 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : protobuf-3.14.0-17.el9.x86_64 34/102 2026-03-09T20:57:03.376 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : protobuf-compiler-3.14.0-17.el9.x86_64 35/102 2026-03-09T20:57:03.376 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-asyncssh-2.13.2-5.el9.noarch 36/102 2026-03-09T20:57:03.376 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-autocommand-2.2.2-8.el9.noarch 37/102 2026-03-09T20:57:03.376 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-babel-2.9.1-2.el9.noarch 38/102 2026-03-09T20:57:03.376 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-backports-tarfile-1.2.0-1.el9.noarch 39/102 2026-03-09T20:57:03.376 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-bcrypt-3.2.2-1.el9.x86_64 40/102 2026-03-09T20:57:03.376 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-cachetools-4.2.4-1.el9.noarch 41/102 2026-03-09T20:57:03.376 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-ceph-common-2:19.2.3-678.ge911bdeb.el9.x 42/102 2026-03-09T20:57:03.376 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-certifi-2023.05.07-4.el9.noarch 43/102 2026-03-09T20:57:03.376 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-cffi-1.14.5-5.el9.x86_64 44/102 2026-03-09T20:57:03.376 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-chardet-4.0.0-5.el9.noarch 45/102 2026-03-09T20:57:03.376 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-cheroot-10.0.1-4.el9.noarch 46/102 2026-03-09T20:57:03.376 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-cherrypy-18.6.1-2.el9.noarch 47/102 2026-03-09T20:57:03.376 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-cryptography-36.0.1-5.el9.x86_64 48/102 2026-03-09T20:57:03.376 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-devel-3.9.25-3.el9.x86_64 49/102 2026-03-09T20:57:03.376 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-google-auth-1:2.45.0-1.el9.noarch 50/102 2026-03-09T20:57:03.376 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-grpcio-1.46.7-10.el9.x86_64 51/102 2026-03-09T20:57:03.376 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-grpcio-tools-1.46.7-10.el9.x86_64 52/102 2026-03-09T20:57:03.376 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-idna-2.10-7.el9.1.noarch 53/102 2026-03-09T20:57:03.376 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-jaraco-8.2.1-3.el9.noarch 54/102 2026-03-09T20:57:03.376 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-jaraco-classes-3.2.1-5.el9.noarch 55/102 2026-03-09T20:57:03.376 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-jaraco-collections-3.0.0-8.el9.noarch 56/102 2026-03-09T20:57:03.376 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-jaraco-context-6.0.1-3.el9.noarch 57/102 2026-03-09T20:57:03.376 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-jaraco-functools-3.5.0-2.el9.noarch 58/102 2026-03-09T20:57:03.376 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-jaraco-text-4.0.0-2.el9.noarch 59/102 2026-03-09T20:57:03.376 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-jinja2-2.11.3-8.el9.noarch 60/102 2026-03-09T20:57:03.376 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-jsonpatch-1.21-16.el9.noarch 61/102 2026-03-09T20:57:03.376 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-jsonpointer-2.0-4.el9.noarch 62/102 2026-03-09T20:57:03.376 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-kubernetes-1:26.1.0-3.el9.noarch 63/102 2026-03-09T20:57:03.376 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-libstoragemgmt-1.10.1-1.el9.x86_64 64/102 2026-03-09T20:57:03.376 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-logutils-0.3.5-21.el9.noarch 65/102 2026-03-09T20:57:03.376 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-mako-1.1.4-6.el9.noarch 66/102 2026-03-09T20:57:03.376 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-markupsafe-1.1.1-12.el9.x86_64 67/102 2026-03-09T20:57:03.377 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-more-itertools-8.12.0-2.el9.noarch 68/102 2026-03-09T20:57:03.377 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-natsort-7.1.1-5.el9.noarch 69/102 2026-03-09T20:57:03.377 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-numpy-1:1.23.5-2.el9.x86_64 70/102 2026-03-09T20:57:03.377 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-numpy-f2py-1:1.23.5-2.el9.x86_64 71/102 2026-03-09T20:57:03.377 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-oauthlib-3.1.1-5.el9.noarch 72/102 2026-03-09T20:57:03.377 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-packaging-20.9-5.el9.noarch 73/102 2026-03-09T20:57:03.377 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-pecan-1.4.2-3.el9.noarch 74/102 2026-03-09T20:57:03.377 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-ply-3.11-14.el9.noarch 75/102 2026-03-09T20:57:03.377 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-portend-3.1.0-2.el9.noarch 76/102 2026-03-09T20:57:03.377 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-prettytable-0.7.2-27.el9.noarch 77/102 2026-03-09T20:57:03.377 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-protobuf-3.14.0-17.el9.noarch 78/102 2026-03-09T20:57:03.377 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-pyOpenSSL-21.0.0-1.el9.noarch 79/102 2026-03-09T20:57:03.377 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-pyasn1-0.4.8-7.el9.noarch 80/102 2026-03-09T20:57:03.377 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-pyasn1-modules-0.4.8-7.el9.noarch 81/102 2026-03-09T20:57:03.377 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-pycparser-2.20-6.el9.noarch 82/102 2026-03-09T20:57:03.377 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-pysocks-1.7.1-12.el9.noarch 83/102 2026-03-09T20:57:03.377 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-pytz-2021.1-5.el9.noarch 84/102 2026-03-09T20:57:03.377 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-repoze-lru-0.7-16.el9.noarch 85/102 2026-03-09T20:57:03.377 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-requests-2.25.1-10.el9.noarch 86/102 2026-03-09T20:57:03.377 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-requests-oauthlib-1.3.0-12.el9.noarch 87/102 2026-03-09T20:57:03.377 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-routes-2.5.1-5.el9.noarch 88/102 2026-03-09T20:57:03.377 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-rsa-4.9-2.el9.noarch 89/102 2026-03-09T20:57:03.377 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-scipy-1.9.3-2.el9.x86_64 90/102 2026-03-09T20:57:03.377 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-tempora-5.0.0-2.el9.noarch 91/102 2026-03-09T20:57:03.377 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-toml-0.10.2-6.el9.noarch 92/102 2026-03-09T20:57:03.377 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-typing-extensions-4.15.0-1.el9.noarch 93/102 2026-03-09T20:57:03.377 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-urllib3-1.26.5-7.el9.noarch 94/102 2026-03-09T20:57:03.377 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-webob-1.8.8-2.el9.noarch 95/102 2026-03-09T20:57:03.377 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-websocket-client-1.2.3-2.el9.noarch 96/102 2026-03-09T20:57:03.377 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-werkzeug-2.0.3-3.el9.1.noarch 97/102 2026-03-09T20:57:03.377 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-zc-lockfile-2.0-10.el9.noarch 98/102 2026-03-09T20:57:03.377 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : qatlib-25.08.0-2.el9.x86_64 99/102 2026-03-09T20:57:03.377 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : qatlib-service-25.08.0-2.el9.x86_64 100/102 2026-03-09T20:57:03.378 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : qatzip-libs-1.3.1-1.el9.x86_64 101/102 2026-03-09T20:57:03.403 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : rbd-mirror-2:19.2.3-678.ge911bdeb.el9.x86_64 102/102 2026-03-09T20:57:03.403 INFO:teuthology.orchestra.run.vm08.stdout: 2026-03-09T20:57:03.403 INFO:teuthology.orchestra.run.vm08.stdout:Removed: 2026-03-09T20:57:03.403 INFO:teuthology.orchestra.run.vm08.stdout: abseil-cpp-20211102.0-4.el9.x86_64 2026-03-09T20:57:03.403 INFO:teuthology.orchestra.run.vm08.stdout: ceph-base-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T20:57:03.403 INFO:teuthology.orchestra.run.vm08.stdout: ceph-common-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T20:57:03.403 INFO:teuthology.orchestra.run.vm08.stdout: ceph-grafana-dashboards-2:19.2.3-678.ge911bdeb.el9.noarch 2026-03-09T20:57:03.404 INFO:teuthology.orchestra.run.vm08.stdout: ceph-immutable-object-cache-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T20:57:03.404 INFO:teuthology.orchestra.run.vm08.stdout: ceph-mgr-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T20:57:03.404 INFO:teuthology.orchestra.run.vm08.stdout: ceph-mgr-cephadm-2:19.2.3-678.ge911bdeb.el9.noarch 2026-03-09T20:57:03.404 INFO:teuthology.orchestra.run.vm08.stdout: ceph-mgr-dashboard-2:19.2.3-678.ge911bdeb.el9.noarch 2026-03-09T20:57:03.404 INFO:teuthology.orchestra.run.vm08.stdout: ceph-mgr-diskprediction-local-2:19.2.3-678.ge911bdeb.el9.noarch 2026-03-09T20:57:03.404 INFO:teuthology.orchestra.run.vm08.stdout: ceph-mgr-modules-core-2:19.2.3-678.ge911bdeb.el9.noarch 2026-03-09T20:57:03.404 INFO:teuthology.orchestra.run.vm08.stdout: ceph-mgr-rook-2:19.2.3-678.ge911bdeb.el9.noarch 2026-03-09T20:57:03.404 INFO:teuthology.orchestra.run.vm08.stdout: ceph-osd-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T20:57:03.404 INFO:teuthology.orchestra.run.vm08.stdout: ceph-prometheus-alerts-2:19.2.3-678.ge911bdeb.el9.noarch 2026-03-09T20:57:03.404 INFO:teuthology.orchestra.run.vm08.stdout: ceph-selinux-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T20:57:03.404 INFO:teuthology.orchestra.run.vm08.stdout: ceph-volume-2:19.2.3-678.ge911bdeb.el9.noarch 2026-03-09T20:57:03.404 INFO:teuthology.orchestra.run.vm08.stdout: cryptsetup-2.8.1-3.el9.x86_64 2026-03-09T20:57:03.404 INFO:teuthology.orchestra.run.vm08.stdout: flexiblas-3.0.4-9.el9.x86_64 2026-03-09T20:57:03.404 INFO:teuthology.orchestra.run.vm08.stdout: flexiblas-netlib-3.0.4-9.el9.x86_64 2026-03-09T20:57:03.404 INFO:teuthology.orchestra.run.vm08.stdout: flexiblas-openblas-openmp-3.0.4-9.el9.x86_64 2026-03-09T20:57:03.404 INFO:teuthology.orchestra.run.vm08.stdout: gperftools-libs-2.9.1-3.el9.x86_64 2026-03-09T20:57:03.404 INFO:teuthology.orchestra.run.vm08.stdout: grpc-data-1.46.7-10.el9.noarch 2026-03-09T20:57:03.404 INFO:teuthology.orchestra.run.vm08.stdout: ledmon-libs-1.1.0-3.el9.x86_64 2026-03-09T20:57:03.404 INFO:teuthology.orchestra.run.vm08.stdout: libcephsqlite-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T20:57:03.404 INFO:teuthology.orchestra.run.vm08.stdout: libconfig-1.7.2-9.el9.x86_64 2026-03-09T20:57:03.404 INFO:teuthology.orchestra.run.vm08.stdout: libgfortran-11.5.0-14.el9.x86_64 2026-03-09T20:57:03.404 INFO:teuthology.orchestra.run.vm08.stdout: liboath-2.6.12-1.el9.x86_64 2026-03-09T20:57:03.404 INFO:teuthology.orchestra.run.vm08.stdout: libquadmath-11.5.0-14.el9.x86_64 2026-03-09T20:57:03.404 INFO:teuthology.orchestra.run.vm08.stdout: libradosstriper1-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T20:57:03.404 INFO:teuthology.orchestra.run.vm08.stdout: libstoragemgmt-1.10.1-1.el9.x86_64 2026-03-09T20:57:03.404 INFO:teuthology.orchestra.run.vm08.stdout: libunwind-1.6.2-1.el9.x86_64 2026-03-09T20:57:03.404 INFO:teuthology.orchestra.run.vm08.stdout: openblas-0.3.29-1.el9.x86_64 2026-03-09T20:57:03.404 INFO:teuthology.orchestra.run.vm08.stdout: openblas-openmp-0.3.29-1.el9.x86_64 2026-03-09T20:57:03.404 INFO:teuthology.orchestra.run.vm08.stdout: pciutils-3.7.0-7.el9.x86_64 2026-03-09T20:57:03.404 INFO:teuthology.orchestra.run.vm08.stdout: protobuf-3.14.0-17.el9.x86_64 2026-03-09T20:57:03.404 INFO:teuthology.orchestra.run.vm08.stdout: protobuf-compiler-3.14.0-17.el9.x86_64 2026-03-09T20:57:03.404 INFO:teuthology.orchestra.run.vm08.stdout: python3-asyncssh-2.13.2-5.el9.noarch 2026-03-09T20:57:03.404 INFO:teuthology.orchestra.run.vm08.stdout: python3-autocommand-2.2.2-8.el9.noarch 2026-03-09T20:57:03.404 INFO:teuthology.orchestra.run.vm08.stdout: python3-babel-2.9.1-2.el9.noarch 2026-03-09T20:57:03.404 INFO:teuthology.orchestra.run.vm08.stdout: python3-backports-tarfile-1.2.0-1.el9.noarch 2026-03-09T20:57:03.404 INFO:teuthology.orchestra.run.vm08.stdout: python3-bcrypt-3.2.2-1.el9.x86_64 2026-03-09T20:57:03.404 INFO:teuthology.orchestra.run.vm08.stdout: python3-cachetools-4.2.4-1.el9.noarch 2026-03-09T20:57:03.404 INFO:teuthology.orchestra.run.vm08.stdout: python3-ceph-common-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T20:57:03.404 INFO:teuthology.orchestra.run.vm08.stdout: python3-certifi-2023.05.07-4.el9.noarch 2026-03-09T20:57:03.404 INFO:teuthology.orchestra.run.vm08.stdout: python3-cffi-1.14.5-5.el9.x86_64 2026-03-09T20:57:03.404 INFO:teuthology.orchestra.run.vm08.stdout: python3-chardet-4.0.0-5.el9.noarch 2026-03-09T20:57:03.404 INFO:teuthology.orchestra.run.vm08.stdout: python3-cheroot-10.0.1-4.el9.noarch 2026-03-09T20:57:03.404 INFO:teuthology.orchestra.run.vm08.stdout: python3-cherrypy-18.6.1-2.el9.noarch 2026-03-09T20:57:03.404 INFO:teuthology.orchestra.run.vm08.stdout: python3-cryptography-36.0.1-5.el9.x86_64 2026-03-09T20:57:03.404 INFO:teuthology.orchestra.run.vm08.stdout: python3-devel-3.9.25-3.el9.x86_64 2026-03-09T20:57:03.404 INFO:teuthology.orchestra.run.vm08.stdout: python3-google-auth-1:2.45.0-1.el9.noarch 2026-03-09T20:57:03.404 INFO:teuthology.orchestra.run.vm08.stdout: python3-grpcio-1.46.7-10.el9.x86_64 2026-03-09T20:57:03.404 INFO:teuthology.orchestra.run.vm08.stdout: python3-grpcio-tools-1.46.7-10.el9.x86_64 2026-03-09T20:57:03.404 INFO:teuthology.orchestra.run.vm08.stdout: python3-idna-2.10-7.el9.1.noarch 2026-03-09T20:57:03.404 INFO:teuthology.orchestra.run.vm08.stdout: python3-jaraco-8.2.1-3.el9.noarch 2026-03-09T20:57:03.404 INFO:teuthology.orchestra.run.vm08.stdout: python3-jaraco-classes-3.2.1-5.el9.noarch 2026-03-09T20:57:03.404 INFO:teuthology.orchestra.run.vm08.stdout: python3-jaraco-collections-3.0.0-8.el9.noarch 2026-03-09T20:57:03.404 INFO:teuthology.orchestra.run.vm08.stdout: python3-jaraco-context-6.0.1-3.el9.noarch 2026-03-09T20:57:03.404 INFO:teuthology.orchestra.run.vm08.stdout: python3-jaraco-functools-3.5.0-2.el9.noarch 2026-03-09T20:57:03.404 INFO:teuthology.orchestra.run.vm08.stdout: python3-jaraco-text-4.0.0-2.el9.noarch 2026-03-09T20:57:03.404 INFO:teuthology.orchestra.run.vm08.stdout: python3-jinja2-2.11.3-8.el9.noarch 2026-03-09T20:57:03.404 INFO:teuthology.orchestra.run.vm08.stdout: python3-jsonpatch-1.21-16.el9.noarch 2026-03-09T20:57:03.404 INFO:teuthology.orchestra.run.vm08.stdout: python3-jsonpointer-2.0-4.el9.noarch 2026-03-09T20:57:03.404 INFO:teuthology.orchestra.run.vm08.stdout: python3-kubernetes-1:26.1.0-3.el9.noarch 2026-03-09T20:57:03.405 INFO:teuthology.orchestra.run.vm08.stdout: python3-libstoragemgmt-1.10.1-1.el9.x86_64 2026-03-09T20:57:03.405 INFO:teuthology.orchestra.run.vm08.stdout: python3-logutils-0.3.5-21.el9.noarch 2026-03-09T20:57:03.405 INFO:teuthology.orchestra.run.vm08.stdout: python3-mako-1.1.4-6.el9.noarch 2026-03-09T20:57:03.405 INFO:teuthology.orchestra.run.vm08.stdout: python3-markupsafe-1.1.1-12.el9.x86_64 2026-03-09T20:57:03.405 INFO:teuthology.orchestra.run.vm08.stdout: python3-more-itertools-8.12.0-2.el9.noarch 2026-03-09T20:57:03.405 INFO:teuthology.orchestra.run.vm08.stdout: python3-natsort-7.1.1-5.el9.noarch 2026-03-09T20:57:03.405 INFO:teuthology.orchestra.run.vm08.stdout: python3-numpy-1:1.23.5-2.el9.x86_64 2026-03-09T20:57:03.405 INFO:teuthology.orchestra.run.vm08.stdout: python3-numpy-f2py-1:1.23.5-2.el9.x86_64 2026-03-09T20:57:03.405 INFO:teuthology.orchestra.run.vm08.stdout: python3-oauthlib-3.1.1-5.el9.noarch 2026-03-09T20:57:03.405 INFO:teuthology.orchestra.run.vm08.stdout: python3-packaging-20.9-5.el9.noarch 2026-03-09T20:57:03.405 INFO:teuthology.orchestra.run.vm08.stdout: python3-pecan-1.4.2-3.el9.noarch 2026-03-09T20:57:03.405 INFO:teuthology.orchestra.run.vm08.stdout: python3-ply-3.11-14.el9.noarch 2026-03-09T20:57:03.405 INFO:teuthology.orchestra.run.vm08.stdout: python3-portend-3.1.0-2.el9.noarch 2026-03-09T20:57:03.405 INFO:teuthology.orchestra.run.vm08.stdout: python3-prettytable-0.7.2-27.el9.noarch 2026-03-09T20:57:03.405 INFO:teuthology.orchestra.run.vm08.stdout: python3-protobuf-3.14.0-17.el9.noarch 2026-03-09T20:57:03.405 INFO:teuthology.orchestra.run.vm08.stdout: python3-pyOpenSSL-21.0.0-1.el9.noarch 2026-03-09T20:57:03.405 INFO:teuthology.orchestra.run.vm08.stdout: python3-pyasn1-0.4.8-7.el9.noarch 2026-03-09T20:57:03.405 INFO:teuthology.orchestra.run.vm08.stdout: python3-pyasn1-modules-0.4.8-7.el9.noarch 2026-03-09T20:57:03.405 INFO:teuthology.orchestra.run.vm08.stdout: python3-pycparser-2.20-6.el9.noarch 2026-03-09T20:57:03.405 INFO:teuthology.orchestra.run.vm08.stdout: python3-pysocks-1.7.1-12.el9.noarch 2026-03-09T20:57:03.405 INFO:teuthology.orchestra.run.vm08.stdout: python3-pytz-2021.1-5.el9.noarch 2026-03-09T20:57:03.405 INFO:teuthology.orchestra.run.vm08.stdout: python3-repoze-lru-0.7-16.el9.noarch 2026-03-09T20:57:03.405 INFO:teuthology.orchestra.run.vm08.stdout: python3-requests-2.25.1-10.el9.noarch 2026-03-09T20:57:03.405 INFO:teuthology.orchestra.run.vm08.stdout: python3-requests-oauthlib-1.3.0-12.el9.noarch 2026-03-09T20:57:03.405 INFO:teuthology.orchestra.run.vm08.stdout: python3-routes-2.5.1-5.el9.noarch 2026-03-09T20:57:03.405 INFO:teuthology.orchestra.run.vm08.stdout: python3-rsa-4.9-2.el9.noarch 2026-03-09T20:57:03.405 INFO:teuthology.orchestra.run.vm08.stdout: python3-scipy-1.9.3-2.el9.x86_64 2026-03-09T20:57:03.405 INFO:teuthology.orchestra.run.vm08.stdout: python3-tempora-5.0.0-2.el9.noarch 2026-03-09T20:57:03.405 INFO:teuthology.orchestra.run.vm08.stdout: python3-toml-0.10.2-6.el9.noarch 2026-03-09T20:57:03.405 INFO:teuthology.orchestra.run.vm08.stdout: python3-typing-extensions-4.15.0-1.el9.noarch 2026-03-09T20:57:03.405 INFO:teuthology.orchestra.run.vm08.stdout: python3-urllib3-1.26.5-7.el9.noarch 2026-03-09T20:57:03.405 INFO:teuthology.orchestra.run.vm08.stdout: python3-webob-1.8.8-2.el9.noarch 2026-03-09T20:57:03.405 INFO:teuthology.orchestra.run.vm08.stdout: python3-websocket-client-1.2.3-2.el9.noarch 2026-03-09T20:57:03.405 INFO:teuthology.orchestra.run.vm08.stdout: python3-werkzeug-2.0.3-3.el9.1.noarch 2026-03-09T20:57:03.405 INFO:teuthology.orchestra.run.vm08.stdout: python3-zc-lockfile-2.0-10.el9.noarch 2026-03-09T20:57:03.405 INFO:teuthology.orchestra.run.vm08.stdout: qatlib-25.08.0-2.el9.x86_64 2026-03-09T20:57:03.405 INFO:teuthology.orchestra.run.vm08.stdout: qatlib-service-25.08.0-2.el9.x86_64 2026-03-09T20:57:03.405 INFO:teuthology.orchestra.run.vm08.stdout: qatzip-libs-1.3.1-1.el9.x86_64 2026-03-09T20:57:03.405 INFO:teuthology.orchestra.run.vm08.stdout: rbd-mirror-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T20:57:03.405 INFO:teuthology.orchestra.run.vm08.stdout: 2026-03-09T20:57:03.405 INFO:teuthology.orchestra.run.vm08.stdout:Complete! 2026-03-09T20:57:03.447 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : rbd-mirror-2:19.2.3-678.ge911bdeb.el9.x86_64 102/102 2026-03-09T20:57:03.448 INFO:teuthology.orchestra.run.vm06.stdout: 2026-03-09T20:57:03.448 INFO:teuthology.orchestra.run.vm06.stdout:Removed: 2026-03-09T20:57:03.448 INFO:teuthology.orchestra.run.vm06.stdout: abseil-cpp-20211102.0-4.el9.x86_64 2026-03-09T20:57:03.448 INFO:teuthology.orchestra.run.vm06.stdout: ceph-base-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T20:57:03.448 INFO:teuthology.orchestra.run.vm06.stdout: ceph-common-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T20:57:03.448 INFO:teuthology.orchestra.run.vm06.stdout: ceph-grafana-dashboards-2:19.2.3-678.ge911bdeb.el9.noarch 2026-03-09T20:57:03.448 INFO:teuthology.orchestra.run.vm06.stdout: ceph-immutable-object-cache-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T20:57:03.448 INFO:teuthology.orchestra.run.vm06.stdout: ceph-mgr-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T20:57:03.448 INFO:teuthology.orchestra.run.vm06.stdout: ceph-mgr-cephadm-2:19.2.3-678.ge911bdeb.el9.noarch 2026-03-09T20:57:03.448 INFO:teuthology.orchestra.run.vm06.stdout: ceph-mgr-dashboard-2:19.2.3-678.ge911bdeb.el9.noarch 2026-03-09T20:57:03.448 INFO:teuthology.orchestra.run.vm06.stdout: ceph-mgr-diskprediction-local-2:19.2.3-678.ge911bdeb.el9.noarch 2026-03-09T20:57:03.448 INFO:teuthology.orchestra.run.vm06.stdout: ceph-mgr-modules-core-2:19.2.3-678.ge911bdeb.el9.noarch 2026-03-09T20:57:03.448 INFO:teuthology.orchestra.run.vm06.stdout: ceph-mgr-rook-2:19.2.3-678.ge911bdeb.el9.noarch 2026-03-09T20:57:03.448 INFO:teuthology.orchestra.run.vm06.stdout: ceph-osd-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T20:57:03.448 INFO:teuthology.orchestra.run.vm06.stdout: ceph-prometheus-alerts-2:19.2.3-678.ge911bdeb.el9.noarch 2026-03-09T20:57:03.448 INFO:teuthology.orchestra.run.vm06.stdout: ceph-selinux-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T20:57:03.448 INFO:teuthology.orchestra.run.vm06.stdout: ceph-volume-2:19.2.3-678.ge911bdeb.el9.noarch 2026-03-09T20:57:03.448 INFO:teuthology.orchestra.run.vm06.stdout: cryptsetup-2.8.1-3.el9.x86_64 2026-03-09T20:57:03.448 INFO:teuthology.orchestra.run.vm06.stdout: flexiblas-3.0.4-9.el9.x86_64 2026-03-09T20:57:03.448 INFO:teuthology.orchestra.run.vm06.stdout: flexiblas-netlib-3.0.4-9.el9.x86_64 2026-03-09T20:57:03.448 INFO:teuthology.orchestra.run.vm06.stdout: flexiblas-openblas-openmp-3.0.4-9.el9.x86_64 2026-03-09T20:57:03.448 INFO:teuthology.orchestra.run.vm06.stdout: gperftools-libs-2.9.1-3.el9.x86_64 2026-03-09T20:57:03.448 INFO:teuthology.orchestra.run.vm06.stdout: grpc-data-1.46.7-10.el9.noarch 2026-03-09T20:57:03.448 INFO:teuthology.orchestra.run.vm06.stdout: ledmon-libs-1.1.0-3.el9.x86_64 2026-03-09T20:57:03.448 INFO:teuthology.orchestra.run.vm06.stdout: libcephsqlite-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T20:57:03.448 INFO:teuthology.orchestra.run.vm06.stdout: libconfig-1.7.2-9.el9.x86_64 2026-03-09T20:57:03.448 INFO:teuthology.orchestra.run.vm06.stdout: libgfortran-11.5.0-14.el9.x86_64 2026-03-09T20:57:03.448 INFO:teuthology.orchestra.run.vm06.stdout: liboath-2.6.12-1.el9.x86_64 2026-03-09T20:57:03.448 INFO:teuthology.orchestra.run.vm06.stdout: libquadmath-11.5.0-14.el9.x86_64 2026-03-09T20:57:03.448 INFO:teuthology.orchestra.run.vm06.stdout: libradosstriper1-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T20:57:03.448 INFO:teuthology.orchestra.run.vm06.stdout: libstoragemgmt-1.10.1-1.el9.x86_64 2026-03-09T20:57:03.448 INFO:teuthology.orchestra.run.vm06.stdout: libunwind-1.6.2-1.el9.x86_64 2026-03-09T20:57:03.448 INFO:teuthology.orchestra.run.vm06.stdout: openblas-0.3.29-1.el9.x86_64 2026-03-09T20:57:03.448 INFO:teuthology.orchestra.run.vm06.stdout: openblas-openmp-0.3.29-1.el9.x86_64 2026-03-09T20:57:03.448 INFO:teuthology.orchestra.run.vm06.stdout: pciutils-3.7.0-7.el9.x86_64 2026-03-09T20:57:03.448 INFO:teuthology.orchestra.run.vm06.stdout: protobuf-3.14.0-17.el9.x86_64 2026-03-09T20:57:03.448 INFO:teuthology.orchestra.run.vm06.stdout: protobuf-compiler-3.14.0-17.el9.x86_64 2026-03-09T20:57:03.448 INFO:teuthology.orchestra.run.vm06.stdout: python3-asyncssh-2.13.2-5.el9.noarch 2026-03-09T20:57:03.448 INFO:teuthology.orchestra.run.vm06.stdout: python3-autocommand-2.2.2-8.el9.noarch 2026-03-09T20:57:03.448 INFO:teuthology.orchestra.run.vm06.stdout: python3-babel-2.9.1-2.el9.noarch 2026-03-09T20:57:03.448 INFO:teuthology.orchestra.run.vm06.stdout: python3-backports-tarfile-1.2.0-1.el9.noarch 2026-03-09T20:57:03.448 INFO:teuthology.orchestra.run.vm06.stdout: python3-bcrypt-3.2.2-1.el9.x86_64 2026-03-09T20:57:03.448 INFO:teuthology.orchestra.run.vm06.stdout: python3-cachetools-4.2.4-1.el9.noarch 2026-03-09T20:57:03.448 INFO:teuthology.orchestra.run.vm06.stdout: python3-ceph-common-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T20:57:03.448 INFO:teuthology.orchestra.run.vm06.stdout: python3-certifi-2023.05.07-4.el9.noarch 2026-03-09T20:57:03.448 INFO:teuthology.orchestra.run.vm06.stdout: python3-cffi-1.14.5-5.el9.x86_64 2026-03-09T20:57:03.448 INFO:teuthology.orchestra.run.vm06.stdout: python3-chardet-4.0.0-5.el9.noarch 2026-03-09T20:57:03.448 INFO:teuthology.orchestra.run.vm06.stdout: python3-cheroot-10.0.1-4.el9.noarch 2026-03-09T20:57:03.448 INFO:teuthology.orchestra.run.vm06.stdout: python3-cherrypy-18.6.1-2.el9.noarch 2026-03-09T20:57:03.448 INFO:teuthology.orchestra.run.vm06.stdout: python3-cryptography-36.0.1-5.el9.x86_64 2026-03-09T20:57:03.448 INFO:teuthology.orchestra.run.vm06.stdout: python3-devel-3.9.25-3.el9.x86_64 2026-03-09T20:57:03.448 INFO:teuthology.orchestra.run.vm06.stdout: python3-google-auth-1:2.45.0-1.el9.noarch 2026-03-09T20:57:03.448 INFO:teuthology.orchestra.run.vm06.stdout: python3-grpcio-1.46.7-10.el9.x86_64 2026-03-09T20:57:03.448 INFO:teuthology.orchestra.run.vm06.stdout: python3-grpcio-tools-1.46.7-10.el9.x86_64 2026-03-09T20:57:03.448 INFO:teuthology.orchestra.run.vm06.stdout: python3-idna-2.10-7.el9.1.noarch 2026-03-09T20:57:03.448 INFO:teuthology.orchestra.run.vm06.stdout: python3-jaraco-8.2.1-3.el9.noarch 2026-03-09T20:57:03.448 INFO:teuthology.orchestra.run.vm06.stdout: python3-jaraco-classes-3.2.1-5.el9.noarch 2026-03-09T20:57:03.449 INFO:teuthology.orchestra.run.vm06.stdout: python3-jaraco-collections-3.0.0-8.el9.noarch 2026-03-09T20:57:03.449 INFO:teuthology.orchestra.run.vm06.stdout: python3-jaraco-context-6.0.1-3.el9.noarch 2026-03-09T20:57:03.449 INFO:teuthology.orchestra.run.vm06.stdout: python3-jaraco-functools-3.5.0-2.el9.noarch 2026-03-09T20:57:03.449 INFO:teuthology.orchestra.run.vm06.stdout: python3-jaraco-text-4.0.0-2.el9.noarch 2026-03-09T20:57:03.449 INFO:teuthology.orchestra.run.vm06.stdout: python3-jinja2-2.11.3-8.el9.noarch 2026-03-09T20:57:03.449 INFO:teuthology.orchestra.run.vm06.stdout: python3-jsonpatch-1.21-16.el9.noarch 2026-03-09T20:57:03.449 INFO:teuthology.orchestra.run.vm06.stdout: python3-jsonpointer-2.0-4.el9.noarch 2026-03-09T20:57:03.449 INFO:teuthology.orchestra.run.vm06.stdout: python3-kubernetes-1:26.1.0-3.el9.noarch 2026-03-09T20:57:03.449 INFO:teuthology.orchestra.run.vm06.stdout: python3-libstoragemgmt-1.10.1-1.el9.x86_64 2026-03-09T20:57:03.449 INFO:teuthology.orchestra.run.vm06.stdout: python3-logutils-0.3.5-21.el9.noarch 2026-03-09T20:57:03.449 INFO:teuthology.orchestra.run.vm06.stdout: python3-mako-1.1.4-6.el9.noarch 2026-03-09T20:57:03.449 INFO:teuthology.orchestra.run.vm06.stdout: python3-markupsafe-1.1.1-12.el9.x86_64 2026-03-09T20:57:03.449 INFO:teuthology.orchestra.run.vm06.stdout: python3-more-itertools-8.12.0-2.el9.noarch 2026-03-09T20:57:03.449 INFO:teuthology.orchestra.run.vm06.stdout: python3-natsort-7.1.1-5.el9.noarch 2026-03-09T20:57:03.449 INFO:teuthology.orchestra.run.vm06.stdout: python3-numpy-1:1.23.5-2.el9.x86_64 2026-03-09T20:57:03.449 INFO:teuthology.orchestra.run.vm06.stdout: python3-numpy-f2py-1:1.23.5-2.el9.x86_64 2026-03-09T20:57:03.449 INFO:teuthology.orchestra.run.vm06.stdout: python3-oauthlib-3.1.1-5.el9.noarch 2026-03-09T20:57:03.449 INFO:teuthology.orchestra.run.vm06.stdout: python3-packaging-20.9-5.el9.noarch 2026-03-09T20:57:03.449 INFO:teuthology.orchestra.run.vm06.stdout: python3-pecan-1.4.2-3.el9.noarch 2026-03-09T20:57:03.449 INFO:teuthology.orchestra.run.vm06.stdout: python3-ply-3.11-14.el9.noarch 2026-03-09T20:57:03.449 INFO:teuthology.orchestra.run.vm06.stdout: python3-portend-3.1.0-2.el9.noarch 2026-03-09T20:57:03.449 INFO:teuthology.orchestra.run.vm06.stdout: python3-prettytable-0.7.2-27.el9.noarch 2026-03-09T20:57:03.449 INFO:teuthology.orchestra.run.vm06.stdout: python3-protobuf-3.14.0-17.el9.noarch 2026-03-09T20:57:03.449 INFO:teuthology.orchestra.run.vm06.stdout: python3-pyOpenSSL-21.0.0-1.el9.noarch 2026-03-09T20:57:03.449 INFO:teuthology.orchestra.run.vm06.stdout: python3-pyasn1-0.4.8-7.el9.noarch 2026-03-09T20:57:03.449 INFO:teuthology.orchestra.run.vm06.stdout: python3-pyasn1-modules-0.4.8-7.el9.noarch 2026-03-09T20:57:03.449 INFO:teuthology.orchestra.run.vm06.stdout: python3-pycparser-2.20-6.el9.noarch 2026-03-09T20:57:03.449 INFO:teuthology.orchestra.run.vm06.stdout: python3-pysocks-1.7.1-12.el9.noarch 2026-03-09T20:57:03.449 INFO:teuthology.orchestra.run.vm06.stdout: python3-pytz-2021.1-5.el9.noarch 2026-03-09T20:57:03.449 INFO:teuthology.orchestra.run.vm06.stdout: python3-repoze-lru-0.7-16.el9.noarch 2026-03-09T20:57:03.449 INFO:teuthology.orchestra.run.vm06.stdout: python3-requests-2.25.1-10.el9.noarch 2026-03-09T20:57:03.449 INFO:teuthology.orchestra.run.vm06.stdout: python3-requests-oauthlib-1.3.0-12.el9.noarch 2026-03-09T20:57:03.449 INFO:teuthology.orchestra.run.vm06.stdout: python3-routes-2.5.1-5.el9.noarch 2026-03-09T20:57:03.449 INFO:teuthology.orchestra.run.vm06.stdout: python3-rsa-4.9-2.el9.noarch 2026-03-09T20:57:03.449 INFO:teuthology.orchestra.run.vm06.stdout: python3-scipy-1.9.3-2.el9.x86_64 2026-03-09T20:57:03.449 INFO:teuthology.orchestra.run.vm06.stdout: python3-tempora-5.0.0-2.el9.noarch 2026-03-09T20:57:03.449 INFO:teuthology.orchestra.run.vm06.stdout: python3-toml-0.10.2-6.el9.noarch 2026-03-09T20:57:03.449 INFO:teuthology.orchestra.run.vm06.stdout: python3-typing-extensions-4.15.0-1.el9.noarch 2026-03-09T20:57:03.449 INFO:teuthology.orchestra.run.vm06.stdout: python3-urllib3-1.26.5-7.el9.noarch 2026-03-09T20:57:03.449 INFO:teuthology.orchestra.run.vm06.stdout: python3-webob-1.8.8-2.el9.noarch 2026-03-09T20:57:03.449 INFO:teuthology.orchestra.run.vm06.stdout: python3-websocket-client-1.2.3-2.el9.noarch 2026-03-09T20:57:03.449 INFO:teuthology.orchestra.run.vm06.stdout: python3-werkzeug-2.0.3-3.el9.1.noarch 2026-03-09T20:57:03.449 INFO:teuthology.orchestra.run.vm06.stdout: python3-zc-lockfile-2.0-10.el9.noarch 2026-03-09T20:57:03.449 INFO:teuthology.orchestra.run.vm06.stdout: qatlib-25.08.0-2.el9.x86_64 2026-03-09T20:57:03.449 INFO:teuthology.orchestra.run.vm06.stdout: qatlib-service-25.08.0-2.el9.x86_64 2026-03-09T20:57:03.449 INFO:teuthology.orchestra.run.vm06.stdout: qatzip-libs-1.3.1-1.el9.x86_64 2026-03-09T20:57:03.449 INFO:teuthology.orchestra.run.vm06.stdout: rbd-mirror-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T20:57:03.449 INFO:teuthology.orchestra.run.vm06.stdout: 2026-03-09T20:57:03.449 INFO:teuthology.orchestra.run.vm06.stdout:Complete! 2026-03-09T20:57:03.468 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : rbd-mirror-2:19.2.3-678.ge911bdeb.el9.x86_64 102/102 2026-03-09T20:57:03.469 INFO:teuthology.orchestra.run.vm01.stdout: 2026-03-09T20:57:03.469 INFO:teuthology.orchestra.run.vm01.stdout:Removed: 2026-03-09T20:57:03.469 INFO:teuthology.orchestra.run.vm01.stdout: abseil-cpp-20211102.0-4.el9.x86_64 2026-03-09T20:57:03.469 INFO:teuthology.orchestra.run.vm01.stdout: ceph-base-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T20:57:03.469 INFO:teuthology.orchestra.run.vm01.stdout: ceph-common-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T20:57:03.469 INFO:teuthology.orchestra.run.vm01.stdout: ceph-grafana-dashboards-2:19.2.3-678.ge911bdeb.el9.noarch 2026-03-09T20:57:03.469 INFO:teuthology.orchestra.run.vm01.stdout: ceph-immutable-object-cache-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T20:57:03.469 INFO:teuthology.orchestra.run.vm01.stdout: ceph-mgr-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T20:57:03.469 INFO:teuthology.orchestra.run.vm01.stdout: ceph-mgr-cephadm-2:19.2.3-678.ge911bdeb.el9.noarch 2026-03-09T20:57:03.469 INFO:teuthology.orchestra.run.vm01.stdout: ceph-mgr-dashboard-2:19.2.3-678.ge911bdeb.el9.noarch 2026-03-09T20:57:03.469 INFO:teuthology.orchestra.run.vm01.stdout: ceph-mgr-diskprediction-local-2:19.2.3-678.ge911bdeb.el9.noarch 2026-03-09T20:57:03.469 INFO:teuthology.orchestra.run.vm01.stdout: ceph-mgr-modules-core-2:19.2.3-678.ge911bdeb.el9.noarch 2026-03-09T20:57:03.469 INFO:teuthology.orchestra.run.vm01.stdout: ceph-mgr-rook-2:19.2.3-678.ge911bdeb.el9.noarch 2026-03-09T20:57:03.469 INFO:teuthology.orchestra.run.vm01.stdout: ceph-osd-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T20:57:03.469 INFO:teuthology.orchestra.run.vm01.stdout: ceph-prometheus-alerts-2:19.2.3-678.ge911bdeb.el9.noarch 2026-03-09T20:57:03.469 INFO:teuthology.orchestra.run.vm01.stdout: ceph-selinux-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T20:57:03.469 INFO:teuthology.orchestra.run.vm01.stdout: ceph-volume-2:19.2.3-678.ge911bdeb.el9.noarch 2026-03-09T20:57:03.469 INFO:teuthology.orchestra.run.vm01.stdout: cryptsetup-2.8.1-3.el9.x86_64 2026-03-09T20:57:03.469 INFO:teuthology.orchestra.run.vm01.stdout: flexiblas-3.0.4-9.el9.x86_64 2026-03-09T20:57:03.469 INFO:teuthology.orchestra.run.vm01.stdout: flexiblas-netlib-3.0.4-9.el9.x86_64 2026-03-09T20:57:03.469 INFO:teuthology.orchestra.run.vm01.stdout: flexiblas-openblas-openmp-3.0.4-9.el9.x86_64 2026-03-09T20:57:03.469 INFO:teuthology.orchestra.run.vm01.stdout: gperftools-libs-2.9.1-3.el9.x86_64 2026-03-09T20:57:03.469 INFO:teuthology.orchestra.run.vm01.stdout: grpc-data-1.46.7-10.el9.noarch 2026-03-09T20:57:03.469 INFO:teuthology.orchestra.run.vm01.stdout: ledmon-libs-1.1.0-3.el9.x86_64 2026-03-09T20:57:03.469 INFO:teuthology.orchestra.run.vm01.stdout: libcephsqlite-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T20:57:03.469 INFO:teuthology.orchestra.run.vm01.stdout: libconfig-1.7.2-9.el9.x86_64 2026-03-09T20:57:03.469 INFO:teuthology.orchestra.run.vm01.stdout: libgfortran-11.5.0-14.el9.x86_64 2026-03-09T20:57:03.469 INFO:teuthology.orchestra.run.vm01.stdout: liboath-2.6.12-1.el9.x86_64 2026-03-09T20:57:03.469 INFO:teuthology.orchestra.run.vm01.stdout: libquadmath-11.5.0-14.el9.x86_64 2026-03-09T20:57:03.469 INFO:teuthology.orchestra.run.vm01.stdout: libradosstriper1-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T20:57:03.469 INFO:teuthology.orchestra.run.vm01.stdout: libstoragemgmt-1.10.1-1.el9.x86_64 2026-03-09T20:57:03.469 INFO:teuthology.orchestra.run.vm01.stdout: libunwind-1.6.2-1.el9.x86_64 2026-03-09T20:57:03.469 INFO:teuthology.orchestra.run.vm01.stdout: openblas-0.3.29-1.el9.x86_64 2026-03-09T20:57:03.469 INFO:teuthology.orchestra.run.vm01.stdout: openblas-openmp-0.3.29-1.el9.x86_64 2026-03-09T20:57:03.469 INFO:teuthology.orchestra.run.vm01.stdout: pciutils-3.7.0-7.el9.x86_64 2026-03-09T20:57:03.469 INFO:teuthology.orchestra.run.vm01.stdout: protobuf-3.14.0-17.el9.x86_64 2026-03-09T20:57:03.469 INFO:teuthology.orchestra.run.vm01.stdout: protobuf-compiler-3.14.0-17.el9.x86_64 2026-03-09T20:57:03.469 INFO:teuthology.orchestra.run.vm01.stdout: python3-asyncssh-2.13.2-5.el9.noarch 2026-03-09T20:57:03.469 INFO:teuthology.orchestra.run.vm01.stdout: python3-autocommand-2.2.2-8.el9.noarch 2026-03-09T20:57:03.469 INFO:teuthology.orchestra.run.vm01.stdout: python3-babel-2.9.1-2.el9.noarch 2026-03-09T20:57:03.469 INFO:teuthology.orchestra.run.vm01.stdout: python3-backports-tarfile-1.2.0-1.el9.noarch 2026-03-09T20:57:03.469 INFO:teuthology.orchestra.run.vm01.stdout: python3-bcrypt-3.2.2-1.el9.x86_64 2026-03-09T20:57:03.469 INFO:teuthology.orchestra.run.vm01.stdout: python3-cachetools-4.2.4-1.el9.noarch 2026-03-09T20:57:03.469 INFO:teuthology.orchestra.run.vm01.stdout: python3-ceph-common-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T20:57:03.469 INFO:teuthology.orchestra.run.vm01.stdout: python3-certifi-2023.05.07-4.el9.noarch 2026-03-09T20:57:03.469 INFO:teuthology.orchestra.run.vm01.stdout: python3-cffi-1.14.5-5.el9.x86_64 2026-03-09T20:57:03.469 INFO:teuthology.orchestra.run.vm01.stdout: python3-chardet-4.0.0-5.el9.noarch 2026-03-09T20:57:03.469 INFO:teuthology.orchestra.run.vm01.stdout: python3-cheroot-10.0.1-4.el9.noarch 2026-03-09T20:57:03.469 INFO:teuthology.orchestra.run.vm01.stdout: python3-cherrypy-18.6.1-2.el9.noarch 2026-03-09T20:57:03.469 INFO:teuthology.orchestra.run.vm01.stdout: python3-cryptography-36.0.1-5.el9.x86_64 2026-03-09T20:57:03.469 INFO:teuthology.orchestra.run.vm01.stdout: python3-devel-3.9.25-3.el9.x86_64 2026-03-09T20:57:03.469 INFO:teuthology.orchestra.run.vm01.stdout: python3-google-auth-1:2.45.0-1.el9.noarch 2026-03-09T20:57:03.469 INFO:teuthology.orchestra.run.vm01.stdout: python3-grpcio-1.46.7-10.el9.x86_64 2026-03-09T20:57:03.469 INFO:teuthology.orchestra.run.vm01.stdout: python3-grpcio-tools-1.46.7-10.el9.x86_64 2026-03-09T20:57:03.469 INFO:teuthology.orchestra.run.vm01.stdout: python3-idna-2.10-7.el9.1.noarch 2026-03-09T20:57:03.470 INFO:teuthology.orchestra.run.vm01.stdout: python3-jaraco-8.2.1-3.el9.noarch 2026-03-09T20:57:03.470 INFO:teuthology.orchestra.run.vm01.stdout: python3-jaraco-classes-3.2.1-5.el9.noarch 2026-03-09T20:57:03.470 INFO:teuthology.orchestra.run.vm01.stdout: python3-jaraco-collections-3.0.0-8.el9.noarch 2026-03-09T20:57:03.470 INFO:teuthology.orchestra.run.vm01.stdout: python3-jaraco-context-6.0.1-3.el9.noarch 2026-03-09T20:57:03.470 INFO:teuthology.orchestra.run.vm01.stdout: python3-jaraco-functools-3.5.0-2.el9.noarch 2026-03-09T20:57:03.470 INFO:teuthology.orchestra.run.vm01.stdout: python3-jaraco-text-4.0.0-2.el9.noarch 2026-03-09T20:57:03.470 INFO:teuthology.orchestra.run.vm01.stdout: python3-jinja2-2.11.3-8.el9.noarch 2026-03-09T20:57:03.470 INFO:teuthology.orchestra.run.vm01.stdout: python3-jsonpatch-1.21-16.el9.noarch 2026-03-09T20:57:03.470 INFO:teuthology.orchestra.run.vm01.stdout: python3-jsonpointer-2.0-4.el9.noarch 2026-03-09T20:57:03.470 INFO:teuthology.orchestra.run.vm01.stdout: python3-kubernetes-1:26.1.0-3.el9.noarch 2026-03-09T20:57:03.470 INFO:teuthology.orchestra.run.vm01.stdout: python3-libstoragemgmt-1.10.1-1.el9.x86_64 2026-03-09T20:57:03.470 INFO:teuthology.orchestra.run.vm01.stdout: python3-logutils-0.3.5-21.el9.noarch 2026-03-09T20:57:03.470 INFO:teuthology.orchestra.run.vm01.stdout: python3-mako-1.1.4-6.el9.noarch 2026-03-09T20:57:03.470 INFO:teuthology.orchestra.run.vm01.stdout: python3-markupsafe-1.1.1-12.el9.x86_64 2026-03-09T20:57:03.470 INFO:teuthology.orchestra.run.vm01.stdout: python3-more-itertools-8.12.0-2.el9.noarch 2026-03-09T20:57:03.470 INFO:teuthology.orchestra.run.vm01.stdout: python3-natsort-7.1.1-5.el9.noarch 2026-03-09T20:57:03.470 INFO:teuthology.orchestra.run.vm01.stdout: python3-numpy-1:1.23.5-2.el9.x86_64 2026-03-09T20:57:03.470 INFO:teuthology.orchestra.run.vm01.stdout: python3-numpy-f2py-1:1.23.5-2.el9.x86_64 2026-03-09T20:57:03.470 INFO:teuthology.orchestra.run.vm01.stdout: python3-oauthlib-3.1.1-5.el9.noarch 2026-03-09T20:57:03.470 INFO:teuthology.orchestra.run.vm01.stdout: python3-packaging-20.9-5.el9.noarch 2026-03-09T20:57:03.470 INFO:teuthology.orchestra.run.vm01.stdout: python3-pecan-1.4.2-3.el9.noarch 2026-03-09T20:57:03.470 INFO:teuthology.orchestra.run.vm01.stdout: python3-ply-3.11-14.el9.noarch 2026-03-09T20:57:03.470 INFO:teuthology.orchestra.run.vm01.stdout: python3-portend-3.1.0-2.el9.noarch 2026-03-09T20:57:03.470 INFO:teuthology.orchestra.run.vm01.stdout: python3-prettytable-0.7.2-27.el9.noarch 2026-03-09T20:57:03.470 INFO:teuthology.orchestra.run.vm01.stdout: python3-protobuf-3.14.0-17.el9.noarch 2026-03-09T20:57:03.470 INFO:teuthology.orchestra.run.vm01.stdout: python3-pyOpenSSL-21.0.0-1.el9.noarch 2026-03-09T20:57:03.470 INFO:teuthology.orchestra.run.vm01.stdout: python3-pyasn1-0.4.8-7.el9.noarch 2026-03-09T20:57:03.470 INFO:teuthology.orchestra.run.vm01.stdout: python3-pyasn1-modules-0.4.8-7.el9.noarch 2026-03-09T20:57:03.470 INFO:teuthology.orchestra.run.vm01.stdout: python3-pycparser-2.20-6.el9.noarch 2026-03-09T20:57:03.470 INFO:teuthology.orchestra.run.vm01.stdout: python3-pysocks-1.7.1-12.el9.noarch 2026-03-09T20:57:03.470 INFO:teuthology.orchestra.run.vm01.stdout: python3-pytz-2021.1-5.el9.noarch 2026-03-09T20:57:03.470 INFO:teuthology.orchestra.run.vm01.stdout: python3-repoze-lru-0.7-16.el9.noarch 2026-03-09T20:57:03.470 INFO:teuthology.orchestra.run.vm01.stdout: python3-requests-2.25.1-10.el9.noarch 2026-03-09T20:57:03.470 INFO:teuthology.orchestra.run.vm01.stdout: python3-requests-oauthlib-1.3.0-12.el9.noarch 2026-03-09T20:57:03.470 INFO:teuthology.orchestra.run.vm01.stdout: python3-routes-2.5.1-5.el9.noarch 2026-03-09T20:57:03.470 INFO:teuthology.orchestra.run.vm01.stdout: python3-rsa-4.9-2.el9.noarch 2026-03-09T20:57:03.470 INFO:teuthology.orchestra.run.vm01.stdout: python3-scipy-1.9.3-2.el9.x86_64 2026-03-09T20:57:03.470 INFO:teuthology.orchestra.run.vm01.stdout: python3-tempora-5.0.0-2.el9.noarch 2026-03-09T20:57:03.470 INFO:teuthology.orchestra.run.vm01.stdout: python3-toml-0.10.2-6.el9.noarch 2026-03-09T20:57:03.470 INFO:teuthology.orchestra.run.vm01.stdout: python3-typing-extensions-4.15.0-1.el9.noarch 2026-03-09T20:57:03.470 INFO:teuthology.orchestra.run.vm01.stdout: python3-urllib3-1.26.5-7.el9.noarch 2026-03-09T20:57:03.470 INFO:teuthology.orchestra.run.vm01.stdout: python3-webob-1.8.8-2.el9.noarch 2026-03-09T20:57:03.470 INFO:teuthology.orchestra.run.vm01.stdout: python3-websocket-client-1.2.3-2.el9.noarch 2026-03-09T20:57:03.470 INFO:teuthology.orchestra.run.vm01.stdout: python3-werkzeug-2.0.3-3.el9.1.noarch 2026-03-09T20:57:03.470 INFO:teuthology.orchestra.run.vm01.stdout: python3-zc-lockfile-2.0-10.el9.noarch 2026-03-09T20:57:03.470 INFO:teuthology.orchestra.run.vm01.stdout: qatlib-25.08.0-2.el9.x86_64 2026-03-09T20:57:03.470 INFO:teuthology.orchestra.run.vm01.stdout: qatlib-service-25.08.0-2.el9.x86_64 2026-03-09T20:57:03.470 INFO:teuthology.orchestra.run.vm01.stdout: qatzip-libs-1.3.1-1.el9.x86_64 2026-03-09T20:57:03.471 INFO:teuthology.orchestra.run.vm01.stdout: rbd-mirror-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T20:57:03.471 INFO:teuthology.orchestra.run.vm01.stdout: 2026-03-09T20:57:03.471 INFO:teuthology.orchestra.run.vm01.stdout:Complete! 2026-03-09T20:57:03.634 INFO:teuthology.orchestra.run.vm08.stdout:Dependencies resolved. 2026-03-09T20:57:03.634 INFO:teuthology.orchestra.run.vm08.stdout:================================================================================ 2026-03-09T20:57:03.634 INFO:teuthology.orchestra.run.vm08.stdout: Package Arch Version Repository Size 2026-03-09T20:57:03.634 INFO:teuthology.orchestra.run.vm08.stdout:================================================================================ 2026-03-09T20:57:03.634 INFO:teuthology.orchestra.run.vm08.stdout:Removing: 2026-03-09T20:57:03.634 INFO:teuthology.orchestra.run.vm08.stdout: cephadm noarch 2:19.2.3-678.ge911bdeb.el9 @ceph-noarch 775 k 2026-03-09T20:57:03.634 INFO:teuthology.orchestra.run.vm08.stdout: 2026-03-09T20:57:03.634 INFO:teuthology.orchestra.run.vm08.stdout:Transaction Summary 2026-03-09T20:57:03.634 INFO:teuthology.orchestra.run.vm08.stdout:================================================================================ 2026-03-09T20:57:03.634 INFO:teuthology.orchestra.run.vm08.stdout:Remove 1 Package 2026-03-09T20:57:03.634 INFO:teuthology.orchestra.run.vm08.stdout: 2026-03-09T20:57:03.634 INFO:teuthology.orchestra.run.vm08.stdout:Freed space: 775 k 2026-03-09T20:57:03.635 INFO:teuthology.orchestra.run.vm08.stdout:Running transaction check 2026-03-09T20:57:03.636 INFO:teuthology.orchestra.run.vm08.stdout:Transaction check succeeded. 2026-03-09T20:57:03.636 INFO:teuthology.orchestra.run.vm08.stdout:Running transaction test 2026-03-09T20:57:03.638 INFO:teuthology.orchestra.run.vm08.stdout:Transaction test succeeded. 2026-03-09T20:57:03.638 INFO:teuthology.orchestra.run.vm08.stdout:Running transaction 2026-03-09T20:57:03.654 INFO:teuthology.orchestra.run.vm08.stdout: Preparing : 1/1 2026-03-09T20:57:03.655 INFO:teuthology.orchestra.run.vm08.stdout: Erasing : cephadm-2:19.2.3-678.ge911bdeb.el9.noarch 1/1 2026-03-09T20:57:03.664 INFO:teuthology.orchestra.run.vm06.stdout:Dependencies resolved. 2026-03-09T20:57:03.664 INFO:teuthology.orchestra.run.vm06.stdout:================================================================================ 2026-03-09T20:57:03.665 INFO:teuthology.orchestra.run.vm06.stdout: Package Arch Version Repository Size 2026-03-09T20:57:03.665 INFO:teuthology.orchestra.run.vm06.stdout:================================================================================ 2026-03-09T20:57:03.665 INFO:teuthology.orchestra.run.vm06.stdout:Removing: 2026-03-09T20:57:03.665 INFO:teuthology.orchestra.run.vm06.stdout: cephadm noarch 2:19.2.3-678.ge911bdeb.el9 @ceph-noarch 775 k 2026-03-09T20:57:03.665 INFO:teuthology.orchestra.run.vm06.stdout: 2026-03-09T20:57:03.665 INFO:teuthology.orchestra.run.vm06.stdout:Transaction Summary 2026-03-09T20:57:03.665 INFO:teuthology.orchestra.run.vm06.stdout:================================================================================ 2026-03-09T20:57:03.665 INFO:teuthology.orchestra.run.vm06.stdout:Remove 1 Package 2026-03-09T20:57:03.665 INFO:teuthology.orchestra.run.vm06.stdout: 2026-03-09T20:57:03.665 INFO:teuthology.orchestra.run.vm06.stdout:Freed space: 775 k 2026-03-09T20:57:03.665 INFO:teuthology.orchestra.run.vm06.stdout:Running transaction check 2026-03-09T20:57:03.666 INFO:teuthology.orchestra.run.vm06.stdout:Transaction check succeeded. 2026-03-09T20:57:03.666 INFO:teuthology.orchestra.run.vm06.stdout:Running transaction test 2026-03-09T20:57:03.668 INFO:teuthology.orchestra.run.vm06.stdout:Transaction test succeeded. 2026-03-09T20:57:03.668 INFO:teuthology.orchestra.run.vm06.stdout:Running transaction 2026-03-09T20:57:03.684 INFO:teuthology.orchestra.run.vm06.stdout: Preparing : 1/1 2026-03-09T20:57:03.684 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : cephadm-2:19.2.3-678.ge911bdeb.el9.noarch 1/1 2026-03-09T20:57:03.693 INFO:teuthology.orchestra.run.vm01.stdout:Dependencies resolved. 2026-03-09T20:57:03.694 INFO:teuthology.orchestra.run.vm01.stdout:================================================================================ 2026-03-09T20:57:03.694 INFO:teuthology.orchestra.run.vm01.stdout: Package Arch Version Repository Size 2026-03-09T20:57:03.694 INFO:teuthology.orchestra.run.vm01.stdout:================================================================================ 2026-03-09T20:57:03.694 INFO:teuthology.orchestra.run.vm01.stdout:Removing: 2026-03-09T20:57:03.694 INFO:teuthology.orchestra.run.vm01.stdout: cephadm noarch 2:19.2.3-678.ge911bdeb.el9 @ceph-noarch 775 k 2026-03-09T20:57:03.694 INFO:teuthology.orchestra.run.vm01.stdout: 2026-03-09T20:57:03.694 INFO:teuthology.orchestra.run.vm01.stdout:Transaction Summary 2026-03-09T20:57:03.694 INFO:teuthology.orchestra.run.vm01.stdout:================================================================================ 2026-03-09T20:57:03.694 INFO:teuthology.orchestra.run.vm01.stdout:Remove 1 Package 2026-03-09T20:57:03.694 INFO:teuthology.orchestra.run.vm01.stdout: 2026-03-09T20:57:03.694 INFO:teuthology.orchestra.run.vm01.stdout:Freed space: 775 k 2026-03-09T20:57:03.694 INFO:teuthology.orchestra.run.vm01.stdout:Running transaction check 2026-03-09T20:57:03.695 INFO:teuthology.orchestra.run.vm01.stdout:Transaction check succeeded. 2026-03-09T20:57:03.696 INFO:teuthology.orchestra.run.vm01.stdout:Running transaction test 2026-03-09T20:57:03.697 INFO:teuthology.orchestra.run.vm01.stdout:Transaction test succeeded. 2026-03-09T20:57:03.697 INFO:teuthology.orchestra.run.vm01.stdout:Running transaction 2026-03-09T20:57:03.714 INFO:teuthology.orchestra.run.vm01.stdout: Preparing : 1/1 2026-03-09T20:57:03.714 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : cephadm-2:19.2.3-678.ge911bdeb.el9.noarch 1/1 2026-03-09T20:57:03.763 INFO:teuthology.orchestra.run.vm08.stdout: Running scriptlet: cephadm-2:19.2.3-678.ge911bdeb.el9.noarch 1/1 2026-03-09T20:57:03.798 INFO:teuthology.orchestra.run.vm06.stdout: Running scriptlet: cephadm-2:19.2.3-678.ge911bdeb.el9.noarch 1/1 2026-03-09T20:57:03.809 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : cephadm-2:19.2.3-678.ge911bdeb.el9.noarch 1/1 2026-03-09T20:57:03.809 INFO:teuthology.orchestra.run.vm08.stdout: 2026-03-09T20:57:03.809 INFO:teuthology.orchestra.run.vm08.stdout:Removed: 2026-03-09T20:57:03.809 INFO:teuthology.orchestra.run.vm08.stdout: cephadm-2:19.2.3-678.ge911bdeb.el9.noarch 2026-03-09T20:57:03.809 INFO:teuthology.orchestra.run.vm08.stdout: 2026-03-09T20:57:03.809 INFO:teuthology.orchestra.run.vm08.stdout:Complete! 2026-03-09T20:57:03.843 INFO:teuthology.orchestra.run.vm01.stdout: Running scriptlet: cephadm-2:19.2.3-678.ge911bdeb.el9.noarch 1/1 2026-03-09T20:57:03.850 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : cephadm-2:19.2.3-678.ge911bdeb.el9.noarch 1/1 2026-03-09T20:57:03.850 INFO:teuthology.orchestra.run.vm06.stdout: 2026-03-09T20:57:03.850 INFO:teuthology.orchestra.run.vm06.stdout:Removed: 2026-03-09T20:57:03.850 INFO:teuthology.orchestra.run.vm06.stdout: cephadm-2:19.2.3-678.ge911bdeb.el9.noarch 2026-03-09T20:57:03.850 INFO:teuthology.orchestra.run.vm06.stdout: 2026-03-09T20:57:03.850 INFO:teuthology.orchestra.run.vm06.stdout:Complete! 2026-03-09T20:57:03.898 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : cephadm-2:19.2.3-678.ge911bdeb.el9.noarch 1/1 2026-03-09T20:57:03.898 INFO:teuthology.orchestra.run.vm01.stdout: 2026-03-09T20:57:03.898 INFO:teuthology.orchestra.run.vm01.stdout:Removed: 2026-03-09T20:57:03.898 INFO:teuthology.orchestra.run.vm01.stdout: cephadm-2:19.2.3-678.ge911bdeb.el9.noarch 2026-03-09T20:57:03.898 INFO:teuthology.orchestra.run.vm01.stdout: 2026-03-09T20:57:03.898 INFO:teuthology.orchestra.run.vm01.stdout:Complete! 2026-03-09T20:57:04.034 INFO:teuthology.orchestra.run.vm08.stdout:No match for argument: ceph-immutable-object-cache 2026-03-09T20:57:04.034 INFO:teuthology.orchestra.run.vm08.stderr:No packages marked for removal. 2026-03-09T20:57:04.037 INFO:teuthology.orchestra.run.vm08.stdout:Dependencies resolved. 2026-03-09T20:57:04.038 INFO:teuthology.orchestra.run.vm08.stdout:Nothing to do. 2026-03-09T20:57:04.038 INFO:teuthology.orchestra.run.vm08.stdout:Complete! 2026-03-09T20:57:04.049 INFO:teuthology.orchestra.run.vm06.stdout:No match for argument: ceph-immutable-object-cache 2026-03-09T20:57:04.049 INFO:teuthology.orchestra.run.vm06.stderr:No packages marked for removal. 2026-03-09T20:57:04.053 INFO:teuthology.orchestra.run.vm06.stdout:Dependencies resolved. 2026-03-09T20:57:04.053 INFO:teuthology.orchestra.run.vm06.stdout:Nothing to do. 2026-03-09T20:57:04.053 INFO:teuthology.orchestra.run.vm06.stdout:Complete! 2026-03-09T20:57:04.089 INFO:teuthology.orchestra.run.vm01.stdout:No match for argument: ceph-immutable-object-cache 2026-03-09T20:57:04.089 INFO:teuthology.orchestra.run.vm01.stderr:No packages marked for removal. 2026-03-09T20:57:04.092 INFO:teuthology.orchestra.run.vm01.stdout:Dependencies resolved. 2026-03-09T20:57:04.093 INFO:teuthology.orchestra.run.vm01.stdout:Nothing to do. 2026-03-09T20:57:04.093 INFO:teuthology.orchestra.run.vm01.stdout:Complete! 2026-03-09T20:57:04.210 INFO:teuthology.orchestra.run.vm08.stdout:No match for argument: ceph-mgr 2026-03-09T20:57:04.210 INFO:teuthology.orchestra.run.vm08.stderr:No packages marked for removal. 2026-03-09T20:57:04.213 INFO:teuthology.orchestra.run.vm08.stdout:Dependencies resolved. 2026-03-09T20:57:04.214 INFO:teuthology.orchestra.run.vm08.stdout:Nothing to do. 2026-03-09T20:57:04.214 INFO:teuthology.orchestra.run.vm08.stdout:Complete! 2026-03-09T20:57:04.238 INFO:teuthology.orchestra.run.vm06.stdout:No match for argument: ceph-mgr 2026-03-09T20:57:04.238 INFO:teuthology.orchestra.run.vm06.stderr:No packages marked for removal. 2026-03-09T20:57:04.242 INFO:teuthology.orchestra.run.vm06.stdout:Dependencies resolved. 2026-03-09T20:57:04.243 INFO:teuthology.orchestra.run.vm06.stdout:Nothing to do. 2026-03-09T20:57:04.243 INFO:teuthology.orchestra.run.vm06.stdout:Complete! 2026-03-09T20:57:04.272 INFO:teuthology.orchestra.run.vm01.stdout:No match for argument: ceph-mgr 2026-03-09T20:57:04.272 INFO:teuthology.orchestra.run.vm01.stderr:No packages marked for removal. 2026-03-09T20:57:04.276 INFO:teuthology.orchestra.run.vm01.stdout:Dependencies resolved. 2026-03-09T20:57:04.276 INFO:teuthology.orchestra.run.vm01.stdout:Nothing to do. 2026-03-09T20:57:04.276 INFO:teuthology.orchestra.run.vm01.stdout:Complete! 2026-03-09T20:57:04.394 INFO:teuthology.orchestra.run.vm08.stdout:No match for argument: ceph-mgr-dashboard 2026-03-09T20:57:04.394 INFO:teuthology.orchestra.run.vm08.stderr:No packages marked for removal. 2026-03-09T20:57:04.398 INFO:teuthology.orchestra.run.vm08.stdout:Dependencies resolved. 2026-03-09T20:57:04.398 INFO:teuthology.orchestra.run.vm08.stdout:Nothing to do. 2026-03-09T20:57:04.398 INFO:teuthology.orchestra.run.vm08.stdout:Complete! 2026-03-09T20:57:04.419 INFO:teuthology.orchestra.run.vm06.stdout:No match for argument: ceph-mgr-dashboard 2026-03-09T20:57:04.419 INFO:teuthology.orchestra.run.vm06.stderr:No packages marked for removal. 2026-03-09T20:57:04.423 INFO:teuthology.orchestra.run.vm06.stdout:Dependencies resolved. 2026-03-09T20:57:04.423 INFO:teuthology.orchestra.run.vm06.stdout:Nothing to do. 2026-03-09T20:57:04.423 INFO:teuthology.orchestra.run.vm06.stdout:Complete! 2026-03-09T20:57:04.448 INFO:teuthology.orchestra.run.vm01.stdout:No match for argument: ceph-mgr-dashboard 2026-03-09T20:57:04.448 INFO:teuthology.orchestra.run.vm01.stderr:No packages marked for removal. 2026-03-09T20:57:04.451 INFO:teuthology.orchestra.run.vm01.stdout:Dependencies resolved. 2026-03-09T20:57:04.452 INFO:teuthology.orchestra.run.vm01.stdout:Nothing to do. 2026-03-09T20:57:04.452 INFO:teuthology.orchestra.run.vm01.stdout:Complete! 2026-03-09T20:57:04.574 INFO:teuthology.orchestra.run.vm08.stdout:No match for argument: ceph-mgr-diskprediction-local 2026-03-09T20:57:04.575 INFO:teuthology.orchestra.run.vm08.stderr:No packages marked for removal. 2026-03-09T20:57:04.578 INFO:teuthology.orchestra.run.vm08.stdout:Dependencies resolved. 2026-03-09T20:57:04.579 INFO:teuthology.orchestra.run.vm08.stdout:Nothing to do. 2026-03-09T20:57:04.579 INFO:teuthology.orchestra.run.vm08.stdout:Complete! 2026-03-09T20:57:04.598 INFO:teuthology.orchestra.run.vm06.stdout:No match for argument: ceph-mgr-diskprediction-local 2026-03-09T20:57:04.598 INFO:teuthology.orchestra.run.vm06.stderr:No packages marked for removal. 2026-03-09T20:57:04.601 INFO:teuthology.orchestra.run.vm06.stdout:Dependencies resolved. 2026-03-09T20:57:04.601 INFO:teuthology.orchestra.run.vm06.stdout:Nothing to do. 2026-03-09T20:57:04.601 INFO:teuthology.orchestra.run.vm06.stdout:Complete! 2026-03-09T20:57:04.627 INFO:teuthology.orchestra.run.vm01.stdout:No match for argument: ceph-mgr-diskprediction-local 2026-03-09T20:57:04.627 INFO:teuthology.orchestra.run.vm01.stderr:No packages marked for removal. 2026-03-09T20:57:04.630 INFO:teuthology.orchestra.run.vm01.stdout:Dependencies resolved. 2026-03-09T20:57:04.631 INFO:teuthology.orchestra.run.vm01.stdout:Nothing to do. 2026-03-09T20:57:04.631 INFO:teuthology.orchestra.run.vm01.stdout:Complete! 2026-03-09T20:57:04.758 INFO:teuthology.orchestra.run.vm08.stdout:No match for argument: ceph-mgr-rook 2026-03-09T20:57:04.758 INFO:teuthology.orchestra.run.vm08.stderr:No packages marked for removal. 2026-03-09T20:57:04.762 INFO:teuthology.orchestra.run.vm08.stdout:Dependencies resolved. 2026-03-09T20:57:04.762 INFO:teuthology.orchestra.run.vm08.stdout:Nothing to do. 2026-03-09T20:57:04.762 INFO:teuthology.orchestra.run.vm08.stdout:Complete! 2026-03-09T20:57:04.774 INFO:teuthology.orchestra.run.vm06.stdout:No match for argument: ceph-mgr-rook 2026-03-09T20:57:04.775 INFO:teuthology.orchestra.run.vm06.stderr:No packages marked for removal. 2026-03-09T20:57:04.778 INFO:teuthology.orchestra.run.vm06.stdout:Dependencies resolved. 2026-03-09T20:57:04.778 INFO:teuthology.orchestra.run.vm06.stdout:Nothing to do. 2026-03-09T20:57:04.778 INFO:teuthology.orchestra.run.vm06.stdout:Complete! 2026-03-09T20:57:04.812 INFO:teuthology.orchestra.run.vm01.stdout:No match for argument: ceph-mgr-rook 2026-03-09T20:57:04.812 INFO:teuthology.orchestra.run.vm01.stderr:No packages marked for removal. 2026-03-09T20:57:04.816 INFO:teuthology.orchestra.run.vm01.stdout:Dependencies resolved. 2026-03-09T20:57:04.816 INFO:teuthology.orchestra.run.vm01.stdout:Nothing to do. 2026-03-09T20:57:04.816 INFO:teuthology.orchestra.run.vm01.stdout:Complete! 2026-03-09T20:57:04.943 INFO:teuthology.orchestra.run.vm08.stdout:No match for argument: ceph-mgr-cephadm 2026-03-09T20:57:04.944 INFO:teuthology.orchestra.run.vm08.stderr:No packages marked for removal. 2026-03-09T20:57:04.947 INFO:teuthology.orchestra.run.vm08.stdout:Dependencies resolved. 2026-03-09T20:57:04.947 INFO:teuthology.orchestra.run.vm08.stdout:Nothing to do. 2026-03-09T20:57:04.947 INFO:teuthology.orchestra.run.vm08.stdout:Complete! 2026-03-09T20:57:04.964 INFO:teuthology.orchestra.run.vm06.stdout:No match for argument: ceph-mgr-cephadm 2026-03-09T20:57:04.964 INFO:teuthology.orchestra.run.vm06.stderr:No packages marked for removal. 2026-03-09T20:57:04.968 INFO:teuthology.orchestra.run.vm06.stdout:Dependencies resolved. 2026-03-09T20:57:04.969 INFO:teuthology.orchestra.run.vm06.stdout:Nothing to do. 2026-03-09T20:57:04.969 INFO:teuthology.orchestra.run.vm06.stdout:Complete! 2026-03-09T20:57:05.014 INFO:teuthology.orchestra.run.vm01.stdout:No match for argument: ceph-mgr-cephadm 2026-03-09T20:57:05.014 INFO:teuthology.orchestra.run.vm01.stderr:No packages marked for removal. 2026-03-09T20:57:05.017 INFO:teuthology.orchestra.run.vm01.stdout:Dependencies resolved. 2026-03-09T20:57:05.017 INFO:teuthology.orchestra.run.vm01.stdout:Nothing to do. 2026-03-09T20:57:05.018 INFO:teuthology.orchestra.run.vm01.stdout:Complete! 2026-03-09T20:57:05.139 INFO:teuthology.orchestra.run.vm08.stdout:Dependencies resolved. 2026-03-09T20:57:05.139 INFO:teuthology.orchestra.run.vm08.stdout:================================================================================ 2026-03-09T20:57:05.139 INFO:teuthology.orchestra.run.vm08.stdout: Package Arch Version Repository Size 2026-03-09T20:57:05.139 INFO:teuthology.orchestra.run.vm08.stdout:================================================================================ 2026-03-09T20:57:05.139 INFO:teuthology.orchestra.run.vm08.stdout:Removing: 2026-03-09T20:57:05.139 INFO:teuthology.orchestra.run.vm08.stdout: ceph-fuse x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 3.6 M 2026-03-09T20:57:05.140 INFO:teuthology.orchestra.run.vm08.stdout: 2026-03-09T20:57:05.140 INFO:teuthology.orchestra.run.vm08.stdout:Transaction Summary 2026-03-09T20:57:05.140 INFO:teuthology.orchestra.run.vm08.stdout:================================================================================ 2026-03-09T20:57:05.140 INFO:teuthology.orchestra.run.vm08.stdout:Remove 1 Package 2026-03-09T20:57:05.140 INFO:teuthology.orchestra.run.vm08.stdout: 2026-03-09T20:57:05.140 INFO:teuthology.orchestra.run.vm08.stdout:Freed space: 3.6 M 2026-03-09T20:57:05.140 INFO:teuthology.orchestra.run.vm08.stdout:Running transaction check 2026-03-09T20:57:05.141 INFO:teuthology.orchestra.run.vm08.stdout:Transaction check succeeded. 2026-03-09T20:57:05.142 INFO:teuthology.orchestra.run.vm08.stdout:Running transaction test 2026-03-09T20:57:05.151 INFO:teuthology.orchestra.run.vm08.stdout:Transaction test succeeded. 2026-03-09T20:57:05.151 INFO:teuthology.orchestra.run.vm08.stdout:Running transaction 2026-03-09T20:57:05.155 INFO:teuthology.orchestra.run.vm06.stdout:Dependencies resolved. 2026-03-09T20:57:05.156 INFO:teuthology.orchestra.run.vm06.stdout:================================================================================ 2026-03-09T20:57:05.156 INFO:teuthology.orchestra.run.vm06.stdout: Package Arch Version Repository Size 2026-03-09T20:57:05.156 INFO:teuthology.orchestra.run.vm06.stdout:================================================================================ 2026-03-09T20:57:05.156 INFO:teuthology.orchestra.run.vm06.stdout:Removing: 2026-03-09T20:57:05.156 INFO:teuthology.orchestra.run.vm06.stdout: ceph-fuse x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 3.6 M 2026-03-09T20:57:05.156 INFO:teuthology.orchestra.run.vm06.stdout: 2026-03-09T20:57:05.156 INFO:teuthology.orchestra.run.vm06.stdout:Transaction Summary 2026-03-09T20:57:05.156 INFO:teuthology.orchestra.run.vm06.stdout:================================================================================ 2026-03-09T20:57:05.156 INFO:teuthology.orchestra.run.vm06.stdout:Remove 1 Package 2026-03-09T20:57:05.156 INFO:teuthology.orchestra.run.vm06.stdout: 2026-03-09T20:57:05.156 INFO:teuthology.orchestra.run.vm06.stdout:Freed space: 3.6 M 2026-03-09T20:57:05.156 INFO:teuthology.orchestra.run.vm06.stdout:Running transaction check 2026-03-09T20:57:05.158 INFO:teuthology.orchestra.run.vm06.stdout:Transaction check succeeded. 2026-03-09T20:57:05.158 INFO:teuthology.orchestra.run.vm06.stdout:Running transaction test 2026-03-09T20:57:05.168 INFO:teuthology.orchestra.run.vm06.stdout:Transaction test succeeded. 2026-03-09T20:57:05.169 INFO:teuthology.orchestra.run.vm06.stdout:Running transaction 2026-03-09T20:57:05.176 INFO:teuthology.orchestra.run.vm08.stdout: Preparing : 1/1 2026-03-09T20:57:05.191 INFO:teuthology.orchestra.run.vm08.stdout: Erasing : ceph-fuse-2:19.2.3-678.ge911bdeb.el9.x86_64 1/1 2026-03-09T20:57:05.193 INFO:teuthology.orchestra.run.vm06.stdout: Preparing : 1/1 2026-03-09T20:57:05.207 INFO:teuthology.orchestra.run.vm01.stdout:Dependencies resolved. 2026-03-09T20:57:05.208 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : ceph-fuse-2:19.2.3-678.ge911bdeb.el9.x86_64 1/1 2026-03-09T20:57:05.208 INFO:teuthology.orchestra.run.vm01.stdout:================================================================================ 2026-03-09T20:57:05.208 INFO:teuthology.orchestra.run.vm01.stdout: Package Arch Version Repository Size 2026-03-09T20:57:05.208 INFO:teuthology.orchestra.run.vm01.stdout:================================================================================ 2026-03-09T20:57:05.208 INFO:teuthology.orchestra.run.vm01.stdout:Removing: 2026-03-09T20:57:05.208 INFO:teuthology.orchestra.run.vm01.stdout: ceph-fuse x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 3.6 M 2026-03-09T20:57:05.208 INFO:teuthology.orchestra.run.vm01.stdout: 2026-03-09T20:57:05.208 INFO:teuthology.orchestra.run.vm01.stdout:Transaction Summary 2026-03-09T20:57:05.208 INFO:teuthology.orchestra.run.vm01.stdout:================================================================================ 2026-03-09T20:57:05.208 INFO:teuthology.orchestra.run.vm01.stdout:Remove 1 Package 2026-03-09T20:57:05.208 INFO:teuthology.orchestra.run.vm01.stdout: 2026-03-09T20:57:05.208 INFO:teuthology.orchestra.run.vm01.stdout:Freed space: 3.6 M 2026-03-09T20:57:05.208 INFO:teuthology.orchestra.run.vm01.stdout:Running transaction check 2026-03-09T20:57:05.210 INFO:teuthology.orchestra.run.vm01.stdout:Transaction check succeeded. 2026-03-09T20:57:05.210 INFO:teuthology.orchestra.run.vm01.stdout:Running transaction test 2026-03-09T20:57:05.220 INFO:teuthology.orchestra.run.vm01.stdout:Transaction test succeeded. 2026-03-09T20:57:05.220 INFO:teuthology.orchestra.run.vm01.stdout:Running transaction 2026-03-09T20:57:05.246 INFO:teuthology.orchestra.run.vm01.stdout: Preparing : 1/1 2026-03-09T20:57:05.257 INFO:teuthology.orchestra.run.vm08.stdout: Running scriptlet: ceph-fuse-2:19.2.3-678.ge911bdeb.el9.x86_64 1/1 2026-03-09T20:57:05.260 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : ceph-fuse-2:19.2.3-678.ge911bdeb.el9.x86_64 1/1 2026-03-09T20:57:05.280 INFO:teuthology.orchestra.run.vm06.stdout: Running scriptlet: ceph-fuse-2:19.2.3-678.ge911bdeb.el9.x86_64 1/1 2026-03-09T20:57:05.302 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : ceph-fuse-2:19.2.3-678.ge911bdeb.el9.x86_64 1/1 2026-03-09T20:57:05.303 INFO:teuthology.orchestra.run.vm08.stdout: 2026-03-09T20:57:05.303 INFO:teuthology.orchestra.run.vm08.stdout:Removed: 2026-03-09T20:57:05.303 INFO:teuthology.orchestra.run.vm08.stdout: ceph-fuse-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T20:57:05.303 INFO:teuthology.orchestra.run.vm08.stdout: 2026-03-09T20:57:05.303 INFO:teuthology.orchestra.run.vm08.stdout:Complete! 2026-03-09T20:57:05.327 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : ceph-fuse-2:19.2.3-678.ge911bdeb.el9.x86_64 1/1 2026-03-09T20:57:05.327 INFO:teuthology.orchestra.run.vm06.stdout: 2026-03-09T20:57:05.327 INFO:teuthology.orchestra.run.vm06.stdout:Removed: 2026-03-09T20:57:05.327 INFO:teuthology.orchestra.run.vm06.stdout: ceph-fuse-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T20:57:05.327 INFO:teuthology.orchestra.run.vm06.stdout: 2026-03-09T20:57:05.327 INFO:teuthology.orchestra.run.vm06.stdout:Complete! 2026-03-09T20:57:05.341 INFO:teuthology.orchestra.run.vm01.stdout: Running scriptlet: ceph-fuse-2:19.2.3-678.ge911bdeb.el9.x86_64 1/1 2026-03-09T20:57:05.384 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : ceph-fuse-2:19.2.3-678.ge911bdeb.el9.x86_64 1/1 2026-03-09T20:57:05.384 INFO:teuthology.orchestra.run.vm01.stdout: 2026-03-09T20:57:05.384 INFO:teuthology.orchestra.run.vm01.stdout:Removed: 2026-03-09T20:57:05.384 INFO:teuthology.orchestra.run.vm01.stdout: ceph-fuse-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T20:57:05.384 INFO:teuthology.orchestra.run.vm01.stdout: 2026-03-09T20:57:05.384 INFO:teuthology.orchestra.run.vm01.stdout:Complete! 2026-03-09T20:57:05.502 INFO:teuthology.orchestra.run.vm08.stdout:No match for argument: ceph-volume 2026-03-09T20:57:05.502 INFO:teuthology.orchestra.run.vm08.stderr:No packages marked for removal. 2026-03-09T20:57:05.505 INFO:teuthology.orchestra.run.vm08.stdout:Dependencies resolved. 2026-03-09T20:57:05.506 INFO:teuthology.orchestra.run.vm08.stdout:Nothing to do. 2026-03-09T20:57:05.506 INFO:teuthology.orchestra.run.vm08.stdout:Complete! 2026-03-09T20:57:05.515 INFO:teuthology.orchestra.run.vm06.stdout:No match for argument: ceph-volume 2026-03-09T20:57:05.515 INFO:teuthology.orchestra.run.vm06.stderr:No packages marked for removal. 2026-03-09T20:57:05.518 INFO:teuthology.orchestra.run.vm06.stdout:Dependencies resolved. 2026-03-09T20:57:05.519 INFO:teuthology.orchestra.run.vm06.stdout:Nothing to do. 2026-03-09T20:57:05.519 INFO:teuthology.orchestra.run.vm06.stdout:Complete! 2026-03-09T20:57:05.576 INFO:teuthology.orchestra.run.vm01.stdout:No match for argument: ceph-volume 2026-03-09T20:57:05.577 INFO:teuthology.orchestra.run.vm01.stderr:No packages marked for removal. 2026-03-09T20:57:05.580 INFO:teuthology.orchestra.run.vm01.stdout:Dependencies resolved. 2026-03-09T20:57:05.580 INFO:teuthology.orchestra.run.vm01.stdout:Nothing to do. 2026-03-09T20:57:05.580 INFO:teuthology.orchestra.run.vm01.stdout:Complete! 2026-03-09T20:57:05.687 INFO:teuthology.orchestra.run.vm08.stdout:Dependencies resolved. 2026-03-09T20:57:05.688 INFO:teuthology.orchestra.run.vm08.stdout:================================================================================ 2026-03-09T20:57:05.688 INFO:teuthology.orchestra.run.vm08.stdout: Package Arch Version Repo Size 2026-03-09T20:57:05.688 INFO:teuthology.orchestra.run.vm08.stdout:================================================================================ 2026-03-09T20:57:05.688 INFO:teuthology.orchestra.run.vm08.stdout:Removing: 2026-03-09T20:57:05.688 INFO:teuthology.orchestra.run.vm08.stdout: librados-devel x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 456 k 2026-03-09T20:57:05.688 INFO:teuthology.orchestra.run.vm08.stdout:Removing dependent packages: 2026-03-09T20:57:05.688 INFO:teuthology.orchestra.run.vm08.stdout: libcephfs-devel x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 153 k 2026-03-09T20:57:05.688 INFO:teuthology.orchestra.run.vm08.stdout: 2026-03-09T20:57:05.688 INFO:teuthology.orchestra.run.vm08.stdout:Transaction Summary 2026-03-09T20:57:05.688 INFO:teuthology.orchestra.run.vm08.stdout:================================================================================ 2026-03-09T20:57:05.688 INFO:teuthology.orchestra.run.vm08.stdout:Remove 2 Packages 2026-03-09T20:57:05.688 INFO:teuthology.orchestra.run.vm08.stdout: 2026-03-09T20:57:05.688 INFO:teuthology.orchestra.run.vm08.stdout:Freed space: 610 k 2026-03-09T20:57:05.688 INFO:teuthology.orchestra.run.vm08.stdout:Running transaction check 2026-03-09T20:57:05.690 INFO:teuthology.orchestra.run.vm08.stdout:Transaction check succeeded. 2026-03-09T20:57:05.690 INFO:teuthology.orchestra.run.vm08.stdout:Running transaction test 2026-03-09T20:57:05.700 INFO:teuthology.orchestra.run.vm08.stdout:Transaction test succeeded. 2026-03-09T20:57:05.700 INFO:teuthology.orchestra.run.vm08.stdout:Running transaction 2026-03-09T20:57:05.700 INFO:teuthology.orchestra.run.vm06.stdout:Dependencies resolved. 2026-03-09T20:57:05.701 INFO:teuthology.orchestra.run.vm06.stdout:================================================================================ 2026-03-09T20:57:05.701 INFO:teuthology.orchestra.run.vm06.stdout: Package Arch Version Repo Size 2026-03-09T20:57:05.701 INFO:teuthology.orchestra.run.vm06.stdout:================================================================================ 2026-03-09T20:57:05.701 INFO:teuthology.orchestra.run.vm06.stdout:Removing: 2026-03-09T20:57:05.701 INFO:teuthology.orchestra.run.vm06.stdout: librados-devel x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 456 k 2026-03-09T20:57:05.701 INFO:teuthology.orchestra.run.vm06.stdout:Removing dependent packages: 2026-03-09T20:57:05.701 INFO:teuthology.orchestra.run.vm06.stdout: libcephfs-devel x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 153 k 2026-03-09T20:57:05.701 INFO:teuthology.orchestra.run.vm06.stdout: 2026-03-09T20:57:05.701 INFO:teuthology.orchestra.run.vm06.stdout:Transaction Summary 2026-03-09T20:57:05.701 INFO:teuthology.orchestra.run.vm06.stdout:================================================================================ 2026-03-09T20:57:05.701 INFO:teuthology.orchestra.run.vm06.stdout:Remove 2 Packages 2026-03-09T20:57:05.701 INFO:teuthology.orchestra.run.vm06.stdout: 2026-03-09T20:57:05.701 INFO:teuthology.orchestra.run.vm06.stdout:Freed space: 610 k 2026-03-09T20:57:05.701 INFO:teuthology.orchestra.run.vm06.stdout:Running transaction check 2026-03-09T20:57:05.703 INFO:teuthology.orchestra.run.vm06.stdout:Transaction check succeeded. 2026-03-09T20:57:05.703 INFO:teuthology.orchestra.run.vm06.stdout:Running transaction test 2026-03-09T20:57:05.713 INFO:teuthology.orchestra.run.vm06.stdout:Transaction test succeeded. 2026-03-09T20:57:05.714 INFO:teuthology.orchestra.run.vm06.stdout:Running transaction 2026-03-09T20:57:05.726 INFO:teuthology.orchestra.run.vm08.stdout: Preparing : 1/1 2026-03-09T20:57:05.728 INFO:teuthology.orchestra.run.vm08.stdout: Erasing : libcephfs-devel-2:19.2.3-678.ge911bdeb.el9.x86_64 1/2 2026-03-09T20:57:05.739 INFO:teuthology.orchestra.run.vm06.stdout: Preparing : 1/1 2026-03-09T20:57:05.741 INFO:teuthology.orchestra.run.vm08.stdout: Erasing : librados-devel-2:19.2.3-678.ge911bdeb.el9.x86_64 2/2 2026-03-09T20:57:05.741 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : libcephfs-devel-2:19.2.3-678.ge911bdeb.el9.x86_64 1/2 2026-03-09T20:57:05.755 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : librados-devel-2:19.2.3-678.ge911bdeb.el9.x86_64 2/2 2026-03-09T20:57:05.765 INFO:teuthology.orchestra.run.vm01.stdout:Dependencies resolved. 2026-03-09T20:57:05.766 INFO:teuthology.orchestra.run.vm01.stdout:================================================================================ 2026-03-09T20:57:05.766 INFO:teuthology.orchestra.run.vm01.stdout: Package Arch Version Repo Size 2026-03-09T20:57:05.766 INFO:teuthology.orchestra.run.vm01.stdout:================================================================================ 2026-03-09T20:57:05.766 INFO:teuthology.orchestra.run.vm01.stdout:Removing: 2026-03-09T20:57:05.766 INFO:teuthology.orchestra.run.vm01.stdout: librados-devel x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 456 k 2026-03-09T20:57:05.766 INFO:teuthology.orchestra.run.vm01.stdout:Removing dependent packages: 2026-03-09T20:57:05.766 INFO:teuthology.orchestra.run.vm01.stdout: libcephfs-devel x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 153 k 2026-03-09T20:57:05.766 INFO:teuthology.orchestra.run.vm01.stdout: 2026-03-09T20:57:05.766 INFO:teuthology.orchestra.run.vm01.stdout:Transaction Summary 2026-03-09T20:57:05.766 INFO:teuthology.orchestra.run.vm01.stdout:================================================================================ 2026-03-09T20:57:05.766 INFO:teuthology.orchestra.run.vm01.stdout:Remove 2 Packages 2026-03-09T20:57:05.766 INFO:teuthology.orchestra.run.vm01.stdout: 2026-03-09T20:57:05.766 INFO:teuthology.orchestra.run.vm01.stdout:Freed space: 610 k 2026-03-09T20:57:05.766 INFO:teuthology.orchestra.run.vm01.stdout:Running transaction check 2026-03-09T20:57:05.768 INFO:teuthology.orchestra.run.vm01.stdout:Transaction check succeeded. 2026-03-09T20:57:05.768 INFO:teuthology.orchestra.run.vm01.stdout:Running transaction test 2026-03-09T20:57:05.779 INFO:teuthology.orchestra.run.vm01.stdout:Transaction test succeeded. 2026-03-09T20:57:05.779 INFO:teuthology.orchestra.run.vm01.stdout:Running transaction 2026-03-09T20:57:05.804 INFO:teuthology.orchestra.run.vm08.stdout: Running scriptlet: librados-devel-2:19.2.3-678.ge911bdeb.el9.x86_64 2/2 2026-03-09T20:57:05.804 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : libcephfs-devel-2:19.2.3-678.ge911bdeb.el9.x86_64 1/2 2026-03-09T20:57:05.804 INFO:teuthology.orchestra.run.vm01.stdout: Preparing : 1/1 2026-03-09T20:57:05.807 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : libcephfs-devel-2:19.2.3-678.ge911bdeb.el9.x86_64 1/2 2026-03-09T20:57:05.820 INFO:teuthology.orchestra.run.vm06.stdout: Running scriptlet: librados-devel-2:19.2.3-678.ge911bdeb.el9.x86_64 2/2 2026-03-09T20:57:05.820 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : libcephfs-devel-2:19.2.3-678.ge911bdeb.el9.x86_64 1/2 2026-03-09T20:57:05.820 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : librados-devel-2:19.2.3-678.ge911bdeb.el9.x86_64 2/2 2026-03-09T20:57:05.851 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : librados-devel-2:19.2.3-678.ge911bdeb.el9.x86_64 2/2 2026-03-09T20:57:05.852 INFO:teuthology.orchestra.run.vm08.stdout: 2026-03-09T20:57:05.852 INFO:teuthology.orchestra.run.vm08.stdout:Removed: 2026-03-09T20:57:05.852 INFO:teuthology.orchestra.run.vm08.stdout: libcephfs-devel-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T20:57:05.852 INFO:teuthology.orchestra.run.vm08.stdout: librados-devel-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T20:57:05.852 INFO:teuthology.orchestra.run.vm08.stdout: 2026-03-09T20:57:05.852 INFO:teuthology.orchestra.run.vm08.stdout:Complete! 2026-03-09T20:57:05.870 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : librados-devel-2:19.2.3-678.ge911bdeb.el9.x86_64 2/2 2026-03-09T20:57:05.870 INFO:teuthology.orchestra.run.vm06.stdout: 2026-03-09T20:57:05.870 INFO:teuthology.orchestra.run.vm06.stdout:Removed: 2026-03-09T20:57:05.870 INFO:teuthology.orchestra.run.vm06.stdout: libcephfs-devel-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T20:57:05.870 INFO:teuthology.orchestra.run.vm06.stdout: librados-devel-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T20:57:05.870 INFO:teuthology.orchestra.run.vm06.stdout: 2026-03-09T20:57:05.870 INFO:teuthology.orchestra.run.vm06.stdout:Complete! 2026-03-09T20:57:05.886 INFO:teuthology.orchestra.run.vm01.stdout: Running scriptlet: librados-devel-2:19.2.3-678.ge911bdeb.el9.x86_64 2/2 2026-03-09T20:57:05.886 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : libcephfs-devel-2:19.2.3-678.ge911bdeb.el9.x86_64 1/2 2026-03-09T20:57:05.931 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : librados-devel-2:19.2.3-678.ge911bdeb.el9.x86_64 2/2 2026-03-09T20:57:05.931 INFO:teuthology.orchestra.run.vm01.stdout: 2026-03-09T20:57:05.931 INFO:teuthology.orchestra.run.vm01.stdout:Removed: 2026-03-09T20:57:05.931 INFO:teuthology.orchestra.run.vm01.stdout: libcephfs-devel-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T20:57:05.931 INFO:teuthology.orchestra.run.vm01.stdout: librados-devel-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T20:57:05.931 INFO:teuthology.orchestra.run.vm01.stdout: 2026-03-09T20:57:05.931 INFO:teuthology.orchestra.run.vm01.stdout:Complete! 2026-03-09T20:57:06.047 INFO:teuthology.orchestra.run.vm08.stdout:Dependencies resolved. 2026-03-09T20:57:06.048 INFO:teuthology.orchestra.run.vm08.stdout:================================================================================ 2026-03-09T20:57:06.048 INFO:teuthology.orchestra.run.vm08.stdout: Package Arch Version Repo Size 2026-03-09T20:57:06.048 INFO:teuthology.orchestra.run.vm08.stdout:================================================================================ 2026-03-09T20:57:06.048 INFO:teuthology.orchestra.run.vm08.stdout:Removing: 2026-03-09T20:57:06.048 INFO:teuthology.orchestra.run.vm08.stdout: libcephfs2 x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 3.0 M 2026-03-09T20:57:06.048 INFO:teuthology.orchestra.run.vm08.stdout:Removing dependent packages: 2026-03-09T20:57:06.048 INFO:teuthology.orchestra.run.vm08.stdout: python3-cephfs x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 514 k 2026-03-09T20:57:06.048 INFO:teuthology.orchestra.run.vm08.stdout:Removing unused dependencies: 2026-03-09T20:57:06.048 INFO:teuthology.orchestra.run.vm08.stdout: python3-ceph-argparse x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 187 k 2026-03-09T20:57:06.048 INFO:teuthology.orchestra.run.vm08.stdout: 2026-03-09T20:57:06.048 INFO:teuthology.orchestra.run.vm08.stdout:Transaction Summary 2026-03-09T20:57:06.048 INFO:teuthology.orchestra.run.vm08.stdout:================================================================================ 2026-03-09T20:57:06.048 INFO:teuthology.orchestra.run.vm08.stdout:Remove 3 Packages 2026-03-09T20:57:06.048 INFO:teuthology.orchestra.run.vm08.stdout: 2026-03-09T20:57:06.048 INFO:teuthology.orchestra.run.vm08.stdout:Freed space: 3.7 M 2026-03-09T20:57:06.048 INFO:teuthology.orchestra.run.vm08.stdout:Running transaction check 2026-03-09T20:57:06.050 INFO:teuthology.orchestra.run.vm08.stdout:Transaction check succeeded. 2026-03-09T20:57:06.050 INFO:teuthology.orchestra.run.vm08.stdout:Running transaction test 2026-03-09T20:57:06.067 INFO:teuthology.orchestra.run.vm06.stdout:Dependencies resolved. 2026-03-09T20:57:06.067 INFO:teuthology.orchestra.run.vm08.stdout:Transaction test succeeded. 2026-03-09T20:57:06.067 INFO:teuthology.orchestra.run.vm08.stdout:Running transaction 2026-03-09T20:57:06.068 INFO:teuthology.orchestra.run.vm06.stdout:================================================================================ 2026-03-09T20:57:06.068 INFO:teuthology.orchestra.run.vm06.stdout: Package Arch Version Repo Size 2026-03-09T20:57:06.068 INFO:teuthology.orchestra.run.vm06.stdout:================================================================================ 2026-03-09T20:57:06.068 INFO:teuthology.orchestra.run.vm06.stdout:Removing: 2026-03-09T20:57:06.068 INFO:teuthology.orchestra.run.vm06.stdout: libcephfs2 x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 3.0 M 2026-03-09T20:57:06.068 INFO:teuthology.orchestra.run.vm06.stdout:Removing dependent packages: 2026-03-09T20:57:06.068 INFO:teuthology.orchestra.run.vm06.stdout: python3-cephfs x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 514 k 2026-03-09T20:57:06.068 INFO:teuthology.orchestra.run.vm06.stdout:Removing unused dependencies: 2026-03-09T20:57:06.068 INFO:teuthology.orchestra.run.vm06.stdout: python3-ceph-argparse x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 187 k 2026-03-09T20:57:06.068 INFO:teuthology.orchestra.run.vm06.stdout: 2026-03-09T20:57:06.068 INFO:teuthology.orchestra.run.vm06.stdout:Transaction Summary 2026-03-09T20:57:06.068 INFO:teuthology.orchestra.run.vm06.stdout:================================================================================ 2026-03-09T20:57:06.068 INFO:teuthology.orchestra.run.vm06.stdout:Remove 3 Packages 2026-03-09T20:57:06.068 INFO:teuthology.orchestra.run.vm06.stdout: 2026-03-09T20:57:06.068 INFO:teuthology.orchestra.run.vm06.stdout:Freed space: 3.7 M 2026-03-09T20:57:06.068 INFO:teuthology.orchestra.run.vm06.stdout:Running transaction check 2026-03-09T20:57:06.070 INFO:teuthology.orchestra.run.vm06.stdout:Transaction check succeeded. 2026-03-09T20:57:06.070 INFO:teuthology.orchestra.run.vm06.stdout:Running transaction test 2026-03-09T20:57:06.086 INFO:teuthology.orchestra.run.vm06.stdout:Transaction test succeeded. 2026-03-09T20:57:06.086 INFO:teuthology.orchestra.run.vm06.stdout:Running transaction 2026-03-09T20:57:06.123 INFO:teuthology.orchestra.run.vm01.stdout:Dependencies resolved. 2026-03-09T20:57:06.123 INFO:teuthology.orchestra.run.vm01.stdout:================================================================================ 2026-03-09T20:57:06.123 INFO:teuthology.orchestra.run.vm01.stdout: Package Arch Version Repo Size 2026-03-09T20:57:06.123 INFO:teuthology.orchestra.run.vm01.stdout:================================================================================ 2026-03-09T20:57:06.123 INFO:teuthology.orchestra.run.vm01.stdout:Removing: 2026-03-09T20:57:06.123 INFO:teuthology.orchestra.run.vm01.stdout: libcephfs2 x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 3.0 M 2026-03-09T20:57:06.123 INFO:teuthology.orchestra.run.vm01.stdout:Removing dependent packages: 2026-03-09T20:57:06.123 INFO:teuthology.orchestra.run.vm01.stdout: python3-cephfs x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 514 k 2026-03-09T20:57:06.123 INFO:teuthology.orchestra.run.vm01.stdout:Removing unused dependencies: 2026-03-09T20:57:06.123 INFO:teuthology.orchestra.run.vm01.stdout: python3-ceph-argparse x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 187 k 2026-03-09T20:57:06.123 INFO:teuthology.orchestra.run.vm01.stdout: 2026-03-09T20:57:06.123 INFO:teuthology.orchestra.run.vm01.stdout:Transaction Summary 2026-03-09T20:57:06.123 INFO:teuthology.orchestra.run.vm01.stdout:================================================================================ 2026-03-09T20:57:06.123 INFO:teuthology.orchestra.run.vm01.stdout:Remove 3 Packages 2026-03-09T20:57:06.123 INFO:teuthology.orchestra.run.vm01.stdout: 2026-03-09T20:57:06.123 INFO:teuthology.orchestra.run.vm01.stdout:Freed space: 3.7 M 2026-03-09T20:57:06.123 INFO:teuthology.orchestra.run.vm01.stdout:Running transaction check 2026-03-09T20:57:06.125 INFO:teuthology.orchestra.run.vm08.stdout: Preparing : 1/1 2026-03-09T20:57:06.125 INFO:teuthology.orchestra.run.vm01.stdout:Transaction check succeeded. 2026-03-09T20:57:06.125 INFO:teuthology.orchestra.run.vm01.stdout:Running transaction test 2026-03-09T20:57:06.139 INFO:teuthology.orchestra.run.vm08.stdout: Erasing : python3-cephfs-2:19.2.3-678.ge911bdeb.el9.x86_64 1/3 2026-03-09T20:57:06.140 INFO:teuthology.orchestra.run.vm01.stdout:Transaction test succeeded. 2026-03-09T20:57:06.141 INFO:teuthology.orchestra.run.vm08.stdout: Erasing : python3-ceph-argparse-2:19.2.3-678.ge911bdeb.el9.x86 2/3 2026-03-09T20:57:06.141 INFO:teuthology.orchestra.run.vm08.stdout: Erasing : libcephfs2-2:19.2.3-678.ge911bdeb.el9.x86_64 3/3 2026-03-09T20:57:06.141 INFO:teuthology.orchestra.run.vm01.stdout:Running transaction 2026-03-09T20:57:06.166 INFO:teuthology.orchestra.run.vm06.stdout: Preparing : 1/1 2026-03-09T20:57:06.168 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : python3-cephfs-2:19.2.3-678.ge911bdeb.el9.x86_64 1/3 2026-03-09T20:57:06.169 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : python3-ceph-argparse-2:19.2.3-678.ge911bdeb.el9.x86 2/3 2026-03-09T20:57:06.169 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : libcephfs2-2:19.2.3-678.ge911bdeb.el9.x86_64 3/3 2026-03-09T20:57:06.172 INFO:teuthology.orchestra.run.vm01.stdout: Preparing : 1/1 2026-03-09T20:57:06.175 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : python3-cephfs-2:19.2.3-678.ge911bdeb.el9.x86_64 1/3 2026-03-09T20:57:06.176 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : python3-ceph-argparse-2:19.2.3-678.ge911bdeb.el9.x86 2/3 2026-03-09T20:57:06.177 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : libcephfs2-2:19.2.3-678.ge911bdeb.el9.x86_64 3/3 2026-03-09T20:57:06.202 INFO:teuthology.orchestra.run.vm08.stdout: Running scriptlet: libcephfs2-2:19.2.3-678.ge911bdeb.el9.x86_64 3/3 2026-03-09T20:57:06.202 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : libcephfs2-2:19.2.3-678.ge911bdeb.el9.x86_64 1/3 2026-03-09T20:57:06.202 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : python3-ceph-argparse-2:19.2.3-678.ge911bdeb.el9.x86 2/3 2026-03-09T20:57:06.237 INFO:teuthology.orchestra.run.vm06.stdout: Running scriptlet: libcephfs2-2:19.2.3-678.ge911bdeb.el9.x86_64 3/3 2026-03-09T20:57:06.238 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : libcephfs2-2:19.2.3-678.ge911bdeb.el9.x86_64 1/3 2026-03-09T20:57:06.238 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-ceph-argparse-2:19.2.3-678.ge911bdeb.el9.x86 2/3 2026-03-09T20:57:06.240 INFO:teuthology.orchestra.run.vm01.stdout: Running scriptlet: libcephfs2-2:19.2.3-678.ge911bdeb.el9.x86_64 3/3 2026-03-09T20:57:06.240 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : libcephfs2-2:19.2.3-678.ge911bdeb.el9.x86_64 1/3 2026-03-09T20:57:06.240 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-ceph-argparse-2:19.2.3-678.ge911bdeb.el9.x86 2/3 2026-03-09T20:57:06.254 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : python3-cephfs-2:19.2.3-678.ge911bdeb.el9.x86_64 3/3 2026-03-09T20:57:06.254 INFO:teuthology.orchestra.run.vm08.stdout: 2026-03-09T20:57:06.254 INFO:teuthology.orchestra.run.vm08.stdout:Removed: 2026-03-09T20:57:06.254 INFO:teuthology.orchestra.run.vm08.stdout: libcephfs2-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T20:57:06.254 INFO:teuthology.orchestra.run.vm08.stdout: python3-ceph-argparse-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T20:57:06.254 INFO:teuthology.orchestra.run.vm08.stdout: python3-cephfs-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T20:57:06.254 INFO:teuthology.orchestra.run.vm08.stdout: 2026-03-09T20:57:06.254 INFO:teuthology.orchestra.run.vm08.stdout:Complete! 2026-03-09T20:57:06.326 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-cephfs-2:19.2.3-678.ge911bdeb.el9.x86_64 3/3 2026-03-09T20:57:06.326 INFO:teuthology.orchestra.run.vm06.stdout: 2026-03-09T20:57:06.326 INFO:teuthology.orchestra.run.vm06.stdout:Removed: 2026-03-09T20:57:06.326 INFO:teuthology.orchestra.run.vm06.stdout: libcephfs2-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T20:57:06.326 INFO:teuthology.orchestra.run.vm06.stdout: python3-ceph-argparse-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T20:57:06.326 INFO:teuthology.orchestra.run.vm06.stdout: python3-cephfs-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T20:57:06.326 INFO:teuthology.orchestra.run.vm06.stdout: 2026-03-09T20:57:06.326 INFO:teuthology.orchestra.run.vm06.stdout:Complete! 2026-03-09T20:57:06.326 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-cephfs-2:19.2.3-678.ge911bdeb.el9.x86_64 3/3 2026-03-09T20:57:06.326 INFO:teuthology.orchestra.run.vm01.stdout: 2026-03-09T20:57:06.326 INFO:teuthology.orchestra.run.vm01.stdout:Removed: 2026-03-09T20:57:06.326 INFO:teuthology.orchestra.run.vm01.stdout: libcephfs2-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T20:57:06.326 INFO:teuthology.orchestra.run.vm01.stdout: python3-ceph-argparse-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T20:57:06.326 INFO:teuthology.orchestra.run.vm01.stdout: python3-cephfs-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T20:57:06.326 INFO:teuthology.orchestra.run.vm01.stdout: 2026-03-09T20:57:06.326 INFO:teuthology.orchestra.run.vm01.stdout:Complete! 2026-03-09T20:57:06.499 INFO:teuthology.orchestra.run.vm08.stdout:No match for argument: libcephfs-devel 2026-03-09T20:57:06.499 INFO:teuthology.orchestra.run.vm08.stderr:No packages marked for removal. 2026-03-09T20:57:06.503 INFO:teuthology.orchestra.run.vm08.stdout:Dependencies resolved. 2026-03-09T20:57:06.504 INFO:teuthology.orchestra.run.vm08.stdout:Nothing to do. 2026-03-09T20:57:06.504 INFO:teuthology.orchestra.run.vm08.stdout:Complete! 2026-03-09T20:57:06.523 INFO:teuthology.orchestra.run.vm01.stdout:No match for argument: libcephfs-devel 2026-03-09T20:57:06.523 INFO:teuthology.orchestra.run.vm01.stderr:No packages marked for removal. 2026-03-09T20:57:06.526 INFO:teuthology.orchestra.run.vm01.stdout:Dependencies resolved. 2026-03-09T20:57:06.527 INFO:teuthology.orchestra.run.vm01.stdout:Nothing to do. 2026-03-09T20:57:06.527 INFO:teuthology.orchestra.run.vm01.stdout:Complete! 2026-03-09T20:57:06.535 INFO:teuthology.orchestra.run.vm06.stdout:No match for argument: libcephfs-devel 2026-03-09T20:57:06.535 INFO:teuthology.orchestra.run.vm06.stderr:No packages marked for removal. 2026-03-09T20:57:06.539 INFO:teuthology.orchestra.run.vm06.stdout:Dependencies resolved. 2026-03-09T20:57:06.539 INFO:teuthology.orchestra.run.vm06.stdout:Nothing to do. 2026-03-09T20:57:06.539 INFO:teuthology.orchestra.run.vm06.stdout:Complete! 2026-03-09T20:57:06.703 INFO:teuthology.orchestra.run.vm08.stdout:Dependencies resolved. 2026-03-09T20:57:06.705 INFO:teuthology.orchestra.run.vm08.stdout:================================================================================ 2026-03-09T20:57:06.705 INFO:teuthology.orchestra.run.vm08.stdout: Package Arch Version Repository Size 2026-03-09T20:57:06.705 INFO:teuthology.orchestra.run.vm08.stdout:================================================================================ 2026-03-09T20:57:06.705 INFO:teuthology.orchestra.run.vm08.stdout:Removing: 2026-03-09T20:57:06.705 INFO:teuthology.orchestra.run.vm08.stdout: librados2 x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 12 M 2026-03-09T20:57:06.705 INFO:teuthology.orchestra.run.vm08.stdout:Removing dependent packages: 2026-03-09T20:57:06.705 INFO:teuthology.orchestra.run.vm08.stdout: python3-rados x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 1.1 M 2026-03-09T20:57:06.705 INFO:teuthology.orchestra.run.vm08.stdout: python3-rbd x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 1.1 M 2026-03-09T20:57:06.705 INFO:teuthology.orchestra.run.vm08.stdout: python3-rgw x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 265 k 2026-03-09T20:57:06.705 INFO:teuthology.orchestra.run.vm08.stdout: qemu-kvm-block-rbd x86_64 17:10.1.0-15.el9 @appstream 37 k 2026-03-09T20:57:06.705 INFO:teuthology.orchestra.run.vm08.stdout: rbd-fuse x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 227 k 2026-03-09T20:57:06.705 INFO:teuthology.orchestra.run.vm08.stdout: rbd-nbd x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 490 k 2026-03-09T20:57:06.705 INFO:teuthology.orchestra.run.vm08.stdout:Removing unused dependencies: 2026-03-09T20:57:06.705 INFO:teuthology.orchestra.run.vm08.stdout: boost-program-options x86_64 1.75.0-13.el9 @appstream 276 k 2026-03-09T20:57:06.705 INFO:teuthology.orchestra.run.vm08.stdout: libarrow x86_64 9.0.0-15.el9 @epel 18 M 2026-03-09T20:57:06.705 INFO:teuthology.orchestra.run.vm08.stdout: libarrow-doc noarch 9.0.0-15.el9 @epel 122 k 2026-03-09T20:57:06.705 INFO:teuthology.orchestra.run.vm08.stdout: libnbd x86_64 1.20.3-4.el9 @appstream 453 k 2026-03-09T20:57:06.705 INFO:teuthology.orchestra.run.vm08.stdout: libpmemobj x86_64 1.12.1-1.el9 @appstream 383 k 2026-03-09T20:57:06.705 INFO:teuthology.orchestra.run.vm08.stdout: librabbitmq x86_64 0.11.0-7.el9 @appstream 102 k 2026-03-09T20:57:06.705 INFO:teuthology.orchestra.run.vm08.stdout: librbd1 x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 13 M 2026-03-09T20:57:06.705 INFO:teuthology.orchestra.run.vm08.stdout: librdkafka x86_64 1.6.1-102.el9 @appstream 2.0 M 2026-03-09T20:57:06.705 INFO:teuthology.orchestra.run.vm08.stdout: librgw2 x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 19 M 2026-03-09T20:57:06.705 INFO:teuthology.orchestra.run.vm08.stdout: lttng-ust x86_64 2.12.0-6.el9 @appstream 1.0 M 2026-03-09T20:57:06.705 INFO:teuthology.orchestra.run.vm08.stdout: parquet-libs x86_64 9.0.0-15.el9 @epel 2.8 M 2026-03-09T20:57:06.705 INFO:teuthology.orchestra.run.vm08.stdout: re2 x86_64 1:20211101-20.el9 @epel 472 k 2026-03-09T20:57:06.705 INFO:teuthology.orchestra.run.vm08.stdout: thrift x86_64 0.15.0-4.el9 @epel 4.8 M 2026-03-09T20:57:06.705 INFO:teuthology.orchestra.run.vm08.stdout: 2026-03-09T20:57:06.706 INFO:teuthology.orchestra.run.vm08.stdout:Transaction Summary 2026-03-09T20:57:06.706 INFO:teuthology.orchestra.run.vm08.stdout:================================================================================ 2026-03-09T20:57:06.706 INFO:teuthology.orchestra.run.vm08.stdout:Remove 20 Packages 2026-03-09T20:57:06.706 INFO:teuthology.orchestra.run.vm08.stdout: 2026-03-09T20:57:06.706 INFO:teuthology.orchestra.run.vm08.stdout:Freed space: 79 M 2026-03-09T20:57:06.706 INFO:teuthology.orchestra.run.vm08.stdout:Running transaction check 2026-03-09T20:57:06.710 INFO:teuthology.orchestra.run.vm08.stdout:Transaction check succeeded. 2026-03-09T20:57:06.710 INFO:teuthology.orchestra.run.vm08.stdout:Running transaction test 2026-03-09T20:57:06.725 INFO:teuthology.orchestra.run.vm01.stdout:Dependencies resolved. 2026-03-09T20:57:06.726 INFO:teuthology.orchestra.run.vm01.stdout:================================================================================ 2026-03-09T20:57:06.726 INFO:teuthology.orchestra.run.vm01.stdout: Package Arch Version Repository Size 2026-03-09T20:57:06.726 INFO:teuthology.orchestra.run.vm01.stdout:================================================================================ 2026-03-09T20:57:06.727 INFO:teuthology.orchestra.run.vm01.stdout:Removing: 2026-03-09T20:57:06.727 INFO:teuthology.orchestra.run.vm01.stdout: librados2 x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 12 M 2026-03-09T20:57:06.727 INFO:teuthology.orchestra.run.vm01.stdout:Removing dependent packages: 2026-03-09T20:57:06.727 INFO:teuthology.orchestra.run.vm01.stdout: python3-rados x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 1.1 M 2026-03-09T20:57:06.727 INFO:teuthology.orchestra.run.vm01.stdout: python3-rbd x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 1.1 M 2026-03-09T20:57:06.727 INFO:teuthology.orchestra.run.vm01.stdout: python3-rgw x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 265 k 2026-03-09T20:57:06.727 INFO:teuthology.orchestra.run.vm01.stdout: qemu-kvm-block-rbd x86_64 17:10.1.0-15.el9 @appstream 37 k 2026-03-09T20:57:06.727 INFO:teuthology.orchestra.run.vm01.stdout: rbd-fuse x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 227 k 2026-03-09T20:57:06.727 INFO:teuthology.orchestra.run.vm01.stdout: rbd-nbd x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 490 k 2026-03-09T20:57:06.727 INFO:teuthology.orchestra.run.vm01.stdout:Removing unused dependencies: 2026-03-09T20:57:06.727 INFO:teuthology.orchestra.run.vm01.stdout: boost-program-options x86_64 1.75.0-13.el9 @appstream 276 k 2026-03-09T20:57:06.727 INFO:teuthology.orchestra.run.vm01.stdout: libarrow x86_64 9.0.0-15.el9 @epel 18 M 2026-03-09T20:57:06.727 INFO:teuthology.orchestra.run.vm01.stdout: libarrow-doc noarch 9.0.0-15.el9 @epel 122 k 2026-03-09T20:57:06.727 INFO:teuthology.orchestra.run.vm01.stdout: libnbd x86_64 1.20.3-4.el9 @appstream 453 k 2026-03-09T20:57:06.727 INFO:teuthology.orchestra.run.vm01.stdout: libpmemobj x86_64 1.12.1-1.el9 @appstream 383 k 2026-03-09T20:57:06.727 INFO:teuthology.orchestra.run.vm01.stdout: librabbitmq x86_64 0.11.0-7.el9 @appstream 102 k 2026-03-09T20:57:06.727 INFO:teuthology.orchestra.run.vm01.stdout: librbd1 x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 13 M 2026-03-09T20:57:06.727 INFO:teuthology.orchestra.run.vm01.stdout: librdkafka x86_64 1.6.1-102.el9 @appstream 2.0 M 2026-03-09T20:57:06.727 INFO:teuthology.orchestra.run.vm01.stdout: librgw2 x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 19 M 2026-03-09T20:57:06.727 INFO:teuthology.orchestra.run.vm01.stdout: lttng-ust x86_64 2.12.0-6.el9 @appstream 1.0 M 2026-03-09T20:57:06.727 INFO:teuthology.orchestra.run.vm01.stdout: parquet-libs x86_64 9.0.0-15.el9 @epel 2.8 M 2026-03-09T20:57:06.727 INFO:teuthology.orchestra.run.vm01.stdout: re2 x86_64 1:20211101-20.el9 @epel 472 k 2026-03-09T20:57:06.727 INFO:teuthology.orchestra.run.vm01.stdout: thrift x86_64 0.15.0-4.el9 @epel 4.8 M 2026-03-09T20:57:06.727 INFO:teuthology.orchestra.run.vm01.stdout: 2026-03-09T20:57:06.727 INFO:teuthology.orchestra.run.vm01.stdout:Transaction Summary 2026-03-09T20:57:06.727 INFO:teuthology.orchestra.run.vm01.stdout:================================================================================ 2026-03-09T20:57:06.727 INFO:teuthology.orchestra.run.vm01.stdout:Remove 20 Packages 2026-03-09T20:57:06.727 INFO:teuthology.orchestra.run.vm01.stdout: 2026-03-09T20:57:06.727 INFO:teuthology.orchestra.run.vm01.stdout:Freed space: 79 M 2026-03-09T20:57:06.727 INFO:teuthology.orchestra.run.vm01.stdout:Running transaction check 2026-03-09T20:57:06.731 INFO:teuthology.orchestra.run.vm01.stdout:Transaction check succeeded. 2026-03-09T20:57:06.731 INFO:teuthology.orchestra.run.vm01.stdout:Running transaction test 2026-03-09T20:57:06.734 INFO:teuthology.orchestra.run.vm08.stdout:Transaction test succeeded. 2026-03-09T20:57:06.734 INFO:teuthology.orchestra.run.vm08.stdout:Running transaction 2026-03-09T20:57:06.735 INFO:teuthology.orchestra.run.vm06.stdout:Dependencies resolved. 2026-03-09T20:57:06.737 INFO:teuthology.orchestra.run.vm06.stdout:================================================================================ 2026-03-09T20:57:06.737 INFO:teuthology.orchestra.run.vm06.stdout: Package Arch Version Repository Size 2026-03-09T20:57:06.737 INFO:teuthology.orchestra.run.vm06.stdout:================================================================================ 2026-03-09T20:57:06.737 INFO:teuthology.orchestra.run.vm06.stdout:Removing: 2026-03-09T20:57:06.737 INFO:teuthology.orchestra.run.vm06.stdout: librados2 x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 12 M 2026-03-09T20:57:06.737 INFO:teuthology.orchestra.run.vm06.stdout:Removing dependent packages: 2026-03-09T20:57:06.737 INFO:teuthology.orchestra.run.vm06.stdout: python3-rados x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 1.1 M 2026-03-09T20:57:06.737 INFO:teuthology.orchestra.run.vm06.stdout: python3-rbd x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 1.1 M 2026-03-09T20:57:06.737 INFO:teuthology.orchestra.run.vm06.stdout: python3-rgw x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 265 k 2026-03-09T20:57:06.737 INFO:teuthology.orchestra.run.vm06.stdout: qemu-kvm-block-rbd x86_64 17:10.1.0-15.el9 @appstream 37 k 2026-03-09T20:57:06.738 INFO:teuthology.orchestra.run.vm06.stdout: rbd-fuse x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 227 k 2026-03-09T20:57:06.738 INFO:teuthology.orchestra.run.vm06.stdout: rbd-nbd x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 490 k 2026-03-09T20:57:06.738 INFO:teuthology.orchestra.run.vm06.stdout:Removing unused dependencies: 2026-03-09T20:57:06.738 INFO:teuthology.orchestra.run.vm06.stdout: boost-program-options x86_64 1.75.0-13.el9 @appstream 276 k 2026-03-09T20:57:06.738 INFO:teuthology.orchestra.run.vm06.stdout: libarrow x86_64 9.0.0-15.el9 @epel 18 M 2026-03-09T20:57:06.738 INFO:teuthology.orchestra.run.vm06.stdout: libarrow-doc noarch 9.0.0-15.el9 @epel 122 k 2026-03-09T20:57:06.738 INFO:teuthology.orchestra.run.vm06.stdout: libnbd x86_64 1.20.3-4.el9 @appstream 453 k 2026-03-09T20:57:06.738 INFO:teuthology.orchestra.run.vm06.stdout: libpmemobj x86_64 1.12.1-1.el9 @appstream 383 k 2026-03-09T20:57:06.738 INFO:teuthology.orchestra.run.vm06.stdout: librabbitmq x86_64 0.11.0-7.el9 @appstream 102 k 2026-03-09T20:57:06.738 INFO:teuthology.orchestra.run.vm06.stdout: librbd1 x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 13 M 2026-03-09T20:57:06.738 INFO:teuthology.orchestra.run.vm06.stdout: librdkafka x86_64 1.6.1-102.el9 @appstream 2.0 M 2026-03-09T20:57:06.738 INFO:teuthology.orchestra.run.vm06.stdout: librgw2 x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 19 M 2026-03-09T20:57:06.738 INFO:teuthology.orchestra.run.vm06.stdout: lttng-ust x86_64 2.12.0-6.el9 @appstream 1.0 M 2026-03-09T20:57:06.738 INFO:teuthology.orchestra.run.vm06.stdout: parquet-libs x86_64 9.0.0-15.el9 @epel 2.8 M 2026-03-09T20:57:06.738 INFO:teuthology.orchestra.run.vm06.stdout: re2 x86_64 1:20211101-20.el9 @epel 472 k 2026-03-09T20:57:06.738 INFO:teuthology.orchestra.run.vm06.stdout: thrift x86_64 0.15.0-4.el9 @epel 4.8 M 2026-03-09T20:57:06.738 INFO:teuthology.orchestra.run.vm06.stdout: 2026-03-09T20:57:06.738 INFO:teuthology.orchestra.run.vm06.stdout:Transaction Summary 2026-03-09T20:57:06.738 INFO:teuthology.orchestra.run.vm06.stdout:================================================================================ 2026-03-09T20:57:06.738 INFO:teuthology.orchestra.run.vm06.stdout:Remove 20 Packages 2026-03-09T20:57:06.738 INFO:teuthology.orchestra.run.vm06.stdout: 2026-03-09T20:57:06.738 INFO:teuthology.orchestra.run.vm06.stdout:Freed space: 79 M 2026-03-09T20:57:06.738 INFO:teuthology.orchestra.run.vm06.stdout:Running transaction check 2026-03-09T20:57:06.742 INFO:teuthology.orchestra.run.vm06.stdout:Transaction check succeeded. 2026-03-09T20:57:06.742 INFO:teuthology.orchestra.run.vm06.stdout:Running transaction test 2026-03-09T20:57:06.756 INFO:teuthology.orchestra.run.vm01.stdout:Transaction test succeeded. 2026-03-09T20:57:06.756 INFO:teuthology.orchestra.run.vm01.stdout:Running transaction 2026-03-09T20:57:06.766 INFO:teuthology.orchestra.run.vm06.stdout:Transaction test succeeded. 2026-03-09T20:57:06.767 INFO:teuthology.orchestra.run.vm06.stdout:Running transaction 2026-03-09T20:57:06.777 INFO:teuthology.orchestra.run.vm08.stdout: Preparing : 1/1 2026-03-09T20:57:06.780 INFO:teuthology.orchestra.run.vm08.stdout: Erasing : rbd-nbd-2:19.2.3-678.ge911bdeb.el9.x86_64 1/20 2026-03-09T20:57:06.782 INFO:teuthology.orchestra.run.vm08.stdout: Erasing : rbd-fuse-2:19.2.3-678.ge911bdeb.el9.x86_64 2/20 2026-03-09T20:57:06.786 INFO:teuthology.orchestra.run.vm08.stdout: Erasing : python3-rgw-2:19.2.3-678.ge911bdeb.el9.x86_64 3/20 2026-03-09T20:57:06.786 INFO:teuthology.orchestra.run.vm08.stdout: Erasing : librgw2-2:19.2.3-678.ge911bdeb.el9.x86_64 4/20 2026-03-09T20:57:06.799 INFO:teuthology.orchestra.run.vm01.stdout: Preparing : 1/1 2026-03-09T20:57:06.800 INFO:teuthology.orchestra.run.vm08.stdout: Running scriptlet: librgw2-2:19.2.3-678.ge911bdeb.el9.x86_64 4/20 2026-03-09T20:57:06.806 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : rbd-nbd-2:19.2.3-678.ge911bdeb.el9.x86_64 1/20 2026-03-09T20:57:06.810 INFO:teuthology.orchestra.run.vm06.stdout: Preparing : 1/1 2026-03-09T20:57:06.823 INFO:teuthology.orchestra.run.vm08.stdout: Erasing : parquet-libs-9.0.0-15.el9.x86_64 5/20 2026-03-09T20:57:06.830 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : rbd-fuse-2:19.2.3-678.ge911bdeb.el9.x86_64 2/20 2026-03-09T20:57:06.851 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : python3-rgw-2:19.2.3-678.ge911bdeb.el9.x86_64 3/20 2026-03-09T20:57:06.851 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : librgw2-2:19.2.3-678.ge911bdeb.el9.x86_64 4/20 2026-03-09T20:57:06.851 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : rbd-nbd-2:19.2.3-678.ge911bdeb.el9.x86_64 1/20 2026-03-09T20:57:06.851 INFO:teuthology.orchestra.run.vm08.stdout: Erasing : python3-rbd-2:19.2.3-678.ge911bdeb.el9.x86_64 6/20 2026-03-09T20:57:06.901 INFO:teuthology.orchestra.run.vm08.stdout: Erasing : python3-rados-2:19.2.3-678.ge911bdeb.el9.x86_64 7/20 2026-03-09T20:57:06.919 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : rbd-fuse-2:19.2.3-678.ge911bdeb.el9.x86_64 2/20 2026-03-09T20:57:06.928 INFO:teuthology.orchestra.run.vm08.stdout: Erasing : qemu-kvm-block-rbd-17:10.1.0-15.el9.x86_64 8/20 2026-03-09T20:57:06.955 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : python3-rgw-2:19.2.3-678.ge911bdeb.el9.x86_64 3/20 2026-03-09T20:57:06.967 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : librgw2-2:19.2.3-678.ge911bdeb.el9.x86_64 4/20 2026-03-09T20:57:06.969 INFO:teuthology.orchestra.run.vm01.stdout: Running scriptlet: librgw2-2:19.2.3-678.ge911bdeb.el9.x86_64 4/20 2026-03-09T20:57:06.970 INFO:teuthology.orchestra.run.vm08.stdout: Erasing : libarrow-doc-9.0.0-15.el9.noarch 9/20 2026-03-09T20:57:06.970 INFO:teuthology.orchestra.run.vm08.stdout: Erasing : librbd1-2:19.2.3-678.ge911bdeb.el9.x86_64 10/20 2026-03-09T20:57:06.975 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : parquet-libs-9.0.0-15.el9.x86_64 5/20 2026-03-09T20:57:06.978 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : python3-rbd-2:19.2.3-678.ge911bdeb.el9.x86_64 6/20 2026-03-09T20:57:06.979 INFO:teuthology.orchestra.run.vm06.stdout: Running scriptlet: librgw2-2:19.2.3-678.ge911bdeb.el9.x86_64 4/20 2026-03-09T20:57:06.980 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : python3-rados-2:19.2.3-678.ge911bdeb.el9.x86_64 7/20 2026-03-09T20:57:06.982 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : qemu-kvm-block-rbd-17:10.1.0-15.el9.x86_64 8/20 2026-03-09T20:57:06.982 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : parquet-libs-9.0.0-15.el9.x86_64 5/20 2026-03-09T20:57:06.985 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : python3-rbd-2:19.2.3-678.ge911bdeb.el9.x86_64 6/20 2026-03-09T20:57:06.986 INFO:teuthology.orchestra.run.vm08.stdout: Running scriptlet: librbd1-2:19.2.3-678.ge911bdeb.el9.x86_64 10/20 2026-03-09T20:57:06.986 INFO:teuthology.orchestra.run.vm08.stdout: Erasing : librados2-2:19.2.3-678.ge911bdeb.el9.x86_64 11/20 2026-03-09T20:57:06.986 INFO:teuthology.orchestra.run.vm08.stdout:warning: file /etc/ceph: remove failed: No such file or directory 2026-03-09T20:57:06.986 INFO:teuthology.orchestra.run.vm08.stdout: 2026-03-09T20:57:06.986 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : libarrow-doc-9.0.0-15.el9.noarch 9/20 2026-03-09T20:57:06.986 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : librbd1-2:19.2.3-678.ge911bdeb.el9.x86_64 10/20 2026-03-09T20:57:06.987 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : python3-rados-2:19.2.3-678.ge911bdeb.el9.x86_64 7/20 2026-03-09T20:57:06.990 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : qemu-kvm-block-rbd-17:10.1.0-15.el9.x86_64 8/20 2026-03-09T20:57:06.993 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : libarrow-doc-9.0.0-15.el9.noarch 9/20 2026-03-09T20:57:06.993 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : librbd1-2:19.2.3-678.ge911bdeb.el9.x86_64 10/20 2026-03-09T20:57:07.001 INFO:teuthology.orchestra.run.vm08.stdout: Running scriptlet: librados2-2:19.2.3-678.ge911bdeb.el9.x86_64 11/20 2026-03-09T20:57:07.003 INFO:teuthology.orchestra.run.vm01.stdout: Running scriptlet: librbd1-2:19.2.3-678.ge911bdeb.el9.x86_64 10/20 2026-03-09T20:57:07.003 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : librados2-2:19.2.3-678.ge911bdeb.el9.x86_64 11/20 2026-03-09T20:57:07.003 INFO:teuthology.orchestra.run.vm01.stdout:warning: file /etc/ceph: remove failed: No such file or directory 2026-03-09T20:57:07.003 INFO:teuthology.orchestra.run.vm01.stdout: 2026-03-09T20:57:07.003 INFO:teuthology.orchestra.run.vm08.stdout: Erasing : libarrow-9.0.0-15.el9.x86_64 12/20 2026-03-09T20:57:07.007 INFO:teuthology.orchestra.run.vm08.stdout: Erasing : re2-1:20211101-20.el9.x86_64 13/20 2026-03-09T20:57:07.009 INFO:teuthology.orchestra.run.vm06.stdout: Running scriptlet: librbd1-2:19.2.3-678.ge911bdeb.el9.x86_64 10/20 2026-03-09T20:57:07.009 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : librados2-2:19.2.3-678.ge911bdeb.el9.x86_64 11/20 2026-03-09T20:57:07.009 INFO:teuthology.orchestra.run.vm06.stdout:warning: file /etc/ceph: remove failed: No such file or directory 2026-03-09T20:57:07.009 INFO:teuthology.orchestra.run.vm06.stdout: 2026-03-09T20:57:07.011 INFO:teuthology.orchestra.run.vm08.stdout: Erasing : lttng-ust-2.12.0-6.el9.x86_64 14/20 2026-03-09T20:57:07.014 INFO:teuthology.orchestra.run.vm08.stdout: Erasing : thrift-0.15.0-4.el9.x86_64 15/20 2026-03-09T20:57:07.017 INFO:teuthology.orchestra.run.vm08.stdout: Erasing : libnbd-1.20.3-4.el9.x86_64 16/20 2026-03-09T20:57:07.019 INFO:teuthology.orchestra.run.vm08.stdout: Erasing : libpmemobj-1.12.1-1.el9.x86_64 17/20 2026-03-09T20:57:07.019 INFO:teuthology.orchestra.run.vm01.stdout: Running scriptlet: librados2-2:19.2.3-678.ge911bdeb.el9.x86_64 11/20 2026-03-09T20:57:07.021 INFO:teuthology.orchestra.run.vm08.stdout: Erasing : boost-program-options-1.75.0-13.el9.x86_64 18/20 2026-03-09T20:57:07.022 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : libarrow-9.0.0-15.el9.x86_64 12/20 2026-03-09T20:57:07.023 INFO:teuthology.orchestra.run.vm08.stdout: Erasing : librabbitmq-0.11.0-7.el9.x86_64 19/20 2026-03-09T20:57:07.025 INFO:teuthology.orchestra.run.vm06.stdout: Running scriptlet: librados2-2:19.2.3-678.ge911bdeb.el9.x86_64 11/20 2026-03-09T20:57:07.025 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : re2-1:20211101-20.el9.x86_64 13/20 2026-03-09T20:57:07.028 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : libarrow-9.0.0-15.el9.x86_64 12/20 2026-03-09T20:57:07.029 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : lttng-ust-2.12.0-6.el9.x86_64 14/20 2026-03-09T20:57:07.032 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : thrift-0.15.0-4.el9.x86_64 15/20 2026-03-09T20:57:07.032 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : re2-1:20211101-20.el9.x86_64 13/20 2026-03-09T20:57:07.035 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : libnbd-1.20.3-4.el9.x86_64 16/20 2026-03-09T20:57:07.036 INFO:teuthology.orchestra.run.vm08.stdout: Erasing : librdkafka-1.6.1-102.el9.x86_64 20/20 2026-03-09T20:57:07.036 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : lttng-ust-2.12.0-6.el9.x86_64 14/20 2026-03-09T20:57:07.037 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : libpmemobj-1.12.1-1.el9.x86_64 17/20 2026-03-09T20:57:07.039 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : thrift-0.15.0-4.el9.x86_64 15/20 2026-03-09T20:57:07.039 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : boost-program-options-1.75.0-13.el9.x86_64 18/20 2026-03-09T20:57:07.042 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : librabbitmq-0.11.0-7.el9.x86_64 19/20 2026-03-09T20:57:07.042 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : libnbd-1.20.3-4.el9.x86_64 16/20 2026-03-09T20:57:07.045 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : libpmemobj-1.12.1-1.el9.x86_64 17/20 2026-03-09T20:57:07.047 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : boost-program-options-1.75.0-13.el9.x86_64 18/20 2026-03-09T20:57:07.049 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : librabbitmq-0.11.0-7.el9.x86_64 19/20 2026-03-09T20:57:07.057 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : librdkafka-1.6.1-102.el9.x86_64 20/20 2026-03-09T20:57:07.063 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : librdkafka-1.6.1-102.el9.x86_64 20/20 2026-03-09T20:57:07.097 INFO:teuthology.orchestra.run.vm08.stdout: Running scriptlet: librdkafka-1.6.1-102.el9.x86_64 20/20 2026-03-09T20:57:07.097 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : boost-program-options-1.75.0-13.el9.x86_64 1/20 2026-03-09T20:57:07.097 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : libarrow-9.0.0-15.el9.x86_64 2/20 2026-03-09T20:57:07.097 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : libarrow-doc-9.0.0-15.el9.noarch 3/20 2026-03-09T20:57:07.097 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : libnbd-1.20.3-4.el9.x86_64 4/20 2026-03-09T20:57:07.097 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : libpmemobj-1.12.1-1.el9.x86_64 5/20 2026-03-09T20:57:07.097 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : librabbitmq-0.11.0-7.el9.x86_64 6/20 2026-03-09T20:57:07.097 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : librados2-2:19.2.3-678.ge911bdeb.el9.x86_64 7/20 2026-03-09T20:57:07.097 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : librbd1-2:19.2.3-678.ge911bdeb.el9.x86_64 8/20 2026-03-09T20:57:07.097 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : librdkafka-1.6.1-102.el9.x86_64 9/20 2026-03-09T20:57:07.098 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : librgw2-2:19.2.3-678.ge911bdeb.el9.x86_64 10/20 2026-03-09T20:57:07.098 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : lttng-ust-2.12.0-6.el9.x86_64 11/20 2026-03-09T20:57:07.098 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : parquet-libs-9.0.0-15.el9.x86_64 12/20 2026-03-09T20:57:07.098 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : python3-rados-2:19.2.3-678.ge911bdeb.el9.x86_64 13/20 2026-03-09T20:57:07.098 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : python3-rbd-2:19.2.3-678.ge911bdeb.el9.x86_64 14/20 2026-03-09T20:57:07.098 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : python3-rgw-2:19.2.3-678.ge911bdeb.el9.x86_64 15/20 2026-03-09T20:57:07.098 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : qemu-kvm-block-rbd-17:10.1.0-15.el9.x86_64 16/20 2026-03-09T20:57:07.098 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : rbd-fuse-2:19.2.3-678.ge911bdeb.el9.x86_64 17/20 2026-03-09T20:57:07.098 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : rbd-nbd-2:19.2.3-678.ge911bdeb.el9.x86_64 18/20 2026-03-09T20:57:07.098 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : re2-1:20211101-20.el9.x86_64 19/20 2026-03-09T20:57:07.121 INFO:teuthology.orchestra.run.vm01.stdout: Running scriptlet: librdkafka-1.6.1-102.el9.x86_64 20/20 2026-03-09T20:57:07.121 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : boost-program-options-1.75.0-13.el9.x86_64 1/20 2026-03-09T20:57:07.121 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : libarrow-9.0.0-15.el9.x86_64 2/20 2026-03-09T20:57:07.121 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : libarrow-doc-9.0.0-15.el9.noarch 3/20 2026-03-09T20:57:07.121 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : libnbd-1.20.3-4.el9.x86_64 4/20 2026-03-09T20:57:07.121 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : libpmemobj-1.12.1-1.el9.x86_64 5/20 2026-03-09T20:57:07.121 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : librabbitmq-0.11.0-7.el9.x86_64 6/20 2026-03-09T20:57:07.121 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : librados2-2:19.2.3-678.ge911bdeb.el9.x86_64 7/20 2026-03-09T20:57:07.121 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : librbd1-2:19.2.3-678.ge911bdeb.el9.x86_64 8/20 2026-03-09T20:57:07.121 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : librdkafka-1.6.1-102.el9.x86_64 9/20 2026-03-09T20:57:07.121 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : librgw2-2:19.2.3-678.ge911bdeb.el9.x86_64 10/20 2026-03-09T20:57:07.121 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : lttng-ust-2.12.0-6.el9.x86_64 11/20 2026-03-09T20:57:07.121 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : parquet-libs-9.0.0-15.el9.x86_64 12/20 2026-03-09T20:57:07.121 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-rados-2:19.2.3-678.ge911bdeb.el9.x86_64 13/20 2026-03-09T20:57:07.121 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-rbd-2:19.2.3-678.ge911bdeb.el9.x86_64 14/20 2026-03-09T20:57:07.121 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-rgw-2:19.2.3-678.ge911bdeb.el9.x86_64 15/20 2026-03-09T20:57:07.121 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : qemu-kvm-block-rbd-17:10.1.0-15.el9.x86_64 16/20 2026-03-09T20:57:07.121 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : rbd-fuse-2:19.2.3-678.ge911bdeb.el9.x86_64 17/20 2026-03-09T20:57:07.121 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : rbd-nbd-2:19.2.3-678.ge911bdeb.el9.x86_64 18/20 2026-03-09T20:57:07.121 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : re2-1:20211101-20.el9.x86_64 19/20 2026-03-09T20:57:07.128 INFO:teuthology.orchestra.run.vm06.stdout: Running scriptlet: librdkafka-1.6.1-102.el9.x86_64 20/20 2026-03-09T20:57:07.128 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : boost-program-options-1.75.0-13.el9.x86_64 1/20 2026-03-09T20:57:07.128 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : libarrow-9.0.0-15.el9.x86_64 2/20 2026-03-09T20:57:07.128 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : libarrow-doc-9.0.0-15.el9.noarch 3/20 2026-03-09T20:57:07.128 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : libnbd-1.20.3-4.el9.x86_64 4/20 2026-03-09T20:57:07.128 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : libpmemobj-1.12.1-1.el9.x86_64 5/20 2026-03-09T20:57:07.128 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : librabbitmq-0.11.0-7.el9.x86_64 6/20 2026-03-09T20:57:07.128 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : librados2-2:19.2.3-678.ge911bdeb.el9.x86_64 7/20 2026-03-09T20:57:07.128 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : librbd1-2:19.2.3-678.ge911bdeb.el9.x86_64 8/20 2026-03-09T20:57:07.128 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : librdkafka-1.6.1-102.el9.x86_64 9/20 2026-03-09T20:57:07.129 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : librgw2-2:19.2.3-678.ge911bdeb.el9.x86_64 10/20 2026-03-09T20:57:07.129 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : lttng-ust-2.12.0-6.el9.x86_64 11/20 2026-03-09T20:57:07.129 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : parquet-libs-9.0.0-15.el9.x86_64 12/20 2026-03-09T20:57:07.129 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-rados-2:19.2.3-678.ge911bdeb.el9.x86_64 13/20 2026-03-09T20:57:07.129 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-rbd-2:19.2.3-678.ge911bdeb.el9.x86_64 14/20 2026-03-09T20:57:07.129 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-rgw-2:19.2.3-678.ge911bdeb.el9.x86_64 15/20 2026-03-09T20:57:07.129 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : qemu-kvm-block-rbd-17:10.1.0-15.el9.x86_64 16/20 2026-03-09T20:57:07.129 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : rbd-fuse-2:19.2.3-678.ge911bdeb.el9.x86_64 17/20 2026-03-09T20:57:07.129 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : rbd-nbd-2:19.2.3-678.ge911bdeb.el9.x86_64 18/20 2026-03-09T20:57:07.129 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : re2-1:20211101-20.el9.x86_64 19/20 2026-03-09T20:57:07.137 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : thrift-0.15.0-4.el9.x86_64 20/20 2026-03-09T20:57:07.137 INFO:teuthology.orchestra.run.vm08.stdout: 2026-03-09T20:57:07.137 INFO:teuthology.orchestra.run.vm08.stdout:Removed: 2026-03-09T20:57:07.137 INFO:teuthology.orchestra.run.vm08.stdout: boost-program-options-1.75.0-13.el9.x86_64 2026-03-09T20:57:07.137 INFO:teuthology.orchestra.run.vm08.stdout: libarrow-9.0.0-15.el9.x86_64 2026-03-09T20:57:07.137 INFO:teuthology.orchestra.run.vm08.stdout: libarrow-doc-9.0.0-15.el9.noarch 2026-03-09T20:57:07.137 INFO:teuthology.orchestra.run.vm08.stdout: libnbd-1.20.3-4.el9.x86_64 2026-03-09T20:57:07.137 INFO:teuthology.orchestra.run.vm08.stdout: libpmemobj-1.12.1-1.el9.x86_64 2026-03-09T20:57:07.137 INFO:teuthology.orchestra.run.vm08.stdout: librabbitmq-0.11.0-7.el9.x86_64 2026-03-09T20:57:07.137 INFO:teuthology.orchestra.run.vm08.stdout: librados2-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T20:57:07.137 INFO:teuthology.orchestra.run.vm08.stdout: librbd1-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T20:57:07.137 INFO:teuthology.orchestra.run.vm08.stdout: librdkafka-1.6.1-102.el9.x86_64 2026-03-09T20:57:07.137 INFO:teuthology.orchestra.run.vm08.stdout: librgw2-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T20:57:07.137 INFO:teuthology.orchestra.run.vm08.stdout: lttng-ust-2.12.0-6.el9.x86_64 2026-03-09T20:57:07.137 INFO:teuthology.orchestra.run.vm08.stdout: parquet-libs-9.0.0-15.el9.x86_64 2026-03-09T20:57:07.137 INFO:teuthology.orchestra.run.vm08.stdout: python3-rados-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T20:57:07.137 INFO:teuthology.orchestra.run.vm08.stdout: python3-rbd-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T20:57:07.137 INFO:teuthology.orchestra.run.vm08.stdout: python3-rgw-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T20:57:07.137 INFO:teuthology.orchestra.run.vm08.stdout: qemu-kvm-block-rbd-17:10.1.0-15.el9.x86_64 2026-03-09T20:57:07.137 INFO:teuthology.orchestra.run.vm08.stdout: rbd-fuse-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T20:57:07.138 INFO:teuthology.orchestra.run.vm08.stdout: rbd-nbd-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T20:57:07.138 INFO:teuthology.orchestra.run.vm08.stdout: re2-1:20211101-20.el9.x86_64 2026-03-09T20:57:07.138 INFO:teuthology.orchestra.run.vm08.stdout: thrift-0.15.0-4.el9.x86_64 2026-03-09T20:57:07.138 INFO:teuthology.orchestra.run.vm08.stdout: 2026-03-09T20:57:07.138 INFO:teuthology.orchestra.run.vm08.stdout:Complete! 2026-03-09T20:57:07.172 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : thrift-0.15.0-4.el9.x86_64 20/20 2026-03-09T20:57:07.172 INFO:teuthology.orchestra.run.vm01.stdout: 2026-03-09T20:57:07.172 INFO:teuthology.orchestra.run.vm01.stdout:Removed: 2026-03-09T20:57:07.172 INFO:teuthology.orchestra.run.vm01.stdout: boost-program-options-1.75.0-13.el9.x86_64 2026-03-09T20:57:07.172 INFO:teuthology.orchestra.run.vm01.stdout: libarrow-9.0.0-15.el9.x86_64 2026-03-09T20:57:07.172 INFO:teuthology.orchestra.run.vm01.stdout: libarrow-doc-9.0.0-15.el9.noarch 2026-03-09T20:57:07.172 INFO:teuthology.orchestra.run.vm01.stdout: libnbd-1.20.3-4.el9.x86_64 2026-03-09T20:57:07.172 INFO:teuthology.orchestra.run.vm01.stdout: libpmemobj-1.12.1-1.el9.x86_64 2026-03-09T20:57:07.172 INFO:teuthology.orchestra.run.vm01.stdout: librabbitmq-0.11.0-7.el9.x86_64 2026-03-09T20:57:07.172 INFO:teuthology.orchestra.run.vm01.stdout: librados2-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T20:57:07.172 INFO:teuthology.orchestra.run.vm01.stdout: librbd1-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T20:57:07.172 INFO:teuthology.orchestra.run.vm01.stdout: librdkafka-1.6.1-102.el9.x86_64 2026-03-09T20:57:07.172 INFO:teuthology.orchestra.run.vm01.stdout: librgw2-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T20:57:07.172 INFO:teuthology.orchestra.run.vm01.stdout: lttng-ust-2.12.0-6.el9.x86_64 2026-03-09T20:57:07.172 INFO:teuthology.orchestra.run.vm01.stdout: parquet-libs-9.0.0-15.el9.x86_64 2026-03-09T20:57:07.172 INFO:teuthology.orchestra.run.vm01.stdout: python3-rados-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T20:57:07.172 INFO:teuthology.orchestra.run.vm01.stdout: python3-rbd-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T20:57:07.173 INFO:teuthology.orchestra.run.vm01.stdout: python3-rgw-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T20:57:07.173 INFO:teuthology.orchestra.run.vm01.stdout: qemu-kvm-block-rbd-17:10.1.0-15.el9.x86_64 2026-03-09T20:57:07.173 INFO:teuthology.orchestra.run.vm01.stdout: rbd-fuse-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T20:57:07.173 INFO:teuthology.orchestra.run.vm01.stdout: rbd-nbd-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T20:57:07.173 INFO:teuthology.orchestra.run.vm01.stdout: re2-1:20211101-20.el9.x86_64 2026-03-09T20:57:07.173 INFO:teuthology.orchestra.run.vm01.stdout: thrift-0.15.0-4.el9.x86_64 2026-03-09T20:57:07.173 INFO:teuthology.orchestra.run.vm01.stdout: 2026-03-09T20:57:07.173 INFO:teuthology.orchestra.run.vm01.stdout:Complete! 2026-03-09T20:57:07.177 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : thrift-0.15.0-4.el9.x86_64 20/20 2026-03-09T20:57:07.177 INFO:teuthology.orchestra.run.vm06.stdout: 2026-03-09T20:57:07.177 INFO:teuthology.orchestra.run.vm06.stdout:Removed: 2026-03-09T20:57:07.177 INFO:teuthology.orchestra.run.vm06.stdout: boost-program-options-1.75.0-13.el9.x86_64 2026-03-09T20:57:07.177 INFO:teuthology.orchestra.run.vm06.stdout: libarrow-9.0.0-15.el9.x86_64 2026-03-09T20:57:07.177 INFO:teuthology.orchestra.run.vm06.stdout: libarrow-doc-9.0.0-15.el9.noarch 2026-03-09T20:57:07.177 INFO:teuthology.orchestra.run.vm06.stdout: libnbd-1.20.3-4.el9.x86_64 2026-03-09T20:57:07.177 INFO:teuthology.orchestra.run.vm06.stdout: libpmemobj-1.12.1-1.el9.x86_64 2026-03-09T20:57:07.177 INFO:teuthology.orchestra.run.vm06.stdout: librabbitmq-0.11.0-7.el9.x86_64 2026-03-09T20:57:07.177 INFO:teuthology.orchestra.run.vm06.stdout: librados2-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T20:57:07.177 INFO:teuthology.orchestra.run.vm06.stdout: librbd1-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T20:57:07.177 INFO:teuthology.orchestra.run.vm06.stdout: librdkafka-1.6.1-102.el9.x86_64 2026-03-09T20:57:07.177 INFO:teuthology.orchestra.run.vm06.stdout: librgw2-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T20:57:07.177 INFO:teuthology.orchestra.run.vm06.stdout: lttng-ust-2.12.0-6.el9.x86_64 2026-03-09T20:57:07.177 INFO:teuthology.orchestra.run.vm06.stdout: parquet-libs-9.0.0-15.el9.x86_64 2026-03-09T20:57:07.178 INFO:teuthology.orchestra.run.vm06.stdout: python3-rados-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T20:57:07.178 INFO:teuthology.orchestra.run.vm06.stdout: python3-rbd-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T20:57:07.178 INFO:teuthology.orchestra.run.vm06.stdout: python3-rgw-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T20:57:07.178 INFO:teuthology.orchestra.run.vm06.stdout: qemu-kvm-block-rbd-17:10.1.0-15.el9.x86_64 2026-03-09T20:57:07.178 INFO:teuthology.orchestra.run.vm06.stdout: rbd-fuse-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T20:57:07.178 INFO:teuthology.orchestra.run.vm06.stdout: rbd-nbd-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T20:57:07.178 INFO:teuthology.orchestra.run.vm06.stdout: re2-1:20211101-20.el9.x86_64 2026-03-09T20:57:07.178 INFO:teuthology.orchestra.run.vm06.stdout: thrift-0.15.0-4.el9.x86_64 2026-03-09T20:57:07.178 INFO:teuthology.orchestra.run.vm06.stdout: 2026-03-09T20:57:07.178 INFO:teuthology.orchestra.run.vm06.stdout:Complete! 2026-03-09T20:57:07.351 INFO:teuthology.orchestra.run.vm08.stdout:No match for argument: librbd1 2026-03-09T20:57:07.351 INFO:teuthology.orchestra.run.vm08.stderr:No packages marked for removal. 2026-03-09T20:57:07.354 INFO:teuthology.orchestra.run.vm08.stdout:Dependencies resolved. 2026-03-09T20:57:07.355 INFO:teuthology.orchestra.run.vm08.stdout:Nothing to do. 2026-03-09T20:57:07.355 INFO:teuthology.orchestra.run.vm08.stdout:Complete! 2026-03-09T20:57:07.385 INFO:teuthology.orchestra.run.vm01.stdout:No match for argument: librbd1 2026-03-09T20:57:07.385 INFO:teuthology.orchestra.run.vm01.stderr:No packages marked for removal. 2026-03-09T20:57:07.388 INFO:teuthology.orchestra.run.vm01.stdout:Dependencies resolved. 2026-03-09T20:57:07.388 INFO:teuthology.orchestra.run.vm01.stdout:Nothing to do. 2026-03-09T20:57:07.389 INFO:teuthology.orchestra.run.vm01.stdout:Complete! 2026-03-09T20:57:07.391 INFO:teuthology.orchestra.run.vm06.stdout:No match for argument: librbd1 2026-03-09T20:57:07.391 INFO:teuthology.orchestra.run.vm06.stderr:No packages marked for removal. 2026-03-09T20:57:07.394 INFO:teuthology.orchestra.run.vm06.stdout:Dependencies resolved. 2026-03-09T20:57:07.395 INFO:teuthology.orchestra.run.vm06.stdout:Nothing to do. 2026-03-09T20:57:07.395 INFO:teuthology.orchestra.run.vm06.stdout:Complete! 2026-03-09T20:57:07.542 INFO:teuthology.orchestra.run.vm08.stdout:No match for argument: python3-rados 2026-03-09T20:57:07.542 INFO:teuthology.orchestra.run.vm08.stderr:No packages marked for removal. 2026-03-09T20:57:07.545 INFO:teuthology.orchestra.run.vm08.stdout:Dependencies resolved. 2026-03-09T20:57:07.545 INFO:teuthology.orchestra.run.vm08.stdout:Nothing to do. 2026-03-09T20:57:07.545 INFO:teuthology.orchestra.run.vm08.stdout:Complete! 2026-03-09T20:57:07.574 INFO:teuthology.orchestra.run.vm01.stdout:No match for argument: python3-rados 2026-03-09T20:57:07.574 INFO:teuthology.orchestra.run.vm01.stderr:No packages marked for removal. 2026-03-09T20:57:07.577 INFO:teuthology.orchestra.run.vm01.stdout:Dependencies resolved. 2026-03-09T20:57:07.577 INFO:teuthology.orchestra.run.vm01.stdout:Nothing to do. 2026-03-09T20:57:07.578 INFO:teuthology.orchestra.run.vm01.stdout:Complete! 2026-03-09T20:57:07.588 INFO:teuthology.orchestra.run.vm06.stdout:No match for argument: python3-rados 2026-03-09T20:57:07.588 INFO:teuthology.orchestra.run.vm06.stderr:No packages marked for removal. 2026-03-09T20:57:07.590 INFO:teuthology.orchestra.run.vm06.stdout:Dependencies resolved. 2026-03-09T20:57:07.591 INFO:teuthology.orchestra.run.vm06.stdout:Nothing to do. 2026-03-09T20:57:07.591 INFO:teuthology.orchestra.run.vm06.stdout:Complete! 2026-03-09T20:57:07.713 INFO:teuthology.orchestra.run.vm08.stdout:No match for argument: python3-rgw 2026-03-09T20:57:07.713 INFO:teuthology.orchestra.run.vm08.stderr:No packages marked for removal. 2026-03-09T20:57:07.715 INFO:teuthology.orchestra.run.vm08.stdout:Dependencies resolved. 2026-03-09T20:57:07.715 INFO:teuthology.orchestra.run.vm08.stdout:Nothing to do. 2026-03-09T20:57:07.715 INFO:teuthology.orchestra.run.vm08.stdout:Complete! 2026-03-09T20:57:07.747 INFO:teuthology.orchestra.run.vm01.stdout:No match for argument: python3-rgw 2026-03-09T20:57:07.747 INFO:teuthology.orchestra.run.vm01.stderr:No packages marked for removal. 2026-03-09T20:57:07.749 INFO:teuthology.orchestra.run.vm01.stdout:Dependencies resolved. 2026-03-09T20:57:07.750 INFO:teuthology.orchestra.run.vm01.stdout:Nothing to do. 2026-03-09T20:57:07.750 INFO:teuthology.orchestra.run.vm01.stdout:Complete! 2026-03-09T20:57:07.753 INFO:teuthology.orchestra.run.vm06.stdout:No match for argument: python3-rgw 2026-03-09T20:57:07.753 INFO:teuthology.orchestra.run.vm06.stderr:No packages marked for removal. 2026-03-09T20:57:07.755 INFO:teuthology.orchestra.run.vm06.stdout:Dependencies resolved. 2026-03-09T20:57:07.756 INFO:teuthology.orchestra.run.vm06.stdout:Nothing to do. 2026-03-09T20:57:07.756 INFO:teuthology.orchestra.run.vm06.stdout:Complete! 2026-03-09T20:57:07.880 INFO:teuthology.orchestra.run.vm08.stdout:No match for argument: python3-cephfs 2026-03-09T20:57:07.880 INFO:teuthology.orchestra.run.vm08.stderr:No packages marked for removal. 2026-03-09T20:57:07.883 INFO:teuthology.orchestra.run.vm08.stdout:Dependencies resolved. 2026-03-09T20:57:07.883 INFO:teuthology.orchestra.run.vm08.stdout:Nothing to do. 2026-03-09T20:57:07.883 INFO:teuthology.orchestra.run.vm08.stdout:Complete! 2026-03-09T20:57:07.929 INFO:teuthology.orchestra.run.vm01.stdout:No match for argument: python3-cephfs 2026-03-09T20:57:07.929 INFO:teuthology.orchestra.run.vm01.stderr:No packages marked for removal. 2026-03-09T20:57:07.930 INFO:teuthology.orchestra.run.vm06.stdout:No match for argument: python3-cephfs 2026-03-09T20:57:07.930 INFO:teuthology.orchestra.run.vm06.stderr:No packages marked for removal. 2026-03-09T20:57:07.931 INFO:teuthology.orchestra.run.vm01.stdout:Dependencies resolved. 2026-03-09T20:57:07.932 INFO:teuthology.orchestra.run.vm01.stdout:Nothing to do. 2026-03-09T20:57:07.932 INFO:teuthology.orchestra.run.vm01.stdout:Complete! 2026-03-09T20:57:07.933 INFO:teuthology.orchestra.run.vm06.stdout:Dependencies resolved. 2026-03-09T20:57:07.933 INFO:teuthology.orchestra.run.vm06.stdout:Nothing to do. 2026-03-09T20:57:07.934 INFO:teuthology.orchestra.run.vm06.stdout:Complete! 2026-03-09T20:57:08.087 INFO:teuthology.orchestra.run.vm08.stdout:No match for argument: python3-rbd 2026-03-09T20:57:08.088 INFO:teuthology.orchestra.run.vm08.stderr:No packages marked for removal. 2026-03-09T20:57:08.092 INFO:teuthology.orchestra.run.vm08.stdout:Dependencies resolved. 2026-03-09T20:57:08.092 INFO:teuthology.orchestra.run.vm08.stdout:Nothing to do. 2026-03-09T20:57:08.092 INFO:teuthology.orchestra.run.vm08.stdout:Complete! 2026-03-09T20:57:08.119 INFO:teuthology.orchestra.run.vm06.stdout:No match for argument: python3-rbd 2026-03-09T20:57:08.119 INFO:teuthology.orchestra.run.vm06.stderr:No packages marked for removal. 2026-03-09T20:57:08.121 INFO:teuthology.orchestra.run.vm06.stdout:Dependencies resolved. 2026-03-09T20:57:08.122 INFO:teuthology.orchestra.run.vm06.stdout:Nothing to do. 2026-03-09T20:57:08.122 INFO:teuthology.orchestra.run.vm06.stdout:Complete! 2026-03-09T20:57:08.122 INFO:teuthology.orchestra.run.vm01.stdout:No match for argument: python3-rbd 2026-03-09T20:57:08.122 INFO:teuthology.orchestra.run.vm01.stderr:No packages marked for removal. 2026-03-09T20:57:08.124 INFO:teuthology.orchestra.run.vm01.stdout:Dependencies resolved. 2026-03-09T20:57:08.125 INFO:teuthology.orchestra.run.vm01.stdout:Nothing to do. 2026-03-09T20:57:08.125 INFO:teuthology.orchestra.run.vm01.stdout:Complete! 2026-03-09T20:57:08.319 INFO:teuthology.orchestra.run.vm08.stdout:No match for argument: rbd-fuse 2026-03-09T20:57:08.319 INFO:teuthology.orchestra.run.vm08.stderr:No packages marked for removal. 2026-03-09T20:57:08.321 INFO:teuthology.orchestra.run.vm08.stdout:Dependencies resolved. 2026-03-09T20:57:08.322 INFO:teuthology.orchestra.run.vm08.stdout:Nothing to do. 2026-03-09T20:57:08.322 INFO:teuthology.orchestra.run.vm08.stdout:Complete! 2026-03-09T20:57:08.325 INFO:teuthology.orchestra.run.vm06.stdout:No match for argument: rbd-fuse 2026-03-09T20:57:08.325 INFO:teuthology.orchestra.run.vm06.stderr:No packages marked for removal. 2026-03-09T20:57:08.327 INFO:teuthology.orchestra.run.vm06.stdout:Dependencies resolved. 2026-03-09T20:57:08.328 INFO:teuthology.orchestra.run.vm06.stdout:Nothing to do. 2026-03-09T20:57:08.328 INFO:teuthology.orchestra.run.vm06.stdout:Complete! 2026-03-09T20:57:08.333 INFO:teuthology.orchestra.run.vm01.stdout:No match for argument: rbd-fuse 2026-03-09T20:57:08.333 INFO:teuthology.orchestra.run.vm01.stderr:No packages marked for removal. 2026-03-09T20:57:08.335 INFO:teuthology.orchestra.run.vm01.stdout:Dependencies resolved. 2026-03-09T20:57:08.336 INFO:teuthology.orchestra.run.vm01.stdout:Nothing to do. 2026-03-09T20:57:08.336 INFO:teuthology.orchestra.run.vm01.stdout:Complete! 2026-03-09T20:57:08.493 INFO:teuthology.orchestra.run.vm08.stdout:No match for argument: rbd-mirror 2026-03-09T20:57:08.493 INFO:teuthology.orchestra.run.vm08.stderr:No packages marked for removal. 2026-03-09T20:57:08.496 INFO:teuthology.orchestra.run.vm08.stdout:Dependencies resolved. 2026-03-09T20:57:08.497 INFO:teuthology.orchestra.run.vm08.stdout:Nothing to do. 2026-03-09T20:57:08.497 INFO:teuthology.orchestra.run.vm08.stdout:Complete! 2026-03-09T20:57:08.503 INFO:teuthology.orchestra.run.vm06.stdout:No match for argument: rbd-mirror 2026-03-09T20:57:08.503 INFO:teuthology.orchestra.run.vm06.stderr:No packages marked for removal. 2026-03-09T20:57:08.505 INFO:teuthology.orchestra.run.vm06.stdout:Dependencies resolved. 2026-03-09T20:57:08.506 INFO:teuthology.orchestra.run.vm06.stdout:Nothing to do. 2026-03-09T20:57:08.506 INFO:teuthology.orchestra.run.vm06.stdout:Complete! 2026-03-09T20:57:08.508 INFO:teuthology.orchestra.run.vm01.stdout:No match for argument: rbd-mirror 2026-03-09T20:57:08.509 INFO:teuthology.orchestra.run.vm01.stderr:No packages marked for removal. 2026-03-09T20:57:08.511 INFO:teuthology.orchestra.run.vm01.stdout:Dependencies resolved. 2026-03-09T20:57:08.512 INFO:teuthology.orchestra.run.vm01.stdout:Nothing to do. 2026-03-09T20:57:08.512 INFO:teuthology.orchestra.run.vm01.stdout:Complete! 2026-03-09T20:57:08.666 INFO:teuthology.orchestra.run.vm08.stdout:No match for argument: rbd-nbd 2026-03-09T20:57:08.666 INFO:teuthology.orchestra.run.vm08.stderr:No packages marked for removal. 2026-03-09T20:57:08.668 INFO:teuthology.orchestra.run.vm08.stdout:Dependencies resolved. 2026-03-09T20:57:08.668 INFO:teuthology.orchestra.run.vm08.stdout:Nothing to do. 2026-03-09T20:57:08.668 INFO:teuthology.orchestra.run.vm08.stdout:Complete! 2026-03-09T20:57:08.676 INFO:teuthology.orchestra.run.vm06.stdout:No match for argument: rbd-nbd 2026-03-09T20:57:08.676 INFO:teuthology.orchestra.run.vm06.stderr:No packages marked for removal. 2026-03-09T20:57:08.678 INFO:teuthology.orchestra.run.vm01.stdout:No match for argument: rbd-nbd 2026-03-09T20:57:08.679 INFO:teuthology.orchestra.run.vm01.stderr:No packages marked for removal. 2026-03-09T20:57:08.679 INFO:teuthology.orchestra.run.vm06.stdout:Dependencies resolved. 2026-03-09T20:57:08.679 INFO:teuthology.orchestra.run.vm06.stdout:Nothing to do. 2026-03-09T20:57:08.679 INFO:teuthology.orchestra.run.vm06.stdout:Complete! 2026-03-09T20:57:08.681 INFO:teuthology.orchestra.run.vm01.stdout:Dependencies resolved. 2026-03-09T20:57:08.681 INFO:teuthology.orchestra.run.vm01.stdout:Nothing to do. 2026-03-09T20:57:08.681 INFO:teuthology.orchestra.run.vm01.stdout:Complete! 2026-03-09T20:57:08.692 DEBUG:teuthology.orchestra.run.vm08:> sudo yum clean all 2026-03-09T20:57:08.699 DEBUG:teuthology.orchestra.run.vm06:> sudo yum clean all 2026-03-09T20:57:08.703 DEBUG:teuthology.orchestra.run.vm01:> sudo yum clean all 2026-03-09T20:57:08.817 INFO:teuthology.orchestra.run.vm08.stdout:56 files removed 2026-03-09T20:57:08.827 INFO:teuthology.orchestra.run.vm06.stdout:56 files removed 2026-03-09T20:57:08.830 INFO:teuthology.orchestra.run.vm01.stdout:56 files removed 2026-03-09T20:57:08.835 DEBUG:teuthology.orchestra.run.vm08:> sudo rm -f /etc/yum.repos.d/ceph.repo 2026-03-09T20:57:08.855 DEBUG:teuthology.orchestra.run.vm01:> sudo rm -f /etc/yum.repos.d/ceph.repo 2026-03-09T20:57:08.855 DEBUG:teuthology.orchestra.run.vm06:> sudo rm -f /etc/yum.repos.d/ceph.repo 2026-03-09T20:57:08.862 DEBUG:teuthology.orchestra.run.vm08:> sudo yum clean expire-cache 2026-03-09T20:57:08.879 DEBUG:teuthology.orchestra.run.vm01:> sudo yum clean expire-cache 2026-03-09T20:57:08.882 DEBUG:teuthology.orchestra.run.vm06:> sudo yum clean expire-cache 2026-03-09T20:57:09.016 INFO:teuthology.orchestra.run.vm08.stdout:Cache was expired 2026-03-09T20:57:09.016 INFO:teuthology.orchestra.run.vm08.stdout:0 files removed 2026-03-09T20:57:09.032 INFO:teuthology.orchestra.run.vm01.stdout:Cache was expired 2026-03-09T20:57:09.032 INFO:teuthology.orchestra.run.vm01.stdout:0 files removed 2026-03-09T20:57:09.035 DEBUG:teuthology.parallel:result is None 2026-03-09T20:57:09.045 INFO:teuthology.orchestra.run.vm06.stdout:Cache was expired 2026-03-09T20:57:09.045 INFO:teuthology.orchestra.run.vm06.stdout:0 files removed 2026-03-09T20:57:09.051 DEBUG:teuthology.parallel:result is None 2026-03-09T20:57:09.065 DEBUG:teuthology.parallel:result is None 2026-03-09T20:57:09.065 INFO:teuthology.task.install:Removing ceph sources lists on ubuntu@vm01.local 2026-03-09T20:57:09.065 INFO:teuthology.task.install:Removing ceph sources lists on ubuntu@vm06.local 2026-03-09T20:57:09.065 INFO:teuthology.task.install:Removing ceph sources lists on ubuntu@vm08.local 2026-03-09T20:57:09.065 DEBUG:teuthology.orchestra.run.vm01:> sudo rm -f /etc/yum.repos.d/ceph.repo 2026-03-09T20:57:09.065 DEBUG:teuthology.orchestra.run.vm06:> sudo rm -f /etc/yum.repos.d/ceph.repo 2026-03-09T20:57:09.065 DEBUG:teuthology.orchestra.run.vm08:> sudo rm -f /etc/yum.repos.d/ceph.repo 2026-03-09T20:57:09.091 DEBUG:teuthology.orchestra.run.vm06:> sudo mv -f /etc/yum/pluginconf.d/priorities.conf.orig /etc/yum/pluginconf.d/priorities.conf 2026-03-09T20:57:09.092 DEBUG:teuthology.orchestra.run.vm01:> sudo mv -f /etc/yum/pluginconf.d/priorities.conf.orig /etc/yum/pluginconf.d/priorities.conf 2026-03-09T20:57:09.094 DEBUG:teuthology.orchestra.run.vm08:> sudo mv -f /etc/yum/pluginconf.d/priorities.conf.orig /etc/yum/pluginconf.d/priorities.conf 2026-03-09T20:57:09.161 DEBUG:teuthology.parallel:result is None 2026-03-09T20:57:09.161 DEBUG:teuthology.parallel:result is None 2026-03-09T20:57:09.162 DEBUG:teuthology.parallel:result is None 2026-03-09T20:57:09.162 DEBUG:teuthology.run_tasks:Unwinding manager clock 2026-03-09T20:57:09.165 INFO:teuthology.task.clock:Checking final clock skew... 2026-03-09T20:57:09.165 DEBUG:teuthology.orchestra.run.vm01:> PATH=/usr/bin:/usr/sbin ntpq -p || PATH=/usr/bin:/usr/sbin chronyc sources || true 2026-03-09T20:57:09.204 DEBUG:teuthology.orchestra.run.vm06:> PATH=/usr/bin:/usr/sbin ntpq -p || PATH=/usr/bin:/usr/sbin chronyc sources || true 2026-03-09T20:57:09.205 DEBUG:teuthology.orchestra.run.vm08:> PATH=/usr/bin:/usr/sbin ntpq -p || PATH=/usr/bin:/usr/sbin chronyc sources || true 2026-03-09T20:57:09.216 INFO:teuthology.orchestra.run.vm01.stderr:bash: line 1: ntpq: command not found 2026-03-09T20:57:09.219 INFO:teuthology.orchestra.run.vm01.stdout:MS Name/IP address Stratum Poll Reach LastRx Last sample 2026-03-09T20:57:09.219 INFO:teuthology.orchestra.run.vm01.stdout:=============================================================================== 2026-03-09T20:57:09.219 INFO:teuthology.orchestra.run.vm01.stdout:^* ntp2.uni-ulm.de 2 6 377 22 +838us[ +836us] +/- 15ms 2026-03-09T20:57:09.219 INFO:teuthology.orchestra.run.vm01.stdout:^+ static.119.109.140.128.c> 2 6 377 21 -536us[ -536us] +/- 47ms 2026-03-09T20:57:09.219 INFO:teuthology.orchestra.run.vm01.stdout:^+ x1.ncomputers.org 2 6 377 21 +1048us[+1048us] +/- 42ms 2026-03-09T20:57:09.219 INFO:teuthology.orchestra.run.vm01.stdout:^+ fa.gnudb.org 2 6 377 21 -608us[ -608us] +/- 41ms 2026-03-09T20:57:09.220 INFO:teuthology.orchestra.run.vm06.stderr:bash: line 1: ntpq: command not found 2026-03-09T20:57:09.221 INFO:teuthology.orchestra.run.vm08.stderr:bash: line 1: ntpq: command not found 2026-03-09T20:57:09.223 INFO:teuthology.orchestra.run.vm06.stdout:MS Name/IP address Stratum Poll Reach LastRx Last sample 2026-03-09T20:57:09.223 INFO:teuthology.orchestra.run.vm06.stdout:=============================================================================== 2026-03-09T20:57:09.223 INFO:teuthology.orchestra.run.vm06.stdout:^+ fa.gnudb.org 2 6 377 20 -739us[ -739us] +/- 41ms 2026-03-09T20:57:09.223 INFO:teuthology.orchestra.run.vm06.stdout:^* ntp2.uni-ulm.de 2 6 377 21 +791us[ +784us] +/- 15ms 2026-03-09T20:57:09.223 INFO:teuthology.orchestra.run.vm06.stdout:^+ static.119.109.140.128.c> 2 6 377 23 -641us[ -648us] +/- 47ms 2026-03-09T20:57:09.223 INFO:teuthology.orchestra.run.vm06.stdout:^+ x1.ncomputers.org 2 6 377 21 -865us[ -865us] +/- 40ms 2026-03-09T20:57:09.224 INFO:teuthology.orchestra.run.vm08.stdout:MS Name/IP address Stratum Poll Reach LastRx Last sample 2026-03-09T20:57:09.224 INFO:teuthology.orchestra.run.vm08.stdout:=============================================================================== 2026-03-09T20:57:09.224 INFO:teuthology.orchestra.run.vm08.stdout:^+ static.119.109.140.128.c> 2 6 377 22 -656us[ -656us] +/- 47ms 2026-03-09T20:57:09.224 INFO:teuthology.orchestra.run.vm08.stdout:^+ x1.ncomputers.org 2 6 377 21 +819us[ +819us] +/- 42ms 2026-03-09T20:57:09.224 INFO:teuthology.orchestra.run.vm08.stdout:^+ fa.gnudb.org 2 6 377 23 -857us[ -857us] +/- 42ms 2026-03-09T20:57:09.224 INFO:teuthology.orchestra.run.vm08.stdout:^* ntp2.uni-ulm.de 2 6 377 23 +464us[ +470us] +/- 15ms 2026-03-09T20:57:09.225 DEBUG:teuthology.run_tasks:Unwinding manager ansible.cephlab 2026-03-09T20:57:09.227 INFO:teuthology.task.ansible:Skipping ansible cleanup... 2026-03-09T20:57:09.227 DEBUG:teuthology.run_tasks:Unwinding manager selinux 2026-03-09T20:57:09.241 DEBUG:teuthology.run_tasks:Unwinding manager pcp 2026-03-09T20:57:09.251 DEBUG:teuthology.run_tasks:Unwinding manager internal.timer 2026-03-09T20:57:09.253 INFO:teuthology.task.internal:Duration was 618.017398 seconds 2026-03-09T20:57:09.253 DEBUG:teuthology.run_tasks:Unwinding manager internal.syslog 2026-03-09T20:57:09.255 INFO:teuthology.task.internal.syslog:Shutting down syslog monitoring... 2026-03-09T20:57:09.255 DEBUG:teuthology.orchestra.run.vm01:> sudo rm -f -- /etc/rsyslog.d/80-cephtest.conf && sudo service rsyslog restart 2026-03-09T20:57:09.261 DEBUG:teuthology.orchestra.run.vm06:> sudo rm -f -- /etc/rsyslog.d/80-cephtest.conf && sudo service rsyslog restart 2026-03-09T20:57:09.266 DEBUG:teuthology.orchestra.run.vm08:> sudo rm -f -- /etc/rsyslog.d/80-cephtest.conf && sudo service rsyslog restart 2026-03-09T20:57:09.299 INFO:teuthology.orchestra.run.vm01.stderr:Redirecting to /bin/systemctl restart rsyslog.service 2026-03-09T20:57:09.308 INFO:teuthology.orchestra.run.vm08.stderr:Redirecting to /bin/systemctl restart rsyslog.service 2026-03-09T20:57:09.310 INFO:teuthology.orchestra.run.vm06.stderr:Redirecting to /bin/systemctl restart rsyslog.service 2026-03-09T20:57:09.574 INFO:teuthology.task.internal.syslog:Checking logs for errors... 2026-03-09T20:57:09.574 DEBUG:teuthology.task.internal.syslog:Checking ubuntu@vm01.local 2026-03-09T20:57:09.574 DEBUG:teuthology.orchestra.run.vm01:> grep -E --binary-files=text '\bBUG\b|\bINFO\b|\bDEADLOCK\b' /home/ubuntu/cephtest/archive/syslog/kern.log | grep -v 'task .* blocked for more than .* seconds' | grep -v 'lockdep is turned off' | grep -v 'trying to register non-static key' | grep -v 'DEBUG: fsize' | grep -v CRON | grep -v 'BUG: bad unlock balance detected' | grep -v 'inconsistent lock state' | grep -v '*** DEADLOCK ***' | grep -v 'INFO: possible irq lock inversion dependency detected' | grep -v 'INFO: NMI handler (perf_event_nmi_handler) took too long to run' | grep -v 'INFO: recovery required on readonly' | grep -v 'ceph-create-keys: INFO' | grep -v INFO:ceph-create-keys | grep -v 'Loaded datasource DataSourceOpenStack' | grep -v 'container-storage-setup: INFO: Volume group backing root filesystem could not be determined' | grep -E -v '\bsalt-master\b|\bsalt-minion\b|\bsalt-api\b' | grep -v ceph-crash | grep -E -v '\btcmu-runner\b.*\bINFO\b' | head -n 1 2026-03-09T20:57:09.597 DEBUG:teuthology.task.internal.syslog:Checking ubuntu@vm06.local 2026-03-09T20:57:09.598 DEBUG:teuthology.orchestra.run.vm06:> grep -E --binary-files=text '\bBUG\b|\bINFO\b|\bDEADLOCK\b' /home/ubuntu/cephtest/archive/syslog/kern.log | grep -v 'task .* blocked for more than .* seconds' | grep -v 'lockdep is turned off' | grep -v 'trying to register non-static key' | grep -v 'DEBUG: fsize' | grep -v CRON | grep -v 'BUG: bad unlock balance detected' | grep -v 'inconsistent lock state' | grep -v '*** DEADLOCK ***' | grep -v 'INFO: possible irq lock inversion dependency detected' | grep -v 'INFO: NMI handler (perf_event_nmi_handler) took too long to run' | grep -v 'INFO: recovery required on readonly' | grep -v 'ceph-create-keys: INFO' | grep -v INFO:ceph-create-keys | grep -v 'Loaded datasource DataSourceOpenStack' | grep -v 'container-storage-setup: INFO: Volume group backing root filesystem could not be determined' | grep -E -v '\bsalt-master\b|\bsalt-minion\b|\bsalt-api\b' | grep -v ceph-crash | grep -E -v '\btcmu-runner\b.*\bINFO\b' | head -n 1 2026-03-09T20:57:09.620 DEBUG:teuthology.task.internal.syslog:Checking ubuntu@vm08.local 2026-03-09T20:57:09.620 DEBUG:teuthology.orchestra.run.vm08:> grep -E --binary-files=text '\bBUG\b|\bINFO\b|\bDEADLOCK\b' /home/ubuntu/cephtest/archive/syslog/kern.log | grep -v 'task .* blocked for more than .* seconds' | grep -v 'lockdep is turned off' | grep -v 'trying to register non-static key' | grep -v 'DEBUG: fsize' | grep -v CRON | grep -v 'BUG: bad unlock balance detected' | grep -v 'inconsistent lock state' | grep -v '*** DEADLOCK ***' | grep -v 'INFO: possible irq lock inversion dependency detected' | grep -v 'INFO: NMI handler (perf_event_nmi_handler) took too long to run' | grep -v 'INFO: recovery required on readonly' | grep -v 'ceph-create-keys: INFO' | grep -v INFO:ceph-create-keys | grep -v 'Loaded datasource DataSourceOpenStack' | grep -v 'container-storage-setup: INFO: Volume group backing root filesystem could not be determined' | grep -E -v '\bsalt-master\b|\bsalt-minion\b|\bsalt-api\b' | grep -v ceph-crash | grep -E -v '\btcmu-runner\b.*\bINFO\b' | head -n 1 2026-03-09T20:57:09.644 INFO:teuthology.task.internal.syslog:Gathering journactl... 2026-03-09T20:57:09.644 DEBUG:teuthology.orchestra.run.vm01:> sudo journalctl > /home/ubuntu/cephtest/archive/syslog/journalctl.log 2026-03-09T20:57:09.646 DEBUG:teuthology.orchestra.run.vm06:> sudo journalctl > /home/ubuntu/cephtest/archive/syslog/journalctl.log 2026-03-09T20:57:09.663 DEBUG:teuthology.orchestra.run.vm08:> sudo journalctl > /home/ubuntu/cephtest/archive/syslog/journalctl.log 2026-03-09T20:57:10.079 INFO:teuthology.task.internal.syslog:Compressing syslogs... 2026-03-09T20:57:10.079 DEBUG:teuthology.orchestra.run.vm01:> find /home/ubuntu/cephtest/archive/syslog -name '*.log' -print0 | sudo xargs -0 --max-args=1 --max-procs=0 --verbose --no-run-if-empty -- gzip -5 --verbose -- 2026-03-09T20:57:10.080 DEBUG:teuthology.orchestra.run.vm06:> find /home/ubuntu/cephtest/archive/syslog -name '*.log' -print0 | sudo xargs -0 --max-args=1 --max-procs=0 --verbose --no-run-if-empty -- gzip -5 --verbose -- 2026-03-09T20:57:10.082 DEBUG:teuthology.orchestra.run.vm08:> find /home/ubuntu/cephtest/archive/syslog -name '*.log' -print0 | sudo xargs -0 --max-args=1 --max-procs=0 --verbose --no-run-if-empty -- gzip -5 --verbose -- 2026-03-09T20:57:10.101 INFO:teuthology.orchestra.run.vm01.stderr:gzip -5 --verbose -- /home/ubuntu/cephtest/archive/syslog/kern.log 2026-03-09T20:57:10.102 INFO:teuthology.orchestra.run.vm01.stderr:gzip -5 --verbose -- /home/ubuntu/cephtest/archive/syslog/misc.log 2026-03-09T20:57:10.102 INFO:teuthology.orchestra.run.vm01.stderr:gzip -5 --verbose -- /home/ubuntu/cephtest/archive/syslog/journalctl.log 2026-03-09T20:57:10.102 INFO:teuthology.orchestra.run.vm01.stderr:/home/ubuntu/cephtest/archive/syslog/kern.log: 0.0% -- replaced with /home/ubuntu/cephtest/archive/syslog/kern.log.gz 2026-03-09T20:57:10.102 INFO:teuthology.orchestra.run.vm01.stderr:/home/ubuntu/cephtest/archive/syslog/misc.log: 0.0% -- replaced with /home/ubuntu/cephtest/archive/syslog/misc.log.gz 2026-03-09T20:57:10.107 INFO:teuthology.orchestra.run.vm06.stderr:gzip -5 --verbose -- /home/ubuntu/cephtest/archive/syslog/kern.log 2026-03-09T20:57:10.107 INFO:teuthology.orchestra.run.vm06.stderr:gzip -5 --verbose -- /home/ubuntu/cephtest/archive/syslog/misc.log 2026-03-09T20:57:10.108 INFO:teuthology.orchestra.run.vm06.stderr:/home/ubuntu/cephtest/archive/syslog/kern.log: gzip -5 --verbose -- 0.0% -- replaced with /home/ubuntu/cephtest/archive/syslog/kern.log.gz 2026-03-09T20:57:10.108 INFO:teuthology.orchestra.run.vm06.stderr: /home/ubuntu/cephtest/archive/syslog/journalctl.log 2026-03-09T20:57:10.108 INFO:teuthology.orchestra.run.vm06.stderr:/home/ubuntu/cephtest/archive/syslog/misc.log: 0.0% -- replaced with /home/ubuntu/cephtest/archive/syslog/misc.log.gz 2026-03-09T20:57:10.108 INFO:teuthology.orchestra.run.vm08.stderr:gzip -5 --verbose -- /home/ubuntu/cephtest/archive/syslog/kern.log 2026-03-09T20:57:10.109 INFO:teuthology.orchestra.run.vm08.stderr:gzip -5 --verbose -- /home/ubuntu/cephtest/archive/syslog/misc.log 2026-03-09T20:57:10.109 INFO:teuthology.orchestra.run.vm08.stderr:/home/ubuntu/cephtest/archive/syslog/kern.log: 0.0% -- replaced with /home/ubuntu/cephtest/archive/syslog/kern.log.gz 2026-03-09T20:57:10.109 INFO:teuthology.orchestra.run.vm08.stderr:gzip -5 --verbose -- /home/ubuntu/cephtest/archive/syslog/journalctl.log 2026-03-09T20:57:10.110 INFO:teuthology.orchestra.run.vm08.stderr:/home/ubuntu/cephtest/archive/syslog/misc.log: 0.0% -- replaced with /home/ubuntu/cephtest/archive/syslog/misc.log.gz/home/ubuntu/cephtest/archive/syslog/journalctl.log: 2026-03-09T20:57:10.213 INFO:teuthology.orchestra.run.vm06.stderr:/home/ubuntu/cephtest/archive/syslog/journalctl.log: 98.3% -- replaced with /home/ubuntu/cephtest/archive/syslog/journalctl.log.gz 2026-03-09T20:57:10.230 INFO:teuthology.orchestra.run.vm08.stderr: 98.3% -- replaced with /home/ubuntu/cephtest/archive/syslog/journalctl.log.gz 2026-03-09T20:57:10.247 INFO:teuthology.orchestra.run.vm01.stderr:/home/ubuntu/cephtest/archive/syslog/journalctl.log: 98.2% -- replaced with /home/ubuntu/cephtest/archive/syslog/journalctl.log.gz 2026-03-09T20:57:10.249 DEBUG:teuthology.run_tasks:Unwinding manager internal.sudo 2026-03-09T20:57:10.252 INFO:teuthology.task.internal:Restoring /etc/sudoers... 2026-03-09T20:57:10.252 DEBUG:teuthology.orchestra.run.vm01:> sudo mv -f /etc/sudoers.orig.teuthology /etc/sudoers 2026-03-09T20:57:10.318 DEBUG:teuthology.orchestra.run.vm06:> sudo mv -f /etc/sudoers.orig.teuthology /etc/sudoers 2026-03-09T20:57:10.342 DEBUG:teuthology.orchestra.run.vm08:> sudo mv -f /etc/sudoers.orig.teuthology /etc/sudoers 2026-03-09T20:57:10.366 DEBUG:teuthology.run_tasks:Unwinding manager internal.coredump 2026-03-09T20:57:10.369 DEBUG:teuthology.orchestra.run.vm01:> sudo sysctl -w kernel.core_pattern=core && sudo bash -c 'for f in `find /home/ubuntu/cephtest/archive/coredump -type f`; do file $f | grep -q systemd-sysusers && rm $f || true ; done' && rmdir --ignore-fail-on-non-empty -- /home/ubuntu/cephtest/archive/coredump 2026-03-09T20:57:10.371 DEBUG:teuthology.orchestra.run.vm06:> sudo sysctl -w kernel.core_pattern=core && sudo bash -c 'for f in `find /home/ubuntu/cephtest/archive/coredump -type f`; do file $f | grep -q systemd-sysusers && rm $f || true ; done' && rmdir --ignore-fail-on-non-empty -- /home/ubuntu/cephtest/archive/coredump 2026-03-09T20:57:10.384 DEBUG:teuthology.orchestra.run.vm08:> sudo sysctl -w kernel.core_pattern=core && sudo bash -c 'for f in `find /home/ubuntu/cephtest/archive/coredump -type f`; do file $f | grep -q systemd-sysusers && rm $f || true ; done' && rmdir --ignore-fail-on-non-empty -- /home/ubuntu/cephtest/archive/coredump 2026-03-09T20:57:10.397 INFO:teuthology.orchestra.run.vm01.stdout:kernel.core_pattern = core 2026-03-09T20:57:10.407 INFO:teuthology.orchestra.run.vm06.stdout:kernel.core_pattern = core 2026-03-09T20:57:10.432 INFO:teuthology.orchestra.run.vm08.stdout:kernel.core_pattern = core 2026-03-09T20:57:10.447 DEBUG:teuthology.orchestra.run.vm01:> test -e /home/ubuntu/cephtest/archive/coredump 2026-03-09T20:57:10.465 DEBUG:teuthology.orchestra.run:got remote process result: 1 2026-03-09T20:57:10.465 DEBUG:teuthology.orchestra.run.vm06:> test -e /home/ubuntu/cephtest/archive/coredump 2026-03-09T20:57:10.480 DEBUG:teuthology.orchestra.run:got remote process result: 1 2026-03-09T20:57:10.480 DEBUG:teuthology.orchestra.run.vm08:> test -e /home/ubuntu/cephtest/archive/coredump 2026-03-09T20:57:10.501 DEBUG:teuthology.orchestra.run:got remote process result: 1 2026-03-09T20:57:10.502 DEBUG:teuthology.run_tasks:Unwinding manager internal.archive 2026-03-09T20:57:10.504 INFO:teuthology.task.internal:Transferring archived files... 2026-03-09T20:57:10.504 DEBUG:teuthology.misc:Transferring archived files from vm01:/home/ubuntu/cephtest/archive to /archive/kyr-2026-03-09_11:23:05-orch-squid-none-default-vps/653/remote/vm01 2026-03-09T20:57:10.504 DEBUG:teuthology.orchestra.run.vm01:> sudo tar c -f - -C /home/ubuntu/cephtest/archive -- . 2026-03-09T20:57:10.533 DEBUG:teuthology.misc:Transferring archived files from vm06:/home/ubuntu/cephtest/archive to /archive/kyr-2026-03-09_11:23:05-orch-squid-none-default-vps/653/remote/vm06 2026-03-09T20:57:10.533 DEBUG:teuthology.orchestra.run.vm06:> sudo tar c -f - -C /home/ubuntu/cephtest/archive -- . 2026-03-09T20:57:10.563 DEBUG:teuthology.misc:Transferring archived files from vm08:/home/ubuntu/cephtest/archive to /archive/kyr-2026-03-09_11:23:05-orch-squid-none-default-vps/653/remote/vm08 2026-03-09T20:57:10.563 DEBUG:teuthology.orchestra.run.vm08:> sudo tar c -f - -C /home/ubuntu/cephtest/archive -- . 2026-03-09T20:57:10.590 INFO:teuthology.task.internal:Removing archive directory... 2026-03-09T20:57:10.590 DEBUG:teuthology.orchestra.run.vm01:> rm -rf -- /home/ubuntu/cephtest/archive 2026-03-09T20:57:10.592 DEBUG:teuthology.orchestra.run.vm06:> rm -rf -- /home/ubuntu/cephtest/archive 2026-03-09T20:57:10.604 DEBUG:teuthology.orchestra.run.vm08:> rm -rf -- /home/ubuntu/cephtest/archive 2026-03-09T20:57:10.645 DEBUG:teuthology.run_tasks:Unwinding manager internal.archive_upload 2026-03-09T20:57:10.648 INFO:teuthology.task.internal:Not uploading archives. 2026-03-09T20:57:10.648 DEBUG:teuthology.run_tasks:Unwinding manager internal.base 2026-03-09T20:57:10.650 INFO:teuthology.task.internal:Tidying up after the test... 2026-03-09T20:57:10.650 DEBUG:teuthology.orchestra.run.vm01:> find /home/ubuntu/cephtest -ls ; rmdir -- /home/ubuntu/cephtest 2026-03-09T20:57:10.652 DEBUG:teuthology.orchestra.run.vm06:> find /home/ubuntu/cephtest -ls ; rmdir -- /home/ubuntu/cephtest 2026-03-09T20:57:10.659 DEBUG:teuthology.orchestra.run.vm08:> find /home/ubuntu/cephtest -ls ; rmdir -- /home/ubuntu/cephtest 2026-03-09T20:57:10.667 INFO:teuthology.orchestra.run.vm01.stdout: 8532144 0 drwxr-xr-x 2 ubuntu ubuntu 6 Mar 9 20:57 /home/ubuntu/cephtest 2026-03-09T20:57:10.674 INFO:teuthology.orchestra.run.vm06.stdout: 8532138 0 drwxr-xr-x 2 ubuntu ubuntu 6 Mar 9 20:57 /home/ubuntu/cephtest 2026-03-09T20:57:10.701 INFO:teuthology.orchestra.run.vm08.stdout: 8532144 0 drwxr-xr-x 2 ubuntu ubuntu 6 Mar 9 20:57 /home/ubuntu/cephtest 2026-03-09T20:57:10.702 DEBUG:teuthology.run_tasks:Unwinding manager console_log 2026-03-09T20:57:10.708 INFO:teuthology.run:Summary data: description: orch/cephadm/workunits/{0-distro/centos_9.stream_runc agent/off mon_election/classic task/test_set_mon_crush_locations} duration: 618.0173983573914 flavor: default owner: kyr success: true 2026-03-09T20:57:10.708 DEBUG:teuthology.report:Pushing job info to http://localhost:8080 2026-03-09T20:57:10.726 INFO:teuthology.run:pass