2026-03-09T15:08:09.368 INFO:root:teuthology version: 1.2.4.dev6+g1c580df7a 2026-03-09T15:08:09.372 DEBUG:teuthology.report:Pushing job info to http://localhost:8080 2026-03-09T15:08:09.391 INFO:teuthology.run:Config: archive_path: /archive/kyr-2026-03-09_11:23:05-orch-squid-none-default-vps/520 branch: squid description: orch/cephadm/workunits/{0-distro/centos_9.stream agent/off mon_election/classic task/test_set_mon_crush_locations} email: null first_in_suite: false flavor: default job_id: '520' last_in_suite: false machine_type: vps name: kyr-2026-03-09_11:23:05-orch-squid-none-default-vps no_nested_subset: false os_type: centos os_version: 9.stream overrides: admin_socket: branch: squid ansible.cephlab: branch: main skip_tags: nagios,monitoring-scripts,hostname,pubkeys,zap,sudoers,kerberos,ntp-client,resolvconf,cpan,nfs vars: timezone: UTC ceph: conf: global: mon election default strategy: 1 mgr: debug mgr: 20 debug ms: 1 mgr/cephadm/use_agent: false mon: debug mon: 20 debug ms: 1 debug paxos: 20 osd: debug ms: 1 debug osd: 20 osd mclock iops capacity threshold hdd: 49000 flavor: default log-ignorelist: - \(MDS_ALL_DOWN\) - \(MDS_UP_LESS_THAN_MAX\) - MON_DOWN - POOL_APP_NOT_ENABLED - mon down - mons down - out of quorum - CEPHADM_FAILED_DAEMON log-only-match: - CEPHADM_ sha1: e911bdebe5c8faa3800735d1568fcdca65db60df ceph-deploy: conf: client: log file: /var/log/ceph/ceph-$name.$pid.log mon: {} install: ceph: flavor: default sha1: e911bdebe5c8faa3800735d1568fcdca65db60df extra_system_packages: deb: - python3-xmltodict - python3-jmespath rpm: - bzip2 - perl-Test-Harness - python3-xmltodict - python3-jmespath selinux: allowlist: - scontext=system_u:system_r:logrotate_t:s0 - scontext=system_u:system_r:getty_t:s0 workunit: branch: tt-squid sha1: 569c3e99c9b32a51b4eaf08731c728f4513ed589 owner: kyr priority: 1000 repo: https://github.com/ceph/ceph.git roles: - - host.a - osd.0 - mon.a - mgr.a - - host.b - osd.1 - mon.b - mgr.b - - host.c - osd.2 - mon.c seed: 3443 sha1: e911bdebe5c8faa3800735d1568fcdca65db60df sleep_before_teardown: 0 subset: 1/64 suite: orch suite_branch: tt-squid suite_path: /home/teuthos/src/github.com_kshtsk_ceph_569c3e99c9b32a51b4eaf08731c728f4513ed589/qa suite_relpath: qa suite_repo: https://github.com/kshtsk/ceph.git suite_sha1: 569c3e99c9b32a51b4eaf08731c728f4513ed589 targets: vm03.local: ecdsa-sha2-nistp256 AAAAE2VjZHNhLXNoYTItbmlzdHAyNTYAAAAIbmlzdHAyNTYAAABBBBqL8H6Ny3L7dnamOIj6JOFR5SfEN5DN/adwm0HDFBzgP2qcTVra/scVuimrcP8ynB+V2OoU5toUSTc5QGdS38Q= vm04.local: ecdsa-sha2-nistp256 AAAAE2VjZHNhLXNoYTItbmlzdHAyNTYAAAAIbmlzdHAyNTYAAABBBB4Q+WZcyXfJJaCF0FnANgSeXbnVf+6yJC3ELqzvEpms7y3qlS1u+m0Pn0ZbJQ9y0l2MLVDS5J6jZOQQVvEjVmc= vm06.local: ecdsa-sha2-nistp256 AAAAE2VjZHNhLXNoYTItbmlzdHAyNTYAAAAIbmlzdHAyNTYAAABBBKvnohRPbAvOWFsioyWUed7254O2Gdrrx6i0zJ9WROWcqAp23hKrpZv9AFGf3Waj08xGiSGxmtX+udBHfMBg9AY= tasks: - pexec: all: - sudo dnf remove nvme-cli -y - sudo dnf install nvmetcli nvme-cli -y - install: null - cephadm: null - cephadm.apply: specs: - placement: count: 3 service_id: foo service_type: mon spec: crush_locations: host.a: - datacenter=a host.b: - datacenter=b - rack=2 host.c: - datacenter=a - rack=3 - cephadm.shell: host.a: - "set -ex\n# since we don't know the real hostnames before the test, the next\n\ # bit is in order to replace the fake hostnames \"host.a/b/c\" with\n# the actual\ \ names cephadm knows the host by within the mon spec\nceph orch host ls --format\ \ json | jq -r '.[] | .hostname' > realnames\necho $'host.a\\nhost.b\\nhost.c'\ \ > fakenames\necho $'a\\nb\\nc' > mon_ids\necho $'{datacenter=a}\\n{datacenter=b,rack=2}\\\ n{datacenter=a,rack=3}' > crush_locs\nceph orch ls --service-name mon --export\ \ > mon.yaml\nMONSPEC=`cat mon.yaml`\necho \"$MONSPEC\"\nwhile read realname\ \ <&3 && read fakename <&4; do\n MONSPEC=\"${MONSPEC//$fakename/$realname}\"\ \ndone 3 mon.yaml\ncat mon.yaml\n\ # now the spec should have the real hostnames, so let's re-apply\nceph orch\ \ apply -i mon.yaml\nsleep 90\nceph orch ps --refresh\nceph orch ls --service-name\ \ mon --export > mon.yaml; ceph orch apply -i mon.yaml\nsleep 90\nceph mon dump\n\ ceph mon dump --format json\n# verify all the crush locations got set from \"\ ceph mon dump\" output\nwhile read monid <&3 && read crushloc <&4; do\n ceph\ \ mon dump --format json | jq --arg monid \"$monid\" --arg crushloc \"$crushloc\"\ \ -e '.mons | .[] | select(.name == $monid) | .crush_location == $crushloc'\n\ done 3, func=.kill_console_loggers at 0x7f139f772170>, signals=[15]) 2026-03-09T15:08:10.230 INFO:teuthology.run_tasks:Running task internal.connect... 2026-03-09T15:08:10.230 INFO:teuthology.task.internal:Opening connections... 2026-03-09T15:08:10.230 DEBUG:teuthology.task.internal:connecting to ubuntu@vm03.local 2026-03-09T15:08:10.231 DEBUG:teuthology.orchestra.connection:{'hostname': 'vm03.local', 'username': 'ubuntu', 'timeout': 60} 2026-03-09T15:08:10.291 DEBUG:teuthology.task.internal:connecting to ubuntu@vm04.local 2026-03-09T15:08:10.292 DEBUG:teuthology.orchestra.connection:{'hostname': 'vm04.local', 'username': 'ubuntu', 'timeout': 60} 2026-03-09T15:08:10.349 DEBUG:teuthology.task.internal:connecting to ubuntu@vm06.local 2026-03-09T15:08:10.349 DEBUG:teuthology.orchestra.connection:{'hostname': 'vm06.local', 'username': 'ubuntu', 'timeout': 60} 2026-03-09T15:08:10.406 INFO:teuthology.run_tasks:Running task internal.push_inventory... 2026-03-09T15:08:10.407 DEBUG:teuthology.orchestra.run.vm03:> uname -m 2026-03-09T15:08:10.421 INFO:teuthology.orchestra.run.vm03.stdout:x86_64 2026-03-09T15:08:10.421 DEBUG:teuthology.orchestra.run.vm03:> cat /etc/os-release 2026-03-09T15:08:10.477 INFO:teuthology.orchestra.run.vm03.stdout:NAME="CentOS Stream" 2026-03-09T15:08:10.477 INFO:teuthology.orchestra.run.vm03.stdout:VERSION="9" 2026-03-09T15:08:10.477 INFO:teuthology.orchestra.run.vm03.stdout:ID="centos" 2026-03-09T15:08:10.477 INFO:teuthology.orchestra.run.vm03.stdout:ID_LIKE="rhel fedora" 2026-03-09T15:08:10.477 INFO:teuthology.orchestra.run.vm03.stdout:VERSION_ID="9" 2026-03-09T15:08:10.477 INFO:teuthology.orchestra.run.vm03.stdout:PLATFORM_ID="platform:el9" 2026-03-09T15:08:10.477 INFO:teuthology.orchestra.run.vm03.stdout:PRETTY_NAME="CentOS Stream 9" 2026-03-09T15:08:10.477 INFO:teuthology.orchestra.run.vm03.stdout:ANSI_COLOR="0;31" 2026-03-09T15:08:10.477 INFO:teuthology.orchestra.run.vm03.stdout:LOGO="fedora-logo-icon" 2026-03-09T15:08:10.477 INFO:teuthology.orchestra.run.vm03.stdout:CPE_NAME="cpe:/o:centos:centos:9" 2026-03-09T15:08:10.477 INFO:teuthology.orchestra.run.vm03.stdout:HOME_URL="https://centos.org/" 2026-03-09T15:08:10.477 INFO:teuthology.orchestra.run.vm03.stdout:BUG_REPORT_URL="https://issues.redhat.com/" 2026-03-09T15:08:10.477 INFO:teuthology.orchestra.run.vm03.stdout:REDHAT_SUPPORT_PRODUCT="Red Hat Enterprise Linux 9" 2026-03-09T15:08:10.477 INFO:teuthology.orchestra.run.vm03.stdout:REDHAT_SUPPORT_PRODUCT_VERSION="CentOS Stream" 2026-03-09T15:08:10.478 INFO:teuthology.lock.ops:Updating vm03.local on lock server 2026-03-09T15:08:10.482 DEBUG:teuthology.orchestra.run.vm04:> uname -m 2026-03-09T15:08:10.496 INFO:teuthology.orchestra.run.vm04.stdout:x86_64 2026-03-09T15:08:10.496 DEBUG:teuthology.orchestra.run.vm04:> cat /etc/os-release 2026-03-09T15:08:10.549 INFO:teuthology.orchestra.run.vm04.stdout:NAME="CentOS Stream" 2026-03-09T15:08:10.549 INFO:teuthology.orchestra.run.vm04.stdout:VERSION="9" 2026-03-09T15:08:10.549 INFO:teuthology.orchestra.run.vm04.stdout:ID="centos" 2026-03-09T15:08:10.549 INFO:teuthology.orchestra.run.vm04.stdout:ID_LIKE="rhel fedora" 2026-03-09T15:08:10.549 INFO:teuthology.orchestra.run.vm04.stdout:VERSION_ID="9" 2026-03-09T15:08:10.549 INFO:teuthology.orchestra.run.vm04.stdout:PLATFORM_ID="platform:el9" 2026-03-09T15:08:10.549 INFO:teuthology.orchestra.run.vm04.stdout:PRETTY_NAME="CentOS Stream 9" 2026-03-09T15:08:10.549 INFO:teuthology.orchestra.run.vm04.stdout:ANSI_COLOR="0;31" 2026-03-09T15:08:10.549 INFO:teuthology.orchestra.run.vm04.stdout:LOGO="fedora-logo-icon" 2026-03-09T15:08:10.549 INFO:teuthology.orchestra.run.vm04.stdout:CPE_NAME="cpe:/o:centos:centos:9" 2026-03-09T15:08:10.549 INFO:teuthology.orchestra.run.vm04.stdout:HOME_URL="https://centos.org/" 2026-03-09T15:08:10.549 INFO:teuthology.orchestra.run.vm04.stdout:BUG_REPORT_URL="https://issues.redhat.com/" 2026-03-09T15:08:10.549 INFO:teuthology.orchestra.run.vm04.stdout:REDHAT_SUPPORT_PRODUCT="Red Hat Enterprise Linux 9" 2026-03-09T15:08:10.549 INFO:teuthology.orchestra.run.vm04.stdout:REDHAT_SUPPORT_PRODUCT_VERSION="CentOS Stream" 2026-03-09T15:08:10.549 INFO:teuthology.lock.ops:Updating vm04.local on lock server 2026-03-09T15:08:10.554 DEBUG:teuthology.orchestra.run.vm06:> uname -m 2026-03-09T15:08:10.567 INFO:teuthology.orchestra.run.vm06.stdout:x86_64 2026-03-09T15:08:10.567 DEBUG:teuthology.orchestra.run.vm06:> cat /etc/os-release 2026-03-09T15:08:10.620 INFO:teuthology.orchestra.run.vm06.stdout:NAME="CentOS Stream" 2026-03-09T15:08:10.620 INFO:teuthology.orchestra.run.vm06.stdout:VERSION="9" 2026-03-09T15:08:10.620 INFO:teuthology.orchestra.run.vm06.stdout:ID="centos" 2026-03-09T15:08:10.621 INFO:teuthology.orchestra.run.vm06.stdout:ID_LIKE="rhel fedora" 2026-03-09T15:08:10.621 INFO:teuthology.orchestra.run.vm06.stdout:VERSION_ID="9" 2026-03-09T15:08:10.621 INFO:teuthology.orchestra.run.vm06.stdout:PLATFORM_ID="platform:el9" 2026-03-09T15:08:10.621 INFO:teuthology.orchestra.run.vm06.stdout:PRETTY_NAME="CentOS Stream 9" 2026-03-09T15:08:10.621 INFO:teuthology.orchestra.run.vm06.stdout:ANSI_COLOR="0;31" 2026-03-09T15:08:10.621 INFO:teuthology.orchestra.run.vm06.stdout:LOGO="fedora-logo-icon" 2026-03-09T15:08:10.621 INFO:teuthology.orchestra.run.vm06.stdout:CPE_NAME="cpe:/o:centos:centos:9" 2026-03-09T15:08:10.621 INFO:teuthology.orchestra.run.vm06.stdout:HOME_URL="https://centos.org/" 2026-03-09T15:08:10.621 INFO:teuthology.orchestra.run.vm06.stdout:BUG_REPORT_URL="https://issues.redhat.com/" 2026-03-09T15:08:10.621 INFO:teuthology.orchestra.run.vm06.stdout:REDHAT_SUPPORT_PRODUCT="Red Hat Enterprise Linux 9" 2026-03-09T15:08:10.621 INFO:teuthology.orchestra.run.vm06.stdout:REDHAT_SUPPORT_PRODUCT_VERSION="CentOS Stream" 2026-03-09T15:08:10.621 INFO:teuthology.lock.ops:Updating vm06.local on lock server 2026-03-09T15:08:10.625 INFO:teuthology.run_tasks:Running task internal.serialize_remote_roles... 2026-03-09T15:08:10.627 INFO:teuthology.run_tasks:Running task internal.check_conflict... 2026-03-09T15:08:10.627 INFO:teuthology.task.internal:Checking for old test directory... 2026-03-09T15:08:10.627 DEBUG:teuthology.orchestra.run.vm03:> test '!' -e /home/ubuntu/cephtest 2026-03-09T15:08:10.629 DEBUG:teuthology.orchestra.run.vm04:> test '!' -e /home/ubuntu/cephtest 2026-03-09T15:08:10.631 DEBUG:teuthology.orchestra.run.vm06:> test '!' -e /home/ubuntu/cephtest 2026-03-09T15:08:10.674 INFO:teuthology.run_tasks:Running task internal.check_ceph_data... 2026-03-09T15:08:10.675 INFO:teuthology.task.internal:Checking for non-empty /var/lib/ceph... 2026-03-09T15:08:10.675 DEBUG:teuthology.orchestra.run.vm03:> test -z $(ls -A /var/lib/ceph) 2026-03-09T15:08:10.685 DEBUG:teuthology.orchestra.run.vm04:> test -z $(ls -A /var/lib/ceph) 2026-03-09T15:08:10.687 DEBUG:teuthology.orchestra.run.vm06:> test -z $(ls -A /var/lib/ceph) 2026-03-09T15:08:10.701 INFO:teuthology.orchestra.run.vm03.stderr:ls: cannot access '/var/lib/ceph': No such file or directory 2026-03-09T15:08:10.702 INFO:teuthology.orchestra.run.vm04.stderr:ls: cannot access '/var/lib/ceph': No such file or directory 2026-03-09T15:08:10.729 INFO:teuthology.orchestra.run.vm06.stderr:ls: cannot access '/var/lib/ceph': No such file or directory 2026-03-09T15:08:10.729 INFO:teuthology.run_tasks:Running task internal.vm_setup... 2026-03-09T15:08:10.736 DEBUG:teuthology.orchestra.run.vm03:> test -e /ceph-qa-ready 2026-03-09T15:08:10.754 DEBUG:teuthology.orchestra.run:got remote process result: 1 2026-03-09T15:08:10.954 DEBUG:teuthology.orchestra.run.vm04:> test -e /ceph-qa-ready 2026-03-09T15:08:10.968 DEBUG:teuthology.orchestra.run:got remote process result: 1 2026-03-09T15:08:11.154 DEBUG:teuthology.orchestra.run.vm06:> test -e /ceph-qa-ready 2026-03-09T15:08:11.168 DEBUG:teuthology.orchestra.run:got remote process result: 1 2026-03-09T15:08:11.343 INFO:teuthology.run_tasks:Running task internal.base... 2026-03-09T15:08:11.344 INFO:teuthology.task.internal:Creating test directory... 2026-03-09T15:08:11.344 DEBUG:teuthology.orchestra.run.vm03:> mkdir -p -m0755 -- /home/ubuntu/cephtest 2026-03-09T15:08:11.346 DEBUG:teuthology.orchestra.run.vm04:> mkdir -p -m0755 -- /home/ubuntu/cephtest 2026-03-09T15:08:11.348 DEBUG:teuthology.orchestra.run.vm06:> mkdir -p -m0755 -- /home/ubuntu/cephtest 2026-03-09T15:08:11.363 INFO:teuthology.run_tasks:Running task internal.archive_upload... 2026-03-09T15:08:11.365 INFO:teuthology.run_tasks:Running task internal.archive... 2026-03-09T15:08:11.366 INFO:teuthology.task.internal:Creating archive directory... 2026-03-09T15:08:11.366 DEBUG:teuthology.orchestra.run.vm03:> install -d -m0755 -- /home/ubuntu/cephtest/archive 2026-03-09T15:08:11.401 DEBUG:teuthology.orchestra.run.vm04:> install -d -m0755 -- /home/ubuntu/cephtest/archive 2026-03-09T15:08:11.405 DEBUG:teuthology.orchestra.run.vm06:> install -d -m0755 -- /home/ubuntu/cephtest/archive 2026-03-09T15:08:11.421 INFO:teuthology.run_tasks:Running task internal.coredump... 2026-03-09T15:08:11.422 INFO:teuthology.task.internal:Enabling coredump saving... 2026-03-09T15:08:11.423 DEBUG:teuthology.orchestra.run.vm03:> test -f /run/.containerenv -o -f /.dockerenv 2026-03-09T15:08:11.468 DEBUG:teuthology.orchestra.run:got remote process result: 1 2026-03-09T15:08:11.468 DEBUG:teuthology.orchestra.run.vm04:> test -f /run/.containerenv -o -f /.dockerenv 2026-03-09T15:08:11.481 DEBUG:teuthology.orchestra.run:got remote process result: 1 2026-03-09T15:08:11.481 DEBUG:teuthology.orchestra.run.vm06:> test -f /run/.containerenv -o -f /.dockerenv 2026-03-09T15:08:11.495 DEBUG:teuthology.orchestra.run:got remote process result: 1 2026-03-09T15:08:11.495 DEBUG:teuthology.orchestra.run.vm03:> install -d -m0755 -- /home/ubuntu/cephtest/archive/coredump && sudo sysctl -w kernel.core_pattern=/home/ubuntu/cephtest/archive/coredump/%t.%p.core && echo kernel.core_pattern=/home/ubuntu/cephtest/archive/coredump/%t.%p.core | sudo tee -a /etc/sysctl.conf 2026-03-09T15:08:11.510 DEBUG:teuthology.orchestra.run.vm04:> install -d -m0755 -- /home/ubuntu/cephtest/archive/coredump && sudo sysctl -w kernel.core_pattern=/home/ubuntu/cephtest/archive/coredump/%t.%p.core && echo kernel.core_pattern=/home/ubuntu/cephtest/archive/coredump/%t.%p.core | sudo tee -a /etc/sysctl.conf 2026-03-09T15:08:11.522 DEBUG:teuthology.orchestra.run.vm06:> install -d -m0755 -- /home/ubuntu/cephtest/archive/coredump && sudo sysctl -w kernel.core_pattern=/home/ubuntu/cephtest/archive/coredump/%t.%p.core && echo kernel.core_pattern=/home/ubuntu/cephtest/archive/coredump/%t.%p.core | sudo tee -a /etc/sysctl.conf 2026-03-09T15:08:11.531 INFO:teuthology.orchestra.run.vm03.stdout:kernel.core_pattern = /home/ubuntu/cephtest/archive/coredump/%t.%p.core 2026-03-09T15:08:11.539 INFO:teuthology.orchestra.run.vm03.stdout:kernel.core_pattern=/home/ubuntu/cephtest/archive/coredump/%t.%p.core 2026-03-09T15:08:11.543 INFO:teuthology.orchestra.run.vm04.stdout:kernel.core_pattern = /home/ubuntu/cephtest/archive/coredump/%t.%p.core 2026-03-09T15:08:11.552 INFO:teuthology.orchestra.run.vm04.stdout:kernel.core_pattern=/home/ubuntu/cephtest/archive/coredump/%t.%p.core 2026-03-09T15:08:11.558 INFO:teuthology.orchestra.run.vm06.stdout:kernel.core_pattern = /home/ubuntu/cephtest/archive/coredump/%t.%p.core 2026-03-09T15:08:11.566 INFO:teuthology.orchestra.run.vm06.stdout:kernel.core_pattern=/home/ubuntu/cephtest/archive/coredump/%t.%p.core 2026-03-09T15:08:11.567 INFO:teuthology.run_tasks:Running task internal.sudo... 2026-03-09T15:08:11.569 INFO:teuthology.task.internal:Configuring sudo... 2026-03-09T15:08:11.569 DEBUG:teuthology.orchestra.run.vm03:> sudo sed -i.orig.teuthology -e 's/^\([^#]*\) \(requiretty\)/\1 !\2/g' -e 's/^\([^#]*\) !\(visiblepw\)/\1 \2/g' /etc/sudoers 2026-03-09T15:08:11.582 DEBUG:teuthology.orchestra.run.vm04:> sudo sed -i.orig.teuthology -e 's/^\([^#]*\) \(requiretty\)/\1 !\2/g' -e 's/^\([^#]*\) !\(visiblepw\)/\1 \2/g' /etc/sudoers 2026-03-09T15:08:11.595 DEBUG:teuthology.orchestra.run.vm06:> sudo sed -i.orig.teuthology -e 's/^\([^#]*\) \(requiretty\)/\1 !\2/g' -e 's/^\([^#]*\) !\(visiblepw\)/\1 \2/g' /etc/sudoers 2026-03-09T15:08:11.630 INFO:teuthology.run_tasks:Running task internal.syslog... 2026-03-09T15:08:11.633 INFO:teuthology.task.internal.syslog:Starting syslog monitoring... 2026-03-09T15:08:11.633 DEBUG:teuthology.orchestra.run.vm03:> mkdir -p -m0755 -- /home/ubuntu/cephtest/archive/syslog 2026-03-09T15:08:11.644 DEBUG:teuthology.orchestra.run.vm04:> mkdir -p -m0755 -- /home/ubuntu/cephtest/archive/syslog 2026-03-09T15:08:11.657 DEBUG:teuthology.orchestra.run.vm06:> mkdir -p -m0755 -- /home/ubuntu/cephtest/archive/syslog 2026-03-09T15:08:11.684 DEBUG:teuthology.orchestra.run.vm03:> install -m 666 /dev/null /home/ubuntu/cephtest/archive/syslog/kern.log 2026-03-09T15:08:11.718 DEBUG:teuthology.orchestra.run.vm03:> install -m 666 /dev/null /home/ubuntu/cephtest/archive/syslog/misc.log 2026-03-09T15:08:11.771 DEBUG:teuthology.orchestra.run.vm03:> set -ex 2026-03-09T15:08:11.771 DEBUG:teuthology.orchestra.run.vm03:> sudo dd of=/etc/rsyslog.d/80-cephtest.conf 2026-03-09T15:08:11.829 DEBUG:teuthology.orchestra.run.vm04:> install -m 666 /dev/null /home/ubuntu/cephtest/archive/syslog/kern.log 2026-03-09T15:08:11.849 DEBUG:teuthology.orchestra.run.vm04:> install -m 666 /dev/null /home/ubuntu/cephtest/archive/syslog/misc.log 2026-03-09T15:08:11.904 DEBUG:teuthology.orchestra.run.vm04:> set -ex 2026-03-09T15:08:11.904 DEBUG:teuthology.orchestra.run.vm04:> sudo dd of=/etc/rsyslog.d/80-cephtest.conf 2026-03-09T15:08:11.962 DEBUG:teuthology.orchestra.run.vm06:> install -m 666 /dev/null /home/ubuntu/cephtest/archive/syslog/kern.log 2026-03-09T15:08:11.985 DEBUG:teuthology.orchestra.run.vm06:> install -m 666 /dev/null /home/ubuntu/cephtest/archive/syslog/misc.log 2026-03-09T15:08:12.040 DEBUG:teuthology.orchestra.run.vm06:> set -ex 2026-03-09T15:08:12.040 DEBUG:teuthology.orchestra.run.vm06:> sudo dd of=/etc/rsyslog.d/80-cephtest.conf 2026-03-09T15:08:12.098 DEBUG:teuthology.orchestra.run.vm03:> sudo service rsyslog restart 2026-03-09T15:08:12.099 DEBUG:teuthology.orchestra.run.vm04:> sudo service rsyslog restart 2026-03-09T15:08:12.100 DEBUG:teuthology.orchestra.run.vm06:> sudo service rsyslog restart 2026-03-09T15:08:12.126 INFO:teuthology.orchestra.run.vm03.stderr:Redirecting to /bin/systemctl restart rsyslog.service 2026-03-09T15:08:12.127 INFO:teuthology.orchestra.run.vm04.stderr:Redirecting to /bin/systemctl restart rsyslog.service 2026-03-09T15:08:12.163 INFO:teuthology.orchestra.run.vm06.stderr:Redirecting to /bin/systemctl restart rsyslog.service 2026-03-09T15:08:12.619 INFO:teuthology.run_tasks:Running task internal.timer... 2026-03-09T15:08:12.621 INFO:teuthology.task.internal:Starting timer... 2026-03-09T15:08:12.621 INFO:teuthology.run_tasks:Running task pcp... 2026-03-09T15:08:12.624 INFO:teuthology.run_tasks:Running task selinux... 2026-03-09T15:08:12.626 DEBUG:teuthology.task:Applying overrides for task selinux: {'allowlist': ['scontext=system_u:system_r:logrotate_t:s0', 'scontext=system_u:system_r:getty_t:s0']} 2026-03-09T15:08:12.626 INFO:teuthology.task.selinux:Excluding vm03: VMs are not yet supported 2026-03-09T15:08:12.626 INFO:teuthology.task.selinux:Excluding vm04: VMs are not yet supported 2026-03-09T15:08:12.626 INFO:teuthology.task.selinux:Excluding vm06: VMs are not yet supported 2026-03-09T15:08:12.626 DEBUG:teuthology.task.selinux:Getting current SELinux state 2026-03-09T15:08:12.626 DEBUG:teuthology.task.selinux:Existing SELinux modes: {} 2026-03-09T15:08:12.626 INFO:teuthology.task.selinux:Putting SELinux into permissive mode 2026-03-09T15:08:12.626 INFO:teuthology.run_tasks:Running task ansible.cephlab... 2026-03-09T15:08:12.628 DEBUG:teuthology.task:Applying overrides for task ansible.cephlab: {'branch': 'main', 'skip_tags': 'nagios,monitoring-scripts,hostname,pubkeys,zap,sudoers,kerberos,ntp-client,resolvconf,cpan,nfs', 'vars': {'timezone': 'UTC'}} 2026-03-09T15:08:12.628 DEBUG:teuthology.repo_utils:Setting repo remote to https://github.com/ceph/ceph-cm-ansible.git 2026-03-09T15:08:12.630 INFO:teuthology.repo_utils:Fetching github.com_ceph_ceph-cm-ansible_main from origin 2026-03-09T15:08:13.303 DEBUG:teuthology.repo_utils:Resetting repo at /home/teuthos/src/github.com_ceph_ceph-cm-ansible_main to origin/main 2026-03-09T15:08:13.308 INFO:teuthology.task.ansible:Playbook: [{'import_playbook': 'ansible_managed.yml'}, {'import_playbook': 'teuthology.yml'}, {'hosts': 'testnodes', 'tasks': [{'set_fact': {'ran_from_cephlab_playbook': True}}]}, {'import_playbook': 'testnodes.yml'}, {'import_playbook': 'container-host.yml'}, {'import_playbook': 'cobbler.yml'}, {'import_playbook': 'paddles.yml'}, {'import_playbook': 'pulpito.yml'}, {'hosts': 'testnodes', 'become': True, 'tasks': [{'name': 'Touch /ceph-qa-ready', 'file': {'path': '/ceph-qa-ready', 'state': 'touch'}, 'when': 'ran_from_cephlab_playbook|bool'}]}] 2026-03-09T15:08:13.309 DEBUG:teuthology.task.ansible:Running ansible-playbook -v --extra-vars '{"ansible_ssh_user": "ubuntu", "timezone": "UTC"}' -i /tmp/teuth_ansible_inventoryl78kfue0 --limit vm03.local,vm04.local,vm06.local /home/teuthos/src/github.com_ceph_ceph-cm-ansible_main/cephlab.yml --skip-tags nagios,monitoring-scripts,hostname,pubkeys,zap,sudoers,kerberos,ntp-client,resolvconf,cpan,nfs 2026-03-09T15:10:25.032 DEBUG:teuthology.task.ansible:Reconnecting to [Remote(name='ubuntu@vm03.local'), Remote(name='ubuntu@vm04.local'), Remote(name='ubuntu@vm06.local')] 2026-03-09T15:10:25.033 INFO:teuthology.orchestra.remote:Trying to reconnect to host 'ubuntu@vm03.local' 2026-03-09T15:10:25.033 DEBUG:teuthology.orchestra.connection:{'hostname': 'vm03.local', 'username': 'ubuntu', 'timeout': 60} 2026-03-09T15:10:25.100 DEBUG:teuthology.orchestra.run.vm03:> true 2026-03-09T15:10:25.179 INFO:teuthology.orchestra.remote:Successfully reconnected to host 'ubuntu@vm03.local' 2026-03-09T15:10:25.179 INFO:teuthology.orchestra.remote:Trying to reconnect to host 'ubuntu@vm04.local' 2026-03-09T15:10:25.179 DEBUG:teuthology.orchestra.connection:{'hostname': 'vm04.local', 'username': 'ubuntu', 'timeout': 60} 2026-03-09T15:10:25.243 DEBUG:teuthology.orchestra.run.vm04:> true 2026-03-09T15:10:25.323 INFO:teuthology.orchestra.remote:Successfully reconnected to host 'ubuntu@vm04.local' 2026-03-09T15:10:25.323 INFO:teuthology.orchestra.remote:Trying to reconnect to host 'ubuntu@vm06.local' 2026-03-09T15:10:25.324 DEBUG:teuthology.orchestra.connection:{'hostname': 'vm06.local', 'username': 'ubuntu', 'timeout': 60} 2026-03-09T15:10:25.389 DEBUG:teuthology.orchestra.run.vm06:> true 2026-03-09T15:10:25.470 INFO:teuthology.orchestra.remote:Successfully reconnected to host 'ubuntu@vm06.local' 2026-03-09T15:10:25.470 INFO:teuthology.run_tasks:Running task clock... 2026-03-09T15:10:25.472 INFO:teuthology.task.clock:Syncing clocks and checking initial clock skew... 2026-03-09T15:10:25.473 INFO:teuthology.orchestra.run:Running command with timeout 360 2026-03-09T15:10:25.473 DEBUG:teuthology.orchestra.run.vm03:> sudo systemctl stop ntp.service || sudo systemctl stop ntpd.service || sudo systemctl stop chronyd.service ; sudo ntpd -gq || sudo chronyc makestep ; sudo systemctl start ntp.service || sudo systemctl start ntpd.service || sudo systemctl start chronyd.service ; PATH=/usr/bin:/usr/sbin ntpq -p || PATH=/usr/bin:/usr/sbin chronyc sources || true 2026-03-09T15:10:25.477 INFO:teuthology.orchestra.run:Running command with timeout 360 2026-03-09T15:10:25.477 DEBUG:teuthology.orchestra.run.vm04:> sudo systemctl stop ntp.service || sudo systemctl stop ntpd.service || sudo systemctl stop chronyd.service ; sudo ntpd -gq || sudo chronyc makestep ; sudo systemctl start ntp.service || sudo systemctl start ntpd.service || sudo systemctl start chronyd.service ; PATH=/usr/bin:/usr/sbin ntpq -p || PATH=/usr/bin:/usr/sbin chronyc sources || true 2026-03-09T15:10:25.481 INFO:teuthology.orchestra.run:Running command with timeout 360 2026-03-09T15:10:25.481 DEBUG:teuthology.orchestra.run.vm06:> sudo systemctl stop ntp.service || sudo systemctl stop ntpd.service || sudo systemctl stop chronyd.service ; sudo ntpd -gq || sudo chronyc makestep ; sudo systemctl start ntp.service || sudo systemctl start ntpd.service || sudo systemctl start chronyd.service ; PATH=/usr/bin:/usr/sbin ntpq -p || PATH=/usr/bin:/usr/sbin chronyc sources || true 2026-03-09T15:10:25.506 INFO:teuthology.orchestra.run.vm03.stderr:Failed to stop ntp.service: Unit ntp.service not loaded. 2026-03-09T15:10:25.516 INFO:teuthology.orchestra.run.vm04.stderr:Failed to stop ntp.service: Unit ntp.service not loaded. 2026-03-09T15:10:25.521 INFO:teuthology.orchestra.run.vm03.stderr:Failed to stop ntpd.service: Unit ntpd.service not loaded. 2026-03-09T15:10:25.532 INFO:teuthology.orchestra.run.vm04.stderr:Failed to stop ntpd.service: Unit ntpd.service not loaded. 2026-03-09T15:10:25.547 INFO:teuthology.orchestra.run.vm06.stderr:Failed to stop ntp.service: Unit ntp.service not loaded. 2026-03-09T15:10:25.551 INFO:teuthology.orchestra.run.vm03.stderr:sudo: ntpd: command not found 2026-03-09T15:10:25.558 INFO:teuthology.orchestra.run.vm04.stderr:sudo: ntpd: command not found 2026-03-09T15:10:25.563 INFO:teuthology.orchestra.run.vm03.stdout:506 Cannot talk to daemon 2026-03-09T15:10:25.564 INFO:teuthology.orchestra.run.vm06.stderr:Failed to stop ntpd.service: Unit ntpd.service not loaded. 2026-03-09T15:10:25.570 INFO:teuthology.orchestra.run.vm04.stdout:506 Cannot talk to daemon 2026-03-09T15:10:25.577 INFO:teuthology.orchestra.run.vm03.stderr:Failed to start ntp.service: Unit ntp.service not found. 2026-03-09T15:10:25.584 INFO:teuthology.orchestra.run.vm04.stderr:Failed to start ntp.service: Unit ntp.service not found. 2026-03-09T15:10:25.591 INFO:teuthology.orchestra.run.vm03.stderr:Failed to start ntpd.service: Unit ntpd.service not found. 2026-03-09T15:10:25.598 INFO:teuthology.orchestra.run.vm04.stderr:Failed to start ntpd.service: Unit ntpd.service not found. 2026-03-09T15:10:25.599 INFO:teuthology.orchestra.run.vm06.stderr:sudo: ntpd: command not found 2026-03-09T15:10:25.614 INFO:teuthology.orchestra.run.vm06.stdout:506 Cannot talk to daemon 2026-03-09T15:10:25.636 INFO:teuthology.orchestra.run.vm06.stderr:Failed to start ntp.service: Unit ntp.service not found. 2026-03-09T15:10:25.637 INFO:teuthology.orchestra.run.vm03.stderr:bash: line 1: ntpq: command not found 2026-03-09T15:10:25.639 INFO:teuthology.orchestra.run.vm03.stdout:MS Name/IP address Stratum Poll Reach LastRx Last sample 2026-03-09T15:10:25.639 INFO:teuthology.orchestra.run.vm03.stdout:=============================================================================== 2026-03-09T15:10:25.643 INFO:teuthology.orchestra.run.vm04.stderr:bash: line 1: ntpq: command not found 2026-03-09T15:10:25.650 INFO:teuthology.orchestra.run.vm04.stdout:MS Name/IP address Stratum Poll Reach LastRx Last sample 2026-03-09T15:10:25.650 INFO:teuthology.orchestra.run.vm04.stdout:=============================================================================== 2026-03-09T15:10:25.658 INFO:teuthology.orchestra.run.vm06.stderr:Failed to start ntpd.service: Unit ntpd.service not found. 2026-03-09T15:10:25.707 INFO:teuthology.orchestra.run.vm06.stderr:bash: line 1: ntpq: command not found 2026-03-09T15:10:25.708 INFO:teuthology.orchestra.run.vm06.stdout:MS Name/IP address Stratum Poll Reach LastRx Last sample 2026-03-09T15:10:25.709 INFO:teuthology.orchestra.run.vm06.stdout:=============================================================================== 2026-03-09T15:10:25.709 INFO:teuthology.run_tasks:Running task pexec... 2026-03-09T15:10:25.711 INFO:teuthology.task.pexec:Executing custom commands... 2026-03-09T15:10:25.712 DEBUG:teuthology.orchestra.run.vm03:> TESTDIR=/home/ubuntu/cephtest bash -s 2026-03-09T15:10:25.712 DEBUG:teuthology.orchestra.run.vm04:> TESTDIR=/home/ubuntu/cephtest bash -s 2026-03-09T15:10:25.712 DEBUG:teuthology.orchestra.run.vm06:> TESTDIR=/home/ubuntu/cephtest bash -s 2026-03-09T15:10:25.715 DEBUG:teuthology.task.pexec:ubuntu@vm03.local< sudo dnf remove nvme-cli -y 2026-03-09T15:10:25.715 DEBUG:teuthology.task.pexec:ubuntu@vm03.local< sudo dnf install nvmetcli nvme-cli -y 2026-03-09T15:10:25.715 INFO:teuthology.task.pexec:Running commands on host ubuntu@vm03.local 2026-03-09T15:10:25.715 INFO:teuthology.task.pexec:sudo dnf remove nvme-cli -y 2026-03-09T15:10:25.715 INFO:teuthology.task.pexec:sudo dnf install nvmetcli nvme-cli -y 2026-03-09T15:10:25.715 DEBUG:teuthology.task.pexec:ubuntu@vm04.local< sudo dnf remove nvme-cli -y 2026-03-09T15:10:25.715 DEBUG:teuthology.task.pexec:ubuntu@vm04.local< sudo dnf install nvmetcli nvme-cli -y 2026-03-09T15:10:25.715 INFO:teuthology.task.pexec:Running commands on host ubuntu@vm04.local 2026-03-09T15:10:25.715 INFO:teuthology.task.pexec:sudo dnf remove nvme-cli -y 2026-03-09T15:10:25.715 INFO:teuthology.task.pexec:sudo dnf install nvmetcli nvme-cli -y 2026-03-09T15:10:25.751 DEBUG:teuthology.task.pexec:ubuntu@vm06.local< sudo dnf remove nvme-cli -y 2026-03-09T15:10:25.751 DEBUG:teuthology.task.pexec:ubuntu@vm06.local< sudo dnf install nvmetcli nvme-cli -y 2026-03-09T15:10:25.751 INFO:teuthology.task.pexec:Running commands on host ubuntu@vm06.local 2026-03-09T15:10:25.751 INFO:teuthology.task.pexec:sudo dnf remove nvme-cli -y 2026-03-09T15:10:25.751 INFO:teuthology.task.pexec:sudo dnf install nvmetcli nvme-cli -y 2026-03-09T15:10:25.910 INFO:teuthology.orchestra.run.vm03.stdout:No match for argument: nvme-cli 2026-03-09T15:10:25.910 INFO:teuthology.orchestra.run.vm03.stderr:No packages marked for removal. 2026-03-09T15:10:25.913 INFO:teuthology.orchestra.run.vm03.stdout:Dependencies resolved. 2026-03-09T15:10:25.914 INFO:teuthology.orchestra.run.vm03.stdout:Nothing to do. 2026-03-09T15:10:25.914 INFO:teuthology.orchestra.run.vm03.stdout:Complete! 2026-03-09T15:10:25.925 INFO:teuthology.orchestra.run.vm04.stdout:No match for argument: nvme-cli 2026-03-09T15:10:25.925 INFO:teuthology.orchestra.run.vm04.stderr:No packages marked for removal. 2026-03-09T15:10:25.929 INFO:teuthology.orchestra.run.vm04.stdout:Dependencies resolved. 2026-03-09T15:10:25.929 INFO:teuthology.orchestra.run.vm04.stdout:Nothing to do. 2026-03-09T15:10:25.929 INFO:teuthology.orchestra.run.vm04.stdout:Complete! 2026-03-09T15:10:26.008 INFO:teuthology.orchestra.run.vm06.stdout:No match for argument: nvme-cli 2026-03-09T15:10:26.008 INFO:teuthology.orchestra.run.vm06.stderr:No packages marked for removal. 2026-03-09T15:10:26.011 INFO:teuthology.orchestra.run.vm06.stdout:Dependencies resolved. 2026-03-09T15:10:26.012 INFO:teuthology.orchestra.run.vm06.stdout:Nothing to do. 2026-03-09T15:10:26.012 INFO:teuthology.orchestra.run.vm06.stdout:Complete! 2026-03-09T15:10:26.309 INFO:teuthology.orchestra.run.vm03.stdout:Last metadata expiration check: 0:01:36 ago on Mon 09 Mar 2026 03:08:50 PM UTC. 2026-03-09T15:10:26.344 INFO:teuthology.orchestra.run.vm04.stdout:Last metadata expiration check: 0:01:24 ago on Mon 09 Mar 2026 03:09:02 PM UTC. 2026-03-09T15:10:26.404 INFO:teuthology.orchestra.run.vm03.stdout:Dependencies resolved. 2026-03-09T15:10:26.404 INFO:teuthology.orchestra.run.vm03.stdout:================================================================================ 2026-03-09T15:10:26.404 INFO:teuthology.orchestra.run.vm03.stdout: Package Architecture Version Repository Size 2026-03-09T15:10:26.404 INFO:teuthology.orchestra.run.vm03.stdout:================================================================================ 2026-03-09T15:10:26.404 INFO:teuthology.orchestra.run.vm03.stdout:Installing: 2026-03-09T15:10:26.404 INFO:teuthology.orchestra.run.vm03.stdout: nvme-cli x86_64 2.16-1.el9 baseos 1.2 M 2026-03-09T15:10:26.404 INFO:teuthology.orchestra.run.vm03.stdout: nvmetcli noarch 0.8-3.el9 baseos 44 k 2026-03-09T15:10:26.404 INFO:teuthology.orchestra.run.vm03.stdout:Installing dependencies: 2026-03-09T15:10:26.404 INFO:teuthology.orchestra.run.vm03.stdout: python3-configshell noarch 1:1.1.30-1.el9 baseos 72 k 2026-03-09T15:10:26.404 INFO:teuthology.orchestra.run.vm03.stdout: python3-kmod x86_64 0.9-32.el9 baseos 84 k 2026-03-09T15:10:26.404 INFO:teuthology.orchestra.run.vm03.stdout: python3-pyparsing noarch 2.4.7-9.el9 baseos 150 k 2026-03-09T15:10:26.404 INFO:teuthology.orchestra.run.vm03.stdout: python3-urwid x86_64 2.1.2-4.el9 baseos 837 k 2026-03-09T15:10:26.404 INFO:teuthology.orchestra.run.vm03.stdout: 2026-03-09T15:10:26.404 INFO:teuthology.orchestra.run.vm03.stdout:Transaction Summary 2026-03-09T15:10:26.404 INFO:teuthology.orchestra.run.vm03.stdout:================================================================================ 2026-03-09T15:10:26.404 INFO:teuthology.orchestra.run.vm03.stdout:Install 6 Packages 2026-03-09T15:10:26.404 INFO:teuthology.orchestra.run.vm03.stdout: 2026-03-09T15:10:26.404 INFO:teuthology.orchestra.run.vm03.stdout:Total download size: 2.3 M 2026-03-09T15:10:26.405 INFO:teuthology.orchestra.run.vm03.stdout:Installed size: 11 M 2026-03-09T15:10:26.405 INFO:teuthology.orchestra.run.vm03.stdout:Downloading Packages: 2026-03-09T15:10:26.438 INFO:teuthology.orchestra.run.vm04.stdout:Dependencies resolved. 2026-03-09T15:10:26.438 INFO:teuthology.orchestra.run.vm04.stdout:================================================================================ 2026-03-09T15:10:26.438 INFO:teuthology.orchestra.run.vm04.stdout: Package Architecture Version Repository Size 2026-03-09T15:10:26.438 INFO:teuthology.orchestra.run.vm04.stdout:================================================================================ 2026-03-09T15:10:26.438 INFO:teuthology.orchestra.run.vm04.stdout:Installing: 2026-03-09T15:10:26.438 INFO:teuthology.orchestra.run.vm04.stdout: nvme-cli x86_64 2.16-1.el9 baseos 1.2 M 2026-03-09T15:10:26.438 INFO:teuthology.orchestra.run.vm04.stdout: nvmetcli noarch 0.8-3.el9 baseos 44 k 2026-03-09T15:10:26.438 INFO:teuthology.orchestra.run.vm04.stdout:Installing dependencies: 2026-03-09T15:10:26.438 INFO:teuthology.orchestra.run.vm04.stdout: python3-configshell noarch 1:1.1.30-1.el9 baseos 72 k 2026-03-09T15:10:26.438 INFO:teuthology.orchestra.run.vm04.stdout: python3-kmod x86_64 0.9-32.el9 baseos 84 k 2026-03-09T15:10:26.438 INFO:teuthology.orchestra.run.vm04.stdout: python3-pyparsing noarch 2.4.7-9.el9 baseos 150 k 2026-03-09T15:10:26.438 INFO:teuthology.orchestra.run.vm04.stdout: python3-urwid x86_64 2.1.2-4.el9 baseos 837 k 2026-03-09T15:10:26.438 INFO:teuthology.orchestra.run.vm04.stdout: 2026-03-09T15:10:26.438 INFO:teuthology.orchestra.run.vm04.stdout:Transaction Summary 2026-03-09T15:10:26.438 INFO:teuthology.orchestra.run.vm04.stdout:================================================================================ 2026-03-09T15:10:26.438 INFO:teuthology.orchestra.run.vm04.stdout:Install 6 Packages 2026-03-09T15:10:26.438 INFO:teuthology.orchestra.run.vm04.stdout: 2026-03-09T15:10:26.439 INFO:teuthology.orchestra.run.vm04.stdout:Total download size: 2.3 M 2026-03-09T15:10:26.439 INFO:teuthology.orchestra.run.vm04.stdout:Installed size: 11 M 2026-03-09T15:10:26.439 INFO:teuthology.orchestra.run.vm04.stdout:Downloading Packages: 2026-03-09T15:10:26.564 INFO:teuthology.orchestra.run.vm06.stdout:Last metadata expiration check: 0:01:29 ago on Mon 09 Mar 2026 03:08:57 PM UTC. 2026-03-09T15:10:26.689 INFO:teuthology.orchestra.run.vm06.stdout:Dependencies resolved. 2026-03-09T15:10:26.689 INFO:teuthology.orchestra.run.vm06.stdout:================================================================================ 2026-03-09T15:10:26.689 INFO:teuthology.orchestra.run.vm06.stdout: Package Architecture Version Repository Size 2026-03-09T15:10:26.689 INFO:teuthology.orchestra.run.vm06.stdout:================================================================================ 2026-03-09T15:10:26.689 INFO:teuthology.orchestra.run.vm06.stdout:Installing: 2026-03-09T15:10:26.690 INFO:teuthology.orchestra.run.vm06.stdout: nvme-cli x86_64 2.16-1.el9 baseos 1.2 M 2026-03-09T15:10:26.690 INFO:teuthology.orchestra.run.vm06.stdout: nvmetcli noarch 0.8-3.el9 baseos 44 k 2026-03-09T15:10:26.690 INFO:teuthology.orchestra.run.vm06.stdout:Installing dependencies: 2026-03-09T15:10:26.690 INFO:teuthology.orchestra.run.vm06.stdout: python3-configshell noarch 1:1.1.30-1.el9 baseos 72 k 2026-03-09T15:10:26.690 INFO:teuthology.orchestra.run.vm06.stdout: python3-kmod x86_64 0.9-32.el9 baseos 84 k 2026-03-09T15:10:26.690 INFO:teuthology.orchestra.run.vm06.stdout: python3-pyparsing noarch 2.4.7-9.el9 baseos 150 k 2026-03-09T15:10:26.690 INFO:teuthology.orchestra.run.vm06.stdout: python3-urwid x86_64 2.1.2-4.el9 baseos 837 k 2026-03-09T15:10:26.690 INFO:teuthology.orchestra.run.vm06.stdout: 2026-03-09T15:10:26.690 INFO:teuthology.orchestra.run.vm06.stdout:Transaction Summary 2026-03-09T15:10:26.690 INFO:teuthology.orchestra.run.vm06.stdout:================================================================================ 2026-03-09T15:10:26.690 INFO:teuthology.orchestra.run.vm06.stdout:Install 6 Packages 2026-03-09T15:10:26.690 INFO:teuthology.orchestra.run.vm06.stdout: 2026-03-09T15:10:26.690 INFO:teuthology.orchestra.run.vm06.stdout:Total download size: 2.3 M 2026-03-09T15:10:26.690 INFO:teuthology.orchestra.run.vm06.stdout:Installed size: 11 M 2026-03-09T15:10:26.690 INFO:teuthology.orchestra.run.vm06.stdout:Downloading Packages: 2026-03-09T15:10:26.728 INFO:teuthology.orchestra.run.vm03.stdout:(1/6): nvmetcli-0.8-3.el9.noarch.rpm 233 kB/s | 44 kB 00:00 2026-03-09T15:10:26.749 INFO:teuthology.orchestra.run.vm03.stdout:(2/6): python3-configshell-1.1.30-1.el9.noarch. 345 kB/s | 72 kB 00:00 2026-03-09T15:10:26.830 INFO:teuthology.orchestra.run.vm04.stdout:(1/6): python3-configshell-1.1.30-1.el9.noarch. 280 kB/s | 72 kB 00:00 2026-03-09T15:10:26.834 INFO:teuthology.orchestra.run.vm04.stdout:(2/6): nvmetcli-0.8-3.el9.noarch.rpm 168 kB/s | 44 kB 00:00 2026-03-09T15:10:26.840 INFO:teuthology.orchestra.run.vm03.stdout:(3/6): python3-kmod-0.9-32.el9.x86_64.rpm 754 kB/s | 84 kB 00:00 2026-03-09T15:10:26.894 INFO:teuthology.orchestra.run.vm03.stdout:(4/6): python3-pyparsing-2.4.7-9.el9.noarch.rpm 1.0 MB/s | 150 kB 00:00 2026-03-09T15:10:26.900 INFO:teuthology.orchestra.run.vm04.stdout:(3/6): python3-kmod-0.9-32.el9.x86_64.rpm 1.2 MB/s | 84 kB 00:00 2026-03-09T15:10:26.970 INFO:teuthology.orchestra.run.vm04.stdout:(4/6): python3-pyparsing-2.4.7-9.el9.noarch.rpm 1.1 MB/s | 150 kB 00:00 2026-03-09T15:10:27.088 INFO:teuthology.orchestra.run.vm04.stdout:(5/6): nvme-cli-2.16-1.el9.x86_64.rpm 2.2 MB/s | 1.2 MB 00:00 2026-03-09T15:10:27.099 INFO:teuthology.orchestra.run.vm04.stdout:(6/6): python3-urwid-2.1.2-4.el9.x86_64.rpm 4.1 MB/s | 837 kB 00:00 2026-03-09T15:10:27.099 INFO:teuthology.orchestra.run.vm04.stdout:-------------------------------------------------------------------------------- 2026-03-09T15:10:27.099 INFO:teuthology.orchestra.run.vm04.stdout:Total 3.5 MB/s | 2.3 MB 00:00 2026-03-09T15:10:27.160 INFO:teuthology.orchestra.run.vm04.stdout:Running transaction check 2026-03-09T15:10:27.168 INFO:teuthology.orchestra.run.vm04.stdout:Transaction check succeeded. 2026-03-09T15:10:27.168 INFO:teuthology.orchestra.run.vm04.stdout:Running transaction test 2026-03-09T15:10:27.186 INFO:teuthology.orchestra.run.vm03.stdout:(5/6): nvme-cli-2.16-1.el9.x86_64.rpm 1.8 MB/s | 1.2 MB 00:00 2026-03-09T15:10:27.218 INFO:teuthology.orchestra.run.vm04.stdout:Transaction test succeeded. 2026-03-09T15:10:27.219 INFO:teuthology.orchestra.run.vm04.stdout:Running transaction 2026-03-09T15:10:27.265 INFO:teuthology.orchestra.run.vm03.stdout:(6/6): python3-urwid-2.1.2-4.el9.x86_64.rpm 1.9 MB/s | 837 kB 00:00 2026-03-09T15:10:27.265 INFO:teuthology.orchestra.run.vm03.stdout:-------------------------------------------------------------------------------- 2026-03-09T15:10:27.265 INFO:teuthology.orchestra.run.vm03.stdout:Total 2.7 MB/s | 2.3 MB 00:00 2026-03-09T15:10:27.328 INFO:teuthology.orchestra.run.vm03.stdout:Running transaction check 2026-03-09T15:10:27.335 INFO:teuthology.orchestra.run.vm03.stdout:Transaction check succeeded. 2026-03-09T15:10:27.335 INFO:teuthology.orchestra.run.vm03.stdout:Running transaction test 2026-03-09T15:10:27.373 INFO:teuthology.orchestra.run.vm04.stdout: Preparing : 1/1 2026-03-09T15:10:27.384 INFO:teuthology.orchestra.run.vm03.stdout:Transaction test succeeded. 2026-03-09T15:10:27.384 INFO:teuthology.orchestra.run.vm03.stdout:Running transaction 2026-03-09T15:10:27.385 INFO:teuthology.orchestra.run.vm04.stdout: Installing : python3-urwid-2.1.2-4.el9.x86_64 1/6 2026-03-09T15:10:27.396 INFO:teuthology.orchestra.run.vm04.stdout: Installing : python3-pyparsing-2.4.7-9.el9.noarch 2/6 2026-03-09T15:10:27.403 INFO:teuthology.orchestra.run.vm04.stdout: Installing : python3-configshell-1:1.1.30-1.el9.noarch 3/6 2026-03-09T15:10:27.411 INFO:teuthology.orchestra.run.vm04.stdout: Installing : python3-kmod-0.9-32.el9.x86_64 4/6 2026-03-09T15:10:27.412 INFO:teuthology.orchestra.run.vm04.stdout: Installing : nvmetcli-0.8-3.el9.noarch 5/6 2026-03-09T15:10:27.495 INFO:teuthology.orchestra.run.vm06.stdout:(1/6): nvmetcli-0.8-3.el9.noarch.rpm 125 kB/s | 44 kB 00:00 2026-03-09T15:10:27.519 INFO:teuthology.orchestra.run.vm06.stdout:(2/6): python3-configshell-1.1.30-1.el9.noarch. 192 kB/s | 72 kB 00:00 2026-03-09T15:10:27.528 INFO:teuthology.orchestra.run.vm03.stdout: Preparing : 1/1 2026-03-09T15:10:27.539 INFO:teuthology.orchestra.run.vm03.stdout: Installing : python3-urwid-2.1.2-4.el9.x86_64 1/6 2026-03-09T15:10:27.553 INFO:teuthology.orchestra.run.vm03.stdout: Installing : python3-pyparsing-2.4.7-9.el9.noarch 2/6 2026-03-09T15:10:27.560 INFO:teuthology.orchestra.run.vm03.stdout: Installing : python3-configshell-1:1.1.30-1.el9.noarch 3/6 2026-03-09T15:10:27.567 INFO:teuthology.orchestra.run.vm03.stdout: Installing : python3-kmod-0.9-32.el9.x86_64 4/6 2026-03-09T15:10:27.568 INFO:teuthology.orchestra.run.vm03.stdout: Installing : nvmetcli-0.8-3.el9.noarch 5/6 2026-03-09T15:10:27.575 INFO:teuthology.orchestra.run.vm04.stdout: Running scriptlet: nvmetcli-0.8-3.el9.noarch 5/6 2026-03-09T15:10:27.580 INFO:teuthology.orchestra.run.vm04.stdout: Installing : nvme-cli-2.16-1.el9.x86_64 6/6 2026-03-09T15:10:27.649 INFO:teuthology.orchestra.run.vm06.stdout:(3/6): python3-kmod-0.9-32.el9.x86_64.rpm 550 kB/s | 84 kB 00:00 2026-03-09T15:10:27.670 INFO:teuthology.orchestra.run.vm06.stdout:(4/6): python3-pyparsing-2.4.7-9.el9.noarch.rpm 998 kB/s | 150 kB 00:00 2026-03-09T15:10:27.713 INFO:teuthology.orchestra.run.vm03.stdout: Running scriptlet: nvmetcli-0.8-3.el9.noarch 5/6 2026-03-09T15:10:27.717 INFO:teuthology.orchestra.run.vm03.stdout: Installing : nvme-cli-2.16-1.el9.x86_64 6/6 2026-03-09T15:10:27.777 INFO:teuthology.orchestra.run.vm06.stdout:(5/6): nvme-cli-2.16-1.el9.x86_64.rpm 1.8 MB/s | 1.2 MB 00:00 2026-03-09T15:10:27.916 INFO:teuthology.orchestra.run.vm04.stdout: Running scriptlet: nvme-cli-2.16-1.el9.x86_64 6/6 2026-03-09T15:10:27.916 INFO:teuthology.orchestra.run.vm04.stdout:Created symlink /etc/systemd/system/default.target.wants/nvmefc-boot-connections.service → /usr/lib/systemd/system/nvmefc-boot-connections.service. 2026-03-09T15:10:27.916 INFO:teuthology.orchestra.run.vm04.stdout: 2026-03-09T15:10:27.934 INFO:teuthology.orchestra.run.vm06.stdout:(6/6): python3-urwid-2.1.2-4.el9.x86_64.rpm 2.9 MB/s | 837 kB 00:00 2026-03-09T15:10:27.936 INFO:teuthology.orchestra.run.vm06.stdout:-------------------------------------------------------------------------------- 2026-03-09T15:10:27.936 INFO:teuthology.orchestra.run.vm06.stdout:Total 1.9 MB/s | 2.3 MB 00:01 2026-03-09T15:10:28.021 INFO:teuthology.orchestra.run.vm06.stdout:Running transaction check 2026-03-09T15:10:28.032 INFO:teuthology.orchestra.run.vm06.stdout:Transaction check succeeded. 2026-03-09T15:10:28.032 INFO:teuthology.orchestra.run.vm06.stdout:Running transaction test 2026-03-09T15:10:28.071 INFO:teuthology.orchestra.run.vm03.stdout: Running scriptlet: nvme-cli-2.16-1.el9.x86_64 6/6 2026-03-09T15:10:28.071 INFO:teuthology.orchestra.run.vm03.stdout:Created symlink /etc/systemd/system/default.target.wants/nvmefc-boot-connections.service → /usr/lib/systemd/system/nvmefc-boot-connections.service. 2026-03-09T15:10:28.071 INFO:teuthology.orchestra.run.vm03.stdout: 2026-03-09T15:10:28.101 INFO:teuthology.orchestra.run.vm06.stdout:Transaction test succeeded. 2026-03-09T15:10:28.101 INFO:teuthology.orchestra.run.vm06.stdout:Running transaction 2026-03-09T15:10:28.297 INFO:teuthology.orchestra.run.vm06.stdout: Preparing : 1/1 2026-03-09T15:10:28.314 INFO:teuthology.orchestra.run.vm06.stdout: Installing : python3-urwid-2.1.2-4.el9.x86_64 1/6 2026-03-09T15:10:28.329 INFO:teuthology.orchestra.run.vm06.stdout: Installing : python3-pyparsing-2.4.7-9.el9.noarch 2/6 2026-03-09T15:10:28.340 INFO:teuthology.orchestra.run.vm06.stdout: Installing : python3-configshell-1:1.1.30-1.el9.noarch 3/6 2026-03-09T15:10:28.348 INFO:teuthology.orchestra.run.vm06.stdout: Installing : python3-kmod-0.9-32.el9.x86_64 4/6 2026-03-09T15:10:28.350 INFO:teuthology.orchestra.run.vm06.stdout: Installing : nvmetcli-0.8-3.el9.noarch 5/6 2026-03-09T15:10:28.499 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : nvme-cli-2.16-1.el9.x86_64 1/6 2026-03-09T15:10:28.499 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : nvmetcli-0.8-3.el9.noarch 2/6 2026-03-09T15:10:28.499 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-configshell-1:1.1.30-1.el9.noarch 3/6 2026-03-09T15:10:28.499 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-kmod-0.9-32.el9.x86_64 4/6 2026-03-09T15:10:28.499 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-pyparsing-2.4.7-9.el9.noarch 5/6 2026-03-09T15:10:28.553 INFO:teuthology.orchestra.run.vm06.stdout: Running scriptlet: nvmetcli-0.8-3.el9.noarch 5/6 2026-03-09T15:10:28.559 INFO:teuthology.orchestra.run.vm06.stdout: Installing : nvme-cli-2.16-1.el9.x86_64 6/6 2026-03-09T15:10:28.618 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-urwid-2.1.2-4.el9.x86_64 6/6 2026-03-09T15:10:28.619 INFO:teuthology.orchestra.run.vm04.stdout: 2026-03-09T15:10:28.619 INFO:teuthology.orchestra.run.vm04.stdout:Installed: 2026-03-09T15:10:28.619 INFO:teuthology.orchestra.run.vm04.stdout: nvme-cli-2.16-1.el9.x86_64 nvmetcli-0.8-3.el9.noarch 2026-03-09T15:10:28.619 INFO:teuthology.orchestra.run.vm04.stdout: python3-configshell-1:1.1.30-1.el9.noarch python3-kmod-0.9-32.el9.x86_64 2026-03-09T15:10:28.619 INFO:teuthology.orchestra.run.vm04.stdout: python3-pyparsing-2.4.7-9.el9.noarch python3-urwid-2.1.2-4.el9.x86_64 2026-03-09T15:10:28.619 INFO:teuthology.orchestra.run.vm04.stdout: 2026-03-09T15:10:28.619 INFO:teuthology.orchestra.run.vm04.stdout:Complete! 2026-03-09T15:10:28.656 INFO:teuthology.orchestra.run.vm03.stdout: Verifying : nvme-cli-2.16-1.el9.x86_64 1/6 2026-03-09T15:10:28.656 INFO:teuthology.orchestra.run.vm03.stdout: Verifying : nvmetcli-0.8-3.el9.noarch 2/6 2026-03-09T15:10:28.656 INFO:teuthology.orchestra.run.vm03.stdout: Verifying : python3-configshell-1:1.1.30-1.el9.noarch 3/6 2026-03-09T15:10:28.657 INFO:teuthology.orchestra.run.vm03.stdout: Verifying : python3-kmod-0.9-32.el9.x86_64 4/6 2026-03-09T15:10:28.657 INFO:teuthology.orchestra.run.vm03.stdout: Verifying : python3-pyparsing-2.4.7-9.el9.noarch 5/6 2026-03-09T15:10:28.703 DEBUG:teuthology.parallel:result is None 2026-03-09T15:10:28.749 INFO:teuthology.orchestra.run.vm03.stdout: Verifying : python3-urwid-2.1.2-4.el9.x86_64 6/6 2026-03-09T15:10:28.749 INFO:teuthology.orchestra.run.vm03.stdout: 2026-03-09T15:10:28.749 INFO:teuthology.orchestra.run.vm03.stdout:Installed: 2026-03-09T15:10:28.749 INFO:teuthology.orchestra.run.vm03.stdout: nvme-cli-2.16-1.el9.x86_64 nvmetcli-0.8-3.el9.noarch 2026-03-09T15:10:28.749 INFO:teuthology.orchestra.run.vm03.stdout: python3-configshell-1:1.1.30-1.el9.noarch python3-kmod-0.9-32.el9.x86_64 2026-03-09T15:10:28.749 INFO:teuthology.orchestra.run.vm03.stdout: python3-pyparsing-2.4.7-9.el9.noarch python3-urwid-2.1.2-4.el9.x86_64 2026-03-09T15:10:28.749 INFO:teuthology.orchestra.run.vm03.stdout: 2026-03-09T15:10:28.749 INFO:teuthology.orchestra.run.vm03.stdout:Complete! 2026-03-09T15:10:28.808 DEBUG:teuthology.parallel:result is None 2026-03-09T15:10:28.984 INFO:teuthology.orchestra.run.vm06.stdout: Running scriptlet: nvme-cli-2.16-1.el9.x86_64 6/6 2026-03-09T15:10:28.984 INFO:teuthology.orchestra.run.vm06.stdout:Created symlink /etc/systemd/system/default.target.wants/nvmefc-boot-connections.service → /usr/lib/systemd/system/nvmefc-boot-connections.service. 2026-03-09T15:10:28.984 INFO:teuthology.orchestra.run.vm06.stdout: 2026-03-09T15:10:29.594 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : nvme-cli-2.16-1.el9.x86_64 1/6 2026-03-09T15:10:29.594 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : nvmetcli-0.8-3.el9.noarch 2/6 2026-03-09T15:10:29.594 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-configshell-1:1.1.30-1.el9.noarch 3/6 2026-03-09T15:10:29.594 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-kmod-0.9-32.el9.x86_64 4/6 2026-03-09T15:10:29.594 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-pyparsing-2.4.7-9.el9.noarch 5/6 2026-03-09T15:10:29.691 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-urwid-2.1.2-4.el9.x86_64 6/6 2026-03-09T15:10:29.691 INFO:teuthology.orchestra.run.vm06.stdout: 2026-03-09T15:10:29.691 INFO:teuthology.orchestra.run.vm06.stdout:Installed: 2026-03-09T15:10:29.691 INFO:teuthology.orchestra.run.vm06.stdout: nvme-cli-2.16-1.el9.x86_64 nvmetcli-0.8-3.el9.noarch 2026-03-09T15:10:29.691 INFO:teuthology.orchestra.run.vm06.stdout: python3-configshell-1:1.1.30-1.el9.noarch python3-kmod-0.9-32.el9.x86_64 2026-03-09T15:10:29.691 INFO:teuthology.orchestra.run.vm06.stdout: python3-pyparsing-2.4.7-9.el9.noarch python3-urwid-2.1.2-4.el9.x86_64 2026-03-09T15:10:29.691 INFO:teuthology.orchestra.run.vm06.stdout: 2026-03-09T15:10:29.691 INFO:teuthology.orchestra.run.vm06.stdout:Complete! 2026-03-09T15:10:29.744 DEBUG:teuthology.parallel:result is None 2026-03-09T15:10:29.744 INFO:teuthology.run_tasks:Running task install... 2026-03-09T15:10:29.747 DEBUG:teuthology.task.install:project ceph 2026-03-09T15:10:29.747 DEBUG:teuthology.task.install:INSTALL overrides: {'ceph': {'flavor': 'default', 'sha1': 'e911bdebe5c8faa3800735d1568fcdca65db60df'}, 'extra_system_packages': {'deb': ['python3-xmltodict', 'python3-jmespath'], 'rpm': ['bzip2', 'perl-Test-Harness', 'python3-xmltodict', 'python3-jmespath']}} 2026-03-09T15:10:29.747 DEBUG:teuthology.task.install:config {'flavor': 'default', 'sha1': 'e911bdebe5c8faa3800735d1568fcdca65db60df', 'extra_system_packages': {'deb': ['python3-xmltodict', 'python3-jmespath'], 'rpm': ['bzip2', 'perl-Test-Harness', 'python3-xmltodict', 'python3-jmespath']}} 2026-03-09T15:10:29.747 INFO:teuthology.task.install:Using flavor: default 2026-03-09T15:10:29.749 DEBUG:teuthology.task.install:Package list is: {'deb': ['ceph', 'cephadm', 'ceph-mds', 'ceph-mgr', 'ceph-common', 'ceph-fuse', 'ceph-test', 'ceph-volume', 'radosgw', 'python3-rados', 'python3-rgw', 'python3-cephfs', 'python3-rbd', 'libcephfs2', 'libcephfs-dev', 'librados2', 'librbd1', 'rbd-fuse'], 'rpm': ['ceph-radosgw', 'ceph-test', 'ceph', 'ceph-base', 'cephadm', 'ceph-immutable-object-cache', 'ceph-mgr', 'ceph-mgr-dashboard', 'ceph-mgr-diskprediction-local', 'ceph-mgr-rook', 'ceph-mgr-cephadm', 'ceph-fuse', 'ceph-volume', 'librados-devel', 'libcephfs2', 'libcephfs-devel', 'librados2', 'librbd1', 'python3-rados', 'python3-rgw', 'python3-cephfs', 'python3-rbd', 'rbd-fuse', 'rbd-mirror', 'rbd-nbd']} 2026-03-09T15:10:29.749 INFO:teuthology.task.install:extra packages: [] 2026-03-09T15:10:29.749 DEBUG:teuthology.task.install.rpm:_update_package_list_and_install: config is {'branch': None, 'cleanup': None, 'debuginfo': None, 'downgrade_packages': [], 'exclude_packages': [], 'extra_packages': [], 'extra_system_packages': {'deb': ['python3-xmltodict', 'python3-jmespath'], 'rpm': ['bzip2', 'perl-Test-Harness', 'python3-xmltodict', 'python3-jmespath']}, 'extras': None, 'enable_coprs': [], 'flavor': 'default', 'install_ceph_packages': True, 'packages': {}, 'project': 'ceph', 'repos_only': False, 'sha1': 'e911bdebe5c8faa3800735d1568fcdca65db60df', 'tag': None, 'wait_for_package': False} 2026-03-09T15:10:29.750 DEBUG:teuthology.packaging:Querying https://shaman.ceph.com/api/search?status=ready&project=ceph&flavor=default&distros=centos%2F9%2Fx86_64&sha1=e911bdebe5c8faa3800735d1568fcdca65db60df 2026-03-09T15:10:29.750 DEBUG:teuthology.task.install.rpm:_update_package_list_and_install: config is {'branch': None, 'cleanup': None, 'debuginfo': None, 'downgrade_packages': [], 'exclude_packages': [], 'extra_packages': [], 'extra_system_packages': {'deb': ['python3-xmltodict', 'python3-jmespath'], 'rpm': ['bzip2', 'perl-Test-Harness', 'python3-xmltodict', 'python3-jmespath']}, 'extras': None, 'enable_coprs': [], 'flavor': 'default', 'install_ceph_packages': True, 'packages': {}, 'project': 'ceph', 'repos_only': False, 'sha1': 'e911bdebe5c8faa3800735d1568fcdca65db60df', 'tag': None, 'wait_for_package': False} 2026-03-09T15:10:29.750 DEBUG:teuthology.packaging:Querying https://shaman.ceph.com/api/search?status=ready&project=ceph&flavor=default&distros=centos%2F9%2Fx86_64&sha1=e911bdebe5c8faa3800735d1568fcdca65db60df 2026-03-09T15:10:29.751 DEBUG:teuthology.task.install.rpm:_update_package_list_and_install: config is {'branch': None, 'cleanup': None, 'debuginfo': None, 'downgrade_packages': [], 'exclude_packages': [], 'extra_packages': [], 'extra_system_packages': {'deb': ['python3-xmltodict', 'python3-jmespath'], 'rpm': ['bzip2', 'perl-Test-Harness', 'python3-xmltodict', 'python3-jmespath']}, 'extras': None, 'enable_coprs': [], 'flavor': 'default', 'install_ceph_packages': True, 'packages': {}, 'project': 'ceph', 'repos_only': False, 'sha1': 'e911bdebe5c8faa3800735d1568fcdca65db60df', 'tag': None, 'wait_for_package': False} 2026-03-09T15:10:29.751 DEBUG:teuthology.packaging:Querying https://shaman.ceph.com/api/search?status=ready&project=ceph&flavor=default&distros=centos%2F9%2Fx86_64&sha1=e911bdebe5c8faa3800735d1568fcdca65db60df 2026-03-09T15:10:30.355 INFO:teuthology.task.install.rpm:Pulling from https://3.chacra.ceph.com/r/ceph/squid/e911bdebe5c8faa3800735d1568fcdca65db60df/centos/9/flavors/default/ 2026-03-09T15:10:30.355 INFO:teuthology.task.install.rpm:Package version is 19.2.3-678.ge911bdeb 2026-03-09T15:10:30.386 INFO:teuthology.task.install.rpm:Pulling from https://3.chacra.ceph.com/r/ceph/squid/e911bdebe5c8faa3800735d1568fcdca65db60df/centos/9/flavors/default/ 2026-03-09T15:10:30.386 INFO:teuthology.task.install.rpm:Package version is 19.2.3-678.ge911bdeb 2026-03-09T15:10:30.422 INFO:teuthology.task.install.rpm:Pulling from https://3.chacra.ceph.com/r/ceph/squid/e911bdebe5c8faa3800735d1568fcdca65db60df/centos/9/flavors/default/ 2026-03-09T15:10:30.422 INFO:teuthology.task.install.rpm:Package version is 19.2.3-678.ge911bdeb 2026-03-09T15:10:30.893 INFO:teuthology.packaging:Writing yum repo: [ceph] name=ceph packages for $basearch baseurl=https://3.chacra.ceph.com/r/ceph/squid/e911bdebe5c8faa3800735d1568fcdca65db60df/centos/9/flavors/default/$basearch enabled=1 gpgcheck=0 type=rpm-md [ceph-noarch] name=ceph noarch packages baseurl=https://3.chacra.ceph.com/r/ceph/squid/e911bdebe5c8faa3800735d1568fcdca65db60df/centos/9/flavors/default/noarch enabled=1 gpgcheck=0 type=rpm-md [ceph-source] name=ceph source packages baseurl=https://3.chacra.ceph.com/r/ceph/squid/e911bdebe5c8faa3800735d1568fcdca65db60df/centos/9/flavors/default/SRPMS enabled=1 gpgcheck=0 type=rpm-md 2026-03-09T15:10:30.894 DEBUG:teuthology.orchestra.run.vm04:> set -ex 2026-03-09T15:10:30.894 DEBUG:teuthology.orchestra.run.vm04:> sudo dd of=/etc/yum.repos.d/ceph.repo 2026-03-09T15:10:30.902 INFO:teuthology.packaging:Writing yum repo: [ceph] name=ceph packages for $basearch baseurl=https://3.chacra.ceph.com/r/ceph/squid/e911bdebe5c8faa3800735d1568fcdca65db60df/centos/9/flavors/default/$basearch enabled=1 gpgcheck=0 type=rpm-md [ceph-noarch] name=ceph noarch packages baseurl=https://3.chacra.ceph.com/r/ceph/squid/e911bdebe5c8faa3800735d1568fcdca65db60df/centos/9/flavors/default/noarch enabled=1 gpgcheck=0 type=rpm-md [ceph-source] name=ceph source packages baseurl=https://3.chacra.ceph.com/r/ceph/squid/e911bdebe5c8faa3800735d1568fcdca65db60df/centos/9/flavors/default/SRPMS enabled=1 gpgcheck=0 type=rpm-md 2026-03-09T15:10:30.902 DEBUG:teuthology.orchestra.run.vm06:> set -ex 2026-03-09T15:10:30.902 DEBUG:teuthology.orchestra.run.vm06:> sudo dd of=/etc/yum.repos.d/ceph.repo 2026-03-09T15:10:30.905 INFO:teuthology.packaging:Writing yum repo: [ceph] name=ceph packages for $basearch baseurl=https://3.chacra.ceph.com/r/ceph/squid/e911bdebe5c8faa3800735d1568fcdca65db60df/centos/9/flavors/default/$basearch enabled=1 gpgcheck=0 type=rpm-md [ceph-noarch] name=ceph noarch packages baseurl=https://3.chacra.ceph.com/r/ceph/squid/e911bdebe5c8faa3800735d1568fcdca65db60df/centos/9/flavors/default/noarch enabled=1 gpgcheck=0 type=rpm-md [ceph-source] name=ceph source packages baseurl=https://3.chacra.ceph.com/r/ceph/squid/e911bdebe5c8faa3800735d1568fcdca65db60df/centos/9/flavors/default/SRPMS enabled=1 gpgcheck=0 type=rpm-md 2026-03-09T15:10:30.905 DEBUG:teuthology.orchestra.run.vm03:> set -ex 2026-03-09T15:10:30.905 DEBUG:teuthology.orchestra.run.vm03:> sudo dd of=/etc/yum.repos.d/ceph.repo 2026-03-09T15:10:30.928 INFO:teuthology.task.install.rpm:Installing packages: ceph-radosgw, ceph-test, ceph, ceph-base, cephadm, ceph-immutable-object-cache, ceph-mgr, ceph-mgr-dashboard, ceph-mgr-diskprediction-local, ceph-mgr-rook, ceph-mgr-cephadm, ceph-fuse, ceph-volume, librados-devel, libcephfs2, libcephfs-devel, librados2, librbd1, python3-rados, python3-rgw, python3-cephfs, python3-rbd, rbd-fuse, rbd-mirror, rbd-nbd, bzip2, perl-Test-Harness, python3-xmltodict, python3-jmespath on remote rpm x86_64 2026-03-09T15:10:30.928 DEBUG:teuthology.orchestra.run.vm04:> if test -f /etc/yum.repos.d/ceph.repo ; then sudo sed -i -e ':a;N;$!ba;s/enabled=1\ngpg/enabled=1\npriority=1\ngpg/g' -e 's;ref/[a-zA-Z0-9_-]*/;sha1/e911bdebe5c8faa3800735d1568fcdca65db60df/;g' /etc/yum.repos.d/ceph.repo ; fi 2026-03-09T15:10:30.935 INFO:teuthology.task.install.rpm:Installing packages: ceph-radosgw, ceph-test, ceph, ceph-base, cephadm, ceph-immutable-object-cache, ceph-mgr, ceph-mgr-dashboard, ceph-mgr-diskprediction-local, ceph-mgr-rook, ceph-mgr-cephadm, ceph-fuse, ceph-volume, librados-devel, libcephfs2, libcephfs-devel, librados2, librbd1, python3-rados, python3-rgw, python3-cephfs, python3-rbd, rbd-fuse, rbd-mirror, rbd-nbd, bzip2, perl-Test-Harness, python3-xmltodict, python3-jmespath on remote rpm x86_64 2026-03-09T15:10:30.935 DEBUG:teuthology.orchestra.run.vm03:> if test -f /etc/yum.repos.d/ceph.repo ; then sudo sed -i -e ':a;N;$!ba;s/enabled=1\ngpg/enabled=1\npriority=1\ngpg/g' -e 's;ref/[a-zA-Z0-9_-]*/;sha1/e911bdebe5c8faa3800735d1568fcdca65db60df/;g' /etc/yum.repos.d/ceph.repo ; fi 2026-03-09T15:10:30.942 INFO:teuthology.task.install.rpm:Installing packages: ceph-radosgw, ceph-test, ceph, ceph-base, cephadm, ceph-immutable-object-cache, ceph-mgr, ceph-mgr-dashboard, ceph-mgr-diskprediction-local, ceph-mgr-rook, ceph-mgr-cephadm, ceph-fuse, ceph-volume, librados-devel, libcephfs2, libcephfs-devel, librados2, librbd1, python3-rados, python3-rgw, python3-cephfs, python3-rbd, rbd-fuse, rbd-mirror, rbd-nbd, bzip2, perl-Test-Harness, python3-xmltodict, python3-jmespath on remote rpm x86_64 2026-03-09T15:10:30.942 DEBUG:teuthology.orchestra.run.vm06:> if test -f /etc/yum.repos.d/ceph.repo ; then sudo sed -i -e ':a;N;$!ba;s/enabled=1\ngpg/enabled=1\npriority=1\ngpg/g' -e 's;ref/[a-zA-Z0-9_-]*/;sha1/e911bdebe5c8faa3800735d1568fcdca65db60df/;g' /etc/yum.repos.d/ceph.repo ; fi 2026-03-09T15:10:31.001 DEBUG:teuthology.orchestra.run.vm04:> sudo touch -a /etc/yum/pluginconf.d/priorities.conf ; test -e /etc/yum/pluginconf.d/priorities.conf.orig || sudo cp -af /etc/yum/pluginconf.d/priorities.conf /etc/yum/pluginconf.d/priorities.conf.orig 2026-03-09T15:10:31.003 DEBUG:teuthology.orchestra.run.vm03:> sudo touch -a /etc/yum/pluginconf.d/priorities.conf ; test -e /etc/yum/pluginconf.d/priorities.conf.orig || sudo cp -af /etc/yum/pluginconf.d/priorities.conf /etc/yum/pluginconf.d/priorities.conf.orig 2026-03-09T15:10:31.016 DEBUG:teuthology.orchestra.run.vm06:> sudo touch -a /etc/yum/pluginconf.d/priorities.conf ; test -e /etc/yum/pluginconf.d/priorities.conf.orig || sudo cp -af /etc/yum/pluginconf.d/priorities.conf /etc/yum/pluginconf.d/priorities.conf.orig 2026-03-09T15:10:31.081 DEBUG:teuthology.orchestra.run.vm04:> grep check_obsoletes /etc/yum/pluginconf.d/priorities.conf && sudo sed -i 's/check_obsoletes.*0/check_obsoletes = 1/g' /etc/yum/pluginconf.d/priorities.conf || echo 'check_obsoletes = 1' | sudo tee -a /etc/yum/pluginconf.d/priorities.conf 2026-03-09T15:10:31.084 DEBUG:teuthology.orchestra.run.vm03:> grep check_obsoletes /etc/yum/pluginconf.d/priorities.conf && sudo sed -i 's/check_obsoletes.*0/check_obsoletes = 1/g' /etc/yum/pluginconf.d/priorities.conf || echo 'check_obsoletes = 1' | sudo tee -a /etc/yum/pluginconf.d/priorities.conf 2026-03-09T15:10:31.101 DEBUG:teuthology.orchestra.run.vm06:> grep check_obsoletes /etc/yum/pluginconf.d/priorities.conf && sudo sed -i 's/check_obsoletes.*0/check_obsoletes = 1/g' /etc/yum/pluginconf.d/priorities.conf || echo 'check_obsoletes = 1' | sudo tee -a /etc/yum/pluginconf.d/priorities.conf 2026-03-09T15:10:31.148 INFO:teuthology.orchestra.run.vm04.stdout:check_obsoletes = 1 2026-03-09T15:10:31.150 DEBUG:teuthology.orchestra.run.vm04:> sudo yum clean all 2026-03-09T15:10:31.152 INFO:teuthology.orchestra.run.vm03.stdout:check_obsoletes = 1 2026-03-09T15:10:31.154 DEBUG:teuthology.orchestra.run.vm03:> sudo yum clean all 2026-03-09T15:10:31.171 INFO:teuthology.orchestra.run.vm06.stdout:check_obsoletes = 1 2026-03-09T15:10:31.174 DEBUG:teuthology.orchestra.run.vm06:> sudo yum clean all 2026-03-09T15:10:31.325 INFO:teuthology.orchestra.run.vm03.stdout:41 files removed 2026-03-09T15:10:31.337 INFO:teuthology.orchestra.run.vm04.stdout:41 files removed 2026-03-09T15:10:31.354 DEBUG:teuthology.orchestra.run.vm03:> sudo yum -y install ceph-radosgw ceph-test ceph ceph-base cephadm ceph-immutable-object-cache ceph-mgr ceph-mgr-dashboard ceph-mgr-diskprediction-local ceph-mgr-rook ceph-mgr-cephadm ceph-fuse ceph-volume librados-devel libcephfs2 libcephfs-devel librados2 librbd1 python3-rados python3-rgw python3-cephfs python3-rbd rbd-fuse rbd-mirror rbd-nbd bzip2 perl-Test-Harness python3-xmltodict python3-jmespath 2026-03-09T15:10:31.371 DEBUG:teuthology.orchestra.run.vm04:> sudo yum -y install ceph-radosgw ceph-test ceph ceph-base cephadm ceph-immutable-object-cache ceph-mgr ceph-mgr-dashboard ceph-mgr-diskprediction-local ceph-mgr-rook ceph-mgr-cephadm ceph-fuse ceph-volume librados-devel libcephfs2 libcephfs-devel librados2 librbd1 python3-rados python3-rgw python3-cephfs python3-rbd rbd-fuse rbd-mirror rbd-nbd bzip2 perl-Test-Harness python3-xmltodict python3-jmespath 2026-03-09T15:10:31.384 INFO:teuthology.orchestra.run.vm06.stdout:41 files removed 2026-03-09T15:10:31.410 DEBUG:teuthology.orchestra.run.vm06:> sudo yum -y install ceph-radosgw ceph-test ceph ceph-base cephadm ceph-immutable-object-cache ceph-mgr ceph-mgr-dashboard ceph-mgr-diskprediction-local ceph-mgr-rook ceph-mgr-cephadm ceph-fuse ceph-volume librados-devel libcephfs2 libcephfs-devel librados2 librbd1 python3-rados python3-rgw python3-cephfs python3-rbd rbd-fuse rbd-mirror rbd-nbd bzip2 perl-Test-Harness python3-xmltodict python3-jmespath 2026-03-09T15:10:32.699 INFO:teuthology.orchestra.run.vm03.stdout:ceph packages for x86_64 70 kB/s | 84 kB 00:01 2026-03-09T15:10:32.717 INFO:teuthology.orchestra.run.vm04.stdout:ceph packages for x86_64 72 kB/s | 84 kB 00:01 2026-03-09T15:10:32.767 INFO:teuthology.orchestra.run.vm06.stdout:ceph packages for x86_64 73 kB/s | 84 kB 00:01 2026-03-09T15:10:33.652 INFO:teuthology.orchestra.run.vm03.stdout:ceph noarch packages 13 kB/s | 12 kB 00:00 2026-03-09T15:10:33.660 INFO:teuthology.orchestra.run.vm04.stdout:ceph noarch packages 13 kB/s | 12 kB 00:00 2026-03-09T15:10:33.757 INFO:teuthology.orchestra.run.vm06.stdout:ceph noarch packages 12 kB/s | 12 kB 00:00 2026-03-09T15:10:34.596 INFO:teuthology.orchestra.run.vm03.stdout:ceph source packages 2.0 kB/s | 1.9 kB 00:00 2026-03-09T15:10:34.633 INFO:teuthology.orchestra.run.vm04.stdout:ceph source packages 2.0 kB/s | 1.9 kB 00:00 2026-03-09T15:10:34.706 INFO:teuthology.orchestra.run.vm06.stdout:ceph source packages 2.0 kB/s | 1.9 kB 00:00 2026-03-09T15:10:35.376 INFO:teuthology.orchestra.run.vm04.stdout:CentOS Stream 9 - BaseOS 12 MB/s | 8.9 MB 00:00 2026-03-09T15:10:35.772 INFO:teuthology.orchestra.run.vm06.stdout:CentOS Stream 9 - BaseOS 8.5 MB/s | 8.9 MB 00:01 2026-03-09T15:10:36.862 INFO:teuthology.orchestra.run.vm06.stdout:CentOS Stream 9 - AppStream 56 MB/s | 27 MB 00:00 2026-03-09T15:10:40.191 INFO:teuthology.orchestra.run.vm04.stdout:CentOS Stream 9 - AppStream 6.5 MB/s | 27 MB 00:04 2026-03-09T15:10:40.284 INFO:teuthology.orchestra.run.vm03.stdout:CentOS Stream 9 - BaseOS 1.6 MB/s | 8.9 MB 00:05 2026-03-09T15:10:40.641 INFO:teuthology.orchestra.run.vm06.stdout:CentOS Stream 9 - CRB 8.1 MB/s | 8.0 MB 00:00 2026-03-09T15:10:42.545 INFO:teuthology.orchestra.run.vm06.stdout:CentOS Stream 9 - Extras packages 20 kB/s | 20 kB 00:01 2026-03-09T15:10:42.575 INFO:teuthology.orchestra.run.vm03.stdout:CentOS Stream 9 - AppStream 17 MB/s | 27 MB 00:01 2026-03-09T15:10:43.411 INFO:teuthology.orchestra.run.vm06.stdout:Extra Packages for Enterprise Linux 26 MB/s | 20 MB 00:00 2026-03-09T15:10:43.889 INFO:teuthology.orchestra.run.vm04.stdout:CentOS Stream 9 - CRB 8.3 MB/s | 8.0 MB 00:00 2026-03-09T15:10:45.400 INFO:teuthology.orchestra.run.vm04.stdout:CentOS Stream 9 - Extras packages 30 kB/s | 20 kB 00:00 2026-03-09T15:10:45.865 INFO:teuthology.orchestra.run.vm04.stdout:Extra Packages for Enterprise Linux 54 MB/s | 20 MB 00:00 2026-03-09T15:10:48.082 INFO:teuthology.orchestra.run.vm06.stdout:lab-extras 64 kB/s | 50 kB 00:00 2026-03-09T15:10:48.788 INFO:teuthology.orchestra.run.vm03.stdout:CentOS Stream 9 - CRB 2.3 MB/s | 8.0 MB 00:03 2026-03-09T15:10:49.435 INFO:teuthology.orchestra.run.vm06.stdout:Package librados2-2:16.2.4-5.el9.x86_64 is already installed. 2026-03-09T15:10:49.435 INFO:teuthology.orchestra.run.vm06.stdout:Package librbd1-2:16.2.4-5.el9.x86_64 is already installed. 2026-03-09T15:10:49.439 INFO:teuthology.orchestra.run.vm06.stdout:Package bzip2-1.0.8-11.el9.x86_64 is already installed. 2026-03-09T15:10:49.439 INFO:teuthology.orchestra.run.vm06.stdout:Package perl-Test-Harness-1:3.42-461.el9.noarch is already installed. 2026-03-09T15:10:49.467 INFO:teuthology.orchestra.run.vm06.stdout:Dependencies resolved. 2026-03-09T15:10:49.471 INFO:teuthology.orchestra.run.vm06.stdout:====================================================================================== 2026-03-09T15:10:49.471 INFO:teuthology.orchestra.run.vm06.stdout: Package Arch Version Repository Size 2026-03-09T15:10:49.471 INFO:teuthology.orchestra.run.vm06.stdout:====================================================================================== 2026-03-09T15:10:49.471 INFO:teuthology.orchestra.run.vm06.stdout:Installing: 2026-03-09T15:10:49.471 INFO:teuthology.orchestra.run.vm06.stdout: ceph x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 6.5 k 2026-03-09T15:10:49.471 INFO:teuthology.orchestra.run.vm06.stdout: ceph-base x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 5.5 M 2026-03-09T15:10:49.471 INFO:teuthology.orchestra.run.vm06.stdout: ceph-fuse x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 1.2 M 2026-03-09T15:10:49.471 INFO:teuthology.orchestra.run.vm06.stdout: ceph-immutable-object-cache x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 145 k 2026-03-09T15:10:49.471 INFO:teuthology.orchestra.run.vm06.stdout: ceph-mgr x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 1.1 M 2026-03-09T15:10:49.471 INFO:teuthology.orchestra.run.vm06.stdout: ceph-mgr-cephadm noarch 2:19.2.3-678.ge911bdeb.el9 ceph-noarch 150 k 2026-03-09T15:10:49.471 INFO:teuthology.orchestra.run.vm06.stdout: ceph-mgr-dashboard noarch 2:19.2.3-678.ge911bdeb.el9 ceph-noarch 3.8 M 2026-03-09T15:10:49.471 INFO:teuthology.orchestra.run.vm06.stdout: ceph-mgr-diskprediction-local noarch 2:19.2.3-678.ge911bdeb.el9 ceph-noarch 7.4 M 2026-03-09T15:10:49.471 INFO:teuthology.orchestra.run.vm06.stdout: ceph-mgr-rook noarch 2:19.2.3-678.ge911bdeb.el9 ceph-noarch 49 k 2026-03-09T15:10:49.471 INFO:teuthology.orchestra.run.vm06.stdout: ceph-radosgw x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 11 M 2026-03-09T15:10:49.471 INFO:teuthology.orchestra.run.vm06.stdout: ceph-test x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 50 M 2026-03-09T15:10:49.471 INFO:teuthology.orchestra.run.vm06.stdout: ceph-volume noarch 2:19.2.3-678.ge911bdeb.el9 ceph-noarch 299 k 2026-03-09T15:10:49.471 INFO:teuthology.orchestra.run.vm06.stdout: cephadm noarch 2:19.2.3-678.ge911bdeb.el9 ceph-noarch 769 k 2026-03-09T15:10:49.471 INFO:teuthology.orchestra.run.vm06.stdout: libcephfs-devel x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 34 k 2026-03-09T15:10:49.471 INFO:teuthology.orchestra.run.vm06.stdout: libcephfs2 x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 1.0 M 2026-03-09T15:10:49.471 INFO:teuthology.orchestra.run.vm06.stdout: librados-devel x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 127 k 2026-03-09T15:10:49.471 INFO:teuthology.orchestra.run.vm06.stdout: python3-cephfs x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 165 k 2026-03-09T15:10:49.471 INFO:teuthology.orchestra.run.vm06.stdout: python3-jmespath noarch 1.0.1-1.el9 appstream 48 k 2026-03-09T15:10:49.471 INFO:teuthology.orchestra.run.vm06.stdout: python3-rados x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 323 k 2026-03-09T15:10:49.471 INFO:teuthology.orchestra.run.vm06.stdout: python3-rbd x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 303 k 2026-03-09T15:10:49.471 INFO:teuthology.orchestra.run.vm06.stdout: python3-rgw x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 100 k 2026-03-09T15:10:49.471 INFO:teuthology.orchestra.run.vm06.stdout: python3-xmltodict noarch 0.12.0-15.el9 epel 22 k 2026-03-09T15:10:49.471 INFO:teuthology.orchestra.run.vm06.stdout: rbd-fuse x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 85 k 2026-03-09T15:10:49.471 INFO:teuthology.orchestra.run.vm06.stdout: rbd-mirror x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 3.1 M 2026-03-09T15:10:49.471 INFO:teuthology.orchestra.run.vm06.stdout: rbd-nbd x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 171 k 2026-03-09T15:10:49.471 INFO:teuthology.orchestra.run.vm06.stdout:Upgrading: 2026-03-09T15:10:49.471 INFO:teuthology.orchestra.run.vm06.stdout: librados2 x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 3.4 M 2026-03-09T15:10:49.471 INFO:teuthology.orchestra.run.vm06.stdout: librbd1 x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 3.2 M 2026-03-09T15:10:49.471 INFO:teuthology.orchestra.run.vm06.stdout:Installing dependencies: 2026-03-09T15:10:49.471 INFO:teuthology.orchestra.run.vm06.stdout: abseil-cpp x86_64 20211102.0-4.el9 epel 551 k 2026-03-09T15:10:49.471 INFO:teuthology.orchestra.run.vm06.stdout: boost-program-options x86_64 1.75.0-13.el9 appstream 104 k 2026-03-09T15:10:49.471 INFO:teuthology.orchestra.run.vm06.stdout: ceph-common x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 22 M 2026-03-09T15:10:49.472 INFO:teuthology.orchestra.run.vm06.stdout: ceph-grafana-dashboards noarch 2:19.2.3-678.ge911bdeb.el9 ceph-noarch 31 k 2026-03-09T15:10:49.472 INFO:teuthology.orchestra.run.vm06.stdout: ceph-mds x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 2.4 M 2026-03-09T15:10:49.472 INFO:teuthology.orchestra.run.vm06.stdout: ceph-mgr-modules-core noarch 2:19.2.3-678.ge911bdeb.el9 ceph-noarch 253 k 2026-03-09T15:10:49.472 INFO:teuthology.orchestra.run.vm06.stdout: ceph-mon x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 4.7 M 2026-03-09T15:10:49.472 INFO:teuthology.orchestra.run.vm06.stdout: ceph-osd x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 17 M 2026-03-09T15:10:49.472 INFO:teuthology.orchestra.run.vm06.stdout: ceph-prometheus-alerts noarch 2:19.2.3-678.ge911bdeb.el9 ceph-noarch 17 k 2026-03-09T15:10:49.472 INFO:teuthology.orchestra.run.vm06.stdout: ceph-selinux x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 25 k 2026-03-09T15:10:49.472 INFO:teuthology.orchestra.run.vm06.stdout: cryptsetup x86_64 2.8.1-3.el9 baseos 351 k 2026-03-09T15:10:49.472 INFO:teuthology.orchestra.run.vm06.stdout: flexiblas x86_64 3.0.4-9.el9 appstream 30 k 2026-03-09T15:10:49.472 INFO:teuthology.orchestra.run.vm06.stdout: flexiblas-netlib x86_64 3.0.4-9.el9 appstream 3.0 M 2026-03-09T15:10:49.472 INFO:teuthology.orchestra.run.vm06.stdout: flexiblas-openblas-openmp x86_64 3.0.4-9.el9 appstream 15 k 2026-03-09T15:10:49.472 INFO:teuthology.orchestra.run.vm06.stdout: gperftools-libs x86_64 2.9.1-3.el9 epel 308 k 2026-03-09T15:10:49.472 INFO:teuthology.orchestra.run.vm06.stdout: grpc-data noarch 1.46.7-10.el9 epel 19 k 2026-03-09T15:10:49.472 INFO:teuthology.orchestra.run.vm06.stdout: ledmon-libs x86_64 1.1.0-3.el9 baseos 40 k 2026-03-09T15:10:49.472 INFO:teuthology.orchestra.run.vm06.stdout: libarrow x86_64 9.0.0-15.el9 epel 4.4 M 2026-03-09T15:10:49.472 INFO:teuthology.orchestra.run.vm06.stdout: libarrow-doc noarch 9.0.0-15.el9 epel 25 k 2026-03-09T15:10:49.472 INFO:teuthology.orchestra.run.vm06.stdout: libcephsqlite x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 163 k 2026-03-09T15:10:49.472 INFO:teuthology.orchestra.run.vm06.stdout: libconfig x86_64 1.7.2-9.el9 baseos 72 k 2026-03-09T15:10:49.472 INFO:teuthology.orchestra.run.vm06.stdout: libgfortran x86_64 11.5.0-14.el9 baseos 794 k 2026-03-09T15:10:49.472 INFO:teuthology.orchestra.run.vm06.stdout: libnbd x86_64 1.20.3-4.el9 appstream 164 k 2026-03-09T15:10:49.472 INFO:teuthology.orchestra.run.vm06.stdout: liboath x86_64 2.6.12-1.el9 epel 49 k 2026-03-09T15:10:49.472 INFO:teuthology.orchestra.run.vm06.stdout: libpmemobj x86_64 1.12.1-1.el9 appstream 160 k 2026-03-09T15:10:49.472 INFO:teuthology.orchestra.run.vm06.stdout: libquadmath x86_64 11.5.0-14.el9 baseos 184 k 2026-03-09T15:10:49.472 INFO:teuthology.orchestra.run.vm06.stdout: librabbitmq x86_64 0.11.0-7.el9 appstream 45 k 2026-03-09T15:10:49.472 INFO:teuthology.orchestra.run.vm06.stdout: libradosstriper1 x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 503 k 2026-03-09T15:10:49.472 INFO:teuthology.orchestra.run.vm06.stdout: librdkafka x86_64 1.6.1-102.el9 appstream 662 k 2026-03-09T15:10:49.472 INFO:teuthology.orchestra.run.vm06.stdout: librgw2 x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 5.4 M 2026-03-09T15:10:49.472 INFO:teuthology.orchestra.run.vm06.stdout: libstoragemgmt x86_64 1.10.1-1.el9 appstream 246 k 2026-03-09T15:10:49.472 INFO:teuthology.orchestra.run.vm06.stdout: libunwind x86_64 1.6.2-1.el9 epel 67 k 2026-03-09T15:10:49.472 INFO:teuthology.orchestra.run.vm06.stdout: libxslt x86_64 1.1.34-12.el9 appstream 233 k 2026-03-09T15:10:49.472 INFO:teuthology.orchestra.run.vm06.stdout: lttng-ust x86_64 2.12.0-6.el9 appstream 292 k 2026-03-09T15:10:49.472 INFO:teuthology.orchestra.run.vm06.stdout: lua x86_64 5.4.4-4.el9 appstream 188 k 2026-03-09T15:10:49.472 INFO:teuthology.orchestra.run.vm06.stdout: lua-devel x86_64 5.4.4-4.el9 crb 22 k 2026-03-09T15:10:49.472 INFO:teuthology.orchestra.run.vm06.stdout: luarocks noarch 3.9.2-5.el9 epel 151 k 2026-03-09T15:10:49.472 INFO:teuthology.orchestra.run.vm06.stdout: mailcap noarch 2.1.49-5.el9 baseos 33 k 2026-03-09T15:10:49.472 INFO:teuthology.orchestra.run.vm06.stdout: openblas x86_64 0.3.29-1.el9 appstream 42 k 2026-03-09T15:10:49.472 INFO:teuthology.orchestra.run.vm06.stdout: openblas-openmp x86_64 0.3.29-1.el9 appstream 5.3 M 2026-03-09T15:10:49.472 INFO:teuthology.orchestra.run.vm06.stdout: parquet-libs x86_64 9.0.0-15.el9 epel 838 k 2026-03-09T15:10:49.472 INFO:teuthology.orchestra.run.vm06.stdout: pciutils x86_64 3.7.0-7.el9 baseos 93 k 2026-03-09T15:10:49.472 INFO:teuthology.orchestra.run.vm06.stdout: protobuf x86_64 3.14.0-17.el9 appstream 1.0 M 2026-03-09T15:10:49.472 INFO:teuthology.orchestra.run.vm06.stdout: protobuf-compiler x86_64 3.14.0-17.el9 crb 862 k 2026-03-09T15:10:49.472 INFO:teuthology.orchestra.run.vm06.stdout: python3-asyncssh noarch 2.13.2-5.el9 epel 548 k 2026-03-09T15:10:49.472 INFO:teuthology.orchestra.run.vm06.stdout: python3-autocommand noarch 2.2.2-8.el9 epel 29 k 2026-03-09T15:10:49.472 INFO:teuthology.orchestra.run.vm06.stdout: python3-babel noarch 2.9.1-2.el9 appstream 6.0 M 2026-03-09T15:10:49.472 INFO:teuthology.orchestra.run.vm06.stdout: python3-backports-tarfile noarch 1.2.0-1.el9 epel 60 k 2026-03-09T15:10:49.472 INFO:teuthology.orchestra.run.vm06.stdout: python3-bcrypt x86_64 3.2.2-1.el9 epel 43 k 2026-03-09T15:10:49.472 INFO:teuthology.orchestra.run.vm06.stdout: python3-cachetools noarch 4.2.4-1.el9 epel 32 k 2026-03-09T15:10:49.472 INFO:teuthology.orchestra.run.vm06.stdout: python3-ceph-argparse x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 45 k 2026-03-09T15:10:49.472 INFO:teuthology.orchestra.run.vm06.stdout: python3-ceph-common x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 142 k 2026-03-09T15:10:49.472 INFO:teuthology.orchestra.run.vm06.stdout: python3-certifi noarch 2023.05.07-4.el9 epel 14 k 2026-03-09T15:10:49.472 INFO:teuthology.orchestra.run.vm06.stdout: python3-cffi x86_64 1.14.5-5.el9 baseos 253 k 2026-03-09T15:10:49.472 INFO:teuthology.orchestra.run.vm06.stdout: python3-cheroot noarch 10.0.1-4.el9 epel 173 k 2026-03-09T15:10:49.472 INFO:teuthology.orchestra.run.vm06.stdout: python3-cherrypy noarch 18.6.1-2.el9 epel 358 k 2026-03-09T15:10:49.472 INFO:teuthology.orchestra.run.vm06.stdout: python3-cryptography x86_64 36.0.1-5.el9 baseos 1.2 M 2026-03-09T15:10:49.472 INFO:teuthology.orchestra.run.vm06.stdout: python3-devel x86_64 3.9.25-3.el9 appstream 244 k 2026-03-09T15:10:49.472 INFO:teuthology.orchestra.run.vm06.stdout: python3-google-auth noarch 1:2.45.0-1.el9 epel 254 k 2026-03-09T15:10:49.472 INFO:teuthology.orchestra.run.vm06.stdout: python3-grpcio x86_64 1.46.7-10.el9 epel 2.0 M 2026-03-09T15:10:49.472 INFO:teuthology.orchestra.run.vm06.stdout: python3-grpcio-tools x86_64 1.46.7-10.el9 epel 144 k 2026-03-09T15:10:49.472 INFO:teuthology.orchestra.run.vm06.stdout: python3-jaraco noarch 8.2.1-3.el9 epel 11 k 2026-03-09T15:10:49.472 INFO:teuthology.orchestra.run.vm06.stdout: python3-jaraco-classes noarch 3.2.1-5.el9 epel 18 k 2026-03-09T15:10:49.472 INFO:teuthology.orchestra.run.vm06.stdout: python3-jaraco-collections noarch 3.0.0-8.el9 epel 23 k 2026-03-09T15:10:49.473 INFO:teuthology.orchestra.run.vm06.stdout: python3-jaraco-context noarch 6.0.1-3.el9 epel 20 k 2026-03-09T15:10:49.473 INFO:teuthology.orchestra.run.vm06.stdout: python3-jaraco-functools noarch 3.5.0-2.el9 epel 19 k 2026-03-09T15:10:49.473 INFO:teuthology.orchestra.run.vm06.stdout: python3-jaraco-text noarch 4.0.0-2.el9 epel 26 k 2026-03-09T15:10:49.473 INFO:teuthology.orchestra.run.vm06.stdout: python3-jinja2 noarch 2.11.3-8.el9 appstream 249 k 2026-03-09T15:10:49.473 INFO:teuthology.orchestra.run.vm06.stdout: python3-kubernetes noarch 1:26.1.0-3.el9 epel 1.0 M 2026-03-09T15:10:49.473 INFO:teuthology.orchestra.run.vm06.stdout: python3-libstoragemgmt x86_64 1.10.1-1.el9 appstream 177 k 2026-03-09T15:10:49.473 INFO:teuthology.orchestra.run.vm06.stdout: python3-logutils noarch 0.3.5-21.el9 epel 46 k 2026-03-09T15:10:49.473 INFO:teuthology.orchestra.run.vm06.stdout: python3-mako noarch 1.1.4-6.el9 appstream 172 k 2026-03-09T15:10:49.473 INFO:teuthology.orchestra.run.vm06.stdout: python3-markupsafe x86_64 1.1.1-12.el9 appstream 35 k 2026-03-09T15:10:49.473 INFO:teuthology.orchestra.run.vm06.stdout: python3-more-itertools noarch 8.12.0-2.el9 epel 79 k 2026-03-09T15:10:49.473 INFO:teuthology.orchestra.run.vm06.stdout: python3-natsort noarch 7.1.1-5.el9 epel 58 k 2026-03-09T15:10:49.473 INFO:teuthology.orchestra.run.vm06.stdout: python3-numpy x86_64 1:1.23.5-2.el9 appstream 6.1 M 2026-03-09T15:10:49.473 INFO:teuthology.orchestra.run.vm06.stdout: python3-numpy-f2py x86_64 1:1.23.5-2.el9 appstream 442 k 2026-03-09T15:10:49.473 INFO:teuthology.orchestra.run.vm06.stdout: python3-packaging noarch 20.9-5.el9 appstream 77 k 2026-03-09T15:10:49.473 INFO:teuthology.orchestra.run.vm06.stdout: python3-pecan noarch 1.4.2-3.el9 epel 272 k 2026-03-09T15:10:49.473 INFO:teuthology.orchestra.run.vm06.stdout: python3-ply noarch 3.11-14.el9 baseos 106 k 2026-03-09T15:10:49.473 INFO:teuthology.orchestra.run.vm06.stdout: python3-portend noarch 3.1.0-2.el9 epel 16 k 2026-03-09T15:10:49.473 INFO:teuthology.orchestra.run.vm06.stdout: python3-protobuf noarch 3.14.0-17.el9 appstream 267 k 2026-03-09T15:10:49.473 INFO:teuthology.orchestra.run.vm06.stdout: python3-pyOpenSSL noarch 21.0.0-1.el9 epel 90 k 2026-03-09T15:10:49.473 INFO:teuthology.orchestra.run.vm06.stdout: python3-pyasn1 noarch 0.4.8-7.el9 appstream 157 k 2026-03-09T15:10:49.473 INFO:teuthology.orchestra.run.vm06.stdout: python3-pyasn1-modules noarch 0.4.8-7.el9 appstream 277 k 2026-03-09T15:10:49.473 INFO:teuthology.orchestra.run.vm06.stdout: python3-pycparser noarch 2.20-6.el9 baseos 135 k 2026-03-09T15:10:49.473 INFO:teuthology.orchestra.run.vm06.stdout: python3-repoze-lru noarch 0.7-16.el9 epel 31 k 2026-03-09T15:10:49.473 INFO:teuthology.orchestra.run.vm06.stdout: python3-requests noarch 2.25.1-10.el9 baseos 126 k 2026-03-09T15:10:49.473 INFO:teuthology.orchestra.run.vm06.stdout: python3-requests-oauthlib noarch 1.3.0-12.el9 appstream 54 k 2026-03-09T15:10:49.473 INFO:teuthology.orchestra.run.vm06.stdout: python3-routes noarch 2.5.1-5.el9 epel 188 k 2026-03-09T15:10:49.473 INFO:teuthology.orchestra.run.vm06.stdout: python3-rsa noarch 4.9-2.el9 epel 59 k 2026-03-09T15:10:49.473 INFO:teuthology.orchestra.run.vm06.stdout: python3-scipy x86_64 1.9.3-2.el9 appstream 19 M 2026-03-09T15:10:49.473 INFO:teuthology.orchestra.run.vm06.stdout: python3-tempora noarch 5.0.0-2.el9 epel 36 k 2026-03-09T15:10:49.473 INFO:teuthology.orchestra.run.vm06.stdout: python3-toml noarch 0.10.2-6.el9 appstream 42 k 2026-03-09T15:10:49.473 INFO:teuthology.orchestra.run.vm06.stdout: python3-typing-extensions noarch 4.15.0-1.el9 epel 86 k 2026-03-09T15:10:49.473 INFO:teuthology.orchestra.run.vm06.stdout: python3-urllib3 noarch 1.26.5-7.el9 baseos 218 k 2026-03-09T15:10:49.473 INFO:teuthology.orchestra.run.vm06.stdout: python3-webob noarch 1.8.8-2.el9 epel 230 k 2026-03-09T15:10:49.473 INFO:teuthology.orchestra.run.vm06.stdout: python3-websocket-client noarch 1.2.3-2.el9 epel 90 k 2026-03-09T15:10:49.473 INFO:teuthology.orchestra.run.vm06.stdout: python3-werkzeug noarch 2.0.3-3.el9.1 epel 427 k 2026-03-09T15:10:49.473 INFO:teuthology.orchestra.run.vm06.stdout: python3-zc-lockfile noarch 2.0-10.el9 epel 20 k 2026-03-09T15:10:49.473 INFO:teuthology.orchestra.run.vm06.stdout: qatlib x86_64 25.08.0-2.el9 appstream 240 k 2026-03-09T15:10:49.473 INFO:teuthology.orchestra.run.vm06.stdout: qatzip-libs x86_64 1.3.1-1.el9 appstream 66 k 2026-03-09T15:10:49.473 INFO:teuthology.orchestra.run.vm06.stdout: re2 x86_64 1:20211101-20.el9 epel 191 k 2026-03-09T15:10:49.473 INFO:teuthology.orchestra.run.vm06.stdout: socat x86_64 1.7.4.1-8.el9 appstream 303 k 2026-03-09T15:10:49.473 INFO:teuthology.orchestra.run.vm06.stdout: thrift x86_64 0.15.0-4.el9 epel 1.6 M 2026-03-09T15:10:49.473 INFO:teuthology.orchestra.run.vm06.stdout: unzip x86_64 6.0-59.el9 baseos 182 k 2026-03-09T15:10:49.473 INFO:teuthology.orchestra.run.vm06.stdout: xmlstarlet x86_64 1.6.1-20.el9 appstream 64 k 2026-03-09T15:10:49.473 INFO:teuthology.orchestra.run.vm06.stdout: zip x86_64 3.0-35.el9 baseos 266 k 2026-03-09T15:10:49.473 INFO:teuthology.orchestra.run.vm06.stdout:Installing weak dependencies: 2026-03-09T15:10:49.473 INFO:teuthology.orchestra.run.vm06.stdout: qatlib-service x86_64 25.08.0-2.el9 appstream 37 k 2026-03-09T15:10:49.473 INFO:teuthology.orchestra.run.vm06.stdout: 2026-03-09T15:10:49.473 INFO:teuthology.orchestra.run.vm06.stdout:Transaction Summary 2026-03-09T15:10:49.473 INFO:teuthology.orchestra.run.vm06.stdout:====================================================================================== 2026-03-09T15:10:49.473 INFO:teuthology.orchestra.run.vm06.stdout:Install 134 Packages 2026-03-09T15:10:49.473 INFO:teuthology.orchestra.run.vm06.stdout:Upgrade 2 Packages 2026-03-09T15:10:49.473 INFO:teuthology.orchestra.run.vm06.stdout: 2026-03-09T15:10:49.473 INFO:teuthology.orchestra.run.vm06.stdout:Total download size: 210 M 2026-03-09T15:10:49.474 INFO:teuthology.orchestra.run.vm06.stdout:Downloading Packages: 2026-03-09T15:10:50.386 INFO:teuthology.orchestra.run.vm04.stdout:lab-extras 65 kB/s | 50 kB 00:00 2026-03-09T15:10:50.982 INFO:teuthology.orchestra.run.vm03.stdout:CentOS Stream 9 - Extras packages 16 kB/s | 20 kB 00:01 2026-03-09T15:10:51.203 INFO:teuthology.orchestra.run.vm06.stdout:(1/136): ceph-19.2.3-678.ge911bdeb.el9.x86_64.r 14 kB/s | 6.5 kB 00:00 2026-03-09T15:10:51.712 INFO:teuthology.orchestra.run.vm04.stdout:Package librados2-2:16.2.4-5.el9.x86_64 is already installed. 2026-03-09T15:10:51.712 INFO:teuthology.orchestra.run.vm04.stdout:Package librbd1-2:16.2.4-5.el9.x86_64 is already installed. 2026-03-09T15:10:51.716 INFO:teuthology.orchestra.run.vm04.stdout:Package bzip2-1.0.8-11.el9.x86_64 is already installed. 2026-03-09T15:10:51.717 INFO:teuthology.orchestra.run.vm04.stdout:Package perl-Test-Harness-1:3.42-461.el9.noarch is already installed. 2026-03-09T15:10:51.745 INFO:teuthology.orchestra.run.vm04.stdout:Dependencies resolved. 2026-03-09T15:10:51.749 INFO:teuthology.orchestra.run.vm04.stdout:====================================================================================== 2026-03-09T15:10:51.749 INFO:teuthology.orchestra.run.vm04.stdout: Package Arch Version Repository Size 2026-03-09T15:10:51.749 INFO:teuthology.orchestra.run.vm04.stdout:====================================================================================== 2026-03-09T15:10:51.749 INFO:teuthology.orchestra.run.vm04.stdout:Installing: 2026-03-09T15:10:51.749 INFO:teuthology.orchestra.run.vm04.stdout: ceph x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 6.5 k 2026-03-09T15:10:51.749 INFO:teuthology.orchestra.run.vm04.stdout: ceph-base x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 5.5 M 2026-03-09T15:10:51.749 INFO:teuthology.orchestra.run.vm04.stdout: ceph-fuse x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 1.2 M 2026-03-09T15:10:51.749 INFO:teuthology.orchestra.run.vm04.stdout: ceph-immutable-object-cache x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 145 k 2026-03-09T15:10:51.749 INFO:teuthology.orchestra.run.vm04.stdout: ceph-mgr x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 1.1 M 2026-03-09T15:10:51.749 INFO:teuthology.orchestra.run.vm04.stdout: ceph-mgr-cephadm noarch 2:19.2.3-678.ge911bdeb.el9 ceph-noarch 150 k 2026-03-09T15:10:51.749 INFO:teuthology.orchestra.run.vm04.stdout: ceph-mgr-dashboard noarch 2:19.2.3-678.ge911bdeb.el9 ceph-noarch 3.8 M 2026-03-09T15:10:51.749 INFO:teuthology.orchestra.run.vm04.stdout: ceph-mgr-diskprediction-local noarch 2:19.2.3-678.ge911bdeb.el9 ceph-noarch 7.4 M 2026-03-09T15:10:51.749 INFO:teuthology.orchestra.run.vm04.stdout: ceph-mgr-rook noarch 2:19.2.3-678.ge911bdeb.el9 ceph-noarch 49 k 2026-03-09T15:10:51.749 INFO:teuthology.orchestra.run.vm04.stdout: ceph-radosgw x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 11 M 2026-03-09T15:10:51.749 INFO:teuthology.orchestra.run.vm04.stdout: ceph-test x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 50 M 2026-03-09T15:10:51.750 INFO:teuthology.orchestra.run.vm04.stdout: ceph-volume noarch 2:19.2.3-678.ge911bdeb.el9 ceph-noarch 299 k 2026-03-09T15:10:51.750 INFO:teuthology.orchestra.run.vm04.stdout: cephadm noarch 2:19.2.3-678.ge911bdeb.el9 ceph-noarch 769 k 2026-03-09T15:10:51.750 INFO:teuthology.orchestra.run.vm04.stdout: libcephfs-devel x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 34 k 2026-03-09T15:10:51.750 INFO:teuthology.orchestra.run.vm04.stdout: libcephfs2 x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 1.0 M 2026-03-09T15:10:51.750 INFO:teuthology.orchestra.run.vm04.stdout: librados-devel x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 127 k 2026-03-09T15:10:51.750 INFO:teuthology.orchestra.run.vm04.stdout: python3-cephfs x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 165 k 2026-03-09T15:10:51.750 INFO:teuthology.orchestra.run.vm04.stdout: python3-jmespath noarch 1.0.1-1.el9 appstream 48 k 2026-03-09T15:10:51.750 INFO:teuthology.orchestra.run.vm04.stdout: python3-rados x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 323 k 2026-03-09T15:10:51.750 INFO:teuthology.orchestra.run.vm04.stdout: python3-rbd x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 303 k 2026-03-09T15:10:51.750 INFO:teuthology.orchestra.run.vm04.stdout: python3-rgw x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 100 k 2026-03-09T15:10:51.750 INFO:teuthology.orchestra.run.vm04.stdout: python3-xmltodict noarch 0.12.0-15.el9 epel 22 k 2026-03-09T15:10:51.750 INFO:teuthology.orchestra.run.vm04.stdout: rbd-fuse x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 85 k 2026-03-09T15:10:51.750 INFO:teuthology.orchestra.run.vm04.stdout: rbd-mirror x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 3.1 M 2026-03-09T15:10:51.750 INFO:teuthology.orchestra.run.vm04.stdout: rbd-nbd x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 171 k 2026-03-09T15:10:51.750 INFO:teuthology.orchestra.run.vm04.stdout:Upgrading: 2026-03-09T15:10:51.750 INFO:teuthology.orchestra.run.vm04.stdout: librados2 x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 3.4 M 2026-03-09T15:10:51.750 INFO:teuthology.orchestra.run.vm04.stdout: librbd1 x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 3.2 M 2026-03-09T15:10:51.750 INFO:teuthology.orchestra.run.vm04.stdout:Installing dependencies: 2026-03-09T15:10:51.750 INFO:teuthology.orchestra.run.vm04.stdout: abseil-cpp x86_64 20211102.0-4.el9 epel 551 k 2026-03-09T15:10:51.750 INFO:teuthology.orchestra.run.vm04.stdout: boost-program-options x86_64 1.75.0-13.el9 appstream 104 k 2026-03-09T15:10:51.750 INFO:teuthology.orchestra.run.vm04.stdout: ceph-common x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 22 M 2026-03-09T15:10:51.750 INFO:teuthology.orchestra.run.vm04.stdout: ceph-grafana-dashboards noarch 2:19.2.3-678.ge911bdeb.el9 ceph-noarch 31 k 2026-03-09T15:10:51.750 INFO:teuthology.orchestra.run.vm04.stdout: ceph-mds x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 2.4 M 2026-03-09T15:10:51.750 INFO:teuthology.orchestra.run.vm04.stdout: ceph-mgr-modules-core noarch 2:19.2.3-678.ge911bdeb.el9 ceph-noarch 253 k 2026-03-09T15:10:51.750 INFO:teuthology.orchestra.run.vm04.stdout: ceph-mon x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 4.7 M 2026-03-09T15:10:51.750 INFO:teuthology.orchestra.run.vm04.stdout: ceph-osd x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 17 M 2026-03-09T15:10:51.750 INFO:teuthology.orchestra.run.vm04.stdout: ceph-prometheus-alerts noarch 2:19.2.3-678.ge911bdeb.el9 ceph-noarch 17 k 2026-03-09T15:10:51.750 INFO:teuthology.orchestra.run.vm04.stdout: ceph-selinux x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 25 k 2026-03-09T15:10:51.750 INFO:teuthology.orchestra.run.vm04.stdout: cryptsetup x86_64 2.8.1-3.el9 baseos 351 k 2026-03-09T15:10:51.750 INFO:teuthology.orchestra.run.vm04.stdout: flexiblas x86_64 3.0.4-9.el9 appstream 30 k 2026-03-09T15:10:51.750 INFO:teuthology.orchestra.run.vm04.stdout: flexiblas-netlib x86_64 3.0.4-9.el9 appstream 3.0 M 2026-03-09T15:10:51.750 INFO:teuthology.orchestra.run.vm04.stdout: flexiblas-openblas-openmp x86_64 3.0.4-9.el9 appstream 15 k 2026-03-09T15:10:51.750 INFO:teuthology.orchestra.run.vm04.stdout: gperftools-libs x86_64 2.9.1-3.el9 epel 308 k 2026-03-09T15:10:51.750 INFO:teuthology.orchestra.run.vm04.stdout: grpc-data noarch 1.46.7-10.el9 epel 19 k 2026-03-09T15:10:51.750 INFO:teuthology.orchestra.run.vm04.stdout: ledmon-libs x86_64 1.1.0-3.el9 baseos 40 k 2026-03-09T15:10:51.750 INFO:teuthology.orchestra.run.vm04.stdout: libarrow x86_64 9.0.0-15.el9 epel 4.4 M 2026-03-09T15:10:51.750 INFO:teuthology.orchestra.run.vm04.stdout: libarrow-doc noarch 9.0.0-15.el9 epel 25 k 2026-03-09T15:10:51.750 INFO:teuthology.orchestra.run.vm04.stdout: libcephsqlite x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 163 k 2026-03-09T15:10:51.750 INFO:teuthology.orchestra.run.vm04.stdout: libconfig x86_64 1.7.2-9.el9 baseos 72 k 2026-03-09T15:10:51.750 INFO:teuthology.orchestra.run.vm04.stdout: libgfortran x86_64 11.5.0-14.el9 baseos 794 k 2026-03-09T15:10:51.750 INFO:teuthology.orchestra.run.vm04.stdout: libnbd x86_64 1.20.3-4.el9 appstream 164 k 2026-03-09T15:10:51.750 INFO:teuthology.orchestra.run.vm04.stdout: liboath x86_64 2.6.12-1.el9 epel 49 k 2026-03-09T15:10:51.750 INFO:teuthology.orchestra.run.vm04.stdout: libpmemobj x86_64 1.12.1-1.el9 appstream 160 k 2026-03-09T15:10:51.750 INFO:teuthology.orchestra.run.vm04.stdout: libquadmath x86_64 11.5.0-14.el9 baseos 184 k 2026-03-09T15:10:51.750 INFO:teuthology.orchestra.run.vm04.stdout: librabbitmq x86_64 0.11.0-7.el9 appstream 45 k 2026-03-09T15:10:51.750 INFO:teuthology.orchestra.run.vm04.stdout: libradosstriper1 x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 503 k 2026-03-09T15:10:51.750 INFO:teuthology.orchestra.run.vm04.stdout: librdkafka x86_64 1.6.1-102.el9 appstream 662 k 2026-03-09T15:10:51.750 INFO:teuthology.orchestra.run.vm04.stdout: librgw2 x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 5.4 M 2026-03-09T15:10:51.750 INFO:teuthology.orchestra.run.vm04.stdout: libstoragemgmt x86_64 1.10.1-1.el9 appstream 246 k 2026-03-09T15:10:51.750 INFO:teuthology.orchestra.run.vm04.stdout: libunwind x86_64 1.6.2-1.el9 epel 67 k 2026-03-09T15:10:51.750 INFO:teuthology.orchestra.run.vm04.stdout: libxslt x86_64 1.1.34-12.el9 appstream 233 k 2026-03-09T15:10:51.750 INFO:teuthology.orchestra.run.vm04.stdout: lttng-ust x86_64 2.12.0-6.el9 appstream 292 k 2026-03-09T15:10:51.750 INFO:teuthology.orchestra.run.vm04.stdout: lua x86_64 5.4.4-4.el9 appstream 188 k 2026-03-09T15:10:51.750 INFO:teuthology.orchestra.run.vm04.stdout: lua-devel x86_64 5.4.4-4.el9 crb 22 k 2026-03-09T15:10:51.750 INFO:teuthology.orchestra.run.vm04.stdout: luarocks noarch 3.9.2-5.el9 epel 151 k 2026-03-09T15:10:51.750 INFO:teuthology.orchestra.run.vm04.stdout: mailcap noarch 2.1.49-5.el9 baseos 33 k 2026-03-09T15:10:51.750 INFO:teuthology.orchestra.run.vm04.stdout: openblas x86_64 0.3.29-1.el9 appstream 42 k 2026-03-09T15:10:51.750 INFO:teuthology.orchestra.run.vm04.stdout: openblas-openmp x86_64 0.3.29-1.el9 appstream 5.3 M 2026-03-09T15:10:51.751 INFO:teuthology.orchestra.run.vm04.stdout: parquet-libs x86_64 9.0.0-15.el9 epel 838 k 2026-03-09T15:10:51.751 INFO:teuthology.orchestra.run.vm04.stdout: pciutils x86_64 3.7.0-7.el9 baseos 93 k 2026-03-09T15:10:51.751 INFO:teuthology.orchestra.run.vm04.stdout: protobuf x86_64 3.14.0-17.el9 appstream 1.0 M 2026-03-09T15:10:51.751 INFO:teuthology.orchestra.run.vm04.stdout: protobuf-compiler x86_64 3.14.0-17.el9 crb 862 k 2026-03-09T15:10:51.751 INFO:teuthology.orchestra.run.vm04.stdout: python3-asyncssh noarch 2.13.2-5.el9 epel 548 k 2026-03-09T15:10:51.751 INFO:teuthology.orchestra.run.vm04.stdout: python3-autocommand noarch 2.2.2-8.el9 epel 29 k 2026-03-09T15:10:51.751 INFO:teuthology.orchestra.run.vm04.stdout: python3-babel noarch 2.9.1-2.el9 appstream 6.0 M 2026-03-09T15:10:51.751 INFO:teuthology.orchestra.run.vm04.stdout: python3-backports-tarfile noarch 1.2.0-1.el9 epel 60 k 2026-03-09T15:10:51.751 INFO:teuthology.orchestra.run.vm04.stdout: python3-bcrypt x86_64 3.2.2-1.el9 epel 43 k 2026-03-09T15:10:51.751 INFO:teuthology.orchestra.run.vm04.stdout: python3-cachetools noarch 4.2.4-1.el9 epel 32 k 2026-03-09T15:10:51.751 INFO:teuthology.orchestra.run.vm04.stdout: python3-ceph-argparse x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 45 k 2026-03-09T15:10:51.751 INFO:teuthology.orchestra.run.vm04.stdout: python3-ceph-common x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 142 k 2026-03-09T15:10:51.751 INFO:teuthology.orchestra.run.vm04.stdout: python3-certifi noarch 2023.05.07-4.el9 epel 14 k 2026-03-09T15:10:51.751 INFO:teuthology.orchestra.run.vm04.stdout: python3-cffi x86_64 1.14.5-5.el9 baseos 253 k 2026-03-09T15:10:51.751 INFO:teuthology.orchestra.run.vm04.stdout: python3-cheroot noarch 10.0.1-4.el9 epel 173 k 2026-03-09T15:10:51.751 INFO:teuthology.orchestra.run.vm04.stdout: python3-cherrypy noarch 18.6.1-2.el9 epel 358 k 2026-03-09T15:10:51.751 INFO:teuthology.orchestra.run.vm04.stdout: python3-cryptography x86_64 36.0.1-5.el9 baseos 1.2 M 2026-03-09T15:10:51.751 INFO:teuthology.orchestra.run.vm04.stdout: python3-devel x86_64 3.9.25-3.el9 appstream 244 k 2026-03-09T15:10:51.751 INFO:teuthology.orchestra.run.vm04.stdout: python3-google-auth noarch 1:2.45.0-1.el9 epel 254 k 2026-03-09T15:10:51.751 INFO:teuthology.orchestra.run.vm04.stdout: python3-grpcio x86_64 1.46.7-10.el9 epel 2.0 M 2026-03-09T15:10:51.751 INFO:teuthology.orchestra.run.vm04.stdout: python3-grpcio-tools x86_64 1.46.7-10.el9 epel 144 k 2026-03-09T15:10:51.751 INFO:teuthology.orchestra.run.vm04.stdout: python3-jaraco noarch 8.2.1-3.el9 epel 11 k 2026-03-09T15:10:51.751 INFO:teuthology.orchestra.run.vm04.stdout: python3-jaraco-classes noarch 3.2.1-5.el9 epel 18 k 2026-03-09T15:10:51.751 INFO:teuthology.orchestra.run.vm04.stdout: python3-jaraco-collections noarch 3.0.0-8.el9 epel 23 k 2026-03-09T15:10:51.751 INFO:teuthology.orchestra.run.vm04.stdout: python3-jaraco-context noarch 6.0.1-3.el9 epel 20 k 2026-03-09T15:10:51.751 INFO:teuthology.orchestra.run.vm04.stdout: python3-jaraco-functools noarch 3.5.0-2.el9 epel 19 k 2026-03-09T15:10:51.751 INFO:teuthology.orchestra.run.vm04.stdout: python3-jaraco-text noarch 4.0.0-2.el9 epel 26 k 2026-03-09T15:10:51.751 INFO:teuthology.orchestra.run.vm04.stdout: python3-jinja2 noarch 2.11.3-8.el9 appstream 249 k 2026-03-09T15:10:51.751 INFO:teuthology.orchestra.run.vm04.stdout: python3-kubernetes noarch 1:26.1.0-3.el9 epel 1.0 M 2026-03-09T15:10:51.751 INFO:teuthology.orchestra.run.vm04.stdout: python3-libstoragemgmt x86_64 1.10.1-1.el9 appstream 177 k 2026-03-09T15:10:51.751 INFO:teuthology.orchestra.run.vm04.stdout: python3-logutils noarch 0.3.5-21.el9 epel 46 k 2026-03-09T15:10:51.751 INFO:teuthology.orchestra.run.vm04.stdout: python3-mako noarch 1.1.4-6.el9 appstream 172 k 2026-03-09T15:10:51.751 INFO:teuthology.orchestra.run.vm04.stdout: python3-markupsafe x86_64 1.1.1-12.el9 appstream 35 k 2026-03-09T15:10:51.751 INFO:teuthology.orchestra.run.vm04.stdout: python3-more-itertools noarch 8.12.0-2.el9 epel 79 k 2026-03-09T15:10:51.751 INFO:teuthology.orchestra.run.vm04.stdout: python3-natsort noarch 7.1.1-5.el9 epel 58 k 2026-03-09T15:10:51.751 INFO:teuthology.orchestra.run.vm04.stdout: python3-numpy x86_64 1:1.23.5-2.el9 appstream 6.1 M 2026-03-09T15:10:51.751 INFO:teuthology.orchestra.run.vm04.stdout: python3-numpy-f2py x86_64 1:1.23.5-2.el9 appstream 442 k 2026-03-09T15:10:51.751 INFO:teuthology.orchestra.run.vm04.stdout: python3-packaging noarch 20.9-5.el9 appstream 77 k 2026-03-09T15:10:51.751 INFO:teuthology.orchestra.run.vm04.stdout: python3-pecan noarch 1.4.2-3.el9 epel 272 k 2026-03-09T15:10:51.751 INFO:teuthology.orchestra.run.vm04.stdout: python3-ply noarch 3.11-14.el9 baseos 106 k 2026-03-09T15:10:51.751 INFO:teuthology.orchestra.run.vm04.stdout: python3-portend noarch 3.1.0-2.el9 epel 16 k 2026-03-09T15:10:51.751 INFO:teuthology.orchestra.run.vm04.stdout: python3-protobuf noarch 3.14.0-17.el9 appstream 267 k 2026-03-09T15:10:51.751 INFO:teuthology.orchestra.run.vm04.stdout: python3-pyOpenSSL noarch 21.0.0-1.el9 epel 90 k 2026-03-09T15:10:51.751 INFO:teuthology.orchestra.run.vm04.stdout: python3-pyasn1 noarch 0.4.8-7.el9 appstream 157 k 2026-03-09T15:10:51.751 INFO:teuthology.orchestra.run.vm04.stdout: python3-pyasn1-modules noarch 0.4.8-7.el9 appstream 277 k 2026-03-09T15:10:51.751 INFO:teuthology.orchestra.run.vm04.stdout: python3-pycparser noarch 2.20-6.el9 baseos 135 k 2026-03-09T15:10:51.751 INFO:teuthology.orchestra.run.vm04.stdout: python3-repoze-lru noarch 0.7-16.el9 epel 31 k 2026-03-09T15:10:51.751 INFO:teuthology.orchestra.run.vm04.stdout: python3-requests noarch 2.25.1-10.el9 baseos 126 k 2026-03-09T15:10:51.751 INFO:teuthology.orchestra.run.vm04.stdout: python3-requests-oauthlib noarch 1.3.0-12.el9 appstream 54 k 2026-03-09T15:10:51.751 INFO:teuthology.orchestra.run.vm04.stdout: python3-routes noarch 2.5.1-5.el9 epel 188 k 2026-03-09T15:10:51.751 INFO:teuthology.orchestra.run.vm04.stdout: python3-rsa noarch 4.9-2.el9 epel 59 k 2026-03-09T15:10:51.751 INFO:teuthology.orchestra.run.vm04.stdout: python3-scipy x86_64 1.9.3-2.el9 appstream 19 M 2026-03-09T15:10:51.751 INFO:teuthology.orchestra.run.vm04.stdout: python3-tempora noarch 5.0.0-2.el9 epel 36 k 2026-03-09T15:10:51.751 INFO:teuthology.orchestra.run.vm04.stdout: python3-toml noarch 0.10.2-6.el9 appstream 42 k 2026-03-09T15:10:51.751 INFO:teuthology.orchestra.run.vm04.stdout: python3-typing-extensions noarch 4.15.0-1.el9 epel 86 k 2026-03-09T15:10:51.751 INFO:teuthology.orchestra.run.vm04.stdout: python3-urllib3 noarch 1.26.5-7.el9 baseos 218 k 2026-03-09T15:10:51.751 INFO:teuthology.orchestra.run.vm04.stdout: python3-webob noarch 1.8.8-2.el9 epel 230 k 2026-03-09T15:10:51.751 INFO:teuthology.orchestra.run.vm04.stdout: python3-websocket-client noarch 1.2.3-2.el9 epel 90 k 2026-03-09T15:10:51.751 INFO:teuthology.orchestra.run.vm04.stdout: python3-werkzeug noarch 2.0.3-3.el9.1 epel 427 k 2026-03-09T15:10:51.751 INFO:teuthology.orchestra.run.vm04.stdout: python3-zc-lockfile noarch 2.0-10.el9 epel 20 k 2026-03-09T15:10:51.751 INFO:teuthology.orchestra.run.vm04.stdout: qatlib x86_64 25.08.0-2.el9 appstream 240 k 2026-03-09T15:10:51.752 INFO:teuthology.orchestra.run.vm04.stdout: qatzip-libs x86_64 1.3.1-1.el9 appstream 66 k 2026-03-09T15:10:51.752 INFO:teuthology.orchestra.run.vm04.stdout: re2 x86_64 1:20211101-20.el9 epel 191 k 2026-03-09T15:10:51.752 INFO:teuthology.orchestra.run.vm04.stdout: socat x86_64 1.7.4.1-8.el9 appstream 303 k 2026-03-09T15:10:51.752 INFO:teuthology.orchestra.run.vm04.stdout: thrift x86_64 0.15.0-4.el9 epel 1.6 M 2026-03-09T15:10:51.752 INFO:teuthology.orchestra.run.vm04.stdout: unzip x86_64 6.0-59.el9 baseos 182 k 2026-03-09T15:10:51.752 INFO:teuthology.orchestra.run.vm04.stdout: xmlstarlet x86_64 1.6.1-20.el9 appstream 64 k 2026-03-09T15:10:51.752 INFO:teuthology.orchestra.run.vm04.stdout: zip x86_64 3.0-35.el9 baseos 266 k 2026-03-09T15:10:51.752 INFO:teuthology.orchestra.run.vm04.stdout:Installing weak dependencies: 2026-03-09T15:10:51.752 INFO:teuthology.orchestra.run.vm04.stdout: qatlib-service x86_64 25.08.0-2.el9 appstream 37 k 2026-03-09T15:10:51.752 INFO:teuthology.orchestra.run.vm04.stdout: 2026-03-09T15:10:51.752 INFO:teuthology.orchestra.run.vm04.stdout:Transaction Summary 2026-03-09T15:10:51.752 INFO:teuthology.orchestra.run.vm04.stdout:====================================================================================== 2026-03-09T15:10:51.752 INFO:teuthology.orchestra.run.vm04.stdout:Install 134 Packages 2026-03-09T15:10:51.752 INFO:teuthology.orchestra.run.vm04.stdout:Upgrade 2 Packages 2026-03-09T15:10:51.752 INFO:teuthology.orchestra.run.vm04.stdout: 2026-03-09T15:10:51.752 INFO:teuthology.orchestra.run.vm04.stdout:Total download size: 210 M 2026-03-09T15:10:51.752 INFO:teuthology.orchestra.run.vm04.stdout:Downloading Packages: 2026-03-09T15:10:52.020 INFO:teuthology.orchestra.run.vm06.stdout:(2/136): ceph-fuse-19.2.3-678.ge911bdeb.el9.x86 1.4 MB/s | 1.2 MB 00:00 2026-03-09T15:10:52.138 INFO:teuthology.orchestra.run.vm06.stdout:(3/136): ceph-immutable-object-cache-19.2.3-678 1.2 MB/s | 145 kB 00:00 2026-03-09T15:10:52.168 INFO:teuthology.orchestra.run.vm03.stdout:Extra Packages for Enterprise Linux 19 MB/s | 20 MB 00:01 2026-03-09T15:10:52.840 INFO:teuthology.orchestra.run.vm06.stdout:(4/136): ceph-mds-19.2.3-678.ge911bdeb.el9.x86_ 3.4 MB/s | 2.4 MB 00:00 2026-03-09T15:10:53.190 INFO:teuthology.orchestra.run.vm06.stdout:(5/136): ceph-mgr-19.2.3-678.ge911bdeb.el9.x86_ 3.1 MB/s | 1.1 MB 00:00 2026-03-09T15:10:53.369 INFO:teuthology.orchestra.run.vm04.stdout:(1/136): ceph-19.2.3-678.ge911bdeb.el9.x86_64.r 14 kB/s | 6.5 kB 00:00 2026-03-09T15:10:54.232 INFO:teuthology.orchestra.run.vm04.stdout:(2/136): ceph-fuse-19.2.3-678.ge911bdeb.el9.x86 1.3 MB/s | 1.2 MB 00:00 2026-03-09T15:10:54.352 INFO:teuthology.orchestra.run.vm04.stdout:(3/136): ceph-immutable-object-cache-19.2.3-678 1.2 MB/s | 145 kB 00:00 2026-03-09T15:10:54.470 INFO:teuthology.orchestra.run.vm06.stdout:(6/136): ceph-mon-19.2.3-678.ge911bdeb.el9.x86_ 3.7 MB/s | 4.7 MB 00:01 2026-03-09T15:10:54.493 INFO:teuthology.orchestra.run.vm06.stdout:(7/136): ceph-base-19.2.3-678.ge911bdeb.el9.x86 1.5 MB/s | 5.5 MB 00:03 2026-03-09T15:10:55.299 INFO:teuthology.orchestra.run.vm04.stdout:(4/136): ceph-mds-19.2.3-678.ge911bdeb.el9.x86_ 2.6 MB/s | 2.4 MB 00:00 2026-03-09T15:10:55.771 INFO:teuthology.orchestra.run.vm04.stdout:(5/136): ceph-mgr-19.2.3-678.ge911bdeb.el9.x86_ 2.3 MB/s | 1.1 MB 00:00 2026-03-09T15:10:55.849 INFO:teuthology.orchestra.run.vm04.stdout:(6/136): ceph-base-19.2.3-678.ge911bdeb.el9.x86 1.9 MB/s | 5.5 MB 00:02 2026-03-09T15:10:56.640 INFO:teuthology.orchestra.run.vm03.stdout:lab-extras 65 kB/s | 50 kB 00:00 2026-03-09T15:10:57.307 INFO:teuthology.orchestra.run.vm04.stdout:(7/136): ceph-mon-19.2.3-678.ge911bdeb.el9.x86_ 3.1 MB/s | 4.7 MB 00:01 2026-03-09T15:10:57.972 INFO:teuthology.orchestra.run.vm03.stdout:Package librados2-2:16.2.4-5.el9.x86_64 is already installed. 2026-03-09T15:10:57.973 INFO:teuthology.orchestra.run.vm03.stdout:Package librbd1-2:16.2.4-5.el9.x86_64 is already installed. 2026-03-09T15:10:57.976 INFO:teuthology.orchestra.run.vm03.stdout:Package bzip2-1.0.8-11.el9.x86_64 is already installed. 2026-03-09T15:10:57.977 INFO:teuthology.orchestra.run.vm03.stdout:Package perl-Test-Harness-1:3.42-461.el9.noarch is already installed. 2026-03-09T15:10:58.005 INFO:teuthology.orchestra.run.vm03.stdout:Dependencies resolved. 2026-03-09T15:10:58.009 INFO:teuthology.orchestra.run.vm03.stdout:====================================================================================== 2026-03-09T15:10:58.009 INFO:teuthology.orchestra.run.vm03.stdout: Package Arch Version Repository Size 2026-03-09T15:10:58.009 INFO:teuthology.orchestra.run.vm03.stdout:====================================================================================== 2026-03-09T15:10:58.009 INFO:teuthology.orchestra.run.vm03.stdout:Installing: 2026-03-09T15:10:58.009 INFO:teuthology.orchestra.run.vm03.stdout: ceph x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 6.5 k 2026-03-09T15:10:58.009 INFO:teuthology.orchestra.run.vm03.stdout: ceph-base x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 5.5 M 2026-03-09T15:10:58.009 INFO:teuthology.orchestra.run.vm03.stdout: ceph-fuse x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 1.2 M 2026-03-09T15:10:58.010 INFO:teuthology.orchestra.run.vm03.stdout: ceph-immutable-object-cache x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 145 k 2026-03-09T15:10:58.010 INFO:teuthology.orchestra.run.vm03.stdout: ceph-mgr x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 1.1 M 2026-03-09T15:10:58.010 INFO:teuthology.orchestra.run.vm03.stdout: ceph-mgr-cephadm noarch 2:19.2.3-678.ge911bdeb.el9 ceph-noarch 150 k 2026-03-09T15:10:58.010 INFO:teuthology.orchestra.run.vm03.stdout: ceph-mgr-dashboard noarch 2:19.2.3-678.ge911bdeb.el9 ceph-noarch 3.8 M 2026-03-09T15:10:58.010 INFO:teuthology.orchestra.run.vm03.stdout: ceph-mgr-diskprediction-local noarch 2:19.2.3-678.ge911bdeb.el9 ceph-noarch 7.4 M 2026-03-09T15:10:58.010 INFO:teuthology.orchestra.run.vm03.stdout: ceph-mgr-rook noarch 2:19.2.3-678.ge911bdeb.el9 ceph-noarch 49 k 2026-03-09T15:10:58.010 INFO:teuthology.orchestra.run.vm03.stdout: ceph-radosgw x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 11 M 2026-03-09T15:10:58.010 INFO:teuthology.orchestra.run.vm03.stdout: ceph-test x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 50 M 2026-03-09T15:10:58.010 INFO:teuthology.orchestra.run.vm03.stdout: ceph-volume noarch 2:19.2.3-678.ge911bdeb.el9 ceph-noarch 299 k 2026-03-09T15:10:58.010 INFO:teuthology.orchestra.run.vm03.stdout: cephadm noarch 2:19.2.3-678.ge911bdeb.el9 ceph-noarch 769 k 2026-03-09T15:10:58.010 INFO:teuthology.orchestra.run.vm03.stdout: libcephfs-devel x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 34 k 2026-03-09T15:10:58.010 INFO:teuthology.orchestra.run.vm03.stdout: libcephfs2 x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 1.0 M 2026-03-09T15:10:58.010 INFO:teuthology.orchestra.run.vm03.stdout: librados-devel x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 127 k 2026-03-09T15:10:58.010 INFO:teuthology.orchestra.run.vm03.stdout: python3-cephfs x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 165 k 2026-03-09T15:10:58.010 INFO:teuthology.orchestra.run.vm03.stdout: python3-jmespath noarch 1.0.1-1.el9 appstream 48 k 2026-03-09T15:10:58.010 INFO:teuthology.orchestra.run.vm03.stdout: python3-rados x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 323 k 2026-03-09T15:10:58.010 INFO:teuthology.orchestra.run.vm03.stdout: python3-rbd x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 303 k 2026-03-09T15:10:58.010 INFO:teuthology.orchestra.run.vm03.stdout: python3-rgw x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 100 k 2026-03-09T15:10:58.010 INFO:teuthology.orchestra.run.vm03.stdout: python3-xmltodict noarch 0.12.0-15.el9 epel 22 k 2026-03-09T15:10:58.010 INFO:teuthology.orchestra.run.vm03.stdout: rbd-fuse x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 85 k 2026-03-09T15:10:58.010 INFO:teuthology.orchestra.run.vm03.stdout: rbd-mirror x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 3.1 M 2026-03-09T15:10:58.010 INFO:teuthology.orchestra.run.vm03.stdout: rbd-nbd x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 171 k 2026-03-09T15:10:58.010 INFO:teuthology.orchestra.run.vm03.stdout:Upgrading: 2026-03-09T15:10:58.010 INFO:teuthology.orchestra.run.vm03.stdout: librados2 x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 3.4 M 2026-03-09T15:10:58.010 INFO:teuthology.orchestra.run.vm03.stdout: librbd1 x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 3.2 M 2026-03-09T15:10:58.010 INFO:teuthology.orchestra.run.vm03.stdout:Installing dependencies: 2026-03-09T15:10:58.010 INFO:teuthology.orchestra.run.vm03.stdout: abseil-cpp x86_64 20211102.0-4.el9 epel 551 k 2026-03-09T15:10:58.010 INFO:teuthology.orchestra.run.vm03.stdout: boost-program-options x86_64 1.75.0-13.el9 appstream 104 k 2026-03-09T15:10:58.010 INFO:teuthology.orchestra.run.vm03.stdout: ceph-common x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 22 M 2026-03-09T15:10:58.010 INFO:teuthology.orchestra.run.vm03.stdout: ceph-grafana-dashboards noarch 2:19.2.3-678.ge911bdeb.el9 ceph-noarch 31 k 2026-03-09T15:10:58.010 INFO:teuthology.orchestra.run.vm03.stdout: ceph-mds x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 2.4 M 2026-03-09T15:10:58.010 INFO:teuthology.orchestra.run.vm03.stdout: ceph-mgr-modules-core noarch 2:19.2.3-678.ge911bdeb.el9 ceph-noarch 253 k 2026-03-09T15:10:58.010 INFO:teuthology.orchestra.run.vm03.stdout: ceph-mon x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 4.7 M 2026-03-09T15:10:58.010 INFO:teuthology.orchestra.run.vm03.stdout: ceph-osd x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 17 M 2026-03-09T15:10:58.010 INFO:teuthology.orchestra.run.vm03.stdout: ceph-prometheus-alerts noarch 2:19.2.3-678.ge911bdeb.el9 ceph-noarch 17 k 2026-03-09T15:10:58.010 INFO:teuthology.orchestra.run.vm03.stdout: ceph-selinux x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 25 k 2026-03-09T15:10:58.010 INFO:teuthology.orchestra.run.vm03.stdout: cryptsetup x86_64 2.8.1-3.el9 baseos 351 k 2026-03-09T15:10:58.010 INFO:teuthology.orchestra.run.vm03.stdout: flexiblas x86_64 3.0.4-9.el9 appstream 30 k 2026-03-09T15:10:58.010 INFO:teuthology.orchestra.run.vm03.stdout: flexiblas-netlib x86_64 3.0.4-9.el9 appstream 3.0 M 2026-03-09T15:10:58.010 INFO:teuthology.orchestra.run.vm03.stdout: flexiblas-openblas-openmp x86_64 3.0.4-9.el9 appstream 15 k 2026-03-09T15:10:58.010 INFO:teuthology.orchestra.run.vm03.stdout: gperftools-libs x86_64 2.9.1-3.el9 epel 308 k 2026-03-09T15:10:58.010 INFO:teuthology.orchestra.run.vm03.stdout: grpc-data noarch 1.46.7-10.el9 epel 19 k 2026-03-09T15:10:58.010 INFO:teuthology.orchestra.run.vm03.stdout: ledmon-libs x86_64 1.1.0-3.el9 baseos 40 k 2026-03-09T15:10:58.010 INFO:teuthology.orchestra.run.vm03.stdout: libarrow x86_64 9.0.0-15.el9 epel 4.4 M 2026-03-09T15:10:58.010 INFO:teuthology.orchestra.run.vm03.stdout: libarrow-doc noarch 9.0.0-15.el9 epel 25 k 2026-03-09T15:10:58.010 INFO:teuthology.orchestra.run.vm03.stdout: libcephsqlite x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 163 k 2026-03-09T15:10:58.010 INFO:teuthology.orchestra.run.vm03.stdout: libconfig x86_64 1.7.2-9.el9 baseos 72 k 2026-03-09T15:10:58.010 INFO:teuthology.orchestra.run.vm03.stdout: libgfortran x86_64 11.5.0-14.el9 baseos 794 k 2026-03-09T15:10:58.010 INFO:teuthology.orchestra.run.vm03.stdout: libnbd x86_64 1.20.3-4.el9 appstream 164 k 2026-03-09T15:10:58.010 INFO:teuthology.orchestra.run.vm03.stdout: liboath x86_64 2.6.12-1.el9 epel 49 k 2026-03-09T15:10:58.010 INFO:teuthology.orchestra.run.vm03.stdout: libpmemobj x86_64 1.12.1-1.el9 appstream 160 k 2026-03-09T15:10:58.010 INFO:teuthology.orchestra.run.vm03.stdout: libquadmath x86_64 11.5.0-14.el9 baseos 184 k 2026-03-09T15:10:58.010 INFO:teuthology.orchestra.run.vm03.stdout: librabbitmq x86_64 0.11.0-7.el9 appstream 45 k 2026-03-09T15:10:58.010 INFO:teuthology.orchestra.run.vm03.stdout: libradosstriper1 x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 503 k 2026-03-09T15:10:58.010 INFO:teuthology.orchestra.run.vm03.stdout: librdkafka x86_64 1.6.1-102.el9 appstream 662 k 2026-03-09T15:10:58.010 INFO:teuthology.orchestra.run.vm03.stdout: librgw2 x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 5.4 M 2026-03-09T15:10:58.011 INFO:teuthology.orchestra.run.vm03.stdout: libstoragemgmt x86_64 1.10.1-1.el9 appstream 246 k 2026-03-09T15:10:58.011 INFO:teuthology.orchestra.run.vm03.stdout: libunwind x86_64 1.6.2-1.el9 epel 67 k 2026-03-09T15:10:58.011 INFO:teuthology.orchestra.run.vm03.stdout: libxslt x86_64 1.1.34-12.el9 appstream 233 k 2026-03-09T15:10:58.011 INFO:teuthology.orchestra.run.vm03.stdout: lttng-ust x86_64 2.12.0-6.el9 appstream 292 k 2026-03-09T15:10:58.011 INFO:teuthology.orchestra.run.vm03.stdout: lua x86_64 5.4.4-4.el9 appstream 188 k 2026-03-09T15:10:58.011 INFO:teuthology.orchestra.run.vm03.stdout: lua-devel x86_64 5.4.4-4.el9 crb 22 k 2026-03-09T15:10:58.011 INFO:teuthology.orchestra.run.vm03.stdout: luarocks noarch 3.9.2-5.el9 epel 151 k 2026-03-09T15:10:58.011 INFO:teuthology.orchestra.run.vm03.stdout: mailcap noarch 2.1.49-5.el9 baseos 33 k 2026-03-09T15:10:58.011 INFO:teuthology.orchestra.run.vm03.stdout: openblas x86_64 0.3.29-1.el9 appstream 42 k 2026-03-09T15:10:58.011 INFO:teuthology.orchestra.run.vm03.stdout: openblas-openmp x86_64 0.3.29-1.el9 appstream 5.3 M 2026-03-09T15:10:58.011 INFO:teuthology.orchestra.run.vm03.stdout: parquet-libs x86_64 9.0.0-15.el9 epel 838 k 2026-03-09T15:10:58.011 INFO:teuthology.orchestra.run.vm03.stdout: pciutils x86_64 3.7.0-7.el9 baseos 93 k 2026-03-09T15:10:58.011 INFO:teuthology.orchestra.run.vm03.stdout: protobuf x86_64 3.14.0-17.el9 appstream 1.0 M 2026-03-09T15:10:58.011 INFO:teuthology.orchestra.run.vm03.stdout: protobuf-compiler x86_64 3.14.0-17.el9 crb 862 k 2026-03-09T15:10:58.011 INFO:teuthology.orchestra.run.vm03.stdout: python3-asyncssh noarch 2.13.2-5.el9 epel 548 k 2026-03-09T15:10:58.011 INFO:teuthology.orchestra.run.vm03.stdout: python3-autocommand noarch 2.2.2-8.el9 epel 29 k 2026-03-09T15:10:58.011 INFO:teuthology.orchestra.run.vm03.stdout: python3-babel noarch 2.9.1-2.el9 appstream 6.0 M 2026-03-09T15:10:58.011 INFO:teuthology.orchestra.run.vm03.stdout: python3-backports-tarfile noarch 1.2.0-1.el9 epel 60 k 2026-03-09T15:10:58.011 INFO:teuthology.orchestra.run.vm03.stdout: python3-bcrypt x86_64 3.2.2-1.el9 epel 43 k 2026-03-09T15:10:58.011 INFO:teuthology.orchestra.run.vm03.stdout: python3-cachetools noarch 4.2.4-1.el9 epel 32 k 2026-03-09T15:10:58.011 INFO:teuthology.orchestra.run.vm03.stdout: python3-ceph-argparse x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 45 k 2026-03-09T15:10:58.011 INFO:teuthology.orchestra.run.vm03.stdout: python3-ceph-common x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 142 k 2026-03-09T15:10:58.011 INFO:teuthology.orchestra.run.vm03.stdout: python3-certifi noarch 2023.05.07-4.el9 epel 14 k 2026-03-09T15:10:58.011 INFO:teuthology.orchestra.run.vm03.stdout: python3-cffi x86_64 1.14.5-5.el9 baseos 253 k 2026-03-09T15:10:58.011 INFO:teuthology.orchestra.run.vm03.stdout: python3-cheroot noarch 10.0.1-4.el9 epel 173 k 2026-03-09T15:10:58.011 INFO:teuthology.orchestra.run.vm03.stdout: python3-cherrypy noarch 18.6.1-2.el9 epel 358 k 2026-03-09T15:10:58.011 INFO:teuthology.orchestra.run.vm03.stdout: python3-cryptography x86_64 36.0.1-5.el9 baseos 1.2 M 2026-03-09T15:10:58.011 INFO:teuthology.orchestra.run.vm03.stdout: python3-devel x86_64 3.9.25-3.el9 appstream 244 k 2026-03-09T15:10:58.011 INFO:teuthology.orchestra.run.vm03.stdout: python3-google-auth noarch 1:2.45.0-1.el9 epel 254 k 2026-03-09T15:10:58.011 INFO:teuthology.orchestra.run.vm03.stdout: python3-grpcio x86_64 1.46.7-10.el9 epel 2.0 M 2026-03-09T15:10:58.011 INFO:teuthology.orchestra.run.vm03.stdout: python3-grpcio-tools x86_64 1.46.7-10.el9 epel 144 k 2026-03-09T15:10:58.011 INFO:teuthology.orchestra.run.vm03.stdout: python3-jaraco noarch 8.2.1-3.el9 epel 11 k 2026-03-09T15:10:58.011 INFO:teuthology.orchestra.run.vm03.stdout: python3-jaraco-classes noarch 3.2.1-5.el9 epel 18 k 2026-03-09T15:10:58.011 INFO:teuthology.orchestra.run.vm03.stdout: python3-jaraco-collections noarch 3.0.0-8.el9 epel 23 k 2026-03-09T15:10:58.011 INFO:teuthology.orchestra.run.vm03.stdout: python3-jaraco-context noarch 6.0.1-3.el9 epel 20 k 2026-03-09T15:10:58.011 INFO:teuthology.orchestra.run.vm03.stdout: python3-jaraco-functools noarch 3.5.0-2.el9 epel 19 k 2026-03-09T15:10:58.011 INFO:teuthology.orchestra.run.vm03.stdout: python3-jaraco-text noarch 4.0.0-2.el9 epel 26 k 2026-03-09T15:10:58.011 INFO:teuthology.orchestra.run.vm03.stdout: python3-jinja2 noarch 2.11.3-8.el9 appstream 249 k 2026-03-09T15:10:58.011 INFO:teuthology.orchestra.run.vm03.stdout: python3-kubernetes noarch 1:26.1.0-3.el9 epel 1.0 M 2026-03-09T15:10:58.011 INFO:teuthology.orchestra.run.vm03.stdout: python3-libstoragemgmt x86_64 1.10.1-1.el9 appstream 177 k 2026-03-09T15:10:58.011 INFO:teuthology.orchestra.run.vm03.stdout: python3-logutils noarch 0.3.5-21.el9 epel 46 k 2026-03-09T15:10:58.011 INFO:teuthology.orchestra.run.vm03.stdout: python3-mako noarch 1.1.4-6.el9 appstream 172 k 2026-03-09T15:10:58.011 INFO:teuthology.orchestra.run.vm03.stdout: python3-markupsafe x86_64 1.1.1-12.el9 appstream 35 k 2026-03-09T15:10:58.011 INFO:teuthology.orchestra.run.vm03.stdout: python3-more-itertools noarch 8.12.0-2.el9 epel 79 k 2026-03-09T15:10:58.011 INFO:teuthology.orchestra.run.vm03.stdout: python3-natsort noarch 7.1.1-5.el9 epel 58 k 2026-03-09T15:10:58.011 INFO:teuthology.orchestra.run.vm03.stdout: python3-numpy x86_64 1:1.23.5-2.el9 appstream 6.1 M 2026-03-09T15:10:58.011 INFO:teuthology.orchestra.run.vm03.stdout: python3-numpy-f2py x86_64 1:1.23.5-2.el9 appstream 442 k 2026-03-09T15:10:58.011 INFO:teuthology.orchestra.run.vm03.stdout: python3-packaging noarch 20.9-5.el9 appstream 77 k 2026-03-09T15:10:58.011 INFO:teuthology.orchestra.run.vm03.stdout: python3-pecan noarch 1.4.2-3.el9 epel 272 k 2026-03-09T15:10:58.011 INFO:teuthology.orchestra.run.vm03.stdout: python3-ply noarch 3.11-14.el9 baseos 106 k 2026-03-09T15:10:58.011 INFO:teuthology.orchestra.run.vm03.stdout: python3-portend noarch 3.1.0-2.el9 epel 16 k 2026-03-09T15:10:58.011 INFO:teuthology.orchestra.run.vm03.stdout: python3-protobuf noarch 3.14.0-17.el9 appstream 267 k 2026-03-09T15:10:58.011 INFO:teuthology.orchestra.run.vm03.stdout: python3-pyOpenSSL noarch 21.0.0-1.el9 epel 90 k 2026-03-09T15:10:58.011 INFO:teuthology.orchestra.run.vm03.stdout: python3-pyasn1 noarch 0.4.8-7.el9 appstream 157 k 2026-03-09T15:10:58.011 INFO:teuthology.orchestra.run.vm03.stdout: python3-pyasn1-modules noarch 0.4.8-7.el9 appstream 277 k 2026-03-09T15:10:58.011 INFO:teuthology.orchestra.run.vm03.stdout: python3-pycparser noarch 2.20-6.el9 baseos 135 k 2026-03-09T15:10:58.011 INFO:teuthology.orchestra.run.vm03.stdout: python3-repoze-lru noarch 0.7-16.el9 epel 31 k 2026-03-09T15:10:58.011 INFO:teuthology.orchestra.run.vm03.stdout: python3-requests noarch 2.25.1-10.el9 baseos 126 k 2026-03-09T15:10:58.011 INFO:teuthology.orchestra.run.vm03.stdout: python3-requests-oauthlib noarch 1.3.0-12.el9 appstream 54 k 2026-03-09T15:10:58.011 INFO:teuthology.orchestra.run.vm03.stdout: python3-routes noarch 2.5.1-5.el9 epel 188 k 2026-03-09T15:10:58.011 INFO:teuthology.orchestra.run.vm03.stdout: python3-rsa noarch 4.9-2.el9 epel 59 k 2026-03-09T15:10:58.011 INFO:teuthology.orchestra.run.vm03.stdout: python3-scipy x86_64 1.9.3-2.el9 appstream 19 M 2026-03-09T15:10:58.011 INFO:teuthology.orchestra.run.vm03.stdout: python3-tempora noarch 5.0.0-2.el9 epel 36 k 2026-03-09T15:10:58.011 INFO:teuthology.orchestra.run.vm03.stdout: python3-toml noarch 0.10.2-6.el9 appstream 42 k 2026-03-09T15:10:58.012 INFO:teuthology.orchestra.run.vm03.stdout: python3-typing-extensions noarch 4.15.0-1.el9 epel 86 k 2026-03-09T15:10:58.012 INFO:teuthology.orchestra.run.vm03.stdout: python3-urllib3 noarch 1.26.5-7.el9 baseos 218 k 2026-03-09T15:10:58.012 INFO:teuthology.orchestra.run.vm03.stdout: python3-webob noarch 1.8.8-2.el9 epel 230 k 2026-03-09T15:10:58.012 INFO:teuthology.orchestra.run.vm03.stdout: python3-websocket-client noarch 1.2.3-2.el9 epel 90 k 2026-03-09T15:10:58.012 INFO:teuthology.orchestra.run.vm03.stdout: python3-werkzeug noarch 2.0.3-3.el9.1 epel 427 k 2026-03-09T15:10:58.012 INFO:teuthology.orchestra.run.vm03.stdout: python3-zc-lockfile noarch 2.0-10.el9 epel 20 k 2026-03-09T15:10:58.012 INFO:teuthology.orchestra.run.vm03.stdout: qatlib x86_64 25.08.0-2.el9 appstream 240 k 2026-03-09T15:10:58.012 INFO:teuthology.orchestra.run.vm03.stdout: qatzip-libs x86_64 1.3.1-1.el9 appstream 66 k 2026-03-09T15:10:58.012 INFO:teuthology.orchestra.run.vm03.stdout: re2 x86_64 1:20211101-20.el9 epel 191 k 2026-03-09T15:10:58.012 INFO:teuthology.orchestra.run.vm03.stdout: socat x86_64 1.7.4.1-8.el9 appstream 303 k 2026-03-09T15:10:58.012 INFO:teuthology.orchestra.run.vm03.stdout: thrift x86_64 0.15.0-4.el9 epel 1.6 M 2026-03-09T15:10:58.012 INFO:teuthology.orchestra.run.vm03.stdout: unzip x86_64 6.0-59.el9 baseos 182 k 2026-03-09T15:10:58.012 INFO:teuthology.orchestra.run.vm03.stdout: xmlstarlet x86_64 1.6.1-20.el9 appstream 64 k 2026-03-09T15:10:58.012 INFO:teuthology.orchestra.run.vm03.stdout: zip x86_64 3.0-35.el9 baseos 266 k 2026-03-09T15:10:58.012 INFO:teuthology.orchestra.run.vm03.stdout:Installing weak dependencies: 2026-03-09T15:10:58.012 INFO:teuthology.orchestra.run.vm03.stdout: qatlib-service x86_64 25.08.0-2.el9 appstream 37 k 2026-03-09T15:10:58.012 INFO:teuthology.orchestra.run.vm03.stdout: 2026-03-09T15:10:58.012 INFO:teuthology.orchestra.run.vm03.stdout:Transaction Summary 2026-03-09T15:10:58.012 INFO:teuthology.orchestra.run.vm03.stdout:====================================================================================== 2026-03-09T15:10:58.012 INFO:teuthology.orchestra.run.vm03.stdout:Install 134 Packages 2026-03-09T15:10:58.012 INFO:teuthology.orchestra.run.vm03.stdout:Upgrade 2 Packages 2026-03-09T15:10:58.012 INFO:teuthology.orchestra.run.vm03.stdout: 2026-03-09T15:10:58.012 INFO:teuthology.orchestra.run.vm03.stdout:Total download size: 210 M 2026-03-09T15:10:58.012 INFO:teuthology.orchestra.run.vm03.stdout:Downloading Packages: 2026-03-09T15:10:59.404 INFO:teuthology.orchestra.run.vm06.stdout:(8/136): ceph-radosgw-19.2.3-678.ge911bdeb.el9. 2.2 MB/s | 11 MB 00:04 2026-03-09T15:10:59.525 INFO:teuthology.orchestra.run.vm06.stdout:(9/136): ceph-selinux-19.2.3-678.ge911bdeb.el9. 208 kB/s | 25 kB 00:00 2026-03-09T15:10:59.704 INFO:teuthology.orchestra.run.vm06.stdout:(10/136): ceph-osd-19.2.3-678.ge911bdeb.el9.x86 3.3 MB/s | 17 MB 00:05 2026-03-09T15:10:59.780 INFO:teuthology.orchestra.run.vm03.stdout:(1/136): ceph-19.2.3-678.ge911bdeb.el9.x86_64.r 15 kB/s | 6.5 kB 00:00 2026-03-09T15:10:59.820 INFO:teuthology.orchestra.run.vm06.stdout:(11/136): libcephfs-devel-19.2.3-678.ge911bdeb. 288 kB/s | 34 kB 00:00 2026-03-09T15:11:00.171 INFO:teuthology.orchestra.run.vm06.stdout:(12/136): libcephfs2-19.2.3-678.ge911bdeb.el9.x 2.8 MB/s | 1.0 MB 00:00 2026-03-09T15:11:00.289 INFO:teuthology.orchestra.run.vm06.stdout:(13/136): libcephsqlite-19.2.3-678.ge911bdeb.el 1.4 MB/s | 163 kB 00:00 2026-03-09T15:11:00.406 INFO:teuthology.orchestra.run.vm06.stdout:(14/136): librados-devel-19.2.3-678.ge911bdeb.e 1.1 MB/s | 127 kB 00:00 2026-03-09T15:11:00.613 INFO:teuthology.orchestra.run.vm04.stdout:(8/136): ceph-radosgw-19.2.3-678.ge911bdeb.el9. 3.2 MB/s | 11 MB 00:03 2026-03-09T15:11:00.639 INFO:teuthology.orchestra.run.vm06.stdout:(15/136): libradosstriper1-19.2.3-678.ge911bdeb 2.1 MB/s | 503 kB 00:00 2026-03-09T15:11:00.645 INFO:teuthology.orchestra.run.vm03.stdout:(2/136): ceph-fuse-19.2.3-678.ge911bdeb.el9.x86 1.3 MB/s | 1.2 MB 00:00 2026-03-09T15:11:00.732 INFO:teuthology.orchestra.run.vm04.stdout:(9/136): ceph-selinux-19.2.3-678.ge911bdeb.el9. 212 kB/s | 25 kB 00:00 2026-03-09T15:11:00.762 INFO:teuthology.orchestra.run.vm03.stdout:(3/136): ceph-immutable-object-cache-19.2.3-678 1.2 MB/s | 145 kB 00:00 2026-03-09T15:11:01.426 INFO:teuthology.orchestra.run.vm03.stdout:(4/136): ceph-mds-19.2.3-678.ge911bdeb.el9.x86_ 3.6 MB/s | 2.4 MB 00:00 2026-03-09T15:11:01.478 INFO:teuthology.orchestra.run.vm04.stdout:(10/136): ceph-osd-19.2.3-678.ge911bdeb.el9.x86 3.0 MB/s | 17 MB 00:05 2026-03-09T15:11:01.653 INFO:teuthology.orchestra.run.vm04.stdout:(11/136): ceph-common-19.2.3-678.ge911bdeb.el9. 2.5 MB/s | 22 MB 00:08 2026-03-09T15:11:01.654 INFO:teuthology.orchestra.run.vm04.stdout:(12/136): libcephfs-devel-19.2.3-678.ge911bdeb. 191 kB/s | 34 kB 00:00 2026-03-09T15:11:01.675 INFO:teuthology.orchestra.run.vm03.stdout:(5/136): ceph-base-19.2.3-678.ge911bdeb.el9.x86 2.4 MB/s | 5.5 MB 00:02 2026-03-09T15:11:01.752 INFO:teuthology.orchestra.run.vm03.stdout:(6/136): ceph-mgr-19.2.3-678.ge911bdeb.el9.x86_ 3.3 MB/s | 1.1 MB 00:00 2026-03-09T15:11:01.776 INFO:teuthology.orchestra.run.vm04.stdout:(13/136): libcephsqlite-19.2.3-678.ge911bdeb.el 1.3 MB/s | 163 kB 00:00 2026-03-09T15:11:01.897 INFO:teuthology.orchestra.run.vm04.stdout:(14/136): librados-devel-19.2.3-678.ge911bdeb.e 1.0 MB/s | 127 kB 00:00 2026-03-09T15:11:02.011 INFO:teuthology.orchestra.run.vm04.stdout:(15/136): libcephfs2-19.2.3-678.ge911bdeb.el9.x 2.7 MB/s | 1.0 MB 00:00 2026-03-09T15:11:02.135 INFO:teuthology.orchestra.run.vm04.stdout:(16/136): libradosstriper1-19.2.3-678.ge911bdeb 2.1 MB/s | 503 kB 00:00 2026-03-09T15:11:02.253 INFO:teuthology.orchestra.run.vm04.stdout:(17/136): python3-ceph-argparse-19.2.3-678.ge91 381 kB/s | 45 kB 00:00 2026-03-09T15:11:02.283 INFO:teuthology.orchestra.run.vm06.stdout:(16/136): librgw2-19.2.3-678.ge911bdeb.el9.x86_ 3.3 MB/s | 5.4 MB 00:01 2026-03-09T15:11:02.373 INFO:teuthology.orchestra.run.vm04.stdout:(18/136): python3-ceph-common-19.2.3-678.ge911b 1.2 MB/s | 142 kB 00:00 2026-03-09T15:11:02.401 INFO:teuthology.orchestra.run.vm06.stdout:(17/136): python3-ceph-argparse-19.2.3-678.ge91 382 kB/s | 45 kB 00:00 2026-03-09T15:11:02.494 INFO:teuthology.orchestra.run.vm04.stdout:(19/136): python3-cephfs-19.2.3-678.ge911bdeb.e 1.3 MB/s | 165 kB 00:00 2026-03-09T15:11:02.505 INFO:teuthology.orchestra.run.vm03.stdout:(7/136): ceph-mon-19.2.3-678.ge911bdeb.el9.x86_ 5.7 MB/s | 4.7 MB 00:00 2026-03-09T15:11:02.519 INFO:teuthology.orchestra.run.vm06.stdout:(18/136): python3-ceph-common-19.2.3-678.ge911b 1.2 MB/s | 142 kB 00:00 2026-03-09T15:11:02.615 INFO:teuthology.orchestra.run.vm04.stdout:(20/136): python3-rados-19.2.3-678.ge911bdeb.el 2.6 MB/s | 323 kB 00:00 2026-03-09T15:11:02.644 INFO:teuthology.orchestra.run.vm06.stdout:(19/136): python3-cephfs-19.2.3-678.ge911bdeb.e 1.3 MB/s | 165 kB 00:00 2026-03-09T15:11:02.738 INFO:teuthology.orchestra.run.vm04.stdout:(21/136): python3-rbd-19.2.3-678.ge911bdeb.el9. 2.4 MB/s | 303 kB 00:00 2026-03-09T15:11:02.764 INFO:teuthology.orchestra.run.vm06.stdout:(20/136): python3-rados-19.2.3-678.ge911bdeb.el 2.6 MB/s | 323 kB 00:00 2026-03-09T15:11:02.858 INFO:teuthology.orchestra.run.vm04.stdout:(22/136): python3-rgw-19.2.3-678.ge911bdeb.el9. 833 kB/s | 100 kB 00:00 2026-03-09T15:11:02.883 INFO:teuthology.orchestra.run.vm06.stdout:(21/136): python3-rbd-19.2.3-678.ge911bdeb.el9. 2.5 MB/s | 303 kB 00:00 2026-03-09T15:11:02.977 INFO:teuthology.orchestra.run.vm04.stdout:(23/136): rbd-fuse-19.2.3-678.ge911bdeb.el9.x86 714 kB/s | 85 kB 00:00 2026-03-09T15:11:03.001 INFO:teuthology.orchestra.run.vm06.stdout:(22/136): python3-rgw-19.2.3-678.ge911bdeb.el9. 845 kB/s | 100 kB 00:00 2026-03-09T15:11:03.119 INFO:teuthology.orchestra.run.vm06.stdout:(23/136): rbd-fuse-19.2.3-678.ge911bdeb.el9.x86 726 kB/s | 85 kB 00:00 2026-03-09T15:11:03.672 INFO:teuthology.orchestra.run.vm03.stdout:(8/136): ceph-common-19.2.3-678.ge911bdeb.el9.x 5.0 MB/s | 22 MB 00:04 2026-03-09T15:11:03.828 INFO:teuthology.orchestra.run.vm03.stdout:(9/136): ceph-selinux-19.2.3-678.ge911bdeb.el9. 161 kB/s | 25 kB 00:00 2026-03-09T15:11:03.828 INFO:teuthology.orchestra.run.vm04.stdout:(24/136): librgw2-19.2.3-678.ge911bdeb.el9.x86_ 3.0 MB/s | 5.4 MB 00:01 2026-03-09T15:11:03.990 INFO:teuthology.orchestra.run.vm04.stdout:(25/136): rbd-nbd-19.2.3-678.ge911bdeb.el9.x86_ 1.0 MB/s | 171 kB 00:00 2026-03-09T15:11:04.053 INFO:teuthology.orchestra.run.vm06.stdout:(24/136): rbd-mirror-19.2.3-678.ge911bdeb.el9.x 3.3 MB/s | 3.1 MB 00:00 2026-03-09T15:11:04.091 INFO:teuthology.orchestra.run.vm04.stdout:(26/136): rbd-mirror-19.2.3-678.ge911bdeb.el9.x 2.8 MB/s | 3.1 MB 00:01 2026-03-09T15:11:04.108 INFO:teuthology.orchestra.run.vm04.stdout:(27/136): ceph-grafana-dashboards-19.2.3-678.ge 263 kB/s | 31 kB 00:00 2026-03-09T15:11:04.171 INFO:teuthology.orchestra.run.vm06.stdout:(25/136): rbd-nbd-19.2.3-678.ge911bdeb.el9.x86_ 1.4 MB/s | 171 kB 00:00 2026-03-09T15:11:04.215 INFO:teuthology.orchestra.run.vm04.stdout:(28/136): ceph-mgr-cephadm-19.2.3-678.ge911bdeb 1.2 MB/s | 150 kB 00:00 2026-03-09T15:11:04.288 INFO:teuthology.orchestra.run.vm06.stdout:(26/136): ceph-grafana-dashboards-19.2.3-678.ge 268 kB/s | 31 kB 00:00 2026-03-09T15:11:04.406 INFO:teuthology.orchestra.run.vm06.stdout:(27/136): ceph-mgr-cephadm-19.2.3-678.ge911bdeb 1.2 MB/s | 150 kB 00:00 2026-03-09T15:11:04.445 INFO:teuthology.orchestra.run.vm03.stdout:(10/136): ceph-radosgw-19.2.3-678.ge911bdeb.el9 5.5 MB/s | 11 MB 00:01 2026-03-09T15:11:04.521 INFO:teuthology.orchestra.run.vm03.stdout:(11/136): ceph-osd-19.2.3-678.ge911bdeb.el9.x86 6.2 MB/s | 17 MB 00:02 2026-03-09T15:11:04.569 INFO:teuthology.orchestra.run.vm03.stdout:(12/136): libcephfs-devel-19.2.3-678.ge911bdeb. 273 kB/s | 34 kB 00:00 2026-03-09T15:11:04.688 INFO:teuthology.orchestra.run.vm03.stdout:(13/136): libcephsqlite-19.2.3-678.ge911bdeb.el 1.3 MB/s | 163 kB 00:00 2026-03-09T15:11:04.742 INFO:teuthology.orchestra.run.vm03.stdout:(14/136): libcephfs2-19.2.3-678.ge911bdeb.el9.x 4.4 MB/s | 1.0 MB 00:00 2026-03-09T15:11:04.806 INFO:teuthology.orchestra.run.vm03.stdout:(15/136): librados-devel-19.2.3-678.ge911bdeb.e 1.0 MB/s | 127 kB 00:00 2026-03-09T15:11:04.856 INFO:teuthology.orchestra.run.vm03.stdout:(16/136): libradosstriper1-19.2.3-678.ge911bdeb 4.3 MB/s | 503 kB 00:00 2026-03-09T15:11:04.965 INFO:teuthology.orchestra.run.vm03.stdout:(17/136): python3-ceph-argparse-19.2.3-678.ge91 415 kB/s | 45 kB 00:00 2026-03-09T15:11:05.075 INFO:teuthology.orchestra.run.vm03.stdout:(18/136): python3-ceph-common-19.2.3-678.ge911b 1.3 MB/s | 142 kB 00:00 2026-03-09T15:11:05.185 INFO:teuthology.orchestra.run.vm03.stdout:(19/136): python3-cephfs-19.2.3-678.ge911bdeb.e 1.5 MB/s | 165 kB 00:00 2026-03-09T15:11:05.300 INFO:teuthology.orchestra.run.vm03.stdout:(20/136): python3-rados-19.2.3-678.ge911bdeb.el 2.7 MB/s | 323 kB 00:00 2026-03-09T15:11:05.301 INFO:teuthology.orchestra.run.vm04.stdout:(29/136): ceph-mgr-dashboard-19.2.3-678.ge911bd 3.2 MB/s | 3.8 MB 00:01 2026-03-09T15:11:05.412 INFO:teuthology.orchestra.run.vm03.stdout:(21/136): python3-rbd-19.2.3-678.ge911bdeb.el9. 2.7 MB/s | 303 kB 00:00 2026-03-09T15:11:05.422 INFO:teuthology.orchestra.run.vm04.stdout:(30/136): ceph-mgr-modules-core-19.2.3-678.ge91 2.0 MB/s | 253 kB 00:00 2026-03-09T15:11:05.521 INFO:teuthology.orchestra.run.vm03.stdout:(22/136): python3-rgw-19.2.3-678.ge911bdeb.el9. 914 kB/s | 100 kB 00:00 2026-03-09T15:11:05.541 INFO:teuthology.orchestra.run.vm04.stdout:(31/136): ceph-mgr-rook-19.2.3-678.ge911bdeb.el 416 kB/s | 49 kB 00:00 2026-03-09T15:11:05.571 INFO:teuthology.orchestra.run.vm06.stdout:(28/136): ceph-mgr-dashboard-19.2.3-678.ge911bd 3.3 MB/s | 3.8 MB 00:01 2026-03-09T15:11:05.638 INFO:teuthology.orchestra.run.vm03.stdout:(23/136): librgw2-19.2.3-678.ge911bdeb.el9.x86_ 6.5 MB/s | 5.4 MB 00:00 2026-03-09T15:11:05.640 INFO:teuthology.orchestra.run.vm03.stdout:(24/136): rbd-fuse-19.2.3-678.ge911bdeb.el9.x86 716 kB/s | 85 kB 00:00 2026-03-09T15:11:05.659 INFO:teuthology.orchestra.run.vm04.stdout:(32/136): ceph-prometheus-alerts-19.2.3-678.ge9 141 kB/s | 17 kB 00:00 2026-03-09T15:11:05.751 INFO:teuthology.orchestra.run.vm03.stdout:(25/136): rbd-nbd-19.2.3-678.ge911bdeb.el9.x86_ 1.5 MB/s | 171 kB 00:00 2026-03-09T15:11:05.782 INFO:teuthology.orchestra.run.vm04.stdout:(33/136): ceph-volume-19.2.3-678.ge911bdeb.el9. 2.4 MB/s | 299 kB 00:00 2026-03-09T15:11:05.836 INFO:teuthology.orchestra.run.vm06.stdout:(29/136): ceph-common-19.2.3-678.ge911bdeb.el9. 1.4 MB/s | 22 MB 00:15 2026-03-09T15:11:05.859 INFO:teuthology.orchestra.run.vm03.stdout:(26/136): ceph-grafana-dashboards-19.2.3-678.ge 287 kB/s | 31 kB 00:00 2026-03-09T15:11:05.961 INFO:teuthology.orchestra.run.vm06.stdout:(30/136): ceph-mgr-modules-core-19.2.3-678.ge91 2.0 MB/s | 253 kB 00:00 2026-03-09T15:11:05.969 INFO:teuthology.orchestra.run.vm03.stdout:(27/136): ceph-mgr-cephadm-19.2.3-678.ge911bdeb 1.3 MB/s | 150 kB 00:00 2026-03-09T15:11:06.025 INFO:teuthology.orchestra.run.vm04.stdout:(34/136): cephadm-19.2.3-678.ge911bdeb.el9.noar 3.1 MB/s | 769 kB 00:00 2026-03-09T15:11:06.074 INFO:teuthology.orchestra.run.vm06.stdout:(31/136): ceph-mgr-rook-19.2.3-678.ge911bdeb.el 435 kB/s | 49 kB 00:00 2026-03-09T15:11:06.120 INFO:teuthology.orchestra.run.vm03.stdout:(28/136): rbd-mirror-19.2.3-678.ge911bdeb.el9.x 6.5 MB/s | 3.1 MB 00:00 2026-03-09T15:11:06.193 INFO:teuthology.orchestra.run.vm06.stdout:(32/136): ceph-prometheus-alerts-19.2.3-678.ge9 141 kB/s | 17 kB 00:00 2026-03-09T15:11:06.228 INFO:teuthology.orchestra.run.vm04.stdout:(35/136): cryptsetup-2.8.1-3.el9.x86_64.rpm 1.7 MB/s | 351 kB 00:00 2026-03-09T15:11:06.262 INFO:teuthology.orchestra.run.vm04.stdout:(36/136): ledmon-libs-1.1.0-3.el9.x86_64.rpm 1.2 MB/s | 40 kB 00:00 2026-03-09T15:11:06.300 INFO:teuthology.orchestra.run.vm04.stdout:(37/136): libconfig-1.7.2-9.el9.x86_64.rpm 1.8 MB/s | 72 kB 00:00 2026-03-09T15:11:06.320 INFO:teuthology.orchestra.run.vm06.stdout:(33/136): ceph-volume-19.2.3-678.ge911bdeb.el9. 2.3 MB/s | 299 kB 00:00 2026-03-09T15:11:06.368 INFO:teuthology.orchestra.run.vm04.stdout:(38/136): libgfortran-11.5.0-14.el9.x86_64.rpm 11 MB/s | 794 kB 00:00 2026-03-09T15:11:06.405 INFO:teuthology.orchestra.run.vm04.stdout:(39/136): libquadmath-11.5.0-14.el9.x86_64.rpm 5.0 MB/s | 184 kB 00:00 2026-03-09T15:11:06.436 INFO:teuthology.orchestra.run.vm04.stdout:(40/136): mailcap-2.1.49-5.el9.noarch.rpm 1.0 MB/s | 33 kB 00:00 2026-03-09T15:11:06.474 INFO:teuthology.orchestra.run.vm04.stdout:(41/136): ceph-mgr-diskprediction-local-19.2.3- 3.3 MB/s | 7.4 MB 00:02 2026-03-09T15:11:06.475 INFO:teuthology.orchestra.run.vm04.stdout:(42/136): pciutils-3.7.0-7.el9.x86_64.rpm 2.3 MB/s | 93 kB 00:00 2026-03-09T15:11:06.548 INFO:teuthology.orchestra.run.vm06.stdout:(34/136): cephadm-19.2.3-678.ge911bdeb.el9.noar 3.3 MB/s | 769 kB 00:00 2026-03-09T15:11:06.568 INFO:teuthology.orchestra.run.vm04.stdout:(43/136): python3-cryptography-36.0.1-5.el9.x86 13 MB/s | 1.2 MB 00:00 2026-03-09T15:11:06.599 INFO:teuthology.orchestra.run.vm04.stdout:(44/136): python3-ply-3.11-14.el9.noarch.rpm 3.4 MB/s | 106 kB 00:00 2026-03-09T15:11:06.627 INFO:teuthology.orchestra.run.vm03.stdout:(29/136): ceph-mgr-dashboard-19.2.3-678.ge911bd 5.8 MB/s | 3.8 MB 00:00 2026-03-09T15:11:06.630 INFO:teuthology.orchestra.run.vm04.stdout:(45/136): python3-pycparser-2.20-6.el9.noarch.r 4.3 MB/s | 135 kB 00:00 2026-03-09T15:11:06.648 INFO:teuthology.orchestra.run.vm04.stdout:(46/136): python3-cffi-1.14.5-5.el9.x86_64.rpm 1.4 MB/s | 253 kB 00:00 2026-03-09T15:11:06.662 INFO:teuthology.orchestra.run.vm04.stdout:(47/136): python3-requests-2.25.1-10.el9.noarch 3.9 MB/s | 126 kB 00:00 2026-03-09T15:11:06.680 INFO:teuthology.orchestra.run.vm04.stdout:(48/136): python3-urllib3-1.26.5-7.el9.noarch.r 6.6 MB/s | 218 kB 00:00 2026-03-09T15:11:06.693 INFO:teuthology.orchestra.run.vm04.stdout:(49/136): unzip-6.0-59.el9.x86_64.rpm 5.7 MB/s | 182 kB 00:00 2026-03-09T15:11:06.713 INFO:teuthology.orchestra.run.vm04.stdout:(50/136): zip-3.0-35.el9.x86_64.rpm 8.0 MB/s | 266 kB 00:00 2026-03-09T15:11:06.738 INFO:teuthology.orchestra.run.vm03.stdout:(30/136): ceph-mgr-modules-core-19.2.3-678.ge91 2.2 MB/s | 253 kB 00:00 2026-03-09T15:11:06.763 INFO:teuthology.orchestra.run.vm04.stdout:(51/136): flexiblas-3.0.4-9.el9.x86_64.rpm 589 kB/s | 30 kB 00:00 2026-03-09T15:11:06.788 INFO:teuthology.orchestra.run.vm04.stdout:(52/136): boost-program-options-1.75.0-13.el9.x 1.1 MB/s | 104 kB 00:00 2026-03-09T15:11:06.803 INFO:teuthology.orchestra.run.vm04.stdout:(53/136): flexiblas-openblas-openmp-3.0.4-9.el9 965 kB/s | 15 kB 00:00 2026-03-09T15:11:06.835 INFO:teuthology.orchestra.run.vm04.stdout:(54/136): libnbd-1.20.3-4.el9.x86_64.rpm 5.2 MB/s | 164 kB 00:00 2026-03-09T15:11:06.848 INFO:teuthology.orchestra.run.vm03.stdout:(31/136): ceph-mgr-rook-19.2.3-678.ge911bdeb.el 448 kB/s | 49 kB 00:00 2026-03-09T15:11:06.857 INFO:teuthology.orchestra.run.vm04.stdout:(55/136): libpmemobj-1.12.1-1.el9.x86_64.rpm 7.1 MB/s | 160 kB 00:00 2026-03-09T15:11:06.876 INFO:teuthology.orchestra.run.vm04.stdout:(56/136): librabbitmq-0.11.0-7.el9.x86_64.rpm 2.4 MB/s | 45 kB 00:00 2026-03-09T15:11:06.908 INFO:teuthology.orchestra.run.vm04.stdout:(57/136): flexiblas-netlib-3.0.4-9.el9.x86_64.r 21 MB/s | 3.0 MB 00:00 2026-03-09T15:11:06.914 INFO:teuthology.orchestra.run.vm04.stdout:(58/136): librdkafka-1.6.1-102.el9.x86_64.rpm 17 MB/s | 662 kB 00:00 2026-03-09T15:11:06.925 INFO:teuthology.orchestra.run.vm04.stdout:(59/136): libstoragemgmt-1.10.1-1.el9.x86_64.rp 14 MB/s | 246 kB 00:00 2026-03-09T15:11:06.934 INFO:teuthology.orchestra.run.vm04.stdout:(60/136): libxslt-1.1.34-12.el9.x86_64.rpm 12 MB/s | 233 kB 00:00 2026-03-09T15:11:06.944 INFO:teuthology.orchestra.run.vm04.stdout:(61/136): lttng-ust-2.12.0-6.el9.x86_64.rpm 16 MB/s | 292 kB 00:00 2026-03-09T15:11:06.951 INFO:teuthology.orchestra.run.vm04.stdout:(62/136): lua-5.4.4-4.el9.x86_64.rpm 11 MB/s | 188 kB 00:00 2026-03-09T15:11:06.957 INFO:teuthology.orchestra.run.vm03.stdout:(32/136): ceph-prometheus-alerts-19.2.3-678.ge9 155 kB/s | 17 kB 00:00 2026-03-09T15:11:06.958 INFO:teuthology.orchestra.run.vm04.stdout:(63/136): openblas-0.3.29-1.el9.x86_64.rpm 2.8 MB/s | 42 kB 00:00 2026-03-09T15:11:06.967 INFO:teuthology.orchestra.run.vm06.stdout:(35/136): cryptsetup-2.8.1-3.el9.x86_64.rpm 840 kB/s | 351 kB 00:00 2026-03-09T15:11:07.013 INFO:teuthology.orchestra.run.vm04.stdout:(64/136): protobuf-3.14.0-17.el9.x86_64.rpm 18 MB/s | 1.0 MB 00:00 2026-03-09T15:11:07.111 INFO:teuthology.orchestra.run.vm04.stdout:(65/136): openblas-openmp-0.3.29-1.el9.x86_64.r 33 MB/s | 5.3 MB 00:00 2026-03-09T15:11:07.125 INFO:teuthology.orchestra.run.vm03.stdout:(33/136): ceph-test-19.2.3-678.ge911bdeb.el9.x8 15 MB/s | 50 MB 00:03 2026-03-09T15:11:07.127 INFO:teuthology.orchestra.run.vm03.stdout:(34/136): ceph-volume-19.2.3-678.ge911bdeb.el9. 1.7 MB/s | 299 kB 00:00 2026-03-09T15:11:07.129 INFO:teuthology.orchestra.run.vm04.stdout:(66/136): python3-devel-3.9.25-3.el9.x86_64.rpm 13 MB/s | 244 kB 00:00 2026-03-09T15:11:07.147 INFO:teuthology.orchestra.run.vm04.stdout:(67/136): python3-babel-2.9.1-2.el9.noarch.rpm 45 MB/s | 6.0 MB 00:00 2026-03-09T15:11:07.149 INFO:teuthology.orchestra.run.vm04.stdout:(68/136): python3-jinja2-2.11.3-8.el9.noarch.rp 13 MB/s | 249 kB 00:00 2026-03-09T15:11:07.162 INFO:teuthology.orchestra.run.vm04.stdout:(69/136): python3-jmespath-1.0.1-1.el9.noarch.r 3.2 MB/s | 48 kB 00:00 2026-03-09T15:11:07.165 INFO:teuthology.orchestra.run.vm04.stdout:(70/136): python3-libstoragemgmt-1.10.1-1.el9.x 11 MB/s | 177 kB 00:00 2026-03-09T15:11:07.179 INFO:teuthology.orchestra.run.vm04.stdout:(71/136): python3-mako-1.1.4-6.el9.noarch.rpm 10 MB/s | 172 kB 00:00 2026-03-09T15:11:07.181 INFO:teuthology.orchestra.run.vm04.stdout:(72/136): python3-markupsafe-1.1.1-12.el9.x86_6 2.3 MB/s | 35 kB 00:00 2026-03-09T15:11:07.209 INFO:teuthology.orchestra.run.vm04.stdout:(73/136): python3-numpy-f2py-1.23.5-2.el9.x86_6 16 MB/s | 442 kB 00:00 2026-03-09T15:11:07.225 INFO:teuthology.orchestra.run.vm04.stdout:(74/136): python3-packaging-20.9-5.el9.noarch.r 4.8 MB/s | 77 kB 00:00 2026-03-09T15:11:07.266 INFO:teuthology.orchestra.run.vm06.stdout:(36/136): ledmon-libs-1.1.0-3.el9.x86_64.rpm 135 kB/s | 40 kB 00:00 2026-03-09T15:11:07.293 INFO:teuthology.orchestra.run.vm04.stdout:(75/136): python3-numpy-1.23.5-2.el9.x86_64.rpm 54 MB/s | 6.1 MB 00:00 2026-03-09T15:11:07.305 INFO:teuthology.orchestra.run.vm03.stdout:(35/136): ceph-mgr-diskprediction-local-19.2.3- 6.2 MB/s | 7.4 MB 00:01 2026-03-09T15:11:07.309 INFO:teuthology.orchestra.run.vm04.stdout:(76/136): python3-pyasn1-0.4.8-7.el9.noarch.rpm 9.4 MB/s | 157 kB 00:00 2026-03-09T15:11:07.328 INFO:teuthology.orchestra.run.vm04.stdout:(77/136): python3-pyasn1-modules-0.4.8-7.el9.no 15 MB/s | 277 kB 00:00 2026-03-09T15:11:07.334 INFO:teuthology.orchestra.run.vm04.stdout:(78/136): python3-protobuf-3.14.0-17.el9.noarch 2.4 MB/s | 267 kB 00:00 2026-03-09T15:11:07.343 INFO:teuthology.orchestra.run.vm04.stdout:(79/136): python3-requests-oauthlib-1.3.0-12.el 3.4 MB/s | 54 kB 00:00 2026-03-09T15:11:07.359 INFO:teuthology.orchestra.run.vm04.stdout:(80/136): python3-toml-0.10.2-6.el9.noarch.rpm 2.7 MB/s | 42 kB 00:00 2026-03-09T15:11:07.376 INFO:teuthology.orchestra.run.vm04.stdout:(81/136): qatlib-25.08.0-2.el9.x86_64.rpm 13 MB/s | 240 kB 00:00 2026-03-09T15:11:07.386 INFO:teuthology.orchestra.run.vm06.stdout:(37/136): libconfig-1.7.2-9.el9.x86_64.rpm 601 kB/s | 72 kB 00:00 2026-03-09T15:11:07.392 INFO:teuthology.orchestra.run.vm04.stdout:(82/136): qatlib-service-25.08.0-2.el9.x86_64.r 2.3 MB/s | 37 kB 00:00 2026-03-09T15:11:07.408 INFO:teuthology.orchestra.run.vm04.stdout:(83/136): qatzip-libs-1.3.1-1.el9.x86_64.rpm 4.2 MB/s | 66 kB 00:00 2026-03-09T15:11:07.427 INFO:teuthology.orchestra.run.vm04.stdout:(84/136): socat-1.7.4.1-8.el9.x86_64.rpm 16 MB/s | 303 kB 00:00 2026-03-09T15:11:07.443 INFO:teuthology.orchestra.run.vm04.stdout:(85/136): xmlstarlet-1.6.1-20.el9.x86_64.rpm 3.9 MB/s | 64 kB 00:00 2026-03-09T15:11:07.465 INFO:teuthology.orchestra.run.vm03.stdout:(36/136): ledmon-libs-1.1.0-3.el9.x86_64.rpm 253 kB/s | 40 kB 00:00 2026-03-09T15:11:07.506 INFO:teuthology.orchestra.run.vm03.stdout:(37/136): cryptsetup-2.8.1-3.el9.x86_64.rpm 928 kB/s | 351 kB 00:00 2026-03-09T15:11:07.513 INFO:teuthology.orchestra.run.vm06.stdout:(38/136): libgfortran-11.5.0-14.el9.x86_64.rpm 6.1 MB/s | 794 kB 00:00 2026-03-09T15:11:07.571 INFO:teuthology.orchestra.run.vm04.stdout:(86/136): lua-devel-5.4.4-4.el9.x86_64.rpm 174 kB/s | 22 kB 00:00 2026-03-09T15:11:07.574 INFO:teuthology.orchestra.run.vm03.stdout:(38/136): libconfig-1.7.2-9.el9.x86_64.rpm 662 kB/s | 72 kB 00:00 2026-03-09T15:11:07.599 INFO:teuthology.orchestra.run.vm03.stdout:(39/136): cephadm-19.2.3-678.ge911bdeb.el9.noar 1.6 MB/s | 769 kB 00:00 2026-03-09T15:11:07.618 INFO:teuthology.orchestra.run.vm06.stdout:(39/136): ceph-mgr-diskprediction-local-19.2.3- 3.6 MB/s | 7.4 MB 00:02 2026-03-09T15:11:07.619 INFO:teuthology.orchestra.run.vm06.stdout:(40/136): libquadmath-11.5.0-14.el9.x86_64.rpm 1.7 MB/s | 184 kB 00:00 2026-03-09T15:11:07.626 INFO:teuthology.orchestra.run.vm03.stdout:(40/136): libgfortran-11.5.0-14.el9.x86_64.rpm 6.5 MB/s | 794 kB 00:00 2026-03-09T15:11:07.684 INFO:teuthology.orchestra.run.vm03.stdout:(41/136): pciutils-3.7.0-7.el9.x86_64.rpm 1.6 MB/s | 93 kB 00:00 2026-03-09T15:11:07.685 INFO:teuthology.orchestra.run.vm03.stdout:(42/136): libquadmath-11.5.0-14.el9.x86_64.rpm 1.6 MB/s | 184 kB 00:00 2026-03-09T15:11:07.694 INFO:teuthology.orchestra.run.vm06.stdout:(41/136): pciutils-3.7.0-7.el9.x86_64.rpm 1.2 MB/s | 93 kB 00:00 2026-03-09T15:11:07.738 INFO:teuthology.orchestra.run.vm04.stdout:(87/136): python3-scipy-1.9.3-2.el9.x86_64.rpm 48 MB/s | 19 MB 00:00 2026-03-09T15:11:07.744 INFO:teuthology.orchestra.run.vm03.stdout:(43/136): python3-cffi-1.14.5-5.el9.x86_64.rpm 4.1 MB/s | 253 kB 00:00 2026-03-09T15:11:07.754 INFO:teuthology.orchestra.run.vm04.stdout:(88/136): abseil-cpp-20211102.0-4.el9.x86_64.rp 34 MB/s | 551 kB 00:00 2026-03-09T15:11:07.760 INFO:teuthology.orchestra.run.vm03.stdout:(44/136): mailcap-2.1.49-5.el9.noarch.rpm 207 kB/s | 33 kB 00:00 2026-03-09T15:11:07.761 INFO:teuthology.orchestra.run.vm04.stdout:(89/136): gperftools-libs-2.9.1-3.el9.x86_64.rp 48 MB/s | 308 kB 00:00 2026-03-09T15:11:07.763 INFO:teuthology.orchestra.run.vm04.stdout:(90/136): grpc-data-1.46.7-10.el9.noarch.rpm 8.6 MB/s | 19 kB 00:00 2026-03-09T15:11:07.778 INFO:teuthology.orchestra.run.vm06.stdout:(42/136): python3-cffi-1.14.5-5.el9.x86_64.rpm 3.0 MB/s | 253 kB 00:00 2026-03-09T15:11:07.782 INFO:teuthology.orchestra.run.vm06.stdout:(43/136): mailcap-2.1.49-5.el9.noarch.rpm 203 kB/s | 33 kB 00:00 2026-03-09T15:11:07.804 INFO:teuthology.orchestra.run.vm03.stdout:(45/136): python3-ply-3.11-14.el9.noarch.rpm 1.7 MB/s | 106 kB 00:00 2026-03-09T15:11:07.829 INFO:teuthology.orchestra.run.vm04.stdout:(91/136): libarrow-9.0.0-15.el9.x86_64.rpm 68 MB/s | 4.4 MB 00:00 2026-03-09T15:11:07.831 INFO:teuthology.orchestra.run.vm04.stdout:(92/136): libarrow-doc-9.0.0-15.el9.noarch.rpm 10 MB/s | 25 kB 00:00 2026-03-09T15:11:07.834 INFO:teuthology.orchestra.run.vm04.stdout:(93/136): liboath-2.6.12-1.el9.x86_64.rpm 20 MB/s | 49 kB 00:00 2026-03-09T15:11:07.837 INFO:teuthology.orchestra.run.vm04.stdout:(94/136): libunwind-1.6.2-1.el9.x86_64.rpm 20 MB/s | 67 kB 00:00 2026-03-09T15:11:07.842 INFO:teuthology.orchestra.run.vm04.stdout:(95/136): luarocks-3.9.2-5.el9.noarch.rpm 35 MB/s | 151 kB 00:00 2026-03-09T15:11:07.861 INFO:teuthology.orchestra.run.vm04.stdout:(96/136): parquet-libs-9.0.0-15.el9.x86_64.rpm 43 MB/s | 838 kB 00:00 2026-03-09T15:11:07.862 INFO:teuthology.orchestra.run.vm03.stdout:(46/136): python3-requests-2.25.1-10.el9.noarch 2.2 MB/s | 126 kB 00:00 2026-03-09T15:11:07.870 INFO:teuthology.orchestra.run.vm04.stdout:(97/136): python3-asyncssh-2.13.2-5.el9.noarch. 61 MB/s | 548 kB 00:00 2026-03-09T15:11:07.873 INFO:teuthology.orchestra.run.vm04.stdout:(98/136): python3-autocommand-2.2.2-8.el9.noarc 13 MB/s | 29 kB 00:00 2026-03-09T15:11:07.876 INFO:teuthology.orchestra.run.vm04.stdout:(99/136): python3-backports-tarfile-1.2.0-1.el9 22 MB/s | 60 kB 00:00 2026-03-09T15:11:07.878 INFO:teuthology.orchestra.run.vm04.stdout:(100/136): python3-bcrypt-3.2.2-1.el9.x86_64.rp 17 MB/s | 43 kB 00:00 2026-03-09T15:11:07.881 INFO:teuthology.orchestra.run.vm04.stdout:(101/136): python3-cachetools-4.2.4-1.el9.noarc 11 MB/s | 32 kB 00:00 2026-03-09T15:11:07.883 INFO:teuthology.orchestra.run.vm04.stdout:(102/136): python3-certifi-2023.05.07-4.el9.noa 6.6 MB/s | 14 kB 00:00 2026-03-09T15:11:07.888 INFO:teuthology.orchestra.run.vm04.stdout:(103/136): python3-cheroot-10.0.1-4.el9.noarch. 36 MB/s | 173 kB 00:00 2026-03-09T15:11:07.895 INFO:teuthology.orchestra.run.vm04.stdout:(104/136): python3-cherrypy-18.6.1-2.el9.noarch 51 MB/s | 358 kB 00:00 2026-03-09T15:11:07.901 INFO:teuthology.orchestra.run.vm04.stdout:(105/136): python3-google-auth-2.45.0-1.el9.noa 48 MB/s | 254 kB 00:00 2026-03-09T15:11:07.901 INFO:teuthology.orchestra.run.vm06.stdout:(44/136): python3-ply-3.11-14.el9.noarch.rpm 898 kB/s | 106 kB 00:00 2026-03-09T15:11:07.911 INFO:teuthology.orchestra.run.vm03.stdout:(47/136): python3-cryptography-36.0.1-5.el9.x86 5.5 MB/s | 1.2 MB 00:00 2026-03-09T15:11:07.921 INFO:teuthology.orchestra.run.vm03.stdout:(48/136): python3-urllib3-1.26.5-7.el9.noarch.r 3.6 MB/s | 218 kB 00:00 2026-03-09T15:11:07.932 INFO:teuthology.orchestra.run.vm04.stdout:(106/136): python3-grpcio-1.46.7-10.el9.x86_64. 65 MB/s | 2.0 MB 00:00 2026-03-09T15:11:07.937 INFO:teuthology.orchestra.run.vm04.stdout:(107/136): python3-grpcio-tools-1.46.7-10.el9.x 31 MB/s | 144 kB 00:00 2026-03-09T15:11:07.939 INFO:teuthology.orchestra.run.vm04.stdout:(108/136): python3-jaraco-8.2.1-3.el9.noarch.rp 5.2 MB/s | 11 kB 00:00 2026-03-09T15:11:07.942 INFO:teuthology.orchestra.run.vm04.stdout:(109/136): python3-jaraco-classes-3.2.1-5.el9.n 7.1 MB/s | 18 kB 00:00 2026-03-09T15:11:07.944 INFO:teuthology.orchestra.run.vm04.stdout:(110/136): python3-jaraco-collections-3.0.0-8.e 10 MB/s | 23 kB 00:00 2026-03-09T15:11:07.947 INFO:teuthology.orchestra.run.vm04.stdout:(111/136): python3-jaraco-context-6.0.1-3.el9.n 9.1 MB/s | 20 kB 00:00 2026-03-09T15:11:07.949 INFO:teuthology.orchestra.run.vm04.stdout:(112/136): python3-jaraco-functools-3.5.0-2.el9 8.8 MB/s | 19 kB 00:00 2026-03-09T15:11:07.951 INFO:teuthology.orchestra.run.vm04.stdout:(113/136): python3-jaraco-text-4.0.0-2.el9.noar 12 MB/s | 26 kB 00:00 2026-03-09T15:11:07.966 INFO:teuthology.orchestra.run.vm04.stdout:(114/136): python3-kubernetes-26.1.0-3.el9.noar 69 MB/s | 1.0 MB 00:00 2026-03-09T15:11:07.967 INFO:teuthology.orchestra.run.vm03.stdout:(49/136): unzip-6.0-59.el9.x86_64.rpm 3.2 MB/s | 182 kB 00:00 2026-03-09T15:11:07.970 INFO:teuthology.orchestra.run.vm04.stdout:(115/136): python3-logutils-0.3.5-21.el9.noarch 14 MB/s | 46 kB 00:00 2026-03-09T15:11:07.973 INFO:teuthology.orchestra.run.vm04.stdout:(116/136): python3-more-itertools-8.12.0-2.el9. 27 MB/s | 79 kB 00:00 2026-03-09T15:11:07.977 INFO:teuthology.orchestra.run.vm03.stdout:(50/136): python3-pycparser-2.20-6.el9.noarch.r 622 kB/s | 135 kB 00:00 2026-03-09T15:11:07.977 INFO:teuthology.orchestra.run.vm04.stdout:(117/136): python3-natsort-7.1.1-5.el9.noarch.r 13 MB/s | 58 kB 00:00 2026-03-09T15:11:07.977 INFO:teuthology.orchestra.run.vm06.stdout:(45/136): python3-pycparser-2.20-6.el9.noarch.r 1.7 MB/s | 135 kB 00:00 2026-03-09T15:11:07.981 INFO:teuthology.orchestra.run.vm03.stdout:(51/136): zip-3.0-35.el9.x86_64.rpm 4.4 MB/s | 266 kB 00:00 2026-03-09T15:11:07.984 INFO:teuthology.orchestra.run.vm04.stdout:(118/136): python3-pecan-1.4.2-3.el9.noarch.rpm 39 MB/s | 272 kB 00:00 2026-03-09T15:11:07.988 INFO:teuthology.orchestra.run.vm04.stdout:(119/136): python3-portend-3.1.0-2.el9.noarch.r 3.8 MB/s | 16 kB 00:00 2026-03-09T15:11:07.992 INFO:teuthology.orchestra.run.vm04.stdout:(120/136): python3-pyOpenSSL-21.0.0-1.el9.noarc 23 MB/s | 90 kB 00:00 2026-03-09T15:11:07.993 INFO:teuthology.orchestra.run.vm03.stdout:(52/136): boost-program-options-1.75.0-13.el9.x 3.9 MB/s | 104 kB 00:00 2026-03-09T15:11:07.995 INFO:teuthology.orchestra.run.vm04.stdout:(121/136): python3-repoze-lru-0.7-16.el9.noarch 13 MB/s | 31 kB 00:00 2026-03-09T15:11:08.000 INFO:teuthology.orchestra.run.vm04.stdout:(122/136): python3-routes-2.5.1-5.el9.noarch.rp 41 MB/s | 188 kB 00:00 2026-03-09T15:11:08.002 INFO:teuthology.orchestra.run.vm04.stdout:(123/136): python3-rsa-4.9-2.el9.noarch.rpm 21 MB/s | 59 kB 00:00 2026-03-09T15:11:08.004 INFO:teuthology.orchestra.run.vm03.stdout:(53/136): flexiblas-3.0.4-9.el9.x86_64.rpm 1.1 MB/s | 30 kB 00:00 2026-03-09T15:11:08.005 INFO:teuthology.orchestra.run.vm04.stdout:(124/136): python3-tempora-5.0.0-2.el9.noarch.r 14 MB/s | 36 kB 00:00 2026-03-09T15:11:08.016 INFO:teuthology.orchestra.run.vm03.stdout:(54/136): flexiblas-openblas-openmp-3.0.4-9.el9 653 kB/s | 15 kB 00:00 2026-03-09T15:11:08.027 INFO:teuthology.orchestra.run.vm04.stdout:(125/136): python3-typing-extensions-4.15.0-1.e 3.8 MB/s | 86 kB 00:00 2026-03-09T15:11:08.032 INFO:teuthology.orchestra.run.vm04.stdout:(126/136): python3-webob-1.8.8-2.el9.noarch.rpm 49 MB/s | 230 kB 00:00 2026-03-09T15:11:08.036 INFO:teuthology.orchestra.run.vm04.stdout:(127/136): python3-websocket-client-1.2.3-2.el9 22 MB/s | 90 kB 00:00 2026-03-09T15:11:08.046 INFO:teuthology.orchestra.run.vm04.stdout:(128/136): python3-werkzeug-2.0.3-3.el9.1.noarc 45 MB/s | 427 kB 00:00 2026-03-09T15:11:08.048 INFO:teuthology.orchestra.run.vm04.stdout:(129/136): python3-xmltodict-0.12.0-15.el9.noar 10 MB/s | 22 kB 00:00 2026-03-09T15:11:08.052 INFO:teuthology.orchestra.run.vm04.stdout:(130/136): python3-zc-lockfile-2.0-10.el9.noarc 5.0 MB/s | 20 kB 00:00 2026-03-09T15:11:08.055 INFO:teuthology.orchestra.run.vm06.stdout:(46/136): python3-requests-2.25.1-10.el9.noarch 1.6 MB/s | 126 kB 00:00 2026-03-09T15:11:08.068 INFO:teuthology.orchestra.run.vm03.stdout:(55/136): libnbd-1.20.3-4.el9.x86_64.rpm 2.5 MB/s | 164 kB 00:00 2026-03-09T15:11:08.074 INFO:teuthology.orchestra.run.vm04.stdout:(131/136): re2-20211101-20.el9.x86_64.rpm 8.7 MB/s | 191 kB 00:00 2026-03-09T15:11:08.089 INFO:teuthology.orchestra.run.vm03.stdout:(56/136): librabbitmq-0.11.0-7.el9.x86_64.rpm 2.1 MB/s | 45 kB 00:00 2026-03-09T15:11:08.096 INFO:teuthology.orchestra.run.vm04.stdout:(132/136): thrift-0.15.0-4.el9.x86_64.rpm 72 MB/s | 1.6 MB 00:00 2026-03-09T15:11:08.130 INFO:teuthology.orchestra.run.vm03.stdout:(57/136): librdkafka-1.6.1-102.el9.x86_64.rpm 16 MB/s | 662 kB 00:00 2026-03-09T15:11:08.135 INFO:teuthology.orchestra.run.vm06.stdout:(47/136): python3-urllib3-1.26.5-7.el9.noarch.r 2.7 MB/s | 218 kB 00:00 2026-03-09T15:11:08.141 INFO:teuthology.orchestra.run.vm03.stdout:(58/136): flexiblas-netlib-3.0.4-9.el9.x86_64.r 19 MB/s | 3.0 MB 00:00 2026-03-09T15:11:08.157 INFO:teuthology.orchestra.run.vm03.stdout:(59/136): libpmemobj-1.12.1-1.el9.x86_64.rpm 1.1 MB/s | 160 kB 00:00 2026-03-09T15:11:08.187 INFO:teuthology.orchestra.run.vm03.stdout:(60/136): lttng-ust-2.12.0-6.el9.x86_64.rpm 9.8 MB/s | 292 kB 00:00 2026-03-09T15:11:08.188 INFO:teuthology.orchestra.run.vm03.stdout:(61/136): libxslt-1.1.34-12.el9.x86_64.rpm 4.9 MB/s | 233 kB 00:00 2026-03-09T15:11:08.190 INFO:teuthology.orchestra.run.vm03.stdout:(62/136): libstoragemgmt-1.10.1-1.el9.x86_64.rp 4.0 MB/s | 246 kB 00:00 2026-03-09T15:11:08.202 INFO:teuthology.orchestra.run.vm03.stdout:(63/136): openblas-0.3.29-1.el9.x86_64.rpm 3.0 MB/s | 42 kB 00:00 2026-03-09T15:11:08.216 INFO:teuthology.orchestra.run.vm06.stdout:(48/136): unzip-6.0-59.el9.x86_64.rpm 2.2 MB/s | 182 kB 00:00 2026-03-09T15:11:08.222 INFO:teuthology.orchestra.run.vm03.stdout:(64/136): lua-5.4.4-4.el9.x86_64.rpm 5.3 MB/s | 188 kB 00:00 2026-03-09T15:11:08.273 INFO:teuthology.orchestra.run.vm03.stdout:(65/136): protobuf-3.14.0-17.el9.x86_64.rpm 14 MB/s | 1.0 MB 00:00 2026-03-09T15:11:08.319 INFO:teuthology.orchestra.run.vm03.stdout:(66/136): python3-devel-3.9.25-3.el9.x86_64.rpm 5.2 MB/s | 244 kB 00:00 2026-03-09T15:11:08.373 INFO:teuthology.orchestra.run.vm03.stdout:(67/136): openblas-openmp-0.3.29-1.el9.x86_64.r 29 MB/s | 5.3 MB 00:00 2026-03-09T15:11:08.374 INFO:teuthology.orchestra.run.vm03.stdout:(68/136): python3-jinja2-2.11.3-8.el9.noarch.rp 4.4 MB/s | 249 kB 00:00 2026-03-09T15:11:08.385 INFO:teuthology.orchestra.run.vm03.stdout:(69/136): python3-jmespath-1.0.1-1.el9.noarch.r 3.8 MB/s | 48 kB 00:00 2026-03-09T15:11:08.402 INFO:teuthology.orchestra.run.vm03.stdout:(70/136): python3-babel-2.9.1-2.el9.noarch.rpm 33 MB/s | 6.0 MB 00:00 2026-03-09T15:11:08.403 INFO:teuthology.orchestra.run.vm03.stdout:(71/136): python3-libstoragemgmt-1.10.1-1.el9.x 6.0 MB/s | 177 kB 00:00 2026-03-09T15:11:08.423 INFO:teuthology.orchestra.run.vm03.stdout:(72/136): python3-markupsafe-1.1.1-12.el9.x86_6 1.6 MB/s | 35 kB 00:00 2026-03-09T15:11:08.451 INFO:teuthology.orchestra.run.vm03.stdout:(73/136): python3-mako-1.1.4-6.el9.noarch.rpm 2.6 MB/s | 172 kB 00:00 2026-03-09T15:11:08.506 INFO:teuthology.orchestra.run.vm03.stdout:(74/136): python3-numpy-f2py-1.23.5-2.el9.x86_6 5.2 MB/s | 442 kB 00:00 2026-03-09T15:11:08.507 INFO:teuthology.orchestra.run.vm03.stdout:(75/136): python3-packaging-20.9-5.el9.noarch.r 1.3 MB/s | 77 kB 00:00 2026-03-09T15:11:08.553 INFO:teuthology.orchestra.run.vm03.stdout:(76/136): python3-numpy-1.23.5-2.el9.x86_64.rpm 41 MB/s | 6.1 MB 00:00 2026-03-09T15:11:08.554 INFO:teuthology.orchestra.run.vm03.stdout:(77/136): python3-protobuf-3.14.0-17.el9.noarch 5.5 MB/s | 267 kB 00:00 2026-03-09T15:11:08.555 INFO:teuthology.orchestra.run.vm03.stdout:(78/136): python3-pyasn1-0.4.8-7.el9.noarch.rpm 3.3 MB/s | 157 kB 00:00 2026-03-09T15:11:08.576 INFO:teuthology.orchestra.run.vm03.stdout:(79/136): python3-requests-oauthlib-1.3.0-12.el 2.5 MB/s | 54 kB 00:00 2026-03-09T15:11:08.585 INFO:teuthology.orchestra.run.vm06.stdout:(49/136): python3-cryptography-36.0.1-5.el9.x86 1.5 MB/s | 1.2 MB 00:00 2026-03-09T15:11:08.593 INFO:teuthology.orchestra.run.vm03.stdout:(80/136): python3-toml-0.10.2-6.el9.noarch.rpm 2.4 MB/s | 42 kB 00:00 2026-03-09T15:11:08.619 INFO:teuthology.orchestra.run.vm03.stdout:(81/136): python3-pyasn1-modules-0.4.8-7.el9.no 4.1 MB/s | 277 kB 00:00 2026-03-09T15:11:08.629 INFO:teuthology.orchestra.run.vm03.stdout:(82/136): qatlib-25.08.0-2.el9.x86_64.rpm 6.4 MB/s | 240 kB 00:00 2026-03-09T15:11:08.641 INFO:teuthology.orchestra.run.vm03.stdout:(83/136): qatlib-service-25.08.0-2.el9.x86_64.r 1.7 MB/s | 37 kB 00:00 2026-03-09T15:11:08.668 INFO:teuthology.orchestra.run.vm03.stdout:(84/136): qatzip-libs-1.3.1-1.el9.x86_64.rpm 1.7 MB/s | 66 kB 00:00 2026-03-09T15:11:08.697 INFO:teuthology.orchestra.run.vm03.stdout:(85/136): socat-1.7.4.1-8.el9.x86_64.rpm 5.3 MB/s | 303 kB 00:00 2026-03-09T15:11:08.716 INFO:teuthology.orchestra.run.vm03.stdout:(86/136): xmlstarlet-1.6.1-20.el9.x86_64.rpm 1.3 MB/s | 64 kB 00:00 2026-03-09T15:11:08.835 INFO:teuthology.orchestra.run.vm03.stdout:(87/136): python3-scipy-1.9.3-2.el9.x86_64.rpm 69 MB/s | 19 MB 00:00 2026-03-09T15:11:08.836 INFO:teuthology.orchestra.run.vm03.stdout:(88/136): lua-devel-5.4.4-4.el9.x86_64.rpm 160 kB/s | 22 kB 00:00 2026-03-09T15:11:08.855 INFO:teuthology.orchestra.run.vm04.stdout:(133/136): protobuf-compiler-3.14.0-17.el9.x86_ 672 kB/s | 862 kB 00:01 2026-03-09T15:11:08.907 INFO:teuthology.orchestra.run.vm03.stdout:(89/136): gperftools-libs-2.9.1-3.el9.x86_64.rp 4.2 MB/s | 308 kB 00:00 2026-03-09T15:11:08.910 INFO:teuthology.orchestra.run.vm03.stdout:(90/136): abseil-cpp-20211102.0-4.el9.x86_64.rp 7.2 MB/s | 551 kB 00:00 2026-03-09T15:11:08.911 INFO:teuthology.orchestra.run.vm03.stdout:(91/136): grpc-data-1.46.7-10.el9.noarch.rpm 5.9 MB/s | 19 kB 00:00 2026-03-09T15:11:08.915 INFO:teuthology.orchestra.run.vm03.stdout:(92/136): libarrow-doc-9.0.0-15.el9.noarch.rpm 6.0 MB/s | 25 kB 00:00 2026-03-09T15:11:08.919 INFO:teuthology.orchestra.run.vm03.stdout:(93/136): liboath-2.6.12-1.el9.x86_64.rpm 12 MB/s | 49 kB 00:00 2026-03-09T15:11:08.924 INFO:teuthology.orchestra.run.vm03.stdout:(94/136): libunwind-1.6.2-1.el9.x86_64.rpm 14 MB/s | 67 kB 00:00 2026-03-09T15:11:08.930 INFO:teuthology.orchestra.run.vm03.stdout:(95/136): luarocks-3.9.2-5.el9.noarch.rpm 25 MB/s | 151 kB 00:00 2026-03-09T15:11:08.955 INFO:teuthology.orchestra.run.vm03.stdout:(96/136): parquet-libs-9.0.0-15.el9.x86_64.rpm 33 MB/s | 838 kB 00:00 2026-03-09T15:11:08.980 INFO:teuthology.orchestra.run.vm03.stdout:(97/136): libarrow-9.0.0-15.el9.x86_64.rpm 63 MB/s | 4.4 MB 00:00 2026-03-09T15:11:08.982 INFO:teuthology.orchestra.run.vm03.stdout:(98/136): python3-asyncssh-2.13.2-5.el9.noarch. 20 MB/s | 548 kB 00:00 2026-03-09T15:11:08.983 INFO:teuthology.orchestra.run.vm03.stdout:(99/136): python3-autocommand-2.2.2-8.el9.noarc 10 MB/s | 29 kB 00:00 2026-03-09T15:11:08.985 INFO:teuthology.orchestra.run.vm03.stdout:(100/136): python3-backports-tarfile-1.2.0-1.el 26 MB/s | 60 kB 00:00 2026-03-09T15:11:08.985 INFO:teuthology.orchestra.run.vm03.stdout:(101/136): python3-bcrypt-3.2.2-1.el9.x86_64.rp 18 MB/s | 43 kB 00:00 2026-03-09T15:11:08.987 INFO:teuthology.orchestra.run.vm03.stdout:(102/136): python3-cachetools-4.2.4-1.el9.noarc 15 MB/s | 32 kB 00:00 2026-03-09T15:11:08.988 INFO:teuthology.orchestra.run.vm03.stdout:(103/136): python3-certifi-2023.05.07-4.el9.noa 6.7 MB/s | 14 kB 00:00 2026-03-09T15:11:08.991 INFO:teuthology.orchestra.run.vm03.stdout:(104/136): python3-cheroot-10.0.1-4.el9.noarch. 45 MB/s | 173 kB 00:00 2026-03-09T15:11:08.995 INFO:teuthology.orchestra.run.vm03.stdout:(105/136): python3-cherrypy-18.6.1-2.el9.noarch 47 MB/s | 358 kB 00:00 2026-03-09T15:11:08.997 INFO:teuthology.orchestra.run.vm03.stdout:(106/136): python3-google-auth-2.45.0-1.el9.noa 42 MB/s | 254 kB 00:00 2026-03-09T15:11:09.002 INFO:teuthology.orchestra.run.vm03.stdout:(107/136): python3-grpcio-tools-1.46.7-10.el9.x 28 MB/s | 144 kB 00:00 2026-03-09T15:11:09.006 INFO:teuthology.orchestra.run.vm03.stdout:(108/136): python3-jaraco-8.2.1-3.el9.noarch.rp 2.7 MB/s | 11 kB 00:00 2026-03-09T15:11:09.011 INFO:teuthology.orchestra.run.vm03.stdout:(109/136): python3-jaraco-classes-3.2.1-5.el9.n 3.7 MB/s | 18 kB 00:00 2026-03-09T15:11:09.016 INFO:teuthology.orchestra.run.vm03.stdout:(110/136): python3-jaraco-collections-3.0.0-8.e 4.4 MB/s | 23 kB 00:00 2026-03-09T15:11:09.024 INFO:teuthology.orchestra.run.vm03.stdout:(111/136): python3-grpcio-1.46.7-10.el9.x86_64. 72 MB/s | 2.0 MB 00:00 2026-03-09T15:11:09.024 INFO:teuthology.orchestra.run.vm03.stdout:(112/136): python3-jaraco-context-6.0.1-3.el9.n 2.5 MB/s | 20 kB 00:00 2026-03-09T15:11:09.026 INFO:teuthology.orchestra.run.vm03.stdout:(113/136): python3-jaraco-functools-3.5.0-2.el9 9.1 MB/s | 19 kB 00:00 2026-03-09T15:11:09.026 INFO:teuthology.orchestra.run.vm03.stdout:(114/136): python3-jaraco-text-4.0.0-2.el9.noar 12 MB/s | 26 kB 00:00 2026-03-09T15:11:09.032 INFO:teuthology.orchestra.run.vm03.stdout:(115/136): python3-logutils-0.3.5-21.el9.noarch 8.3 MB/s | 46 kB 00:00 2026-03-09T15:11:09.038 INFO:teuthology.orchestra.run.vm03.stdout:(116/136): python3-more-itertools-8.12.0-2.el9. 13 MB/s | 79 kB 00:00 2026-03-09T15:11:09.041 INFO:teuthology.orchestra.run.vm03.stdout:(117/136): protobuf-compiler-3.14.0-17.el9.x86_ 2.6 MB/s | 862 kB 00:00 2026-03-09T15:11:09.042 INFO:teuthology.orchestra.run.vm03.stdout:(118/136): python3-natsort-7.1.1-5.el9.noarch.r 16 MB/s | 58 kB 00:00 2026-03-09T15:11:09.046 INFO:teuthology.orchestra.run.vm03.stdout:(119/136): python3-kubernetes-26.1.0-3.el9.noar 52 MB/s | 1.0 MB 00:00 2026-03-09T15:11:09.046 INFO:teuthology.orchestra.run.vm03.stdout:(120/136): python3-portend-3.1.0-2.el9.noarch.r 3.8 MB/s | 16 kB 00:00 2026-03-09T15:11:09.049 INFO:teuthology.orchestra.run.vm03.stdout:(121/136): python3-pyOpenSSL-21.0.0-1.el9.noarc 28 MB/s | 90 kB 00:00 2026-03-09T15:11:09.050 INFO:teuthology.orchestra.run.vm03.stdout:(122/136): python3-repoze-lru-0.7-16.el9.noarch 9.7 MB/s | 31 kB 00:00 2026-03-09T15:11:09.052 INFO:teuthology.orchestra.run.vm03.stdout:(123/136): python3-pecan-1.4.2-3.el9.noarch.rpm 25 MB/s | 272 kB 00:00 2026-03-09T15:11:09.053 INFO:teuthology.orchestra.run.vm03.stdout:(124/136): python3-routes-2.5.1-5.el9.noarch.rp 48 MB/s | 188 kB 00:00 2026-03-09T15:11:09.054 INFO:teuthology.orchestra.run.vm03.stdout:(125/136): python3-rsa-4.9-2.el9.noarch.rpm 15 MB/s | 59 kB 00:00 2026-03-09T15:11:09.054 INFO:teuthology.orchestra.run.vm03.stdout:(126/136): python3-tempora-5.0.0-2.el9.noarch.r 13 MB/s | 36 kB 00:00 2026-03-09T15:11:09.056 INFO:teuthology.orchestra.run.vm03.stdout:(127/136): python3-typing-extensions-4.15.0-1.e 32 MB/s | 86 kB 00:00 2026-03-09T15:11:09.058 INFO:teuthology.orchestra.run.vm03.stdout:(128/136): python3-webob-1.8.8-2.el9.noarch.rpm 52 MB/s | 230 kB 00:00 2026-03-09T15:11:09.060 INFO:teuthology.orchestra.run.vm03.stdout:(129/136): python3-websocket-client-1.2.3-2.el9 17 MB/s | 90 kB 00:00 2026-03-09T15:11:09.063 INFO:teuthology.orchestra.run.vm03.stdout:(130/136): python3-werkzeug-2.0.3-3.el9.1.noarc 57 MB/s | 427 kB 00:00 2026-03-09T15:11:09.064 INFO:teuthology.orchestra.run.vm03.stdout:(131/136): python3-xmltodict-0.12.0-15.el9.noar 3.9 MB/s | 22 kB 00:00 2026-03-09T15:11:09.064 INFO:teuthology.orchestra.run.vm03.stdout:(132/136): python3-zc-lockfile-2.0-10.el9.noarc 4.3 MB/s | 20 kB 00:00 2026-03-09T15:11:09.068 INFO:teuthology.orchestra.run.vm03.stdout:(133/136): re2-20211101-20.el9.x86_64.rpm 47 MB/s | 191 kB 00:00 2026-03-09T15:11:09.085 INFO:teuthology.orchestra.run.vm03.stdout:(134/136): thrift-0.15.0-4.el9.x86_64.rpm 76 MB/s | 1.6 MB 00:00 2026-03-09T15:11:09.092 INFO:teuthology.orchestra.run.vm06.stdout:(50/136): zip-3.0-35.el9.x86_64.rpm 304 kB/s | 266 kB 00:00 2026-03-09T15:11:09.201 INFO:teuthology.orchestra.run.vm06.stdout:(51/136): flexiblas-3.0.4-9.el9.x86_64.rpm 273 kB/s | 30 kB 00:00 2026-03-09T15:11:09.564 INFO:teuthology.orchestra.run.vm06.stdout:(52/136): boost-program-options-1.75.0-13.el9.x 106 kB/s | 104 kB 00:00 2026-03-09T15:11:09.650 INFO:teuthology.orchestra.run.vm06.stdout:(53/136): flexiblas-openblas-openmp-3.0.4-9.el9 173 kB/s | 15 kB 00:00 2026-03-09T15:11:09.887 INFO:teuthology.orchestra.run.vm03.stdout:(135/136): librados2-19.2.3-678.ge911bdeb.el9.x 4.2 MB/s | 3.4 MB 00:00 2026-03-09T15:11:10.063 INFO:teuthology.orchestra.run.vm06.stdout:(54/136): libnbd-1.20.3-4.el9.x86_64.rpm 397 kB/s | 164 kB 00:00 2026-03-09T15:11:10.267 INFO:teuthology.orchestra.run.vm04.stdout:(134/136): librbd1-19.2.3-678.ge911bdeb.el9.x86 2.2 MB/s | 3.2 MB 00:01 2026-03-09T15:11:10.270 INFO:teuthology.orchestra.run.vm06.stdout:(55/136): libpmemobj-1.12.1-1.el9.x86_64.rpm 773 kB/s | 160 kB 00:00 2026-03-09T15:11:10.381 INFO:teuthology.orchestra.run.vm06.stdout:(56/136): librabbitmq-0.11.0-7.el9.x86_64.rpm 409 kB/s | 45 kB 00:00 2026-03-09T15:11:10.586 INFO:teuthology.orchestra.run.vm03.stdout:(136/136): librbd1-19.2.3-678.ge911bdeb.el9.x86 2.1 MB/s | 3.2 MB 00:01 2026-03-09T15:11:10.587 INFO:teuthology.orchestra.run.vm03.stdout:-------------------------------------------------------------------------------- 2026-03-09T15:11:10.588 INFO:teuthology.orchestra.run.vm03.stdout:Total 17 MB/s | 210 MB 00:12 2026-03-09T15:11:10.921 INFO:teuthology.orchestra.run.vm06.stdout:(57/136): flexiblas-netlib-3.0.4-9.el9.x86_64.r 1.7 MB/s | 3.0 MB 00:01 2026-03-09T15:11:10.954 INFO:teuthology.orchestra.run.vm06.stdout:(58/136): librdkafka-1.6.1-102.el9.x86_64.rpm 1.1 MB/s | 662 kB 00:00 2026-03-09T15:11:11.004 INFO:teuthology.orchestra.run.vm06.stdout:(59/136): libstoragemgmt-1.10.1-1.el9.x86_64.rp 2.9 MB/s | 246 kB 00:00 2026-03-09T15:11:11.068 INFO:teuthology.orchestra.run.vm03.stdout:Running transaction check 2026-03-09T15:11:11.090 INFO:teuthology.orchestra.run.vm06.stdout:(60/136): lttng-ust-2.12.0-6.el9.x86_64.rpm 3.3 MB/s | 292 kB 00:00 2026-03-09T15:11:11.111 INFO:teuthology.orchestra.run.vm06.stdout:(61/136): libxslt-1.1.34-12.el9.x86_64.rpm 1.5 MB/s | 233 kB 00:00 2026-03-09T15:11:11.114 INFO:teuthology.orchestra.run.vm03.stdout:Transaction check succeeded. 2026-03-09T15:11:11.114 INFO:teuthology.orchestra.run.vm03.stdout:Running transaction test 2026-03-09T15:11:11.153 INFO:teuthology.orchestra.run.vm06.stdout:(62/136): lua-5.4.4-4.el9.x86_64.rpm 2.9 MB/s | 188 kB 00:00 2026-03-09T15:11:11.162 INFO:teuthology.orchestra.run.vm06.stdout:(63/136): openblas-0.3.29-1.el9.x86_64.rpm 824 kB/s | 42 kB 00:00 2026-03-09T15:11:11.771 INFO:teuthology.orchestra.run.vm06.stdout:(64/136): protobuf-3.14.0-17.el9.x86_64.rpm 1.6 MB/s | 1.0 MB 00:00 2026-03-09T15:11:11.927 INFO:teuthology.orchestra.run.vm03.stdout:Transaction test succeeded. 2026-03-09T15:11:11.928 INFO:teuthology.orchestra.run.vm03.stdout:Running transaction 2026-03-09T15:11:12.094 INFO:teuthology.orchestra.run.vm06.stdout:(65/136): openblas-openmp-0.3.29-1.el9.x86_64.r 5.6 MB/s | 5.3 MB 00:00 2026-03-09T15:11:12.178 INFO:teuthology.orchestra.run.vm06.stdout:(66/136): python3-devel-3.9.25-3.el9.x86_64.rpm 2.8 MB/s | 244 kB 00:00 2026-03-09T15:11:12.259 INFO:teuthology.orchestra.run.vm06.stdout:(67/136): python3-jinja2-2.11.3-8.el9.noarch.rp 3.0 MB/s | 249 kB 00:00 2026-03-09T15:11:12.318 INFO:teuthology.orchestra.run.vm06.stdout:(68/136): python3-jmespath-1.0.1-1.el9.noarch.r 810 kB/s | 48 kB 00:00 2026-03-09T15:11:12.415 INFO:teuthology.orchestra.run.vm06.stdout:(69/136): python3-libstoragemgmt-1.10.1-1.el9.x 1.8 MB/s | 177 kB 00:00 2026-03-09T15:11:12.482 INFO:teuthology.orchestra.run.vm06.stdout:(70/136): python3-mako-1.1.4-6.el9.noarch.rpm 2.5 MB/s | 172 kB 00:00 2026-03-09T15:11:12.526 INFO:teuthology.orchestra.run.vm06.stdout:(71/136): python3-markupsafe-1.1.1-12.el9.x86_6 799 kB/s | 35 kB 00:00 2026-03-09T15:11:12.806 INFO:teuthology.orchestra.run.vm03.stdout: Preparing : 1/1 2026-03-09T15:11:12.828 INFO:teuthology.orchestra.run.vm03.stdout: Installing : python3-more-itertools-8.12.0-2.el9.noarch 1/138 2026-03-09T15:11:12.839 INFO:teuthology.orchestra.run.vm03.stdout: Installing : thrift-0.15.0-4.el9.x86_64 2/138 2026-03-09T15:11:13.005 INFO:teuthology.orchestra.run.vm03.stdout: Installing : lttng-ust-2.12.0-6.el9.x86_64 3/138 2026-03-09T15:11:13.007 INFO:teuthology.orchestra.run.vm03.stdout: Upgrading : librados2-2:19.2.3-678.ge911bdeb.el9.x86_64 4/138 2026-03-09T15:11:13.069 INFO:teuthology.orchestra.run.vm03.stdout: Running scriptlet: librados2-2:19.2.3-678.ge911bdeb.el9.x86_64 4/138 2026-03-09T15:11:13.070 INFO:teuthology.orchestra.run.vm03.stdout: Installing : libcephfs2-2:19.2.3-678.ge911bdeb.el9.x86_64 5/138 2026-03-09T15:11:13.100 INFO:teuthology.orchestra.run.vm03.stdout: Running scriptlet: libcephfs2-2:19.2.3-678.ge911bdeb.el9.x86_64 5/138 2026-03-09T15:11:13.108 INFO:teuthology.orchestra.run.vm03.stdout: Installing : python3-rados-2:19.2.3-678.ge911bdeb.el9.x86_64 6/138 2026-03-09T15:11:13.111 INFO:teuthology.orchestra.run.vm03.stdout: Installing : librdkafka-1.6.1-102.el9.x86_64 7/138 2026-03-09T15:11:13.114 INFO:teuthology.orchestra.run.vm03.stdout: Installing : librabbitmq-0.11.0-7.el9.x86_64 8/138 2026-03-09T15:11:13.119 INFO:teuthology.orchestra.run.vm03.stdout: Installing : python3-jaraco-8.2.1-3.el9.noarch 9/138 2026-03-09T15:11:13.128 INFO:teuthology.orchestra.run.vm03.stdout: Installing : libnbd-1.20.3-4.el9.x86_64 10/138 2026-03-09T15:11:13.130 INFO:teuthology.orchestra.run.vm03.stdout: Installing : libcephsqlite-2:19.2.3-678.ge911bdeb.el9.x86_64 11/138 2026-03-09T15:11:13.165 INFO:teuthology.orchestra.run.vm03.stdout: Running scriptlet: libcephsqlite-2:19.2.3-678.ge911bdeb.el9.x86_64 11/138 2026-03-09T15:11:13.167 INFO:teuthology.orchestra.run.vm03.stdout: Installing : libradosstriper1-2:19.2.3-678.ge911bdeb.el9.x86_ 12/138 2026-03-09T15:11:13.182 INFO:teuthology.orchestra.run.vm03.stdout: Running scriptlet: libradosstriper1-2:19.2.3-678.ge911bdeb.el9.x86_ 12/138 2026-03-09T15:11:13.214 INFO:teuthology.orchestra.run.vm03.stdout: Installing : re2-1:20211101-20.el9.x86_64 13/138 2026-03-09T15:11:13.249 INFO:teuthology.orchestra.run.vm03.stdout: Installing : libarrow-9.0.0-15.el9.x86_64 14/138 2026-03-09T15:11:13.255 INFO:teuthology.orchestra.run.vm03.stdout: Installing : python3-werkzeug-2.0.3-3.el9.1.noarch 15/138 2026-03-09T15:11:13.278 INFO:teuthology.orchestra.run.vm03.stdout: Installing : liboath-2.6.12-1.el9.x86_64 16/138 2026-03-09T15:11:13.292 INFO:teuthology.orchestra.run.vm03.stdout: Installing : python3-pyasn1-0.4.8-7.el9.noarch 17/138 2026-03-09T15:11:13.299 INFO:teuthology.orchestra.run.vm03.stdout: Installing : python3-packaging-20.9-5.el9.noarch 18/138 2026-03-09T15:11:13.309 INFO:teuthology.orchestra.run.vm03.stdout: Installing : python3-markupsafe-1.1.1-12.el9.x86_64 19/138 2026-03-09T15:11:13.315 INFO:teuthology.orchestra.run.vm03.stdout: Installing : protobuf-3.14.0-17.el9.x86_64 20/138 2026-03-09T15:11:13.319 INFO:teuthology.orchestra.run.vm03.stdout: Installing : lua-5.4.4-4.el9.x86_64 21/138 2026-03-09T15:11:13.324 INFO:teuthology.orchestra.run.vm03.stdout: Installing : flexiblas-3.0.4-9.el9.x86_64 22/138 2026-03-09T15:11:13.339 INFO:teuthology.orchestra.run.vm06.stdout:(72/136): python3-numpy-1.23.5-2.el9.x86_64.rpm 7.5 MB/s | 6.1 MB 00:00 2026-03-09T15:11:13.350 INFO:teuthology.orchestra.run.vm03.stdout: Installing : unzip-6.0-59.el9.x86_64 23/138 2026-03-09T15:11:13.369 INFO:teuthology.orchestra.run.vm03.stdout: Installing : python3-urllib3-1.26.5-7.el9.noarch 24/138 2026-03-09T15:11:13.373 INFO:teuthology.orchestra.run.vm03.stdout: Installing : python3-requests-2.25.1-10.el9.noarch 25/138 2026-03-09T15:11:13.380 INFO:teuthology.orchestra.run.vm03.stdout: Installing : libquadmath-11.5.0-14.el9.x86_64 26/138 2026-03-09T15:11:13.382 INFO:teuthology.orchestra.run.vm03.stdout: Installing : libgfortran-11.5.0-14.el9.x86_64 27/138 2026-03-09T15:11:13.415 INFO:teuthology.orchestra.run.vm03.stdout: Installing : ledmon-libs-1.1.0-3.el9.x86_64 28/138 2026-03-09T15:11:13.422 INFO:teuthology.orchestra.run.vm06.stdout:(73/136): python3-numpy-f2py-1.23.5-2.el9.x86_6 5.3 MB/s | 442 kB 00:00 2026-03-09T15:11:13.422 INFO:teuthology.orchestra.run.vm03.stdout: Installing : python3-ceph-common-2:19.2.3-678.ge911bdeb.el9.x 29/138 2026-03-09T15:11:13.434 INFO:teuthology.orchestra.run.vm03.stdout: Installing : python3-ceph-argparse-2:19.2.3-678.ge911bdeb.el9 30/138 2026-03-09T15:11:13.449 INFO:teuthology.orchestra.run.vm03.stdout: Installing : python3-cephfs-2:19.2.3-678.ge911bdeb.el9.x86_64 31/138 2026-03-09T15:11:13.454 INFO:teuthology.orchestra.run.vm06.stdout:(74/136): python3-packaging-20.9-5.el9.noarch.r 2.4 MB/s | 77 kB 00:00 2026-03-09T15:11:13.457 INFO:teuthology.orchestra.run.vm03.stdout: Installing : python3-requests-oauthlib-1.3.0-12.el9.noarch 32/138 2026-03-09T15:11:13.487 INFO:teuthology.orchestra.run.vm03.stdout: Installing : zip-3.0-35.el9.x86_64 33/138 2026-03-09T15:11:13.493 INFO:teuthology.orchestra.run.vm03.stdout: Installing : luarocks-3.9.2-5.el9.noarch 34/138 2026-03-09T15:11:13.502 INFO:teuthology.orchestra.run.vm03.stdout: Installing : lua-devel-5.4.4-4.el9.x86_64 35/138 2026-03-09T15:11:13.504 INFO:teuthology.orchestra.run.vm06.stdout:(75/136): python3-protobuf-3.14.0-17.el9.noarch 5.3 MB/s | 267 kB 00:00 2026-03-09T15:11:13.533 INFO:teuthology.orchestra.run.vm03.stdout: Installing : protobuf-compiler-3.14.0-17.el9.x86_64 36/138 2026-03-09T15:11:13.571 INFO:teuthology.orchestra.run.vm06.stdout:(76/136): python3-pyasn1-0.4.8-7.el9.noarch.rpm 2.3 MB/s | 157 kB 00:00 2026-03-09T15:11:13.591 INFO:teuthology.orchestra.run.vm03.stdout: Installing : python3-mako-1.1.4-6.el9.noarch 37/138 2026-03-09T15:11:13.608 INFO:teuthology.orchestra.run.vm03.stdout: Installing : python3-pyasn1-modules-0.4.8-7.el9.noarch 38/138 2026-03-09T15:11:13.615 INFO:teuthology.orchestra.run.vm03.stdout: Installing : python3-rsa-4.9-2.el9.noarch 39/138 2026-03-09T15:11:13.624 INFO:teuthology.orchestra.run.vm03.stdout: Installing : python3-jaraco-classes-3.2.1-5.el9.noarch 40/138 2026-03-09T15:11:13.630 INFO:teuthology.orchestra.run.vm06.stdout:(77/136): python3-pyasn1-modules-0.4.8-7.el9.no 4.6 MB/s | 277 kB 00:00 2026-03-09T15:11:13.630 INFO:teuthology.orchestra.run.vm03.stdout: Installing : librados-devel-2:19.2.3-678.ge911bdeb.el9.x86_64 41/138 2026-03-09T15:11:13.634 INFO:teuthology.orchestra.run.vm03.stdout: Installing : python3-zc-lockfile-2.0-10.el9.noarch 42/138 2026-03-09T15:11:13.651 INFO:teuthology.orchestra.run.vm03.stdout: Installing : python3-xmltodict-0.12.0-15.el9.noarch 43/138 2026-03-09T15:11:13.676 INFO:teuthology.orchestra.run.vm03.stdout: Installing : python3-websocket-client-1.2.3-2.el9.noarch 44/138 2026-03-09T15:11:13.682 INFO:teuthology.orchestra.run.vm03.stdout: Installing : python3-webob-1.8.8-2.el9.noarch 45/138 2026-03-09T15:11:13.688 INFO:teuthology.orchestra.run.vm03.stdout: Installing : python3-typing-extensions-4.15.0-1.el9.noarch 46/138 2026-03-09T15:11:13.693 INFO:teuthology.orchestra.run.vm06.stdout:(78/136): python3-requests-oauthlib-1.3.0-12.el 847 kB/s | 54 kB 00:00 2026-03-09T15:11:13.702 INFO:teuthology.orchestra.run.vm03.stdout: Installing : python3-repoze-lru-0.7-16.el9.noarch 47/138 2026-03-09T15:11:13.713 INFO:teuthology.orchestra.run.vm03.stdout: Installing : python3-routes-2.5.1-5.el9.noarch 48/138 2026-03-09T15:11:13.724 INFO:teuthology.orchestra.run.vm03.stdout: Installing : python3-natsort-7.1.1-5.el9.noarch 49/138 2026-03-09T15:11:13.785 INFO:teuthology.orchestra.run.vm03.stdout: Installing : python3-logutils-0.3.5-21.el9.noarch 50/138 2026-03-09T15:11:13.793 INFO:teuthology.orchestra.run.vm03.stdout: Installing : python3-pecan-1.4.2-3.el9.noarch 51/138 2026-03-09T15:11:13.803 INFO:teuthology.orchestra.run.vm03.stdout: Installing : python3-certifi-2023.05.07-4.el9.noarch 52/138 2026-03-09T15:11:13.853 INFO:teuthology.orchestra.run.vm03.stdout: Installing : python3-cachetools-4.2.4-1.el9.noarch 53/138 2026-03-09T15:11:14.001 INFO:teuthology.orchestra.run.vm06.stdout:(79/136): python3-babel-2.9.1-2.el9.noarch.rpm 2.7 MB/s | 6.0 MB 00:02 2026-03-09T15:11:14.048 INFO:teuthology.orchestra.run.vm06.stdout:(80/136): python3-toml-0.10.2-6.el9.noarch.rpm 890 kB/s | 42 kB 00:00 2026-03-09T15:11:14.136 INFO:teuthology.orchestra.run.vm06.stdout:(81/136): qatlib-25.08.0-2.el9.x86_64.rpm 2.7 MB/s | 240 kB 00:00 2026-03-09T15:11:14.189 INFO:teuthology.orchestra.run.vm06.stdout:(82/136): qatlib-service-25.08.0-2.el9.x86_64.r 692 kB/s | 37 kB 00:00 2026-03-09T15:11:14.231 INFO:teuthology.orchestra.run.vm03.stdout: Installing : python3-google-auth-1:2.45.0-1.el9.noarch 54/138 2026-03-09T15:11:14.244 INFO:teuthology.orchestra.run.vm06.stdout:(83/136): qatzip-libs-1.3.1-1.el9.x86_64.rpm 1.2 MB/s | 66 kB 00:00 2026-03-09T15:11:14.248 INFO:teuthology.orchestra.run.vm03.stdout: Installing : python3-kubernetes-1:26.1.0-3.el9.noarch 55/138 2026-03-09T15:11:14.254 INFO:teuthology.orchestra.run.vm03.stdout: Installing : python3-backports-tarfile-1.2.0-1.el9.noarch 56/138 2026-03-09T15:11:14.262 INFO:teuthology.orchestra.run.vm03.stdout: Installing : python3-jaraco-context-6.0.1-3.el9.noarch 57/138 2026-03-09T15:11:14.267 INFO:teuthology.orchestra.run.vm03.stdout: Installing : python3-autocommand-2.2.2-8.el9.noarch 58/138 2026-03-09T15:11:14.274 INFO:teuthology.orchestra.run.vm03.stdout: Installing : libunwind-1.6.2-1.el9.x86_64 59/138 2026-03-09T15:11:14.278 INFO:teuthology.orchestra.run.vm03.stdout: Installing : gperftools-libs-2.9.1-3.el9.x86_64 60/138 2026-03-09T15:11:14.280 INFO:teuthology.orchestra.run.vm03.stdout: Installing : libarrow-doc-9.0.0-15.el9.noarch 61/138 2026-03-09T15:11:14.310 INFO:teuthology.orchestra.run.vm03.stdout: Installing : grpc-data-1.46.7-10.el9.noarch 62/138 2026-03-09T15:11:14.358 INFO:teuthology.orchestra.run.vm06.stdout:(84/136): socat-1.7.4.1-8.el9.x86_64.rpm 2.6 MB/s | 303 kB 00:00 2026-03-09T15:11:14.361 INFO:teuthology.orchestra.run.vm03.stdout: Installing : abseil-cpp-20211102.0-4.el9.x86_64 63/138 2026-03-09T15:11:14.374 INFO:teuthology.orchestra.run.vm03.stdout: Installing : python3-grpcio-1.46.7-10.el9.x86_64 64/138 2026-03-09T15:11:14.382 INFO:teuthology.orchestra.run.vm03.stdout: Installing : socat-1.7.4.1-8.el9.x86_64 65/138 2026-03-09T15:11:14.388 INFO:teuthology.orchestra.run.vm03.stdout: Installing : python3-toml-0.10.2-6.el9.noarch 66/138 2026-03-09T15:11:14.395 INFO:teuthology.orchestra.run.vm03.stdout: Installing : python3-jaraco-functools-3.5.0-2.el9.noarch 67/138 2026-03-09T15:11:14.400 INFO:teuthology.orchestra.run.vm03.stdout: Installing : python3-jaraco-text-4.0.0-2.el9.noarch 68/138 2026-03-09T15:11:14.409 INFO:teuthology.orchestra.run.vm03.stdout: Installing : python3-jaraco-collections-3.0.0-8.el9.noarch 69/138 2026-03-09T15:11:14.412 INFO:teuthology.orchestra.run.vm06.stdout:(85/136): xmlstarlet-1.6.1-20.el9.x86_64.rpm 1.2 MB/s | 64 kB 00:00 2026-03-09T15:11:14.415 INFO:teuthology.orchestra.run.vm03.stdout: Installing : python3-tempora-5.0.0-2.el9.noarch 70/138 2026-03-09T15:11:14.448 INFO:teuthology.orchestra.run.vm03.stdout: Installing : python3-portend-3.1.0-2.el9.noarch 71/138 2026-03-09T15:11:14.461 INFO:teuthology.orchestra.run.vm03.stdout: Installing : python3-protobuf-3.14.0-17.el9.noarch 72/138 2026-03-09T15:11:14.511 INFO:teuthology.orchestra.run.vm03.stdout: Installing : python3-grpcio-tools-1.46.7-10.el9.x86_64 73/138 2026-03-09T15:11:14.576 INFO:teuthology.orchestra.run.vm06.stdout:(86/136): lua-devel-5.4.4-4.el9.x86_64.rpm 136 kB/s | 22 kB 00:00 2026-03-09T15:11:14.774 INFO:teuthology.orchestra.run.vm03.stdout: Installing : python3-devel-3.9.25-3.el9.x86_64 74/138 2026-03-09T15:11:14.806 INFO:teuthology.orchestra.run.vm03.stdout: Installing : python3-babel-2.9.1-2.el9.noarch 75/138 2026-03-09T15:11:14.812 INFO:teuthology.orchestra.run.vm03.stdout: Installing : python3-jinja2-2.11.3-8.el9.noarch 76/138 2026-03-09T15:11:14.850 INFO:teuthology.orchestra.run.vm06.stdout:(87/136): protobuf-compiler-3.14.0-17.el9.x86_6 3.1 MB/s | 862 kB 00:00 2026-03-09T15:11:14.869 INFO:teuthology.orchestra.run.vm06.stdout:(88/136): abseil-cpp-20211102.0-4.el9.x86_64.rp 29 MB/s | 551 kB 00:00 2026-03-09T15:11:14.871 INFO:teuthology.orchestra.run.vm03.stdout: Installing : openblas-0.3.29-1.el9.x86_64 77/138 2026-03-09T15:11:14.874 INFO:teuthology.orchestra.run.vm03.stdout: Installing : openblas-openmp-0.3.29-1.el9.x86_64 78/138 2026-03-09T15:11:14.879 INFO:teuthology.orchestra.run.vm06.stdout:(89/136): gperftools-libs-2.9.1-3.el9.x86_64.rp 29 MB/s | 308 kB 00:00 2026-03-09T15:11:14.882 INFO:teuthology.orchestra.run.vm06.stdout:(90/136): grpc-data-1.46.7-10.el9.noarch.rpm 9.6 MB/s | 19 kB 00:00 2026-03-09T15:11:14.899 INFO:teuthology.orchestra.run.vm03.stdout: Installing : flexiblas-openblas-openmp-3.0.4-9.el9.x86_64 79/138 2026-03-09T15:11:14.943 INFO:teuthology.orchestra.run.vm06.stdout:(91/136): libarrow-9.0.0-15.el9.x86_64.rpm 73 MB/s | 4.4 MB 00:00 2026-03-09T15:11:14.945 INFO:teuthology.orchestra.run.vm06.stdout:(92/136): libarrow-doc-9.0.0-15.el9.noarch.rpm 12 MB/s | 25 kB 00:00 2026-03-09T15:11:14.947 INFO:teuthology.orchestra.run.vm06.stdout:(93/136): liboath-2.6.12-1.el9.x86_64.rpm 21 MB/s | 49 kB 00:00 2026-03-09T15:11:14.950 INFO:teuthology.orchestra.run.vm06.stdout:(94/136): libunwind-1.6.2-1.el9.x86_64.rpm 25 MB/s | 67 kB 00:00 2026-03-09T15:11:14.954 INFO:teuthology.orchestra.run.vm06.stdout:(95/136): luarocks-3.9.2-5.el9.noarch.rpm 40 MB/s | 151 kB 00:00 2026-03-09T15:11:14.965 INFO:teuthology.orchestra.run.vm06.stdout:(96/136): parquet-libs-9.0.0-15.el9.x86_64.rpm 75 MB/s | 838 kB 00:00 2026-03-09T15:11:14.975 INFO:teuthology.orchestra.run.vm06.stdout:(97/136): python3-asyncssh-2.13.2-5.el9.noarch. 56 MB/s | 548 kB 00:00 2026-03-09T15:11:14.977 INFO:teuthology.orchestra.run.vm06.stdout:(98/136): python3-autocommand-2.2.2-8.el9.noarc 14 MB/s | 29 kB 00:00 2026-03-09T15:11:14.980 INFO:teuthology.orchestra.run.vm06.stdout:(99/136): python3-backports-tarfile-1.2.0-1.el9 21 MB/s | 60 kB 00:00 2026-03-09T15:11:14.983 INFO:teuthology.orchestra.run.vm06.stdout:(100/136): python3-bcrypt-3.2.2-1.el9.x86_64.rp 19 MB/s | 43 kB 00:00 2026-03-09T15:11:14.985 INFO:teuthology.orchestra.run.vm06.stdout:(101/136): python3-cachetools-4.2.4-1.el9.noarc 15 MB/s | 32 kB 00:00 2026-03-09T15:11:14.987 INFO:teuthology.orchestra.run.vm06.stdout:(102/136): python3-certifi-2023.05.07-4.el9.noa 7.1 MB/s | 14 kB 00:00 2026-03-09T15:11:14.992 INFO:teuthology.orchestra.run.vm06.stdout:(103/136): python3-cheroot-10.0.1-4.el9.noarch. 39 MB/s | 173 kB 00:00 2026-03-09T15:11:14.998 INFO:teuthology.orchestra.run.vm06.stdout:(104/136): python3-cherrypy-18.6.1-2.el9.noarch 62 MB/s | 358 kB 00:00 2026-03-09T15:11:15.002 INFO:teuthology.orchestra.run.vm06.stdout:(105/136): python3-google-auth-2.45.0-1.el9.noa 56 MB/s | 254 kB 00:00 2026-03-09T15:11:15.031 INFO:teuthology.orchestra.run.vm06.stdout:(106/136): python3-grpcio-1.46.7-10.el9.x86_64. 71 MB/s | 2.0 MB 00:00 2026-03-09T15:11:15.035 INFO:teuthology.orchestra.run.vm06.stdout:(107/136): python3-grpcio-tools-1.46.7-10.el9.x 36 MB/s | 144 kB 00:00 2026-03-09T15:11:15.037 INFO:teuthology.orchestra.run.vm06.stdout:(108/136): python3-jaraco-8.2.1-3.el9.noarch.rp 5.9 MB/s | 11 kB 00:00 2026-03-09T15:11:15.040 INFO:teuthology.orchestra.run.vm06.stdout:(109/136): python3-jaraco-classes-3.2.1-5.el9.n 6.7 MB/s | 18 kB 00:00 2026-03-09T15:11:15.042 INFO:teuthology.orchestra.run.vm06.stdout:(110/136): python3-jaraco-collections-3.0.0-8.e 11 MB/s | 23 kB 00:00 2026-03-09T15:11:15.045 INFO:teuthology.orchestra.run.vm06.stdout:(111/136): python3-jaraco-context-6.0.1-3.el9.n 9.2 MB/s | 20 kB 00:00 2026-03-09T15:11:15.047 INFO:teuthology.orchestra.run.vm06.stdout:(112/136): python3-jaraco-functools-3.5.0-2.el9 8.7 MB/s | 19 kB 00:00 2026-03-09T15:11:15.049 INFO:teuthology.orchestra.run.vm06.stdout:(113/136): python3-jaraco-text-4.0.0-2.el9.noar 11 MB/s | 26 kB 00:00 2026-03-09T15:11:15.063 INFO:teuthology.orchestra.run.vm06.stdout:(114/136): python3-kubernetes-26.1.0-3.el9.noar 77 MB/s | 1.0 MB 00:00 2026-03-09T15:11:15.070 INFO:teuthology.orchestra.run.vm06.stdout:(115/136): python3-logutils-0.3.5-21.el9.noarch 6.6 MB/s | 46 kB 00:00 2026-03-09T15:11:15.073 INFO:teuthology.orchestra.run.vm06.stdout:(116/136): python3-more-itertools-8.12.0-2.el9. 26 MB/s | 79 kB 00:00 2026-03-09T15:11:15.077 INFO:teuthology.orchestra.run.vm06.stdout:(117/136): python3-natsort-7.1.1-5.el9.noarch.r 19 MB/s | 58 kB 00:00 2026-03-09T15:11:15.084 INFO:teuthology.orchestra.run.vm06.stdout:(118/136): python3-pecan-1.4.2-3.el9.noarch.rpm 35 MB/s | 272 kB 00:00 2026-03-09T15:11:15.089 INFO:teuthology.orchestra.run.vm06.stdout:(119/136): python3-portend-3.1.0-2.el9.noarch.r 3.8 MB/s | 16 kB 00:00 2026-03-09T15:11:15.093 INFO:teuthology.orchestra.run.vm06.stdout:(120/136): python3-pyOpenSSL-21.0.0-1.el9.noarc 24 MB/s | 90 kB 00:00 2026-03-09T15:11:15.095 INFO:teuthology.orchestra.run.vm06.stdout:(121/136): python3-repoze-lru-0.7-16.el9.noarch 13 MB/s | 31 kB 00:00 2026-03-09T15:11:15.099 INFO:teuthology.orchestra.run.vm06.stdout:(122/136): python3-routes-2.5.1-5.el9.noarch.rp 48 MB/s | 188 kB 00:00 2026-03-09T15:11:15.102 INFO:teuthology.orchestra.run.vm06.stdout:(123/136): python3-rsa-4.9-2.el9.noarch.rpm 24 MB/s | 59 kB 00:00 2026-03-09T15:11:15.104 INFO:teuthology.orchestra.run.vm06.stdout:(124/136): python3-tempora-5.0.0-2.el9.noarch.r 15 MB/s | 36 kB 00:00 2026-03-09T15:11:15.107 INFO:teuthology.orchestra.run.vm06.stdout:(125/136): python3-typing-extensions-4.15.0-1.e 32 MB/s | 86 kB 00:00 2026-03-09T15:11:15.112 INFO:teuthology.orchestra.run.vm06.stdout:(126/136): python3-webob-1.8.8-2.el9.noarch.rpm 51 MB/s | 230 kB 00:00 2026-03-09T15:11:15.115 INFO:teuthology.orchestra.run.vm06.stdout:(127/136): python3-websocket-client-1.2.3-2.el9 32 MB/s | 90 kB 00:00 2026-03-09T15:11:15.124 INFO:teuthology.orchestra.run.vm06.stdout:(128/136): python3-werkzeug-2.0.3-3.el9.1.noarc 47 MB/s | 427 kB 00:00 2026-03-09T15:11:15.126 INFO:teuthology.orchestra.run.vm06.stdout:(129/136): python3-xmltodict-0.12.0-15.el9.noar 11 MB/s | 22 kB 00:00 2026-03-09T15:11:15.128 INFO:teuthology.orchestra.run.vm06.stdout:(130/136): python3-zc-lockfile-2.0-10.el9.noarc 8.6 MB/s | 20 kB 00:00 2026-03-09T15:11:15.132 INFO:teuthology.orchestra.run.vm06.stdout:(131/136): re2-20211101-20.el9.x86_64.rpm 50 MB/s | 191 kB 00:00 2026-03-09T15:11:15.153 INFO:teuthology.orchestra.run.vm06.stdout:(132/136): thrift-0.15.0-4.el9.x86_64.rpm 78 MB/s | 1.6 MB 00:00 2026-03-09T15:11:15.279 INFO:teuthology.orchestra.run.vm03.stdout: Installing : flexiblas-netlib-3.0.4-9.el9.x86_64 80/138 2026-03-09T15:11:15.368 INFO:teuthology.orchestra.run.vm03.stdout: Installing : python3-numpy-1:1.23.5-2.el9.x86_64 81/138 2026-03-09T15:11:15.743 INFO:teuthology.orchestra.run.vm04.stdout:(135/136): librados2-19.2.3-678.ge911bdeb.el9.x 459 kB/s | 3.4 MB 00:07 2026-03-09T15:11:15.743 INFO:teuthology.orchestra.run.vm06.stdout:(133/136): python3-scipy-1.9.3-2.el9.x86_64.rpm 9.4 MB/s | 19 MB 00:02 2026-03-09T15:11:16.150 INFO:teuthology.orchestra.run.vm03.stdout: Installing : python3-numpy-f2py-1:1.23.5-2.el9.x86_64 82/138 2026-03-09T15:11:16.177 INFO:teuthology.orchestra.run.vm03.stdout: Installing : python3-scipy-1.9.3-2.el9.x86_64 83/138 2026-03-09T15:11:16.185 INFO:teuthology.orchestra.run.vm03.stdout: Installing : libxslt-1.1.34-12.el9.x86_64 84/138 2026-03-09T15:11:16.189 INFO:teuthology.orchestra.run.vm03.stdout: Installing : xmlstarlet-1.6.1-20.el9.x86_64 85/138 2026-03-09T15:11:16.344 INFO:teuthology.orchestra.run.vm03.stdout: Installing : libpmemobj-1.12.1-1.el9.x86_64 86/138 2026-03-09T15:11:16.347 INFO:teuthology.orchestra.run.vm03.stdout: Upgrading : librbd1-2:19.2.3-678.ge911bdeb.el9.x86_64 87/138 2026-03-09T15:11:16.380 INFO:teuthology.orchestra.run.vm03.stdout: Running scriptlet: librbd1-2:19.2.3-678.ge911bdeb.el9.x86_64 87/138 2026-03-09T15:11:16.385 INFO:teuthology.orchestra.run.vm03.stdout: Installing : python3-rbd-2:19.2.3-678.ge911bdeb.el9.x86_64 88/138 2026-03-09T15:11:16.392 INFO:teuthology.orchestra.run.vm03.stdout: Installing : boost-program-options-1.75.0-13.el9.x86_64 89/138 2026-03-09T15:11:16.654 INFO:teuthology.orchestra.run.vm03.stdout: Installing : parquet-libs-9.0.0-15.el9.x86_64 90/138 2026-03-09T15:11:16.657 INFO:teuthology.orchestra.run.vm03.stdout: Installing : librgw2-2:19.2.3-678.ge911bdeb.el9.x86_64 91/138 2026-03-09T15:11:16.676 INFO:teuthology.orchestra.run.vm03.stdout: Running scriptlet: librgw2-2:19.2.3-678.ge911bdeb.el9.x86_64 91/138 2026-03-09T15:11:16.677 INFO:teuthology.orchestra.run.vm06.stdout:(134/136): librados2-19.2.3-678.ge911bdeb.el9.x 2.3 MB/s | 3.4 MB 00:01 2026-03-09T15:11:16.679 INFO:teuthology.orchestra.run.vm03.stdout: Installing : python3-rgw-2:19.2.3-678.ge911bdeb.el9.x86_64 92/138 2026-03-09T15:11:16.768 INFO:teuthology.orchestra.run.vm06.stdout:(135/136): librbd1-19.2.3-678.ge911bdeb.el9.x86 3.1 MB/s | 3.2 MB 00:01 2026-03-09T15:11:17.784 INFO:teuthology.orchestra.run.vm03.stdout: Running scriptlet: ceph-common-2:19.2.3-678.ge911bdeb.el9.x86_64 93/138 2026-03-09T15:11:17.788 INFO:teuthology.orchestra.run.vm03.stdout: Installing : ceph-common-2:19.2.3-678.ge911bdeb.el9.x86_64 93/138 2026-03-09T15:11:17.809 INFO:teuthology.orchestra.run.vm03.stdout: Running scriptlet: ceph-common-2:19.2.3-678.ge911bdeb.el9.x86_64 93/138 2026-03-09T15:11:17.825 INFO:teuthology.orchestra.run.vm03.stdout: Installing : python3-ply-3.11-14.el9.noarch 94/138 2026-03-09T15:11:17.844 INFO:teuthology.orchestra.run.vm03.stdout: Installing : python3-pycparser-2.20-6.el9.noarch 95/138 2026-03-09T15:11:17.930 INFO:teuthology.orchestra.run.vm03.stdout: Installing : python3-cffi-1.14.5-5.el9.x86_64 96/138 2026-03-09T15:11:17.944 INFO:teuthology.orchestra.run.vm03.stdout: Installing : python3-cryptography-36.0.1-5.el9.x86_64 97/138 2026-03-09T15:11:17.971 INFO:teuthology.orchestra.run.vm03.stdout: Installing : python3-pyOpenSSL-21.0.0-1.el9.noarch 98/138 2026-03-09T15:11:18.006 INFO:teuthology.orchestra.run.vm03.stdout: Installing : python3-cheroot-10.0.1-4.el9.noarch 99/138 2026-03-09T15:11:18.065 INFO:teuthology.orchestra.run.vm03.stdout: Installing : python3-cherrypy-18.6.1-2.el9.noarch 100/138 2026-03-09T15:11:18.074 INFO:teuthology.orchestra.run.vm03.stdout: Installing : python3-asyncssh-2.13.2-5.el9.noarch 101/138 2026-03-09T15:11:18.080 INFO:teuthology.orchestra.run.vm03.stdout: Installing : python3-bcrypt-3.2.2-1.el9.x86_64 102/138 2026-03-09T15:11:18.086 INFO:teuthology.orchestra.run.vm03.stdout: Installing : pciutils-3.7.0-7.el9.x86_64 103/138 2026-03-09T15:11:18.090 INFO:teuthology.orchestra.run.vm03.stdout: Installing : qatlib-25.08.0-2.el9.x86_64 104/138 2026-03-09T15:11:18.092 INFO:teuthology.orchestra.run.vm03.stdout: Installing : qatlib-service-25.08.0-2.el9.x86_64 105/138 2026-03-09T15:11:18.109 INFO:teuthology.orchestra.run.vm03.stdout: Running scriptlet: qatlib-service-25.08.0-2.el9.x86_64 105/138 2026-03-09T15:11:18.253 INFO:teuthology.orchestra.run.vm06.stdout:(136/136): ceph-test-19.2.3-678.ge911bdeb.el9.x 2.6 MB/s | 50 MB 00:18 2026-03-09T15:11:18.255 INFO:teuthology.orchestra.run.vm06.stdout:-------------------------------------------------------------------------------- 2026-03-09T15:11:18.255 INFO:teuthology.orchestra.run.vm06.stdout:Total 7.3 MB/s | 210 MB 00:28 2026-03-09T15:11:18.557 INFO:teuthology.orchestra.run.vm03.stdout: Installing : qatzip-libs-1.3.1-1.el9.x86_64 106/138 2026-03-09T15:11:18.562 INFO:teuthology.orchestra.run.vm03.stdout: Installing : ceph-base-2:19.2.3-678.ge911bdeb.el9.x86_64 107/138 2026-03-09T15:11:18.607 INFO:teuthology.orchestra.run.vm03.stdout: Running scriptlet: ceph-base-2:19.2.3-678.ge911bdeb.el9.x86_64 107/138 2026-03-09T15:11:18.607 INFO:teuthology.orchestra.run.vm03.stdout:Created symlink /etc/systemd/system/multi-user.target.wants/ceph.target → /usr/lib/systemd/system/ceph.target. 2026-03-09T15:11:18.607 INFO:teuthology.orchestra.run.vm03.stdout:Created symlink /etc/systemd/system/ceph.target.wants/ceph-crash.service → /usr/lib/systemd/system/ceph-crash.service. 2026-03-09T15:11:18.607 INFO:teuthology.orchestra.run.vm03.stdout: 2026-03-09T15:11:18.611 INFO:teuthology.orchestra.run.vm03.stdout: Installing : ceph-selinux-2:19.2.3-678.ge911bdeb.el9.x86_64 108/138 2026-03-09T15:11:18.734 INFO:teuthology.orchestra.run.vm06.stdout:Running transaction check 2026-03-09T15:11:18.781 INFO:teuthology.orchestra.run.vm06.stdout:Transaction check succeeded. 2026-03-09T15:11:18.781 INFO:teuthology.orchestra.run.vm06.stdout:Running transaction test 2026-03-09T15:11:19.595 INFO:teuthology.orchestra.run.vm06.stdout:Transaction test succeeded. 2026-03-09T15:11:19.595 INFO:teuthology.orchestra.run.vm06.stdout:Running transaction 2026-03-09T15:11:20.476 INFO:teuthology.orchestra.run.vm06.stdout: Preparing : 1/1 2026-03-09T15:11:20.489 INFO:teuthology.orchestra.run.vm06.stdout: Installing : python3-more-itertools-8.12.0-2.el9.noarch 1/138 2026-03-09T15:11:20.500 INFO:teuthology.orchestra.run.vm06.stdout: Installing : thrift-0.15.0-4.el9.x86_64 2/138 2026-03-09T15:11:20.665 INFO:teuthology.orchestra.run.vm06.stdout: Installing : lttng-ust-2.12.0-6.el9.x86_64 3/138 2026-03-09T15:11:20.668 INFO:teuthology.orchestra.run.vm06.stdout: Upgrading : librados2-2:19.2.3-678.ge911bdeb.el9.x86_64 4/138 2026-03-09T15:11:20.726 INFO:teuthology.orchestra.run.vm06.stdout: Running scriptlet: librados2-2:19.2.3-678.ge911bdeb.el9.x86_64 4/138 2026-03-09T15:11:20.728 INFO:teuthology.orchestra.run.vm06.stdout: Installing : libcephfs2-2:19.2.3-678.ge911bdeb.el9.x86_64 5/138 2026-03-09T15:11:20.758 INFO:teuthology.orchestra.run.vm06.stdout: Running scriptlet: libcephfs2-2:19.2.3-678.ge911bdeb.el9.x86_64 5/138 2026-03-09T15:11:20.767 INFO:teuthology.orchestra.run.vm06.stdout: Installing : python3-rados-2:19.2.3-678.ge911bdeb.el9.x86_64 6/138 2026-03-09T15:11:20.770 INFO:teuthology.orchestra.run.vm06.stdout: Installing : librdkafka-1.6.1-102.el9.x86_64 7/138 2026-03-09T15:11:20.773 INFO:teuthology.orchestra.run.vm06.stdout: Installing : librabbitmq-0.11.0-7.el9.x86_64 8/138 2026-03-09T15:11:20.778 INFO:teuthology.orchestra.run.vm06.stdout: Installing : python3-jaraco-8.2.1-3.el9.noarch 9/138 2026-03-09T15:11:20.790 INFO:teuthology.orchestra.run.vm06.stdout: Installing : libnbd-1.20.3-4.el9.x86_64 10/138 2026-03-09T15:11:20.791 INFO:teuthology.orchestra.run.vm06.stdout: Installing : libcephsqlite-2:19.2.3-678.ge911bdeb.el9.x86_64 11/138 2026-03-09T15:11:20.827 INFO:teuthology.orchestra.run.vm06.stdout: Running scriptlet: libcephsqlite-2:19.2.3-678.ge911bdeb.el9.x86_64 11/138 2026-03-09T15:11:20.829 INFO:teuthology.orchestra.run.vm06.stdout: Installing : libradosstriper1-2:19.2.3-678.ge911bdeb.el9.x86_ 12/138 2026-03-09T15:11:20.842 INFO:teuthology.orchestra.run.vm06.stdout: Running scriptlet: libradosstriper1-2:19.2.3-678.ge911bdeb.el9.x86_ 12/138 2026-03-09T15:11:20.880 INFO:teuthology.orchestra.run.vm06.stdout: Installing : re2-1:20211101-20.el9.x86_64 13/138 2026-03-09T15:11:20.918 INFO:teuthology.orchestra.run.vm06.stdout: Installing : libarrow-9.0.0-15.el9.x86_64 14/138 2026-03-09T15:11:20.923 INFO:teuthology.orchestra.run.vm06.stdout: Installing : python3-werkzeug-2.0.3-3.el9.1.noarch 15/138 2026-03-09T15:11:20.947 INFO:teuthology.orchestra.run.vm06.stdout: Installing : liboath-2.6.12-1.el9.x86_64 16/138 2026-03-09T15:11:20.961 INFO:teuthology.orchestra.run.vm06.stdout: Installing : python3-pyasn1-0.4.8-7.el9.noarch 17/138 2026-03-09T15:11:20.969 INFO:teuthology.orchestra.run.vm06.stdout: Installing : python3-packaging-20.9-5.el9.noarch 18/138 2026-03-09T15:11:20.979 INFO:teuthology.orchestra.run.vm06.stdout: Installing : python3-markupsafe-1.1.1-12.el9.x86_64 19/138 2026-03-09T15:11:20.985 INFO:teuthology.orchestra.run.vm06.stdout: Installing : protobuf-3.14.0-17.el9.x86_64 20/138 2026-03-09T15:11:20.988 INFO:teuthology.orchestra.run.vm06.stdout: Installing : lua-5.4.4-4.el9.x86_64 21/138 2026-03-09T15:11:20.994 INFO:teuthology.orchestra.run.vm06.stdout: Installing : flexiblas-3.0.4-9.el9.x86_64 22/138 2026-03-09T15:11:21.021 INFO:teuthology.orchestra.run.vm06.stdout: Installing : unzip-6.0-59.el9.x86_64 23/138 2026-03-09T15:11:21.037 INFO:teuthology.orchestra.run.vm06.stdout: Installing : python3-urllib3-1.26.5-7.el9.noarch 24/138 2026-03-09T15:11:21.041 INFO:teuthology.orchestra.run.vm06.stdout: Installing : python3-requests-2.25.1-10.el9.noarch 25/138 2026-03-09T15:11:21.048 INFO:teuthology.orchestra.run.vm06.stdout: Installing : libquadmath-11.5.0-14.el9.x86_64 26/138 2026-03-09T15:11:21.050 INFO:teuthology.orchestra.run.vm06.stdout: Installing : libgfortran-11.5.0-14.el9.x86_64 27/138 2026-03-09T15:11:21.080 INFO:teuthology.orchestra.run.vm06.stdout: Installing : ledmon-libs-1.1.0-3.el9.x86_64 28/138 2026-03-09T15:11:21.087 INFO:teuthology.orchestra.run.vm06.stdout: Installing : python3-ceph-common-2:19.2.3-678.ge911bdeb.el9.x 29/138 2026-03-09T15:11:21.096 INFO:teuthology.orchestra.run.vm06.stdout: Installing : python3-ceph-argparse-2:19.2.3-678.ge911bdeb.el9 30/138 2026-03-09T15:11:21.110 INFO:teuthology.orchestra.run.vm06.stdout: Installing : python3-cephfs-2:19.2.3-678.ge911bdeb.el9.x86_64 31/138 2026-03-09T15:11:21.118 INFO:teuthology.orchestra.run.vm06.stdout: Installing : python3-requests-oauthlib-1.3.0-12.el9.noarch 32/138 2026-03-09T15:11:21.145 INFO:teuthology.orchestra.run.vm06.stdout: Installing : zip-3.0-35.el9.x86_64 33/138 2026-03-09T15:11:21.150 INFO:teuthology.orchestra.run.vm06.stdout: Installing : luarocks-3.9.2-5.el9.noarch 34/138 2026-03-09T15:11:21.158 INFO:teuthology.orchestra.run.vm06.stdout: Installing : lua-devel-5.4.4-4.el9.x86_64 35/138 2026-03-09T15:11:21.186 INFO:teuthology.orchestra.run.vm06.stdout: Installing : protobuf-compiler-3.14.0-17.el9.x86_64 36/138 2026-03-09T15:11:21.245 INFO:teuthology.orchestra.run.vm06.stdout: Installing : python3-mako-1.1.4-6.el9.noarch 37/138 2026-03-09T15:11:21.261 INFO:teuthology.orchestra.run.vm06.stdout: Installing : python3-pyasn1-modules-0.4.8-7.el9.noarch 38/138 2026-03-09T15:11:21.269 INFO:teuthology.orchestra.run.vm06.stdout: Installing : python3-rsa-4.9-2.el9.noarch 39/138 2026-03-09T15:11:21.278 INFO:teuthology.orchestra.run.vm06.stdout: Installing : python3-jaraco-classes-3.2.1-5.el9.noarch 40/138 2026-03-09T15:11:21.284 INFO:teuthology.orchestra.run.vm06.stdout: Installing : librados-devel-2:19.2.3-678.ge911bdeb.el9.x86_64 41/138 2026-03-09T15:11:21.289 INFO:teuthology.orchestra.run.vm06.stdout: Installing : python3-zc-lockfile-2.0-10.el9.noarch 42/138 2026-03-09T15:11:21.306 INFO:teuthology.orchestra.run.vm06.stdout: Installing : python3-xmltodict-0.12.0-15.el9.noarch 43/138 2026-03-09T15:11:21.331 INFO:teuthology.orchestra.run.vm06.stdout: Installing : python3-websocket-client-1.2.3-2.el9.noarch 44/138 2026-03-09T15:11:21.337 INFO:teuthology.orchestra.run.vm06.stdout: Installing : python3-webob-1.8.8-2.el9.noarch 45/138 2026-03-09T15:11:21.344 INFO:teuthology.orchestra.run.vm06.stdout: Installing : python3-typing-extensions-4.15.0-1.el9.noarch 46/138 2026-03-09T15:11:21.357 INFO:teuthology.orchestra.run.vm06.stdout: Installing : python3-repoze-lru-0.7-16.el9.noarch 47/138 2026-03-09T15:11:21.369 INFO:teuthology.orchestra.run.vm06.stdout: Installing : python3-routes-2.5.1-5.el9.noarch 48/138 2026-03-09T15:11:21.380 INFO:teuthology.orchestra.run.vm06.stdout: Installing : python3-natsort-7.1.1-5.el9.noarch 49/138 2026-03-09T15:11:21.441 INFO:teuthology.orchestra.run.vm06.stdout: Installing : python3-logutils-0.3.5-21.el9.noarch 50/138 2026-03-09T15:11:21.448 INFO:teuthology.orchestra.run.vm06.stdout: Installing : python3-pecan-1.4.2-3.el9.noarch 51/138 2026-03-09T15:11:21.457 INFO:teuthology.orchestra.run.vm06.stdout: Installing : python3-certifi-2023.05.07-4.el9.noarch 52/138 2026-03-09T15:11:21.504 INFO:teuthology.orchestra.run.vm06.stdout: Installing : python3-cachetools-4.2.4-1.el9.noarch 53/138 2026-03-09T15:11:21.896 INFO:teuthology.orchestra.run.vm06.stdout: Installing : python3-google-auth-1:2.45.0-1.el9.noarch 54/138 2026-03-09T15:11:21.955 INFO:teuthology.orchestra.run.vm04.stdout:(136/136): ceph-test-19.2.3-678.ge911bdeb.el9.x 2.3 MB/s | 50 MB 00:21 2026-03-09T15:11:21.957 INFO:teuthology.orchestra.run.vm04.stdout:-------------------------------------------------------------------------------- 2026-03-09T15:11:21.957 INFO:teuthology.orchestra.run.vm04.stdout:Total 7.0 MB/s | 210 MB 00:30 2026-03-09T15:11:21.961 INFO:teuthology.orchestra.run.vm06.stdout: Installing : python3-kubernetes-1:26.1.0-3.el9.noarch 55/138 2026-03-09T15:11:21.967 INFO:teuthology.orchestra.run.vm06.stdout: Installing : python3-backports-tarfile-1.2.0-1.el9.noarch 56/138 2026-03-09T15:11:21.975 INFO:teuthology.orchestra.run.vm06.stdout: Installing : python3-jaraco-context-6.0.1-3.el9.noarch 57/138 2026-03-09T15:11:21.981 INFO:teuthology.orchestra.run.vm06.stdout: Installing : python3-autocommand-2.2.2-8.el9.noarch 58/138 2026-03-09T15:11:21.989 INFO:teuthology.orchestra.run.vm06.stdout: Installing : libunwind-1.6.2-1.el9.x86_64 59/138 2026-03-09T15:11:21.993 INFO:teuthology.orchestra.run.vm06.stdout: Installing : gperftools-libs-2.9.1-3.el9.x86_64 60/138 2026-03-09T15:11:21.995 INFO:teuthology.orchestra.run.vm06.stdout: Installing : libarrow-doc-9.0.0-15.el9.noarch 61/138 2026-03-09T15:11:22.027 INFO:teuthology.orchestra.run.vm06.stdout: Installing : grpc-data-1.46.7-10.el9.noarch 62/138 2026-03-09T15:11:22.081 INFO:teuthology.orchestra.run.vm06.stdout: Installing : abseil-cpp-20211102.0-4.el9.x86_64 63/138 2026-03-09T15:11:22.098 INFO:teuthology.orchestra.run.vm06.stdout: Installing : python3-grpcio-1.46.7-10.el9.x86_64 64/138 2026-03-09T15:11:22.107 INFO:teuthology.orchestra.run.vm06.stdout: Installing : socat-1.7.4.1-8.el9.x86_64 65/138 2026-03-09T15:11:22.113 INFO:teuthology.orchestra.run.vm06.stdout: Installing : python3-toml-0.10.2-6.el9.noarch 66/138 2026-03-09T15:11:22.121 INFO:teuthology.orchestra.run.vm06.stdout: Installing : python3-jaraco-functools-3.5.0-2.el9.noarch 67/138 2026-03-09T15:11:22.128 INFO:teuthology.orchestra.run.vm06.stdout: Installing : python3-jaraco-text-4.0.0-2.el9.noarch 68/138 2026-03-09T15:11:22.140 INFO:teuthology.orchestra.run.vm06.stdout: Installing : python3-jaraco-collections-3.0.0-8.el9.noarch 69/138 2026-03-09T15:11:22.146 INFO:teuthology.orchestra.run.vm06.stdout: Installing : python3-tempora-5.0.0-2.el9.noarch 70/138 2026-03-09T15:11:22.183 INFO:teuthology.orchestra.run.vm06.stdout: Installing : python3-portend-3.1.0-2.el9.noarch 71/138 2026-03-09T15:11:22.197 INFO:teuthology.orchestra.run.vm06.stdout: Installing : python3-protobuf-3.14.0-17.el9.noarch 72/138 2026-03-09T15:11:22.241 INFO:teuthology.orchestra.run.vm06.stdout: Installing : python3-grpcio-tools-1.46.7-10.el9.x86_64 73/138 2026-03-09T15:11:22.512 INFO:teuthology.orchestra.run.vm06.stdout: Installing : python3-devel-3.9.25-3.el9.x86_64 74/138 2026-03-09T15:11:22.516 INFO:teuthology.orchestra.run.vm04.stdout:Running transaction check 2026-03-09T15:11:22.544 INFO:teuthology.orchestra.run.vm06.stdout: Installing : python3-babel-2.9.1-2.el9.noarch 75/138 2026-03-09T15:11:22.552 INFO:teuthology.orchestra.run.vm06.stdout: Installing : python3-jinja2-2.11.3-8.el9.noarch 76/138 2026-03-09T15:11:22.565 INFO:teuthology.orchestra.run.vm04.stdout:Transaction check succeeded. 2026-03-09T15:11:22.566 INFO:teuthology.orchestra.run.vm04.stdout:Running transaction test 2026-03-09T15:11:22.612 INFO:teuthology.orchestra.run.vm06.stdout: Installing : openblas-0.3.29-1.el9.x86_64 77/138 2026-03-09T15:11:22.616 INFO:teuthology.orchestra.run.vm06.stdout: Installing : openblas-openmp-0.3.29-1.el9.x86_64 78/138 2026-03-09T15:11:22.638 INFO:teuthology.orchestra.run.vm06.stdout: Installing : flexiblas-openblas-openmp-3.0.4-9.el9.x86_64 79/138 2026-03-09T15:11:23.026 INFO:teuthology.orchestra.run.vm06.stdout: Installing : flexiblas-netlib-3.0.4-9.el9.x86_64 80/138 2026-03-09T15:11:23.117 INFO:teuthology.orchestra.run.vm06.stdout: Installing : python3-numpy-1:1.23.5-2.el9.x86_64 81/138 2026-03-09T15:11:23.399 INFO:teuthology.orchestra.run.vm04.stdout:Transaction test succeeded. 2026-03-09T15:11:23.400 INFO:teuthology.orchestra.run.vm04.stdout:Running transaction 2026-03-09T15:11:23.895 INFO:teuthology.orchestra.run.vm06.stdout: Installing : python3-numpy-f2py-1:1.23.5-2.el9.x86_64 82/138 2026-03-09T15:11:24.137 INFO:teuthology.orchestra.run.vm06.stdout: Installing : python3-scipy-1.9.3-2.el9.x86_64 83/138 2026-03-09T15:11:24.190 INFO:teuthology.orchestra.run.vm06.stdout: Installing : libxslt-1.1.34-12.el9.x86_64 84/138 2026-03-09T15:11:24.194 INFO:teuthology.orchestra.run.vm06.stdout: Installing : xmlstarlet-1.6.1-20.el9.x86_64 85/138 2026-03-09T15:11:24.351 INFO:teuthology.orchestra.run.vm06.stdout: Installing : libpmemobj-1.12.1-1.el9.x86_64 86/138 2026-03-09T15:11:24.354 INFO:teuthology.orchestra.run.vm06.stdout: Upgrading : librbd1-2:19.2.3-678.ge911bdeb.el9.x86_64 87/138 2026-03-09T15:11:24.386 INFO:teuthology.orchestra.run.vm06.stdout: Running scriptlet: librbd1-2:19.2.3-678.ge911bdeb.el9.x86_64 87/138 2026-03-09T15:11:24.390 INFO:teuthology.orchestra.run.vm06.stdout: Installing : python3-rbd-2:19.2.3-678.ge911bdeb.el9.x86_64 88/138 2026-03-09T15:11:24.397 INFO:teuthology.orchestra.run.vm06.stdout: Installing : boost-program-options-1.75.0-13.el9.x86_64 89/138 2026-03-09T15:11:24.640 INFO:teuthology.orchestra.run.vm06.stdout: Installing : parquet-libs-9.0.0-15.el9.x86_64 90/138 2026-03-09T15:11:24.642 INFO:teuthology.orchestra.run.vm06.stdout: Installing : librgw2-2:19.2.3-678.ge911bdeb.el9.x86_64 91/138 2026-03-09T15:11:24.661 INFO:teuthology.orchestra.run.vm03.stdout: Running scriptlet: ceph-selinux-2:19.2.3-678.ge911bdeb.el9.x86_64 108/138 2026-03-09T15:11:24.661 INFO:teuthology.orchestra.run.vm03.stdout:skipping the directory /sys 2026-03-09T15:11:24.661 INFO:teuthology.orchestra.run.vm03.stdout:skipping the directory /proc 2026-03-09T15:11:24.661 INFO:teuthology.orchestra.run.vm03.stdout:skipping the directory /mnt 2026-03-09T15:11:24.661 INFO:teuthology.orchestra.run.vm03.stdout:skipping the directory /var/tmp 2026-03-09T15:11:24.661 INFO:teuthology.orchestra.run.vm03.stdout:skipping the directory /home 2026-03-09T15:11:24.661 INFO:teuthology.orchestra.run.vm03.stdout:skipping the directory /root 2026-03-09T15:11:24.661 INFO:teuthology.orchestra.run.vm03.stdout:skipping the directory /tmp 2026-03-09T15:11:24.661 INFO:teuthology.orchestra.run.vm03.stdout: 2026-03-09T15:11:24.665 INFO:teuthology.orchestra.run.vm06.stdout: Running scriptlet: librgw2-2:19.2.3-678.ge911bdeb.el9.x86_64 91/138 2026-03-09T15:11:24.667 INFO:teuthology.orchestra.run.vm06.stdout: Installing : python3-rgw-2:19.2.3-678.ge911bdeb.el9.x86_64 92/138 2026-03-09T15:11:24.784 INFO:teuthology.orchestra.run.vm03.stdout: Installing : ceph-mds-2:19.2.3-678.ge911bdeb.el9.x86_64 109/138 2026-03-09T15:11:24.806 INFO:teuthology.orchestra.run.vm03.stdout: Running scriptlet: ceph-mds-2:19.2.3-678.ge911bdeb.el9.x86_64 109/138 2026-03-09T15:11:24.806 INFO:teuthology.orchestra.run.vm03.stdout:Glob pattern passed to enable, but globs are not supported for this. 2026-03-09T15:11:24.806 INFO:teuthology.orchestra.run.vm03.stdout:Invalid unit name "ceph-mds@*.service" escaped as "ceph-mds@\x2a.service". 2026-03-09T15:11:24.806 INFO:teuthology.orchestra.run.vm03.stdout:Created symlink /etc/systemd/system/multi-user.target.wants/ceph-mds.target → /usr/lib/systemd/system/ceph-mds.target. 2026-03-09T15:11:24.806 INFO:teuthology.orchestra.run.vm03.stdout:Created symlink /etc/systemd/system/ceph.target.wants/ceph-mds.target → /usr/lib/systemd/system/ceph-mds.target. 2026-03-09T15:11:24.806 INFO:teuthology.orchestra.run.vm03.stdout: 2026-03-09T15:11:25.040 INFO:teuthology.orchestra.run.vm03.stdout: Installing : ceph-mon-2:19.2.3-678.ge911bdeb.el9.x86_64 110/138 2026-03-09T15:11:25.060 INFO:teuthology.orchestra.run.vm03.stdout: Running scriptlet: ceph-mon-2:19.2.3-678.ge911bdeb.el9.x86_64 110/138 2026-03-09T15:11:25.061 INFO:teuthology.orchestra.run.vm03.stdout:Glob pattern passed to enable, but globs are not supported for this. 2026-03-09T15:11:25.061 INFO:teuthology.orchestra.run.vm03.stdout:Invalid unit name "ceph-mon@*.service" escaped as "ceph-mon@\x2a.service". 2026-03-09T15:11:25.061 INFO:teuthology.orchestra.run.vm03.stdout:Created symlink /etc/systemd/system/multi-user.target.wants/ceph-mon.target → /usr/lib/systemd/system/ceph-mon.target. 2026-03-09T15:11:25.061 INFO:teuthology.orchestra.run.vm03.stdout:Created symlink /etc/systemd/system/ceph.target.wants/ceph-mon.target → /usr/lib/systemd/system/ceph-mon.target. 2026-03-09T15:11:25.061 INFO:teuthology.orchestra.run.vm03.stdout: 2026-03-09T15:11:25.068 INFO:teuthology.orchestra.run.vm03.stdout: Installing : mailcap-2.1.49-5.el9.noarch 111/138 2026-03-09T15:11:25.071 INFO:teuthology.orchestra.run.vm03.stdout: Installing : libconfig-1.7.2-9.el9.x86_64 112/138 2026-03-09T15:11:25.072 INFO:teuthology.orchestra.run.vm04.stdout: Preparing : 1/1 2026-03-09T15:11:25.086 INFO:teuthology.orchestra.run.vm04.stdout: Installing : python3-more-itertools-8.12.0-2.el9.noarch 1/138 2026-03-09T15:11:25.089 INFO:teuthology.orchestra.run.vm03.stdout: Running scriptlet: libstoragemgmt-1.10.1-1.el9.x86_64 113/138 2026-03-09T15:11:25.089 INFO:teuthology.orchestra.run.vm03.stdout:Creating group 'qat' with GID 994. 2026-03-09T15:11:25.089 INFO:teuthology.orchestra.run.vm03.stdout:Creating group 'libstoragemgmt' with GID 993. 2026-03-09T15:11:25.089 INFO:teuthology.orchestra.run.vm03.stdout:Creating user 'libstoragemgmt' (daemon account for libstoragemgmt) with UID 993 and GID 993. 2026-03-09T15:11:25.089 INFO:teuthology.orchestra.run.vm03.stdout: 2026-03-09T15:11:25.098 INFO:teuthology.orchestra.run.vm04.stdout: Installing : thrift-0.15.0-4.el9.x86_64 2/138 2026-03-09T15:11:25.101 INFO:teuthology.orchestra.run.vm03.stdout: Installing : libstoragemgmt-1.10.1-1.el9.x86_64 113/138 2026-03-09T15:11:25.129 INFO:teuthology.orchestra.run.vm03.stdout: Running scriptlet: libstoragemgmt-1.10.1-1.el9.x86_64 113/138 2026-03-09T15:11:25.129 INFO:teuthology.orchestra.run.vm03.stdout:Created symlink /etc/systemd/system/multi-user.target.wants/libstoragemgmt.service → /usr/lib/systemd/system/libstoragemgmt.service. 2026-03-09T15:11:25.129 INFO:teuthology.orchestra.run.vm03.stdout: 2026-03-09T15:11:25.172 INFO:teuthology.orchestra.run.vm03.stdout: Installing : python3-libstoragemgmt-1.10.1-1.el9.x86_64 114/138 2026-03-09T15:11:25.248 INFO:teuthology.orchestra.run.vm03.stdout: Installing : cryptsetup-2.8.1-3.el9.x86_64 115/138 2026-03-09T15:11:25.253 INFO:teuthology.orchestra.run.vm03.stdout: Installing : ceph-volume-2:19.2.3-678.ge911bdeb.el9.noarch 116/138 2026-03-09T15:11:25.267 INFO:teuthology.orchestra.run.vm03.stdout: Running scriptlet: ceph-volume-2:19.2.3-678.ge911bdeb.el9.noarch 116/138 2026-03-09T15:11:25.267 INFO:teuthology.orchestra.run.vm03.stdout:Glob pattern passed to enable, but globs are not supported for this. 2026-03-09T15:11:25.267 INFO:teuthology.orchestra.run.vm03.stdout:Invalid unit name "ceph-volume@*.service" escaped as "ceph-volume@\x2a.service". 2026-03-09T15:11:25.267 INFO:teuthology.orchestra.run.vm03.stdout: 2026-03-09T15:11:25.267 INFO:teuthology.orchestra.run.vm04.stdout: Installing : lttng-ust-2.12.0-6.el9.x86_64 3/138 2026-03-09T15:11:25.270 INFO:teuthology.orchestra.run.vm04.stdout: Upgrading : librados2-2:19.2.3-678.ge911bdeb.el9.x86_64 4/138 2026-03-09T15:11:25.328 INFO:teuthology.orchestra.run.vm04.stdout: Running scriptlet: librados2-2:19.2.3-678.ge911bdeb.el9.x86_64 4/138 2026-03-09T15:11:25.330 INFO:teuthology.orchestra.run.vm04.stdout: Installing : libcephfs2-2:19.2.3-678.ge911bdeb.el9.x86_64 5/138 2026-03-09T15:11:25.359 INFO:teuthology.orchestra.run.vm04.stdout: Running scriptlet: libcephfs2-2:19.2.3-678.ge911bdeb.el9.x86_64 5/138 2026-03-09T15:11:25.367 INFO:teuthology.orchestra.run.vm04.stdout: Installing : python3-rados-2:19.2.3-678.ge911bdeb.el9.x86_64 6/138 2026-03-09T15:11:25.370 INFO:teuthology.orchestra.run.vm04.stdout: Installing : librdkafka-1.6.1-102.el9.x86_64 7/138 2026-03-09T15:11:25.373 INFO:teuthology.orchestra.run.vm04.stdout: Installing : librabbitmq-0.11.0-7.el9.x86_64 8/138 2026-03-09T15:11:25.378 INFO:teuthology.orchestra.run.vm04.stdout: Installing : python3-jaraco-8.2.1-3.el9.noarch 9/138 2026-03-09T15:11:25.387 INFO:teuthology.orchestra.run.vm04.stdout: Installing : libnbd-1.20.3-4.el9.x86_64 10/138 2026-03-09T15:11:25.388 INFO:teuthology.orchestra.run.vm04.stdout: Installing : libcephsqlite-2:19.2.3-678.ge911bdeb.el9.x86_64 11/138 2026-03-09T15:11:25.424 INFO:teuthology.orchestra.run.vm04.stdout: Running scriptlet: libcephsqlite-2:19.2.3-678.ge911bdeb.el9.x86_64 11/138 2026-03-09T15:11:25.426 INFO:teuthology.orchestra.run.vm04.stdout: Installing : libradosstriper1-2:19.2.3-678.ge911bdeb.el9.x86_ 12/138 2026-03-09T15:11:25.439 INFO:teuthology.orchestra.run.vm04.stdout: Running scriptlet: libradosstriper1-2:19.2.3-678.ge911bdeb.el9.x86_ 12/138 2026-03-09T15:11:25.471 INFO:teuthology.orchestra.run.vm04.stdout: Installing : re2-1:20211101-20.el9.x86_64 13/138 2026-03-09T15:11:25.508 INFO:teuthology.orchestra.run.vm04.stdout: Installing : libarrow-9.0.0-15.el9.x86_64 14/138 2026-03-09T15:11:25.512 INFO:teuthology.orchestra.run.vm04.stdout: Installing : python3-werkzeug-2.0.3-3.el9.1.noarch 15/138 2026-03-09T15:11:25.537 INFO:teuthology.orchestra.run.vm04.stdout: Installing : liboath-2.6.12-1.el9.x86_64 16/138 2026-03-09T15:11:25.553 INFO:teuthology.orchestra.run.vm04.stdout: Installing : python3-pyasn1-0.4.8-7.el9.noarch 17/138 2026-03-09T15:11:25.560 INFO:teuthology.orchestra.run.vm04.stdout: Installing : python3-packaging-20.9-5.el9.noarch 18/138 2026-03-09T15:11:25.570 INFO:teuthology.orchestra.run.vm04.stdout: Installing : python3-markupsafe-1.1.1-12.el9.x86_64 19/138 2026-03-09T15:11:25.576 INFO:teuthology.orchestra.run.vm04.stdout: Installing : protobuf-3.14.0-17.el9.x86_64 20/138 2026-03-09T15:11:25.580 INFO:teuthology.orchestra.run.vm04.stdout: Installing : lua-5.4.4-4.el9.x86_64 21/138 2026-03-09T15:11:25.585 INFO:teuthology.orchestra.run.vm04.stdout: Installing : flexiblas-3.0.4-9.el9.x86_64 22/138 2026-03-09T15:11:25.617 INFO:teuthology.orchestra.run.vm04.stdout: Installing : unzip-6.0-59.el9.x86_64 23/138 2026-03-09T15:11:25.632 INFO:teuthology.orchestra.run.vm04.stdout: Installing : python3-urllib3-1.26.5-7.el9.noarch 24/138 2026-03-09T15:11:25.637 INFO:teuthology.orchestra.run.vm04.stdout: Installing : python3-requests-2.25.1-10.el9.noarch 25/138 2026-03-09T15:11:25.644 INFO:teuthology.orchestra.run.vm04.stdout: Installing : libquadmath-11.5.0-14.el9.x86_64 26/138 2026-03-09T15:11:25.646 INFO:teuthology.orchestra.run.vm04.stdout: Installing : libgfortran-11.5.0-14.el9.x86_64 27/138 2026-03-09T15:11:25.678 INFO:teuthology.orchestra.run.vm04.stdout: Installing : ledmon-libs-1.1.0-3.el9.x86_64 28/138 2026-03-09T15:11:25.684 INFO:teuthology.orchestra.run.vm04.stdout: Installing : python3-ceph-common-2:19.2.3-678.ge911bdeb.el9.x 29/138 2026-03-09T15:11:25.694 INFO:teuthology.orchestra.run.vm04.stdout: Installing : python3-ceph-argparse-2:19.2.3-678.ge911bdeb.el9 30/138 2026-03-09T15:11:25.707 INFO:teuthology.orchestra.run.vm04.stdout: Installing : python3-cephfs-2:19.2.3-678.ge911bdeb.el9.x86_64 31/138 2026-03-09T15:11:25.715 INFO:teuthology.orchestra.run.vm04.stdout: Installing : python3-requests-oauthlib-1.3.0-12.el9.noarch 32/138 2026-03-09T15:11:25.746 INFO:teuthology.orchestra.run.vm04.stdout: Installing : zip-3.0-35.el9.x86_64 33/138 2026-03-09T15:11:25.751 INFO:teuthology.orchestra.run.vm04.stdout: Installing : luarocks-3.9.2-5.el9.noarch 34/138 2026-03-09T15:11:25.765 INFO:teuthology.orchestra.run.vm04.stdout: Installing : lua-devel-5.4.4-4.el9.x86_64 35/138 2026-03-09T15:11:25.787 INFO:teuthology.orchestra.run.vm06.stdout: Running scriptlet: ceph-common-2:19.2.3-678.ge911bdeb.el9.x86_64 93/138 2026-03-09T15:11:25.793 INFO:teuthology.orchestra.run.vm06.stdout: Installing : ceph-common-2:19.2.3-678.ge911bdeb.el9.x86_64 93/138 2026-03-09T15:11:25.795 INFO:teuthology.orchestra.run.vm04.stdout: Installing : protobuf-compiler-3.14.0-17.el9.x86_64 36/138 2026-03-09T15:11:25.815 INFO:teuthology.orchestra.run.vm06.stdout: Running scriptlet: ceph-common-2:19.2.3-678.ge911bdeb.el9.x86_64 93/138 2026-03-09T15:11:25.833 INFO:teuthology.orchestra.run.vm06.stdout: Installing : python3-ply-3.11-14.el9.noarch 94/138 2026-03-09T15:11:25.855 INFO:teuthology.orchestra.run.vm04.stdout: Installing : python3-mako-1.1.4-6.el9.noarch 37/138 2026-03-09T15:11:25.855 INFO:teuthology.orchestra.run.vm06.stdout: Installing : python3-pycparser-2.20-6.el9.noarch 95/138 2026-03-09T15:11:25.871 INFO:teuthology.orchestra.run.vm04.stdout: Installing : python3-pyasn1-modules-0.4.8-7.el9.noarch 38/138 2026-03-09T15:11:25.879 INFO:teuthology.orchestra.run.vm04.stdout: Installing : python3-rsa-4.9-2.el9.noarch 39/138 2026-03-09T15:11:25.888 INFO:teuthology.orchestra.run.vm04.stdout: Installing : python3-jaraco-classes-3.2.1-5.el9.noarch 40/138 2026-03-09T15:11:25.894 INFO:teuthology.orchestra.run.vm04.stdout: Installing : librados-devel-2:19.2.3-678.ge911bdeb.el9.x86_64 41/138 2026-03-09T15:11:25.898 INFO:teuthology.orchestra.run.vm04.stdout: Installing : python3-zc-lockfile-2.0-10.el9.noarch 42/138 2026-03-09T15:11:25.915 INFO:teuthology.orchestra.run.vm04.stdout: Installing : python3-xmltodict-0.12.0-15.el9.noarch 43/138 2026-03-09T15:11:25.939 INFO:teuthology.orchestra.run.vm04.stdout: Installing : python3-websocket-client-1.2.3-2.el9.noarch 44/138 2026-03-09T15:11:25.943 INFO:teuthology.orchestra.run.vm06.stdout: Installing : python3-cffi-1.14.5-5.el9.x86_64 96/138 2026-03-09T15:11:25.945 INFO:teuthology.orchestra.run.vm04.stdout: Installing : python3-webob-1.8.8-2.el9.noarch 45/138 2026-03-09T15:11:25.952 INFO:teuthology.orchestra.run.vm04.stdout: Installing : python3-typing-extensions-4.15.0-1.el9.noarch 46/138 2026-03-09T15:11:25.958 INFO:teuthology.orchestra.run.vm06.stdout: Installing : python3-cryptography-36.0.1-5.el9.x86_64 97/138 2026-03-09T15:11:25.965 INFO:teuthology.orchestra.run.vm04.stdout: Installing : python3-repoze-lru-0.7-16.el9.noarch 47/138 2026-03-09T15:11:25.977 INFO:teuthology.orchestra.run.vm04.stdout: Installing : python3-routes-2.5.1-5.el9.noarch 48/138 2026-03-09T15:11:25.986 INFO:teuthology.orchestra.run.vm06.stdout: Installing : python3-pyOpenSSL-21.0.0-1.el9.noarch 98/138 2026-03-09T15:11:25.988 INFO:teuthology.orchestra.run.vm04.stdout: Installing : python3-natsort-7.1.1-5.el9.noarch 49/138 2026-03-09T15:11:26.027 INFO:teuthology.orchestra.run.vm06.stdout: Installing : python3-cheroot-10.0.1-4.el9.noarch 99/138 2026-03-09T15:11:26.049 INFO:teuthology.orchestra.run.vm04.stdout: Installing : python3-logutils-0.3.5-21.el9.noarch 50/138 2026-03-09T15:11:26.057 INFO:teuthology.orchestra.run.vm04.stdout: Installing : python3-pecan-1.4.2-3.el9.noarch 51/138 2026-03-09T15:11:26.066 INFO:teuthology.orchestra.run.vm04.stdout: Installing : python3-certifi-2023.05.07-4.el9.noarch 52/138 2026-03-09T15:11:26.067 INFO:teuthology.orchestra.run.vm03.stdout: Installing : ceph-osd-2:19.2.3-678.ge911bdeb.el9.x86_64 117/138 2026-03-09T15:11:26.090 INFO:teuthology.orchestra.run.vm06.stdout: Installing : python3-cherrypy-18.6.1-2.el9.noarch 100/138 2026-03-09T15:11:26.090 INFO:teuthology.orchestra.run.vm03.stdout: Running scriptlet: ceph-osd-2:19.2.3-678.ge911bdeb.el9.x86_64 117/138 2026-03-09T15:11:26.090 INFO:teuthology.orchestra.run.vm03.stdout:Glob pattern passed to enable, but globs are not supported for this. 2026-03-09T15:11:26.090 INFO:teuthology.orchestra.run.vm03.stdout:Invalid unit name "ceph-osd@*.service" escaped as "ceph-osd@\x2a.service". 2026-03-09T15:11:26.091 INFO:teuthology.orchestra.run.vm03.stdout:Created symlink /etc/systemd/system/multi-user.target.wants/ceph-osd.target → /usr/lib/systemd/system/ceph-osd.target. 2026-03-09T15:11:26.091 INFO:teuthology.orchestra.run.vm03.stdout:Created symlink /etc/systemd/system/ceph.target.wants/ceph-osd.target → /usr/lib/systemd/system/ceph-osd.target. 2026-03-09T15:11:26.091 INFO:teuthology.orchestra.run.vm03.stdout: 2026-03-09T15:11:26.100 INFO:teuthology.orchestra.run.vm06.stdout: Installing : python3-asyncssh-2.13.2-5.el9.noarch 101/138 2026-03-09T15:11:26.106 INFO:teuthology.orchestra.run.vm06.stdout: Installing : python3-bcrypt-3.2.2-1.el9.x86_64 102/138 2026-03-09T15:11:26.113 INFO:teuthology.orchestra.run.vm04.stdout: Installing : python3-cachetools-4.2.4-1.el9.noarch 53/138 2026-03-09T15:11:26.114 INFO:teuthology.orchestra.run.vm06.stdout: Installing : pciutils-3.7.0-7.el9.x86_64 103/138 2026-03-09T15:11:26.118 INFO:teuthology.orchestra.run.vm06.stdout: Installing : qatlib-25.08.0-2.el9.x86_64 104/138 2026-03-09T15:11:26.121 INFO:teuthology.orchestra.run.vm06.stdout: Installing : qatlib-service-25.08.0-2.el9.x86_64 105/138 2026-03-09T15:11:26.139 INFO:teuthology.orchestra.run.vm06.stdout: Running scriptlet: qatlib-service-25.08.0-2.el9.x86_64 105/138 2026-03-09T15:11:26.152 INFO:teuthology.orchestra.run.vm03.stdout: Running scriptlet: cephadm-2:19.2.3-678.ge911bdeb.el9.noarch 118/138 2026-03-09T15:11:26.155 INFO:teuthology.orchestra.run.vm03.stdout: Installing : cephadm-2:19.2.3-678.ge911bdeb.el9.noarch 118/138 2026-03-09T15:11:26.162 INFO:teuthology.orchestra.run.vm03.stdout: Installing : ceph-prometheus-alerts-2:19.2.3-678.ge911bdeb.el 119/138 2026-03-09T15:11:26.184 INFO:teuthology.orchestra.run.vm03.stdout: Installing : ceph-grafana-dashboards-2:19.2.3-678.ge911bdeb.e 120/138 2026-03-09T15:11:26.187 INFO:teuthology.orchestra.run.vm03.stdout: Installing : ceph-mgr-cephadm-2:19.2.3-678.ge911bdeb.el9.noar 121/138 2026-03-09T15:11:26.439 INFO:teuthology.orchestra.run.vm06.stdout: Installing : qatzip-libs-1.3.1-1.el9.x86_64 106/138 2026-03-09T15:11:26.470 INFO:teuthology.orchestra.run.vm06.stdout: Installing : ceph-base-2:19.2.3-678.ge911bdeb.el9.x86_64 107/138 2026-03-09T15:11:26.476 INFO:teuthology.orchestra.run.vm04.stdout: Installing : python3-google-auth-1:2.45.0-1.el9.noarch 54/138 2026-03-09T15:11:26.492 INFO:teuthology.orchestra.run.vm04.stdout: Installing : python3-kubernetes-1:26.1.0-3.el9.noarch 55/138 2026-03-09T15:11:26.499 INFO:teuthology.orchestra.run.vm04.stdout: Installing : python3-backports-tarfile-1.2.0-1.el9.noarch 56/138 2026-03-09T15:11:26.506 INFO:teuthology.orchestra.run.vm04.stdout: Installing : python3-jaraco-context-6.0.1-3.el9.noarch 57/138 2026-03-09T15:11:26.511 INFO:teuthology.orchestra.run.vm04.stdout: Installing : python3-autocommand-2.2.2-8.el9.noarch 58/138 2026-03-09T15:11:26.517 INFO:teuthology.orchestra.run.vm06.stdout: Running scriptlet: ceph-base-2:19.2.3-678.ge911bdeb.el9.x86_64 107/138 2026-03-09T15:11:26.517 INFO:teuthology.orchestra.run.vm06.stdout:Created symlink /etc/systemd/system/multi-user.target.wants/ceph.target → /usr/lib/systemd/system/ceph.target. 2026-03-09T15:11:26.517 INFO:teuthology.orchestra.run.vm06.stdout:Created symlink /etc/systemd/system/ceph.target.wants/ceph-crash.service → /usr/lib/systemd/system/ceph-crash.service. 2026-03-09T15:11:26.517 INFO:teuthology.orchestra.run.vm06.stdout: 2026-03-09T15:11:26.519 INFO:teuthology.orchestra.run.vm04.stdout: Installing : libunwind-1.6.2-1.el9.x86_64 59/138 2026-03-09T15:11:26.523 INFO:teuthology.orchestra.run.vm06.stdout: Installing : ceph-selinux-2:19.2.3-678.ge911bdeb.el9.x86_64 108/138 2026-03-09T15:11:26.523 INFO:teuthology.orchestra.run.vm04.stdout: Installing : gperftools-libs-2.9.1-3.el9.x86_64 60/138 2026-03-09T15:11:26.525 INFO:teuthology.orchestra.run.vm04.stdout: Installing : libarrow-doc-9.0.0-15.el9.noarch 61/138 2026-03-09T15:11:26.556 INFO:teuthology.orchestra.run.vm04.stdout: Installing : grpc-data-1.46.7-10.el9.noarch 62/138 2026-03-09T15:11:26.606 INFO:teuthology.orchestra.run.vm04.stdout: Installing : abseil-cpp-20211102.0-4.el9.x86_64 63/138 2026-03-09T15:11:26.619 INFO:teuthology.orchestra.run.vm04.stdout: Installing : python3-grpcio-1.46.7-10.el9.x86_64 64/138 2026-03-09T15:11:26.627 INFO:teuthology.orchestra.run.vm04.stdout: Installing : socat-1.7.4.1-8.el9.x86_64 65/138 2026-03-09T15:11:26.632 INFO:teuthology.orchestra.run.vm04.stdout: Installing : python3-toml-0.10.2-6.el9.noarch 66/138 2026-03-09T15:11:26.640 INFO:teuthology.orchestra.run.vm04.stdout: Installing : python3-jaraco-functools-3.5.0-2.el9.noarch 67/138 2026-03-09T15:11:26.645 INFO:teuthology.orchestra.run.vm04.stdout: Installing : python3-jaraco-text-4.0.0-2.el9.noarch 68/138 2026-03-09T15:11:26.655 INFO:teuthology.orchestra.run.vm04.stdout: Installing : python3-jaraco-collections-3.0.0-8.el9.noarch 69/138 2026-03-09T15:11:26.660 INFO:teuthology.orchestra.run.vm04.stdout: Installing : python3-tempora-5.0.0-2.el9.noarch 70/138 2026-03-09T15:11:26.694 INFO:teuthology.orchestra.run.vm04.stdout: Installing : python3-portend-3.1.0-2.el9.noarch 71/138 2026-03-09T15:11:26.708 INFO:teuthology.orchestra.run.vm04.stdout: Installing : python3-protobuf-3.14.0-17.el9.noarch 72/138 2026-03-09T15:11:26.724 INFO:teuthology.orchestra.run.vm03.stdout: Running scriptlet: ceph-mgr-cephadm-2:19.2.3-678.ge911bdeb.el9.noar 121/138 2026-03-09T15:11:26.730 INFO:teuthology.orchestra.run.vm03.stdout: Installing : ceph-mgr-dashboard-2:19.2.3-678.ge911bdeb.el9.no 122/138 2026-03-09T15:11:26.752 INFO:teuthology.orchestra.run.vm04.stdout: Installing : python3-grpcio-tools-1.46.7-10.el9.x86_64 73/138 2026-03-09T15:11:27.004 INFO:teuthology.orchestra.run.vm04.stdout: Installing : python3-devel-3.9.25-3.el9.x86_64 74/138 2026-03-09T15:11:27.034 INFO:teuthology.orchestra.run.vm04.stdout: Installing : python3-babel-2.9.1-2.el9.noarch 75/138 2026-03-09T15:11:27.041 INFO:teuthology.orchestra.run.vm04.stdout: Installing : python3-jinja2-2.11.3-8.el9.noarch 76/138 2026-03-09T15:11:27.100 INFO:teuthology.orchestra.run.vm04.stdout: Installing : openblas-0.3.29-1.el9.x86_64 77/138 2026-03-09T15:11:27.103 INFO:teuthology.orchestra.run.vm04.stdout: Installing : openblas-openmp-0.3.29-1.el9.x86_64 78/138 2026-03-09T15:11:27.126 INFO:teuthology.orchestra.run.vm04.stdout: Installing : flexiblas-openblas-openmp-3.0.4-9.el9.x86_64 79/138 2026-03-09T15:11:27.242 INFO:teuthology.orchestra.run.vm03.stdout: Running scriptlet: ceph-mgr-dashboard-2:19.2.3-678.ge911bdeb.el9.no 122/138 2026-03-09T15:11:27.245 INFO:teuthology.orchestra.run.vm03.stdout: Installing : ceph-mgr-diskprediction-local-2:19.2.3-678.ge911 123/138 2026-03-09T15:11:27.309 INFO:teuthology.orchestra.run.vm03.stdout: Running scriptlet: ceph-mgr-diskprediction-local-2:19.2.3-678.ge911 123/138 2026-03-09T15:11:27.372 INFO:teuthology.orchestra.run.vm03.stdout: Installing : ceph-mgr-modules-core-2:19.2.3-678.ge911bdeb.el9 124/138 2026-03-09T15:11:27.375 INFO:teuthology.orchestra.run.vm03.stdout: Installing : ceph-mgr-2:19.2.3-678.ge911bdeb.el9.x86_64 125/138 2026-03-09T15:11:27.400 INFO:teuthology.orchestra.run.vm03.stdout: Running scriptlet: ceph-mgr-2:19.2.3-678.ge911bdeb.el9.x86_64 125/138 2026-03-09T15:11:27.400 INFO:teuthology.orchestra.run.vm03.stdout:Glob pattern passed to enable, but globs are not supported for this. 2026-03-09T15:11:27.400 INFO:teuthology.orchestra.run.vm03.stdout:Invalid unit name "ceph-mgr@*.service" escaped as "ceph-mgr@\x2a.service". 2026-03-09T15:11:27.400 INFO:teuthology.orchestra.run.vm03.stdout:Created symlink /etc/systemd/system/multi-user.target.wants/ceph-mgr.target → /usr/lib/systemd/system/ceph-mgr.target. 2026-03-09T15:11:27.400 INFO:teuthology.orchestra.run.vm03.stdout:Created symlink /etc/systemd/system/ceph.target.wants/ceph-mgr.target → /usr/lib/systemd/system/ceph-mgr.target. 2026-03-09T15:11:27.400 INFO:teuthology.orchestra.run.vm03.stdout: 2026-03-09T15:11:27.414 INFO:teuthology.orchestra.run.vm03.stdout: Installing : ceph-mgr-rook-2:19.2.3-678.ge911bdeb.el9.noarch 126/138 2026-03-09T15:11:27.426 INFO:teuthology.orchestra.run.vm03.stdout: Running scriptlet: ceph-mgr-rook-2:19.2.3-678.ge911bdeb.el9.noarch 126/138 2026-03-09T15:11:27.494 INFO:teuthology.orchestra.run.vm04.stdout: Installing : flexiblas-netlib-3.0.4-9.el9.x86_64 80/138 2026-03-09T15:11:27.585 INFO:teuthology.orchestra.run.vm04.stdout: Installing : python3-numpy-1:1.23.5-2.el9.x86_64 81/138 2026-03-09T15:11:27.942 INFO:teuthology.orchestra.run.vm03.stdout: Installing : ceph-2:19.2.3-678.ge911bdeb.el9.x86_64 127/138 2026-03-09T15:11:27.946 INFO:teuthology.orchestra.run.vm03.stdout: Installing : ceph-radosgw-2:19.2.3-678.ge911bdeb.el9.x86_64 128/138 2026-03-09T15:11:27.967 INFO:teuthology.orchestra.run.vm03.stdout: Running scriptlet: ceph-radosgw-2:19.2.3-678.ge911bdeb.el9.x86_64 128/138 2026-03-09T15:11:27.967 INFO:teuthology.orchestra.run.vm03.stdout:Glob pattern passed to enable, but globs are not supported for this. 2026-03-09T15:11:27.967 INFO:teuthology.orchestra.run.vm03.stdout:Invalid unit name "ceph-radosgw@*.service" escaped as "ceph-radosgw@\x2a.service". 2026-03-09T15:11:27.967 INFO:teuthology.orchestra.run.vm03.stdout:Created symlink /etc/systemd/system/multi-user.target.wants/ceph-radosgw.target → /usr/lib/systemd/system/ceph-radosgw.target. 2026-03-09T15:11:27.967 INFO:teuthology.orchestra.run.vm03.stdout:Created symlink /etc/systemd/system/ceph.target.wants/ceph-radosgw.target → /usr/lib/systemd/system/ceph-radosgw.target. 2026-03-09T15:11:27.967 INFO:teuthology.orchestra.run.vm03.stdout: 2026-03-09T15:11:27.978 INFO:teuthology.orchestra.run.vm03.stdout: Installing : ceph-immutable-object-cache-2:19.2.3-678.ge911bd 129/138 2026-03-09T15:11:28.002 INFO:teuthology.orchestra.run.vm03.stdout: Running scriptlet: ceph-immutable-object-cache-2:19.2.3-678.ge911bd 129/138 2026-03-09T15:11:28.002 INFO:teuthology.orchestra.run.vm03.stdout:Glob pattern passed to enable, but globs are not supported for this. 2026-03-09T15:11:28.002 INFO:teuthology.orchestra.run.vm03.stdout:Invalid unit name "ceph-immutable-object-cache@*.service" escaped as "ceph-immutable-object-cache@\x2a.service". 2026-03-09T15:11:28.002 INFO:teuthology.orchestra.run.vm03.stdout: 2026-03-09T15:11:28.164 INFO:teuthology.orchestra.run.vm03.stdout: Installing : rbd-mirror-2:19.2.3-678.ge911bdeb.el9.x86_64 130/138 2026-03-09T15:11:28.187 INFO:teuthology.orchestra.run.vm03.stdout: Running scriptlet: rbd-mirror-2:19.2.3-678.ge911bdeb.el9.x86_64 130/138 2026-03-09T15:11:28.188 INFO:teuthology.orchestra.run.vm03.stdout:Glob pattern passed to enable, but globs are not supported for this. 2026-03-09T15:11:28.188 INFO:teuthology.orchestra.run.vm03.stdout:Invalid unit name "ceph-rbd-mirror@*.service" escaped as "ceph-rbd-mirror@\x2a.service". 2026-03-09T15:11:28.188 INFO:teuthology.orchestra.run.vm03.stdout:Created symlink /etc/systemd/system/multi-user.target.wants/ceph-rbd-mirror.target → /usr/lib/systemd/system/ceph-rbd-mirror.target. 2026-03-09T15:11:28.188 INFO:teuthology.orchestra.run.vm03.stdout:Created symlink /etc/systemd/system/ceph.target.wants/ceph-rbd-mirror.target → /usr/lib/systemd/system/ceph-rbd-mirror.target. 2026-03-09T15:11:28.188 INFO:teuthology.orchestra.run.vm03.stdout: 2026-03-09T15:11:28.384 INFO:teuthology.orchestra.run.vm04.stdout: Installing : python3-numpy-f2py-1:1.23.5-2.el9.x86_64 82/138 2026-03-09T15:11:28.413 INFO:teuthology.orchestra.run.vm04.stdout: Installing : python3-scipy-1.9.3-2.el9.x86_64 83/138 2026-03-09T15:11:28.419 INFO:teuthology.orchestra.run.vm04.stdout: Installing : libxslt-1.1.34-12.el9.x86_64 84/138 2026-03-09T15:11:28.424 INFO:teuthology.orchestra.run.vm04.stdout: Installing : xmlstarlet-1.6.1-20.el9.x86_64 85/138 2026-03-09T15:11:28.583 INFO:teuthology.orchestra.run.vm04.stdout: Installing : libpmemobj-1.12.1-1.el9.x86_64 86/138 2026-03-09T15:11:28.586 INFO:teuthology.orchestra.run.vm04.stdout: Upgrading : librbd1-2:19.2.3-678.ge911bdeb.el9.x86_64 87/138 2026-03-09T15:11:28.619 INFO:teuthology.orchestra.run.vm04.stdout: Running scriptlet: librbd1-2:19.2.3-678.ge911bdeb.el9.x86_64 87/138 2026-03-09T15:11:28.623 INFO:teuthology.orchestra.run.vm04.stdout: Installing : python3-rbd-2:19.2.3-678.ge911bdeb.el9.x86_64 88/138 2026-03-09T15:11:28.632 INFO:teuthology.orchestra.run.vm04.stdout: Installing : boost-program-options-1.75.0-13.el9.x86_64 89/138 2026-03-09T15:11:28.898 INFO:teuthology.orchestra.run.vm04.stdout: Installing : parquet-libs-9.0.0-15.el9.x86_64 90/138 2026-03-09T15:11:28.900 INFO:teuthology.orchestra.run.vm04.stdout: Installing : librgw2-2:19.2.3-678.ge911bdeb.el9.x86_64 91/138 2026-03-09T15:11:28.920 INFO:teuthology.orchestra.run.vm04.stdout: Running scriptlet: librgw2-2:19.2.3-678.ge911bdeb.el9.x86_64 91/138 2026-03-09T15:11:28.922 INFO:teuthology.orchestra.run.vm04.stdout: Installing : python3-rgw-2:19.2.3-678.ge911bdeb.el9.x86_64 92/138 2026-03-09T15:11:30.053 INFO:teuthology.orchestra.run.vm04.stdout: Running scriptlet: ceph-common-2:19.2.3-678.ge911bdeb.el9.x86_64 93/138 2026-03-09T15:11:30.058 INFO:teuthology.orchestra.run.vm04.stdout: Installing : ceph-common-2:19.2.3-678.ge911bdeb.el9.x86_64 93/138 2026-03-09T15:11:30.081 INFO:teuthology.orchestra.run.vm04.stdout: Running scriptlet: ceph-common-2:19.2.3-678.ge911bdeb.el9.x86_64 93/138 2026-03-09T15:11:30.097 INFO:teuthology.orchestra.run.vm04.stdout: Installing : python3-ply-3.11-14.el9.noarch 94/138 2026-03-09T15:11:30.117 INFO:teuthology.orchestra.run.vm04.stdout: Installing : python3-pycparser-2.20-6.el9.noarch 95/138 2026-03-09T15:11:30.208 INFO:teuthology.orchestra.run.vm04.stdout: Installing : python3-cffi-1.14.5-5.el9.x86_64 96/138 2026-03-09T15:11:30.222 INFO:teuthology.orchestra.run.vm04.stdout: Installing : python3-cryptography-36.0.1-5.el9.x86_64 97/138 2026-03-09T15:11:30.253 INFO:teuthology.orchestra.run.vm04.stdout: Installing : python3-pyOpenSSL-21.0.0-1.el9.noarch 98/138 2026-03-09T15:11:30.295 INFO:teuthology.orchestra.run.vm04.stdout: Installing : python3-cheroot-10.0.1-4.el9.noarch 99/138 2026-03-09T15:11:30.355 INFO:teuthology.orchestra.run.vm04.stdout: Installing : python3-cherrypy-18.6.1-2.el9.noarch 100/138 2026-03-09T15:11:30.364 INFO:teuthology.orchestra.run.vm04.stdout: Installing : python3-asyncssh-2.13.2-5.el9.noarch 101/138 2026-03-09T15:11:30.370 INFO:teuthology.orchestra.run.vm04.stdout: Installing : python3-bcrypt-3.2.2-1.el9.x86_64 102/138 2026-03-09T15:11:30.376 INFO:teuthology.orchestra.run.vm04.stdout: Installing : pciutils-3.7.0-7.el9.x86_64 103/138 2026-03-09T15:11:30.380 INFO:teuthology.orchestra.run.vm04.stdout: Installing : qatlib-25.08.0-2.el9.x86_64 104/138 2026-03-09T15:11:30.382 INFO:teuthology.orchestra.run.vm04.stdout: Installing : qatlib-service-25.08.0-2.el9.x86_64 105/138 2026-03-09T15:11:30.400 INFO:teuthology.orchestra.run.vm04.stdout: Running scriptlet: qatlib-service-25.08.0-2.el9.x86_64 105/138 2026-03-09T15:11:30.710 INFO:teuthology.orchestra.run.vm04.stdout: Installing : qatzip-libs-1.3.1-1.el9.x86_64 106/138 2026-03-09T15:11:30.716 INFO:teuthology.orchestra.run.vm04.stdout: Installing : ceph-base-2:19.2.3-678.ge911bdeb.el9.x86_64 107/138 2026-03-09T15:11:30.758 INFO:teuthology.orchestra.run.vm04.stdout: Running scriptlet: ceph-base-2:19.2.3-678.ge911bdeb.el9.x86_64 107/138 2026-03-09T15:11:30.758 INFO:teuthology.orchestra.run.vm04.stdout:Created symlink /etc/systemd/system/multi-user.target.wants/ceph.target → /usr/lib/systemd/system/ceph.target. 2026-03-09T15:11:30.758 INFO:teuthology.orchestra.run.vm04.stdout:Created symlink /etc/systemd/system/ceph.target.wants/ceph-crash.service → /usr/lib/systemd/system/ceph-crash.service. 2026-03-09T15:11:30.758 INFO:teuthology.orchestra.run.vm04.stdout: 2026-03-09T15:11:30.763 INFO:teuthology.orchestra.run.vm04.stdout: Installing : ceph-selinux-2:19.2.3-678.ge911bdeb.el9.x86_64 108/138 2026-03-09T15:11:30.772 INFO:teuthology.orchestra.run.vm03.stdout: Installing : ceph-test-2:19.2.3-678.ge911bdeb.el9.x86_64 131/138 2026-03-09T15:11:30.783 INFO:teuthology.orchestra.run.vm03.stdout: Installing : rbd-fuse-2:19.2.3-678.ge911bdeb.el9.x86_64 132/138 2026-03-09T15:11:30.789 INFO:teuthology.orchestra.run.vm03.stdout: Installing : rbd-nbd-2:19.2.3-678.ge911bdeb.el9.x86_64 133/138 2026-03-09T15:11:30.848 INFO:teuthology.orchestra.run.vm03.stdout: Installing : libcephfs-devel-2:19.2.3-678.ge911bdeb.el9.x86_6 134/138 2026-03-09T15:11:30.857 INFO:teuthology.orchestra.run.vm03.stdout: Installing : ceph-fuse-2:19.2.3-678.ge911bdeb.el9.x86_64 135/138 2026-03-09T15:11:30.860 INFO:teuthology.orchestra.run.vm03.stdout: Installing : python3-jmespath-1.0.1-1.el9.noarch 136/138 2026-03-09T15:11:30.860 INFO:teuthology.orchestra.run.vm03.stdout: Cleanup : librbd1-2:16.2.4-5.el9.x86_64 137/138 2026-03-09T15:11:30.875 INFO:teuthology.orchestra.run.vm03.stdout: Running scriptlet: librbd1-2:16.2.4-5.el9.x86_64 137/138 2026-03-09T15:11:30.875 INFO:teuthology.orchestra.run.vm03.stdout: Cleanup : librados2-2:16.2.4-5.el9.x86_64 138/138 2026-03-09T15:11:32.528 INFO:teuthology.orchestra.run.vm03.stdout: Running scriptlet: librados2-2:16.2.4-5.el9.x86_64 138/138 2026-03-09T15:11:32.528 INFO:teuthology.orchestra.run.vm03.stdout: Verifying : ceph-2:19.2.3-678.ge911bdeb.el9.x86_64 1/138 2026-03-09T15:11:32.528 INFO:teuthology.orchestra.run.vm03.stdout: Verifying : ceph-base-2:19.2.3-678.ge911bdeb.el9.x86_64 2/138 2026-03-09T15:11:32.528 INFO:teuthology.orchestra.run.vm03.stdout: Verifying : ceph-common-2:19.2.3-678.ge911bdeb.el9.x86_64 3/138 2026-03-09T15:11:32.528 INFO:teuthology.orchestra.run.vm03.stdout: Verifying : ceph-fuse-2:19.2.3-678.ge911bdeb.el9.x86_64 4/138 2026-03-09T15:11:32.528 INFO:teuthology.orchestra.run.vm03.stdout: Verifying : ceph-immutable-object-cache-2:19.2.3-678.ge911bd 5/138 2026-03-09T15:11:32.528 INFO:teuthology.orchestra.run.vm03.stdout: Verifying : ceph-mds-2:19.2.3-678.ge911bdeb.el9.x86_64 6/138 2026-03-09T15:11:32.528 INFO:teuthology.orchestra.run.vm03.stdout: Verifying : ceph-mgr-2:19.2.3-678.ge911bdeb.el9.x86_64 7/138 2026-03-09T15:11:32.528 INFO:teuthology.orchestra.run.vm03.stdout: Verifying : ceph-mon-2:19.2.3-678.ge911bdeb.el9.x86_64 8/138 2026-03-09T15:11:32.528 INFO:teuthology.orchestra.run.vm03.stdout: Verifying : ceph-osd-2:19.2.3-678.ge911bdeb.el9.x86_64 9/138 2026-03-09T15:11:32.528 INFO:teuthology.orchestra.run.vm03.stdout: Verifying : ceph-radosgw-2:19.2.3-678.ge911bdeb.el9.x86_64 10/138 2026-03-09T15:11:32.528 INFO:teuthology.orchestra.run.vm03.stdout: Verifying : ceph-selinux-2:19.2.3-678.ge911bdeb.el9.x86_64 11/138 2026-03-09T15:11:32.528 INFO:teuthology.orchestra.run.vm03.stdout: Verifying : ceph-test-2:19.2.3-678.ge911bdeb.el9.x86_64 12/138 2026-03-09T15:11:32.528 INFO:teuthology.orchestra.run.vm03.stdout: Verifying : libcephfs-devel-2:19.2.3-678.ge911bdeb.el9.x86_6 13/138 2026-03-09T15:11:32.528 INFO:teuthology.orchestra.run.vm03.stdout: Verifying : libcephfs2-2:19.2.3-678.ge911bdeb.el9.x86_64 14/138 2026-03-09T15:11:32.528 INFO:teuthology.orchestra.run.vm03.stdout: Verifying : libcephsqlite-2:19.2.3-678.ge911bdeb.el9.x86_64 15/138 2026-03-09T15:11:32.528 INFO:teuthology.orchestra.run.vm03.stdout: Verifying : librados-devel-2:19.2.3-678.ge911bdeb.el9.x86_64 16/138 2026-03-09T15:11:32.528 INFO:teuthology.orchestra.run.vm03.stdout: Verifying : libradosstriper1-2:19.2.3-678.ge911bdeb.el9.x86_ 17/138 2026-03-09T15:11:32.529 INFO:teuthology.orchestra.run.vm03.stdout: Verifying : librgw2-2:19.2.3-678.ge911bdeb.el9.x86_64 18/138 2026-03-09T15:11:32.529 INFO:teuthology.orchestra.run.vm03.stdout: Verifying : python3-ceph-argparse-2:19.2.3-678.ge911bdeb.el9 19/138 2026-03-09T15:11:32.529 INFO:teuthology.orchestra.run.vm03.stdout: Verifying : python3-ceph-common-2:19.2.3-678.ge911bdeb.el9.x 20/138 2026-03-09T15:11:32.529 INFO:teuthology.orchestra.run.vm03.stdout: Verifying : python3-cephfs-2:19.2.3-678.ge911bdeb.el9.x86_64 21/138 2026-03-09T15:11:32.529 INFO:teuthology.orchestra.run.vm03.stdout: Verifying : python3-rados-2:19.2.3-678.ge911bdeb.el9.x86_64 22/138 2026-03-09T15:11:32.529 INFO:teuthology.orchestra.run.vm03.stdout: Verifying : python3-rbd-2:19.2.3-678.ge911bdeb.el9.x86_64 23/138 2026-03-09T15:11:32.529 INFO:teuthology.orchestra.run.vm03.stdout: Verifying : python3-rgw-2:19.2.3-678.ge911bdeb.el9.x86_64 24/138 2026-03-09T15:11:32.529 INFO:teuthology.orchestra.run.vm03.stdout: Verifying : rbd-fuse-2:19.2.3-678.ge911bdeb.el9.x86_64 25/138 2026-03-09T15:11:32.529 INFO:teuthology.orchestra.run.vm03.stdout: Verifying : rbd-mirror-2:19.2.3-678.ge911bdeb.el9.x86_64 26/138 2026-03-09T15:11:32.529 INFO:teuthology.orchestra.run.vm03.stdout: Verifying : rbd-nbd-2:19.2.3-678.ge911bdeb.el9.x86_64 27/138 2026-03-09T15:11:32.529 INFO:teuthology.orchestra.run.vm03.stdout: Verifying : ceph-grafana-dashboards-2:19.2.3-678.ge911bdeb.e 28/138 2026-03-09T15:11:32.529 INFO:teuthology.orchestra.run.vm03.stdout: Verifying : ceph-mgr-cephadm-2:19.2.3-678.ge911bdeb.el9.noar 29/138 2026-03-09T15:11:32.529 INFO:teuthology.orchestra.run.vm03.stdout: Verifying : ceph-mgr-dashboard-2:19.2.3-678.ge911bdeb.el9.no 30/138 2026-03-09T15:11:32.529 INFO:teuthology.orchestra.run.vm03.stdout: Verifying : ceph-mgr-diskprediction-local-2:19.2.3-678.ge911 31/138 2026-03-09T15:11:32.529 INFO:teuthology.orchestra.run.vm03.stdout: Verifying : ceph-mgr-modules-core-2:19.2.3-678.ge911bdeb.el9 32/138 2026-03-09T15:11:32.529 INFO:teuthology.orchestra.run.vm03.stdout: Verifying : ceph-mgr-rook-2:19.2.3-678.ge911bdeb.el9.noarch 33/138 2026-03-09T15:11:32.529 INFO:teuthology.orchestra.run.vm03.stdout: Verifying : ceph-prometheus-alerts-2:19.2.3-678.ge911bdeb.el 34/138 2026-03-09T15:11:32.529 INFO:teuthology.orchestra.run.vm03.stdout: Verifying : ceph-volume-2:19.2.3-678.ge911bdeb.el9.noarch 35/138 2026-03-09T15:11:32.529 INFO:teuthology.orchestra.run.vm03.stdout: Verifying : cephadm-2:19.2.3-678.ge911bdeb.el9.noarch 36/138 2026-03-09T15:11:32.529 INFO:teuthology.orchestra.run.vm03.stdout: Verifying : cryptsetup-2.8.1-3.el9.x86_64 37/138 2026-03-09T15:11:32.529 INFO:teuthology.orchestra.run.vm03.stdout: Verifying : ledmon-libs-1.1.0-3.el9.x86_64 38/138 2026-03-09T15:11:32.529 INFO:teuthology.orchestra.run.vm03.stdout: Verifying : libconfig-1.7.2-9.el9.x86_64 39/138 2026-03-09T15:11:32.529 INFO:teuthology.orchestra.run.vm03.stdout: Verifying : libgfortran-11.5.0-14.el9.x86_64 40/138 2026-03-09T15:11:32.529 INFO:teuthology.orchestra.run.vm03.stdout: Verifying : libquadmath-11.5.0-14.el9.x86_64 41/138 2026-03-09T15:11:32.529 INFO:teuthology.orchestra.run.vm03.stdout: Verifying : mailcap-2.1.49-5.el9.noarch 42/138 2026-03-09T15:11:32.529 INFO:teuthology.orchestra.run.vm03.stdout: Verifying : pciutils-3.7.0-7.el9.x86_64 43/138 2026-03-09T15:11:32.529 INFO:teuthology.orchestra.run.vm03.stdout: Verifying : python3-cffi-1.14.5-5.el9.x86_64 44/138 2026-03-09T15:11:32.529 INFO:teuthology.orchestra.run.vm03.stdout: Verifying : python3-cryptography-36.0.1-5.el9.x86_64 45/138 2026-03-09T15:11:32.530 INFO:teuthology.orchestra.run.vm03.stdout: Verifying : python3-ply-3.11-14.el9.noarch 46/138 2026-03-09T15:11:32.531 INFO:teuthology.orchestra.run.vm03.stdout: Verifying : python3-pycparser-2.20-6.el9.noarch 47/138 2026-03-09T15:11:32.531 INFO:teuthology.orchestra.run.vm03.stdout: Verifying : python3-requests-2.25.1-10.el9.noarch 48/138 2026-03-09T15:11:32.531 INFO:teuthology.orchestra.run.vm03.stdout: Verifying : python3-urllib3-1.26.5-7.el9.noarch 49/138 2026-03-09T15:11:32.531 INFO:teuthology.orchestra.run.vm03.stdout: Verifying : unzip-6.0-59.el9.x86_64 50/138 2026-03-09T15:11:32.531 INFO:teuthology.orchestra.run.vm03.stdout: Verifying : zip-3.0-35.el9.x86_64 51/138 2026-03-09T15:11:32.531 INFO:teuthology.orchestra.run.vm03.stdout: Verifying : boost-program-options-1.75.0-13.el9.x86_64 52/138 2026-03-09T15:11:32.531 INFO:teuthology.orchestra.run.vm03.stdout: Verifying : flexiblas-3.0.4-9.el9.x86_64 53/138 2026-03-09T15:11:32.531 INFO:teuthology.orchestra.run.vm03.stdout: Verifying : flexiblas-netlib-3.0.4-9.el9.x86_64 54/138 2026-03-09T15:11:32.531 INFO:teuthology.orchestra.run.vm03.stdout: Verifying : flexiblas-openblas-openmp-3.0.4-9.el9.x86_64 55/138 2026-03-09T15:11:32.531 INFO:teuthology.orchestra.run.vm03.stdout: Verifying : libnbd-1.20.3-4.el9.x86_64 56/138 2026-03-09T15:11:32.531 INFO:teuthology.orchestra.run.vm03.stdout: Verifying : libpmemobj-1.12.1-1.el9.x86_64 57/138 2026-03-09T15:11:32.531 INFO:teuthology.orchestra.run.vm03.stdout: Verifying : librabbitmq-0.11.0-7.el9.x86_64 58/138 2026-03-09T15:11:32.531 INFO:teuthology.orchestra.run.vm03.stdout: Verifying : librdkafka-1.6.1-102.el9.x86_64 59/138 2026-03-09T15:11:32.531 INFO:teuthology.orchestra.run.vm03.stdout: Verifying : libstoragemgmt-1.10.1-1.el9.x86_64 60/138 2026-03-09T15:11:32.531 INFO:teuthology.orchestra.run.vm03.stdout: Verifying : libxslt-1.1.34-12.el9.x86_64 61/138 2026-03-09T15:11:32.531 INFO:teuthology.orchestra.run.vm03.stdout: Verifying : lttng-ust-2.12.0-6.el9.x86_64 62/138 2026-03-09T15:11:32.531 INFO:teuthology.orchestra.run.vm03.stdout: Verifying : lua-5.4.4-4.el9.x86_64 63/138 2026-03-09T15:11:32.531 INFO:teuthology.orchestra.run.vm03.stdout: Verifying : openblas-0.3.29-1.el9.x86_64 64/138 2026-03-09T15:11:32.531 INFO:teuthology.orchestra.run.vm03.stdout: Verifying : openblas-openmp-0.3.29-1.el9.x86_64 65/138 2026-03-09T15:11:32.531 INFO:teuthology.orchestra.run.vm03.stdout: Verifying : protobuf-3.14.0-17.el9.x86_64 66/138 2026-03-09T15:11:32.531 INFO:teuthology.orchestra.run.vm03.stdout: Verifying : python3-babel-2.9.1-2.el9.noarch 67/138 2026-03-09T15:11:32.531 INFO:teuthology.orchestra.run.vm03.stdout: Verifying : python3-devel-3.9.25-3.el9.x86_64 68/138 2026-03-09T15:11:32.531 INFO:teuthology.orchestra.run.vm03.stdout: Verifying : python3-jinja2-2.11.3-8.el9.noarch 69/138 2026-03-09T15:11:32.531 INFO:teuthology.orchestra.run.vm03.stdout: Verifying : python3-jmespath-1.0.1-1.el9.noarch 70/138 2026-03-09T15:11:32.531 INFO:teuthology.orchestra.run.vm03.stdout: Verifying : python3-libstoragemgmt-1.10.1-1.el9.x86_64 71/138 2026-03-09T15:11:32.531 INFO:teuthology.orchestra.run.vm03.stdout: Verifying : python3-mako-1.1.4-6.el9.noarch 72/138 2026-03-09T15:11:32.531 INFO:teuthology.orchestra.run.vm03.stdout: Verifying : python3-markupsafe-1.1.1-12.el9.x86_64 73/138 2026-03-09T15:11:32.531 INFO:teuthology.orchestra.run.vm03.stdout: Verifying : python3-numpy-1:1.23.5-2.el9.x86_64 74/138 2026-03-09T15:11:32.531 INFO:teuthology.orchestra.run.vm03.stdout: Verifying : python3-numpy-f2py-1:1.23.5-2.el9.x86_64 75/138 2026-03-09T15:11:32.531 INFO:teuthology.orchestra.run.vm03.stdout: Verifying : python3-packaging-20.9-5.el9.noarch 76/138 2026-03-09T15:11:32.531 INFO:teuthology.orchestra.run.vm03.stdout: Verifying : python3-protobuf-3.14.0-17.el9.noarch 77/138 2026-03-09T15:11:32.531 INFO:teuthology.orchestra.run.vm03.stdout: Verifying : python3-pyasn1-0.4.8-7.el9.noarch 78/138 2026-03-09T15:11:32.531 INFO:teuthology.orchestra.run.vm03.stdout: Verifying : python3-pyasn1-modules-0.4.8-7.el9.noarch 79/138 2026-03-09T15:11:32.531 INFO:teuthology.orchestra.run.vm03.stdout: Verifying : python3-requests-oauthlib-1.3.0-12.el9.noarch 80/138 2026-03-09T15:11:32.531 INFO:teuthology.orchestra.run.vm03.stdout: Verifying : python3-scipy-1.9.3-2.el9.x86_64 81/138 2026-03-09T15:11:32.533 INFO:teuthology.orchestra.run.vm03.stdout: Verifying : python3-toml-0.10.2-6.el9.noarch 82/138 2026-03-09T15:11:32.533 INFO:teuthology.orchestra.run.vm03.stdout: Verifying : qatlib-25.08.0-2.el9.x86_64 83/138 2026-03-09T15:11:32.533 INFO:teuthology.orchestra.run.vm03.stdout: Verifying : qatlib-service-25.08.0-2.el9.x86_64 84/138 2026-03-09T15:11:32.533 INFO:teuthology.orchestra.run.vm03.stdout: Verifying : qatzip-libs-1.3.1-1.el9.x86_64 85/138 2026-03-09T15:11:32.533 INFO:teuthology.orchestra.run.vm03.stdout: Verifying : socat-1.7.4.1-8.el9.x86_64 86/138 2026-03-09T15:11:32.533 INFO:teuthology.orchestra.run.vm03.stdout: Verifying : xmlstarlet-1.6.1-20.el9.x86_64 87/138 2026-03-09T15:11:32.533 INFO:teuthology.orchestra.run.vm03.stdout: Verifying : lua-devel-5.4.4-4.el9.x86_64 88/138 2026-03-09T15:11:32.533 INFO:teuthology.orchestra.run.vm03.stdout: Verifying : protobuf-compiler-3.14.0-17.el9.x86_64 89/138 2026-03-09T15:11:32.533 INFO:teuthology.orchestra.run.vm03.stdout: Verifying : abseil-cpp-20211102.0-4.el9.x86_64 90/138 2026-03-09T15:11:32.533 INFO:teuthology.orchestra.run.vm03.stdout: Verifying : gperftools-libs-2.9.1-3.el9.x86_64 91/138 2026-03-09T15:11:32.533 INFO:teuthology.orchestra.run.vm03.stdout: Verifying : grpc-data-1.46.7-10.el9.noarch 92/138 2026-03-09T15:11:32.533 INFO:teuthology.orchestra.run.vm03.stdout: Verifying : libarrow-9.0.0-15.el9.x86_64 93/138 2026-03-09T15:11:32.533 INFO:teuthology.orchestra.run.vm03.stdout: Verifying : libarrow-doc-9.0.0-15.el9.noarch 94/138 2026-03-09T15:11:32.533 INFO:teuthology.orchestra.run.vm03.stdout: Verifying : liboath-2.6.12-1.el9.x86_64 95/138 2026-03-09T15:11:32.533 INFO:teuthology.orchestra.run.vm03.stdout: Verifying : libunwind-1.6.2-1.el9.x86_64 96/138 2026-03-09T15:11:32.533 INFO:teuthology.orchestra.run.vm03.stdout: Verifying : luarocks-3.9.2-5.el9.noarch 97/138 2026-03-09T15:11:32.533 INFO:teuthology.orchestra.run.vm03.stdout: Verifying : parquet-libs-9.0.0-15.el9.x86_64 98/138 2026-03-09T15:11:32.533 INFO:teuthology.orchestra.run.vm03.stdout: Verifying : python3-asyncssh-2.13.2-5.el9.noarch 99/138 2026-03-09T15:11:32.533 INFO:teuthology.orchestra.run.vm03.stdout: Verifying : python3-autocommand-2.2.2-8.el9.noarch 100/138 2026-03-09T15:11:32.533 INFO:teuthology.orchestra.run.vm03.stdout: Verifying : python3-backports-tarfile-1.2.0-1.el9.noarch 101/138 2026-03-09T15:11:32.533 INFO:teuthology.orchestra.run.vm03.stdout: Verifying : python3-bcrypt-3.2.2-1.el9.x86_64 102/138 2026-03-09T15:11:32.533 INFO:teuthology.orchestra.run.vm03.stdout: Verifying : python3-cachetools-4.2.4-1.el9.noarch 103/138 2026-03-09T15:11:32.533 INFO:teuthology.orchestra.run.vm03.stdout: Verifying : python3-certifi-2023.05.07-4.el9.noarch 104/138 2026-03-09T15:11:32.533 INFO:teuthology.orchestra.run.vm03.stdout: Verifying : python3-cheroot-10.0.1-4.el9.noarch 105/138 2026-03-09T15:11:32.533 INFO:teuthology.orchestra.run.vm03.stdout: Verifying : python3-cherrypy-18.6.1-2.el9.noarch 106/138 2026-03-09T15:11:32.533 INFO:teuthology.orchestra.run.vm03.stdout: Verifying : python3-google-auth-1:2.45.0-1.el9.noarch 107/138 2026-03-09T15:11:32.533 INFO:teuthology.orchestra.run.vm03.stdout: Verifying : python3-grpcio-1.46.7-10.el9.x86_64 108/138 2026-03-09T15:11:32.533 INFO:teuthology.orchestra.run.vm03.stdout: Verifying : python3-grpcio-tools-1.46.7-10.el9.x86_64 109/138 2026-03-09T15:11:32.533 INFO:teuthology.orchestra.run.vm03.stdout: Verifying : python3-jaraco-8.2.1-3.el9.noarch 110/138 2026-03-09T15:11:32.533 INFO:teuthology.orchestra.run.vm03.stdout: Verifying : python3-jaraco-classes-3.2.1-5.el9.noarch 111/138 2026-03-09T15:11:32.533 INFO:teuthology.orchestra.run.vm03.stdout: Verifying : python3-jaraco-collections-3.0.0-8.el9.noarch 112/138 2026-03-09T15:11:32.533 INFO:teuthology.orchestra.run.vm03.stdout: Verifying : python3-jaraco-context-6.0.1-3.el9.noarch 113/138 2026-03-09T15:11:32.533 INFO:teuthology.orchestra.run.vm03.stdout: Verifying : python3-jaraco-functools-3.5.0-2.el9.noarch 114/138 2026-03-09T15:11:32.533 INFO:teuthology.orchestra.run.vm03.stdout: Verifying : python3-jaraco-text-4.0.0-2.el9.noarch 115/138 2026-03-09T15:11:32.533 INFO:teuthology.orchestra.run.vm03.stdout: Verifying : python3-kubernetes-1:26.1.0-3.el9.noarch 116/138 2026-03-09T15:11:32.533 INFO:teuthology.orchestra.run.vm03.stdout: Verifying : python3-logutils-0.3.5-21.el9.noarch 117/138 2026-03-09T15:11:32.533 INFO:teuthology.orchestra.run.vm03.stdout: Verifying : python3-more-itertools-8.12.0-2.el9.noarch 118/138 2026-03-09T15:11:32.533 INFO:teuthology.orchestra.run.vm03.stdout: Verifying : python3-natsort-7.1.1-5.el9.noarch 119/138 2026-03-09T15:11:32.533 INFO:teuthology.orchestra.run.vm03.stdout: Verifying : python3-pecan-1.4.2-3.el9.noarch 120/138 2026-03-09T15:11:32.533 INFO:teuthology.orchestra.run.vm03.stdout: Verifying : python3-portend-3.1.0-2.el9.noarch 121/138 2026-03-09T15:11:32.533 INFO:teuthology.orchestra.run.vm03.stdout: Verifying : python3-pyOpenSSL-21.0.0-1.el9.noarch 122/138 2026-03-09T15:11:32.533 INFO:teuthology.orchestra.run.vm03.stdout: Verifying : python3-repoze-lru-0.7-16.el9.noarch 123/138 2026-03-09T15:11:32.533 INFO:teuthology.orchestra.run.vm03.stdout: Verifying : python3-routes-2.5.1-5.el9.noarch 124/138 2026-03-09T15:11:32.533 INFO:teuthology.orchestra.run.vm03.stdout: Verifying : python3-rsa-4.9-2.el9.noarch 125/138 2026-03-09T15:11:32.533 INFO:teuthology.orchestra.run.vm03.stdout: Verifying : python3-tempora-5.0.0-2.el9.noarch 126/138 2026-03-09T15:11:32.533 INFO:teuthology.orchestra.run.vm03.stdout: Verifying : python3-typing-extensions-4.15.0-1.el9.noarch 127/138 2026-03-09T15:11:32.533 INFO:teuthology.orchestra.run.vm03.stdout: Verifying : python3-webob-1.8.8-2.el9.noarch 128/138 2026-03-09T15:11:32.533 INFO:teuthology.orchestra.run.vm03.stdout: Verifying : python3-websocket-client-1.2.3-2.el9.noarch 129/138 2026-03-09T15:11:32.533 INFO:teuthology.orchestra.run.vm03.stdout: Verifying : python3-werkzeug-2.0.3-3.el9.1.noarch 130/138 2026-03-09T15:11:32.534 INFO:teuthology.orchestra.run.vm03.stdout: Verifying : python3-xmltodict-0.12.0-15.el9.noarch 131/138 2026-03-09T15:11:32.534 INFO:teuthology.orchestra.run.vm03.stdout: Verifying : python3-zc-lockfile-2.0-10.el9.noarch 132/138 2026-03-09T15:11:32.534 INFO:teuthology.orchestra.run.vm03.stdout: Verifying : re2-1:20211101-20.el9.x86_64 133/138 2026-03-09T15:11:32.534 INFO:teuthology.orchestra.run.vm03.stdout: Verifying : thrift-0.15.0-4.el9.x86_64 134/138 2026-03-09T15:11:32.534 INFO:teuthology.orchestra.run.vm03.stdout: Verifying : librados2-2:19.2.3-678.ge911bdeb.el9.x86_64 135/138 2026-03-09T15:11:32.534 INFO:teuthology.orchestra.run.vm03.stdout: Verifying : librados2-2:16.2.4-5.el9.x86_64 136/138 2026-03-09T15:11:32.534 INFO:teuthology.orchestra.run.vm03.stdout: Verifying : librbd1-2:19.2.3-678.ge911bdeb.el9.x86_64 137/138 2026-03-09T15:11:32.961 INFO:teuthology.orchestra.run.vm03.stdout: Verifying : librbd1-2:16.2.4-5.el9.x86_64 138/138 2026-03-09T15:11:32.962 INFO:teuthology.orchestra.run.vm03.stdout: 2026-03-09T15:11:32.962 INFO:teuthology.orchestra.run.vm03.stdout:Upgraded: 2026-03-09T15:11:32.962 INFO:teuthology.orchestra.run.vm03.stdout: librados2-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T15:11:32.962 INFO:teuthology.orchestra.run.vm03.stdout: librbd1-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T15:11:32.962 INFO:teuthology.orchestra.run.vm03.stdout:Installed: 2026-03-09T15:11:32.962 INFO:teuthology.orchestra.run.vm03.stdout: abseil-cpp-20211102.0-4.el9.x86_64 2026-03-09T15:11:32.962 INFO:teuthology.orchestra.run.vm03.stdout: boost-program-options-1.75.0-13.el9.x86_64 2026-03-09T15:11:32.962 INFO:teuthology.orchestra.run.vm03.stdout: ceph-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T15:11:32.962 INFO:teuthology.orchestra.run.vm03.stdout: ceph-base-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T15:11:32.962 INFO:teuthology.orchestra.run.vm03.stdout: ceph-common-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T15:11:32.962 INFO:teuthology.orchestra.run.vm03.stdout: ceph-fuse-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T15:11:32.962 INFO:teuthology.orchestra.run.vm03.stdout: ceph-grafana-dashboards-2:19.2.3-678.ge911bdeb.el9.noarch 2026-03-09T15:11:32.962 INFO:teuthology.orchestra.run.vm03.stdout: ceph-immutable-object-cache-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T15:11:32.962 INFO:teuthology.orchestra.run.vm03.stdout: ceph-mds-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T15:11:32.962 INFO:teuthology.orchestra.run.vm03.stdout: ceph-mgr-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T15:11:32.962 INFO:teuthology.orchestra.run.vm03.stdout: ceph-mgr-cephadm-2:19.2.3-678.ge911bdeb.el9.noarch 2026-03-09T15:11:32.962 INFO:teuthology.orchestra.run.vm03.stdout: ceph-mgr-dashboard-2:19.2.3-678.ge911bdeb.el9.noarch 2026-03-09T15:11:32.962 INFO:teuthology.orchestra.run.vm03.stdout: ceph-mgr-diskprediction-local-2:19.2.3-678.ge911bdeb.el9.noarch 2026-03-09T15:11:32.962 INFO:teuthology.orchestra.run.vm03.stdout: ceph-mgr-modules-core-2:19.2.3-678.ge911bdeb.el9.noarch 2026-03-09T15:11:32.962 INFO:teuthology.orchestra.run.vm03.stdout: ceph-mgr-rook-2:19.2.3-678.ge911bdeb.el9.noarch 2026-03-09T15:11:32.962 INFO:teuthology.orchestra.run.vm03.stdout: ceph-mon-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T15:11:32.962 INFO:teuthology.orchestra.run.vm03.stdout: ceph-osd-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T15:11:32.962 INFO:teuthology.orchestra.run.vm03.stdout: ceph-prometheus-alerts-2:19.2.3-678.ge911bdeb.el9.noarch 2026-03-09T15:11:32.962 INFO:teuthology.orchestra.run.vm03.stdout: ceph-radosgw-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T15:11:32.962 INFO:teuthology.orchestra.run.vm03.stdout: ceph-selinux-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T15:11:32.962 INFO:teuthology.orchestra.run.vm03.stdout: ceph-test-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T15:11:32.962 INFO:teuthology.orchestra.run.vm03.stdout: ceph-volume-2:19.2.3-678.ge911bdeb.el9.noarch 2026-03-09T15:11:32.962 INFO:teuthology.orchestra.run.vm03.stdout: cephadm-2:19.2.3-678.ge911bdeb.el9.noarch 2026-03-09T15:11:32.962 INFO:teuthology.orchestra.run.vm03.stdout: cryptsetup-2.8.1-3.el9.x86_64 2026-03-09T15:11:32.962 INFO:teuthology.orchestra.run.vm03.stdout: flexiblas-3.0.4-9.el9.x86_64 2026-03-09T15:11:32.962 INFO:teuthology.orchestra.run.vm03.stdout: flexiblas-netlib-3.0.4-9.el9.x86_64 2026-03-09T15:11:32.962 INFO:teuthology.orchestra.run.vm03.stdout: flexiblas-openblas-openmp-3.0.4-9.el9.x86_64 2026-03-09T15:11:32.962 INFO:teuthology.orchestra.run.vm03.stdout: gperftools-libs-2.9.1-3.el9.x86_64 2026-03-09T15:11:32.962 INFO:teuthology.orchestra.run.vm03.stdout: grpc-data-1.46.7-10.el9.noarch 2026-03-09T15:11:32.962 INFO:teuthology.orchestra.run.vm03.stdout: ledmon-libs-1.1.0-3.el9.x86_64 2026-03-09T15:11:32.962 INFO:teuthology.orchestra.run.vm03.stdout: libarrow-9.0.0-15.el9.x86_64 2026-03-09T15:11:32.962 INFO:teuthology.orchestra.run.vm03.stdout: libarrow-doc-9.0.0-15.el9.noarch 2026-03-09T15:11:32.962 INFO:teuthology.orchestra.run.vm03.stdout: libcephfs-devel-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T15:11:32.962 INFO:teuthology.orchestra.run.vm03.stdout: libcephfs2-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T15:11:32.962 INFO:teuthology.orchestra.run.vm03.stdout: libcephsqlite-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T15:11:32.963 INFO:teuthology.orchestra.run.vm03.stdout: libconfig-1.7.2-9.el9.x86_64 2026-03-09T15:11:32.963 INFO:teuthology.orchestra.run.vm03.stdout: libgfortran-11.5.0-14.el9.x86_64 2026-03-09T15:11:32.963 INFO:teuthology.orchestra.run.vm03.stdout: libnbd-1.20.3-4.el9.x86_64 2026-03-09T15:11:32.963 INFO:teuthology.orchestra.run.vm03.stdout: liboath-2.6.12-1.el9.x86_64 2026-03-09T15:11:32.963 INFO:teuthology.orchestra.run.vm03.stdout: libpmemobj-1.12.1-1.el9.x86_64 2026-03-09T15:11:32.963 INFO:teuthology.orchestra.run.vm03.stdout: libquadmath-11.5.0-14.el9.x86_64 2026-03-09T15:11:32.963 INFO:teuthology.orchestra.run.vm03.stdout: librabbitmq-0.11.0-7.el9.x86_64 2026-03-09T15:11:32.963 INFO:teuthology.orchestra.run.vm03.stdout: librados-devel-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T15:11:32.963 INFO:teuthology.orchestra.run.vm03.stdout: libradosstriper1-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T15:11:32.963 INFO:teuthology.orchestra.run.vm03.stdout: librdkafka-1.6.1-102.el9.x86_64 2026-03-09T15:11:32.963 INFO:teuthology.orchestra.run.vm03.stdout: librgw2-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T15:11:32.963 INFO:teuthology.orchestra.run.vm03.stdout: libstoragemgmt-1.10.1-1.el9.x86_64 2026-03-09T15:11:32.963 INFO:teuthology.orchestra.run.vm03.stdout: libunwind-1.6.2-1.el9.x86_64 2026-03-09T15:11:32.963 INFO:teuthology.orchestra.run.vm03.stdout: libxslt-1.1.34-12.el9.x86_64 2026-03-09T15:11:32.963 INFO:teuthology.orchestra.run.vm03.stdout: lttng-ust-2.12.0-6.el9.x86_64 2026-03-09T15:11:32.963 INFO:teuthology.orchestra.run.vm03.stdout: lua-5.4.4-4.el9.x86_64 2026-03-09T15:11:32.963 INFO:teuthology.orchestra.run.vm03.stdout: lua-devel-5.4.4-4.el9.x86_64 2026-03-09T15:11:32.963 INFO:teuthology.orchestra.run.vm03.stdout: luarocks-3.9.2-5.el9.noarch 2026-03-09T15:11:32.963 INFO:teuthology.orchestra.run.vm03.stdout: mailcap-2.1.49-5.el9.noarch 2026-03-09T15:11:32.963 INFO:teuthology.orchestra.run.vm03.stdout: openblas-0.3.29-1.el9.x86_64 2026-03-09T15:11:32.963 INFO:teuthology.orchestra.run.vm03.stdout: openblas-openmp-0.3.29-1.el9.x86_64 2026-03-09T15:11:32.963 INFO:teuthology.orchestra.run.vm03.stdout: parquet-libs-9.0.0-15.el9.x86_64 2026-03-09T15:11:32.963 INFO:teuthology.orchestra.run.vm03.stdout: pciutils-3.7.0-7.el9.x86_64 2026-03-09T15:11:32.963 INFO:teuthology.orchestra.run.vm03.stdout: protobuf-3.14.0-17.el9.x86_64 2026-03-09T15:11:32.963 INFO:teuthology.orchestra.run.vm03.stdout: protobuf-compiler-3.14.0-17.el9.x86_64 2026-03-09T15:11:32.963 INFO:teuthology.orchestra.run.vm03.stdout: python3-asyncssh-2.13.2-5.el9.noarch 2026-03-09T15:11:32.963 INFO:teuthology.orchestra.run.vm03.stdout: python3-autocommand-2.2.2-8.el9.noarch 2026-03-09T15:11:32.963 INFO:teuthology.orchestra.run.vm03.stdout: python3-babel-2.9.1-2.el9.noarch 2026-03-09T15:11:32.963 INFO:teuthology.orchestra.run.vm03.stdout: python3-backports-tarfile-1.2.0-1.el9.noarch 2026-03-09T15:11:32.963 INFO:teuthology.orchestra.run.vm03.stdout: python3-bcrypt-3.2.2-1.el9.x86_64 2026-03-09T15:11:32.963 INFO:teuthology.orchestra.run.vm03.stdout: python3-cachetools-4.2.4-1.el9.noarch 2026-03-09T15:11:32.963 INFO:teuthology.orchestra.run.vm03.stdout: python3-ceph-argparse-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T15:11:32.963 INFO:teuthology.orchestra.run.vm03.stdout: python3-ceph-common-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T15:11:32.963 INFO:teuthology.orchestra.run.vm03.stdout: python3-cephfs-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T15:11:32.963 INFO:teuthology.orchestra.run.vm03.stdout: python3-certifi-2023.05.07-4.el9.noarch 2026-03-09T15:11:32.963 INFO:teuthology.orchestra.run.vm03.stdout: python3-cffi-1.14.5-5.el9.x86_64 2026-03-09T15:11:32.963 INFO:teuthology.orchestra.run.vm03.stdout: python3-cheroot-10.0.1-4.el9.noarch 2026-03-09T15:11:32.963 INFO:teuthology.orchestra.run.vm03.stdout: python3-cherrypy-18.6.1-2.el9.noarch 2026-03-09T15:11:32.963 INFO:teuthology.orchestra.run.vm03.stdout: python3-cryptography-36.0.1-5.el9.x86_64 2026-03-09T15:11:32.963 INFO:teuthology.orchestra.run.vm03.stdout: python3-devel-3.9.25-3.el9.x86_64 2026-03-09T15:11:32.963 INFO:teuthology.orchestra.run.vm03.stdout: python3-google-auth-1:2.45.0-1.el9.noarch 2026-03-09T15:11:32.963 INFO:teuthology.orchestra.run.vm03.stdout: python3-grpcio-1.46.7-10.el9.x86_64 2026-03-09T15:11:32.963 INFO:teuthology.orchestra.run.vm03.stdout: python3-grpcio-tools-1.46.7-10.el9.x86_64 2026-03-09T15:11:32.963 INFO:teuthology.orchestra.run.vm03.stdout: python3-jaraco-8.2.1-3.el9.noarch 2026-03-09T15:11:32.963 INFO:teuthology.orchestra.run.vm03.stdout: python3-jaraco-classes-3.2.1-5.el9.noarch 2026-03-09T15:11:32.963 INFO:teuthology.orchestra.run.vm03.stdout: python3-jaraco-collections-3.0.0-8.el9.noarch 2026-03-09T15:11:32.963 INFO:teuthology.orchestra.run.vm03.stdout: python3-jaraco-context-6.0.1-3.el9.noarch 2026-03-09T15:11:32.963 INFO:teuthology.orchestra.run.vm03.stdout: python3-jaraco-functools-3.5.0-2.el9.noarch 2026-03-09T15:11:32.963 INFO:teuthology.orchestra.run.vm03.stdout: python3-jaraco-text-4.0.0-2.el9.noarch 2026-03-09T15:11:32.963 INFO:teuthology.orchestra.run.vm03.stdout: python3-jinja2-2.11.3-8.el9.noarch 2026-03-09T15:11:32.963 INFO:teuthology.orchestra.run.vm03.stdout: python3-jmespath-1.0.1-1.el9.noarch 2026-03-09T15:11:32.963 INFO:teuthology.orchestra.run.vm03.stdout: python3-kubernetes-1:26.1.0-3.el9.noarch 2026-03-09T15:11:32.963 INFO:teuthology.orchestra.run.vm03.stdout: python3-libstoragemgmt-1.10.1-1.el9.x86_64 2026-03-09T15:11:32.963 INFO:teuthology.orchestra.run.vm03.stdout: python3-logutils-0.3.5-21.el9.noarch 2026-03-09T15:11:32.963 INFO:teuthology.orchestra.run.vm03.stdout: python3-mako-1.1.4-6.el9.noarch 2026-03-09T15:11:32.963 INFO:teuthology.orchestra.run.vm03.stdout: python3-markupsafe-1.1.1-12.el9.x86_64 2026-03-09T15:11:32.963 INFO:teuthology.orchestra.run.vm03.stdout: python3-more-itertools-8.12.0-2.el9.noarch 2026-03-09T15:11:32.963 INFO:teuthology.orchestra.run.vm03.stdout: python3-natsort-7.1.1-5.el9.noarch 2026-03-09T15:11:32.963 INFO:teuthology.orchestra.run.vm03.stdout: python3-numpy-1:1.23.5-2.el9.x86_64 2026-03-09T15:11:32.963 INFO:teuthology.orchestra.run.vm03.stdout: python3-numpy-f2py-1:1.23.5-2.el9.x86_64 2026-03-09T15:11:32.964 INFO:teuthology.orchestra.run.vm03.stdout: python3-packaging-20.9-5.el9.noarch 2026-03-09T15:11:32.964 INFO:teuthology.orchestra.run.vm03.stdout: python3-pecan-1.4.2-3.el9.noarch 2026-03-09T15:11:32.964 INFO:teuthology.orchestra.run.vm03.stdout: python3-ply-3.11-14.el9.noarch 2026-03-09T15:11:32.964 INFO:teuthology.orchestra.run.vm03.stdout: python3-portend-3.1.0-2.el9.noarch 2026-03-09T15:11:32.964 INFO:teuthology.orchestra.run.vm03.stdout: python3-protobuf-3.14.0-17.el9.noarch 2026-03-09T15:11:32.964 INFO:teuthology.orchestra.run.vm03.stdout: python3-pyOpenSSL-21.0.0-1.el9.noarch 2026-03-09T15:11:32.964 INFO:teuthology.orchestra.run.vm03.stdout: python3-pyasn1-0.4.8-7.el9.noarch 2026-03-09T15:11:32.964 INFO:teuthology.orchestra.run.vm03.stdout: python3-pyasn1-modules-0.4.8-7.el9.noarch 2026-03-09T15:11:32.964 INFO:teuthology.orchestra.run.vm03.stdout: python3-pycparser-2.20-6.el9.noarch 2026-03-09T15:11:32.964 INFO:teuthology.orchestra.run.vm03.stdout: python3-rados-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T15:11:32.964 INFO:teuthology.orchestra.run.vm03.stdout: python3-rbd-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T15:11:32.964 INFO:teuthology.orchestra.run.vm03.stdout: python3-repoze-lru-0.7-16.el9.noarch 2026-03-09T15:11:32.964 INFO:teuthology.orchestra.run.vm03.stdout: python3-requests-2.25.1-10.el9.noarch 2026-03-09T15:11:32.964 INFO:teuthology.orchestra.run.vm03.stdout: python3-requests-oauthlib-1.3.0-12.el9.noarch 2026-03-09T15:11:32.964 INFO:teuthology.orchestra.run.vm03.stdout: python3-rgw-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T15:11:32.964 INFO:teuthology.orchestra.run.vm03.stdout: python3-routes-2.5.1-5.el9.noarch 2026-03-09T15:11:32.964 INFO:teuthology.orchestra.run.vm03.stdout: python3-rsa-4.9-2.el9.noarch 2026-03-09T15:11:32.964 INFO:teuthology.orchestra.run.vm03.stdout: python3-scipy-1.9.3-2.el9.x86_64 2026-03-09T15:11:32.964 INFO:teuthology.orchestra.run.vm03.stdout: python3-tempora-5.0.0-2.el9.noarch 2026-03-09T15:11:32.964 INFO:teuthology.orchestra.run.vm03.stdout: python3-toml-0.10.2-6.el9.noarch 2026-03-09T15:11:32.964 INFO:teuthology.orchestra.run.vm03.stdout: python3-typing-extensions-4.15.0-1.el9.noarch 2026-03-09T15:11:32.964 INFO:teuthology.orchestra.run.vm03.stdout: python3-urllib3-1.26.5-7.el9.noarch 2026-03-09T15:11:32.964 INFO:teuthology.orchestra.run.vm03.stdout: python3-webob-1.8.8-2.el9.noarch 2026-03-09T15:11:32.964 INFO:teuthology.orchestra.run.vm03.stdout: python3-websocket-client-1.2.3-2.el9.noarch 2026-03-09T15:11:32.964 INFO:teuthology.orchestra.run.vm03.stdout: python3-werkzeug-2.0.3-3.el9.1.noarch 2026-03-09T15:11:32.964 INFO:teuthology.orchestra.run.vm03.stdout: python3-xmltodict-0.12.0-15.el9.noarch 2026-03-09T15:11:32.964 INFO:teuthology.orchestra.run.vm03.stdout: python3-zc-lockfile-2.0-10.el9.noarch 2026-03-09T15:11:32.964 INFO:teuthology.orchestra.run.vm03.stdout: qatlib-25.08.0-2.el9.x86_64 2026-03-09T15:11:32.964 INFO:teuthology.orchestra.run.vm03.stdout: qatlib-service-25.08.0-2.el9.x86_64 2026-03-09T15:11:32.964 INFO:teuthology.orchestra.run.vm03.stdout: qatzip-libs-1.3.1-1.el9.x86_64 2026-03-09T15:11:32.964 INFO:teuthology.orchestra.run.vm03.stdout: rbd-fuse-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T15:11:32.964 INFO:teuthology.orchestra.run.vm03.stdout: rbd-mirror-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T15:11:32.964 INFO:teuthology.orchestra.run.vm03.stdout: rbd-nbd-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T15:11:32.964 INFO:teuthology.orchestra.run.vm03.stdout: re2-1:20211101-20.el9.x86_64 2026-03-09T15:11:32.964 INFO:teuthology.orchestra.run.vm03.stdout: socat-1.7.4.1-8.el9.x86_64 2026-03-09T15:11:32.964 INFO:teuthology.orchestra.run.vm03.stdout: thrift-0.15.0-4.el9.x86_64 2026-03-09T15:11:32.964 INFO:teuthology.orchestra.run.vm03.stdout: unzip-6.0-59.el9.x86_64 2026-03-09T15:11:32.964 INFO:teuthology.orchestra.run.vm03.stdout: xmlstarlet-1.6.1-20.el9.x86_64 2026-03-09T15:11:32.964 INFO:teuthology.orchestra.run.vm03.stdout: zip-3.0-35.el9.x86_64 2026-03-09T15:11:32.964 INFO:teuthology.orchestra.run.vm03.stdout: 2026-03-09T15:11:32.964 INFO:teuthology.orchestra.run.vm03.stdout:Complete! 2026-03-09T15:11:33.056 DEBUG:teuthology.parallel:result is None 2026-03-09T15:11:33.274 INFO:teuthology.orchestra.run.vm06.stdout: Running scriptlet: ceph-selinux-2:19.2.3-678.ge911bdeb.el9.x86_64 108/138 2026-03-09T15:11:33.274 INFO:teuthology.orchestra.run.vm06.stdout:skipping the directory /sys 2026-03-09T15:11:33.274 INFO:teuthology.orchestra.run.vm06.stdout:skipping the directory /proc 2026-03-09T15:11:33.274 INFO:teuthology.orchestra.run.vm06.stdout:skipping the directory /mnt 2026-03-09T15:11:33.274 INFO:teuthology.orchestra.run.vm06.stdout:skipping the directory /var/tmp 2026-03-09T15:11:33.274 INFO:teuthology.orchestra.run.vm06.stdout:skipping the directory /home 2026-03-09T15:11:33.274 INFO:teuthology.orchestra.run.vm06.stdout:skipping the directory /root 2026-03-09T15:11:33.274 INFO:teuthology.orchestra.run.vm06.stdout:skipping the directory /tmp 2026-03-09T15:11:33.274 INFO:teuthology.orchestra.run.vm06.stdout: 2026-03-09T15:11:33.401 INFO:teuthology.orchestra.run.vm06.stdout: Installing : ceph-mds-2:19.2.3-678.ge911bdeb.el9.x86_64 109/138 2026-03-09T15:11:33.425 INFO:teuthology.orchestra.run.vm06.stdout: Running scriptlet: ceph-mds-2:19.2.3-678.ge911bdeb.el9.x86_64 109/138 2026-03-09T15:11:33.425 INFO:teuthology.orchestra.run.vm06.stdout:Glob pattern passed to enable, but globs are not supported for this. 2026-03-09T15:11:33.425 INFO:teuthology.orchestra.run.vm06.stdout:Invalid unit name "ceph-mds@*.service" escaped as "ceph-mds@\x2a.service". 2026-03-09T15:11:33.425 INFO:teuthology.orchestra.run.vm06.stdout:Created symlink /etc/systemd/system/multi-user.target.wants/ceph-mds.target → /usr/lib/systemd/system/ceph-mds.target. 2026-03-09T15:11:33.425 INFO:teuthology.orchestra.run.vm06.stdout:Created symlink /etc/systemd/system/ceph.target.wants/ceph-mds.target → /usr/lib/systemd/system/ceph-mds.target. 2026-03-09T15:11:33.425 INFO:teuthology.orchestra.run.vm06.stdout: 2026-03-09T15:11:33.659 INFO:teuthology.orchestra.run.vm06.stdout: Installing : ceph-mon-2:19.2.3-678.ge911bdeb.el9.x86_64 110/138 2026-03-09T15:11:33.680 INFO:teuthology.orchestra.run.vm06.stdout: Running scriptlet: ceph-mon-2:19.2.3-678.ge911bdeb.el9.x86_64 110/138 2026-03-09T15:11:33.680 INFO:teuthology.orchestra.run.vm06.stdout:Glob pattern passed to enable, but globs are not supported for this. 2026-03-09T15:11:33.680 INFO:teuthology.orchestra.run.vm06.stdout:Invalid unit name "ceph-mon@*.service" escaped as "ceph-mon@\x2a.service". 2026-03-09T15:11:33.680 INFO:teuthology.orchestra.run.vm06.stdout:Created symlink /etc/systemd/system/multi-user.target.wants/ceph-mon.target → /usr/lib/systemd/system/ceph-mon.target. 2026-03-09T15:11:33.680 INFO:teuthology.orchestra.run.vm06.stdout:Created symlink /etc/systemd/system/ceph.target.wants/ceph-mon.target → /usr/lib/systemd/system/ceph-mon.target. 2026-03-09T15:11:33.680 INFO:teuthology.orchestra.run.vm06.stdout: 2026-03-09T15:11:33.690 INFO:teuthology.orchestra.run.vm06.stdout: Installing : mailcap-2.1.49-5.el9.noarch 111/138 2026-03-09T15:11:33.692 INFO:teuthology.orchestra.run.vm06.stdout: Installing : libconfig-1.7.2-9.el9.x86_64 112/138 2026-03-09T15:11:33.710 INFO:teuthology.orchestra.run.vm06.stdout: Running scriptlet: libstoragemgmt-1.10.1-1.el9.x86_64 113/138 2026-03-09T15:11:33.710 INFO:teuthology.orchestra.run.vm06.stdout:Creating group 'qat' with GID 994. 2026-03-09T15:11:33.710 INFO:teuthology.orchestra.run.vm06.stdout:Creating group 'libstoragemgmt' with GID 993. 2026-03-09T15:11:33.710 INFO:teuthology.orchestra.run.vm06.stdout:Creating user 'libstoragemgmt' (daemon account for libstoragemgmt) with UID 993 and GID 993. 2026-03-09T15:11:33.710 INFO:teuthology.orchestra.run.vm06.stdout: 2026-03-09T15:11:33.721 INFO:teuthology.orchestra.run.vm06.stdout: Installing : libstoragemgmt-1.10.1-1.el9.x86_64 113/138 2026-03-09T15:11:33.747 INFO:teuthology.orchestra.run.vm06.stdout: Running scriptlet: libstoragemgmt-1.10.1-1.el9.x86_64 113/138 2026-03-09T15:11:33.747 INFO:teuthology.orchestra.run.vm06.stdout:Created symlink /etc/systemd/system/multi-user.target.wants/libstoragemgmt.service → /usr/lib/systemd/system/libstoragemgmt.service. 2026-03-09T15:11:33.747 INFO:teuthology.orchestra.run.vm06.stdout: 2026-03-09T15:11:33.793 INFO:teuthology.orchestra.run.vm06.stdout: Installing : python3-libstoragemgmt-1.10.1-1.el9.x86_64 114/138 2026-03-09T15:11:33.872 INFO:teuthology.orchestra.run.vm06.stdout: Installing : cryptsetup-2.8.1-3.el9.x86_64 115/138 2026-03-09T15:11:33.878 INFO:teuthology.orchestra.run.vm06.stdout: Installing : ceph-volume-2:19.2.3-678.ge911bdeb.el9.noarch 116/138 2026-03-09T15:11:33.892 INFO:teuthology.orchestra.run.vm06.stdout: Running scriptlet: ceph-volume-2:19.2.3-678.ge911bdeb.el9.noarch 116/138 2026-03-09T15:11:33.892 INFO:teuthology.orchestra.run.vm06.stdout:Glob pattern passed to enable, but globs are not supported for this. 2026-03-09T15:11:33.892 INFO:teuthology.orchestra.run.vm06.stdout:Invalid unit name "ceph-volume@*.service" escaped as "ceph-volume@\x2a.service". 2026-03-09T15:11:33.892 INFO:teuthology.orchestra.run.vm06.stdout: 2026-03-09T15:11:34.705 INFO:teuthology.orchestra.run.vm06.stdout: Installing : ceph-osd-2:19.2.3-678.ge911bdeb.el9.x86_64 117/138 2026-03-09T15:11:34.731 INFO:teuthology.orchestra.run.vm06.stdout: Running scriptlet: ceph-osd-2:19.2.3-678.ge911bdeb.el9.x86_64 117/138 2026-03-09T15:11:34.731 INFO:teuthology.orchestra.run.vm06.stdout:Glob pattern passed to enable, but globs are not supported for this. 2026-03-09T15:11:34.732 INFO:teuthology.orchestra.run.vm06.stdout:Invalid unit name "ceph-osd@*.service" escaped as "ceph-osd@\x2a.service". 2026-03-09T15:11:34.732 INFO:teuthology.orchestra.run.vm06.stdout:Created symlink /etc/systemd/system/multi-user.target.wants/ceph-osd.target → /usr/lib/systemd/system/ceph-osd.target. 2026-03-09T15:11:34.732 INFO:teuthology.orchestra.run.vm06.stdout:Created symlink /etc/systemd/system/ceph.target.wants/ceph-osd.target → /usr/lib/systemd/system/ceph-osd.target. 2026-03-09T15:11:34.732 INFO:teuthology.orchestra.run.vm06.stdout: 2026-03-09T15:11:34.789 INFO:teuthology.orchestra.run.vm06.stdout: Running scriptlet: cephadm-2:19.2.3-678.ge911bdeb.el9.noarch 118/138 2026-03-09T15:11:34.793 INFO:teuthology.orchestra.run.vm06.stdout: Installing : cephadm-2:19.2.3-678.ge911bdeb.el9.noarch 118/138 2026-03-09T15:11:34.799 INFO:teuthology.orchestra.run.vm06.stdout: Installing : ceph-prometheus-alerts-2:19.2.3-678.ge911bdeb.el 119/138 2026-03-09T15:11:34.821 INFO:teuthology.orchestra.run.vm06.stdout: Installing : ceph-grafana-dashboards-2:19.2.3-678.ge911bdeb.e 120/138 2026-03-09T15:11:34.823 INFO:teuthology.orchestra.run.vm06.stdout: Installing : ceph-mgr-cephadm-2:19.2.3-678.ge911bdeb.el9.noar 121/138 2026-03-09T15:11:35.347 INFO:teuthology.orchestra.run.vm06.stdout: Running scriptlet: ceph-mgr-cephadm-2:19.2.3-678.ge911bdeb.el9.noar 121/138 2026-03-09T15:11:35.353 INFO:teuthology.orchestra.run.vm06.stdout: Installing : ceph-mgr-dashboard-2:19.2.3-678.ge911bdeb.el9.no 122/138 2026-03-09T15:11:35.881 INFO:teuthology.orchestra.run.vm06.stdout: Running scriptlet: ceph-mgr-dashboard-2:19.2.3-678.ge911bdeb.el9.no 122/138 2026-03-09T15:11:35.926 INFO:teuthology.orchestra.run.vm06.stdout: Installing : ceph-mgr-diskprediction-local-2:19.2.3-678.ge911 123/138 2026-03-09T15:11:35.990 INFO:teuthology.orchestra.run.vm06.stdout: Running scriptlet: ceph-mgr-diskprediction-local-2:19.2.3-678.ge911 123/138 2026-03-09T15:11:36.075 INFO:teuthology.orchestra.run.vm06.stdout: Installing : ceph-mgr-modules-core-2:19.2.3-678.ge911bdeb.el9 124/138 2026-03-09T15:11:36.138 INFO:teuthology.orchestra.run.vm06.stdout: Installing : ceph-mgr-2:19.2.3-678.ge911bdeb.el9.x86_64 125/138 2026-03-09T15:11:36.158 INFO:teuthology.orchestra.run.vm06.stdout: Running scriptlet: ceph-mgr-2:19.2.3-678.ge911bdeb.el9.x86_64 125/138 2026-03-09T15:11:36.159 INFO:teuthology.orchestra.run.vm06.stdout:Glob pattern passed to enable, but globs are not supported for this. 2026-03-09T15:11:36.159 INFO:teuthology.orchestra.run.vm06.stdout:Invalid unit name "ceph-mgr@*.service" escaped as "ceph-mgr@\x2a.service". 2026-03-09T15:11:36.159 INFO:teuthology.orchestra.run.vm06.stdout:Created symlink /etc/systemd/system/multi-user.target.wants/ceph-mgr.target → /usr/lib/systemd/system/ceph-mgr.target. 2026-03-09T15:11:36.159 INFO:teuthology.orchestra.run.vm06.stdout:Created symlink /etc/systemd/system/ceph.target.wants/ceph-mgr.target → /usr/lib/systemd/system/ceph-mgr.target. 2026-03-09T15:11:36.159 INFO:teuthology.orchestra.run.vm06.stdout: 2026-03-09T15:11:36.240 INFO:teuthology.orchestra.run.vm06.stdout: Installing : ceph-mgr-rook-2:19.2.3-678.ge911bdeb.el9.noarch 126/138 2026-03-09T15:11:36.251 INFO:teuthology.orchestra.run.vm06.stdout: Running scriptlet: ceph-mgr-rook-2:19.2.3-678.ge911bdeb.el9.noarch 126/138 2026-03-09T15:11:36.879 INFO:teuthology.orchestra.run.vm06.stdout: Installing : ceph-2:19.2.3-678.ge911bdeb.el9.x86_64 127/138 2026-03-09T15:11:36.882 INFO:teuthology.orchestra.run.vm06.stdout: Installing : ceph-radosgw-2:19.2.3-678.ge911bdeb.el9.x86_64 128/138 2026-03-09T15:11:36.904 INFO:teuthology.orchestra.run.vm06.stdout: Running scriptlet: ceph-radosgw-2:19.2.3-678.ge911bdeb.el9.x86_64 128/138 2026-03-09T15:11:36.904 INFO:teuthology.orchestra.run.vm06.stdout:Glob pattern passed to enable, but globs are not supported for this. 2026-03-09T15:11:36.904 INFO:teuthology.orchestra.run.vm06.stdout:Invalid unit name "ceph-radosgw@*.service" escaped as "ceph-radosgw@\x2a.service". 2026-03-09T15:11:36.904 INFO:teuthology.orchestra.run.vm06.stdout:Created symlink /etc/systemd/system/multi-user.target.wants/ceph-radosgw.target → /usr/lib/systemd/system/ceph-radosgw.target. 2026-03-09T15:11:36.904 INFO:teuthology.orchestra.run.vm06.stdout:Created symlink /etc/systemd/system/ceph.target.wants/ceph-radosgw.target → /usr/lib/systemd/system/ceph-radosgw.target. 2026-03-09T15:11:36.904 INFO:teuthology.orchestra.run.vm06.stdout: 2026-03-09T15:11:36.925 INFO:teuthology.orchestra.run.vm04.stdout: Running scriptlet: ceph-selinux-2:19.2.3-678.ge911bdeb.el9.x86_64 108/138 2026-03-09T15:11:36.925 INFO:teuthology.orchestra.run.vm04.stdout:skipping the directory /sys 2026-03-09T15:11:36.925 INFO:teuthology.orchestra.run.vm04.stdout:skipping the directory /proc 2026-03-09T15:11:36.925 INFO:teuthology.orchestra.run.vm04.stdout:skipping the directory /mnt 2026-03-09T15:11:36.925 INFO:teuthology.orchestra.run.vm04.stdout:skipping the directory /var/tmp 2026-03-09T15:11:36.925 INFO:teuthology.orchestra.run.vm04.stdout:skipping the directory /home 2026-03-09T15:11:36.925 INFO:teuthology.orchestra.run.vm04.stdout:skipping the directory /root 2026-03-09T15:11:36.925 INFO:teuthology.orchestra.run.vm04.stdout:skipping the directory /tmp 2026-03-09T15:11:36.925 INFO:teuthology.orchestra.run.vm04.stdout: 2026-03-09T15:11:36.931 INFO:teuthology.orchestra.run.vm06.stdout: Installing : ceph-immutable-object-cache-2:19.2.3-678.ge911bd 129/138 2026-03-09T15:11:36.953 INFO:teuthology.orchestra.run.vm06.stdout: Running scriptlet: ceph-immutable-object-cache-2:19.2.3-678.ge911bd 129/138 2026-03-09T15:11:36.953 INFO:teuthology.orchestra.run.vm06.stdout:Glob pattern passed to enable, but globs are not supported for this. 2026-03-09T15:11:36.953 INFO:teuthology.orchestra.run.vm06.stdout:Invalid unit name "ceph-immutable-object-cache@*.service" escaped as "ceph-immutable-object-cache@\x2a.service". 2026-03-09T15:11:36.953 INFO:teuthology.orchestra.run.vm06.stdout: 2026-03-09T15:11:37.058 INFO:teuthology.orchestra.run.vm04.stdout: Installing : ceph-mds-2:19.2.3-678.ge911bdeb.el9.x86_64 109/138 2026-03-09T15:11:37.081 INFO:teuthology.orchestra.run.vm04.stdout: Running scriptlet: ceph-mds-2:19.2.3-678.ge911bdeb.el9.x86_64 109/138 2026-03-09T15:11:37.081 INFO:teuthology.orchestra.run.vm04.stdout:Glob pattern passed to enable, but globs are not supported for this. 2026-03-09T15:11:37.081 INFO:teuthology.orchestra.run.vm04.stdout:Invalid unit name "ceph-mds@*.service" escaped as "ceph-mds@\x2a.service". 2026-03-09T15:11:37.081 INFO:teuthology.orchestra.run.vm04.stdout:Created symlink /etc/systemd/system/multi-user.target.wants/ceph-mds.target → /usr/lib/systemd/system/ceph-mds.target. 2026-03-09T15:11:37.081 INFO:teuthology.orchestra.run.vm04.stdout:Created symlink /etc/systemd/system/ceph.target.wants/ceph-mds.target → /usr/lib/systemd/system/ceph-mds.target. 2026-03-09T15:11:37.081 INFO:teuthology.orchestra.run.vm04.stdout: 2026-03-09T15:11:37.116 INFO:teuthology.orchestra.run.vm06.stdout: Installing : rbd-mirror-2:19.2.3-678.ge911bdeb.el9.x86_64 130/138 2026-03-09T15:11:37.136 INFO:teuthology.orchestra.run.vm06.stdout: Running scriptlet: rbd-mirror-2:19.2.3-678.ge911bdeb.el9.x86_64 130/138 2026-03-09T15:11:37.136 INFO:teuthology.orchestra.run.vm06.stdout:Glob pattern passed to enable, but globs are not supported for this. 2026-03-09T15:11:37.136 INFO:teuthology.orchestra.run.vm06.stdout:Invalid unit name "ceph-rbd-mirror@*.service" escaped as "ceph-rbd-mirror@\x2a.service". 2026-03-09T15:11:37.136 INFO:teuthology.orchestra.run.vm06.stdout:Created symlink /etc/systemd/system/multi-user.target.wants/ceph-rbd-mirror.target → /usr/lib/systemd/system/ceph-rbd-mirror.target. 2026-03-09T15:11:37.136 INFO:teuthology.orchestra.run.vm06.stdout:Created symlink /etc/systemd/system/ceph.target.wants/ceph-rbd-mirror.target → /usr/lib/systemd/system/ceph-rbd-mirror.target. 2026-03-09T15:11:37.136 INFO:teuthology.orchestra.run.vm06.stdout: 2026-03-09T15:11:37.313 INFO:teuthology.orchestra.run.vm04.stdout: Installing : ceph-mon-2:19.2.3-678.ge911bdeb.el9.x86_64 110/138 2026-03-09T15:11:37.334 INFO:teuthology.orchestra.run.vm04.stdout: Running scriptlet: ceph-mon-2:19.2.3-678.ge911bdeb.el9.x86_64 110/138 2026-03-09T15:11:37.334 INFO:teuthology.orchestra.run.vm04.stdout:Glob pattern passed to enable, but globs are not supported for this. 2026-03-09T15:11:37.334 INFO:teuthology.orchestra.run.vm04.stdout:Invalid unit name "ceph-mon@*.service" escaped as "ceph-mon@\x2a.service". 2026-03-09T15:11:37.334 INFO:teuthology.orchestra.run.vm04.stdout:Created symlink /etc/systemd/system/multi-user.target.wants/ceph-mon.target → /usr/lib/systemd/system/ceph-mon.target. 2026-03-09T15:11:37.334 INFO:teuthology.orchestra.run.vm04.stdout:Created symlink /etc/systemd/system/ceph.target.wants/ceph-mon.target → /usr/lib/systemd/system/ceph-mon.target. 2026-03-09T15:11:37.334 INFO:teuthology.orchestra.run.vm04.stdout: 2026-03-09T15:11:37.342 INFO:teuthology.orchestra.run.vm04.stdout: Installing : mailcap-2.1.49-5.el9.noarch 111/138 2026-03-09T15:11:37.345 INFO:teuthology.orchestra.run.vm04.stdout: Installing : libconfig-1.7.2-9.el9.x86_64 112/138 2026-03-09T15:11:37.363 INFO:teuthology.orchestra.run.vm04.stdout: Running scriptlet: libstoragemgmt-1.10.1-1.el9.x86_64 113/138 2026-03-09T15:11:37.363 INFO:teuthology.orchestra.run.vm04.stdout:Creating group 'qat' with GID 994. 2026-03-09T15:11:37.363 INFO:teuthology.orchestra.run.vm04.stdout:Creating group 'libstoragemgmt' with GID 993. 2026-03-09T15:11:37.363 INFO:teuthology.orchestra.run.vm04.stdout:Creating user 'libstoragemgmt' (daemon account for libstoragemgmt) with UID 993 and GID 993. 2026-03-09T15:11:37.363 INFO:teuthology.orchestra.run.vm04.stdout: 2026-03-09T15:11:37.374 INFO:teuthology.orchestra.run.vm04.stdout: Installing : libstoragemgmt-1.10.1-1.el9.x86_64 113/138 2026-03-09T15:11:37.404 INFO:teuthology.orchestra.run.vm04.stdout: Running scriptlet: libstoragemgmt-1.10.1-1.el9.x86_64 113/138 2026-03-09T15:11:37.404 INFO:teuthology.orchestra.run.vm04.stdout:Created symlink /etc/systemd/system/multi-user.target.wants/libstoragemgmt.service → /usr/lib/systemd/system/libstoragemgmt.service. 2026-03-09T15:11:37.404 INFO:teuthology.orchestra.run.vm04.stdout: 2026-03-09T15:11:37.452 INFO:teuthology.orchestra.run.vm04.stdout: Installing : python3-libstoragemgmt-1.10.1-1.el9.x86_64 114/138 2026-03-09T15:11:37.530 INFO:teuthology.orchestra.run.vm04.stdout: Installing : cryptsetup-2.8.1-3.el9.x86_64 115/138 2026-03-09T15:11:37.535 INFO:teuthology.orchestra.run.vm04.stdout: Installing : ceph-volume-2:19.2.3-678.ge911bdeb.el9.noarch 116/138 2026-03-09T15:11:37.548 INFO:teuthology.orchestra.run.vm04.stdout: Running scriptlet: ceph-volume-2:19.2.3-678.ge911bdeb.el9.noarch 116/138 2026-03-09T15:11:37.549 INFO:teuthology.orchestra.run.vm04.stdout:Glob pattern passed to enable, but globs are not supported for this. 2026-03-09T15:11:37.549 INFO:teuthology.orchestra.run.vm04.stdout:Invalid unit name "ceph-volume@*.service" escaped as "ceph-volume@\x2a.service". 2026-03-09T15:11:37.549 INFO:teuthology.orchestra.run.vm04.stdout: 2026-03-09T15:11:38.347 INFO:teuthology.orchestra.run.vm04.stdout: Installing : ceph-osd-2:19.2.3-678.ge911bdeb.el9.x86_64 117/138 2026-03-09T15:11:38.372 INFO:teuthology.orchestra.run.vm04.stdout: Running scriptlet: ceph-osd-2:19.2.3-678.ge911bdeb.el9.x86_64 117/138 2026-03-09T15:11:38.372 INFO:teuthology.orchestra.run.vm04.stdout:Glob pattern passed to enable, but globs are not supported for this. 2026-03-09T15:11:38.372 INFO:teuthology.orchestra.run.vm04.stdout:Invalid unit name "ceph-osd@*.service" escaped as "ceph-osd@\x2a.service". 2026-03-09T15:11:38.372 INFO:teuthology.orchestra.run.vm04.stdout:Created symlink /etc/systemd/system/multi-user.target.wants/ceph-osd.target → /usr/lib/systemd/system/ceph-osd.target. 2026-03-09T15:11:38.372 INFO:teuthology.orchestra.run.vm04.stdout:Created symlink /etc/systemd/system/ceph.target.wants/ceph-osd.target → /usr/lib/systemd/system/ceph-osd.target. 2026-03-09T15:11:38.372 INFO:teuthology.orchestra.run.vm04.stdout: 2026-03-09T15:11:38.434 INFO:teuthology.orchestra.run.vm04.stdout: Running scriptlet: cephadm-2:19.2.3-678.ge911bdeb.el9.noarch 118/138 2026-03-09T15:11:38.437 INFO:teuthology.orchestra.run.vm04.stdout: Installing : cephadm-2:19.2.3-678.ge911bdeb.el9.noarch 118/138 2026-03-09T15:11:38.443 INFO:teuthology.orchestra.run.vm04.stdout: Installing : ceph-prometheus-alerts-2:19.2.3-678.ge911bdeb.el 119/138 2026-03-09T15:11:38.466 INFO:teuthology.orchestra.run.vm04.stdout: Installing : ceph-grafana-dashboards-2:19.2.3-678.ge911bdeb.e 120/138 2026-03-09T15:11:38.470 INFO:teuthology.orchestra.run.vm04.stdout: Installing : ceph-mgr-cephadm-2:19.2.3-678.ge911bdeb.el9.noar 121/138 2026-03-09T15:11:39.013 INFO:teuthology.orchestra.run.vm04.stdout: Running scriptlet: ceph-mgr-cephadm-2:19.2.3-678.ge911bdeb.el9.noar 121/138 2026-03-09T15:11:39.020 INFO:teuthology.orchestra.run.vm04.stdout: Installing : ceph-mgr-dashboard-2:19.2.3-678.ge911bdeb.el9.no 122/138 2026-03-09T15:11:39.541 INFO:teuthology.orchestra.run.vm04.stdout: Running scriptlet: ceph-mgr-dashboard-2:19.2.3-678.ge911bdeb.el9.no 122/138 2026-03-09T15:11:39.543 INFO:teuthology.orchestra.run.vm04.stdout: Installing : ceph-mgr-diskprediction-local-2:19.2.3-678.ge911 123/138 2026-03-09T15:11:39.603 INFO:teuthology.orchestra.run.vm04.stdout: Running scriptlet: ceph-mgr-diskprediction-local-2:19.2.3-678.ge911 123/138 2026-03-09T15:11:39.625 INFO:teuthology.orchestra.run.vm06.stdout: Installing : ceph-test-2:19.2.3-678.ge911bdeb.el9.x86_64 131/138 2026-03-09T15:11:39.635 INFO:teuthology.orchestra.run.vm06.stdout: Installing : rbd-fuse-2:19.2.3-678.ge911bdeb.el9.x86_64 132/138 2026-03-09T15:11:39.640 INFO:teuthology.orchestra.run.vm06.stdout: Installing : rbd-nbd-2:19.2.3-678.ge911bdeb.el9.x86_64 133/138 2026-03-09T15:11:39.661 INFO:teuthology.orchestra.run.vm04.stdout: Installing : ceph-mgr-modules-core-2:19.2.3-678.ge911bdeb.el9 124/138 2026-03-09T15:11:39.663 INFO:teuthology.orchestra.run.vm04.stdout: Installing : ceph-mgr-2:19.2.3-678.ge911bdeb.el9.x86_64 125/138 2026-03-09T15:11:39.683 INFO:teuthology.orchestra.run.vm04.stdout: Running scriptlet: ceph-mgr-2:19.2.3-678.ge911bdeb.el9.x86_64 125/138 2026-03-09T15:11:39.683 INFO:teuthology.orchestra.run.vm04.stdout:Glob pattern passed to enable, but globs are not supported for this. 2026-03-09T15:11:39.683 INFO:teuthology.orchestra.run.vm04.stdout:Invalid unit name "ceph-mgr@*.service" escaped as "ceph-mgr@\x2a.service". 2026-03-09T15:11:39.683 INFO:teuthology.orchestra.run.vm04.stdout:Created symlink /etc/systemd/system/multi-user.target.wants/ceph-mgr.target → /usr/lib/systemd/system/ceph-mgr.target. 2026-03-09T15:11:39.683 INFO:teuthology.orchestra.run.vm04.stdout:Created symlink /etc/systemd/system/ceph.target.wants/ceph-mgr.target → /usr/lib/systemd/system/ceph-mgr.target. 2026-03-09T15:11:39.683 INFO:teuthology.orchestra.run.vm04.stdout: 2026-03-09T15:11:39.694 INFO:teuthology.orchestra.run.vm06.stdout: Installing : libcephfs-devel-2:19.2.3-678.ge911bdeb.el9.x86_6 134/138 2026-03-09T15:11:39.698 INFO:teuthology.orchestra.run.vm04.stdout: Installing : ceph-mgr-rook-2:19.2.3-678.ge911bdeb.el9.noarch 126/138 2026-03-09T15:11:39.704 INFO:teuthology.orchestra.run.vm06.stdout: Installing : ceph-fuse-2:19.2.3-678.ge911bdeb.el9.x86_64 135/138 2026-03-09T15:11:39.708 INFO:teuthology.orchestra.run.vm06.stdout: Installing : python3-jmespath-1.0.1-1.el9.noarch 136/138 2026-03-09T15:11:39.708 INFO:teuthology.orchestra.run.vm06.stdout: Cleanup : librbd1-2:16.2.4-5.el9.x86_64 137/138 2026-03-09T15:11:39.709 INFO:teuthology.orchestra.run.vm04.stdout: Running scriptlet: ceph-mgr-rook-2:19.2.3-678.ge911bdeb.el9.noarch 126/138 2026-03-09T15:11:39.723 INFO:teuthology.orchestra.run.vm06.stdout: Running scriptlet: librbd1-2:16.2.4-5.el9.x86_64 137/138 2026-03-09T15:11:39.723 INFO:teuthology.orchestra.run.vm06.stdout: Cleanup : librados2-2:16.2.4-5.el9.x86_64 138/138 2026-03-09T15:11:40.270 INFO:teuthology.orchestra.run.vm04.stdout: Installing : ceph-2:19.2.3-678.ge911bdeb.el9.x86_64 127/138 2026-03-09T15:11:40.277 INFO:teuthology.orchestra.run.vm04.stdout: Installing : ceph-radosgw-2:19.2.3-678.ge911bdeb.el9.x86_64 128/138 2026-03-09T15:11:40.298 INFO:teuthology.orchestra.run.vm04.stdout: Running scriptlet: ceph-radosgw-2:19.2.3-678.ge911bdeb.el9.x86_64 128/138 2026-03-09T15:11:40.298 INFO:teuthology.orchestra.run.vm04.stdout:Glob pattern passed to enable, but globs are not supported for this. 2026-03-09T15:11:40.298 INFO:teuthology.orchestra.run.vm04.stdout:Invalid unit name "ceph-radosgw@*.service" escaped as "ceph-radosgw@\x2a.service". 2026-03-09T15:11:40.298 INFO:teuthology.orchestra.run.vm04.stdout:Created symlink /etc/systemd/system/multi-user.target.wants/ceph-radosgw.target → /usr/lib/systemd/system/ceph-radosgw.target. 2026-03-09T15:11:40.298 INFO:teuthology.orchestra.run.vm04.stdout:Created symlink /etc/systemd/system/ceph.target.wants/ceph-radosgw.target → /usr/lib/systemd/system/ceph-radosgw.target. 2026-03-09T15:11:40.298 INFO:teuthology.orchestra.run.vm04.stdout: 2026-03-09T15:11:40.322 INFO:teuthology.orchestra.run.vm04.stdout: Installing : ceph-immutable-object-cache-2:19.2.3-678.ge911bd 129/138 2026-03-09T15:11:40.345 INFO:teuthology.orchestra.run.vm04.stdout: Running scriptlet: ceph-immutable-object-cache-2:19.2.3-678.ge911bd 129/138 2026-03-09T15:11:40.345 INFO:teuthology.orchestra.run.vm04.stdout:Glob pattern passed to enable, but globs are not supported for this. 2026-03-09T15:11:40.345 INFO:teuthology.orchestra.run.vm04.stdout:Invalid unit name "ceph-immutable-object-cache@*.service" escaped as "ceph-immutable-object-cache@\x2a.service". 2026-03-09T15:11:40.345 INFO:teuthology.orchestra.run.vm04.stdout: 2026-03-09T15:11:40.573 INFO:teuthology.orchestra.run.vm04.stdout: Installing : rbd-mirror-2:19.2.3-678.ge911bdeb.el9.x86_64 130/138 2026-03-09T15:11:40.594 INFO:teuthology.orchestra.run.vm04.stdout: Running scriptlet: rbd-mirror-2:19.2.3-678.ge911bdeb.el9.x86_64 130/138 2026-03-09T15:11:40.595 INFO:teuthology.orchestra.run.vm04.stdout:Glob pattern passed to enable, but globs are not supported for this. 2026-03-09T15:11:40.595 INFO:teuthology.orchestra.run.vm04.stdout:Invalid unit name "ceph-rbd-mirror@*.service" escaped as "ceph-rbd-mirror@\x2a.service". 2026-03-09T15:11:40.595 INFO:teuthology.orchestra.run.vm04.stdout:Created symlink /etc/systemd/system/multi-user.target.wants/ceph-rbd-mirror.target → /usr/lib/systemd/system/ceph-rbd-mirror.target. 2026-03-09T15:11:40.595 INFO:teuthology.orchestra.run.vm04.stdout:Created symlink /etc/systemd/system/ceph.target.wants/ceph-rbd-mirror.target → /usr/lib/systemd/system/ceph-rbd-mirror.target. 2026-03-09T15:11:40.595 INFO:teuthology.orchestra.run.vm04.stdout: 2026-03-09T15:11:40.990 INFO:teuthology.orchestra.run.vm06.stdout: Running scriptlet: librados2-2:16.2.4-5.el9.x86_64 138/138 2026-03-09T15:11:40.990 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : ceph-2:19.2.3-678.ge911bdeb.el9.x86_64 1/138 2026-03-09T15:11:40.990 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : ceph-base-2:19.2.3-678.ge911bdeb.el9.x86_64 2/138 2026-03-09T15:11:40.990 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : ceph-common-2:19.2.3-678.ge911bdeb.el9.x86_64 3/138 2026-03-09T15:11:40.990 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : ceph-fuse-2:19.2.3-678.ge911bdeb.el9.x86_64 4/138 2026-03-09T15:11:40.990 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : ceph-immutable-object-cache-2:19.2.3-678.ge911bd 5/138 2026-03-09T15:11:40.990 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : ceph-mds-2:19.2.3-678.ge911bdeb.el9.x86_64 6/138 2026-03-09T15:11:40.990 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : ceph-mgr-2:19.2.3-678.ge911bdeb.el9.x86_64 7/138 2026-03-09T15:11:40.990 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : ceph-mon-2:19.2.3-678.ge911bdeb.el9.x86_64 8/138 2026-03-09T15:11:40.990 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : ceph-osd-2:19.2.3-678.ge911bdeb.el9.x86_64 9/138 2026-03-09T15:11:40.990 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : ceph-radosgw-2:19.2.3-678.ge911bdeb.el9.x86_64 10/138 2026-03-09T15:11:40.990 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : ceph-selinux-2:19.2.3-678.ge911bdeb.el9.x86_64 11/138 2026-03-09T15:11:40.990 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : ceph-test-2:19.2.3-678.ge911bdeb.el9.x86_64 12/138 2026-03-09T15:11:40.990 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : libcephfs-devel-2:19.2.3-678.ge911bdeb.el9.x86_6 13/138 2026-03-09T15:11:40.990 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : libcephfs2-2:19.2.3-678.ge911bdeb.el9.x86_64 14/138 2026-03-09T15:11:40.990 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : libcephsqlite-2:19.2.3-678.ge911bdeb.el9.x86_64 15/138 2026-03-09T15:11:40.990 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : librados-devel-2:19.2.3-678.ge911bdeb.el9.x86_64 16/138 2026-03-09T15:11:40.990 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : libradosstriper1-2:19.2.3-678.ge911bdeb.el9.x86_ 17/138 2026-03-09T15:11:40.990 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : librgw2-2:19.2.3-678.ge911bdeb.el9.x86_64 18/138 2026-03-09T15:11:40.990 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-ceph-argparse-2:19.2.3-678.ge911bdeb.el9 19/138 2026-03-09T15:11:40.990 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-ceph-common-2:19.2.3-678.ge911bdeb.el9.x 20/138 2026-03-09T15:11:40.990 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-cephfs-2:19.2.3-678.ge911bdeb.el9.x86_64 21/138 2026-03-09T15:11:40.990 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-rados-2:19.2.3-678.ge911bdeb.el9.x86_64 22/138 2026-03-09T15:11:40.990 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-rbd-2:19.2.3-678.ge911bdeb.el9.x86_64 23/138 2026-03-09T15:11:40.990 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-rgw-2:19.2.3-678.ge911bdeb.el9.x86_64 24/138 2026-03-09T15:11:40.990 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : rbd-fuse-2:19.2.3-678.ge911bdeb.el9.x86_64 25/138 2026-03-09T15:11:40.990 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : rbd-mirror-2:19.2.3-678.ge911bdeb.el9.x86_64 26/138 2026-03-09T15:11:40.990 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : rbd-nbd-2:19.2.3-678.ge911bdeb.el9.x86_64 27/138 2026-03-09T15:11:40.990 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : ceph-grafana-dashboards-2:19.2.3-678.ge911bdeb.e 28/138 2026-03-09T15:11:40.990 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : ceph-mgr-cephadm-2:19.2.3-678.ge911bdeb.el9.noar 29/138 2026-03-09T15:11:40.990 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : ceph-mgr-dashboard-2:19.2.3-678.ge911bdeb.el9.no 30/138 2026-03-09T15:11:40.990 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : ceph-mgr-diskprediction-local-2:19.2.3-678.ge911 31/138 2026-03-09T15:11:40.990 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : ceph-mgr-modules-core-2:19.2.3-678.ge911bdeb.el9 32/138 2026-03-09T15:11:40.990 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : ceph-mgr-rook-2:19.2.3-678.ge911bdeb.el9.noarch 33/138 2026-03-09T15:11:40.990 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : ceph-prometheus-alerts-2:19.2.3-678.ge911bdeb.el 34/138 2026-03-09T15:11:40.990 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : ceph-volume-2:19.2.3-678.ge911bdeb.el9.noarch 35/138 2026-03-09T15:11:40.990 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : cephadm-2:19.2.3-678.ge911bdeb.el9.noarch 36/138 2026-03-09T15:11:40.990 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : cryptsetup-2.8.1-3.el9.x86_64 37/138 2026-03-09T15:11:40.990 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : ledmon-libs-1.1.0-3.el9.x86_64 38/138 2026-03-09T15:11:40.990 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : libconfig-1.7.2-9.el9.x86_64 39/138 2026-03-09T15:11:40.992 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : libgfortran-11.5.0-14.el9.x86_64 40/138 2026-03-09T15:11:40.992 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : libquadmath-11.5.0-14.el9.x86_64 41/138 2026-03-09T15:11:40.992 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : mailcap-2.1.49-5.el9.noarch 42/138 2026-03-09T15:11:40.992 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : pciutils-3.7.0-7.el9.x86_64 43/138 2026-03-09T15:11:40.992 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-cffi-1.14.5-5.el9.x86_64 44/138 2026-03-09T15:11:40.992 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-cryptography-36.0.1-5.el9.x86_64 45/138 2026-03-09T15:11:40.992 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-ply-3.11-14.el9.noarch 46/138 2026-03-09T15:11:40.992 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-pycparser-2.20-6.el9.noarch 47/138 2026-03-09T15:11:40.992 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-requests-2.25.1-10.el9.noarch 48/138 2026-03-09T15:11:40.992 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-urllib3-1.26.5-7.el9.noarch 49/138 2026-03-09T15:11:40.992 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : unzip-6.0-59.el9.x86_64 50/138 2026-03-09T15:11:40.992 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : zip-3.0-35.el9.x86_64 51/138 2026-03-09T15:11:40.992 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : boost-program-options-1.75.0-13.el9.x86_64 52/138 2026-03-09T15:11:40.992 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : flexiblas-3.0.4-9.el9.x86_64 53/138 2026-03-09T15:11:40.992 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : flexiblas-netlib-3.0.4-9.el9.x86_64 54/138 2026-03-09T15:11:40.992 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : flexiblas-openblas-openmp-3.0.4-9.el9.x86_64 55/138 2026-03-09T15:11:40.992 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : libnbd-1.20.3-4.el9.x86_64 56/138 2026-03-09T15:11:40.992 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : libpmemobj-1.12.1-1.el9.x86_64 57/138 2026-03-09T15:11:40.992 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : librabbitmq-0.11.0-7.el9.x86_64 58/138 2026-03-09T15:11:40.992 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : librdkafka-1.6.1-102.el9.x86_64 59/138 2026-03-09T15:11:40.992 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : libstoragemgmt-1.10.1-1.el9.x86_64 60/138 2026-03-09T15:11:40.992 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : libxslt-1.1.34-12.el9.x86_64 61/138 2026-03-09T15:11:40.992 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : lttng-ust-2.12.0-6.el9.x86_64 62/138 2026-03-09T15:11:40.992 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : lua-5.4.4-4.el9.x86_64 63/138 2026-03-09T15:11:40.992 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : openblas-0.3.29-1.el9.x86_64 64/138 2026-03-09T15:11:40.992 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : openblas-openmp-0.3.29-1.el9.x86_64 65/138 2026-03-09T15:11:40.992 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : protobuf-3.14.0-17.el9.x86_64 66/138 2026-03-09T15:11:40.992 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-babel-2.9.1-2.el9.noarch 67/138 2026-03-09T15:11:40.992 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-devel-3.9.25-3.el9.x86_64 68/138 2026-03-09T15:11:40.992 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-jinja2-2.11.3-8.el9.noarch 69/138 2026-03-09T15:11:40.992 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-jmespath-1.0.1-1.el9.noarch 70/138 2026-03-09T15:11:40.992 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-libstoragemgmt-1.10.1-1.el9.x86_64 71/138 2026-03-09T15:11:40.992 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-mako-1.1.4-6.el9.noarch 72/138 2026-03-09T15:11:40.992 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-markupsafe-1.1.1-12.el9.x86_64 73/138 2026-03-09T15:11:40.992 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-numpy-1:1.23.5-2.el9.x86_64 74/138 2026-03-09T15:11:40.992 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-numpy-f2py-1:1.23.5-2.el9.x86_64 75/138 2026-03-09T15:11:40.992 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-packaging-20.9-5.el9.noarch 76/138 2026-03-09T15:11:40.992 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-protobuf-3.14.0-17.el9.noarch 77/138 2026-03-09T15:11:40.992 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-pyasn1-0.4.8-7.el9.noarch 78/138 2026-03-09T15:11:40.992 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-pyasn1-modules-0.4.8-7.el9.noarch 79/138 2026-03-09T15:11:40.992 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-requests-oauthlib-1.3.0-12.el9.noarch 80/138 2026-03-09T15:11:40.992 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-scipy-1.9.3-2.el9.x86_64 81/138 2026-03-09T15:11:40.992 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-toml-0.10.2-6.el9.noarch 82/138 2026-03-09T15:11:40.992 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : qatlib-25.08.0-2.el9.x86_64 83/138 2026-03-09T15:11:40.992 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : qatlib-service-25.08.0-2.el9.x86_64 84/138 2026-03-09T15:11:40.992 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : qatzip-libs-1.3.1-1.el9.x86_64 85/138 2026-03-09T15:11:40.992 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : socat-1.7.4.1-8.el9.x86_64 86/138 2026-03-09T15:11:40.992 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : xmlstarlet-1.6.1-20.el9.x86_64 87/138 2026-03-09T15:11:40.992 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : lua-devel-5.4.4-4.el9.x86_64 88/138 2026-03-09T15:11:40.993 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : protobuf-compiler-3.14.0-17.el9.x86_64 89/138 2026-03-09T15:11:40.993 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : abseil-cpp-20211102.0-4.el9.x86_64 90/138 2026-03-09T15:11:40.993 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : gperftools-libs-2.9.1-3.el9.x86_64 91/138 2026-03-09T15:11:40.993 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : grpc-data-1.46.7-10.el9.noarch 92/138 2026-03-09T15:11:40.993 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : libarrow-9.0.0-15.el9.x86_64 93/138 2026-03-09T15:11:40.993 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : libarrow-doc-9.0.0-15.el9.noarch 94/138 2026-03-09T15:11:40.993 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : liboath-2.6.12-1.el9.x86_64 95/138 2026-03-09T15:11:40.993 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : libunwind-1.6.2-1.el9.x86_64 96/138 2026-03-09T15:11:40.993 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : luarocks-3.9.2-5.el9.noarch 97/138 2026-03-09T15:11:40.993 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : parquet-libs-9.0.0-15.el9.x86_64 98/138 2026-03-09T15:11:40.993 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-asyncssh-2.13.2-5.el9.noarch 99/138 2026-03-09T15:11:40.993 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-autocommand-2.2.2-8.el9.noarch 100/138 2026-03-09T15:11:40.993 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-backports-tarfile-1.2.0-1.el9.noarch 101/138 2026-03-09T15:11:40.993 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-bcrypt-3.2.2-1.el9.x86_64 102/138 2026-03-09T15:11:40.993 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-cachetools-4.2.4-1.el9.noarch 103/138 2026-03-09T15:11:40.993 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-certifi-2023.05.07-4.el9.noarch 104/138 2026-03-09T15:11:40.994 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-cheroot-10.0.1-4.el9.noarch 105/138 2026-03-09T15:11:40.994 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-cherrypy-18.6.1-2.el9.noarch 106/138 2026-03-09T15:11:40.994 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-google-auth-1:2.45.0-1.el9.noarch 107/138 2026-03-09T15:11:40.994 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-grpcio-1.46.7-10.el9.x86_64 108/138 2026-03-09T15:11:40.994 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-grpcio-tools-1.46.7-10.el9.x86_64 109/138 2026-03-09T15:11:40.994 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-jaraco-8.2.1-3.el9.noarch 110/138 2026-03-09T15:11:40.994 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-jaraco-classes-3.2.1-5.el9.noarch 111/138 2026-03-09T15:11:40.994 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-jaraco-collections-3.0.0-8.el9.noarch 112/138 2026-03-09T15:11:40.994 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-jaraco-context-6.0.1-3.el9.noarch 113/138 2026-03-09T15:11:40.994 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-jaraco-functools-3.5.0-2.el9.noarch 114/138 2026-03-09T15:11:40.994 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-jaraco-text-4.0.0-2.el9.noarch 115/138 2026-03-09T15:11:40.994 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-kubernetes-1:26.1.0-3.el9.noarch 116/138 2026-03-09T15:11:40.994 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-logutils-0.3.5-21.el9.noarch 117/138 2026-03-09T15:11:40.994 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-more-itertools-8.12.0-2.el9.noarch 118/138 2026-03-09T15:11:40.994 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-natsort-7.1.1-5.el9.noarch 119/138 2026-03-09T15:11:40.994 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-pecan-1.4.2-3.el9.noarch 120/138 2026-03-09T15:11:40.994 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-portend-3.1.0-2.el9.noarch 121/138 2026-03-09T15:11:40.994 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-pyOpenSSL-21.0.0-1.el9.noarch 122/138 2026-03-09T15:11:40.994 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-repoze-lru-0.7-16.el9.noarch 123/138 2026-03-09T15:11:40.994 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-routes-2.5.1-5.el9.noarch 124/138 2026-03-09T15:11:40.994 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-rsa-4.9-2.el9.noarch 125/138 2026-03-09T15:11:40.994 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-tempora-5.0.0-2.el9.noarch 126/138 2026-03-09T15:11:40.994 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-typing-extensions-4.15.0-1.el9.noarch 127/138 2026-03-09T15:11:40.994 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-webob-1.8.8-2.el9.noarch 128/138 2026-03-09T15:11:40.994 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-websocket-client-1.2.3-2.el9.noarch 129/138 2026-03-09T15:11:40.994 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-werkzeug-2.0.3-3.el9.1.noarch 130/138 2026-03-09T15:11:40.994 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-xmltodict-0.12.0-15.el9.noarch 131/138 2026-03-09T15:11:40.994 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-zc-lockfile-2.0-10.el9.noarch 132/138 2026-03-09T15:11:40.994 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : re2-1:20211101-20.el9.x86_64 133/138 2026-03-09T15:11:40.994 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : thrift-0.15.0-4.el9.x86_64 134/138 2026-03-09T15:11:40.994 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : librados2-2:19.2.3-678.ge911bdeb.el9.x86_64 135/138 2026-03-09T15:11:40.994 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : librados2-2:16.2.4-5.el9.x86_64 136/138 2026-03-09T15:11:40.994 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : librbd1-2:19.2.3-678.ge911bdeb.el9.x86_64 137/138 2026-03-09T15:11:41.360 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : librbd1-2:16.2.4-5.el9.x86_64 138/138 2026-03-09T15:11:41.360 INFO:teuthology.orchestra.run.vm06.stdout: 2026-03-09T15:11:41.360 INFO:teuthology.orchestra.run.vm06.stdout:Upgraded: 2026-03-09T15:11:41.360 INFO:teuthology.orchestra.run.vm06.stdout: librados2-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T15:11:41.360 INFO:teuthology.orchestra.run.vm06.stdout: librbd1-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T15:11:41.360 INFO:teuthology.orchestra.run.vm06.stdout:Installed: 2026-03-09T15:11:41.360 INFO:teuthology.orchestra.run.vm06.stdout: abseil-cpp-20211102.0-4.el9.x86_64 2026-03-09T15:11:41.360 INFO:teuthology.orchestra.run.vm06.stdout: boost-program-options-1.75.0-13.el9.x86_64 2026-03-09T15:11:41.360 INFO:teuthology.orchestra.run.vm06.stdout: ceph-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T15:11:41.360 INFO:teuthology.orchestra.run.vm06.stdout: ceph-base-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T15:11:41.360 INFO:teuthology.orchestra.run.vm06.stdout: ceph-common-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T15:11:41.360 INFO:teuthology.orchestra.run.vm06.stdout: ceph-fuse-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T15:11:41.361 INFO:teuthology.orchestra.run.vm06.stdout: ceph-grafana-dashboards-2:19.2.3-678.ge911bdeb.el9.noarch 2026-03-09T15:11:41.361 INFO:teuthology.orchestra.run.vm06.stdout: ceph-immutable-object-cache-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T15:11:41.361 INFO:teuthology.orchestra.run.vm06.stdout: ceph-mds-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T15:11:41.361 INFO:teuthology.orchestra.run.vm06.stdout: ceph-mgr-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T15:11:41.361 INFO:teuthology.orchestra.run.vm06.stdout: ceph-mgr-cephadm-2:19.2.3-678.ge911bdeb.el9.noarch 2026-03-09T15:11:41.361 INFO:teuthology.orchestra.run.vm06.stdout: ceph-mgr-dashboard-2:19.2.3-678.ge911bdeb.el9.noarch 2026-03-09T15:11:41.361 INFO:teuthology.orchestra.run.vm06.stdout: ceph-mgr-diskprediction-local-2:19.2.3-678.ge911bdeb.el9.noarch 2026-03-09T15:11:41.361 INFO:teuthology.orchestra.run.vm06.stdout: ceph-mgr-modules-core-2:19.2.3-678.ge911bdeb.el9.noarch 2026-03-09T15:11:41.361 INFO:teuthology.orchestra.run.vm06.stdout: ceph-mgr-rook-2:19.2.3-678.ge911bdeb.el9.noarch 2026-03-09T15:11:41.361 INFO:teuthology.orchestra.run.vm06.stdout: ceph-mon-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T15:11:41.361 INFO:teuthology.orchestra.run.vm06.stdout: ceph-osd-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T15:11:41.361 INFO:teuthology.orchestra.run.vm06.stdout: ceph-prometheus-alerts-2:19.2.3-678.ge911bdeb.el9.noarch 2026-03-09T15:11:41.361 INFO:teuthology.orchestra.run.vm06.stdout: ceph-radosgw-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T15:11:41.361 INFO:teuthology.orchestra.run.vm06.stdout: ceph-selinux-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T15:11:41.361 INFO:teuthology.orchestra.run.vm06.stdout: ceph-test-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T15:11:41.361 INFO:teuthology.orchestra.run.vm06.stdout: ceph-volume-2:19.2.3-678.ge911bdeb.el9.noarch 2026-03-09T15:11:41.361 INFO:teuthology.orchestra.run.vm06.stdout: cephadm-2:19.2.3-678.ge911bdeb.el9.noarch 2026-03-09T15:11:41.361 INFO:teuthology.orchestra.run.vm06.stdout: cryptsetup-2.8.1-3.el9.x86_64 2026-03-09T15:11:41.361 INFO:teuthology.orchestra.run.vm06.stdout: flexiblas-3.0.4-9.el9.x86_64 2026-03-09T15:11:41.361 INFO:teuthology.orchestra.run.vm06.stdout: flexiblas-netlib-3.0.4-9.el9.x86_64 2026-03-09T15:11:41.361 INFO:teuthology.orchestra.run.vm06.stdout: flexiblas-openblas-openmp-3.0.4-9.el9.x86_64 2026-03-09T15:11:41.361 INFO:teuthology.orchestra.run.vm06.stdout: gperftools-libs-2.9.1-3.el9.x86_64 2026-03-09T15:11:41.361 INFO:teuthology.orchestra.run.vm06.stdout: grpc-data-1.46.7-10.el9.noarch 2026-03-09T15:11:41.361 INFO:teuthology.orchestra.run.vm06.stdout: ledmon-libs-1.1.0-3.el9.x86_64 2026-03-09T15:11:41.361 INFO:teuthology.orchestra.run.vm06.stdout: libarrow-9.0.0-15.el9.x86_64 2026-03-09T15:11:41.361 INFO:teuthology.orchestra.run.vm06.stdout: libarrow-doc-9.0.0-15.el9.noarch 2026-03-09T15:11:41.361 INFO:teuthology.orchestra.run.vm06.stdout: libcephfs-devel-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T15:11:41.361 INFO:teuthology.orchestra.run.vm06.stdout: libcephfs2-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T15:11:41.361 INFO:teuthology.orchestra.run.vm06.stdout: libcephsqlite-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T15:11:41.361 INFO:teuthology.orchestra.run.vm06.stdout: libconfig-1.7.2-9.el9.x86_64 2026-03-09T15:11:41.361 INFO:teuthology.orchestra.run.vm06.stdout: libgfortran-11.5.0-14.el9.x86_64 2026-03-09T15:11:41.361 INFO:teuthology.orchestra.run.vm06.stdout: libnbd-1.20.3-4.el9.x86_64 2026-03-09T15:11:41.361 INFO:teuthology.orchestra.run.vm06.stdout: liboath-2.6.12-1.el9.x86_64 2026-03-09T15:11:41.361 INFO:teuthology.orchestra.run.vm06.stdout: libpmemobj-1.12.1-1.el9.x86_64 2026-03-09T15:11:41.361 INFO:teuthology.orchestra.run.vm06.stdout: libquadmath-11.5.0-14.el9.x86_64 2026-03-09T15:11:41.361 INFO:teuthology.orchestra.run.vm06.stdout: librabbitmq-0.11.0-7.el9.x86_64 2026-03-09T15:11:41.361 INFO:teuthology.orchestra.run.vm06.stdout: librados-devel-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T15:11:41.361 INFO:teuthology.orchestra.run.vm06.stdout: libradosstriper1-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T15:11:41.361 INFO:teuthology.orchestra.run.vm06.stdout: librdkafka-1.6.1-102.el9.x86_64 2026-03-09T15:11:41.361 INFO:teuthology.orchestra.run.vm06.stdout: librgw2-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T15:11:41.361 INFO:teuthology.orchestra.run.vm06.stdout: libstoragemgmt-1.10.1-1.el9.x86_64 2026-03-09T15:11:41.361 INFO:teuthology.orchestra.run.vm06.stdout: libunwind-1.6.2-1.el9.x86_64 2026-03-09T15:11:41.361 INFO:teuthology.orchestra.run.vm06.stdout: libxslt-1.1.34-12.el9.x86_64 2026-03-09T15:11:41.361 INFO:teuthology.orchestra.run.vm06.stdout: lttng-ust-2.12.0-6.el9.x86_64 2026-03-09T15:11:41.361 INFO:teuthology.orchestra.run.vm06.stdout: lua-5.4.4-4.el9.x86_64 2026-03-09T15:11:41.361 INFO:teuthology.orchestra.run.vm06.stdout: lua-devel-5.4.4-4.el9.x86_64 2026-03-09T15:11:41.361 INFO:teuthology.orchestra.run.vm06.stdout: luarocks-3.9.2-5.el9.noarch 2026-03-09T15:11:41.361 INFO:teuthology.orchestra.run.vm06.stdout: mailcap-2.1.49-5.el9.noarch 2026-03-09T15:11:41.361 INFO:teuthology.orchestra.run.vm06.stdout: openblas-0.3.29-1.el9.x86_64 2026-03-09T15:11:41.361 INFO:teuthology.orchestra.run.vm06.stdout: openblas-openmp-0.3.29-1.el9.x86_64 2026-03-09T15:11:41.361 INFO:teuthology.orchestra.run.vm06.stdout: parquet-libs-9.0.0-15.el9.x86_64 2026-03-09T15:11:41.361 INFO:teuthology.orchestra.run.vm06.stdout: pciutils-3.7.0-7.el9.x86_64 2026-03-09T15:11:41.361 INFO:teuthology.orchestra.run.vm06.stdout: protobuf-3.14.0-17.el9.x86_64 2026-03-09T15:11:41.361 INFO:teuthology.orchestra.run.vm06.stdout: protobuf-compiler-3.14.0-17.el9.x86_64 2026-03-09T15:11:41.361 INFO:teuthology.orchestra.run.vm06.stdout: python3-asyncssh-2.13.2-5.el9.noarch 2026-03-09T15:11:41.361 INFO:teuthology.orchestra.run.vm06.stdout: python3-autocommand-2.2.2-8.el9.noarch 2026-03-09T15:11:41.361 INFO:teuthology.orchestra.run.vm06.stdout: python3-babel-2.9.1-2.el9.noarch 2026-03-09T15:11:41.361 INFO:teuthology.orchestra.run.vm06.stdout: python3-backports-tarfile-1.2.0-1.el9.noarch 2026-03-09T15:11:41.361 INFO:teuthology.orchestra.run.vm06.stdout: python3-bcrypt-3.2.2-1.el9.x86_64 2026-03-09T15:11:41.361 INFO:teuthology.orchestra.run.vm06.stdout: python3-cachetools-4.2.4-1.el9.noarch 2026-03-09T15:11:41.361 INFO:teuthology.orchestra.run.vm06.stdout: python3-ceph-argparse-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T15:11:41.362 INFO:teuthology.orchestra.run.vm06.stdout: python3-ceph-common-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T15:11:41.362 INFO:teuthology.orchestra.run.vm06.stdout: python3-cephfs-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T15:11:41.362 INFO:teuthology.orchestra.run.vm06.stdout: python3-certifi-2023.05.07-4.el9.noarch 2026-03-09T15:11:41.362 INFO:teuthology.orchestra.run.vm06.stdout: python3-cffi-1.14.5-5.el9.x86_64 2026-03-09T15:11:41.362 INFO:teuthology.orchestra.run.vm06.stdout: python3-cheroot-10.0.1-4.el9.noarch 2026-03-09T15:11:41.362 INFO:teuthology.orchestra.run.vm06.stdout: python3-cherrypy-18.6.1-2.el9.noarch 2026-03-09T15:11:41.362 INFO:teuthology.orchestra.run.vm06.stdout: python3-cryptography-36.0.1-5.el9.x86_64 2026-03-09T15:11:41.362 INFO:teuthology.orchestra.run.vm06.stdout: python3-devel-3.9.25-3.el9.x86_64 2026-03-09T15:11:41.362 INFO:teuthology.orchestra.run.vm06.stdout: python3-google-auth-1:2.45.0-1.el9.noarch 2026-03-09T15:11:41.362 INFO:teuthology.orchestra.run.vm06.stdout: python3-grpcio-1.46.7-10.el9.x86_64 2026-03-09T15:11:41.362 INFO:teuthology.orchestra.run.vm06.stdout: python3-grpcio-tools-1.46.7-10.el9.x86_64 2026-03-09T15:11:41.362 INFO:teuthology.orchestra.run.vm06.stdout: python3-jaraco-8.2.1-3.el9.noarch 2026-03-09T15:11:41.362 INFO:teuthology.orchestra.run.vm06.stdout: python3-jaraco-classes-3.2.1-5.el9.noarch 2026-03-09T15:11:41.362 INFO:teuthology.orchestra.run.vm06.stdout: python3-jaraco-collections-3.0.0-8.el9.noarch 2026-03-09T15:11:41.362 INFO:teuthology.orchestra.run.vm06.stdout: python3-jaraco-context-6.0.1-3.el9.noarch 2026-03-09T15:11:41.362 INFO:teuthology.orchestra.run.vm06.stdout: python3-jaraco-functools-3.5.0-2.el9.noarch 2026-03-09T15:11:41.362 INFO:teuthology.orchestra.run.vm06.stdout: python3-jaraco-text-4.0.0-2.el9.noarch 2026-03-09T15:11:41.362 INFO:teuthology.orchestra.run.vm06.stdout: python3-jinja2-2.11.3-8.el9.noarch 2026-03-09T15:11:41.362 INFO:teuthology.orchestra.run.vm06.stdout: python3-jmespath-1.0.1-1.el9.noarch 2026-03-09T15:11:41.362 INFO:teuthology.orchestra.run.vm06.stdout: python3-kubernetes-1:26.1.0-3.el9.noarch 2026-03-09T15:11:41.362 INFO:teuthology.orchestra.run.vm06.stdout: python3-libstoragemgmt-1.10.1-1.el9.x86_64 2026-03-09T15:11:41.362 INFO:teuthology.orchestra.run.vm06.stdout: python3-logutils-0.3.5-21.el9.noarch 2026-03-09T15:11:41.362 INFO:teuthology.orchestra.run.vm06.stdout: python3-mako-1.1.4-6.el9.noarch 2026-03-09T15:11:41.362 INFO:teuthology.orchestra.run.vm06.stdout: python3-markupsafe-1.1.1-12.el9.x86_64 2026-03-09T15:11:41.362 INFO:teuthology.orchestra.run.vm06.stdout: python3-more-itertools-8.12.0-2.el9.noarch 2026-03-09T15:11:41.362 INFO:teuthology.orchestra.run.vm06.stdout: python3-natsort-7.1.1-5.el9.noarch 2026-03-09T15:11:41.362 INFO:teuthology.orchestra.run.vm06.stdout: python3-numpy-1:1.23.5-2.el9.x86_64 2026-03-09T15:11:41.362 INFO:teuthology.orchestra.run.vm06.stdout: python3-numpy-f2py-1:1.23.5-2.el9.x86_64 2026-03-09T15:11:41.362 INFO:teuthology.orchestra.run.vm06.stdout: python3-packaging-20.9-5.el9.noarch 2026-03-09T15:11:41.362 INFO:teuthology.orchestra.run.vm06.stdout: python3-pecan-1.4.2-3.el9.noarch 2026-03-09T15:11:41.362 INFO:teuthology.orchestra.run.vm06.stdout: python3-ply-3.11-14.el9.noarch 2026-03-09T15:11:41.362 INFO:teuthology.orchestra.run.vm06.stdout: python3-portend-3.1.0-2.el9.noarch 2026-03-09T15:11:41.362 INFO:teuthology.orchestra.run.vm06.stdout: python3-protobuf-3.14.0-17.el9.noarch 2026-03-09T15:11:41.362 INFO:teuthology.orchestra.run.vm06.stdout: python3-pyOpenSSL-21.0.0-1.el9.noarch 2026-03-09T15:11:41.362 INFO:teuthology.orchestra.run.vm06.stdout: python3-pyasn1-0.4.8-7.el9.noarch 2026-03-09T15:11:41.362 INFO:teuthology.orchestra.run.vm06.stdout: python3-pyasn1-modules-0.4.8-7.el9.noarch 2026-03-09T15:11:41.362 INFO:teuthology.orchestra.run.vm06.stdout: python3-pycparser-2.20-6.el9.noarch 2026-03-09T15:11:41.362 INFO:teuthology.orchestra.run.vm06.stdout: python3-rados-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T15:11:41.362 INFO:teuthology.orchestra.run.vm06.stdout: python3-rbd-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T15:11:41.362 INFO:teuthology.orchestra.run.vm06.stdout: python3-repoze-lru-0.7-16.el9.noarch 2026-03-09T15:11:41.362 INFO:teuthology.orchestra.run.vm06.stdout: python3-requests-2.25.1-10.el9.noarch 2026-03-09T15:11:41.362 INFO:teuthology.orchestra.run.vm06.stdout: python3-requests-oauthlib-1.3.0-12.el9.noarch 2026-03-09T15:11:41.362 INFO:teuthology.orchestra.run.vm06.stdout: python3-rgw-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T15:11:41.362 INFO:teuthology.orchestra.run.vm06.stdout: python3-routes-2.5.1-5.el9.noarch 2026-03-09T15:11:41.362 INFO:teuthology.orchestra.run.vm06.stdout: python3-rsa-4.9-2.el9.noarch 2026-03-09T15:11:41.362 INFO:teuthology.orchestra.run.vm06.stdout: python3-scipy-1.9.3-2.el9.x86_64 2026-03-09T15:11:41.362 INFO:teuthology.orchestra.run.vm06.stdout: python3-tempora-5.0.0-2.el9.noarch 2026-03-09T15:11:41.362 INFO:teuthology.orchestra.run.vm06.stdout: python3-toml-0.10.2-6.el9.noarch 2026-03-09T15:11:41.362 INFO:teuthology.orchestra.run.vm06.stdout: python3-typing-extensions-4.15.0-1.el9.noarch 2026-03-09T15:11:41.362 INFO:teuthology.orchestra.run.vm06.stdout: python3-urllib3-1.26.5-7.el9.noarch 2026-03-09T15:11:41.362 INFO:teuthology.orchestra.run.vm06.stdout: python3-webob-1.8.8-2.el9.noarch 2026-03-09T15:11:41.362 INFO:teuthology.orchestra.run.vm06.stdout: python3-websocket-client-1.2.3-2.el9.noarch 2026-03-09T15:11:41.362 INFO:teuthology.orchestra.run.vm06.stdout: python3-werkzeug-2.0.3-3.el9.1.noarch 2026-03-09T15:11:41.362 INFO:teuthology.orchestra.run.vm06.stdout: python3-xmltodict-0.12.0-15.el9.noarch 2026-03-09T15:11:41.362 INFO:teuthology.orchestra.run.vm06.stdout: python3-zc-lockfile-2.0-10.el9.noarch 2026-03-09T15:11:41.362 INFO:teuthology.orchestra.run.vm06.stdout: qatlib-25.08.0-2.el9.x86_64 2026-03-09T15:11:41.362 INFO:teuthology.orchestra.run.vm06.stdout: qatlib-service-25.08.0-2.el9.x86_64 2026-03-09T15:11:41.362 INFO:teuthology.orchestra.run.vm06.stdout: qatzip-libs-1.3.1-1.el9.x86_64 2026-03-09T15:11:41.362 INFO:teuthology.orchestra.run.vm06.stdout: rbd-fuse-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T15:11:41.362 INFO:teuthology.orchestra.run.vm06.stdout: rbd-mirror-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T15:11:41.362 INFO:teuthology.orchestra.run.vm06.stdout: rbd-nbd-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T15:11:41.362 INFO:teuthology.orchestra.run.vm06.stdout: re2-1:20211101-20.el9.x86_64 2026-03-09T15:11:41.362 INFO:teuthology.orchestra.run.vm06.stdout: socat-1.7.4.1-8.el9.x86_64 2026-03-09T15:11:41.363 INFO:teuthology.orchestra.run.vm06.stdout: thrift-0.15.0-4.el9.x86_64 2026-03-09T15:11:41.363 INFO:teuthology.orchestra.run.vm06.stdout: unzip-6.0-59.el9.x86_64 2026-03-09T15:11:41.363 INFO:teuthology.orchestra.run.vm06.stdout: xmlstarlet-1.6.1-20.el9.x86_64 2026-03-09T15:11:41.363 INFO:teuthology.orchestra.run.vm06.stdout: zip-3.0-35.el9.x86_64 2026-03-09T15:11:41.363 INFO:teuthology.orchestra.run.vm06.stdout: 2026-03-09T15:11:41.363 INFO:teuthology.orchestra.run.vm06.stdout:Complete! 2026-03-09T15:11:41.443 DEBUG:teuthology.parallel:result is None 2026-03-09T15:11:43.253 INFO:teuthology.orchestra.run.vm04.stdout: Installing : ceph-test-2:19.2.3-678.ge911bdeb.el9.x86_64 131/138 2026-03-09T15:11:43.265 INFO:teuthology.orchestra.run.vm04.stdout: Installing : rbd-fuse-2:19.2.3-678.ge911bdeb.el9.x86_64 132/138 2026-03-09T15:11:43.272 INFO:teuthology.orchestra.run.vm04.stdout: Installing : rbd-nbd-2:19.2.3-678.ge911bdeb.el9.x86_64 133/138 2026-03-09T15:11:43.329 INFO:teuthology.orchestra.run.vm04.stdout: Installing : libcephfs-devel-2:19.2.3-678.ge911bdeb.el9.x86_6 134/138 2026-03-09T15:11:43.339 INFO:teuthology.orchestra.run.vm04.stdout: Installing : ceph-fuse-2:19.2.3-678.ge911bdeb.el9.x86_64 135/138 2026-03-09T15:11:43.342 INFO:teuthology.orchestra.run.vm04.stdout: Installing : python3-jmespath-1.0.1-1.el9.noarch 136/138 2026-03-09T15:11:43.342 INFO:teuthology.orchestra.run.vm04.stdout: Cleanup : librbd1-2:16.2.4-5.el9.x86_64 137/138 2026-03-09T15:11:43.358 INFO:teuthology.orchestra.run.vm04.stdout: Running scriptlet: librbd1-2:16.2.4-5.el9.x86_64 137/138 2026-03-09T15:11:43.358 INFO:teuthology.orchestra.run.vm04.stdout: Cleanup : librados2-2:16.2.4-5.el9.x86_64 138/138 2026-03-09T15:11:44.602 INFO:teuthology.orchestra.run.vm04.stdout: Running scriptlet: librados2-2:16.2.4-5.el9.x86_64 138/138 2026-03-09T15:11:44.602 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : ceph-2:19.2.3-678.ge911bdeb.el9.x86_64 1/138 2026-03-09T15:11:44.602 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : ceph-base-2:19.2.3-678.ge911bdeb.el9.x86_64 2/138 2026-03-09T15:11:44.602 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : ceph-common-2:19.2.3-678.ge911bdeb.el9.x86_64 3/138 2026-03-09T15:11:44.602 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : ceph-fuse-2:19.2.3-678.ge911bdeb.el9.x86_64 4/138 2026-03-09T15:11:44.602 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : ceph-immutable-object-cache-2:19.2.3-678.ge911bd 5/138 2026-03-09T15:11:44.602 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : ceph-mds-2:19.2.3-678.ge911bdeb.el9.x86_64 6/138 2026-03-09T15:11:44.602 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : ceph-mgr-2:19.2.3-678.ge911bdeb.el9.x86_64 7/138 2026-03-09T15:11:44.602 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : ceph-mon-2:19.2.3-678.ge911bdeb.el9.x86_64 8/138 2026-03-09T15:11:44.602 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : ceph-osd-2:19.2.3-678.ge911bdeb.el9.x86_64 9/138 2026-03-09T15:11:44.603 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : ceph-radosgw-2:19.2.3-678.ge911bdeb.el9.x86_64 10/138 2026-03-09T15:11:44.603 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : ceph-selinux-2:19.2.3-678.ge911bdeb.el9.x86_64 11/138 2026-03-09T15:11:44.603 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : ceph-test-2:19.2.3-678.ge911bdeb.el9.x86_64 12/138 2026-03-09T15:11:44.603 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : libcephfs-devel-2:19.2.3-678.ge911bdeb.el9.x86_6 13/138 2026-03-09T15:11:44.603 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : libcephfs2-2:19.2.3-678.ge911bdeb.el9.x86_64 14/138 2026-03-09T15:11:44.603 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : libcephsqlite-2:19.2.3-678.ge911bdeb.el9.x86_64 15/138 2026-03-09T15:11:44.603 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : librados-devel-2:19.2.3-678.ge911bdeb.el9.x86_64 16/138 2026-03-09T15:11:44.603 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : libradosstriper1-2:19.2.3-678.ge911bdeb.el9.x86_ 17/138 2026-03-09T15:11:44.603 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : librgw2-2:19.2.3-678.ge911bdeb.el9.x86_64 18/138 2026-03-09T15:11:44.603 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-ceph-argparse-2:19.2.3-678.ge911bdeb.el9 19/138 2026-03-09T15:11:44.603 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-ceph-common-2:19.2.3-678.ge911bdeb.el9.x 20/138 2026-03-09T15:11:44.603 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-cephfs-2:19.2.3-678.ge911bdeb.el9.x86_64 21/138 2026-03-09T15:11:44.603 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-rados-2:19.2.3-678.ge911bdeb.el9.x86_64 22/138 2026-03-09T15:11:44.603 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-rbd-2:19.2.3-678.ge911bdeb.el9.x86_64 23/138 2026-03-09T15:11:44.603 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-rgw-2:19.2.3-678.ge911bdeb.el9.x86_64 24/138 2026-03-09T15:11:44.603 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : rbd-fuse-2:19.2.3-678.ge911bdeb.el9.x86_64 25/138 2026-03-09T15:11:44.603 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : rbd-mirror-2:19.2.3-678.ge911bdeb.el9.x86_64 26/138 2026-03-09T15:11:44.603 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : rbd-nbd-2:19.2.3-678.ge911bdeb.el9.x86_64 27/138 2026-03-09T15:11:44.603 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : ceph-grafana-dashboards-2:19.2.3-678.ge911bdeb.e 28/138 2026-03-09T15:11:44.603 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : ceph-mgr-cephadm-2:19.2.3-678.ge911bdeb.el9.noar 29/138 2026-03-09T15:11:44.603 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : ceph-mgr-dashboard-2:19.2.3-678.ge911bdeb.el9.no 30/138 2026-03-09T15:11:44.603 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : ceph-mgr-diskprediction-local-2:19.2.3-678.ge911 31/138 2026-03-09T15:11:44.603 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : ceph-mgr-modules-core-2:19.2.3-678.ge911bdeb.el9 32/138 2026-03-09T15:11:44.603 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : ceph-mgr-rook-2:19.2.3-678.ge911bdeb.el9.noarch 33/138 2026-03-09T15:11:44.603 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : ceph-prometheus-alerts-2:19.2.3-678.ge911bdeb.el 34/138 2026-03-09T15:11:44.603 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : ceph-volume-2:19.2.3-678.ge911bdeb.el9.noarch 35/138 2026-03-09T15:11:44.603 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : cephadm-2:19.2.3-678.ge911bdeb.el9.noarch 36/138 2026-03-09T15:11:44.603 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : cryptsetup-2.8.1-3.el9.x86_64 37/138 2026-03-09T15:11:44.603 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : ledmon-libs-1.1.0-3.el9.x86_64 38/138 2026-03-09T15:11:44.604 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : libconfig-1.7.2-9.el9.x86_64 39/138 2026-03-09T15:11:44.604 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : libgfortran-11.5.0-14.el9.x86_64 40/138 2026-03-09T15:11:44.604 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : libquadmath-11.5.0-14.el9.x86_64 41/138 2026-03-09T15:11:44.604 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : mailcap-2.1.49-5.el9.noarch 42/138 2026-03-09T15:11:44.604 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : pciutils-3.7.0-7.el9.x86_64 43/138 2026-03-09T15:11:44.604 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-cffi-1.14.5-5.el9.x86_64 44/138 2026-03-09T15:11:44.604 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-cryptography-36.0.1-5.el9.x86_64 45/138 2026-03-09T15:11:44.604 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-ply-3.11-14.el9.noarch 46/138 2026-03-09T15:11:44.604 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-pycparser-2.20-6.el9.noarch 47/138 2026-03-09T15:11:44.604 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-requests-2.25.1-10.el9.noarch 48/138 2026-03-09T15:11:44.604 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-urllib3-1.26.5-7.el9.noarch 49/138 2026-03-09T15:11:44.604 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : unzip-6.0-59.el9.x86_64 50/138 2026-03-09T15:11:44.604 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : zip-3.0-35.el9.x86_64 51/138 2026-03-09T15:11:44.604 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : boost-program-options-1.75.0-13.el9.x86_64 52/138 2026-03-09T15:11:44.604 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : flexiblas-3.0.4-9.el9.x86_64 53/138 2026-03-09T15:11:44.604 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : flexiblas-netlib-3.0.4-9.el9.x86_64 54/138 2026-03-09T15:11:44.604 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : flexiblas-openblas-openmp-3.0.4-9.el9.x86_64 55/138 2026-03-09T15:11:44.604 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : libnbd-1.20.3-4.el9.x86_64 56/138 2026-03-09T15:11:44.604 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : libpmemobj-1.12.1-1.el9.x86_64 57/138 2026-03-09T15:11:44.604 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : librabbitmq-0.11.0-7.el9.x86_64 58/138 2026-03-09T15:11:44.604 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : librdkafka-1.6.1-102.el9.x86_64 59/138 2026-03-09T15:11:44.604 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : libstoragemgmt-1.10.1-1.el9.x86_64 60/138 2026-03-09T15:11:44.604 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : libxslt-1.1.34-12.el9.x86_64 61/138 2026-03-09T15:11:44.604 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : lttng-ust-2.12.0-6.el9.x86_64 62/138 2026-03-09T15:11:44.604 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : lua-5.4.4-4.el9.x86_64 63/138 2026-03-09T15:11:44.605 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : openblas-0.3.29-1.el9.x86_64 64/138 2026-03-09T15:11:44.605 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : openblas-openmp-0.3.29-1.el9.x86_64 65/138 2026-03-09T15:11:44.605 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : protobuf-3.14.0-17.el9.x86_64 66/138 2026-03-09T15:11:44.605 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-babel-2.9.1-2.el9.noarch 67/138 2026-03-09T15:11:44.605 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-devel-3.9.25-3.el9.x86_64 68/138 2026-03-09T15:11:44.605 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-jinja2-2.11.3-8.el9.noarch 69/138 2026-03-09T15:11:44.605 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-jmespath-1.0.1-1.el9.noarch 70/138 2026-03-09T15:11:44.605 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-libstoragemgmt-1.10.1-1.el9.x86_64 71/138 2026-03-09T15:11:44.605 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-mako-1.1.4-6.el9.noarch 72/138 2026-03-09T15:11:44.605 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-markupsafe-1.1.1-12.el9.x86_64 73/138 2026-03-09T15:11:44.605 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-numpy-1:1.23.5-2.el9.x86_64 74/138 2026-03-09T15:11:44.605 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-numpy-f2py-1:1.23.5-2.el9.x86_64 75/138 2026-03-09T15:11:44.605 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-packaging-20.9-5.el9.noarch 76/138 2026-03-09T15:11:44.605 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-protobuf-3.14.0-17.el9.noarch 77/138 2026-03-09T15:11:44.605 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-pyasn1-0.4.8-7.el9.noarch 78/138 2026-03-09T15:11:44.605 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-pyasn1-modules-0.4.8-7.el9.noarch 79/138 2026-03-09T15:11:44.605 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-requests-oauthlib-1.3.0-12.el9.noarch 80/138 2026-03-09T15:11:44.605 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-scipy-1.9.3-2.el9.x86_64 81/138 2026-03-09T15:11:44.605 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-toml-0.10.2-6.el9.noarch 82/138 2026-03-09T15:11:44.605 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : qatlib-25.08.0-2.el9.x86_64 83/138 2026-03-09T15:11:44.606 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : qatlib-service-25.08.0-2.el9.x86_64 84/138 2026-03-09T15:11:44.606 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : qatzip-libs-1.3.1-1.el9.x86_64 85/138 2026-03-09T15:11:44.606 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : socat-1.7.4.1-8.el9.x86_64 86/138 2026-03-09T15:11:44.606 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : xmlstarlet-1.6.1-20.el9.x86_64 87/138 2026-03-09T15:11:44.606 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : lua-devel-5.4.4-4.el9.x86_64 88/138 2026-03-09T15:11:44.606 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : protobuf-compiler-3.14.0-17.el9.x86_64 89/138 2026-03-09T15:11:44.606 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : abseil-cpp-20211102.0-4.el9.x86_64 90/138 2026-03-09T15:11:44.606 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : gperftools-libs-2.9.1-3.el9.x86_64 91/138 2026-03-09T15:11:44.606 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : grpc-data-1.46.7-10.el9.noarch 92/138 2026-03-09T15:11:44.606 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : libarrow-9.0.0-15.el9.x86_64 93/138 2026-03-09T15:11:44.606 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : libarrow-doc-9.0.0-15.el9.noarch 94/138 2026-03-09T15:11:44.606 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : liboath-2.6.12-1.el9.x86_64 95/138 2026-03-09T15:11:44.606 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : libunwind-1.6.2-1.el9.x86_64 96/138 2026-03-09T15:11:44.606 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : luarocks-3.9.2-5.el9.noarch 97/138 2026-03-09T15:11:44.606 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : parquet-libs-9.0.0-15.el9.x86_64 98/138 2026-03-09T15:11:44.606 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-asyncssh-2.13.2-5.el9.noarch 99/138 2026-03-09T15:11:44.606 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-autocommand-2.2.2-8.el9.noarch 100/138 2026-03-09T15:11:44.606 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-backports-tarfile-1.2.0-1.el9.noarch 101/138 2026-03-09T15:11:44.606 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-bcrypt-3.2.2-1.el9.x86_64 102/138 2026-03-09T15:11:44.606 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-cachetools-4.2.4-1.el9.noarch 103/138 2026-03-09T15:11:44.606 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-certifi-2023.05.07-4.el9.noarch 104/138 2026-03-09T15:11:44.606 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-cheroot-10.0.1-4.el9.noarch 105/138 2026-03-09T15:11:44.606 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-cherrypy-18.6.1-2.el9.noarch 106/138 2026-03-09T15:11:44.606 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-google-auth-1:2.45.0-1.el9.noarch 107/138 2026-03-09T15:11:44.607 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-grpcio-1.46.7-10.el9.x86_64 108/138 2026-03-09T15:11:44.607 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-grpcio-tools-1.46.7-10.el9.x86_64 109/138 2026-03-09T15:11:44.607 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-jaraco-8.2.1-3.el9.noarch 110/138 2026-03-09T15:11:44.607 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-jaraco-classes-3.2.1-5.el9.noarch 111/138 2026-03-09T15:11:44.607 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-jaraco-collections-3.0.0-8.el9.noarch 112/138 2026-03-09T15:11:44.607 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-jaraco-context-6.0.1-3.el9.noarch 113/138 2026-03-09T15:11:44.607 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-jaraco-functools-3.5.0-2.el9.noarch 114/138 2026-03-09T15:11:44.607 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-jaraco-text-4.0.0-2.el9.noarch 115/138 2026-03-09T15:11:44.607 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-kubernetes-1:26.1.0-3.el9.noarch 116/138 2026-03-09T15:11:44.607 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-logutils-0.3.5-21.el9.noarch 117/138 2026-03-09T15:11:44.607 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-more-itertools-8.12.0-2.el9.noarch 118/138 2026-03-09T15:11:44.607 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-natsort-7.1.1-5.el9.noarch 119/138 2026-03-09T15:11:44.607 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-pecan-1.4.2-3.el9.noarch 120/138 2026-03-09T15:11:44.607 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-portend-3.1.0-2.el9.noarch 121/138 2026-03-09T15:11:44.607 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-pyOpenSSL-21.0.0-1.el9.noarch 122/138 2026-03-09T15:11:44.607 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-repoze-lru-0.7-16.el9.noarch 123/138 2026-03-09T15:11:44.607 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-routes-2.5.1-5.el9.noarch 124/138 2026-03-09T15:11:44.607 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-rsa-4.9-2.el9.noarch 125/138 2026-03-09T15:11:44.607 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-tempora-5.0.0-2.el9.noarch 126/138 2026-03-09T15:11:44.607 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-typing-extensions-4.15.0-1.el9.noarch 127/138 2026-03-09T15:11:44.607 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-webob-1.8.8-2.el9.noarch 128/138 2026-03-09T15:11:44.607 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-websocket-client-1.2.3-2.el9.noarch 129/138 2026-03-09T15:11:44.607 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-werkzeug-2.0.3-3.el9.1.noarch 130/138 2026-03-09T15:11:44.607 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-xmltodict-0.12.0-15.el9.noarch 131/138 2026-03-09T15:11:44.607 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-zc-lockfile-2.0-10.el9.noarch 132/138 2026-03-09T15:11:44.608 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : re2-1:20211101-20.el9.x86_64 133/138 2026-03-09T15:11:44.608 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : thrift-0.15.0-4.el9.x86_64 134/138 2026-03-09T15:11:44.608 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : librados2-2:19.2.3-678.ge911bdeb.el9.x86_64 135/138 2026-03-09T15:11:44.608 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : librados2-2:16.2.4-5.el9.x86_64 136/138 2026-03-09T15:11:44.608 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : librbd1-2:19.2.3-678.ge911bdeb.el9.x86_64 137/138 2026-03-09T15:11:44.708 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : librbd1-2:16.2.4-5.el9.x86_64 138/138 2026-03-09T15:11:44.708 INFO:teuthology.orchestra.run.vm04.stdout: 2026-03-09T15:11:44.709 INFO:teuthology.orchestra.run.vm04.stdout:Upgraded: 2026-03-09T15:11:44.709 INFO:teuthology.orchestra.run.vm04.stdout: librados2-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T15:11:44.709 INFO:teuthology.orchestra.run.vm04.stdout: librbd1-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T15:11:44.709 INFO:teuthology.orchestra.run.vm04.stdout:Installed: 2026-03-09T15:11:44.709 INFO:teuthology.orchestra.run.vm04.stdout: abseil-cpp-20211102.0-4.el9.x86_64 2026-03-09T15:11:44.709 INFO:teuthology.orchestra.run.vm04.stdout: boost-program-options-1.75.0-13.el9.x86_64 2026-03-09T15:11:44.709 INFO:teuthology.orchestra.run.vm04.stdout: ceph-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T15:11:44.709 INFO:teuthology.orchestra.run.vm04.stdout: ceph-base-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T15:11:44.709 INFO:teuthology.orchestra.run.vm04.stdout: ceph-common-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T15:11:44.709 INFO:teuthology.orchestra.run.vm04.stdout: ceph-fuse-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T15:11:44.709 INFO:teuthology.orchestra.run.vm04.stdout: ceph-grafana-dashboards-2:19.2.3-678.ge911bdeb.el9.noarch 2026-03-09T15:11:44.709 INFO:teuthology.orchestra.run.vm04.stdout: ceph-immutable-object-cache-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T15:11:44.709 INFO:teuthology.orchestra.run.vm04.stdout: ceph-mds-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T15:11:44.709 INFO:teuthology.orchestra.run.vm04.stdout: ceph-mgr-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T15:11:44.709 INFO:teuthology.orchestra.run.vm04.stdout: ceph-mgr-cephadm-2:19.2.3-678.ge911bdeb.el9.noarch 2026-03-09T15:11:44.709 INFO:teuthology.orchestra.run.vm04.stdout: ceph-mgr-dashboard-2:19.2.3-678.ge911bdeb.el9.noarch 2026-03-09T15:11:44.709 INFO:teuthology.orchestra.run.vm04.stdout: ceph-mgr-diskprediction-local-2:19.2.3-678.ge911bdeb.el9.noarch 2026-03-09T15:11:44.709 INFO:teuthology.orchestra.run.vm04.stdout: ceph-mgr-modules-core-2:19.2.3-678.ge911bdeb.el9.noarch 2026-03-09T15:11:44.709 INFO:teuthology.orchestra.run.vm04.stdout: ceph-mgr-rook-2:19.2.3-678.ge911bdeb.el9.noarch 2026-03-09T15:11:44.709 INFO:teuthology.orchestra.run.vm04.stdout: ceph-mon-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T15:11:44.709 INFO:teuthology.orchestra.run.vm04.stdout: ceph-osd-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T15:11:44.709 INFO:teuthology.orchestra.run.vm04.stdout: ceph-prometheus-alerts-2:19.2.3-678.ge911bdeb.el9.noarch 2026-03-09T15:11:44.709 INFO:teuthology.orchestra.run.vm04.stdout: ceph-radosgw-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T15:11:44.709 INFO:teuthology.orchestra.run.vm04.stdout: ceph-selinux-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T15:11:44.709 INFO:teuthology.orchestra.run.vm04.stdout: ceph-test-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T15:11:44.709 INFO:teuthology.orchestra.run.vm04.stdout: ceph-volume-2:19.2.3-678.ge911bdeb.el9.noarch 2026-03-09T15:11:44.709 INFO:teuthology.orchestra.run.vm04.stdout: cephadm-2:19.2.3-678.ge911bdeb.el9.noarch 2026-03-09T15:11:44.709 INFO:teuthology.orchestra.run.vm04.stdout: cryptsetup-2.8.1-3.el9.x86_64 2026-03-09T15:11:44.709 INFO:teuthology.orchestra.run.vm04.stdout: flexiblas-3.0.4-9.el9.x86_64 2026-03-09T15:11:44.709 INFO:teuthology.orchestra.run.vm04.stdout: flexiblas-netlib-3.0.4-9.el9.x86_64 2026-03-09T15:11:44.709 INFO:teuthology.orchestra.run.vm04.stdout: flexiblas-openblas-openmp-3.0.4-9.el9.x86_64 2026-03-09T15:11:44.709 INFO:teuthology.orchestra.run.vm04.stdout: gperftools-libs-2.9.1-3.el9.x86_64 2026-03-09T15:11:44.709 INFO:teuthology.orchestra.run.vm04.stdout: grpc-data-1.46.7-10.el9.noarch 2026-03-09T15:11:44.709 INFO:teuthology.orchestra.run.vm04.stdout: ledmon-libs-1.1.0-3.el9.x86_64 2026-03-09T15:11:44.709 INFO:teuthology.orchestra.run.vm04.stdout: libarrow-9.0.0-15.el9.x86_64 2026-03-09T15:11:44.709 INFO:teuthology.orchestra.run.vm04.stdout: libarrow-doc-9.0.0-15.el9.noarch 2026-03-09T15:11:44.709 INFO:teuthology.orchestra.run.vm04.stdout: libcephfs-devel-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T15:11:44.709 INFO:teuthology.orchestra.run.vm04.stdout: libcephfs2-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T15:11:44.709 INFO:teuthology.orchestra.run.vm04.stdout: libcephsqlite-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T15:11:44.709 INFO:teuthology.orchestra.run.vm04.stdout: libconfig-1.7.2-9.el9.x86_64 2026-03-09T15:11:44.709 INFO:teuthology.orchestra.run.vm04.stdout: libgfortran-11.5.0-14.el9.x86_64 2026-03-09T15:11:44.709 INFO:teuthology.orchestra.run.vm04.stdout: libnbd-1.20.3-4.el9.x86_64 2026-03-09T15:11:44.709 INFO:teuthology.orchestra.run.vm04.stdout: liboath-2.6.12-1.el9.x86_64 2026-03-09T15:11:44.709 INFO:teuthology.orchestra.run.vm04.stdout: libpmemobj-1.12.1-1.el9.x86_64 2026-03-09T15:11:44.709 INFO:teuthology.orchestra.run.vm04.stdout: libquadmath-11.5.0-14.el9.x86_64 2026-03-09T15:11:44.709 INFO:teuthology.orchestra.run.vm04.stdout: librabbitmq-0.11.0-7.el9.x86_64 2026-03-09T15:11:44.709 INFO:teuthology.orchestra.run.vm04.stdout: librados-devel-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T15:11:44.709 INFO:teuthology.orchestra.run.vm04.stdout: libradosstriper1-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T15:11:44.709 INFO:teuthology.orchestra.run.vm04.stdout: librdkafka-1.6.1-102.el9.x86_64 2026-03-09T15:11:44.709 INFO:teuthology.orchestra.run.vm04.stdout: librgw2-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T15:11:44.709 INFO:teuthology.orchestra.run.vm04.stdout: libstoragemgmt-1.10.1-1.el9.x86_64 2026-03-09T15:11:44.709 INFO:teuthology.orchestra.run.vm04.stdout: libunwind-1.6.2-1.el9.x86_64 2026-03-09T15:11:44.710 INFO:teuthology.orchestra.run.vm04.stdout: libxslt-1.1.34-12.el9.x86_64 2026-03-09T15:11:44.710 INFO:teuthology.orchestra.run.vm04.stdout: lttng-ust-2.12.0-6.el9.x86_64 2026-03-09T15:11:44.710 INFO:teuthology.orchestra.run.vm04.stdout: lua-5.4.4-4.el9.x86_64 2026-03-09T15:11:44.710 INFO:teuthology.orchestra.run.vm04.stdout: lua-devel-5.4.4-4.el9.x86_64 2026-03-09T15:11:44.710 INFO:teuthology.orchestra.run.vm04.stdout: luarocks-3.9.2-5.el9.noarch 2026-03-09T15:11:44.710 INFO:teuthology.orchestra.run.vm04.stdout: mailcap-2.1.49-5.el9.noarch 2026-03-09T15:11:44.710 INFO:teuthology.orchestra.run.vm04.stdout: openblas-0.3.29-1.el9.x86_64 2026-03-09T15:11:44.710 INFO:teuthology.orchestra.run.vm04.stdout: openblas-openmp-0.3.29-1.el9.x86_64 2026-03-09T15:11:44.710 INFO:teuthology.orchestra.run.vm04.stdout: parquet-libs-9.0.0-15.el9.x86_64 2026-03-09T15:11:44.710 INFO:teuthology.orchestra.run.vm04.stdout: pciutils-3.7.0-7.el9.x86_64 2026-03-09T15:11:44.710 INFO:teuthology.orchestra.run.vm04.stdout: protobuf-3.14.0-17.el9.x86_64 2026-03-09T15:11:44.710 INFO:teuthology.orchestra.run.vm04.stdout: protobuf-compiler-3.14.0-17.el9.x86_64 2026-03-09T15:11:44.710 INFO:teuthology.orchestra.run.vm04.stdout: python3-asyncssh-2.13.2-5.el9.noarch 2026-03-09T15:11:44.710 INFO:teuthology.orchestra.run.vm04.stdout: python3-autocommand-2.2.2-8.el9.noarch 2026-03-09T15:11:44.710 INFO:teuthology.orchestra.run.vm04.stdout: python3-babel-2.9.1-2.el9.noarch 2026-03-09T15:11:44.710 INFO:teuthology.orchestra.run.vm04.stdout: python3-backports-tarfile-1.2.0-1.el9.noarch 2026-03-09T15:11:44.710 INFO:teuthology.orchestra.run.vm04.stdout: python3-bcrypt-3.2.2-1.el9.x86_64 2026-03-09T15:11:44.710 INFO:teuthology.orchestra.run.vm04.stdout: python3-cachetools-4.2.4-1.el9.noarch 2026-03-09T15:11:44.710 INFO:teuthology.orchestra.run.vm04.stdout: python3-ceph-argparse-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T15:11:44.710 INFO:teuthology.orchestra.run.vm04.stdout: python3-ceph-common-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T15:11:44.710 INFO:teuthology.orchestra.run.vm04.stdout: python3-cephfs-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T15:11:44.710 INFO:teuthology.orchestra.run.vm04.stdout: python3-certifi-2023.05.07-4.el9.noarch 2026-03-09T15:11:44.710 INFO:teuthology.orchestra.run.vm04.stdout: python3-cffi-1.14.5-5.el9.x86_64 2026-03-09T15:11:44.710 INFO:teuthology.orchestra.run.vm04.stdout: python3-cheroot-10.0.1-4.el9.noarch 2026-03-09T15:11:44.710 INFO:teuthology.orchestra.run.vm04.stdout: python3-cherrypy-18.6.1-2.el9.noarch 2026-03-09T15:11:44.710 INFO:teuthology.orchestra.run.vm04.stdout: python3-cryptography-36.0.1-5.el9.x86_64 2026-03-09T15:11:44.710 INFO:teuthology.orchestra.run.vm04.stdout: python3-devel-3.9.25-3.el9.x86_64 2026-03-09T15:11:44.710 INFO:teuthology.orchestra.run.vm04.stdout: python3-google-auth-1:2.45.0-1.el9.noarch 2026-03-09T15:11:44.710 INFO:teuthology.orchestra.run.vm04.stdout: python3-grpcio-1.46.7-10.el9.x86_64 2026-03-09T15:11:44.710 INFO:teuthology.orchestra.run.vm04.stdout: python3-grpcio-tools-1.46.7-10.el9.x86_64 2026-03-09T15:11:44.710 INFO:teuthology.orchestra.run.vm04.stdout: python3-jaraco-8.2.1-3.el9.noarch 2026-03-09T15:11:44.710 INFO:teuthology.orchestra.run.vm04.stdout: python3-jaraco-classes-3.2.1-5.el9.noarch 2026-03-09T15:11:44.710 INFO:teuthology.orchestra.run.vm04.stdout: python3-jaraco-collections-3.0.0-8.el9.noarch 2026-03-09T15:11:44.710 INFO:teuthology.orchestra.run.vm04.stdout: python3-jaraco-context-6.0.1-3.el9.noarch 2026-03-09T15:11:44.710 INFO:teuthology.orchestra.run.vm04.stdout: python3-jaraco-functools-3.5.0-2.el9.noarch 2026-03-09T15:11:44.710 INFO:teuthology.orchestra.run.vm04.stdout: python3-jaraco-text-4.0.0-2.el9.noarch 2026-03-09T15:11:44.710 INFO:teuthology.orchestra.run.vm04.stdout: python3-jinja2-2.11.3-8.el9.noarch 2026-03-09T15:11:44.710 INFO:teuthology.orchestra.run.vm04.stdout: python3-jmespath-1.0.1-1.el9.noarch 2026-03-09T15:11:44.710 INFO:teuthology.orchestra.run.vm04.stdout: python3-kubernetes-1:26.1.0-3.el9.noarch 2026-03-09T15:11:44.710 INFO:teuthology.orchestra.run.vm04.stdout: python3-libstoragemgmt-1.10.1-1.el9.x86_64 2026-03-09T15:11:44.710 INFO:teuthology.orchestra.run.vm04.stdout: python3-logutils-0.3.5-21.el9.noarch 2026-03-09T15:11:44.710 INFO:teuthology.orchestra.run.vm04.stdout: python3-mako-1.1.4-6.el9.noarch 2026-03-09T15:11:44.710 INFO:teuthology.orchestra.run.vm04.stdout: python3-markupsafe-1.1.1-12.el9.x86_64 2026-03-09T15:11:44.710 INFO:teuthology.orchestra.run.vm04.stdout: python3-more-itertools-8.12.0-2.el9.noarch 2026-03-09T15:11:44.710 INFO:teuthology.orchestra.run.vm04.stdout: python3-natsort-7.1.1-5.el9.noarch 2026-03-09T15:11:44.710 INFO:teuthology.orchestra.run.vm04.stdout: python3-numpy-1:1.23.5-2.el9.x86_64 2026-03-09T15:11:44.710 INFO:teuthology.orchestra.run.vm04.stdout: python3-numpy-f2py-1:1.23.5-2.el9.x86_64 2026-03-09T15:11:44.710 INFO:teuthology.orchestra.run.vm04.stdout: python3-packaging-20.9-5.el9.noarch 2026-03-09T15:11:44.710 INFO:teuthology.orchestra.run.vm04.stdout: python3-pecan-1.4.2-3.el9.noarch 2026-03-09T15:11:44.710 INFO:teuthology.orchestra.run.vm04.stdout: python3-ply-3.11-14.el9.noarch 2026-03-09T15:11:44.710 INFO:teuthology.orchestra.run.vm04.stdout: python3-portend-3.1.0-2.el9.noarch 2026-03-09T15:11:44.710 INFO:teuthology.orchestra.run.vm04.stdout: python3-protobuf-3.14.0-17.el9.noarch 2026-03-09T15:11:44.710 INFO:teuthology.orchestra.run.vm04.stdout: python3-pyOpenSSL-21.0.0-1.el9.noarch 2026-03-09T15:11:44.710 INFO:teuthology.orchestra.run.vm04.stdout: python3-pyasn1-0.4.8-7.el9.noarch 2026-03-09T15:11:44.710 INFO:teuthology.orchestra.run.vm04.stdout: python3-pyasn1-modules-0.4.8-7.el9.noarch 2026-03-09T15:11:44.710 INFO:teuthology.orchestra.run.vm04.stdout: python3-pycparser-2.20-6.el9.noarch 2026-03-09T15:11:44.710 INFO:teuthology.orchestra.run.vm04.stdout: python3-rados-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T15:11:44.710 INFO:teuthology.orchestra.run.vm04.stdout: python3-rbd-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T15:11:44.710 INFO:teuthology.orchestra.run.vm04.stdout: python3-repoze-lru-0.7-16.el9.noarch 2026-03-09T15:11:44.710 INFO:teuthology.orchestra.run.vm04.stdout: python3-requests-2.25.1-10.el9.noarch 2026-03-09T15:11:44.710 INFO:teuthology.orchestra.run.vm04.stdout: python3-requests-oauthlib-1.3.0-12.el9.noarch 2026-03-09T15:11:44.710 INFO:teuthology.orchestra.run.vm04.stdout: python3-rgw-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T15:11:44.711 INFO:teuthology.orchestra.run.vm04.stdout: python3-routes-2.5.1-5.el9.noarch 2026-03-09T15:11:44.711 INFO:teuthology.orchestra.run.vm04.stdout: python3-rsa-4.9-2.el9.noarch 2026-03-09T15:11:44.711 INFO:teuthology.orchestra.run.vm04.stdout: python3-scipy-1.9.3-2.el9.x86_64 2026-03-09T15:11:44.711 INFO:teuthology.orchestra.run.vm04.stdout: python3-tempora-5.0.0-2.el9.noarch 2026-03-09T15:11:44.711 INFO:teuthology.orchestra.run.vm04.stdout: python3-toml-0.10.2-6.el9.noarch 2026-03-09T15:11:44.711 INFO:teuthology.orchestra.run.vm04.stdout: python3-typing-extensions-4.15.0-1.el9.noarch 2026-03-09T15:11:44.711 INFO:teuthology.orchestra.run.vm04.stdout: python3-urllib3-1.26.5-7.el9.noarch 2026-03-09T15:11:44.711 INFO:teuthology.orchestra.run.vm04.stdout: python3-webob-1.8.8-2.el9.noarch 2026-03-09T15:11:44.711 INFO:teuthology.orchestra.run.vm04.stdout: python3-websocket-client-1.2.3-2.el9.noarch 2026-03-09T15:11:44.711 INFO:teuthology.orchestra.run.vm04.stdout: python3-werkzeug-2.0.3-3.el9.1.noarch 2026-03-09T15:11:44.711 INFO:teuthology.orchestra.run.vm04.stdout: python3-xmltodict-0.12.0-15.el9.noarch 2026-03-09T15:11:44.711 INFO:teuthology.orchestra.run.vm04.stdout: python3-zc-lockfile-2.0-10.el9.noarch 2026-03-09T15:11:44.711 INFO:teuthology.orchestra.run.vm04.stdout: qatlib-25.08.0-2.el9.x86_64 2026-03-09T15:11:44.711 INFO:teuthology.orchestra.run.vm04.stdout: qatlib-service-25.08.0-2.el9.x86_64 2026-03-09T15:11:44.711 INFO:teuthology.orchestra.run.vm04.stdout: qatzip-libs-1.3.1-1.el9.x86_64 2026-03-09T15:11:44.711 INFO:teuthology.orchestra.run.vm04.stdout: rbd-fuse-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T15:11:44.711 INFO:teuthology.orchestra.run.vm04.stdout: rbd-mirror-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T15:11:44.711 INFO:teuthology.orchestra.run.vm04.stdout: rbd-nbd-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T15:11:44.711 INFO:teuthology.orchestra.run.vm04.stdout: re2-1:20211101-20.el9.x86_64 2026-03-09T15:11:44.711 INFO:teuthology.orchestra.run.vm04.stdout: socat-1.7.4.1-8.el9.x86_64 2026-03-09T15:11:44.711 INFO:teuthology.orchestra.run.vm04.stdout: thrift-0.15.0-4.el9.x86_64 2026-03-09T15:11:44.711 INFO:teuthology.orchestra.run.vm04.stdout: unzip-6.0-59.el9.x86_64 2026-03-09T15:11:44.711 INFO:teuthology.orchestra.run.vm04.stdout: xmlstarlet-1.6.1-20.el9.x86_64 2026-03-09T15:11:44.711 INFO:teuthology.orchestra.run.vm04.stdout: zip-3.0-35.el9.x86_64 2026-03-09T15:11:44.711 INFO:teuthology.orchestra.run.vm04.stdout: 2026-03-09T15:11:44.711 INFO:teuthology.orchestra.run.vm04.stdout:Complete! 2026-03-09T15:11:44.806 DEBUG:teuthology.parallel:result is None 2026-03-09T15:11:44.806 DEBUG:teuthology.packaging:Querying https://shaman.ceph.com/api/search?status=ready&project=ceph&flavor=default&distros=centos%2F9%2Fx86_64&sha1=e911bdebe5c8faa3800735d1568fcdca65db60df 2026-03-09T15:11:45.375 DEBUG:teuthology.orchestra.run.vm03:> rpm -q ceph --qf '%{VERSION}-%{RELEASE}' 2026-03-09T15:11:45.396 INFO:teuthology.orchestra.run.vm03.stdout:19.2.3-678.ge911bdeb.el9 2026-03-09T15:11:45.397 INFO:teuthology.packaging:The installed version of ceph is 19.2.3-678.ge911bdeb.el9 2026-03-09T15:11:45.397 INFO:teuthology.task.install:The correct ceph version 19.2.3-678.ge911bdeb is installed. 2026-03-09T15:11:45.398 DEBUG:teuthology.packaging:Querying https://shaman.ceph.com/api/search?status=ready&project=ceph&flavor=default&distros=centos%2F9%2Fx86_64&sha1=e911bdebe5c8faa3800735d1568fcdca65db60df 2026-03-09T15:11:46.064 DEBUG:teuthology.orchestra.run.vm04:> rpm -q ceph --qf '%{VERSION}-%{RELEASE}' 2026-03-09T15:11:46.084 INFO:teuthology.orchestra.run.vm04.stdout:19.2.3-678.ge911bdeb.el9 2026-03-09T15:11:46.084 INFO:teuthology.packaging:The installed version of ceph is 19.2.3-678.ge911bdeb.el9 2026-03-09T15:11:46.084 INFO:teuthology.task.install:The correct ceph version 19.2.3-678.ge911bdeb is installed. 2026-03-09T15:11:46.085 DEBUG:teuthology.packaging:Querying https://shaman.ceph.com/api/search?status=ready&project=ceph&flavor=default&distros=centos%2F9%2Fx86_64&sha1=e911bdebe5c8faa3800735d1568fcdca65db60df 2026-03-09T15:11:46.705 DEBUG:teuthology.orchestra.run.vm06:> rpm -q ceph --qf '%{VERSION}-%{RELEASE}' 2026-03-09T15:11:46.724 INFO:teuthology.orchestra.run.vm06.stdout:19.2.3-678.ge911bdeb.el9 2026-03-09T15:11:46.724 INFO:teuthology.packaging:The installed version of ceph is 19.2.3-678.ge911bdeb.el9 2026-03-09T15:11:46.724 INFO:teuthology.task.install:The correct ceph version 19.2.3-678.ge911bdeb is installed. 2026-03-09T15:11:46.725 INFO:teuthology.task.install.util:Shipping valgrind.supp... 2026-03-09T15:11:46.725 DEBUG:teuthology.orchestra.run.vm03:> set -ex 2026-03-09T15:11:46.725 DEBUG:teuthology.orchestra.run.vm03:> sudo dd of=/home/ubuntu/cephtest/valgrind.supp 2026-03-09T15:11:46.751 DEBUG:teuthology.orchestra.run.vm04:> set -ex 2026-03-09T15:11:46.751 DEBUG:teuthology.orchestra.run.vm04:> sudo dd of=/home/ubuntu/cephtest/valgrind.supp 2026-03-09T15:11:46.778 DEBUG:teuthology.orchestra.run.vm06:> set -ex 2026-03-09T15:11:46.778 DEBUG:teuthology.orchestra.run.vm06:> sudo dd of=/home/ubuntu/cephtest/valgrind.supp 2026-03-09T15:11:46.805 INFO:teuthology.task.install.util:Shipping 'daemon-helper'... 2026-03-09T15:11:46.805 DEBUG:teuthology.orchestra.run.vm03:> set -ex 2026-03-09T15:11:46.805 DEBUG:teuthology.orchestra.run.vm03:> sudo dd of=/usr/bin/daemon-helper 2026-03-09T15:11:46.831 DEBUG:teuthology.orchestra.run.vm03:> sudo chmod a=rx -- /usr/bin/daemon-helper 2026-03-09T15:11:46.897 DEBUG:teuthology.orchestra.run.vm04:> set -ex 2026-03-09T15:11:46.897 DEBUG:teuthology.orchestra.run.vm04:> sudo dd of=/usr/bin/daemon-helper 2026-03-09T15:11:46.922 DEBUG:teuthology.orchestra.run.vm04:> sudo chmod a=rx -- /usr/bin/daemon-helper 2026-03-09T15:11:46.988 DEBUG:teuthology.orchestra.run.vm06:> set -ex 2026-03-09T15:11:46.988 DEBUG:teuthology.orchestra.run.vm06:> sudo dd of=/usr/bin/daemon-helper 2026-03-09T15:11:47.012 DEBUG:teuthology.orchestra.run.vm06:> sudo chmod a=rx -- /usr/bin/daemon-helper 2026-03-09T15:11:47.076 INFO:teuthology.task.install.util:Shipping 'adjust-ulimits'... 2026-03-09T15:11:47.076 DEBUG:teuthology.orchestra.run.vm03:> set -ex 2026-03-09T15:11:47.076 DEBUG:teuthology.orchestra.run.vm03:> sudo dd of=/usr/bin/adjust-ulimits 2026-03-09T15:11:47.100 DEBUG:teuthology.orchestra.run.vm03:> sudo chmod a=rx -- /usr/bin/adjust-ulimits 2026-03-09T15:11:47.164 DEBUG:teuthology.orchestra.run.vm04:> set -ex 2026-03-09T15:11:47.164 DEBUG:teuthology.orchestra.run.vm04:> sudo dd of=/usr/bin/adjust-ulimits 2026-03-09T15:11:47.189 DEBUG:teuthology.orchestra.run.vm04:> sudo chmod a=rx -- /usr/bin/adjust-ulimits 2026-03-09T15:11:47.256 DEBUG:teuthology.orchestra.run.vm06:> set -ex 2026-03-09T15:11:47.257 DEBUG:teuthology.orchestra.run.vm06:> sudo dd of=/usr/bin/adjust-ulimits 2026-03-09T15:11:47.281 DEBUG:teuthology.orchestra.run.vm06:> sudo chmod a=rx -- /usr/bin/adjust-ulimits 2026-03-09T15:11:47.345 INFO:teuthology.task.install.util:Shipping 'stdin-killer'... 2026-03-09T15:11:47.346 DEBUG:teuthology.orchestra.run.vm03:> set -ex 2026-03-09T15:11:47.346 DEBUG:teuthology.orchestra.run.vm03:> sudo dd of=/usr/bin/stdin-killer 2026-03-09T15:11:47.368 DEBUG:teuthology.orchestra.run.vm03:> sudo chmod a=rx -- /usr/bin/stdin-killer 2026-03-09T15:11:47.431 DEBUG:teuthology.orchestra.run.vm04:> set -ex 2026-03-09T15:11:47.431 DEBUG:teuthology.orchestra.run.vm04:> sudo dd of=/usr/bin/stdin-killer 2026-03-09T15:11:47.456 DEBUG:teuthology.orchestra.run.vm04:> sudo chmod a=rx -- /usr/bin/stdin-killer 2026-03-09T15:11:47.521 DEBUG:teuthology.orchestra.run.vm06:> set -ex 2026-03-09T15:11:47.521 DEBUG:teuthology.orchestra.run.vm06:> sudo dd of=/usr/bin/stdin-killer 2026-03-09T15:11:47.546 DEBUG:teuthology.orchestra.run.vm06:> sudo chmod a=rx -- /usr/bin/stdin-killer 2026-03-09T15:11:47.611 INFO:teuthology.run_tasks:Running task cephadm... 2026-03-09T15:11:47.656 INFO:tasks.cephadm:Config: {'conf': {'global': {'mon election default strategy': 1}, 'mgr': {'debug mgr': 20, 'debug ms': 1, 'mgr/cephadm/use_agent': False}, 'mon': {'debug mon': 20, 'debug ms': 1, 'debug paxos': 20}, 'osd': {'debug ms': 1, 'debug osd': 20, 'osd mclock iops capacity threshold hdd': 49000}}, 'flavor': 'default', 'log-ignorelist': ['\\(MDS_ALL_DOWN\\)', '\\(MDS_UP_LESS_THAN_MAX\\)', 'MON_DOWN', 'POOL_APP_NOT_ENABLED', 'mon down', 'mons down', 'out of quorum', 'CEPHADM_FAILED_DAEMON'], 'log-only-match': ['CEPHADM_'], 'sha1': 'e911bdebe5c8faa3800735d1568fcdca65db60df'} 2026-03-09T15:11:47.656 INFO:tasks.cephadm:Cluster image is quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df 2026-03-09T15:11:47.656 INFO:tasks.cephadm:Cluster fsid is 4abede80-1bca-11f1-bcd8-9b1da3c03ce5 2026-03-09T15:11:47.656 INFO:tasks.cephadm:Choosing monitor IPs and ports... 2026-03-09T15:11:47.656 INFO:tasks.cephadm:Monitor IPs: {'mon.a': '192.168.123.103', 'mon.b': '192.168.123.104', 'mon.c': '192.168.123.106'} 2026-03-09T15:11:47.656 INFO:tasks.cephadm:First mon is mon.a on vm03 2026-03-09T15:11:47.656 INFO:tasks.cephadm:First mgr is a 2026-03-09T15:11:47.656 INFO:tasks.cephadm:Normalizing hostnames... 2026-03-09T15:11:47.656 DEBUG:teuthology.orchestra.run.vm03:> sudo hostname $(hostname -s) 2026-03-09T15:11:47.680 DEBUG:teuthology.orchestra.run.vm04:> sudo hostname $(hostname -s) 2026-03-09T15:11:47.704 DEBUG:teuthology.orchestra.run.vm06:> sudo hostname $(hostname -s) 2026-03-09T15:11:47.728 INFO:tasks.cephadm:Downloading "compiled" cephadm from cachra 2026-03-09T15:11:47.728 DEBUG:teuthology.packaging:Querying https://shaman.ceph.com/api/search?status=ready&project=ceph&flavor=default&distros=centos%2F9%2Fx86_64&sha1=e911bdebe5c8faa3800735d1568fcdca65db60df 2026-03-09T15:11:48.354 INFO:tasks.cephadm:builder_project result: [{'url': 'https://3.chacra.ceph.com/r/ceph/squid/e911bdebe5c8faa3800735d1568fcdca65db60df/centos/9/flavors/default/', 'chacra_url': 'https://3.chacra.ceph.com/repos/ceph/squid/e911bdebe5c8faa3800735d1568fcdca65db60df/centos/9/flavors/default/', 'ref': 'squid', 'sha1': 'e911bdebe5c8faa3800735d1568fcdca65db60df', 'distro': 'centos', 'distro_version': '9', 'distro_codename': None, 'modified': '2026-02-25 18:55:15.146628', 'status': 'ready', 'flavor': 'default', 'project': 'ceph', 'archs': ['source', 'x86_64'], 'extra': {'version': '19.2.3-678-ge911bdeb', 'package_manager_version': '19.2.3-678.ge911bdeb', 'build_url': 'https://jenkins.ceph.com/job/ceph-dev-pipeline/3275/', 'root_build_cause': '', 'node_name': '10.20.192.26+soko16', 'job_name': 'ceph-dev-pipeline'}}] 2026-03-09T15:11:48.938 INFO:tasks.util.chacra:got chacra host 3.chacra.ceph.com, ref squid, sha1 e911bdebe5c8faa3800735d1568fcdca65db60df from https://shaman.ceph.com/api/search/?project=ceph&distros=centos%2F9%2Fx86_64&flavor=default&sha1=e911bdebe5c8faa3800735d1568fcdca65db60df 2026-03-09T15:11:48.939 INFO:tasks.cephadm:Discovered cachra url: https://3.chacra.ceph.com/binaries/ceph/squid/e911bdebe5c8faa3800735d1568fcdca65db60df/centos/9/x86_64/flavors/default/cephadm 2026-03-09T15:11:48.939 INFO:tasks.cephadm:Downloading cephadm from url: https://3.chacra.ceph.com/binaries/ceph/squid/e911bdebe5c8faa3800735d1568fcdca65db60df/centos/9/x86_64/flavors/default/cephadm 2026-03-09T15:11:48.939 DEBUG:teuthology.orchestra.run.vm03:> curl --silent -L https://3.chacra.ceph.com/binaries/ceph/squid/e911bdebe5c8faa3800735d1568fcdca65db60df/centos/9/x86_64/flavors/default/cephadm > /home/ubuntu/cephtest/cephadm && ls -l /home/ubuntu/cephtest/cephadm 2026-03-09T15:11:50.317 INFO:teuthology.orchestra.run.vm03.stdout:-rw-r--r--. 1 ubuntu ubuntu 788355 Mar 9 15:11 /home/ubuntu/cephtest/cephadm 2026-03-09T15:11:50.317 DEBUG:teuthology.orchestra.run.vm04:> curl --silent -L https://3.chacra.ceph.com/binaries/ceph/squid/e911bdebe5c8faa3800735d1568fcdca65db60df/centos/9/x86_64/flavors/default/cephadm > /home/ubuntu/cephtest/cephadm && ls -l /home/ubuntu/cephtest/cephadm 2026-03-09T15:11:51.712 INFO:teuthology.orchestra.run.vm04.stdout:-rw-r--r--. 1 ubuntu ubuntu 788355 Mar 9 15:11 /home/ubuntu/cephtest/cephadm 2026-03-09T15:11:51.713 DEBUG:teuthology.orchestra.run.vm06:> curl --silent -L https://3.chacra.ceph.com/binaries/ceph/squid/e911bdebe5c8faa3800735d1568fcdca65db60df/centos/9/x86_64/flavors/default/cephadm > /home/ubuntu/cephtest/cephadm && ls -l /home/ubuntu/cephtest/cephadm 2026-03-09T15:11:53.147 INFO:teuthology.orchestra.run.vm06.stdout:-rw-r--r--. 1 ubuntu ubuntu 788355 Mar 9 15:11 /home/ubuntu/cephtest/cephadm 2026-03-09T15:11:53.147 DEBUG:teuthology.orchestra.run.vm03:> test -s /home/ubuntu/cephtest/cephadm && test $(stat -c%s /home/ubuntu/cephtest/cephadm) -gt 1000 && chmod +x /home/ubuntu/cephtest/cephadm 2026-03-09T15:11:53.162 DEBUG:teuthology.orchestra.run.vm04:> test -s /home/ubuntu/cephtest/cephadm && test $(stat -c%s /home/ubuntu/cephtest/cephadm) -gt 1000 && chmod +x /home/ubuntu/cephtest/cephadm 2026-03-09T15:11:53.176 DEBUG:teuthology.orchestra.run.vm06:> test -s /home/ubuntu/cephtest/cephadm && test $(stat -c%s /home/ubuntu/cephtest/cephadm) -gt 1000 && chmod +x /home/ubuntu/cephtest/cephadm 2026-03-09T15:11:53.196 INFO:tasks.cephadm:Pulling image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df on all hosts... 2026-03-09T15:11:53.196 DEBUG:teuthology.orchestra.run.vm03:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df pull 2026-03-09T15:11:53.203 DEBUG:teuthology.orchestra.run.vm04:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df pull 2026-03-09T15:11:53.218 DEBUG:teuthology.orchestra.run.vm06:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df pull 2026-03-09T15:11:53.363 INFO:teuthology.orchestra.run.vm03.stderr:Pulling container image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df... 2026-03-09T15:11:53.375 INFO:teuthology.orchestra.run.vm04.stderr:Pulling container image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df... 2026-03-09T15:11:53.400 INFO:teuthology.orchestra.run.vm06.stderr:Pulling container image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df... 2026-03-09T15:12:37.384 INFO:teuthology.orchestra.run.vm06.stdout:{ 2026-03-09T15:12:37.384 INFO:teuthology.orchestra.run.vm06.stdout: "ceph_version": "ceph version 19.2.3-678-ge911bdeb (e911bdebe5c8faa3800735d1568fcdca65db60df) squid (stable)", 2026-03-09T15:12:37.384 INFO:teuthology.orchestra.run.vm06.stdout: "image_id": "654f31e6858eb235bbece362255b685a945f2b6a367e2b88c4930c984fbb214c", 2026-03-09T15:12:37.384 INFO:teuthology.orchestra.run.vm06.stdout: "repo_digests": [ 2026-03-09T15:12:37.384 INFO:teuthology.orchestra.run.vm06.stdout: "quay.ceph.io/ceph-ci/ceph@sha256:8fda260ab1d2d3118a1622f7df75f44f285dfe74e71793626152a711c12bf2cc" 2026-03-09T15:12:37.384 INFO:teuthology.orchestra.run.vm06.stdout: ] 2026-03-09T15:12:37.384 INFO:teuthology.orchestra.run.vm06.stdout:} 2026-03-09T15:12:37.392 INFO:teuthology.orchestra.run.vm04.stdout:{ 2026-03-09T15:12:37.392 INFO:teuthology.orchestra.run.vm04.stdout: "ceph_version": "ceph version 19.2.3-678-ge911bdeb (e911bdebe5c8faa3800735d1568fcdca65db60df) squid (stable)", 2026-03-09T15:12:37.392 INFO:teuthology.orchestra.run.vm04.stdout: "image_id": "654f31e6858eb235bbece362255b685a945f2b6a367e2b88c4930c984fbb214c", 2026-03-09T15:12:37.392 INFO:teuthology.orchestra.run.vm04.stdout: "repo_digests": [ 2026-03-09T15:12:37.392 INFO:teuthology.orchestra.run.vm04.stdout: "quay.ceph.io/ceph-ci/ceph@sha256:8fda260ab1d2d3118a1622f7df75f44f285dfe74e71793626152a711c12bf2cc" 2026-03-09T15:12:37.392 INFO:teuthology.orchestra.run.vm04.stdout: ] 2026-03-09T15:12:37.392 INFO:teuthology.orchestra.run.vm04.stdout:} 2026-03-09T15:12:38.689 INFO:teuthology.orchestra.run.vm03.stdout:{ 2026-03-09T15:12:38.689 INFO:teuthology.orchestra.run.vm03.stdout: "ceph_version": "ceph version 19.2.3-678-ge911bdeb (e911bdebe5c8faa3800735d1568fcdca65db60df) squid (stable)", 2026-03-09T15:12:38.689 INFO:teuthology.orchestra.run.vm03.stdout: "image_id": "654f31e6858eb235bbece362255b685a945f2b6a367e2b88c4930c984fbb214c", 2026-03-09T15:12:38.689 INFO:teuthology.orchestra.run.vm03.stdout: "repo_digests": [ 2026-03-09T15:12:38.689 INFO:teuthology.orchestra.run.vm03.stdout: "quay.ceph.io/ceph-ci/ceph@sha256:8fda260ab1d2d3118a1622f7df75f44f285dfe74e71793626152a711c12bf2cc" 2026-03-09T15:12:38.689 INFO:teuthology.orchestra.run.vm03.stdout: ] 2026-03-09T15:12:38.689 INFO:teuthology.orchestra.run.vm03.stdout:} 2026-03-09T15:12:38.705 DEBUG:teuthology.orchestra.run.vm03:> sudo mkdir -p /etc/ceph 2026-03-09T15:12:38.729 DEBUG:teuthology.orchestra.run.vm04:> sudo mkdir -p /etc/ceph 2026-03-09T15:12:38.754 DEBUG:teuthology.orchestra.run.vm06:> sudo mkdir -p /etc/ceph 2026-03-09T15:12:38.781 DEBUG:teuthology.orchestra.run.vm03:> sudo chmod 777 /etc/ceph 2026-03-09T15:12:38.803 DEBUG:teuthology.orchestra.run.vm04:> sudo chmod 777 /etc/ceph 2026-03-09T15:12:38.826 DEBUG:teuthology.orchestra.run.vm06:> sudo chmod 777 /etc/ceph 2026-03-09T15:12:38.851 INFO:tasks.cephadm:Writing seed config... 2026-03-09T15:12:38.851 INFO:tasks.cephadm: override: [global] mon election default strategy = 1 2026-03-09T15:12:38.851 INFO:tasks.cephadm: override: [mgr] debug mgr = 20 2026-03-09T15:12:38.852 INFO:tasks.cephadm: override: [mgr] debug ms = 1 2026-03-09T15:12:38.852 INFO:tasks.cephadm: override: [mgr] mgr/cephadm/use_agent = False 2026-03-09T15:12:38.852 INFO:tasks.cephadm: override: [mon] debug mon = 20 2026-03-09T15:12:38.852 INFO:tasks.cephadm: override: [mon] debug ms = 1 2026-03-09T15:12:38.852 INFO:tasks.cephadm: override: [mon] debug paxos = 20 2026-03-09T15:12:38.852 INFO:tasks.cephadm: override: [osd] debug ms = 1 2026-03-09T15:12:38.852 INFO:tasks.cephadm: override: [osd] debug osd = 20 2026-03-09T15:12:38.852 INFO:tasks.cephadm: override: [osd] osd mclock iops capacity threshold hdd = 49000 2026-03-09T15:12:38.852 DEBUG:teuthology.orchestra.run.vm03:> set -ex 2026-03-09T15:12:38.852 DEBUG:teuthology.orchestra.run.vm03:> dd of=/home/ubuntu/cephtest/seed.ceph.conf 2026-03-09T15:12:38.866 DEBUG:tasks.cephadm:Final config: [global] # make logging friendly to teuthology log_to_file = true log_to_stderr = false log to journald = false mon cluster log to file = true mon cluster log file level = debug mon clock drift allowed = 1.000 # replicate across OSDs, not hosts osd crush chooseleaf type = 0 #osd pool default size = 2 osd pool default erasure code profile = plugin=jerasure technique=reed_sol_van k=2 m=1 crush-failure-domain=osd # enable some debugging auth debug = true ms die on old message = true ms die on bug = true debug asserts on shutdown = true # adjust warnings mon max pg per osd = 10000# >= luminous mon pg warn max object skew = 0 mon osd allow primary affinity = true mon osd allow pg remap = true mon warn on legacy crush tunables = false mon warn on crush straw calc version zero = false mon warn on no sortbitwise = false mon warn on osd down out interval zero = false mon warn on too few osds = false mon_warn_on_pool_pg_num_not_power_of_two = false # disable pg_autoscaler by default for new pools osd_pool_default_pg_autoscale_mode = off # tests delete pools mon allow pool delete = true fsid = 4abede80-1bca-11f1-bcd8-9b1da3c03ce5 mon election default strategy = 1 [osd] osd scrub load threshold = 5.0 osd scrub max interval = 600 osd mclock profile = high_recovery_ops osd recover clone overlap = true osd recovery max chunk = 1048576 osd deep scrub update digest min age = 30 osd map max advance = 10 osd memory target autotune = true # debugging osd debug shutdown = true osd debug op order = true osd debug verify stray on activate = true osd debug pg log writeout = true osd debug verify cached snaps = true osd debug verify missing on start = true osd debug misdirected ops = true osd op queue = debug_random osd op queue cut off = debug_random osd shutdown pgref assert = true bdev debug aio = true osd sloppy crc = true debug ms = 1 debug osd = 20 osd mclock iops capacity threshold hdd = 49000 [mgr] mon reweight min pgs per osd = 4 mon reweight min bytes per osd = 10 mgr/telemetry/nag = false debug mgr = 20 debug ms = 1 mgr/cephadm/use_agent = False [mon] mon data avail warn = 5 mon mgr mkfs grace = 240 mon reweight min pgs per osd = 4 mon osd reporter subtree level = osd mon osd prime pg temp = true mon reweight min bytes per osd = 10 # rotate auth tickets quickly to exercise renewal paths auth mon ticket ttl = 660# 11m auth service ticket ttl = 240# 4m # don't complain about global id reclaim mon_warn_on_insecure_global_id_reclaim = false mon_warn_on_insecure_global_id_reclaim_allowed = false debug mon = 20 debug ms = 1 debug paxos = 20 [client.rgw] rgw cache enabled = true rgw enable ops log = true rgw enable usage log = true 2026-03-09T15:12:38.866 DEBUG:teuthology.orchestra.run.vm03:mon.a> sudo journalctl -f -n 0 -u ceph-4abede80-1bca-11f1-bcd8-9b1da3c03ce5@mon.a.service 2026-03-09T15:12:38.907 DEBUG:teuthology.orchestra.run.vm03:mgr.a> sudo journalctl -f -n 0 -u ceph-4abede80-1bca-11f1-bcd8-9b1da3c03ce5@mgr.a.service 2026-03-09T15:12:38.949 INFO:tasks.cephadm:Bootstrapping... 2026-03-09T15:12:38.949 DEBUG:teuthology.orchestra.run.vm03:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df -v bootstrap --fsid 4abede80-1bca-11f1-bcd8-9b1da3c03ce5 --config /home/ubuntu/cephtest/seed.ceph.conf --output-config /etc/ceph/ceph.conf --output-keyring /etc/ceph/ceph.client.admin.keyring --output-pub-ssh-key /home/ubuntu/cephtest/ceph.pub --mon-id a --mgr-id a --orphan-initial-daemons --skip-monitoring-stack --mon-ip 192.168.123.103 --skip-admin-label && sudo chmod +r /etc/ceph/ceph.client.admin.keyring 2026-03-09T15:12:39.083 INFO:teuthology.orchestra.run.vm03.stdout:-------------------------------------------------------------------------------- 2026-03-09T15:12:39.083 INFO:teuthology.orchestra.run.vm03.stdout:cephadm ['--image', 'quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df', '-v', 'bootstrap', '--fsid', '4abede80-1bca-11f1-bcd8-9b1da3c03ce5', '--config', '/home/ubuntu/cephtest/seed.ceph.conf', '--output-config', '/etc/ceph/ceph.conf', '--output-keyring', '/etc/ceph/ceph.client.admin.keyring', '--output-pub-ssh-key', '/home/ubuntu/cephtest/ceph.pub', '--mon-id', 'a', '--mgr-id', 'a', '--orphan-initial-daemons', '--skip-monitoring-stack', '--mon-ip', '192.168.123.103', '--skip-admin-label'] 2026-03-09T15:12:39.083 INFO:teuthology.orchestra.run.vm03.stderr:Specifying an fsid for your cluster offers no advantages and may increase the likelihood of fsid conflicts. 2026-03-09T15:12:39.083 INFO:teuthology.orchestra.run.vm03.stdout:Verifying podman|docker is present... 2026-03-09T15:12:39.101 INFO:teuthology.orchestra.run.vm03.stdout:/bin/podman: stdout 5.8.0 2026-03-09T15:12:39.102 INFO:teuthology.orchestra.run.vm03.stdout:Verifying lvm2 is present... 2026-03-09T15:12:39.102 INFO:teuthology.orchestra.run.vm03.stdout:Verifying time synchronization is in place... 2026-03-09T15:12:39.108 INFO:teuthology.orchestra.run.vm03.stdout:Non-zero exit code 1 from systemctl is-enabled chrony.service 2026-03-09T15:12:39.108 INFO:teuthology.orchestra.run.vm03.stdout:systemctl: stderr Failed to get unit file state for chrony.service: No such file or directory 2026-03-09T15:12:39.112 INFO:teuthology.orchestra.run.vm03.stdout:Non-zero exit code 3 from systemctl is-active chrony.service 2026-03-09T15:12:39.112 INFO:teuthology.orchestra.run.vm03.stdout:systemctl: stdout inactive 2026-03-09T15:12:39.117 INFO:teuthology.orchestra.run.vm03.stdout:systemctl: stdout enabled 2026-03-09T15:12:39.121 INFO:teuthology.orchestra.run.vm03.stdout:systemctl: stdout active 2026-03-09T15:12:39.121 INFO:teuthology.orchestra.run.vm03.stdout:Unit chronyd.service is enabled and running 2026-03-09T15:12:39.121 INFO:teuthology.orchestra.run.vm03.stdout:Repeating the final host check... 2026-03-09T15:12:39.137 INFO:teuthology.orchestra.run.vm03.stdout:/bin/podman: stdout 5.8.0 2026-03-09T15:12:39.137 INFO:teuthology.orchestra.run.vm03.stdout:podman (/bin/podman) version 5.8.0 is present 2026-03-09T15:12:39.137 INFO:teuthology.orchestra.run.vm03.stdout:systemctl is present 2026-03-09T15:12:39.137 INFO:teuthology.orchestra.run.vm03.stdout:lvcreate is present 2026-03-09T15:12:39.142 INFO:teuthology.orchestra.run.vm03.stdout:Non-zero exit code 1 from systemctl is-enabled chrony.service 2026-03-09T15:12:39.142 INFO:teuthology.orchestra.run.vm03.stdout:systemctl: stderr Failed to get unit file state for chrony.service: No such file or directory 2026-03-09T15:12:39.147 INFO:teuthology.orchestra.run.vm03.stdout:Non-zero exit code 3 from systemctl is-active chrony.service 2026-03-09T15:12:39.147 INFO:teuthology.orchestra.run.vm03.stdout:systemctl: stdout inactive 2026-03-09T15:12:39.152 INFO:teuthology.orchestra.run.vm03.stdout:systemctl: stdout enabled 2026-03-09T15:12:39.156 INFO:teuthology.orchestra.run.vm03.stdout:systemctl: stdout active 2026-03-09T15:12:39.156 INFO:teuthology.orchestra.run.vm03.stdout:Unit chronyd.service is enabled and running 2026-03-09T15:12:39.156 INFO:teuthology.orchestra.run.vm03.stdout:Host looks OK 2026-03-09T15:12:39.156 INFO:teuthology.orchestra.run.vm03.stdout:Cluster fsid: 4abede80-1bca-11f1-bcd8-9b1da3c03ce5 2026-03-09T15:12:39.156 INFO:teuthology.orchestra.run.vm03.stdout:Acquiring lock 139943286901488 on /run/cephadm/4abede80-1bca-11f1-bcd8-9b1da3c03ce5.lock 2026-03-09T15:12:39.156 INFO:teuthology.orchestra.run.vm03.stdout:Lock 139943286901488 acquired on /run/cephadm/4abede80-1bca-11f1-bcd8-9b1da3c03ce5.lock 2026-03-09T15:12:39.156 INFO:teuthology.orchestra.run.vm03.stdout:Verifying IP 192.168.123.103 port 3300 ... 2026-03-09T15:12:39.157 INFO:teuthology.orchestra.run.vm03.stdout:Verifying IP 192.168.123.103 port 6789 ... 2026-03-09T15:12:39.157 INFO:teuthology.orchestra.run.vm03.stdout:Base mon IP(s) is [192.168.123.103:3300, 192.168.123.103:6789], mon addrv is [v2:192.168.123.103:3300,v1:192.168.123.103:6789] 2026-03-09T15:12:39.159 INFO:teuthology.orchestra.run.vm03.stdout:/sbin/ip: stdout default via 192.168.123.1 dev eth0 proto dhcp src 192.168.123.103 metric 100 2026-03-09T15:12:39.159 INFO:teuthology.orchestra.run.vm03.stdout:/sbin/ip: stdout 192.168.123.0/24 dev eth0 proto kernel scope link src 192.168.123.103 metric 100 2026-03-09T15:12:39.161 INFO:teuthology.orchestra.run.vm03.stdout:/sbin/ip: stdout ::1 dev lo proto kernel metric 256 pref medium 2026-03-09T15:12:39.161 INFO:teuthology.orchestra.run.vm03.stdout:/sbin/ip: stdout fe80::/64 dev eth0 proto kernel metric 1024 pref medium 2026-03-09T15:12:39.163 INFO:teuthology.orchestra.run.vm03.stdout:/sbin/ip: stdout 1: lo: mtu 65536 state UNKNOWN qlen 1000 2026-03-09T15:12:39.163 INFO:teuthology.orchestra.run.vm03.stdout:/sbin/ip: stdout inet6 ::1/128 scope host 2026-03-09T15:12:39.163 INFO:teuthology.orchestra.run.vm03.stdout:/sbin/ip: stdout valid_lft forever preferred_lft forever 2026-03-09T15:12:39.163 INFO:teuthology.orchestra.run.vm03.stdout:/sbin/ip: stdout 2: eth0: mtu 1500 state UP qlen 1000 2026-03-09T15:12:39.163 INFO:teuthology.orchestra.run.vm03.stdout:/sbin/ip: stdout inet6 fe80::5055:ff:fe00:3/64 scope link noprefixroute 2026-03-09T15:12:39.163 INFO:teuthology.orchestra.run.vm03.stdout:/sbin/ip: stdout valid_lft forever preferred_lft forever 2026-03-09T15:12:39.163 INFO:teuthology.orchestra.run.vm03.stdout:Mon IP `192.168.123.103` is in CIDR network `192.168.123.0/24` 2026-03-09T15:12:39.163 INFO:teuthology.orchestra.run.vm03.stdout:Mon IP `192.168.123.103` is in CIDR network `192.168.123.0/24` 2026-03-09T15:12:39.164 INFO:teuthology.orchestra.run.vm03.stdout:Inferred mon public CIDR from local network configuration ['192.168.123.0/24', '192.168.123.0/24'] 2026-03-09T15:12:39.164 INFO:teuthology.orchestra.run.vm03.stdout:Internal network (--cluster-network) has not been provided, OSD replication will default to the public_network 2026-03-09T15:12:39.164 INFO:teuthology.orchestra.run.vm03.stdout:Pulling container image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df... 2026-03-09T15:12:40.373 INFO:teuthology.orchestra.run.vm03.stdout:/bin/podman: stdout 654f31e6858eb235bbece362255b685a945f2b6a367e2b88c4930c984fbb214c 2026-03-09T15:12:40.373 INFO:teuthology.orchestra.run.vm03.stdout:/bin/podman: stderr Trying to pull quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df... 2026-03-09T15:12:40.373 INFO:teuthology.orchestra.run.vm03.stdout:/bin/podman: stderr Getting image source signatures 2026-03-09T15:12:40.373 INFO:teuthology.orchestra.run.vm03.stdout:/bin/podman: stderr Copying blob sha256:1752b8d01aa0dd33bbe0ab24e8316174c94fbdcd5d26252e2680bba0624747a7 2026-03-09T15:12:40.373 INFO:teuthology.orchestra.run.vm03.stdout:/bin/podman: stderr Copying blob sha256:8e380faede39ebd4286247457b408d979ab568aafd8389c42ec304b8cfba4e92 2026-03-09T15:12:40.373 INFO:teuthology.orchestra.run.vm03.stdout:/bin/podman: stderr Copying config sha256:654f31e6858eb235bbece362255b685a945f2b6a367e2b88c4930c984fbb214c 2026-03-09T15:12:40.373 INFO:teuthology.orchestra.run.vm03.stdout:/bin/podman: stderr Writing manifest to image destination 2026-03-09T15:12:40.505 INFO:teuthology.orchestra.run.vm03.stdout:ceph: stdout ceph version 19.2.3-678-ge911bdeb (e911bdebe5c8faa3800735d1568fcdca65db60df) squid (stable) 2026-03-09T15:12:40.505 INFO:teuthology.orchestra.run.vm03.stdout:Ceph version: ceph version 19.2.3-678-ge911bdeb (e911bdebe5c8faa3800735d1568fcdca65db60df) squid (stable) 2026-03-09T15:12:40.505 INFO:teuthology.orchestra.run.vm03.stdout:Extracting ceph user uid/gid from container image... 2026-03-09T15:12:40.582 INFO:teuthology.orchestra.run.vm03.stdout:stat: stdout 167 167 2026-03-09T15:12:40.582 INFO:teuthology.orchestra.run.vm03.stdout:Creating initial keys... 2026-03-09T15:12:40.664 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph-authtool: stdout AQBo465prmKwJhAAK8adNdBRRgVqheDsu3X9vg== 2026-03-09T15:12:40.787 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph-authtool: stdout AQBo465prwjKLBAAECzyuCGbZEHcdrIE+cPrVg== 2026-03-09T15:12:40.862 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph-authtool: stdout AQBo465pGXhCMhAAsGTyIf5/0eytbXLt3AGNQw== 2026-03-09T15:12:40.864 INFO:teuthology.orchestra.run.vm03.stdout:Creating initial monmap... 2026-03-09T15:12:40.985 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/monmaptool: stdout /usr/bin/monmaptool: monmap file /tmp/monmap 2026-03-09T15:12:40.985 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/monmaptool: stdout setting min_mon_release = quincy 2026-03-09T15:12:40.985 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/monmaptool: stdout /usr/bin/monmaptool: set fsid to 4abede80-1bca-11f1-bcd8-9b1da3c03ce5 2026-03-09T15:12:40.985 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/monmaptool: stdout /usr/bin/monmaptool: writing epoch 0 to /tmp/monmap (1 monitors) 2026-03-09T15:12:40.985 INFO:teuthology.orchestra.run.vm03.stdout:monmaptool for a [v2:192.168.123.103:3300,v1:192.168.123.103:6789] on /usr/bin/monmaptool: monmap file /tmp/monmap 2026-03-09T15:12:40.985 INFO:teuthology.orchestra.run.vm03.stdout:setting min_mon_release = quincy 2026-03-09T15:12:40.985 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/monmaptool: set fsid to 4abede80-1bca-11f1-bcd8-9b1da3c03ce5 2026-03-09T15:12:40.985 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/monmaptool: writing epoch 0 to /tmp/monmap (1 monitors) 2026-03-09T15:12:40.985 INFO:teuthology.orchestra.run.vm03.stdout: 2026-03-09T15:12:40.985 INFO:teuthology.orchestra.run.vm03.stdout:Creating mon... 2026-03-09T15:12:41.393 INFO:teuthology.orchestra.run.vm03.stdout:create mon.a on 2026-03-09T15:12:41.539 INFO:teuthology.orchestra.run.vm03.stdout:systemctl: stderr Removed "/etc/systemd/system/multi-user.target.wants/ceph.target". 2026-03-09T15:12:41.655 INFO:teuthology.orchestra.run.vm03.stdout:systemctl: stderr Created symlink /etc/systemd/system/multi-user.target.wants/ceph.target → /etc/systemd/system/ceph.target. 2026-03-09T15:12:41.777 INFO:teuthology.orchestra.run.vm03.stdout:systemctl: stderr Created symlink /etc/systemd/system/multi-user.target.wants/ceph-4abede80-1bca-11f1-bcd8-9b1da3c03ce5.target → /etc/systemd/system/ceph-4abede80-1bca-11f1-bcd8-9b1da3c03ce5.target. 2026-03-09T15:12:41.777 INFO:teuthology.orchestra.run.vm03.stdout:systemctl: stderr Created symlink /etc/systemd/system/ceph.target.wants/ceph-4abede80-1bca-11f1-bcd8-9b1da3c03ce5.target → /etc/systemd/system/ceph-4abede80-1bca-11f1-bcd8-9b1da3c03ce5.target. 2026-03-09T15:12:41.916 INFO:teuthology.orchestra.run.vm03.stdout:Non-zero exit code 1 from systemctl reset-failed ceph-4abede80-1bca-11f1-bcd8-9b1da3c03ce5@mon.a 2026-03-09T15:12:41.916 INFO:teuthology.orchestra.run.vm03.stdout:systemctl: stderr Failed to reset failed state of unit ceph-4abede80-1bca-11f1-bcd8-9b1da3c03ce5@mon.a.service: Unit ceph-4abede80-1bca-11f1-bcd8-9b1da3c03ce5@mon.a.service not loaded. 2026-03-09T15:12:42.047 INFO:teuthology.orchestra.run.vm03.stdout:systemctl: stderr Created symlink /etc/systemd/system/ceph-4abede80-1bca-11f1-bcd8-9b1da3c03ce5.target.wants/ceph-4abede80-1bca-11f1-bcd8-9b1da3c03ce5@mon.a.service → /etc/systemd/system/ceph-4abede80-1bca-11f1-bcd8-9b1da3c03ce5@.service. 2026-03-09T15:12:42.207 INFO:teuthology.orchestra.run.vm03.stdout:firewalld does not appear to be present 2026-03-09T15:12:42.207 INFO:teuthology.orchestra.run.vm03.stdout:Not possible to enable service . firewalld.service is not available 2026-03-09T15:12:42.207 INFO:teuthology.orchestra.run.vm03.stdout:Waiting for mon to start... 2026-03-09T15:12:42.207 INFO:teuthology.orchestra.run.vm03.stdout:Waiting for mon... 2026-03-09T15:12:42.415 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout cluster: 2026-03-09T15:12:42.415 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout id: 4abede80-1bca-11f1-bcd8-9b1da3c03ce5 2026-03-09T15:12:42.415 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout health: HEALTH_OK 2026-03-09T15:12:42.415 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout 2026-03-09T15:12:42.416 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout services: 2026-03-09T15:12:42.416 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout mon: 1 daemons, quorum a (age 0.140939s) 2026-03-09T15:12:42.416 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout mgr: no daemons active 2026-03-09T15:12:42.416 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout osd: 0 osds: 0 up, 0 in 2026-03-09T15:12:42.416 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout 2026-03-09T15:12:42.416 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout data: 2026-03-09T15:12:42.416 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout pools: 0 pools, 0 pgs 2026-03-09T15:12:42.416 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout objects: 0 objects, 0 B 2026-03-09T15:12:42.416 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout usage: 0 B used, 0 B / 0 B avail 2026-03-09T15:12:42.416 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout pgs: 2026-03-09T15:12:42.416 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout 2026-03-09T15:12:42.416 INFO:teuthology.orchestra.run.vm03.stdout:mon is available 2026-03-09T15:12:42.416 INFO:teuthology.orchestra.run.vm03.stdout:Assimilating anything we can from ceph.conf... 2026-03-09T15:12:42.598 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout 2026-03-09T15:12:42.598 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout [global] 2026-03-09T15:12:42.598 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout fsid = 4abede80-1bca-11f1-bcd8-9b1da3c03ce5 2026-03-09T15:12:42.598 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout mon_cluster_log_file_level = debug 2026-03-09T15:12:42.598 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout mon_host = [v2:192.168.123.103:3300,v1:192.168.123.103:6789] 2026-03-09T15:12:42.598 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout mon_osd_allow_pg_remap = true 2026-03-09T15:12:42.598 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout mon_osd_allow_primary_affinity = true 2026-03-09T15:12:42.598 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout mon_warn_on_no_sortbitwise = false 2026-03-09T15:12:42.598 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout osd_crush_chooseleaf_type = 0 2026-03-09T15:12:42.598 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout 2026-03-09T15:12:42.598 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout [mgr] 2026-03-09T15:12:42.598 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout mgr/cephadm/use_agent = False 2026-03-09T15:12:42.598 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout mgr/telemetry/nag = false 2026-03-09T15:12:42.598 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout 2026-03-09T15:12:42.598 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout [osd] 2026-03-09T15:12:42.598 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout osd_map_max_advance = 10 2026-03-09T15:12:42.598 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout osd_sloppy_crc = true 2026-03-09T15:12:42.598 INFO:teuthology.orchestra.run.vm03.stdout:Generating new minimal ceph.conf... 2026-03-09T15:12:42.797 INFO:teuthology.orchestra.run.vm03.stdout:Restarting the monitor... 2026-03-09T15:12:42.875 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:12:42 vm03 systemd[1]: Stopping Ceph mon.a for 4abede80-1bca-11f1-bcd8-9b1da3c03ce5... 2026-03-09T15:12:43.146 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:12:42 vm03 ceph-4abede80-1bca-11f1-bcd8-9b1da3c03ce5-mon-a[51476]: 2026-03-09T15:12:42.873+0000 7f02dc1a3640 -1 received signal: Terminated from /run/podman-init -- /usr/bin/ceph-mon -n mon.a -f --setuser ceph --setgroup ceph --default-log-to-file=false --default-log-to-journald=true --default-log-to-stderr=false --default-mon-cluster-log-to-file=false --default-mon-cluster-log-to-journald=true --default-mon-cluster-log-to-stderr=false (PID: 1) UID: 0 2026-03-09T15:12:43.146 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:12:42 vm03 ceph-4abede80-1bca-11f1-bcd8-9b1da3c03ce5-mon-a[51476]: 2026-03-09T15:12:42.873+0000 7f02dc1a3640 -1 mon.a@0(leader) e1 *** Got Signal Terminated *** 2026-03-09T15:12:43.146 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:12:42 vm03 podman[51681]: 2026-03-09 15:12:42.936691454 +0000 UTC m=+0.076520247 container died 4ed1bb90c9b2db0638bfdd4ab46e6a852d93f0a17267e76ce928d82d15e30ef7 (image=quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df, name=ceph-4abede80-1bca-11f1-bcd8-9b1da3c03ce5-mon-a, org.opencontainers.image.authors=Ceph Release Team , ceph=True, org.label-schema.build-date=20260223, GANESHA_REPO_BASEURL=https://buildlogs.centos.org/centos/$releasever-stream/storage/$basearch/nfsganesha-5/, org.opencontainers.image.documentation=https://docs.ceph.com/, CEPH_SHA1=e911bdebe5c8faa3800735d1568fcdca65db60df, CEPH_REF=squid, org.label-schema.schema-version=1.0, OSD_FLAVOR=default, CEPH_GIT_REPO=https://github.com/ceph/ceph-ci.git, FROM_IMAGE=quay.io/centos/centos:stream9, org.label-schema.license=GPLv2, io.buildah.version=1.41.3, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.vendor=CentOS) 2026-03-09T15:12:43.146 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:12:42 vm03 podman[51681]: 2026-03-09 15:12:42.959086709 +0000 UTC m=+0.098915502 container remove 4ed1bb90c9b2db0638bfdd4ab46e6a852d93f0a17267e76ce928d82d15e30ef7 (image=quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df, name=ceph-4abede80-1bca-11f1-bcd8-9b1da3c03ce5-mon-a, CEPH_REF=squid, io.buildah.version=1.41.3, ceph=True, org.opencontainers.image.documentation=https://docs.ceph.com/, org.label-schema.vendor=CentOS, FROM_IMAGE=quay.io/centos/centos:stream9, org.label-schema.schema-version=1.0, org.label-schema.license=GPLv2, org.opencontainers.image.authors=Ceph Release Team , GANESHA_REPO_BASEURL=https://buildlogs.centos.org/centos/$releasever-stream/storage/$basearch/nfsganesha-5/, org.label-schema.build-date=20260223, CEPH_SHA1=e911bdebe5c8faa3800735d1568fcdca65db60df, org.label-schema.name=CentOS Stream 9 Base Image, CEPH_GIT_REPO=https://github.com/ceph/ceph-ci.git, OSD_FLAVOR=default) 2026-03-09T15:12:43.146 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:12:42 vm03 bash[51681]: ceph-4abede80-1bca-11f1-bcd8-9b1da3c03ce5-mon-a 2026-03-09T15:12:43.146 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:12:43 vm03 systemd[1]: ceph-4abede80-1bca-11f1-bcd8-9b1da3c03ce5@mon.a.service: Deactivated successfully. 2026-03-09T15:12:43.146 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:12:43 vm03 systemd[1]: Stopped Ceph mon.a for 4abede80-1bca-11f1-bcd8-9b1da3c03ce5. 2026-03-09T15:12:43.146 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:12:43 vm03 systemd[1]: Starting Ceph mon.a for 4abede80-1bca-11f1-bcd8-9b1da3c03ce5... 2026-03-09T15:12:43.146 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:12:43 vm03 podman[51748]: 2026-03-09 15:12:43.118336135 +0000 UTC m=+0.015778165 container create 6be9db7b0806bfc81f0a9ffb9f29a88b12342c6adf3210897061af2f62ab50d0 (image=quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df, name=ceph-4abede80-1bca-11f1-bcd8-9b1da3c03ce5-mon-a, CEPH_GIT_REPO=https://github.com/ceph/ceph-ci.git, FROM_IMAGE=quay.io/centos/centos:stream9, org.label-schema.name=CentOS Stream 9 Base Image, CEPH_SHA1=e911bdebe5c8faa3800735d1568fcdca65db60df, org.label-schema.license=GPLv2, GANESHA_REPO_BASEURL=https://buildlogs.centos.org/centos/$releasever-stream/storage/$basearch/nfsganesha-5/, org.opencontainers.image.authors=Ceph Release Team , org.opencontainers.image.documentation=https://docs.ceph.com/, org.label-schema.build-date=20260223, OSD_FLAVOR=default, CEPH_REF=squid, org.label-schema.schema-version=1.0, ceph=True, io.buildah.version=1.41.3, org.label-schema.vendor=CentOS) 2026-03-09T15:12:43.166 INFO:teuthology.orchestra.run.vm03.stdout:Setting public_network to 192.168.123.0/24 in mon config section 2026-03-09T15:12:43.351 INFO:teuthology.orchestra.run.vm03.stdout:Wrote config to /etc/ceph/ceph.conf 2026-03-09T15:12:43.353 INFO:teuthology.orchestra.run.vm03.stdout:Wrote keyring to /etc/ceph/ceph.client.admin.keyring 2026-03-09T15:12:43.353 INFO:teuthology.orchestra.run.vm03.stdout:Creating mgr... 2026-03-09T15:12:43.353 INFO:teuthology.orchestra.run.vm03.stdout:Verifying port 0.0.0.0:9283 ... 2026-03-09T15:12:43.354 INFO:teuthology.orchestra.run.vm03.stdout:Verifying port 0.0.0.0:8765 ... 2026-03-09T15:12:43.423 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:12:43 vm03 podman[51748]: 2026-03-09 15:12:43.152584309 +0000 UTC m=+0.050026339 container init 6be9db7b0806bfc81f0a9ffb9f29a88b12342c6adf3210897061af2f62ab50d0 (image=quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df, name=ceph-4abede80-1bca-11f1-bcd8-9b1da3c03ce5-mon-a, OSD_FLAVOR=default, io.buildah.version=1.41.3, org.label-schema.build-date=20260223, org.label-schema.license=GPLv2, FROM_IMAGE=quay.io/centos/centos:stream9, org.label-schema.vendor=CentOS, org.label-schema.name=CentOS Stream 9 Base Image, CEPH_SHA1=e911bdebe5c8faa3800735d1568fcdca65db60df, CEPH_GIT_REPO=https://github.com/ceph/ceph-ci.git, GANESHA_REPO_BASEURL=https://buildlogs.centos.org/centos/$releasever-stream/storage/$basearch/nfsganesha-5/, ceph=True, org.opencontainers.image.authors=Ceph Release Team , CEPH_REF=squid, org.label-schema.schema-version=1.0, org.opencontainers.image.documentation=https://docs.ceph.com/) 2026-03-09T15:12:43.424 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:12:43 vm03 podman[51748]: 2026-03-09 15:12:43.15510766 +0000 UTC m=+0.052549690 container start 6be9db7b0806bfc81f0a9ffb9f29a88b12342c6adf3210897061af2f62ab50d0 (image=quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df, name=ceph-4abede80-1bca-11f1-bcd8-9b1da3c03ce5-mon-a, ceph=True, org.opencontainers.image.authors=Ceph Release Team , io.buildah.version=1.41.3, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.schema-version=1.0, org.label-schema.build-date=20260223, OSD_FLAVOR=default, GANESHA_REPO_BASEURL=https://buildlogs.centos.org/centos/$releasever-stream/storage/$basearch/nfsganesha-5/, CEPH_REF=squid, org.label-schema.vendor=CentOS, CEPH_GIT_REPO=https://github.com/ceph/ceph-ci.git, org.label-schema.license=GPLv2, FROM_IMAGE=quay.io/centos/centos:stream9, org.opencontainers.image.documentation=https://docs.ceph.com/, CEPH_SHA1=e911bdebe5c8faa3800735d1568fcdca65db60df) 2026-03-09T15:12:43.424 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:12:43 vm03 bash[51748]: 6be9db7b0806bfc81f0a9ffb9f29a88b12342c6adf3210897061af2f62ab50d0 2026-03-09T15:12:43.424 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:12:43 vm03 podman[51748]: 2026-03-09 15:12:43.112146215 +0000 UTC m=+0.009588255 image pull 654f31e6858eb235bbece362255b685a945f2b6a367e2b88c4930c984fbb214c quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df 2026-03-09T15:12:43.424 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:12:43 vm03 systemd[1]: Started Ceph mon.a for 4abede80-1bca-11f1-bcd8-9b1da3c03ce5. 2026-03-09T15:12:43.424 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:12:43 vm03 ceph-mon[51762]: set uid:gid to 167:167 (ceph:ceph) 2026-03-09T15:12:43.424 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:12:43 vm03 ceph-mon[51762]: ceph version 19.2.3-678-ge911bdeb (e911bdebe5c8faa3800735d1568fcdca65db60df) squid (stable), process ceph-mon, pid 2 2026-03-09T15:12:43.424 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:12:43 vm03 ceph-mon[51762]: pidfile_write: ignore empty --pid-file 2026-03-09T15:12:43.424 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:12:43 vm03 ceph-mon[51762]: load: jerasure load: lrc 2026-03-09T15:12:43.424 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:12:43 vm03 ceph-mon[51762]: rocksdb: RocksDB version: 7.9.2 2026-03-09T15:12:43.424 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:12:43 vm03 ceph-mon[51762]: rocksdb: Git sha 0 2026-03-09T15:12:43.424 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:12:43 vm03 ceph-mon[51762]: rocksdb: Compile date 2026-02-25 18:11:04 2026-03-09T15:12:43.424 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:12:43 vm03 ceph-mon[51762]: rocksdb: DB SUMMARY 2026-03-09T15:12:43.424 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:12:43 vm03 ceph-mon[51762]: rocksdb: DB Session ID: UDJN936ALV6C96O9R81L 2026-03-09T15:12:43.424 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:12:43 vm03 ceph-mon[51762]: rocksdb: CURRENT file: CURRENT 2026-03-09T15:12:43.424 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:12:43 vm03 ceph-mon[51762]: rocksdb: IDENTITY file: IDENTITY 2026-03-09T15:12:43.424 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:12:43 vm03 ceph-mon[51762]: rocksdb: MANIFEST file: MANIFEST-000010 size: 179 Bytes 2026-03-09T15:12:43.424 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:12:43 vm03 ceph-mon[51762]: rocksdb: SST files in /var/lib/ceph/mon/ceph-a/store.db dir, Total Num: 1, files: 000008.sst 2026-03-09T15:12:43.424 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:12:43 vm03 ceph-mon[51762]: rocksdb: Write Ahead Log file in /var/lib/ceph/mon/ceph-a/store.db: 000009.log size: 75535 ; 2026-03-09T15:12:43.424 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:12:43 vm03 ceph-mon[51762]: rocksdb: Options.error_if_exists: 0 2026-03-09T15:12:43.424 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:12:43 vm03 ceph-mon[51762]: rocksdb: Options.create_if_missing: 0 2026-03-09T15:12:43.424 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:12:43 vm03 ceph-mon[51762]: rocksdb: Options.paranoid_checks: 1 2026-03-09T15:12:43.424 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:12:43 vm03 ceph-mon[51762]: rocksdb: Options.flush_verify_memtable_count: 1 2026-03-09T15:12:43.424 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:12:43 vm03 ceph-mon[51762]: rocksdb: Options.track_and_verify_wals_in_manifest: 0 2026-03-09T15:12:43.424 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:12:43 vm03 ceph-mon[51762]: rocksdb: Options.verify_sst_unique_id_in_manifest: 1 2026-03-09T15:12:43.424 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:12:43 vm03 ceph-mon[51762]: rocksdb: Options.env: 0x560e9662adc0 2026-03-09T15:12:43.424 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:12:43 vm03 ceph-mon[51762]: rocksdb: Options.fs: PosixFileSystem 2026-03-09T15:12:43.424 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:12:43 vm03 ceph-mon[51762]: rocksdb: Options.info_log: 0x560e989f4700 2026-03-09T15:12:43.424 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:12:43 vm03 ceph-mon[51762]: rocksdb: Options.max_file_opening_threads: 16 2026-03-09T15:12:43.424 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:12:43 vm03 ceph-mon[51762]: rocksdb: Options.statistics: (nil) 2026-03-09T15:12:43.424 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:12:43 vm03 ceph-mon[51762]: rocksdb: Options.use_fsync: 0 2026-03-09T15:12:43.424 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:12:43 vm03 ceph-mon[51762]: rocksdb: Options.max_log_file_size: 0 2026-03-09T15:12:43.424 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:12:43 vm03 ceph-mon[51762]: rocksdb: Options.max_manifest_file_size: 1073741824 2026-03-09T15:12:43.424 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:12:43 vm03 ceph-mon[51762]: rocksdb: Options.log_file_time_to_roll: 0 2026-03-09T15:12:43.424 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:12:43 vm03 ceph-mon[51762]: rocksdb: Options.keep_log_file_num: 1000 2026-03-09T15:12:43.424 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:12:43 vm03 ceph-mon[51762]: rocksdb: Options.recycle_log_file_num: 0 2026-03-09T15:12:43.424 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:12:43 vm03 ceph-mon[51762]: rocksdb: Options.allow_fallocate: 1 2026-03-09T15:12:43.424 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:12:43 vm03 ceph-mon[51762]: rocksdb: Options.allow_mmap_reads: 0 2026-03-09T15:12:43.424 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:12:43 vm03 ceph-mon[51762]: rocksdb: Options.allow_mmap_writes: 0 2026-03-09T15:12:43.424 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:12:43 vm03 ceph-mon[51762]: rocksdb: Options.use_direct_reads: 0 2026-03-09T15:12:43.424 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:12:43 vm03 ceph-mon[51762]: rocksdb: Options.use_direct_io_for_flush_and_compaction: 0 2026-03-09T15:12:43.424 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:12:43 vm03 ceph-mon[51762]: rocksdb: Options.create_missing_column_families: 0 2026-03-09T15:12:43.424 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:12:43 vm03 ceph-mon[51762]: rocksdb: Options.db_log_dir: 2026-03-09T15:12:43.424 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:12:43 vm03 ceph-mon[51762]: rocksdb: Options.wal_dir: 2026-03-09T15:12:43.424 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:12:43 vm03 ceph-mon[51762]: rocksdb: Options.table_cache_numshardbits: 6 2026-03-09T15:12:43.424 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:12:43 vm03 ceph-mon[51762]: rocksdb: Options.WAL_ttl_seconds: 0 2026-03-09T15:12:43.425 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:12:43 vm03 ceph-mon[51762]: rocksdb: Options.WAL_size_limit_MB: 0 2026-03-09T15:12:43.425 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:12:43 vm03 ceph-mon[51762]: rocksdb: Options.max_write_batch_group_size_bytes: 1048576 2026-03-09T15:12:43.425 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:12:43 vm03 ceph-mon[51762]: rocksdb: Options.manifest_preallocation_size: 4194304 2026-03-09T15:12:43.425 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:12:43 vm03 ceph-mon[51762]: rocksdb: Options.is_fd_close_on_exec: 1 2026-03-09T15:12:43.425 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:12:43 vm03 ceph-mon[51762]: rocksdb: Options.advise_random_on_open: 1 2026-03-09T15:12:43.425 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:12:43 vm03 ceph-mon[51762]: rocksdb: Options.db_write_buffer_size: 0 2026-03-09T15:12:43.425 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:12:43 vm03 ceph-mon[51762]: rocksdb: Options.write_buffer_manager: 0x560e989f9900 2026-03-09T15:12:43.425 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:12:43 vm03 ceph-mon[51762]: rocksdb: Options.access_hint_on_compaction_start: 1 2026-03-09T15:12:43.425 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:12:43 vm03 ceph-mon[51762]: rocksdb: Options.random_access_max_buffer_size: 1048576 2026-03-09T15:12:43.425 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:12:43 vm03 ceph-mon[51762]: rocksdb: Options.use_adaptive_mutex: 0 2026-03-09T15:12:43.425 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:12:43 vm03 ceph-mon[51762]: rocksdb: Options.rate_limiter: (nil) 2026-03-09T15:12:43.425 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:12:43 vm03 ceph-mon[51762]: rocksdb: Options.sst_file_manager.rate_bytes_per_sec: 0 2026-03-09T15:12:43.425 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:12:43 vm03 ceph-mon[51762]: rocksdb: Options.wal_recovery_mode: 2 2026-03-09T15:12:43.425 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:12:43 vm03 ceph-mon[51762]: rocksdb: Options.enable_thread_tracking: 0 2026-03-09T15:12:43.425 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:12:43 vm03 ceph-mon[51762]: rocksdb: Options.enable_pipelined_write: 0 2026-03-09T15:12:43.425 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:12:43 vm03 ceph-mon[51762]: rocksdb: Options.unordered_write: 0 2026-03-09T15:12:43.425 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:12:43 vm03 ceph-mon[51762]: rocksdb: Options.allow_concurrent_memtable_write: 1 2026-03-09T15:12:43.425 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:12:43 vm03 ceph-mon[51762]: rocksdb: Options.enable_write_thread_adaptive_yield: 1 2026-03-09T15:12:43.425 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:12:43 vm03 ceph-mon[51762]: rocksdb: Options.write_thread_max_yield_usec: 100 2026-03-09T15:12:43.425 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:12:43 vm03 ceph-mon[51762]: rocksdb: Options.write_thread_slow_yield_usec: 3 2026-03-09T15:12:43.425 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:12:43 vm03 ceph-mon[51762]: rocksdb: Options.row_cache: None 2026-03-09T15:12:43.425 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:12:43 vm03 ceph-mon[51762]: rocksdb: Options.wal_filter: None 2026-03-09T15:12:43.425 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:12:43 vm03 ceph-mon[51762]: rocksdb: Options.avoid_flush_during_recovery: 0 2026-03-09T15:12:43.425 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:12:43 vm03 ceph-mon[51762]: rocksdb: Options.allow_ingest_behind: 0 2026-03-09T15:12:43.425 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:12:43 vm03 ceph-mon[51762]: rocksdb: Options.two_write_queues: 0 2026-03-09T15:12:43.425 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:12:43 vm03 ceph-mon[51762]: rocksdb: Options.manual_wal_flush: 0 2026-03-09T15:12:43.425 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:12:43 vm03 ceph-mon[51762]: rocksdb: Options.wal_compression: 0 2026-03-09T15:12:43.425 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:12:43 vm03 ceph-mon[51762]: rocksdb: Options.atomic_flush: 0 2026-03-09T15:12:43.425 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:12:43 vm03 ceph-mon[51762]: rocksdb: Options.avoid_unnecessary_blocking_io: 0 2026-03-09T15:12:43.425 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:12:43 vm03 ceph-mon[51762]: rocksdb: Options.persist_stats_to_disk: 0 2026-03-09T15:12:43.425 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:12:43 vm03 ceph-mon[51762]: rocksdb: Options.write_dbid_to_manifest: 0 2026-03-09T15:12:43.425 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:12:43 vm03 ceph-mon[51762]: rocksdb: Options.log_readahead_size: 0 2026-03-09T15:12:43.425 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:12:43 vm03 ceph-mon[51762]: rocksdb: Options.file_checksum_gen_factory: Unknown 2026-03-09T15:12:43.425 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:12:43 vm03 ceph-mon[51762]: rocksdb: Options.best_efforts_recovery: 0 2026-03-09T15:12:43.425 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:12:43 vm03 ceph-mon[51762]: rocksdb: Options.max_bgerror_resume_count: 2147483647 2026-03-09T15:12:43.425 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:12:43 vm03 ceph-mon[51762]: rocksdb: Options.bgerror_resume_retry_interval: 1000000 2026-03-09T15:12:43.425 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:12:43 vm03 ceph-mon[51762]: rocksdb: Options.allow_data_in_errors: 0 2026-03-09T15:12:43.425 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:12:43 vm03 ceph-mon[51762]: rocksdb: Options.db_host_id: __hostname__ 2026-03-09T15:12:43.425 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:12:43 vm03 ceph-mon[51762]: rocksdb: Options.enforce_single_del_contracts: true 2026-03-09T15:12:43.425 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:12:43 vm03 ceph-mon[51762]: rocksdb: Options.max_background_jobs: 2 2026-03-09T15:12:43.425 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:12:43 vm03 ceph-mon[51762]: rocksdb: Options.max_background_compactions: -1 2026-03-09T15:12:43.425 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:12:43 vm03 ceph-mon[51762]: rocksdb: Options.max_subcompactions: 1 2026-03-09T15:12:43.425 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:12:43 vm03 ceph-mon[51762]: rocksdb: Options.avoid_flush_during_shutdown: 0 2026-03-09T15:12:43.425 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:12:43 vm03 ceph-mon[51762]: rocksdb: Options.writable_file_max_buffer_size: 1048576 2026-03-09T15:12:43.425 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:12:43 vm03 ceph-mon[51762]: rocksdb: Options.delayed_write_rate : 16777216 2026-03-09T15:12:43.425 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:12:43 vm03 ceph-mon[51762]: rocksdb: Options.max_total_wal_size: 0 2026-03-09T15:12:43.425 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:12:43 vm03 ceph-mon[51762]: rocksdb: Options.delete_obsolete_files_period_micros: 21600000000 2026-03-09T15:12:43.425 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:12:43 vm03 ceph-mon[51762]: rocksdb: Options.stats_dump_period_sec: 600 2026-03-09T15:12:43.425 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:12:43 vm03 ceph-mon[51762]: rocksdb: Options.stats_persist_period_sec: 600 2026-03-09T15:12:43.425 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:12:43 vm03 ceph-mon[51762]: rocksdb: Options.stats_history_buffer_size: 1048576 2026-03-09T15:12:43.425 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:12:43 vm03 ceph-mon[51762]: rocksdb: Options.max_open_files: -1 2026-03-09T15:12:43.425 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:12:43 vm03 ceph-mon[51762]: rocksdb: Options.bytes_per_sync: 0 2026-03-09T15:12:43.425 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:12:43 vm03 ceph-mon[51762]: rocksdb: Options.wal_bytes_per_sync: 0 2026-03-09T15:12:43.425 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:12:43 vm03 ceph-mon[51762]: rocksdb: Options.strict_bytes_per_sync: 0 2026-03-09T15:12:43.425 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:12:43 vm03 ceph-mon[51762]: rocksdb: Options.compaction_readahead_size: 0 2026-03-09T15:12:43.425 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:12:43 vm03 ceph-mon[51762]: rocksdb: Options.max_background_flushes: -1 2026-03-09T15:12:43.425 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:12:43 vm03 ceph-mon[51762]: rocksdb: Compression algorithms supported: 2026-03-09T15:12:43.425 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:12:43 vm03 ceph-mon[51762]: rocksdb: kZSTD supported: 0 2026-03-09T15:12:43.425 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:12:43 vm03 ceph-mon[51762]: rocksdb: kXpressCompression supported: 0 2026-03-09T15:12:43.425 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:12:43 vm03 ceph-mon[51762]: rocksdb: kBZip2Compression supported: 0 2026-03-09T15:12:43.426 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:12:43 vm03 ceph-mon[51762]: rocksdb: kZSTDNotFinalCompression supported: 0 2026-03-09T15:12:43.426 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:12:43 vm03 ceph-mon[51762]: rocksdb: kLZ4Compression supported: 1 2026-03-09T15:12:43.426 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:12:43 vm03 ceph-mon[51762]: rocksdb: kZlibCompression supported: 1 2026-03-09T15:12:43.426 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:12:43 vm03 ceph-mon[51762]: rocksdb: kLZ4HCCompression supported: 1 2026-03-09T15:12:43.426 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:12:43 vm03 ceph-mon[51762]: rocksdb: kSnappyCompression supported: 1 2026-03-09T15:12:43.426 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:12:43 vm03 ceph-mon[51762]: rocksdb: Fast CRC32 supported: Supported on x86 2026-03-09T15:12:43.426 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:12:43 vm03 ceph-mon[51762]: rocksdb: DMutex implementation: pthread_mutex_t 2026-03-09T15:12:43.426 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:12:43 vm03 ceph-mon[51762]: rocksdb: [db/version_set.cc:5527] Recovering from manifest file: /var/lib/ceph/mon/ceph-a/store.db/MANIFEST-000010 2026-03-09T15:12:43.426 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:12:43 vm03 ceph-mon[51762]: rocksdb: [db/column_family.cc:630] --------------- Options for column family [default]: 2026-03-09T15:12:43.426 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:12:43 vm03 ceph-mon[51762]: rocksdb: Options.comparator: leveldb.BytewiseComparator 2026-03-09T15:12:43.426 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:12:43 vm03 ceph-mon[51762]: rocksdb: Options.merge_operator: 2026-03-09T15:12:43.426 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:12:43 vm03 ceph-mon[51762]: rocksdb: Options.compaction_filter: None 2026-03-09T15:12:43.426 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:12:43 vm03 ceph-mon[51762]: rocksdb: Options.compaction_filter_factory: None 2026-03-09T15:12:43.426 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:12:43 vm03 ceph-mon[51762]: rocksdb: Options.sst_partitioner_factory: None 2026-03-09T15:12:43.426 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:12:43 vm03 ceph-mon[51762]: rocksdb: Options.memtable_factory: SkipListFactory 2026-03-09T15:12:43.426 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:12:43 vm03 ceph-mon[51762]: rocksdb: Options.table_factory: BlockBasedTable 2026-03-09T15:12:43.426 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:12:43 vm03 ceph-mon[51762]: rocksdb: table_factory options: flush_block_policy_factory: FlushBlockBySizePolicyFactory (0x560e989f4640) 2026-03-09T15:12:43.426 INFO:journalctl@ceph.mon.a.vm03.stdout: cache_index_and_filter_blocks: 1 2026-03-09T15:12:43.426 INFO:journalctl@ceph.mon.a.vm03.stdout: cache_index_and_filter_blocks_with_high_priority: 0 2026-03-09T15:12:43.426 INFO:journalctl@ceph.mon.a.vm03.stdout: pin_l0_filter_and_index_blocks_in_cache: 0 2026-03-09T15:12:43.426 INFO:journalctl@ceph.mon.a.vm03.stdout: pin_top_level_index_and_filter: 1 2026-03-09T15:12:43.426 INFO:journalctl@ceph.mon.a.vm03.stdout: index_type: 0 2026-03-09T15:12:43.426 INFO:journalctl@ceph.mon.a.vm03.stdout: data_block_index_type: 0 2026-03-09T15:12:43.426 INFO:journalctl@ceph.mon.a.vm03.stdout: index_shortening: 1 2026-03-09T15:12:43.426 INFO:journalctl@ceph.mon.a.vm03.stdout: data_block_hash_table_util_ratio: 0.750000 2026-03-09T15:12:43.426 INFO:journalctl@ceph.mon.a.vm03.stdout: checksum: 4 2026-03-09T15:12:43.426 INFO:journalctl@ceph.mon.a.vm03.stdout: no_block_cache: 0 2026-03-09T15:12:43.426 INFO:journalctl@ceph.mon.a.vm03.stdout: block_cache: 0x560e98a19350 2026-03-09T15:12:43.426 INFO:journalctl@ceph.mon.a.vm03.stdout: block_cache_name: BinnedLRUCache 2026-03-09T15:12:43.426 INFO:journalctl@ceph.mon.a.vm03.stdout: block_cache_options: 2026-03-09T15:12:43.426 INFO:journalctl@ceph.mon.a.vm03.stdout: capacity : 536870912 2026-03-09T15:12:43.426 INFO:journalctl@ceph.mon.a.vm03.stdout: num_shard_bits : 4 2026-03-09T15:12:43.426 INFO:journalctl@ceph.mon.a.vm03.stdout: strict_capacity_limit : 0 2026-03-09T15:12:43.426 INFO:journalctl@ceph.mon.a.vm03.stdout: high_pri_pool_ratio: 0.000 2026-03-09T15:12:43.426 INFO:journalctl@ceph.mon.a.vm03.stdout: block_cache_compressed: (nil) 2026-03-09T15:12:43.426 INFO:journalctl@ceph.mon.a.vm03.stdout: persistent_cache: (nil) 2026-03-09T15:12:43.426 INFO:journalctl@ceph.mon.a.vm03.stdout: block_size: 4096 2026-03-09T15:12:43.426 INFO:journalctl@ceph.mon.a.vm03.stdout: block_size_deviation: 10 2026-03-09T15:12:43.426 INFO:journalctl@ceph.mon.a.vm03.stdout: block_restart_interval: 16 2026-03-09T15:12:43.426 INFO:journalctl@ceph.mon.a.vm03.stdout: index_block_restart_interval: 1 2026-03-09T15:12:43.426 INFO:journalctl@ceph.mon.a.vm03.stdout: metadata_block_size: 4096 2026-03-09T15:12:43.426 INFO:journalctl@ceph.mon.a.vm03.stdout: partition_filters: 0 2026-03-09T15:12:43.426 INFO:journalctl@ceph.mon.a.vm03.stdout: use_delta_encoding: 1 2026-03-09T15:12:43.426 INFO:journalctl@ceph.mon.a.vm03.stdout: filter_policy: bloomfilter 2026-03-09T15:12:43.426 INFO:journalctl@ceph.mon.a.vm03.stdout: whole_key_filtering: 1 2026-03-09T15:12:43.426 INFO:journalctl@ceph.mon.a.vm03.stdout: verify_compression: 0 2026-03-09T15:12:43.426 INFO:journalctl@ceph.mon.a.vm03.stdout: read_amp_bytes_per_bit: 0 2026-03-09T15:12:43.426 INFO:journalctl@ceph.mon.a.vm03.stdout: format_version: 5 2026-03-09T15:12:43.426 INFO:journalctl@ceph.mon.a.vm03.stdout: enable_index_compression: 1 2026-03-09T15:12:43.426 INFO:journalctl@ceph.mon.a.vm03.stdout: block_align: 0 2026-03-09T15:12:43.426 INFO:journalctl@ceph.mon.a.vm03.stdout: max_auto_readahead_size: 262144 2026-03-09T15:12:43.426 INFO:journalctl@ceph.mon.a.vm03.stdout: prepopulate_block_cache: 0 2026-03-09T15:12:43.426 INFO:journalctl@ceph.mon.a.vm03.stdout: initial_auto_readahead_size: 8192 2026-03-09T15:12:43.426 INFO:journalctl@ceph.mon.a.vm03.stdout: num_file_reads_for_auto_readahead: 2 2026-03-09T15:12:43.426 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:12:43 vm03 ceph-mon[51762]: rocksdb: Options.write_buffer_size: 33554432 2026-03-09T15:12:43.426 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:12:43 vm03 ceph-mon[51762]: rocksdb: Options.max_write_buffer_number: 2 2026-03-09T15:12:43.426 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:12:43 vm03 ceph-mon[51762]: rocksdb: Options.compression: NoCompression 2026-03-09T15:12:43.426 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:12:43 vm03 ceph-mon[51762]: rocksdb: Options.bottommost_compression: Disabled 2026-03-09T15:12:43.426 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:12:43 vm03 ceph-mon[51762]: rocksdb: Options.prefix_extractor: nullptr 2026-03-09T15:12:43.427 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:12:43 vm03 ceph-mon[51762]: rocksdb: Options.memtable_insert_with_hint_prefix_extractor: nullptr 2026-03-09T15:12:43.427 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:12:43 vm03 ceph-mon[51762]: rocksdb: Options.num_levels: 7 2026-03-09T15:12:43.427 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:12:43 vm03 ceph-mon[51762]: rocksdb: Options.min_write_buffer_number_to_merge: 1 2026-03-09T15:12:43.427 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:12:43 vm03 ceph-mon[51762]: rocksdb: Options.max_write_buffer_number_to_maintain: 0 2026-03-09T15:12:43.427 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:12:43 vm03 ceph-mon[51762]: rocksdb: Options.max_write_buffer_size_to_maintain: 0 2026-03-09T15:12:43.427 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:12:43 vm03 ceph-mon[51762]: rocksdb: Options.bottommost_compression_opts.window_bits: -14 2026-03-09T15:12:43.427 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:12:43 vm03 ceph-mon[51762]: rocksdb: Options.bottommost_compression_opts.level: 32767 2026-03-09T15:12:43.427 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:12:43 vm03 ceph-mon[51762]: rocksdb: Options.bottommost_compression_opts.strategy: 0 2026-03-09T15:12:43.427 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:12:43 vm03 ceph-mon[51762]: rocksdb: Options.bottommost_compression_opts.max_dict_bytes: 0 2026-03-09T15:12:43.427 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:12:43 vm03 ceph-mon[51762]: rocksdb: Options.bottommost_compression_opts.zstd_max_train_bytes: 0 2026-03-09T15:12:43.427 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:12:43 vm03 ceph-mon[51762]: rocksdb: Options.bottommost_compression_opts.parallel_threads: 1 2026-03-09T15:12:43.427 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:12:43 vm03 ceph-mon[51762]: rocksdb: Options.bottommost_compression_opts.enabled: false 2026-03-09T15:12:43.427 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:12:43 vm03 ceph-mon[51762]: rocksdb: Options.bottommost_compression_opts.max_dict_buffer_bytes: 0 2026-03-09T15:12:43.427 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:12:43 vm03 ceph-mon[51762]: rocksdb: Options.bottommost_compression_opts.use_zstd_dict_trainer: true 2026-03-09T15:12:43.427 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:12:43 vm03 ceph-mon[51762]: rocksdb: Options.compression_opts.window_bits: -14 2026-03-09T15:12:43.427 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:12:43 vm03 ceph-mon[51762]: rocksdb: Options.compression_opts.level: 32767 2026-03-09T15:12:43.427 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:12:43 vm03 ceph-mon[51762]: rocksdb: Options.compression_opts.strategy: 0 2026-03-09T15:12:43.427 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:12:43 vm03 ceph-mon[51762]: rocksdb: Options.compression_opts.max_dict_bytes: 0 2026-03-09T15:12:43.427 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:12:43 vm03 ceph-mon[51762]: rocksdb: Options.compression_opts.zstd_max_train_bytes: 0 2026-03-09T15:12:43.427 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:12:43 vm03 ceph-mon[51762]: rocksdb: Options.compression_opts.use_zstd_dict_trainer: true 2026-03-09T15:12:43.427 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:12:43 vm03 ceph-mon[51762]: rocksdb: Options.compression_opts.parallel_threads: 1 2026-03-09T15:12:43.427 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:12:43 vm03 ceph-mon[51762]: rocksdb: Options.compression_opts.enabled: false 2026-03-09T15:12:43.427 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:12:43 vm03 ceph-mon[51762]: rocksdb: Options.compression_opts.max_dict_buffer_bytes: 0 2026-03-09T15:12:43.427 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:12:43 vm03 ceph-mon[51762]: rocksdb: Options.level0_file_num_compaction_trigger: 4 2026-03-09T15:12:43.427 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:12:43 vm03 ceph-mon[51762]: rocksdb: Options.level0_slowdown_writes_trigger: 20 2026-03-09T15:12:43.427 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:12:43 vm03 ceph-mon[51762]: rocksdb: Options.level0_stop_writes_trigger: 36 2026-03-09T15:12:43.427 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:12:43 vm03 ceph-mon[51762]: rocksdb: Options.target_file_size_base: 67108864 2026-03-09T15:12:43.427 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:12:43 vm03 ceph-mon[51762]: rocksdb: Options.target_file_size_multiplier: 1 2026-03-09T15:12:43.427 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:12:43 vm03 ceph-mon[51762]: rocksdb: Options.max_bytes_for_level_base: 268435456 2026-03-09T15:12:43.427 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:12:43 vm03 ceph-mon[51762]: rocksdb: Options.level_compaction_dynamic_level_bytes: 1 2026-03-09T15:12:43.427 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:12:43 vm03 ceph-mon[51762]: rocksdb: Options.max_bytes_for_level_multiplier: 10.000000 2026-03-09T15:12:43.427 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:12:43 vm03 ceph-mon[51762]: rocksdb: Options.max_bytes_for_level_multiplier_addtl[0]: 1 2026-03-09T15:12:43.427 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:12:43 vm03 ceph-mon[51762]: rocksdb: Options.max_bytes_for_level_multiplier_addtl[1]: 1 2026-03-09T15:12:43.427 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:12:43 vm03 ceph-mon[51762]: rocksdb: Options.max_bytes_for_level_multiplier_addtl[2]: 1 2026-03-09T15:12:43.427 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:12:43 vm03 ceph-mon[51762]: rocksdb: Options.max_bytes_for_level_multiplier_addtl[3]: 1 2026-03-09T15:12:43.427 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:12:43 vm03 ceph-mon[51762]: rocksdb: Options.max_bytes_for_level_multiplier_addtl[4]: 1 2026-03-09T15:12:43.427 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:12:43 vm03 ceph-mon[51762]: rocksdb: Options.max_bytes_for_level_multiplier_addtl[5]: 1 2026-03-09T15:12:43.427 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:12:43 vm03 ceph-mon[51762]: rocksdb: Options.max_bytes_for_level_multiplier_addtl[6]: 1 2026-03-09T15:12:43.427 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:12:43 vm03 ceph-mon[51762]: rocksdb: Options.max_sequential_skip_in_iterations: 8 2026-03-09T15:12:43.427 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:12:43 vm03 ceph-mon[51762]: rocksdb: Options.max_compaction_bytes: 1677721600 2026-03-09T15:12:43.427 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:12:43 vm03 ceph-mon[51762]: rocksdb: Options.ignore_max_compaction_bytes_for_input: true 2026-03-09T15:12:43.427 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:12:43 vm03 ceph-mon[51762]: rocksdb: Options.arena_block_size: 1048576 2026-03-09T15:12:43.427 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:12:43 vm03 ceph-mon[51762]: rocksdb: Options.soft_pending_compaction_bytes_limit: 68719476736 2026-03-09T15:12:43.427 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:12:43 vm03 ceph-mon[51762]: rocksdb: Options.hard_pending_compaction_bytes_limit: 274877906944 2026-03-09T15:12:43.427 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:12:43 vm03 ceph-mon[51762]: rocksdb: Options.disable_auto_compactions: 0 2026-03-09T15:12:43.427 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:12:43 vm03 ceph-mon[51762]: rocksdb: Options.compaction_style: kCompactionStyleLevel 2026-03-09T15:12:43.427 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:12:43 vm03 ceph-mon[51762]: rocksdb: Options.compaction_pri: kMinOverlappingRatio 2026-03-09T15:12:43.427 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:12:43 vm03 ceph-mon[51762]: rocksdb: Options.compaction_options_universal.size_ratio: 1 2026-03-09T15:12:43.427 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:12:43 vm03 ceph-mon[51762]: rocksdb: Options.compaction_options_universal.min_merge_width: 2 2026-03-09T15:12:43.427 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:12:43 vm03 ceph-mon[51762]: rocksdb: Options.compaction_options_universal.max_merge_width: 4294967295 2026-03-09T15:12:43.427 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:12:43 vm03 ceph-mon[51762]: rocksdb: Options.compaction_options_universal.max_size_amplification_percent: 200 2026-03-09T15:12:43.428 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:12:43 vm03 ceph-mon[51762]: rocksdb: Options.compaction_options_universal.compression_size_percent: -1 2026-03-09T15:12:43.428 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:12:43 vm03 ceph-mon[51762]: rocksdb: Options.compaction_options_universal.stop_style: kCompactionStopStyleTotalSize 2026-03-09T15:12:43.428 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:12:43 vm03 ceph-mon[51762]: rocksdb: Options.compaction_options_fifo.max_table_files_size: 1073741824 2026-03-09T15:12:43.428 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:12:43 vm03 ceph-mon[51762]: rocksdb: Options.compaction_options_fifo.allow_compaction: 0 2026-03-09T15:12:43.428 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:12:43 vm03 ceph-mon[51762]: rocksdb: Options.table_properties_collectors: CompactOnDeletionCollector (Sliding window size = 32768 Deletion trigger = 16384 Deletion ratio = 0); 2026-03-09T15:12:43.428 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:12:43 vm03 ceph-mon[51762]: rocksdb: Options.inplace_update_support: 0 2026-03-09T15:12:43.428 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:12:43 vm03 ceph-mon[51762]: rocksdb: Options.inplace_update_num_locks: 10000 2026-03-09T15:12:43.428 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:12:43 vm03 ceph-mon[51762]: rocksdb: Options.memtable_prefix_bloom_size_ratio: 0.000000 2026-03-09T15:12:43.428 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:12:43 vm03 ceph-mon[51762]: rocksdb: Options.memtable_whole_key_filtering: 0 2026-03-09T15:12:43.428 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:12:43 vm03 ceph-mon[51762]: rocksdb: Options.memtable_huge_page_size: 0 2026-03-09T15:12:43.428 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:12:43 vm03 ceph-mon[51762]: rocksdb: Options.bloom_locality: 0 2026-03-09T15:12:43.428 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:12:43 vm03 ceph-mon[51762]: rocksdb: Options.max_successive_merges: 0 2026-03-09T15:12:43.428 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:12:43 vm03 ceph-mon[51762]: rocksdb: Options.optimize_filters_for_hits: 0 2026-03-09T15:12:43.428 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:12:43 vm03 ceph-mon[51762]: rocksdb: Options.paranoid_file_checks: 0 2026-03-09T15:12:43.428 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:12:43 vm03 ceph-mon[51762]: rocksdb: Options.force_consistency_checks: 1 2026-03-09T15:12:43.428 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:12:43 vm03 ceph-mon[51762]: rocksdb: Options.report_bg_io_stats: 0 2026-03-09T15:12:43.428 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:12:43 vm03 ceph-mon[51762]: rocksdb: Options.ttl: 2592000 2026-03-09T15:12:43.428 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:12:43 vm03 ceph-mon[51762]: rocksdb: Options.periodic_compaction_seconds: 0 2026-03-09T15:12:43.428 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:12:43 vm03 ceph-mon[51762]: rocksdb: Options.preclude_last_level_data_seconds: 0 2026-03-09T15:12:43.428 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:12:43 vm03 ceph-mon[51762]: rocksdb: Options.preserve_internal_time_seconds: 0 2026-03-09T15:12:43.428 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:12:43 vm03 ceph-mon[51762]: rocksdb: Options.enable_blob_files: false 2026-03-09T15:12:43.428 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:12:43 vm03 ceph-mon[51762]: rocksdb: Options.min_blob_size: 0 2026-03-09T15:12:43.428 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:12:43 vm03 ceph-mon[51762]: rocksdb: Options.blob_file_size: 268435456 2026-03-09T15:12:43.428 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:12:43 vm03 ceph-mon[51762]: rocksdb: Options.blob_compression_type: NoCompression 2026-03-09T15:12:43.428 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:12:43 vm03 ceph-mon[51762]: rocksdb: Options.enable_blob_garbage_collection: false 2026-03-09T15:12:43.428 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:12:43 vm03 ceph-mon[51762]: rocksdb: Options.blob_garbage_collection_age_cutoff: 0.250000 2026-03-09T15:12:43.428 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:12:43 vm03 ceph-mon[51762]: rocksdb: Options.blob_garbage_collection_force_threshold: 1.000000 2026-03-09T15:12:43.428 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:12:43 vm03 ceph-mon[51762]: rocksdb: Options.blob_compaction_readahead_size: 0 2026-03-09T15:12:43.428 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:12:43 vm03 ceph-mon[51762]: rocksdb: Options.blob_file_starting_level: 0 2026-03-09T15:12:43.428 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:12:43 vm03 ceph-mon[51762]: rocksdb: Options.experimental_mempurge_threshold: 0.000000 2026-03-09T15:12:43.428 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:12:43 vm03 ceph-mon[51762]: rocksdb: [db/version_set.cc:5566] Recovered from manifest file:/var/lib/ceph/mon/ceph-a/store.db/MANIFEST-000010 succeeded,manifest_file_number is 10, next_file_number is 12, last_sequence is 5, log_number is 5,prev_log_number is 0,max_column_family is 0,min_log_number_to_keep is 5 2026-03-09T15:12:43.428 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:12:43 vm03 ceph-mon[51762]: rocksdb: [db/version_set.cc:5581] Column family [default] (ID 0), log number is 5 2026-03-09T15:12:43.428 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:12:43 vm03 ceph-mon[51762]: rocksdb: [db/db_impl/db_impl_open.cc:539] DB ID: 1f0d4009-dabc-455e-b229-bea10d2de78c 2026-03-09T15:12:43.428 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:12:43 vm03 ceph-mon[51762]: rocksdb: EVENT_LOG_v1 {"time_micros": 1773069163183732, "job": 1, "event": "recovery_started", "wal_files": [9]} 2026-03-09T15:12:43.428 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:12:43 vm03 ceph-mon[51762]: rocksdb: [db/db_impl/db_impl_open.cc:1043] Recovering log #9 mode 2 2026-03-09T15:12:43.428 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:12:43 vm03 ceph-mon[51762]: rocksdb: EVENT_LOG_v1 {"time_micros": 1773069163185290, "cf_name": "default", "job": 1, "event": "table_file_creation", "file_number": 13, "file_size": 72616, "file_checksum": "", "file_checksum_func_name": "Unknown", "smallest_seqno": 8, "largest_seqno": 225, "table_properties": {"data_size": 70895, "index_size": 174, "index_partitions": 0, "top_level_index_size": 0, "index_key_is_user_key": 1, "index_value_is_delta_encoded": 1, "filter_size": 517, "raw_key_size": 9705, "raw_average_key_size": 49, "raw_value_size": 65374, "raw_average_value_size": 333, "num_data_blocks": 8, "num_entries": 196, "num_filter_entries": 196, "num_deletions": 3, "num_merge_operands": 0, "num_range_deletions": 0, "format_version": 0, "fixed_key_len": 0, "filter_policy": "bloomfilter", "column_family_name": "default", "column_family_id": 0, "comparator": "leveldb.BytewiseComparator", "merge_operator": "", "prefix_extractor_name": "nullptr", "property_collectors": "[CompactOnDeletionCollector]", "compression": "NoCompression", "compression_options": "window_bits=-14; level=32767; strategy=0; max_dict_bytes=0; zstd_max_train_bytes=0; enabled=0; max_dict_buffer_bytes=0; use_zstd_dict_trainer=1; ", "creation_time": 1773069163, "oldest_key_time": 0, "file_creation_time": 0, "slow_compression_estimated_data_size": 0, "fast_compression_estimated_data_size": 0, "db_id": "1f0d4009-dabc-455e-b229-bea10d2de78c", "db_session_id": "UDJN936ALV6C96O9R81L", "orig_file_number": 13, "seqno_to_time_mapping": "N/A"}} 2026-03-09T15:12:43.428 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:12:43 vm03 ceph-mon[51762]: rocksdb: EVENT_LOG_v1 {"time_micros": 1773069163185347, "job": 1, "event": "recovery_finished"} 2026-03-09T15:12:43.428 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:12:43 vm03 ceph-mon[51762]: rocksdb: [db/version_set.cc:5047] Creating manifest 15 2026-03-09T15:12:43.428 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:12:43 vm03 ceph-mon[51762]: rocksdb: [file/delete_scheduler.cc:74] Deleted file /var/lib/ceph/mon/ceph-a/store.db/000009.log immediately, rate_bytes_per_sec 0, total_trash_size 0 max_trash_db_ratio 0.250000 2026-03-09T15:12:43.428 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:12:43 vm03 ceph-mon[51762]: rocksdb: [db/db_impl/db_impl_open.cc:1987] SstFileManager instance 0x560e98a1ae00 2026-03-09T15:12:43.428 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:12:43 vm03 ceph-mon[51762]: rocksdb: DB pointer 0x560e98b30000 2026-03-09T15:12:43.428 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:12:43 vm03 ceph-mon[51762]: rocksdb: [db/db_impl/db_impl.cc:1109] ------- DUMPING STATS ------- 2026-03-09T15:12:43.428 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:12:43 vm03 ceph-mon[51762]: rocksdb: [db/db_impl/db_impl.cc:1111] 2026-03-09T15:12:43.428 INFO:journalctl@ceph.mon.a.vm03.stdout: ** DB Stats ** 2026-03-09T15:12:43.428 INFO:journalctl@ceph.mon.a.vm03.stdout: Uptime(secs): 0.0 total, 0.0 interval 2026-03-09T15:12:43.428 INFO:journalctl@ceph.mon.a.vm03.stdout: Cumulative writes: 0 writes, 0 keys, 0 commit groups, 0.0 writes per commit group, ingest: 0.00 GB, 0.00 MB/s 2026-03-09T15:12:43.428 INFO:journalctl@ceph.mon.a.vm03.stdout: Cumulative WAL: 0 writes, 0 syncs, 0.00 writes per sync, written: 0.00 GB, 0.00 MB/s 2026-03-09T15:12:43.428 INFO:journalctl@ceph.mon.a.vm03.stdout: Cumulative stall: 00:00:0.000 H:M:S, 0.0 percent 2026-03-09T15:12:43.428 INFO:journalctl@ceph.mon.a.vm03.stdout: Interval writes: 0 writes, 0 keys, 0 commit groups, 0.0 writes per commit group, ingest: 0.00 MB, 0.00 MB/s 2026-03-09T15:12:43.428 INFO:journalctl@ceph.mon.a.vm03.stdout: Interval WAL: 0 writes, 0 syncs, 0.00 writes per sync, written: 0.00 GB, 0.00 MB/s 2026-03-09T15:12:43.428 INFO:journalctl@ceph.mon.a.vm03.stdout: Interval stall: 00:00:0.000 H:M:S, 0.0 percent 2026-03-09T15:12:43.428 INFO:journalctl@ceph.mon.a.vm03.stdout: 2026-03-09T15:12:43.428 INFO:journalctl@ceph.mon.a.vm03.stdout: ** Compaction Stats [default] ** 2026-03-09T15:12:43.428 INFO:journalctl@ceph.mon.a.vm03.stdout: Level Files Size Score Read(GB) Rn(GB) Rnp1(GB) Write(GB) Wnew(GB) Moved(GB) W-Amp Rd(MB/s) Wr(MB/s) Comp(sec) CompMergeCPU(sec) Comp(cnt) Avg(sec) KeyIn KeyDrop Rblob(GB) Wblob(GB) 2026-03-09T15:12:43.428 INFO:journalctl@ceph.mon.a.vm03.stdout: ------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------ 2026-03-09T15:12:43.428 INFO:journalctl@ceph.mon.a.vm03.stdout: L0 2/0 72.77 KB 0.5 0.0 0.0 0.0 0.0 0.0 0.0 1.0 0.0 52.5 0.00 0.00 1 0.001 0 0 0.0 0.0 2026-03-09T15:12:43.428 INFO:journalctl@ceph.mon.a.vm03.stdout: Sum 2/0 72.77 KB 0.0 0.0 0.0 0.0 0.0 0.0 0.0 1.0 0.0 52.5 0.00 0.00 1 0.001 0 0 0.0 0.0 2026-03-09T15:12:43.428 INFO:journalctl@ceph.mon.a.vm03.stdout: Int 0/0 0.00 KB 0.0 0.0 0.0 0.0 0.0 0.0 0.0 1.0 0.0 52.5 0.00 0.00 1 0.001 0 0 0.0 0.0 2026-03-09T15:12:43.428 INFO:journalctl@ceph.mon.a.vm03.stdout: 2026-03-09T15:12:43.429 INFO:journalctl@ceph.mon.a.vm03.stdout: ** Compaction Stats [default] ** 2026-03-09T15:12:43.429 INFO:journalctl@ceph.mon.a.vm03.stdout: Priority Files Size Score Read(GB) Rn(GB) Rnp1(GB) Write(GB) Wnew(GB) Moved(GB) W-Amp Rd(MB/s) Wr(MB/s) Comp(sec) CompMergeCPU(sec) Comp(cnt) Avg(sec) KeyIn KeyDrop Rblob(GB) Wblob(GB) 2026-03-09T15:12:43.429 INFO:journalctl@ceph.mon.a.vm03.stdout: --------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------- 2026-03-09T15:12:43.429 INFO:journalctl@ceph.mon.a.vm03.stdout: User 0/0 0.00 KB 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0.0 52.5 0.00 0.00 1 0.001 0 0 0.0 0.0 2026-03-09T15:12:43.429 INFO:journalctl@ceph.mon.a.vm03.stdout: 2026-03-09T15:12:43.429 INFO:journalctl@ceph.mon.a.vm03.stdout: Blob file count: 0, total size: 0.0 GB, garbage size: 0.0 GB, space amp: 0.0 2026-03-09T15:12:43.429 INFO:journalctl@ceph.mon.a.vm03.stdout: 2026-03-09T15:12:43.429 INFO:journalctl@ceph.mon.a.vm03.stdout: Uptime(secs): 0.0 total, 0.0 interval 2026-03-09T15:12:43.429 INFO:journalctl@ceph.mon.a.vm03.stdout: Flush(GB): cumulative 0.000, interval 0.000 2026-03-09T15:12:43.429 INFO:journalctl@ceph.mon.a.vm03.stdout: AddFile(GB): cumulative 0.000, interval 0.000 2026-03-09T15:12:43.429 INFO:journalctl@ceph.mon.a.vm03.stdout: AddFile(Total Files): cumulative 0, interval 0 2026-03-09T15:12:43.429 INFO:journalctl@ceph.mon.a.vm03.stdout: AddFile(L0 Files): cumulative 0, interval 0 2026-03-09T15:12:43.429 INFO:journalctl@ceph.mon.a.vm03.stdout: AddFile(Keys): cumulative 0, interval 0 2026-03-09T15:12:43.429 INFO:journalctl@ceph.mon.a.vm03.stdout: Cumulative compaction: 0.00 GB write, 5.89 MB/s write, 0.00 GB read, 0.00 MB/s read, 0.0 seconds 2026-03-09T15:12:43.429 INFO:journalctl@ceph.mon.a.vm03.stdout: Interval compaction: 0.00 GB write, 5.89 MB/s write, 0.00 GB read, 0.00 MB/s read, 0.0 seconds 2026-03-09T15:12:43.429 INFO:journalctl@ceph.mon.a.vm03.stdout: Stalls(count): 0 level0_slowdown, 0 level0_slowdown_with_compaction, 0 level0_numfiles, 0 level0_numfiles_with_compaction, 0 stop for pending_compaction_bytes, 0 slowdown for pending_compaction_bytes, 0 memtable_compaction, 0 memtable_slowdown, interval 0 total count 2026-03-09T15:12:43.429 INFO:journalctl@ceph.mon.a.vm03.stdout: Block cache BinnedLRUCache@0x560e98a19350#2 capacity: 512.00 MB usage: 26.17 KB table_size: 0 occupancy: 18446744073709551615 collections: 1 last_copies: 0 last_secs: 6e-06 secs_since: 0 2026-03-09T15:12:43.429 INFO:journalctl@ceph.mon.a.vm03.stdout: Block cache entry stats(count,size,portion): DataBlock(3,25.11 KB,0.00478923%) FilterBlock(2,0.70 KB,0.00013411%) IndexBlock(2,0.36 KB,6.85453e-05%) Misc(1,0.00 KB,0%) 2026-03-09T15:12:43.429 INFO:journalctl@ceph.mon.a.vm03.stdout: 2026-03-09T15:12:43.429 INFO:journalctl@ceph.mon.a.vm03.stdout: ** File Read Latency Histogram By Level [default] ** 2026-03-09T15:12:43.429 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:12:43 vm03 ceph-mon[51762]: starting mon.a rank 0 at public addrs [v2:192.168.123.103:3300/0,v1:192.168.123.103:6789/0] at bind addrs [v2:192.168.123.103:3300/0,v1:192.168.123.103:6789/0] mon_data /var/lib/ceph/mon/ceph-a fsid 4abede80-1bca-11f1-bcd8-9b1da3c03ce5 2026-03-09T15:12:43.429 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:12:43 vm03 ceph-mon[51762]: mon.a@-1(???) e1 preinit fsid 4abede80-1bca-11f1-bcd8-9b1da3c03ce5 2026-03-09T15:12:43.429 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:12:43 vm03 ceph-mon[51762]: mon.a@-1(???).mds e1 new map 2026-03-09T15:12:43.429 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:12:43 vm03 ceph-mon[51762]: mon.a@-1(???).mds e1 print_map 2026-03-09T15:12:43.429 INFO:journalctl@ceph.mon.a.vm03.stdout: e1 2026-03-09T15:12:43.429 INFO:journalctl@ceph.mon.a.vm03.stdout: btime 2026-03-09T15:12:42:233343+0000 2026-03-09T15:12:43.429 INFO:journalctl@ceph.mon.a.vm03.stdout: enable_multiple, ever_enabled_multiple: 1,1 2026-03-09T15:12:43.429 INFO:journalctl@ceph.mon.a.vm03.stdout: default compat: compat={},rocompat={},incompat={1=base v0.20,2=client writeable ranges,3=default file layouts on dirs,4=dir inode in separate object,5=mds uses versioned encoding,6=dirfrag is stored in omap,8=no anchor table,9=file layout v2,10=snaprealm v2,11=minor log segments,12=quiesce subvolumes} 2026-03-09T15:12:43.429 INFO:journalctl@ceph.mon.a.vm03.stdout: legacy client fscid: -1 2026-03-09T15:12:43.429 INFO:journalctl@ceph.mon.a.vm03.stdout: 2026-03-09T15:12:43.429 INFO:journalctl@ceph.mon.a.vm03.stdout: No filesystems configured 2026-03-09T15:12:43.429 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:12:43 vm03 ceph-mon[51762]: mon.a@-1(???).osd e1 crush map has features 3314932999778484224, adjusting msgr requires 2026-03-09T15:12:43.429 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:12:43 vm03 ceph-mon[51762]: mon.a@-1(???).osd e1 crush map has features 288514050185494528, adjusting msgr requires 2026-03-09T15:12:43.429 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:12:43 vm03 ceph-mon[51762]: mon.a@-1(???).osd e1 crush map has features 288514050185494528, adjusting msgr requires 2026-03-09T15:12:43.429 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:12:43 vm03 ceph-mon[51762]: mon.a@-1(???).osd e1 crush map has features 288514050185494528, adjusting msgr requires 2026-03-09T15:12:43.429 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:12:43 vm03 ceph-mon[51762]: mon.a@-1(???).paxosservice(auth 1..2) refresh upgraded, format 0 -> 3 2026-03-09T15:12:43.429 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:12:43 vm03 ceph-mon[51762]: mon.a is new leader, mons a in quorum (ranks 0) 2026-03-09T15:12:43.429 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:12:43 vm03 ceph-mon[51762]: monmap epoch 1 2026-03-09T15:12:43.429 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:12:43 vm03 ceph-mon[51762]: fsid 4abede80-1bca-11f1-bcd8-9b1da3c03ce5 2026-03-09T15:12:43.429 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:12:43 vm03 ceph-mon[51762]: last_changed 2026-03-09T15:12:40.941469+0000 2026-03-09T15:12:43.429 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:12:43 vm03 ceph-mon[51762]: created 2026-03-09T15:12:40.941469+0000 2026-03-09T15:12:43.429 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:12:43 vm03 ceph-mon[51762]: min_mon_release 19 (squid) 2026-03-09T15:12:43.429 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:12:43 vm03 ceph-mon[51762]: election_strategy: 1 2026-03-09T15:12:43.429 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:12:43 vm03 ceph-mon[51762]: 0: [v2:192.168.123.103:3300/0,v1:192.168.123.103:6789/0] mon.a 2026-03-09T15:12:43.429 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:12:43 vm03 ceph-mon[51762]: fsmap 2026-03-09T15:12:43.429 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:12:43 vm03 ceph-mon[51762]: osdmap e1: 0 total, 0 up, 0 in 2026-03-09T15:12:43.429 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:12:43 vm03 ceph-mon[51762]: mgrmap e1: no daemons active 2026-03-09T15:12:43.497 INFO:teuthology.orchestra.run.vm03.stdout:Non-zero exit code 1 from systemctl reset-failed ceph-4abede80-1bca-11f1-bcd8-9b1da3c03ce5@mgr.a 2026-03-09T15:12:43.497 INFO:teuthology.orchestra.run.vm03.stdout:systemctl: stderr Failed to reset failed state of unit ceph-4abede80-1bca-11f1-bcd8-9b1da3c03ce5@mgr.a.service: Unit ceph-4abede80-1bca-11f1-bcd8-9b1da3c03ce5@mgr.a.service not loaded. 2026-03-09T15:12:43.613 INFO:teuthology.orchestra.run.vm03.stdout:systemctl: stderr Created symlink /etc/systemd/system/ceph-4abede80-1bca-11f1-bcd8-9b1da3c03ce5.target.wants/ceph-4abede80-1bca-11f1-bcd8-9b1da3c03ce5@mgr.a.service → /etc/systemd/system/ceph-4abede80-1bca-11f1-bcd8-9b1da3c03ce5@.service. 2026-03-09T15:12:43.781 INFO:teuthology.orchestra.run.vm03.stdout:firewalld does not appear to be present 2026-03-09T15:12:43.781 INFO:teuthology.orchestra.run.vm03.stdout:Not possible to enable service . firewalld.service is not available 2026-03-09T15:12:43.781 INFO:teuthology.orchestra.run.vm03.stdout:firewalld does not appear to be present 2026-03-09T15:12:43.781 INFO:teuthology.orchestra.run.vm03.stdout:Not possible to open ports <[9283, 8765]>. firewalld.service is not available 2026-03-09T15:12:43.781 INFO:teuthology.orchestra.run.vm03.stdout:Waiting for mgr to start... 2026-03-09T15:12:43.781 INFO:teuthology.orchestra.run.vm03.stdout:Waiting for mgr... 2026-03-09T15:12:43.984 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout 2026-03-09T15:12:43.984 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout { 2026-03-09T15:12:43.984 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "fsid": "4abede80-1bca-11f1-bcd8-9b1da3c03ce5", 2026-03-09T15:12:43.984 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "health": { 2026-03-09T15:12:43.984 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "status": "HEALTH_OK", 2026-03-09T15:12:43.984 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "checks": {}, 2026-03-09T15:12:43.984 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "mutes": [] 2026-03-09T15:12:43.984 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout }, 2026-03-09T15:12:43.984 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "election_epoch": 5, 2026-03-09T15:12:43.984 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "quorum": [ 2026-03-09T15:12:43.984 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout 0 2026-03-09T15:12:43.984 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout ], 2026-03-09T15:12:43.984 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "quorum_names": [ 2026-03-09T15:12:43.984 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "a" 2026-03-09T15:12:43.985 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout ], 2026-03-09T15:12:43.985 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "quorum_age": 0, 2026-03-09T15:12:43.985 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "monmap": { 2026-03-09T15:12:43.985 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "epoch": 1, 2026-03-09T15:12:43.985 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "min_mon_release_name": "squid", 2026-03-09T15:12:43.985 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "num_mons": 1 2026-03-09T15:12:43.986 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout }, 2026-03-09T15:12:43.986 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "osdmap": { 2026-03-09T15:12:43.986 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "epoch": 1, 2026-03-09T15:12:43.986 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "num_osds": 0, 2026-03-09T15:12:43.986 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "num_up_osds": 0, 2026-03-09T15:12:43.986 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "osd_up_since": 0, 2026-03-09T15:12:43.986 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "num_in_osds": 0, 2026-03-09T15:12:43.986 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "osd_in_since": 0, 2026-03-09T15:12:43.986 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "num_remapped_pgs": 0 2026-03-09T15:12:43.986 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout }, 2026-03-09T15:12:43.986 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "pgmap": { 2026-03-09T15:12:43.986 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "pgs_by_state": [], 2026-03-09T15:12:43.986 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "num_pgs": 0, 2026-03-09T15:12:43.986 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "num_pools": 0, 2026-03-09T15:12:43.986 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "num_objects": 0, 2026-03-09T15:12:43.986 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "data_bytes": 0, 2026-03-09T15:12:43.986 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "bytes_used": 0, 2026-03-09T15:12:43.986 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "bytes_avail": 0, 2026-03-09T15:12:43.986 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "bytes_total": 0 2026-03-09T15:12:43.986 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout }, 2026-03-09T15:12:43.986 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "fsmap": { 2026-03-09T15:12:43.986 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "epoch": 1, 2026-03-09T15:12:43.986 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "btime": "2026-03-09T15:12:42:233343+0000", 2026-03-09T15:12:43.986 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "by_rank": [], 2026-03-09T15:12:43.986 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "up:standby": 0 2026-03-09T15:12:43.986 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout }, 2026-03-09T15:12:43.986 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "mgrmap": { 2026-03-09T15:12:43.986 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "available": false, 2026-03-09T15:12:43.986 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "num_standbys": 0, 2026-03-09T15:12:43.986 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "modules": [ 2026-03-09T15:12:43.986 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "iostat", 2026-03-09T15:12:43.986 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "nfs", 2026-03-09T15:12:43.986 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "restful" 2026-03-09T15:12:43.986 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout ], 2026-03-09T15:12:43.986 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "services": {} 2026-03-09T15:12:43.986 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout }, 2026-03-09T15:12:43.986 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "servicemap": { 2026-03-09T15:12:43.986 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "epoch": 1, 2026-03-09T15:12:43.986 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "modified": "2026-03-09T15:12:42.233938+0000", 2026-03-09T15:12:43.986 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "services": {} 2026-03-09T15:12:43.986 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout }, 2026-03-09T15:12:43.986 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "progress_events": {} 2026-03-09T15:12:43.986 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout } 2026-03-09T15:12:43.986 INFO:teuthology.orchestra.run.vm03.stdout:mgr not available, waiting (1/15)... 2026-03-09T15:12:44.676 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:12:44 vm03 ceph-mon[51762]: from='client.? 192.168.123.103:0/756017194' entity='client.admin' 2026-03-09T15:12:44.677 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:12:44 vm03 ceph-mon[51762]: from='client.? 192.168.123.103:0/2218557366' entity='client.admin' cmd=[{"prefix": "status", "format": "json-pretty"}]: dispatch 2026-03-09T15:12:44.677 INFO:journalctl@ceph.mgr.a.vm03.stdout:Mar 09 15:12:44 vm03 ceph-4abede80-1bca-11f1-bcd8-9b1da3c03ce5-mgr-a[51970]: 2026-03-09T15:12:44.358+0000 7f5cea541140 -1 mgr[py] Module crash has missing NOTIFY_TYPES member 2026-03-09T15:12:45.050 INFO:journalctl@ceph.mgr.a.vm03.stdout:Mar 09 15:12:44 vm03 ceph-4abede80-1bca-11f1-bcd8-9b1da3c03ce5-mgr-a[51970]: 2026-03-09T15:12:44.675+0000 7f5cea541140 -1 mgr[py] Module devicehealth has missing NOTIFY_TYPES member 2026-03-09T15:12:45.050 INFO:journalctl@ceph.mgr.a.vm03.stdout:Mar 09 15:12:44 vm03 ceph-4abede80-1bca-11f1-bcd8-9b1da3c03ce5-mgr-a[51970]: /lib64/python3.9/site-packages/scipy/__init__.py:73: UserWarning: NumPy was imported from a Python sub-interpreter but NumPy does not properly support sub-interpreters. This will likely work for most users but might cause hard to track down issues or subtle bugs. A common user of the rare sub-interpreter feature is wsgi which also allows single-interpreter mode. 2026-03-09T15:12:45.050 INFO:journalctl@ceph.mgr.a.vm03.stdout:Mar 09 15:12:44 vm03 ceph-4abede80-1bca-11f1-bcd8-9b1da3c03ce5-mgr-a[51970]: Improvements in the case of bugs are welcome, but is not on the NumPy roadmap, and full support may require significant effort to achieve. 2026-03-09T15:12:45.050 INFO:journalctl@ceph.mgr.a.vm03.stdout:Mar 09 15:12:44 vm03 ceph-4abede80-1bca-11f1-bcd8-9b1da3c03ce5-mgr-a[51970]: from numpy import show_config as show_numpy_config 2026-03-09T15:12:45.050 INFO:journalctl@ceph.mgr.a.vm03.stdout:Mar 09 15:12:44 vm03 ceph-4abede80-1bca-11f1-bcd8-9b1da3c03ce5-mgr-a[51970]: 2026-03-09T15:12:44.759+0000 7f5cea541140 -1 mgr[py] Module diskprediction_local has missing NOTIFY_TYPES member 2026-03-09T15:12:45.050 INFO:journalctl@ceph.mgr.a.vm03.stdout:Mar 09 15:12:44 vm03 ceph-4abede80-1bca-11f1-bcd8-9b1da3c03ce5-mgr-a[51970]: 2026-03-09T15:12:44.793+0000 7f5cea541140 -1 mgr[py] Module influx has missing NOTIFY_TYPES member 2026-03-09T15:12:45.050 INFO:journalctl@ceph.mgr.a.vm03.stdout:Mar 09 15:12:44 vm03 ceph-4abede80-1bca-11f1-bcd8-9b1da3c03ce5-mgr-a[51970]: 2026-03-09T15:12:44.860+0000 7f5cea541140 -1 mgr[py] Module iostat has missing NOTIFY_TYPES member 2026-03-09T15:12:45.597 INFO:journalctl@ceph.mgr.a.vm03.stdout:Mar 09 15:12:45 vm03 ceph-4abede80-1bca-11f1-bcd8-9b1da3c03ce5-mgr-a[51970]: 2026-03-09T15:12:45.334+0000 7f5cea541140 -1 mgr[py] Module nfs has missing NOTIFY_TYPES member 2026-03-09T15:12:45.597 INFO:journalctl@ceph.mgr.a.vm03.stdout:Mar 09 15:12:45 vm03 ceph-4abede80-1bca-11f1-bcd8-9b1da3c03ce5-mgr-a[51970]: 2026-03-09T15:12:45.444+0000 7f5cea541140 -1 mgr[py] Module orchestrator has missing NOTIFY_TYPES member 2026-03-09T15:12:45.597 INFO:journalctl@ceph.mgr.a.vm03.stdout:Mar 09 15:12:45 vm03 ceph-4abede80-1bca-11f1-bcd8-9b1da3c03ce5-mgr-a[51970]: 2026-03-09T15:12:45.483+0000 7f5cea541140 -1 mgr[py] Module osd_perf_query has missing NOTIFY_TYPES member 2026-03-09T15:12:45.597 INFO:journalctl@ceph.mgr.a.vm03.stdout:Mar 09 15:12:45 vm03 ceph-4abede80-1bca-11f1-bcd8-9b1da3c03ce5-mgr-a[51970]: 2026-03-09T15:12:45.517+0000 7f5cea541140 -1 mgr[py] Module osd_support has missing NOTIFY_TYPES member 2026-03-09T15:12:45.597 INFO:journalctl@ceph.mgr.a.vm03.stdout:Mar 09 15:12:45 vm03 ceph-4abede80-1bca-11f1-bcd8-9b1da3c03ce5-mgr-a[51970]: 2026-03-09T15:12:45.558+0000 7f5cea541140 -1 mgr[py] Module pg_autoscaler has missing NOTIFY_TYPES member 2026-03-09T15:12:45.597 INFO:journalctl@ceph.mgr.a.vm03.stdout:Mar 09 15:12:45 vm03 ceph-4abede80-1bca-11f1-bcd8-9b1da3c03ce5-mgr-a[51970]: 2026-03-09T15:12:45.595+0000 7f5cea541140 -1 mgr[py] Module progress has missing NOTIFY_TYPES member 2026-03-09T15:12:46.024 INFO:journalctl@ceph.mgr.a.vm03.stdout:Mar 09 15:12:45 vm03 ceph-4abede80-1bca-11f1-bcd8-9b1da3c03ce5-mgr-a[51970]: 2026-03-09T15:12:45.762+0000 7f5cea541140 -1 mgr[py] Module prometheus has missing NOTIFY_TYPES member 2026-03-09T15:12:46.024 INFO:journalctl@ceph.mgr.a.vm03.stdout:Mar 09 15:12:45 vm03 ceph-4abede80-1bca-11f1-bcd8-9b1da3c03ce5-mgr-a[51970]: 2026-03-09T15:12:45.810+0000 7f5cea541140 -1 mgr[py] Module rbd_support has missing NOTIFY_TYPES member 2026-03-09T15:12:46.205 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout 2026-03-09T15:12:46.205 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout { 2026-03-09T15:12:46.205 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "fsid": "4abede80-1bca-11f1-bcd8-9b1da3c03ce5", 2026-03-09T15:12:46.205 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "health": { 2026-03-09T15:12:46.205 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "status": "HEALTH_OK", 2026-03-09T15:12:46.205 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "checks": {}, 2026-03-09T15:12:46.205 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "mutes": [] 2026-03-09T15:12:46.205 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout }, 2026-03-09T15:12:46.205 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "election_epoch": 5, 2026-03-09T15:12:46.205 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "quorum": [ 2026-03-09T15:12:46.205 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout 0 2026-03-09T15:12:46.205 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout ], 2026-03-09T15:12:46.205 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "quorum_names": [ 2026-03-09T15:12:46.205 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "a" 2026-03-09T15:12:46.205 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout ], 2026-03-09T15:12:46.205 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "quorum_age": 2, 2026-03-09T15:12:46.205 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "monmap": { 2026-03-09T15:12:46.205 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "epoch": 1, 2026-03-09T15:12:46.205 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "min_mon_release_name": "squid", 2026-03-09T15:12:46.205 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "num_mons": 1 2026-03-09T15:12:46.206 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout }, 2026-03-09T15:12:46.206 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "osdmap": { 2026-03-09T15:12:46.206 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "epoch": 1, 2026-03-09T15:12:46.206 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "num_osds": 0, 2026-03-09T15:12:46.206 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "num_up_osds": 0, 2026-03-09T15:12:46.206 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "osd_up_since": 0, 2026-03-09T15:12:46.206 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "num_in_osds": 0, 2026-03-09T15:12:46.206 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "osd_in_since": 0, 2026-03-09T15:12:46.206 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "num_remapped_pgs": 0 2026-03-09T15:12:46.206 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout }, 2026-03-09T15:12:46.206 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "pgmap": { 2026-03-09T15:12:46.206 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "pgs_by_state": [], 2026-03-09T15:12:46.206 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "num_pgs": 0, 2026-03-09T15:12:46.206 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "num_pools": 0, 2026-03-09T15:12:46.206 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "num_objects": 0, 2026-03-09T15:12:46.206 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "data_bytes": 0, 2026-03-09T15:12:46.206 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "bytes_used": 0, 2026-03-09T15:12:46.206 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "bytes_avail": 0, 2026-03-09T15:12:46.206 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "bytes_total": 0 2026-03-09T15:12:46.206 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout }, 2026-03-09T15:12:46.206 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "fsmap": { 2026-03-09T15:12:46.206 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "epoch": 1, 2026-03-09T15:12:46.206 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "btime": "2026-03-09T15:12:42:233343+0000", 2026-03-09T15:12:46.206 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "by_rank": [], 2026-03-09T15:12:46.206 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "up:standby": 0 2026-03-09T15:12:46.206 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout }, 2026-03-09T15:12:46.206 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "mgrmap": { 2026-03-09T15:12:46.206 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "available": false, 2026-03-09T15:12:46.206 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "num_standbys": 0, 2026-03-09T15:12:46.206 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "modules": [ 2026-03-09T15:12:46.206 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "iostat", 2026-03-09T15:12:46.206 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "nfs", 2026-03-09T15:12:46.206 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "restful" 2026-03-09T15:12:46.206 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout ], 2026-03-09T15:12:46.206 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "services": {} 2026-03-09T15:12:46.206 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout }, 2026-03-09T15:12:46.206 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "servicemap": { 2026-03-09T15:12:46.206 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "epoch": 1, 2026-03-09T15:12:46.206 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "modified": "2026-03-09T15:12:42.233938+0000", 2026-03-09T15:12:46.206 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "services": {} 2026-03-09T15:12:46.206 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout }, 2026-03-09T15:12:46.206 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "progress_events": {} 2026-03-09T15:12:46.206 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout } 2026-03-09T15:12:46.206 INFO:teuthology.orchestra.run.vm03.stdout:mgr not available, waiting (2/15)... 2026-03-09T15:12:46.301 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:12:46 vm03 ceph-mon[51762]: from='client.? 192.168.123.103:0/3884293640' entity='client.admin' cmd=[{"prefix": "status", "format": "json-pretty"}]: dispatch 2026-03-09T15:12:46.301 INFO:journalctl@ceph.mgr.a.vm03.stdout:Mar 09 15:12:46 vm03 ceph-4abede80-1bca-11f1-bcd8-9b1da3c03ce5-mgr-a[51970]: 2026-03-09T15:12:46.030+0000 7f5cea541140 -1 mgr[py] Module rgw has missing NOTIFY_TYPES member 2026-03-09T15:12:46.609 INFO:journalctl@ceph.mgr.a.vm03.stdout:Mar 09 15:12:46 vm03 ceph-4abede80-1bca-11f1-bcd8-9b1da3c03ce5-mgr-a[51970]: 2026-03-09T15:12:46.342+0000 7f5cea541140 -1 mgr[py] Module rook has missing NOTIFY_TYPES member 2026-03-09T15:12:46.609 INFO:journalctl@ceph.mgr.a.vm03.stdout:Mar 09 15:12:46 vm03 ceph-4abede80-1bca-11f1-bcd8-9b1da3c03ce5-mgr-a[51970]: 2026-03-09T15:12:46.380+0000 7f5cea541140 -1 mgr[py] Module selftest has missing NOTIFY_TYPES member 2026-03-09T15:12:46.609 INFO:journalctl@ceph.mgr.a.vm03.stdout:Mar 09 15:12:46 vm03 ceph-4abede80-1bca-11f1-bcd8-9b1da3c03ce5-mgr-a[51970]: 2026-03-09T15:12:46.420+0000 7f5cea541140 -1 mgr[py] Module snap_schedule has missing NOTIFY_TYPES member 2026-03-09T15:12:46.609 INFO:journalctl@ceph.mgr.a.vm03.stdout:Mar 09 15:12:46 vm03 ceph-4abede80-1bca-11f1-bcd8-9b1da3c03ce5-mgr-a[51970]: 2026-03-09T15:12:46.495+0000 7f5cea541140 -1 mgr[py] Module status has missing NOTIFY_TYPES member 2026-03-09T15:12:46.609 INFO:journalctl@ceph.mgr.a.vm03.stdout:Mar 09 15:12:46 vm03 ceph-4abede80-1bca-11f1-bcd8-9b1da3c03ce5-mgr-a[51970]: 2026-03-09T15:12:46.531+0000 7f5cea541140 -1 mgr[py] Module telegraf has missing NOTIFY_TYPES member 2026-03-09T15:12:46.886 INFO:journalctl@ceph.mgr.a.vm03.stdout:Mar 09 15:12:46 vm03 ceph-4abede80-1bca-11f1-bcd8-9b1da3c03ce5-mgr-a[51970]: 2026-03-09T15:12:46.607+0000 7f5cea541140 -1 mgr[py] Module telemetry has missing NOTIFY_TYPES member 2026-03-09T15:12:46.886 INFO:journalctl@ceph.mgr.a.vm03.stdout:Mar 09 15:12:46 vm03 ceph-4abede80-1bca-11f1-bcd8-9b1da3c03ce5-mgr-a[51970]: 2026-03-09T15:12:46.717+0000 7f5cea541140 -1 mgr[py] Module test_orchestrator has missing NOTIFY_TYPES member 2026-03-09T15:12:46.886 INFO:journalctl@ceph.mgr.a.vm03.stdout:Mar 09 15:12:46 vm03 ceph-4abede80-1bca-11f1-bcd8-9b1da3c03ce5-mgr-a[51970]: 2026-03-09T15:12:46.850+0000 7f5cea541140 -1 mgr[py] Module volumes has missing NOTIFY_TYPES member 2026-03-09T15:12:47.217 INFO:journalctl@ceph.mgr.a.vm03.stdout:Mar 09 15:12:46 vm03 ceph-4abede80-1bca-11f1-bcd8-9b1da3c03ce5-mgr-a[51970]: 2026-03-09T15:12:46.884+0000 7f5cea541140 -1 mgr[py] Module zabbix has missing NOTIFY_TYPES member 2026-03-09T15:12:47.550 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:12:47 vm03 ceph-mon[51762]: Activating manager daemon a 2026-03-09T15:12:47.550 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:12:47 vm03 ceph-mon[51762]: mgrmap e2: a(active, starting, since 0.00340831s) 2026-03-09T15:12:47.550 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:12:47 vm03 ceph-mon[51762]: from='mgr.14100 192.168.123.103:0/1455497336' entity='mgr.a' cmd=[{"prefix": "mds metadata"}]: dispatch 2026-03-09T15:12:47.550 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:12:47 vm03 ceph-mon[51762]: from='mgr.14100 192.168.123.103:0/1455497336' entity='mgr.a' cmd=[{"prefix": "osd metadata"}]: dispatch 2026-03-09T15:12:47.550 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:12:47 vm03 ceph-mon[51762]: from='mgr.14100 192.168.123.103:0/1455497336' entity='mgr.a' cmd=[{"prefix": "mon metadata"}]: dispatch 2026-03-09T15:12:47.550 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:12:47 vm03 ceph-mon[51762]: from='mgr.14100 192.168.123.103:0/1455497336' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "a"}]: dispatch 2026-03-09T15:12:47.550 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:12:47 vm03 ceph-mon[51762]: from='mgr.14100 192.168.123.103:0/1455497336' entity='mgr.a' cmd=[{"prefix": "mgr metadata", "who": "a", "id": "a"}]: dispatch 2026-03-09T15:12:47.550 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:12:47 vm03 ceph-mon[51762]: Manager daemon a is now available 2026-03-09T15:12:47.550 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:12:47 vm03 ceph-mon[51762]: from='mgr.14100 192.168.123.103:0/1455497336' entity='mgr.a' 2026-03-09T15:12:47.550 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:12:47 vm03 ceph-mon[51762]: from='mgr.14100 192.168.123.103:0/1455497336' entity='mgr.a' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/a/mirror_snapshot_schedule"}]: dispatch 2026-03-09T15:12:47.550 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:12:47 vm03 ceph-mon[51762]: from='mgr.14100 192.168.123.103:0/1455497336' entity='mgr.a' 2026-03-09T15:12:47.550 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:12:47 vm03 ceph-mon[51762]: from='mgr.14100 192.168.123.103:0/1455497336' entity='mgr.a' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/a/trash_purge_schedule"}]: dispatch 2026-03-09T15:12:47.550 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:12:47 vm03 ceph-mon[51762]: from='mgr.14100 192.168.123.103:0/1455497336' entity='mgr.a' 2026-03-09T15:12:48.498 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout 2026-03-09T15:12:48.498 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout { 2026-03-09T15:12:48.498 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "fsid": "4abede80-1bca-11f1-bcd8-9b1da3c03ce5", 2026-03-09T15:12:48.498 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "health": { 2026-03-09T15:12:48.498 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "status": "HEALTH_OK", 2026-03-09T15:12:48.498 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "checks": {}, 2026-03-09T15:12:48.498 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "mutes": [] 2026-03-09T15:12:48.499 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout }, 2026-03-09T15:12:48.499 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "election_epoch": 5, 2026-03-09T15:12:48.499 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "quorum": [ 2026-03-09T15:12:48.499 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout 0 2026-03-09T15:12:48.499 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout ], 2026-03-09T15:12:48.499 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "quorum_names": [ 2026-03-09T15:12:48.499 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "a" 2026-03-09T15:12:48.499 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout ], 2026-03-09T15:12:48.499 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "quorum_age": 5, 2026-03-09T15:12:48.499 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "monmap": { 2026-03-09T15:12:48.499 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "epoch": 1, 2026-03-09T15:12:48.499 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "min_mon_release_name": "squid", 2026-03-09T15:12:48.499 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "num_mons": 1 2026-03-09T15:12:48.499 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout }, 2026-03-09T15:12:48.499 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "osdmap": { 2026-03-09T15:12:48.499 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "epoch": 1, 2026-03-09T15:12:48.499 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "num_osds": 0, 2026-03-09T15:12:48.499 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "num_up_osds": 0, 2026-03-09T15:12:48.499 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "osd_up_since": 0, 2026-03-09T15:12:48.499 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "num_in_osds": 0, 2026-03-09T15:12:48.499 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "osd_in_since": 0, 2026-03-09T15:12:48.499 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "num_remapped_pgs": 0 2026-03-09T15:12:48.500 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout }, 2026-03-09T15:12:48.500 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "pgmap": { 2026-03-09T15:12:48.500 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "pgs_by_state": [], 2026-03-09T15:12:48.500 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "num_pgs": 0, 2026-03-09T15:12:48.500 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "num_pools": 0, 2026-03-09T15:12:48.500 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "num_objects": 0, 2026-03-09T15:12:48.500 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "data_bytes": 0, 2026-03-09T15:12:48.500 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "bytes_used": 0, 2026-03-09T15:12:48.500 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "bytes_avail": 0, 2026-03-09T15:12:48.500 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "bytes_total": 0 2026-03-09T15:12:48.500 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout }, 2026-03-09T15:12:48.500 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "fsmap": { 2026-03-09T15:12:48.500 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "epoch": 1, 2026-03-09T15:12:48.500 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "btime": "2026-03-09T15:12:42:233343+0000", 2026-03-09T15:12:48.500 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "by_rank": [], 2026-03-09T15:12:48.500 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "up:standby": 0 2026-03-09T15:12:48.500 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout }, 2026-03-09T15:12:48.500 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "mgrmap": { 2026-03-09T15:12:48.500 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "available": true, 2026-03-09T15:12:48.500 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "num_standbys": 0, 2026-03-09T15:12:48.500 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "modules": [ 2026-03-09T15:12:48.500 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "iostat", 2026-03-09T15:12:48.500 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "nfs", 2026-03-09T15:12:48.500 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "restful" 2026-03-09T15:12:48.500 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout ], 2026-03-09T15:12:48.500 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "services": {} 2026-03-09T15:12:48.500 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout }, 2026-03-09T15:12:48.500 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "servicemap": { 2026-03-09T15:12:48.500 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "epoch": 1, 2026-03-09T15:12:48.500 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "modified": "2026-03-09T15:12:42.233938+0000", 2026-03-09T15:12:48.500 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "services": {} 2026-03-09T15:12:48.500 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout }, 2026-03-09T15:12:48.500 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "progress_events": {} 2026-03-09T15:12:48.500 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout } 2026-03-09T15:12:48.500 INFO:teuthology.orchestra.run.vm03.stdout:mgr is available 2026-03-09T15:12:48.755 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout 2026-03-09T15:12:48.756 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout [global] 2026-03-09T15:12:48.756 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout fsid = 4abede80-1bca-11f1-bcd8-9b1da3c03ce5 2026-03-09T15:12:48.756 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout mon_cluster_log_file_level = debug 2026-03-09T15:12:48.756 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout mon_host = [v2:192.168.123.103:3300,v1:192.168.123.103:6789] 2026-03-09T15:12:48.756 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout mon_osd_allow_pg_remap = true 2026-03-09T15:12:48.756 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout mon_osd_allow_primary_affinity = true 2026-03-09T15:12:48.756 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout mon_warn_on_no_sortbitwise = false 2026-03-09T15:12:48.756 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout osd_crush_chooseleaf_type = 0 2026-03-09T15:12:48.756 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout 2026-03-09T15:12:48.756 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout [mgr] 2026-03-09T15:12:48.756 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout mgr/telemetry/nag = false 2026-03-09T15:12:48.756 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout 2026-03-09T15:12:48.756 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout [osd] 2026-03-09T15:12:48.756 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout osd_map_max_advance = 10 2026-03-09T15:12:48.756 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout osd_sloppy_crc = true 2026-03-09T15:12:48.756 INFO:teuthology.orchestra.run.vm03.stdout:Enabling cephadm module... 2026-03-09T15:12:49.049 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:12:48 vm03 ceph-mon[51762]: mgrmap e3: a(active, since 1.00853s) 2026-03-09T15:12:49.050 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:12:48 vm03 ceph-mon[51762]: from='client.? 192.168.123.103:0/2318789674' entity='client.admin' cmd=[{"prefix": "status", "format": "json-pretty"}]: dispatch 2026-03-09T15:12:49.050 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:12:48 vm03 ceph-mon[51762]: from='client.? 192.168.123.103:0/1926176375' entity='client.admin' cmd=[{"prefix": "config assimilate-conf"}]: dispatch 2026-03-09T15:12:49.050 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:12:48 vm03 ceph-mon[51762]: from='client.? 192.168.123.103:0/1926176375' entity='client.admin' cmd='[{"prefix": "config assimilate-conf"}]': finished 2026-03-09T15:12:50.009 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:12:49 vm03 ceph-mon[51762]: from='client.? 192.168.123.103:0/3209012445' entity='client.admin' cmd=[{"prefix": "mgr module enable", "module": "cephadm"}]: dispatch 2026-03-09T15:12:50.009 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:12:49 vm03 ceph-mon[51762]: from='client.? 192.168.123.103:0/3209012445' entity='client.admin' cmd='[{"prefix": "mgr module enable", "module": "cephadm"}]': finished 2026-03-09T15:12:50.009 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:12:49 vm03 ceph-mon[51762]: mgrmap e4: a(active, since 2s) 2026-03-09T15:12:50.009 INFO:journalctl@ceph.mgr.a.vm03.stdout:Mar 09 15:12:49 vm03 ceph-4abede80-1bca-11f1-bcd8-9b1da3c03ce5-mgr-a[51970]: ignoring --setuser ceph since I am not root 2026-03-09T15:12:50.009 INFO:journalctl@ceph.mgr.a.vm03.stdout:Mar 09 15:12:49 vm03 ceph-4abede80-1bca-11f1-bcd8-9b1da3c03ce5-mgr-a[51970]: ignoring --setgroup ceph since I am not root 2026-03-09T15:12:50.009 INFO:journalctl@ceph.mgr.a.vm03.stdout:Mar 09 15:12:49 vm03 ceph-4abede80-1bca-11f1-bcd8-9b1da3c03ce5-mgr-a[51970]: 2026-03-09T15:12:49.829+0000 7fcaa928f140 -1 mgr[py] Module alerts has missing NOTIFY_TYPES member 2026-03-09T15:12:50.009 INFO:journalctl@ceph.mgr.a.vm03.stdout:Mar 09 15:12:49 vm03 ceph-4abede80-1bca-11f1-bcd8-9b1da3c03ce5-mgr-a[51970]: 2026-03-09T15:12:49.886+0000 7fcaa928f140 -1 mgr[py] Module balancer has missing NOTIFY_TYPES member 2026-03-09T15:12:50.051 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout { 2026-03-09T15:12:50.051 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "epoch": 4, 2026-03-09T15:12:50.051 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "available": true, 2026-03-09T15:12:50.051 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "active_name": "a", 2026-03-09T15:12:50.051 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "num_standby": 0 2026-03-09T15:12:50.051 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout } 2026-03-09T15:12:50.051 INFO:teuthology.orchestra.run.vm03.stdout:Waiting for the mgr to restart... 2026-03-09T15:12:50.051 INFO:teuthology.orchestra.run.vm03.stdout:Waiting for mgr epoch 4... 2026-03-09T15:12:50.601 INFO:journalctl@ceph.mgr.a.vm03.stdout:Mar 09 15:12:50 vm03 ceph-4abede80-1bca-11f1-bcd8-9b1da3c03ce5-mgr-a[51970]: 2026-03-09T15:12:50.300+0000 7fcaa928f140 -1 mgr[py] Module crash has missing NOTIFY_TYPES member 2026-03-09T15:12:50.898 INFO:journalctl@ceph.mgr.a.vm03.stdout:Mar 09 15:12:50 vm03 ceph-4abede80-1bca-11f1-bcd8-9b1da3c03ce5-mgr-a[51970]: 2026-03-09T15:12:50.600+0000 7fcaa928f140 -1 mgr[py] Module devicehealth has missing NOTIFY_TYPES member 2026-03-09T15:12:50.898 INFO:journalctl@ceph.mgr.a.vm03.stdout:Mar 09 15:12:50 vm03 ceph-4abede80-1bca-11f1-bcd8-9b1da3c03ce5-mgr-a[51970]: /lib64/python3.9/site-packages/scipy/__init__.py:73: UserWarning: NumPy was imported from a Python sub-interpreter but NumPy does not properly support sub-interpreters. This will likely work for most users but might cause hard to track down issues or subtle bugs. A common user of the rare sub-interpreter feature is wsgi which also allows single-interpreter mode. 2026-03-09T15:12:50.899 INFO:journalctl@ceph.mgr.a.vm03.stdout:Mar 09 15:12:50 vm03 ceph-4abede80-1bca-11f1-bcd8-9b1da3c03ce5-mgr-a[51970]: Improvements in the case of bugs are welcome, but is not on the NumPy roadmap, and full support may require significant effort to achieve. 2026-03-09T15:12:50.899 INFO:journalctl@ceph.mgr.a.vm03.stdout:Mar 09 15:12:50 vm03 ceph-4abede80-1bca-11f1-bcd8-9b1da3c03ce5-mgr-a[51970]: from numpy import show_config as show_numpy_config 2026-03-09T15:12:50.899 INFO:journalctl@ceph.mgr.a.vm03.stdout:Mar 09 15:12:50 vm03 ceph-4abede80-1bca-11f1-bcd8-9b1da3c03ce5-mgr-a[51970]: 2026-03-09T15:12:50.681+0000 7fcaa928f140 -1 mgr[py] Module diskprediction_local has missing NOTIFY_TYPES member 2026-03-09T15:12:50.899 INFO:journalctl@ceph.mgr.a.vm03.stdout:Mar 09 15:12:50 vm03 ceph-4abede80-1bca-11f1-bcd8-9b1da3c03ce5-mgr-a[51970]: 2026-03-09T15:12:50.719+0000 7fcaa928f140 -1 mgr[py] Module influx has missing NOTIFY_TYPES member 2026-03-09T15:12:50.899 INFO:journalctl@ceph.mgr.a.vm03.stdout:Mar 09 15:12:50 vm03 ceph-4abede80-1bca-11f1-bcd8-9b1da3c03ce5-mgr-a[51970]: 2026-03-09T15:12:50.790+0000 7fcaa928f140 -1 mgr[py] Module iostat has missing NOTIFY_TYPES member 2026-03-09T15:12:51.259 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:12:50 vm03 ceph-mon[51762]: from='client.? 192.168.123.103:0/171617166' entity='client.admin' cmd=[{"prefix": "mgr stat"}]: dispatch 2026-03-09T15:12:51.550 INFO:journalctl@ceph.mgr.a.vm03.stdout:Mar 09 15:12:51 vm03 ceph-4abede80-1bca-11f1-bcd8-9b1da3c03ce5-mgr-a[51970]: 2026-03-09T15:12:51.257+0000 7fcaa928f140 -1 mgr[py] Module nfs has missing NOTIFY_TYPES member 2026-03-09T15:12:51.550 INFO:journalctl@ceph.mgr.a.vm03.stdout:Mar 09 15:12:51 vm03 ceph-4abede80-1bca-11f1-bcd8-9b1da3c03ce5-mgr-a[51970]: 2026-03-09T15:12:51.360+0000 7fcaa928f140 -1 mgr[py] Module orchestrator has missing NOTIFY_TYPES member 2026-03-09T15:12:51.550 INFO:journalctl@ceph.mgr.a.vm03.stdout:Mar 09 15:12:51 vm03 ceph-4abede80-1bca-11f1-bcd8-9b1da3c03ce5-mgr-a[51970]: 2026-03-09T15:12:51.397+0000 7fcaa928f140 -1 mgr[py] Module osd_perf_query has missing NOTIFY_TYPES member 2026-03-09T15:12:51.550 INFO:journalctl@ceph.mgr.a.vm03.stdout:Mar 09 15:12:51 vm03 ceph-4abede80-1bca-11f1-bcd8-9b1da3c03ce5-mgr-a[51970]: 2026-03-09T15:12:51.429+0000 7fcaa928f140 -1 mgr[py] Module osd_support has missing NOTIFY_TYPES member 2026-03-09T15:12:51.550 INFO:journalctl@ceph.mgr.a.vm03.stdout:Mar 09 15:12:51 vm03 ceph-4abede80-1bca-11f1-bcd8-9b1da3c03ce5-mgr-a[51970]: 2026-03-09T15:12:51.466+0000 7fcaa928f140 -1 mgr[py] Module pg_autoscaler has missing NOTIFY_TYPES member 2026-03-09T15:12:51.550 INFO:journalctl@ceph.mgr.a.vm03.stdout:Mar 09 15:12:51 vm03 ceph-4abede80-1bca-11f1-bcd8-9b1da3c03ce5-mgr-a[51970]: 2026-03-09T15:12:51.501+0000 7fcaa928f140 -1 mgr[py] Module progress has missing NOTIFY_TYPES member 2026-03-09T15:12:51.919 INFO:journalctl@ceph.mgr.a.vm03.stdout:Mar 09 15:12:51 vm03 ceph-4abede80-1bca-11f1-bcd8-9b1da3c03ce5-mgr-a[51970]: 2026-03-09T15:12:51.661+0000 7fcaa928f140 -1 mgr[py] Module prometheus has missing NOTIFY_TYPES member 2026-03-09T15:12:51.919 INFO:journalctl@ceph.mgr.a.vm03.stdout:Mar 09 15:12:51 vm03 ceph-4abede80-1bca-11f1-bcd8-9b1da3c03ce5-mgr-a[51970]: 2026-03-09T15:12:51.707+0000 7fcaa928f140 -1 mgr[py] Module rbd_support has missing NOTIFY_TYPES member 2026-03-09T15:12:52.187 INFO:journalctl@ceph.mgr.a.vm03.stdout:Mar 09 15:12:51 vm03 ceph-4abede80-1bca-11f1-bcd8-9b1da3c03ce5-mgr-a[51970]: 2026-03-09T15:12:51.917+0000 7fcaa928f140 -1 mgr[py] Module rgw has missing NOTIFY_TYPES member 2026-03-09T15:12:52.549 INFO:journalctl@ceph.mgr.a.vm03.stdout:Mar 09 15:12:52 vm03 ceph-4abede80-1bca-11f1-bcd8-9b1da3c03ce5-mgr-a[51970]: 2026-03-09T15:12:52.185+0000 7fcaa928f140 -1 mgr[py] Module rook has missing NOTIFY_TYPES member 2026-03-09T15:12:52.550 INFO:journalctl@ceph.mgr.a.vm03.stdout:Mar 09 15:12:52 vm03 ceph-4abede80-1bca-11f1-bcd8-9b1da3c03ce5-mgr-a[51970]: 2026-03-09T15:12:52.220+0000 7fcaa928f140 -1 mgr[py] Module selftest has missing NOTIFY_TYPES member 2026-03-09T15:12:52.550 INFO:journalctl@ceph.mgr.a.vm03.stdout:Mar 09 15:12:52 vm03 ceph-4abede80-1bca-11f1-bcd8-9b1da3c03ce5-mgr-a[51970]: 2026-03-09T15:12:52.258+0000 7fcaa928f140 -1 mgr[py] Module snap_schedule has missing NOTIFY_TYPES member 2026-03-09T15:12:52.550 INFO:journalctl@ceph.mgr.a.vm03.stdout:Mar 09 15:12:52 vm03 ceph-4abede80-1bca-11f1-bcd8-9b1da3c03ce5-mgr-a[51970]: 2026-03-09T15:12:52.330+0000 7fcaa928f140 -1 mgr[py] Module status has missing NOTIFY_TYPES member 2026-03-09T15:12:52.550 INFO:journalctl@ceph.mgr.a.vm03.stdout:Mar 09 15:12:52 vm03 ceph-4abede80-1bca-11f1-bcd8-9b1da3c03ce5-mgr-a[51970]: 2026-03-09T15:12:52.363+0000 7fcaa928f140 -1 mgr[py] Module telegraf has missing NOTIFY_TYPES member 2026-03-09T15:12:52.550 INFO:journalctl@ceph.mgr.a.vm03.stdout:Mar 09 15:12:52 vm03 ceph-4abede80-1bca-11f1-bcd8-9b1da3c03ce5-mgr-a[51970]: 2026-03-09T15:12:52.436+0000 7fcaa928f140 -1 mgr[py] Module telemetry has missing NOTIFY_TYPES member 2026-03-09T15:12:53.050 INFO:journalctl@ceph.mgr.a.vm03.stdout:Mar 09 15:12:52 vm03 ceph-4abede80-1bca-11f1-bcd8-9b1da3c03ce5-mgr-a[51970]: 2026-03-09T15:12:52.557+0000 7fcaa928f140 -1 mgr[py] Module test_orchestrator has missing NOTIFY_TYPES member 2026-03-09T15:12:53.050 INFO:journalctl@ceph.mgr.a.vm03.stdout:Mar 09 15:12:52 vm03 ceph-4abede80-1bca-11f1-bcd8-9b1da3c03ce5-mgr-a[51970]: 2026-03-09T15:12:52.698+0000 7fcaa928f140 -1 mgr[py] Module volumes has missing NOTIFY_TYPES member 2026-03-09T15:12:53.050 INFO:journalctl@ceph.mgr.a.vm03.stdout:Mar 09 15:12:52 vm03 ceph-4abede80-1bca-11f1-bcd8-9b1da3c03ce5-mgr-a[51970]: 2026-03-09T15:12:52.732+0000 7fcaa928f140 -1 mgr[py] Module zabbix has missing NOTIFY_TYPES member 2026-03-09T15:12:53.050 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:12:52 vm03 ceph-mon[51762]: Active manager daemon a restarted 2026-03-09T15:12:53.050 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:12:52 vm03 ceph-mon[51762]: Activating manager daemon a 2026-03-09T15:12:53.050 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:12:52 vm03 ceph-mon[51762]: osdmap e2: 0 total, 0 up, 0 in 2026-03-09T15:12:53.050 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:12:52 vm03 ceph-mon[51762]: mgrmap e5: a(active, starting, since 0.00565066s) 2026-03-09T15:12:53.050 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:12:52 vm03 ceph-mon[51762]: from='mgr.14118 192.168.123.103:0/4279538088' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "a"}]: dispatch 2026-03-09T15:12:53.050 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:12:52 vm03 ceph-mon[51762]: from='mgr.14118 192.168.123.103:0/4279538088' entity='mgr.a' cmd=[{"prefix": "mgr metadata", "who": "a", "id": "a"}]: dispatch 2026-03-09T15:12:53.050 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:12:52 vm03 ceph-mon[51762]: from='mgr.14118 192.168.123.103:0/4279538088' entity='mgr.a' cmd=[{"prefix": "mds metadata"}]: dispatch 2026-03-09T15:12:53.050 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:12:52 vm03 ceph-mon[51762]: from='mgr.14118 192.168.123.103:0/4279538088' entity='mgr.a' cmd=[{"prefix": "osd metadata"}]: dispatch 2026-03-09T15:12:53.050 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:12:52 vm03 ceph-mon[51762]: from='mgr.14118 192.168.123.103:0/4279538088' entity='mgr.a' cmd=[{"prefix": "mon metadata"}]: dispatch 2026-03-09T15:12:53.050 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:12:52 vm03 ceph-mon[51762]: Manager daemon a is now available 2026-03-09T15:12:53.050 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:12:52 vm03 ceph-mon[51762]: Found migration_current of "None". Setting to last migration. 2026-03-09T15:12:53.050 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:12:52 vm03 ceph-mon[51762]: from='mgr.14118 192.168.123.103:0/4279538088' entity='mgr.a' 2026-03-09T15:12:53.050 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:12:52 vm03 ceph-mon[51762]: from='mgr.14118 192.168.123.103:0/4279538088' entity='mgr.a' 2026-03-09T15:12:53.050 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:12:52 vm03 ceph-mon[51762]: from='mgr.14118 192.168.123.103:0/4279538088' entity='mgr.a' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/a/mirror_snapshot_schedule"}]: dispatch 2026-03-09T15:12:53.050 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:12:52 vm03 ceph-mon[51762]: from='mgr.14118 192.168.123.103:0/4279538088' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T15:12:53.050 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:12:52 vm03 ceph-mon[51762]: from='mgr.14118 192.168.123.103:0/4279538088' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T15:12:53.050 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:12:52 vm03 ceph-mon[51762]: from='mgr.14118 192.168.123.103:0/4279538088' entity='mgr.a' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/a/trash_purge_schedule"}]: dispatch 2026-03-09T15:12:53.777 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout { 2026-03-09T15:12:53.777 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "mgrmap_epoch": 6, 2026-03-09T15:12:53.777 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "initialized": true 2026-03-09T15:12:53.777 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout } 2026-03-09T15:12:53.777 INFO:teuthology.orchestra.run.vm03.stdout:mgr epoch 4 is available 2026-03-09T15:12:53.777 INFO:teuthology.orchestra.run.vm03.stdout:Setting orchestrator backend to cephadm... 2026-03-09T15:12:54.271 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout value unchanged 2026-03-09T15:12:54.271 INFO:teuthology.orchestra.run.vm03.stdout:Generating ssh key... 2026-03-09T15:12:54.517 INFO:journalctl@ceph.mgr.a.vm03.stdout:Mar 09 15:12:54 vm03 ceph-4abede80-1bca-11f1-bcd8-9b1da3c03ce5-mgr-a[51970]: Generating public/private ed25519 key pair. 2026-03-09T15:12:54.517 INFO:journalctl@ceph.mgr.a.vm03.stdout:Mar 09 15:12:54 vm03 ceph-4abede80-1bca-11f1-bcd8-9b1da3c03ce5-mgr-a[51970]: Your identification has been saved in /tmp/tmpfpziai5s/key 2026-03-09T15:12:54.517 INFO:journalctl@ceph.mgr.a.vm03.stdout:Mar 09 15:12:54 vm03 ceph-4abede80-1bca-11f1-bcd8-9b1da3c03ce5-mgr-a[51970]: Your public key has been saved in /tmp/tmpfpziai5s/key.pub 2026-03-09T15:12:54.517 INFO:journalctl@ceph.mgr.a.vm03.stdout:Mar 09 15:12:54 vm03 ceph-4abede80-1bca-11f1-bcd8-9b1da3c03ce5-mgr-a[51970]: The key fingerprint is: 2026-03-09T15:12:54.517 INFO:journalctl@ceph.mgr.a.vm03.stdout:Mar 09 15:12:54 vm03 ceph-4abede80-1bca-11f1-bcd8-9b1da3c03ce5-mgr-a[51970]: SHA256:KNCYo2UMv8jwBewDXUiDPyYaq0a/JSrzmJrSi5UVGmo ceph-4abede80-1bca-11f1-bcd8-9b1da3c03ce5 2026-03-09T15:12:54.517 INFO:journalctl@ceph.mgr.a.vm03.stdout:Mar 09 15:12:54 vm03 ceph-4abede80-1bca-11f1-bcd8-9b1da3c03ce5-mgr-a[51970]: The key's randomart image is: 2026-03-09T15:12:54.517 INFO:journalctl@ceph.mgr.a.vm03.stdout:Mar 09 15:12:54 vm03 ceph-4abede80-1bca-11f1-bcd8-9b1da3c03ce5-mgr-a[51970]: +--[ED25519 256]--+ 2026-03-09T15:12:54.517 INFO:journalctl@ceph.mgr.a.vm03.stdout:Mar 09 15:12:54 vm03 ceph-4abede80-1bca-11f1-bcd8-9b1da3c03ce5-mgr-a[51970]: |.=+o. | 2026-03-09T15:12:54.517 INFO:journalctl@ceph.mgr.a.vm03.stdout:Mar 09 15:12:54 vm03 ceph-4abede80-1bca-11f1-bcd8-9b1da3c03ce5-mgr-a[51970]: |o++* | 2026-03-09T15:12:54.517 INFO:journalctl@ceph.mgr.a.vm03.stdout:Mar 09 15:12:54 vm03 ceph-4abede80-1bca-11f1-bcd8-9b1da3c03ce5-mgr-a[51970]: |.+@ + | 2026-03-09T15:12:54.517 INFO:journalctl@ceph.mgr.a.vm03.stdout:Mar 09 15:12:54 vm03 ceph-4abede80-1bca-11f1-bcd8-9b1da3c03ce5-mgr-a[51970]: |=B*B . . | 2026-03-09T15:12:54.517 INFO:journalctl@ceph.mgr.a.vm03.stdout:Mar 09 15:12:54 vm03 ceph-4abede80-1bca-11f1-bcd8-9b1da3c03ce5-mgr-a[51970]: |=E+oo . S | 2026-03-09T15:12:54.517 INFO:journalctl@ceph.mgr.a.vm03.stdout:Mar 09 15:12:54 vm03 ceph-4abede80-1bca-11f1-bcd8-9b1da3c03ce5-mgr-a[51970]: |+. o . | 2026-03-09T15:12:54.517 INFO:journalctl@ceph.mgr.a.vm03.stdout:Mar 09 15:12:54 vm03 ceph-4abede80-1bca-11f1-bcd8-9b1da3c03ce5-mgr-a[51970]: |o.+. . | 2026-03-09T15:12:54.518 INFO:journalctl@ceph.mgr.a.vm03.stdout:Mar 09 15:12:54 vm03 ceph-4abede80-1bca-11f1-bcd8-9b1da3c03ce5-mgr-a[51970]: |=Bo.o | 2026-03-09T15:12:54.518 INFO:journalctl@ceph.mgr.a.vm03.stdout:Mar 09 15:12:54 vm03 ceph-4abede80-1bca-11f1-bcd8-9b1da3c03ce5-mgr-a[51970]: |@+oo. | 2026-03-09T15:12:54.518 INFO:journalctl@ceph.mgr.a.vm03.stdout:Mar 09 15:12:54 vm03 ceph-4abede80-1bca-11f1-bcd8-9b1da3c03ce5-mgr-a[51970]: +----[SHA256]-----+ 2026-03-09T15:12:54.518 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:12:54 vm03 ceph-mon[51762]: from='mgr.14118 192.168.123.103:0/4279538088' entity='mgr.a' 2026-03-09T15:12:54.518 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:12:54 vm03 ceph-mon[51762]: from='mgr.14118 192.168.123.103:0/4279538088' entity='mgr.a' 2026-03-09T15:12:54.518 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:12:54 vm03 ceph-mon[51762]: mgrmap e6: a(active, since 1.00912s) 2026-03-09T15:12:54.518 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:12:54 vm03 ceph-mon[51762]: from='client.14122 -' entity='client.admin' cmd=[{"prefix": "get_command_descriptions"}]: dispatch 2026-03-09T15:12:54.518 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:12:54 vm03 ceph-mon[51762]: from='client.14122 -' entity='client.admin' cmd=[{"prefix": "mgr_status"}]: dispatch 2026-03-09T15:12:54.518 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:12:54 vm03 ceph-mon[51762]: from='mgr.14118 192.168.123.103:0/4279538088' entity='mgr.a' 2026-03-09T15:12:54.518 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:12:54 vm03 ceph-mon[51762]: from='mgr.14118 192.168.123.103:0/4279538088' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T15:12:54.518 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:12:54 vm03 ceph-mon[51762]: from='mgr.14118 192.168.123.103:0/4279538088' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T15:12:54.783 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout ssh-ed25519 AAAAC3NzaC1lZDI1NTE5AAAAIHztYoUUgHiZBo3FQs6g84Zq5oMZLtsz+P/jQKy8iZjB ceph-4abede80-1bca-11f1-bcd8-9b1da3c03ce5 2026-03-09T15:12:54.784 INFO:teuthology.orchestra.run.vm03.stdout:Wrote public SSH key to /home/ubuntu/cephtest/ceph.pub 2026-03-09T15:12:54.784 INFO:teuthology.orchestra.run.vm03.stdout:Adding key to root@localhost authorized_keys... 2026-03-09T15:12:54.784 INFO:teuthology.orchestra.run.vm03.stdout:Adding host vm03... 2026-03-09T15:12:55.620 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:12:55 vm03 ceph-mon[51762]: from='client.14130 -' entity='client.admin' cmd=[{"prefix": "orch set backend", "module_name": "cephadm", "target": ["mon-mgr", ""]}]: dispatch 2026-03-09T15:12:55.620 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:12:55 vm03 ceph-mon[51762]: [09/Mar/2026:15:12:53] ENGINE Bus STARTING 2026-03-09T15:12:55.620 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:12:55 vm03 ceph-mon[51762]: [09/Mar/2026:15:12:54] ENGINE Serving on https://192.168.123.103:7150 2026-03-09T15:12:55.620 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:12:55 vm03 ceph-mon[51762]: [09/Mar/2026:15:12:54] ENGINE Client ('192.168.123.103', 51794) lost — peer dropped the TLS connection suddenly, during handshake: (6, 'TLS/SSL connection has been closed (EOF) (_ssl.c:1147)') 2026-03-09T15:12:55.620 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:12:55 vm03 ceph-mon[51762]: [09/Mar/2026:15:12:54] ENGINE Serving on http://192.168.123.103:8765 2026-03-09T15:12:55.620 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:12:55 vm03 ceph-mon[51762]: [09/Mar/2026:15:12:54] ENGINE Bus STARTED 2026-03-09T15:12:55.620 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:12:55 vm03 ceph-mon[51762]: from='client.14132 -' entity='client.admin' cmd=[{"prefix": "cephadm set-user", "user": "root", "target": ["mon-mgr", ""]}]: dispatch 2026-03-09T15:12:55.620 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:12:55 vm03 ceph-mon[51762]: from='client.14134 -' entity='client.admin' cmd=[{"prefix": "cephadm generate-key", "target": ["mon-mgr", ""]}]: dispatch 2026-03-09T15:12:55.620 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:12:55 vm03 ceph-mon[51762]: Generating ssh key... 2026-03-09T15:12:55.620 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:12:55 vm03 ceph-mon[51762]: from='mgr.14118 192.168.123.103:0/4279538088' entity='mgr.a' 2026-03-09T15:12:55.620 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:12:55 vm03 ceph-mon[51762]: from='mgr.14118 192.168.123.103:0/4279538088' entity='mgr.a' 2026-03-09T15:12:55.620 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:12:55 vm03 ceph-mon[51762]: from='client.14136 -' entity='client.admin' cmd=[{"prefix": "cephadm get-pub-key", "target": ["mon-mgr", ""]}]: dispatch 2026-03-09T15:12:56.528 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout Added host 'vm03' with addr '192.168.123.103' 2026-03-09T15:12:56.528 INFO:teuthology.orchestra.run.vm03.stdout:Deploying unmanaged mon service... 2026-03-09T15:12:56.774 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:12:56 vm03 ceph-mon[51762]: from='client.14138 -' entity='client.admin' cmd=[{"prefix": "orch host add", "hostname": "vm03", "addr": "192.168.123.103", "target": ["mon-mgr", ""]}]: dispatch 2026-03-09T15:12:56.774 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:12:56 vm03 ceph-mon[51762]: Deploying cephadm binary to vm03 2026-03-09T15:12:56.774 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:12:56 vm03 ceph-mon[51762]: mgrmap e7: a(active, since 2s) 2026-03-09T15:12:56.774 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:12:56 vm03 ceph-mon[51762]: from='mgr.14118 192.168.123.103:0/4279538088' entity='mgr.a' 2026-03-09T15:12:56.774 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:12:56 vm03 ceph-mon[51762]: from='mgr.14118 192.168.123.103:0/4279538088' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T15:12:56.804 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout Scheduled mon update... 2026-03-09T15:12:56.804 INFO:teuthology.orchestra.run.vm03.stdout:Deploying unmanaged mgr service... 2026-03-09T15:12:57.082 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout Scheduled mgr update... 2026-03-09T15:12:57.600 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:12:57 vm03 ceph-mon[51762]: Added host vm03 2026-03-09T15:12:57.600 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:12:57 vm03 ceph-mon[51762]: from='mgr.14118 192.168.123.103:0/4279538088' entity='mgr.a' 2026-03-09T15:12:57.600 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:12:57 vm03 ceph-mon[51762]: from='mgr.14118 192.168.123.103:0/4279538088' entity='mgr.a' 2026-03-09T15:12:57.600 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:12:57 vm03 ceph-mon[51762]: from='client.? 192.168.123.103:0/43776022' entity='client.admin' 2026-03-09T15:12:57.621 INFO:teuthology.orchestra.run.vm03.stdout:Enabling the dashboard module... 2026-03-09T15:12:59.644 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:12:59 vm03 ceph-mon[51762]: from='client.14140 -' entity='client.admin' cmd=[{"prefix": "orch apply", "service_type": "mon", "unmanaged": true, "target": ["mon-mgr", ""]}]: dispatch 2026-03-09T15:12:59.644 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:12:59 vm03 ceph-mon[51762]: Saving service mon spec with placement count:5 2026-03-09T15:12:59.644 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:12:59 vm03 ceph-mon[51762]: from='client.14142 -' entity='client.admin' cmd=[{"prefix": "orch apply", "service_type": "mgr", "unmanaged": true, "target": ["mon-mgr", ""]}]: dispatch 2026-03-09T15:12:59.644 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:12:59 vm03 ceph-mon[51762]: Saving service mgr spec with placement count:2 2026-03-09T15:12:59.644 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:12:59 vm03 ceph-mon[51762]: from='client.? 192.168.123.103:0/1213236296' entity='client.admin' 2026-03-09T15:12:59.644 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:12:59 vm03 ceph-mon[51762]: from='client.? 192.168.123.103:0/1946284992' entity='client.admin' cmd=[{"prefix": "mgr module enable", "module": "dashboard"}]: dispatch 2026-03-09T15:12:59.644 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:12:59 vm03 ceph-mon[51762]: from='mgr.14118 192.168.123.103:0/4279538088' entity='mgr.a' 2026-03-09T15:12:59.644 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:12:59 vm03 ceph-mon[51762]: from='mgr.14118 192.168.123.103:0/4279538088' entity='mgr.a' 2026-03-09T15:12:59.644 INFO:journalctl@ceph.mgr.a.vm03.stdout:Mar 09 15:12:59 vm03 ceph-4abede80-1bca-11f1-bcd8-9b1da3c03ce5-mgr-a[51970]: ignoring --setuser ceph since I am not root 2026-03-09T15:12:59.644 INFO:journalctl@ceph.mgr.a.vm03.stdout:Mar 09 15:12:59 vm03 ceph-4abede80-1bca-11f1-bcd8-9b1da3c03ce5-mgr-a[51970]: ignoring --setgroup ceph since I am not root 2026-03-09T15:12:59.644 INFO:journalctl@ceph.mgr.a.vm03.stdout:Mar 09 15:12:59 vm03 ceph-4abede80-1bca-11f1-bcd8-9b1da3c03ce5-mgr-a[51970]: 2026-03-09T15:12:59.590+0000 7fa580bc4140 -1 mgr[py] Module alerts has missing NOTIFY_TYPES member 2026-03-09T15:12:59.644 INFO:journalctl@ceph.mgr.a.vm03.stdout:Mar 09 15:12:59 vm03 ceph-4abede80-1bca-11f1-bcd8-9b1da3c03ce5-mgr-a[51970]: 2026-03-09T15:12:59.642+0000 7fa580bc4140 -1 mgr[py] Module balancer has missing NOTIFY_TYPES member 2026-03-09T15:12:59.851 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout { 2026-03-09T15:12:59.851 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "epoch": 8, 2026-03-09T15:12:59.851 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "available": true, 2026-03-09T15:12:59.851 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "active_name": "a", 2026-03-09T15:12:59.851 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "num_standby": 0 2026-03-09T15:12:59.851 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout } 2026-03-09T15:12:59.851 INFO:teuthology.orchestra.run.vm03.stdout:Waiting for the mgr to restart... 2026-03-09T15:12:59.851 INFO:teuthology.orchestra.run.vm03.stdout:Waiting for mgr epoch 8... 2026-03-09T15:13:00.300 INFO:journalctl@ceph.mgr.a.vm03.stdout:Mar 09 15:13:00 vm03 ceph-4abede80-1bca-11f1-bcd8-9b1da3c03ce5-mgr-a[51970]: 2026-03-09T15:13:00.037+0000 7fa580bc4140 -1 mgr[py] Module crash has missing NOTIFY_TYPES member 2026-03-09T15:13:00.799 INFO:journalctl@ceph.mgr.a.vm03.stdout:Mar 09 15:13:00 vm03 ceph-4abede80-1bca-11f1-bcd8-9b1da3c03ce5-mgr-a[51970]: 2026-03-09T15:13:00.347+0000 7fa580bc4140 -1 mgr[py] Module devicehealth has missing NOTIFY_TYPES member 2026-03-09T15:13:00.800 INFO:journalctl@ceph.mgr.a.vm03.stdout:Mar 09 15:13:00 vm03 ceph-4abede80-1bca-11f1-bcd8-9b1da3c03ce5-mgr-a[51970]: /lib64/python3.9/site-packages/scipy/__init__.py:73: UserWarning: NumPy was imported from a Python sub-interpreter but NumPy does not properly support sub-interpreters. This will likely work for most users but might cause hard to track down issues or subtle bugs. A common user of the rare sub-interpreter feature is wsgi which also allows single-interpreter mode. 2026-03-09T15:13:00.800 INFO:journalctl@ceph.mgr.a.vm03.stdout:Mar 09 15:13:00 vm03 ceph-4abede80-1bca-11f1-bcd8-9b1da3c03ce5-mgr-a[51970]: Improvements in the case of bugs are welcome, but is not on the NumPy roadmap, and full support may require significant effort to achieve. 2026-03-09T15:13:00.800 INFO:journalctl@ceph.mgr.a.vm03.stdout:Mar 09 15:13:00 vm03 ceph-4abede80-1bca-11f1-bcd8-9b1da3c03ce5-mgr-a[51970]: from numpy import show_config as show_numpy_config 2026-03-09T15:13:00.800 INFO:journalctl@ceph.mgr.a.vm03.stdout:Mar 09 15:13:00 vm03 ceph-4abede80-1bca-11f1-bcd8-9b1da3c03ce5-mgr-a[51970]: 2026-03-09T15:13:00.432+0000 7fa580bc4140 -1 mgr[py] Module diskprediction_local has missing NOTIFY_TYPES member 2026-03-09T15:13:00.800 INFO:journalctl@ceph.mgr.a.vm03.stdout:Mar 09 15:13:00 vm03 ceph-4abede80-1bca-11f1-bcd8-9b1da3c03ce5-mgr-a[51970]: 2026-03-09T15:13:00.469+0000 7fa580bc4140 -1 mgr[py] Module influx has missing NOTIFY_TYPES member 2026-03-09T15:13:00.800 INFO:journalctl@ceph.mgr.a.vm03.stdout:Mar 09 15:13:00 vm03 ceph-4abede80-1bca-11f1-bcd8-9b1da3c03ce5-mgr-a[51970]: 2026-03-09T15:13:00.540+0000 7fa580bc4140 -1 mgr[py] Module iostat has missing NOTIFY_TYPES member 2026-03-09T15:13:00.800 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:13:00 vm03 ceph-mon[51762]: from='client.? 192.168.123.103:0/1946284992' entity='client.admin' cmd='[{"prefix": "mgr module enable", "module": "dashboard"}]': finished 2026-03-09T15:13:00.800 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:13:00 vm03 ceph-mon[51762]: mgrmap e8: a(active, since 6s) 2026-03-09T15:13:00.800 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:13:00 vm03 ceph-mon[51762]: from='client.? 192.168.123.103:0/1206920631' entity='client.admin' cmd=[{"prefix": "mgr stat"}]: dispatch 2026-03-09T15:13:01.300 INFO:journalctl@ceph.mgr.a.vm03.stdout:Mar 09 15:13:01 vm03 ceph-4abede80-1bca-11f1-bcd8-9b1da3c03ce5-mgr-a[51970]: 2026-03-09T15:13:01.016+0000 7fa580bc4140 -1 mgr[py] Module nfs has missing NOTIFY_TYPES member 2026-03-09T15:13:01.300 INFO:journalctl@ceph.mgr.a.vm03.stdout:Mar 09 15:13:01 vm03 ceph-4abede80-1bca-11f1-bcd8-9b1da3c03ce5-mgr-a[51970]: 2026-03-09T15:13:01.121+0000 7fa580bc4140 -1 mgr[py] Module orchestrator has missing NOTIFY_TYPES member 2026-03-09T15:13:01.300 INFO:journalctl@ceph.mgr.a.vm03.stdout:Mar 09 15:13:01 vm03 ceph-4abede80-1bca-11f1-bcd8-9b1da3c03ce5-mgr-a[51970]: 2026-03-09T15:13:01.159+0000 7fa580bc4140 -1 mgr[py] Module osd_perf_query has missing NOTIFY_TYPES member 2026-03-09T15:13:01.300 INFO:journalctl@ceph.mgr.a.vm03.stdout:Mar 09 15:13:01 vm03 ceph-4abede80-1bca-11f1-bcd8-9b1da3c03ce5-mgr-a[51970]: 2026-03-09T15:13:01.192+0000 7fa580bc4140 -1 mgr[py] Module osd_support has missing NOTIFY_TYPES member 2026-03-09T15:13:01.300 INFO:journalctl@ceph.mgr.a.vm03.stdout:Mar 09 15:13:01 vm03 ceph-4abede80-1bca-11f1-bcd8-9b1da3c03ce5-mgr-a[51970]: 2026-03-09T15:13:01.230+0000 7fa580bc4140 -1 mgr[py] Module pg_autoscaler has missing NOTIFY_TYPES member 2026-03-09T15:13:01.300 INFO:journalctl@ceph.mgr.a.vm03.stdout:Mar 09 15:13:01 vm03 ceph-4abede80-1bca-11f1-bcd8-9b1da3c03ce5-mgr-a[51970]: 2026-03-09T15:13:01.265+0000 7fa580bc4140 -1 mgr[py] Module progress has missing NOTIFY_TYPES member 2026-03-09T15:13:01.680 INFO:journalctl@ceph.mgr.a.vm03.stdout:Mar 09 15:13:01 vm03 ceph-4abede80-1bca-11f1-bcd8-9b1da3c03ce5-mgr-a[51970]: 2026-03-09T15:13:01.424+0000 7fa580bc4140 -1 mgr[py] Module prometheus has missing NOTIFY_TYPES member 2026-03-09T15:13:01.680 INFO:journalctl@ceph.mgr.a.vm03.stdout:Mar 09 15:13:01 vm03 ceph-4abede80-1bca-11f1-bcd8-9b1da3c03ce5-mgr-a[51970]: 2026-03-09T15:13:01.472+0000 7fa580bc4140 -1 mgr[py] Module rbd_support has missing NOTIFY_TYPES member 2026-03-09T15:13:01.944 INFO:journalctl@ceph.mgr.a.vm03.stdout:Mar 09 15:13:01 vm03 ceph-4abede80-1bca-11f1-bcd8-9b1da3c03ce5-mgr-a[51970]: 2026-03-09T15:13:01.680+0000 7fa580bc4140 -1 mgr[py] Module rgw has missing NOTIFY_TYPES member 2026-03-09T15:13:02.210 INFO:journalctl@ceph.mgr.a.vm03.stdout:Mar 09 15:13:01 vm03 ceph-4abede80-1bca-11f1-bcd8-9b1da3c03ce5-mgr-a[51970]: 2026-03-09T15:13:01.943+0000 7fa580bc4140 -1 mgr[py] Module rook has missing NOTIFY_TYPES member 2026-03-09T15:13:02.210 INFO:journalctl@ceph.mgr.a.vm03.stdout:Mar 09 15:13:01 vm03 ceph-4abede80-1bca-11f1-bcd8-9b1da3c03ce5-mgr-a[51970]: 2026-03-09T15:13:01.980+0000 7fa580bc4140 -1 mgr[py] Module selftest has missing NOTIFY_TYPES member 2026-03-09T15:13:02.210 INFO:journalctl@ceph.mgr.a.vm03.stdout:Mar 09 15:13:02 vm03 ceph-4abede80-1bca-11f1-bcd8-9b1da3c03ce5-mgr-a[51970]: 2026-03-09T15:13:02.020+0000 7fa580bc4140 -1 mgr[py] Module snap_schedule has missing NOTIFY_TYPES member 2026-03-09T15:13:02.210 INFO:journalctl@ceph.mgr.a.vm03.stdout:Mar 09 15:13:02 vm03 ceph-4abede80-1bca-11f1-bcd8-9b1da3c03ce5-mgr-a[51970]: 2026-03-09T15:13:02.096+0000 7fa580bc4140 -1 mgr[py] Module status has missing NOTIFY_TYPES member 2026-03-09T15:13:02.210 INFO:journalctl@ceph.mgr.a.vm03.stdout:Mar 09 15:13:02 vm03 ceph-4abede80-1bca-11f1-bcd8-9b1da3c03ce5-mgr-a[51970]: 2026-03-09T15:13:02.132+0000 7fa580bc4140 -1 mgr[py] Module telegraf has missing NOTIFY_TYPES member 2026-03-09T15:13:02.476 INFO:journalctl@ceph.mgr.a.vm03.stdout:Mar 09 15:13:02 vm03 ceph-4abede80-1bca-11f1-bcd8-9b1da3c03ce5-mgr-a[51970]: 2026-03-09T15:13:02.209+0000 7fa580bc4140 -1 mgr[py] Module telemetry has missing NOTIFY_TYPES member 2026-03-09T15:13:02.476 INFO:journalctl@ceph.mgr.a.vm03.stdout:Mar 09 15:13:02 vm03 ceph-4abede80-1bca-11f1-bcd8-9b1da3c03ce5-mgr-a[51970]: 2026-03-09T15:13:02.315+0000 7fa580bc4140 -1 mgr[py] Module test_orchestrator has missing NOTIFY_TYPES member 2026-03-09T15:13:02.476 INFO:journalctl@ceph.mgr.a.vm03.stdout:Mar 09 15:13:02 vm03 ceph-4abede80-1bca-11f1-bcd8-9b1da3c03ce5-mgr-a[51970]: 2026-03-09T15:13:02.442+0000 7fa580bc4140 -1 mgr[py] Module volumes has missing NOTIFY_TYPES member 2026-03-09T15:13:02.800 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:13:02 vm03 ceph-mon[51762]: Active manager daemon a restarted 2026-03-09T15:13:02.800 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:13:02 vm03 ceph-mon[51762]: Activating manager daemon a 2026-03-09T15:13:02.800 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:13:02 vm03 ceph-mon[51762]: osdmap e3: 0 total, 0 up, 0 in 2026-03-09T15:13:02.800 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:13:02 vm03 ceph-mon[51762]: mgrmap e9: a(active, starting, since 0.00641125s) 2026-03-09T15:13:02.800 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:13:02 vm03 ceph-mon[51762]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "a"}]: dispatch 2026-03-09T15:13:02.800 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:13:02 vm03 ceph-mon[51762]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' cmd=[{"prefix": "mgr metadata", "who": "a", "id": "a"}]: dispatch 2026-03-09T15:13:02.800 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:13:02 vm03 ceph-mon[51762]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' cmd=[{"prefix": "mds metadata"}]: dispatch 2026-03-09T15:13:02.800 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:13:02 vm03 ceph-mon[51762]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' cmd=[{"prefix": "osd metadata"}]: dispatch 2026-03-09T15:13:02.800 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:13:02 vm03 ceph-mon[51762]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' cmd=[{"prefix": "mon metadata"}]: dispatch 2026-03-09T15:13:02.800 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:13:02 vm03 ceph-mon[51762]: Manager daemon a is now available 2026-03-09T15:13:02.800 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:13:02 vm03 ceph-mon[51762]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T15:13:02.800 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:13:02 vm03 ceph-mon[51762]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/a/mirror_snapshot_schedule"}]: dispatch 2026-03-09T15:13:02.800 INFO:journalctl@ceph.mgr.a.vm03.stdout:Mar 09 15:13:02 vm03 ceph-4abede80-1bca-11f1-bcd8-9b1da3c03ce5-mgr-a[51970]: 2026-03-09T15:13:02.476+0000 7fa580bc4140 -1 mgr[py] Module zabbix has missing NOTIFY_TYPES member 2026-03-09T15:13:03.543 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout { 2026-03-09T15:13:03.544 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "mgrmap_epoch": 10, 2026-03-09T15:13:03.544 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "initialized": true 2026-03-09T15:13:03.544 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout } 2026-03-09T15:13:03.544 INFO:teuthology.orchestra.run.vm03.stdout:mgr epoch 8 is available 2026-03-09T15:13:03.544 INFO:teuthology.orchestra.run.vm03.stdout:Generating a dashboard self-signed certificate... 2026-03-09T15:13:03.776 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:13:03 vm03 ceph-mon[51762]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/a/trash_purge_schedule"}]: dispatch 2026-03-09T15:13:03.776 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:13:03 vm03 ceph-mon[51762]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' 2026-03-09T15:13:03.776 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:13:03 vm03 ceph-mon[51762]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' 2026-03-09T15:13:03.776 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:13:03 vm03 ceph-mon[51762]: from='client.14154 -' entity='client.admin' cmd=[{"prefix": "get_command_descriptions"}]: dispatch 2026-03-09T15:13:03.776 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:13:03 vm03 ceph-mon[51762]: mgrmap e10: a(active, since 1.01224s) 2026-03-09T15:13:03.776 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:13:03 vm03 ceph-mon[51762]: from='client.14154 -' entity='client.admin' cmd=[{"prefix": "mgr_status"}]: dispatch 2026-03-09T15:13:03.967 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout Self-signed certificate created 2026-03-09T15:13:03.967 INFO:teuthology.orchestra.run.vm03.stdout:Creating initial admin user... 2026-03-09T15:13:04.354 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout {"username": "admin", "password": "$2b$12$gWlBPQvMQ/3eWhXp.kPn6uXpT0cjNv0.xWsQw/Tpq5j/gZWT6trD2", "roles": ["administrator"], "name": null, "email": null, "lastUpdate": 1773069184, "enabled": true, "pwdExpirationDate": null, "pwdUpdateRequired": true} 2026-03-09T15:13:04.354 INFO:teuthology.orchestra.run.vm03.stdout:Fetching dashboard port number... 2026-03-09T15:13:04.607 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout 8443 2026-03-09T15:13:04.607 INFO:teuthology.orchestra.run.vm03.stdout:firewalld does not appear to be present 2026-03-09T15:13:04.607 INFO:teuthology.orchestra.run.vm03.stdout:Not possible to open ports <[8443]>. firewalld.service is not available 2026-03-09T15:13:04.608 INFO:teuthology.orchestra.run.vm03.stdout:Ceph Dashboard is now available at: 2026-03-09T15:13:04.608 INFO:teuthology.orchestra.run.vm03.stdout: 2026-03-09T15:13:04.608 INFO:teuthology.orchestra.run.vm03.stdout: URL: https://vm03.local:8443/ 2026-03-09T15:13:04.608 INFO:teuthology.orchestra.run.vm03.stdout: User: admin 2026-03-09T15:13:04.608 INFO:teuthology.orchestra.run.vm03.stdout: Password: baz8p7r1so 2026-03-09T15:13:04.608 INFO:teuthology.orchestra.run.vm03.stdout: 2026-03-09T15:13:04.608 INFO:teuthology.orchestra.run.vm03.stdout:Saving cluster configuration to /var/lib/ceph/4abede80-1bca-11f1-bcd8-9b1da3c03ce5/config directory 2026-03-09T15:13:04.939 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:13:04 vm03 ceph-mon[51762]: [09/Mar/2026:15:13:03] ENGINE Bus STARTING 2026-03-09T15:13:05.025 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stderr set mgr/dashboard/cluster/status 2026-03-09T15:13:05.025 INFO:teuthology.orchestra.run.vm03.stdout:You can access the Ceph CLI as following in case of multi-cluster or non-default config: 2026-03-09T15:13:05.025 INFO:teuthology.orchestra.run.vm03.stdout: 2026-03-09T15:13:05.025 INFO:teuthology.orchestra.run.vm03.stdout: sudo /home/ubuntu/cephtest/cephadm shell --fsid 4abede80-1bca-11f1-bcd8-9b1da3c03ce5 -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring 2026-03-09T15:13:05.025 INFO:teuthology.orchestra.run.vm03.stdout: 2026-03-09T15:13:05.025 INFO:teuthology.orchestra.run.vm03.stdout:Or, if you are only running a single cluster on this host: 2026-03-09T15:13:05.025 INFO:teuthology.orchestra.run.vm03.stdout: 2026-03-09T15:13:05.025 INFO:teuthology.orchestra.run.vm03.stdout: sudo /home/ubuntu/cephtest/cephadm shell 2026-03-09T15:13:05.025 INFO:teuthology.orchestra.run.vm03.stdout: 2026-03-09T15:13:05.025 INFO:teuthology.orchestra.run.vm03.stdout:Please consider enabling telemetry to help improve Ceph: 2026-03-09T15:13:05.025 INFO:teuthology.orchestra.run.vm03.stdout: 2026-03-09T15:13:05.025 INFO:teuthology.orchestra.run.vm03.stdout: ceph telemetry on 2026-03-09T15:13:05.025 INFO:teuthology.orchestra.run.vm03.stdout: 2026-03-09T15:13:05.025 INFO:teuthology.orchestra.run.vm03.stdout:For more information see: 2026-03-09T15:13:05.025 INFO:teuthology.orchestra.run.vm03.stdout: 2026-03-09T15:13:05.025 INFO:teuthology.orchestra.run.vm03.stdout: https://docs.ceph.com/en/latest/mgr/telemetry/ 2026-03-09T15:13:05.025 INFO:teuthology.orchestra.run.vm03.stdout: 2026-03-09T15:13:05.025 INFO:teuthology.orchestra.run.vm03.stdout:Bootstrap complete. 2026-03-09T15:13:05.057 INFO:tasks.cephadm:Fetching config... 2026-03-09T15:13:05.057 DEBUG:teuthology.orchestra.run.vm03:> set -ex 2026-03-09T15:13:05.057 DEBUG:teuthology.orchestra.run.vm03:> dd if=/etc/ceph/ceph.conf of=/dev/stdout 2026-03-09T15:13:05.076 INFO:tasks.cephadm:Fetching client.admin keyring... 2026-03-09T15:13:05.076 DEBUG:teuthology.orchestra.run.vm03:> set -ex 2026-03-09T15:13:05.076 DEBUG:teuthology.orchestra.run.vm03:> dd if=/etc/ceph/ceph.client.admin.keyring of=/dev/stdout 2026-03-09T15:13:05.144 INFO:tasks.cephadm:Fetching mon keyring... 2026-03-09T15:13:05.145 DEBUG:teuthology.orchestra.run.vm03:> set -ex 2026-03-09T15:13:05.145 DEBUG:teuthology.orchestra.run.vm03:> sudo dd if=/var/lib/ceph/4abede80-1bca-11f1-bcd8-9b1da3c03ce5/mon.a/keyring of=/dev/stdout 2026-03-09T15:13:05.205 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:13:04 vm03 ceph-mon[51762]: [09/Mar/2026:15:13:03] ENGINE Serving on https://192.168.123.103:7150 2026-03-09T15:13:05.206 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:13:04 vm03 ceph-mon[51762]: [09/Mar/2026:15:13:03] ENGINE Client ('192.168.123.103', 34762) lost — peer dropped the TLS connection suddenly, during handshake: (6, 'TLS/SSL connection has been closed (EOF) (_ssl.c:1147)') 2026-03-09T15:13:05.206 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:13:04 vm03 ceph-mon[51762]: from='client.14162 -' entity='client.admin' cmd=[{"prefix": "dashboard create-self-signed-cert", "target": ["mon-mgr", ""]}]: dispatch 2026-03-09T15:13:05.206 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:13:04 vm03 ceph-mon[51762]: [09/Mar/2026:15:13:03] ENGINE Serving on http://192.168.123.103:8765 2026-03-09T15:13:05.206 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:13:04 vm03 ceph-mon[51762]: [09/Mar/2026:15:13:03] ENGINE Bus STARTED 2026-03-09T15:13:05.206 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:13:04 vm03 ceph-mon[51762]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' 2026-03-09T15:13:05.206 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:13:04 vm03 ceph-mon[51762]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' 2026-03-09T15:13:05.206 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:13:04 vm03 ceph-mon[51762]: from='client.14164 -' entity='client.admin' cmd=[{"prefix": "dashboard ac-user-create", "username": "admin", "rolename": "administrator", "force_password": true, "pwd_update_required": true, "target": ["mon-mgr", ""]}]: dispatch 2026-03-09T15:13:05.206 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:13:04 vm03 ceph-mon[51762]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' 2026-03-09T15:13:05.206 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:13:04 vm03 ceph-mon[51762]: from='client.? 192.168.123.103:0/3023107133' entity='client.admin' cmd=[{"prefix": "config get", "who": "mgr", "key": "mgr/dashboard/ssl_server_port"}]: dispatch 2026-03-09T15:13:05.213 INFO:tasks.cephadm:Fetching pub ssh key... 2026-03-09T15:13:05.213 DEBUG:teuthology.orchestra.run.vm03:> set -ex 2026-03-09T15:13:05.213 DEBUG:teuthology.orchestra.run.vm03:> dd if=/home/ubuntu/cephtest/ceph.pub of=/dev/stdout 2026-03-09T15:13:05.271 INFO:tasks.cephadm:Installing pub ssh key for root users... 2026-03-09T15:13:05.271 DEBUG:teuthology.orchestra.run.vm03:> sudo install -d -m 0700 /root/.ssh && echo 'ssh-ed25519 AAAAC3NzaC1lZDI1NTE5AAAAIHztYoUUgHiZBo3FQs6g84Zq5oMZLtsz+P/jQKy8iZjB ceph-4abede80-1bca-11f1-bcd8-9b1da3c03ce5' | sudo tee -a /root/.ssh/authorized_keys && sudo chmod 0600 /root/.ssh/authorized_keys 2026-03-09T15:13:05.368 INFO:teuthology.orchestra.run.vm03.stdout:ssh-ed25519 AAAAC3NzaC1lZDI1NTE5AAAAIHztYoUUgHiZBo3FQs6g84Zq5oMZLtsz+P/jQKy8iZjB ceph-4abede80-1bca-11f1-bcd8-9b1da3c03ce5 2026-03-09T15:13:05.378 DEBUG:teuthology.orchestra.run.vm04:> sudo install -d -m 0700 /root/.ssh && echo 'ssh-ed25519 AAAAC3NzaC1lZDI1NTE5AAAAIHztYoUUgHiZBo3FQs6g84Zq5oMZLtsz+P/jQKy8iZjB ceph-4abede80-1bca-11f1-bcd8-9b1da3c03ce5' | sudo tee -a /root/.ssh/authorized_keys && sudo chmod 0600 /root/.ssh/authorized_keys 2026-03-09T15:13:05.409 INFO:teuthology.orchestra.run.vm04.stdout:ssh-ed25519 AAAAC3NzaC1lZDI1NTE5AAAAIHztYoUUgHiZBo3FQs6g84Zq5oMZLtsz+P/jQKy8iZjB ceph-4abede80-1bca-11f1-bcd8-9b1da3c03ce5 2026-03-09T15:13:05.419 DEBUG:teuthology.orchestra.run.vm06:> sudo install -d -m 0700 /root/.ssh && echo 'ssh-ed25519 AAAAC3NzaC1lZDI1NTE5AAAAIHztYoUUgHiZBo3FQs6g84Zq5oMZLtsz+P/jQKy8iZjB ceph-4abede80-1bca-11f1-bcd8-9b1da3c03ce5' | sudo tee -a /root/.ssh/authorized_keys && sudo chmod 0600 /root/.ssh/authorized_keys 2026-03-09T15:13:05.455 INFO:teuthology.orchestra.run.vm06.stdout:ssh-ed25519 AAAAC3NzaC1lZDI1NTE5AAAAIHztYoUUgHiZBo3FQs6g84Zq5oMZLtsz+P/jQKy8iZjB ceph-4abede80-1bca-11f1-bcd8-9b1da3c03ce5 2026-03-09T15:13:05.466 DEBUG:teuthology.orchestra.run.vm03:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 4abede80-1bca-11f1-bcd8-9b1da3c03ce5 -- ceph config set mgr mgr/cephadm/allow_ptrace true 2026-03-09T15:13:05.631 INFO:teuthology.orchestra.run.vm03.stderr:Inferring config /var/lib/ceph/4abede80-1bca-11f1-bcd8-9b1da3c03ce5/mon.a/config 2026-03-09T15:13:05.978 INFO:tasks.cephadm:Distributing conf and client.admin keyring to all hosts + 0755 2026-03-09T15:13:05.978 DEBUG:teuthology.orchestra.run.vm03:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 4abede80-1bca-11f1-bcd8-9b1da3c03ce5 -- ceph orch client-keyring set client.admin '*' --mode 0755 2026-03-09T15:13:06.003 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:13:05 vm03 ceph-mon[51762]: from='client.? 192.168.123.103:0/1982347819' entity='client.admin' 2026-03-09T15:13:06.004 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:13:05 vm03 ceph-mon[51762]: mgrmap e11: a(active, since 2s) 2026-03-09T15:13:06.004 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:13:05 vm03 ceph-mon[51762]: from='client.? 192.168.123.103:0/101570896' entity='client.admin' 2026-03-09T15:13:06.174 INFO:teuthology.orchestra.run.vm03.stderr:Inferring config /var/lib/ceph/4abede80-1bca-11f1-bcd8-9b1da3c03ce5/mon.a/config 2026-03-09T15:13:06.483 INFO:tasks.cephadm:Writing (initial) conf and keyring to vm04 2026-03-09T15:13:06.483 DEBUG:teuthology.orchestra.run.vm04:> set -ex 2026-03-09T15:13:06.483 DEBUG:teuthology.orchestra.run.vm04:> dd of=/etc/ceph/ceph.conf 2026-03-09T15:13:06.497 DEBUG:teuthology.orchestra.run.vm04:> set -ex 2026-03-09T15:13:06.497 DEBUG:teuthology.orchestra.run.vm04:> dd of=/etc/ceph/ceph.client.admin.keyring 2026-03-09T15:13:06.551 INFO:tasks.cephadm:Adding host vm04 to orchestrator... 2026-03-09T15:13:06.552 DEBUG:teuthology.orchestra.run.vm03:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 4abede80-1bca-11f1-bcd8-9b1da3c03ce5 -- ceph orch host add vm04 2026-03-09T15:13:06.724 INFO:teuthology.orchestra.run.vm03.stderr:Inferring config /var/lib/ceph/4abede80-1bca-11f1-bcd8-9b1da3c03ce5/mon.a/config 2026-03-09T15:13:07.102 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:13:07 vm03 ceph-mon[51762]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' 2026-03-09T15:13:07.102 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:13:07 vm03 ceph-mon[51762]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' 2026-03-09T15:13:07.102 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:13:07 vm03 ceph-mon[51762]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' cmd=[{"prefix": "config rm", "who": "osd/host:vm03", "name": "osd_memory_target"}]: dispatch 2026-03-09T15:13:07.102 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:13:07 vm03 ceph-mon[51762]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' 2026-03-09T15:13:07.102 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:13:07 vm03 ceph-mon[51762]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T15:13:07.102 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:13:07 vm03 ceph-mon[51762]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' 2026-03-09T15:13:07.103 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:13:07 vm03 ceph-mon[51762]: from='client.14172 -' entity='client.admin' cmd=[{"prefix": "orch client-keyring set", "entity": "client.admin", "placement": "*", "mode": "0755", "target": ["mon-mgr", ""]}]: dispatch 2026-03-09T15:13:07.103 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:13:07 vm03 ceph-mon[51762]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' 2026-03-09T15:13:07.103 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:13:07 vm03 ceph-mon[51762]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T15:13:07.103 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:13:07 vm03 ceph-mon[51762]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T15:13:07.103 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:13:07 vm03 ceph-mon[51762]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T15:13:07.103 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:13:07 vm03 ceph-mon[51762]: Updating vm03:/etc/ceph/ceph.conf 2026-03-09T15:13:07.103 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:13:07 vm03 ceph-mon[51762]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' 2026-03-09T15:13:07.103 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:13:07 vm03 ceph-mon[51762]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' 2026-03-09T15:13:07.103 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:13:07 vm03 ceph-mon[51762]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' 2026-03-09T15:13:08.521 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:13:08 vm03 ceph-mon[51762]: Updating vm03:/var/lib/ceph/4abede80-1bca-11f1-bcd8-9b1da3c03ce5/config/ceph.conf 2026-03-09T15:13:08.522 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:13:08 vm03 ceph-mon[51762]: Updating vm03:/etc/ceph/ceph.client.admin.keyring 2026-03-09T15:13:08.522 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:13:08 vm03 ceph-mon[51762]: Updating vm03:/var/lib/ceph/4abede80-1bca-11f1-bcd8-9b1da3c03ce5/config/ceph.client.admin.keyring 2026-03-09T15:13:08.522 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:13:08 vm03 ceph-mon[51762]: from='client.14174 -' entity='client.admin' cmd=[{"prefix": "orch host add", "hostname": "vm04", "target": ["mon-mgr", ""]}]: dispatch 2026-03-09T15:13:08.522 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:13:08 vm03 ceph-mon[51762]: Deploying cephadm binary to vm04 2026-03-09T15:13:08.522 INFO:teuthology.orchestra.run.vm03.stdout:Added host 'vm04' with addr '192.168.123.104' 2026-03-09T15:13:08.573 DEBUG:teuthology.orchestra.run.vm03:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 4abede80-1bca-11f1-bcd8-9b1da3c03ce5 -- ceph orch host ls --format=json 2026-03-09T15:13:08.739 INFO:teuthology.orchestra.run.vm03.stderr:Inferring config /var/lib/ceph/4abede80-1bca-11f1-bcd8-9b1da3c03ce5/mon.a/config 2026-03-09T15:13:08.958 INFO:teuthology.orchestra.run.vm03.stdout: 2026-03-09T15:13:08.958 INFO:teuthology.orchestra.run.vm03.stdout:[{"addr": "192.168.123.103", "hostname": "vm03", "labels": [], "status": ""}, {"addr": "192.168.123.104", "hostname": "vm04", "labels": [], "status": ""}] 2026-03-09T15:13:09.018 INFO:tasks.cephadm:Writing (initial) conf and keyring to vm06 2026-03-09T15:13:09.018 DEBUG:teuthology.orchestra.run.vm06:> set -ex 2026-03-09T15:13:09.018 DEBUG:teuthology.orchestra.run.vm06:> dd of=/etc/ceph/ceph.conf 2026-03-09T15:13:09.032 DEBUG:teuthology.orchestra.run.vm06:> set -ex 2026-03-09T15:13:09.032 DEBUG:teuthology.orchestra.run.vm06:> dd of=/etc/ceph/ceph.client.admin.keyring 2026-03-09T15:13:09.087 INFO:tasks.cephadm:Adding host vm06 to orchestrator... 2026-03-09T15:13:09.087 DEBUG:teuthology.orchestra.run.vm03:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 4abede80-1bca-11f1-bcd8-9b1da3c03ce5 -- ceph orch host add vm06 2026-03-09T15:13:09.242 INFO:teuthology.orchestra.run.vm03.stderr:Inferring config /var/lib/ceph/4abede80-1bca-11f1-bcd8-9b1da3c03ce5/mon.a/config 2026-03-09T15:13:09.521 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:13:09 vm03 ceph-mon[51762]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' 2026-03-09T15:13:09.799 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:13:09 vm03 ceph-mon[51762]: Added host vm04 2026-03-09T15:13:09.799 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:13:09 vm03 ceph-mon[51762]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T15:13:09.799 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:13:09 vm03 ceph-mon[51762]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' 2026-03-09T15:13:09.799 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:13:09 vm03 ceph-mon[51762]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' 2026-03-09T15:13:09.799 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:13:09 vm03 ceph-mon[51762]: from='client.14176 -' entity='client.admin' cmd=[{"prefix": "orch host ls", "target": ["mon-mgr", ""], "format": "json"}]: dispatch 2026-03-09T15:13:09.799 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:13:09 vm03 ceph-mon[51762]: mgrmap e12: a(active, since 6s) 2026-03-09T15:13:09.799 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:13:09 vm03 ceph-mon[51762]: from='client.14178 -' entity='client.admin' cmd=[{"prefix": "orch host add", "hostname": "vm06", "target": ["mon-mgr", ""]}]: dispatch 2026-03-09T15:13:10.920 INFO:teuthology.orchestra.run.vm03.stdout:Added host 'vm06' with addr '192.168.123.106' 2026-03-09T15:13:10.983 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:13:10 vm03 ceph-mon[51762]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' 2026-03-09T15:13:10.983 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:13:10 vm03 ceph-mon[51762]: Deploying cephadm binary to vm06 2026-03-09T15:13:10.985 DEBUG:teuthology.orchestra.run.vm03:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 4abede80-1bca-11f1-bcd8-9b1da3c03ce5 -- ceph orch host ls --format=json 2026-03-09T15:13:11.164 INFO:teuthology.orchestra.run.vm03.stderr:Inferring config /var/lib/ceph/4abede80-1bca-11f1-bcd8-9b1da3c03ce5/mon.a/config 2026-03-09T15:13:11.406 INFO:teuthology.orchestra.run.vm03.stdout: 2026-03-09T15:13:11.407 INFO:teuthology.orchestra.run.vm03.stdout:[{"addr": "192.168.123.103", "hostname": "vm03", "labels": [], "status": ""}, {"addr": "192.168.123.104", "hostname": "vm04", "labels": [], "status": ""}, {"addr": "192.168.123.106", "hostname": "vm06", "labels": [], "status": ""}] 2026-03-09T15:13:11.483 INFO:tasks.cephadm:Setting crush tunables to default 2026-03-09T15:13:11.483 DEBUG:teuthology.orchestra.run.vm03:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 4abede80-1bca-11f1-bcd8-9b1da3c03ce5 -- ceph osd crush tunables default 2026-03-09T15:13:11.659 INFO:teuthology.orchestra.run.vm03.stderr:Inferring config /var/lib/ceph/4abede80-1bca-11f1-bcd8-9b1da3c03ce5/mon.a/config 2026-03-09T15:13:12.050 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:13:11 vm03 ceph-mon[51762]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' 2026-03-09T15:13:12.050 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:13:11 vm03 ceph-mon[51762]: Added host vm06 2026-03-09T15:13:12.050 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:13:11 vm03 ceph-mon[51762]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' 2026-03-09T15:13:12.050 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:13:11 vm03 ceph-mon[51762]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' 2026-03-09T15:13:12.050 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:13:11 vm03 ceph-mon[51762]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' 2026-03-09T15:13:12.050 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:13:11 vm03 ceph-mon[51762]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' 2026-03-09T15:13:12.050 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:13:11 vm03 ceph-mon[51762]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' cmd=[{"prefix": "config rm", "who": "osd/host:vm04", "name": "osd_memory_target"}]: dispatch 2026-03-09T15:13:12.050 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:13:11 vm03 ceph-mon[51762]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T15:13:12.050 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:13:11 vm03 ceph-mon[51762]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T15:13:12.050 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:13:11 vm03 ceph-mon[51762]: Updating vm04:/etc/ceph/ceph.conf 2026-03-09T15:13:12.050 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:13:11 vm03 ceph-mon[51762]: Updating vm04:/var/lib/ceph/4abede80-1bca-11f1-bcd8-9b1da3c03ce5/config/ceph.conf 2026-03-09T15:13:12.050 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:13:11 vm03 ceph-mon[51762]: Updating vm04:/etc/ceph/ceph.client.admin.keyring 2026-03-09T15:13:12.050 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:13:11 vm03 ceph-mon[51762]: Updating vm04:/var/lib/ceph/4abede80-1bca-11f1-bcd8-9b1da3c03ce5/config/ceph.client.admin.keyring 2026-03-09T15:13:12.050 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:13:11 vm03 ceph-mon[51762]: from='client.14180 -' entity='client.admin' cmd=[{"prefix": "orch host ls", "target": ["mon-mgr", ""], "format": "json"}]: dispatch 2026-03-09T15:13:12.050 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:13:11 vm03 ceph-mon[51762]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' 2026-03-09T15:13:12.050 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:13:11 vm03 ceph-mon[51762]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' 2026-03-09T15:13:12.050 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:13:11 vm03 ceph-mon[51762]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' 2026-03-09T15:13:12.050 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:13:11 vm03 ceph-mon[51762]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T15:13:12.050 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:13:11 vm03 ceph-mon[51762]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' 2026-03-09T15:13:12.050 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:13:11 vm03 ceph-mon[51762]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' 2026-03-09T15:13:12.050 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:13:11 vm03 ceph-mon[51762]: from='client.? 192.168.123.103:0/411604909' entity='client.admin' cmd=[{"prefix": "osd crush tunables", "profile": "default"}]: dispatch 2026-03-09T15:13:12.870 INFO:teuthology.orchestra.run.vm03.stderr:adjusted tunables profile to default 2026-03-09T15:13:12.922 INFO:tasks.cephadm:Adding mon.a on vm03 2026-03-09T15:13:12.922 INFO:tasks.cephadm:Adding mon.b on vm04 2026-03-09T15:13:12.922 INFO:tasks.cephadm:Adding mon.c on vm06 2026-03-09T15:13:12.922 DEBUG:teuthology.orchestra.run.vm06:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 4abede80-1bca-11f1-bcd8-9b1da3c03ce5 -- ceph orch apply mon '3;vm03:192.168.123.103=a;vm04:192.168.123.104=b;vm06:192.168.123.106=c' 2026-03-09T15:13:13.117 INFO:teuthology.orchestra.run.vm06.stderr:Inferring config /etc/ceph/ceph.conf 2026-03-09T15:13:13.159 INFO:teuthology.orchestra.run.vm06.stderr:Inferring config /etc/ceph/ceph.conf 2026-03-09T15:13:13.412 INFO:teuthology.orchestra.run.vm06.stdout:Scheduled mon update... 2026-03-09T15:13:13.510 DEBUG:teuthology.orchestra.run.vm04:mon.b> sudo journalctl -f -n 0 -u ceph-4abede80-1bca-11f1-bcd8-9b1da3c03ce5@mon.b.service 2026-03-09T15:13:13.512 DEBUG:teuthology.orchestra.run.vm06:mon.c> sudo journalctl -f -n 0 -u ceph-4abede80-1bca-11f1-bcd8-9b1da3c03ce5@mon.c.service 2026-03-09T15:13:13.514 INFO:tasks.cephadm:Waiting for 3 mons in monmap... 2026-03-09T15:13:13.514 DEBUG:teuthology.orchestra.run.vm06:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 4abede80-1bca-11f1-bcd8-9b1da3c03ce5 -- ceph mon dump -f json 2026-03-09T15:13:13.549 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:13:13 vm03 ceph-mon[51762]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' 2026-03-09T15:13:13.549 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:13:13 vm03 ceph-mon[51762]: from='client.? 192.168.123.103:0/411604909' entity='client.admin' cmd='[{"prefix": "osd crush tunables", "profile": "default"}]': finished 2026-03-09T15:13:13.549 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:13:13 vm03 ceph-mon[51762]: osdmap e4: 0 total, 0 up, 0 in 2026-03-09T15:13:13.741 INFO:teuthology.orchestra.run.vm06.stderr:Inferring config /etc/ceph/ceph.conf 2026-03-09T15:13:13.794 INFO:teuthology.orchestra.run.vm06.stderr:Inferring config /etc/ceph/ceph.conf 2026-03-09T15:13:14.059 INFO:teuthology.orchestra.run.vm06.stdout: 2026-03-09T15:13:14.059 INFO:teuthology.orchestra.run.vm06.stdout:{"epoch":1,"fsid":"4abede80-1bca-11f1-bcd8-9b1da3c03ce5","modified":"2026-03-09T15:12:40.941469Z","created":"2026-03-09T15:12:40.941469Z","min_mon_release":19,"min_mon_release_name":"squid","election_strategy":1,"disallowed_leaders":"","stretch_mode":false,"tiebreaker_mon":"","removed_ranks":"","features":{"persistent":["kraken","luminous","mimic","osdmap-prune","nautilus","octopus","pacific","elector-pinging","quincy","reef","squid"],"optional":[]},"mons":[{"rank":0,"name":"a","public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.103:3300","nonce":0},{"type":"v1","addr":"192.168.123.103:6789","nonce":0}]},"addr":"192.168.123.103:6789/0","public_addr":"192.168.123.103:6789/0","priority":0,"weight":0,"crush_location":"{}"}],"quorum":[0]} 2026-03-09T15:13:14.059 INFO:teuthology.orchestra.run.vm06.stderr:dumped monmap epoch 1 2026-03-09T15:13:14.799 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:13:14 vm03 ceph-mon[51762]: from='client.14184 -' entity='client.admin' cmd=[{"prefix": "orch apply", "service_type": "mon", "placement": "3;vm03:192.168.123.103=a;vm04:192.168.123.104=b;vm06:192.168.123.106=c", "target": ["mon-mgr", ""]}]: dispatch 2026-03-09T15:13:14.800 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:13:14 vm03 ceph-mon[51762]: Saving service mon spec with placement vm03:192.168.123.103=a;vm04:192.168.123.104=b;vm06:192.168.123.106=c;count:3 2026-03-09T15:13:14.800 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:13:14 vm03 ceph-mon[51762]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' 2026-03-09T15:13:14.800 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:13:14 vm03 ceph-mon[51762]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' 2026-03-09T15:13:14.800 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:13:14 vm03 ceph-mon[51762]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' 2026-03-09T15:13:14.800 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:13:14 vm03 ceph-mon[51762]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' 2026-03-09T15:13:14.800 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:13:14 vm03 ceph-mon[51762]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' 2026-03-09T15:13:14.800 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:13:14 vm03 ceph-mon[51762]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' cmd=[{"prefix": "config rm", "who": "osd/host:vm06", "name": "osd_memory_target"}]: dispatch 2026-03-09T15:13:14.800 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:13:14 vm03 ceph-mon[51762]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T15:13:14.800 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:13:14 vm03 ceph-mon[51762]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T15:13:14.800 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:13:14 vm03 ceph-mon[51762]: from='client.? 192.168.123.106:0/1264792745' entity='client.admin' cmd=[{"prefix": "mon dump", "format": "json"}]: dispatch 2026-03-09T15:13:14.800 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:13:14 vm03 ceph-mon[51762]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' 2026-03-09T15:13:14.800 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:13:14 vm03 ceph-mon[51762]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' 2026-03-09T15:13:14.800 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:13:14 vm03 ceph-mon[51762]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' 2026-03-09T15:13:14.800 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:13:14 vm03 ceph-mon[51762]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "mon."}]: dispatch 2026-03-09T15:13:14.800 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:13:14 vm03 ceph-mon[51762]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T15:13:15.119 INFO:tasks.cephadm:Waiting for 3 mons in monmap... 2026-03-09T15:13:15.119 DEBUG:teuthology.orchestra.run.vm06:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 4abede80-1bca-11f1-bcd8-9b1da3c03ce5 -- ceph mon dump -f json 2026-03-09T15:13:15.421 INFO:teuthology.orchestra.run.vm06.stderr:Inferring config /var/lib/ceph/4abede80-1bca-11f1-bcd8-9b1da3c03ce5/mon.c/config 2026-03-09T15:13:15.486 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 09 15:13:15 vm06 ceph-mon[56506]: Deploying daemon mon.c on vm06 2026-03-09T15:13:15.767 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 09 15:13:15 vm06 ceph-mon[56506]: mon.c@-1(synchronizing).paxosservice(auth 1..3) refresh upgraded, format 0 -> 3 2026-03-09T15:13:15.799 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:13:15 vm03 ceph-mon[51762]: Updating vm06:/etc/ceph/ceph.conf 2026-03-09T15:13:15.799 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:13:15 vm03 ceph-mon[51762]: Updating vm06:/var/lib/ceph/4abede80-1bca-11f1-bcd8-9b1da3c03ce5/config/ceph.conf 2026-03-09T15:13:15.799 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:13:15 vm03 ceph-mon[51762]: Updating vm06:/etc/ceph/ceph.client.admin.keyring 2026-03-09T15:13:15.799 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:13:15 vm03 ceph-mon[51762]: Updating vm06:/var/lib/ceph/4abede80-1bca-11f1-bcd8-9b1da3c03ce5/config/ceph.client.admin.keyring 2026-03-09T15:13:15.799 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:13:15 vm03 ceph-mon[51762]: Deploying daemon mon.c on vm06 2026-03-09T15:13:17.001 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 15:13:16 vm04 ceph-mon[53998]: mon.b@-1(synchronizing).paxosservice(auth 1..3) refresh upgraded, format 0 -> 3 2026-03-09T15:13:20.825 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 09 15:13:20 vm06 ceph-mon[56506]: Deploying daemon mon.b on vm04 2026-03-09T15:13:20.825 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 09 15:13:20 vm06 ceph-mon[56506]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "a"}]: dispatch 2026-03-09T15:13:20.825 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 09 15:13:20 vm06 ceph-mon[56506]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "c"}]: dispatch 2026-03-09T15:13:20.825 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 09 15:13:20 vm06 ceph-mon[56506]: mon.a calling monitor election 2026-03-09T15:13:20.825 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 09 15:13:20 vm06 ceph-mon[56506]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "c"}]: dispatch 2026-03-09T15:13:20.825 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 09 15:13:20 vm06 ceph-mon[56506]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-09T15:13:20.825 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 09 15:13:20 vm06 ceph-mon[56506]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "c"}]: dispatch 2026-03-09T15:13:20.825 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 09 15:13:20 vm06 ceph-mon[56506]: mon.c calling monitor election 2026-03-09T15:13:20.825 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 09 15:13:20 vm06 ceph-mon[56506]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-09T15:13:20.825 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 09 15:13:20 vm06 ceph-mon[56506]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "c"}]: dispatch 2026-03-09T15:13:20.825 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 09 15:13:20 vm06 ceph-mon[56506]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-09T15:13:20.825 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 09 15:13:20 vm06 ceph-mon[56506]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "c"}]: dispatch 2026-03-09T15:13:20.825 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 09 15:13:20 vm06 ceph-mon[56506]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-09T15:13:20.825 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 09 15:13:20 vm06 ceph-mon[56506]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "c"}]: dispatch 2026-03-09T15:13:20.825 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 09 15:13:20 vm06 ceph-mon[56506]: mon.a is new leader, mons a,c in quorum (ranks 0,1) 2026-03-09T15:13:20.825 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 09 15:13:20 vm06 ceph-mon[56506]: monmap epoch 2 2026-03-09T15:13:20.825 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 09 15:13:20 vm06 ceph-mon[56506]: fsid 4abede80-1bca-11f1-bcd8-9b1da3c03ce5 2026-03-09T15:13:20.825 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 09 15:13:20 vm06 ceph-mon[56506]: last_changed 2026-03-09T15:13:15.488831+0000 2026-03-09T15:13:20.825 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 09 15:13:20 vm06 ceph-mon[56506]: created 2026-03-09T15:12:40.941469+0000 2026-03-09T15:13:20.825 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 09 15:13:20 vm06 ceph-mon[56506]: min_mon_release 19 (squid) 2026-03-09T15:13:20.825 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 09 15:13:20 vm06 ceph-mon[56506]: election_strategy: 1 2026-03-09T15:13:20.825 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 09 15:13:20 vm06 ceph-mon[56506]: 0: [v2:192.168.123.103:3300/0,v1:192.168.123.103:6789/0] mon.a 2026-03-09T15:13:20.825 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 09 15:13:20 vm06 ceph-mon[56506]: 1: [v2:192.168.123.106:3300/0,v1:192.168.123.106:6789/0] mon.c 2026-03-09T15:13:20.825 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 09 15:13:20 vm06 ceph-mon[56506]: fsmap 2026-03-09T15:13:20.825 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 09 15:13:20 vm06 ceph-mon[56506]: osdmap e4: 0 total, 0 up, 0 in 2026-03-09T15:13:20.825 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 09 15:13:20 vm06 ceph-mon[56506]: mgrmap e12: a(active, since 18s) 2026-03-09T15:13:20.825 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 09 15:13:20 vm06 ceph-mon[56506]: overall HEALTH_OK 2026-03-09T15:13:20.825 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 09 15:13:20 vm06 ceph-mon[56506]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' 2026-03-09T15:13:20.825 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 09 15:13:20 vm06 ceph-mon[56506]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' 2026-03-09T15:13:20.825 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 09 15:13:20 vm06 ceph-mon[56506]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' 2026-03-09T15:13:20.825 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 09 15:13:20 vm06 ceph-mon[56506]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' 2026-03-09T15:13:21.049 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:13:20 vm03 ceph-mon[51762]: Deploying daemon mon.b on vm04 2026-03-09T15:13:21.049 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:13:20 vm03 ceph-mon[51762]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "a"}]: dispatch 2026-03-09T15:13:21.049 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:13:20 vm03 ceph-mon[51762]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "c"}]: dispatch 2026-03-09T15:13:21.049 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:13:20 vm03 ceph-mon[51762]: mon.a calling monitor election 2026-03-09T15:13:21.049 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:13:20 vm03 ceph-mon[51762]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "c"}]: dispatch 2026-03-09T15:13:21.049 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:13:20 vm03 ceph-mon[51762]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-09T15:13:21.050 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:13:20 vm03 ceph-mon[51762]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "c"}]: dispatch 2026-03-09T15:13:21.050 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:13:20 vm03 ceph-mon[51762]: mon.c calling monitor election 2026-03-09T15:13:21.050 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:13:20 vm03 ceph-mon[51762]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-09T15:13:21.050 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:13:20 vm03 ceph-mon[51762]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "c"}]: dispatch 2026-03-09T15:13:21.050 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:13:20 vm03 ceph-mon[51762]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-09T15:13:21.050 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:13:20 vm03 ceph-mon[51762]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "c"}]: dispatch 2026-03-09T15:13:21.050 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:13:20 vm03 ceph-mon[51762]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-09T15:13:21.050 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:13:20 vm03 ceph-mon[51762]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "c"}]: dispatch 2026-03-09T15:13:21.050 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:13:20 vm03 ceph-mon[51762]: mon.a is new leader, mons a,c in quorum (ranks 0,1) 2026-03-09T15:13:21.050 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:13:20 vm03 ceph-mon[51762]: monmap epoch 2 2026-03-09T15:13:21.050 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:13:20 vm03 ceph-mon[51762]: fsid 4abede80-1bca-11f1-bcd8-9b1da3c03ce5 2026-03-09T15:13:21.050 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:13:20 vm03 ceph-mon[51762]: last_changed 2026-03-09T15:13:15.488831+0000 2026-03-09T15:13:21.050 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:13:20 vm03 ceph-mon[51762]: created 2026-03-09T15:12:40.941469+0000 2026-03-09T15:13:21.050 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:13:20 vm03 ceph-mon[51762]: min_mon_release 19 (squid) 2026-03-09T15:13:21.050 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:13:20 vm03 ceph-mon[51762]: election_strategy: 1 2026-03-09T15:13:21.050 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:13:20 vm03 ceph-mon[51762]: 0: [v2:192.168.123.103:3300/0,v1:192.168.123.103:6789/0] mon.a 2026-03-09T15:13:21.050 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:13:20 vm03 ceph-mon[51762]: 1: [v2:192.168.123.106:3300/0,v1:192.168.123.106:6789/0] mon.c 2026-03-09T15:13:21.050 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:13:20 vm03 ceph-mon[51762]: fsmap 2026-03-09T15:13:21.050 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:13:20 vm03 ceph-mon[51762]: osdmap e4: 0 total, 0 up, 0 in 2026-03-09T15:13:21.050 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:13:20 vm03 ceph-mon[51762]: mgrmap e12: a(active, since 18s) 2026-03-09T15:13:21.050 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:13:20 vm03 ceph-mon[51762]: overall HEALTH_OK 2026-03-09T15:13:21.050 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:13:20 vm03 ceph-mon[51762]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' 2026-03-09T15:13:21.050 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:13:20 vm03 ceph-mon[51762]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' 2026-03-09T15:13:21.050 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:13:20 vm03 ceph-mon[51762]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' 2026-03-09T15:13:21.050 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:13:20 vm03 ceph-mon[51762]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' 2026-03-09T15:13:21.799 INFO:journalctl@ceph.mgr.a.vm03.stdout:Mar 09 15:13:21 vm03 ceph-4abede80-1bca-11f1-bcd8-9b1da3c03ce5-mgr-a[51970]: 2026-03-09T15:13:21.486+0000 7fa54cf23640 -1 mgr.server handle_report got status from non-daemon mon.c 2026-03-09T15:13:26.145 INFO:teuthology.orchestra.run.vm06.stdout: 2026-03-09T15:13:26.145 INFO:teuthology.orchestra.run.vm06.stdout:{"epoch":3,"fsid":"4abede80-1bca-11f1-bcd8-9b1da3c03ce5","modified":"2026-03-09T15:13:20.804785Z","created":"2026-03-09T15:12:40.941469Z","min_mon_release":19,"min_mon_release_name":"squid","election_strategy":1,"disallowed_leaders":"","stretch_mode":false,"tiebreaker_mon":"","removed_ranks":"","features":{"persistent":["kraken","luminous","mimic","osdmap-prune","nautilus","octopus","pacific","elector-pinging","quincy","reef","squid"],"optional":[]},"mons":[{"rank":0,"name":"a","public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.103:3300","nonce":0},{"type":"v1","addr":"192.168.123.103:6789","nonce":0}]},"addr":"192.168.123.103:6789/0","public_addr":"192.168.123.103:6789/0","priority":0,"weight":0,"crush_location":"{}"},{"rank":1,"name":"c","public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.106:3300","nonce":0},{"type":"v1","addr":"192.168.123.106:6789","nonce":0}]},"addr":"192.168.123.106:6789/0","public_addr":"192.168.123.106:6789/0","priority":0,"weight":0,"crush_location":"{}"},{"rank":2,"name":"b","public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.104:3300","nonce":0},{"type":"v1","addr":"192.168.123.104:6789","nonce":0}]},"addr":"192.168.123.104:6789/0","public_addr":"192.168.123.104:6789/0","priority":0,"weight":0,"crush_location":"{}"}],"quorum":[0,1]} 2026-03-09T15:13:26.145 INFO:teuthology.orchestra.run.vm06.stderr:dumped monmap epoch 3 2026-03-09T15:13:26.145 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 09 15:13:25 vm06 ceph-mon[56506]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "a"}]: dispatch 2026-03-09T15:13:26.145 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 09 15:13:25 vm06 ceph-mon[56506]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-09T15:13:26.145 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 09 15:13:25 vm06 ceph-mon[56506]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "c"}]: dispatch 2026-03-09T15:13:26.145 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 09 15:13:25 vm06 ceph-mon[56506]: mon.a calling monitor election 2026-03-09T15:13:26.146 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 09 15:13:25 vm06 ceph-mon[56506]: mon.c calling monitor election 2026-03-09T15:13:26.146 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 09 15:13:25 vm06 ceph-mon[56506]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-09T15:13:26.146 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 09 15:13:25 vm06 ceph-mon[56506]: pgmap v4: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-09T15:13:26.146 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 09 15:13:25 vm06 ceph-mon[56506]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-09T15:13:26.146 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 09 15:13:25 vm06 ceph-mon[56506]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-09T15:13:26.146 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 09 15:13:25 vm06 ceph-mon[56506]: pgmap v5: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-09T15:13:26.146 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 09 15:13:25 vm06 ceph-mon[56506]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-09T15:13:26.146 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 09 15:13:25 vm06 ceph-mon[56506]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-09T15:13:26.146 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 09 15:13:25 vm06 ceph-mon[56506]: mon.a is new leader, mons a,c in quorum (ranks 0,1) 2026-03-09T15:13:26.146 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 09 15:13:25 vm06 ceph-mon[56506]: monmap epoch 3 2026-03-09T15:13:26.146 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 09 15:13:25 vm06 ceph-mon[56506]: fsid 4abede80-1bca-11f1-bcd8-9b1da3c03ce5 2026-03-09T15:13:26.146 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 09 15:13:25 vm06 ceph-mon[56506]: last_changed 2026-03-09T15:13:20.804785+0000 2026-03-09T15:13:26.146 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 09 15:13:25 vm06 ceph-mon[56506]: created 2026-03-09T15:12:40.941469+0000 2026-03-09T15:13:26.146 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 09 15:13:25 vm06 ceph-mon[56506]: min_mon_release 19 (squid) 2026-03-09T15:13:26.146 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 09 15:13:25 vm06 ceph-mon[56506]: election_strategy: 1 2026-03-09T15:13:26.146 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 09 15:13:25 vm06 ceph-mon[56506]: 0: [v2:192.168.123.103:3300/0,v1:192.168.123.103:6789/0] mon.a 2026-03-09T15:13:26.146 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 09 15:13:25 vm06 ceph-mon[56506]: 1: [v2:192.168.123.106:3300/0,v1:192.168.123.106:6789/0] mon.c 2026-03-09T15:13:26.146 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 09 15:13:25 vm06 ceph-mon[56506]: 2: [v2:192.168.123.104:3300/0,v1:192.168.123.104:6789/0] mon.b 2026-03-09T15:13:26.146 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 09 15:13:25 vm06 ceph-mon[56506]: fsmap 2026-03-09T15:13:26.146 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 09 15:13:25 vm06 ceph-mon[56506]: osdmap e4: 0 total, 0 up, 0 in 2026-03-09T15:13:26.146 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 09 15:13:25 vm06 ceph-mon[56506]: mgrmap e12: a(active, since 23s) 2026-03-09T15:13:26.146 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 09 15:13:25 vm06 ceph-mon[56506]: overall HEALTH_OK 2026-03-09T15:13:26.146 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 09 15:13:25 vm06 ceph-mon[56506]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' 2026-03-09T15:13:26.146 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 09 15:13:25 vm06 ceph-mon[56506]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' 2026-03-09T15:13:26.146 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 09 15:13:25 vm06 ceph-mon[56506]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' 2026-03-09T15:13:26.146 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 09 15:13:25 vm06 ceph-mon[56506]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' 2026-03-09T15:13:26.146 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 09 15:13:25 vm06 ceph-mon[56506]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' 2026-03-09T15:13:26.146 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 09 15:13:25 vm06 ceph-mon[56506]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T15:13:26.146 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 09 15:13:25 vm06 ceph-mon[56506]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T15:13:26.201 INFO:tasks.cephadm:Generating final ceph.conf file... 2026-03-09T15:13:26.201 DEBUG:teuthology.orchestra.run.vm03:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 4abede80-1bca-11f1-bcd8-9b1da3c03ce5 -- ceph config generate-minimal-conf 2026-03-09T15:13:26.233 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:13:25 vm03 ceph-mon[51762]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "a"}]: dispatch 2026-03-09T15:13:26.234 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:13:25 vm03 ceph-mon[51762]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-09T15:13:26.234 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:13:25 vm03 ceph-mon[51762]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "c"}]: dispatch 2026-03-09T15:13:26.234 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:13:25 vm03 ceph-mon[51762]: mon.a calling monitor election 2026-03-09T15:13:26.234 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:13:25 vm03 ceph-mon[51762]: mon.c calling monitor election 2026-03-09T15:13:26.234 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:13:25 vm03 ceph-mon[51762]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-09T15:13:26.234 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:13:25 vm03 ceph-mon[51762]: pgmap v4: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-09T15:13:26.234 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:13:25 vm03 ceph-mon[51762]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-09T15:13:26.234 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:13:25 vm03 ceph-mon[51762]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-09T15:13:26.234 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:13:25 vm03 ceph-mon[51762]: pgmap v5: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-09T15:13:26.234 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:13:25 vm03 ceph-mon[51762]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-09T15:13:26.234 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:13:25 vm03 ceph-mon[51762]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-09T15:13:26.234 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:13:25 vm03 ceph-mon[51762]: mon.a is new leader, mons a,c in quorum (ranks 0,1) 2026-03-09T15:13:26.234 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:13:25 vm03 ceph-mon[51762]: monmap epoch 3 2026-03-09T15:13:26.234 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:13:25 vm03 ceph-mon[51762]: fsid 4abede80-1bca-11f1-bcd8-9b1da3c03ce5 2026-03-09T15:13:26.234 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:13:25 vm03 ceph-mon[51762]: last_changed 2026-03-09T15:13:20.804785+0000 2026-03-09T15:13:26.234 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:13:25 vm03 ceph-mon[51762]: created 2026-03-09T15:12:40.941469+0000 2026-03-09T15:13:26.234 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:13:25 vm03 ceph-mon[51762]: min_mon_release 19 (squid) 2026-03-09T15:13:26.234 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:13:25 vm03 ceph-mon[51762]: election_strategy: 1 2026-03-09T15:13:26.234 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:13:25 vm03 ceph-mon[51762]: 0: [v2:192.168.123.103:3300/0,v1:192.168.123.103:6789/0] mon.a 2026-03-09T15:13:26.234 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:13:25 vm03 ceph-mon[51762]: 1: [v2:192.168.123.106:3300/0,v1:192.168.123.106:6789/0] mon.c 2026-03-09T15:13:26.234 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:13:25 vm03 ceph-mon[51762]: 2: [v2:192.168.123.104:3300/0,v1:192.168.123.104:6789/0] mon.b 2026-03-09T15:13:26.234 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:13:25 vm03 ceph-mon[51762]: fsmap 2026-03-09T15:13:26.234 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:13:25 vm03 ceph-mon[51762]: osdmap e4: 0 total, 0 up, 0 in 2026-03-09T15:13:26.234 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:13:25 vm03 ceph-mon[51762]: mgrmap e12: a(active, since 23s) 2026-03-09T15:13:26.234 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:13:25 vm03 ceph-mon[51762]: overall HEALTH_OK 2026-03-09T15:13:26.234 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:13:25 vm03 ceph-mon[51762]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' 2026-03-09T15:13:26.234 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:13:25 vm03 ceph-mon[51762]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' 2026-03-09T15:13:26.234 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:13:25 vm03 ceph-mon[51762]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' 2026-03-09T15:13:26.234 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:13:25 vm03 ceph-mon[51762]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' 2026-03-09T15:13:26.234 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:13:25 vm03 ceph-mon[51762]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' 2026-03-09T15:13:26.234 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:13:25 vm03 ceph-mon[51762]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T15:13:26.234 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:13:25 vm03 ceph-mon[51762]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T15:13:26.389 INFO:teuthology.orchestra.run.vm03.stderr:Inferring config /var/lib/ceph/4abede80-1bca-11f1-bcd8-9b1da3c03ce5/mon.a/config 2026-03-09T15:13:26.633 INFO:teuthology.orchestra.run.vm03.stdout:# minimal ceph.conf for 4abede80-1bca-11f1-bcd8-9b1da3c03ce5 2026-03-09T15:13:26.633 INFO:teuthology.orchestra.run.vm03.stdout:[global] 2026-03-09T15:13:26.633 INFO:teuthology.orchestra.run.vm03.stdout: fsid = 4abede80-1bca-11f1-bcd8-9b1da3c03ce5 2026-03-09T15:13:26.633 INFO:teuthology.orchestra.run.vm03.stdout: mon_host = [v2:192.168.123.103:3300/0,v1:192.168.123.103:6789/0] [v2:192.168.123.104:3300/0,v1:192.168.123.104:6789/0] [v2:192.168.123.106:3300/0,v1:192.168.123.106:6789/0] 2026-03-09T15:13:26.700 INFO:tasks.cephadm:Distributing (final) config and client.admin keyring... 2026-03-09T15:13:26.700 DEBUG:teuthology.orchestra.run.vm03:> set -ex 2026-03-09T15:13:26.700 DEBUG:teuthology.orchestra.run.vm03:> sudo dd of=/etc/ceph/ceph.conf 2026-03-09T15:13:26.770 DEBUG:teuthology.orchestra.run.vm03:> set -ex 2026-03-09T15:13:26.770 DEBUG:teuthology.orchestra.run.vm03:> sudo dd of=/etc/ceph/ceph.client.admin.keyring 2026-03-09T15:13:26.838 DEBUG:teuthology.orchestra.run.vm04:> set -ex 2026-03-09T15:13:26.838 DEBUG:teuthology.orchestra.run.vm04:> sudo dd of=/etc/ceph/ceph.conf 2026-03-09T15:13:26.863 DEBUG:teuthology.orchestra.run.vm04:> set -ex 2026-03-09T15:13:26.863 DEBUG:teuthology.orchestra.run.vm04:> sudo dd of=/etc/ceph/ceph.client.admin.keyring 2026-03-09T15:13:26.926 DEBUG:teuthology.orchestra.run.vm06:> set -ex 2026-03-09T15:13:26.926 DEBUG:teuthology.orchestra.run.vm06:> sudo dd of=/etc/ceph/ceph.conf 2026-03-09T15:13:26.959 DEBUG:teuthology.orchestra.run.vm06:> set -ex 2026-03-09T15:13:26.959 DEBUG:teuthology.orchestra.run.vm06:> sudo dd of=/etc/ceph/ceph.client.admin.keyring 2026-03-09T15:13:27.044 INFO:tasks.cephadm:Adding mgr.a on vm03 2026-03-09T15:13:27.044 INFO:tasks.cephadm:Adding mgr.b on vm04 2026-03-09T15:13:27.044 DEBUG:teuthology.orchestra.run.vm06:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 4abede80-1bca-11f1-bcd8-9b1da3c03ce5 -- ceph orch apply mgr '2;vm03=a;vm04=b' 2026-03-09T15:13:27.242 INFO:teuthology.orchestra.run.vm06.stderr:Inferring config /var/lib/ceph/4abede80-1bca-11f1-bcd8-9b1da3c03ce5/mon.c/config 2026-03-09T15:13:27.265 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 09 15:13:27 vm06 ceph-mon[56506]: Updating vm03:/etc/ceph/ceph.conf 2026-03-09T15:13:27.265 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 09 15:13:27 vm06 ceph-mon[56506]: Updating vm04:/etc/ceph/ceph.conf 2026-03-09T15:13:27.265 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 09 15:13:27 vm06 ceph-mon[56506]: Updating vm06:/etc/ceph/ceph.conf 2026-03-09T15:13:27.266 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 09 15:13:27 vm06 ceph-mon[56506]: Updating vm03:/var/lib/ceph/4abede80-1bca-11f1-bcd8-9b1da3c03ce5/config/ceph.conf 2026-03-09T15:13:27.266 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 09 15:13:27 vm06 ceph-mon[56506]: Updating vm06:/var/lib/ceph/4abede80-1bca-11f1-bcd8-9b1da3c03ce5/config/ceph.conf 2026-03-09T15:13:27.266 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 09 15:13:27 vm06 ceph-mon[56506]: Updating vm04:/var/lib/ceph/4abede80-1bca-11f1-bcd8-9b1da3c03ce5/config/ceph.conf 2026-03-09T15:13:27.266 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 09 15:13:27 vm06 ceph-mon[56506]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' 2026-03-09T15:13:27.266 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 09 15:13:27 vm06 ceph-mon[56506]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' 2026-03-09T15:13:27.266 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 09 15:13:27 vm06 ceph-mon[56506]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' 2026-03-09T15:13:27.266 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 09 15:13:27 vm06 ceph-mon[56506]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' 2026-03-09T15:13:27.266 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 09 15:13:27 vm06 ceph-mon[56506]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' 2026-03-09T15:13:27.266 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 09 15:13:27 vm06 ceph-mon[56506]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' 2026-03-09T15:13:27.266 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 09 15:13:27 vm06 ceph-mon[56506]: from='client.? 192.168.123.106:0/1702195684' entity='client.admin' cmd=[{"prefix": "mon dump", "format": "json"}]: dispatch 2026-03-09T15:13:27.266 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 09 15:13:27 vm06 ceph-mon[56506]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' 2026-03-09T15:13:27.266 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 09 15:13:27 vm06 ceph-mon[56506]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' 2026-03-09T15:13:27.266 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 09 15:13:27 vm06 ceph-mon[56506]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' 2026-03-09T15:13:27.266 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 09 15:13:27 vm06 ceph-mon[56506]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' 2026-03-09T15:13:27.266 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 09 15:13:27 vm06 ceph-mon[56506]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' 2026-03-09T15:13:27.266 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 09 15:13:27 vm06 ceph-mon[56506]: Reconfiguring mon.a (unknown last config time)... 2026-03-09T15:13:27.266 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 09 15:13:27 vm06 ceph-mon[56506]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "mon."}]: dispatch 2026-03-09T15:13:27.266 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 09 15:13:27 vm06 ceph-mon[56506]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' cmd=[{"prefix": "config get", "who": "mon", "key": "public_network"}]: dispatch 2026-03-09T15:13:27.266 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 09 15:13:27 vm06 ceph-mon[56506]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T15:13:27.266 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 09 15:13:27 vm06 ceph-mon[56506]: Reconfiguring daemon mon.a on vm03 2026-03-09T15:13:27.266 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 09 15:13:27 vm06 ceph-mon[56506]: pgmap v6: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-09T15:13:27.266 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 09 15:13:27 vm06 ceph-mon[56506]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' 2026-03-09T15:13:27.266 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 09 15:13:27 vm06 ceph-mon[56506]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' 2026-03-09T15:13:27.266 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 09 15:13:27 vm06 ceph-mon[56506]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "mon."}]: dispatch 2026-03-09T15:13:27.266 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 09 15:13:27 vm06 ceph-mon[56506]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' cmd=[{"prefix": "config get", "who": "mon", "key": "public_network"}]: dispatch 2026-03-09T15:13:27.266 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 09 15:13:27 vm06 ceph-mon[56506]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T15:13:27.266 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 09 15:13:27 vm06 ceph-mon[56506]: from='client.? 192.168.123.103:0/1490553382' entity='client.admin' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T15:13:27.266 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 09 15:13:27 vm06 ceph-mon[56506]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' 2026-03-09T15:13:27.266 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 09 15:13:27 vm06 ceph-mon[56506]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' 2026-03-09T15:13:27.266 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 09 15:13:27 vm06 ceph-mon[56506]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-09T15:13:27.266 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 09 15:13:27 vm06 ceph-mon[56506]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "mon."}]: dispatch 2026-03-09T15:13:27.266 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 09 15:13:27 vm06 ceph-mon[56506]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' cmd=[{"prefix": "config get", "who": "mon", "key": "public_network"}]: dispatch 2026-03-09T15:13:27.266 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 09 15:13:27 vm06 ceph-mon[56506]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T15:13:27.266 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 09 15:13:27 vm06 ceph-mon[56506]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' 2026-03-09T15:13:27.266 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 09 15:13:27 vm06 ceph-mon[56506]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' 2026-03-09T15:13:27.266 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 09 15:13:27 vm06 ceph-mon[56506]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T15:13:27.266 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 09 15:13:27 vm06 ceph-mon[56506]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T15:13:27.266 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 09 15:13:27 vm06 ceph-mon[56506]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T15:13:27.266 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 09 15:13:27 vm06 ceph-mon[56506]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' 2026-03-09T15:13:27.470 INFO:teuthology.orchestra.run.vm06.stdout:Scheduled mgr update... 2026-03-09T15:13:27.527 DEBUG:teuthology.orchestra.run.vm04:mgr.b> sudo journalctl -f -n 0 -u ceph-4abede80-1bca-11f1-bcd8-9b1da3c03ce5@mgr.b.service 2026-03-09T15:13:27.528 INFO:tasks.cephadm:Deploying OSDs... 2026-03-09T15:13:27.529 DEBUG:teuthology.orchestra.run.vm03:> set -ex 2026-03-09T15:13:27.529 DEBUG:teuthology.orchestra.run.vm03:> dd if=/scratch_devs of=/dev/stdout 2026-03-09T15:13:27.544 DEBUG:teuthology.orchestra.run:got remote process result: 1 2026-03-09T15:13:27.544 DEBUG:teuthology.orchestra.run.vm03:> ls /dev/[sv]d? 2026-03-09T15:13:27.549 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:13:27 vm03 ceph-mon[51762]: Updating vm03:/etc/ceph/ceph.conf 2026-03-09T15:13:27.549 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:13:27 vm03 ceph-mon[51762]: Updating vm04:/etc/ceph/ceph.conf 2026-03-09T15:13:27.549 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:13:27 vm03 ceph-mon[51762]: Updating vm06:/etc/ceph/ceph.conf 2026-03-09T15:13:27.549 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:13:27 vm03 ceph-mon[51762]: Updating vm03:/var/lib/ceph/4abede80-1bca-11f1-bcd8-9b1da3c03ce5/config/ceph.conf 2026-03-09T15:13:27.549 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:13:27 vm03 ceph-mon[51762]: Updating vm06:/var/lib/ceph/4abede80-1bca-11f1-bcd8-9b1da3c03ce5/config/ceph.conf 2026-03-09T15:13:27.549 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:13:27 vm03 ceph-mon[51762]: Updating vm04:/var/lib/ceph/4abede80-1bca-11f1-bcd8-9b1da3c03ce5/config/ceph.conf 2026-03-09T15:13:27.549 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:13:27 vm03 ceph-mon[51762]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' 2026-03-09T15:13:27.549 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:13:27 vm03 ceph-mon[51762]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' 2026-03-09T15:13:27.549 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:13:27 vm03 ceph-mon[51762]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' 2026-03-09T15:13:27.549 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:13:27 vm03 ceph-mon[51762]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' 2026-03-09T15:13:27.549 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:13:27 vm03 ceph-mon[51762]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' 2026-03-09T15:13:27.549 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:13:27 vm03 ceph-mon[51762]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' 2026-03-09T15:13:27.549 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:13:27 vm03 ceph-mon[51762]: from='client.? 192.168.123.106:0/1702195684' entity='client.admin' cmd=[{"prefix": "mon dump", "format": "json"}]: dispatch 2026-03-09T15:13:27.549 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:13:27 vm03 ceph-mon[51762]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' 2026-03-09T15:13:27.549 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:13:27 vm03 ceph-mon[51762]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' 2026-03-09T15:13:27.549 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:13:27 vm03 ceph-mon[51762]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' 2026-03-09T15:13:27.549 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:13:27 vm03 ceph-mon[51762]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' 2026-03-09T15:13:27.549 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:13:27 vm03 ceph-mon[51762]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' 2026-03-09T15:13:27.550 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:13:27 vm03 ceph-mon[51762]: Reconfiguring mon.a (unknown last config time)... 2026-03-09T15:13:27.550 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:13:27 vm03 ceph-mon[51762]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "mon."}]: dispatch 2026-03-09T15:13:27.550 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:13:27 vm03 ceph-mon[51762]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' cmd=[{"prefix": "config get", "who": "mon", "key": "public_network"}]: dispatch 2026-03-09T15:13:27.550 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:13:27 vm03 ceph-mon[51762]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T15:13:27.550 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:13:27 vm03 ceph-mon[51762]: Reconfiguring daemon mon.a on vm03 2026-03-09T15:13:27.550 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:13:27 vm03 ceph-mon[51762]: pgmap v6: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-09T15:13:27.550 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:13:27 vm03 ceph-mon[51762]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' 2026-03-09T15:13:27.550 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:13:27 vm03 ceph-mon[51762]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' 2026-03-09T15:13:27.550 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:13:27 vm03 ceph-mon[51762]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "mon."}]: dispatch 2026-03-09T15:13:27.550 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:13:27 vm03 ceph-mon[51762]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' cmd=[{"prefix": "config get", "who": "mon", "key": "public_network"}]: dispatch 2026-03-09T15:13:27.550 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:13:27 vm03 ceph-mon[51762]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T15:13:27.550 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:13:27 vm03 ceph-mon[51762]: from='client.? 192.168.123.103:0/1490553382' entity='client.admin' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T15:13:27.550 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:13:27 vm03 ceph-mon[51762]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' 2026-03-09T15:13:27.550 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:13:27 vm03 ceph-mon[51762]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' 2026-03-09T15:13:27.550 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:13:27 vm03 ceph-mon[51762]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-09T15:13:27.550 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:13:27 vm03 ceph-mon[51762]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "mon."}]: dispatch 2026-03-09T15:13:27.550 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:13:27 vm03 ceph-mon[51762]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' cmd=[{"prefix": "config get", "who": "mon", "key": "public_network"}]: dispatch 2026-03-09T15:13:27.550 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:13:27 vm03 ceph-mon[51762]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T15:13:27.550 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:13:27 vm03 ceph-mon[51762]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' 2026-03-09T15:13:27.550 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:13:27 vm03 ceph-mon[51762]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' 2026-03-09T15:13:27.550 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:13:27 vm03 ceph-mon[51762]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T15:13:27.550 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:13:27 vm03 ceph-mon[51762]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T15:13:27.550 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:13:27 vm03 ceph-mon[51762]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T15:13:27.550 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:13:27 vm03 ceph-mon[51762]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' 2026-03-09T15:13:27.564 INFO:teuthology.orchestra.run.vm03.stdout:/dev/vda 2026-03-09T15:13:27.564 INFO:teuthology.orchestra.run.vm03.stdout:/dev/vdb 2026-03-09T15:13:27.564 INFO:teuthology.orchestra.run.vm03.stdout:/dev/vdc 2026-03-09T15:13:27.564 INFO:teuthology.orchestra.run.vm03.stdout:/dev/vdd 2026-03-09T15:13:27.564 INFO:teuthology.orchestra.run.vm03.stdout:/dev/vde 2026-03-09T15:13:27.564 WARNING:teuthology.misc:Removing root device: /dev/vda from device list 2026-03-09T15:13:27.564 DEBUG:teuthology.misc:devs=['/dev/vdb', '/dev/vdc', '/dev/vdd', '/dev/vde'] 2026-03-09T15:13:27.564 DEBUG:teuthology.orchestra.run.vm03:> stat /dev/vdb 2026-03-09T15:13:27.622 INFO:teuthology.orchestra.run.vm03.stdout: File: /dev/vdb 2026-03-09T15:13:27.622 INFO:teuthology.orchestra.run.vm03.stdout: Size: 0 Blocks: 0 IO Block: 512 block special file 2026-03-09T15:13:27.622 INFO:teuthology.orchestra.run.vm03.stdout:Device: 6h/6d Inode: 254 Links: 1 Device type: fc,10 2026-03-09T15:13:27.622 INFO:teuthology.orchestra.run.vm03.stdout:Access: (0660/brw-rw----) Uid: ( 0/ root) Gid: ( 6/ disk) 2026-03-09T15:13:27.622 INFO:teuthology.orchestra.run.vm03.stdout:Context: system_u:object_r:fixed_disk_device_t:s0 2026-03-09T15:13:27.622 INFO:teuthology.orchestra.run.vm03.stdout:Access: 2026-03-09 15:13:05.882062116 +0000 2026-03-09T15:13:27.622 INFO:teuthology.orchestra.run.vm03.stdout:Modify: 2026-03-09 15:10:28.121559362 +0000 2026-03-09T15:13:27.622 INFO:teuthology.orchestra.run.vm03.stdout:Change: 2026-03-09 15:10:28.121559362 +0000 2026-03-09T15:13:27.622 INFO:teuthology.orchestra.run.vm03.stdout: Birth: 2026-03-09 15:06:31.227000000 +0000 2026-03-09T15:13:27.622 DEBUG:teuthology.orchestra.run.vm03:> sudo dd if=/dev/vdb of=/dev/null count=1 2026-03-09T15:13:27.685 INFO:teuthology.orchestra.run.vm03.stderr:1+0 records in 2026-03-09T15:13:27.685 INFO:teuthology.orchestra.run.vm03.stderr:1+0 records out 2026-03-09T15:13:27.685 INFO:teuthology.orchestra.run.vm03.stderr:512 bytes copied, 0.000185119 s, 2.8 MB/s 2026-03-09T15:13:27.686 DEBUG:teuthology.orchestra.run.vm03:> ! mount | grep -v devtmpfs | grep -q /dev/vdb 2026-03-09T15:13:27.742 DEBUG:teuthology.orchestra.run.vm03:> stat /dev/vdc 2026-03-09T15:13:27.801 INFO:teuthology.orchestra.run.vm03.stdout: File: /dev/vdc 2026-03-09T15:13:27.801 INFO:teuthology.orchestra.run.vm03.stdout: Size: 0 Blocks: 0 IO Block: 512 block special file 2026-03-09T15:13:27.801 INFO:teuthology.orchestra.run.vm03.stdout:Device: 6h/6d Inode: 255 Links: 1 Device type: fc,20 2026-03-09T15:13:27.801 INFO:teuthology.orchestra.run.vm03.stdout:Access: (0660/brw-rw----) Uid: ( 0/ root) Gid: ( 6/ disk) 2026-03-09T15:13:27.801 INFO:teuthology.orchestra.run.vm03.stdout:Context: system_u:object_r:fixed_disk_device_t:s0 2026-03-09T15:13:27.801 INFO:teuthology.orchestra.run.vm03.stdout:Access: 2026-03-09 15:13:05.924062455 +0000 2026-03-09T15:13:27.801 INFO:teuthology.orchestra.run.vm03.stdout:Modify: 2026-03-09 15:10:28.157559405 +0000 2026-03-09T15:13:27.801 INFO:teuthology.orchestra.run.vm03.stdout:Change: 2026-03-09 15:10:28.157559405 +0000 2026-03-09T15:13:27.801 INFO:teuthology.orchestra.run.vm03.stdout: Birth: 2026-03-09 15:06:31.235000000 +0000 2026-03-09T15:13:27.802 DEBUG:teuthology.orchestra.run.vm03:> sudo dd if=/dev/vdc of=/dev/null count=1 2026-03-09T15:13:27.867 INFO:teuthology.orchestra.run.vm03.stderr:1+0 records in 2026-03-09T15:13:27.867 INFO:teuthology.orchestra.run.vm03.stderr:1+0 records out 2026-03-09T15:13:27.867 INFO:teuthology.orchestra.run.vm03.stderr:512 bytes copied, 0.000156013 s, 3.3 MB/s 2026-03-09T15:13:27.868 DEBUG:teuthology.orchestra.run.vm03:> ! mount | grep -v devtmpfs | grep -q /dev/vdc 2026-03-09T15:13:27.927 DEBUG:teuthology.orchestra.run.vm03:> stat /dev/vdd 2026-03-09T15:13:27.986 INFO:teuthology.orchestra.run.vm03.stdout: File: /dev/vdd 2026-03-09T15:13:27.986 INFO:teuthology.orchestra.run.vm03.stdout: Size: 0 Blocks: 0 IO Block: 512 block special file 2026-03-09T15:13:27.986 INFO:teuthology.orchestra.run.vm03.stdout:Device: 6h/6d Inode: 256 Links: 1 Device type: fc,30 2026-03-09T15:13:27.986 INFO:teuthology.orchestra.run.vm03.stdout:Access: (0660/brw-rw----) Uid: ( 0/ root) Gid: ( 6/ disk) 2026-03-09T15:13:27.986 INFO:teuthology.orchestra.run.vm03.stdout:Context: system_u:object_r:fixed_disk_device_t:s0 2026-03-09T15:13:27.986 INFO:teuthology.orchestra.run.vm03.stdout:Access: 2026-03-09 15:13:05.989062981 +0000 2026-03-09T15:13:27.986 INFO:teuthology.orchestra.run.vm03.stdout:Modify: 2026-03-09 15:10:28.136559380 +0000 2026-03-09T15:13:27.986 INFO:teuthology.orchestra.run.vm03.stdout:Change: 2026-03-09 15:10:28.136559380 +0000 2026-03-09T15:13:27.986 INFO:teuthology.orchestra.run.vm03.stdout: Birth: 2026-03-09 15:06:31.239000000 +0000 2026-03-09T15:13:27.986 DEBUG:teuthology.orchestra.run.vm03:> sudo dd if=/dev/vdd of=/dev/null count=1 2026-03-09T15:13:28.052 INFO:teuthology.orchestra.run.vm03.stderr:1+0 records in 2026-03-09T15:13:28.052 INFO:teuthology.orchestra.run.vm03.stderr:1+0 records out 2026-03-09T15:13:28.052 INFO:teuthology.orchestra.run.vm03.stderr:512 bytes copied, 0.000160773 s, 3.2 MB/s 2026-03-09T15:13:28.053 DEBUG:teuthology.orchestra.run.vm03:> ! mount | grep -v devtmpfs | grep -q /dev/vdd 2026-03-09T15:13:28.112 DEBUG:teuthology.orchestra.run.vm03:> stat /dev/vde 2026-03-09T15:13:28.137 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 15:13:27 vm04 ceph-mon[53998]: Deploying daemon mon.b on vm04 2026-03-09T15:13:28.137 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 15:13:27 vm04 ceph-mon[53998]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "a"}]: dispatch 2026-03-09T15:13:28.137 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 15:13:27 vm04 ceph-mon[53998]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "c"}]: dispatch 2026-03-09T15:13:28.137 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 15:13:27 vm04 ceph-mon[53998]: mon.a calling monitor election 2026-03-09T15:13:28.137 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 15:13:27 vm04 ceph-mon[53998]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "c"}]: dispatch 2026-03-09T15:13:28.137 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 15:13:27 vm04 ceph-mon[53998]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-09T15:13:28.137 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 15:13:27 vm04 ceph-mon[53998]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "c"}]: dispatch 2026-03-09T15:13:28.137 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 15:13:27 vm04 ceph-mon[53998]: mon.c calling monitor election 2026-03-09T15:13:28.137 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 15:13:27 vm04 ceph-mon[53998]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-09T15:13:28.137 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 15:13:27 vm04 ceph-mon[53998]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "c"}]: dispatch 2026-03-09T15:13:28.137 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 15:13:27 vm04 ceph-mon[53998]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-09T15:13:28.137 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 15:13:27 vm04 ceph-mon[53998]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "c"}]: dispatch 2026-03-09T15:13:28.137 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 15:13:27 vm04 ceph-mon[53998]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-09T15:13:28.137 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 15:13:27 vm04 ceph-mon[53998]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "c"}]: dispatch 2026-03-09T15:13:28.137 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 15:13:27 vm04 ceph-mon[53998]: mon.a is new leader, mons a,c in quorum (ranks 0,1) 2026-03-09T15:13:28.137 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 15:13:27 vm04 ceph-mon[53998]: monmap epoch 2 2026-03-09T15:13:28.137 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 15:13:27 vm04 ceph-mon[53998]: fsid 4abede80-1bca-11f1-bcd8-9b1da3c03ce5 2026-03-09T15:13:28.137 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 15:13:27 vm04 ceph-mon[53998]: last_changed 2026-03-09T15:13:15.488831+0000 2026-03-09T15:13:28.137 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 15:13:27 vm04 ceph-mon[53998]: created 2026-03-09T15:12:40.941469+0000 2026-03-09T15:13:28.137 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 15:13:27 vm04 ceph-mon[53998]: min_mon_release 19 (squid) 2026-03-09T15:13:28.137 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 15:13:27 vm04 ceph-mon[53998]: election_strategy: 1 2026-03-09T15:13:28.137 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 15:13:27 vm04 ceph-mon[53998]: 0: [v2:192.168.123.103:3300/0,v1:192.168.123.103:6789/0] mon.a 2026-03-09T15:13:28.137 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 15:13:27 vm04 ceph-mon[53998]: 1: [v2:192.168.123.106:3300/0,v1:192.168.123.106:6789/0] mon.c 2026-03-09T15:13:28.137 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 15:13:27 vm04 ceph-mon[53998]: fsmap 2026-03-09T15:13:28.137 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 15:13:27 vm04 ceph-mon[53998]: osdmap e4: 0 total, 0 up, 0 in 2026-03-09T15:13:28.137 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 15:13:27 vm04 ceph-mon[53998]: mgrmap e12: a(active, since 18s) 2026-03-09T15:13:28.137 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 15:13:27 vm04 ceph-mon[53998]: overall HEALTH_OK 2026-03-09T15:13:28.137 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 15:13:27 vm04 ceph-mon[53998]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' 2026-03-09T15:13:28.137 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 15:13:27 vm04 ceph-mon[53998]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' 2026-03-09T15:13:28.137 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 15:13:27 vm04 ceph-mon[53998]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' 2026-03-09T15:13:28.137 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 15:13:27 vm04 ceph-mon[53998]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' 2026-03-09T15:13:28.137 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 15:13:27 vm04 ceph-mon[53998]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "a"}]: dispatch 2026-03-09T15:13:28.137 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 15:13:27 vm04 ceph-mon[53998]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-09T15:13:28.137 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 15:13:27 vm04 ceph-mon[53998]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "c"}]: dispatch 2026-03-09T15:13:28.137 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 15:13:27 vm04 ceph-mon[53998]: mon.a calling monitor election 2026-03-09T15:13:28.137 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 15:13:27 vm04 ceph-mon[53998]: mon.c calling monitor election 2026-03-09T15:13:28.137 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 15:13:27 vm04 ceph-mon[53998]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-09T15:13:28.137 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 15:13:27 vm04 ceph-mon[53998]: pgmap v4: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-09T15:13:28.137 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 15:13:27 vm04 ceph-mon[53998]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-09T15:13:28.137 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 15:13:27 vm04 ceph-mon[53998]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-09T15:13:28.137 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 15:13:27 vm04 ceph-mon[53998]: pgmap v5: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-09T15:13:28.137 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 15:13:27 vm04 ceph-mon[53998]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-09T15:13:28.137 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 15:13:27 vm04 ceph-mon[53998]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-09T15:13:28.137 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 15:13:27 vm04 ceph-mon[53998]: mon.a is new leader, mons a,c in quorum (ranks 0,1) 2026-03-09T15:13:28.137 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 15:13:27 vm04 ceph-mon[53998]: monmap epoch 3 2026-03-09T15:13:28.138 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 15:13:27 vm04 ceph-mon[53998]: fsid 4abede80-1bca-11f1-bcd8-9b1da3c03ce5 2026-03-09T15:13:28.138 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 15:13:27 vm04 ceph-mon[53998]: last_changed 2026-03-09T15:13:20.804785+0000 2026-03-09T15:13:28.138 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 15:13:27 vm04 ceph-mon[53998]: created 2026-03-09T15:12:40.941469+0000 2026-03-09T15:13:28.138 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 15:13:27 vm04 ceph-mon[53998]: min_mon_release 19 (squid) 2026-03-09T15:13:28.138 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 15:13:27 vm04 ceph-mon[53998]: election_strategy: 1 2026-03-09T15:13:28.138 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 15:13:27 vm04 ceph-mon[53998]: 0: [v2:192.168.123.103:3300/0,v1:192.168.123.103:6789/0] mon.a 2026-03-09T15:13:28.138 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 15:13:27 vm04 ceph-mon[53998]: 1: [v2:192.168.123.106:3300/0,v1:192.168.123.106:6789/0] mon.c 2026-03-09T15:13:28.138 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 15:13:27 vm04 ceph-mon[53998]: 2: [v2:192.168.123.104:3300/0,v1:192.168.123.104:6789/0] mon.b 2026-03-09T15:13:28.138 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 15:13:27 vm04 ceph-mon[53998]: fsmap 2026-03-09T15:13:28.138 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 15:13:27 vm04 ceph-mon[53998]: osdmap e4: 0 total, 0 up, 0 in 2026-03-09T15:13:28.138 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 15:13:27 vm04 ceph-mon[53998]: mgrmap e12: a(active, since 23s) 2026-03-09T15:13:28.138 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 15:13:27 vm04 ceph-mon[53998]: overall HEALTH_OK 2026-03-09T15:13:28.138 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 15:13:27 vm04 ceph-mon[53998]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' 2026-03-09T15:13:28.138 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 15:13:27 vm04 ceph-mon[53998]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' 2026-03-09T15:13:28.138 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 15:13:27 vm04 ceph-mon[53998]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' 2026-03-09T15:13:28.138 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 15:13:27 vm04 ceph-mon[53998]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' 2026-03-09T15:13:28.138 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 15:13:27 vm04 ceph-mon[53998]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' 2026-03-09T15:13:28.138 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 15:13:27 vm04 ceph-mon[53998]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T15:13:28.138 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 15:13:27 vm04 ceph-mon[53998]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T15:13:28.138 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 15:13:27 vm04 ceph-mon[53998]: Updating vm03:/etc/ceph/ceph.conf 2026-03-09T15:13:28.138 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 15:13:27 vm04 ceph-mon[53998]: Updating vm04:/etc/ceph/ceph.conf 2026-03-09T15:13:28.138 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 15:13:27 vm04 ceph-mon[53998]: Updating vm06:/etc/ceph/ceph.conf 2026-03-09T15:13:28.138 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 15:13:27 vm04 ceph-mon[53998]: Updating vm03:/var/lib/ceph/4abede80-1bca-11f1-bcd8-9b1da3c03ce5/config/ceph.conf 2026-03-09T15:13:28.138 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 15:13:27 vm04 ceph-mon[53998]: Updating vm06:/var/lib/ceph/4abede80-1bca-11f1-bcd8-9b1da3c03ce5/config/ceph.conf 2026-03-09T15:13:28.138 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 15:13:27 vm04 ceph-mon[53998]: Updating vm04:/var/lib/ceph/4abede80-1bca-11f1-bcd8-9b1da3c03ce5/config/ceph.conf 2026-03-09T15:13:28.138 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 15:13:27 vm04 ceph-mon[53998]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' 2026-03-09T15:13:28.138 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 15:13:27 vm04 ceph-mon[53998]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' 2026-03-09T15:13:28.138 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 15:13:27 vm04 ceph-mon[53998]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' 2026-03-09T15:13:28.138 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 15:13:27 vm04 ceph-mon[53998]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' 2026-03-09T15:13:28.138 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 15:13:27 vm04 ceph-mon[53998]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' 2026-03-09T15:13:28.138 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 15:13:27 vm04 ceph-mon[53998]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' 2026-03-09T15:13:28.138 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 15:13:27 vm04 ceph-mon[53998]: from='client.? 192.168.123.106:0/1702195684' entity='client.admin' cmd=[{"prefix": "mon dump", "format": "json"}]: dispatch 2026-03-09T15:13:28.138 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 15:13:27 vm04 ceph-mon[53998]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' 2026-03-09T15:13:28.138 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 15:13:27 vm04 ceph-mon[53998]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' 2026-03-09T15:13:28.138 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 15:13:27 vm04 ceph-mon[53998]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' 2026-03-09T15:13:28.138 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 15:13:27 vm04 ceph-mon[53998]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' 2026-03-09T15:13:28.138 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 15:13:27 vm04 ceph-mon[53998]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' 2026-03-09T15:13:28.138 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 15:13:27 vm04 ceph-mon[53998]: Reconfiguring mon.a (unknown last config time)... 2026-03-09T15:13:28.138 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 15:13:27 vm04 ceph-mon[53998]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "mon."}]: dispatch 2026-03-09T15:13:28.138 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 15:13:27 vm04 ceph-mon[53998]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' cmd=[{"prefix": "config get", "who": "mon", "key": "public_network"}]: dispatch 2026-03-09T15:13:28.138 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 15:13:27 vm04 ceph-mon[53998]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T15:13:28.138 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 15:13:27 vm04 ceph-mon[53998]: Reconfiguring daemon mon.a on vm03 2026-03-09T15:13:28.138 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 15:13:27 vm04 ceph-mon[53998]: pgmap v6: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-09T15:13:28.138 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 15:13:27 vm04 ceph-mon[53998]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' 2026-03-09T15:13:28.138 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 15:13:27 vm04 ceph-mon[53998]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' 2026-03-09T15:13:28.138 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 15:13:27 vm04 ceph-mon[53998]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "mon."}]: dispatch 2026-03-09T15:13:28.138 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 15:13:27 vm04 ceph-mon[53998]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' cmd=[{"prefix": "config get", "who": "mon", "key": "public_network"}]: dispatch 2026-03-09T15:13:28.138 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 15:13:27 vm04 ceph-mon[53998]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T15:13:28.138 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 15:13:27 vm04 ceph-mon[53998]: from='client.? 192.168.123.103:0/1490553382' entity='client.admin' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T15:13:28.138 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 15:13:27 vm04 ceph-mon[53998]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' 2026-03-09T15:13:28.138 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 15:13:27 vm04 ceph-mon[53998]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' 2026-03-09T15:13:28.138 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 15:13:27 vm04 ceph-mon[53998]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-09T15:13:28.138 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 15:13:27 vm04 ceph-mon[53998]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "mon."}]: dispatch 2026-03-09T15:13:28.138 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 15:13:27 vm04 ceph-mon[53998]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' cmd=[{"prefix": "config get", "who": "mon", "key": "public_network"}]: dispatch 2026-03-09T15:13:28.138 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 15:13:27 vm04 ceph-mon[53998]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T15:13:28.138 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 15:13:27 vm04 ceph-mon[53998]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' 2026-03-09T15:13:28.138 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 15:13:27 vm04 ceph-mon[53998]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' 2026-03-09T15:13:28.138 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 15:13:27 vm04 ceph-mon[53998]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T15:13:28.138 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 15:13:27 vm04 ceph-mon[53998]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T15:13:28.138 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 15:13:27 vm04 ceph-mon[53998]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T15:13:28.139 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 15:13:27 vm04 ceph-mon[53998]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' 2026-03-09T15:13:28.139 INFO:journalctl@ceph.mgr.b.vm04.stdout:Mar 09 15:13:28 vm04 systemd[1]: Starting Ceph mgr.b for 4abede80-1bca-11f1-bcd8-9b1da3c03ce5... 2026-03-09T15:13:28.169 INFO:teuthology.orchestra.run.vm03.stdout: File: /dev/vde 2026-03-09T15:13:28.169 INFO:teuthology.orchestra.run.vm03.stdout: Size: 0 Blocks: 0 IO Block: 512 block special file 2026-03-09T15:13:28.169 INFO:teuthology.orchestra.run.vm03.stdout:Device: 6h/6d Inode: 257 Links: 1 Device type: fc,40 2026-03-09T15:13:28.169 INFO:teuthology.orchestra.run.vm03.stdout:Access: (0660/brw-rw----) Uid: ( 0/ root) Gid: ( 6/ disk) 2026-03-09T15:13:28.169 INFO:teuthology.orchestra.run.vm03.stdout:Context: system_u:object_r:fixed_disk_device_t:s0 2026-03-09T15:13:28.169 INFO:teuthology.orchestra.run.vm03.stdout:Access: 2026-03-09 15:13:06.026063280 +0000 2026-03-09T15:13:28.169 INFO:teuthology.orchestra.run.vm03.stdout:Modify: 2026-03-09 15:10:28.162559411 +0000 2026-03-09T15:13:28.169 INFO:teuthology.orchestra.run.vm03.stdout:Change: 2026-03-09 15:10:28.162559411 +0000 2026-03-09T15:13:28.169 INFO:teuthology.orchestra.run.vm03.stdout: Birth: 2026-03-09 15:06:31.292000000 +0000 2026-03-09T15:13:28.169 DEBUG:teuthology.orchestra.run.vm03:> sudo dd if=/dev/vde of=/dev/null count=1 2026-03-09T15:13:28.242 INFO:teuthology.orchestra.run.vm03.stderr:1+0 records in 2026-03-09T15:13:28.242 INFO:teuthology.orchestra.run.vm03.stderr:1+0 records out 2026-03-09T15:13:28.242 INFO:teuthology.orchestra.run.vm03.stderr:512 bytes copied, 0.000120126 s, 4.3 MB/s 2026-03-09T15:13:28.243 DEBUG:teuthology.orchestra.run.vm03:> ! mount | grep -v devtmpfs | grep -q /dev/vde 2026-03-09T15:13:28.299 DEBUG:teuthology.orchestra.run.vm04:> set -ex 2026-03-09T15:13:28.299 DEBUG:teuthology.orchestra.run.vm04:> dd if=/scratch_devs of=/dev/stdout 2026-03-09T15:13:28.405 DEBUG:teuthology.orchestra.run:got remote process result: 1 2026-03-09T15:13:28.405 DEBUG:teuthology.orchestra.run.vm04:> ls /dev/[sv]d? 2026-03-09T15:13:28.451 INFO:teuthology.orchestra.run.vm04.stdout:/dev/vda 2026-03-09T15:13:28.451 INFO:teuthology.orchestra.run.vm04.stdout:/dev/vdb 2026-03-09T15:13:28.451 INFO:teuthology.orchestra.run.vm04.stdout:/dev/vdc 2026-03-09T15:13:28.451 INFO:teuthology.orchestra.run.vm04.stdout:/dev/vdd 2026-03-09T15:13:28.451 INFO:teuthology.orchestra.run.vm04.stdout:/dev/vde 2026-03-09T15:13:28.451 WARNING:teuthology.misc:Removing root device: /dev/vda from device list 2026-03-09T15:13:28.451 DEBUG:teuthology.misc:devs=['/dev/vdb', '/dev/vdc', '/dev/vdd', '/dev/vde'] 2026-03-09T15:13:28.451 DEBUG:teuthology.orchestra.run.vm04:> stat /dev/vdb 2026-03-09T15:13:28.480 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:13:28 vm03 ceph-mon[51762]: mon.b calling monitor election 2026-03-09T15:13:28.480 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 15:13:28 vm04 ceph-mon[53998]: mon.b calling monitor election 2026-03-09T15:13:28.480 INFO:journalctl@ceph.mgr.b.vm04.stdout:Mar 09 15:13:28 vm04 podman[54922]: 2026-03-09 15:13:28.135751609 +0000 UTC m=+0.015526800 container create 518720cdb99228601b5cbb6224235a125b9cc19b8ea3654608475ff01d469936 (image=quay.ceph.io/ceph-ci/ceph@sha256:8fda260ab1d2d3118a1622f7df75f44f285dfe74e71793626152a711c12bf2cc, name=ceph-4abede80-1bca-11f1-bcd8-9b1da3c03ce5-mgr-b, CEPH_REF=squid, org.label-schema.build-date=20260223, ceph=True, org.label-schema.vendor=CentOS, CEPH_GIT_REPO=https://github.com/ceph/ceph-ci.git, org.label-schema.name=CentOS Stream 9 Base Image, org.opencontainers.image.authors=Ceph Release Team , org.label-schema.schema-version=1.0, io.buildah.version=1.41.3, org.opencontainers.image.documentation=https://docs.ceph.com/, FROM_IMAGE=quay.io/centos/centos:stream9, OSD_FLAVOR=default, GANESHA_REPO_BASEURL=https://buildlogs.centos.org/centos/$releasever-stream/storage/$basearch/nfsganesha-5/, org.label-schema.license=GPLv2, CEPH_SHA1=e911bdebe5c8faa3800735d1568fcdca65db60df) 2026-03-09T15:13:28.480 INFO:journalctl@ceph.mgr.b.vm04.stdout:Mar 09 15:13:28 vm04 podman[54922]: 2026-03-09 15:13:28.170298945 +0000 UTC m=+0.050074136 container init 518720cdb99228601b5cbb6224235a125b9cc19b8ea3654608475ff01d469936 (image=quay.ceph.io/ceph-ci/ceph@sha256:8fda260ab1d2d3118a1622f7df75f44f285dfe74e71793626152a711c12bf2cc, name=ceph-4abede80-1bca-11f1-bcd8-9b1da3c03ce5-mgr-b, FROM_IMAGE=quay.io/centos/centos:stream9, io.buildah.version=1.41.3, org.opencontainers.image.documentation=https://docs.ceph.com/, ceph=True, org.label-schema.schema-version=1.0, GANESHA_REPO_BASEURL=https://buildlogs.centos.org/centos/$releasever-stream/storage/$basearch/nfsganesha-5/, CEPH_REF=squid, org.label-schema.name=CentOS Stream 9 Base Image, org.opencontainers.image.authors=Ceph Release Team , OSD_FLAVOR=default, org.label-schema.build-date=20260223, org.label-schema.vendor=CentOS, org.label-schema.license=GPLv2, CEPH_SHA1=e911bdebe5c8faa3800735d1568fcdca65db60df, CEPH_GIT_REPO=https://github.com/ceph/ceph-ci.git) 2026-03-09T15:13:28.480 INFO:journalctl@ceph.mgr.b.vm04.stdout:Mar 09 15:13:28 vm04 podman[54922]: 2026-03-09 15:13:28.173061424 +0000 UTC m=+0.052836615 container start 518720cdb99228601b5cbb6224235a125b9cc19b8ea3654608475ff01d469936 (image=quay.ceph.io/ceph-ci/ceph@sha256:8fda260ab1d2d3118a1622f7df75f44f285dfe74e71793626152a711c12bf2cc, name=ceph-4abede80-1bca-11f1-bcd8-9b1da3c03ce5-mgr-b, ceph=True, CEPH_SHA1=e911bdebe5c8faa3800735d1568fcdca65db60df, OSD_FLAVOR=default, org.label-schema.license=GPLv2, org.label-schema.schema-version=1.0, org.label-schema.vendor=CentOS, FROM_IMAGE=quay.io/centos/centos:stream9, io.buildah.version=1.41.3, org.opencontainers.image.documentation=https://docs.ceph.com/, GANESHA_REPO_BASEURL=https://buildlogs.centos.org/centos/$releasever-stream/storage/$basearch/nfsganesha-5/, CEPH_REF=squid, org.label-schema.build-date=20260223, CEPH_GIT_REPO=https://github.com/ceph/ceph-ci.git, org.label-schema.name=CentOS Stream 9 Base Image, org.opencontainers.image.authors=Ceph Release Team ) 2026-03-09T15:13:28.480 INFO:journalctl@ceph.mgr.b.vm04.stdout:Mar 09 15:13:28 vm04 bash[54922]: 518720cdb99228601b5cbb6224235a125b9cc19b8ea3654608475ff01d469936 2026-03-09T15:13:28.480 INFO:journalctl@ceph.mgr.b.vm04.stdout:Mar 09 15:13:28 vm04 podman[54922]: 2026-03-09 15:13:28.129060336 +0000 UTC m=+0.008835527 image pull 654f31e6858eb235bbece362255b685a945f2b6a367e2b88c4930c984fbb214c quay.ceph.io/ceph-ci/ceph@sha256:8fda260ab1d2d3118a1622f7df75f44f285dfe74e71793626152a711c12bf2cc 2026-03-09T15:13:28.480 INFO:journalctl@ceph.mgr.b.vm04.stdout:Mar 09 15:13:28 vm04 systemd[1]: Started Ceph mgr.b for 4abede80-1bca-11f1-bcd8-9b1da3c03ce5. 2026-03-09T15:13:28.480 INFO:journalctl@ceph.mgr.b.vm04.stdout:Mar 09 15:13:28 vm04 ceph-4abede80-1bca-11f1-bcd8-9b1da3c03ce5-mgr-b[54932]: 2026-03-09T15:13:28.272+0000 7f6d64ff6140 -1 mgr[py] Module alerts has missing NOTIFY_TYPES member 2026-03-09T15:13:28.480 INFO:journalctl@ceph.mgr.b.vm04.stdout:Mar 09 15:13:28 vm04 ceph-4abede80-1bca-11f1-bcd8-9b1da3c03ce5-mgr-b[54932]: 2026-03-09T15:13:28.316+0000 7f6d64ff6140 -1 mgr[py] Module balancer has missing NOTIFY_TYPES member 2026-03-09T15:13:28.534 INFO:teuthology.orchestra.run.vm04.stdout: File: /dev/vdb 2026-03-09T15:13:28.534 INFO:teuthology.orchestra.run.vm04.stdout: Size: 0 Blocks: 0 IO Block: 512 block special file 2026-03-09T15:13:28.534 INFO:teuthology.orchestra.run.vm04.stdout:Device: 6h/6d Inode: 254 Links: 1 Device type: fc,10 2026-03-09T15:13:28.534 INFO:teuthology.orchestra.run.vm04.stdout:Access: (0660/brw-rw----) Uid: ( 0/ root) Gid: ( 6/ disk) 2026-03-09T15:13:28.534 INFO:teuthology.orchestra.run.vm04.stdout:Context: system_u:object_r:fixed_disk_device_t:s0 2026-03-09T15:13:28.534 INFO:teuthology.orchestra.run.vm04.stdout:Access: 2026-03-09 15:13:10.799500503 +0000 2026-03-09T15:13:28.534 INFO:teuthology.orchestra.run.vm04.stdout:Modify: 2026-03-09 15:10:27.993013807 +0000 2026-03-09T15:13:28.534 INFO:teuthology.orchestra.run.vm04.stdout:Change: 2026-03-09 15:10:27.993013807 +0000 2026-03-09T15:13:28.534 INFO:teuthology.orchestra.run.vm04.stdout: Birth: 2026-03-09 15:07:02.254000000 +0000 2026-03-09T15:13:28.534 DEBUG:teuthology.orchestra.run.vm04:> sudo dd if=/dev/vdb of=/dev/null count=1 2026-03-09T15:13:28.560 INFO:teuthology.orchestra.run.vm04.stderr:1+0 records in 2026-03-09T15:13:28.560 INFO:teuthology.orchestra.run.vm04.stderr:1+0 records out 2026-03-09T15:13:28.560 INFO:teuthology.orchestra.run.vm04.stderr:512 bytes copied, 0.000183223 s, 2.8 MB/s 2026-03-09T15:13:28.561 DEBUG:teuthology.orchestra.run.vm04:> ! mount | grep -v devtmpfs | grep -q /dev/vdb 2026-03-09T15:13:28.642 DEBUG:teuthology.orchestra.run.vm04:> stat /dev/vdc 2026-03-09T15:13:28.726 INFO:teuthology.orchestra.run.vm04.stdout: File: /dev/vdc 2026-03-09T15:13:28.726 INFO:teuthology.orchestra.run.vm04.stdout: Size: 0 Blocks: 0 IO Block: 512 block special file 2026-03-09T15:13:28.726 INFO:teuthology.orchestra.run.vm04.stdout:Device: 6h/6d Inode: 255 Links: 1 Device type: fc,20 2026-03-09T15:13:28.726 INFO:teuthology.orchestra.run.vm04.stdout:Access: (0660/brw-rw----) Uid: ( 0/ root) Gid: ( 6/ disk) 2026-03-09T15:13:28.726 INFO:teuthology.orchestra.run.vm04.stdout:Context: system_u:object_r:fixed_disk_device_t:s0 2026-03-09T15:13:28.726 INFO:teuthology.orchestra.run.vm04.stdout:Access: 2026-03-09 15:13:10.827500525 +0000 2026-03-09T15:13:28.726 INFO:teuthology.orchestra.run.vm04.stdout:Modify: 2026-03-09 15:10:27.997013814 +0000 2026-03-09T15:13:28.726 INFO:teuthology.orchestra.run.vm04.stdout:Change: 2026-03-09 15:10:27.997013814 +0000 2026-03-09T15:13:28.726 INFO:teuthology.orchestra.run.vm04.stdout: Birth: 2026-03-09 15:07:02.259000000 +0000 2026-03-09T15:13:28.726 DEBUG:teuthology.orchestra.run.vm04:> sudo dd if=/dev/vdc of=/dev/null count=1 2026-03-09T15:13:28.750 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 15:13:28 vm04 ceph-mon[53998]: mon.b calling monitor election 2026-03-09T15:13:28.750 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 15:13:28 vm04 ceph-mon[53998]: mon.c calling monitor election 2026-03-09T15:13:28.750 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 15:13:28 vm04 ceph-mon[53998]: mon.a calling monitor election 2026-03-09T15:13:28.750 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 15:13:28 vm04 ceph-mon[53998]: mon.a is new leader, mons a,c,b in quorum (ranks 0,1,2) 2026-03-09T15:13:28.750 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 15:13:28 vm04 ceph-mon[53998]: monmap epoch 3 2026-03-09T15:13:28.750 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 15:13:28 vm04 ceph-mon[53998]: fsid 4abede80-1bca-11f1-bcd8-9b1da3c03ce5 2026-03-09T15:13:28.750 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 15:13:28 vm04 ceph-mon[53998]: last_changed 2026-03-09T15:13:20.804785+0000 2026-03-09T15:13:28.750 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 15:13:28 vm04 ceph-mon[53998]: created 2026-03-09T15:12:40.941469+0000 2026-03-09T15:13:28.750 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 15:13:28 vm04 ceph-mon[53998]: min_mon_release 19 (squid) 2026-03-09T15:13:28.750 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 15:13:28 vm04 ceph-mon[53998]: election_strategy: 1 2026-03-09T15:13:28.750 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 15:13:28 vm04 ceph-mon[53998]: 0: [v2:192.168.123.103:3300/0,v1:192.168.123.103:6789/0] mon.a 2026-03-09T15:13:28.750 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 15:13:28 vm04 ceph-mon[53998]: 1: [v2:192.168.123.106:3300/0,v1:192.168.123.106:6789/0] mon.c 2026-03-09T15:13:28.750 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 15:13:28 vm04 ceph-mon[53998]: 2: [v2:192.168.123.104:3300/0,v1:192.168.123.104:6789/0] mon.b 2026-03-09T15:13:28.750 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 15:13:28 vm04 ceph-mon[53998]: fsmap 2026-03-09T15:13:28.750 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 15:13:28 vm04 ceph-mon[53998]: osdmap e4: 0 total, 0 up, 0 in 2026-03-09T15:13:28.750 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 15:13:28 vm04 ceph-mon[53998]: mgrmap e12: a(active, since 25s) 2026-03-09T15:13:28.750 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 15:13:28 vm04 ceph-mon[53998]: overall HEALTH_OK 2026-03-09T15:13:28.750 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 15:13:28 vm04 ceph-mon[53998]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' 2026-03-09T15:13:28.750 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 15:13:28 vm04 ceph-mon[53998]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' 2026-03-09T15:13:28.750 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 15:13:28 vm04 ceph-mon[53998]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' 2026-03-09T15:13:28.750 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 15:13:28 vm04 ceph-mon[53998]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' 2026-03-09T15:13:28.750 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 15:13:28 vm04 ceph-mon[53998]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T15:13:28.750 INFO:journalctl@ceph.mgr.b.vm04.stdout:Mar 09 15:13:28 vm04 ceph-4abede80-1bca-11f1-bcd8-9b1da3c03ce5-mgr-b[54932]: 2026-03-09T15:13:28.747+0000 7f6d64ff6140 -1 mgr[py] Module crash has missing NOTIFY_TYPES member 2026-03-09T15:13:28.767 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 09 15:13:28 vm06 ceph-mon[56506]: mon.b calling monitor election 2026-03-09T15:13:28.767 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 09 15:13:28 vm06 ceph-mon[56506]: mon.b calling monitor election 2026-03-09T15:13:28.767 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 09 15:13:28 vm06 ceph-mon[56506]: mon.c calling monitor election 2026-03-09T15:13:28.767 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 09 15:13:28 vm06 ceph-mon[56506]: mon.a calling monitor election 2026-03-09T15:13:28.767 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 09 15:13:28 vm06 ceph-mon[56506]: mon.a is new leader, mons a,c,b in quorum (ranks 0,1,2) 2026-03-09T15:13:28.767 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 09 15:13:28 vm06 ceph-mon[56506]: monmap epoch 3 2026-03-09T15:13:28.767 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 09 15:13:28 vm06 ceph-mon[56506]: fsid 4abede80-1bca-11f1-bcd8-9b1da3c03ce5 2026-03-09T15:13:28.767 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 09 15:13:28 vm06 ceph-mon[56506]: last_changed 2026-03-09T15:13:20.804785+0000 2026-03-09T15:13:28.767 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 09 15:13:28 vm06 ceph-mon[56506]: created 2026-03-09T15:12:40.941469+0000 2026-03-09T15:13:28.767 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 09 15:13:28 vm06 ceph-mon[56506]: min_mon_release 19 (squid) 2026-03-09T15:13:28.767 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 09 15:13:28 vm06 ceph-mon[56506]: election_strategy: 1 2026-03-09T15:13:28.767 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 09 15:13:28 vm06 ceph-mon[56506]: 0: [v2:192.168.123.103:3300/0,v1:192.168.123.103:6789/0] mon.a 2026-03-09T15:13:28.767 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 09 15:13:28 vm06 ceph-mon[56506]: 1: [v2:192.168.123.106:3300/0,v1:192.168.123.106:6789/0] mon.c 2026-03-09T15:13:28.767 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 09 15:13:28 vm06 ceph-mon[56506]: 2: [v2:192.168.123.104:3300/0,v1:192.168.123.104:6789/0] mon.b 2026-03-09T15:13:28.767 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 09 15:13:28 vm06 ceph-mon[56506]: fsmap 2026-03-09T15:13:28.767 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 09 15:13:28 vm06 ceph-mon[56506]: osdmap e4: 0 total, 0 up, 0 in 2026-03-09T15:13:28.767 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 09 15:13:28 vm06 ceph-mon[56506]: mgrmap e12: a(active, since 25s) 2026-03-09T15:13:28.767 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 09 15:13:28 vm06 ceph-mon[56506]: overall HEALTH_OK 2026-03-09T15:13:28.767 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 09 15:13:28 vm06 ceph-mon[56506]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' 2026-03-09T15:13:28.767 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 09 15:13:28 vm06 ceph-mon[56506]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' 2026-03-09T15:13:28.767 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 09 15:13:28 vm06 ceph-mon[56506]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' 2026-03-09T15:13:28.767 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 09 15:13:28 vm06 ceph-mon[56506]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' 2026-03-09T15:13:28.767 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 09 15:13:28 vm06 ceph-mon[56506]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T15:13:28.781 INFO:teuthology.orchestra.run.vm04.stderr:1+0 records in 2026-03-09T15:13:28.781 INFO:teuthology.orchestra.run.vm04.stderr:1+0 records out 2026-03-09T15:13:28.781 INFO:teuthology.orchestra.run.vm04.stderr:512 bytes copied, 0.000608689 s, 841 kB/s 2026-03-09T15:13:28.782 DEBUG:teuthology.orchestra.run.vm04:> ! mount | grep -v devtmpfs | grep -q /dev/vdc 2026-03-09T15:13:28.799 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:13:28 vm03 ceph-mon[51762]: mon.b calling monitor election 2026-03-09T15:13:28.799 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:13:28 vm03 ceph-mon[51762]: mon.c calling monitor election 2026-03-09T15:13:28.799 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:13:28 vm03 ceph-mon[51762]: mon.a calling monitor election 2026-03-09T15:13:28.799 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:13:28 vm03 ceph-mon[51762]: mon.a is new leader, mons a,c,b in quorum (ranks 0,1,2) 2026-03-09T15:13:28.799 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:13:28 vm03 ceph-mon[51762]: monmap epoch 3 2026-03-09T15:13:28.799 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:13:28 vm03 ceph-mon[51762]: fsid 4abede80-1bca-11f1-bcd8-9b1da3c03ce5 2026-03-09T15:13:28.799 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:13:28 vm03 ceph-mon[51762]: last_changed 2026-03-09T15:13:20.804785+0000 2026-03-09T15:13:28.799 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:13:28 vm03 ceph-mon[51762]: created 2026-03-09T15:12:40.941469+0000 2026-03-09T15:13:28.799 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:13:28 vm03 ceph-mon[51762]: min_mon_release 19 (squid) 2026-03-09T15:13:28.799 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:13:28 vm03 ceph-mon[51762]: election_strategy: 1 2026-03-09T15:13:28.799 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:13:28 vm03 ceph-mon[51762]: 0: [v2:192.168.123.103:3300/0,v1:192.168.123.103:6789/0] mon.a 2026-03-09T15:13:28.799 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:13:28 vm03 ceph-mon[51762]: 1: [v2:192.168.123.106:3300/0,v1:192.168.123.106:6789/0] mon.c 2026-03-09T15:13:28.799 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:13:28 vm03 ceph-mon[51762]: 2: [v2:192.168.123.104:3300/0,v1:192.168.123.104:6789/0] mon.b 2026-03-09T15:13:28.799 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:13:28 vm03 ceph-mon[51762]: fsmap 2026-03-09T15:13:28.799 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:13:28 vm03 ceph-mon[51762]: osdmap e4: 0 total, 0 up, 0 in 2026-03-09T15:13:28.799 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:13:28 vm03 ceph-mon[51762]: mgrmap e12: a(active, since 25s) 2026-03-09T15:13:28.800 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:13:28 vm03 ceph-mon[51762]: overall HEALTH_OK 2026-03-09T15:13:28.800 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:13:28 vm03 ceph-mon[51762]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' 2026-03-09T15:13:28.800 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:13:28 vm03 ceph-mon[51762]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' 2026-03-09T15:13:28.800 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:13:28 vm03 ceph-mon[51762]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' 2026-03-09T15:13:28.800 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:13:28 vm03 ceph-mon[51762]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' 2026-03-09T15:13:28.800 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:13:28 vm03 ceph-mon[51762]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T15:13:28.847 DEBUG:teuthology.orchestra.run.vm04:> stat /dev/vdd 2026-03-09T15:13:28.906 INFO:teuthology.orchestra.run.vm04.stdout: File: /dev/vdd 2026-03-09T15:13:28.906 INFO:teuthology.orchestra.run.vm04.stdout: Size: 0 Blocks: 0 IO Block: 512 block special file 2026-03-09T15:13:28.906 INFO:teuthology.orchestra.run.vm04.stdout:Device: 6h/6d Inode: 256 Links: 1 Device type: fc,30 2026-03-09T15:13:28.906 INFO:teuthology.orchestra.run.vm04.stdout:Access: (0660/brw-rw----) Uid: ( 0/ root) Gid: ( 6/ disk) 2026-03-09T15:13:28.906 INFO:teuthology.orchestra.run.vm04.stdout:Context: system_u:object_r:fixed_disk_device_t:s0 2026-03-09T15:13:28.906 INFO:teuthology.orchestra.run.vm04.stdout:Access: 2026-03-09 15:13:10.854500547 +0000 2026-03-09T15:13:28.906 INFO:teuthology.orchestra.run.vm04.stdout:Modify: 2026-03-09 15:10:28.000013819 +0000 2026-03-09T15:13:28.906 INFO:teuthology.orchestra.run.vm04.stdout:Change: 2026-03-09 15:10:28.000013819 +0000 2026-03-09T15:13:28.906 INFO:teuthology.orchestra.run.vm04.stdout: Birth: 2026-03-09 15:07:02.263000000 +0000 2026-03-09T15:13:28.906 DEBUG:teuthology.orchestra.run.vm04:> sudo dd if=/dev/vdd of=/dev/null count=1 2026-03-09T15:13:28.971 INFO:teuthology.orchestra.run.vm04.stderr:1+0 records in 2026-03-09T15:13:28.971 INFO:teuthology.orchestra.run.vm04.stderr:1+0 records out 2026-03-09T15:13:28.971 INFO:teuthology.orchestra.run.vm04.stderr:512 bytes copied, 0.00018673 s, 2.7 MB/s 2026-03-09T15:13:28.972 DEBUG:teuthology.orchestra.run.vm04:> ! mount | grep -v devtmpfs | grep -q /dev/vdd 2026-03-09T15:13:29.030 DEBUG:teuthology.orchestra.run.vm04:> stat /dev/vde 2026-03-09T15:13:29.092 INFO:teuthology.orchestra.run.vm04.stdout: File: /dev/vde 2026-03-09T15:13:29.092 INFO:teuthology.orchestra.run.vm04.stdout: Size: 0 Blocks: 0 IO Block: 512 block special file 2026-03-09T15:13:29.092 INFO:teuthology.orchestra.run.vm04.stdout:Device: 6h/6d Inode: 257 Links: 1 Device type: fc,40 2026-03-09T15:13:29.092 INFO:teuthology.orchestra.run.vm04.stdout:Access: (0660/brw-rw----) Uid: ( 0/ root) Gid: ( 6/ disk) 2026-03-09T15:13:29.092 INFO:teuthology.orchestra.run.vm04.stdout:Context: system_u:object_r:fixed_disk_device_t:s0 2026-03-09T15:13:29.092 INFO:teuthology.orchestra.run.vm04.stdout:Access: 2026-03-09 15:13:10.882500570 +0000 2026-03-09T15:13:29.092 INFO:teuthology.orchestra.run.vm04.stdout:Modify: 2026-03-09 15:10:28.001013820 +0000 2026-03-09T15:13:29.092 INFO:teuthology.orchestra.run.vm04.stdout:Change: 2026-03-09 15:10:28.001013820 +0000 2026-03-09T15:13:29.092 INFO:teuthology.orchestra.run.vm04.stdout: Birth: 2026-03-09 15:07:02.288000000 +0000 2026-03-09T15:13:29.092 DEBUG:teuthology.orchestra.run.vm04:> sudo dd if=/dev/vde of=/dev/null count=1 2026-03-09T15:13:29.157 INFO:teuthology.orchestra.run.vm04.stderr:1+0 records in 2026-03-09T15:13:29.157 INFO:teuthology.orchestra.run.vm04.stderr:1+0 records out 2026-03-09T15:13:29.157 INFO:teuthology.orchestra.run.vm04.stderr:512 bytes copied, 0.000165299 s, 3.1 MB/s 2026-03-09T15:13:29.158 DEBUG:teuthology.orchestra.run.vm04:> ! mount | grep -v devtmpfs | grep -q /dev/vde 2026-03-09T15:13:29.225 DEBUG:teuthology.orchestra.run.vm06:> set -ex 2026-03-09T15:13:29.225 DEBUG:teuthology.orchestra.run.vm06:> dd if=/scratch_devs of=/dev/stdout 2026-03-09T15:13:29.243 DEBUG:teuthology.orchestra.run:got remote process result: 1 2026-03-09T15:13:29.243 DEBUG:teuthology.orchestra.run.vm06:> ls /dev/[sv]d? 2026-03-09T15:13:29.300 INFO:teuthology.orchestra.run.vm06.stdout:/dev/vda 2026-03-09T15:13:29.300 INFO:teuthology.orchestra.run.vm06.stdout:/dev/vdb 2026-03-09T15:13:29.300 INFO:teuthology.orchestra.run.vm06.stdout:/dev/vdc 2026-03-09T15:13:29.300 INFO:teuthology.orchestra.run.vm06.stdout:/dev/vdd 2026-03-09T15:13:29.300 INFO:teuthology.orchestra.run.vm06.stdout:/dev/vde 2026-03-09T15:13:29.300 WARNING:teuthology.misc:Removing root device: /dev/vda from device list 2026-03-09T15:13:29.300 DEBUG:teuthology.misc:devs=['/dev/vdb', '/dev/vdc', '/dev/vdd', '/dev/vde'] 2026-03-09T15:13:29.300 DEBUG:teuthology.orchestra.run.vm06:> stat /dev/vdb 2026-03-09T15:13:29.357 INFO:teuthology.orchestra.run.vm06.stdout: File: /dev/vdb 2026-03-09T15:13:29.357 INFO:teuthology.orchestra.run.vm06.stdout: Size: 0 Blocks: 0 IO Block: 512 block special file 2026-03-09T15:13:29.357 INFO:teuthology.orchestra.run.vm06.stdout:Device: 6h/6d Inode: 254 Links: 1 Device type: fc,10 2026-03-09T15:13:29.357 INFO:teuthology.orchestra.run.vm06.stdout:Access: (0660/brw-rw----) Uid: ( 0/ root) Gid: ( 6/ disk) 2026-03-09T15:13:29.357 INFO:teuthology.orchestra.run.vm06.stdout:Context: system_u:object_r:fixed_disk_device_t:s0 2026-03-09T15:13:29.357 INFO:teuthology.orchestra.run.vm06.stdout:Access: 2026-03-09 15:13:13.338215540 +0000 2026-03-09T15:13:29.357 INFO:teuthology.orchestra.run.vm06.stdout:Modify: 2026-03-09 15:10:29.030007155 +0000 2026-03-09T15:13:29.357 INFO:teuthology.orchestra.run.vm06.stdout:Change: 2026-03-09 15:10:29.030007155 +0000 2026-03-09T15:13:29.357 INFO:teuthology.orchestra.run.vm06.stdout: Birth: 2026-03-09 15:07:32.219000000 +0000 2026-03-09T15:13:29.357 DEBUG:teuthology.orchestra.run.vm06:> sudo dd if=/dev/vdb of=/dev/null count=1 2026-03-09T15:13:29.420 INFO:teuthology.orchestra.run.vm06.stderr:1+0 records in 2026-03-09T15:13:29.420 INFO:teuthology.orchestra.run.vm06.stderr:1+0 records out 2026-03-09T15:13:29.420 INFO:teuthology.orchestra.run.vm06.stderr:512 bytes copied, 0.000161021 s, 3.2 MB/s 2026-03-09T15:13:29.421 DEBUG:teuthology.orchestra.run.vm06:> ! mount | grep -v devtmpfs | grep -q /dev/vdb 2026-03-09T15:13:29.477 DEBUG:teuthology.orchestra.run.vm06:> stat /dev/vdc 2026-03-09T15:13:29.485 INFO:journalctl@ceph.mgr.b.vm04.stdout:Mar 09 15:13:29 vm04 ceph-4abede80-1bca-11f1-bcd8-9b1da3c03ce5-mgr-b[54932]: 2026-03-09T15:13:29.087+0000 7f6d64ff6140 -1 mgr[py] Module devicehealth has missing NOTIFY_TYPES member 2026-03-09T15:13:29.485 INFO:journalctl@ceph.mgr.b.vm04.stdout:Mar 09 15:13:29 vm04 ceph-4abede80-1bca-11f1-bcd8-9b1da3c03ce5-mgr-b[54932]: /lib64/python3.9/site-packages/scipy/__init__.py:73: UserWarning: NumPy was imported from a Python sub-interpreter but NumPy does not properly support sub-interpreters. This will likely work for most users but might cause hard to track down issues or subtle bugs. A common user of the rare sub-interpreter feature is wsgi which also allows single-interpreter mode. 2026-03-09T15:13:29.485 INFO:journalctl@ceph.mgr.b.vm04.stdout:Mar 09 15:13:29 vm04 ceph-4abede80-1bca-11f1-bcd8-9b1da3c03ce5-mgr-b[54932]: Improvements in the case of bugs are welcome, but is not on the NumPy roadmap, and full support may require significant effort to achieve. 2026-03-09T15:13:29.485 INFO:journalctl@ceph.mgr.b.vm04.stdout:Mar 09 15:13:29 vm04 ceph-4abede80-1bca-11f1-bcd8-9b1da3c03ce5-mgr-b[54932]: from numpy import show_config as show_numpy_config 2026-03-09T15:13:29.485 INFO:journalctl@ceph.mgr.b.vm04.stdout:Mar 09 15:13:29 vm04 ceph-4abede80-1bca-11f1-bcd8-9b1da3c03ce5-mgr-b[54932]: 2026-03-09T15:13:29.180+0000 7f6d64ff6140 -1 mgr[py] Module diskprediction_local has missing NOTIFY_TYPES member 2026-03-09T15:13:29.485 INFO:journalctl@ceph.mgr.b.vm04.stdout:Mar 09 15:13:29 vm04 ceph-4abede80-1bca-11f1-bcd8-9b1da3c03ce5-mgr-b[54932]: 2026-03-09T15:13:29.216+0000 7f6d64ff6140 -1 mgr[py] Module influx has missing NOTIFY_TYPES member 2026-03-09T15:13:29.485 INFO:journalctl@ceph.mgr.b.vm04.stdout:Mar 09 15:13:29 vm04 ceph-4abede80-1bca-11f1-bcd8-9b1da3c03ce5-mgr-b[54932]: 2026-03-09T15:13:29.281+0000 7f6d64ff6140 -1 mgr[py] Module iostat has missing NOTIFY_TYPES member 2026-03-09T15:13:29.533 INFO:teuthology.orchestra.run.vm06.stdout: File: /dev/vdc 2026-03-09T15:13:29.533 INFO:teuthology.orchestra.run.vm06.stdout: Size: 0 Blocks: 0 IO Block: 512 block special file 2026-03-09T15:13:29.533 INFO:teuthology.orchestra.run.vm06.stdout:Device: 6h/6d Inode: 255 Links: 1 Device type: fc,20 2026-03-09T15:13:29.533 INFO:teuthology.orchestra.run.vm06.stdout:Access: (0660/brw-rw----) Uid: ( 0/ root) Gid: ( 6/ disk) 2026-03-09T15:13:29.533 INFO:teuthology.orchestra.run.vm06.stdout:Context: system_u:object_r:fixed_disk_device_t:s0 2026-03-09T15:13:29.533 INFO:teuthology.orchestra.run.vm06.stdout:Access: 2026-03-09 15:13:13.369215588 +0000 2026-03-09T15:13:29.533 INFO:teuthology.orchestra.run.vm06.stdout:Modify: 2026-03-09 15:10:29.037007167 +0000 2026-03-09T15:13:29.533 INFO:teuthology.orchestra.run.vm06.stdout:Change: 2026-03-09 15:10:29.037007167 +0000 2026-03-09T15:13:29.533 INFO:teuthology.orchestra.run.vm06.stdout: Birth: 2026-03-09 15:07:32.229000000 +0000 2026-03-09T15:13:29.533 DEBUG:teuthology.orchestra.run.vm06:> sudo dd if=/dev/vdc of=/dev/null count=1 2026-03-09T15:13:29.595 INFO:teuthology.orchestra.run.vm06.stderr:1+0 records in 2026-03-09T15:13:29.595 INFO:teuthology.orchestra.run.vm06.stderr:1+0 records out 2026-03-09T15:13:29.595 INFO:teuthology.orchestra.run.vm06.stderr:512 bytes copied, 0.000179106 s, 2.9 MB/s 2026-03-09T15:13:29.596 DEBUG:teuthology.orchestra.run.vm06:> ! mount | grep -v devtmpfs | grep -q /dev/vdc 2026-03-09T15:13:29.652 DEBUG:teuthology.orchestra.run.vm06:> stat /dev/vdd 2026-03-09T15:13:29.708 INFO:teuthology.orchestra.run.vm06.stdout: File: /dev/vdd 2026-03-09T15:13:29.708 INFO:teuthology.orchestra.run.vm06.stdout: Size: 0 Blocks: 0 IO Block: 512 block special file 2026-03-09T15:13:29.708 INFO:teuthology.orchestra.run.vm06.stdout:Device: 6h/6d Inode: 256 Links: 1 Device type: fc,30 2026-03-09T15:13:29.708 INFO:teuthology.orchestra.run.vm06.stdout:Access: (0660/brw-rw----) Uid: ( 0/ root) Gid: ( 6/ disk) 2026-03-09T15:13:29.708 INFO:teuthology.orchestra.run.vm06.stdout:Context: system_u:object_r:fixed_disk_device_t:s0 2026-03-09T15:13:29.708 INFO:teuthology.orchestra.run.vm06.stdout:Access: 2026-03-09 15:13:13.398215634 +0000 2026-03-09T15:13:29.708 INFO:teuthology.orchestra.run.vm06.stdout:Modify: 2026-03-09 15:10:29.063007213 +0000 2026-03-09T15:13:29.708 INFO:teuthology.orchestra.run.vm06.stdout:Change: 2026-03-09 15:10:29.063007213 +0000 2026-03-09T15:13:29.708 INFO:teuthology.orchestra.run.vm06.stdout: Birth: 2026-03-09 15:07:32.238000000 +0000 2026-03-09T15:13:29.708 DEBUG:teuthology.orchestra.run.vm06:> sudo dd if=/dev/vdd of=/dev/null count=1 2026-03-09T15:13:29.751 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 15:13:29 vm04 ceph-mon[53998]: pgmap v7: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-09T15:13:29.751 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 15:13:29 vm04 ceph-mon[53998]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-09T15:13:29.751 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 15:13:29 vm04 ceph-mon[53998]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' 2026-03-09T15:13:29.751 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 15:13:29 vm04 ceph-mon[53998]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' 2026-03-09T15:13:29.751 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 15:13:29 vm04 ceph-mon[53998]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T15:13:29.751 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 15:13:29 vm04 ceph-mon[53998]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T15:13:29.751 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 15:13:29 vm04 ceph-mon[53998]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' 2026-03-09T15:13:29.751 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 15:13:29 vm04 ceph-mon[53998]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' cmd=[{"prefix": "auth get-or-create", "entity": "mgr.a", "caps": ["mon", "profile mgr", "osd", "allow *", "mds", "allow *"]}]: dispatch 2026-03-09T15:13:29.751 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 15:13:29 vm04 ceph-mon[53998]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' cmd=[{"prefix": "mgr services"}]: dispatch 2026-03-09T15:13:29.751 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 15:13:29 vm04 ceph-mon[53998]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T15:13:29.751 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 15:13:29 vm04 ceph-mon[53998]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' 2026-03-09T15:13:29.751 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 15:13:29 vm04 ceph-mon[53998]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' 2026-03-09T15:13:29.751 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 15:13:29 vm04 ceph-mon[53998]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T15:13:29.751 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 15:13:29 vm04 ceph-mon[53998]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T15:13:29.751 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 15:13:29 vm04 ceph-mon[53998]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T15:13:29.751 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 15:13:29 vm04 ceph-mon[53998]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' 2026-03-09T15:13:29.767 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 09 15:13:29 vm06 ceph-mon[56506]: pgmap v7: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-09T15:13:29.767 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 09 15:13:29 vm06 ceph-mon[56506]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-09T15:13:29.767 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 09 15:13:29 vm06 ceph-mon[56506]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' 2026-03-09T15:13:29.767 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 09 15:13:29 vm06 ceph-mon[56506]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' 2026-03-09T15:13:29.767 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 09 15:13:29 vm06 ceph-mon[56506]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T15:13:29.767 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 09 15:13:29 vm06 ceph-mon[56506]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T15:13:29.767 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 09 15:13:29 vm06 ceph-mon[56506]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' 2026-03-09T15:13:29.767 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 09 15:13:29 vm06 ceph-mon[56506]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' cmd=[{"prefix": "auth get-or-create", "entity": "mgr.a", "caps": ["mon", "profile mgr", "osd", "allow *", "mds", "allow *"]}]: dispatch 2026-03-09T15:13:29.767 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 09 15:13:29 vm06 ceph-mon[56506]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' cmd=[{"prefix": "mgr services"}]: dispatch 2026-03-09T15:13:29.767 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 09 15:13:29 vm06 ceph-mon[56506]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T15:13:29.767 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 09 15:13:29 vm06 ceph-mon[56506]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' 2026-03-09T15:13:29.767 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 09 15:13:29 vm06 ceph-mon[56506]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' 2026-03-09T15:13:29.767 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 09 15:13:29 vm06 ceph-mon[56506]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T15:13:29.767 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 09 15:13:29 vm06 ceph-mon[56506]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T15:13:29.767 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 09 15:13:29 vm06 ceph-mon[56506]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T15:13:29.767 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 09 15:13:29 vm06 ceph-mon[56506]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' 2026-03-09T15:13:29.770 INFO:teuthology.orchestra.run.vm06.stderr:1+0 records in 2026-03-09T15:13:29.770 INFO:teuthology.orchestra.run.vm06.stderr:1+0 records out 2026-03-09T15:13:29.770 INFO:teuthology.orchestra.run.vm06.stderr:512 bytes copied, 0.000143708 s, 3.6 MB/s 2026-03-09T15:13:29.771 DEBUG:teuthology.orchestra.run.vm06:> ! mount | grep -v devtmpfs | grep -q /dev/vdd 2026-03-09T15:13:29.799 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:13:29 vm03 ceph-mon[51762]: pgmap v7: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-09T15:13:29.799 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:13:29 vm03 ceph-mon[51762]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-09T15:13:29.799 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:13:29 vm03 ceph-mon[51762]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' 2026-03-09T15:13:29.799 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:13:29 vm03 ceph-mon[51762]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' 2026-03-09T15:13:29.799 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:13:29 vm03 ceph-mon[51762]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T15:13:29.799 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:13:29 vm03 ceph-mon[51762]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T15:13:29.799 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:13:29 vm03 ceph-mon[51762]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' 2026-03-09T15:13:29.799 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:13:29 vm03 ceph-mon[51762]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' cmd=[{"prefix": "auth get-or-create", "entity": "mgr.a", "caps": ["mon", "profile mgr", "osd", "allow *", "mds", "allow *"]}]: dispatch 2026-03-09T15:13:29.799 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:13:29 vm03 ceph-mon[51762]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' cmd=[{"prefix": "mgr services"}]: dispatch 2026-03-09T15:13:29.799 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:13:29 vm03 ceph-mon[51762]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T15:13:29.799 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:13:29 vm03 ceph-mon[51762]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' 2026-03-09T15:13:29.800 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:13:29 vm03 ceph-mon[51762]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' 2026-03-09T15:13:29.800 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:13:29 vm03 ceph-mon[51762]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T15:13:29.800 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:13:29 vm03 ceph-mon[51762]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T15:13:29.800 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:13:29 vm03 ceph-mon[51762]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T15:13:29.800 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:13:29 vm03 ceph-mon[51762]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' 2026-03-09T15:13:29.826 DEBUG:teuthology.orchestra.run.vm06:> stat /dev/vde 2026-03-09T15:13:29.884 INFO:teuthology.orchestra.run.vm06.stdout: File: /dev/vde 2026-03-09T15:13:29.885 INFO:teuthology.orchestra.run.vm06.stdout: Size: 0 Blocks: 0 IO Block: 512 block special file 2026-03-09T15:13:29.885 INFO:teuthology.orchestra.run.vm06.stdout:Device: 6h/6d Inode: 257 Links: 1 Device type: fc,40 2026-03-09T15:13:29.885 INFO:teuthology.orchestra.run.vm06.stdout:Access: (0660/brw-rw----) Uid: ( 0/ root) Gid: ( 6/ disk) 2026-03-09T15:13:29.885 INFO:teuthology.orchestra.run.vm06.stdout:Context: system_u:object_r:fixed_disk_device_t:s0 2026-03-09T15:13:29.885 INFO:teuthology.orchestra.run.vm06.stdout:Access: 2026-03-09 15:13:13.433215688 +0000 2026-03-09T15:13:29.885 INFO:teuthology.orchestra.run.vm06.stdout:Modify: 2026-03-09 15:10:29.056007201 +0000 2026-03-09T15:13:29.885 INFO:teuthology.orchestra.run.vm06.stdout:Change: 2026-03-09 15:10:29.056007201 +0000 2026-03-09T15:13:29.885 INFO:teuthology.orchestra.run.vm06.stdout: Birth: 2026-03-09 15:07:32.241000000 +0000 2026-03-09T15:13:29.885 DEBUG:teuthology.orchestra.run.vm06:> sudo dd if=/dev/vde of=/dev/null count=1 2026-03-09T15:13:29.949 INFO:teuthology.orchestra.run.vm06.stderr:1+0 records in 2026-03-09T15:13:29.949 INFO:teuthology.orchestra.run.vm06.stderr:1+0 records out 2026-03-09T15:13:29.949 INFO:teuthology.orchestra.run.vm06.stderr:512 bytes copied, 0.000163255 s, 3.1 MB/s 2026-03-09T15:13:29.950 DEBUG:teuthology.orchestra.run.vm06:> ! mount | grep -v devtmpfs | grep -q /dev/vde 2026-03-09T15:13:30.008 INFO:tasks.cephadm:Deploying osd.0 on vm03 with /dev/vde... 2026-03-09T15:13:30.008 DEBUG:teuthology.orchestra.run.vm03:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df ceph-volume -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 4abede80-1bca-11f1-bcd8-9b1da3c03ce5 -- lvm zap /dev/vde 2026-03-09T15:13:30.166 INFO:journalctl@ceph.mgr.b.vm04.stdout:Mar 09 15:13:29 vm04 ceph-4abede80-1bca-11f1-bcd8-9b1da3c03ce5-mgr-b[54932]: 2026-03-09T15:13:29.753+0000 7f6d64ff6140 -1 mgr[py] Module nfs has missing NOTIFY_TYPES member 2026-03-09T15:13:30.167 INFO:journalctl@ceph.mgr.b.vm04.stdout:Mar 09 15:13:29 vm04 ceph-4abede80-1bca-11f1-bcd8-9b1da3c03ce5-mgr-b[54932]: 2026-03-09T15:13:29.856+0000 7f6d64ff6140 -1 mgr[py] Module orchestrator has missing NOTIFY_TYPES member 2026-03-09T15:13:30.167 INFO:journalctl@ceph.mgr.b.vm04.stdout:Mar 09 15:13:29 vm04 ceph-4abede80-1bca-11f1-bcd8-9b1da3c03ce5-mgr-b[54932]: 2026-03-09T15:13:29.895+0000 7f6d64ff6140 -1 mgr[py] Module osd_perf_query has missing NOTIFY_TYPES member 2026-03-09T15:13:30.167 INFO:journalctl@ceph.mgr.b.vm04.stdout:Mar 09 15:13:29 vm04 ceph-4abede80-1bca-11f1-bcd8-9b1da3c03ce5-mgr-b[54932]: 2026-03-09T15:13:29.929+0000 7f6d64ff6140 -1 mgr[py] Module osd_support has missing NOTIFY_TYPES member 2026-03-09T15:13:30.167 INFO:journalctl@ceph.mgr.b.vm04.stdout:Mar 09 15:13:29 vm04 ceph-4abede80-1bca-11f1-bcd8-9b1da3c03ce5-mgr-b[54932]: 2026-03-09T15:13:29.968+0000 7f6d64ff6140 -1 mgr[py] Module pg_autoscaler has missing NOTIFY_TYPES member 2026-03-09T15:13:30.167 INFO:journalctl@ceph.mgr.b.vm04.stdout:Mar 09 15:13:30 vm04 ceph-4abede80-1bca-11f1-bcd8-9b1da3c03ce5-mgr-b[54932]: 2026-03-09T15:13:30.002+0000 7f6d64ff6140 -1 mgr[py] Module progress has missing NOTIFY_TYPES member 2026-03-09T15:13:30.167 INFO:journalctl@ceph.mgr.b.vm04.stdout:Mar 09 15:13:30 vm04 ceph-4abede80-1bca-11f1-bcd8-9b1da3c03ce5-mgr-b[54932]: 2026-03-09T15:13:30.165+0000 7f6d64ff6140 -1 mgr[py] Module prometheus has missing NOTIFY_TYPES member 2026-03-09T15:13:30.170 INFO:teuthology.orchestra.run.vm03.stderr:Inferring config /var/lib/ceph/4abede80-1bca-11f1-bcd8-9b1da3c03ce5/mon.a/config 2026-03-09T15:13:30.190 INFO:journalctl@ceph.mgr.a.vm03.stdout:Mar 09 15:13:29 vm03 ceph-4abede80-1bca-11f1-bcd8-9b1da3c03ce5-mgr-a[51970]: 2026-03-09T15:13:29.805+0000 7fa54cf23640 -1 mgr.server handle_report got status from non-daemon mon.b 2026-03-09T15:13:30.422 INFO:journalctl@ceph.mgr.b.vm04.stdout:Mar 09 15:13:30 vm04 ceph-4abede80-1bca-11f1-bcd8-9b1da3c03ce5-mgr-b[54932]: 2026-03-09T15:13:30.213+0000 7f6d64ff6140 -1 mgr[py] Module rbd_support has missing NOTIFY_TYPES member 2026-03-09T15:13:30.487 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:13:30 vm03 ceph-mon[51762]: Reconfiguring mgr.a (unknown last config time)... 2026-03-09T15:13:30.487 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:13:30 vm03 ceph-mon[51762]: Reconfiguring daemon mgr.a on vm03 2026-03-09T15:13:30.675 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 15:13:30 vm04 ceph-mon[53998]: Reconfiguring mgr.a (unknown last config time)... 2026-03-09T15:13:30.675 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 15:13:30 vm04 ceph-mon[53998]: Reconfiguring daemon mgr.a on vm03 2026-03-09T15:13:30.675 INFO:journalctl@ceph.mgr.b.vm04.stdout:Mar 09 15:13:30 vm04 ceph-4abede80-1bca-11f1-bcd8-9b1da3c03ce5-mgr-b[54932]: 2026-03-09T15:13:30.421+0000 7f6d64ff6140 -1 mgr[py] Module rgw has missing NOTIFY_TYPES member 2026-03-09T15:13:30.676 INFO:journalctl@ceph.mgr.b.vm04.stdout:Mar 09 15:13:30 vm04 ceph-4abede80-1bca-11f1-bcd8-9b1da3c03ce5-mgr-b[54932]: 2026-03-09T15:13:30.674+0000 7f6d64ff6140 -1 mgr[py] Module rook has missing NOTIFY_TYPES member 2026-03-09T15:13:30.767 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 09 15:13:30 vm06 ceph-mon[56506]: Reconfiguring mgr.a (unknown last config time)... 2026-03-09T15:13:30.767 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 09 15:13:30 vm06 ceph-mon[56506]: Reconfiguring daemon mgr.a on vm03 2026-03-09T15:13:30.922 INFO:teuthology.orchestra.run.vm03.stdout: 2026-03-09T15:13:30.927 INFO:journalctl@ceph.mgr.b.vm04.stdout:Mar 09 15:13:30 vm04 ceph-4abede80-1bca-11f1-bcd8-9b1da3c03ce5-mgr-b[54932]: 2026-03-09T15:13:30.707+0000 7f6d64ff6140 -1 mgr[py] Module selftest has missing NOTIFY_TYPES member 2026-03-09T15:13:30.927 INFO:journalctl@ceph.mgr.b.vm04.stdout:Mar 09 15:13:30 vm04 ceph-4abede80-1bca-11f1-bcd8-9b1da3c03ce5-mgr-b[54932]: 2026-03-09T15:13:30.745+0000 7f6d64ff6140 -1 mgr[py] Module snap_schedule has missing NOTIFY_TYPES member 2026-03-09T15:13:30.927 INFO:journalctl@ceph.mgr.b.vm04.stdout:Mar 09 15:13:30 vm04 ceph-4abede80-1bca-11f1-bcd8-9b1da3c03ce5-mgr-b[54932]: 2026-03-09T15:13:30.814+0000 7f6d64ff6140 -1 mgr[py] Module status has missing NOTIFY_TYPES member 2026-03-09T15:13:30.927 INFO:journalctl@ceph.mgr.b.vm04.stdout:Mar 09 15:13:30 vm04 ceph-4abede80-1bca-11f1-bcd8-9b1da3c03ce5-mgr-b[54932]: 2026-03-09T15:13:30.850+0000 7f6d64ff6140 -1 mgr[py] Module telegraf has missing NOTIFY_TYPES member 2026-03-09T15:13:30.927 INFO:journalctl@ceph.mgr.b.vm04.stdout:Mar 09 15:13:30 vm04 ceph-4abede80-1bca-11f1-bcd8-9b1da3c03ce5-mgr-b[54932]: 2026-03-09T15:13:30.926+0000 7f6d64ff6140 -1 mgr[py] Module telemetry has missing NOTIFY_TYPES member 2026-03-09T15:13:30.938 DEBUG:teuthology.orchestra.run.vm03:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 4abede80-1bca-11f1-bcd8-9b1da3c03ce5 -- ceph orch daemon add osd vm03:/dev/vde 2026-03-09T15:13:31.091 INFO:teuthology.orchestra.run.vm03.stderr:Inferring config /var/lib/ceph/4abede80-1bca-11f1-bcd8-9b1da3c03ce5/mon.a/config 2026-03-09T15:13:31.191 INFO:journalctl@ceph.mgr.b.vm04.stdout:Mar 09 15:13:31 vm04 ceph-4abede80-1bca-11f1-bcd8-9b1da3c03ce5-mgr-b[54932]: 2026-03-09T15:13:31.029+0000 7f6d64ff6140 -1 mgr[py] Module test_orchestrator has missing NOTIFY_TYPES member 2026-03-09T15:13:31.191 INFO:journalctl@ceph.mgr.b.vm04.stdout:Mar 09 15:13:31 vm04 ceph-4abede80-1bca-11f1-bcd8-9b1da3c03ce5-mgr-b[54932]: 2026-03-09T15:13:31.156+0000 7f6d64ff6140 -1 mgr[py] Module volumes has missing NOTIFY_TYPES member 2026-03-09T15:13:31.191 INFO:journalctl@ceph.mgr.b.vm04.stdout:Mar 09 15:13:31 vm04 ceph-4abede80-1bca-11f1-bcd8-9b1da3c03ce5-mgr-b[54932]: 2026-03-09T15:13:31.189+0000 7f6d64ff6140 -1 mgr[py] Module zabbix has missing NOTIFY_TYPES member 2026-03-09T15:13:31.751 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 15:13:31 vm04 ceph-mon[53998]: pgmap v8: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-09T15:13:31.751 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 15:13:31 vm04 ceph-mon[53998]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' 2026-03-09T15:13:31.751 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 15:13:31 vm04 ceph-mon[53998]: Standby manager daemon b started 2026-03-09T15:13:31.751 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 15:13:31 vm04 ceph-mon[53998]: from='mgr.? 192.168.123.104:0/3034536622' entity='mgr.b' cmd=[{"prefix": "config-key get", "key": "mgr/dashboard/b/crt"}]: dispatch 2026-03-09T15:13:31.751 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 15:13:31 vm04 ceph-mon[53998]: from='mgr.? 192.168.123.104:0/3034536622' entity='mgr.b' cmd=[{"prefix": "config-key get", "key": "mgr/dashboard/crt"}]: dispatch 2026-03-09T15:13:31.751 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 15:13:31 vm04 ceph-mon[53998]: from='mgr.? 192.168.123.104:0/3034536622' entity='mgr.b' cmd=[{"prefix": "config-key get", "key": "mgr/dashboard/b/key"}]: dispatch 2026-03-09T15:13:31.751 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 15:13:31 vm04 ceph-mon[53998]: from='mgr.? 192.168.123.104:0/3034536622' entity='mgr.b' cmd=[{"prefix": "config-key get", "key": "mgr/dashboard/key"}]: dispatch 2026-03-09T15:13:31.751 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 15:13:31 vm04 ceph-mon[53998]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' cmd=[{"prefix": "osd tree", "states": ["destroyed"], "format": "json"}]: dispatch 2026-03-09T15:13:31.751 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 15:13:31 vm04 ceph-mon[53998]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.bootstrap-osd"}]: dispatch 2026-03-09T15:13:31.751 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 15:13:31 vm04 ceph-mon[53998]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T15:13:31.767 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 09 15:13:31 vm06 ceph-mon[56506]: pgmap v8: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-09T15:13:31.767 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 09 15:13:31 vm06 ceph-mon[56506]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' 2026-03-09T15:13:31.767 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 09 15:13:31 vm06 ceph-mon[56506]: Standby manager daemon b started 2026-03-09T15:13:31.767 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 09 15:13:31 vm06 ceph-mon[56506]: from='mgr.? 192.168.123.104:0/3034536622' entity='mgr.b' cmd=[{"prefix": "config-key get", "key": "mgr/dashboard/b/crt"}]: dispatch 2026-03-09T15:13:31.767 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 09 15:13:31 vm06 ceph-mon[56506]: from='mgr.? 192.168.123.104:0/3034536622' entity='mgr.b' cmd=[{"prefix": "config-key get", "key": "mgr/dashboard/crt"}]: dispatch 2026-03-09T15:13:31.767 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 09 15:13:31 vm06 ceph-mon[56506]: from='mgr.? 192.168.123.104:0/3034536622' entity='mgr.b' cmd=[{"prefix": "config-key get", "key": "mgr/dashboard/b/key"}]: dispatch 2026-03-09T15:13:31.767 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 09 15:13:31 vm06 ceph-mon[56506]: from='mgr.? 192.168.123.104:0/3034536622' entity='mgr.b' cmd=[{"prefix": "config-key get", "key": "mgr/dashboard/key"}]: dispatch 2026-03-09T15:13:31.767 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 09 15:13:31 vm06 ceph-mon[56506]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' cmd=[{"prefix": "osd tree", "states": ["destroyed"], "format": "json"}]: dispatch 2026-03-09T15:13:31.767 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 09 15:13:31 vm06 ceph-mon[56506]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.bootstrap-osd"}]: dispatch 2026-03-09T15:13:31.767 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 09 15:13:31 vm06 ceph-mon[56506]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T15:13:31.799 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:13:31 vm03 ceph-mon[51762]: pgmap v8: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-09T15:13:31.800 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:13:31 vm03 ceph-mon[51762]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' 2026-03-09T15:13:31.800 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:13:31 vm03 ceph-mon[51762]: Standby manager daemon b started 2026-03-09T15:13:31.800 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:13:31 vm03 ceph-mon[51762]: from='mgr.? 192.168.123.104:0/3034536622' entity='mgr.b' cmd=[{"prefix": "config-key get", "key": "mgr/dashboard/b/crt"}]: dispatch 2026-03-09T15:13:31.800 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:13:31 vm03 ceph-mon[51762]: from='mgr.? 192.168.123.104:0/3034536622' entity='mgr.b' cmd=[{"prefix": "config-key get", "key": "mgr/dashboard/crt"}]: dispatch 2026-03-09T15:13:31.800 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:13:31 vm03 ceph-mon[51762]: from='mgr.? 192.168.123.104:0/3034536622' entity='mgr.b' cmd=[{"prefix": "config-key get", "key": "mgr/dashboard/b/key"}]: dispatch 2026-03-09T15:13:31.800 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:13:31 vm03 ceph-mon[51762]: from='mgr.? 192.168.123.104:0/3034536622' entity='mgr.b' cmd=[{"prefix": "config-key get", "key": "mgr/dashboard/key"}]: dispatch 2026-03-09T15:13:31.800 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:13:31 vm03 ceph-mon[51762]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' cmd=[{"prefix": "osd tree", "states": ["destroyed"], "format": "json"}]: dispatch 2026-03-09T15:13:31.800 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:13:31 vm03 ceph-mon[51762]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.bootstrap-osd"}]: dispatch 2026-03-09T15:13:31.800 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:13:31 vm03 ceph-mon[51762]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T15:13:32.751 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 15:13:32 vm04 ceph-mon[53998]: from='client.14214 -' entity='client.admin' cmd=[{"prefix": "orch daemon add osd", "svc_arg": "vm03:/dev/vde", "target": ["mon-mgr", ""]}]: dispatch 2026-03-09T15:13:32.751 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 15:13:32 vm04 ceph-mon[53998]: mgrmap e13: a(active, since 29s), standbys: b 2026-03-09T15:13:32.751 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 15:13:32 vm04 ceph-mon[53998]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' cmd=[{"prefix": "mgr metadata", "who": "b", "id": "b"}]: dispatch 2026-03-09T15:13:32.751 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 15:13:32 vm04 ceph-mon[53998]: from='client.? 192.168.123.103:0/2978807971' entity='client.bootstrap-osd' cmd=[{"prefix": "osd new", "uuid": "913171bb-10e5-4020-a31f-9550e7c31d7b"}]: dispatch 2026-03-09T15:13:32.751 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 15:13:32 vm04 ceph-mon[53998]: from='client.? ' entity='client.bootstrap-osd' cmd=[{"prefix": "osd new", "uuid": "913171bb-10e5-4020-a31f-9550e7c31d7b"}]: dispatch 2026-03-09T15:13:32.751 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 15:13:32 vm04 ceph-mon[53998]: from='client.? ' entity='client.bootstrap-osd' cmd='[{"prefix": "osd new", "uuid": "913171bb-10e5-4020-a31f-9550e7c31d7b"}]': finished 2026-03-09T15:13:32.751 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 15:13:32 vm04 ceph-mon[53998]: osdmap e5: 1 total, 0 up, 1 in 2026-03-09T15:13:32.751 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 15:13:32 vm04 ceph-mon[53998]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 0}]: dispatch 2026-03-09T15:13:32.751 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 15:13:32 vm04 ceph-mon[53998]: from='client.? 192.168.123.103:0/2505746070' entity='client.bootstrap-osd' cmd=[{"prefix": "mon getmap"}]: dispatch 2026-03-09T15:13:32.767 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 09 15:13:32 vm06 ceph-mon[56506]: from='client.14214 -' entity='client.admin' cmd=[{"prefix": "orch daemon add osd", "svc_arg": "vm03:/dev/vde", "target": ["mon-mgr", ""]}]: dispatch 2026-03-09T15:13:32.767 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 09 15:13:32 vm06 ceph-mon[56506]: mgrmap e13: a(active, since 29s), standbys: b 2026-03-09T15:13:32.767 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 09 15:13:32 vm06 ceph-mon[56506]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' cmd=[{"prefix": "mgr metadata", "who": "b", "id": "b"}]: dispatch 2026-03-09T15:13:32.767 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 09 15:13:32 vm06 ceph-mon[56506]: from='client.? 192.168.123.103:0/2978807971' entity='client.bootstrap-osd' cmd=[{"prefix": "osd new", "uuid": "913171bb-10e5-4020-a31f-9550e7c31d7b"}]: dispatch 2026-03-09T15:13:32.767 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 09 15:13:32 vm06 ceph-mon[56506]: from='client.? ' entity='client.bootstrap-osd' cmd=[{"prefix": "osd new", "uuid": "913171bb-10e5-4020-a31f-9550e7c31d7b"}]: dispatch 2026-03-09T15:13:32.767 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 09 15:13:32 vm06 ceph-mon[56506]: from='client.? ' entity='client.bootstrap-osd' cmd='[{"prefix": "osd new", "uuid": "913171bb-10e5-4020-a31f-9550e7c31d7b"}]': finished 2026-03-09T15:13:32.767 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 09 15:13:32 vm06 ceph-mon[56506]: osdmap e5: 1 total, 0 up, 1 in 2026-03-09T15:13:32.767 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 09 15:13:32 vm06 ceph-mon[56506]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 0}]: dispatch 2026-03-09T15:13:32.767 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 09 15:13:32 vm06 ceph-mon[56506]: from='client.? 192.168.123.103:0/2505746070' entity='client.bootstrap-osd' cmd=[{"prefix": "mon getmap"}]: dispatch 2026-03-09T15:13:32.799 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:13:32 vm03 ceph-mon[51762]: from='client.14214 -' entity='client.admin' cmd=[{"prefix": "orch daemon add osd", "svc_arg": "vm03:/dev/vde", "target": ["mon-mgr", ""]}]: dispatch 2026-03-09T15:13:32.799 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:13:32 vm03 ceph-mon[51762]: mgrmap e13: a(active, since 29s), standbys: b 2026-03-09T15:13:32.799 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:13:32 vm03 ceph-mon[51762]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' cmd=[{"prefix": "mgr metadata", "who": "b", "id": "b"}]: dispatch 2026-03-09T15:13:32.799 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:13:32 vm03 ceph-mon[51762]: from='client.? 192.168.123.103:0/2978807971' entity='client.bootstrap-osd' cmd=[{"prefix": "osd new", "uuid": "913171bb-10e5-4020-a31f-9550e7c31d7b"}]: dispatch 2026-03-09T15:13:32.799 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:13:32 vm03 ceph-mon[51762]: from='client.? ' entity='client.bootstrap-osd' cmd=[{"prefix": "osd new", "uuid": "913171bb-10e5-4020-a31f-9550e7c31d7b"}]: dispatch 2026-03-09T15:13:32.799 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:13:32 vm03 ceph-mon[51762]: from='client.? ' entity='client.bootstrap-osd' cmd='[{"prefix": "osd new", "uuid": "913171bb-10e5-4020-a31f-9550e7c31d7b"}]': finished 2026-03-09T15:13:32.799 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:13:32 vm03 ceph-mon[51762]: osdmap e5: 1 total, 0 up, 1 in 2026-03-09T15:13:32.799 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:13:32 vm03 ceph-mon[51762]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 0}]: dispatch 2026-03-09T15:13:32.799 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:13:32 vm03 ceph-mon[51762]: from='client.? 192.168.123.103:0/2505746070' entity='client.bootstrap-osd' cmd=[{"prefix": "mon getmap"}]: dispatch 2026-03-09T15:13:33.751 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 15:13:33 vm04 ceph-mon[53998]: pgmap v10: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-09T15:13:33.767 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 09 15:13:33 vm06 ceph-mon[56506]: pgmap v10: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-09T15:13:33.799 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:13:33 vm03 ceph-mon[51762]: pgmap v10: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-09T15:13:35.719 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:13:35 vm03 ceph-mon[51762]: pgmap v11: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-09T15:13:35.767 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 09 15:13:35 vm06 ceph-mon[56506]: pgmap v11: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-09T15:13:36.001 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 15:13:35 vm04 ceph-mon[53998]: pgmap v11: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-09T15:13:36.508 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:13:36 vm03 ceph-mon[51762]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "osd.0"}]: dispatch 2026-03-09T15:13:36.508 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:13:36 vm03 ceph-mon[51762]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T15:13:36.767 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 09 15:13:36 vm06 ceph-mon[56506]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "osd.0"}]: dispatch 2026-03-09T15:13:36.767 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 09 15:13:36 vm06 ceph-mon[56506]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T15:13:37.001 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 15:13:36 vm04 ceph-mon[53998]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "osd.0"}]: dispatch 2026-03-09T15:13:37.001 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 15:13:36 vm04 ceph-mon[53998]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T15:13:37.767 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 09 15:13:37 vm06 ceph-mon[56506]: Deploying daemon osd.0 on vm03 2026-03-09T15:13:37.767 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 09 15:13:37 vm06 ceph-mon[56506]: pgmap v12: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-09T15:13:37.799 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:13:37 vm03 ceph-mon[51762]: Deploying daemon osd.0 on vm03 2026-03-09T15:13:37.799 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:13:37 vm03 ceph-mon[51762]: pgmap v12: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-09T15:13:38.001 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 15:13:37 vm04 ceph-mon[53998]: Deploying daemon osd.0 on vm03 2026-03-09T15:13:38.001 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 15:13:37 vm04 ceph-mon[53998]: pgmap v12: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-09T15:13:38.757 INFO:teuthology.orchestra.run.vm03.stdout:Created osd(s) 0 on host 'vm03' 2026-03-09T15:13:38.767 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 09 15:13:38 vm06 ceph-mon[56506]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T15:13:38.767 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 09 15:13:38 vm06 ceph-mon[56506]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' 2026-03-09T15:13:38.767 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 09 15:13:38 vm06 ceph-mon[56506]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' 2026-03-09T15:13:38.767 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 09 15:13:38 vm06 ceph-mon[56506]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' 2026-03-09T15:13:38.767 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 09 15:13:38 vm06 ceph-mon[56506]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' 2026-03-09T15:13:38.767 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 09 15:13:38 vm06 ceph-mon[56506]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T15:13:38.767 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 09 15:13:38 vm06 ceph-mon[56506]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T15:13:38.767 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 09 15:13:38 vm06 ceph-mon[56506]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' 2026-03-09T15:13:38.771 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:13:38 vm03 ceph-mon[51762]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T15:13:38.771 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:13:38 vm03 ceph-mon[51762]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' 2026-03-09T15:13:38.771 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:13:38 vm03 ceph-mon[51762]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' 2026-03-09T15:13:38.771 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:13:38 vm03 ceph-mon[51762]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' 2026-03-09T15:13:38.771 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:13:38 vm03 ceph-mon[51762]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' 2026-03-09T15:13:38.772 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:13:38 vm03 ceph-mon[51762]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T15:13:38.772 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:13:38 vm03 ceph-mon[51762]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T15:13:38.772 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:13:38 vm03 ceph-mon[51762]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' 2026-03-09T15:13:38.830 DEBUG:teuthology.orchestra.run.vm03:osd.0> sudo journalctl -f -n 0 -u ceph-4abede80-1bca-11f1-bcd8-9b1da3c03ce5@osd.0.service 2026-03-09T15:13:38.832 INFO:tasks.cephadm:Deploying osd.1 on vm04 with /dev/vde... 2026-03-09T15:13:38.832 DEBUG:teuthology.orchestra.run.vm04:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df ceph-volume -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 4abede80-1bca-11f1-bcd8-9b1da3c03ce5 -- lvm zap /dev/vde 2026-03-09T15:13:38.853 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 15:13:38 vm04 ceph-mon[53998]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T15:13:38.853 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 15:13:38 vm04 ceph-mon[53998]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' 2026-03-09T15:13:38.853 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 15:13:38 vm04 ceph-mon[53998]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' 2026-03-09T15:13:38.853 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 15:13:38 vm04 ceph-mon[53998]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' 2026-03-09T15:13:38.853 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 15:13:38 vm04 ceph-mon[53998]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' 2026-03-09T15:13:38.853 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 15:13:38 vm04 ceph-mon[53998]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T15:13:38.853 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 15:13:38 vm04 ceph-mon[53998]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T15:13:38.853 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 15:13:38 vm04 ceph-mon[53998]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' 2026-03-09T15:13:39.000 INFO:teuthology.orchestra.run.vm04.stderr:Inferring config /var/lib/ceph/4abede80-1bca-11f1-bcd8-9b1da3c03ce5/mon.b/config 2026-03-09T15:13:39.694 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 15:13:39 vm04 ceph-mon[53998]: pgmap v13: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-09T15:13:39.695 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 15:13:39 vm04 ceph-mon[53998]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T15:13:39.695 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 15:13:39 vm04 ceph-mon[53998]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' 2026-03-09T15:13:39.695 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 15:13:39 vm04 ceph-mon[53998]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' 2026-03-09T15:13:39.767 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 09 15:13:39 vm06 ceph-mon[56506]: pgmap v13: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-09T15:13:39.767 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 09 15:13:39 vm06 ceph-mon[56506]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T15:13:39.767 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 09 15:13:39 vm06 ceph-mon[56506]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' 2026-03-09T15:13:39.767 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 09 15:13:39 vm06 ceph-mon[56506]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' 2026-03-09T15:13:39.786 INFO:teuthology.orchestra.run.vm04.stdout: 2026-03-09T15:13:39.800 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:13:39 vm03 ceph-mon[51762]: pgmap v13: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-09T15:13:39.800 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:13:39 vm03 ceph-mon[51762]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T15:13:39.800 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:13:39 vm03 ceph-mon[51762]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' 2026-03-09T15:13:39.800 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:13:39 vm03 ceph-mon[51762]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' 2026-03-09T15:13:39.801 INFO:journalctl@ceph.osd.0.vm03.stdout:Mar 09 15:13:39 vm03 ceph-4abede80-1bca-11f1-bcd8-9b1da3c03ce5-osd-0[61316]: 2026-03-09T15:13:39.623+0000 7f89ea5f3740 -1 osd.0 0 log_to_monitors true 2026-03-09T15:13:39.802 DEBUG:teuthology.orchestra.run.vm04:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 4abede80-1bca-11f1-bcd8-9b1da3c03ce5 -- ceph orch daemon add osd vm04:/dev/vde 2026-03-09T15:13:39.997 INFO:teuthology.orchestra.run.vm04.stderr:Inferring config /var/lib/ceph/4abede80-1bca-11f1-bcd8-9b1da3c03ce5/mon.b/config 2026-03-09T15:13:40.752 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 15:13:40 vm04 ceph-mon[53998]: from='osd.0 [v2:192.168.123.103:6802/615876468,v1:192.168.123.103:6803/615876468]' entity='osd.0' cmd=[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["0"]}]: dispatch 2026-03-09T15:13:40.752 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 15:13:40 vm04 ceph-mon[53998]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' 2026-03-09T15:13:40.752 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 15:13:40 vm04 ceph-mon[53998]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' 2026-03-09T15:13:40.752 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 15:13:40 vm04 ceph-mon[53998]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' cmd=[{"prefix": "config rm", "who": "osd.0", "name": "osd_memory_target"}]: dispatch 2026-03-09T15:13:40.752 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 15:13:40 vm04 ceph-mon[53998]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T15:13:40.752 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 15:13:40 vm04 ceph-mon[53998]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T15:13:40.752 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 15:13:40 vm04 ceph-mon[53998]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' 2026-03-09T15:13:40.752 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 15:13:40 vm04 ceph-mon[53998]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' cmd=[{"prefix": "osd tree", "states": ["destroyed"], "format": "json"}]: dispatch 2026-03-09T15:13:40.752 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 15:13:40 vm04 ceph-mon[53998]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.bootstrap-osd"}]: dispatch 2026-03-09T15:13:40.752 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 15:13:40 vm04 ceph-mon[53998]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T15:13:40.799 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:13:40 vm03 ceph-mon[51762]: from='osd.0 [v2:192.168.123.103:6802/615876468,v1:192.168.123.103:6803/615876468]' entity='osd.0' cmd=[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["0"]}]: dispatch 2026-03-09T15:13:40.799 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:13:40 vm03 ceph-mon[51762]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' 2026-03-09T15:13:40.799 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:13:40 vm03 ceph-mon[51762]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' 2026-03-09T15:13:40.799 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:13:40 vm03 ceph-mon[51762]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' cmd=[{"prefix": "config rm", "who": "osd.0", "name": "osd_memory_target"}]: dispatch 2026-03-09T15:13:40.799 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:13:40 vm03 ceph-mon[51762]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T15:13:40.799 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:13:40 vm03 ceph-mon[51762]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T15:13:40.799 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:13:40 vm03 ceph-mon[51762]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' 2026-03-09T15:13:40.799 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:13:40 vm03 ceph-mon[51762]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' cmd=[{"prefix": "osd tree", "states": ["destroyed"], "format": "json"}]: dispatch 2026-03-09T15:13:40.799 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:13:40 vm03 ceph-mon[51762]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.bootstrap-osd"}]: dispatch 2026-03-09T15:13:40.799 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:13:40 vm03 ceph-mon[51762]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T15:13:41.017 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 09 15:13:40 vm06 ceph-mon[56506]: from='osd.0 [v2:192.168.123.103:6802/615876468,v1:192.168.123.103:6803/615876468]' entity='osd.0' cmd=[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["0"]}]: dispatch 2026-03-09T15:13:41.017 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 09 15:13:40 vm06 ceph-mon[56506]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' 2026-03-09T15:13:41.017 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 09 15:13:40 vm06 ceph-mon[56506]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' 2026-03-09T15:13:41.017 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 09 15:13:40 vm06 ceph-mon[56506]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' cmd=[{"prefix": "config rm", "who": "osd.0", "name": "osd_memory_target"}]: dispatch 2026-03-09T15:13:41.017 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 09 15:13:40 vm06 ceph-mon[56506]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T15:13:41.017 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 09 15:13:40 vm06 ceph-mon[56506]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T15:13:41.017 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 09 15:13:40 vm06 ceph-mon[56506]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' 2026-03-09T15:13:41.017 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 09 15:13:40 vm06 ceph-mon[56506]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' cmd=[{"prefix": "osd tree", "states": ["destroyed"], "format": "json"}]: dispatch 2026-03-09T15:13:41.017 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 09 15:13:40 vm06 ceph-mon[56506]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.bootstrap-osd"}]: dispatch 2026-03-09T15:13:41.017 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 09 15:13:40 vm06 ceph-mon[56506]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T15:13:41.752 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 15:13:41 vm04 ceph-mon[53998]: Detected new or changed devices on vm03 2026-03-09T15:13:41.752 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 15:13:41 vm04 ceph-mon[53998]: Adjusting osd_memory_target on vm03 to 257.0M 2026-03-09T15:13:41.752 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 15:13:41 vm04 ceph-mon[53998]: Unable to set osd_memory_target on vm03 to 269536460: error parsing value: Value '269536460' is below minimum 939524096 2026-03-09T15:13:41.752 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 15:13:41 vm04 ceph-mon[53998]: from='client.24125 -' entity='client.admin' cmd=[{"prefix": "orch daemon add osd", "svc_arg": "vm04:/dev/vde", "target": ["mon-mgr", ""]}]: dispatch 2026-03-09T15:13:41.752 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 15:13:41 vm04 ceph-mon[53998]: pgmap v14: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-09T15:13:41.752 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 15:13:41 vm04 ceph-mon[53998]: from='osd.0 [v2:192.168.123.103:6802/615876468,v1:192.168.123.103:6803/615876468]' entity='osd.0' cmd='[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["0"]}]': finished 2026-03-09T15:13:41.752 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 15:13:41 vm04 ceph-mon[53998]: osdmap e6: 1 total, 0 up, 1 in 2026-03-09T15:13:41.752 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 15:13:41 vm04 ceph-mon[53998]: from='osd.0 [v2:192.168.123.103:6802/615876468,v1:192.168.123.103:6803/615876468]' entity='osd.0' cmd=[{"prefix": "osd crush create-or-move", "id": 0, "weight":0.0195, "args": ["host=vm03", "root=default"]}]: dispatch 2026-03-09T15:13:41.752 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 15:13:41 vm04 ceph-mon[53998]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 0}]: dispatch 2026-03-09T15:13:41.752 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 15:13:41 vm04 ceph-mon[53998]: from='client.? 192.168.123.104:0/2177275976' entity='client.bootstrap-osd' cmd=[{"prefix": "osd new", "uuid": "62ee7f4e-6ba5-481b-9cde-4c62baacf802"}]: dispatch 2026-03-09T15:13:41.752 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 15:13:41 vm04 ceph-mon[53998]: from='osd.0 [v2:192.168.123.103:6802/615876468,v1:192.168.123.103:6803/615876468]' entity='osd.0' cmd='[{"prefix": "osd crush create-or-move", "id": 0, "weight":0.0195, "args": ["host=vm03", "root=default"]}]': finished 2026-03-09T15:13:41.752 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 15:13:41 vm04 ceph-mon[53998]: from='client.? 192.168.123.104:0/2177275976' entity='client.bootstrap-osd' cmd='[{"prefix": "osd new", "uuid": "62ee7f4e-6ba5-481b-9cde-4c62baacf802"}]': finished 2026-03-09T15:13:41.752 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 15:13:41 vm04 ceph-mon[53998]: osdmap e7: 2 total, 0 up, 2 in 2026-03-09T15:13:41.752 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 15:13:41 vm04 ceph-mon[53998]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 0}]: dispatch 2026-03-09T15:13:41.752 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 15:13:41 vm04 ceph-mon[53998]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 1}]: dispatch 2026-03-09T15:13:41.752 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 15:13:41 vm04 ceph-mon[53998]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 0}]: dispatch 2026-03-09T15:13:41.752 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 15:13:41 vm04 ceph-mon[53998]: from='client.? 192.168.123.104:0/1526452863' entity='client.bootstrap-osd' cmd=[{"prefix": "mon getmap"}]: dispatch 2026-03-09T15:13:41.799 INFO:journalctl@ceph.osd.0.vm03.stdout:Mar 09 15:13:41 vm03 ceph-4abede80-1bca-11f1-bcd8-9b1da3c03ce5-osd-0[61316]: 2026-03-09T15:13:41.758+0000 7f89e6574640 -1 osd.0 0 waiting for initial osdmap 2026-03-09T15:13:41.799 INFO:journalctl@ceph.osd.0.vm03.stdout:Mar 09 15:13:41 vm03 ceph-4abede80-1bca-11f1-bcd8-9b1da3c03ce5-osd-0[61316]: 2026-03-09T15:13:41.763+0000 7f89e1b9d640 -1 osd.0 7 set_numa_affinity unable to identify public interface '' numa node: (2) No such file or directory 2026-03-09T15:13:41.799 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:13:41 vm03 ceph-mon[51762]: Detected new or changed devices on vm03 2026-03-09T15:13:41.799 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:13:41 vm03 ceph-mon[51762]: Adjusting osd_memory_target on vm03 to 257.0M 2026-03-09T15:13:41.799 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:13:41 vm03 ceph-mon[51762]: Unable to set osd_memory_target on vm03 to 269536460: error parsing value: Value '269536460' is below minimum 939524096 2026-03-09T15:13:41.799 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:13:41 vm03 ceph-mon[51762]: from='client.24125 -' entity='client.admin' cmd=[{"prefix": "orch daemon add osd", "svc_arg": "vm04:/dev/vde", "target": ["mon-mgr", ""]}]: dispatch 2026-03-09T15:13:41.799 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:13:41 vm03 ceph-mon[51762]: pgmap v14: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-09T15:13:41.799 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:13:41 vm03 ceph-mon[51762]: from='osd.0 [v2:192.168.123.103:6802/615876468,v1:192.168.123.103:6803/615876468]' entity='osd.0' cmd='[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["0"]}]': finished 2026-03-09T15:13:41.799 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:13:41 vm03 ceph-mon[51762]: osdmap e6: 1 total, 0 up, 1 in 2026-03-09T15:13:41.799 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:13:41 vm03 ceph-mon[51762]: from='osd.0 [v2:192.168.123.103:6802/615876468,v1:192.168.123.103:6803/615876468]' entity='osd.0' cmd=[{"prefix": "osd crush create-or-move", "id": 0, "weight":0.0195, "args": ["host=vm03", "root=default"]}]: dispatch 2026-03-09T15:13:41.799 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:13:41 vm03 ceph-mon[51762]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 0}]: dispatch 2026-03-09T15:13:41.799 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:13:41 vm03 ceph-mon[51762]: from='client.? 192.168.123.104:0/2177275976' entity='client.bootstrap-osd' cmd=[{"prefix": "osd new", "uuid": "62ee7f4e-6ba5-481b-9cde-4c62baacf802"}]: dispatch 2026-03-09T15:13:41.799 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:13:41 vm03 ceph-mon[51762]: from='osd.0 [v2:192.168.123.103:6802/615876468,v1:192.168.123.103:6803/615876468]' entity='osd.0' cmd='[{"prefix": "osd crush create-or-move", "id": 0, "weight":0.0195, "args": ["host=vm03", "root=default"]}]': finished 2026-03-09T15:13:41.799 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:13:41 vm03 ceph-mon[51762]: from='client.? 192.168.123.104:0/2177275976' entity='client.bootstrap-osd' cmd='[{"prefix": "osd new", "uuid": "62ee7f4e-6ba5-481b-9cde-4c62baacf802"}]': finished 2026-03-09T15:13:41.800 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:13:41 vm03 ceph-mon[51762]: osdmap e7: 2 total, 0 up, 2 in 2026-03-09T15:13:41.800 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:13:41 vm03 ceph-mon[51762]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 0}]: dispatch 2026-03-09T15:13:41.800 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:13:41 vm03 ceph-mon[51762]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 1}]: dispatch 2026-03-09T15:13:41.800 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:13:41 vm03 ceph-mon[51762]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 0}]: dispatch 2026-03-09T15:13:41.800 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:13:41 vm03 ceph-mon[51762]: from='client.? 192.168.123.104:0/1526452863' entity='client.bootstrap-osd' cmd=[{"prefix": "mon getmap"}]: dispatch 2026-03-09T15:13:42.017 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 09 15:13:41 vm06 ceph-mon[56506]: Detected new or changed devices on vm03 2026-03-09T15:13:42.017 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 09 15:13:41 vm06 ceph-mon[56506]: Adjusting osd_memory_target on vm03 to 257.0M 2026-03-09T15:13:42.017 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 09 15:13:41 vm06 ceph-mon[56506]: Unable to set osd_memory_target on vm03 to 269536460: error parsing value: Value '269536460' is below minimum 939524096 2026-03-09T15:13:42.017 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 09 15:13:41 vm06 ceph-mon[56506]: from='client.24125 -' entity='client.admin' cmd=[{"prefix": "orch daemon add osd", "svc_arg": "vm04:/dev/vde", "target": ["mon-mgr", ""]}]: dispatch 2026-03-09T15:13:42.017 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 09 15:13:41 vm06 ceph-mon[56506]: pgmap v14: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-09T15:13:42.017 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 09 15:13:41 vm06 ceph-mon[56506]: from='osd.0 [v2:192.168.123.103:6802/615876468,v1:192.168.123.103:6803/615876468]' entity='osd.0' cmd='[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["0"]}]': finished 2026-03-09T15:13:42.017 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 09 15:13:41 vm06 ceph-mon[56506]: osdmap e6: 1 total, 0 up, 1 in 2026-03-09T15:13:42.017 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 09 15:13:41 vm06 ceph-mon[56506]: from='osd.0 [v2:192.168.123.103:6802/615876468,v1:192.168.123.103:6803/615876468]' entity='osd.0' cmd=[{"prefix": "osd crush create-or-move", "id": 0, "weight":0.0195, "args": ["host=vm03", "root=default"]}]: dispatch 2026-03-09T15:13:42.017 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 09 15:13:41 vm06 ceph-mon[56506]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 0}]: dispatch 2026-03-09T15:13:42.017 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 09 15:13:41 vm06 ceph-mon[56506]: from='client.? 192.168.123.104:0/2177275976' entity='client.bootstrap-osd' cmd=[{"prefix": "osd new", "uuid": "62ee7f4e-6ba5-481b-9cde-4c62baacf802"}]: dispatch 2026-03-09T15:13:42.017 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 09 15:13:41 vm06 ceph-mon[56506]: from='osd.0 [v2:192.168.123.103:6802/615876468,v1:192.168.123.103:6803/615876468]' entity='osd.0' cmd='[{"prefix": "osd crush create-or-move", "id": 0, "weight":0.0195, "args": ["host=vm03", "root=default"]}]': finished 2026-03-09T15:13:42.017 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 09 15:13:41 vm06 ceph-mon[56506]: from='client.? 192.168.123.104:0/2177275976' entity='client.bootstrap-osd' cmd='[{"prefix": "osd new", "uuid": "62ee7f4e-6ba5-481b-9cde-4c62baacf802"}]': finished 2026-03-09T15:13:42.017 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 09 15:13:41 vm06 ceph-mon[56506]: osdmap e7: 2 total, 0 up, 2 in 2026-03-09T15:13:42.017 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 09 15:13:41 vm06 ceph-mon[56506]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 0}]: dispatch 2026-03-09T15:13:42.017 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 09 15:13:41 vm06 ceph-mon[56506]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 1}]: dispatch 2026-03-09T15:13:42.017 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 09 15:13:41 vm06 ceph-mon[56506]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 0}]: dispatch 2026-03-09T15:13:42.017 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 09 15:13:41 vm06 ceph-mon[56506]: from='client.? 192.168.123.104:0/1526452863' entity='client.bootstrap-osd' cmd=[{"prefix": "mon getmap"}]: dispatch 2026-03-09T15:13:43.049 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:13:42 vm03 ceph-mon[51762]: purged_snaps scrub starts 2026-03-09T15:13:43.049 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:13:42 vm03 ceph-mon[51762]: purged_snaps scrub ok 2026-03-09T15:13:43.049 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:13:42 vm03 ceph-mon[51762]: from='osd.0 [v2:192.168.123.103:6802/615876468,v1:192.168.123.103:6803/615876468]' entity='osd.0' 2026-03-09T15:13:43.050 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:13:42 vm03 ceph-mon[51762]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 0}]: dispatch 2026-03-09T15:13:43.050 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:13:42 vm03 ceph-mon[51762]: pgmap v17: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-09T15:13:43.251 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 15:13:42 vm04 ceph-mon[53998]: purged_snaps scrub starts 2026-03-09T15:13:43.251 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 15:13:42 vm04 ceph-mon[53998]: purged_snaps scrub ok 2026-03-09T15:13:43.251 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 15:13:42 vm04 ceph-mon[53998]: from='osd.0 [v2:192.168.123.103:6802/615876468,v1:192.168.123.103:6803/615876468]' entity='osd.0' 2026-03-09T15:13:43.251 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 15:13:42 vm04 ceph-mon[53998]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 0}]: dispatch 2026-03-09T15:13:43.251 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 15:13:42 vm04 ceph-mon[53998]: pgmap v17: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-09T15:13:43.267 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 09 15:13:42 vm06 ceph-mon[56506]: purged_snaps scrub starts 2026-03-09T15:13:43.267 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 09 15:13:42 vm06 ceph-mon[56506]: purged_snaps scrub ok 2026-03-09T15:13:43.267 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 09 15:13:42 vm06 ceph-mon[56506]: from='osd.0 [v2:192.168.123.103:6802/615876468,v1:192.168.123.103:6803/615876468]' entity='osd.0' 2026-03-09T15:13:43.267 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 09 15:13:42 vm06 ceph-mon[56506]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 0}]: dispatch 2026-03-09T15:13:43.267 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 09 15:13:42 vm06 ceph-mon[56506]: pgmap v17: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-09T15:13:44.049 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:13:43 vm03 ceph-mon[51762]: osd.0 [v2:192.168.123.103:6802/615876468,v1:192.168.123.103:6803/615876468] boot 2026-03-09T15:13:44.049 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:13:43 vm03 ceph-mon[51762]: osdmap e8: 2 total, 1 up, 2 in 2026-03-09T15:13:44.049 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:13:43 vm03 ceph-mon[51762]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 0}]: dispatch 2026-03-09T15:13:44.049 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:13:43 vm03 ceph-mon[51762]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 1}]: dispatch 2026-03-09T15:13:44.077 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 15:13:43 vm04 ceph-mon[53998]: osd.0 [v2:192.168.123.103:6802/615876468,v1:192.168.123.103:6803/615876468] boot 2026-03-09T15:13:44.077 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 15:13:43 vm04 ceph-mon[53998]: osdmap e8: 2 total, 1 up, 2 in 2026-03-09T15:13:44.077 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 15:13:43 vm04 ceph-mon[53998]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 0}]: dispatch 2026-03-09T15:13:44.077 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 15:13:43 vm04 ceph-mon[53998]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 1}]: dispatch 2026-03-09T15:13:44.267 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 09 15:13:43 vm06 ceph-mon[56506]: osd.0 [v2:192.168.123.103:6802/615876468,v1:192.168.123.103:6803/615876468] boot 2026-03-09T15:13:44.267 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 09 15:13:43 vm06 ceph-mon[56506]: osdmap e8: 2 total, 1 up, 2 in 2026-03-09T15:13:44.267 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 09 15:13:43 vm06 ceph-mon[56506]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 0}]: dispatch 2026-03-09T15:13:44.267 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 09 15:13:43 vm06 ceph-mon[56506]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 1}]: dispatch 2026-03-09T15:13:44.917 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 15:13:44 vm04 ceph-mon[53998]: osdmap e9: 2 total, 1 up, 2 in 2026-03-09T15:13:44.918 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 15:13:44 vm04 ceph-mon[53998]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 1}]: dispatch 2026-03-09T15:13:44.918 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 15:13:44 vm04 ceph-mon[53998]: pgmap v20: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-09T15:13:45.049 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:13:44 vm03 ceph-mon[51762]: osdmap e9: 2 total, 1 up, 2 in 2026-03-09T15:13:45.049 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:13:44 vm03 ceph-mon[51762]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 1}]: dispatch 2026-03-09T15:13:45.049 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:13:44 vm03 ceph-mon[51762]: pgmap v20: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-09T15:13:45.267 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 09 15:13:44 vm06 ceph-mon[56506]: osdmap e9: 2 total, 1 up, 2 in 2026-03-09T15:13:45.267 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 09 15:13:44 vm06 ceph-mon[56506]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 1}]: dispatch 2026-03-09T15:13:45.267 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 09 15:13:44 vm06 ceph-mon[56506]: pgmap v20: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-09T15:13:46.049 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:13:45 vm03 ceph-mon[51762]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "osd.1"}]: dispatch 2026-03-09T15:13:46.049 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:13:45 vm03 ceph-mon[51762]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T15:13:46.049 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:13:45 vm03 ceph-mon[51762]: Deploying daemon osd.1 on vm04 2026-03-09T15:13:46.052 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 15:13:45 vm04 ceph-mon[53998]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "osd.1"}]: dispatch 2026-03-09T15:13:46.052 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 15:13:45 vm04 ceph-mon[53998]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T15:13:46.052 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 15:13:45 vm04 ceph-mon[53998]: Deploying daemon osd.1 on vm04 2026-03-09T15:13:46.267 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 09 15:13:45 vm06 ceph-mon[56506]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "osd.1"}]: dispatch 2026-03-09T15:13:46.267 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 09 15:13:45 vm06 ceph-mon[56506]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T15:13:46.267 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 09 15:13:45 vm06 ceph-mon[56506]: Deploying daemon osd.1 on vm04 2026-03-09T15:13:47.049 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:13:46 vm03 ceph-mon[51762]: pgmap v21: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-09T15:13:47.208 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 15:13:46 vm04 ceph-mon[53998]: pgmap v21: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-09T15:13:47.267 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 09 15:13:46 vm06 ceph-mon[56506]: pgmap v21: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-09T15:13:47.976 INFO:teuthology.orchestra.run.vm04.stdout:Created osd(s) 1 on host 'vm04' 2026-03-09T15:13:48.041 DEBUG:teuthology.orchestra.run.vm04:osd.1> sudo journalctl -f -n 0 -u ceph-4abede80-1bca-11f1-bcd8-9b1da3c03ce5@osd.1.service 2026-03-09T15:13:48.042 INFO:tasks.cephadm:Deploying osd.2 on vm06 with /dev/vde... 2026-03-09T15:13:48.042 DEBUG:teuthology.orchestra.run.vm06:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df ceph-volume -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 4abede80-1bca-11f1-bcd8-9b1da3c03ce5 -- lvm zap /dev/vde 2026-03-09T15:13:48.062 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 09 15:13:47 vm06 ceph-mon[56506]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T15:13:48.062 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 09 15:13:47 vm06 ceph-mon[56506]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' 2026-03-09T15:13:48.062 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 09 15:13:47 vm06 ceph-mon[56506]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' 2026-03-09T15:13:48.062 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 09 15:13:47 vm06 ceph-mon[56506]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' 2026-03-09T15:13:48.063 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 09 15:13:47 vm06 ceph-mon[56506]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' 2026-03-09T15:13:48.063 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 09 15:13:47 vm06 ceph-mon[56506]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T15:13:48.063 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 09 15:13:47 vm06 ceph-mon[56506]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T15:13:48.063 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 09 15:13:47 vm06 ceph-mon[56506]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' 2026-03-09T15:13:48.071 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 15:13:47 vm04 ceph-mon[53998]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T15:13:48.071 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 15:13:47 vm04 ceph-mon[53998]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' 2026-03-09T15:13:48.071 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 15:13:47 vm04 ceph-mon[53998]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' 2026-03-09T15:13:48.071 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 15:13:47 vm04 ceph-mon[53998]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' 2026-03-09T15:13:48.071 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 15:13:47 vm04 ceph-mon[53998]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' 2026-03-09T15:13:48.071 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 15:13:47 vm04 ceph-mon[53998]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T15:13:48.071 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 15:13:47 vm04 ceph-mon[53998]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T15:13:48.071 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 15:13:47 vm04 ceph-mon[53998]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' 2026-03-09T15:13:48.203 INFO:teuthology.orchestra.run.vm06.stderr:Inferring config /var/lib/ceph/4abede80-1bca-11f1-bcd8-9b1da3c03ce5/mon.c/config 2026-03-09T15:13:48.299 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:13:47 vm03 ceph-mon[51762]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T15:13:48.299 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:13:47 vm03 ceph-mon[51762]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' 2026-03-09T15:13:48.299 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:13:47 vm03 ceph-mon[51762]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' 2026-03-09T15:13:48.299 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:13:47 vm03 ceph-mon[51762]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' 2026-03-09T15:13:48.299 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:13:47 vm03 ceph-mon[51762]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' 2026-03-09T15:13:48.299 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:13:47 vm03 ceph-mon[51762]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T15:13:48.299 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:13:47 vm03 ceph-mon[51762]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T15:13:48.299 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:13:47 vm03 ceph-mon[51762]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' 2026-03-09T15:13:48.923 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 09 15:13:48 vm06 ceph-mon[56506]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T15:13:48.924 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 09 15:13:48 vm06 ceph-mon[56506]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' 2026-03-09T15:13:48.924 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 09 15:13:48 vm06 ceph-mon[56506]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' 2026-03-09T15:13:48.924 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 09 15:13:48 vm06 ceph-mon[56506]: pgmap v22: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-09T15:13:49.021 INFO:teuthology.orchestra.run.vm06.stdout: 2026-03-09T15:13:49.037 DEBUG:teuthology.orchestra.run.vm06:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 4abede80-1bca-11f1-bcd8-9b1da3c03ce5 -- ceph orch daemon add osd vm06:/dev/vde 2026-03-09T15:13:49.238 INFO:teuthology.orchestra.run.vm06.stderr:Inferring config /var/lib/ceph/4abede80-1bca-11f1-bcd8-9b1da3c03ce5/mon.c/config 2026-03-09T15:13:49.251 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 15:13:48 vm04 ceph-mon[53998]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T15:13:49.251 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 15:13:48 vm04 ceph-mon[53998]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' 2026-03-09T15:13:49.251 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 15:13:48 vm04 ceph-mon[53998]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' 2026-03-09T15:13:49.251 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 15:13:48 vm04 ceph-mon[53998]: pgmap v22: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-09T15:13:49.251 INFO:journalctl@ceph.osd.1.vm04.stdout:Mar 09 15:13:48 vm04 ceph-4abede80-1bca-11f1-bcd8-9b1da3c03ce5-osd-1[57595]: 2026-03-09T15:13:48.857+0000 7f583fd75740 -1 osd.1 0 log_to_monitors true 2026-03-09T15:13:49.299 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:13:48 vm03 ceph-mon[51762]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T15:13:49.299 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:13:48 vm03 ceph-mon[51762]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' 2026-03-09T15:13:49.299 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:13:48 vm03 ceph-mon[51762]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' 2026-03-09T15:13:49.299 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:13:48 vm03 ceph-mon[51762]: pgmap v22: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-09T15:13:50.017 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 09 15:13:49 vm06 ceph-mon[56506]: from='osd.1 [v2:192.168.123.104:6800/1573535831,v1:192.168.123.104:6801/1573535831]' entity='osd.1' cmd=[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["1"]}]: dispatch 2026-03-09T15:13:50.017 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 09 15:13:49 vm06 ceph-mon[56506]: Detected new or changed devices on vm04 2026-03-09T15:13:50.017 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 09 15:13:49 vm06 ceph-mon[56506]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' 2026-03-09T15:13:50.017 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 09 15:13:49 vm06 ceph-mon[56506]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' 2026-03-09T15:13:50.017 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 09 15:13:49 vm06 ceph-mon[56506]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' cmd=[{"prefix": "config rm", "who": "osd.1", "name": "osd_memory_target"}]: dispatch 2026-03-09T15:13:50.017 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 09 15:13:49 vm06 ceph-mon[56506]: Adjusting osd_memory_target on vm04 to 257.0M 2026-03-09T15:13:50.017 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 09 15:13:49 vm06 ceph-mon[56506]: Unable to set osd_memory_target on vm04 to 269530726: error parsing value: Value '269530726' is below minimum 939524096 2026-03-09T15:13:50.017 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 09 15:13:49 vm06 ceph-mon[56506]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T15:13:50.017 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 09 15:13:49 vm06 ceph-mon[56506]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T15:13:50.017 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 09 15:13:49 vm06 ceph-mon[56506]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' 2026-03-09T15:13:50.017 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 09 15:13:49 vm06 ceph-mon[56506]: from='client.14262 -' entity='client.admin' cmd=[{"prefix": "orch daemon add osd", "svc_arg": "vm06:/dev/vde", "target": ["mon-mgr", ""]}]: dispatch 2026-03-09T15:13:50.017 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 09 15:13:49 vm06 ceph-mon[56506]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' cmd=[{"prefix": "osd tree", "states": ["destroyed"], "format": "json"}]: dispatch 2026-03-09T15:13:50.017 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 09 15:13:49 vm06 ceph-mon[56506]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.bootstrap-osd"}]: dispatch 2026-03-09T15:13:50.017 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 09 15:13:49 vm06 ceph-mon[56506]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T15:13:50.204 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 15:13:49 vm04 ceph-mon[53998]: from='osd.1 [v2:192.168.123.104:6800/1573535831,v1:192.168.123.104:6801/1573535831]' entity='osd.1' cmd=[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["1"]}]: dispatch 2026-03-09T15:13:50.204 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 15:13:49 vm04 ceph-mon[53998]: Detected new or changed devices on vm04 2026-03-09T15:13:50.204 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 15:13:49 vm04 ceph-mon[53998]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' 2026-03-09T15:13:50.204 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 15:13:49 vm04 ceph-mon[53998]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' 2026-03-09T15:13:50.204 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 15:13:49 vm04 ceph-mon[53998]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' cmd=[{"prefix": "config rm", "who": "osd.1", "name": "osd_memory_target"}]: dispatch 2026-03-09T15:13:50.204 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 15:13:49 vm04 ceph-mon[53998]: Adjusting osd_memory_target on vm04 to 257.0M 2026-03-09T15:13:50.204 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 15:13:49 vm04 ceph-mon[53998]: Unable to set osd_memory_target on vm04 to 269530726: error parsing value: Value '269530726' is below minimum 939524096 2026-03-09T15:13:50.204 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 15:13:49 vm04 ceph-mon[53998]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T15:13:50.204 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 15:13:49 vm04 ceph-mon[53998]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T15:13:50.204 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 15:13:49 vm04 ceph-mon[53998]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' 2026-03-09T15:13:50.204 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 15:13:49 vm04 ceph-mon[53998]: from='client.14262 -' entity='client.admin' cmd=[{"prefix": "orch daemon add osd", "svc_arg": "vm06:/dev/vde", "target": ["mon-mgr", ""]}]: dispatch 2026-03-09T15:13:50.204 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 15:13:49 vm04 ceph-mon[53998]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' cmd=[{"prefix": "osd tree", "states": ["destroyed"], "format": "json"}]: dispatch 2026-03-09T15:13:50.204 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 15:13:49 vm04 ceph-mon[53998]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.bootstrap-osd"}]: dispatch 2026-03-09T15:13:50.204 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 15:13:49 vm04 ceph-mon[53998]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T15:13:50.299 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:13:49 vm03 ceph-mon[51762]: from='osd.1 [v2:192.168.123.104:6800/1573535831,v1:192.168.123.104:6801/1573535831]' entity='osd.1' cmd=[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["1"]}]: dispatch 2026-03-09T15:13:50.299 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:13:49 vm03 ceph-mon[51762]: Detected new or changed devices on vm04 2026-03-09T15:13:50.299 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:13:49 vm03 ceph-mon[51762]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' 2026-03-09T15:13:50.299 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:13:49 vm03 ceph-mon[51762]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' 2026-03-09T15:13:50.299 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:13:49 vm03 ceph-mon[51762]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' cmd=[{"prefix": "config rm", "who": "osd.1", "name": "osd_memory_target"}]: dispatch 2026-03-09T15:13:50.299 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:13:49 vm03 ceph-mon[51762]: Adjusting osd_memory_target on vm04 to 257.0M 2026-03-09T15:13:50.299 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:13:49 vm03 ceph-mon[51762]: Unable to set osd_memory_target on vm04 to 269530726: error parsing value: Value '269530726' is below minimum 939524096 2026-03-09T15:13:50.299 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:13:49 vm03 ceph-mon[51762]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T15:13:50.299 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:13:49 vm03 ceph-mon[51762]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T15:13:50.299 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:13:49 vm03 ceph-mon[51762]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' 2026-03-09T15:13:50.299 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:13:49 vm03 ceph-mon[51762]: from='client.14262 -' entity='client.admin' cmd=[{"prefix": "orch daemon add osd", "svc_arg": "vm06:/dev/vde", "target": ["mon-mgr", ""]}]: dispatch 2026-03-09T15:13:50.299 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:13:49 vm03 ceph-mon[51762]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' cmd=[{"prefix": "osd tree", "states": ["destroyed"], "format": "json"}]: dispatch 2026-03-09T15:13:50.299 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:13:49 vm03 ceph-mon[51762]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.bootstrap-osd"}]: dispatch 2026-03-09T15:13:50.299 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:13:49 vm03 ceph-mon[51762]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T15:13:50.501 INFO:journalctl@ceph.osd.1.vm04.stdout:Mar 09 15:13:50 vm04 ceph-4abede80-1bca-11f1-bcd8-9b1da3c03ce5-osd-1[57595]: 2026-03-09T15:13:50.202+0000 7f583c509640 -1 osd.1 0 waiting for initial osdmap 2026-03-09T15:13:50.501 INFO:journalctl@ceph.osd.1.vm04.stdout:Mar 09 15:13:50 vm04 ceph-4abede80-1bca-11f1-bcd8-9b1da3c03ce5-osd-1[57595]: 2026-03-09T15:13:50.206+0000 7f5837b20640 -1 osd.1 11 set_numa_affinity unable to identify public interface '' numa node: (2) No such file or directory 2026-03-09T15:13:51.018 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 09 15:13:50 vm06 ceph-mon[56506]: from='osd.1 [v2:192.168.123.104:6800/1573535831,v1:192.168.123.104:6801/1573535831]' entity='osd.1' cmd='[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["1"]}]': finished 2026-03-09T15:13:51.018 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 09 15:13:50 vm06 ceph-mon[56506]: osdmap e10: 2 total, 1 up, 2 in 2026-03-09T15:13:51.018 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 09 15:13:50 vm06 ceph-mon[56506]: from='osd.1 [v2:192.168.123.104:6800/1573535831,v1:192.168.123.104:6801/1573535831]' entity='osd.1' cmd=[{"prefix": "osd crush create-or-move", "id": 1, "weight":0.0195, "args": ["host=vm04", "root=default"]}]: dispatch 2026-03-09T15:13:51.018 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 09 15:13:50 vm06 ceph-mon[56506]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 1}]: dispatch 2026-03-09T15:13:51.018 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 09 15:13:50 vm06 ceph-mon[56506]: from='client.? 192.168.123.106:0/2372401307' entity='client.bootstrap-osd' cmd=[{"prefix": "osd new", "uuid": "d53d63dc-a0c2-402c-8257-919cdd6b513c"}]: dispatch 2026-03-09T15:13:51.018 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 09 15:13:50 vm06 ceph-mon[56506]: from='client.? ' entity='client.bootstrap-osd' cmd=[{"prefix": "osd new", "uuid": "d53d63dc-a0c2-402c-8257-919cdd6b513c"}]: dispatch 2026-03-09T15:13:51.018 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 09 15:13:50 vm06 ceph-mon[56506]: from='osd.1 [v2:192.168.123.104:6800/1573535831,v1:192.168.123.104:6801/1573535831]' entity='osd.1' cmd='[{"prefix": "osd crush create-or-move", "id": 1, "weight":0.0195, "args": ["host=vm04", "root=default"]}]': finished 2026-03-09T15:13:51.018 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 09 15:13:50 vm06 ceph-mon[56506]: from='client.? ' entity='client.bootstrap-osd' cmd='[{"prefix": "osd new", "uuid": "d53d63dc-a0c2-402c-8257-919cdd6b513c"}]': finished 2026-03-09T15:13:51.018 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 09 15:13:50 vm06 ceph-mon[56506]: osdmap e11: 3 total, 1 up, 3 in 2026-03-09T15:13:51.018 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 09 15:13:50 vm06 ceph-mon[56506]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 1}]: dispatch 2026-03-09T15:13:51.018 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 09 15:13:50 vm06 ceph-mon[56506]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 2}]: dispatch 2026-03-09T15:13:51.018 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 09 15:13:50 vm06 ceph-mon[56506]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 1}]: dispatch 2026-03-09T15:13:51.018 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 09 15:13:50 vm06 ceph-mon[56506]: pgmap v25: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-09T15:13:51.018 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 09 15:13:50 vm06 ceph-mon[56506]: from='client.? 192.168.123.106:0/2261190342' entity='client.bootstrap-osd' cmd=[{"prefix": "mon getmap"}]: dispatch 2026-03-09T15:13:51.251 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 15:13:50 vm04 ceph-mon[53998]: from='osd.1 [v2:192.168.123.104:6800/1573535831,v1:192.168.123.104:6801/1573535831]' entity='osd.1' cmd='[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["1"]}]': finished 2026-03-09T15:13:51.251 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 15:13:50 vm04 ceph-mon[53998]: osdmap e10: 2 total, 1 up, 2 in 2026-03-09T15:13:51.251 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 15:13:50 vm04 ceph-mon[53998]: from='osd.1 [v2:192.168.123.104:6800/1573535831,v1:192.168.123.104:6801/1573535831]' entity='osd.1' cmd=[{"prefix": "osd crush create-or-move", "id": 1, "weight":0.0195, "args": ["host=vm04", "root=default"]}]: dispatch 2026-03-09T15:13:51.251 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 15:13:50 vm04 ceph-mon[53998]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 1}]: dispatch 2026-03-09T15:13:51.251 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 15:13:50 vm04 ceph-mon[53998]: from='client.? 192.168.123.106:0/2372401307' entity='client.bootstrap-osd' cmd=[{"prefix": "osd new", "uuid": "d53d63dc-a0c2-402c-8257-919cdd6b513c"}]: dispatch 2026-03-09T15:13:51.251 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 15:13:50 vm04 ceph-mon[53998]: from='client.? ' entity='client.bootstrap-osd' cmd=[{"prefix": "osd new", "uuid": "d53d63dc-a0c2-402c-8257-919cdd6b513c"}]: dispatch 2026-03-09T15:13:51.251 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 15:13:50 vm04 ceph-mon[53998]: from='osd.1 [v2:192.168.123.104:6800/1573535831,v1:192.168.123.104:6801/1573535831]' entity='osd.1' cmd='[{"prefix": "osd crush create-or-move", "id": 1, "weight":0.0195, "args": ["host=vm04", "root=default"]}]': finished 2026-03-09T15:13:51.251 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 15:13:50 vm04 ceph-mon[53998]: from='client.? ' entity='client.bootstrap-osd' cmd='[{"prefix": "osd new", "uuid": "d53d63dc-a0c2-402c-8257-919cdd6b513c"}]': finished 2026-03-09T15:13:51.251 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 15:13:50 vm04 ceph-mon[53998]: osdmap e11: 3 total, 1 up, 3 in 2026-03-09T15:13:51.251 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 15:13:50 vm04 ceph-mon[53998]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 1}]: dispatch 2026-03-09T15:13:51.251 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 15:13:50 vm04 ceph-mon[53998]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 2}]: dispatch 2026-03-09T15:13:51.251 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 15:13:50 vm04 ceph-mon[53998]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 1}]: dispatch 2026-03-09T15:13:51.251 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 15:13:50 vm04 ceph-mon[53998]: pgmap v25: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-09T15:13:51.251 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 15:13:50 vm04 ceph-mon[53998]: from='client.? 192.168.123.106:0/2261190342' entity='client.bootstrap-osd' cmd=[{"prefix": "mon getmap"}]: dispatch 2026-03-09T15:13:51.299 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:13:50 vm03 ceph-mon[51762]: from='osd.1 [v2:192.168.123.104:6800/1573535831,v1:192.168.123.104:6801/1573535831]' entity='osd.1' cmd='[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["1"]}]': finished 2026-03-09T15:13:51.299 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:13:50 vm03 ceph-mon[51762]: osdmap e10: 2 total, 1 up, 2 in 2026-03-09T15:13:51.299 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:13:50 vm03 ceph-mon[51762]: from='osd.1 [v2:192.168.123.104:6800/1573535831,v1:192.168.123.104:6801/1573535831]' entity='osd.1' cmd=[{"prefix": "osd crush create-or-move", "id": 1, "weight":0.0195, "args": ["host=vm04", "root=default"]}]: dispatch 2026-03-09T15:13:51.299 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:13:50 vm03 ceph-mon[51762]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 1}]: dispatch 2026-03-09T15:13:51.300 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:13:50 vm03 ceph-mon[51762]: from='client.? 192.168.123.106:0/2372401307' entity='client.bootstrap-osd' cmd=[{"prefix": "osd new", "uuid": "d53d63dc-a0c2-402c-8257-919cdd6b513c"}]: dispatch 2026-03-09T15:13:51.300 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:13:50 vm03 ceph-mon[51762]: from='client.? ' entity='client.bootstrap-osd' cmd=[{"prefix": "osd new", "uuid": "d53d63dc-a0c2-402c-8257-919cdd6b513c"}]: dispatch 2026-03-09T15:13:51.300 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:13:50 vm03 ceph-mon[51762]: from='osd.1 [v2:192.168.123.104:6800/1573535831,v1:192.168.123.104:6801/1573535831]' entity='osd.1' cmd='[{"prefix": "osd crush create-or-move", "id": 1, "weight":0.0195, "args": ["host=vm04", "root=default"]}]': finished 2026-03-09T15:13:51.300 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:13:50 vm03 ceph-mon[51762]: from='client.? ' entity='client.bootstrap-osd' cmd='[{"prefix": "osd new", "uuid": "d53d63dc-a0c2-402c-8257-919cdd6b513c"}]': finished 2026-03-09T15:13:51.300 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:13:50 vm03 ceph-mon[51762]: osdmap e11: 3 total, 1 up, 3 in 2026-03-09T15:13:51.300 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:13:50 vm03 ceph-mon[51762]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 1}]: dispatch 2026-03-09T15:13:51.300 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:13:50 vm03 ceph-mon[51762]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 2}]: dispatch 2026-03-09T15:13:51.300 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:13:50 vm03 ceph-mon[51762]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 1}]: dispatch 2026-03-09T15:13:51.300 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:13:50 vm03 ceph-mon[51762]: pgmap v25: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-09T15:13:51.300 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:13:50 vm03 ceph-mon[51762]: from='client.? 192.168.123.106:0/2261190342' entity='client.bootstrap-osd' cmd=[{"prefix": "mon getmap"}]: dispatch 2026-03-09T15:13:52.501 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 15:13:52 vm04 ceph-mon[53998]: purged_snaps scrub starts 2026-03-09T15:13:52.501 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 15:13:52 vm04 ceph-mon[53998]: purged_snaps scrub ok 2026-03-09T15:13:52.501 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 15:13:52 vm04 ceph-mon[53998]: osd.1 [v2:192.168.123.104:6800/1573535831,v1:192.168.123.104:6801/1573535831] boot 2026-03-09T15:13:52.501 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 15:13:52 vm04 ceph-mon[53998]: osdmap e12: 3 total, 2 up, 3 in 2026-03-09T15:13:52.501 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 15:13:52 vm04 ceph-mon[53998]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 1}]: dispatch 2026-03-09T15:13:52.501 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 15:13:52 vm04 ceph-mon[53998]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 2}]: dispatch 2026-03-09T15:13:52.517 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 09 15:13:52 vm06 ceph-mon[56506]: purged_snaps scrub starts 2026-03-09T15:13:52.517 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 09 15:13:52 vm06 ceph-mon[56506]: purged_snaps scrub ok 2026-03-09T15:13:52.517 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 09 15:13:52 vm06 ceph-mon[56506]: osd.1 [v2:192.168.123.104:6800/1573535831,v1:192.168.123.104:6801/1573535831] boot 2026-03-09T15:13:52.517 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 09 15:13:52 vm06 ceph-mon[56506]: osdmap e12: 3 total, 2 up, 3 in 2026-03-09T15:13:52.517 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 09 15:13:52 vm06 ceph-mon[56506]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 1}]: dispatch 2026-03-09T15:13:52.517 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 09 15:13:52 vm06 ceph-mon[56506]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 2}]: dispatch 2026-03-09T15:13:52.549 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:13:52 vm03 ceph-mon[51762]: purged_snaps scrub starts 2026-03-09T15:13:52.549 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:13:52 vm03 ceph-mon[51762]: purged_snaps scrub ok 2026-03-09T15:13:52.549 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:13:52 vm03 ceph-mon[51762]: osd.1 [v2:192.168.123.104:6800/1573535831,v1:192.168.123.104:6801/1573535831] boot 2026-03-09T15:13:52.549 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:13:52 vm03 ceph-mon[51762]: osdmap e12: 3 total, 2 up, 3 in 2026-03-09T15:13:52.549 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:13:52 vm03 ceph-mon[51762]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 1}]: dispatch 2026-03-09T15:13:52.549 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:13:52 vm03 ceph-mon[51762]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 2}]: dispatch 2026-03-09T15:13:53.501 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 15:13:53 vm04 ceph-mon[53998]: osdmap e13: 3 total, 2 up, 3 in 2026-03-09T15:13:53.501 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 15:13:53 vm04 ceph-mon[53998]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 2}]: dispatch 2026-03-09T15:13:53.501 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 15:13:53 vm04 ceph-mon[53998]: pgmap v28: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-09T15:13:53.517 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 09 15:13:53 vm06 ceph-mon[56506]: osdmap e13: 3 total, 2 up, 3 in 2026-03-09T15:13:53.517 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 09 15:13:53 vm06 ceph-mon[56506]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 2}]: dispatch 2026-03-09T15:13:53.517 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 09 15:13:53 vm06 ceph-mon[56506]: pgmap v28: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-09T15:13:53.549 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:13:53 vm03 ceph-mon[51762]: osdmap e13: 3 total, 2 up, 3 in 2026-03-09T15:13:53.549 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:13:53 vm03 ceph-mon[51762]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 2}]: dispatch 2026-03-09T15:13:53.549 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:13:53 vm03 ceph-mon[51762]: pgmap v28: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-09T15:13:54.462 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 09 15:13:54 vm06 ceph-mon[56506]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "osd.2"}]: dispatch 2026-03-09T15:13:54.462 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 09 15:13:54 vm06 ceph-mon[56506]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T15:13:54.751 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 15:13:54 vm04 ceph-mon[53998]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "osd.2"}]: dispatch 2026-03-09T15:13:54.751 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 15:13:54 vm04 ceph-mon[53998]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T15:13:54.799 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:13:54 vm03 ceph-mon[51762]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "osd.2"}]: dispatch 2026-03-09T15:13:54.799 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:13:54 vm03 ceph-mon[51762]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T15:13:55.518 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 09 15:13:55 vm06 ceph-mon[56506]: Deploying daemon osd.2 on vm06 2026-03-09T15:13:55.518 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 09 15:13:55 vm06 ceph-mon[56506]: pgmap v29: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-09T15:13:55.751 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 15:13:55 vm04 ceph-mon[53998]: Deploying daemon osd.2 on vm06 2026-03-09T15:13:55.751 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 15:13:55 vm04 ceph-mon[53998]: pgmap v29: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-09T15:13:55.799 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:13:55 vm03 ceph-mon[51762]: Deploying daemon osd.2 on vm06 2026-03-09T15:13:55.799 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:13:55 vm03 ceph-mon[51762]: pgmap v29: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-09T15:13:56.597 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 09 15:13:56 vm06 ceph-mon[56506]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T15:13:56.597 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 09 15:13:56 vm06 ceph-mon[56506]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' 2026-03-09T15:13:56.597 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 09 15:13:56 vm06 ceph-mon[56506]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' 2026-03-09T15:13:56.751 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 15:13:56 vm04 ceph-mon[53998]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T15:13:56.751 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 15:13:56 vm04 ceph-mon[53998]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' 2026-03-09T15:13:56.751 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 15:13:56 vm04 ceph-mon[53998]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' 2026-03-09T15:13:56.799 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:13:56 vm03 ceph-mon[51762]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T15:13:56.799 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:13:56 vm03 ceph-mon[51762]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' 2026-03-09T15:13:56.799 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:13:56 vm03 ceph-mon[51762]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' 2026-03-09T15:13:57.228 INFO:teuthology.orchestra.run.vm06.stdout:Created osd(s) 2 on host 'vm06' 2026-03-09T15:13:57.287 DEBUG:teuthology.orchestra.run.vm06:osd.2> sudo journalctl -f -n 0 -u ceph-4abede80-1bca-11f1-bcd8-9b1da3c03ce5@osd.2.service 2026-03-09T15:13:57.289 INFO:tasks.cephadm:Waiting for 3 OSDs to come up... 2026-03-09T15:13:57.289 DEBUG:teuthology.orchestra.run.vm03:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 4abede80-1bca-11f1-bcd8-9b1da3c03ce5 -- ceph osd stat -f json 2026-03-09T15:13:57.466 INFO:teuthology.orchestra.run.vm03.stderr:Inferring config /var/lib/ceph/4abede80-1bca-11f1-bcd8-9b1da3c03ce5/mon.a/config 2026-03-09T15:13:57.581 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:13:57 vm03 ceph-mon[51762]: pgmap v30: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-09T15:13:57.581 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:13:57 vm03 ceph-mon[51762]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' 2026-03-09T15:13:57.581 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:13:57 vm03 ceph-mon[51762]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' 2026-03-09T15:13:57.581 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:13:57 vm03 ceph-mon[51762]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T15:13:57.581 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:13:57 vm03 ceph-mon[51762]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T15:13:57.581 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:13:57 vm03 ceph-mon[51762]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' 2026-03-09T15:13:57.581 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:13:57 vm03 ceph-mon[51762]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T15:13:57.581 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:13:57 vm03 ceph-mon[51762]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' 2026-03-09T15:13:57.581 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:13:57 vm03 ceph-mon[51762]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' 2026-03-09T15:13:57.585 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 09 15:13:57 vm06 ceph-mon[56506]: pgmap v30: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-09T15:13:57.585 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 09 15:13:57 vm06 ceph-mon[56506]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' 2026-03-09T15:13:57.585 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 09 15:13:57 vm06 ceph-mon[56506]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' 2026-03-09T15:13:57.586 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 09 15:13:57 vm06 ceph-mon[56506]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T15:13:57.586 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 09 15:13:57 vm06 ceph-mon[56506]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T15:13:57.586 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 09 15:13:57 vm06 ceph-mon[56506]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' 2026-03-09T15:13:57.586 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 09 15:13:57 vm06 ceph-mon[56506]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T15:13:57.586 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 09 15:13:57 vm06 ceph-mon[56506]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' 2026-03-09T15:13:57.586 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 09 15:13:57 vm06 ceph-mon[56506]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' 2026-03-09T15:13:57.685 INFO:teuthology.orchestra.run.vm03.stdout: 2026-03-09T15:13:57.746 INFO:teuthology.orchestra.run.vm03.stdout:{"epoch":13,"num_osds":3,"num_up_osds":2,"osd_up_since":1773069231,"num_in_osds":3,"osd_in_since":1773069230,"num_remapped_pgs":0} 2026-03-09T15:13:57.751 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 15:13:57 vm04 ceph-mon[53998]: pgmap v30: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-09T15:13:57.751 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 15:13:57 vm04 ceph-mon[53998]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' 2026-03-09T15:13:57.751 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 15:13:57 vm04 ceph-mon[53998]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' 2026-03-09T15:13:57.751 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 15:13:57 vm04 ceph-mon[53998]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T15:13:57.751 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 15:13:57 vm04 ceph-mon[53998]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T15:13:57.751 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 15:13:57 vm04 ceph-mon[53998]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' 2026-03-09T15:13:57.751 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 15:13:57 vm04 ceph-mon[53998]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T15:13:57.751 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 15:13:57 vm04 ceph-mon[53998]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' 2026-03-09T15:13:57.751 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 15:13:57 vm04 ceph-mon[53998]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' 2026-03-09T15:13:58.192 INFO:journalctl@ceph.osd.2.vm06.stdout:Mar 09 15:13:57 vm06 ceph-4abede80-1bca-11f1-bcd8-9b1da3c03ce5-osd-2[59842]: 2026-03-09T15:13:57.842+0000 7ff436b8a740 -1 osd.2 0 log_to_monitors true 2026-03-09T15:13:58.517 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 09 15:13:58 vm06 ceph-mon[56506]: from='client.? 192.168.123.103:0/1222053211' entity='client.admin' cmd=[{"prefix": "osd stat", "format": "json"}]: dispatch 2026-03-09T15:13:58.517 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 09 15:13:58 vm06 ceph-mon[56506]: from='osd.2 [v2:192.168.123.106:6800/3722863356,v1:192.168.123.106:6801/3722863356]' entity='osd.2' cmd=[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["2"]}]: dispatch 2026-03-09T15:13:58.517 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 09 15:13:58 vm06 ceph-mon[56506]: from='osd.2 ' entity='osd.2' cmd=[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["2"]}]: dispatch 2026-03-09T15:13:58.517 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 09 15:13:58 vm06 ceph-mon[56506]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' 2026-03-09T15:13:58.517 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 09 15:13:58 vm06 ceph-mon[56506]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' 2026-03-09T15:13:58.517 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 09 15:13:58 vm06 ceph-mon[56506]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' cmd=[{"prefix": "config rm", "who": "osd.2", "name": "osd_memory_target"}]: dispatch 2026-03-09T15:13:58.517 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 09 15:13:58 vm06 ceph-mon[56506]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' 2026-03-09T15:13:58.517 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 09 15:13:58 vm06 ceph-mon[56506]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T15:13:58.517 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 09 15:13:58 vm06 ceph-mon[56506]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T15:13:58.517 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 09 15:13:58 vm06 ceph-mon[56506]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' 2026-03-09T15:13:58.747 DEBUG:teuthology.orchestra.run.vm03:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 4abede80-1bca-11f1-bcd8-9b1da3c03ce5 -- ceph osd stat -f json 2026-03-09T15:13:58.751 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 15:13:58 vm04 ceph-mon[53998]: from='client.? 192.168.123.103:0/1222053211' entity='client.admin' cmd=[{"prefix": "osd stat", "format": "json"}]: dispatch 2026-03-09T15:13:58.751 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 15:13:58 vm04 ceph-mon[53998]: from='osd.2 [v2:192.168.123.106:6800/3722863356,v1:192.168.123.106:6801/3722863356]' entity='osd.2' cmd=[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["2"]}]: dispatch 2026-03-09T15:13:58.751 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 15:13:58 vm04 ceph-mon[53998]: from='osd.2 ' entity='osd.2' cmd=[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["2"]}]: dispatch 2026-03-09T15:13:58.751 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 15:13:58 vm04 ceph-mon[53998]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' 2026-03-09T15:13:58.751 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 15:13:58 vm04 ceph-mon[53998]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' 2026-03-09T15:13:58.751 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 15:13:58 vm04 ceph-mon[53998]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' cmd=[{"prefix": "config rm", "who": "osd.2", "name": "osd_memory_target"}]: dispatch 2026-03-09T15:13:58.751 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 15:13:58 vm04 ceph-mon[53998]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' 2026-03-09T15:13:58.751 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 15:13:58 vm04 ceph-mon[53998]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T15:13:58.751 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 15:13:58 vm04 ceph-mon[53998]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T15:13:58.751 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 15:13:58 vm04 ceph-mon[53998]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' 2026-03-09T15:13:58.772 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:13:58 vm03 ceph-mon[51762]: from='client.? 192.168.123.103:0/1222053211' entity='client.admin' cmd=[{"prefix": "osd stat", "format": "json"}]: dispatch 2026-03-09T15:13:58.772 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:13:58 vm03 ceph-mon[51762]: from='osd.2 [v2:192.168.123.106:6800/3722863356,v1:192.168.123.106:6801/3722863356]' entity='osd.2' cmd=[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["2"]}]: dispatch 2026-03-09T15:13:58.772 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:13:58 vm03 ceph-mon[51762]: from='osd.2 ' entity='osd.2' cmd=[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["2"]}]: dispatch 2026-03-09T15:13:58.773 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:13:58 vm03 ceph-mon[51762]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' 2026-03-09T15:13:58.773 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:13:58 vm03 ceph-mon[51762]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' 2026-03-09T15:13:58.773 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:13:58 vm03 ceph-mon[51762]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' cmd=[{"prefix": "config rm", "who": "osd.2", "name": "osd_memory_target"}]: dispatch 2026-03-09T15:13:58.773 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:13:58 vm03 ceph-mon[51762]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' 2026-03-09T15:13:58.773 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:13:58 vm03 ceph-mon[51762]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T15:13:58.773 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:13:58 vm03 ceph-mon[51762]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T15:13:58.773 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:13:58 vm03 ceph-mon[51762]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' 2026-03-09T15:13:58.923 INFO:teuthology.orchestra.run.vm03.stderr:Inferring config /var/lib/ceph/4abede80-1bca-11f1-bcd8-9b1da3c03ce5/mon.a/config 2026-03-09T15:13:59.155 INFO:teuthology.orchestra.run.vm03.stdout: 2026-03-09T15:13:59.203 INFO:teuthology.orchestra.run.vm03.stdout:{"epoch":14,"num_osds":3,"num_up_osds":2,"osd_up_since":1773069231,"num_in_osds":3,"osd_in_since":1773069230,"num_remapped_pgs":0} 2026-03-09T15:13:59.751 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 15:13:59 vm04 ceph-mon[53998]: Detected new or changed devices on vm06 2026-03-09T15:13:59.751 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 15:13:59 vm04 ceph-mon[53998]: Adjusting osd_memory_target on vm06 to 4353M 2026-03-09T15:13:59.751 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 15:13:59 vm04 ceph-mon[53998]: from='osd.2 ' entity='osd.2' cmd='[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["2"]}]': finished 2026-03-09T15:13:59.751 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 15:13:59 vm04 ceph-mon[53998]: from='osd.2 [v2:192.168.123.106:6800/3722863356,v1:192.168.123.106:6801/3722863356]' entity='osd.2' cmd=[{"prefix": "osd crush create-or-move", "id": 2, "weight":0.0195, "args": ["host=vm06", "root=default"]}]: dispatch 2026-03-09T15:13:59.751 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 15:13:59 vm04 ceph-mon[53998]: osdmap e14: 3 total, 2 up, 3 in 2026-03-09T15:13:59.751 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 15:13:59 vm04 ceph-mon[53998]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 2}]: dispatch 2026-03-09T15:13:59.751 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 15:13:59 vm04 ceph-mon[53998]: from='osd.2 ' entity='osd.2' cmd=[{"prefix": "osd crush create-or-move", "id": 2, "weight":0.0195, "args": ["host=vm06", "root=default"]}]: dispatch 2026-03-09T15:13:59.751 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 15:13:59 vm04 ceph-mon[53998]: pgmap v32: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-09T15:13:59.751 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 15:13:59 vm04 ceph-mon[53998]: from='client.? 192.168.123.103:0/2837263933' entity='client.admin' cmd=[{"prefix": "osd stat", "format": "json"}]: dispatch 2026-03-09T15:13:59.751 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 15:13:59 vm04 ceph-mon[53998]: from='osd.2 ' entity='osd.2' cmd='[{"prefix": "osd crush create-or-move", "id": 2, "weight":0.0195, "args": ["host=vm06", "root=default"]}]': finished 2026-03-09T15:13:59.751 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 15:13:59 vm04 ceph-mon[53998]: osdmap e15: 3 total, 2 up, 3 in 2026-03-09T15:13:59.767 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 09 15:13:59 vm06 ceph-mon[56506]: Detected new or changed devices on vm06 2026-03-09T15:13:59.767 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 09 15:13:59 vm06 ceph-mon[56506]: Adjusting osd_memory_target on vm06 to 4353M 2026-03-09T15:13:59.767 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 09 15:13:59 vm06 ceph-mon[56506]: from='osd.2 ' entity='osd.2' cmd='[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["2"]}]': finished 2026-03-09T15:13:59.767 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 09 15:13:59 vm06 ceph-mon[56506]: from='osd.2 [v2:192.168.123.106:6800/3722863356,v1:192.168.123.106:6801/3722863356]' entity='osd.2' cmd=[{"prefix": "osd crush create-or-move", "id": 2, "weight":0.0195, "args": ["host=vm06", "root=default"]}]: dispatch 2026-03-09T15:13:59.767 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 09 15:13:59 vm06 ceph-mon[56506]: osdmap e14: 3 total, 2 up, 3 in 2026-03-09T15:13:59.767 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 09 15:13:59 vm06 ceph-mon[56506]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 2}]: dispatch 2026-03-09T15:13:59.767 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 09 15:13:59 vm06 ceph-mon[56506]: from='osd.2 ' entity='osd.2' cmd=[{"prefix": "osd crush create-or-move", "id": 2, "weight":0.0195, "args": ["host=vm06", "root=default"]}]: dispatch 2026-03-09T15:13:59.767 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 09 15:13:59 vm06 ceph-mon[56506]: pgmap v32: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-09T15:13:59.767 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 09 15:13:59 vm06 ceph-mon[56506]: from='client.? 192.168.123.103:0/2837263933' entity='client.admin' cmd=[{"prefix": "osd stat", "format": "json"}]: dispatch 2026-03-09T15:13:59.767 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 09 15:13:59 vm06 ceph-mon[56506]: from='osd.2 ' entity='osd.2' cmd='[{"prefix": "osd crush create-or-move", "id": 2, "weight":0.0195, "args": ["host=vm06", "root=default"]}]': finished 2026-03-09T15:13:59.767 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 09 15:13:59 vm06 ceph-mon[56506]: osdmap e15: 3 total, 2 up, 3 in 2026-03-09T15:13:59.767 INFO:journalctl@ceph.osd.2.vm06.stdout:Mar 09 15:13:59 vm06 ceph-4abede80-1bca-11f1-bcd8-9b1da3c03ce5-osd-2[59842]: 2026-03-09T15:13:59.347+0000 7ff432b0b640 -1 osd.2 0 waiting for initial osdmap 2026-03-09T15:13:59.767 INFO:journalctl@ceph.osd.2.vm06.stdout:Mar 09 15:13:59 vm06 ceph-4abede80-1bca-11f1-bcd8-9b1da3c03ce5-osd-2[59842]: 2026-03-09T15:13:59.352+0000 7ff42e935640 -1 osd.2 15 set_numa_affinity unable to identify public interface '' numa node: (2) No such file or directory 2026-03-09T15:13:59.799 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:13:59 vm03 ceph-mon[51762]: Detected new or changed devices on vm06 2026-03-09T15:13:59.799 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:13:59 vm03 ceph-mon[51762]: Adjusting osd_memory_target on vm06 to 4353M 2026-03-09T15:13:59.799 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:13:59 vm03 ceph-mon[51762]: from='osd.2 ' entity='osd.2' cmd='[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["2"]}]': finished 2026-03-09T15:13:59.799 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:13:59 vm03 ceph-mon[51762]: from='osd.2 [v2:192.168.123.106:6800/3722863356,v1:192.168.123.106:6801/3722863356]' entity='osd.2' cmd=[{"prefix": "osd crush create-or-move", "id": 2, "weight":0.0195, "args": ["host=vm06", "root=default"]}]: dispatch 2026-03-09T15:13:59.799 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:13:59 vm03 ceph-mon[51762]: osdmap e14: 3 total, 2 up, 3 in 2026-03-09T15:13:59.799 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:13:59 vm03 ceph-mon[51762]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 2}]: dispatch 2026-03-09T15:13:59.799 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:13:59 vm03 ceph-mon[51762]: from='osd.2 ' entity='osd.2' cmd=[{"prefix": "osd crush create-or-move", "id": 2, "weight":0.0195, "args": ["host=vm06", "root=default"]}]: dispatch 2026-03-09T15:13:59.799 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:13:59 vm03 ceph-mon[51762]: pgmap v32: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-09T15:13:59.799 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:13:59 vm03 ceph-mon[51762]: from='client.? 192.168.123.103:0/2837263933' entity='client.admin' cmd=[{"prefix": "osd stat", "format": "json"}]: dispatch 2026-03-09T15:13:59.799 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:13:59 vm03 ceph-mon[51762]: from='osd.2 ' entity='osd.2' cmd='[{"prefix": "osd crush create-or-move", "id": 2, "weight":0.0195, "args": ["host=vm06", "root=default"]}]': finished 2026-03-09T15:13:59.799 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:13:59 vm03 ceph-mon[51762]: osdmap e15: 3 total, 2 up, 3 in 2026-03-09T15:14:00.204 DEBUG:teuthology.orchestra.run.vm03:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 4abede80-1bca-11f1-bcd8-9b1da3c03ce5 -- ceph osd stat -f json 2026-03-09T15:14:00.371 INFO:teuthology.orchestra.run.vm03.stderr:Inferring config /var/lib/ceph/4abede80-1bca-11f1-bcd8-9b1da3c03ce5/mon.a/config 2026-03-09T15:14:00.486 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:14:00 vm03 ceph-mon[51762]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 2}]: dispatch 2026-03-09T15:14:00.486 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:14:00 vm03 ceph-mon[51762]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 2}]: dispatch 2026-03-09T15:14:00.486 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:14:00 vm03 ceph-mon[51762]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 2}]: dispatch 2026-03-09T15:14:00.486 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:14:00 vm03 ceph-mon[51762]: osd.2 [v2:192.168.123.106:6800/3722863356,v1:192.168.123.106:6801/3722863356] boot 2026-03-09T15:14:00.486 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:14:00 vm03 ceph-mon[51762]: osdmap e16: 3 total, 3 up, 3 in 2026-03-09T15:14:00.486 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:14:00 vm03 ceph-mon[51762]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 2}]: dispatch 2026-03-09T15:14:00.594 INFO:teuthology.orchestra.run.vm03.stdout: 2026-03-09T15:14:00.653 INFO:teuthology.orchestra.run.vm03.stdout:{"epoch":16,"num_osds":3,"num_up_osds":3,"osd_up_since":1773069240,"num_in_osds":3,"osd_in_since":1773069230,"num_remapped_pgs":0} 2026-03-09T15:14:00.653 DEBUG:teuthology.orchestra.run.vm03:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell --fsid 4abede80-1bca-11f1-bcd8-9b1da3c03ce5 -- ceph osd dump --format=json 2026-03-09T15:14:00.751 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 15:14:00 vm04 ceph-mon[53998]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 2}]: dispatch 2026-03-09T15:14:00.751 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 15:14:00 vm04 ceph-mon[53998]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 2}]: dispatch 2026-03-09T15:14:00.751 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 15:14:00 vm04 ceph-mon[53998]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 2}]: dispatch 2026-03-09T15:14:00.751 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 15:14:00 vm04 ceph-mon[53998]: osd.2 [v2:192.168.123.106:6800/3722863356,v1:192.168.123.106:6801/3722863356] boot 2026-03-09T15:14:00.751 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 15:14:00 vm04 ceph-mon[53998]: osdmap e16: 3 total, 3 up, 3 in 2026-03-09T15:14:00.751 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 15:14:00 vm04 ceph-mon[53998]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 2}]: dispatch 2026-03-09T15:14:00.767 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 09 15:14:00 vm06 ceph-mon[56506]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 2}]: dispatch 2026-03-09T15:14:00.767 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 09 15:14:00 vm06 ceph-mon[56506]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 2}]: dispatch 2026-03-09T15:14:00.767 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 09 15:14:00 vm06 ceph-mon[56506]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 2}]: dispatch 2026-03-09T15:14:00.767 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 09 15:14:00 vm06 ceph-mon[56506]: osd.2 [v2:192.168.123.106:6800/3722863356,v1:192.168.123.106:6801/3722863356] boot 2026-03-09T15:14:00.767 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 09 15:14:00 vm06 ceph-mon[56506]: osdmap e16: 3 total, 3 up, 3 in 2026-03-09T15:14:00.767 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 09 15:14:00 vm06 ceph-mon[56506]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 2}]: dispatch 2026-03-09T15:14:00.811 INFO:teuthology.orchestra.run.vm03.stderr:Inferring config /var/lib/ceph/4abede80-1bca-11f1-bcd8-9b1da3c03ce5/mon.a/config 2026-03-09T15:14:01.022 INFO:teuthology.orchestra.run.vm03.stdout: 2026-03-09T15:14:01.023 INFO:teuthology.orchestra.run.vm03.stdout:{"epoch":16,"fsid":"4abede80-1bca-11f1-bcd8-9b1da3c03ce5","created":"2026-03-09T15:12:42.233631+0000","modified":"2026-03-09T15:14:00.353477+0000","last_up_change":"2026-03-09T15:14:00.353477+0000","last_in_change":"2026-03-09T15:13:50.194917+0000","flags":"sortbitwise,recovery_deletes,purged_snapdirs,pglog_hardlimit","flags_num":5799936,"flags_set":["pglog_hardlimit","purged_snapdirs","recovery_deletes","sortbitwise"],"crush_version":8,"full_ratio":0.94999998807907104,"backfillfull_ratio":0.89999997615814209,"nearfull_ratio":0.85000002384185791,"cluster_snapshot":"","pool_max":0,"max_osd":3,"require_min_compat_client":"luminous","min_compat_client":"jewel","require_osd_release":"squid","allow_crimson":false,"pools":[],"osds":[{"osd":0,"uuid":"913171bb-10e5-4020-a31f-9550e7c31d7b","up":1,"in":1,"weight":1,"primary_affinity":1,"last_clean_begin":0,"last_clean_end":0,"up_from":8,"up_thru":0,"down_at":0,"lost_at":0,"public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.103:6802","nonce":615876468},{"type":"v1","addr":"192.168.123.103:6803","nonce":615876468}]},"cluster_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.103:6804","nonce":615876468},{"type":"v1","addr":"192.168.123.103:6805","nonce":615876468}]},"heartbeat_back_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.103:6808","nonce":615876468},{"type":"v1","addr":"192.168.123.103:6809","nonce":615876468}]},"heartbeat_front_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.103:6806","nonce":615876468},{"type":"v1","addr":"192.168.123.103:6807","nonce":615876468}]},"public_addr":"192.168.123.103:6803/615876468","cluster_addr":"192.168.123.103:6805/615876468","heartbeat_back_addr":"192.168.123.103:6809/615876468","heartbeat_front_addr":"192.168.123.103:6807/615876468","state":["exists","up"]},{"osd":1,"uuid":"62ee7f4e-6ba5-481b-9cde-4c62baacf802","up":1,"in":1,"weight":1,"primary_affinity":1,"last_clean_begin":0,"last_clean_end":0,"up_from":12,"up_thru":0,"down_at":0,"lost_at":0,"public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.104:6800","nonce":1573535831},{"type":"v1","addr":"192.168.123.104:6801","nonce":1573535831}]},"cluster_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.104:6802","nonce":1573535831},{"type":"v1","addr":"192.168.123.104:6803","nonce":1573535831}]},"heartbeat_back_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.104:6806","nonce":1573535831},{"type":"v1","addr":"192.168.123.104:6807","nonce":1573535831}]},"heartbeat_front_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.104:6804","nonce":1573535831},{"type":"v1","addr":"192.168.123.104:6805","nonce":1573535831}]},"public_addr":"192.168.123.104:6801/1573535831","cluster_addr":"192.168.123.104:6803/1573535831","heartbeat_back_addr":"192.168.123.104:6807/1573535831","heartbeat_front_addr":"192.168.123.104:6805/1573535831","state":["exists","up"]},{"osd":2,"uuid":"d53d63dc-a0c2-402c-8257-919cdd6b513c","up":1,"in":1,"weight":1,"primary_affinity":1,"last_clean_begin":0,"last_clean_end":0,"up_from":16,"up_thru":0,"down_at":0,"lost_at":0,"public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.106:6800","nonce":3722863356},{"type":"v1","addr":"192.168.123.106:6801","nonce":3722863356}]},"cluster_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.106:6802","nonce":3722863356},{"type":"v1","addr":"192.168.123.106:6803","nonce":3722863356}]},"heartbeat_back_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.106:6806","nonce":3722863356},{"type":"v1","addr":"192.168.123.106:6807","nonce":3722863356}]},"heartbeat_front_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.106:6804","nonce":3722863356},{"type":"v1","addr":"192.168.123.106:6805","nonce":3722863356}]},"public_addr":"192.168.123.106:6801/3722863356","cluster_addr":"192.168.123.106:6803/3722863356","heartbeat_back_addr":"192.168.123.106:6807/3722863356","heartbeat_front_addr":"192.168.123.106:6805/3722863356","state":["exists","up"]}],"osd_xinfo":[{"osd":0,"down_stamp":"0.000000","laggy_probability":0,"laggy_interval":0,"features":4540701547738038271,"old_weight":0,"last_purged_snaps_scrub":"2026-03-09T15:13:40.595321+0000","dead_epoch":0},{"osd":1,"down_stamp":"0.000000","laggy_probability":0,"laggy_interval":0,"features":4540701547738038271,"old_weight":0,"last_purged_snaps_scrub":"2026-03-09T15:13:49.850182+0000","dead_epoch":0},{"osd":2,"down_stamp":"0.000000","laggy_probability":0,"laggy_interval":0,"features":4540701547738038271,"old_weight":0,"last_purged_snaps_scrub":"0.000000","dead_epoch":0}],"pg_upmap":[],"pg_upmap_items":[],"pg_upmap_primaries":[],"pg_temp":[],"primary_temp":[],"blocklist":{"192.168.123.103:6801/3510874031":"2026-03-10T15:13:02.478314+0000","192.168.123.103:6800/3510874031":"2026-03-10T15:13:02.478314+0000","192.168.123.103:0/1665269914":"2026-03-10T15:13:02.478314+0000","192.168.123.103:0/4246730401":"2026-03-10T15:13:02.478314+0000","192.168.123.103:0/983012295":"2026-03-10T15:12:52.735172+0000","192.168.123.103:0/3380842377":"2026-03-10T15:13:02.478314+0000","192.168.123.103:0/1826640327":"2026-03-10T15:12:52.735172+0000","192.168.123.103:6800/2441273787":"2026-03-10T15:12:52.735172+0000","192.168.123.103:6801/2441273787":"2026-03-10T15:12:52.735172+0000","192.168.123.103:0/4187493480":"2026-03-10T15:12:52.735172+0000"},"range_blocklist":{},"erasure_code_profiles":{"default":{"crush-failure-domain":"osd","k":"2","m":"1","plugin":"jerasure","technique":"reed_sol_van"}},"removed_snaps_queue":[],"new_removed_snaps":[],"new_purged_snaps":[],"crush_node_flags":{},"device_class_flags":{},"stretch_mode":{"stretch_mode_enabled":false,"stretch_bucket_count":0,"degraded_stretch_mode":0,"recovering_stretch_mode":0,"stretch_mode_bucket":0}} 2026-03-09T15:14:01.082 INFO:tasks.cephadm.ceph_manager.ceph:[] 2026-03-09T15:14:01.083 INFO:tasks.cephadm:Setting up client nodes... 2026-03-09T15:14:01.083 INFO:tasks.ceph:Waiting until ceph daemons up and pgs clean... 2026-03-09T15:14:01.083 INFO:tasks.cephadm.ceph_manager.ceph:waiting for mgr available 2026-03-09T15:14:01.083 DEBUG:teuthology.orchestra.run.vm03:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell --fsid 4abede80-1bca-11f1-bcd8-9b1da3c03ce5 -- ceph mgr dump --format=json 2026-03-09T15:14:01.236 INFO:teuthology.orchestra.run.vm03.stderr:Inferring config /var/lib/ceph/4abede80-1bca-11f1-bcd8-9b1da3c03ce5/mon.a/config 2026-03-09T15:14:01.372 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:14:01 vm03 ceph-mon[51762]: purged_snaps scrub starts 2026-03-09T15:14:01.373 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:14:01 vm03 ceph-mon[51762]: purged_snaps scrub ok 2026-03-09T15:14:01.373 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:14:01 vm03 ceph-mon[51762]: pgmap v35: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-09T15:14:01.373 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:14:01 vm03 ceph-mon[51762]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' cmd=[{"prefix": "osd pool create", "format": "json", "pool": ".mgr", "pg_num": 1, "pg_num_min": 1, "pg_num_max": 32, "yes_i_really_mean_it": true}]: dispatch 2026-03-09T15:14:01.373 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:14:01 vm03 ceph-mon[51762]: from='client.? 192.168.123.103:0/2604958013' entity='client.admin' cmd=[{"prefix": "osd stat", "format": "json"}]: dispatch 2026-03-09T15:14:01.373 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:14:01 vm03 ceph-mon[51762]: from='client.? 192.168.123.103:0/3444384227' entity='client.admin' cmd=[{"prefix": "osd dump", "format": "json"}]: dispatch 2026-03-09T15:14:01.467 INFO:teuthology.orchestra.run.vm03.stdout: 2026-03-09T15:14:01.514 INFO:teuthology.orchestra.run.vm03.stdout:{"epoch":13,"flags":0,"active_gid":14150,"active_name":"a","active_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.103:6800","nonce":3524444278},{"type":"v1","addr":"192.168.123.103:6801","nonce":3524444278}]},"active_addr":"192.168.123.103:6801/3524444278","active_change":"2026-03-09T15:13:02.478411+0000","active_mgr_features":4540701547738038271,"available":true,"standbys":[{"gid":14208,"name":"b","mgr_features":4540701547738038271,"available_modules":[{"name":"alerts","can_run":true,"error_string":"","module_options":{"interval":{"name":"interval","type":"secs","level":"advanced","flags":1,"default_value":"60","min":"","max":"","enum_allowed":[],"desc":"How frequently to reexamine health status","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"smtp_destination":{"name":"smtp_destination","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"Email address to send alerts to","long_desc":"","tags":[],"see_also":[]},"smtp_from_name":{"name":"smtp_from_name","type":"str","level":"advanced","flags":1,"default_value":"Ceph","min":"","max":"","enum_allowed":[],"desc":"Email From: name","long_desc":"","tags":[],"see_also":[]},"smtp_host":{"name":"smtp_host","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"SMTP server","long_desc":"","tags":[],"see_also":[]},"smtp_password":{"name":"smtp_password","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"Password to authenticate with","long_desc":"","tags":[],"see_also":[]},"smtp_port":{"name":"smtp_port","type":"int","level":"advanced","flags":1,"default_value":"465","min":"","max":"","enum_allowed":[],"desc":"SMTP port","long_desc":"","tags":[],"see_also":[]},"smtp_sender":{"name":"smtp_sender","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"SMTP envelope sender","long_desc":"","tags":[],"see_also":[]},"smtp_ssl":{"name":"smtp_ssl","type":"bool","level":"advanced","flags":1,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"Use SSL to connect to SMTP server","long_desc":"","tags":[],"see_also":[]},"smtp_user":{"name":"smtp_user","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"User to authenticate as","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"balancer","can_run":true,"error_string":"","module_options":{"active":{"name":"active","type":"bool","level":"advanced","flags":1,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"automatically balance PGs across cluster","long_desc":"","tags":[],"see_also":[]},"begin_time":{"name":"begin_time","type":"str","level":"advanced","flags":1,"default_value":"0000","min":"","max":"","enum_allowed":[],"desc":"beginning time of day to automatically balance","long_desc":"This is a time of day in the format HHMM.","tags":[],"see_also":[]},"begin_weekday":{"name":"begin_weekday","type":"uint","level":"advanced","flags":1,"default_value":"0","min":"0","max":"6","enum_allowed":[],"desc":"Restrict automatic balancing to this day of the week or later","long_desc":"0 = Sunday, 1 = Monday, etc.","tags":[],"see_also":[]},"crush_compat_max_iterations":{"name":"crush_compat_max_iterations","type":"uint","level":"advanced","flags":1,"default_value":"25","min":"1","max":"250","enum_allowed":[],"desc":"maximum number of iterations to attempt optimization","long_desc":"","tags":[],"see_also":[]},"crush_compat_metrics":{"name":"crush_compat_metrics","type":"str","level":"advanced","flags":1,"default_value":"pgs,objects,bytes","min":"","max":"","enum_allowed":[],"desc":"metrics with which to calculate OSD utilization","long_desc":"Value is a list of one or more of \"pgs\", \"objects\", or \"bytes\", and indicates which metrics to use to balance utilization.","tags":[],"see_also":[]},"crush_compat_step":{"name":"crush_compat_step","type":"float","level":"advanced","flags":1,"default_value":"0.5","min":"0.001","max":"0.999","enum_allowed":[],"desc":"aggressiveness of optimization","long_desc":".99 is very aggressive, .01 is less aggressive","tags":[],"see_also":[]},"end_time":{"name":"end_time","type":"str","level":"advanced","flags":1,"default_value":"2359","min":"","max":"","enum_allowed":[],"desc":"ending time of day to automatically balance","long_desc":"This is a time of day in the format HHMM.","tags":[],"see_also":[]},"end_weekday":{"name":"end_weekday","type":"uint","level":"advanced","flags":1,"default_value":"0","min":"0","max":"6","enum_allowed":[],"desc":"Restrict automatic balancing to days of the week earlier than this","long_desc":"0 = Sunday, 1 = Monday, etc.","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"min_score":{"name":"min_score","type":"float","level":"advanced","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"minimum score, below which no optimization is attempted","long_desc":"","tags":[],"see_also":[]},"mode":{"name":"mode","type":"str","level":"advanced","flags":1,"default_value":"upmap","min":"","max":"","enum_allowed":["crush-compat","none","read","upmap","upmap-read"],"desc":"Balancer mode","long_desc":"","tags":[],"see_also":[]},"pool_ids":{"name":"pool_ids","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"pools which the automatic balancing will be limited to","long_desc":"","tags":[],"see_also":[]},"sleep_interval":{"name":"sleep_interval","type":"secs","level":"advanced","flags":1,"default_value":"60","min":"","max":"","enum_allowed":[],"desc":"how frequently to wake up and attempt optimization","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"update_pg_upmap_activity":{"name":"update_pg_upmap_activity","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Updates pg_upmap activity stats to be used in `balancer status detail`","long_desc":"","tags":[],"see_also":[]},"upmap_max_deviation":{"name":"upmap_max_deviation","type":"int","level":"advanced","flags":1,"default_value":"5","min":"1","max":"","enum_allowed":[],"desc":"deviation below which no optimization is attempted","long_desc":"If the number of PGs are within this count then no optimization is attempted","tags":[],"see_also":[]},"upmap_max_optimizations":{"name":"upmap_max_optimizations","type":"uint","level":"advanced","flags":1,"default_value":"10","min":"","max":"","enum_allowed":[],"desc":"maximum upmap optimizations to make per attempt","long_desc":"","tags":[],"see_also":[]}}},{"name":"cephadm","can_run":true,"error_string":"","module_options":{"agent_down_multiplier":{"name":"agent_down_multiplier","type":"float","level":"advanced","flags":0,"default_value":"3.0","min":"","max":"","enum_allowed":[],"desc":"Multiplied by agent refresh rate to calculate how long agent must not report before being marked down","long_desc":"","tags":[],"see_also":[]},"agent_refresh_rate":{"name":"agent_refresh_rate","type":"secs","level":"advanced","flags":0,"default_value":"20","min":"","max":"","enum_allowed":[],"desc":"How often agent on each host will try to gather and send metadata","long_desc":"","tags":[],"see_also":[]},"agent_starting_port":{"name":"agent_starting_port","type":"int","level":"advanced","flags":0,"default_value":"4721","min":"","max":"","enum_allowed":[],"desc":"First port agent will try to bind to (will also try up to next 1000 subsequent ports if blocked)","long_desc":"","tags":[],"see_also":[]},"allow_ptrace":{"name":"allow_ptrace","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"allow SYS_PTRACE capability on ceph containers","long_desc":"The SYS_PTRACE capability is needed to attach to a process with gdb or strace. Enabling this options can allow debugging daemons that encounter problems at runtime.","tags":[],"see_also":[]},"autotune_interval":{"name":"autotune_interval","type":"secs","level":"advanced","flags":0,"default_value":"600","min":"","max":"","enum_allowed":[],"desc":"how frequently to autotune daemon memory","long_desc":"","tags":[],"see_also":[]},"autotune_memory_target_ratio":{"name":"autotune_memory_target_ratio","type":"float","level":"advanced","flags":0,"default_value":"0.7","min":"","max":"","enum_allowed":[],"desc":"ratio of total system memory to divide amongst autotuned daemons","long_desc":"","tags":[],"see_also":[]},"cephadm_log_destination":{"name":"cephadm_log_destination","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":["file","file,syslog","syslog"],"desc":"Destination for cephadm command's persistent logging","long_desc":"","tags":[],"see_also":[]},"cgroups_split":{"name":"cgroups_split","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"Pass --cgroups=split when cephadm creates containers (currently podman only)","long_desc":"","tags":[],"see_also":[]},"config_checks_enabled":{"name":"config_checks_enabled","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Enable or disable the cephadm configuration analysis","long_desc":"","tags":[],"see_also":[]},"config_dashboard":{"name":"config_dashboard","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"manage configs like API endpoints in Dashboard.","long_desc":"","tags":[],"see_also":[]},"container_image_alertmanager":{"name":"container_image_alertmanager","type":"str","level":"advanced","flags":0,"default_value":"quay.io/prometheus/alertmanager:v0.25.0","min":"","max":"","enum_allowed":[],"desc":"Prometheus container image","long_desc":"","tags":[],"see_also":[]},"container_image_base":{"name":"container_image_base","type":"str","level":"advanced","flags":1,"default_value":"quay.io/ceph/ceph","min":"","max":"","enum_allowed":[],"desc":"Container image name, without the tag","long_desc":"","tags":[],"see_also":[]},"container_image_elasticsearch":{"name":"container_image_elasticsearch","type":"str","level":"advanced","flags":0,"default_value":"quay.io/omrizeneva/elasticsearch:6.8.23","min":"","max":"","enum_allowed":[],"desc":"elasticsearch container image","long_desc":"","tags":[],"see_also":[]},"container_image_grafana":{"name":"container_image_grafana","type":"str","level":"advanced","flags":0,"default_value":"quay.io/ceph/grafana:10.4.0","min":"","max":"","enum_allowed":[],"desc":"Prometheus container image","long_desc":"","tags":[],"see_also":[]},"container_image_haproxy":{"name":"container_image_haproxy","type":"str","level":"advanced","flags":0,"default_value":"quay.io/ceph/haproxy:2.3","min":"","max":"","enum_allowed":[],"desc":"HAproxy container image","long_desc":"","tags":[],"see_also":[]},"container_image_jaeger_agent":{"name":"container_image_jaeger_agent","type":"str","level":"advanced","flags":0,"default_value":"quay.io/jaegertracing/jaeger-agent:1.29","min":"","max":"","enum_allowed":[],"desc":"Jaeger agent container image","long_desc":"","tags":[],"see_also":[]},"container_image_jaeger_collector":{"name":"container_image_jaeger_collector","type":"str","level":"advanced","flags":0,"default_value":"quay.io/jaegertracing/jaeger-collector:1.29","min":"","max":"","enum_allowed":[],"desc":"Jaeger collector container image","long_desc":"","tags":[],"see_also":[]},"container_image_jaeger_query":{"name":"container_image_jaeger_query","type":"str","level":"advanced","flags":0,"default_value":"quay.io/jaegertracing/jaeger-query:1.29","min":"","max":"","enum_allowed":[],"desc":"Jaeger query container image","long_desc":"","tags":[],"see_also":[]},"container_image_keepalived":{"name":"container_image_keepalived","type":"str","level":"advanced","flags":0,"default_value":"quay.io/ceph/keepalived:2.2.4","min":"","max":"","enum_allowed":[],"desc":"Keepalived container image","long_desc":"","tags":[],"see_also":[]},"container_image_loki":{"name":"container_image_loki","type":"str","level":"advanced","flags":0,"default_value":"quay.io/ceph/loki:3.0.0","min":"","max":"","enum_allowed":[],"desc":"Loki container image","long_desc":"","tags":[],"see_also":[]},"container_image_node_exporter":{"name":"container_image_node_exporter","type":"str","level":"advanced","flags":0,"default_value":"quay.io/prometheus/node-exporter:v1.7.0","min":"","max":"","enum_allowed":[],"desc":"Prometheus container image","long_desc":"","tags":[],"see_also":[]},"container_image_nvmeof":{"name":"container_image_nvmeof","type":"str","level":"advanced","flags":0,"default_value":"quay.io/ceph/nvmeof:1.2.5","min":"","max":"","enum_allowed":[],"desc":"Nvme-of container image","long_desc":"","tags":[],"see_also":[]},"container_image_prometheus":{"name":"container_image_prometheus","type":"str","level":"advanced","flags":0,"default_value":"quay.io/prometheus/prometheus:v2.51.0","min":"","max":"","enum_allowed":[],"desc":"Prometheus container image","long_desc":"","tags":[],"see_also":[]},"container_image_promtail":{"name":"container_image_promtail","type":"str","level":"advanced","flags":0,"default_value":"quay.io/ceph/promtail:3.0.0","min":"","max":"","enum_allowed":[],"desc":"Promtail container image","long_desc":"","tags":[],"see_also":[]},"container_image_samba":{"name":"container_image_samba","type":"str","level":"advanced","flags":0,"default_value":"quay.io/samba.org/samba-server:devbuilds-centos-amd64","min":"","max":"","enum_allowed":[],"desc":"Samba/SMB container image","long_desc":"","tags":[],"see_also":[]},"container_image_snmp_gateway":{"name":"container_image_snmp_gateway","type":"str","level":"advanced","flags":0,"default_value":"quay.io/ceph/snmp-notifier:v1.2.1","min":"","max":"","enum_allowed":[],"desc":"SNMP Gateway container image","long_desc":"","tags":[],"see_also":[]},"container_init":{"name":"container_init","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"Run podman/docker with `--init`","long_desc":"","tags":[],"see_also":[]},"daemon_cache_timeout":{"name":"daemon_cache_timeout","type":"secs","level":"advanced","flags":0,"default_value":"600","min":"","max":"","enum_allowed":[],"desc":"seconds to cache service (daemon) inventory","long_desc":"","tags":[],"see_also":[]},"default_cephadm_command_timeout":{"name":"default_cephadm_command_timeout","type":"int","level":"advanced","flags":0,"default_value":"900","min":"","max":"","enum_allowed":[],"desc":"Default timeout applied to cephadm commands run directly on the host (in seconds)","long_desc":"","tags":[],"see_also":[]},"default_registry":{"name":"default_registry","type":"str","level":"advanced","flags":0,"default_value":"quay.io","min":"","max":"","enum_allowed":[],"desc":"Search-registry to which we should normalize unqualified image names. This is not the default registry","long_desc":"","tags":[],"see_also":[]},"device_cache_timeout":{"name":"device_cache_timeout","type":"secs","level":"advanced","flags":0,"default_value":"1800","min":"","max":"","enum_allowed":[],"desc":"seconds to cache device inventory","long_desc":"","tags":[],"see_also":[]},"device_enhanced_scan":{"name":"device_enhanced_scan","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Use libstoragemgmt during device scans","long_desc":"","tags":[],"see_also":[]},"facts_cache_timeout":{"name":"facts_cache_timeout","type":"secs","level":"advanced","flags":0,"default_value":"60","min":"","max":"","enum_allowed":[],"desc":"seconds to cache host facts data","long_desc":"","tags":[],"see_also":[]},"grafana_dashboards_path":{"name":"grafana_dashboards_path","type":"str","level":"advanced","flags":0,"default_value":"/etc/grafana/dashboards/ceph-dashboard/","min":"","max":"","enum_allowed":[],"desc":"location of dashboards to include in grafana deployments","long_desc":"","tags":[],"see_also":[]},"host_check_interval":{"name":"host_check_interval","type":"secs","level":"advanced","flags":0,"default_value":"600","min":"","max":"","enum_allowed":[],"desc":"how frequently to perform a host check","long_desc":"","tags":[],"see_also":[]},"hw_monitoring":{"name":"hw_monitoring","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Deploy hw monitoring daemon on every host.","long_desc":"","tags":[],"see_also":[]},"inventory_list_all":{"name":"inventory_list_all","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Whether ceph-volume inventory should report more devices (mostly mappers (LVs / mpaths), partitions...)","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_refresh_metadata":{"name":"log_refresh_metadata","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Log all refresh metadata. Includes daemon, device, and host info collected regularly. Only has effect if logging at debug level","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"log to the \"cephadm\" cluster log channel\"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"manage_etc_ceph_ceph_conf":{"name":"manage_etc_ceph_ceph_conf","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Manage and own /etc/ceph/ceph.conf on the hosts.","long_desc":"","tags":[],"see_also":[]},"manage_etc_ceph_ceph_conf_hosts":{"name":"manage_etc_ceph_ceph_conf_hosts","type":"str","level":"advanced","flags":0,"default_value":"*","min":"","max":"","enum_allowed":[],"desc":"PlacementSpec describing on which hosts to manage /etc/ceph/ceph.conf","long_desc":"","tags":[],"see_also":[]},"max_count_per_host":{"name":"max_count_per_host","type":"int","level":"advanced","flags":0,"default_value":"10","min":"","max":"","enum_allowed":[],"desc":"max number of daemons per service per host","long_desc":"","tags":[],"see_also":[]},"max_osd_draining_count":{"name":"max_osd_draining_count","type":"int","level":"advanced","flags":0,"default_value":"10","min":"","max":"","enum_allowed":[],"desc":"max number of osds that will be drained simultaneously when osds are removed","long_desc":"","tags":[],"see_also":[]},"migration_current":{"name":"migration_current","type":"int","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"internal - do not modify","long_desc":"","tags":[],"see_also":[]},"mode":{"name":"mode","type":"str","level":"advanced","flags":0,"default_value":"root","min":"","max":"","enum_allowed":["cephadm-package","root"],"desc":"mode for remote execution of cephadm","long_desc":"","tags":[],"see_also":[]},"oob_default_addr":{"name":"oob_default_addr","type":"str","level":"advanced","flags":0,"default_value":"169.254.1.1","min":"","max":"","enum_allowed":[],"desc":"Default address for RedFish API (oob management).","long_desc":"","tags":[],"see_also":[]},"prometheus_alerts_path":{"name":"prometheus_alerts_path","type":"str","level":"advanced","flags":0,"default_value":"/etc/prometheus/ceph/ceph_default_alerts.yml","min":"","max":"","enum_allowed":[],"desc":"location of alerts to include in prometheus deployments","long_desc":"","tags":[],"see_also":[]},"registry_insecure":{"name":"registry_insecure","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Registry is to be considered insecure (no TLS available). Only for development purposes.","long_desc":"","tags":[],"see_also":[]},"registry_password":{"name":"registry_password","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"Custom repository password. Only used for logging into a registry.","long_desc":"","tags":[],"see_also":[]},"registry_url":{"name":"registry_url","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"Registry url for login purposes. This is not the default registry","long_desc":"","tags":[],"see_also":[]},"registry_username":{"name":"registry_username","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"Custom repository username. Only used for logging into a registry.","long_desc":"","tags":[],"see_also":[]},"secure_monitoring_stack":{"name":"secure_monitoring_stack","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Enable TLS security for all the monitoring stack daemons","long_desc":"","tags":[],"see_also":[]},"service_discovery_port":{"name":"service_discovery_port","type":"int","level":"advanced","flags":0,"default_value":"8765","min":"","max":"","enum_allowed":[],"desc":"cephadm service discovery port","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"ssh_config_file":{"name":"ssh_config_file","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"customized SSH config file to connect to managed hosts","long_desc":"","tags":[],"see_also":[]},"ssh_keepalive_count_max":{"name":"ssh_keepalive_count_max","type":"int","level":"advanced","flags":0,"default_value":"3","min":"","max":"","enum_allowed":[],"desc":"How many times ssh connections can fail liveness checks before the host is marked offline","long_desc":"","tags":[],"see_also":[]},"ssh_keepalive_interval":{"name":"ssh_keepalive_interval","type":"int","level":"advanced","flags":0,"default_value":"7","min":"","max":"","enum_allowed":[],"desc":"How often ssh connections are checked for liveness","long_desc":"","tags":[],"see_also":[]},"use_agent":{"name":"use_agent","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Use cephadm agent on each host to gather and send metadata","long_desc":"","tags":[],"see_also":[]},"use_repo_digest":{"name":"use_repo_digest","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"Automatically convert image tags to image digest. Make sure all daemons use the same image","long_desc":"","tags":[],"see_also":[]},"warn_on_failed_host_check":{"name":"warn_on_failed_host_check","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"raise a health warning if the host check fails","long_desc":"","tags":[],"see_also":[]},"warn_on_stray_daemons":{"name":"warn_on_stray_daemons","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"raise a health warning if daemons are detected that are not managed by cephadm","long_desc":"","tags":[],"see_also":[]},"warn_on_stray_hosts":{"name":"warn_on_stray_hosts","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"raise a health warning if daemons are detected on a host that is not managed by cephadm","long_desc":"","tags":[],"see_also":[]}}},{"name":"crash","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"retain_interval":{"name":"retain_interval","type":"secs","level":"advanced","flags":1,"default_value":"31536000","min":"","max":"","enum_allowed":[],"desc":"how long to retain crashes before pruning them","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"warn_recent_interval":{"name":"warn_recent_interval","type":"secs","level":"advanced","flags":1,"default_value":"1209600","min":"","max":"","enum_allowed":[],"desc":"time interval in which to warn about recent crashes","long_desc":"","tags":[],"see_also":[]}}},{"name":"dashboard","can_run":true,"error_string":"","module_options":{"ACCOUNT_LOCKOUT_ATTEMPTS":{"name":"ACCOUNT_LOCKOUT_ATTEMPTS","type":"int","level":"advanced","flags":0,"default_value":"10","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"ALERTMANAGER_API_HOST":{"name":"ALERTMANAGER_API_HOST","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"ALERTMANAGER_API_SSL_VERIFY":{"name":"ALERTMANAGER_API_SSL_VERIFY","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"AUDIT_API_ENABLED":{"name":"AUDIT_API_ENABLED","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"AUDIT_API_LOG_PAYLOAD":{"name":"AUDIT_API_LOG_PAYLOAD","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"ENABLE_BROWSABLE_API":{"name":"ENABLE_BROWSABLE_API","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"FEATURE_TOGGLE_CEPHFS":{"name":"FEATURE_TOGGLE_CEPHFS","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"FEATURE_TOGGLE_DASHBOARD":{"name":"FEATURE_TOGGLE_DASHBOARD","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"FEATURE_TOGGLE_ISCSI":{"name":"FEATURE_TOGGLE_ISCSI","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"FEATURE_TOGGLE_MIRRORING":{"name":"FEATURE_TOGGLE_MIRRORING","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"FEATURE_TOGGLE_NFS":{"name":"FEATURE_TOGGLE_NFS","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"FEATURE_TOGGLE_RBD":{"name":"FEATURE_TOGGLE_RBD","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"FEATURE_TOGGLE_RGW":{"name":"FEATURE_TOGGLE_RGW","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"GANESHA_CLUSTERS_RADOS_POOL_NAMESPACE":{"name":"GANESHA_CLUSTERS_RADOS_POOL_NAMESPACE","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"GRAFANA_API_PASSWORD":{"name":"GRAFANA_API_PASSWORD","type":"str","level":"advanced","flags":0,"default_value":"admin","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"GRAFANA_API_SSL_VERIFY":{"name":"GRAFANA_API_SSL_VERIFY","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"GRAFANA_API_URL":{"name":"GRAFANA_API_URL","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"GRAFANA_API_USERNAME":{"name":"GRAFANA_API_USERNAME","type":"str","level":"advanced","flags":0,"default_value":"admin","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"GRAFANA_FRONTEND_API_URL":{"name":"GRAFANA_FRONTEND_API_URL","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"GRAFANA_UPDATE_DASHBOARDS":{"name":"GRAFANA_UPDATE_DASHBOARDS","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"ISCSI_API_SSL_VERIFICATION":{"name":"ISCSI_API_SSL_VERIFICATION","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"ISSUE_TRACKER_API_KEY":{"name":"ISSUE_TRACKER_API_KEY","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PROMETHEUS_API_HOST":{"name":"PROMETHEUS_API_HOST","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PROMETHEUS_API_SSL_VERIFY":{"name":"PROMETHEUS_API_SSL_VERIFY","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PWD_POLICY_CHECK_COMPLEXITY_ENABLED":{"name":"PWD_POLICY_CHECK_COMPLEXITY_ENABLED","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PWD_POLICY_CHECK_EXCLUSION_LIST_ENABLED":{"name":"PWD_POLICY_CHECK_EXCLUSION_LIST_ENABLED","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PWD_POLICY_CHECK_LENGTH_ENABLED":{"name":"PWD_POLICY_CHECK_LENGTH_ENABLED","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PWD_POLICY_CHECK_OLDPWD_ENABLED":{"name":"PWD_POLICY_CHECK_OLDPWD_ENABLED","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PWD_POLICY_CHECK_REPETITIVE_CHARS_ENABLED":{"name":"PWD_POLICY_CHECK_REPETITIVE_CHARS_ENABLED","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PWD_POLICY_CHECK_SEQUENTIAL_CHARS_ENABLED":{"name":"PWD_POLICY_CHECK_SEQUENTIAL_CHARS_ENABLED","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PWD_POLICY_CHECK_USERNAME_ENABLED":{"name":"PWD_POLICY_CHECK_USERNAME_ENABLED","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PWD_POLICY_ENABLED":{"name":"PWD_POLICY_ENABLED","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PWD_POLICY_EXCLUSION_LIST":{"name":"PWD_POLICY_EXCLUSION_LIST","type":"str","level":"advanced","flags":0,"default_value":"osd,host,dashboard,pool,block,nfs,ceph,monitors,gateway,logs,crush,maps","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PWD_POLICY_MIN_COMPLEXITY":{"name":"PWD_POLICY_MIN_COMPLEXITY","type":"int","level":"advanced","flags":0,"default_value":"10","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PWD_POLICY_MIN_LENGTH":{"name":"PWD_POLICY_MIN_LENGTH","type":"int","level":"advanced","flags":0,"default_value":"8","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"REST_REQUESTS_TIMEOUT":{"name":"REST_REQUESTS_TIMEOUT","type":"int","level":"advanced","flags":0,"default_value":"45","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"RGW_API_ACCESS_KEY":{"name":"RGW_API_ACCESS_KEY","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"RGW_API_ADMIN_RESOURCE":{"name":"RGW_API_ADMIN_RESOURCE","type":"str","level":"advanced","flags":0,"default_value":"admin","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"RGW_API_SECRET_KEY":{"name":"RGW_API_SECRET_KEY","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"RGW_API_SSL_VERIFY":{"name":"RGW_API_SSL_VERIFY","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"USER_PWD_EXPIRATION_SPAN":{"name":"USER_PWD_EXPIRATION_SPAN","type":"int","level":"advanced","flags":0,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"USER_PWD_EXPIRATION_WARNING_1":{"name":"USER_PWD_EXPIRATION_WARNING_1","type":"int","level":"advanced","flags":0,"default_value":"10","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"USER_PWD_EXPIRATION_WARNING_2":{"name":"USER_PWD_EXPIRATION_WARNING_2","type":"int","level":"advanced","flags":0,"default_value":"5","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"cross_origin_url":{"name":"cross_origin_url","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"crt_file":{"name":"crt_file","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"debug":{"name":"debug","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Enable/disable debug options","long_desc":"","tags":[],"see_also":[]},"jwt_token_ttl":{"name":"jwt_token_ttl","type":"int","level":"advanced","flags":0,"default_value":"28800","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"key_file":{"name":"key_file","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"motd":{"name":"motd","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"The message of the day","long_desc":"","tags":[],"see_also":[]},"redirect_resolve_ip_addr":{"name":"redirect_resolve_ip_addr","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"server_addr":{"name":"server_addr","type":"str","level":"advanced","flags":0,"default_value":"::","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"server_port":{"name":"server_port","type":"int","level":"advanced","flags":0,"default_value":"8080","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"ssl":{"name":"ssl","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"ssl_server_port":{"name":"ssl_server_port","type":"int","level":"advanced","flags":0,"default_value":"8443","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"standby_behaviour":{"name":"standby_behaviour","type":"str","level":"advanced","flags":0,"default_value":"redirect","min":"","max":"","enum_allowed":["error","redirect"],"desc":"","long_desc":"","tags":[],"see_also":[]},"standby_error_status_code":{"name":"standby_error_status_code","type":"int","level":"advanced","flags":0,"default_value":"500","min":"400","max":"599","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"url_prefix":{"name":"url_prefix","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"devicehealth","can_run":true,"error_string":"","module_options":{"enable_monitoring":{"name":"enable_monitoring","type":"bool","level":"advanced","flags":1,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"monitor device health metrics","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"mark_out_threshold":{"name":"mark_out_threshold","type":"secs","level":"advanced","flags":1,"default_value":"2419200","min":"","max":"","enum_allowed":[],"desc":"automatically mark OSD if it may fail before this long","long_desc":"","tags":[],"see_also":[]},"pool_name":{"name":"pool_name","type":"str","level":"advanced","flags":1,"default_value":"device_health_metrics","min":"","max":"","enum_allowed":[],"desc":"name of pool in which to store device health metrics","long_desc":"","tags":[],"see_also":[]},"retention_period":{"name":"retention_period","type":"secs","level":"advanced","flags":1,"default_value":"15552000","min":"","max":"","enum_allowed":[],"desc":"how long to retain device health metrics","long_desc":"","tags":[],"see_also":[]},"scrape_frequency":{"name":"scrape_frequency","type":"secs","level":"advanced","flags":1,"default_value":"86400","min":"","max":"","enum_allowed":[],"desc":"how frequently to scrape device health metrics","long_desc":"","tags":[],"see_also":[]},"self_heal":{"name":"self_heal","type":"bool","level":"advanced","flags":1,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"preemptively heal cluster around devices that may fail","long_desc":"","tags":[],"see_also":[]},"sleep_interval":{"name":"sleep_interval","type":"secs","level":"advanced","flags":1,"default_value":"600","min":"","max":"","enum_allowed":[],"desc":"how frequently to wake up and check device health","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"warn_threshold":{"name":"warn_threshold","type":"secs","level":"advanced","flags":1,"default_value":"7257600","min":"","max":"","enum_allowed":[],"desc":"raise health warning if OSD may fail before this long","long_desc":"","tags":[],"see_also":[]}}},{"name":"diskprediction_local","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"predict_interval":{"name":"predict_interval","type":"str","level":"advanced","flags":0,"default_value":"86400","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"predictor_model":{"name":"predictor_model","type":"str","level":"advanced","flags":0,"default_value":"prophetstor","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sleep_interval":{"name":"sleep_interval","type":"str","level":"advanced","flags":0,"default_value":"600","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"influx","can_run":false,"error_string":"influxdb python module not found","module_options":{"batch_size":{"name":"batch_size","type":"int","level":"advanced","flags":0,"default_value":"5000","min":"","max":"","enum_allowed":[],"desc":"How big batches of data points should be when sending to InfluxDB.","long_desc":"","tags":[],"see_also":[]},"database":{"name":"database","type":"str","level":"advanced","flags":0,"default_value":"ceph","min":"","max":"","enum_allowed":[],"desc":"InfluxDB database name. You will need to create this database and grant write privileges to the configured username or the username must have admin privileges to create it.","long_desc":"","tags":[],"see_also":[]},"hostname":{"name":"hostname","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"InfluxDB server hostname","long_desc":"","tags":[],"see_also":[]},"interval":{"name":"interval","type":"secs","level":"advanced","flags":0,"default_value":"30","min":"5","max":"","enum_allowed":[],"desc":"Time between reports to InfluxDB. Default 30 seconds.","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"password":{"name":"password","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"password of InfluxDB server user","long_desc":"","tags":[],"see_also":[]},"port":{"name":"port","type":"int","level":"advanced","flags":0,"default_value":"8086","min":"","max":"","enum_allowed":[],"desc":"InfluxDB server port","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"ssl":{"name":"ssl","type":"str","level":"advanced","flags":0,"default_value":"false","min":"","max":"","enum_allowed":[],"desc":"Use https connection for InfluxDB server. Use \"true\" or \"false\".","long_desc":"","tags":[],"see_also":[]},"threads":{"name":"threads","type":"int","level":"advanced","flags":0,"default_value":"5","min":"1","max":"32","enum_allowed":[],"desc":"How many worker threads should be spawned for sending data to InfluxDB.","long_desc":"","tags":[],"see_also":[]},"username":{"name":"username","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"username of InfluxDB server user","long_desc":"","tags":[],"see_also":[]},"verify_ssl":{"name":"verify_ssl","type":"str","level":"advanced","flags":0,"default_value":"true","min":"","max":"","enum_allowed":[],"desc":"Verify https cert for InfluxDB server. Use \"true\" or \"false\".","long_desc":"","tags":[],"see_also":[]}}},{"name":"insights","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"iostat","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"k8sevents","can_run":true,"error_string":"","module_options":{"ceph_event_retention_days":{"name":"ceph_event_retention_days","type":"int","level":"advanced","flags":0,"default_value":"7","min":"","max":"","enum_allowed":[],"desc":"Days to hold ceph event information within local cache","long_desc":"","tags":[],"see_also":[]},"config_check_secs":{"name":"config_check_secs","type":"int","level":"advanced","flags":0,"default_value":"10","min":"10","max":"","enum_allowed":[],"desc":"interval (secs) to check for cluster configuration changes","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"localpool","can_run":true,"error_string":"","module_options":{"failure_domain":{"name":"failure_domain","type":"str","level":"advanced","flags":1,"default_value":"host","min":"","max":"","enum_allowed":[],"desc":"failure domain for any created local pool","long_desc":"what failure domain we should separate data replicas across.","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"min_size":{"name":"min_size","type":"int","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"default min_size for any created local pool","long_desc":"value to set min_size to (unchanged from Ceph's default if this option is not set)","tags":[],"see_also":[]},"num_rep":{"name":"num_rep","type":"int","level":"advanced","flags":1,"default_value":"3","min":"","max":"","enum_allowed":[],"desc":"default replica count for any created local pool","long_desc":"","tags":[],"see_also":[]},"pg_num":{"name":"pg_num","type":"int","level":"advanced","flags":1,"default_value":"128","min":"","max":"","enum_allowed":[],"desc":"default pg_num for any created local pool","long_desc":"","tags":[],"see_also":[]},"prefix":{"name":"prefix","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"name prefix for any created local pool","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"subtree":{"name":"subtree","type":"str","level":"advanced","flags":1,"default_value":"rack","min":"","max":"","enum_allowed":[],"desc":"CRUSH level for which to create a local pool","long_desc":"which CRUSH subtree type the module should create a pool for.","tags":[],"see_also":[]}}},{"name":"mds_autoscaler","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"mirroring","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"nfs","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"orchestrator","can_run":true,"error_string":"","module_options":{"fail_fs":{"name":"fail_fs","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Fail filesystem for rapid multi-rank mds upgrade","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"orchestrator":{"name":"orchestrator","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["cephadm","rook","test_orchestrator"],"desc":"Orchestrator backend","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"osd_perf_query","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"osd_support","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"pg_autoscaler","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sleep_interval":{"name":"sleep_interval","type":"secs","level":"advanced","flags":0,"default_value":"60","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"threshold":{"name":"threshold","type":"float","level":"advanced","flags":0,"default_value":"3.0","min":"1.0","max":"","enum_allowed":[],"desc":"scaling threshold","long_desc":"The factor by which the `NEW PG_NUM` must vary from the current`PG_NUM` before being accepted. Cannot be less than 1.0","tags":[],"see_also":[]}}},{"name":"progress","can_run":true,"error_string":"","module_options":{"allow_pg_recovery_event":{"name":"allow_pg_recovery_event","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"allow the module to show pg recovery progress","long_desc":"","tags":[],"see_also":[]},"enabled":{"name":"enabled","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"max_completed_events":{"name":"max_completed_events","type":"int","level":"advanced","flags":1,"default_value":"50","min":"","max":"","enum_allowed":[],"desc":"number of past completed events to remember","long_desc":"","tags":[],"see_also":[]},"sleep_interval":{"name":"sleep_interval","type":"secs","level":"advanced","flags":1,"default_value":"5","min":"","max":"","enum_allowed":[],"desc":"how long the module is going to sleep","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"prometheus","can_run":true,"error_string":"","module_options":{"cache":{"name":"cache","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"exclude_perf_counters":{"name":"exclude_perf_counters","type":"bool","level":"advanced","flags":1,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"Do not include perf-counters in the metrics output","long_desc":"Gathering perf-counters from a single Prometheus exporter can degrade ceph-mgr performance, especially in large clusters. Instead, Ceph-exporter daemons are now used by default for perf-counter gathering. This should only be disabled when no ceph-exporters are deployed.","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"rbd_stats_pools":{"name":"rbd_stats_pools","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"rbd_stats_pools_refresh_interval":{"name":"rbd_stats_pools_refresh_interval","type":"int","level":"advanced","flags":0,"default_value":"300","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"scrape_interval":{"name":"scrape_interval","type":"float","level":"advanced","flags":0,"default_value":"15.0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"server_addr":{"name":"server_addr","type":"str","level":"advanced","flags":0,"default_value":"::","min":"","max":"","enum_allowed":[],"desc":"the IPv4 or IPv6 address on which the module listens for HTTP requests","long_desc":"","tags":[],"see_also":[]},"server_port":{"name":"server_port","type":"int","level":"advanced","flags":1,"default_value":"9283","min":"","max":"","enum_allowed":[],"desc":"the port on which the module listens for HTTP requests","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"stale_cache_strategy":{"name":"stale_cache_strategy","type":"str","level":"advanced","flags":0,"default_value":"log","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"standby_behaviour":{"name":"standby_behaviour","type":"str","level":"advanced","flags":1,"default_value":"default","min":"","max":"","enum_allowed":["default","error"],"desc":"","long_desc":"","tags":[],"see_also":[]},"standby_error_status_code":{"name":"standby_error_status_code","type":"int","level":"advanced","flags":1,"default_value":"500","min":"400","max":"599","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"rbd_support","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"max_concurrent_snap_create":{"name":"max_concurrent_snap_create","type":"int","level":"advanced","flags":0,"default_value":"10","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"mirror_snapshot_schedule":{"name":"mirror_snapshot_schedule","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"trash_purge_schedule":{"name":"trash_purge_schedule","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"restful","can_run":true,"error_string":"","module_options":{"enable_auth":{"name":"enable_auth","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"key_file":{"name":"key_file","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"max_requests":{"name":"max_requests","type":"int","level":"advanced","flags":0,"default_value":"500","min":"","max":"","enum_allowed":[],"desc":"Maximum number of requests to keep in memory. When new request comes in, the oldest request will be removed if the number of requests exceeds the max request number. if un-finished request is removed, error message will be logged in the ceph-mgr log.","long_desc":"","tags":[],"see_also":[]},"server_addr":{"name":"server_addr","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"server_port":{"name":"server_port","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"rgw","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"secondary_zone_period_retry_limit":{"name":"secondary_zone_period_retry_limit","type":"int","level":"advanced","flags":0,"default_value":"5","min":"","max":"","enum_allowed":[],"desc":"RGW module period update retry limit for secondary site","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"rook","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"storage_class":{"name":"storage_class","type":"str","level":"advanced","flags":0,"default_value":"local","min":"","max":"","enum_allowed":[],"desc":"storage class name for LSO-discovered PVs","long_desc":"","tags":[],"see_also":[]}}},{"name":"selftest","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"roption1":{"name":"roption1","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"roption2":{"name":"roption2","type":"str","level":"advanced","flags":0,"default_value":"xyz","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"rwoption1":{"name":"rwoption1","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"rwoption2":{"name":"rwoption2","type":"int","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"rwoption3":{"name":"rwoption3","type":"float","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"rwoption4":{"name":"rwoption4","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"rwoption5":{"name":"rwoption5","type":"bool","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"rwoption6":{"name":"rwoption6","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"rwoption7":{"name":"rwoption7","type":"int","level":"advanced","flags":0,"default_value":"","min":"1","max":"42","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"testkey":{"name":"testkey","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"testlkey":{"name":"testlkey","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"testnewline":{"name":"testnewline","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"snap_schedule","can_run":true,"error_string":"","module_options":{"allow_m_granularity":{"name":"allow_m_granularity","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"allow minute scheduled snapshots","long_desc":"","tags":[],"see_also":[]},"dump_on_update":{"name":"dump_on_update","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"dump database to debug log on update","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"stats","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"status","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"telegraf","can_run":true,"error_string":"","module_options":{"address":{"name":"address","type":"str","level":"advanced","flags":0,"default_value":"unixgram:///tmp/telegraf.sock","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"interval":{"name":"interval","type":"secs","level":"advanced","flags":0,"default_value":"15","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"telemetry","can_run":true,"error_string":"","module_options":{"channel_basic":{"name":"channel_basic","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"Share basic cluster information (size, version)","long_desc":"","tags":[],"see_also":[]},"channel_crash":{"name":"channel_crash","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"Share metadata about Ceph daemon crashes (version, stack straces, etc)","long_desc":"","tags":[],"see_also":[]},"channel_device":{"name":"channel_device","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"Share device health metrics (e.g., SMART data, minus potentially identifying info like serial numbers)","long_desc":"","tags":[],"see_also":[]},"channel_ident":{"name":"channel_ident","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Share a user-provided description and/or contact email for the cluster","long_desc":"","tags":[],"see_also":[]},"channel_perf":{"name":"channel_perf","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Share various performance metrics of a cluster","long_desc":"","tags":[],"see_also":[]},"contact":{"name":"contact","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"description":{"name":"description","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"device_url":{"name":"device_url","type":"str","level":"advanced","flags":0,"default_value":"https://telemetry.ceph.com/device","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"enabled":{"name":"enabled","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"interval":{"name":"interval","type":"int","level":"advanced","flags":0,"default_value":"24","min":"8","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"last_opt_revision":{"name":"last_opt_revision","type":"int","level":"advanced","flags":0,"default_value":"1","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"leaderboard":{"name":"leaderboard","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"leaderboard_description":{"name":"leaderboard_description","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"organization":{"name":"organization","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"proxy":{"name":"proxy","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"url":{"name":"url","type":"str","level":"advanced","flags":0,"default_value":"https://telemetry.ceph.com/report","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"test_orchestrator","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"volumes","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"max_concurrent_clones":{"name":"max_concurrent_clones","type":"int","level":"advanced","flags":0,"default_value":"4","min":"","max":"","enum_allowed":[],"desc":"Number of asynchronous cloner threads","long_desc":"","tags":[],"see_also":[]},"periodic_async_work":{"name":"periodic_async_work","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Periodically check for async work","long_desc":"","tags":[],"see_also":[]},"snapshot_clone_delay":{"name":"snapshot_clone_delay","type":"int","level":"advanced","flags":0,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"Delay clone begin operation by snapshot_clone_delay seconds","long_desc":"","tags":[],"see_also":[]},"snapshot_clone_no_wait":{"name":"snapshot_clone_no_wait","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"Reject subvolume clone request when cloner threads are busy","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"zabbix","can_run":true,"error_string":"","module_options":{"discovery_interval":{"name":"discovery_interval","type":"uint","level":"advanced","flags":0,"default_value":"100","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"identifier":{"name":"identifier","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"interval":{"name":"interval","type":"secs","level":"advanced","flags":0,"default_value":"60","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"zabbix_host":{"name":"zabbix_host","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"zabbix_port":{"name":"zabbix_port","type":"int","level":"advanced","flags":0,"default_value":"10051","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"zabbix_sender":{"name":"zabbix_sender","type":"str","level":"advanced","flags":0,"default_value":"/usr/bin/zabbix_sender","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}}]}],"modules":["cephadm","dashboard","iostat","nfs","restful"],"available_modules":[{"name":"alerts","can_run":true,"error_string":"","module_options":{"interval":{"name":"interval","type":"secs","level":"advanced","flags":1,"default_value":"60","min":"","max":"","enum_allowed":[],"desc":"How frequently to reexamine health status","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"smtp_destination":{"name":"smtp_destination","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"Email address to send alerts to","long_desc":"","tags":[],"see_also":[]},"smtp_from_name":{"name":"smtp_from_name","type":"str","level":"advanced","flags":1,"default_value":"Ceph","min":"","max":"","enum_allowed":[],"desc":"Email From: name","long_desc":"","tags":[],"see_also":[]},"smtp_host":{"name":"smtp_host","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"SMTP server","long_desc":"","tags":[],"see_also":[]},"smtp_password":{"name":"smtp_password","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"Password to authenticate with","long_desc":"","tags":[],"see_also":[]},"smtp_port":{"name":"smtp_port","type":"int","level":"advanced","flags":1,"default_value":"465","min":"","max":"","enum_allowed":[],"desc":"SMTP port","long_desc":"","tags":[],"see_also":[]},"smtp_sender":{"name":"smtp_sender","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"SMTP envelope sender","long_desc":"","tags":[],"see_also":[]},"smtp_ssl":{"name":"smtp_ssl","type":"bool","level":"advanced","flags":1,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"Use SSL to connect to SMTP server","long_desc":"","tags":[],"see_also":[]},"smtp_user":{"name":"smtp_user","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"User to authenticate as","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"balancer","can_run":true,"error_string":"","module_options":{"active":{"name":"active","type":"bool","level":"advanced","flags":1,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"automatically balance PGs across cluster","long_desc":"","tags":[],"see_also":[]},"begin_time":{"name":"begin_time","type":"str","level":"advanced","flags":1,"default_value":"0000","min":"","max":"","enum_allowed":[],"desc":"beginning time of day to automatically balance","long_desc":"This is a time of day in the format HHMM.","tags":[],"see_also":[]},"begin_weekday":{"name":"begin_weekday","type":"uint","level":"advanced","flags":1,"default_value":"0","min":"0","max":"6","enum_allowed":[],"desc":"Restrict automatic balancing to this day of the week or later","long_desc":"0 = Sunday, 1 = Monday, etc.","tags":[],"see_also":[]},"crush_compat_max_iterations":{"name":"crush_compat_max_iterations","type":"uint","level":"advanced","flags":1,"default_value":"25","min":"1","max":"250","enum_allowed":[],"desc":"maximum number of iterations to attempt optimization","long_desc":"","tags":[],"see_also":[]},"crush_compat_metrics":{"name":"crush_compat_metrics","type":"str","level":"advanced","flags":1,"default_value":"pgs,objects,bytes","min":"","max":"","enum_allowed":[],"desc":"metrics with which to calculate OSD utilization","long_desc":"Value is a list of one or more of \"pgs\", \"objects\", or \"bytes\", and indicates which metrics to use to balance utilization.","tags":[],"see_also":[]},"crush_compat_step":{"name":"crush_compat_step","type":"float","level":"advanced","flags":1,"default_value":"0.5","min":"0.001","max":"0.999","enum_allowed":[],"desc":"aggressiveness of optimization","long_desc":".99 is very aggressive, .01 is less aggressive","tags":[],"see_also":[]},"end_time":{"name":"end_time","type":"str","level":"advanced","flags":1,"default_value":"2359","min":"","max":"","enum_allowed":[],"desc":"ending time of day to automatically balance","long_desc":"This is a time of day in the format HHMM.","tags":[],"see_also":[]},"end_weekday":{"name":"end_weekday","type":"uint","level":"advanced","flags":1,"default_value":"0","min":"0","max":"6","enum_allowed":[],"desc":"Restrict automatic balancing to days of the week earlier than this","long_desc":"0 = Sunday, 1 = Monday, etc.","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"min_score":{"name":"min_score","type":"float","level":"advanced","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"minimum score, below which no optimization is attempted","long_desc":"","tags":[],"see_also":[]},"mode":{"name":"mode","type":"str","level":"advanced","flags":1,"default_value":"upmap","min":"","max":"","enum_allowed":["crush-compat","none","read","upmap","upmap-read"],"desc":"Balancer mode","long_desc":"","tags":[],"see_also":[]},"pool_ids":{"name":"pool_ids","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"pools which the automatic balancing will be limited to","long_desc":"","tags":[],"see_also":[]},"sleep_interval":{"name":"sleep_interval","type":"secs","level":"advanced","flags":1,"default_value":"60","min":"","max":"","enum_allowed":[],"desc":"how frequently to wake up and attempt optimization","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"update_pg_upmap_activity":{"name":"update_pg_upmap_activity","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Updates pg_upmap activity stats to be used in `balancer status detail`","long_desc":"","tags":[],"see_also":[]},"upmap_max_deviation":{"name":"upmap_max_deviation","type":"int","level":"advanced","flags":1,"default_value":"5","min":"1","max":"","enum_allowed":[],"desc":"deviation below which no optimization is attempted","long_desc":"If the number of PGs are within this count then no optimization is attempted","tags":[],"see_also":[]},"upmap_max_optimizations":{"name":"upmap_max_optimizations","type":"uint","level":"advanced","flags":1,"default_value":"10","min":"","max":"","enum_allowed":[],"desc":"maximum upmap optimizations to make per attempt","long_desc":"","tags":[],"see_also":[]}}},{"name":"cephadm","can_run":true,"error_string":"","module_options":{"agent_down_multiplier":{"name":"agent_down_multiplier","type":"float","level":"advanced","flags":0,"default_value":"3.0","min":"","max":"","enum_allowed":[],"desc":"Multiplied by agent refresh rate to calculate how long agent must not report before being marked down","long_desc":"","tags":[],"see_also":[]},"agent_refresh_rate":{"name":"agent_refresh_rate","type":"secs","level":"advanced","flags":0,"default_value":"20","min":"","max":"","enum_allowed":[],"desc":"How often agent on each host will try to gather and send metadata","long_desc":"","tags":[],"see_also":[]},"agent_starting_port":{"name":"agent_starting_port","type":"int","level":"advanced","flags":0,"default_value":"4721","min":"","max":"","enum_allowed":[],"desc":"First port agent will try to bind to (will also try up to next 1000 subsequent ports if blocked)","long_desc":"","tags":[],"see_also":[]},"allow_ptrace":{"name":"allow_ptrace","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"allow SYS_PTRACE capability on ceph containers","long_desc":"The SYS_PTRACE capability is needed to attach to a process with gdb or strace. Enabling this options can allow debugging daemons that encounter problems at runtime.","tags":[],"see_also":[]},"autotune_interval":{"name":"autotune_interval","type":"secs","level":"advanced","flags":0,"default_value":"600","min":"","max":"","enum_allowed":[],"desc":"how frequently to autotune daemon memory","long_desc":"","tags":[],"see_also":[]},"autotune_memory_target_ratio":{"name":"autotune_memory_target_ratio","type":"float","level":"advanced","flags":0,"default_value":"0.7","min":"","max":"","enum_allowed":[],"desc":"ratio of total system memory to divide amongst autotuned daemons","long_desc":"","tags":[],"see_also":[]},"cephadm_log_destination":{"name":"cephadm_log_destination","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":["file","file,syslog","syslog"],"desc":"Destination for cephadm command's persistent logging","long_desc":"","tags":[],"see_also":[]},"cgroups_split":{"name":"cgroups_split","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"Pass --cgroups=split when cephadm creates containers (currently podman only)","long_desc":"","tags":[],"see_also":[]},"config_checks_enabled":{"name":"config_checks_enabled","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Enable or disable the cephadm configuration analysis","long_desc":"","tags":[],"see_also":[]},"config_dashboard":{"name":"config_dashboard","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"manage configs like API endpoints in Dashboard.","long_desc":"","tags":[],"see_also":[]},"container_image_alertmanager":{"name":"container_image_alertmanager","type":"str","level":"advanced","flags":0,"default_value":"quay.io/prometheus/alertmanager:v0.25.0","min":"","max":"","enum_allowed":[],"desc":"Prometheus container image","long_desc":"","tags":[],"see_also":[]},"container_image_base":{"name":"container_image_base","type":"str","level":"advanced","flags":1,"default_value":"quay.io/ceph/ceph","min":"","max":"","enum_allowed":[],"desc":"Container image name, without the tag","long_desc":"","tags":[],"see_also":[]},"container_image_elasticsearch":{"name":"container_image_elasticsearch","type":"str","level":"advanced","flags":0,"default_value":"quay.io/omrizeneva/elasticsearch:6.8.23","min":"","max":"","enum_allowed":[],"desc":"elasticsearch container image","long_desc":"","tags":[],"see_also":[]},"container_image_grafana":{"name":"container_image_grafana","type":"str","level":"advanced","flags":0,"default_value":"quay.io/ceph/grafana:10.4.0","min":"","max":"","enum_allowed":[],"desc":"Prometheus container image","long_desc":"","tags":[],"see_also":[]},"container_image_haproxy":{"name":"container_image_haproxy","type":"str","level":"advanced","flags":0,"default_value":"quay.io/ceph/haproxy:2.3","min":"","max":"","enum_allowed":[],"desc":"HAproxy container image","long_desc":"","tags":[],"see_also":[]},"container_image_jaeger_agent":{"name":"container_image_jaeger_agent","type":"str","level":"advanced","flags":0,"default_value":"quay.io/jaegertracing/jaeger-agent:1.29","min":"","max":"","enum_allowed":[],"desc":"Jaeger agent container image","long_desc":"","tags":[],"see_also":[]},"container_image_jaeger_collector":{"name":"container_image_jaeger_collector","type":"str","level":"advanced","flags":0,"default_value":"quay.io/jaegertracing/jaeger-collector:1.29","min":"","max":"","enum_allowed":[],"desc":"Jaeger collector container image","long_desc":"","tags":[],"see_also":[]},"container_image_jaeger_query":{"name":"container_image_jaeger_query","type":"str","level":"advanced","flags":0,"default_value":"quay.io/jaegertracing/jaeger-query:1.29","min":"","max":"","enum_allowed":[],"desc":"Jaeger query container image","long_desc":"","tags":[],"see_also":[]},"container_image_keepalived":{"name":"container_image_keepalived","type":"str","level":"advanced","flags":0,"default_value":"quay.io/ceph/keepalived:2.2.4","min":"","max":"","enum_allowed":[],"desc":"Keepalived container image","long_desc":"","tags":[],"see_also":[]},"container_image_loki":{"name":"container_image_loki","type":"str","level":"advanced","flags":0,"default_value":"quay.io/ceph/loki:3.0.0","min":"","max":"","enum_allowed":[],"desc":"Loki container image","long_desc":"","tags":[],"see_also":[]},"container_image_node_exporter":{"name":"container_image_node_exporter","type":"str","level":"advanced","flags":0,"default_value":"quay.io/prometheus/node-exporter:v1.7.0","min":"","max":"","enum_allowed":[],"desc":"Prometheus container image","long_desc":"","tags":[],"see_also":[]},"container_image_nvmeof":{"name":"container_image_nvmeof","type":"str","level":"advanced","flags":0,"default_value":"quay.io/ceph/nvmeof:1.2.5","min":"","max":"","enum_allowed":[],"desc":"Nvme-of container image","long_desc":"","tags":[],"see_also":[]},"container_image_prometheus":{"name":"container_image_prometheus","type":"str","level":"advanced","flags":0,"default_value":"quay.io/prometheus/prometheus:v2.51.0","min":"","max":"","enum_allowed":[],"desc":"Prometheus container image","long_desc":"","tags":[],"see_also":[]},"container_image_promtail":{"name":"container_image_promtail","type":"str","level":"advanced","flags":0,"default_value":"quay.io/ceph/promtail:3.0.0","min":"","max":"","enum_allowed":[],"desc":"Promtail container image","long_desc":"","tags":[],"see_also":[]},"container_image_samba":{"name":"container_image_samba","type":"str","level":"advanced","flags":0,"default_value":"quay.io/samba.org/samba-server:devbuilds-centos-amd64","min":"","max":"","enum_allowed":[],"desc":"Samba/SMB container image","long_desc":"","tags":[],"see_also":[]},"container_image_snmp_gateway":{"name":"container_image_snmp_gateway","type":"str","level":"advanced","flags":0,"default_value":"quay.io/ceph/snmp-notifier:v1.2.1","min":"","max":"","enum_allowed":[],"desc":"SNMP Gateway container image","long_desc":"","tags":[],"see_also":[]},"container_init":{"name":"container_init","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"Run podman/docker with `--init`","long_desc":"","tags":[],"see_also":[]},"daemon_cache_timeout":{"name":"daemon_cache_timeout","type":"secs","level":"advanced","flags":0,"default_value":"600","min":"","max":"","enum_allowed":[],"desc":"seconds to cache service (daemon) inventory","long_desc":"","tags":[],"see_also":[]},"default_cephadm_command_timeout":{"name":"default_cephadm_command_timeout","type":"int","level":"advanced","flags":0,"default_value":"900","min":"","max":"","enum_allowed":[],"desc":"Default timeout applied to cephadm commands run directly on the host (in seconds)","long_desc":"","tags":[],"see_also":[]},"default_registry":{"name":"default_registry","type":"str","level":"advanced","flags":0,"default_value":"quay.io","min":"","max":"","enum_allowed":[],"desc":"Search-registry to which we should normalize unqualified image names. This is not the default registry","long_desc":"","tags":[],"see_also":[]},"device_cache_timeout":{"name":"device_cache_timeout","type":"secs","level":"advanced","flags":0,"default_value":"1800","min":"","max":"","enum_allowed":[],"desc":"seconds to cache device inventory","long_desc":"","tags":[],"see_also":[]},"device_enhanced_scan":{"name":"device_enhanced_scan","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Use libstoragemgmt during device scans","long_desc":"","tags":[],"see_also":[]},"facts_cache_timeout":{"name":"facts_cache_timeout","type":"secs","level":"advanced","flags":0,"default_value":"60","min":"","max":"","enum_allowed":[],"desc":"seconds to cache host facts data","long_desc":"","tags":[],"see_also":[]},"grafana_dashboards_path":{"name":"grafana_dashboards_path","type":"str","level":"advanced","flags":0,"default_value":"/etc/grafana/dashboards/ceph-dashboard/","min":"","max":"","enum_allowed":[],"desc":"location of dashboards to include in grafana deployments","long_desc":"","tags":[],"see_also":[]},"host_check_interval":{"name":"host_check_interval","type":"secs","level":"advanced","flags":0,"default_value":"600","min":"","max":"","enum_allowed":[],"desc":"how frequently to perform a host check","long_desc":"","tags":[],"see_also":[]},"hw_monitoring":{"name":"hw_monitoring","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Deploy hw monitoring daemon on every host.","long_desc":"","tags":[],"see_also":[]},"inventory_list_all":{"name":"inventory_list_all","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Whether ceph-volume inventory should report more devices (mostly mappers (LVs / mpaths), partitions...)","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_refresh_metadata":{"name":"log_refresh_metadata","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Log all refresh metadata. Includes daemon, device, and host info collected regularly. Only has effect if logging at debug level","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"log to the \"cephadm\" cluster log channel\"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"manage_etc_ceph_ceph_conf":{"name":"manage_etc_ceph_ceph_conf","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Manage and own /etc/ceph/ceph.conf on the hosts.","long_desc":"","tags":[],"see_also":[]},"manage_etc_ceph_ceph_conf_hosts":{"name":"manage_etc_ceph_ceph_conf_hosts","type":"str","level":"advanced","flags":0,"default_value":"*","min":"","max":"","enum_allowed":[],"desc":"PlacementSpec describing on which hosts to manage /etc/ceph/ceph.conf","long_desc":"","tags":[],"see_also":[]},"max_count_per_host":{"name":"max_count_per_host","type":"int","level":"advanced","flags":0,"default_value":"10","min":"","max":"","enum_allowed":[],"desc":"max number of daemons per service per host","long_desc":"","tags":[],"see_also":[]},"max_osd_draining_count":{"name":"max_osd_draining_count","type":"int","level":"advanced","flags":0,"default_value":"10","min":"","max":"","enum_allowed":[],"desc":"max number of osds that will be drained simultaneously when osds are removed","long_desc":"","tags":[],"see_also":[]},"migration_current":{"name":"migration_current","type":"int","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"internal - do not modify","long_desc":"","tags":[],"see_also":[]},"mode":{"name":"mode","type":"str","level":"advanced","flags":0,"default_value":"root","min":"","max":"","enum_allowed":["cephadm-package","root"],"desc":"mode for remote execution of cephadm","long_desc":"","tags":[],"see_also":[]},"oob_default_addr":{"name":"oob_default_addr","type":"str","level":"advanced","flags":0,"default_value":"169.254.1.1","min":"","max":"","enum_allowed":[],"desc":"Default address for RedFish API (oob management).","long_desc":"","tags":[],"see_also":[]},"prometheus_alerts_path":{"name":"prometheus_alerts_path","type":"str","level":"advanced","flags":0,"default_value":"/etc/prometheus/ceph/ceph_default_alerts.yml","min":"","max":"","enum_allowed":[],"desc":"location of alerts to include in prometheus deployments","long_desc":"","tags":[],"see_also":[]},"registry_insecure":{"name":"registry_insecure","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Registry is to be considered insecure (no TLS available). Only for development purposes.","long_desc":"","tags":[],"see_also":[]},"registry_password":{"name":"registry_password","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"Custom repository password. Only used for logging into a registry.","long_desc":"","tags":[],"see_also":[]},"registry_url":{"name":"registry_url","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"Registry url for login purposes. This is not the default registry","long_desc":"","tags":[],"see_also":[]},"registry_username":{"name":"registry_username","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"Custom repository username. Only used for logging into a registry.","long_desc":"","tags":[],"see_also":[]},"secure_monitoring_stack":{"name":"secure_monitoring_stack","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Enable TLS security for all the monitoring stack daemons","long_desc":"","tags":[],"see_also":[]},"service_discovery_port":{"name":"service_discovery_port","type":"int","level":"advanced","flags":0,"default_value":"8765","min":"","max":"","enum_allowed":[],"desc":"cephadm service discovery port","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"ssh_config_file":{"name":"ssh_config_file","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"customized SSH config file to connect to managed hosts","long_desc":"","tags":[],"see_also":[]},"ssh_keepalive_count_max":{"name":"ssh_keepalive_count_max","type":"int","level":"advanced","flags":0,"default_value":"3","min":"","max":"","enum_allowed":[],"desc":"How many times ssh connections can fail liveness checks before the host is marked offline","long_desc":"","tags":[],"see_also":[]},"ssh_keepalive_interval":{"name":"ssh_keepalive_interval","type":"int","level":"advanced","flags":0,"default_value":"7","min":"","max":"","enum_allowed":[],"desc":"How often ssh connections are checked for liveness","long_desc":"","tags":[],"see_also":[]},"use_agent":{"name":"use_agent","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Use cephadm agent on each host to gather and send metadata","long_desc":"","tags":[],"see_also":[]},"use_repo_digest":{"name":"use_repo_digest","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"Automatically convert image tags to image digest. Make sure all daemons use the same image","long_desc":"","tags":[],"see_also":[]},"warn_on_failed_host_check":{"name":"warn_on_failed_host_check","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"raise a health warning if the host check fails","long_desc":"","tags":[],"see_also":[]},"warn_on_stray_daemons":{"name":"warn_on_stray_daemons","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"raise a health warning if daemons are detected that are not managed by cephadm","long_desc":"","tags":[],"see_also":[]},"warn_on_stray_hosts":{"name":"warn_on_stray_hosts","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"raise a health warning if daemons are detected on a host that is not managed by cephadm","long_desc":"","tags":[],"see_also":[]}}},{"name":"crash","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"retain_interval":{"name":"retain_interval","type":"secs","level":"advanced","flags":1,"default_value":"31536000","min":"","max":"","enum_allowed":[],"desc":"how long to retain crashes before pruning them","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"warn_recent_interval":{"name":"warn_recent_interval","type":"secs","level":"advanced","flags":1,"default_value":"1209600","min":"","max":"","enum_allowed":[],"desc":"time interval in which to warn about recent crashes","long_desc":"","tags":[],"see_also":[]}}},{"name":"dashboard","can_run":true,"error_string":"","module_options":{"ACCOUNT_LOCKOUT_ATTEMPTS":{"name":"ACCOUNT_LOCKOUT_ATTEMPTS","type":"int","level":"advanced","flags":0,"default_value":"10","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"ALERTMANAGER_API_HOST":{"name":"ALERTMANAGER_API_HOST","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"ALERTMANAGER_API_SSL_VERIFY":{"name":"ALERTMANAGER_API_SSL_VERIFY","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"AUDIT_API_ENABLED":{"name":"AUDIT_API_ENABLED","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"AUDIT_API_LOG_PAYLOAD":{"name":"AUDIT_API_LOG_PAYLOAD","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"ENABLE_BROWSABLE_API":{"name":"ENABLE_BROWSABLE_API","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"FEATURE_TOGGLE_CEPHFS":{"name":"FEATURE_TOGGLE_CEPHFS","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"FEATURE_TOGGLE_DASHBOARD":{"name":"FEATURE_TOGGLE_DASHBOARD","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"FEATURE_TOGGLE_ISCSI":{"name":"FEATURE_TOGGLE_ISCSI","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"FEATURE_TOGGLE_MIRRORING":{"name":"FEATURE_TOGGLE_MIRRORING","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"FEATURE_TOGGLE_NFS":{"name":"FEATURE_TOGGLE_NFS","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"FEATURE_TOGGLE_RBD":{"name":"FEATURE_TOGGLE_RBD","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"FEATURE_TOGGLE_RGW":{"name":"FEATURE_TOGGLE_RGW","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"GANESHA_CLUSTERS_RADOS_POOL_NAMESPACE":{"name":"GANESHA_CLUSTERS_RADOS_POOL_NAMESPACE","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"GRAFANA_API_PASSWORD":{"name":"GRAFANA_API_PASSWORD","type":"str","level":"advanced","flags":0,"default_value":"admin","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"GRAFANA_API_SSL_VERIFY":{"name":"GRAFANA_API_SSL_VERIFY","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"GRAFANA_API_URL":{"name":"GRAFANA_API_URL","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"GRAFANA_API_USERNAME":{"name":"GRAFANA_API_USERNAME","type":"str","level":"advanced","flags":0,"default_value":"admin","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"GRAFANA_FRONTEND_API_URL":{"name":"GRAFANA_FRONTEND_API_URL","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"GRAFANA_UPDATE_DASHBOARDS":{"name":"GRAFANA_UPDATE_DASHBOARDS","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"ISCSI_API_SSL_VERIFICATION":{"name":"ISCSI_API_SSL_VERIFICATION","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"ISSUE_TRACKER_API_KEY":{"name":"ISSUE_TRACKER_API_KEY","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PROMETHEUS_API_HOST":{"name":"PROMETHEUS_API_HOST","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PROMETHEUS_API_SSL_VERIFY":{"name":"PROMETHEUS_API_SSL_VERIFY","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PWD_POLICY_CHECK_COMPLEXITY_ENABLED":{"name":"PWD_POLICY_CHECK_COMPLEXITY_ENABLED","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PWD_POLICY_CHECK_EXCLUSION_LIST_ENABLED":{"name":"PWD_POLICY_CHECK_EXCLUSION_LIST_ENABLED","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PWD_POLICY_CHECK_LENGTH_ENABLED":{"name":"PWD_POLICY_CHECK_LENGTH_ENABLED","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PWD_POLICY_CHECK_OLDPWD_ENABLED":{"name":"PWD_POLICY_CHECK_OLDPWD_ENABLED","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PWD_POLICY_CHECK_REPETITIVE_CHARS_ENABLED":{"name":"PWD_POLICY_CHECK_REPETITIVE_CHARS_ENABLED","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PWD_POLICY_CHECK_SEQUENTIAL_CHARS_ENABLED":{"name":"PWD_POLICY_CHECK_SEQUENTIAL_CHARS_ENABLED","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PWD_POLICY_CHECK_USERNAME_ENABLED":{"name":"PWD_POLICY_CHECK_USERNAME_ENABLED","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PWD_POLICY_ENABLED":{"name":"PWD_POLICY_ENABLED","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PWD_POLICY_EXCLUSION_LIST":{"name":"PWD_POLICY_EXCLUSION_LIST","type":"str","level":"advanced","flags":0,"default_value":"osd,host,dashboard,pool,block,nfs,ceph,monitors,gateway,logs,crush,maps","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PWD_POLICY_MIN_COMPLEXITY":{"name":"PWD_POLICY_MIN_COMPLEXITY","type":"int","level":"advanced","flags":0,"default_value":"10","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PWD_POLICY_MIN_LENGTH":{"name":"PWD_POLICY_MIN_LENGTH","type":"int","level":"advanced","flags":0,"default_value":"8","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"REST_REQUESTS_TIMEOUT":{"name":"REST_REQUESTS_TIMEOUT","type":"int","level":"advanced","flags":0,"default_value":"45","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"RGW_API_ACCESS_KEY":{"name":"RGW_API_ACCESS_KEY","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"RGW_API_ADMIN_RESOURCE":{"name":"RGW_API_ADMIN_RESOURCE","type":"str","level":"advanced","flags":0,"default_value":"admin","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"RGW_API_SECRET_KEY":{"name":"RGW_API_SECRET_KEY","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"RGW_API_SSL_VERIFY":{"name":"RGW_API_SSL_VERIFY","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"USER_PWD_EXPIRATION_SPAN":{"name":"USER_PWD_EXPIRATION_SPAN","type":"int","level":"advanced","flags":0,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"USER_PWD_EXPIRATION_WARNING_1":{"name":"USER_PWD_EXPIRATION_WARNING_1","type":"int","level":"advanced","flags":0,"default_value":"10","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"USER_PWD_EXPIRATION_WARNING_2":{"name":"USER_PWD_EXPIRATION_WARNING_2","type":"int","level":"advanced","flags":0,"default_value":"5","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"cross_origin_url":{"name":"cross_origin_url","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"crt_file":{"name":"crt_file","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"debug":{"name":"debug","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Enable/disable debug options","long_desc":"","tags":[],"see_also":[]},"jwt_token_ttl":{"name":"jwt_token_ttl","type":"int","level":"advanced","flags":0,"default_value":"28800","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"key_file":{"name":"key_file","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"motd":{"name":"motd","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"The message of the day","long_desc":"","tags":[],"see_also":[]},"redirect_resolve_ip_addr":{"name":"redirect_resolve_ip_addr","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"server_addr":{"name":"server_addr","type":"str","level":"advanced","flags":0,"default_value":"::","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"server_port":{"name":"server_port","type":"int","level":"advanced","flags":0,"default_value":"8080","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"ssl":{"name":"ssl","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"ssl_server_port":{"name":"ssl_server_port","type":"int","level":"advanced","flags":0,"default_value":"8443","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"standby_behaviour":{"name":"standby_behaviour","type":"str","level":"advanced","flags":0,"default_value":"redirect","min":"","max":"","enum_allowed":["error","redirect"],"desc":"","long_desc":"","tags":[],"see_also":[]},"standby_error_status_code":{"name":"standby_error_status_code","type":"int","level":"advanced","flags":0,"default_value":"500","min":"400","max":"599","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"url_prefix":{"name":"url_prefix","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"devicehealth","can_run":true,"error_string":"","module_options":{"enable_monitoring":{"name":"enable_monitoring","type":"bool","level":"advanced","flags":1,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"monitor device health metrics","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"mark_out_threshold":{"name":"mark_out_threshold","type":"secs","level":"advanced","flags":1,"default_value":"2419200","min":"","max":"","enum_allowed":[],"desc":"automatically mark OSD if it may fail before this long","long_desc":"","tags":[],"see_also":[]},"pool_name":{"name":"pool_name","type":"str","level":"advanced","flags":1,"default_value":"device_health_metrics","min":"","max":"","enum_allowed":[],"desc":"name of pool in which to store device health metrics","long_desc":"","tags":[],"see_also":[]},"retention_period":{"name":"retention_period","type":"secs","level":"advanced","flags":1,"default_value":"15552000","min":"","max":"","enum_allowed":[],"desc":"how long to retain device health metrics","long_desc":"","tags":[],"see_also":[]},"scrape_frequency":{"name":"scrape_frequency","type":"secs","level":"advanced","flags":1,"default_value":"86400","min":"","max":"","enum_allowed":[],"desc":"how frequently to scrape device health metrics","long_desc":"","tags":[],"see_also":[]},"self_heal":{"name":"self_heal","type":"bool","level":"advanced","flags":1,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"preemptively heal cluster around devices that may fail","long_desc":"","tags":[],"see_also":[]},"sleep_interval":{"name":"sleep_interval","type":"secs","level":"advanced","flags":1,"default_value":"600","min":"","max":"","enum_allowed":[],"desc":"how frequently to wake up and check device health","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"warn_threshold":{"name":"warn_threshold","type":"secs","level":"advanced","flags":1,"default_value":"7257600","min":"","max":"","enum_allowed":[],"desc":"raise health warning if OSD may fail before this long","long_desc":"","tags":[],"see_also":[]}}},{"name":"diskprediction_local","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"predict_interval":{"name":"predict_interval","type":"str","level":"advanced","flags":0,"default_value":"86400","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"predictor_model":{"name":"predictor_model","type":"str","level":"advanced","flags":0,"default_value":"prophetstor","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sleep_interval":{"name":"sleep_interval","type":"str","level":"advanced","flags":0,"default_value":"600","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"influx","can_run":false,"error_string":"influxdb python module not found","module_options":{"batch_size":{"name":"batch_size","type":"int","level":"advanced","flags":0,"default_value":"5000","min":"","max":"","enum_allowed":[],"desc":"How big batches of data points should be when sending to InfluxDB.","long_desc":"","tags":[],"see_also":[]},"database":{"name":"database","type":"str","level":"advanced","flags":0,"default_value":"ceph","min":"","max":"","enum_allowed":[],"desc":"InfluxDB database name. You will need to create this database and grant write privileges to the configured username or the username must have admin privileges to create it.","long_desc":"","tags":[],"see_also":[]},"hostname":{"name":"hostname","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"InfluxDB server hostname","long_desc":"","tags":[],"see_also":[]},"interval":{"name":"interval","type":"secs","level":"advanced","flags":0,"default_value":"30","min":"5","max":"","enum_allowed":[],"desc":"Time between reports to InfluxDB. Default 30 seconds.","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"password":{"name":"password","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"password of InfluxDB server user","long_desc":"","tags":[],"see_also":[]},"port":{"name":"port","type":"int","level":"advanced","flags":0,"default_value":"8086","min":"","max":"","enum_allowed":[],"desc":"InfluxDB server port","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"ssl":{"name":"ssl","type":"str","level":"advanced","flags":0,"default_value":"false","min":"","max":"","enum_allowed":[],"desc":"Use https connection for InfluxDB server. Use \"true\" or \"false\".","long_desc":"","tags":[],"see_also":[]},"threads":{"name":"threads","type":"int","level":"advanced","flags":0,"default_value":"5","min":"1","max":"32","enum_allowed":[],"desc":"How many worker threads should be spawned for sending data to InfluxDB.","long_desc":"","tags":[],"see_also":[]},"username":{"name":"username","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"username of InfluxDB server user","long_desc":"","tags":[],"see_also":[]},"verify_ssl":{"name":"verify_ssl","type":"str","level":"advanced","flags":0,"default_value":"true","min":"","max":"","enum_allowed":[],"desc":"Verify https cert for InfluxDB server. Use \"true\" or \"false\".","long_desc":"","tags":[],"see_also":[]}}},{"name":"insights","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"iostat","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"k8sevents","can_run":true,"error_string":"","module_options":{"ceph_event_retention_days":{"name":"ceph_event_retention_days","type":"int","level":"advanced","flags":0,"default_value":"7","min":"","max":"","enum_allowed":[],"desc":"Days to hold ceph event information within local cache","long_desc":"","tags":[],"see_also":[]},"config_check_secs":{"name":"config_check_secs","type":"int","level":"advanced","flags":0,"default_value":"10","min":"10","max":"","enum_allowed":[],"desc":"interval (secs) to check for cluster configuration changes","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"localpool","can_run":true,"error_string":"","module_options":{"failure_domain":{"name":"failure_domain","type":"str","level":"advanced","flags":1,"default_value":"host","min":"","max":"","enum_allowed":[],"desc":"failure domain for any created local pool","long_desc":"what failure domain we should separate data replicas across.","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"min_size":{"name":"min_size","type":"int","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"default min_size for any created local pool","long_desc":"value to set min_size to (unchanged from Ceph's default if this option is not set)","tags":[],"see_also":[]},"num_rep":{"name":"num_rep","type":"int","level":"advanced","flags":1,"default_value":"3","min":"","max":"","enum_allowed":[],"desc":"default replica count for any created local pool","long_desc":"","tags":[],"see_also":[]},"pg_num":{"name":"pg_num","type":"int","level":"advanced","flags":1,"default_value":"128","min":"","max":"","enum_allowed":[],"desc":"default pg_num for any created local pool","long_desc":"","tags":[],"see_also":[]},"prefix":{"name":"prefix","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"name prefix for any created local pool","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"subtree":{"name":"subtree","type":"str","level":"advanced","flags":1,"default_value":"rack","min":"","max":"","enum_allowed":[],"desc":"CRUSH level for which to create a local pool","long_desc":"which CRUSH subtree type the module should create a pool for.","tags":[],"see_also":[]}}},{"name":"mds_autoscaler","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"mirroring","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"nfs","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"orchestrator","can_run":true,"error_string":"","module_options":{"fail_fs":{"name":"fail_fs","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Fail filesystem for rapid multi-rank mds upgrade","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"orchestrator":{"name":"orchestrator","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["cephadm","rook","test_orchestrator"],"desc":"Orchestrator backend","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"osd_perf_query","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"osd_support","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"pg_autoscaler","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sleep_interval":{"name":"sleep_interval","type":"secs","level":"advanced","flags":0,"default_value":"60","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"threshold":{"name":"threshold","type":"float","level":"advanced","flags":0,"default_value":"3.0","min":"1.0","max":"","enum_allowed":[],"desc":"scaling threshold","long_desc":"The factor by which the `NEW PG_NUM` must vary from the current`PG_NUM` before being accepted. Cannot be less than 1.0","tags":[],"see_also":[]}}},{"name":"progress","can_run":true,"error_string":"","module_options":{"allow_pg_recovery_event":{"name":"allow_pg_recovery_event","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"allow the module to show pg recovery progress","long_desc":"","tags":[],"see_also":[]},"enabled":{"name":"enabled","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"max_completed_events":{"name":"max_completed_events","type":"int","level":"advanced","flags":1,"default_value":"50","min":"","max":"","enum_allowed":[],"desc":"number of past completed events to remember","long_desc":"","tags":[],"see_also":[]},"sleep_interval":{"name":"sleep_interval","type":"secs","level":"advanced","flags":1,"default_value":"5","min":"","max":"","enum_allowed":[],"desc":"how long the module is going to sleep","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"prometheus","can_run":true,"error_string":"","module_options":{"cache":{"name":"cache","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"exclude_perf_counters":{"name":"exclude_perf_counters","type":"bool","level":"advanced","flags":1,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"Do not include perf-counters in the metrics output","long_desc":"Gathering perf-counters from a single Prometheus exporter can degrade ceph-mgr performance, especially in large clusters. Instead, Ceph-exporter daemons are now used by default for perf-counter gathering. This should only be disabled when no ceph-exporters are deployed.","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"rbd_stats_pools":{"name":"rbd_stats_pools","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"rbd_stats_pools_refresh_interval":{"name":"rbd_stats_pools_refresh_interval","type":"int","level":"advanced","flags":0,"default_value":"300","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"scrape_interval":{"name":"scrape_interval","type":"float","level":"advanced","flags":0,"default_value":"15.0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"server_addr":{"name":"server_addr","type":"str","level":"advanced","flags":0,"default_value":"::","min":"","max":"","enum_allowed":[],"desc":"the IPv4 or IPv6 address on which the module listens for HTTP requests","long_desc":"","tags":[],"see_also":[]},"server_port":{"name":"server_port","type":"int","level":"advanced","flags":1,"default_value":"9283","min":"","max":"","enum_allowed":[],"desc":"the port on which the module listens for HTTP requests","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"stale_cache_strategy":{"name":"stale_cache_strategy","type":"str","level":"advanced","flags":0,"default_value":"log","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"standby_behaviour":{"name":"standby_behaviour","type":"str","level":"advanced","flags":1,"default_value":"default","min":"","max":"","enum_allowed":["default","error"],"desc":"","long_desc":"","tags":[],"see_also":[]},"standby_error_status_code":{"name":"standby_error_status_code","type":"int","level":"advanced","flags":1,"default_value":"500","min":"400","max":"599","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"rbd_support","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"max_concurrent_snap_create":{"name":"max_concurrent_snap_create","type":"int","level":"advanced","flags":0,"default_value":"10","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"mirror_snapshot_schedule":{"name":"mirror_snapshot_schedule","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"trash_purge_schedule":{"name":"trash_purge_schedule","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"restful","can_run":true,"error_string":"","module_options":{"enable_auth":{"name":"enable_auth","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"key_file":{"name":"key_file","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"max_requests":{"name":"max_requests","type":"int","level":"advanced","flags":0,"default_value":"500","min":"","max":"","enum_allowed":[],"desc":"Maximum number of requests to keep in memory. When new request comes in, the oldest request will be removed if the number of requests exceeds the max request number. if un-finished request is removed, error message will be logged in the ceph-mgr log.","long_desc":"","tags":[],"see_also":[]},"server_addr":{"name":"server_addr","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"server_port":{"name":"server_port","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"rgw","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"secondary_zone_period_retry_limit":{"name":"secondary_zone_period_retry_limit","type":"int","level":"advanced","flags":0,"default_value":"5","min":"","max":"","enum_allowed":[],"desc":"RGW module period update retry limit for secondary site","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"rook","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"storage_class":{"name":"storage_class","type":"str","level":"advanced","flags":0,"default_value":"local","min":"","max":"","enum_allowed":[],"desc":"storage class name for LSO-discovered PVs","long_desc":"","tags":[],"see_also":[]}}},{"name":"selftest","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"roption1":{"name":"roption1","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"roption2":{"name":"roption2","type":"str","level":"advanced","flags":0,"default_value":"xyz","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"rwoption1":{"name":"rwoption1","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"rwoption2":{"name":"rwoption2","type":"int","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"rwoption3":{"name":"rwoption3","type":"float","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"rwoption4":{"name":"rwoption4","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"rwoption5":{"name":"rwoption5","type":"bool","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"rwoption6":{"name":"rwoption6","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"rwoption7":{"name":"rwoption7","type":"int","level":"advanced","flags":0,"default_value":"","min":"1","max":"42","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"testkey":{"name":"testkey","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"testlkey":{"name":"testlkey","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"testnewline":{"name":"testnewline","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"snap_schedule","can_run":true,"error_string":"","module_options":{"allow_m_granularity":{"name":"allow_m_granularity","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"allow minute scheduled snapshots","long_desc":"","tags":[],"see_also":[]},"dump_on_update":{"name":"dump_on_update","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"dump database to debug log on update","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"stats","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"status","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"telegraf","can_run":true,"error_string":"","module_options":{"address":{"name":"address","type":"str","level":"advanced","flags":0,"default_value":"unixgram:///tmp/telegraf.sock","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"interval":{"name":"interval","type":"secs","level":"advanced","flags":0,"default_value":"15","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"telemetry","can_run":true,"error_string":"","module_options":{"channel_basic":{"name":"channel_basic","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"Share basic cluster information (size, version)","long_desc":"","tags":[],"see_also":[]},"channel_crash":{"name":"channel_crash","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"Share metadata about Ceph daemon crashes (version, stack straces, etc)","long_desc":"","tags":[],"see_also":[]},"channel_device":{"name":"channel_device","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"Share device health metrics (e.g., SMART data, minus potentially identifying info like serial numbers)","long_desc":"","tags":[],"see_also":[]},"channel_ident":{"name":"channel_ident","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Share a user-provided description and/or contact email for the cluster","long_desc":"","tags":[],"see_also":[]},"channel_perf":{"name":"channel_perf","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Share various performance metrics of a cluster","long_desc":"","tags":[],"see_also":[]},"contact":{"name":"contact","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"description":{"name":"description","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"device_url":{"name":"device_url","type":"str","level":"advanced","flags":0,"default_value":"https://telemetry.ceph.com/device","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"enabled":{"name":"enabled","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"interval":{"name":"interval","type":"int","level":"advanced","flags":0,"default_value":"24","min":"8","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"last_opt_revision":{"name":"last_opt_revision","type":"int","level":"advanced","flags":0,"default_value":"1","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"leaderboard":{"name":"leaderboard","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"leaderboard_description":{"name":"leaderboard_description","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"organization":{"name":"organization","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"proxy":{"name":"proxy","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"url":{"name":"url","type":"str","level":"advanced","flags":0,"default_value":"https://telemetry.ceph.com/report","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"test_orchestrator","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"volumes","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"max_concurrent_clones":{"name":"max_concurrent_clones","type":"int","level":"advanced","flags":0,"default_value":"4","min":"","max":"","enum_allowed":[],"desc":"Number of asynchronous cloner threads","long_desc":"","tags":[],"see_also":[]},"periodic_async_work":{"name":"periodic_async_work","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Periodically check for async work","long_desc":"","tags":[],"see_also":[]},"snapshot_clone_delay":{"name":"snapshot_clone_delay","type":"int","level":"advanced","flags":0,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"Delay clone begin operation by snapshot_clone_delay seconds","long_desc":"","tags":[],"see_also":[]},"snapshot_clone_no_wait":{"name":"snapshot_clone_no_wait","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"Reject subvolume clone request when cloner threads are busy","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"zabbix","can_run":true,"error_string":"","module_options":{"discovery_interval":{"name":"discovery_interval","type":"uint","level":"advanced","flags":0,"default_value":"100","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"identifier":{"name":"identifier","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"interval":{"name":"interval","type":"secs","level":"advanced","flags":0,"default_value":"60","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"zabbix_host":{"name":"zabbix_host","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"zabbix_port":{"name":"zabbix_port","type":"int","level":"advanced","flags":0,"default_value":"10051","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"zabbix_sender":{"name":"zabbix_sender","type":"str","level":"advanced","flags":0,"default_value":"/usr/bin/zabbix_sender","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}}],"services":{"dashboard":"https://192.168.123.103:8443/"},"always_on_modules":{"octopus":["balancer","crash","devicehealth","orchestrator","pg_autoscaler","progress","rbd_support","status","telemetry","volumes"],"pacific":["balancer","crash","devicehealth","orchestrator","pg_autoscaler","progress","rbd_support","status","telemetry","volumes"],"quincy":["balancer","crash","devicehealth","orchestrator","pg_autoscaler","progress","rbd_support","status","telemetry","volumes"],"reef":["balancer","crash","devicehealth","orchestrator","pg_autoscaler","progress","rbd_support","status","telemetry","volumes"],"squid":["balancer","crash","devicehealth","orchestrator","pg_autoscaler","progress","rbd_support","status","telemetry","volumes"]},"force_disabled_modules":{},"last_failure_osd_epoch":3,"active_clients":[{"name":"libcephsqlite","addrvec":[{"type":"v2","addr":"192.168.123.103:0","nonce":4007691705}]},{"name":"rbd_support","addrvec":[{"type":"v2","addr":"192.168.123.103:0","nonce":2413255244}]},{"name":"volumes","addrvec":[{"type":"v2","addr":"192.168.123.103:0","nonce":809873473}]}]} 2026-03-09T15:14:01.516 INFO:tasks.cephadm.ceph_manager.ceph:mgr available! 2026-03-09T15:14:01.516 INFO:tasks.cephadm.ceph_manager.ceph:waiting for all up 2026-03-09T15:14:01.516 DEBUG:teuthology.orchestra.run.vm03:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell --fsid 4abede80-1bca-11f1-bcd8-9b1da3c03ce5 -- ceph osd dump --format=json 2026-03-09T15:14:01.686 INFO:teuthology.orchestra.run.vm03.stderr:Inferring config /var/lib/ceph/4abede80-1bca-11f1-bcd8-9b1da3c03ce5/mon.a/config 2026-03-09T15:14:01.751 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 15:14:01 vm04 ceph-mon[53998]: purged_snaps scrub starts 2026-03-09T15:14:01.751 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 15:14:01 vm04 ceph-mon[53998]: purged_snaps scrub ok 2026-03-09T15:14:01.751 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 15:14:01 vm04 ceph-mon[53998]: pgmap v35: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-09T15:14:01.751 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 15:14:01 vm04 ceph-mon[53998]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' cmd=[{"prefix": "osd pool create", "format": "json", "pool": ".mgr", "pg_num": 1, "pg_num_min": 1, "pg_num_max": 32, "yes_i_really_mean_it": true}]: dispatch 2026-03-09T15:14:01.751 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 15:14:01 vm04 ceph-mon[53998]: from='client.? 192.168.123.103:0/2604958013' entity='client.admin' cmd=[{"prefix": "osd stat", "format": "json"}]: dispatch 2026-03-09T15:14:01.751 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 15:14:01 vm04 ceph-mon[53998]: from='client.? 192.168.123.103:0/3444384227' entity='client.admin' cmd=[{"prefix": "osd dump", "format": "json"}]: dispatch 2026-03-09T15:14:01.768 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 09 15:14:01 vm06 ceph-mon[56506]: purged_snaps scrub starts 2026-03-09T15:14:01.768 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 09 15:14:01 vm06 ceph-mon[56506]: purged_snaps scrub ok 2026-03-09T15:14:01.768 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 09 15:14:01 vm06 ceph-mon[56506]: pgmap v35: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-09T15:14:01.768 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 09 15:14:01 vm06 ceph-mon[56506]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' cmd=[{"prefix": "osd pool create", "format": "json", "pool": ".mgr", "pg_num": 1, "pg_num_min": 1, "pg_num_max": 32, "yes_i_really_mean_it": true}]: dispatch 2026-03-09T15:14:01.768 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 09 15:14:01 vm06 ceph-mon[56506]: from='client.? 192.168.123.103:0/2604958013' entity='client.admin' cmd=[{"prefix": "osd stat", "format": "json"}]: dispatch 2026-03-09T15:14:01.768 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 09 15:14:01 vm06 ceph-mon[56506]: from='client.? 192.168.123.103:0/3444384227' entity='client.admin' cmd=[{"prefix": "osd dump", "format": "json"}]: dispatch 2026-03-09T15:14:01.896 INFO:teuthology.orchestra.run.vm03.stdout: 2026-03-09T15:14:01.897 INFO:teuthology.orchestra.run.vm03.stdout:{"epoch":17,"fsid":"4abede80-1bca-11f1-bcd8-9b1da3c03ce5","created":"2026-03-09T15:12:42.233631+0000","modified":"2026-03-09T15:14:01.364622+0000","last_up_change":"2026-03-09T15:14:00.353477+0000","last_in_change":"2026-03-09T15:13:50.194917+0000","flags":"sortbitwise,recovery_deletes,purged_snapdirs,pglog_hardlimit","flags_num":5799936,"flags_set":["pglog_hardlimit","purged_snapdirs","recovery_deletes","sortbitwise"],"crush_version":8,"full_ratio":0.94999998807907104,"backfillfull_ratio":0.89999997615814209,"nearfull_ratio":0.85000002384185791,"cluster_snapshot":"","pool_max":1,"max_osd":3,"require_min_compat_client":"luminous","min_compat_client":"jewel","require_osd_release":"squid","allow_crimson":false,"pools":[{"pool":1,"pool_name":".mgr","create_time":"2026-03-09T15:14:00.523821+0000","flags":32769,"flags_names":"hashpspool,creating","type":1,"size":3,"min_size":2,"crush_rule":0,"peering_crush_bucket_count":0,"peering_crush_bucket_target":0,"peering_crush_bucket_barrier":0,"peering_crush_bucket_mandatory_member":2147483647,"is_stretch_pool":false,"object_hash":2,"pg_autoscale_mode":"off","pg_num":1,"pg_placement_num":1,"pg_placement_num_target":1,"pg_num_target":1,"pg_num_pending":1,"last_pg_merge_meta":{"source_pgid":"0.0","ready_epoch":0,"last_epoch_started":0,"last_epoch_clean":0,"source_version":"0'0","target_version":"0'0"},"last_change":"17","last_force_op_resend":"0","last_force_op_resend_prenautilus":"0","last_force_op_resend_preluminous":"0","auid":0,"snap_mode":"selfmanaged","snap_seq":0,"snap_epoch":0,"pool_snaps":[],"removed_snaps":"[]","quota_max_bytes":0,"quota_max_objects":0,"tiers":[],"tier_of":-1,"read_tier":-1,"write_tier":-1,"cache_mode":"none","target_max_bytes":0,"target_max_objects":0,"cache_target_dirty_ratio_micro":400000,"cache_target_dirty_high_ratio_micro":600000,"cache_target_full_ratio_micro":800000,"cache_min_flush_age":0,"cache_min_evict_age":0,"erasure_code_profile":"","hit_set_params":{"type":"none"},"hit_set_period":0,"hit_set_count":0,"use_gmt_hitset":true,"min_read_recency_for_promote":0,"min_write_recency_for_promote":0,"hit_set_grade_decay_rate":0,"hit_set_search_last_n":0,"grade_table":[],"stripe_width":0,"expected_num_objects":0,"fast_read":false,"options":{"pg_num_max":32,"pg_num_min":1},"application_metadata":{},"read_balance":{"score_type":"Fair distribution","score_acting":3,"score_stable":3,"optimal_score":1,"raw_score_acting":3,"raw_score_stable":3,"primary_affinity_weighted":1,"average_primary_affinity":1,"average_primary_affinity_weighted":1}}],"osds":[{"osd":0,"uuid":"913171bb-10e5-4020-a31f-9550e7c31d7b","up":1,"in":1,"weight":1,"primary_affinity":1,"last_clean_begin":0,"last_clean_end":0,"up_from":8,"up_thru":0,"down_at":0,"lost_at":0,"public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.103:6802","nonce":615876468},{"type":"v1","addr":"192.168.123.103:6803","nonce":615876468}]},"cluster_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.103:6804","nonce":615876468},{"type":"v1","addr":"192.168.123.103:6805","nonce":615876468}]},"heartbeat_back_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.103:6808","nonce":615876468},{"type":"v1","addr":"192.168.123.103:6809","nonce":615876468}]},"heartbeat_front_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.103:6806","nonce":615876468},{"type":"v1","addr":"192.168.123.103:6807","nonce":615876468}]},"public_addr":"192.168.123.103:6803/615876468","cluster_addr":"192.168.123.103:6805/615876468","heartbeat_back_addr":"192.168.123.103:6809/615876468","heartbeat_front_addr":"192.168.123.103:6807/615876468","state":["exists","up"]},{"osd":1,"uuid":"62ee7f4e-6ba5-481b-9cde-4c62baacf802","up":1,"in":1,"weight":1,"primary_affinity":1,"last_clean_begin":0,"last_clean_end":0,"up_from":12,"up_thru":0,"down_at":0,"lost_at":0,"public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.104:6800","nonce":1573535831},{"type":"v1","addr":"192.168.123.104:6801","nonce":1573535831}]},"cluster_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.104:6802","nonce":1573535831},{"type":"v1","addr":"192.168.123.104:6803","nonce":1573535831}]},"heartbeat_back_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.104:6806","nonce":1573535831},{"type":"v1","addr":"192.168.123.104:6807","nonce":1573535831}]},"heartbeat_front_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.104:6804","nonce":1573535831},{"type":"v1","addr":"192.168.123.104:6805","nonce":1573535831}]},"public_addr":"192.168.123.104:6801/1573535831","cluster_addr":"192.168.123.104:6803/1573535831","heartbeat_back_addr":"192.168.123.104:6807/1573535831","heartbeat_front_addr":"192.168.123.104:6805/1573535831","state":["exists","up"]},{"osd":2,"uuid":"d53d63dc-a0c2-402c-8257-919cdd6b513c","up":1,"in":1,"weight":1,"primary_affinity":1,"last_clean_begin":0,"last_clean_end":0,"up_from":16,"up_thru":0,"down_at":0,"lost_at":0,"public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.106:6800","nonce":3722863356},{"type":"v1","addr":"192.168.123.106:6801","nonce":3722863356}]},"cluster_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.106:6802","nonce":3722863356},{"type":"v1","addr":"192.168.123.106:6803","nonce":3722863356}]},"heartbeat_back_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.106:6806","nonce":3722863356},{"type":"v1","addr":"192.168.123.106:6807","nonce":3722863356}]},"heartbeat_front_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.106:6804","nonce":3722863356},{"type":"v1","addr":"192.168.123.106:6805","nonce":3722863356}]},"public_addr":"192.168.123.106:6801/3722863356","cluster_addr":"192.168.123.106:6803/3722863356","heartbeat_back_addr":"192.168.123.106:6807/3722863356","heartbeat_front_addr":"192.168.123.106:6805/3722863356","state":["exists","up"]}],"osd_xinfo":[{"osd":0,"down_stamp":"0.000000","laggy_probability":0,"laggy_interval":0,"features":4540701547738038271,"old_weight":0,"last_purged_snaps_scrub":"2026-03-09T15:13:40.595321+0000","dead_epoch":0},{"osd":1,"down_stamp":"0.000000","laggy_probability":0,"laggy_interval":0,"features":4540701547738038271,"old_weight":0,"last_purged_snaps_scrub":"2026-03-09T15:13:49.850182+0000","dead_epoch":0},{"osd":2,"down_stamp":"0.000000","laggy_probability":0,"laggy_interval":0,"features":4540701547738038271,"old_weight":0,"last_purged_snaps_scrub":"2026-03-09T15:13:58.860702+0000","dead_epoch":0}],"pg_upmap":[],"pg_upmap_items":[],"pg_upmap_primaries":[],"pg_temp":[],"primary_temp":[],"blocklist":{"192.168.123.103:6801/3510874031":"2026-03-10T15:13:02.478314+0000","192.168.123.103:6800/3510874031":"2026-03-10T15:13:02.478314+0000","192.168.123.103:0/1665269914":"2026-03-10T15:13:02.478314+0000","192.168.123.103:0/4246730401":"2026-03-10T15:13:02.478314+0000","192.168.123.103:0/983012295":"2026-03-10T15:12:52.735172+0000","192.168.123.103:0/3380842377":"2026-03-10T15:13:02.478314+0000","192.168.123.103:0/1826640327":"2026-03-10T15:12:52.735172+0000","192.168.123.103:6800/2441273787":"2026-03-10T15:12:52.735172+0000","192.168.123.103:6801/2441273787":"2026-03-10T15:12:52.735172+0000","192.168.123.103:0/4187493480":"2026-03-10T15:12:52.735172+0000"},"range_blocklist":{},"erasure_code_profiles":{"default":{"crush-failure-domain":"osd","k":"2","m":"1","plugin":"jerasure","technique":"reed_sol_van"}},"removed_snaps_queue":[],"new_removed_snaps":[],"new_purged_snaps":[],"crush_node_flags":{},"device_class_flags":{},"stretch_mode":{"stretch_mode_enabled":false,"stretch_bucket_count":0,"degraded_stretch_mode":0,"recovering_stretch_mode":0,"stretch_mode_bucket":0}} 2026-03-09T15:14:01.957 INFO:tasks.cephadm.ceph_manager.ceph:all up! 2026-03-09T15:14:01.957 DEBUG:teuthology.orchestra.run.vm03:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell --fsid 4abede80-1bca-11f1-bcd8-9b1da3c03ce5 -- ceph osd dump --format=json 2026-03-09T15:14:02.106 INFO:teuthology.orchestra.run.vm03.stderr:Inferring config /var/lib/ceph/4abede80-1bca-11f1-bcd8-9b1da3c03ce5/mon.a/config 2026-03-09T15:14:02.330 INFO:teuthology.orchestra.run.vm03.stdout: 2026-03-09T15:14:02.330 INFO:teuthology.orchestra.run.vm03.stdout:{"epoch":17,"fsid":"4abede80-1bca-11f1-bcd8-9b1da3c03ce5","created":"2026-03-09T15:12:42.233631+0000","modified":"2026-03-09T15:14:01.364622+0000","last_up_change":"2026-03-09T15:14:00.353477+0000","last_in_change":"2026-03-09T15:13:50.194917+0000","flags":"sortbitwise,recovery_deletes,purged_snapdirs,pglog_hardlimit","flags_num":5799936,"flags_set":["pglog_hardlimit","purged_snapdirs","recovery_deletes","sortbitwise"],"crush_version":8,"full_ratio":0.94999998807907104,"backfillfull_ratio":0.89999997615814209,"nearfull_ratio":0.85000002384185791,"cluster_snapshot":"","pool_max":1,"max_osd":3,"require_min_compat_client":"luminous","min_compat_client":"jewel","require_osd_release":"squid","allow_crimson":false,"pools":[{"pool":1,"pool_name":".mgr","create_time":"2026-03-09T15:14:00.523821+0000","flags":32769,"flags_names":"hashpspool,creating","type":1,"size":3,"min_size":2,"crush_rule":0,"peering_crush_bucket_count":0,"peering_crush_bucket_target":0,"peering_crush_bucket_barrier":0,"peering_crush_bucket_mandatory_member":2147483647,"is_stretch_pool":false,"object_hash":2,"pg_autoscale_mode":"off","pg_num":1,"pg_placement_num":1,"pg_placement_num_target":1,"pg_num_target":1,"pg_num_pending":1,"last_pg_merge_meta":{"source_pgid":"0.0","ready_epoch":0,"last_epoch_started":0,"last_epoch_clean":0,"source_version":"0'0","target_version":"0'0"},"last_change":"17","last_force_op_resend":"0","last_force_op_resend_prenautilus":"0","last_force_op_resend_preluminous":"0","auid":0,"snap_mode":"selfmanaged","snap_seq":0,"snap_epoch":0,"pool_snaps":[],"removed_snaps":"[]","quota_max_bytes":0,"quota_max_objects":0,"tiers":[],"tier_of":-1,"read_tier":-1,"write_tier":-1,"cache_mode":"none","target_max_bytes":0,"target_max_objects":0,"cache_target_dirty_ratio_micro":400000,"cache_target_dirty_high_ratio_micro":600000,"cache_target_full_ratio_micro":800000,"cache_min_flush_age":0,"cache_min_evict_age":0,"erasure_code_profile":"","hit_set_params":{"type":"none"},"hit_set_period":0,"hit_set_count":0,"use_gmt_hitset":true,"min_read_recency_for_promote":0,"min_write_recency_for_promote":0,"hit_set_grade_decay_rate":0,"hit_set_search_last_n":0,"grade_table":[],"stripe_width":0,"expected_num_objects":0,"fast_read":false,"options":{"pg_num_max":32,"pg_num_min":1},"application_metadata":{},"read_balance":{"score_type":"Fair distribution","score_acting":3,"score_stable":3,"optimal_score":1,"raw_score_acting":3,"raw_score_stable":3,"primary_affinity_weighted":1,"average_primary_affinity":1,"average_primary_affinity_weighted":1}}],"osds":[{"osd":0,"uuid":"913171bb-10e5-4020-a31f-9550e7c31d7b","up":1,"in":1,"weight":1,"primary_affinity":1,"last_clean_begin":0,"last_clean_end":0,"up_from":8,"up_thru":0,"down_at":0,"lost_at":0,"public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.103:6802","nonce":615876468},{"type":"v1","addr":"192.168.123.103:6803","nonce":615876468}]},"cluster_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.103:6804","nonce":615876468},{"type":"v1","addr":"192.168.123.103:6805","nonce":615876468}]},"heartbeat_back_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.103:6808","nonce":615876468},{"type":"v1","addr":"192.168.123.103:6809","nonce":615876468}]},"heartbeat_front_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.103:6806","nonce":615876468},{"type":"v1","addr":"192.168.123.103:6807","nonce":615876468}]},"public_addr":"192.168.123.103:6803/615876468","cluster_addr":"192.168.123.103:6805/615876468","heartbeat_back_addr":"192.168.123.103:6809/615876468","heartbeat_front_addr":"192.168.123.103:6807/615876468","state":["exists","up"]},{"osd":1,"uuid":"62ee7f4e-6ba5-481b-9cde-4c62baacf802","up":1,"in":1,"weight":1,"primary_affinity":1,"last_clean_begin":0,"last_clean_end":0,"up_from":12,"up_thru":0,"down_at":0,"lost_at":0,"public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.104:6800","nonce":1573535831},{"type":"v1","addr":"192.168.123.104:6801","nonce":1573535831}]},"cluster_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.104:6802","nonce":1573535831},{"type":"v1","addr":"192.168.123.104:6803","nonce":1573535831}]},"heartbeat_back_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.104:6806","nonce":1573535831},{"type":"v1","addr":"192.168.123.104:6807","nonce":1573535831}]},"heartbeat_front_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.104:6804","nonce":1573535831},{"type":"v1","addr":"192.168.123.104:6805","nonce":1573535831}]},"public_addr":"192.168.123.104:6801/1573535831","cluster_addr":"192.168.123.104:6803/1573535831","heartbeat_back_addr":"192.168.123.104:6807/1573535831","heartbeat_front_addr":"192.168.123.104:6805/1573535831","state":["exists","up"]},{"osd":2,"uuid":"d53d63dc-a0c2-402c-8257-919cdd6b513c","up":1,"in":1,"weight":1,"primary_affinity":1,"last_clean_begin":0,"last_clean_end":0,"up_from":16,"up_thru":0,"down_at":0,"lost_at":0,"public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.106:6800","nonce":3722863356},{"type":"v1","addr":"192.168.123.106:6801","nonce":3722863356}]},"cluster_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.106:6802","nonce":3722863356},{"type":"v1","addr":"192.168.123.106:6803","nonce":3722863356}]},"heartbeat_back_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.106:6806","nonce":3722863356},{"type":"v1","addr":"192.168.123.106:6807","nonce":3722863356}]},"heartbeat_front_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.106:6804","nonce":3722863356},{"type":"v1","addr":"192.168.123.106:6805","nonce":3722863356}]},"public_addr":"192.168.123.106:6801/3722863356","cluster_addr":"192.168.123.106:6803/3722863356","heartbeat_back_addr":"192.168.123.106:6807/3722863356","heartbeat_front_addr":"192.168.123.106:6805/3722863356","state":["exists","up"]}],"osd_xinfo":[{"osd":0,"down_stamp":"0.000000","laggy_probability":0,"laggy_interval":0,"features":4540701547738038271,"old_weight":0,"last_purged_snaps_scrub":"2026-03-09T15:13:40.595321+0000","dead_epoch":0},{"osd":1,"down_stamp":"0.000000","laggy_probability":0,"laggy_interval":0,"features":4540701547738038271,"old_weight":0,"last_purged_snaps_scrub":"2026-03-09T15:13:49.850182+0000","dead_epoch":0},{"osd":2,"down_stamp":"0.000000","laggy_probability":0,"laggy_interval":0,"features":4540701547738038271,"old_weight":0,"last_purged_snaps_scrub":"2026-03-09T15:13:58.860702+0000","dead_epoch":0}],"pg_upmap":[],"pg_upmap_items":[],"pg_upmap_primaries":[],"pg_temp":[],"primary_temp":[],"blocklist":{"192.168.123.103:6801/3510874031":"2026-03-10T15:13:02.478314+0000","192.168.123.103:6800/3510874031":"2026-03-10T15:13:02.478314+0000","192.168.123.103:0/1665269914":"2026-03-10T15:13:02.478314+0000","192.168.123.103:0/4246730401":"2026-03-10T15:13:02.478314+0000","192.168.123.103:0/983012295":"2026-03-10T15:12:52.735172+0000","192.168.123.103:0/3380842377":"2026-03-10T15:13:02.478314+0000","192.168.123.103:0/1826640327":"2026-03-10T15:12:52.735172+0000","192.168.123.103:6800/2441273787":"2026-03-10T15:12:52.735172+0000","192.168.123.103:6801/2441273787":"2026-03-10T15:12:52.735172+0000","192.168.123.103:0/4187493480":"2026-03-10T15:12:52.735172+0000"},"range_blocklist":{},"erasure_code_profiles":{"default":{"crush-failure-domain":"osd","k":"2","m":"1","plugin":"jerasure","technique":"reed_sol_van"}},"removed_snaps_queue":[],"new_removed_snaps":[],"new_purged_snaps":[],"crush_node_flags":{},"device_class_flags":{},"stretch_mode":{"stretch_mode_enabled":false,"stretch_bucket_count":0,"degraded_stretch_mode":0,"recovering_stretch_mode":0,"stretch_mode_bucket":0}} 2026-03-09T15:14:02.400 DEBUG:teuthology.orchestra.run.vm03:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell --fsid 4abede80-1bca-11f1-bcd8-9b1da3c03ce5 -- ceph tell osd.0 flush_pg_stats 2026-03-09T15:14:02.400 DEBUG:teuthology.orchestra.run.vm03:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell --fsid 4abede80-1bca-11f1-bcd8-9b1da3c03ce5 -- ceph tell osd.1 flush_pg_stats 2026-03-09T15:14:02.400 DEBUG:teuthology.orchestra.run.vm03:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell --fsid 4abede80-1bca-11f1-bcd8-9b1da3c03ce5 -- ceph tell osd.2 flush_pg_stats 2026-03-09T15:14:02.676 INFO:teuthology.orchestra.run.vm03.stderr:Inferring config /var/lib/ceph/4abede80-1bca-11f1-bcd8-9b1da3c03ce5/mon.a/config 2026-03-09T15:14:02.710 INFO:teuthology.orchestra.run.vm03.stderr:Inferring config /var/lib/ceph/4abede80-1bca-11f1-bcd8-9b1da3c03ce5/mon.a/config 2026-03-09T15:14:02.712 INFO:journalctl@ceph.osd.0.vm03.stdout:Mar 09 15:14:02 vm03 sudo[65356]: ceph : PWD=/ ; USER=root ; COMMAND=/usr/sbin/smartctl -x --json=o /dev/vde 2026-03-09T15:14:02.712 INFO:journalctl@ceph.osd.0.vm03.stdout:Mar 09 15:14:02 vm03 sudo[65356]: pam_systemd(sudo:session): Failed to connect to system bus: No such file or directory 2026-03-09T15:14:02.712 INFO:journalctl@ceph.osd.0.vm03.stdout:Mar 09 15:14:02 vm03 sudo[65356]: pam_unix(sudo:session): session opened for user root(uid=0) by (uid=167) 2026-03-09T15:14:02.712 INFO:journalctl@ceph.osd.0.vm03.stdout:Mar 09 15:14:02 vm03 sudo[65356]: pam_unix(sudo:session): session closed for user root 2026-03-09T15:14:02.713 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:14:02 vm03 ceph-mon[51762]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' cmd='[{"prefix": "osd pool create", "format": "json", "pool": ".mgr", "pg_num": 1, "pg_num_min": 1, "pg_num_max": 32, "yes_i_really_mean_it": true}]': finished 2026-03-09T15:14:02.713 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:14:02 vm03 ceph-mon[51762]: osdmap e17: 3 total, 3 up, 3 in 2026-03-09T15:14:02.713 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:14:02 vm03 ceph-mon[51762]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' cmd=[{"prefix": "osd pool application enable", "format": "json", "pool": ".mgr", "app": "mgr", "yes_i_really_mean_it": true}]: dispatch 2026-03-09T15:14:02.713 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:14:02 vm03 ceph-mon[51762]: from='client.? 192.168.123.103:0/3131817460' entity='client.admin' cmd=[{"prefix": "mgr dump", "format": "json"}]: dispatch 2026-03-09T15:14:02.713 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:14:02 vm03 ceph-mon[51762]: from='client.? 192.168.123.103:0/1654457159' entity='client.admin' cmd=[{"prefix": "osd dump", "format": "json"}]: dispatch 2026-03-09T15:14:02.713 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:14:02 vm03 ceph-mon[51762]: from='client.? 192.168.123.103:0/2317104000' entity='client.admin' cmd=[{"prefix": "osd dump", "format": "json"}]: dispatch 2026-03-09T15:14:02.713 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:14:02 vm03 sudo[65370]: ceph : PWD=/ ; USER=root ; COMMAND=/usr/sbin/smartctl -x --json=o /dev/vda 2026-03-09T15:14:02.713 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:14:02 vm03 sudo[65370]: pam_systemd(sudo:session): Failed to connect to system bus: No such file or directory 2026-03-09T15:14:02.713 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:14:02 vm03 sudo[65370]: pam_unix(sudo:session): session opened for user root(uid=0) by (uid=167) 2026-03-09T15:14:02.713 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:14:02 vm03 sudo[65370]: pam_unix(sudo:session): session closed for user root 2026-03-09T15:14:02.751 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 15:14:02 vm04 ceph-mon[53998]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' cmd='[{"prefix": "osd pool create", "format": "json", "pool": ".mgr", "pg_num": 1, "pg_num_min": 1, "pg_num_max": 32, "yes_i_really_mean_it": true}]': finished 2026-03-09T15:14:02.751 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 15:14:02 vm04 ceph-mon[53998]: osdmap e17: 3 total, 3 up, 3 in 2026-03-09T15:14:02.751 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 15:14:02 vm04 ceph-mon[53998]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' cmd=[{"prefix": "osd pool application enable", "format": "json", "pool": ".mgr", "app": "mgr", "yes_i_really_mean_it": true}]: dispatch 2026-03-09T15:14:02.751 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 15:14:02 vm04 ceph-mon[53998]: from='client.? 192.168.123.103:0/3131817460' entity='client.admin' cmd=[{"prefix": "mgr dump", "format": "json"}]: dispatch 2026-03-09T15:14:02.751 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 15:14:02 vm04 ceph-mon[53998]: from='client.? 192.168.123.103:0/1654457159' entity='client.admin' cmd=[{"prefix": "osd dump", "format": "json"}]: dispatch 2026-03-09T15:14:02.751 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 15:14:02 vm04 ceph-mon[53998]: from='client.? 192.168.123.103:0/2317104000' entity='client.admin' cmd=[{"prefix": "osd dump", "format": "json"}]: dispatch 2026-03-09T15:14:02.751 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 15:14:02 vm04 sudo[60474]: ceph : PWD=/ ; USER=root ; COMMAND=/usr/sbin/smartctl -x --json=o /dev/vda 2026-03-09T15:14:02.751 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 15:14:02 vm04 sudo[60474]: pam_systemd(sudo:session): Failed to connect to system bus: No such file or directory 2026-03-09T15:14:02.751 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 15:14:02 vm04 sudo[60474]: pam_unix(sudo:session): session opened for user root(uid=0) by (uid=167) 2026-03-09T15:14:02.751 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 15:14:02 vm04 sudo[60474]: pam_unix(sudo:session): session closed for user root 2026-03-09T15:14:02.751 INFO:journalctl@ceph.osd.1.vm04.stdout:Mar 09 15:14:02 vm04 sudo[60470]: ceph : PWD=/ ; USER=root ; COMMAND=/usr/sbin/smartctl -x --json=o /dev/vde 2026-03-09T15:14:02.751 INFO:journalctl@ceph.osd.1.vm04.stdout:Mar 09 15:14:02 vm04 sudo[60470]: pam_systemd(sudo:session): Failed to connect to system bus: No such file or directory 2026-03-09T15:14:02.751 INFO:journalctl@ceph.osd.1.vm04.stdout:Mar 09 15:14:02 vm04 sudo[60470]: pam_unix(sudo:session): session opened for user root(uid=0) by (uid=167) 2026-03-09T15:14:02.751 INFO:journalctl@ceph.osd.1.vm04.stdout:Mar 09 15:14:02 vm04 sudo[60470]: pam_unix(sudo:session): session closed for user root 2026-03-09T15:14:02.767 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 09 15:14:02 vm06 ceph-mon[56506]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' cmd='[{"prefix": "osd pool create", "format": "json", "pool": ".mgr", "pg_num": 1, "pg_num_min": 1, "pg_num_max": 32, "yes_i_really_mean_it": true}]': finished 2026-03-09T15:14:02.767 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 09 15:14:02 vm06 ceph-mon[56506]: osdmap e17: 3 total, 3 up, 3 in 2026-03-09T15:14:02.767 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 09 15:14:02 vm06 ceph-mon[56506]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' cmd=[{"prefix": "osd pool application enable", "format": "json", "pool": ".mgr", "app": "mgr", "yes_i_really_mean_it": true}]: dispatch 2026-03-09T15:14:02.767 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 09 15:14:02 vm06 ceph-mon[56506]: from='client.? 192.168.123.103:0/3131817460' entity='client.admin' cmd=[{"prefix": "mgr dump", "format": "json"}]: dispatch 2026-03-09T15:14:02.767 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 09 15:14:02 vm06 ceph-mon[56506]: from='client.? 192.168.123.103:0/1654457159' entity='client.admin' cmd=[{"prefix": "osd dump", "format": "json"}]: dispatch 2026-03-09T15:14:02.767 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 09 15:14:02 vm06 ceph-mon[56506]: from='client.? 192.168.123.103:0/2317104000' entity='client.admin' cmd=[{"prefix": "osd dump", "format": "json"}]: dispatch 2026-03-09T15:14:02.767 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 09 15:14:02 vm06 sudo[62629]: ceph : PWD=/ ; USER=root ; COMMAND=/usr/sbin/smartctl -x --json=o /dev/vda 2026-03-09T15:14:02.767 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 09 15:14:02 vm06 sudo[62629]: pam_systemd(sudo:session): Failed to connect to system bus: No such file or directory 2026-03-09T15:14:02.767 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 09 15:14:02 vm06 sudo[62629]: pam_unix(sudo:session): session opened for user root(uid=0) by (uid=167) 2026-03-09T15:14:02.767 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 09 15:14:02 vm06 sudo[62629]: pam_unix(sudo:session): session closed for user root 2026-03-09T15:14:02.767 INFO:journalctl@ceph.osd.2.vm06.stdout:Mar 09 15:14:02 vm06 sudo[62625]: ceph : PWD=/ ; USER=root ; COMMAND=/usr/sbin/smartctl -x --json=o /dev/vde 2026-03-09T15:14:02.767 INFO:journalctl@ceph.osd.2.vm06.stdout:Mar 09 15:14:02 vm06 sudo[62625]: pam_systemd(sudo:session): Failed to connect to system bus: No such file or directory 2026-03-09T15:14:02.767 INFO:journalctl@ceph.osd.2.vm06.stdout:Mar 09 15:14:02 vm06 sudo[62625]: pam_unix(sudo:session): session opened for user root(uid=0) by (uid=167) 2026-03-09T15:14:02.767 INFO:journalctl@ceph.osd.2.vm06.stdout:Mar 09 15:14:02 vm06 sudo[62625]: pam_unix(sudo:session): session closed for user root 2026-03-09T15:14:02.775 INFO:teuthology.orchestra.run.vm03.stderr:Inferring config /var/lib/ceph/4abede80-1bca-11f1-bcd8-9b1da3c03ce5/mon.a/config 2026-03-09T15:14:02.925 INFO:teuthology.orchestra.run.vm03.stdout:51539607556 2026-03-09T15:14:02.925 DEBUG:teuthology.orchestra.run.vm03:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell --fsid 4abede80-1bca-11f1-bcd8-9b1da3c03ce5 -- ceph osd last-stat-seq osd.1 2026-03-09T15:14:03.003 INFO:teuthology.orchestra.run.vm03.stdout:34359738373 2026-03-09T15:14:03.004 DEBUG:teuthology.orchestra.run.vm03:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell --fsid 4abede80-1bca-11f1-bcd8-9b1da3c03ce5 -- ceph osd last-stat-seq osd.0 2026-03-09T15:14:03.070 INFO:teuthology.orchestra.run.vm03.stdout:68719476738 2026-03-09T15:14:03.070 DEBUG:teuthology.orchestra.run.vm03:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell --fsid 4abede80-1bca-11f1-bcd8-9b1da3c03ce5 -- ceph osd last-stat-seq osd.2 2026-03-09T15:14:03.216 INFO:teuthology.orchestra.run.vm03.stderr:Inferring config /var/lib/ceph/4abede80-1bca-11f1-bcd8-9b1da3c03ce5/mon.a/config 2026-03-09T15:14:03.279 INFO:teuthology.orchestra.run.vm03.stderr:Inferring config /var/lib/ceph/4abede80-1bca-11f1-bcd8-9b1da3c03ce5/mon.a/config 2026-03-09T15:14:03.405 INFO:teuthology.orchestra.run.vm03.stderr:Inferring config /var/lib/ceph/4abede80-1bca-11f1-bcd8-9b1da3c03ce5/mon.a/config 2026-03-09T15:14:03.534 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:14:03 vm03 ceph-mon[51762]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' cmd='[{"prefix": "osd pool application enable", "format": "json", "pool": ".mgr", "app": "mgr", "yes_i_really_mean_it": true}]': finished 2026-03-09T15:14:03.534 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:14:03 vm03 ceph-mon[51762]: osdmap e18: 3 total, 3 up, 3 in 2026-03-09T15:14:03.534 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:14:03 vm03 ceph-mon[51762]: pgmap v38: 1 pgs: 1 unknown; 0 B data, 79 MiB used, 60 GiB / 60 GiB avail 2026-03-09T15:14:03.534 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:14:03 vm03 ceph-mon[51762]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "a"}]: dispatch 2026-03-09T15:14:03.534 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:14:03 vm03 ceph-mon[51762]: from='admin socket' entity='admin socket' cmd='smart' args=[json]: dispatch 2026-03-09T15:14:03.534 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:14:03 vm03 ceph-mon[51762]: from='admin socket' entity='admin socket' cmd=smart args=[json]: finished 2026-03-09T15:14:03.534 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:14:03 vm03 ceph-mon[51762]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-09T15:14:03.534 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:14:03 vm03 ceph-mon[51762]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "c"}]: dispatch 2026-03-09T15:14:03.534 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:14:03 vm03 ceph-mon[51762]: from='admin socket' entity='admin socket' cmd='smart' args=[json]: dispatch 2026-03-09T15:14:03.534 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:14:03 vm03 ceph-mon[51762]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "a"}]: dispatch 2026-03-09T15:14:03.534 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:14:03 vm03 ceph-mon[51762]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-09T15:14:03.534 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:14:03 vm03 ceph-mon[51762]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "c"}]: dispatch 2026-03-09T15:14:03.534 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:14:03 vm03 ceph-mon[51762]: from='admin socket' entity='admin socket' cmd=smart args=[json]: finished 2026-03-09T15:14:03.534 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:14:03 vm03 ceph-mon[51762]: from='admin socket' entity='admin socket' cmd='smart' args=[json]: dispatch 2026-03-09T15:14:03.534 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:14:03 vm03 ceph-mon[51762]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "a"}]: dispatch 2026-03-09T15:14:03.534 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:14:03 vm03 ceph-mon[51762]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-09T15:14:03.535 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:14:03 vm03 ceph-mon[51762]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "c"}]: dispatch 2026-03-09T15:14:03.535 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:14:03 vm03 ceph-mon[51762]: from='admin socket' entity='admin socket' cmd=smart args=[json]: finished 2026-03-09T15:14:03.596 INFO:teuthology.orchestra.run.vm03.stdout:51539607555 2026-03-09T15:14:03.644 INFO:tasks.cephadm.ceph_manager.ceph:need seq 51539607556 got 51539607555 for osd.1 2026-03-09T15:14:03.655 INFO:teuthology.orchestra.run.vm03.stdout:68719476737 2026-03-09T15:14:03.728 INFO:teuthology.orchestra.run.vm03.stdout:34359738372 2026-03-09T15:14:03.732 INFO:tasks.cephadm.ceph_manager.ceph:need seq 68719476738 got 68719476737 for osd.2 2026-03-09T15:14:03.751 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 15:14:03 vm04 ceph-mon[53998]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' cmd='[{"prefix": "osd pool application enable", "format": "json", "pool": ".mgr", "app": "mgr", "yes_i_really_mean_it": true}]': finished 2026-03-09T15:14:03.751 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 15:14:03 vm04 ceph-mon[53998]: osdmap e18: 3 total, 3 up, 3 in 2026-03-09T15:14:03.751 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 15:14:03 vm04 ceph-mon[53998]: pgmap v38: 1 pgs: 1 unknown; 0 B data, 79 MiB used, 60 GiB / 60 GiB avail 2026-03-09T15:14:03.751 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 15:14:03 vm04 ceph-mon[53998]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "a"}]: dispatch 2026-03-09T15:14:03.751 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 15:14:03 vm04 ceph-mon[53998]: from='admin socket' entity='admin socket' cmd='smart' args=[json]: dispatch 2026-03-09T15:14:03.751 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 15:14:03 vm04 ceph-mon[53998]: from='admin socket' entity='admin socket' cmd=smart args=[json]: finished 2026-03-09T15:14:03.752 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 15:14:03 vm04 ceph-mon[53998]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-09T15:14:03.752 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 15:14:03 vm04 ceph-mon[53998]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "c"}]: dispatch 2026-03-09T15:14:03.752 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 15:14:03 vm04 ceph-mon[53998]: from='admin socket' entity='admin socket' cmd='smart' args=[json]: dispatch 2026-03-09T15:14:03.752 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 15:14:03 vm04 ceph-mon[53998]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "a"}]: dispatch 2026-03-09T15:14:03.752 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 15:14:03 vm04 ceph-mon[53998]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-09T15:14:03.752 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 15:14:03 vm04 ceph-mon[53998]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "c"}]: dispatch 2026-03-09T15:14:03.752 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 15:14:03 vm04 ceph-mon[53998]: from='admin socket' entity='admin socket' cmd=smart args=[json]: finished 2026-03-09T15:14:03.752 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 15:14:03 vm04 ceph-mon[53998]: from='admin socket' entity='admin socket' cmd='smart' args=[json]: dispatch 2026-03-09T15:14:03.752 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 15:14:03 vm04 ceph-mon[53998]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "a"}]: dispatch 2026-03-09T15:14:03.752 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 15:14:03 vm04 ceph-mon[53998]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-09T15:14:03.752 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 15:14:03 vm04 ceph-mon[53998]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "c"}]: dispatch 2026-03-09T15:14:03.752 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 15:14:03 vm04 ceph-mon[53998]: from='admin socket' entity='admin socket' cmd=smart args=[json]: finished 2026-03-09T15:14:03.767 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 09 15:14:03 vm06 ceph-mon[56506]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' cmd='[{"prefix": "osd pool application enable", "format": "json", "pool": ".mgr", "app": "mgr", "yes_i_really_mean_it": true}]': finished 2026-03-09T15:14:03.767 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 09 15:14:03 vm06 ceph-mon[56506]: osdmap e18: 3 total, 3 up, 3 in 2026-03-09T15:14:03.767 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 09 15:14:03 vm06 ceph-mon[56506]: pgmap v38: 1 pgs: 1 unknown; 0 B data, 79 MiB used, 60 GiB / 60 GiB avail 2026-03-09T15:14:03.767 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 09 15:14:03 vm06 ceph-mon[56506]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "a"}]: dispatch 2026-03-09T15:14:03.767 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 09 15:14:03 vm06 ceph-mon[56506]: from='admin socket' entity='admin socket' cmd='smart' args=[json]: dispatch 2026-03-09T15:14:03.767 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 09 15:14:03 vm06 ceph-mon[56506]: from='admin socket' entity='admin socket' cmd=smart args=[json]: finished 2026-03-09T15:14:03.767 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 09 15:14:03 vm06 ceph-mon[56506]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-09T15:14:03.767 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 09 15:14:03 vm06 ceph-mon[56506]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "c"}]: dispatch 2026-03-09T15:14:03.767 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 09 15:14:03 vm06 ceph-mon[56506]: from='admin socket' entity='admin socket' cmd='smart' args=[json]: dispatch 2026-03-09T15:14:03.767 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 09 15:14:03 vm06 ceph-mon[56506]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "a"}]: dispatch 2026-03-09T15:14:03.767 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 09 15:14:03 vm06 ceph-mon[56506]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-09T15:14:03.767 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 09 15:14:03 vm06 ceph-mon[56506]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "c"}]: dispatch 2026-03-09T15:14:03.767 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 09 15:14:03 vm06 ceph-mon[56506]: from='admin socket' entity='admin socket' cmd=smart args=[json]: finished 2026-03-09T15:14:03.767 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 09 15:14:03 vm06 ceph-mon[56506]: from='admin socket' entity='admin socket' cmd='smart' args=[json]: dispatch 2026-03-09T15:14:03.767 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 09 15:14:03 vm06 ceph-mon[56506]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "a"}]: dispatch 2026-03-09T15:14:03.767 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 09 15:14:03 vm06 ceph-mon[56506]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-09T15:14:03.767 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 09 15:14:03 vm06 ceph-mon[56506]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "c"}]: dispatch 2026-03-09T15:14:03.767 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 09 15:14:03 vm06 ceph-mon[56506]: from='admin socket' entity='admin socket' cmd=smart args=[json]: finished 2026-03-09T15:14:03.779 INFO:tasks.cephadm.ceph_manager.ceph:need seq 34359738373 got 34359738372 for osd.0 2026-03-09T15:14:04.645 DEBUG:teuthology.orchestra.run.vm03:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell --fsid 4abede80-1bca-11f1-bcd8-9b1da3c03ce5 -- ceph osd last-stat-seq osd.1 2026-03-09T15:14:04.668 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:14:04 vm03 ceph-mon[51762]: osdmap e19: 3 total, 3 up, 3 in 2026-03-09T15:14:04.668 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:14:04 vm03 ceph-mon[51762]: mgrmap e14: a(active, since 60s), standbys: b 2026-03-09T15:14:04.668 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:14:04 vm03 ceph-mon[51762]: from='client.? 192.168.123.103:0/1857526896' entity='client.admin' cmd=[{"prefix": "osd last-stat-seq", "id": 1}]: dispatch 2026-03-09T15:14:04.668 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:14:04 vm03 ceph-mon[51762]: from='client.? 192.168.123.103:0/2826189671' entity='client.admin' cmd=[{"prefix": "osd last-stat-seq", "id": 2}]: dispatch 2026-03-09T15:14:04.668 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:14:04 vm03 ceph-mon[51762]: from='client.? 192.168.123.103:0/132737357' entity='client.admin' cmd=[{"prefix": "osd last-stat-seq", "id": 0}]: dispatch 2026-03-09T15:14:04.733 DEBUG:teuthology.orchestra.run.vm03:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell --fsid 4abede80-1bca-11f1-bcd8-9b1da3c03ce5 -- ceph osd last-stat-seq osd.2 2026-03-09T15:14:04.751 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 15:14:04 vm04 ceph-mon[53998]: osdmap e19: 3 total, 3 up, 3 in 2026-03-09T15:14:04.751 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 15:14:04 vm04 ceph-mon[53998]: mgrmap e14: a(active, since 60s), standbys: b 2026-03-09T15:14:04.751 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 15:14:04 vm04 ceph-mon[53998]: from='client.? 192.168.123.103:0/1857526896' entity='client.admin' cmd=[{"prefix": "osd last-stat-seq", "id": 1}]: dispatch 2026-03-09T15:14:04.751 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 15:14:04 vm04 ceph-mon[53998]: from='client.? 192.168.123.103:0/2826189671' entity='client.admin' cmd=[{"prefix": "osd last-stat-seq", "id": 2}]: dispatch 2026-03-09T15:14:04.751 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 15:14:04 vm04 ceph-mon[53998]: from='client.? 192.168.123.103:0/132737357' entity='client.admin' cmd=[{"prefix": "osd last-stat-seq", "id": 0}]: dispatch 2026-03-09T15:14:04.767 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 09 15:14:04 vm06 ceph-mon[56506]: osdmap e19: 3 total, 3 up, 3 in 2026-03-09T15:14:04.767 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 09 15:14:04 vm06 ceph-mon[56506]: mgrmap e14: a(active, since 60s), standbys: b 2026-03-09T15:14:04.767 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 09 15:14:04 vm06 ceph-mon[56506]: from='client.? 192.168.123.103:0/1857526896' entity='client.admin' cmd=[{"prefix": "osd last-stat-seq", "id": 1}]: dispatch 2026-03-09T15:14:04.767 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 09 15:14:04 vm06 ceph-mon[56506]: from='client.? 192.168.123.103:0/2826189671' entity='client.admin' cmd=[{"prefix": "osd last-stat-seq", "id": 2}]: dispatch 2026-03-09T15:14:04.767 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 09 15:14:04 vm06 ceph-mon[56506]: from='client.? 192.168.123.103:0/132737357' entity='client.admin' cmd=[{"prefix": "osd last-stat-seq", "id": 0}]: dispatch 2026-03-09T15:14:04.780 DEBUG:teuthology.orchestra.run.vm03:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell --fsid 4abede80-1bca-11f1-bcd8-9b1da3c03ce5 -- ceph osd last-stat-seq osd.0 2026-03-09T15:14:04.818 INFO:teuthology.orchestra.run.vm03.stderr:Inferring config /var/lib/ceph/4abede80-1bca-11f1-bcd8-9b1da3c03ce5/mon.a/config 2026-03-09T15:14:05.003 INFO:teuthology.orchestra.run.vm03.stderr:Inferring config /var/lib/ceph/4abede80-1bca-11f1-bcd8-9b1da3c03ce5/mon.a/config 2026-03-09T15:14:05.114 INFO:teuthology.orchestra.run.vm03.stdout:51539607555 2026-03-09T15:14:05.127 INFO:teuthology.orchestra.run.vm03.stderr:Inferring config /var/lib/ceph/4abede80-1bca-11f1-bcd8-9b1da3c03ce5/mon.a/config 2026-03-09T15:14:05.200 INFO:tasks.cephadm.ceph_manager.ceph:need seq 51539607556 got 51539607555 for osd.1 2026-03-09T15:14:05.279 INFO:teuthology.orchestra.run.vm03.stdout:68719476737 2026-03-09T15:14:05.349 INFO:tasks.cephadm.ceph_manager.ceph:need seq 68719476738 got 68719476737 for osd.2 2026-03-09T15:14:05.420 INFO:teuthology.orchestra.run.vm03.stdout:34359738374 2026-03-09T15:14:05.466 INFO:tasks.cephadm.ceph_manager.ceph:need seq 34359738373 got 34359738374 for osd.0 2026-03-09T15:14:05.466 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:14:05 vm03 ceph-mon[51762]: pgmap v40: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-09T15:14:05.466 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:14:05 vm03 ceph-mon[51762]: from='client.? 192.168.123.103:0/2577423008' entity='client.admin' cmd=[{"prefix": "osd last-stat-seq", "id": 1}]: dispatch 2026-03-09T15:14:05.466 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:14:05 vm03 ceph-mon[51762]: from='client.? 192.168.123.103:0/2819421486' entity='client.admin' cmd=[{"prefix": "osd last-stat-seq", "id": 2}]: dispatch 2026-03-09T15:14:05.466 DEBUG:teuthology.parallel:result is None 2026-03-09T15:14:05.751 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 15:14:05 vm04 ceph-mon[53998]: pgmap v40: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-09T15:14:05.751 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 15:14:05 vm04 ceph-mon[53998]: from='client.? 192.168.123.103:0/2577423008' entity='client.admin' cmd=[{"prefix": "osd last-stat-seq", "id": 1}]: dispatch 2026-03-09T15:14:05.751 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 15:14:05 vm04 ceph-mon[53998]: from='client.? 192.168.123.103:0/2819421486' entity='client.admin' cmd=[{"prefix": "osd last-stat-seq", "id": 2}]: dispatch 2026-03-09T15:14:05.767 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 09 15:14:05 vm06 ceph-mon[56506]: pgmap v40: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-09T15:14:05.767 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 09 15:14:05 vm06 ceph-mon[56506]: from='client.? 192.168.123.103:0/2577423008' entity='client.admin' cmd=[{"prefix": "osd last-stat-seq", "id": 1}]: dispatch 2026-03-09T15:14:05.767 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 09 15:14:05 vm06 ceph-mon[56506]: from='client.? 192.168.123.103:0/2819421486' entity='client.admin' cmd=[{"prefix": "osd last-stat-seq", "id": 2}]: dispatch 2026-03-09T15:14:06.201 DEBUG:teuthology.orchestra.run.vm03:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell --fsid 4abede80-1bca-11f1-bcd8-9b1da3c03ce5 -- ceph osd last-stat-seq osd.1 2026-03-09T15:14:06.349 DEBUG:teuthology.orchestra.run.vm03:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell --fsid 4abede80-1bca-11f1-bcd8-9b1da3c03ce5 -- ceph osd last-stat-seq osd.2 2026-03-09T15:14:06.367 INFO:teuthology.orchestra.run.vm03.stderr:Inferring config /var/lib/ceph/4abede80-1bca-11f1-bcd8-9b1da3c03ce5/mon.a/config 2026-03-09T15:14:06.485 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:14:06 vm03 ceph-mon[51762]: from='client.? 192.168.123.103:0/3357546573' entity='client.admin' cmd=[{"prefix": "osd last-stat-seq", "id": 0}]: dispatch 2026-03-09T15:14:06.562 INFO:teuthology.orchestra.run.vm03.stderr:Inferring config /var/lib/ceph/4abede80-1bca-11f1-bcd8-9b1da3c03ce5/mon.a/config 2026-03-09T15:14:06.612 INFO:teuthology.orchestra.run.vm03.stdout:51539607556 2026-03-09T15:14:06.691 INFO:tasks.cephadm.ceph_manager.ceph:need seq 51539607556 got 51539607556 for osd.1 2026-03-09T15:14:06.692 DEBUG:teuthology.parallel:result is None 2026-03-09T15:14:06.751 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 15:14:06 vm04 ceph-mon[53998]: from='client.? 192.168.123.103:0/3357546573' entity='client.admin' cmd=[{"prefix": "osd last-stat-seq", "id": 0}]: dispatch 2026-03-09T15:14:06.767 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 09 15:14:06 vm06 ceph-mon[56506]: from='client.? 192.168.123.103:0/3357546573' entity='client.admin' cmd=[{"prefix": "osd last-stat-seq", "id": 0}]: dispatch 2026-03-09T15:14:06.801 INFO:teuthology.orchestra.run.vm03.stdout:68719476738 2026-03-09T15:14:06.845 INFO:tasks.cephadm.ceph_manager.ceph:need seq 68719476738 got 68719476738 for osd.2 2026-03-09T15:14:06.845 DEBUG:teuthology.parallel:result is None 2026-03-09T15:14:06.845 INFO:tasks.cephadm.ceph_manager.ceph:waiting for clean 2026-03-09T15:14:06.845 DEBUG:teuthology.orchestra.run.vm03:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell --fsid 4abede80-1bca-11f1-bcd8-9b1da3c03ce5 -- ceph pg dump --format=json 2026-03-09T15:14:07.009 INFO:teuthology.orchestra.run.vm03.stderr:Inferring config /var/lib/ceph/4abede80-1bca-11f1-bcd8-9b1da3c03ce5/mon.a/config 2026-03-09T15:14:07.226 INFO:teuthology.orchestra.run.vm03.stdout: 2026-03-09T15:14:07.226 INFO:teuthology.orchestra.run.vm03.stderr:dumped all 2026-03-09T15:14:07.290 INFO:teuthology.orchestra.run.vm03.stdout:{"pg_ready":true,"pg_map":{"version":41,"stamp":"2026-03-09T15:14:06.493749+0000","last_osdmap_epoch":0,"last_pg_scan":0,"pg_stats_sum":{"stat_sum":{"num_bytes":459280,"num_objects":2,"num_object_clones":0,"num_object_copies":6,"num_objects_missing_on_primary":0,"num_objects_missing":0,"num_objects_degraded":0,"num_objects_misplaced":0,"num_objects_unfound":0,"num_objects_dirty":2,"num_whiteouts":0,"num_read":46,"num_read_kb":37,"num_write":57,"num_write_kb":584,"num_scrub_errors":0,"num_shallow_scrub_errors":0,"num_deep_scrub_errors":0,"num_objects_recovered":0,"num_bytes_recovered":0,"num_keys_recovered":0,"num_objects_omap":0,"num_objects_hit_set_archive":0,"num_bytes_hit_set_archive":0,"num_flush":0,"num_flush_kb":0,"num_evict":0,"num_evict_kb":0,"num_promote":0,"num_flush_mode_high":0,"num_flush_mode_low":0,"num_evict_mode_some":0,"num_evict_mode_full":0,"num_objects_pinned":0,"num_legacy_snapsets":0,"num_large_omap_objects":0,"num_objects_manifest":0,"num_omap_bytes":0,"num_omap_keys":0,"num_objects_repaired":0},"store_stats":{"total":0,"available":0,"internally_reserved":0,"allocated":0,"data_stored":0,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":0},"log_size":32,"ondisk_log_size":32,"up":3,"acting":3,"num_store_stats":0},"osd_stats_sum":{"up_from":0,"seq":0,"num_pgs":3,"num_osds":3,"num_per_pool_osds":3,"num_per_pool_omap_osds":3,"kb":62902272,"kb_used":82484,"kb_used_data":1700,"kb_used_omap":4,"kb_used_meta":80443,"kb_avail":62819788,"statfs":{"total":64411926528,"available":64327462912,"internally_reserved":0,"allocated":1740800,"data_stored":1394349,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":4769,"internal_metadata":82373983},"hb_peers":[],"snap_trim_queue_len":0,"num_snap_trimming":0,"num_shards_repaired":0,"op_queue_age_hist":{"histogram":[],"upper_bound":1},"perf_stat":{"commit_latency_ms":0,"apply_latency_ms":0,"commit_latency_ns":0,"apply_latency_ns":0},"alerts":[],"network_ping_times":[]},"pg_stats_delta":{"stat_sum":{"num_bytes":0,"num_objects":0,"num_object_clones":0,"num_object_copies":0,"num_objects_missing_on_primary":0,"num_objects_missing":0,"num_objects_degraded":0,"num_objects_misplaced":0,"num_objects_unfound":0,"num_objects_dirty":0,"num_whiteouts":0,"num_read":0,"num_read_kb":0,"num_write":0,"num_write_kb":0,"num_scrub_errors":0,"num_shallow_scrub_errors":0,"num_deep_scrub_errors":0,"num_objects_recovered":0,"num_bytes_recovered":0,"num_keys_recovered":0,"num_objects_omap":0,"num_objects_hit_set_archive":0,"num_bytes_hit_set_archive":0,"num_flush":0,"num_flush_kb":0,"num_evict":0,"num_evict_kb":0,"num_promote":0,"num_flush_mode_high":0,"num_flush_mode_low":0,"num_evict_mode_some":0,"num_evict_mode_full":0,"num_objects_pinned":0,"num_legacy_snapsets":0,"num_large_omap_objects":0,"num_objects_manifest":0,"num_omap_bytes":0,"num_omap_keys":0,"num_objects_repaired":0},"store_stats":{"total":0,"available":0,"internally_reserved":0,"allocated":0,"data_stored":0,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":0},"log_size":0,"ondisk_log_size":0,"up":0,"acting":0,"num_store_stats":0,"stamp_delta":"3.106349"},"pg_stats":[{"pgid":"1.0","version":"18'32","reported_seq":57,"reported_epoch":19,"state":"active+clean","last_fresh":"2026-03-09T15:14:03.387967+0000","last_change":"2026-03-09T15:14:02.379010+0000","last_active":"2026-03-09T15:14:03.387967+0000","last_peered":"2026-03-09T15:14:03.387967+0000","last_clean":"2026-03-09T15:14:03.387967+0000","last_became_active":"2026-03-09T15:14:02.378816+0000","last_became_peered":"2026-03-09T15:14:02.378816+0000","last_unstale":"2026-03-09T15:14:03.387967+0000","last_undegraded":"2026-03-09T15:14:03.387967+0000","last_fullsized":"2026-03-09T15:14:03.387967+0000","mapping_epoch":17,"log_start":"0'0","ondisk_log_start":"0'0","created":17,"last_epoch_clean":18,"parent":"0.0","parent_split_bits":0,"last_scrub":"0'0","last_scrub_stamp":"2026-03-09T15:14:01.364622+0000","last_deep_scrub":"0'0","last_deep_scrub_stamp":"2026-03-09T15:14:01.364622+0000","last_clean_scrub_stamp":"2026-03-09T15:14:01.364622+0000","objects_scrubbed":0,"log_size":32,"log_dups_size":0,"ondisk_log_size":32,"stats_invalid":false,"dirty_stats_invalid":false,"omap_stats_invalid":false,"hitset_stats_invalid":false,"hitset_bytes_stats_invalid":false,"pin_stats_invalid":false,"manifest_stats_invalid":false,"snaptrimq_len":0,"last_scrub_duration":0,"scrub_schedule":"periodic scrub scheduled @ 2026-03-10T17:24:04.992646+0000","scrub_duration":0,"objects_trimmed":0,"snaptrim_duration":0,"stat_sum":{"num_bytes":459280,"num_objects":2,"num_object_clones":0,"num_object_copies":6,"num_objects_missing_on_primary":0,"num_objects_missing":0,"num_objects_degraded":0,"num_objects_misplaced":0,"num_objects_unfound":0,"num_objects_dirty":2,"num_whiteouts":0,"num_read":46,"num_read_kb":37,"num_write":57,"num_write_kb":584,"num_scrub_errors":0,"num_shallow_scrub_errors":0,"num_deep_scrub_errors":0,"num_objects_recovered":0,"num_bytes_recovered":0,"num_keys_recovered":0,"num_objects_omap":0,"num_objects_hit_set_archive":0,"num_bytes_hit_set_archive":0,"num_flush":0,"num_flush_kb":0,"num_evict":0,"num_evict_kb":0,"num_promote":0,"num_flush_mode_high":0,"num_flush_mode_low":0,"num_evict_mode_some":0,"num_evict_mode_full":0,"num_objects_pinned":0,"num_legacy_snapsets":0,"num_large_omap_objects":0,"num_objects_manifest":0,"num_omap_bytes":0,"num_omap_keys":0,"num_objects_repaired":0},"up":[1,2,0],"acting":[1,2,0],"avail_no_missing":[],"object_location_counts":[],"blocked_by":[],"up_primary":1,"acting_primary":1,"purged_snaps":[]}],"pool_stats":[{"poolid":1,"num_pg":1,"stat_sum":{"num_bytes":459280,"num_objects":2,"num_object_clones":0,"num_object_copies":6,"num_objects_missing_on_primary":0,"num_objects_missing":0,"num_objects_degraded":0,"num_objects_misplaced":0,"num_objects_unfound":0,"num_objects_dirty":2,"num_whiteouts":0,"num_read":46,"num_read_kb":37,"num_write":57,"num_write_kb":584,"num_scrub_errors":0,"num_shallow_scrub_errors":0,"num_deep_scrub_errors":0,"num_objects_recovered":0,"num_bytes_recovered":0,"num_keys_recovered":0,"num_objects_omap":0,"num_objects_hit_set_archive":0,"num_bytes_hit_set_archive":0,"num_flush":0,"num_flush_kb":0,"num_evict":0,"num_evict_kb":0,"num_promote":0,"num_flush_mode_high":0,"num_flush_mode_low":0,"num_evict_mode_some":0,"num_evict_mode_full":0,"num_objects_pinned":0,"num_legacy_snapsets":0,"num_large_omap_objects":0,"num_objects_manifest":0,"num_omap_bytes":0,"num_omap_keys":0,"num_objects_repaired":0},"store_stats":{"total":0,"available":0,"internally_reserved":0,"allocated":1388544,"data_stored":1377840,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":0},"log_size":32,"ondisk_log_size":32,"up":3,"acting":3,"num_store_stats":3}],"osd_stats":[{"osd":2,"up_from":16,"seq":68719476739,"num_pgs":1,"num_osds":1,"num_per_pool_osds":1,"num_per_pool_omap_osds":1,"kb":20967424,"kb_used":27584,"kb_used_data":612,"kb_used_omap":1,"kb_used_meta":26814,"kb_avail":20939840,"statfs":{"total":21470642176,"available":21442396160,"internally_reserved":0,"allocated":626688,"data_stored":509738,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":1590,"internal_metadata":27457994},"hb_peers":[0,1],"snap_trim_queue_len":0,"num_snap_trimming":0,"num_shards_repaired":0,"op_queue_age_hist":{"histogram":[],"upper_bound":1},"perf_stat":{"commit_latency_ms":0,"apply_latency_ms":0,"commit_latency_ns":0,"apply_latency_ns":0},"alerts":[]},{"osd":1,"up_from":12,"seq":51539607557,"num_pgs":1,"num_osds":1,"num_per_pool_osds":1,"num_per_pool_omap_osds":1,"kb":20967424,"kb_used":27584,"kb_used_data":612,"kb_used_omap":1,"kb_used_meta":26814,"kb_avail":20939840,"statfs":{"total":21470642176,"available":21442396160,"internally_reserved":0,"allocated":626688,"data_stored":509738,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":1589,"internal_metadata":27457995},"hb_peers":[0,2],"snap_trim_queue_len":0,"num_snap_trimming":0,"num_shards_repaired":0,"op_queue_age_hist":{"histogram":[],"upper_bound":1},"perf_stat":{"commit_latency_ms":0,"apply_latency_ms":0,"commit_latency_ns":0,"apply_latency_ns":0},"alerts":[]},{"osd":0,"up_from":8,"seq":34359738374,"num_pgs":1,"num_osds":1,"num_per_pool_osds":1,"num_per_pool_omap_osds":1,"kb":20967424,"kb_used":27316,"kb_used_data":476,"kb_used_omap":1,"kb_used_meta":26814,"kb_avail":20940108,"statfs":{"total":21470642176,"available":21442670592,"internally_reserved":0,"allocated":487424,"data_stored":374873,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":1590,"internal_metadata":27457994},"hb_peers":[1,2],"snap_trim_queue_len":0,"num_snap_trimming":0,"num_shards_repaired":0,"op_queue_age_hist":{"histogram":[],"upper_bound":1},"perf_stat":{"commit_latency_ms":0,"apply_latency_ms":0,"commit_latency_ns":0,"apply_latency_ns":0},"alerts":[]}],"pool_statfs":[{"poolid":1,"osd":0,"total":0,"available":0,"internally_reserved":0,"allocated":462848,"data_stored":459280,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":0},{"poolid":1,"osd":1,"total":0,"available":0,"internally_reserved":0,"allocated":462848,"data_stored":459280,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":0},{"poolid":1,"osd":2,"total":0,"available":0,"internally_reserved":0,"allocated":462848,"data_stored":459280,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":0}]}} 2026-03-09T15:14:07.290 DEBUG:teuthology.orchestra.run.vm03:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell --fsid 4abede80-1bca-11f1-bcd8-9b1da3c03ce5 -- ceph pg dump --format=json 2026-03-09T15:14:07.465 INFO:teuthology.orchestra.run.vm03.stderr:Inferring config /var/lib/ceph/4abede80-1bca-11f1-bcd8-9b1da3c03ce5/mon.a/config 2026-03-09T15:14:07.580 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:14:07 vm03 ceph-mon[51762]: pgmap v41: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-09T15:14:07.580 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:14:07 vm03 ceph-mon[51762]: from='client.? 192.168.123.103:0/3140079543' entity='client.admin' cmd=[{"prefix": "osd last-stat-seq", "id": 1}]: dispatch 2026-03-09T15:14:07.580 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:14:07 vm03 ceph-mon[51762]: from='client.? 192.168.123.103:0/2870537684' entity='client.admin' cmd=[{"prefix": "osd last-stat-seq", "id": 2}]: dispatch 2026-03-09T15:14:07.680 INFO:teuthology.orchestra.run.vm03.stdout: 2026-03-09T15:14:07.680 INFO:teuthology.orchestra.run.vm03.stderr:dumped all 2026-03-09T15:14:07.747 INFO:teuthology.orchestra.run.vm03.stdout:{"pg_ready":true,"pg_map":{"version":41,"stamp":"2026-03-09T15:14:06.493749+0000","last_osdmap_epoch":0,"last_pg_scan":0,"pg_stats_sum":{"stat_sum":{"num_bytes":459280,"num_objects":2,"num_object_clones":0,"num_object_copies":6,"num_objects_missing_on_primary":0,"num_objects_missing":0,"num_objects_degraded":0,"num_objects_misplaced":0,"num_objects_unfound":0,"num_objects_dirty":2,"num_whiteouts":0,"num_read":46,"num_read_kb":37,"num_write":57,"num_write_kb":584,"num_scrub_errors":0,"num_shallow_scrub_errors":0,"num_deep_scrub_errors":0,"num_objects_recovered":0,"num_bytes_recovered":0,"num_keys_recovered":0,"num_objects_omap":0,"num_objects_hit_set_archive":0,"num_bytes_hit_set_archive":0,"num_flush":0,"num_flush_kb":0,"num_evict":0,"num_evict_kb":0,"num_promote":0,"num_flush_mode_high":0,"num_flush_mode_low":0,"num_evict_mode_some":0,"num_evict_mode_full":0,"num_objects_pinned":0,"num_legacy_snapsets":0,"num_large_omap_objects":0,"num_objects_manifest":0,"num_omap_bytes":0,"num_omap_keys":0,"num_objects_repaired":0},"store_stats":{"total":0,"available":0,"internally_reserved":0,"allocated":0,"data_stored":0,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":0},"log_size":32,"ondisk_log_size":32,"up":3,"acting":3,"num_store_stats":0},"osd_stats_sum":{"up_from":0,"seq":0,"num_pgs":3,"num_osds":3,"num_per_pool_osds":3,"num_per_pool_omap_osds":3,"kb":62902272,"kb_used":82484,"kb_used_data":1700,"kb_used_omap":4,"kb_used_meta":80443,"kb_avail":62819788,"statfs":{"total":64411926528,"available":64327462912,"internally_reserved":0,"allocated":1740800,"data_stored":1394349,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":4769,"internal_metadata":82373983},"hb_peers":[],"snap_trim_queue_len":0,"num_snap_trimming":0,"num_shards_repaired":0,"op_queue_age_hist":{"histogram":[],"upper_bound":1},"perf_stat":{"commit_latency_ms":0,"apply_latency_ms":0,"commit_latency_ns":0,"apply_latency_ns":0},"alerts":[],"network_ping_times":[]},"pg_stats_delta":{"stat_sum":{"num_bytes":0,"num_objects":0,"num_object_clones":0,"num_object_copies":0,"num_objects_missing_on_primary":0,"num_objects_missing":0,"num_objects_degraded":0,"num_objects_misplaced":0,"num_objects_unfound":0,"num_objects_dirty":0,"num_whiteouts":0,"num_read":0,"num_read_kb":0,"num_write":0,"num_write_kb":0,"num_scrub_errors":0,"num_shallow_scrub_errors":0,"num_deep_scrub_errors":0,"num_objects_recovered":0,"num_bytes_recovered":0,"num_keys_recovered":0,"num_objects_omap":0,"num_objects_hit_set_archive":0,"num_bytes_hit_set_archive":0,"num_flush":0,"num_flush_kb":0,"num_evict":0,"num_evict_kb":0,"num_promote":0,"num_flush_mode_high":0,"num_flush_mode_low":0,"num_evict_mode_some":0,"num_evict_mode_full":0,"num_objects_pinned":0,"num_legacy_snapsets":0,"num_large_omap_objects":0,"num_objects_manifest":0,"num_omap_bytes":0,"num_omap_keys":0,"num_objects_repaired":0},"store_stats":{"total":0,"available":0,"internally_reserved":0,"allocated":0,"data_stored":0,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":0},"log_size":0,"ondisk_log_size":0,"up":0,"acting":0,"num_store_stats":0,"stamp_delta":"3.106349"},"pg_stats":[{"pgid":"1.0","version":"18'32","reported_seq":57,"reported_epoch":19,"state":"active+clean","last_fresh":"2026-03-09T15:14:03.387967+0000","last_change":"2026-03-09T15:14:02.379010+0000","last_active":"2026-03-09T15:14:03.387967+0000","last_peered":"2026-03-09T15:14:03.387967+0000","last_clean":"2026-03-09T15:14:03.387967+0000","last_became_active":"2026-03-09T15:14:02.378816+0000","last_became_peered":"2026-03-09T15:14:02.378816+0000","last_unstale":"2026-03-09T15:14:03.387967+0000","last_undegraded":"2026-03-09T15:14:03.387967+0000","last_fullsized":"2026-03-09T15:14:03.387967+0000","mapping_epoch":17,"log_start":"0'0","ondisk_log_start":"0'0","created":17,"last_epoch_clean":18,"parent":"0.0","parent_split_bits":0,"last_scrub":"0'0","last_scrub_stamp":"2026-03-09T15:14:01.364622+0000","last_deep_scrub":"0'0","last_deep_scrub_stamp":"2026-03-09T15:14:01.364622+0000","last_clean_scrub_stamp":"2026-03-09T15:14:01.364622+0000","objects_scrubbed":0,"log_size":32,"log_dups_size":0,"ondisk_log_size":32,"stats_invalid":false,"dirty_stats_invalid":false,"omap_stats_invalid":false,"hitset_stats_invalid":false,"hitset_bytes_stats_invalid":false,"pin_stats_invalid":false,"manifest_stats_invalid":false,"snaptrimq_len":0,"last_scrub_duration":0,"scrub_schedule":"periodic scrub scheduled @ 2026-03-10T17:24:04.992646+0000","scrub_duration":0,"objects_trimmed":0,"snaptrim_duration":0,"stat_sum":{"num_bytes":459280,"num_objects":2,"num_object_clones":0,"num_object_copies":6,"num_objects_missing_on_primary":0,"num_objects_missing":0,"num_objects_degraded":0,"num_objects_misplaced":0,"num_objects_unfound":0,"num_objects_dirty":2,"num_whiteouts":0,"num_read":46,"num_read_kb":37,"num_write":57,"num_write_kb":584,"num_scrub_errors":0,"num_shallow_scrub_errors":0,"num_deep_scrub_errors":0,"num_objects_recovered":0,"num_bytes_recovered":0,"num_keys_recovered":0,"num_objects_omap":0,"num_objects_hit_set_archive":0,"num_bytes_hit_set_archive":0,"num_flush":0,"num_flush_kb":0,"num_evict":0,"num_evict_kb":0,"num_promote":0,"num_flush_mode_high":0,"num_flush_mode_low":0,"num_evict_mode_some":0,"num_evict_mode_full":0,"num_objects_pinned":0,"num_legacy_snapsets":0,"num_large_omap_objects":0,"num_objects_manifest":0,"num_omap_bytes":0,"num_omap_keys":0,"num_objects_repaired":0},"up":[1,2,0],"acting":[1,2,0],"avail_no_missing":[],"object_location_counts":[],"blocked_by":[],"up_primary":1,"acting_primary":1,"purged_snaps":[]}],"pool_stats":[{"poolid":1,"num_pg":1,"stat_sum":{"num_bytes":459280,"num_objects":2,"num_object_clones":0,"num_object_copies":6,"num_objects_missing_on_primary":0,"num_objects_missing":0,"num_objects_degraded":0,"num_objects_misplaced":0,"num_objects_unfound":0,"num_objects_dirty":2,"num_whiteouts":0,"num_read":46,"num_read_kb":37,"num_write":57,"num_write_kb":584,"num_scrub_errors":0,"num_shallow_scrub_errors":0,"num_deep_scrub_errors":0,"num_objects_recovered":0,"num_bytes_recovered":0,"num_keys_recovered":0,"num_objects_omap":0,"num_objects_hit_set_archive":0,"num_bytes_hit_set_archive":0,"num_flush":0,"num_flush_kb":0,"num_evict":0,"num_evict_kb":0,"num_promote":0,"num_flush_mode_high":0,"num_flush_mode_low":0,"num_evict_mode_some":0,"num_evict_mode_full":0,"num_objects_pinned":0,"num_legacy_snapsets":0,"num_large_omap_objects":0,"num_objects_manifest":0,"num_omap_bytes":0,"num_omap_keys":0,"num_objects_repaired":0},"store_stats":{"total":0,"available":0,"internally_reserved":0,"allocated":1388544,"data_stored":1377840,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":0},"log_size":32,"ondisk_log_size":32,"up":3,"acting":3,"num_store_stats":3}],"osd_stats":[{"osd":2,"up_from":16,"seq":68719476739,"num_pgs":1,"num_osds":1,"num_per_pool_osds":1,"num_per_pool_omap_osds":1,"kb":20967424,"kb_used":27584,"kb_used_data":612,"kb_used_omap":1,"kb_used_meta":26814,"kb_avail":20939840,"statfs":{"total":21470642176,"available":21442396160,"internally_reserved":0,"allocated":626688,"data_stored":509738,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":1590,"internal_metadata":27457994},"hb_peers":[0,1],"snap_trim_queue_len":0,"num_snap_trimming":0,"num_shards_repaired":0,"op_queue_age_hist":{"histogram":[],"upper_bound":1},"perf_stat":{"commit_latency_ms":0,"apply_latency_ms":0,"commit_latency_ns":0,"apply_latency_ns":0},"alerts":[]},{"osd":1,"up_from":12,"seq":51539607557,"num_pgs":1,"num_osds":1,"num_per_pool_osds":1,"num_per_pool_omap_osds":1,"kb":20967424,"kb_used":27584,"kb_used_data":612,"kb_used_omap":1,"kb_used_meta":26814,"kb_avail":20939840,"statfs":{"total":21470642176,"available":21442396160,"internally_reserved":0,"allocated":626688,"data_stored":509738,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":1589,"internal_metadata":27457995},"hb_peers":[0,2],"snap_trim_queue_len":0,"num_snap_trimming":0,"num_shards_repaired":0,"op_queue_age_hist":{"histogram":[],"upper_bound":1},"perf_stat":{"commit_latency_ms":0,"apply_latency_ms":0,"commit_latency_ns":0,"apply_latency_ns":0},"alerts":[]},{"osd":0,"up_from":8,"seq":34359738374,"num_pgs":1,"num_osds":1,"num_per_pool_osds":1,"num_per_pool_omap_osds":1,"kb":20967424,"kb_used":27316,"kb_used_data":476,"kb_used_omap":1,"kb_used_meta":26814,"kb_avail":20940108,"statfs":{"total":21470642176,"available":21442670592,"internally_reserved":0,"allocated":487424,"data_stored":374873,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":1590,"internal_metadata":27457994},"hb_peers":[1,2],"snap_trim_queue_len":0,"num_snap_trimming":0,"num_shards_repaired":0,"op_queue_age_hist":{"histogram":[],"upper_bound":1},"perf_stat":{"commit_latency_ms":0,"apply_latency_ms":0,"commit_latency_ns":0,"apply_latency_ns":0},"alerts":[]}],"pool_statfs":[{"poolid":1,"osd":0,"total":0,"available":0,"internally_reserved":0,"allocated":462848,"data_stored":459280,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":0},{"poolid":1,"osd":1,"total":0,"available":0,"internally_reserved":0,"allocated":462848,"data_stored":459280,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":0},{"poolid":1,"osd":2,"total":0,"available":0,"internally_reserved":0,"allocated":462848,"data_stored":459280,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":0}]}} 2026-03-09T15:14:07.747 INFO:tasks.cephadm.ceph_manager.ceph:clean! 2026-03-09T15:14:07.747 INFO:tasks.ceph:Waiting until ceph cluster ceph is healthy... 2026-03-09T15:14:07.747 INFO:tasks.cephadm.ceph_manager.ceph:wait_until_healthy 2026-03-09T15:14:07.747 DEBUG:teuthology.orchestra.run.vm03:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell --fsid 4abede80-1bca-11f1-bcd8-9b1da3c03ce5 -- ceph health --format=json 2026-03-09T15:14:07.751 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 15:14:07 vm04 ceph-mon[53998]: pgmap v41: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-09T15:14:07.751 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 15:14:07 vm04 ceph-mon[53998]: from='client.? 192.168.123.103:0/3140079543' entity='client.admin' cmd=[{"prefix": "osd last-stat-seq", "id": 1}]: dispatch 2026-03-09T15:14:07.751 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 15:14:07 vm04 ceph-mon[53998]: from='client.? 192.168.123.103:0/2870537684' entity='client.admin' cmd=[{"prefix": "osd last-stat-seq", "id": 2}]: dispatch 2026-03-09T15:14:07.767 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 09 15:14:07 vm06 ceph-mon[56506]: pgmap v41: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-09T15:14:07.767 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 09 15:14:07 vm06 ceph-mon[56506]: from='client.? 192.168.123.103:0/3140079543' entity='client.admin' cmd=[{"prefix": "osd last-stat-seq", "id": 1}]: dispatch 2026-03-09T15:14:07.767 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 09 15:14:07 vm06 ceph-mon[56506]: from='client.? 192.168.123.103:0/2870537684' entity='client.admin' cmd=[{"prefix": "osd last-stat-seq", "id": 2}]: dispatch 2026-03-09T15:14:07.907 INFO:teuthology.orchestra.run.vm03.stderr:Inferring config /var/lib/ceph/4abede80-1bca-11f1-bcd8-9b1da3c03ce5/mon.a/config 2026-03-09T15:14:08.136 INFO:teuthology.orchestra.run.vm03.stdout: 2026-03-09T15:14:08.137 INFO:teuthology.orchestra.run.vm03.stdout:{"status":"HEALTH_OK","checks":{},"mutes":[]} 2026-03-09T15:14:08.182 INFO:tasks.cephadm.ceph_manager.ceph:wait_until_healthy done 2026-03-09T15:14:08.182 INFO:tasks.cephadm:Setup complete, yielding 2026-03-09T15:14:08.182 INFO:teuthology.run_tasks:Running task cephadm.apply... 2026-03-09T15:14:08.185 INFO:tasks.cephadm:Applying spec(s): placement: count: 3 service_id: foo service_type: mon spec: crush_locations: host.a: - datacenter=a host.b: - datacenter=b - rack=2 host.c: - datacenter=a - rack=3 2026-03-09T15:14:08.185 DEBUG:teuthology.orchestra.run.vm03:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 4abede80-1bca-11f1-bcd8-9b1da3c03ce5 -- ceph orch apply -i - 2026-03-09T15:14:08.342 INFO:teuthology.orchestra.run.vm03.stderr:Inferring config /var/lib/ceph/4abede80-1bca-11f1-bcd8-9b1da3c03ce5/mon.a/config 2026-03-09T15:14:08.457 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:14:08 vm03 ceph-mon[51762]: from='client.24238 -' entity='client.admin' cmd=[{"prefix": "pg dump", "target": ["mon-mgr", ""], "format": "json"}]: dispatch 2026-03-09T15:14:08.457 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:14:08 vm03 ceph-mon[51762]: from='client.? 192.168.123.103:0/475234116' entity='client.admin' cmd=[{"prefix": "health", "format": "json"}]: dispatch 2026-03-09T15:14:08.571 INFO:teuthology.orchestra.run.vm03.stdout:Scheduled mon update... 2026-03-09T15:14:08.637 INFO:teuthology.run_tasks:Running task cephadm.shell... 2026-03-09T15:14:08.639 INFO:tasks.cephadm:Running commands on role host.a host ubuntu@vm03.local 2026-03-09T15:14:08.640 DEBUG:teuthology.orchestra.run.vm03:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 4abede80-1bca-11f1-bcd8-9b1da3c03ce5 -- bash -c 'set -ex 2026-03-09T15:14:08.640 DEBUG:teuthology.orchestra.run.vm03:> # since we don'"'"'t know the real hostnames before the test, the next 2026-03-09T15:14:08.640 DEBUG:teuthology.orchestra.run.vm03:> # bit is in order to replace the fake hostnames "host.a/b/c" with 2026-03-09T15:14:08.640 DEBUG:teuthology.orchestra.run.vm03:> # the actual names cephadm knows the host by within the mon spec 2026-03-09T15:14:08.640 DEBUG:teuthology.orchestra.run.vm03:> ceph orch host ls --format json | jq -r '"'"'.[] | .hostname'"'"' > realnames 2026-03-09T15:14:08.640 DEBUG:teuthology.orchestra.run.vm03:> echo $'"'"'host.a\nhost.b\nhost.c'"'"' > fakenames 2026-03-09T15:14:08.640 DEBUG:teuthology.orchestra.run.vm03:> echo $'"'"'a\nb\nc'"'"' > mon_ids 2026-03-09T15:14:08.640 DEBUG:teuthology.orchestra.run.vm03:> echo $'"'"'{datacenter=a}\n{datacenter=b,rack=2}\n{datacenter=a,rack=3}'"'"' > crush_locs 2026-03-09T15:14:08.640 DEBUG:teuthology.orchestra.run.vm03:> ceph orch ls --service-name mon --export > mon.yaml 2026-03-09T15:14:08.640 DEBUG:teuthology.orchestra.run.vm03:> MONSPEC=`cat mon.yaml` 2026-03-09T15:14:08.640 DEBUG:teuthology.orchestra.run.vm03:> echo "$MONSPEC" 2026-03-09T15:14:08.640 DEBUG:teuthology.orchestra.run.vm03:> while read realname <&3 && read fakename <&4; do 2026-03-09T15:14:08.640 DEBUG:teuthology.orchestra.run.vm03:> MONSPEC="${MONSPEC//$fakename/$realname}" 2026-03-09T15:14:08.640 DEBUG:teuthology.orchestra.run.vm03:> done 3 echo "$MONSPEC" > mon.yaml 2026-03-09T15:14:08.640 DEBUG:teuthology.orchestra.run.vm03:> cat mon.yaml 2026-03-09T15:14:08.640 DEBUG:teuthology.orchestra.run.vm03:> # now the spec should have the real hostnames, so let'"'"'s re-apply 2026-03-09T15:14:08.640 DEBUG:teuthology.orchestra.run.vm03:> ceph orch apply -i mon.yaml 2026-03-09T15:14:08.640 DEBUG:teuthology.orchestra.run.vm03:> sleep 90 2026-03-09T15:14:08.640 DEBUG:teuthology.orchestra.run.vm03:> ceph orch ps --refresh 2026-03-09T15:14:08.640 DEBUG:teuthology.orchestra.run.vm03:> ceph orch ls --service-name mon --export > mon.yaml; ceph orch apply -i mon.yaml 2026-03-09T15:14:08.640 DEBUG:teuthology.orchestra.run.vm03:> sleep 90 2026-03-09T15:14:08.640 DEBUG:teuthology.orchestra.run.vm03:> ceph mon dump 2026-03-09T15:14:08.640 DEBUG:teuthology.orchestra.run.vm03:> ceph mon dump --format json 2026-03-09T15:14:08.640 DEBUG:teuthology.orchestra.run.vm03:> # verify all the crush locations got set from "ceph mon dump" output 2026-03-09T15:14:08.640 DEBUG:teuthology.orchestra.run.vm03:> while read monid <&3 && read crushloc <&4; do 2026-03-09T15:14:08.640 DEBUG:teuthology.orchestra.run.vm03:> ceph mon dump --format json | jq --arg monid "$monid" --arg crushloc "$crushloc" -e '"'"'.mons | .[] | select(.name == $monid) | .crush_location == $crushloc'"'"' 2026-03-09T15:14:08.640 DEBUG:teuthology.orchestra.run.vm03:> done 3 ' 2026-03-09T15:14:08.751 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 15:14:08 vm04 ceph-mon[53998]: from='client.24238 -' entity='client.admin' cmd=[{"prefix": "pg dump", "target": ["mon-mgr", ""], "format": "json"}]: dispatch 2026-03-09T15:14:08.751 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 15:14:08 vm04 ceph-mon[53998]: from='client.? 192.168.123.103:0/475234116' entity='client.admin' cmd=[{"prefix": "health", "format": "json"}]: dispatch 2026-03-09T15:14:08.767 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 09 15:14:08 vm06 ceph-mon[56506]: from='client.24238 -' entity='client.admin' cmd=[{"prefix": "pg dump", "target": ["mon-mgr", ""], "format": "json"}]: dispatch 2026-03-09T15:14:08.767 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 09 15:14:08 vm06 ceph-mon[56506]: from='client.? 192.168.123.103:0/475234116' entity='client.admin' cmd=[{"prefix": "health", "format": "json"}]: dispatch 2026-03-09T15:14:08.828 INFO:teuthology.orchestra.run.vm03.stderr:Inferring config /var/lib/ceph/4abede80-1bca-11f1-bcd8-9b1da3c03ce5/mon.a/config 2026-03-09T15:14:08.907 INFO:teuthology.orchestra.run.vm03.stderr:+ ceph orch host ls --format json 2026-03-09T15:14:08.910 INFO:teuthology.orchestra.run.vm03.stderr:+ jq -r '.[] | .hostname' 2026-03-09T15:14:09.071 INFO:teuthology.orchestra.run.vm03.stderr:+ echo 'host.a 2026-03-09T15:14:09.071 INFO:teuthology.orchestra.run.vm03.stderr:host.b 2026-03-09T15:14:09.071 INFO:teuthology.orchestra.run.vm03.stderr:host.c' 2026-03-09T15:14:09.072 INFO:teuthology.orchestra.run.vm03.stderr:+ echo 'a 2026-03-09T15:14:09.072 INFO:teuthology.orchestra.run.vm03.stderr:b 2026-03-09T15:14:09.072 INFO:teuthology.orchestra.run.vm03.stderr:c' 2026-03-09T15:14:09.072 INFO:teuthology.orchestra.run.vm03.stderr:+ echo '{datacenter=a} 2026-03-09T15:14:09.072 INFO:teuthology.orchestra.run.vm03.stderr:{datacenter=b,rack=2} 2026-03-09T15:14:09.072 INFO:teuthology.orchestra.run.vm03.stderr:{datacenter=a,rack=3}' 2026-03-09T15:14:09.072 INFO:teuthology.orchestra.run.vm03.stderr:+ ceph orch ls --service-name mon --export 2026-03-09T15:14:09.231 INFO:teuthology.orchestra.run.vm03.stderr:++ cat mon.yaml 2026-03-09T15:14:09.233 INFO:teuthology.orchestra.run.vm03.stderr:+ MONSPEC='service_type: mon 2026-03-09T15:14:09.233 INFO:teuthology.orchestra.run.vm03.stderr:service_name: mon 2026-03-09T15:14:09.233 INFO:teuthology.orchestra.run.vm03.stderr:placement: 2026-03-09T15:14:09.233 INFO:teuthology.orchestra.run.vm03.stderr: count: 3 2026-03-09T15:14:09.233 INFO:teuthology.orchestra.run.vm03.stderr:spec: 2026-03-09T15:14:09.233 INFO:teuthology.orchestra.run.vm03.stderr: crush_locations: 2026-03-09T15:14:09.233 INFO:teuthology.orchestra.run.vm03.stderr: host.a: 2026-03-09T15:14:09.233 INFO:teuthology.orchestra.run.vm03.stderr: - datacenter=a 2026-03-09T15:14:09.233 INFO:teuthology.orchestra.run.vm03.stderr: host.b: 2026-03-09T15:14:09.233 INFO:teuthology.orchestra.run.vm03.stderr: - datacenter=b 2026-03-09T15:14:09.233 INFO:teuthology.orchestra.run.vm03.stderr: - rack=2 2026-03-09T15:14:09.233 INFO:teuthology.orchestra.run.vm03.stderr: host.c: 2026-03-09T15:14:09.233 INFO:teuthology.orchestra.run.vm03.stderr: - datacenter=a 2026-03-09T15:14:09.233 INFO:teuthology.orchestra.run.vm03.stderr: - rack=3' 2026-03-09T15:14:09.233 INFO:teuthology.orchestra.run.vm03.stderr:+ echo 'service_type: mon 2026-03-09T15:14:09.233 INFO:teuthology.orchestra.run.vm03.stderr:service_name: mon 2026-03-09T15:14:09.233 INFO:teuthology.orchestra.run.vm03.stderr:placement: 2026-03-09T15:14:09.233 INFO:teuthology.orchestra.run.vm03.stderr: count: 3 2026-03-09T15:14:09.233 INFO:teuthology.orchestra.run.vm03.stderr:spec: 2026-03-09T15:14:09.233 INFO:teuthology.orchestra.run.vm03.stderr: crush_locations: 2026-03-09T15:14:09.233 INFO:teuthology.orchestra.run.vm03.stderr: host.a: 2026-03-09T15:14:09.233 INFO:teuthology.orchestra.run.vm03.stderr: - datacenter=a 2026-03-09T15:14:09.233 INFO:teuthology.orchestra.run.vm03.stderr: host.b: 2026-03-09T15:14:09.233 INFO:teuthology.orchestra.run.vm03.stderr: - datacenter=b 2026-03-09T15:14:09.233 INFO:teuthology.orchestra.run.vm03.stderr: - rack=2 2026-03-09T15:14:09.233 INFO:teuthology.orchestra.run.vm03.stderr: host.c: 2026-03-09T15:14:09.233 INFO:teuthology.orchestra.run.vm03.stderr: - datacenter=a 2026-03-09T15:14:09.233 INFO:teuthology.orchestra.run.vm03.stderr: - rack=3' 2026-03-09T15:14:09.233 INFO:teuthology.orchestra.run.vm03.stdout:service_type: mon 2026-03-09T15:14:09.233 INFO:teuthology.orchestra.run.vm03.stdout:service_name: mon 2026-03-09T15:14:09.233 INFO:teuthology.orchestra.run.vm03.stdout:placement: 2026-03-09T15:14:09.233 INFO:teuthology.orchestra.run.vm03.stdout: count: 3 2026-03-09T15:14:09.233 INFO:teuthology.orchestra.run.vm03.stdout:spec: 2026-03-09T15:14:09.233 INFO:teuthology.orchestra.run.vm03.stdout: crush_locations: 2026-03-09T15:14:09.233 INFO:teuthology.orchestra.run.vm03.stdout: host.a: 2026-03-09T15:14:09.233 INFO:teuthology.orchestra.run.vm03.stdout: - datacenter=a 2026-03-09T15:14:09.233 INFO:teuthology.orchestra.run.vm03.stdout: host.b: 2026-03-09T15:14:09.233 INFO:teuthology.orchestra.run.vm03.stdout: - datacenter=b 2026-03-09T15:14:09.233 INFO:teuthology.orchestra.run.vm03.stdout: - rack=2 2026-03-09T15:14:09.233 INFO:teuthology.orchestra.run.vm03.stdout: host.c: 2026-03-09T15:14:09.233 INFO:teuthology.orchestra.run.vm03.stdout: - datacenter=a 2026-03-09T15:14:09.233 INFO:teuthology.orchestra.run.vm03.stdout: - rack=3 2026-03-09T15:14:09.234 INFO:teuthology.orchestra.run.vm03.stderr:+ read realname 2026-03-09T15:14:09.234 INFO:teuthology.orchestra.run.vm03.stderr:+ read fakename 2026-03-09T15:14:09.234 INFO:teuthology.orchestra.run.vm03.stderr:+ MONSPEC='service_type: mon 2026-03-09T15:14:09.234 INFO:teuthology.orchestra.run.vm03.stderr:service_name: mon 2026-03-09T15:14:09.234 INFO:teuthology.orchestra.run.vm03.stderr:placement: 2026-03-09T15:14:09.234 INFO:teuthology.orchestra.run.vm03.stderr: count: 3 2026-03-09T15:14:09.234 INFO:teuthology.orchestra.run.vm03.stderr:spec: 2026-03-09T15:14:09.234 INFO:teuthology.orchestra.run.vm03.stderr: crush_locations: 2026-03-09T15:14:09.234 INFO:teuthology.orchestra.run.vm03.stderr: vm03: 2026-03-09T15:14:09.234 INFO:teuthology.orchestra.run.vm03.stderr: - datacenter=a 2026-03-09T15:14:09.234 INFO:teuthology.orchestra.run.vm03.stderr: host.b: 2026-03-09T15:14:09.234 INFO:teuthology.orchestra.run.vm03.stderr: - datacenter=b 2026-03-09T15:14:09.234 INFO:teuthology.orchestra.run.vm03.stderr: - rack=2 2026-03-09T15:14:09.234 INFO:teuthology.orchestra.run.vm03.stderr: host.c: 2026-03-09T15:14:09.234 INFO:teuthology.orchestra.run.vm03.stderr: - datacenter=a 2026-03-09T15:14:09.234 INFO:teuthology.orchestra.run.vm03.stderr: - rack=3' 2026-03-09T15:14:09.234 INFO:teuthology.orchestra.run.vm03.stderr:+ read realname 2026-03-09T15:14:09.234 INFO:teuthology.orchestra.run.vm03.stderr:+ read fakename 2026-03-09T15:14:09.234 INFO:teuthology.orchestra.run.vm03.stderr:+ MONSPEC='service_type: mon 2026-03-09T15:14:09.234 INFO:teuthology.orchestra.run.vm03.stderr:service_name: mon 2026-03-09T15:14:09.234 INFO:teuthology.orchestra.run.vm03.stderr:placement: 2026-03-09T15:14:09.234 INFO:teuthology.orchestra.run.vm03.stderr: count: 3 2026-03-09T15:14:09.234 INFO:teuthology.orchestra.run.vm03.stderr:spec: 2026-03-09T15:14:09.234 INFO:teuthology.orchestra.run.vm03.stderr: crush_locations: 2026-03-09T15:14:09.234 INFO:teuthology.orchestra.run.vm03.stderr: vm03: 2026-03-09T15:14:09.234 INFO:teuthology.orchestra.run.vm03.stderr: - datacenter=a 2026-03-09T15:14:09.234 INFO:teuthology.orchestra.run.vm03.stderr: vm04: 2026-03-09T15:14:09.234 INFO:teuthology.orchestra.run.vm03.stderr: - datacenter=b 2026-03-09T15:14:09.234 INFO:teuthology.orchestra.run.vm03.stderr: - rack=2 2026-03-09T15:14:09.234 INFO:teuthology.orchestra.run.vm03.stderr: host.c: 2026-03-09T15:14:09.234 INFO:teuthology.orchestra.run.vm03.stderr: - datacenter=a 2026-03-09T15:14:09.234 INFO:teuthology.orchestra.run.vm03.stderr: - rack=3' 2026-03-09T15:14:09.234 INFO:teuthology.orchestra.run.vm03.stderr:+ read realname 2026-03-09T15:14:09.234 INFO:teuthology.orchestra.run.vm03.stderr:+ read fakename 2026-03-09T15:14:09.234 INFO:teuthology.orchestra.run.vm03.stderr:+ MONSPEC='service_type: mon 2026-03-09T15:14:09.234 INFO:teuthology.orchestra.run.vm03.stderr:service_name: mon 2026-03-09T15:14:09.234 INFO:teuthology.orchestra.run.vm03.stderr:placement: 2026-03-09T15:14:09.234 INFO:teuthology.orchestra.run.vm03.stderr: count: 3 2026-03-09T15:14:09.234 INFO:teuthology.orchestra.run.vm03.stderr:spec: 2026-03-09T15:14:09.235 INFO:teuthology.orchestra.run.vm03.stderr: crush_locations: 2026-03-09T15:14:09.235 INFO:teuthology.orchestra.run.vm03.stderr: vm03: 2026-03-09T15:14:09.235 INFO:teuthology.orchestra.run.vm03.stderr: - datacenter=a 2026-03-09T15:14:09.235 INFO:teuthology.orchestra.run.vm03.stderr: vm04: 2026-03-09T15:14:09.235 INFO:teuthology.orchestra.run.vm03.stderr: - datacenter=b 2026-03-09T15:14:09.235 INFO:teuthology.orchestra.run.vm03.stderr: - rack=2 2026-03-09T15:14:09.235 INFO:teuthology.orchestra.run.vm03.stderr: vm06: 2026-03-09T15:14:09.235 INFO:teuthology.orchestra.run.vm03.stderr: - datacenter=a 2026-03-09T15:14:09.235 INFO:teuthology.orchestra.run.vm03.stderr: - rack=3' 2026-03-09T15:14:09.235 INFO:teuthology.orchestra.run.vm03.stderr:+ read realname 2026-03-09T15:14:09.235 INFO:teuthology.orchestra.run.vm03.stderr:+ echo 'service_type: mon 2026-03-09T15:14:09.235 INFO:teuthology.orchestra.run.vm03.stderr:service_name: mon 2026-03-09T15:14:09.235 INFO:teuthology.orchestra.run.vm03.stderr:placement: 2026-03-09T15:14:09.235 INFO:teuthology.orchestra.run.vm03.stderr: count: 3 2026-03-09T15:14:09.235 INFO:teuthology.orchestra.run.vm03.stderr:spec: 2026-03-09T15:14:09.235 INFO:teuthology.orchestra.run.vm03.stderr: crush_locations: 2026-03-09T15:14:09.235 INFO:teuthology.orchestra.run.vm03.stderr: vm03: 2026-03-09T15:14:09.235 INFO:teuthology.orchestra.run.vm03.stderr: - datacenter=a 2026-03-09T15:14:09.235 INFO:teuthology.orchestra.run.vm03.stderr: vm04: 2026-03-09T15:14:09.235 INFO:teuthology.orchestra.run.vm03.stderr: - datacenter=b 2026-03-09T15:14:09.235 INFO:teuthology.orchestra.run.vm03.stderr: - rack=2 2026-03-09T15:14:09.235 INFO:teuthology.orchestra.run.vm03.stderr: vm06: 2026-03-09T15:14:09.235 INFO:teuthology.orchestra.run.vm03.stderr: - datacenter=a 2026-03-09T15:14:09.235 INFO:teuthology.orchestra.run.vm03.stderr: - rack=3' 2026-03-09T15:14:09.235 INFO:teuthology.orchestra.run.vm03.stderr:+ cat mon.yaml 2026-03-09T15:14:09.235 INFO:teuthology.orchestra.run.vm03.stdout:service_type: mon 2026-03-09T15:14:09.235 INFO:teuthology.orchestra.run.vm03.stdout:service_name: mon 2026-03-09T15:14:09.235 INFO:teuthology.orchestra.run.vm03.stdout:placement: 2026-03-09T15:14:09.235 INFO:teuthology.orchestra.run.vm03.stdout: count: 3 2026-03-09T15:14:09.235 INFO:teuthology.orchestra.run.vm03.stdout:spec: 2026-03-09T15:14:09.235 INFO:teuthology.orchestra.run.vm03.stdout: crush_locations: 2026-03-09T15:14:09.235 INFO:teuthology.orchestra.run.vm03.stdout: vm03: 2026-03-09T15:14:09.235 INFO:teuthology.orchestra.run.vm03.stdout: - datacenter=a 2026-03-09T15:14:09.235 INFO:teuthology.orchestra.run.vm03.stdout: vm04: 2026-03-09T15:14:09.235 INFO:teuthology.orchestra.run.vm03.stdout: - datacenter=b 2026-03-09T15:14:09.235 INFO:teuthology.orchestra.run.vm03.stdout: - rack=2 2026-03-09T15:14:09.235 INFO:teuthology.orchestra.run.vm03.stdout: vm06: 2026-03-09T15:14:09.235 INFO:teuthology.orchestra.run.vm03.stdout: - datacenter=a 2026-03-09T15:14:09.235 INFO:teuthology.orchestra.run.vm03.stdout: - rack=3 2026-03-09T15:14:09.235 INFO:teuthology.orchestra.run.vm03.stderr:+ ceph orch apply -i mon.yaml 2026-03-09T15:14:09.389 INFO:teuthology.orchestra.run.vm03.stdout:Scheduled mon update... 2026-03-09T15:14:09.405 INFO:teuthology.orchestra.run.vm03.stderr:+ sleep 90 2026-03-09T15:14:15.549 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:14:15 vm03 ceph-mon[51762]: Setting crush location for mon b to {datacenter=b,rack=2} 2026-03-09T15:14:15.550 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:14:15 vm03 ceph-mon[51762]: Setting crush location for mon c to {datacenter=a,rack=3} 2026-03-09T15:14:15.550 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:14:15 vm03 ceph-mon[51762]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' cmd='[{"prefix": "mon set_location", "name": "c", "args": ["datacenter=a", "rack=3"]}]': finished 2026-03-09T15:14:15.550 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:14:15 vm03 ceph-mon[51762]: mon.b calling monitor election 2026-03-09T15:14:15.550 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:14:15 vm03 ceph-mon[51762]: mon.c calling monitor election 2026-03-09T15:14:15.550 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:14:15 vm03 ceph-mon[51762]: mon.a calling monitor election 2026-03-09T15:14:15.550 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:14:15 vm03 ceph-mon[51762]: mon.a is new leader, mons a,c,b in quorum (ranks 0,1,2) 2026-03-09T15:14:15.550 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:14:15 vm03 ceph-mon[51762]: monmap epoch 6 2026-03-09T15:14:15.550 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:14:15 vm03 ceph-mon[51762]: fsid 4abede80-1bca-11f1-bcd8-9b1da3c03ce5 2026-03-09T15:14:15.550 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:14:15 vm03 ceph-mon[51762]: last_changed 2026-03-09T15:14:14.439856+0000 2026-03-09T15:14:15.550 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:14:15 vm03 ceph-mon[51762]: created 2026-03-09T15:12:40.941469+0000 2026-03-09T15:14:15.550 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:14:15 vm03 ceph-mon[51762]: min_mon_release 19 (squid) 2026-03-09T15:14:15.550 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:14:15 vm03 ceph-mon[51762]: election_strategy: 1 2026-03-09T15:14:15.550 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:14:15 vm03 ceph-mon[51762]: 0: [v2:192.168.123.103:3300/0,v1:192.168.123.103:6789/0] mon.a; crush_location {datacenter=a} 2026-03-09T15:14:15.550 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:14:15 vm03 ceph-mon[51762]: 1: [v2:192.168.123.106:3300/0,v1:192.168.123.106:6789/0] mon.c; crush_location {datacenter=a,rack=3} 2026-03-09T15:14:15.550 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:14:15 vm03 ceph-mon[51762]: 2: [v2:192.168.123.104:3300/0,v1:192.168.123.104:6789/0] mon.b; crush_location {datacenter=b,rack=2} 2026-03-09T15:14:15.550 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:14:15 vm03 ceph-mon[51762]: fsmap 2026-03-09T15:14:15.550 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:14:15 vm03 ceph-mon[51762]: osdmap e19: 3 total, 3 up, 3 in 2026-03-09T15:14:15.550 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:14:15 vm03 ceph-mon[51762]: mgrmap e14: a(active, since 71s), standbys: b 2026-03-09T15:14:15.550 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:14:15 vm03 ceph-mon[51762]: overall HEALTH_OK 2026-03-09T15:14:15.550 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:14:15 vm03 ceph-mon[51762]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "a"}]: dispatch 2026-03-09T15:14:15.550 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:14:15 vm03 ceph-mon[51762]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-09T15:14:15.550 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:14:15 vm03 ceph-mon[51762]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "c"}]: dispatch 2026-03-09T15:14:15.550 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:14:15 vm03 ceph-mon[51762]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' 2026-03-09T15:14:15.550 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:14:15 vm03 ceph-mon[51762]: pgmap v45: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-09T15:14:15.550 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:14:15 vm03 ceph-mon[51762]: Reconfiguring mon.a (monmap changed)... 2026-03-09T15:14:15.550 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:14:15 vm03 ceph-mon[51762]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "mon."}]: dispatch 2026-03-09T15:14:15.550 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:14:15 vm03 ceph-mon[51762]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' cmd=[{"prefix": "config get", "who": "mon", "key": "public_network"}]: dispatch 2026-03-09T15:14:15.550 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:14:15 vm03 ceph-mon[51762]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T15:14:15.550 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:14:15 vm03 ceph-mon[51762]: Reconfiguring daemon mon.a on vm03 2026-03-09T15:14:15.550 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:14:15 vm03 ceph-mon[51762]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' 2026-03-09T15:14:15.550 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:14:15 vm03 ceph-mon[51762]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' 2026-03-09T15:14:15.550 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:14:15 vm03 ceph-mon[51762]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' cmd=[{"prefix": "auth get-or-create", "entity": "mgr.a", "caps": ["mon", "profile mgr", "osd", "allow *", "mds", "allow *"]}]: dispatch 2026-03-09T15:14:15.550 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:14:15 vm03 ceph-mon[51762]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' cmd=[{"prefix": "mgr services"}]: dispatch 2026-03-09T15:14:15.550 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:14:15 vm03 ceph-mon[51762]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T15:14:15.550 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:14:15 vm03 ceph-mon[51762]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' 2026-03-09T15:14:15.550 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:14:15 vm03 ceph-mon[51762]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' 2026-03-09T15:14:15.550 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:14:15 vm03 ceph-mon[51762]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "osd.0"}]: dispatch 2026-03-09T15:14:15.550 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:14:15 vm03 ceph-mon[51762]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T15:14:15.550 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:14:15 vm03 ceph-mon[51762]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' 2026-03-09T15:14:15.550 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:14:15 vm03 ceph-mon[51762]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' 2026-03-09T15:14:15.550 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:14:15 vm03 ceph-mon[51762]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "mon."}]: dispatch 2026-03-09T15:14:15.550 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:14:15 vm03 ceph-mon[51762]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' cmd=[{"prefix": "config get", "who": "mon", "key": "public_network"}]: dispatch 2026-03-09T15:14:15.550 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:14:15 vm03 ceph-mon[51762]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T15:14:15.752 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 15:14:15 vm04 ceph-mon[53998]: Setting crush location for mon b to {datacenter=b,rack=2} 2026-03-09T15:14:15.752 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 15:14:15 vm04 ceph-mon[53998]: Setting crush location for mon c to {datacenter=a,rack=3} 2026-03-09T15:14:15.752 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 15:14:15 vm04 ceph-mon[53998]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' cmd='[{"prefix": "mon set_location", "name": "c", "args": ["datacenter=a", "rack=3"]}]': finished 2026-03-09T15:14:15.752 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 15:14:15 vm04 ceph-mon[53998]: mon.b calling monitor election 2026-03-09T15:14:15.752 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 15:14:15 vm04 ceph-mon[53998]: mon.c calling monitor election 2026-03-09T15:14:15.752 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 15:14:15 vm04 ceph-mon[53998]: mon.a calling monitor election 2026-03-09T15:14:15.752 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 15:14:15 vm04 ceph-mon[53998]: mon.a is new leader, mons a,c,b in quorum (ranks 0,1,2) 2026-03-09T15:14:15.752 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 15:14:15 vm04 ceph-mon[53998]: monmap epoch 6 2026-03-09T15:14:15.752 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 15:14:15 vm04 ceph-mon[53998]: fsid 4abede80-1bca-11f1-bcd8-9b1da3c03ce5 2026-03-09T15:14:15.752 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 15:14:15 vm04 ceph-mon[53998]: last_changed 2026-03-09T15:14:14.439856+0000 2026-03-09T15:14:15.753 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 15:14:15 vm04 ceph-mon[53998]: created 2026-03-09T15:12:40.941469+0000 2026-03-09T15:14:15.753 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 15:14:15 vm04 ceph-mon[53998]: min_mon_release 19 (squid) 2026-03-09T15:14:15.753 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 15:14:15 vm04 ceph-mon[53998]: election_strategy: 1 2026-03-09T15:14:15.753 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 15:14:15 vm04 ceph-mon[53998]: 0: [v2:192.168.123.103:3300/0,v1:192.168.123.103:6789/0] mon.a; crush_location {datacenter=a} 2026-03-09T15:14:15.753 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 15:14:15 vm04 ceph-mon[53998]: 1: [v2:192.168.123.106:3300/0,v1:192.168.123.106:6789/0] mon.c; crush_location {datacenter=a,rack=3} 2026-03-09T15:14:15.753 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 15:14:15 vm04 ceph-mon[53998]: 2: [v2:192.168.123.104:3300/0,v1:192.168.123.104:6789/0] mon.b; crush_location {datacenter=b,rack=2} 2026-03-09T15:14:15.753 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 15:14:15 vm04 ceph-mon[53998]: fsmap 2026-03-09T15:14:15.753 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 15:14:15 vm04 ceph-mon[53998]: osdmap e19: 3 total, 3 up, 3 in 2026-03-09T15:14:15.753 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 15:14:15 vm04 ceph-mon[53998]: mgrmap e14: a(active, since 71s), standbys: b 2026-03-09T15:14:15.753 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 15:14:15 vm04 ceph-mon[53998]: overall HEALTH_OK 2026-03-09T15:14:15.753 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 15:14:15 vm04 ceph-mon[53998]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "a"}]: dispatch 2026-03-09T15:14:15.753 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 15:14:15 vm04 ceph-mon[53998]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-09T15:14:15.753 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 15:14:15 vm04 ceph-mon[53998]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "c"}]: dispatch 2026-03-09T15:14:15.753 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 15:14:15 vm04 ceph-mon[53998]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' 2026-03-09T15:14:15.753 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 15:14:15 vm04 ceph-mon[53998]: pgmap v45: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-09T15:14:15.753 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 15:14:15 vm04 ceph-mon[53998]: Reconfiguring mon.a (monmap changed)... 2026-03-09T15:14:15.753 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 15:14:15 vm04 ceph-mon[53998]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "mon."}]: dispatch 2026-03-09T15:14:15.753 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 15:14:15 vm04 ceph-mon[53998]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' cmd=[{"prefix": "config get", "who": "mon", "key": "public_network"}]: dispatch 2026-03-09T15:14:15.753 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 15:14:15 vm04 ceph-mon[53998]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T15:14:15.753 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 15:14:15 vm04 ceph-mon[53998]: Reconfiguring daemon mon.a on vm03 2026-03-09T15:14:15.753 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 15:14:15 vm04 ceph-mon[53998]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' 2026-03-09T15:14:15.753 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 15:14:15 vm04 ceph-mon[53998]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' 2026-03-09T15:14:15.753 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 15:14:15 vm04 ceph-mon[53998]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' cmd=[{"prefix": "auth get-or-create", "entity": "mgr.a", "caps": ["mon", "profile mgr", "osd", "allow *", "mds", "allow *"]}]: dispatch 2026-03-09T15:14:15.753 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 15:14:15 vm04 ceph-mon[53998]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' cmd=[{"prefix": "mgr services"}]: dispatch 2026-03-09T15:14:15.753 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 15:14:15 vm04 ceph-mon[53998]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T15:14:15.753 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 15:14:15 vm04 ceph-mon[53998]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' 2026-03-09T15:14:15.753 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 15:14:15 vm04 ceph-mon[53998]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' 2026-03-09T15:14:15.753 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 15:14:15 vm04 ceph-mon[53998]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "osd.0"}]: dispatch 2026-03-09T15:14:15.753 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 15:14:15 vm04 ceph-mon[53998]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T15:14:15.753 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 15:14:15 vm04 ceph-mon[53998]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' 2026-03-09T15:14:15.753 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 15:14:15 vm04 ceph-mon[53998]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' 2026-03-09T15:14:15.753 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 15:14:15 vm04 ceph-mon[53998]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "mon."}]: dispatch 2026-03-09T15:14:15.753 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 15:14:15 vm04 ceph-mon[53998]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' cmd=[{"prefix": "config get", "who": "mon", "key": "public_network"}]: dispatch 2026-03-09T15:14:15.753 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 15:14:15 vm04 ceph-mon[53998]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T15:14:15.767 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 09 15:14:15 vm06 ceph-mon[56506]: Setting crush location for mon b to {datacenter=b,rack=2} 2026-03-09T15:14:15.767 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 09 15:14:15 vm06 ceph-mon[56506]: Setting crush location for mon c to {datacenter=a,rack=3} 2026-03-09T15:14:15.767 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 09 15:14:15 vm06 ceph-mon[56506]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' cmd='[{"prefix": "mon set_location", "name": "c", "args": ["datacenter=a", "rack=3"]}]': finished 2026-03-09T15:14:15.767 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 09 15:14:15 vm06 ceph-mon[56506]: mon.b calling monitor election 2026-03-09T15:14:15.767 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 09 15:14:15 vm06 ceph-mon[56506]: mon.c calling monitor election 2026-03-09T15:14:15.767 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 09 15:14:15 vm06 ceph-mon[56506]: mon.a calling monitor election 2026-03-09T15:14:15.767 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 09 15:14:15 vm06 ceph-mon[56506]: mon.a is new leader, mons a,c,b in quorum (ranks 0,1,2) 2026-03-09T15:14:15.767 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 09 15:14:15 vm06 ceph-mon[56506]: monmap epoch 6 2026-03-09T15:14:15.767 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 09 15:14:15 vm06 ceph-mon[56506]: fsid 4abede80-1bca-11f1-bcd8-9b1da3c03ce5 2026-03-09T15:14:15.767 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 09 15:14:15 vm06 ceph-mon[56506]: last_changed 2026-03-09T15:14:14.439856+0000 2026-03-09T15:14:15.767 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 09 15:14:15 vm06 ceph-mon[56506]: created 2026-03-09T15:12:40.941469+0000 2026-03-09T15:14:15.767 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 09 15:14:15 vm06 ceph-mon[56506]: min_mon_release 19 (squid) 2026-03-09T15:14:15.767 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 09 15:14:15 vm06 ceph-mon[56506]: election_strategy: 1 2026-03-09T15:14:15.767 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 09 15:14:15 vm06 ceph-mon[56506]: 0: [v2:192.168.123.103:3300/0,v1:192.168.123.103:6789/0] mon.a; crush_location {datacenter=a} 2026-03-09T15:14:15.767 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 09 15:14:15 vm06 ceph-mon[56506]: 1: [v2:192.168.123.106:3300/0,v1:192.168.123.106:6789/0] mon.c; crush_location {datacenter=a,rack=3} 2026-03-09T15:14:15.767 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 09 15:14:15 vm06 ceph-mon[56506]: 2: [v2:192.168.123.104:3300/0,v1:192.168.123.104:6789/0] mon.b; crush_location {datacenter=b,rack=2} 2026-03-09T15:14:15.767 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 09 15:14:15 vm06 ceph-mon[56506]: fsmap 2026-03-09T15:14:15.767 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 09 15:14:15 vm06 ceph-mon[56506]: osdmap e19: 3 total, 3 up, 3 in 2026-03-09T15:14:15.767 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 09 15:14:15 vm06 ceph-mon[56506]: mgrmap e14: a(active, since 71s), standbys: b 2026-03-09T15:14:15.767 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 09 15:14:15 vm06 ceph-mon[56506]: overall HEALTH_OK 2026-03-09T15:14:15.767 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 09 15:14:15 vm06 ceph-mon[56506]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "a"}]: dispatch 2026-03-09T15:14:15.767 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 09 15:14:15 vm06 ceph-mon[56506]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-09T15:14:15.767 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 09 15:14:15 vm06 ceph-mon[56506]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "c"}]: dispatch 2026-03-09T15:14:15.767 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 09 15:14:15 vm06 ceph-mon[56506]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' 2026-03-09T15:14:15.767 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 09 15:14:15 vm06 ceph-mon[56506]: pgmap v45: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-09T15:14:15.767 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 09 15:14:15 vm06 ceph-mon[56506]: Reconfiguring mon.a (monmap changed)... 2026-03-09T15:14:15.767 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 09 15:14:15 vm06 ceph-mon[56506]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "mon."}]: dispatch 2026-03-09T15:14:15.768 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 09 15:14:15 vm06 ceph-mon[56506]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' cmd=[{"prefix": "config get", "who": "mon", "key": "public_network"}]: dispatch 2026-03-09T15:14:15.768 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 09 15:14:15 vm06 ceph-mon[56506]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T15:14:15.768 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 09 15:14:15 vm06 ceph-mon[56506]: Reconfiguring daemon mon.a on vm03 2026-03-09T15:14:15.768 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 09 15:14:15 vm06 ceph-mon[56506]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' 2026-03-09T15:14:15.768 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 09 15:14:15 vm06 ceph-mon[56506]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' 2026-03-09T15:14:15.768 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 09 15:14:15 vm06 ceph-mon[56506]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' cmd=[{"prefix": "auth get-or-create", "entity": "mgr.a", "caps": ["mon", "profile mgr", "osd", "allow *", "mds", "allow *"]}]: dispatch 2026-03-09T15:14:15.768 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 09 15:14:15 vm06 ceph-mon[56506]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' cmd=[{"prefix": "mgr services"}]: dispatch 2026-03-09T15:14:15.768 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 09 15:14:15 vm06 ceph-mon[56506]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T15:14:15.768 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 09 15:14:15 vm06 ceph-mon[56506]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' 2026-03-09T15:14:15.768 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 09 15:14:15 vm06 ceph-mon[56506]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' 2026-03-09T15:14:15.768 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 09 15:14:15 vm06 ceph-mon[56506]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "osd.0"}]: dispatch 2026-03-09T15:14:15.768 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 09 15:14:15 vm06 ceph-mon[56506]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T15:14:15.768 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 09 15:14:15 vm06 ceph-mon[56506]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' 2026-03-09T15:14:15.768 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 09 15:14:15 vm06 ceph-mon[56506]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' 2026-03-09T15:14:15.768 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 09 15:14:15 vm06 ceph-mon[56506]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "mon."}]: dispatch 2026-03-09T15:14:15.768 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 09 15:14:15 vm06 ceph-mon[56506]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' cmd=[{"prefix": "config get", "who": "mon", "key": "public_network"}]: dispatch 2026-03-09T15:14:15.768 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 09 15:14:15 vm06 ceph-mon[56506]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T15:14:16.471 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 15:14:16 vm04 ceph-mon[53998]: Reconfiguring mgr.a (monmap changed)... 2026-03-09T15:14:16.471 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 15:14:16 vm04 ceph-mon[53998]: Reconfiguring daemon mgr.a on vm03 2026-03-09T15:14:16.471 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 15:14:16 vm04 ceph-mon[53998]: Reconfiguring osd.0 (monmap changed)... 2026-03-09T15:14:16.471 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 15:14:16 vm04 ceph-mon[53998]: Reconfiguring daemon osd.0 on vm03 2026-03-09T15:14:16.471 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 15:14:16 vm04 ceph-mon[53998]: Reconfiguring mon.b (monmap changed)... 2026-03-09T15:14:16.471 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 15:14:16 vm04 ceph-mon[53998]: Reconfiguring daemon mon.b on vm04 2026-03-09T15:14:16.471 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 15:14:16 vm04 ceph-mon[53998]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' 2026-03-09T15:14:16.471 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 15:14:16 vm04 ceph-mon[53998]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' 2026-03-09T15:14:16.471 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 15:14:16 vm04 ceph-mon[53998]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' cmd=[{"prefix": "auth get-or-create", "entity": "mgr.b", "caps": ["mon", "profile mgr", "osd", "allow *", "mds", "allow *"]}]: dispatch 2026-03-09T15:14:16.471 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 15:14:16 vm04 ceph-mon[53998]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' cmd=[{"prefix": "mgr services"}]: dispatch 2026-03-09T15:14:16.471 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 15:14:16 vm04 ceph-mon[53998]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T15:14:16.471 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 15:14:16 vm04 ceph-mon[53998]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' 2026-03-09T15:14:16.471 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 15:14:16 vm04 ceph-mon[53998]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' 2026-03-09T15:14:16.471 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 15:14:16 vm04 ceph-mon[53998]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "osd.1"}]: dispatch 2026-03-09T15:14:16.471 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 15:14:16 vm04 ceph-mon[53998]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T15:14:16.471 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 15:14:16 vm04 ceph-mon[53998]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' 2026-03-09T15:14:16.471 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 15:14:16 vm04 ceph-mon[53998]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' 2026-03-09T15:14:16.471 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 15:14:16 vm04 ceph-mon[53998]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "mon."}]: dispatch 2026-03-09T15:14:16.471 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 15:14:16 vm04 ceph-mon[53998]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' cmd=[{"prefix": "config get", "who": "mon", "key": "public_network"}]: dispatch 2026-03-09T15:14:16.471 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 15:14:16 vm04 ceph-mon[53998]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T15:14:16.471 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 15:14:16 vm04 ceph-mon[53998]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' 2026-03-09T15:14:16.471 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 15:14:16 vm04 ceph-mon[53998]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' 2026-03-09T15:14:16.471 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 15:14:16 vm04 ceph-mon[53998]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "osd.2"}]: dispatch 2026-03-09T15:14:16.471 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 15:14:16 vm04 ceph-mon[53998]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T15:14:16.518 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 09 15:14:16 vm06 ceph-mon[56506]: Reconfiguring mgr.a (monmap changed)... 2026-03-09T15:14:16.518 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 09 15:14:16 vm06 ceph-mon[56506]: Reconfiguring daemon mgr.a on vm03 2026-03-09T15:14:16.518 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 09 15:14:16 vm06 ceph-mon[56506]: Reconfiguring osd.0 (monmap changed)... 2026-03-09T15:14:16.518 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 09 15:14:16 vm06 ceph-mon[56506]: Reconfiguring daemon osd.0 on vm03 2026-03-09T15:14:16.518 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 09 15:14:16 vm06 ceph-mon[56506]: Reconfiguring mon.b (monmap changed)... 2026-03-09T15:14:16.518 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 09 15:14:16 vm06 ceph-mon[56506]: Reconfiguring daemon mon.b on vm04 2026-03-09T15:14:16.518 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 09 15:14:16 vm06 ceph-mon[56506]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' 2026-03-09T15:14:16.518 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 09 15:14:16 vm06 ceph-mon[56506]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' 2026-03-09T15:14:16.518 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 09 15:14:16 vm06 ceph-mon[56506]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' cmd=[{"prefix": "auth get-or-create", "entity": "mgr.b", "caps": ["mon", "profile mgr", "osd", "allow *", "mds", "allow *"]}]: dispatch 2026-03-09T15:14:16.518 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 09 15:14:16 vm06 ceph-mon[56506]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' cmd=[{"prefix": "mgr services"}]: dispatch 2026-03-09T15:14:16.518 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 09 15:14:16 vm06 ceph-mon[56506]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T15:14:16.518 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 09 15:14:16 vm06 ceph-mon[56506]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' 2026-03-09T15:14:16.518 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 09 15:14:16 vm06 ceph-mon[56506]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' 2026-03-09T15:14:16.518 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 09 15:14:16 vm06 ceph-mon[56506]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "osd.1"}]: dispatch 2026-03-09T15:14:16.518 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 09 15:14:16 vm06 ceph-mon[56506]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T15:14:16.518 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 09 15:14:16 vm06 ceph-mon[56506]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' 2026-03-09T15:14:16.518 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 09 15:14:16 vm06 ceph-mon[56506]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' 2026-03-09T15:14:16.518 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 09 15:14:16 vm06 ceph-mon[56506]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "mon."}]: dispatch 2026-03-09T15:14:16.518 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 09 15:14:16 vm06 ceph-mon[56506]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' cmd=[{"prefix": "config get", "who": "mon", "key": "public_network"}]: dispatch 2026-03-09T15:14:16.518 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 09 15:14:16 vm06 ceph-mon[56506]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T15:14:16.518 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 09 15:14:16 vm06 ceph-mon[56506]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' 2026-03-09T15:14:16.518 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 09 15:14:16 vm06 ceph-mon[56506]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' 2026-03-09T15:14:16.518 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 09 15:14:16 vm06 ceph-mon[56506]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "osd.2"}]: dispatch 2026-03-09T15:14:16.518 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 09 15:14:16 vm06 ceph-mon[56506]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T15:14:16.799 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:14:16 vm03 ceph-mon[51762]: Reconfiguring mgr.a (monmap changed)... 2026-03-09T15:14:16.799 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:14:16 vm03 ceph-mon[51762]: Reconfiguring daemon mgr.a on vm03 2026-03-09T15:14:16.799 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:14:16 vm03 ceph-mon[51762]: Reconfiguring osd.0 (monmap changed)... 2026-03-09T15:14:16.799 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:14:16 vm03 ceph-mon[51762]: Reconfiguring daemon osd.0 on vm03 2026-03-09T15:14:16.799 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:14:16 vm03 ceph-mon[51762]: Reconfiguring mon.b (monmap changed)... 2026-03-09T15:14:16.799 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:14:16 vm03 ceph-mon[51762]: Reconfiguring daemon mon.b on vm04 2026-03-09T15:14:16.799 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:14:16 vm03 ceph-mon[51762]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' 2026-03-09T15:14:16.799 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:14:16 vm03 ceph-mon[51762]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' 2026-03-09T15:14:16.799 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:14:16 vm03 ceph-mon[51762]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' cmd=[{"prefix": "auth get-or-create", "entity": "mgr.b", "caps": ["mon", "profile mgr", "osd", "allow *", "mds", "allow *"]}]: dispatch 2026-03-09T15:14:16.799 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:14:16 vm03 ceph-mon[51762]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' cmd=[{"prefix": "mgr services"}]: dispatch 2026-03-09T15:14:16.799 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:14:16 vm03 ceph-mon[51762]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T15:14:16.800 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:14:16 vm03 ceph-mon[51762]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' 2026-03-09T15:14:16.800 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:14:16 vm03 ceph-mon[51762]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' 2026-03-09T15:14:16.800 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:14:16 vm03 ceph-mon[51762]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "osd.1"}]: dispatch 2026-03-09T15:14:16.800 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:14:16 vm03 ceph-mon[51762]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T15:14:16.800 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:14:16 vm03 ceph-mon[51762]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' 2026-03-09T15:14:16.800 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:14:16 vm03 ceph-mon[51762]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' 2026-03-09T15:14:16.800 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:14:16 vm03 ceph-mon[51762]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "mon."}]: dispatch 2026-03-09T15:14:16.800 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:14:16 vm03 ceph-mon[51762]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' cmd=[{"prefix": "config get", "who": "mon", "key": "public_network"}]: dispatch 2026-03-09T15:14:16.800 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:14:16 vm03 ceph-mon[51762]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T15:14:16.800 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:14:16 vm03 ceph-mon[51762]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' 2026-03-09T15:14:16.800 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:14:16 vm03 ceph-mon[51762]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' 2026-03-09T15:14:16.800 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:14:16 vm03 ceph-mon[51762]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "osd.2"}]: dispatch 2026-03-09T15:14:16.800 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:14:16 vm03 ceph-mon[51762]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T15:14:17.751 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 15:14:17 vm04 ceph-mon[53998]: Reconfiguring mgr.b (monmap changed)... 2026-03-09T15:14:17.751 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 15:14:17 vm04 ceph-mon[53998]: Reconfiguring daemon mgr.b on vm04 2026-03-09T15:14:17.751 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 15:14:17 vm04 ceph-mon[53998]: Reconfiguring osd.1 (monmap changed)... 2026-03-09T15:14:17.751 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 15:14:17 vm04 ceph-mon[53998]: Reconfiguring daemon osd.1 on vm04 2026-03-09T15:14:17.751 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 15:14:17 vm04 ceph-mon[53998]: Reconfiguring mon.c (monmap changed)... 2026-03-09T15:14:17.751 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 15:14:17 vm04 ceph-mon[53998]: Reconfiguring daemon mon.c on vm06 2026-03-09T15:14:17.751 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 15:14:17 vm04 ceph-mon[53998]: Reconfiguring osd.2 (monmap changed)... 2026-03-09T15:14:17.751 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 15:14:17 vm04 ceph-mon[53998]: Reconfiguring daemon osd.2 on vm06 2026-03-09T15:14:17.751 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 15:14:17 vm04 ceph-mon[53998]: pgmap v46: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-09T15:14:17.751 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 15:14:17 vm04 ceph-mon[53998]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' 2026-03-09T15:14:17.751 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 15:14:17 vm04 ceph-mon[53998]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' 2026-03-09T15:14:17.767 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 09 15:14:17 vm06 ceph-mon[56506]: Reconfiguring mgr.b (monmap changed)... 2026-03-09T15:14:17.767 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 09 15:14:17 vm06 ceph-mon[56506]: Reconfiguring daemon mgr.b on vm04 2026-03-09T15:14:17.767 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 09 15:14:17 vm06 ceph-mon[56506]: Reconfiguring osd.1 (monmap changed)... 2026-03-09T15:14:17.767 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 09 15:14:17 vm06 ceph-mon[56506]: Reconfiguring daemon osd.1 on vm04 2026-03-09T15:14:17.767 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 09 15:14:17 vm06 ceph-mon[56506]: Reconfiguring mon.c (monmap changed)... 2026-03-09T15:14:17.767 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 09 15:14:17 vm06 ceph-mon[56506]: Reconfiguring daemon mon.c on vm06 2026-03-09T15:14:17.767 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 09 15:14:17 vm06 ceph-mon[56506]: Reconfiguring osd.2 (monmap changed)... 2026-03-09T15:14:17.767 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 09 15:14:17 vm06 ceph-mon[56506]: Reconfiguring daemon osd.2 on vm06 2026-03-09T15:14:17.767 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 09 15:14:17 vm06 ceph-mon[56506]: pgmap v46: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-09T15:14:17.767 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 09 15:14:17 vm06 ceph-mon[56506]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' 2026-03-09T15:14:17.767 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 09 15:14:17 vm06 ceph-mon[56506]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' 2026-03-09T15:14:17.799 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:14:17 vm03 ceph-mon[51762]: Reconfiguring mgr.b (monmap changed)... 2026-03-09T15:14:17.799 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:14:17 vm03 ceph-mon[51762]: Reconfiguring daemon mgr.b on vm04 2026-03-09T15:14:17.799 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:14:17 vm03 ceph-mon[51762]: Reconfiguring osd.1 (monmap changed)... 2026-03-09T15:14:17.799 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:14:17 vm03 ceph-mon[51762]: Reconfiguring daemon osd.1 on vm04 2026-03-09T15:14:17.799 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:14:17 vm03 ceph-mon[51762]: Reconfiguring mon.c (monmap changed)... 2026-03-09T15:14:17.799 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:14:17 vm03 ceph-mon[51762]: Reconfiguring daemon mon.c on vm06 2026-03-09T15:14:17.799 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:14:17 vm03 ceph-mon[51762]: Reconfiguring osd.2 (monmap changed)... 2026-03-09T15:14:17.799 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:14:17 vm03 ceph-mon[51762]: Reconfiguring daemon osd.2 on vm06 2026-03-09T15:14:17.799 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:14:17 vm03 ceph-mon[51762]: pgmap v46: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-09T15:14:17.799 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:14:17 vm03 ceph-mon[51762]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' 2026-03-09T15:14:17.799 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:14:17 vm03 ceph-mon[51762]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' 2026-03-09T15:14:19.001 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 15:14:18 vm04 ceph-mon[53998]: pgmap v47: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-09T15:14:19.017 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 09 15:14:18 vm06 ceph-mon[56506]: pgmap v47: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-09T15:14:19.049 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:14:18 vm03 ceph-mon[51762]: pgmap v47: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-09T15:14:21.001 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 15:14:20 vm04 ceph-mon[53998]: pgmap v48: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-09T15:14:21.017 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 09 15:14:20 vm06 ceph-mon[56506]: pgmap v48: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-09T15:14:21.049 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:14:20 vm03 ceph-mon[51762]: pgmap v48: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-09T15:14:23.001 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 15:14:22 vm04 ceph-mon[53998]: pgmap v49: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-09T15:14:23.017 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 09 15:14:22 vm06 ceph-mon[56506]: pgmap v49: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-09T15:14:23.049 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:14:22 vm03 ceph-mon[51762]: pgmap v49: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-09T15:14:25.001 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 15:14:24 vm04 ceph-mon[53998]: pgmap v50: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-09T15:14:25.017 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 09 15:14:24 vm06 ceph-mon[56506]: pgmap v50: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-09T15:14:25.049 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:14:24 vm03 ceph-mon[51762]: pgmap v50: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-09T15:14:27.001 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 15:14:26 vm04 ceph-mon[53998]: pgmap v51: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-09T15:14:27.017 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 09 15:14:26 vm06 ceph-mon[56506]: pgmap v51: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-09T15:14:27.049 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:14:26 vm03 ceph-mon[51762]: pgmap v51: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-09T15:14:29.001 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 15:14:28 vm04 ceph-mon[53998]: pgmap v52: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-09T15:14:29.017 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 09 15:14:28 vm06 ceph-mon[56506]: pgmap v52: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-09T15:14:29.049 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:14:28 vm03 ceph-mon[51762]: pgmap v52: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-09T15:14:31.001 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 15:14:30 vm04 ceph-mon[53998]: pgmap v53: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-09T15:14:31.017 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 09 15:14:30 vm06 ceph-mon[56506]: pgmap v53: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-09T15:14:31.049 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:14:30 vm03 ceph-mon[51762]: pgmap v53: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-09T15:14:33.001 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 15:14:32 vm04 ceph-mon[53998]: pgmap v54: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-09T15:14:33.017 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 09 15:14:32 vm06 ceph-mon[56506]: pgmap v54: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-09T15:14:33.049 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:14:32 vm03 ceph-mon[51762]: pgmap v54: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-09T15:14:35.001 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 15:14:34 vm04 ceph-mon[53998]: pgmap v55: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-09T15:14:35.017 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 09 15:14:34 vm06 ceph-mon[56506]: pgmap v55: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-09T15:14:35.049 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:14:34 vm03 ceph-mon[51762]: pgmap v55: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-09T15:14:37.001 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 15:14:36 vm04 ceph-mon[53998]: pgmap v56: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-09T15:14:37.017 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 09 15:14:36 vm06 ceph-mon[56506]: pgmap v56: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-09T15:14:37.049 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:14:36 vm03 ceph-mon[51762]: pgmap v56: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-09T15:14:39.001 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 15:14:38 vm04 ceph-mon[53998]: pgmap v57: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-09T15:14:39.017 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 09 15:14:38 vm06 ceph-mon[56506]: pgmap v57: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-09T15:14:39.049 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:14:38 vm03 ceph-mon[51762]: pgmap v57: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-09T15:14:41.001 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 15:14:40 vm04 ceph-mon[53998]: pgmap v58: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-09T15:14:41.017 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 09 15:14:40 vm06 ceph-mon[56506]: pgmap v58: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-09T15:14:41.049 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:14:40 vm03 ceph-mon[51762]: pgmap v58: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-09T15:14:43.001 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 15:14:42 vm04 ceph-mon[53998]: pgmap v59: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-09T15:14:43.017 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 09 15:14:42 vm06 ceph-mon[56506]: pgmap v59: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-09T15:14:43.049 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:14:42 vm03 ceph-mon[51762]: pgmap v59: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-09T15:14:45.001 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 15:14:44 vm04 ceph-mon[53998]: pgmap v60: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-09T15:14:45.017 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 09 15:14:44 vm06 ceph-mon[56506]: pgmap v60: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-09T15:14:45.049 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:14:44 vm03 ceph-mon[51762]: pgmap v60: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-09T15:14:47.001 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 15:14:46 vm04 ceph-mon[53998]: pgmap v61: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-09T15:14:47.017 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 09 15:14:46 vm06 ceph-mon[56506]: pgmap v61: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-09T15:14:47.049 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:14:46 vm03 ceph-mon[51762]: pgmap v61: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-09T15:14:48.956 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 15:14:48 vm04 ceph-mon[53998]: pgmap v62: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-09T15:14:49.017 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 09 15:14:48 vm06 ceph-mon[56506]: pgmap v62: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-09T15:14:49.049 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:14:48 vm03 ceph-mon[51762]: pgmap v62: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-09T15:14:51.001 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 15:14:50 vm04 ceph-mon[53998]: pgmap v63: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-09T15:14:51.017 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 09 15:14:50 vm06 ceph-mon[56506]: pgmap v63: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-09T15:14:51.049 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:14:50 vm03 ceph-mon[51762]: pgmap v63: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-09T15:14:53.001 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 15:14:52 vm04 ceph-mon[53998]: pgmap v64: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-09T15:14:53.017 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 09 15:14:52 vm06 ceph-mon[56506]: pgmap v64: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-09T15:14:53.049 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:14:52 vm03 ceph-mon[51762]: pgmap v64: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-09T15:14:55.001 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 15:14:54 vm04 ceph-mon[53998]: pgmap v65: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-09T15:14:55.017 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 09 15:14:54 vm06 ceph-mon[56506]: pgmap v65: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-09T15:14:55.049 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:14:54 vm03 ceph-mon[51762]: pgmap v65: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-09T15:14:57.001 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 15:14:56 vm04 ceph-mon[53998]: pgmap v66: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-09T15:14:57.017 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 09 15:14:56 vm06 ceph-mon[56506]: pgmap v66: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-09T15:14:57.049 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:14:56 vm03 ceph-mon[51762]: pgmap v66: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-09T15:14:59.001 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 15:14:58 vm04 ceph-mon[53998]: pgmap v67: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-09T15:14:59.017 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 09 15:14:58 vm06 ceph-mon[56506]: pgmap v67: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-09T15:14:59.049 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:14:58 vm03 ceph-mon[51762]: pgmap v67: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-09T15:15:01.001 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 15:15:00 vm04 ceph-mon[53998]: pgmap v68: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-09T15:15:01.017 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 09 15:15:00 vm06 ceph-mon[56506]: pgmap v68: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-09T15:15:01.049 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:15:00 vm03 ceph-mon[51762]: pgmap v68: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-09T15:15:03.001 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 15:15:02 vm04 ceph-mon[53998]: pgmap v69: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-09T15:15:03.017 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 09 15:15:02 vm06 ceph-mon[56506]: pgmap v69: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-09T15:15:03.049 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:15:02 vm03 ceph-mon[51762]: pgmap v69: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-09T15:15:05.001 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 15:15:04 vm04 ceph-mon[53998]: pgmap v70: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-09T15:15:05.017 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 09 15:15:04 vm06 ceph-mon[56506]: pgmap v70: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-09T15:15:05.049 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:15:04 vm03 ceph-mon[51762]: pgmap v70: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-09T15:15:07.001 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 15:15:06 vm04 ceph-mon[53998]: pgmap v71: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-09T15:15:07.017 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 09 15:15:06 vm06 ceph-mon[56506]: pgmap v71: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-09T15:15:07.049 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:15:06 vm03 ceph-mon[51762]: pgmap v71: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-09T15:15:09.001 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 15:15:08 vm04 ceph-mon[53998]: pgmap v72: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-09T15:15:09.017 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 09 15:15:08 vm06 ceph-mon[56506]: pgmap v72: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-09T15:15:09.049 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:15:08 vm03 ceph-mon[51762]: pgmap v72: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-09T15:15:11.001 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 15:15:10 vm04 ceph-mon[53998]: pgmap v73: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-09T15:15:11.017 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 09 15:15:10 vm06 ceph-mon[56506]: pgmap v73: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-09T15:15:11.049 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:15:10 vm03 ceph-mon[51762]: pgmap v73: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-09T15:15:13.001 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 15:15:12 vm04 ceph-mon[53998]: pgmap v74: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-09T15:15:13.017 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 09 15:15:12 vm06 ceph-mon[56506]: pgmap v74: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-09T15:15:13.049 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:15:12 vm03 ceph-mon[51762]: pgmap v74: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-09T15:15:15.001 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 15:15:14 vm04 ceph-mon[53998]: pgmap v75: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-09T15:15:15.017 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 09 15:15:14 vm06 ceph-mon[56506]: pgmap v75: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-09T15:15:15.049 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:15:14 vm03 ceph-mon[51762]: pgmap v75: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-09T15:15:17.001 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 15:15:16 vm04 ceph-mon[53998]: pgmap v76: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-09T15:15:17.017 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 09 15:15:16 vm06 ceph-mon[56506]: pgmap v76: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-09T15:15:17.049 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:15:16 vm03 ceph-mon[51762]: pgmap v76: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-09T15:15:18.001 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 15:15:17 vm04 ceph-mon[53998]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T15:15:18.001 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 15:15:17 vm04 ceph-mon[53998]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T15:15:18.001 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 15:15:17 vm04 ceph-mon[53998]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T15:15:18.001 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 15:15:17 vm04 ceph-mon[53998]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' 2026-03-09T15:15:18.017 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 09 15:15:17 vm06 ceph-mon[56506]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T15:15:18.017 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 09 15:15:17 vm06 ceph-mon[56506]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T15:15:18.017 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 09 15:15:17 vm06 ceph-mon[56506]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T15:15:18.017 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 09 15:15:17 vm06 ceph-mon[56506]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' 2026-03-09T15:15:18.049 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:15:17 vm03 ceph-mon[51762]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T15:15:18.049 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:15:17 vm03 ceph-mon[51762]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T15:15:18.049 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:15:17 vm03 ceph-mon[51762]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T15:15:18.049 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:15:17 vm03 ceph-mon[51762]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' 2026-03-09T15:15:19.001 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 15:15:18 vm04 ceph-mon[53998]: pgmap v77: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-09T15:15:19.017 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 09 15:15:18 vm06 ceph-mon[56506]: pgmap v77: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-09T15:15:19.049 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:15:18 vm03 ceph-mon[51762]: pgmap v77: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-09T15:15:21.001 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 15:15:20 vm04 ceph-mon[53998]: pgmap v78: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-09T15:15:21.017 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 09 15:15:20 vm06 ceph-mon[56506]: pgmap v78: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-09T15:15:21.049 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:15:20 vm03 ceph-mon[51762]: pgmap v78: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-09T15:15:23.001 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 15:15:22 vm04 ceph-mon[53998]: pgmap v79: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-09T15:15:23.017 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 09 15:15:22 vm06 ceph-mon[56506]: pgmap v79: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-09T15:15:23.049 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:15:22 vm03 ceph-mon[51762]: pgmap v79: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-09T15:15:25.001 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 15:15:24 vm04 ceph-mon[53998]: pgmap v80: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-09T15:15:25.017 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 09 15:15:24 vm06 ceph-mon[56506]: pgmap v80: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-09T15:15:25.049 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:15:24 vm03 ceph-mon[51762]: pgmap v80: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-09T15:15:27.001 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 15:15:26 vm04 ceph-mon[53998]: pgmap v81: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-09T15:15:27.017 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 09 15:15:26 vm06 ceph-mon[56506]: pgmap v81: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-09T15:15:27.049 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:15:26 vm03 ceph-mon[51762]: pgmap v81: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-09T15:15:29.001 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 15:15:28 vm04 ceph-mon[53998]: pgmap v82: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-09T15:15:29.017 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 09 15:15:28 vm06 ceph-mon[56506]: pgmap v82: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-09T15:15:29.049 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:15:28 vm03 ceph-mon[51762]: pgmap v82: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-09T15:15:31.001 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 15:15:30 vm04 ceph-mon[53998]: pgmap v83: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-09T15:15:31.017 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 09 15:15:30 vm06 ceph-mon[56506]: pgmap v83: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-09T15:15:31.049 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:15:30 vm03 ceph-mon[51762]: pgmap v83: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-09T15:15:33.001 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 15:15:32 vm04 ceph-mon[53998]: pgmap v84: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-09T15:15:33.017 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 09 15:15:32 vm06 ceph-mon[56506]: pgmap v84: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-09T15:15:33.049 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:15:32 vm03 ceph-mon[51762]: pgmap v84: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-09T15:15:35.001 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 15:15:34 vm04 ceph-mon[53998]: pgmap v85: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-09T15:15:35.017 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 09 15:15:34 vm06 ceph-mon[56506]: pgmap v85: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-09T15:15:35.049 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:15:34 vm03 ceph-mon[51762]: pgmap v85: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-09T15:15:37.001 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 15:15:36 vm04 ceph-mon[53998]: pgmap v86: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-09T15:15:37.017 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 09 15:15:36 vm06 ceph-mon[56506]: pgmap v86: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-09T15:15:37.049 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:15:36 vm03 ceph-mon[51762]: pgmap v86: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-09T15:15:39.001 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 15:15:38 vm04 ceph-mon[53998]: pgmap v87: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-09T15:15:39.017 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 09 15:15:38 vm06 ceph-mon[56506]: pgmap v87: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-09T15:15:39.049 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:15:38 vm03 ceph-mon[51762]: pgmap v87: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-09T15:15:39.407 INFO:teuthology.orchestra.run.vm03.stderr:+ ceph orch ps --refresh 2026-03-09T15:15:39.556 INFO:teuthology.orchestra.run.vm03.stdout:NAME HOST PORTS STATUS REFRESHED AGE MEM USE MEM LIM VERSION IMAGE ID CONTAINER ID 2026-03-09T15:15:39.556 INFO:teuthology.orchestra.run.vm03.stdout:mgr.a vm03 *:9283,8765 running (2m) 2m ago 2m 536M - 19.2.3-678-ge911bdeb 654f31e6858e b94230b6b9cf 2026-03-09T15:15:39.556 INFO:teuthology.orchestra.run.vm03.stdout:mgr.b vm04 *:8443,8765 running (2m) 112s ago 2m 487M - 19.2.3-678-ge911bdeb 654f31e6858e 518720cdb992 2026-03-09T15:15:39.556 INFO:teuthology.orchestra.run.vm03.stdout:mon.a vm03 running (2m) 2m ago 2m 43.3M 2048M 19.2.3-678-ge911bdeb 654f31e6858e 6be9db7b0806 2026-03-09T15:15:39.556 INFO:teuthology.orchestra.run.vm03.stdout:mon.b vm04 running (2m) 112s ago 2m 35.4M 2048M 19.2.3-678-ge911bdeb 654f31e6858e ea2a9cb34548 2026-03-09T15:15:39.556 INFO:teuthology.orchestra.run.vm03.stdout:mon.c vm06 running (2m) 102s ago 2m 37.3M 2048M 19.2.3-678-ge911bdeb 654f31e6858e 9cba3f0dfa7c 2026-03-09T15:15:39.556 INFO:teuthology.orchestra.run.vm03.stdout:osd.0 vm03 running (2m) 2m ago 2m 13.0M 4096M 19.2.3-678-ge911bdeb 654f31e6858e 646ad6643c08 2026-03-09T15:15:39.556 INFO:teuthology.orchestra.run.vm03.stdout:osd.1 vm04 running (112s) 112s ago 112s 12.9M 4096M 19.2.3-678-ge911bdeb 654f31e6858e 91cd0ade9d3c 2026-03-09T15:15:39.556 INFO:teuthology.orchestra.run.vm03.stdout:osd.2 vm06 running (103s) 102s ago 103s 12.8M 4353M 19.2.3-678-ge911bdeb 654f31e6858e 72ec1ee0f143 2026-03-09T15:15:39.573 INFO:teuthology.orchestra.run.vm03.stderr:+ ceph orch ls --service-name mon --export 2026-03-09T15:15:39.755 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:15:39 vm03 ceph-mon[51762]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T15:15:39.755 INFO:teuthology.orchestra.run.vm03.stderr:+ ceph orch apply -i mon.yaml 2026-03-09T15:15:39.842 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 15:15:39 vm04 ceph-mon[53998]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T15:15:39.847 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 09 15:15:39 vm06 ceph-mon[56506]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T15:15:39.938 INFO:teuthology.orchestra.run.vm03.stdout:Scheduled mon update... 2026-03-09T15:15:39.949 INFO:teuthology.orchestra.run.vm03.stderr:+ sleep 90 2026-03-09T15:15:41.251 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 15:15:40 vm04 ceph-mon[53998]: from='client.14439 -' entity='client.admin' cmd=[{"prefix": "orch ps", "refresh": true, "target": ["mon-mgr", ""]}]: dispatch 2026-03-09T15:15:41.251 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 15:15:40 vm04 ceph-mon[53998]: from='client.24284 -' entity='client.admin' cmd=[{"prefix": "orch ls", "service_name": "mon", "export": true, "target": ["mon-mgr", ""]}]: dispatch 2026-03-09T15:15:41.251 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 15:15:40 vm04 ceph-mon[53998]: from='client.24290 -' entity='client.admin' cmd=[{"prefix": "orch apply", "target": ["mon-mgr", ""]}]: dispatch 2026-03-09T15:15:41.251 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 15:15:40 vm04 ceph-mon[53998]: Saving service mon spec with placement count:3 2026-03-09T15:15:41.251 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 15:15:40 vm04 ceph-mon[53998]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' 2026-03-09T15:15:41.251 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 15:15:40 vm04 ceph-mon[53998]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' 2026-03-09T15:15:41.251 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 15:15:40 vm04 ceph-mon[53998]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' 2026-03-09T15:15:41.251 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 15:15:40 vm04 ceph-mon[53998]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' 2026-03-09T15:15:41.251 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 15:15:40 vm04 ceph-mon[53998]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' 2026-03-09T15:15:41.251 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 15:15:40 vm04 ceph-mon[53998]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' 2026-03-09T15:15:41.251 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 15:15:40 vm04 ceph-mon[53998]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' 2026-03-09T15:15:41.251 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 15:15:40 vm04 ceph-mon[53998]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T15:15:41.251 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 15:15:40 vm04 ceph-mon[53998]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T15:15:41.251 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 15:15:40 vm04 ceph-mon[53998]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' 2026-03-09T15:15:41.251 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 15:15:40 vm04 ceph-mon[53998]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' cmd=[{"prefix": "quorum_status"}]: dispatch 2026-03-09T15:15:41.251 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 15:15:40 vm04 ceph-mon[53998]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' 2026-03-09T15:15:41.251 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 15:15:40 vm04 ceph-mon[53998]: pgmap v88: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-09T15:15:41.251 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 15:15:40 vm04 ceph-mon[53998]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T15:15:41.251 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 15:15:40 vm04 ceph-mon[53998]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T15:15:41.251 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 15:15:40 vm04 ceph-mon[53998]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T15:15:41.251 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 15:15:40 vm04 ceph-mon[53998]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' 2026-03-09T15:15:41.267 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 09 15:15:40 vm06 ceph-mon[56506]: from='client.14439 -' entity='client.admin' cmd=[{"prefix": "orch ps", "refresh": true, "target": ["mon-mgr", ""]}]: dispatch 2026-03-09T15:15:41.267 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 09 15:15:40 vm06 ceph-mon[56506]: from='client.24284 -' entity='client.admin' cmd=[{"prefix": "orch ls", "service_name": "mon", "export": true, "target": ["mon-mgr", ""]}]: dispatch 2026-03-09T15:15:41.267 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 09 15:15:40 vm06 ceph-mon[56506]: from='client.24290 -' entity='client.admin' cmd=[{"prefix": "orch apply", "target": ["mon-mgr", ""]}]: dispatch 2026-03-09T15:15:41.267 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 09 15:15:40 vm06 ceph-mon[56506]: Saving service mon spec with placement count:3 2026-03-09T15:15:41.267 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 09 15:15:40 vm06 ceph-mon[56506]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' 2026-03-09T15:15:41.267 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 09 15:15:40 vm06 ceph-mon[56506]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' 2026-03-09T15:15:41.267 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 09 15:15:40 vm06 ceph-mon[56506]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' 2026-03-09T15:15:41.267 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 09 15:15:40 vm06 ceph-mon[56506]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' 2026-03-09T15:15:41.267 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 09 15:15:40 vm06 ceph-mon[56506]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' 2026-03-09T15:15:41.267 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 09 15:15:40 vm06 ceph-mon[56506]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' 2026-03-09T15:15:41.267 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 09 15:15:40 vm06 ceph-mon[56506]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' 2026-03-09T15:15:41.267 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 09 15:15:40 vm06 ceph-mon[56506]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T15:15:41.267 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 09 15:15:40 vm06 ceph-mon[56506]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T15:15:41.267 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 09 15:15:40 vm06 ceph-mon[56506]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' 2026-03-09T15:15:41.267 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 09 15:15:40 vm06 ceph-mon[56506]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' cmd=[{"prefix": "quorum_status"}]: dispatch 2026-03-09T15:15:41.267 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 09 15:15:40 vm06 ceph-mon[56506]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' 2026-03-09T15:15:41.267 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 09 15:15:40 vm06 ceph-mon[56506]: pgmap v88: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-09T15:15:41.267 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 09 15:15:40 vm06 ceph-mon[56506]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T15:15:41.268 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 09 15:15:40 vm06 ceph-mon[56506]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T15:15:41.268 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 09 15:15:40 vm06 ceph-mon[56506]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T15:15:41.268 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 09 15:15:40 vm06 ceph-mon[56506]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' 2026-03-09T15:15:41.299 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:15:40 vm03 ceph-mon[51762]: from='client.14439 -' entity='client.admin' cmd=[{"prefix": "orch ps", "refresh": true, "target": ["mon-mgr", ""]}]: dispatch 2026-03-09T15:15:41.299 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:15:40 vm03 ceph-mon[51762]: from='client.24284 -' entity='client.admin' cmd=[{"prefix": "orch ls", "service_name": "mon", "export": true, "target": ["mon-mgr", ""]}]: dispatch 2026-03-09T15:15:41.299 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:15:40 vm03 ceph-mon[51762]: from='client.24290 -' entity='client.admin' cmd=[{"prefix": "orch apply", "target": ["mon-mgr", ""]}]: dispatch 2026-03-09T15:15:41.299 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:15:40 vm03 ceph-mon[51762]: Saving service mon spec with placement count:3 2026-03-09T15:15:41.299 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:15:40 vm03 ceph-mon[51762]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' 2026-03-09T15:15:41.299 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:15:40 vm03 ceph-mon[51762]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' 2026-03-09T15:15:41.299 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:15:40 vm03 ceph-mon[51762]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' 2026-03-09T15:15:41.299 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:15:40 vm03 ceph-mon[51762]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' 2026-03-09T15:15:41.299 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:15:40 vm03 ceph-mon[51762]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' 2026-03-09T15:15:41.299 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:15:40 vm03 ceph-mon[51762]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' 2026-03-09T15:15:41.299 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:15:40 vm03 ceph-mon[51762]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' 2026-03-09T15:15:41.299 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:15:40 vm03 ceph-mon[51762]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T15:15:41.299 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:15:40 vm03 ceph-mon[51762]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T15:15:41.300 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:15:40 vm03 ceph-mon[51762]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' 2026-03-09T15:15:41.300 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:15:40 vm03 ceph-mon[51762]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' cmd=[{"prefix": "quorum_status"}]: dispatch 2026-03-09T15:15:41.300 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:15:40 vm03 ceph-mon[51762]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' 2026-03-09T15:15:41.300 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:15:40 vm03 ceph-mon[51762]: pgmap v88: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-09T15:15:41.300 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:15:40 vm03 ceph-mon[51762]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T15:15:41.300 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:15:40 vm03 ceph-mon[51762]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T15:15:41.300 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:15:40 vm03 ceph-mon[51762]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T15:15:41.300 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:15:40 vm03 ceph-mon[51762]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' 2026-03-09T15:15:43.001 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 15:15:42 vm04 ceph-mon[53998]: pgmap v89: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-09T15:15:43.017 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 09 15:15:42 vm06 ceph-mon[56506]: pgmap v89: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-09T15:15:43.049 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:15:42 vm03 ceph-mon[51762]: pgmap v89: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-09T15:15:45.001 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 15:15:44 vm04 ceph-mon[53998]: pgmap v90: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-09T15:15:45.017 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 09 15:15:44 vm06 ceph-mon[56506]: pgmap v90: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-09T15:15:45.049 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:15:44 vm03 ceph-mon[51762]: pgmap v90: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-09T15:15:47.001 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 15:15:46 vm04 ceph-mon[53998]: pgmap v91: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-09T15:15:47.017 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 09 15:15:46 vm06 ceph-mon[56506]: pgmap v91: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-09T15:15:47.049 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:15:46 vm03 ceph-mon[51762]: pgmap v91: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-09T15:15:49.001 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 15:15:48 vm04 ceph-mon[53998]: pgmap v92: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-09T15:15:49.017 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 09 15:15:48 vm06 ceph-mon[56506]: pgmap v92: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-09T15:15:49.049 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:15:48 vm03 ceph-mon[51762]: pgmap v92: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-09T15:15:51.001 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 15:15:50 vm04 ceph-mon[53998]: pgmap v93: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-09T15:15:51.017 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 09 15:15:50 vm06 ceph-mon[56506]: pgmap v93: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-09T15:15:51.049 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:15:50 vm03 ceph-mon[51762]: pgmap v93: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-09T15:15:53.001 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 15:15:52 vm04 ceph-mon[53998]: pgmap v94: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-09T15:15:53.017 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 09 15:15:52 vm06 ceph-mon[56506]: pgmap v94: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-09T15:15:53.049 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:15:52 vm03 ceph-mon[51762]: pgmap v94: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-09T15:15:55.001 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 15:15:54 vm04 ceph-mon[53998]: pgmap v95: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-09T15:15:55.017 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 09 15:15:54 vm06 ceph-mon[56506]: pgmap v95: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-09T15:15:55.049 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:15:54 vm03 ceph-mon[51762]: pgmap v95: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-09T15:15:57.001 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 15:15:56 vm04 ceph-mon[53998]: pgmap v96: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-09T15:15:57.017 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 09 15:15:56 vm06 ceph-mon[56506]: pgmap v96: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-09T15:15:57.049 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:15:56 vm03 ceph-mon[51762]: pgmap v96: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-09T15:15:59.001 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 15:15:58 vm04 ceph-mon[53998]: pgmap v97: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-09T15:15:59.017 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 09 15:15:58 vm06 ceph-mon[56506]: pgmap v97: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-09T15:15:59.049 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:15:58 vm03 ceph-mon[51762]: pgmap v97: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-09T15:16:01.001 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 15:16:00 vm04 ceph-mon[53998]: pgmap v98: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-09T15:16:01.017 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 09 15:16:00 vm06 ceph-mon[56506]: pgmap v98: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-09T15:16:01.049 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:16:00 vm03 ceph-mon[51762]: pgmap v98: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-09T15:16:03.001 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 15:16:02 vm04 ceph-mon[53998]: pgmap v99: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-09T15:16:03.017 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 09 15:16:02 vm06 ceph-mon[56506]: pgmap v99: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-09T15:16:03.049 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:16:02 vm03 ceph-mon[51762]: pgmap v99: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-09T15:16:05.001 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 15:16:04 vm04 ceph-mon[53998]: pgmap v100: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-09T15:16:05.017 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 09 15:16:04 vm06 ceph-mon[56506]: pgmap v100: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-09T15:16:05.049 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:16:04 vm03 ceph-mon[51762]: pgmap v100: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-09T15:16:07.001 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 15:16:06 vm04 ceph-mon[53998]: pgmap v101: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-09T15:16:07.017 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 09 15:16:06 vm06 ceph-mon[56506]: pgmap v101: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-09T15:16:07.049 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:16:06 vm03 ceph-mon[51762]: pgmap v101: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-09T15:16:09.001 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 15:16:08 vm04 ceph-mon[53998]: pgmap v102: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-09T15:16:09.017 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 09 15:16:08 vm06 ceph-mon[56506]: pgmap v102: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-09T15:16:09.049 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:16:08 vm03 ceph-mon[51762]: pgmap v102: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-09T15:16:11.001 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 15:16:10 vm04 ceph-mon[53998]: pgmap v103: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-09T15:16:11.017 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 09 15:16:10 vm06 ceph-mon[56506]: pgmap v103: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-09T15:16:11.049 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:16:10 vm03 ceph-mon[51762]: pgmap v103: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-09T15:16:13.001 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 15:16:12 vm04 ceph-mon[53998]: pgmap v104: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-09T15:16:13.017 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 09 15:16:12 vm06 ceph-mon[56506]: pgmap v104: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-09T15:16:13.049 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:16:12 vm03 ceph-mon[51762]: pgmap v104: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-09T15:16:15.001 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 15:16:14 vm04 ceph-mon[53998]: pgmap v105: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-09T15:16:15.017 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 09 15:16:14 vm06 ceph-mon[56506]: pgmap v105: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-09T15:16:15.049 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:16:14 vm03 ceph-mon[51762]: pgmap v105: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-09T15:16:17.001 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 15:16:16 vm04 ceph-mon[53998]: pgmap v106: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-09T15:16:17.017 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 09 15:16:16 vm06 ceph-mon[56506]: pgmap v106: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-09T15:16:17.049 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:16:16 vm03 ceph-mon[51762]: pgmap v106: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-09T15:16:19.001 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 15:16:18 vm04 ceph-mon[53998]: pgmap v107: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-09T15:16:19.017 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 09 15:16:18 vm06 ceph-mon[56506]: pgmap v107: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-09T15:16:19.049 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:16:18 vm03 ceph-mon[51762]: pgmap v107: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-09T15:16:21.001 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 15:16:20 vm04 ceph-mon[53998]: pgmap v108: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-09T15:16:21.017 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 09 15:16:20 vm06 ceph-mon[56506]: pgmap v108: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-09T15:16:21.049 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:16:20 vm03 ceph-mon[51762]: pgmap v108: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-09T15:16:23.001 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 15:16:22 vm04 ceph-mon[53998]: pgmap v109: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-09T15:16:23.017 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 09 15:16:22 vm06 ceph-mon[56506]: pgmap v109: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-09T15:16:23.049 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:16:22 vm03 ceph-mon[51762]: pgmap v109: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-09T15:16:25.001 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 15:16:24 vm04 ceph-mon[53998]: pgmap v110: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-09T15:16:25.017 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 09 15:16:24 vm06 ceph-mon[56506]: pgmap v110: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-09T15:16:25.049 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:16:24 vm03 ceph-mon[51762]: pgmap v110: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-09T15:16:27.001 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 15:16:26 vm04 ceph-mon[53998]: pgmap v111: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-09T15:16:27.017 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 09 15:16:26 vm06 ceph-mon[56506]: pgmap v111: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-09T15:16:27.049 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:16:26 vm03 ceph-mon[51762]: pgmap v111: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-09T15:16:29.001 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 15:16:28 vm04 ceph-mon[53998]: pgmap v112: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-09T15:16:29.017 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 09 15:16:28 vm06 ceph-mon[56506]: pgmap v112: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-09T15:16:29.049 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:16:28 vm03 ceph-mon[51762]: pgmap v112: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-09T15:16:31.001 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 15:16:30 vm04 ceph-mon[53998]: pgmap v113: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-09T15:16:31.017 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 09 15:16:30 vm06 ceph-mon[56506]: pgmap v113: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-09T15:16:31.049 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:16:30 vm03 ceph-mon[51762]: pgmap v113: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-09T15:16:33.001 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 15:16:32 vm04 ceph-mon[53998]: pgmap v114: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-09T15:16:33.017 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 09 15:16:32 vm06 ceph-mon[56506]: pgmap v114: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-09T15:16:33.049 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:16:32 vm03 ceph-mon[51762]: pgmap v114: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-09T15:16:35.001 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 15:16:34 vm04 ceph-mon[53998]: pgmap v115: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-09T15:16:35.017 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 09 15:16:34 vm06 ceph-mon[56506]: pgmap v115: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-09T15:16:35.049 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:16:34 vm03 ceph-mon[51762]: pgmap v115: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-09T15:16:37.001 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 15:16:36 vm04 ceph-mon[53998]: pgmap v116: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-09T15:16:37.017 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 09 15:16:36 vm06 ceph-mon[56506]: pgmap v116: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-09T15:16:37.049 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:16:36 vm03 ceph-mon[51762]: pgmap v116: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-09T15:16:39.001 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 15:16:38 vm04 ceph-mon[53998]: pgmap v117: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-09T15:16:39.017 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 09 15:16:38 vm06 ceph-mon[56506]: pgmap v117: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-09T15:16:39.049 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:16:38 vm03 ceph-mon[51762]: pgmap v117: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-09T15:16:41.001 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 15:16:40 vm04 ceph-mon[53998]: pgmap v118: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-09T15:16:41.017 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 09 15:16:40 vm06 ceph-mon[56506]: pgmap v118: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-09T15:16:41.049 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:16:40 vm03 ceph-mon[51762]: pgmap v118: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-09T15:16:42.001 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 15:16:41 vm04 ceph-mon[53998]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T15:16:42.001 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 15:16:41 vm04 ceph-mon[53998]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T15:16:42.001 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 15:16:41 vm04 ceph-mon[53998]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T15:16:42.001 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 15:16:41 vm04 ceph-mon[53998]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' 2026-03-09T15:16:42.017 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 09 15:16:41 vm06 ceph-mon[56506]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T15:16:42.017 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 09 15:16:41 vm06 ceph-mon[56506]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T15:16:42.017 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 09 15:16:41 vm06 ceph-mon[56506]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T15:16:42.017 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 09 15:16:41 vm06 ceph-mon[56506]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' 2026-03-09T15:16:42.049 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:16:41 vm03 ceph-mon[51762]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T15:16:42.049 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:16:41 vm03 ceph-mon[51762]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T15:16:42.049 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:16:41 vm03 ceph-mon[51762]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T15:16:42.049 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:16:41 vm03 ceph-mon[51762]: from='mgr.14150 192.168.123.103:0/3334899447' entity='mgr.a' 2026-03-09T15:16:43.001 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 15:16:42 vm04 ceph-mon[53998]: pgmap v119: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-09T15:16:43.017 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 09 15:16:42 vm06 ceph-mon[56506]: pgmap v119: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-09T15:16:43.049 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:16:42 vm03 ceph-mon[51762]: pgmap v119: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-09T15:16:45.001 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 15:16:44 vm04 ceph-mon[53998]: pgmap v120: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-09T15:16:45.017 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 09 15:16:44 vm06 ceph-mon[56506]: pgmap v120: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-09T15:16:45.049 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:16:44 vm03 ceph-mon[51762]: pgmap v120: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-09T15:16:47.001 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 15:16:46 vm04 ceph-mon[53998]: pgmap v121: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-09T15:16:47.017 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 09 15:16:46 vm06 ceph-mon[56506]: pgmap v121: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-09T15:16:47.049 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:16:46 vm03 ceph-mon[51762]: pgmap v121: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-09T15:16:49.001 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 15:16:48 vm04 ceph-mon[53998]: pgmap v122: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-09T15:16:49.017 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 09 15:16:48 vm06 ceph-mon[56506]: pgmap v122: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-09T15:16:49.049 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:16:48 vm03 ceph-mon[51762]: pgmap v122: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-09T15:16:51.001 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 15:16:50 vm04 ceph-mon[53998]: pgmap v123: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-09T15:16:51.017 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 09 15:16:50 vm06 ceph-mon[56506]: pgmap v123: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-09T15:16:51.049 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:16:50 vm03 ceph-mon[51762]: pgmap v123: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-09T15:16:53.001 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 15:16:52 vm04 ceph-mon[53998]: pgmap v124: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-09T15:16:53.017 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 09 15:16:52 vm06 ceph-mon[56506]: pgmap v124: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-09T15:16:53.049 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:16:52 vm03 ceph-mon[51762]: pgmap v124: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-09T15:16:55.001 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 15:16:54 vm04 ceph-mon[53998]: pgmap v125: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-09T15:16:55.017 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 09 15:16:54 vm06 ceph-mon[56506]: pgmap v125: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-09T15:16:55.049 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:16:54 vm03 ceph-mon[51762]: pgmap v125: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-09T15:16:57.001 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 15:16:56 vm04 ceph-mon[53998]: pgmap v126: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-09T15:16:57.017 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 09 15:16:56 vm06 ceph-mon[56506]: pgmap v126: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-09T15:16:57.049 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:16:56 vm03 ceph-mon[51762]: pgmap v126: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-09T15:16:59.001 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 15:16:58 vm04 ceph-mon[53998]: pgmap v127: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-09T15:16:59.017 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 09 15:16:58 vm06 ceph-mon[56506]: pgmap v127: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-09T15:16:59.049 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:16:58 vm03 ceph-mon[51762]: pgmap v127: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-09T15:17:01.001 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 15:17:00 vm04 ceph-mon[53998]: pgmap v128: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-09T15:17:01.017 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 09 15:17:00 vm06 ceph-mon[56506]: pgmap v128: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-09T15:17:01.049 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:17:00 vm03 ceph-mon[51762]: pgmap v128: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-09T15:17:03.001 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 15:17:02 vm04 ceph-mon[53998]: pgmap v129: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-09T15:17:03.017 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 09 15:17:02 vm06 ceph-mon[56506]: pgmap v129: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-09T15:17:03.049 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:17:02 vm03 ceph-mon[51762]: pgmap v129: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-09T15:17:05.001 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 15:17:04 vm04 ceph-mon[53998]: pgmap v130: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-09T15:17:05.017 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 09 15:17:04 vm06 ceph-mon[56506]: pgmap v130: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-09T15:17:05.049 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:17:04 vm03 ceph-mon[51762]: pgmap v130: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-09T15:17:07.001 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 15:17:06 vm04 ceph-mon[53998]: pgmap v131: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-09T15:17:07.017 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 09 15:17:06 vm06 ceph-mon[56506]: pgmap v131: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-09T15:17:07.049 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:17:06 vm03 ceph-mon[51762]: pgmap v131: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-09T15:17:09.001 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 15:17:08 vm04 ceph-mon[53998]: pgmap v132: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-09T15:17:09.017 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 09 15:17:08 vm06 ceph-mon[56506]: pgmap v132: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-09T15:17:09.049 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:17:08 vm03 ceph-mon[51762]: pgmap v132: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-09T15:17:09.952 INFO:teuthology.orchestra.run.vm03.stderr:+ ceph mon dump 2026-03-09T15:17:10.129 INFO:teuthology.orchestra.run.vm03.stdout:epoch 6 2026-03-09T15:17:10.129 INFO:teuthology.orchestra.run.vm03.stdout:fsid 4abede80-1bca-11f1-bcd8-9b1da3c03ce5 2026-03-09T15:17:10.129 INFO:teuthology.orchestra.run.vm03.stdout:last_changed 2026-03-09T15:14:14.439856+0000 2026-03-09T15:17:10.129 INFO:teuthology.orchestra.run.vm03.stdout:created 2026-03-09T15:12:40.941469+0000 2026-03-09T15:17:10.129 INFO:teuthology.orchestra.run.vm03.stdout:min_mon_release 19 (squid) 2026-03-09T15:17:10.129 INFO:teuthology.orchestra.run.vm03.stdout:election_strategy: 1 2026-03-09T15:17:10.129 INFO:teuthology.orchestra.run.vm03.stdout:0: [v2:192.168.123.103:3300/0,v1:192.168.123.103:6789/0] mon.a; crush_location {datacenter=a} 2026-03-09T15:17:10.129 INFO:teuthology.orchestra.run.vm03.stdout:1: [v2:192.168.123.106:3300/0,v1:192.168.123.106:6789/0] mon.c; crush_location {datacenter=a,rack=3} 2026-03-09T15:17:10.129 INFO:teuthology.orchestra.run.vm03.stdout:2: [v2:192.168.123.104:3300/0,v1:192.168.123.104:6789/0] mon.b; crush_location {datacenter=b,rack=2} 2026-03-09T15:17:10.129 INFO:teuthology.orchestra.run.vm03.stderr:dumped monmap epoch 6 2026-03-09T15:17:10.138 INFO:teuthology.orchestra.run.vm03.stderr:+ ceph mon dump --format json 2026-03-09T15:17:10.299 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:17:10 vm03 ceph-mon[51762]: from='client.? 192.168.123.103:0/3636280884' entity='client.admin' cmd=[{"prefix": "mon dump"}]: dispatch 2026-03-09T15:17:10.323 INFO:teuthology.orchestra.run.vm03.stdout: 2026-03-09T15:17:10.323 INFO:teuthology.orchestra.run.vm03.stdout:{"epoch":6,"fsid":"4abede80-1bca-11f1-bcd8-9b1da3c03ce5","modified":"2026-03-09T15:14:14.439856Z","created":"2026-03-09T15:12:40.941469Z","min_mon_release":19,"min_mon_release_name":"squid","election_strategy":1,"disallowed_leaders":"","stretch_mode":false,"tiebreaker_mon":"","removed_ranks":"","features":{"persistent":["kraken","luminous","mimic","osdmap-prune","nautilus","octopus","pacific","elector-pinging","quincy","reef","squid"],"optional":[]},"mons":[{"rank":0,"name":"a","public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.103:3300","nonce":0},{"type":"v1","addr":"192.168.123.103:6789","nonce":0}]},"addr":"192.168.123.103:6789/0","public_addr":"192.168.123.103:6789/0","priority":0,"weight":0,"crush_location":"{datacenter=a}"},{"rank":1,"name":"c","public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.106:3300","nonce":0},{"type":"v1","addr":"192.168.123.106:6789","nonce":0}]},"addr":"192.168.123.106:6789/0","public_addr":"192.168.123.106:6789/0","priority":0,"weight":0,"crush_location":"{datacenter=a,rack=3}"},{"rank":2,"name":"b","public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.104:3300","nonce":0},{"type":"v1","addr":"192.168.123.104:6789","nonce":0}]},"addr":"192.168.123.104:6789/0","public_addr":"192.168.123.104:6789/0","priority":0,"weight":0,"crush_location":"{datacenter=b,rack=2}"}],"quorum":[0,1,2]} 2026-03-09T15:17:10.323 INFO:teuthology.orchestra.run.vm03.stderr:dumped monmap epoch 6 2026-03-09T15:17:10.333 INFO:teuthology.orchestra.run.vm03.stderr:+ read monid 2026-03-09T15:17:10.333 INFO:teuthology.orchestra.run.vm03.stderr:+ read crushloc 2026-03-09T15:17:10.334 INFO:teuthology.orchestra.run.vm03.stderr:+ ceph mon dump --format json 2026-03-09T15:17:10.334 INFO:teuthology.orchestra.run.vm03.stderr:+ jq --arg monid a --arg crushloc '{datacenter=a}' -e '.mons | .[] | select(.name == $monid) | .crush_location == $crushloc' 2026-03-09T15:17:10.501 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 15:17:10 vm04 ceph-mon[53998]: from='client.? 192.168.123.103:0/3636280884' entity='client.admin' cmd=[{"prefix": "mon dump"}]: dispatch 2026-03-09T15:17:10.517 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 09 15:17:10 vm06 ceph-mon[56506]: from='client.? 192.168.123.103:0/3636280884' entity='client.admin' cmd=[{"prefix": "mon dump"}]: dispatch 2026-03-09T15:17:10.520 INFO:teuthology.orchestra.run.vm03.stderr:dumped monmap epoch 6 2026-03-09T15:17:10.531 INFO:teuthology.orchestra.run.vm03.stdout:true 2026-03-09T15:17:10.531 INFO:teuthology.orchestra.run.vm03.stderr:+ read monid 2026-03-09T15:17:10.531 INFO:teuthology.orchestra.run.vm03.stderr:+ read crushloc 2026-03-09T15:17:10.531 INFO:teuthology.orchestra.run.vm03.stderr:+ ceph mon dump --format json 2026-03-09T15:17:10.531 INFO:teuthology.orchestra.run.vm03.stderr:+ jq --arg monid b --arg crushloc '{datacenter=b,rack=2}' -e '.mons | .[] | select(.name == $monid) | .crush_location == $crushloc' 2026-03-09T15:17:10.710 INFO:teuthology.orchestra.run.vm03.stderr:dumped monmap epoch 6 2026-03-09T15:17:10.719 INFO:teuthology.orchestra.run.vm03.stdout:true 2026-03-09T15:17:10.719 INFO:teuthology.orchestra.run.vm03.stderr:+ read monid 2026-03-09T15:17:10.720 INFO:teuthology.orchestra.run.vm03.stderr:+ read crushloc 2026-03-09T15:17:10.720 INFO:teuthology.orchestra.run.vm03.stderr:+ ceph mon dump --format json 2026-03-09T15:17:10.720 INFO:teuthology.orchestra.run.vm03.stderr:+ jq --arg monid c --arg crushloc '{datacenter=a,rack=3}' -e '.mons | .[] | select(.name == $monid) | .crush_location == $crushloc' 2026-03-09T15:17:10.894 INFO:teuthology.orchestra.run.vm03.stderr:dumped monmap epoch 6 2026-03-09T15:17:10.902 INFO:teuthology.orchestra.run.vm03.stdout:true 2026-03-09T15:17:10.902 INFO:teuthology.orchestra.run.vm03.stderr:+ read monid 2026-03-09T15:17:10.955 DEBUG:teuthology.run_tasks:Unwinding manager cephadm 2026-03-09T15:17:10.957 INFO:tasks.cephadm:Teardown begin 2026-03-09T15:17:10.957 DEBUG:teuthology.orchestra.run.vm03:> sudo rm -f /etc/ceph/ceph.conf /etc/ceph/ceph.client.admin.keyring 2026-03-09T15:17:10.982 DEBUG:teuthology.orchestra.run.vm04:> sudo rm -f /etc/ceph/ceph.conf /etc/ceph/ceph.client.admin.keyring 2026-03-09T15:17:11.010 DEBUG:teuthology.orchestra.run.vm06:> sudo rm -f /etc/ceph/ceph.conf /etc/ceph/ceph.client.admin.keyring 2026-03-09T15:17:11.039 INFO:tasks.cephadm:Disabling cephadm mgr module 2026-03-09T15:17:11.039 DEBUG:teuthology.orchestra.run.vm03:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 4abede80-1bca-11f1-bcd8-9b1da3c03ce5 -- ceph mgr module disable cephadm 2026-03-09T15:17:11.184 INFO:teuthology.orchestra.run.vm03.stderr:Inferring config /var/lib/ceph/4abede80-1bca-11f1-bcd8-9b1da3c03ce5/mon.a/config 2026-03-09T15:17:11.203 INFO:teuthology.orchestra.run.vm03.stderr:Error: statfs /etc/ceph/ceph.client.admin.keyring: no such file or directory 2026-03-09T15:17:11.221 DEBUG:teuthology.orchestra.run:got remote process result: 125 2026-03-09T15:17:11.221 INFO:tasks.cephadm:Cleaning up testdir ceph.* files... 2026-03-09T15:17:11.221 DEBUG:teuthology.orchestra.run.vm03:> rm -f /home/ubuntu/cephtest/seed.ceph.conf /home/ubuntu/cephtest/ceph.pub 2026-03-09T15:17:11.235 DEBUG:teuthology.orchestra.run.vm04:> rm -f /home/ubuntu/cephtest/seed.ceph.conf /home/ubuntu/cephtest/ceph.pub 2026-03-09T15:17:11.249 DEBUG:teuthology.orchestra.run.vm06:> rm -f /home/ubuntu/cephtest/seed.ceph.conf /home/ubuntu/cephtest/ceph.pub 2026-03-09T15:17:11.263 INFO:tasks.cephadm:Stopping all daemons... 2026-03-09T15:17:11.263 INFO:tasks.cephadm.mon.a:Stopping mon.a... 2026-03-09T15:17:11.263 DEBUG:teuthology.orchestra.run.vm03:> sudo systemctl stop ceph-4abede80-1bca-11f1-bcd8-9b1da3c03ce5@mon.a 2026-03-09T15:17:11.293 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:17:11 vm03 ceph-mon[51762]: from='client.? 192.168.123.103:0/2017300978' entity='client.admin' cmd=[{"prefix": "mon dump", "format": "json"}]: dispatch 2026-03-09T15:17:11.293 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:17:11 vm03 ceph-mon[51762]: pgmap v133: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-09T15:17:11.293 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:17:11 vm03 ceph-mon[51762]: from='client.? 192.168.123.103:0/1456484213' entity='client.admin' cmd=[{"prefix": "mon dump", "format": "json"}]: dispatch 2026-03-09T15:17:11.293 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:17:11 vm03 ceph-mon[51762]: from='client.? 192.168.123.103:0/859253712' entity='client.admin' cmd=[{"prefix": "mon dump", "format": "json"}]: dispatch 2026-03-09T15:17:11.293 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:17:11 vm03 ceph-mon[51762]: from='client.? 192.168.123.103:0/2496366498' entity='client.admin' cmd=[{"prefix": "mon dump", "format": "json"}]: dispatch 2026-03-09T15:17:11.501 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 15:17:11 vm04 ceph-mon[53998]: from='client.? 192.168.123.103:0/2017300978' entity='client.admin' cmd=[{"prefix": "mon dump", "format": "json"}]: dispatch 2026-03-09T15:17:11.501 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 15:17:11 vm04 ceph-mon[53998]: pgmap v133: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-09T15:17:11.501 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 15:17:11 vm04 ceph-mon[53998]: from='client.? 192.168.123.103:0/1456484213' entity='client.admin' cmd=[{"prefix": "mon dump", "format": "json"}]: dispatch 2026-03-09T15:17:11.501 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 15:17:11 vm04 ceph-mon[53998]: from='client.? 192.168.123.103:0/859253712' entity='client.admin' cmd=[{"prefix": "mon dump", "format": "json"}]: dispatch 2026-03-09T15:17:11.501 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 15:17:11 vm04 ceph-mon[53998]: from='client.? 192.168.123.103:0/2496366498' entity='client.admin' cmd=[{"prefix": "mon dump", "format": "json"}]: dispatch 2026-03-09T15:17:11.517 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 09 15:17:11 vm06 ceph-mon[56506]: from='client.? 192.168.123.103:0/2017300978' entity='client.admin' cmd=[{"prefix": "mon dump", "format": "json"}]: dispatch 2026-03-09T15:17:11.517 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 09 15:17:11 vm06 ceph-mon[56506]: pgmap v133: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-09T15:17:11.517 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 09 15:17:11 vm06 ceph-mon[56506]: from='client.? 192.168.123.103:0/1456484213' entity='client.admin' cmd=[{"prefix": "mon dump", "format": "json"}]: dispatch 2026-03-09T15:17:11.517 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 09 15:17:11 vm06 ceph-mon[56506]: from='client.? 192.168.123.103:0/859253712' entity='client.admin' cmd=[{"prefix": "mon dump", "format": "json"}]: dispatch 2026-03-09T15:17:11.517 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 09 15:17:11 vm06 ceph-mon[56506]: from='client.? 192.168.123.103:0/2496366498' entity='client.admin' cmd=[{"prefix": "mon dump", "format": "json"}]: dispatch 2026-03-09T15:17:11.550 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:17:11 vm03 systemd[1]: Stopping Ceph mon.a for 4abede80-1bca-11f1-bcd8-9b1da3c03ce5... 2026-03-09T15:17:11.550 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:17:11 vm03 ceph-4abede80-1bca-11f1-bcd8-9b1da3c03ce5-mon-a[51758]: 2026-03-09T15:17:11.364+0000 7f4f64a5e640 -1 received signal: Terminated from /run/podman-init -- /usr/bin/ceph-mon -n mon.a -f --setuser ceph --setgroup ceph --default-log-to-file=false --default-log-to-journald=true --default-log-to-stderr=false --default-mon-cluster-log-to-file=false --default-mon-cluster-log-to-journald=true --default-mon-cluster-log-to-stderr=false (PID: 1) UID: 0 2026-03-09T15:17:11.550 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 15:17:11 vm03 ceph-4abede80-1bca-11f1-bcd8-9b1da3c03ce5-mon-a[51758]: 2026-03-09T15:17:11.364+0000 7f4f64a5e640 -1 mon.a@0(leader) e6 *** Got Signal Terminated *** 2026-03-09T15:17:11.686 DEBUG:teuthology.orchestra.run.vm03:> sudo pkill -f 'journalctl -f -n 0 -u ceph-4abede80-1bca-11f1-bcd8-9b1da3c03ce5@mon.a.service' 2026-03-09T15:17:11.754 DEBUG:teuthology.orchestra.run:got remote process result: None 2026-03-09T15:17:11.754 INFO:tasks.cephadm.mon.a:Stopped mon.a 2026-03-09T15:17:11.754 INFO:tasks.cephadm.mon.c:Stopping mon.b... 2026-03-09T15:17:11.754 DEBUG:teuthology.orchestra.run.vm04:> sudo systemctl stop ceph-4abede80-1bca-11f1-bcd8-9b1da3c03ce5@mon.b 2026-03-09T15:17:11.963 DEBUG:teuthology.orchestra.run.vm04:> sudo pkill -f 'journalctl -f -n 0 -u ceph-4abede80-1bca-11f1-bcd8-9b1da3c03ce5@mon.b.service' 2026-03-09T15:17:11.992 DEBUG:teuthology.orchestra.run:got remote process result: None 2026-03-09T15:17:11.992 INFO:tasks.cephadm.mon.c:Stopped mon.b 2026-03-09T15:17:11.992 INFO:tasks.cephadm.mon.c:Stopping mon.c... 2026-03-09T15:17:11.992 DEBUG:teuthology.orchestra.run.vm06:> sudo systemctl stop ceph-4abede80-1bca-11f1-bcd8-9b1da3c03ce5@mon.c 2026-03-09T15:17:12.267 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 09 15:17:12 vm06 systemd[1]: Stopping Ceph mon.c for 4abede80-1bca-11f1-bcd8-9b1da3c03ce5... 2026-03-09T15:17:12.267 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 09 15:17:12 vm06 ceph-4abede80-1bca-11f1-bcd8-9b1da3c03ce5-mon-c[56502]: 2026-03-09T15:17:12.084+0000 7f0f6f3d4640 -1 received signal: Terminated from /run/podman-init -- /usr/bin/ceph-mon -n mon.c -f --setuser ceph --setgroup ceph --default-log-to-file=false --default-log-to-journald=true --default-log-to-stderr=false --default-mon-cluster-log-to-file=false --default-mon-cluster-log-to-journald=true --default-mon-cluster-log-to-stderr=false (PID: 1) UID: 0 2026-03-09T15:17:12.267 INFO:journalctl@ceph.mon.c.vm06.stdout:Mar 09 15:17:12 vm06 ceph-4abede80-1bca-11f1-bcd8-9b1da3c03ce5-mon-c[56502]: 2026-03-09T15:17:12.084+0000 7f0f6f3d4640 -1 mon.c@1(peon) e6 *** Got Signal Terminated *** 2026-03-09T15:17:12.416 DEBUG:teuthology.orchestra.run.vm06:> sudo pkill -f 'journalctl -f -n 0 -u ceph-4abede80-1bca-11f1-bcd8-9b1da3c03ce5@mon.c.service' 2026-03-09T15:17:12.446 DEBUG:teuthology.orchestra.run:got remote process result: None 2026-03-09T15:17:12.446 INFO:tasks.cephadm.mon.c:Stopped mon.c 2026-03-09T15:17:12.446 INFO:tasks.cephadm.mgr.a:Stopping mgr.a... 2026-03-09T15:17:12.446 DEBUG:teuthology.orchestra.run.vm03:> sudo systemctl stop ceph-4abede80-1bca-11f1-bcd8-9b1da3c03ce5@mgr.a 2026-03-09T15:17:12.669 DEBUG:teuthology.orchestra.run.vm03:> sudo pkill -f 'journalctl -f -n 0 -u ceph-4abede80-1bca-11f1-bcd8-9b1da3c03ce5@mgr.a.service' 2026-03-09T15:17:12.695 DEBUG:teuthology.orchestra.run:got remote process result: None 2026-03-09T15:17:12.695 INFO:tasks.cephadm.mgr.a:Stopped mgr.a 2026-03-09T15:17:12.695 INFO:tasks.cephadm.mgr.b:Stopping mgr.b... 2026-03-09T15:17:12.695 DEBUG:teuthology.orchestra.run.vm04:> sudo systemctl stop ceph-4abede80-1bca-11f1-bcd8-9b1da3c03ce5@mgr.b 2026-03-09T15:17:12.840 INFO:journalctl@ceph.mgr.b.vm04.stdout:Mar 09 15:17:12 vm04 systemd[1]: Stopping Ceph mgr.b for 4abede80-1bca-11f1-bcd8-9b1da3c03ce5... 2026-03-09T15:17:12.840 INFO:journalctl@ceph.mgr.b.vm04.stdout:Mar 09 15:17:12 vm04 podman[61369]: 2026-03-09 15:17:12.818263761 +0000 UTC m=+0.046795509 container died 518720cdb99228601b5cbb6224235a125b9cc19b8ea3654608475ff01d469936 (image=quay.ceph.io/ceph-ci/ceph@sha256:8fda260ab1d2d3118a1622f7df75f44f285dfe74e71793626152a711c12bf2cc, name=ceph-4abede80-1bca-11f1-bcd8-9b1da3c03ce5-mgr-b, org.opencontainers.image.authors=Ceph Release Team , GANESHA_REPO_BASEURL=https://buildlogs.centos.org/centos/$releasever-stream/storage/$basearch/nfsganesha-5/, org.label-schema.vendor=CentOS, FROM_IMAGE=quay.io/centos/centos:stream9, CEPH_REF=squid, CEPH_GIT_REPO=https://github.com/ceph/ceph-ci.git, io.buildah.version=1.41.3, org.label-schema.license=GPLv2, org.label-schema.name=CentOS Stream 9 Base Image, OSD_FLAVOR=default, org.label-schema.schema-version=1.0, org.label-schema.build-date=20260223, org.opencontainers.image.documentation=https://docs.ceph.com/, CEPH_SHA1=e911bdebe5c8faa3800735d1568fcdca65db60df, ceph=True) 2026-03-09T15:17:12.840 INFO:journalctl@ceph.mgr.b.vm04.stdout:Mar 09 15:17:12 vm04 podman[61369]: 2026-03-09 15:17:12.839687771 +0000 UTC m=+0.068219528 container remove 518720cdb99228601b5cbb6224235a125b9cc19b8ea3654608475ff01d469936 (image=quay.ceph.io/ceph-ci/ceph@sha256:8fda260ab1d2d3118a1622f7df75f44f285dfe74e71793626152a711c12bf2cc, name=ceph-4abede80-1bca-11f1-bcd8-9b1da3c03ce5-mgr-b, io.buildah.version=1.41.3, org.label-schema.build-date=20260223, org.label-schema.vendor=CentOS, org.label-schema.schema-version=1.0, org.opencontainers.image.authors=Ceph Release Team , org.label-schema.name=CentOS Stream 9 Base Image, FROM_IMAGE=quay.io/centos/centos:stream9, OSD_FLAVOR=default, GANESHA_REPO_BASEURL=https://buildlogs.centos.org/centos/$releasever-stream/storage/$basearch/nfsganesha-5/, org.label-schema.license=GPLv2, CEPH_GIT_REPO=https://github.com/ceph/ceph-ci.git, CEPH_SHA1=e911bdebe5c8faa3800735d1568fcdca65db60df, ceph=True, CEPH_REF=squid, org.opencontainers.image.documentation=https://docs.ceph.com/) 2026-03-09T15:17:12.892 DEBUG:teuthology.orchestra.run.vm04:> sudo pkill -f 'journalctl -f -n 0 -u ceph-4abede80-1bca-11f1-bcd8-9b1da3c03ce5@mgr.b.service' 2026-03-09T15:17:12.918 DEBUG:teuthology.orchestra.run:got remote process result: None 2026-03-09T15:17:12.918 INFO:tasks.cephadm.mgr.b:Stopped mgr.b 2026-03-09T15:17:12.918 INFO:tasks.cephadm.osd.0:Stopping osd.0... 2026-03-09T15:17:12.918 DEBUG:teuthology.orchestra.run.vm03:> sudo systemctl stop ceph-4abede80-1bca-11f1-bcd8-9b1da3c03ce5@osd.0 2026-03-09T15:17:13.299 INFO:journalctl@ceph.osd.0.vm03.stdout:Mar 09 15:17:12 vm03 systemd[1]: Stopping Ceph osd.0 for 4abede80-1bca-11f1-bcd8-9b1da3c03ce5... 2026-03-09T15:17:13.299 INFO:journalctl@ceph.osd.0.vm03.stdout:Mar 09 15:17:12 vm03 ceph-4abede80-1bca-11f1-bcd8-9b1da3c03ce5-osd-0[61316]: 2026-03-09T15:17:12.997+0000 7f89e7588640 -1 received signal: Terminated from /run/podman-init -- /usr/bin/ceph-osd -n osd.0 -f --setuser ceph --setgroup ceph --default-log-to-file=false --default-log-to-journald=true --default-log-to-stderr=false (PID: 1) UID: 0 2026-03-09T15:17:13.299 INFO:journalctl@ceph.osd.0.vm03.stdout:Mar 09 15:17:12 vm03 ceph-4abede80-1bca-11f1-bcd8-9b1da3c03ce5-osd-0[61316]: 2026-03-09T15:17:12.997+0000 7f89e7588640 -1 osd.0 19 *** Got signal Terminated *** 2026-03-09T15:17:13.299 INFO:journalctl@ceph.osd.0.vm03.stdout:Mar 09 15:17:12 vm03 ceph-4abede80-1bca-11f1-bcd8-9b1da3c03ce5-osd-0[61316]: 2026-03-09T15:17:12.997+0000 7f89e7588640 -1 osd.0 19 *** Immediate shutdown (osd_fast_shutdown=true) *** 2026-03-09T15:17:18.300 INFO:journalctl@ceph.osd.0.vm03.stdout:Mar 09 15:17:18 vm03 podman[68181]: 2026-03-09 15:17:18.024751675 +0000 UTC m=+5.038648619 container died 646ad6643c084d5e16b64f0237816bbe52c9448a9794281ce284634b8d74211c (image=quay.ceph.io/ceph-ci/ceph@sha256:8fda260ab1d2d3118a1622f7df75f44f285dfe74e71793626152a711c12bf2cc, name=ceph-4abede80-1bca-11f1-bcd8-9b1da3c03ce5-osd-0, org.label-schema.name=CentOS Stream 9 Base Image, CEPH_GIT_REPO=https://github.com/ceph/ceph-ci.git, CEPH_REF=squid, org.label-schema.license=GPLv2, org.opencontainers.image.documentation=https://docs.ceph.com/, io.buildah.version=1.41.3, OSD_FLAVOR=default, org.label-schema.build-date=20260223, GANESHA_REPO_BASEURL=https://buildlogs.centos.org/centos/$releasever-stream/storage/$basearch/nfsganesha-5/, org.label-schema.vendor=CentOS, CEPH_SHA1=e911bdebe5c8faa3800735d1568fcdca65db60df, FROM_IMAGE=quay.io/centos/centos:stream9, ceph=True, org.opencontainers.image.authors=Ceph Release Team , org.label-schema.schema-version=1.0) 2026-03-09T15:17:18.300 INFO:journalctl@ceph.osd.0.vm03.stdout:Mar 09 15:17:18 vm03 podman[68181]: 2026-03-09 15:17:18.041967244 +0000 UTC m=+5.055864198 container remove 646ad6643c084d5e16b64f0237816bbe52c9448a9794281ce284634b8d74211c (image=quay.ceph.io/ceph-ci/ceph@sha256:8fda260ab1d2d3118a1622f7df75f44f285dfe74e71793626152a711c12bf2cc, name=ceph-4abede80-1bca-11f1-bcd8-9b1da3c03ce5-osd-0, org.label-schema.build-date=20260223, GANESHA_REPO_BASEURL=https://buildlogs.centos.org/centos/$releasever-stream/storage/$basearch/nfsganesha-5/, CEPH_SHA1=e911bdebe5c8faa3800735d1568fcdca65db60df, org.label-schema.schema-version=1.0, org.label-schema.vendor=CentOS, CEPH_REF=squid, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.license=GPLv2, OSD_FLAVOR=default, org.opencontainers.image.authors=Ceph Release Team , ceph=True, FROM_IMAGE=quay.io/centos/centos:stream9, CEPH_GIT_REPO=https://github.com/ceph/ceph-ci.git, io.buildah.version=1.41.3, org.opencontainers.image.documentation=https://docs.ceph.com/) 2026-03-09T15:17:18.300 INFO:journalctl@ceph.osd.0.vm03.stdout:Mar 09 15:17:18 vm03 bash[68181]: ceph-4abede80-1bca-11f1-bcd8-9b1da3c03ce5-osd-0 2026-03-09T15:17:18.300 INFO:journalctl@ceph.osd.0.vm03.stdout:Mar 09 15:17:18 vm03 podman[68249]: 2026-03-09 15:17:18.160657553 +0000 UTC m=+0.014645649 container create 3e5ae7b9e1eb6e99d7ece5e2cd47b599f9ddafbaa30b5db23ebb84da3484863b (image=quay.ceph.io/ceph-ci/ceph@sha256:8fda260ab1d2d3118a1622f7df75f44f285dfe74e71793626152a711c12bf2cc, name=ceph-4abede80-1bca-11f1-bcd8-9b1da3c03ce5-osd-0-deactivate, CEPH_SHA1=e911bdebe5c8faa3800735d1568fcdca65db60df, org.label-schema.name=CentOS Stream 9 Base Image, GANESHA_REPO_BASEURL=https://buildlogs.centos.org/centos/$releasever-stream/storage/$basearch/nfsganesha-5/, org.label-schema.build-date=20260223, FROM_IMAGE=quay.io/centos/centos:stream9, org.opencontainers.image.authors=Ceph Release Team , org.label-schema.vendor=CentOS, io.buildah.version=1.41.3, org.label-schema.schema-version=1.0, org.label-schema.license=GPLv2, org.opencontainers.image.documentation=https://docs.ceph.com/, CEPH_REF=squid, OSD_FLAVOR=default, ceph=True, CEPH_GIT_REPO=https://github.com/ceph/ceph-ci.git) 2026-03-09T15:17:18.300 INFO:journalctl@ceph.osd.0.vm03.stdout:Mar 09 15:17:18 vm03 podman[68249]: 2026-03-09 15:17:18.203785635 +0000 UTC m=+0.057773720 container init 3e5ae7b9e1eb6e99d7ece5e2cd47b599f9ddafbaa30b5db23ebb84da3484863b (image=quay.ceph.io/ceph-ci/ceph@sha256:8fda260ab1d2d3118a1622f7df75f44f285dfe74e71793626152a711c12bf2cc, name=ceph-4abede80-1bca-11f1-bcd8-9b1da3c03ce5-osd-0-deactivate, org.opencontainers.image.documentation=https://docs.ceph.com/, ceph=True, FROM_IMAGE=quay.io/centos/centos:stream9, OSD_FLAVOR=default, org.label-schema.build-date=20260223, org.label-schema.schema-version=1.0, CEPH_REF=squid, CEPH_GIT_REPO=https://github.com/ceph/ceph-ci.git, CEPH_SHA1=e911bdebe5c8faa3800735d1568fcdca65db60df, org.label-schema.name=CentOS Stream 9 Base Image, org.opencontainers.image.authors=Ceph Release Team , org.label-schema.vendor=CentOS, GANESHA_REPO_BASEURL=https://buildlogs.centos.org/centos/$releasever-stream/storage/$basearch/nfsganesha-5/, io.buildah.version=1.41.3, org.label-schema.license=GPLv2) 2026-03-09T15:17:18.300 INFO:journalctl@ceph.osd.0.vm03.stdout:Mar 09 15:17:18 vm03 podman[68249]: 2026-03-09 15:17:18.206699087 +0000 UTC m=+0.060687183 container start 3e5ae7b9e1eb6e99d7ece5e2cd47b599f9ddafbaa30b5db23ebb84da3484863b (image=quay.ceph.io/ceph-ci/ceph@sha256:8fda260ab1d2d3118a1622f7df75f44f285dfe74e71793626152a711c12bf2cc, name=ceph-4abede80-1bca-11f1-bcd8-9b1da3c03ce5-osd-0-deactivate, CEPH_GIT_REPO=https://github.com/ceph/ceph-ci.git, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.vendor=CentOS, CEPH_SHA1=e911bdebe5c8faa3800735d1568fcdca65db60df, org.label-schema.build-date=20260223, org.label-schema.schema-version=1.0, FROM_IMAGE=quay.io/centos/centos:stream9, CEPH_REF=squid, org.opencontainers.image.authors=Ceph Release Team , ceph=True, org.label-schema.license=GPLv2, org.opencontainers.image.documentation=https://docs.ceph.com/, GANESHA_REPO_BASEURL=https://buildlogs.centos.org/centos/$releasever-stream/storage/$basearch/nfsganesha-5/, OSD_FLAVOR=default, io.buildah.version=1.41.3) 2026-03-09T15:17:18.300 INFO:journalctl@ceph.osd.0.vm03.stdout:Mar 09 15:17:18 vm03 podman[68249]: 2026-03-09 15:17:18.207604802 +0000 UTC m=+0.061592898 container attach 3e5ae7b9e1eb6e99d7ece5e2cd47b599f9ddafbaa30b5db23ebb84da3484863b (image=quay.ceph.io/ceph-ci/ceph@sha256:8fda260ab1d2d3118a1622f7df75f44f285dfe74e71793626152a711c12bf2cc, name=ceph-4abede80-1bca-11f1-bcd8-9b1da3c03ce5-osd-0-deactivate, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.license=GPLv2, FROM_IMAGE=quay.io/centos/centos:stream9, org.opencontainers.image.documentation=https://docs.ceph.com/, CEPH_REF=squid, org.label-schema.vendor=CentOS, GANESHA_REPO_BASEURL=https://buildlogs.centos.org/centos/$releasever-stream/storage/$basearch/nfsganesha-5/, ceph=True, org.label-schema.schema-version=1.0, CEPH_SHA1=e911bdebe5c8faa3800735d1568fcdca65db60df, org.opencontainers.image.authors=Ceph Release Team , io.buildah.version=1.41.3, CEPH_GIT_REPO=https://github.com/ceph/ceph-ci.git, OSD_FLAVOR=default, org.label-schema.build-date=20260223) 2026-03-09T15:17:18.300 INFO:journalctl@ceph.osd.0.vm03.stdout:Mar 09 15:17:18 vm03 podman[68249]: 2026-03-09 15:17:18.154783531 +0000 UTC m=+0.008771636 image pull 654f31e6858eb235bbece362255b685a945f2b6a367e2b88c4930c984fbb214c quay.ceph.io/ceph-ci/ceph@sha256:8fda260ab1d2d3118a1622f7df75f44f285dfe74e71793626152a711c12bf2cc 2026-03-09T15:17:18.354 DEBUG:teuthology.orchestra.run.vm03:> sudo pkill -f 'journalctl -f -n 0 -u ceph-4abede80-1bca-11f1-bcd8-9b1da3c03ce5@osd.0.service' 2026-03-09T15:17:18.388 DEBUG:teuthology.orchestra.run:got remote process result: None 2026-03-09T15:17:18.388 INFO:tasks.cephadm.osd.0:Stopped osd.0 2026-03-09T15:17:18.388 INFO:tasks.cephadm.osd.1:Stopping osd.1... 2026-03-09T15:17:18.388 DEBUG:teuthology.orchestra.run.vm04:> sudo systemctl stop ceph-4abede80-1bca-11f1-bcd8-9b1da3c03ce5@osd.1 2026-03-09T15:17:18.751 INFO:journalctl@ceph.osd.1.vm04.stdout:Mar 09 15:17:18 vm04 systemd[1]: Stopping Ceph osd.1 for 4abede80-1bca-11f1-bcd8-9b1da3c03ce5... 2026-03-09T15:17:18.751 INFO:journalctl@ceph.osd.1.vm04.stdout:Mar 09 15:17:18 vm04 ceph-4abede80-1bca-11f1-bcd8-9b1da3c03ce5-osd-1[57595]: 2026-03-09T15:17:18.484+0000 7f583cd0a640 -1 received signal: Terminated from /run/podman-init -- /usr/bin/ceph-osd -n osd.1 -f --setuser ceph --setgroup ceph --default-log-to-file=false --default-log-to-journald=true --default-log-to-stderr=false (PID: 1) UID: 0 2026-03-09T15:17:18.751 INFO:journalctl@ceph.osd.1.vm04.stdout:Mar 09 15:17:18 vm04 ceph-4abede80-1bca-11f1-bcd8-9b1da3c03ce5-osd-1[57595]: 2026-03-09T15:17:18.484+0000 7f583cd0a640 -1 osd.1 19 *** Got signal Terminated *** 2026-03-09T15:17:18.751 INFO:journalctl@ceph.osd.1.vm04.stdout:Mar 09 15:17:18 vm04 ceph-4abede80-1bca-11f1-bcd8-9b1da3c03ce5-osd-1[57595]: 2026-03-09T15:17:18.484+0000 7f583cd0a640 -1 osd.1 19 *** Immediate shutdown (osd_fast_shutdown=true) *** 2026-03-09T15:17:23.839 INFO:journalctl@ceph.osd.1.vm04.stdout:Mar 09 15:17:23 vm04 podman[61471]: 2026-03-09 15:17:23.513252933 +0000 UTC m=+5.042628902 container died 91cd0ade9d3c7afaea11099fef06811078cd9630d667d07f496d91208b1c8c02 (image=quay.ceph.io/ceph-ci/ceph@sha256:8fda260ab1d2d3118a1622f7df75f44f285dfe74e71793626152a711c12bf2cc, name=ceph-4abede80-1bca-11f1-bcd8-9b1da3c03ce5-osd-1, org.label-schema.license=GPLv2, CEPH_SHA1=e911bdebe5c8faa3800735d1568fcdca65db60df, org.opencontainers.image.documentation=https://docs.ceph.com/, org.label-schema.build-date=20260223, CEPH_GIT_REPO=https://github.com/ceph/ceph-ci.git, OSD_FLAVOR=default, org.label-schema.name=CentOS Stream 9 Base Image, GANESHA_REPO_BASEURL=https://buildlogs.centos.org/centos/$releasever-stream/storage/$basearch/nfsganesha-5/, org.label-schema.schema-version=1.0, org.label-schema.vendor=CentOS, org.opencontainers.image.authors=Ceph Release Team , ceph=True, FROM_IMAGE=quay.io/centos/centos:stream9, CEPH_REF=squid, io.buildah.version=1.41.3) 2026-03-09T15:17:23.839 INFO:journalctl@ceph.osd.1.vm04.stdout:Mar 09 15:17:23 vm04 podman[61471]: 2026-03-09 15:17:23.538338115 +0000 UTC m=+5.067714083 container remove 91cd0ade9d3c7afaea11099fef06811078cd9630d667d07f496d91208b1c8c02 (image=quay.ceph.io/ceph-ci/ceph@sha256:8fda260ab1d2d3118a1622f7df75f44f285dfe74e71793626152a711c12bf2cc, name=ceph-4abede80-1bca-11f1-bcd8-9b1da3c03ce5-osd-1, org.label-schema.build-date=20260223, org.opencontainers.image.documentation=https://docs.ceph.com/, org.label-schema.schema-version=1.0, FROM_IMAGE=quay.io/centos/centos:stream9, io.buildah.version=1.41.3, org.label-schema.license=GPLv2, CEPH_SHA1=e911bdebe5c8faa3800735d1568fcdca65db60df, GANESHA_REPO_BASEURL=https://buildlogs.centos.org/centos/$releasever-stream/storage/$basearch/nfsganesha-5/, ceph=True, org.label-schema.vendor=CentOS, org.label-schema.name=CentOS Stream 9 Base Image, OSD_FLAVOR=default, org.opencontainers.image.authors=Ceph Release Team , CEPH_GIT_REPO=https://github.com/ceph/ceph-ci.git, CEPH_REF=squid) 2026-03-09T15:17:23.839 INFO:journalctl@ceph.osd.1.vm04.stdout:Mar 09 15:17:23 vm04 bash[61471]: ceph-4abede80-1bca-11f1-bcd8-9b1da3c03ce5-osd-1 2026-03-09T15:17:23.839 INFO:journalctl@ceph.osd.1.vm04.stdout:Mar 09 15:17:23 vm04 podman[61542]: 2026-03-09 15:17:23.65884532 +0000 UTC m=+0.015055300 container create afba58302a92ec12173b9e74ae407d750ef8557629f2e458b64cd8ddd62a9200 (image=quay.ceph.io/ceph-ci/ceph@sha256:8fda260ab1d2d3118a1622f7df75f44f285dfe74e71793626152a711c12bf2cc, name=ceph-4abede80-1bca-11f1-bcd8-9b1da3c03ce5-osd-1-deactivate, io.buildah.version=1.41.3, GANESHA_REPO_BASEURL=https://buildlogs.centos.org/centos/$releasever-stream/storage/$basearch/nfsganesha-5/, org.label-schema.license=GPLv2, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.vendor=CentOS, CEPH_REF=squid, FROM_IMAGE=quay.io/centos/centos:stream9, CEPH_GIT_REPO=https://github.com/ceph/ceph-ci.git, org.label-schema.build-date=20260223, org.opencontainers.image.documentation=https://docs.ceph.com/, org.label-schema.schema-version=1.0, ceph=True, CEPH_SHA1=e911bdebe5c8faa3800735d1568fcdca65db60df, org.opencontainers.image.authors=Ceph Release Team , OSD_FLAVOR=default) 2026-03-09T15:17:23.839 INFO:journalctl@ceph.osd.1.vm04.stdout:Mar 09 15:17:23 vm04 podman[61542]: 2026-03-09 15:17:23.706218328 +0000 UTC m=+0.062428318 container init afba58302a92ec12173b9e74ae407d750ef8557629f2e458b64cd8ddd62a9200 (image=quay.ceph.io/ceph-ci/ceph@sha256:8fda260ab1d2d3118a1622f7df75f44f285dfe74e71793626152a711c12bf2cc, name=ceph-4abede80-1bca-11f1-bcd8-9b1da3c03ce5-osd-1-deactivate, io.buildah.version=1.41.3, org.opencontainers.image.authors=Ceph Release Team , CEPH_REF=squid, GANESHA_REPO_BASEURL=https://buildlogs.centos.org/centos/$releasever-stream/storage/$basearch/nfsganesha-5/, FROM_IMAGE=quay.io/centos/centos:stream9, org.label-schema.license=GPLv2, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.schema-version=1.0, CEPH_SHA1=e911bdebe5c8faa3800735d1568fcdca65db60df, CEPH_GIT_REPO=https://github.com/ceph/ceph-ci.git, org.label-schema.build-date=20260223, org.opencontainers.image.documentation=https://docs.ceph.com/, org.label-schema.vendor=CentOS, OSD_FLAVOR=default, ceph=True) 2026-03-09T15:17:23.839 INFO:journalctl@ceph.osd.1.vm04.stdout:Mar 09 15:17:23 vm04 podman[61542]: 2026-03-09 15:17:23.709004874 +0000 UTC m=+0.065214854 container start afba58302a92ec12173b9e74ae407d750ef8557629f2e458b64cd8ddd62a9200 (image=quay.ceph.io/ceph-ci/ceph@sha256:8fda260ab1d2d3118a1622f7df75f44f285dfe74e71793626152a711c12bf2cc, name=ceph-4abede80-1bca-11f1-bcd8-9b1da3c03ce5-osd-1-deactivate, CEPH_SHA1=e911bdebe5c8faa3800735d1568fcdca65db60df, io.buildah.version=1.41.3, org.opencontainers.image.documentation=https://docs.ceph.com/, org.label-schema.license=GPLv2, org.opencontainers.image.authors=Ceph Release Team , CEPH_GIT_REPO=https://github.com/ceph/ceph-ci.git, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.vendor=CentOS, CEPH_REF=squid, OSD_FLAVOR=default, org.label-schema.schema-version=1.0, ceph=True, GANESHA_REPO_BASEURL=https://buildlogs.centos.org/centos/$releasever-stream/storage/$basearch/nfsganesha-5/, FROM_IMAGE=quay.io/centos/centos:stream9, org.label-schema.build-date=20260223) 2026-03-09T15:17:23.839 INFO:journalctl@ceph.osd.1.vm04.stdout:Mar 09 15:17:23 vm04 podman[61542]: 2026-03-09 15:17:23.713964487 +0000 UTC m=+0.070174467 container attach afba58302a92ec12173b9e74ae407d750ef8557629f2e458b64cd8ddd62a9200 (image=quay.ceph.io/ceph-ci/ceph@sha256:8fda260ab1d2d3118a1622f7df75f44f285dfe74e71793626152a711c12bf2cc, name=ceph-4abede80-1bca-11f1-bcd8-9b1da3c03ce5-osd-1-deactivate, CEPH_GIT_REPO=https://github.com/ceph/ceph-ci.git, org.label-schema.build-date=20260223, ceph=True, CEPH_SHA1=e911bdebe5c8faa3800735d1568fcdca65db60df, FROM_IMAGE=quay.io/centos/centos:stream9, org.opencontainers.image.authors=Ceph Release Team , org.label-schema.name=CentOS Stream 9 Base Image, org.opencontainers.image.documentation=https://docs.ceph.com/, OSD_FLAVOR=default, io.buildah.version=1.41.3, org.label-schema.vendor=CentOS, CEPH_REF=squid, org.label-schema.schema-version=1.0, org.label-schema.license=GPLv2, GANESHA_REPO_BASEURL=https://buildlogs.centos.org/centos/$releasever-stream/storage/$basearch/nfsganesha-5/) 2026-03-09T15:17:23.839 INFO:journalctl@ceph.osd.1.vm04.stdout:Mar 09 15:17:23 vm04 podman[61542]: 2026-03-09 15:17:23.652751985 +0000 UTC m=+0.008961965 image pull 654f31e6858eb235bbece362255b685a945f2b6a367e2b88c4930c984fbb214c quay.ceph.io/ceph-ci/ceph@sha256:8fda260ab1d2d3118a1622f7df75f44f285dfe74e71793626152a711c12bf2cc 2026-03-09T15:17:23.839 INFO:journalctl@ceph.osd.1.vm04.stdout:Mar 09 15:17:23 vm04 podman[61542]: 2026-03-09 15:17:23.837994435 +0000 UTC m=+0.194204415 container died afba58302a92ec12173b9e74ae407d750ef8557629f2e458b64cd8ddd62a9200 (image=quay.ceph.io/ceph-ci/ceph@sha256:8fda260ab1d2d3118a1622f7df75f44f285dfe74e71793626152a711c12bf2cc, name=ceph-4abede80-1bca-11f1-bcd8-9b1da3c03ce5-osd-1-deactivate, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.build-date=20260223, org.label-schema.vendor=CentOS, OSD_FLAVOR=default, FROM_IMAGE=quay.io/centos/centos:stream9, org.label-schema.license=GPLv2, CEPH_SHA1=e911bdebe5c8faa3800735d1568fcdca65db60df, org.opencontainers.image.authors=Ceph Release Team , org.opencontainers.image.documentation=https://docs.ceph.com/, CEPH_REF=squid, GANESHA_REPO_BASEURL=https://buildlogs.centos.org/centos/$releasever-stream/storage/$basearch/nfsganesha-5/, org.label-schema.schema-version=1.0, ceph=True, io.buildah.version=1.41.3, CEPH_GIT_REPO=https://github.com/ceph/ceph-ci.git) 2026-03-09T15:17:23.866 DEBUG:teuthology.orchestra.run.vm04:> sudo pkill -f 'journalctl -f -n 0 -u ceph-4abede80-1bca-11f1-bcd8-9b1da3c03ce5@osd.1.service' 2026-03-09T15:17:23.894 DEBUG:teuthology.orchestra.run:got remote process result: None 2026-03-09T15:17:23.895 INFO:tasks.cephadm.osd.1:Stopped osd.1 2026-03-09T15:17:23.895 INFO:tasks.cephadm.osd.2:Stopping osd.2... 2026-03-09T15:17:23.895 DEBUG:teuthology.orchestra.run.vm06:> sudo systemctl stop ceph-4abede80-1bca-11f1-bcd8-9b1da3c03ce5@osd.2 2026-03-09T15:17:24.267 INFO:journalctl@ceph.osd.2.vm06.stdout:Mar 09 15:17:23 vm06 systemd[1]: Stopping Ceph osd.2 for 4abede80-1bca-11f1-bcd8-9b1da3c03ce5... 2026-03-09T15:17:24.267 INFO:journalctl@ceph.osd.2.vm06.stdout:Mar 09 15:17:23 vm06 ceph-4abede80-1bca-11f1-bcd8-9b1da3c03ce5-osd-2[59842]: 2026-03-09T15:17:23.991+0000 7ff433b1f640 -1 received signal: Terminated from /run/podman-init -- /usr/bin/ceph-osd -n osd.2 -f --setuser ceph --setgroup ceph --default-log-to-file=false --default-log-to-journald=true --default-log-to-stderr=false (PID: 1) UID: 0 2026-03-09T15:17:24.267 INFO:journalctl@ceph.osd.2.vm06.stdout:Mar 09 15:17:23 vm06 ceph-4abede80-1bca-11f1-bcd8-9b1da3c03ce5-osd-2[59842]: 2026-03-09T15:17:23.991+0000 7ff433b1f640 -1 osd.2 19 *** Got signal Terminated *** 2026-03-09T15:17:24.267 INFO:journalctl@ceph.osd.2.vm06.stdout:Mar 09 15:17:23 vm06 ceph-4abede80-1bca-11f1-bcd8-9b1da3c03ce5-osd-2[59842]: 2026-03-09T15:17:23.991+0000 7ff433b1f640 -1 osd.2 19 *** Immediate shutdown (osd_fast_shutdown=true) *** 2026-03-09T15:17:29.267 INFO:journalctl@ceph.osd.2.vm06.stdout:Mar 09 15:17:29 vm06 podman[63427]: 2026-03-09 15:17:29.015190265 +0000 UTC m=+5.036050776 container died 72ec1ee0f14387be9a644b82d136e7be018d3eef6eeec93657c7a138ec72fbc1 (image=quay.ceph.io/ceph-ci/ceph@sha256:8fda260ab1d2d3118a1622f7df75f44f285dfe74e71793626152a711c12bf2cc, name=ceph-4abede80-1bca-11f1-bcd8-9b1da3c03ce5-osd-2, org.label-schema.schema-version=1.0, org.opencontainers.image.authors=Ceph Release Team , CEPH_GIT_REPO=https://github.com/ceph/ceph-ci.git, GANESHA_REPO_BASEURL=https://buildlogs.centos.org/centos/$releasever-stream/storage/$basearch/nfsganesha-5/, org.label-schema.vendor=CentOS, org.opencontainers.image.documentation=https://docs.ceph.com/, CEPH_SHA1=e911bdebe5c8faa3800735d1568fcdca65db60df, org.label-schema.license=GPLv2, org.label-schema.name=CentOS Stream 9 Base Image, FROM_IMAGE=quay.io/centos/centos:stream9, io.buildah.version=1.41.3, OSD_FLAVOR=default, CEPH_REF=squid, org.label-schema.build-date=20260223, ceph=True) 2026-03-09T15:17:29.268 INFO:journalctl@ceph.osd.2.vm06.stdout:Mar 09 15:17:29 vm06 podman[63427]: 2026-03-09 15:17:29.036806438 +0000 UTC m=+5.057666958 container remove 72ec1ee0f14387be9a644b82d136e7be018d3eef6eeec93657c7a138ec72fbc1 (image=quay.ceph.io/ceph-ci/ceph@sha256:8fda260ab1d2d3118a1622f7df75f44f285dfe74e71793626152a711c12bf2cc, name=ceph-4abede80-1bca-11f1-bcd8-9b1da3c03ce5-osd-2, CEPH_GIT_REPO=https://github.com/ceph/ceph-ci.git, FROM_IMAGE=quay.io/centos/centos:stream9, ceph=True, org.label-schema.schema-version=1.0, CEPH_SHA1=e911bdebe5c8faa3800735d1568fcdca65db60df, OSD_FLAVOR=default, io.buildah.version=1.41.3, org.label-schema.license=GPLv2, org.opencontainers.image.authors=Ceph Release Team , GANESHA_REPO_BASEURL=https://buildlogs.centos.org/centos/$releasever-stream/storage/$basearch/nfsganesha-5/, org.label-schema.name=CentOS Stream 9 Base Image, CEPH_REF=squid, org.label-schema.vendor=CentOS, org.label-schema.build-date=20260223, org.opencontainers.image.documentation=https://docs.ceph.com/) 2026-03-09T15:17:29.268 INFO:journalctl@ceph.osd.2.vm06.stdout:Mar 09 15:17:29 vm06 bash[63427]: ceph-4abede80-1bca-11f1-bcd8-9b1da3c03ce5-osd-2 2026-03-09T15:17:29.268 INFO:journalctl@ceph.osd.2.vm06.stdout:Mar 09 15:17:29 vm06 podman[63493]: 2026-03-09 15:17:29.158929159 +0000 UTC m=+0.018068613 container create 08481c083e48113441255ecee003184c4ab50487d5115335773298040c508f78 (image=quay.ceph.io/ceph-ci/ceph@sha256:8fda260ab1d2d3118a1622f7df75f44f285dfe74e71793626152a711c12bf2cc, name=ceph-4abede80-1bca-11f1-bcd8-9b1da3c03ce5-osd-2-deactivate, org.label-schema.build-date=20260223, CEPH_SHA1=e911bdebe5c8faa3800735d1568fcdca65db60df, org.label-schema.schema-version=1.0, FROM_IMAGE=quay.io/centos/centos:stream9, OSD_FLAVOR=default, GANESHA_REPO_BASEURL=https://buildlogs.centos.org/centos/$releasever-stream/storage/$basearch/nfsganesha-5/, org.opencontainers.image.authors=Ceph Release Team , CEPH_GIT_REPO=https://github.com/ceph/ceph-ci.git, org.label-schema.name=CentOS Stream 9 Base Image, io.buildah.version=1.41.3, org.label-schema.license=GPLv2, org.label-schema.vendor=CentOS, org.opencontainers.image.documentation=https://docs.ceph.com/, CEPH_REF=squid, ceph=True) 2026-03-09T15:17:29.268 INFO:journalctl@ceph.osd.2.vm06.stdout:Mar 09 15:17:29 vm06 podman[63493]: 2026-03-09 15:17:29.197194947 +0000 UTC m=+0.056334400 container init 08481c083e48113441255ecee003184c4ab50487d5115335773298040c508f78 (image=quay.ceph.io/ceph-ci/ceph@sha256:8fda260ab1d2d3118a1622f7df75f44f285dfe74e71793626152a711c12bf2cc, name=ceph-4abede80-1bca-11f1-bcd8-9b1da3c03ce5-osd-2-deactivate, org.label-schema.build-date=20260223, FROM_IMAGE=quay.io/centos/centos:stream9, org.opencontainers.image.documentation=https://docs.ceph.com/, org.opencontainers.image.authors=Ceph Release Team , CEPH_REF=squid, CEPH_SHA1=e911bdebe5c8faa3800735d1568fcdca65db60df, org.label-schema.schema-version=1.0, ceph=True, OSD_FLAVOR=default, io.buildah.version=1.41.3, GANESHA_REPO_BASEURL=https://buildlogs.centos.org/centos/$releasever-stream/storage/$basearch/nfsganesha-5/, org.label-schema.license=GPLv2, org.label-schema.vendor=CentOS, CEPH_GIT_REPO=https://github.com/ceph/ceph-ci.git, org.label-schema.name=CentOS Stream 9 Base Image) 2026-03-09T15:17:29.268 INFO:journalctl@ceph.osd.2.vm06.stdout:Mar 09 15:17:29 vm06 podman[63493]: 2026-03-09 15:17:29.204093059 +0000 UTC m=+0.063232522 container start 08481c083e48113441255ecee003184c4ab50487d5115335773298040c508f78 (image=quay.ceph.io/ceph-ci/ceph@sha256:8fda260ab1d2d3118a1622f7df75f44f285dfe74e71793626152a711c12bf2cc, name=ceph-4abede80-1bca-11f1-bcd8-9b1da3c03ce5-osd-2-deactivate, org.label-schema.license=GPLv2, org.label-schema.build-date=20260223, CEPH_GIT_REPO=https://github.com/ceph/ceph-ci.git, OSD_FLAVOR=default, org.opencontainers.image.authors=Ceph Release Team , FROM_IMAGE=quay.io/centos/centos:stream9, CEPH_REF=squid, io.buildah.version=1.41.3, org.label-schema.vendor=CentOS, CEPH_SHA1=e911bdebe5c8faa3800735d1568fcdca65db60df, org.opencontainers.image.documentation=https://docs.ceph.com/, org.label-schema.schema-version=1.0, ceph=True, org.label-schema.name=CentOS Stream 9 Base Image, GANESHA_REPO_BASEURL=https://buildlogs.centos.org/centos/$releasever-stream/storage/$basearch/nfsganesha-5/) 2026-03-09T15:17:29.268 INFO:journalctl@ceph.osd.2.vm06.stdout:Mar 09 15:17:29 vm06 podman[63493]: 2026-03-09 15:17:29.205006208 +0000 UTC m=+0.064145671 container attach 08481c083e48113441255ecee003184c4ab50487d5115335773298040c508f78 (image=quay.ceph.io/ceph-ci/ceph@sha256:8fda260ab1d2d3118a1622f7df75f44f285dfe74e71793626152a711c12bf2cc, name=ceph-4abede80-1bca-11f1-bcd8-9b1da3c03ce5-osd-2-deactivate, CEPH_SHA1=e911bdebe5c8faa3800735d1568fcdca65db60df, org.label-schema.schema-version=1.0, org.label-schema.name=CentOS Stream 9 Base Image, CEPH_REF=squid, org.label-schema.vendor=CentOS, org.opencontainers.image.authors=Ceph Release Team , FROM_IMAGE=quay.io/centos/centos:stream9, org.opencontainers.image.documentation=https://docs.ceph.com/, OSD_FLAVOR=default, org.label-schema.license=GPLv2, ceph=True, CEPH_GIT_REPO=https://github.com/ceph/ceph-ci.git, io.buildah.version=1.41.3, GANESHA_REPO_BASEURL=https://buildlogs.centos.org/centos/$releasever-stream/storage/$basearch/nfsganesha-5/, org.label-schema.build-date=20260223) 2026-03-09T15:17:29.268 INFO:journalctl@ceph.osd.2.vm06.stdout:Mar 09 15:17:29 vm06 podman[63493]: 2026-03-09 15:17:29.149341504 +0000 UTC m=+0.008480977 image pull 654f31e6858eb235bbece362255b685a945f2b6a367e2b88c4930c984fbb214c quay.ceph.io/ceph-ci/ceph@sha256:8fda260ab1d2d3118a1622f7df75f44f285dfe74e71793626152a711c12bf2cc 2026-03-09T15:17:29.360 DEBUG:teuthology.orchestra.run.vm06:> sudo pkill -f 'journalctl -f -n 0 -u ceph-4abede80-1bca-11f1-bcd8-9b1da3c03ce5@osd.2.service' 2026-03-09T15:17:29.388 DEBUG:teuthology.orchestra.run:got remote process result: None 2026-03-09T15:17:29.388 INFO:tasks.cephadm.osd.2:Stopped osd.2 2026-03-09T15:17:29.388 DEBUG:teuthology.orchestra.run.vm03:> sudo /home/ubuntu/cephtest/cephadm rm-cluster --fsid 4abede80-1bca-11f1-bcd8-9b1da3c03ce5 --force --keep-logs 2026-03-09T15:17:29.509 INFO:teuthology.orchestra.run.vm03.stdout:Deleting cluster with fsid: 4abede80-1bca-11f1-bcd8-9b1da3c03ce5 2026-03-09T15:17:30.338 DEBUG:teuthology.orchestra.run.vm04:> sudo /home/ubuntu/cephtest/cephadm rm-cluster --fsid 4abede80-1bca-11f1-bcd8-9b1da3c03ce5 --force --keep-logs 2026-03-09T15:17:30.461 INFO:teuthology.orchestra.run.vm04.stdout:Deleting cluster with fsid: 4abede80-1bca-11f1-bcd8-9b1da3c03ce5 2026-03-09T15:17:31.332 DEBUG:teuthology.orchestra.run.vm06:> sudo /home/ubuntu/cephtest/cephadm rm-cluster --fsid 4abede80-1bca-11f1-bcd8-9b1da3c03ce5 --force --keep-logs 2026-03-09T15:17:31.452 INFO:teuthology.orchestra.run.vm06.stdout:Deleting cluster with fsid: 4abede80-1bca-11f1-bcd8-9b1da3c03ce5 2026-03-09T15:17:32.166 DEBUG:teuthology.orchestra.run.vm03:> sudo rm -f /etc/ceph/ceph.conf /etc/ceph/ceph.client.admin.keyring 2026-03-09T15:17:32.191 DEBUG:teuthology.orchestra.run.vm04:> sudo rm -f /etc/ceph/ceph.conf /etc/ceph/ceph.client.admin.keyring 2026-03-09T15:17:32.215 DEBUG:teuthology.orchestra.run.vm06:> sudo rm -f /etc/ceph/ceph.conf /etc/ceph/ceph.client.admin.keyring 2026-03-09T15:17:32.239 INFO:tasks.cephadm:Archiving crash dumps... 2026-03-09T15:17:32.239 DEBUG:teuthology.misc:Transferring archived files from vm03:/var/lib/ceph/4abede80-1bca-11f1-bcd8-9b1da3c03ce5/crash to /archive/kyr-2026-03-09_11:23:05-orch-squid-none-default-vps/520/remote/vm03/crash 2026-03-09T15:17:32.239 DEBUG:teuthology.orchestra.run.vm03:> sudo tar c -f - -C /var/lib/ceph/4abede80-1bca-11f1-bcd8-9b1da3c03ce5/crash -- . 2026-03-09T15:17:32.264 INFO:teuthology.orchestra.run.vm03.stderr:tar: /var/lib/ceph/4abede80-1bca-11f1-bcd8-9b1da3c03ce5/crash: Cannot open: No such file or directory 2026-03-09T15:17:32.264 INFO:teuthology.orchestra.run.vm03.stderr:tar: Error is not recoverable: exiting now 2026-03-09T15:17:32.265 DEBUG:teuthology.misc:Transferring archived files from vm04:/var/lib/ceph/4abede80-1bca-11f1-bcd8-9b1da3c03ce5/crash to /archive/kyr-2026-03-09_11:23:05-orch-squid-none-default-vps/520/remote/vm04/crash 2026-03-09T15:17:32.265 DEBUG:teuthology.orchestra.run.vm04:> sudo tar c -f - -C /var/lib/ceph/4abede80-1bca-11f1-bcd8-9b1da3c03ce5/crash -- . 2026-03-09T15:17:32.289 INFO:teuthology.orchestra.run.vm04.stderr:tar: /var/lib/ceph/4abede80-1bca-11f1-bcd8-9b1da3c03ce5/crash: Cannot open: No such file or directory 2026-03-09T15:17:32.289 INFO:teuthology.orchestra.run.vm04.stderr:tar: Error is not recoverable: exiting now 2026-03-09T15:17:32.290 DEBUG:teuthology.misc:Transferring archived files from vm06:/var/lib/ceph/4abede80-1bca-11f1-bcd8-9b1da3c03ce5/crash to /archive/kyr-2026-03-09_11:23:05-orch-squid-none-default-vps/520/remote/vm06/crash 2026-03-09T15:17:32.290 DEBUG:teuthology.orchestra.run.vm06:> sudo tar c -f - -C /var/lib/ceph/4abede80-1bca-11f1-bcd8-9b1da3c03ce5/crash -- . 2026-03-09T15:17:32.313 INFO:teuthology.orchestra.run.vm06.stderr:tar: /var/lib/ceph/4abede80-1bca-11f1-bcd8-9b1da3c03ce5/crash: Cannot open: No such file or directory 2026-03-09T15:17:32.314 INFO:teuthology.orchestra.run.vm06.stderr:tar: Error is not recoverable: exiting now 2026-03-09T15:17:32.314 INFO:tasks.cephadm:Checking cluster log for badness... 2026-03-09T15:17:32.314 DEBUG:teuthology.orchestra.run.vm03:> sudo egrep '\[ERR\]|\[WRN\]|\[SEC\]' /var/log/ceph/4abede80-1bca-11f1-bcd8-9b1da3c03ce5/ceph.log | egrep CEPHADM_ | egrep -v '\(MDS_ALL_DOWN\)' | egrep -v '\(MDS_UP_LESS_THAN_MAX\)' | egrep -v MON_DOWN | egrep -v POOL_APP_NOT_ENABLED | egrep -v 'mon down' | egrep -v 'mons down' | egrep -v 'out of quorum' | egrep -v CEPHADM_FAILED_DAEMON | head -n 1 2026-03-09T15:17:32.343 INFO:tasks.cephadm:Compressing logs... 2026-03-09T15:17:32.343 DEBUG:teuthology.orchestra.run.vm03:> time sudo find /var/log/ceph /var/log/rbd-target-api -name '*.log' -print0 | sudo xargs --max-args=1 --max-procs=0 --verbose -0 --no-run-if-empty -- gzip -5 --verbose -- 2026-03-09T15:17:32.385 DEBUG:teuthology.orchestra.run.vm04:> time sudo find /var/log/ceph /var/log/rbd-target-api -name '*.log' -print0 | sudo xargs --max-args=1 --max-procs=0 --verbose -0 --no-run-if-empty -- gzip -5 --verbose -- 2026-03-09T15:17:32.387 DEBUG:teuthology.orchestra.run.vm06:> time sudo find /var/log/ceph /var/log/rbd-target-api -name '*.log' -print0 | sudo xargs --max-args=1 --max-procs=0 --verbose -0 --no-run-if-empty -- gzip -5 --verbose -- 2026-03-09T15:17:32.408 INFO:teuthology.orchestra.run.vm03.stderr:find: gzip -5 --verbose -- /var/log/ceph/cephadm.log 2026-03-09T15:17:32.408 INFO:teuthology.orchestra.run.vm03.stderr:‘/var/log/rbd-target-api’: No such file or directory 2026-03-09T15:17:32.409 INFO:teuthology.orchestra.run.vm03.stderr:gzip -5 --verbose -- /var/log/ceph/4abede80-1bca-11f1-bcd8-9b1da3c03ce5/ceph-mon.a.log 2026-03-09T15:17:32.409 INFO:teuthology.orchestra.run.vm03.stderr:gzip -5 --verbose -- /var/log/ceph/4abede80-1bca-11f1-bcd8-9b1da3c03ce5/ceph.log 2026-03-09T15:17:32.410 INFO:teuthology.orchestra.run.vm03.stderr:/var/log/ceph/cephadm.log: /var/log/ceph/4abede80-1bca-11f1-bcd8-9b1da3c03ce5/ceph-mon.a.log: 89.9% -- replaced with /var/log/ceph/cephadm.log.gz 2026-03-09T15:17:32.410 INFO:teuthology.orchestra.run.vm03.stderr:gzip -5 --verbose -- /var/log/ceph/4abede80-1bca-11f1-bcd8-9b1da3c03ce5/ceph-mgr.a.log 2026-03-09T15:17:32.411 INFO:teuthology.orchestra.run.vm03.stderr:/var/log/ceph/4abede80-1bca-11f1-bcd8-9b1da3c03ce5/ceph.log: 86.9% -- replaced with /var/log/ceph/4abede80-1bca-11f1-bcd8-9b1da3c03ce5/ceph.log.gz 2026-03-09T15:17:32.411 INFO:teuthology.orchestra.run.vm03.stderr:gzip -5 --verbose -- /var/log/ceph/4abede80-1bca-11f1-bcd8-9b1da3c03ce5/ceph.audit.log 2026-03-09T15:17:32.412 INFO:teuthology.orchestra.run.vm04.stderr:gzip -5 --verbose -- /var/log/ceph/cephadm.log 2026-03-09T15:17:32.412 INFO:teuthology.orchestra.run.vm04.stderr:find: ‘/var/log/rbd-target-api’: No such file or directory 2026-03-09T15:17:32.413 INFO:teuthology.orchestra.run.vm04.stderr:gzip -5 --verbose -- /var/log/ceph/4abede80-1bca-11f1-bcd8-9b1da3c03ce5/ceph-volume.log 2026-03-09T15:17:32.413 INFO:teuthology.orchestra.run.vm04.stderr:/var/log/ceph/cephadm.log: gzip -5 --verbose -- /var/log/ceph/4abede80-1bca-11f1-bcd8-9b1da3c03ce5/ceph-mon.b.log 2026-03-09T15:17:32.414 INFO:teuthology.orchestra.run.vm06.stderr:gzip -5 --verbose -- /var/log/ceph/cephadm.log 2026-03-09T15:17:32.414 INFO:teuthology.orchestra.run.vm06.stderr:find: ‘/var/log/rbd-target-api’: No such file or directory 2026-03-09T15:17:32.414 INFO:teuthology.orchestra.run.vm04.stderr: 88.5% -- replaced with /var/log/ceph/cephadm.log.gz 2026-03-09T15:17:32.415 INFO:teuthology.orchestra.run.vm04.stderr:/var/log/ceph/4abede80-1bca-11f1-bcd8-9b1da3c03ce5/ceph-volume.log: gzip -5 --verbose -- /var/log/ceph/4abede80-1bca-11f1-bcd8-9b1da3c03ce5/ceph.cephadm.log 2026-03-09T15:17:32.415 INFO:teuthology.orchestra.run.vm06.stderr:gzip -5 --verbose -- /var/log/ceph/4abede80-1bca-11f1-bcd8-9b1da3c03ce5/ceph-volume.log 2026-03-09T15:17:32.415 INFO:teuthology.orchestra.run.vm06.stderr:gzip -5 --verbose -- /var/log/ceph/4abede80-1bca-11f1-bcd8-9b1da3c03ce5/ceph-mon.c.log 2026-03-09T15:17:32.416 INFO:teuthology.orchestra.run.vm04.stderr:/var/log/ceph/4abede80-1bca-11f1-bcd8-9b1da3c03ce5/ceph-mon.b.log: 94.7% -- replaced with /var/log/ceph/4abede80-1bca-11f1-bcd8-9b1da3c03ce5/ceph-volume.log.gz 2026-03-09T15:17:32.416 INFO:teuthology.orchestra.run.vm06.stderr:/var/log/ceph/cephadm.log: /var/log/ceph/4abede80-1bca-11f1-bcd8-9b1da3c03ce5/ceph-volume.log: 88.5% -- replaced with /var/log/ceph/cephadm.log.gz 2026-03-09T15:17:32.416 INFO:teuthology.orchestra.run.vm03.stderr:/var/log/ceph/4abede80-1bca-11f1-bcd8-9b1da3c03ce5/ceph-mgr.a.log: gzip -5 --verbose -- /var/log/ceph/4abede80-1bca-11f1-bcd8-9b1da3c03ce5/ceph.cephadm.log 2026-03-09T15:17:32.416 INFO:teuthology.orchestra.run.vm04.stderr:gzip -5 --verbose -- /var/log/ceph/4abede80-1bca-11f1-bcd8-9b1da3c03ce5/ceph.audit.log 2026-03-09T15:17:32.416 INFO:teuthology.orchestra.run.vm06.stderr:gzip -5 --verbose -- /var/log/ceph/4abede80-1bca-11f1-bcd8-9b1da3c03ce5/ceph.cephadm.log 2026-03-09T15:17:32.417 INFO:teuthology.orchestra.run.vm04.stderr:/var/log/ceph/4abede80-1bca-11f1-bcd8-9b1da3c03ce5/ceph.cephadm.log: 80.0% -- replaced with /var/log/ceph/4abede80-1bca-11f1-bcd8-9b1da3c03ce5/ceph.cephadm.log.gz 2026-03-09T15:17:32.417 INFO:teuthology.orchestra.run.vm04.stderr:gzip -5 --verbose -- /var/log/ceph/4abede80-1bca-11f1-bcd8-9b1da3c03ce5/ceph.log 2026-03-09T15:17:32.417 INFO:teuthology.orchestra.run.vm06.stderr:/var/log/ceph/4abede80-1bca-11f1-bcd8-9b1da3c03ce5/ceph-mon.c.log: 94.7% -- replaced with /var/log/ceph/4abede80-1bca-11f1-bcd8-9b1da3c03ce5/ceph-volume.log.gz 2026-03-09T15:17:32.418 INFO:teuthology.orchestra.run.vm06.stderr:gzip -5 --verbose -- /var/log/ceph/4abede80-1bca-11f1-bcd8-9b1da3c03ce5/ceph.audit.log 2026-03-09T15:17:32.418 INFO:teuthology.orchestra.run.vm04.stderr:/var/log/ceph/4abede80-1bca-11f1-bcd8-9b1da3c03ce5/ceph.audit.log: 90.2% -- replaced with /var/log/ceph/4abede80-1bca-11f1-bcd8-9b1da3c03ce5/ceph.audit.log.gz 2026-03-09T15:17:32.418 INFO:teuthology.orchestra.run.vm06.stderr:/var/log/ceph/4abede80-1bca-11f1-bcd8-9b1da3c03ce5/ceph.cephadm.log: 80.0% -- replaced with /var/log/ceph/4abede80-1bca-11f1-bcd8-9b1da3c03ce5/ceph.cephadm.log.gz 2026-03-09T15:17:32.418 INFO:teuthology.orchestra.run.vm04.stderr:gzip -5 --verbose -- /var/log/ceph/4abede80-1bca-11f1-bcd8-9b1da3c03ce5/ceph-mgr.b.log 2026-03-09T15:17:32.419 INFO:teuthology.orchestra.run.vm03.stderr:/var/log/ceph/4abede80-1bca-11f1-bcd8-9b1da3c03ce5/ceph.audit.log: 89.8% -- replaced with /var/log/ceph/4abede80-1bca-11f1-bcd8-9b1da3c03ce5/ceph.audit.log.gz 2026-03-09T15:17:32.419 INFO:teuthology.orchestra.run.vm04.stderr:/var/log/ceph/4abede80-1bca-11f1-bcd8-9b1da3c03ce5/ceph.log: 87.2% -- replaced with /var/log/ceph/4abede80-1bca-11f1-bcd8-9b1da3c03ce5/ceph.log.gz 2026-03-09T15:17:32.419 INFO:teuthology.orchestra.run.vm04.stderr:gzip -5 --verbose -- /var/log/ceph/4abede80-1bca-11f1-bcd8-9b1da3c03ce5/ceph-osd.1.log 2026-03-09T15:17:32.421 INFO:teuthology.orchestra.run.vm04.stderr:/var/log/ceph/4abede80-1bca-11f1-bcd8-9b1da3c03ce5/ceph-mgr.b.log: 90.8% -- replaced with /var/log/ceph/4abede80-1bca-11f1-bcd8-9b1da3c03ce5/ceph-mgr.b.log.gz 2026-03-09T15:17:32.422 INFO:teuthology.orchestra.run.vm06.stderr:gzip -5 --verbose -- /var/log/ceph/4abede80-1bca-11f1-bcd8-9b1da3c03ce5/ceph.log 2026-03-09T15:17:32.422 INFO:teuthology.orchestra.run.vm03.stderr:gzip -5 --verbose -- /var/log/ceph/4abede80-1bca-11f1-bcd8-9b1da3c03ce5/ceph-volume.log 2026-03-09T15:17:32.423 INFO:teuthology.orchestra.run.vm03.stderr:/var/log/ceph/4abede80-1bca-11f1-bcd8-9b1da3c03ce5/ceph.cephadm.log: 81.0% -- replaced with /var/log/ceph/4abede80-1bca-11f1-bcd8-9b1da3c03ce5/ceph.cephadm.log.gz 2026-03-09T15:17:32.423 INFO:teuthology.orchestra.run.vm06.stderr:/var/log/ceph/4abede80-1bca-11f1-bcd8-9b1da3c03ce5/ceph.audit.log: 90.2% -- replaced with /var/log/ceph/4abede80-1bca-11f1-bcd8-9b1da3c03ce5/ceph.audit.log.gz 2026-03-09T15:17:32.423 INFO:teuthology.orchestra.run.vm06.stderr:gzip -5 --verbose -- /var/log/ceph/4abede80-1bca-11f1-bcd8-9b1da3c03ce5/ceph-osd.2.log 2026-03-09T15:17:32.423 INFO:teuthology.orchestra.run.vm06.stderr:/var/log/ceph/4abede80-1bca-11f1-bcd8-9b1da3c03ce5/ceph.log: 87.2% -- replaced with /var/log/ceph/4abede80-1bca-11f1-bcd8-9b1da3c03ce5/ceph.log.gz 2026-03-09T15:17:32.427 INFO:teuthology.orchestra.run.vm03.stderr:gzip -5 --verbose -- /var/log/ceph/4abede80-1bca-11f1-bcd8-9b1da3c03ce5/ceph-osd.0.log 2026-03-09T15:17:32.431 INFO:teuthology.orchestra.run.vm03.stderr:/var/log/ceph/4abede80-1bca-11f1-bcd8-9b1da3c03ce5/ceph-volume.log: 94.7% -- replaced with /var/log/ceph/4abede80-1bca-11f1-bcd8-9b1da3c03ce5/ceph-volume.log.gz 2026-03-09T15:17:32.439 INFO:teuthology.orchestra.run.vm06.stderr:/var/log/ceph/4abede80-1bca-11f1-bcd8-9b1da3c03ce5/ceph-osd.2.log: 93.3% -- replaced with /var/log/ceph/4abede80-1bca-11f1-bcd8-9b1da3c03ce5/ceph-osd.2.log.gz 2026-03-09T15:17:32.452 INFO:teuthology.orchestra.run.vm06.stderr: 93.0% -- replaced with /var/log/ceph/4abede80-1bca-11f1-bcd8-9b1da3c03ce5/ceph-mon.c.log.gz 2026-03-09T15:17:32.452 INFO:teuthology.orchestra.run.vm06.stderr: 2026-03-09T15:17:32.452 INFO:teuthology.orchestra.run.vm06.stderr:real 0m0.046s 2026-03-09T15:17:32.452 INFO:teuthology.orchestra.run.vm06.stderr:user 0m0.052s 2026-03-09T15:17:32.452 INFO:teuthology.orchestra.run.vm06.stderr:sys 0m0.019s 2026-03-09T15:17:32.452 INFO:teuthology.orchestra.run.vm04.stderr:/var/log/ceph/4abede80-1bca-11f1-bcd8-9b1da3c03ce5/ceph-osd.1.log: 93.0% -- replaced with /var/log/ceph/4abede80-1bca-11f1-bcd8-9b1da3c03ce5/ceph-mon.b.log.gz 2026-03-09T15:17:32.452 INFO:teuthology.orchestra.run.vm04.stderr: 93.4% -- replaced with /var/log/ceph/4abede80-1bca-11f1-bcd8-9b1da3c03ce5/ceph-osd.1.log.gz 2026-03-09T15:17:32.452 INFO:teuthology.orchestra.run.vm04.stderr: 2026-03-09T15:17:32.452 INFO:teuthology.orchestra.run.vm04.stderr:real 0m0.045s 2026-03-09T15:17:32.452 INFO:teuthology.orchestra.run.vm04.stderr:user 0m0.061s 2026-03-09T15:17:32.452 INFO:teuthology.orchestra.run.vm04.stderr:sys 0m0.019s 2026-03-09T15:17:32.462 INFO:teuthology.orchestra.run.vm03.stderr:/var/log/ceph/4abede80-1bca-11f1-bcd8-9b1da3c03ce5/ceph-osd.0.log: 90.7% -- replaced with /var/log/ceph/4abede80-1bca-11f1-bcd8-9b1da3c03ce5/ceph-mgr.a.log.gz 2026-03-09T15:17:32.462 INFO:teuthology.orchestra.run.vm03.stderr: 93.3% -- replaced with /var/log/ceph/4abede80-1bca-11f1-bcd8-9b1da3c03ce5/ceph-osd.0.log.gz 2026-03-09T15:17:32.510 INFO:teuthology.orchestra.run.vm03.stderr: 91.2% -- replaced with /var/log/ceph/4abede80-1bca-11f1-bcd8-9b1da3c03ce5/ceph-mon.a.log.gz 2026-03-09T15:17:32.511 INFO:teuthology.orchestra.run.vm03.stderr: 2026-03-09T15:17:32.511 INFO:teuthology.orchestra.run.vm03.stderr:real 0m0.113s 2026-03-09T15:17:32.511 INFO:teuthology.orchestra.run.vm03.stderr:user 0m0.149s 2026-03-09T15:17:32.511 INFO:teuthology.orchestra.run.vm03.stderr:sys 0m0.022s 2026-03-09T15:17:32.511 INFO:tasks.cephadm:Archiving logs... 2026-03-09T15:17:32.511 DEBUG:teuthology.misc:Transferring archived files from vm03:/var/log/ceph to /archive/kyr-2026-03-09_11:23:05-orch-squid-none-default-vps/520/remote/vm03/log 2026-03-09T15:17:32.512 DEBUG:teuthology.orchestra.run.vm03:> sudo tar c -f - -C /var/log/ceph -- . 2026-03-09T15:17:32.588 DEBUG:teuthology.misc:Transferring archived files from vm04:/var/log/ceph to /archive/kyr-2026-03-09_11:23:05-orch-squid-none-default-vps/520/remote/vm04/log 2026-03-09T15:17:32.588 DEBUG:teuthology.orchestra.run.vm04:> sudo tar c -f - -C /var/log/ceph -- . 2026-03-09T15:17:32.615 DEBUG:teuthology.misc:Transferring archived files from vm06:/var/log/ceph to /archive/kyr-2026-03-09_11:23:05-orch-squid-none-default-vps/520/remote/vm06/log 2026-03-09T15:17:32.615 DEBUG:teuthology.orchestra.run.vm06:> sudo tar c -f - -C /var/log/ceph -- . 2026-03-09T15:17:32.641 INFO:tasks.cephadm:Removing cluster... 2026-03-09T15:17:32.641 DEBUG:teuthology.orchestra.run.vm03:> sudo /home/ubuntu/cephtest/cephadm rm-cluster --fsid 4abede80-1bca-11f1-bcd8-9b1da3c03ce5 --force 2026-03-09T15:17:32.764 INFO:teuthology.orchestra.run.vm03.stdout:Deleting cluster with fsid: 4abede80-1bca-11f1-bcd8-9b1da3c03ce5 2026-03-09T15:17:32.955 DEBUG:teuthology.orchestra.run.vm04:> sudo /home/ubuntu/cephtest/cephadm rm-cluster --fsid 4abede80-1bca-11f1-bcd8-9b1da3c03ce5 --force 2026-03-09T15:17:33.073 INFO:teuthology.orchestra.run.vm04.stdout:Deleting cluster with fsid: 4abede80-1bca-11f1-bcd8-9b1da3c03ce5 2026-03-09T15:17:33.263 DEBUG:teuthology.orchestra.run.vm06:> sudo /home/ubuntu/cephtest/cephadm rm-cluster --fsid 4abede80-1bca-11f1-bcd8-9b1da3c03ce5 --force 2026-03-09T15:17:33.385 INFO:teuthology.orchestra.run.vm06.stdout:Deleting cluster with fsid: 4abede80-1bca-11f1-bcd8-9b1da3c03ce5 2026-03-09T15:17:33.576 INFO:tasks.cephadm:Removing cephadm ... 2026-03-09T15:17:33.576 DEBUG:teuthology.orchestra.run.vm03:> rm -rf /home/ubuntu/cephtest/cephadm 2026-03-09T15:17:33.591 DEBUG:teuthology.orchestra.run.vm04:> rm -rf /home/ubuntu/cephtest/cephadm 2026-03-09T15:17:33.605 DEBUG:teuthology.orchestra.run.vm06:> rm -rf /home/ubuntu/cephtest/cephadm 2026-03-09T15:17:33.619 INFO:tasks.cephadm:Teardown complete 2026-03-09T15:17:33.619 DEBUG:teuthology.run_tasks:Unwinding manager install 2026-03-09T15:17:33.621 INFO:teuthology.task.install.util:Removing shipped files: /home/ubuntu/cephtest/valgrind.supp /usr/bin/daemon-helper /usr/bin/adjust-ulimits /usr/bin/stdin-killer... 2026-03-09T15:17:33.622 DEBUG:teuthology.orchestra.run.vm03:> sudo rm -f -- /home/ubuntu/cephtest/valgrind.supp /usr/bin/daemon-helper /usr/bin/adjust-ulimits /usr/bin/stdin-killer 2026-03-09T15:17:33.632 DEBUG:teuthology.orchestra.run.vm04:> sudo rm -f -- /home/ubuntu/cephtest/valgrind.supp /usr/bin/daemon-helper /usr/bin/adjust-ulimits /usr/bin/stdin-killer 2026-03-09T15:17:33.647 DEBUG:teuthology.orchestra.run.vm06:> sudo rm -f -- /home/ubuntu/cephtest/valgrind.supp /usr/bin/daemon-helper /usr/bin/adjust-ulimits /usr/bin/stdin-killer 2026-03-09T15:17:33.689 INFO:teuthology.task.install.rpm:Removing packages: ceph-radosgw, ceph-test, ceph, ceph-base, cephadm, ceph-immutable-object-cache, ceph-mgr, ceph-mgr-dashboard, ceph-mgr-diskprediction-local, ceph-mgr-rook, ceph-mgr-cephadm, ceph-fuse, ceph-volume, librados-devel, libcephfs2, libcephfs-devel, librados2, librbd1, python3-rados, python3-rgw, python3-cephfs, python3-rbd, rbd-fuse, rbd-mirror, rbd-nbd on rpm system. 2026-03-09T15:17:33.690 DEBUG:teuthology.orchestra.run.vm03:> 2026-03-09T15:17:33.690 DEBUG:teuthology.orchestra.run.vm03:> for d in ceph-radosgw ceph-test ceph ceph-base cephadm ceph-immutable-object-cache ceph-mgr ceph-mgr-dashboard ceph-mgr-diskprediction-local ceph-mgr-rook ceph-mgr-cephadm ceph-fuse ceph-volume librados-devel libcephfs2 libcephfs-devel librados2 librbd1 python3-rados python3-rgw python3-cephfs python3-rbd rbd-fuse rbd-mirror rbd-nbd ; do 2026-03-09T15:17:33.690 DEBUG:teuthology.orchestra.run.vm03:> sudo yum -y remove $d || true 2026-03-09T15:17:33.690 DEBUG:teuthology.orchestra.run.vm03:> done 2026-03-09T15:17:33.696 INFO:teuthology.task.install.rpm:Removing packages: ceph-radosgw, ceph-test, ceph, ceph-base, cephadm, ceph-immutable-object-cache, ceph-mgr, ceph-mgr-dashboard, ceph-mgr-diskprediction-local, ceph-mgr-rook, ceph-mgr-cephadm, ceph-fuse, ceph-volume, librados-devel, libcephfs2, libcephfs-devel, librados2, librbd1, python3-rados, python3-rgw, python3-cephfs, python3-rbd, rbd-fuse, rbd-mirror, rbd-nbd on rpm system. 2026-03-09T15:17:33.696 DEBUG:teuthology.orchestra.run.vm04:> 2026-03-09T15:17:33.696 DEBUG:teuthology.orchestra.run.vm04:> for d in ceph-radosgw ceph-test ceph ceph-base cephadm ceph-immutable-object-cache ceph-mgr ceph-mgr-dashboard ceph-mgr-diskprediction-local ceph-mgr-rook ceph-mgr-cephadm ceph-fuse ceph-volume librados-devel libcephfs2 libcephfs-devel librados2 librbd1 python3-rados python3-rgw python3-cephfs python3-rbd rbd-fuse rbd-mirror rbd-nbd ; do 2026-03-09T15:17:33.696 DEBUG:teuthology.orchestra.run.vm04:> sudo yum -y remove $d || true 2026-03-09T15:17:33.696 DEBUG:teuthology.orchestra.run.vm04:> done 2026-03-09T15:17:33.702 INFO:teuthology.task.install.rpm:Removing packages: ceph-radosgw, ceph-test, ceph, ceph-base, cephadm, ceph-immutable-object-cache, ceph-mgr, ceph-mgr-dashboard, ceph-mgr-diskprediction-local, ceph-mgr-rook, ceph-mgr-cephadm, ceph-fuse, ceph-volume, librados-devel, libcephfs2, libcephfs-devel, librados2, librbd1, python3-rados, python3-rgw, python3-cephfs, python3-rbd, rbd-fuse, rbd-mirror, rbd-nbd on rpm system. 2026-03-09T15:17:33.702 DEBUG:teuthology.orchestra.run.vm06:> 2026-03-09T15:17:33.702 DEBUG:teuthology.orchestra.run.vm06:> for d in ceph-radosgw ceph-test ceph ceph-base cephadm ceph-immutable-object-cache ceph-mgr ceph-mgr-dashboard ceph-mgr-diskprediction-local ceph-mgr-rook ceph-mgr-cephadm ceph-fuse ceph-volume librados-devel libcephfs2 libcephfs-devel librados2 librbd1 python3-rados python3-rgw python3-cephfs python3-rbd rbd-fuse rbd-mirror rbd-nbd ; do 2026-03-09T15:17:33.702 DEBUG:teuthology.orchestra.run.vm06:> sudo yum -y remove $d || true 2026-03-09T15:17:33.702 DEBUG:teuthology.orchestra.run.vm06:> done 2026-03-09T15:17:33.870 INFO:teuthology.orchestra.run.vm03.stdout:Dependencies resolved. 2026-03-09T15:17:33.871 INFO:teuthology.orchestra.run.vm03.stdout:================================================================================ 2026-03-09T15:17:33.871 INFO:teuthology.orchestra.run.vm03.stdout: Package Arch Version Repository Size 2026-03-09T15:17:33.871 INFO:teuthology.orchestra.run.vm03.stdout:================================================================================ 2026-03-09T15:17:33.871 INFO:teuthology.orchestra.run.vm03.stdout:Removing: 2026-03-09T15:17:33.871 INFO:teuthology.orchestra.run.vm03.stdout: ceph-radosgw x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 39 M 2026-03-09T15:17:33.871 INFO:teuthology.orchestra.run.vm03.stdout:Removing unused dependencies: 2026-03-09T15:17:33.871 INFO:teuthology.orchestra.run.vm03.stdout: mailcap noarch 2.1.49-5.el9 @baseos 78 k 2026-03-09T15:17:33.871 INFO:teuthology.orchestra.run.vm03.stdout: 2026-03-09T15:17:33.871 INFO:teuthology.orchestra.run.vm03.stdout:Transaction Summary 2026-03-09T15:17:33.871 INFO:teuthology.orchestra.run.vm03.stdout:================================================================================ 2026-03-09T15:17:33.871 INFO:teuthology.orchestra.run.vm03.stdout:Remove 2 Packages 2026-03-09T15:17:33.871 INFO:teuthology.orchestra.run.vm03.stdout: 2026-03-09T15:17:33.871 INFO:teuthology.orchestra.run.vm03.stdout:Freed space: 39 M 2026-03-09T15:17:33.871 INFO:teuthology.orchestra.run.vm03.stdout:Running transaction check 2026-03-09T15:17:33.873 INFO:teuthology.orchestra.run.vm03.stdout:Transaction check succeeded. 2026-03-09T15:17:33.873 INFO:teuthology.orchestra.run.vm03.stdout:Running transaction test 2026-03-09T15:17:33.886 INFO:teuthology.orchestra.run.vm03.stdout:Transaction test succeeded. 2026-03-09T15:17:33.886 INFO:teuthology.orchestra.run.vm03.stdout:Running transaction 2026-03-09T15:17:33.888 INFO:teuthology.orchestra.run.vm04.stdout:Dependencies resolved. 2026-03-09T15:17:33.888 INFO:teuthology.orchestra.run.vm04.stdout:================================================================================ 2026-03-09T15:17:33.888 INFO:teuthology.orchestra.run.vm04.stdout: Package Arch Version Repository Size 2026-03-09T15:17:33.889 INFO:teuthology.orchestra.run.vm04.stdout:================================================================================ 2026-03-09T15:17:33.889 INFO:teuthology.orchestra.run.vm04.stdout:Removing: 2026-03-09T15:17:33.889 INFO:teuthology.orchestra.run.vm04.stdout: ceph-radosgw x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 39 M 2026-03-09T15:17:33.889 INFO:teuthology.orchestra.run.vm04.stdout:Removing unused dependencies: 2026-03-09T15:17:33.889 INFO:teuthology.orchestra.run.vm04.stdout: mailcap noarch 2.1.49-5.el9 @baseos 78 k 2026-03-09T15:17:33.889 INFO:teuthology.orchestra.run.vm04.stdout: 2026-03-09T15:17:33.889 INFO:teuthology.orchestra.run.vm04.stdout:Transaction Summary 2026-03-09T15:17:33.889 INFO:teuthology.orchestra.run.vm04.stdout:================================================================================ 2026-03-09T15:17:33.889 INFO:teuthology.orchestra.run.vm04.stdout:Remove 2 Packages 2026-03-09T15:17:33.889 INFO:teuthology.orchestra.run.vm04.stdout: 2026-03-09T15:17:33.889 INFO:teuthology.orchestra.run.vm04.stdout:Freed space: 39 M 2026-03-09T15:17:33.889 INFO:teuthology.orchestra.run.vm04.stdout:Running transaction check 2026-03-09T15:17:33.891 INFO:teuthology.orchestra.run.vm04.stdout:Transaction check succeeded. 2026-03-09T15:17:33.891 INFO:teuthology.orchestra.run.vm04.stdout:Running transaction test 2026-03-09T15:17:33.901 INFO:teuthology.orchestra.run.vm06.stdout:Dependencies resolved. 2026-03-09T15:17:33.902 INFO:teuthology.orchestra.run.vm06.stdout:================================================================================ 2026-03-09T15:17:33.902 INFO:teuthology.orchestra.run.vm06.stdout: Package Arch Version Repository Size 2026-03-09T15:17:33.902 INFO:teuthology.orchestra.run.vm06.stdout:================================================================================ 2026-03-09T15:17:33.902 INFO:teuthology.orchestra.run.vm06.stdout:Removing: 2026-03-09T15:17:33.902 INFO:teuthology.orchestra.run.vm06.stdout: ceph-radosgw x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 39 M 2026-03-09T15:17:33.902 INFO:teuthology.orchestra.run.vm06.stdout:Removing unused dependencies: 2026-03-09T15:17:33.902 INFO:teuthology.orchestra.run.vm06.stdout: mailcap noarch 2.1.49-5.el9 @baseos 78 k 2026-03-09T15:17:33.902 INFO:teuthology.orchestra.run.vm06.stdout: 2026-03-09T15:17:33.902 INFO:teuthology.orchestra.run.vm06.stdout:Transaction Summary 2026-03-09T15:17:33.902 INFO:teuthology.orchestra.run.vm06.stdout:================================================================================ 2026-03-09T15:17:33.902 INFO:teuthology.orchestra.run.vm06.stdout:Remove 2 Packages 2026-03-09T15:17:33.902 INFO:teuthology.orchestra.run.vm06.stdout: 2026-03-09T15:17:33.902 INFO:teuthology.orchestra.run.vm06.stdout:Freed space: 39 M 2026-03-09T15:17:33.902 INFO:teuthology.orchestra.run.vm06.stdout:Running transaction check 2026-03-09T15:17:33.904 INFO:teuthology.orchestra.run.vm04.stdout:Transaction test succeeded. 2026-03-09T15:17:33.904 INFO:teuthology.orchestra.run.vm04.stdout:Running transaction 2026-03-09T15:17:33.904 INFO:teuthology.orchestra.run.vm06.stdout:Transaction check succeeded. 2026-03-09T15:17:33.904 INFO:teuthology.orchestra.run.vm06.stdout:Running transaction test 2026-03-09T15:17:33.916 INFO:teuthology.orchestra.run.vm03.stdout: Preparing : 1/1 2026-03-09T15:17:33.917 INFO:teuthology.orchestra.run.vm06.stdout:Transaction test succeeded. 2026-03-09T15:17:33.917 INFO:teuthology.orchestra.run.vm06.stdout:Running transaction 2026-03-09T15:17:33.934 INFO:teuthology.orchestra.run.vm04.stdout: Preparing : 1/1 2026-03-09T15:17:33.939 INFO:teuthology.orchestra.run.vm03.stdout: Running scriptlet: ceph-radosgw-2:19.2.3-678.ge911bdeb.el9.x86_64 1/2 2026-03-09T15:17:33.939 INFO:teuthology.orchestra.run.vm03.stdout:Glob pattern passed to enable, but globs are not supported for this. 2026-03-09T15:17:33.939 INFO:teuthology.orchestra.run.vm03.stdout:Invalid unit name "ceph-radosgw@*.service" escaped as "ceph-radosgw@\x2a.service". 2026-03-09T15:17:33.939 INFO:teuthology.orchestra.run.vm03.stdout:Removed "/etc/systemd/system/multi-user.target.wants/ceph-radosgw.target". 2026-03-09T15:17:33.939 INFO:teuthology.orchestra.run.vm03.stdout:Removed "/etc/systemd/system/ceph.target.wants/ceph-radosgw.target". 2026-03-09T15:17:33.939 INFO:teuthology.orchestra.run.vm03.stdout: 2026-03-09T15:17:33.942 INFO:teuthology.orchestra.run.vm03.stdout: Erasing : ceph-radosgw-2:19.2.3-678.ge911bdeb.el9.x86_64 1/2 2026-03-09T15:17:33.948 INFO:teuthology.orchestra.run.vm06.stdout: Preparing : 1/1 2026-03-09T15:17:33.950 INFO:teuthology.orchestra.run.vm03.stdout: Running scriptlet: ceph-radosgw-2:19.2.3-678.ge911bdeb.el9.x86_64 1/2 2026-03-09T15:17:33.956 INFO:teuthology.orchestra.run.vm04.stdout: Running scriptlet: ceph-radosgw-2:19.2.3-678.ge911bdeb.el9.x86_64 1/2 2026-03-09T15:17:33.956 INFO:teuthology.orchestra.run.vm04.stdout:Glob pattern passed to enable, but globs are not supported for this. 2026-03-09T15:17:33.956 INFO:teuthology.orchestra.run.vm04.stdout:Invalid unit name "ceph-radosgw@*.service" escaped as "ceph-radosgw@\x2a.service". 2026-03-09T15:17:33.956 INFO:teuthology.orchestra.run.vm04.stdout:Removed "/etc/systemd/system/multi-user.target.wants/ceph-radosgw.target". 2026-03-09T15:17:33.956 INFO:teuthology.orchestra.run.vm04.stdout:Removed "/etc/systemd/system/ceph.target.wants/ceph-radosgw.target". 2026-03-09T15:17:33.956 INFO:teuthology.orchestra.run.vm04.stdout: 2026-03-09T15:17:33.959 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : ceph-radosgw-2:19.2.3-678.ge911bdeb.el9.x86_64 1/2 2026-03-09T15:17:33.965 INFO:teuthology.orchestra.run.vm03.stdout: Erasing : mailcap-2.1.49-5.el9.noarch 2/2 2026-03-09T15:17:33.968 INFO:teuthology.orchestra.run.vm04.stdout: Running scriptlet: ceph-radosgw-2:19.2.3-678.ge911bdeb.el9.x86_64 1/2 2026-03-09T15:17:33.970 INFO:teuthology.orchestra.run.vm06.stdout: Running scriptlet: ceph-radosgw-2:19.2.3-678.ge911bdeb.el9.x86_64 1/2 2026-03-09T15:17:33.970 INFO:teuthology.orchestra.run.vm06.stdout:Glob pattern passed to enable, but globs are not supported for this. 2026-03-09T15:17:33.970 INFO:teuthology.orchestra.run.vm06.stdout:Invalid unit name "ceph-radosgw@*.service" escaped as "ceph-radosgw@\x2a.service". 2026-03-09T15:17:33.970 INFO:teuthology.orchestra.run.vm06.stdout:Removed "/etc/systemd/system/multi-user.target.wants/ceph-radosgw.target". 2026-03-09T15:17:33.970 INFO:teuthology.orchestra.run.vm06.stdout:Removed "/etc/systemd/system/ceph.target.wants/ceph-radosgw.target". 2026-03-09T15:17:33.970 INFO:teuthology.orchestra.run.vm06.stdout: 2026-03-09T15:17:33.973 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : ceph-radosgw-2:19.2.3-678.ge911bdeb.el9.x86_64 1/2 2026-03-09T15:17:33.981 INFO:teuthology.orchestra.run.vm06.stdout: Running scriptlet: ceph-radosgw-2:19.2.3-678.ge911bdeb.el9.x86_64 1/2 2026-03-09T15:17:33.982 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : mailcap-2.1.49-5.el9.noarch 2/2 2026-03-09T15:17:33.996 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : mailcap-2.1.49-5.el9.noarch 2/2 2026-03-09T15:17:34.034 INFO:teuthology.orchestra.run.vm03.stdout: Running scriptlet: mailcap-2.1.49-5.el9.noarch 2/2 2026-03-09T15:17:34.034 INFO:teuthology.orchestra.run.vm03.stdout: Verifying : ceph-radosgw-2:19.2.3-678.ge911bdeb.el9.x86_64 1/2 2026-03-09T15:17:34.057 INFO:teuthology.orchestra.run.vm04.stdout: Running scriptlet: mailcap-2.1.49-5.el9.noarch 2/2 2026-03-09T15:17:34.057 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : ceph-radosgw-2:19.2.3-678.ge911bdeb.el9.x86_64 1/2 2026-03-09T15:17:34.074 INFO:teuthology.orchestra.run.vm06.stdout: Running scriptlet: mailcap-2.1.49-5.el9.noarch 2/2 2026-03-09T15:17:34.074 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : ceph-radosgw-2:19.2.3-678.ge911bdeb.el9.x86_64 1/2 2026-03-09T15:17:34.094 INFO:teuthology.orchestra.run.vm03.stdout: Verifying : mailcap-2.1.49-5.el9.noarch 2/2 2026-03-09T15:17:34.094 INFO:teuthology.orchestra.run.vm03.stdout: 2026-03-09T15:17:34.094 INFO:teuthology.orchestra.run.vm03.stdout:Removed: 2026-03-09T15:17:34.094 INFO:teuthology.orchestra.run.vm03.stdout: ceph-radosgw-2:19.2.3-678.ge911bdeb.el9.x86_64 mailcap-2.1.49-5.el9.noarch 2026-03-09T15:17:34.094 INFO:teuthology.orchestra.run.vm03.stdout: 2026-03-09T15:17:34.094 INFO:teuthology.orchestra.run.vm03.stdout:Complete! 2026-03-09T15:17:34.112 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : mailcap-2.1.49-5.el9.noarch 2/2 2026-03-09T15:17:34.112 INFO:teuthology.orchestra.run.vm04.stdout: 2026-03-09T15:17:34.112 INFO:teuthology.orchestra.run.vm04.stdout:Removed: 2026-03-09T15:17:34.112 INFO:teuthology.orchestra.run.vm04.stdout: ceph-radosgw-2:19.2.3-678.ge911bdeb.el9.x86_64 mailcap-2.1.49-5.el9.noarch 2026-03-09T15:17:34.112 INFO:teuthology.orchestra.run.vm04.stdout: 2026-03-09T15:17:34.112 INFO:teuthology.orchestra.run.vm04.stdout:Complete! 2026-03-09T15:17:34.126 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : mailcap-2.1.49-5.el9.noarch 2/2 2026-03-09T15:17:34.126 INFO:teuthology.orchestra.run.vm06.stdout: 2026-03-09T15:17:34.126 INFO:teuthology.orchestra.run.vm06.stdout:Removed: 2026-03-09T15:17:34.126 INFO:teuthology.orchestra.run.vm06.stdout: ceph-radosgw-2:19.2.3-678.ge911bdeb.el9.x86_64 mailcap-2.1.49-5.el9.noarch 2026-03-09T15:17:34.126 INFO:teuthology.orchestra.run.vm06.stdout: 2026-03-09T15:17:34.126 INFO:teuthology.orchestra.run.vm06.stdout:Complete! 2026-03-09T15:17:34.310 INFO:teuthology.orchestra.run.vm03.stdout:Dependencies resolved. 2026-03-09T15:17:34.311 INFO:teuthology.orchestra.run.vm04.stdout:Dependencies resolved. 2026-03-09T15:17:34.311 INFO:teuthology.orchestra.run.vm03.stdout:================================================================================ 2026-03-09T15:17:34.311 INFO:teuthology.orchestra.run.vm03.stdout: Package Arch Version Repository Size 2026-03-09T15:17:34.311 INFO:teuthology.orchestra.run.vm03.stdout:================================================================================ 2026-03-09T15:17:34.311 INFO:teuthology.orchestra.run.vm03.stdout:Removing: 2026-03-09T15:17:34.311 INFO:teuthology.orchestra.run.vm03.stdout: ceph-test x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 210 M 2026-03-09T15:17:34.311 INFO:teuthology.orchestra.run.vm03.stdout:Removing unused dependencies: 2026-03-09T15:17:34.311 INFO:teuthology.orchestra.run.vm03.stdout: libxslt x86_64 1.1.34-12.el9 @appstream 743 k 2026-03-09T15:17:34.311 INFO:teuthology.orchestra.run.vm03.stdout: socat x86_64 1.7.4.1-8.el9 @appstream 1.1 M 2026-03-09T15:17:34.311 INFO:teuthology.orchestra.run.vm03.stdout: xmlstarlet x86_64 1.6.1-20.el9 @appstream 195 k 2026-03-09T15:17:34.311 INFO:teuthology.orchestra.run.vm03.stdout: 2026-03-09T15:17:34.311 INFO:teuthology.orchestra.run.vm03.stdout:Transaction Summary 2026-03-09T15:17:34.311 INFO:teuthology.orchestra.run.vm03.stdout:================================================================================ 2026-03-09T15:17:34.311 INFO:teuthology.orchestra.run.vm03.stdout:Remove 4 Packages 2026-03-09T15:17:34.311 INFO:teuthology.orchestra.run.vm03.stdout: 2026-03-09T15:17:34.311 INFO:teuthology.orchestra.run.vm03.stdout:Freed space: 212 M 2026-03-09T15:17:34.312 INFO:teuthology.orchestra.run.vm03.stdout:Running transaction check 2026-03-09T15:17:34.312 INFO:teuthology.orchestra.run.vm04.stdout:================================================================================ 2026-03-09T15:17:34.312 INFO:teuthology.orchestra.run.vm04.stdout: Package Arch Version Repository Size 2026-03-09T15:17:34.312 INFO:teuthology.orchestra.run.vm04.stdout:================================================================================ 2026-03-09T15:17:34.312 INFO:teuthology.orchestra.run.vm04.stdout:Removing: 2026-03-09T15:17:34.312 INFO:teuthology.orchestra.run.vm04.stdout: ceph-test x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 210 M 2026-03-09T15:17:34.312 INFO:teuthology.orchestra.run.vm04.stdout:Removing unused dependencies: 2026-03-09T15:17:34.312 INFO:teuthology.orchestra.run.vm04.stdout: libxslt x86_64 1.1.34-12.el9 @appstream 743 k 2026-03-09T15:17:34.312 INFO:teuthology.orchestra.run.vm04.stdout: socat x86_64 1.7.4.1-8.el9 @appstream 1.1 M 2026-03-09T15:17:34.312 INFO:teuthology.orchestra.run.vm04.stdout: xmlstarlet x86_64 1.6.1-20.el9 @appstream 195 k 2026-03-09T15:17:34.312 INFO:teuthology.orchestra.run.vm04.stdout: 2026-03-09T15:17:34.312 INFO:teuthology.orchestra.run.vm04.stdout:Transaction Summary 2026-03-09T15:17:34.312 INFO:teuthology.orchestra.run.vm04.stdout:================================================================================ 2026-03-09T15:17:34.312 INFO:teuthology.orchestra.run.vm04.stdout:Remove 4 Packages 2026-03-09T15:17:34.312 INFO:teuthology.orchestra.run.vm04.stdout: 2026-03-09T15:17:34.312 INFO:teuthology.orchestra.run.vm04.stdout:Freed space: 212 M 2026-03-09T15:17:34.312 INFO:teuthology.orchestra.run.vm04.stdout:Running transaction check 2026-03-09T15:17:34.314 INFO:teuthology.orchestra.run.vm03.stdout:Transaction check succeeded. 2026-03-09T15:17:34.314 INFO:teuthology.orchestra.run.vm03.stdout:Running transaction test 2026-03-09T15:17:34.315 INFO:teuthology.orchestra.run.vm04.stdout:Transaction check succeeded. 2026-03-09T15:17:34.315 INFO:teuthology.orchestra.run.vm04.stdout:Running transaction test 2026-03-09T15:17:34.338 INFO:teuthology.orchestra.run.vm04.stdout:Transaction test succeeded. 2026-03-09T15:17:34.338 INFO:teuthology.orchestra.run.vm04.stdout:Running transaction 2026-03-09T15:17:34.338 INFO:teuthology.orchestra.run.vm03.stdout:Transaction test succeeded. 2026-03-09T15:17:34.339 INFO:teuthology.orchestra.run.vm03.stdout:Running transaction 2026-03-09T15:17:34.355 INFO:teuthology.orchestra.run.vm06.stdout:Dependencies resolved. 2026-03-09T15:17:34.356 INFO:teuthology.orchestra.run.vm06.stdout:================================================================================ 2026-03-09T15:17:34.356 INFO:teuthology.orchestra.run.vm06.stdout: Package Arch Version Repository Size 2026-03-09T15:17:34.356 INFO:teuthology.orchestra.run.vm06.stdout:================================================================================ 2026-03-09T15:17:34.356 INFO:teuthology.orchestra.run.vm06.stdout:Removing: 2026-03-09T15:17:34.356 INFO:teuthology.orchestra.run.vm06.stdout: ceph-test x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 210 M 2026-03-09T15:17:34.356 INFO:teuthology.orchestra.run.vm06.stdout:Removing unused dependencies: 2026-03-09T15:17:34.356 INFO:teuthology.orchestra.run.vm06.stdout: libxslt x86_64 1.1.34-12.el9 @appstream 743 k 2026-03-09T15:17:34.356 INFO:teuthology.orchestra.run.vm06.stdout: socat x86_64 1.7.4.1-8.el9 @appstream 1.1 M 2026-03-09T15:17:34.356 INFO:teuthology.orchestra.run.vm06.stdout: xmlstarlet x86_64 1.6.1-20.el9 @appstream 195 k 2026-03-09T15:17:34.356 INFO:teuthology.orchestra.run.vm06.stdout: 2026-03-09T15:17:34.356 INFO:teuthology.orchestra.run.vm06.stdout:Transaction Summary 2026-03-09T15:17:34.356 INFO:teuthology.orchestra.run.vm06.stdout:================================================================================ 2026-03-09T15:17:34.356 INFO:teuthology.orchestra.run.vm06.stdout:Remove 4 Packages 2026-03-09T15:17:34.356 INFO:teuthology.orchestra.run.vm06.stdout: 2026-03-09T15:17:34.356 INFO:teuthology.orchestra.run.vm06.stdout:Freed space: 212 M 2026-03-09T15:17:34.356 INFO:teuthology.orchestra.run.vm06.stdout:Running transaction check 2026-03-09T15:17:34.359 INFO:teuthology.orchestra.run.vm06.stdout:Transaction check succeeded. 2026-03-09T15:17:34.359 INFO:teuthology.orchestra.run.vm06.stdout:Running transaction test 2026-03-09T15:17:34.380 INFO:teuthology.orchestra.run.vm06.stdout:Transaction test succeeded. 2026-03-09T15:17:34.380 INFO:teuthology.orchestra.run.vm06.stdout:Running transaction 2026-03-09T15:17:34.399 INFO:teuthology.orchestra.run.vm04.stdout: Preparing : 1/1 2026-03-09T15:17:34.401 INFO:teuthology.orchestra.run.vm03.stdout: Preparing : 1/1 2026-03-09T15:17:34.405 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : ceph-test-2:19.2.3-678.ge911bdeb.el9.x86_64 1/4 2026-03-09T15:17:34.407 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : xmlstarlet-1.6.1-20.el9.x86_64 2/4 2026-03-09T15:17:34.407 INFO:teuthology.orchestra.run.vm03.stdout: Erasing : ceph-test-2:19.2.3-678.ge911bdeb.el9.x86_64 1/4 2026-03-09T15:17:34.409 INFO:teuthology.orchestra.run.vm03.stdout: Erasing : xmlstarlet-1.6.1-20.el9.x86_64 2/4 2026-03-09T15:17:34.410 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : libxslt-1.1.34-12.el9.x86_64 3/4 2026-03-09T15:17:34.412 INFO:teuthology.orchestra.run.vm03.stdout: Erasing : libxslt-1.1.34-12.el9.x86_64 3/4 2026-03-09T15:17:34.425 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : socat-1.7.4.1-8.el9.x86_64 4/4 2026-03-09T15:17:34.427 INFO:teuthology.orchestra.run.vm03.stdout: Erasing : socat-1.7.4.1-8.el9.x86_64 4/4 2026-03-09T15:17:34.443 INFO:teuthology.orchestra.run.vm06.stdout: Preparing : 1/1 2026-03-09T15:17:34.450 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : ceph-test-2:19.2.3-678.ge911bdeb.el9.x86_64 1/4 2026-03-09T15:17:34.452 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : xmlstarlet-1.6.1-20.el9.x86_64 2/4 2026-03-09T15:17:34.456 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : libxslt-1.1.34-12.el9.x86_64 3/4 2026-03-09T15:17:34.474 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : socat-1.7.4.1-8.el9.x86_64 4/4 2026-03-09T15:17:34.493 INFO:teuthology.orchestra.run.vm03.stdout: Running scriptlet: socat-1.7.4.1-8.el9.x86_64 4/4 2026-03-09T15:17:34.493 INFO:teuthology.orchestra.run.vm03.stdout: Verifying : ceph-test-2:19.2.3-678.ge911bdeb.el9.x86_64 1/4 2026-03-09T15:17:34.493 INFO:teuthology.orchestra.run.vm03.stdout: Verifying : libxslt-1.1.34-12.el9.x86_64 2/4 2026-03-09T15:17:34.493 INFO:teuthology.orchestra.run.vm03.stdout: Verifying : socat-1.7.4.1-8.el9.x86_64 3/4 2026-03-09T15:17:34.494 INFO:teuthology.orchestra.run.vm04.stdout: Running scriptlet: socat-1.7.4.1-8.el9.x86_64 4/4 2026-03-09T15:17:34.494 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : ceph-test-2:19.2.3-678.ge911bdeb.el9.x86_64 1/4 2026-03-09T15:17:34.494 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : libxslt-1.1.34-12.el9.x86_64 2/4 2026-03-09T15:17:34.494 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : socat-1.7.4.1-8.el9.x86_64 3/4 2026-03-09T15:17:34.542 INFO:teuthology.orchestra.run.vm03.stdout: Verifying : xmlstarlet-1.6.1-20.el9.x86_64 4/4 2026-03-09T15:17:34.543 INFO:teuthology.orchestra.run.vm03.stdout: 2026-03-09T15:17:34.543 INFO:teuthology.orchestra.run.vm03.stdout:Removed: 2026-03-09T15:17:34.543 INFO:teuthology.orchestra.run.vm03.stdout: ceph-test-2:19.2.3-678.ge911bdeb.el9.x86_64 libxslt-1.1.34-12.el9.x86_64 2026-03-09T15:17:34.543 INFO:teuthology.orchestra.run.vm03.stdout: socat-1.7.4.1-8.el9.x86_64 xmlstarlet-1.6.1-20.el9.x86_64 2026-03-09T15:17:34.543 INFO:teuthology.orchestra.run.vm03.stdout: 2026-03-09T15:17:34.543 INFO:teuthology.orchestra.run.vm03.stdout:Complete! 2026-03-09T15:17:34.548 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : xmlstarlet-1.6.1-20.el9.x86_64 4/4 2026-03-09T15:17:34.548 INFO:teuthology.orchestra.run.vm04.stdout: 2026-03-09T15:17:34.548 INFO:teuthology.orchestra.run.vm04.stdout:Removed: 2026-03-09T15:17:34.548 INFO:teuthology.orchestra.run.vm04.stdout: ceph-test-2:19.2.3-678.ge911bdeb.el9.x86_64 libxslt-1.1.34-12.el9.x86_64 2026-03-09T15:17:34.548 INFO:teuthology.orchestra.run.vm04.stdout: socat-1.7.4.1-8.el9.x86_64 xmlstarlet-1.6.1-20.el9.x86_64 2026-03-09T15:17:34.548 INFO:teuthology.orchestra.run.vm04.stdout: 2026-03-09T15:17:34.548 INFO:teuthology.orchestra.run.vm04.stdout:Complete! 2026-03-09T15:17:34.550 INFO:teuthology.orchestra.run.vm06.stdout: Running scriptlet: socat-1.7.4.1-8.el9.x86_64 4/4 2026-03-09T15:17:34.550 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : ceph-test-2:19.2.3-678.ge911bdeb.el9.x86_64 1/4 2026-03-09T15:17:34.550 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : libxslt-1.1.34-12.el9.x86_64 2/4 2026-03-09T15:17:34.550 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : socat-1.7.4.1-8.el9.x86_64 3/4 2026-03-09T15:17:34.600 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : xmlstarlet-1.6.1-20.el9.x86_64 4/4 2026-03-09T15:17:34.600 INFO:teuthology.orchestra.run.vm06.stdout: 2026-03-09T15:17:34.600 INFO:teuthology.orchestra.run.vm06.stdout:Removed: 2026-03-09T15:17:34.600 INFO:teuthology.orchestra.run.vm06.stdout: ceph-test-2:19.2.3-678.ge911bdeb.el9.x86_64 libxslt-1.1.34-12.el9.x86_64 2026-03-09T15:17:34.600 INFO:teuthology.orchestra.run.vm06.stdout: socat-1.7.4.1-8.el9.x86_64 xmlstarlet-1.6.1-20.el9.x86_64 2026-03-09T15:17:34.600 INFO:teuthology.orchestra.run.vm06.stdout: 2026-03-09T15:17:34.600 INFO:teuthology.orchestra.run.vm06.stdout:Complete! 2026-03-09T15:17:34.760 INFO:teuthology.orchestra.run.vm04.stdout:Dependencies resolved. 2026-03-09T15:17:34.760 INFO:teuthology.orchestra.run.vm04.stdout:================================================================================ 2026-03-09T15:17:34.760 INFO:teuthology.orchestra.run.vm04.stdout: Package Arch Version Repository Size 2026-03-09T15:17:34.760 INFO:teuthology.orchestra.run.vm04.stdout:================================================================================ 2026-03-09T15:17:34.760 INFO:teuthology.orchestra.run.vm04.stdout:Removing: 2026-03-09T15:17:34.760 INFO:teuthology.orchestra.run.vm04.stdout: ceph x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 0 2026-03-09T15:17:34.760 INFO:teuthology.orchestra.run.vm04.stdout:Removing unused dependencies: 2026-03-09T15:17:34.760 INFO:teuthology.orchestra.run.vm04.stdout: ceph-mds x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 7.5 M 2026-03-09T15:17:34.760 INFO:teuthology.orchestra.run.vm04.stdout: ceph-mon x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 18 M 2026-03-09T15:17:34.760 INFO:teuthology.orchestra.run.vm04.stdout: lua x86_64 5.4.4-4.el9 @appstream 593 k 2026-03-09T15:17:34.760 INFO:teuthology.orchestra.run.vm04.stdout: lua-devel x86_64 5.4.4-4.el9 @crb 49 k 2026-03-09T15:17:34.760 INFO:teuthology.orchestra.run.vm04.stdout: luarocks noarch 3.9.2-5.el9 @epel 692 k 2026-03-09T15:17:34.760 INFO:teuthology.orchestra.run.vm04.stdout: unzip x86_64 6.0-59.el9 @baseos 389 k 2026-03-09T15:17:34.760 INFO:teuthology.orchestra.run.vm04.stdout: zip x86_64 3.0-35.el9 @baseos 724 k 2026-03-09T15:17:34.760 INFO:teuthology.orchestra.run.vm04.stdout: 2026-03-09T15:17:34.760 INFO:teuthology.orchestra.run.vm04.stdout:Transaction Summary 2026-03-09T15:17:34.761 INFO:teuthology.orchestra.run.vm04.stdout:================================================================================ 2026-03-09T15:17:34.761 INFO:teuthology.orchestra.run.vm04.stdout:Remove 8 Packages 2026-03-09T15:17:34.761 INFO:teuthology.orchestra.run.vm04.stdout: 2026-03-09T15:17:34.761 INFO:teuthology.orchestra.run.vm04.stdout:Freed space: 28 M 2026-03-09T15:17:34.761 INFO:teuthology.orchestra.run.vm04.stdout:Running transaction check 2026-03-09T15:17:34.763 INFO:teuthology.orchestra.run.vm04.stdout:Transaction check succeeded. 2026-03-09T15:17:34.763 INFO:teuthology.orchestra.run.vm04.stdout:Running transaction test 2026-03-09T15:17:34.763 INFO:teuthology.orchestra.run.vm03.stdout:Dependencies resolved. 2026-03-09T15:17:34.764 INFO:teuthology.orchestra.run.vm03.stdout:================================================================================ 2026-03-09T15:17:34.764 INFO:teuthology.orchestra.run.vm03.stdout: Package Arch Version Repository Size 2026-03-09T15:17:34.764 INFO:teuthology.orchestra.run.vm03.stdout:================================================================================ 2026-03-09T15:17:34.764 INFO:teuthology.orchestra.run.vm03.stdout:Removing: 2026-03-09T15:17:34.764 INFO:teuthology.orchestra.run.vm03.stdout: ceph x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 0 2026-03-09T15:17:34.764 INFO:teuthology.orchestra.run.vm03.stdout:Removing unused dependencies: 2026-03-09T15:17:34.764 INFO:teuthology.orchestra.run.vm03.stdout: ceph-mds x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 7.5 M 2026-03-09T15:17:34.764 INFO:teuthology.orchestra.run.vm03.stdout: ceph-mon x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 18 M 2026-03-09T15:17:34.764 INFO:teuthology.orchestra.run.vm03.stdout: lua x86_64 5.4.4-4.el9 @appstream 593 k 2026-03-09T15:17:34.764 INFO:teuthology.orchestra.run.vm03.stdout: lua-devel x86_64 5.4.4-4.el9 @crb 49 k 2026-03-09T15:17:34.764 INFO:teuthology.orchestra.run.vm03.stdout: luarocks noarch 3.9.2-5.el9 @epel 692 k 2026-03-09T15:17:34.764 INFO:teuthology.orchestra.run.vm03.stdout: unzip x86_64 6.0-59.el9 @baseos 389 k 2026-03-09T15:17:34.764 INFO:teuthology.orchestra.run.vm03.stdout: zip x86_64 3.0-35.el9 @baseos 724 k 2026-03-09T15:17:34.764 INFO:teuthology.orchestra.run.vm03.stdout: 2026-03-09T15:17:34.764 INFO:teuthology.orchestra.run.vm03.stdout:Transaction Summary 2026-03-09T15:17:34.765 INFO:teuthology.orchestra.run.vm03.stdout:================================================================================ 2026-03-09T15:17:34.765 INFO:teuthology.orchestra.run.vm03.stdout:Remove 8 Packages 2026-03-09T15:17:34.765 INFO:teuthology.orchestra.run.vm03.stdout: 2026-03-09T15:17:34.765 INFO:teuthology.orchestra.run.vm03.stdout:Freed space: 28 M 2026-03-09T15:17:34.765 INFO:teuthology.orchestra.run.vm03.stdout:Running transaction check 2026-03-09T15:17:34.767 INFO:teuthology.orchestra.run.vm03.stdout:Transaction check succeeded. 2026-03-09T15:17:34.767 INFO:teuthology.orchestra.run.vm03.stdout:Running transaction test 2026-03-09T15:17:34.787 INFO:teuthology.orchestra.run.vm04.stdout:Transaction test succeeded. 2026-03-09T15:17:34.787 INFO:teuthology.orchestra.run.vm04.stdout:Running transaction 2026-03-09T15:17:34.792 INFO:teuthology.orchestra.run.vm03.stdout:Transaction test succeeded. 2026-03-09T15:17:34.792 INFO:teuthology.orchestra.run.vm03.stdout:Running transaction 2026-03-09T15:17:34.803 INFO:teuthology.orchestra.run.vm06.stdout:Dependencies resolved. 2026-03-09T15:17:34.803 INFO:teuthology.orchestra.run.vm06.stdout:================================================================================ 2026-03-09T15:17:34.803 INFO:teuthology.orchestra.run.vm06.stdout: Package Arch Version Repository Size 2026-03-09T15:17:34.803 INFO:teuthology.orchestra.run.vm06.stdout:================================================================================ 2026-03-09T15:17:34.803 INFO:teuthology.orchestra.run.vm06.stdout:Removing: 2026-03-09T15:17:34.803 INFO:teuthology.orchestra.run.vm06.stdout: ceph x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 0 2026-03-09T15:17:34.804 INFO:teuthology.orchestra.run.vm06.stdout:Removing unused dependencies: 2026-03-09T15:17:34.804 INFO:teuthology.orchestra.run.vm06.stdout: ceph-mds x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 7.5 M 2026-03-09T15:17:34.804 INFO:teuthology.orchestra.run.vm06.stdout: ceph-mon x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 18 M 2026-03-09T15:17:34.804 INFO:teuthology.orchestra.run.vm06.stdout: lua x86_64 5.4.4-4.el9 @appstream 593 k 2026-03-09T15:17:34.804 INFO:teuthology.orchestra.run.vm06.stdout: lua-devel x86_64 5.4.4-4.el9 @crb 49 k 2026-03-09T15:17:34.804 INFO:teuthology.orchestra.run.vm06.stdout: luarocks noarch 3.9.2-5.el9 @epel 692 k 2026-03-09T15:17:34.804 INFO:teuthology.orchestra.run.vm06.stdout: unzip x86_64 6.0-59.el9 @baseos 389 k 2026-03-09T15:17:34.804 INFO:teuthology.orchestra.run.vm06.stdout: zip x86_64 3.0-35.el9 @baseos 724 k 2026-03-09T15:17:34.804 INFO:teuthology.orchestra.run.vm06.stdout: 2026-03-09T15:17:34.804 INFO:teuthology.orchestra.run.vm06.stdout:Transaction Summary 2026-03-09T15:17:34.804 INFO:teuthology.orchestra.run.vm06.stdout:================================================================================ 2026-03-09T15:17:34.804 INFO:teuthology.orchestra.run.vm06.stdout:Remove 8 Packages 2026-03-09T15:17:34.804 INFO:teuthology.orchestra.run.vm06.stdout: 2026-03-09T15:17:34.804 INFO:teuthology.orchestra.run.vm06.stdout:Freed space: 28 M 2026-03-09T15:17:34.804 INFO:teuthology.orchestra.run.vm06.stdout:Running transaction check 2026-03-09T15:17:34.806 INFO:teuthology.orchestra.run.vm06.stdout:Transaction check succeeded. 2026-03-09T15:17:34.806 INFO:teuthology.orchestra.run.vm06.stdout:Running transaction test 2026-03-09T15:17:34.829 INFO:teuthology.orchestra.run.vm06.stdout:Transaction test succeeded. 2026-03-09T15:17:34.829 INFO:teuthology.orchestra.run.vm06.stdout:Running transaction 2026-03-09T15:17:34.858 INFO:teuthology.orchestra.run.vm04.stdout: Preparing : 1/1 2026-03-09T15:17:34.861 INFO:teuthology.orchestra.run.vm03.stdout: Preparing : 1/1 2026-03-09T15:17:34.863 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : ceph-2:19.2.3-678.ge911bdeb.el9.x86_64 1/8 2026-03-09T15:17:34.866 INFO:teuthology.orchestra.run.vm03.stdout: Erasing : ceph-2:19.2.3-678.ge911bdeb.el9.x86_64 1/8 2026-03-09T15:17:34.866 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : luarocks-3.9.2-5.el9.noarch 2/8 2026-03-09T15:17:34.868 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : lua-devel-5.4.4-4.el9.x86_64 3/8 2026-03-09T15:17:34.870 INFO:teuthology.orchestra.run.vm06.stdout: Preparing : 1/1 2026-03-09T15:17:34.870 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : zip-3.0-35.el9.x86_64 4/8 2026-03-09T15:17:34.870 INFO:teuthology.orchestra.run.vm03.stdout: Erasing : luarocks-3.9.2-5.el9.noarch 2/8 2026-03-09T15:17:34.872 INFO:teuthology.orchestra.run.vm03.stdout: Erasing : lua-devel-5.4.4-4.el9.x86_64 3/8 2026-03-09T15:17:34.872 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : unzip-6.0-59.el9.x86_64 5/8 2026-03-09T15:17:34.874 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : lua-5.4.4-4.el9.x86_64 6/8 2026-03-09T15:17:34.874 INFO:teuthology.orchestra.run.vm03.stdout: Erasing : zip-3.0-35.el9.x86_64 4/8 2026-03-09T15:17:34.875 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : ceph-2:19.2.3-678.ge911bdeb.el9.x86_64 1/8 2026-03-09T15:17:34.877 INFO:teuthology.orchestra.run.vm03.stdout: Erasing : unzip-6.0-59.el9.x86_64 5/8 2026-03-09T15:17:34.879 INFO:teuthology.orchestra.run.vm03.stdout: Erasing : lua-5.4.4-4.el9.x86_64 6/8 2026-03-09T15:17:34.879 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : luarocks-3.9.2-5.el9.noarch 2/8 2026-03-09T15:17:34.881 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : lua-devel-5.4.4-4.el9.x86_64 3/8 2026-03-09T15:17:34.884 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : zip-3.0-35.el9.x86_64 4/8 2026-03-09T15:17:34.887 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : unzip-6.0-59.el9.x86_64 5/8 2026-03-09T15:17:34.888 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : lua-5.4.4-4.el9.x86_64 6/8 2026-03-09T15:17:34.891 INFO:teuthology.orchestra.run.vm04.stdout: Running scriptlet: ceph-mds-2:19.2.3-678.ge911bdeb.el9.x86_64 7/8 2026-03-09T15:17:34.891 INFO:teuthology.orchestra.run.vm04.stdout:Glob pattern passed to enable, but globs are not supported for this. 2026-03-09T15:17:34.891 INFO:teuthology.orchestra.run.vm04.stdout:Invalid unit name "ceph-mds@*.service" escaped as "ceph-mds@\x2a.service". 2026-03-09T15:17:34.891 INFO:teuthology.orchestra.run.vm04.stdout:Removed "/etc/systemd/system/multi-user.target.wants/ceph-mds.target". 2026-03-09T15:17:34.891 INFO:teuthology.orchestra.run.vm04.stdout:Removed "/etc/systemd/system/ceph.target.wants/ceph-mds.target". 2026-03-09T15:17:34.891 INFO:teuthology.orchestra.run.vm04.stdout: 2026-03-09T15:17:34.892 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : ceph-mds-2:19.2.3-678.ge911bdeb.el9.x86_64 7/8 2026-03-09T15:17:34.896 INFO:teuthology.orchestra.run.vm03.stdout: Running scriptlet: ceph-mds-2:19.2.3-678.ge911bdeb.el9.x86_64 7/8 2026-03-09T15:17:34.896 INFO:teuthology.orchestra.run.vm03.stdout:Glob pattern passed to enable, but globs are not supported for this. 2026-03-09T15:17:34.896 INFO:teuthology.orchestra.run.vm03.stdout:Invalid unit name "ceph-mds@*.service" escaped as "ceph-mds@\x2a.service". 2026-03-09T15:17:34.897 INFO:teuthology.orchestra.run.vm03.stdout:Removed "/etc/systemd/system/multi-user.target.wants/ceph-mds.target". 2026-03-09T15:17:34.897 INFO:teuthology.orchestra.run.vm03.stdout:Removed "/etc/systemd/system/ceph.target.wants/ceph-mds.target". 2026-03-09T15:17:34.897 INFO:teuthology.orchestra.run.vm03.stdout: 2026-03-09T15:17:34.897 INFO:teuthology.orchestra.run.vm03.stdout: Erasing : ceph-mds-2:19.2.3-678.ge911bdeb.el9.x86_64 7/8 2026-03-09T15:17:34.899 INFO:teuthology.orchestra.run.vm04.stdout: Running scriptlet: ceph-mds-2:19.2.3-678.ge911bdeb.el9.x86_64 7/8 2026-03-09T15:17:34.903 INFO:teuthology.orchestra.run.vm03.stdout: Running scriptlet: ceph-mds-2:19.2.3-678.ge911bdeb.el9.x86_64 7/8 2026-03-09T15:17:34.906 INFO:teuthology.orchestra.run.vm06.stdout: Running scriptlet: ceph-mds-2:19.2.3-678.ge911bdeb.el9.x86_64 7/8 2026-03-09T15:17:34.906 INFO:teuthology.orchestra.run.vm06.stdout:Glob pattern passed to enable, but globs are not supported for this. 2026-03-09T15:17:34.906 INFO:teuthology.orchestra.run.vm06.stdout:Invalid unit name "ceph-mds@*.service" escaped as "ceph-mds@\x2a.service". 2026-03-09T15:17:34.906 INFO:teuthology.orchestra.run.vm06.stdout:Removed "/etc/systemd/system/multi-user.target.wants/ceph-mds.target". 2026-03-09T15:17:34.906 INFO:teuthology.orchestra.run.vm06.stdout:Removed "/etc/systemd/system/ceph.target.wants/ceph-mds.target". 2026-03-09T15:17:34.906 INFO:teuthology.orchestra.run.vm06.stdout: 2026-03-09T15:17:34.906 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : ceph-mds-2:19.2.3-678.ge911bdeb.el9.x86_64 7/8 2026-03-09T15:17:34.913 INFO:teuthology.orchestra.run.vm06.stdout: Running scriptlet: ceph-mds-2:19.2.3-678.ge911bdeb.el9.x86_64 7/8 2026-03-09T15:17:34.917 INFO:teuthology.orchestra.run.vm04.stdout: Running scriptlet: ceph-mon-2:19.2.3-678.ge911bdeb.el9.x86_64 8/8 2026-03-09T15:17:34.917 INFO:teuthology.orchestra.run.vm04.stdout:Glob pattern passed to enable, but globs are not supported for this. 2026-03-09T15:17:34.917 INFO:teuthology.orchestra.run.vm04.stdout:Invalid unit name "ceph-mon@*.service" escaped as "ceph-mon@\x2a.service". 2026-03-09T15:17:34.917 INFO:teuthology.orchestra.run.vm04.stdout:Removed "/etc/systemd/system/multi-user.target.wants/ceph-mon.target". 2026-03-09T15:17:34.917 INFO:teuthology.orchestra.run.vm04.stdout:Removed "/etc/systemd/system/ceph.target.wants/ceph-mon.target". 2026-03-09T15:17:34.917 INFO:teuthology.orchestra.run.vm04.stdout: 2026-03-09T15:17:34.919 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : ceph-mon-2:19.2.3-678.ge911bdeb.el9.x86_64 8/8 2026-03-09T15:17:34.921 INFO:teuthology.orchestra.run.vm03.stdout: Running scriptlet: ceph-mon-2:19.2.3-678.ge911bdeb.el9.x86_64 8/8 2026-03-09T15:17:34.921 INFO:teuthology.orchestra.run.vm03.stdout:Glob pattern passed to enable, but globs are not supported for this. 2026-03-09T15:17:34.921 INFO:teuthology.orchestra.run.vm03.stdout:Invalid unit name "ceph-mon@*.service" escaped as "ceph-mon@\x2a.service". 2026-03-09T15:17:34.921 INFO:teuthology.orchestra.run.vm03.stdout:Removed "/etc/systemd/system/multi-user.target.wants/ceph-mon.target". 2026-03-09T15:17:34.921 INFO:teuthology.orchestra.run.vm03.stdout:Removed "/etc/systemd/system/ceph.target.wants/ceph-mon.target". 2026-03-09T15:17:34.921 INFO:teuthology.orchestra.run.vm03.stdout: 2026-03-09T15:17:34.922 INFO:teuthology.orchestra.run.vm03.stdout: Erasing : ceph-mon-2:19.2.3-678.ge911bdeb.el9.x86_64 8/8 2026-03-09T15:17:34.930 INFO:teuthology.orchestra.run.vm06.stdout: Running scriptlet: ceph-mon-2:19.2.3-678.ge911bdeb.el9.x86_64 8/8 2026-03-09T15:17:34.930 INFO:teuthology.orchestra.run.vm06.stdout:Glob pattern passed to enable, but globs are not supported for this. 2026-03-09T15:17:34.930 INFO:teuthology.orchestra.run.vm06.stdout:Invalid unit name "ceph-mon@*.service" escaped as "ceph-mon@\x2a.service". 2026-03-09T15:17:34.930 INFO:teuthology.orchestra.run.vm06.stdout:Removed "/etc/systemd/system/multi-user.target.wants/ceph-mon.target". 2026-03-09T15:17:34.930 INFO:teuthology.orchestra.run.vm06.stdout:Removed "/etc/systemd/system/ceph.target.wants/ceph-mon.target". 2026-03-09T15:17:34.930 INFO:teuthology.orchestra.run.vm06.stdout: 2026-03-09T15:17:34.932 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : ceph-mon-2:19.2.3-678.ge911bdeb.el9.x86_64 8/8 2026-03-09T15:17:35.004 INFO:teuthology.orchestra.run.vm04.stdout: Running scriptlet: ceph-mon-2:19.2.3-678.ge911bdeb.el9.x86_64 8/8 2026-03-09T15:17:35.004 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : ceph-2:19.2.3-678.ge911bdeb.el9.x86_64 1/8 2026-03-09T15:17:35.004 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : ceph-mds-2:19.2.3-678.ge911bdeb.el9.x86_64 2/8 2026-03-09T15:17:35.004 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : ceph-mon-2:19.2.3-678.ge911bdeb.el9.x86_64 3/8 2026-03-09T15:17:35.004 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : lua-5.4.4-4.el9.x86_64 4/8 2026-03-09T15:17:35.004 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : lua-devel-5.4.4-4.el9.x86_64 5/8 2026-03-09T15:17:35.004 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : luarocks-3.9.2-5.el9.noarch 6/8 2026-03-09T15:17:35.004 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : unzip-6.0-59.el9.x86_64 7/8 2026-03-09T15:17:35.009 INFO:teuthology.orchestra.run.vm03.stdout: Running scriptlet: ceph-mon-2:19.2.3-678.ge911bdeb.el9.x86_64 8/8 2026-03-09T15:17:35.009 INFO:teuthology.orchestra.run.vm03.stdout: Verifying : ceph-2:19.2.3-678.ge911bdeb.el9.x86_64 1/8 2026-03-09T15:17:35.009 INFO:teuthology.orchestra.run.vm03.stdout: Verifying : ceph-mds-2:19.2.3-678.ge911bdeb.el9.x86_64 2/8 2026-03-09T15:17:35.009 INFO:teuthology.orchestra.run.vm03.stdout: Verifying : ceph-mon-2:19.2.3-678.ge911bdeb.el9.x86_64 3/8 2026-03-09T15:17:35.009 INFO:teuthology.orchestra.run.vm03.stdout: Verifying : lua-5.4.4-4.el9.x86_64 4/8 2026-03-09T15:17:35.009 INFO:teuthology.orchestra.run.vm03.stdout: Verifying : lua-devel-5.4.4-4.el9.x86_64 5/8 2026-03-09T15:17:35.009 INFO:teuthology.orchestra.run.vm03.stdout: Verifying : luarocks-3.9.2-5.el9.noarch 6/8 2026-03-09T15:17:35.009 INFO:teuthology.orchestra.run.vm03.stdout: Verifying : unzip-6.0-59.el9.x86_64 7/8 2026-03-09T15:17:35.020 INFO:teuthology.orchestra.run.vm06.stdout: Running scriptlet: ceph-mon-2:19.2.3-678.ge911bdeb.el9.x86_64 8/8 2026-03-09T15:17:35.020 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : ceph-2:19.2.3-678.ge911bdeb.el9.x86_64 1/8 2026-03-09T15:17:35.020 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : ceph-mds-2:19.2.3-678.ge911bdeb.el9.x86_64 2/8 2026-03-09T15:17:35.020 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : ceph-mon-2:19.2.3-678.ge911bdeb.el9.x86_64 3/8 2026-03-09T15:17:35.020 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : lua-5.4.4-4.el9.x86_64 4/8 2026-03-09T15:17:35.020 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : lua-devel-5.4.4-4.el9.x86_64 5/8 2026-03-09T15:17:35.020 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : luarocks-3.9.2-5.el9.noarch 6/8 2026-03-09T15:17:35.020 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : unzip-6.0-59.el9.x86_64 7/8 2026-03-09T15:17:35.063 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : zip-3.0-35.el9.x86_64 8/8 2026-03-09T15:17:35.063 INFO:teuthology.orchestra.run.vm04.stdout: 2026-03-09T15:17:35.063 INFO:teuthology.orchestra.run.vm04.stdout:Removed: 2026-03-09T15:17:35.063 INFO:teuthology.orchestra.run.vm04.stdout: ceph-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T15:17:35.063 INFO:teuthology.orchestra.run.vm04.stdout: ceph-mds-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T15:17:35.063 INFO:teuthology.orchestra.run.vm04.stdout: ceph-mon-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T15:17:35.063 INFO:teuthology.orchestra.run.vm04.stdout: lua-5.4.4-4.el9.x86_64 2026-03-09T15:17:35.063 INFO:teuthology.orchestra.run.vm04.stdout: lua-devel-5.4.4-4.el9.x86_64 2026-03-09T15:17:35.063 INFO:teuthology.orchestra.run.vm04.stdout: luarocks-3.9.2-5.el9.noarch 2026-03-09T15:17:35.063 INFO:teuthology.orchestra.run.vm04.stdout: unzip-6.0-59.el9.x86_64 2026-03-09T15:17:35.063 INFO:teuthology.orchestra.run.vm04.stdout: zip-3.0-35.el9.x86_64 2026-03-09T15:17:35.063 INFO:teuthology.orchestra.run.vm04.stdout: 2026-03-09T15:17:35.063 INFO:teuthology.orchestra.run.vm04.stdout:Complete! 2026-03-09T15:17:35.071 INFO:teuthology.orchestra.run.vm03.stdout: Verifying : zip-3.0-35.el9.x86_64 8/8 2026-03-09T15:17:35.072 INFO:teuthology.orchestra.run.vm03.stdout: 2026-03-09T15:17:35.072 INFO:teuthology.orchestra.run.vm03.stdout:Removed: 2026-03-09T15:17:35.072 INFO:teuthology.orchestra.run.vm03.stdout: ceph-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T15:17:35.072 INFO:teuthology.orchestra.run.vm03.stdout: ceph-mds-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T15:17:35.072 INFO:teuthology.orchestra.run.vm03.stdout: ceph-mon-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T15:17:35.072 INFO:teuthology.orchestra.run.vm03.stdout: lua-5.4.4-4.el9.x86_64 2026-03-09T15:17:35.072 INFO:teuthology.orchestra.run.vm03.stdout: lua-devel-5.4.4-4.el9.x86_64 2026-03-09T15:17:35.072 INFO:teuthology.orchestra.run.vm03.stdout: luarocks-3.9.2-5.el9.noarch 2026-03-09T15:17:35.072 INFO:teuthology.orchestra.run.vm03.stdout: unzip-6.0-59.el9.x86_64 2026-03-09T15:17:35.072 INFO:teuthology.orchestra.run.vm03.stdout: zip-3.0-35.el9.x86_64 2026-03-09T15:17:35.072 INFO:teuthology.orchestra.run.vm03.stdout: 2026-03-09T15:17:35.072 INFO:teuthology.orchestra.run.vm03.stdout:Complete! 2026-03-09T15:17:35.075 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : zip-3.0-35.el9.x86_64 8/8 2026-03-09T15:17:35.075 INFO:teuthology.orchestra.run.vm06.stdout: 2026-03-09T15:17:35.075 INFO:teuthology.orchestra.run.vm06.stdout:Removed: 2026-03-09T15:17:35.075 INFO:teuthology.orchestra.run.vm06.stdout: ceph-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T15:17:35.075 INFO:teuthology.orchestra.run.vm06.stdout: ceph-mds-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T15:17:35.076 INFO:teuthology.orchestra.run.vm06.stdout: ceph-mon-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T15:17:35.076 INFO:teuthology.orchestra.run.vm06.stdout: lua-5.4.4-4.el9.x86_64 2026-03-09T15:17:35.076 INFO:teuthology.orchestra.run.vm06.stdout: lua-devel-5.4.4-4.el9.x86_64 2026-03-09T15:17:35.076 INFO:teuthology.orchestra.run.vm06.stdout: luarocks-3.9.2-5.el9.noarch 2026-03-09T15:17:35.076 INFO:teuthology.orchestra.run.vm06.stdout: unzip-6.0-59.el9.x86_64 2026-03-09T15:17:35.076 INFO:teuthology.orchestra.run.vm06.stdout: zip-3.0-35.el9.x86_64 2026-03-09T15:17:35.076 INFO:teuthology.orchestra.run.vm06.stdout: 2026-03-09T15:17:35.076 INFO:teuthology.orchestra.run.vm06.stdout:Complete! 2026-03-09T15:17:35.279 INFO:teuthology.orchestra.run.vm04.stdout:Dependencies resolved. 2026-03-09T15:17:35.280 INFO:teuthology.orchestra.run.vm06.stdout:Dependencies resolved. 2026-03-09T15:17:35.285 INFO:teuthology.orchestra.run.vm04.stdout:=========================================================================================== 2026-03-09T15:17:35.285 INFO:teuthology.orchestra.run.vm04.stdout: Package Arch Version Repository Size 2026-03-09T15:17:35.285 INFO:teuthology.orchestra.run.vm04.stdout:=========================================================================================== 2026-03-09T15:17:35.285 INFO:teuthology.orchestra.run.vm04.stdout:Removing: 2026-03-09T15:17:35.285 INFO:teuthology.orchestra.run.vm04.stdout: ceph-base x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 23 M 2026-03-09T15:17:35.285 INFO:teuthology.orchestra.run.vm04.stdout:Removing dependent packages: 2026-03-09T15:17:35.285 INFO:teuthology.orchestra.run.vm04.stdout: ceph-immutable-object-cache x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 431 k 2026-03-09T15:17:35.285 INFO:teuthology.orchestra.run.vm04.stdout: ceph-mgr x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 3.4 M 2026-03-09T15:17:35.285 INFO:teuthology.orchestra.run.vm04.stdout: ceph-mgr-cephadm noarch 2:19.2.3-678.ge911bdeb.el9 @ceph-noarch 806 k 2026-03-09T15:17:35.285 INFO:teuthology.orchestra.run.vm04.stdout: ceph-mgr-dashboard noarch 2:19.2.3-678.ge911bdeb.el9 @ceph-noarch 88 M 2026-03-09T15:17:35.285 INFO:teuthology.orchestra.run.vm04.stdout: ceph-mgr-diskprediction-local noarch 2:19.2.3-678.ge911bdeb.el9 @ceph-noarch 66 M 2026-03-09T15:17:35.285 INFO:teuthology.orchestra.run.vm04.stdout: ceph-mgr-rook noarch 2:19.2.3-678.ge911bdeb.el9 @ceph-noarch 563 k 2026-03-09T15:17:35.285 INFO:teuthology.orchestra.run.vm04.stdout: ceph-osd x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 59 M 2026-03-09T15:17:35.285 INFO:teuthology.orchestra.run.vm04.stdout: ceph-volume noarch 2:19.2.3-678.ge911bdeb.el9 @ceph-noarch 1.4 M 2026-03-09T15:17:35.285 INFO:teuthology.orchestra.run.vm04.stdout: rbd-mirror x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 13 M 2026-03-09T15:17:35.285 INFO:teuthology.orchestra.run.vm04.stdout:Removing unused dependencies: 2026-03-09T15:17:35.285 INFO:teuthology.orchestra.run.vm04.stdout: abseil-cpp x86_64 20211102.0-4.el9 @epel 1.9 M 2026-03-09T15:17:35.285 INFO:teuthology.orchestra.run.vm04.stdout: ceph-common x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 85 M 2026-03-09T15:17:35.285 INFO:teuthology.orchestra.run.vm04.stdout: ceph-grafana-dashboards noarch 2:19.2.3-678.ge911bdeb.el9 @ceph-noarch 628 k 2026-03-09T15:17:35.285 INFO:teuthology.orchestra.run.vm04.stdout: ceph-mgr-modules-core noarch 2:19.2.3-678.ge911bdeb.el9 @ceph-noarch 1.5 M 2026-03-09T15:17:35.285 INFO:teuthology.orchestra.run.vm04.stdout: ceph-prometheus-alerts noarch 2:19.2.3-678.ge911bdeb.el9 @ceph-noarch 52 k 2026-03-09T15:17:35.285 INFO:teuthology.orchestra.run.vm04.stdout: ceph-selinux x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 138 k 2026-03-09T15:17:35.285 INFO:teuthology.orchestra.run.vm04.stdout: cryptsetup x86_64 2.8.1-3.el9 @baseos 770 k 2026-03-09T15:17:35.285 INFO:teuthology.orchestra.run.vm04.stdout: flexiblas x86_64 3.0.4-9.el9 @appstream 68 k 2026-03-09T15:17:35.285 INFO:teuthology.orchestra.run.vm04.stdout: flexiblas-netlib x86_64 3.0.4-9.el9 @appstream 11 M 2026-03-09T15:17:35.285 INFO:teuthology.orchestra.run.vm04.stdout: flexiblas-openblas-openmp x86_64 3.0.4-9.el9 @appstream 39 k 2026-03-09T15:17:35.285 INFO:teuthology.orchestra.run.vm04.stdout: gperftools-libs x86_64 2.9.1-3.el9 @epel 1.4 M 2026-03-09T15:17:35.285 INFO:teuthology.orchestra.run.vm04.stdout: grpc-data noarch 1.46.7-10.el9 @epel 13 k 2026-03-09T15:17:35.285 INFO:teuthology.orchestra.run.vm04.stdout: ledmon-libs x86_64 1.1.0-3.el9 @baseos 80 k 2026-03-09T15:17:35.285 INFO:teuthology.orchestra.run.vm04.stdout: libcephsqlite x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 425 k 2026-03-09T15:17:35.285 INFO:teuthology.orchestra.run.vm04.stdout: libconfig x86_64 1.7.2-9.el9 @baseos 220 k 2026-03-09T15:17:35.285 INFO:teuthology.orchestra.run.vm04.stdout: libgfortran x86_64 11.5.0-14.el9 @baseos 2.8 M 2026-03-09T15:17:35.285 INFO:teuthology.orchestra.run.vm04.stdout: liboath x86_64 2.6.12-1.el9 @epel 94 k 2026-03-09T15:17:35.285 INFO:teuthology.orchestra.run.vm04.stdout: libquadmath x86_64 11.5.0-14.el9 @baseos 330 k 2026-03-09T15:17:35.285 INFO:teuthology.orchestra.run.vm04.stdout: libradosstriper1 x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 1.6 M 2026-03-09T15:17:35.285 INFO:teuthology.orchestra.run.vm04.stdout: libstoragemgmt x86_64 1.10.1-1.el9 @appstream 685 k 2026-03-09T15:17:35.285 INFO:teuthology.orchestra.run.vm04.stdout: libunwind x86_64 1.6.2-1.el9 @epel 170 k 2026-03-09T15:17:35.285 INFO:teuthology.orchestra.run.vm04.stdout: openblas x86_64 0.3.29-1.el9 @appstream 112 k 2026-03-09T15:17:35.285 INFO:teuthology.orchestra.run.vm04.stdout: openblas-openmp x86_64 0.3.29-1.el9 @appstream 46 M 2026-03-09T15:17:35.285 INFO:teuthology.orchestra.run.vm04.stdout: pciutils x86_64 3.7.0-7.el9 @baseos 216 k 2026-03-09T15:17:35.285 INFO:teuthology.orchestra.run.vm04.stdout: protobuf x86_64 3.14.0-17.el9 @appstream 3.5 M 2026-03-09T15:17:35.285 INFO:teuthology.orchestra.run.vm04.stdout: protobuf-compiler x86_64 3.14.0-17.el9 @crb 2.9 M 2026-03-09T15:17:35.285 INFO:teuthology.orchestra.run.vm04.stdout: python3-asyncssh noarch 2.13.2-5.el9 @epel 3.9 M 2026-03-09T15:17:35.285 INFO:teuthology.orchestra.run.vm04.stdout: python3-autocommand noarch 2.2.2-8.el9 @epel 82 k 2026-03-09T15:17:35.286 INFO:teuthology.orchestra.run.vm04.stdout: python3-babel noarch 2.9.1-2.el9 @appstream 27 M 2026-03-09T15:17:35.286 INFO:teuthology.orchestra.run.vm04.stdout: python3-backports-tarfile noarch 1.2.0-1.el9 @epel 254 k 2026-03-09T15:17:35.286 INFO:teuthology.orchestra.run.vm04.stdout: python3-bcrypt x86_64 3.2.2-1.el9 @epel 87 k 2026-03-09T15:17:35.286 INFO:teuthology.orchestra.run.vm04.stdout: python3-cachetools noarch 4.2.4-1.el9 @epel 93 k 2026-03-09T15:17:35.286 INFO:teuthology.orchestra.run.vm04.stdout: python3-ceph-common x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 702 k 2026-03-09T15:17:35.286 INFO:teuthology.orchestra.run.vm04.stdout: python3-certifi noarch 2023.05.07-4.el9 @epel 6.3 k 2026-03-09T15:17:35.286 INFO:teuthology.orchestra.run.vm04.stdout: python3-cffi x86_64 1.14.5-5.el9 @baseos 1.0 M 2026-03-09T15:17:35.286 INFO:teuthology.orchestra.run.vm04.stdout: python3-chardet noarch 4.0.0-5.el9 @anaconda 1.4 M 2026-03-09T15:17:35.286 INFO:teuthology.orchestra.run.vm04.stdout: python3-cheroot noarch 10.0.1-4.el9 @epel 682 k 2026-03-09T15:17:35.286 INFO:teuthology.orchestra.run.vm04.stdout: python3-cherrypy noarch 18.6.1-2.el9 @epel 1.1 M 2026-03-09T15:17:35.286 INFO:teuthology.orchestra.run.vm04.stdout: python3-cryptography x86_64 36.0.1-5.el9 @baseos 4.5 M 2026-03-09T15:17:35.286 INFO:teuthology.orchestra.run.vm04.stdout: python3-devel x86_64 3.9.25-3.el9 @appstream 765 k 2026-03-09T15:17:35.286 INFO:teuthology.orchestra.run.vm04.stdout: python3-google-auth noarch 1:2.45.0-1.el9 @epel 1.4 M 2026-03-09T15:17:35.286 INFO:teuthology.orchestra.run.vm04.stdout: python3-grpcio x86_64 1.46.7-10.el9 @epel 6.7 M 2026-03-09T15:17:35.286 INFO:teuthology.orchestra.run.vm04.stdout: python3-grpcio-tools x86_64 1.46.7-10.el9 @epel 418 k 2026-03-09T15:17:35.286 INFO:teuthology.orchestra.run.vm04.stdout: python3-idna noarch 2.10-7.el9.1 @anaconda 513 k 2026-03-09T15:17:35.286 INFO:teuthology.orchestra.run.vm04.stdout: python3-jaraco noarch 8.2.1-3.el9 @epel 3.7 k 2026-03-09T15:17:35.286 INFO:teuthology.orchestra.run.vm04.stdout: python3-jaraco-classes noarch 3.2.1-5.el9 @epel 24 k 2026-03-09T15:17:35.286 INFO:teuthology.orchestra.run.vm04.stdout: python3-jaraco-collections noarch 3.0.0-8.el9 @epel 55 k 2026-03-09T15:17:35.286 INFO:teuthology.orchestra.run.vm04.stdout: python3-jaraco-context noarch 6.0.1-3.el9 @epel 31 k 2026-03-09T15:17:35.286 INFO:teuthology.orchestra.run.vm04.stdout: python3-jaraco-functools noarch 3.5.0-2.el9 @epel 33 k 2026-03-09T15:17:35.286 INFO:teuthology.orchestra.run.vm04.stdout: python3-jaraco-text noarch 4.0.0-2.el9 @epel 51 k 2026-03-09T15:17:35.286 INFO:teuthology.orchestra.run.vm04.stdout: python3-jinja2 noarch 2.11.3-8.el9 @appstream 1.1 M 2026-03-09T15:17:35.286 INFO:teuthology.orchestra.run.vm04.stdout: python3-jsonpatch noarch 1.21-16.el9 @koji-override-0 55 k 2026-03-09T15:17:35.286 INFO:teuthology.orchestra.run.vm04.stdout: python3-jsonpointer noarch 2.0-4.el9 @koji-override-0 34 k 2026-03-09T15:17:35.286 INFO:teuthology.orchestra.run.vm04.stdout: python3-kubernetes noarch 1:26.1.0-3.el9 @epel 21 M 2026-03-09T15:17:35.286 INFO:teuthology.orchestra.run.vm04.stdout: python3-libstoragemgmt x86_64 1.10.1-1.el9 @appstream 832 k 2026-03-09T15:17:35.286 INFO:teuthology.orchestra.run.vm04.stdout: python3-logutils noarch 0.3.5-21.el9 @epel 126 k 2026-03-09T15:17:35.286 INFO:teuthology.orchestra.run.vm04.stdout: python3-mako noarch 1.1.4-6.el9 @appstream 534 k 2026-03-09T15:17:35.286 INFO:teuthology.orchestra.run.vm04.stdout: python3-markupsafe x86_64 1.1.1-12.el9 @appstream 60 k 2026-03-09T15:17:35.286 INFO:teuthology.orchestra.run.vm04.stdout: python3-more-itertools noarch 8.12.0-2.el9 @epel 378 k 2026-03-09T15:17:35.286 INFO:teuthology.orchestra.run.vm04.stdout: python3-natsort noarch 7.1.1-5.el9 @epel 215 k 2026-03-09T15:17:35.286 INFO:teuthology.orchestra.run.vm04.stdout: python3-numpy x86_64 1:1.23.5-2.el9 @appstream 30 M 2026-03-09T15:17:35.286 INFO:teuthology.orchestra.run.vm04.stdout: python3-numpy-f2py x86_64 1:1.23.5-2.el9 @appstream 1.7 M 2026-03-09T15:17:35.286 INFO:teuthology.orchestra.run.vm04.stdout: python3-oauthlib noarch 3.1.1-5.el9 @koji-override-0 888 k 2026-03-09T15:17:35.286 INFO:teuthology.orchestra.run.vm04.stdout: python3-packaging noarch 20.9-5.el9 @appstream 248 k 2026-03-09T15:17:35.286 INFO:teuthology.orchestra.run.vm04.stdout: python3-pecan noarch 1.4.2-3.el9 @epel 1.3 M 2026-03-09T15:17:35.286 INFO:teuthology.orchestra.run.vm04.stdout: python3-ply noarch 3.11-14.el9 @baseos 430 k 2026-03-09T15:17:35.286 INFO:teuthology.orchestra.run.vm04.stdout: python3-portend noarch 3.1.0-2.el9 @epel 20 k 2026-03-09T15:17:35.286 INFO:teuthology.orchestra.run.vm04.stdout: python3-prettytable noarch 0.7.2-27.el9 @koji-override-0 166 k 2026-03-09T15:17:35.286 INFO:teuthology.orchestra.run.vm04.stdout: python3-protobuf noarch 3.14.0-17.el9 @appstream 1.4 M 2026-03-09T15:17:35.286 INFO:teuthology.orchestra.run.vm04.stdout: python3-pyOpenSSL noarch 21.0.0-1.el9 @epel 389 k 2026-03-09T15:17:35.286 INFO:teuthology.orchestra.run.vm04.stdout: python3-pyasn1 noarch 0.4.8-7.el9 @appstream 622 k 2026-03-09T15:17:35.286 INFO:teuthology.orchestra.run.vm04.stdout: python3-pyasn1-modules noarch 0.4.8-7.el9 @appstream 1.0 M 2026-03-09T15:17:35.286 INFO:teuthology.orchestra.run.vm04.stdout: python3-pycparser noarch 2.20-6.el9 @baseos 745 k 2026-03-09T15:17:35.286 INFO:teuthology.orchestra.run.vm04.stdout: python3-pysocks noarch 1.7.1-12.el9 @anaconda 88 k 2026-03-09T15:17:35.286 INFO:teuthology.orchestra.run.vm04.stdout: python3-pytz noarch 2021.1-5.el9 @koji-override-0 176 k 2026-03-09T15:17:35.286 INFO:teuthology.orchestra.run.vm04.stdout: python3-repoze-lru noarch 0.7-16.el9 @epel 83 k 2026-03-09T15:17:35.286 INFO:teuthology.orchestra.run.vm04.stdout: python3-requests noarch 2.25.1-10.el9 @baseos 405 k 2026-03-09T15:17:35.286 INFO:teuthology.orchestra.run.vm04.stdout: python3-requests-oauthlib noarch 1.3.0-12.el9 @appstream 119 k 2026-03-09T15:17:35.286 INFO:teuthology.orchestra.run.vm04.stdout: python3-routes noarch 2.5.1-5.el9 @epel 459 k 2026-03-09T15:17:35.286 INFO:teuthology.orchestra.run.vm04.stdout: python3-rsa noarch 4.9-2.el9 @epel 202 k 2026-03-09T15:17:35.286 INFO:teuthology.orchestra.run.vm04.stdout: python3-scipy x86_64 1.9.3-2.el9 @appstream 76 M 2026-03-09T15:17:35.286 INFO:teuthology.orchestra.run.vm04.stdout: python3-tempora noarch 5.0.0-2.el9 @epel 96 k 2026-03-09T15:17:35.286 INFO:teuthology.orchestra.run.vm04.stdout: python3-toml noarch 0.10.2-6.el9 @appstream 99 k 2026-03-09T15:17:35.286 INFO:teuthology.orchestra.run.vm04.stdout: python3-typing-extensions noarch 4.15.0-1.el9 @epel 447 k 2026-03-09T15:17:35.286 INFO:teuthology.orchestra.run.vm04.stdout: python3-urllib3 noarch 1.26.5-7.el9 @baseos 746 k 2026-03-09T15:17:35.286 INFO:teuthology.orchestra.run.vm04.stdout: python3-webob noarch 1.8.8-2.el9 @epel 1.2 M 2026-03-09T15:17:35.286 INFO:teuthology.orchestra.run.vm04.stdout: python3-websocket-client noarch 1.2.3-2.el9 @epel 319 k 2026-03-09T15:17:35.286 INFO:teuthology.orchestra.run.vm04.stdout: python3-werkzeug noarch 2.0.3-3.el9.1 @epel 1.9 M 2026-03-09T15:17:35.286 INFO:teuthology.orchestra.run.vm04.stdout: python3-zc-lockfile noarch 2.0-10.el9 @epel 35 k 2026-03-09T15:17:35.286 INFO:teuthology.orchestra.run.vm04.stdout: qatlib x86_64 25.08.0-2.el9 @appstream 639 k 2026-03-09T15:17:35.287 INFO:teuthology.orchestra.run.vm04.stdout: qatlib-service x86_64 25.08.0-2.el9 @appstream 69 k 2026-03-09T15:17:35.287 INFO:teuthology.orchestra.run.vm04.stdout: qatzip-libs x86_64 1.3.1-1.el9 @appstream 148 k 2026-03-09T15:17:35.287 INFO:teuthology.orchestra.run.vm04.stdout: 2026-03-09T15:17:35.287 INFO:teuthology.orchestra.run.vm04.stdout:Transaction Summary 2026-03-09T15:17:35.287 INFO:teuthology.orchestra.run.vm04.stdout:=========================================================================================== 2026-03-09T15:17:35.287 INFO:teuthology.orchestra.run.vm04.stdout:Remove 102 Packages 2026-03-09T15:17:35.287 INFO:teuthology.orchestra.run.vm04.stdout: 2026-03-09T15:17:35.287 INFO:teuthology.orchestra.run.vm04.stdout:Freed space: 613 M 2026-03-09T15:17:35.287 INFO:teuthology.orchestra.run.vm04.stdout:Running transaction check 2026-03-09T15:17:35.287 INFO:teuthology.orchestra.run.vm06.stdout:=========================================================================================== 2026-03-09T15:17:35.287 INFO:teuthology.orchestra.run.vm06.stdout: Package Arch Version Repository Size 2026-03-09T15:17:35.287 INFO:teuthology.orchestra.run.vm06.stdout:=========================================================================================== 2026-03-09T15:17:35.287 INFO:teuthology.orchestra.run.vm06.stdout:Removing: 2026-03-09T15:17:35.287 INFO:teuthology.orchestra.run.vm06.stdout: ceph-base x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 23 M 2026-03-09T15:17:35.287 INFO:teuthology.orchestra.run.vm06.stdout:Removing dependent packages: 2026-03-09T15:17:35.287 INFO:teuthology.orchestra.run.vm06.stdout: ceph-immutable-object-cache x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 431 k 2026-03-09T15:17:35.287 INFO:teuthology.orchestra.run.vm06.stdout: ceph-mgr x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 3.4 M 2026-03-09T15:17:35.287 INFO:teuthology.orchestra.run.vm06.stdout: ceph-mgr-cephadm noarch 2:19.2.3-678.ge911bdeb.el9 @ceph-noarch 806 k 2026-03-09T15:17:35.287 INFO:teuthology.orchestra.run.vm06.stdout: ceph-mgr-dashboard noarch 2:19.2.3-678.ge911bdeb.el9 @ceph-noarch 88 M 2026-03-09T15:17:35.287 INFO:teuthology.orchestra.run.vm06.stdout: ceph-mgr-diskprediction-local noarch 2:19.2.3-678.ge911bdeb.el9 @ceph-noarch 66 M 2026-03-09T15:17:35.287 INFO:teuthology.orchestra.run.vm06.stdout: ceph-mgr-rook noarch 2:19.2.3-678.ge911bdeb.el9 @ceph-noarch 563 k 2026-03-09T15:17:35.287 INFO:teuthology.orchestra.run.vm06.stdout: ceph-osd x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 59 M 2026-03-09T15:17:35.287 INFO:teuthology.orchestra.run.vm06.stdout: ceph-volume noarch 2:19.2.3-678.ge911bdeb.el9 @ceph-noarch 1.4 M 2026-03-09T15:17:35.287 INFO:teuthology.orchestra.run.vm06.stdout: rbd-mirror x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 13 M 2026-03-09T15:17:35.287 INFO:teuthology.orchestra.run.vm06.stdout:Removing unused dependencies: 2026-03-09T15:17:35.287 INFO:teuthology.orchestra.run.vm06.stdout: abseil-cpp x86_64 20211102.0-4.el9 @epel 1.9 M 2026-03-09T15:17:35.287 INFO:teuthology.orchestra.run.vm06.stdout: ceph-common x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 85 M 2026-03-09T15:17:35.287 INFO:teuthology.orchestra.run.vm06.stdout: ceph-grafana-dashboards noarch 2:19.2.3-678.ge911bdeb.el9 @ceph-noarch 628 k 2026-03-09T15:17:35.287 INFO:teuthology.orchestra.run.vm06.stdout: ceph-mgr-modules-core noarch 2:19.2.3-678.ge911bdeb.el9 @ceph-noarch 1.5 M 2026-03-09T15:17:35.287 INFO:teuthology.orchestra.run.vm06.stdout: ceph-prometheus-alerts noarch 2:19.2.3-678.ge911bdeb.el9 @ceph-noarch 52 k 2026-03-09T15:17:35.287 INFO:teuthology.orchestra.run.vm06.stdout: ceph-selinux x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 138 k 2026-03-09T15:17:35.287 INFO:teuthology.orchestra.run.vm06.stdout: cryptsetup x86_64 2.8.1-3.el9 @baseos 770 k 2026-03-09T15:17:35.287 INFO:teuthology.orchestra.run.vm06.stdout: flexiblas x86_64 3.0.4-9.el9 @appstream 68 k 2026-03-09T15:17:35.287 INFO:teuthology.orchestra.run.vm06.stdout: flexiblas-netlib x86_64 3.0.4-9.el9 @appstream 11 M 2026-03-09T15:17:35.287 INFO:teuthology.orchestra.run.vm06.stdout: flexiblas-openblas-openmp x86_64 3.0.4-9.el9 @appstream 39 k 2026-03-09T15:17:35.287 INFO:teuthology.orchestra.run.vm06.stdout: gperftools-libs x86_64 2.9.1-3.el9 @epel 1.4 M 2026-03-09T15:17:35.287 INFO:teuthology.orchestra.run.vm06.stdout: grpc-data noarch 1.46.7-10.el9 @epel 13 k 2026-03-09T15:17:35.287 INFO:teuthology.orchestra.run.vm06.stdout: ledmon-libs x86_64 1.1.0-3.el9 @baseos 80 k 2026-03-09T15:17:35.287 INFO:teuthology.orchestra.run.vm06.stdout: libcephsqlite x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 425 k 2026-03-09T15:17:35.287 INFO:teuthology.orchestra.run.vm06.stdout: libconfig x86_64 1.7.2-9.el9 @baseos 220 k 2026-03-09T15:17:35.287 INFO:teuthology.orchestra.run.vm06.stdout: libgfortran x86_64 11.5.0-14.el9 @baseos 2.8 M 2026-03-09T15:17:35.287 INFO:teuthology.orchestra.run.vm06.stdout: liboath x86_64 2.6.12-1.el9 @epel 94 k 2026-03-09T15:17:35.287 INFO:teuthology.orchestra.run.vm06.stdout: libquadmath x86_64 11.5.0-14.el9 @baseos 330 k 2026-03-09T15:17:35.287 INFO:teuthology.orchestra.run.vm06.stdout: libradosstriper1 x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 1.6 M 2026-03-09T15:17:35.287 INFO:teuthology.orchestra.run.vm06.stdout: libstoragemgmt x86_64 1.10.1-1.el9 @appstream 685 k 2026-03-09T15:17:35.287 INFO:teuthology.orchestra.run.vm06.stdout: libunwind x86_64 1.6.2-1.el9 @epel 170 k 2026-03-09T15:17:35.287 INFO:teuthology.orchestra.run.vm06.stdout: openblas x86_64 0.3.29-1.el9 @appstream 112 k 2026-03-09T15:17:35.288 INFO:teuthology.orchestra.run.vm06.stdout: openblas-openmp x86_64 0.3.29-1.el9 @appstream 46 M 2026-03-09T15:17:35.288 INFO:teuthology.orchestra.run.vm06.stdout: pciutils x86_64 3.7.0-7.el9 @baseos 216 k 2026-03-09T15:17:35.288 INFO:teuthology.orchestra.run.vm06.stdout: protobuf x86_64 3.14.0-17.el9 @appstream 3.5 M 2026-03-09T15:17:35.288 INFO:teuthology.orchestra.run.vm06.stdout: protobuf-compiler x86_64 3.14.0-17.el9 @crb 2.9 M 2026-03-09T15:17:35.288 INFO:teuthology.orchestra.run.vm06.stdout: python3-asyncssh noarch 2.13.2-5.el9 @epel 3.9 M 2026-03-09T15:17:35.288 INFO:teuthology.orchestra.run.vm06.stdout: python3-autocommand noarch 2.2.2-8.el9 @epel 82 k 2026-03-09T15:17:35.288 INFO:teuthology.orchestra.run.vm06.stdout: python3-babel noarch 2.9.1-2.el9 @appstream 27 M 2026-03-09T15:17:35.288 INFO:teuthology.orchestra.run.vm06.stdout: python3-backports-tarfile noarch 1.2.0-1.el9 @epel 254 k 2026-03-09T15:17:35.288 INFO:teuthology.orchestra.run.vm06.stdout: python3-bcrypt x86_64 3.2.2-1.el9 @epel 87 k 2026-03-09T15:17:35.288 INFO:teuthology.orchestra.run.vm06.stdout: python3-cachetools noarch 4.2.4-1.el9 @epel 93 k 2026-03-09T15:17:35.288 INFO:teuthology.orchestra.run.vm06.stdout: python3-ceph-common x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 702 k 2026-03-09T15:17:35.288 INFO:teuthology.orchestra.run.vm06.stdout: python3-certifi noarch 2023.05.07-4.el9 @epel 6.3 k 2026-03-09T15:17:35.288 INFO:teuthology.orchestra.run.vm06.stdout: python3-cffi x86_64 1.14.5-5.el9 @baseos 1.0 M 2026-03-09T15:17:35.288 INFO:teuthology.orchestra.run.vm06.stdout: python3-chardet noarch 4.0.0-5.el9 @anaconda 1.4 M 2026-03-09T15:17:35.288 INFO:teuthology.orchestra.run.vm06.stdout: python3-cheroot noarch 10.0.1-4.el9 @epel 682 k 2026-03-09T15:17:35.288 INFO:teuthology.orchestra.run.vm06.stdout: python3-cherrypy noarch 18.6.1-2.el9 @epel 1.1 M 2026-03-09T15:17:35.288 INFO:teuthology.orchestra.run.vm06.stdout: python3-cryptography x86_64 36.0.1-5.el9 @baseos 4.5 M 2026-03-09T15:17:35.288 INFO:teuthology.orchestra.run.vm06.stdout: python3-devel x86_64 3.9.25-3.el9 @appstream 765 k 2026-03-09T15:17:35.288 INFO:teuthology.orchestra.run.vm06.stdout: python3-google-auth noarch 1:2.45.0-1.el9 @epel 1.4 M 2026-03-09T15:17:35.288 INFO:teuthology.orchestra.run.vm06.stdout: python3-grpcio x86_64 1.46.7-10.el9 @epel 6.7 M 2026-03-09T15:17:35.288 INFO:teuthology.orchestra.run.vm06.stdout: python3-grpcio-tools x86_64 1.46.7-10.el9 @epel 418 k 2026-03-09T15:17:35.288 INFO:teuthology.orchestra.run.vm06.stdout: python3-idna noarch 2.10-7.el9.1 @anaconda 513 k 2026-03-09T15:17:35.288 INFO:teuthology.orchestra.run.vm06.stdout: python3-jaraco noarch 8.2.1-3.el9 @epel 3.7 k 2026-03-09T15:17:35.288 INFO:teuthology.orchestra.run.vm06.stdout: python3-jaraco-classes noarch 3.2.1-5.el9 @epel 24 k 2026-03-09T15:17:35.288 INFO:teuthology.orchestra.run.vm06.stdout: python3-jaraco-collections noarch 3.0.0-8.el9 @epel 55 k 2026-03-09T15:17:35.288 INFO:teuthology.orchestra.run.vm06.stdout: python3-jaraco-context noarch 6.0.1-3.el9 @epel 31 k 2026-03-09T15:17:35.288 INFO:teuthology.orchestra.run.vm06.stdout: python3-jaraco-functools noarch 3.5.0-2.el9 @epel 33 k 2026-03-09T15:17:35.288 INFO:teuthology.orchestra.run.vm06.stdout: python3-jaraco-text noarch 4.0.0-2.el9 @epel 51 k 2026-03-09T15:17:35.288 INFO:teuthology.orchestra.run.vm06.stdout: python3-jinja2 noarch 2.11.3-8.el9 @appstream 1.1 M 2026-03-09T15:17:35.288 INFO:teuthology.orchestra.run.vm06.stdout: python3-jsonpatch noarch 1.21-16.el9 @koji-override-0 55 k 2026-03-09T15:17:35.288 INFO:teuthology.orchestra.run.vm06.stdout: python3-jsonpointer noarch 2.0-4.el9 @koji-override-0 34 k 2026-03-09T15:17:35.288 INFO:teuthology.orchestra.run.vm06.stdout: python3-kubernetes noarch 1:26.1.0-3.el9 @epel 21 M 2026-03-09T15:17:35.288 INFO:teuthology.orchestra.run.vm06.stdout: python3-libstoragemgmt x86_64 1.10.1-1.el9 @appstream 832 k 2026-03-09T15:17:35.288 INFO:teuthology.orchestra.run.vm06.stdout: python3-logutils noarch 0.3.5-21.el9 @epel 126 k 2026-03-09T15:17:35.288 INFO:teuthology.orchestra.run.vm06.stdout: python3-mako noarch 1.1.4-6.el9 @appstream 534 k 2026-03-09T15:17:35.288 INFO:teuthology.orchestra.run.vm06.stdout: python3-markupsafe x86_64 1.1.1-12.el9 @appstream 60 k 2026-03-09T15:17:35.288 INFO:teuthology.orchestra.run.vm06.stdout: python3-more-itertools noarch 8.12.0-2.el9 @epel 378 k 2026-03-09T15:17:35.288 INFO:teuthology.orchestra.run.vm06.stdout: python3-natsort noarch 7.1.1-5.el9 @epel 215 k 2026-03-09T15:17:35.288 INFO:teuthology.orchestra.run.vm06.stdout: python3-numpy x86_64 1:1.23.5-2.el9 @appstream 30 M 2026-03-09T15:17:35.288 INFO:teuthology.orchestra.run.vm06.stdout: python3-numpy-f2py x86_64 1:1.23.5-2.el9 @appstream 1.7 M 2026-03-09T15:17:35.288 INFO:teuthology.orchestra.run.vm06.stdout: python3-oauthlib noarch 3.1.1-5.el9 @koji-override-0 888 k 2026-03-09T15:17:35.288 INFO:teuthology.orchestra.run.vm06.stdout: python3-packaging noarch 20.9-5.el9 @appstream 248 k 2026-03-09T15:17:35.288 INFO:teuthology.orchestra.run.vm06.stdout: python3-pecan noarch 1.4.2-3.el9 @epel 1.3 M 2026-03-09T15:17:35.288 INFO:teuthology.orchestra.run.vm06.stdout: python3-ply noarch 3.11-14.el9 @baseos 430 k 2026-03-09T15:17:35.288 INFO:teuthology.orchestra.run.vm06.stdout: python3-portend noarch 3.1.0-2.el9 @epel 20 k 2026-03-09T15:17:35.288 INFO:teuthology.orchestra.run.vm06.stdout: python3-prettytable noarch 0.7.2-27.el9 @koji-override-0 166 k 2026-03-09T15:17:35.288 INFO:teuthology.orchestra.run.vm06.stdout: python3-protobuf noarch 3.14.0-17.el9 @appstream 1.4 M 2026-03-09T15:17:35.288 INFO:teuthology.orchestra.run.vm06.stdout: python3-pyOpenSSL noarch 21.0.0-1.el9 @epel 389 k 2026-03-09T15:17:35.288 INFO:teuthology.orchestra.run.vm06.stdout: python3-pyasn1 noarch 0.4.8-7.el9 @appstream 622 k 2026-03-09T15:17:35.288 INFO:teuthology.orchestra.run.vm06.stdout: python3-pyasn1-modules noarch 0.4.8-7.el9 @appstream 1.0 M 2026-03-09T15:17:35.288 INFO:teuthology.orchestra.run.vm06.stdout: python3-pycparser noarch 2.20-6.el9 @baseos 745 k 2026-03-09T15:17:35.288 INFO:teuthology.orchestra.run.vm06.stdout: python3-pysocks noarch 1.7.1-12.el9 @anaconda 88 k 2026-03-09T15:17:35.288 INFO:teuthology.orchestra.run.vm06.stdout: python3-pytz noarch 2021.1-5.el9 @koji-override-0 176 k 2026-03-09T15:17:35.288 INFO:teuthology.orchestra.run.vm06.stdout: python3-repoze-lru noarch 0.7-16.el9 @epel 83 k 2026-03-09T15:17:35.288 INFO:teuthology.orchestra.run.vm06.stdout: python3-requests noarch 2.25.1-10.el9 @baseos 405 k 2026-03-09T15:17:35.288 INFO:teuthology.orchestra.run.vm06.stdout: python3-requests-oauthlib noarch 1.3.0-12.el9 @appstream 119 k 2026-03-09T15:17:35.288 INFO:teuthology.orchestra.run.vm06.stdout: python3-routes noarch 2.5.1-5.el9 @epel 459 k 2026-03-09T15:17:35.288 INFO:teuthology.orchestra.run.vm06.stdout: python3-rsa noarch 4.9-2.el9 @epel 202 k 2026-03-09T15:17:35.288 INFO:teuthology.orchestra.run.vm06.stdout: python3-scipy x86_64 1.9.3-2.el9 @appstream 76 M 2026-03-09T15:17:35.288 INFO:teuthology.orchestra.run.vm06.stdout: python3-tempora noarch 5.0.0-2.el9 @epel 96 k 2026-03-09T15:17:35.288 INFO:teuthology.orchestra.run.vm06.stdout: python3-toml noarch 0.10.2-6.el9 @appstream 99 k 2026-03-09T15:17:35.288 INFO:teuthology.orchestra.run.vm06.stdout: python3-typing-extensions noarch 4.15.0-1.el9 @epel 447 k 2026-03-09T15:17:35.288 INFO:teuthology.orchestra.run.vm06.stdout: python3-urllib3 noarch 1.26.5-7.el9 @baseos 746 k 2026-03-09T15:17:35.288 INFO:teuthology.orchestra.run.vm06.stdout: python3-webob noarch 1.8.8-2.el9 @epel 1.2 M 2026-03-09T15:17:35.289 INFO:teuthology.orchestra.run.vm06.stdout: python3-websocket-client noarch 1.2.3-2.el9 @epel 319 k 2026-03-09T15:17:35.289 INFO:teuthology.orchestra.run.vm06.stdout: python3-werkzeug noarch 2.0.3-3.el9.1 @epel 1.9 M 2026-03-09T15:17:35.289 INFO:teuthology.orchestra.run.vm06.stdout: python3-zc-lockfile noarch 2.0-10.el9 @epel 35 k 2026-03-09T15:17:35.289 INFO:teuthology.orchestra.run.vm06.stdout: qatlib x86_64 25.08.0-2.el9 @appstream 639 k 2026-03-09T15:17:35.289 INFO:teuthology.orchestra.run.vm06.stdout: qatlib-service x86_64 25.08.0-2.el9 @appstream 69 k 2026-03-09T15:17:35.289 INFO:teuthology.orchestra.run.vm06.stdout: qatzip-libs x86_64 1.3.1-1.el9 @appstream 148 k 2026-03-09T15:17:35.289 INFO:teuthology.orchestra.run.vm06.stdout: 2026-03-09T15:17:35.289 INFO:teuthology.orchestra.run.vm06.stdout:Transaction Summary 2026-03-09T15:17:35.289 INFO:teuthology.orchestra.run.vm06.stdout:=========================================================================================== 2026-03-09T15:17:35.289 INFO:teuthology.orchestra.run.vm06.stdout:Remove 102 Packages 2026-03-09T15:17:35.289 INFO:teuthology.orchestra.run.vm06.stdout: 2026-03-09T15:17:35.289 INFO:teuthology.orchestra.run.vm06.stdout:Freed space: 613 M 2026-03-09T15:17:35.289 INFO:teuthology.orchestra.run.vm06.stdout:Running transaction check 2026-03-09T15:17:35.291 INFO:teuthology.orchestra.run.vm03.stdout:Dependencies resolved. 2026-03-09T15:17:35.296 INFO:teuthology.orchestra.run.vm03.stdout:=========================================================================================== 2026-03-09T15:17:35.296 INFO:teuthology.orchestra.run.vm03.stdout: Package Arch Version Repository Size 2026-03-09T15:17:35.296 INFO:teuthology.orchestra.run.vm03.stdout:=========================================================================================== 2026-03-09T15:17:35.296 INFO:teuthology.orchestra.run.vm03.stdout:Removing: 2026-03-09T15:17:35.296 INFO:teuthology.orchestra.run.vm03.stdout: ceph-base x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 23 M 2026-03-09T15:17:35.296 INFO:teuthology.orchestra.run.vm03.stdout:Removing dependent packages: 2026-03-09T15:17:35.296 INFO:teuthology.orchestra.run.vm03.stdout: ceph-immutable-object-cache x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 431 k 2026-03-09T15:17:35.296 INFO:teuthology.orchestra.run.vm03.stdout: ceph-mgr x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 3.4 M 2026-03-09T15:17:35.296 INFO:teuthology.orchestra.run.vm03.stdout: ceph-mgr-cephadm noarch 2:19.2.3-678.ge911bdeb.el9 @ceph-noarch 806 k 2026-03-09T15:17:35.296 INFO:teuthology.orchestra.run.vm03.stdout: ceph-mgr-dashboard noarch 2:19.2.3-678.ge911bdeb.el9 @ceph-noarch 88 M 2026-03-09T15:17:35.296 INFO:teuthology.orchestra.run.vm03.stdout: ceph-mgr-diskprediction-local noarch 2:19.2.3-678.ge911bdeb.el9 @ceph-noarch 66 M 2026-03-09T15:17:35.296 INFO:teuthology.orchestra.run.vm03.stdout: ceph-mgr-rook noarch 2:19.2.3-678.ge911bdeb.el9 @ceph-noarch 563 k 2026-03-09T15:17:35.296 INFO:teuthology.orchestra.run.vm03.stdout: ceph-osd x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 59 M 2026-03-09T15:17:35.296 INFO:teuthology.orchestra.run.vm03.stdout: ceph-volume noarch 2:19.2.3-678.ge911bdeb.el9 @ceph-noarch 1.4 M 2026-03-09T15:17:35.296 INFO:teuthology.orchestra.run.vm03.stdout: rbd-mirror x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 13 M 2026-03-09T15:17:35.296 INFO:teuthology.orchestra.run.vm03.stdout:Removing unused dependencies: 2026-03-09T15:17:35.296 INFO:teuthology.orchestra.run.vm03.stdout: abseil-cpp x86_64 20211102.0-4.el9 @epel 1.9 M 2026-03-09T15:17:35.296 INFO:teuthology.orchestra.run.vm03.stdout: ceph-common x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 85 M 2026-03-09T15:17:35.297 INFO:teuthology.orchestra.run.vm03.stdout: ceph-grafana-dashboards noarch 2:19.2.3-678.ge911bdeb.el9 @ceph-noarch 628 k 2026-03-09T15:17:35.297 INFO:teuthology.orchestra.run.vm03.stdout: ceph-mgr-modules-core noarch 2:19.2.3-678.ge911bdeb.el9 @ceph-noarch 1.5 M 2026-03-09T15:17:35.297 INFO:teuthology.orchestra.run.vm03.stdout: ceph-prometheus-alerts noarch 2:19.2.3-678.ge911bdeb.el9 @ceph-noarch 52 k 2026-03-09T15:17:35.297 INFO:teuthology.orchestra.run.vm03.stdout: ceph-selinux x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 138 k 2026-03-09T15:17:35.297 INFO:teuthology.orchestra.run.vm03.stdout: cryptsetup x86_64 2.8.1-3.el9 @baseos 770 k 2026-03-09T15:17:35.297 INFO:teuthology.orchestra.run.vm03.stdout: flexiblas x86_64 3.0.4-9.el9 @appstream 68 k 2026-03-09T15:17:35.297 INFO:teuthology.orchestra.run.vm03.stdout: flexiblas-netlib x86_64 3.0.4-9.el9 @appstream 11 M 2026-03-09T15:17:35.297 INFO:teuthology.orchestra.run.vm03.stdout: flexiblas-openblas-openmp x86_64 3.0.4-9.el9 @appstream 39 k 2026-03-09T15:17:35.297 INFO:teuthology.orchestra.run.vm03.stdout: gperftools-libs x86_64 2.9.1-3.el9 @epel 1.4 M 2026-03-09T15:17:35.297 INFO:teuthology.orchestra.run.vm03.stdout: grpc-data noarch 1.46.7-10.el9 @epel 13 k 2026-03-09T15:17:35.297 INFO:teuthology.orchestra.run.vm03.stdout: ledmon-libs x86_64 1.1.0-3.el9 @baseos 80 k 2026-03-09T15:17:35.297 INFO:teuthology.orchestra.run.vm03.stdout: libcephsqlite x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 425 k 2026-03-09T15:17:35.297 INFO:teuthology.orchestra.run.vm03.stdout: libconfig x86_64 1.7.2-9.el9 @baseos 220 k 2026-03-09T15:17:35.297 INFO:teuthology.orchestra.run.vm03.stdout: libgfortran x86_64 11.5.0-14.el9 @baseos 2.8 M 2026-03-09T15:17:35.297 INFO:teuthology.orchestra.run.vm03.stdout: liboath x86_64 2.6.12-1.el9 @epel 94 k 2026-03-09T15:17:35.297 INFO:teuthology.orchestra.run.vm03.stdout: libquadmath x86_64 11.5.0-14.el9 @baseos 330 k 2026-03-09T15:17:35.297 INFO:teuthology.orchestra.run.vm03.stdout: libradosstriper1 x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 1.6 M 2026-03-09T15:17:35.297 INFO:teuthology.orchestra.run.vm03.stdout: libstoragemgmt x86_64 1.10.1-1.el9 @appstream 685 k 2026-03-09T15:17:35.297 INFO:teuthology.orchestra.run.vm03.stdout: libunwind x86_64 1.6.2-1.el9 @epel 170 k 2026-03-09T15:17:35.297 INFO:teuthology.orchestra.run.vm03.stdout: openblas x86_64 0.3.29-1.el9 @appstream 112 k 2026-03-09T15:17:35.297 INFO:teuthology.orchestra.run.vm03.stdout: openblas-openmp x86_64 0.3.29-1.el9 @appstream 46 M 2026-03-09T15:17:35.297 INFO:teuthology.orchestra.run.vm03.stdout: pciutils x86_64 3.7.0-7.el9 @baseos 216 k 2026-03-09T15:17:35.297 INFO:teuthology.orchestra.run.vm03.stdout: protobuf x86_64 3.14.0-17.el9 @appstream 3.5 M 2026-03-09T15:17:35.297 INFO:teuthology.orchestra.run.vm03.stdout: protobuf-compiler x86_64 3.14.0-17.el9 @crb 2.9 M 2026-03-09T15:17:35.297 INFO:teuthology.orchestra.run.vm03.stdout: python3-asyncssh noarch 2.13.2-5.el9 @epel 3.9 M 2026-03-09T15:17:35.297 INFO:teuthology.orchestra.run.vm03.stdout: python3-autocommand noarch 2.2.2-8.el9 @epel 82 k 2026-03-09T15:17:35.297 INFO:teuthology.orchestra.run.vm03.stdout: python3-babel noarch 2.9.1-2.el9 @appstream 27 M 2026-03-09T15:17:35.297 INFO:teuthology.orchestra.run.vm03.stdout: python3-backports-tarfile noarch 1.2.0-1.el9 @epel 254 k 2026-03-09T15:17:35.297 INFO:teuthology.orchestra.run.vm03.stdout: python3-bcrypt x86_64 3.2.2-1.el9 @epel 87 k 2026-03-09T15:17:35.297 INFO:teuthology.orchestra.run.vm03.stdout: python3-cachetools noarch 4.2.4-1.el9 @epel 93 k 2026-03-09T15:17:35.297 INFO:teuthology.orchestra.run.vm03.stdout: python3-ceph-common x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 702 k 2026-03-09T15:17:35.297 INFO:teuthology.orchestra.run.vm03.stdout: python3-certifi noarch 2023.05.07-4.el9 @epel 6.3 k 2026-03-09T15:17:35.297 INFO:teuthology.orchestra.run.vm03.stdout: python3-cffi x86_64 1.14.5-5.el9 @baseos 1.0 M 2026-03-09T15:17:35.297 INFO:teuthology.orchestra.run.vm03.stdout: python3-chardet noarch 4.0.0-5.el9 @anaconda 1.4 M 2026-03-09T15:17:35.297 INFO:teuthology.orchestra.run.vm03.stdout: python3-cheroot noarch 10.0.1-4.el9 @epel 682 k 2026-03-09T15:17:35.297 INFO:teuthology.orchestra.run.vm03.stdout: python3-cherrypy noarch 18.6.1-2.el9 @epel 1.1 M 2026-03-09T15:17:35.297 INFO:teuthology.orchestra.run.vm03.stdout: python3-cryptography x86_64 36.0.1-5.el9 @baseos 4.5 M 2026-03-09T15:17:35.297 INFO:teuthology.orchestra.run.vm03.stdout: python3-devel x86_64 3.9.25-3.el9 @appstream 765 k 2026-03-09T15:17:35.297 INFO:teuthology.orchestra.run.vm03.stdout: python3-google-auth noarch 1:2.45.0-1.el9 @epel 1.4 M 2026-03-09T15:17:35.297 INFO:teuthology.orchestra.run.vm03.stdout: python3-grpcio x86_64 1.46.7-10.el9 @epel 6.7 M 2026-03-09T15:17:35.297 INFO:teuthology.orchestra.run.vm03.stdout: python3-grpcio-tools x86_64 1.46.7-10.el9 @epel 418 k 2026-03-09T15:17:35.297 INFO:teuthology.orchestra.run.vm03.stdout: python3-idna noarch 2.10-7.el9.1 @anaconda 513 k 2026-03-09T15:17:35.297 INFO:teuthology.orchestra.run.vm03.stdout: python3-jaraco noarch 8.2.1-3.el9 @epel 3.7 k 2026-03-09T15:17:35.297 INFO:teuthology.orchestra.run.vm03.stdout: python3-jaraco-classes noarch 3.2.1-5.el9 @epel 24 k 2026-03-09T15:17:35.297 INFO:teuthology.orchestra.run.vm03.stdout: python3-jaraco-collections noarch 3.0.0-8.el9 @epel 55 k 2026-03-09T15:17:35.297 INFO:teuthology.orchestra.run.vm03.stdout: python3-jaraco-context noarch 6.0.1-3.el9 @epel 31 k 2026-03-09T15:17:35.297 INFO:teuthology.orchestra.run.vm03.stdout: python3-jaraco-functools noarch 3.5.0-2.el9 @epel 33 k 2026-03-09T15:17:35.297 INFO:teuthology.orchestra.run.vm03.stdout: python3-jaraco-text noarch 4.0.0-2.el9 @epel 51 k 2026-03-09T15:17:35.297 INFO:teuthology.orchestra.run.vm03.stdout: python3-jinja2 noarch 2.11.3-8.el9 @appstream 1.1 M 2026-03-09T15:17:35.297 INFO:teuthology.orchestra.run.vm03.stdout: python3-jsonpatch noarch 1.21-16.el9 @koji-override-0 55 k 2026-03-09T15:17:35.297 INFO:teuthology.orchestra.run.vm03.stdout: python3-jsonpointer noarch 2.0-4.el9 @koji-override-0 34 k 2026-03-09T15:17:35.297 INFO:teuthology.orchestra.run.vm03.stdout: python3-kubernetes noarch 1:26.1.0-3.el9 @epel 21 M 2026-03-09T15:17:35.297 INFO:teuthology.orchestra.run.vm03.stdout: python3-libstoragemgmt x86_64 1.10.1-1.el9 @appstream 832 k 2026-03-09T15:17:35.297 INFO:teuthology.orchestra.run.vm03.stdout: python3-logutils noarch 0.3.5-21.el9 @epel 126 k 2026-03-09T15:17:35.297 INFO:teuthology.orchestra.run.vm03.stdout: python3-mako noarch 1.1.4-6.el9 @appstream 534 k 2026-03-09T15:17:35.297 INFO:teuthology.orchestra.run.vm03.stdout: python3-markupsafe x86_64 1.1.1-12.el9 @appstream 60 k 2026-03-09T15:17:35.297 INFO:teuthology.orchestra.run.vm03.stdout: python3-more-itertools noarch 8.12.0-2.el9 @epel 378 k 2026-03-09T15:17:35.298 INFO:teuthology.orchestra.run.vm03.stdout: python3-natsort noarch 7.1.1-5.el9 @epel 215 k 2026-03-09T15:17:35.298 INFO:teuthology.orchestra.run.vm03.stdout: python3-numpy x86_64 1:1.23.5-2.el9 @appstream 30 M 2026-03-09T15:17:35.298 INFO:teuthology.orchestra.run.vm03.stdout: python3-numpy-f2py x86_64 1:1.23.5-2.el9 @appstream 1.7 M 2026-03-09T15:17:35.298 INFO:teuthology.orchestra.run.vm03.stdout: python3-oauthlib noarch 3.1.1-5.el9 @koji-override-0 888 k 2026-03-09T15:17:35.298 INFO:teuthology.orchestra.run.vm03.stdout: python3-packaging noarch 20.9-5.el9 @appstream 248 k 2026-03-09T15:17:35.298 INFO:teuthology.orchestra.run.vm03.stdout: python3-pecan noarch 1.4.2-3.el9 @epel 1.3 M 2026-03-09T15:17:35.298 INFO:teuthology.orchestra.run.vm03.stdout: python3-ply noarch 3.11-14.el9 @baseos 430 k 2026-03-09T15:17:35.298 INFO:teuthology.orchestra.run.vm03.stdout: python3-portend noarch 3.1.0-2.el9 @epel 20 k 2026-03-09T15:17:35.298 INFO:teuthology.orchestra.run.vm03.stdout: python3-prettytable noarch 0.7.2-27.el9 @koji-override-0 166 k 2026-03-09T15:17:35.298 INFO:teuthology.orchestra.run.vm03.stdout: python3-protobuf noarch 3.14.0-17.el9 @appstream 1.4 M 2026-03-09T15:17:35.298 INFO:teuthology.orchestra.run.vm03.stdout: python3-pyOpenSSL noarch 21.0.0-1.el9 @epel 389 k 2026-03-09T15:17:35.298 INFO:teuthology.orchestra.run.vm03.stdout: python3-pyasn1 noarch 0.4.8-7.el9 @appstream 622 k 2026-03-09T15:17:35.298 INFO:teuthology.orchestra.run.vm03.stdout: python3-pyasn1-modules noarch 0.4.8-7.el9 @appstream 1.0 M 2026-03-09T15:17:35.298 INFO:teuthology.orchestra.run.vm03.stdout: python3-pycparser noarch 2.20-6.el9 @baseos 745 k 2026-03-09T15:17:35.298 INFO:teuthology.orchestra.run.vm03.stdout: python3-pysocks noarch 1.7.1-12.el9 @anaconda 88 k 2026-03-09T15:17:35.298 INFO:teuthology.orchestra.run.vm03.stdout: python3-pytz noarch 2021.1-5.el9 @koji-override-0 176 k 2026-03-09T15:17:35.298 INFO:teuthology.orchestra.run.vm03.stdout: python3-repoze-lru noarch 0.7-16.el9 @epel 83 k 2026-03-09T15:17:35.298 INFO:teuthology.orchestra.run.vm03.stdout: python3-requests noarch 2.25.1-10.el9 @baseos 405 k 2026-03-09T15:17:35.298 INFO:teuthology.orchestra.run.vm03.stdout: python3-requests-oauthlib noarch 1.3.0-12.el9 @appstream 119 k 2026-03-09T15:17:35.298 INFO:teuthology.orchestra.run.vm03.stdout: python3-routes noarch 2.5.1-5.el9 @epel 459 k 2026-03-09T15:17:35.298 INFO:teuthology.orchestra.run.vm03.stdout: python3-rsa noarch 4.9-2.el9 @epel 202 k 2026-03-09T15:17:35.298 INFO:teuthology.orchestra.run.vm03.stdout: python3-scipy x86_64 1.9.3-2.el9 @appstream 76 M 2026-03-09T15:17:35.298 INFO:teuthology.orchestra.run.vm03.stdout: python3-tempora noarch 5.0.0-2.el9 @epel 96 k 2026-03-09T15:17:35.298 INFO:teuthology.orchestra.run.vm03.stdout: python3-toml noarch 0.10.2-6.el9 @appstream 99 k 2026-03-09T15:17:35.298 INFO:teuthology.orchestra.run.vm03.stdout: python3-typing-extensions noarch 4.15.0-1.el9 @epel 447 k 2026-03-09T15:17:35.298 INFO:teuthology.orchestra.run.vm03.stdout: python3-urllib3 noarch 1.26.5-7.el9 @baseos 746 k 2026-03-09T15:17:35.298 INFO:teuthology.orchestra.run.vm03.stdout: python3-webob noarch 1.8.8-2.el9 @epel 1.2 M 2026-03-09T15:17:35.298 INFO:teuthology.orchestra.run.vm03.stdout: python3-websocket-client noarch 1.2.3-2.el9 @epel 319 k 2026-03-09T15:17:35.298 INFO:teuthology.orchestra.run.vm03.stdout: python3-werkzeug noarch 2.0.3-3.el9.1 @epel 1.9 M 2026-03-09T15:17:35.298 INFO:teuthology.orchestra.run.vm03.stdout: python3-zc-lockfile noarch 2.0-10.el9 @epel 35 k 2026-03-09T15:17:35.298 INFO:teuthology.orchestra.run.vm03.stdout: qatlib x86_64 25.08.0-2.el9 @appstream 639 k 2026-03-09T15:17:35.298 INFO:teuthology.orchestra.run.vm03.stdout: qatlib-service x86_64 25.08.0-2.el9 @appstream 69 k 2026-03-09T15:17:35.298 INFO:teuthology.orchestra.run.vm03.stdout: qatzip-libs x86_64 1.3.1-1.el9 @appstream 148 k 2026-03-09T15:17:35.298 INFO:teuthology.orchestra.run.vm03.stdout: 2026-03-09T15:17:35.298 INFO:teuthology.orchestra.run.vm03.stdout:Transaction Summary 2026-03-09T15:17:35.298 INFO:teuthology.orchestra.run.vm03.stdout:=========================================================================================== 2026-03-09T15:17:35.298 INFO:teuthology.orchestra.run.vm03.stdout:Remove 102 Packages 2026-03-09T15:17:35.298 INFO:teuthology.orchestra.run.vm03.stdout: 2026-03-09T15:17:35.298 INFO:teuthology.orchestra.run.vm03.stdout:Freed space: 613 M 2026-03-09T15:17:35.298 INFO:teuthology.orchestra.run.vm03.stdout:Running transaction check 2026-03-09T15:17:35.312 INFO:teuthology.orchestra.run.vm04.stdout:Transaction check succeeded. 2026-03-09T15:17:35.312 INFO:teuthology.orchestra.run.vm04.stdout:Running transaction test 2026-03-09T15:17:35.313 INFO:teuthology.orchestra.run.vm06.stdout:Transaction check succeeded. 2026-03-09T15:17:35.313 INFO:teuthology.orchestra.run.vm06.stdout:Running transaction test 2026-03-09T15:17:35.324 INFO:teuthology.orchestra.run.vm03.stdout:Transaction check succeeded. 2026-03-09T15:17:35.324 INFO:teuthology.orchestra.run.vm03.stdout:Running transaction test 2026-03-09T15:17:35.418 INFO:teuthology.orchestra.run.vm04.stdout:Transaction test succeeded. 2026-03-09T15:17:35.418 INFO:teuthology.orchestra.run.vm04.stdout:Running transaction 2026-03-09T15:17:35.418 INFO:teuthology.orchestra.run.vm06.stdout:Transaction test succeeded. 2026-03-09T15:17:35.418 INFO:teuthology.orchestra.run.vm06.stdout:Running transaction 2026-03-09T15:17:35.430 INFO:teuthology.orchestra.run.vm03.stdout:Transaction test succeeded. 2026-03-09T15:17:35.430 INFO:teuthology.orchestra.run.vm03.stdout:Running transaction 2026-03-09T15:17:35.557 INFO:teuthology.orchestra.run.vm04.stdout: Preparing : 1/1 2026-03-09T15:17:35.557 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : ceph-mgr-rook-2:19.2.3-678.ge911bdeb.el9.noarch 1/102 2026-03-09T15:17:35.558 INFO:teuthology.orchestra.run.vm06.stdout: Preparing : 1/1 2026-03-09T15:17:35.558 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : ceph-mgr-rook-2:19.2.3-678.ge911bdeb.el9.noarch 1/102 2026-03-09T15:17:35.564 INFO:teuthology.orchestra.run.vm04.stdout: Running scriptlet: ceph-mgr-rook-2:19.2.3-678.ge911bdeb.el9.noarch 1/102 2026-03-09T15:17:35.565 INFO:teuthology.orchestra.run.vm06.stdout: Running scriptlet: ceph-mgr-rook-2:19.2.3-678.ge911bdeb.el9.noarch 1/102 2026-03-09T15:17:35.570 INFO:teuthology.orchestra.run.vm03.stdout: Preparing : 1/1 2026-03-09T15:17:35.570 INFO:teuthology.orchestra.run.vm03.stdout: Erasing : ceph-mgr-rook-2:19.2.3-678.ge911bdeb.el9.noarch 1/102 2026-03-09T15:17:35.577 INFO:teuthology.orchestra.run.vm03.stdout: Running scriptlet: ceph-mgr-rook-2:19.2.3-678.ge911bdeb.el9.noarch 1/102 2026-03-09T15:17:35.581 INFO:teuthology.orchestra.run.vm06.stdout: Running scriptlet: ceph-mgr-2:19.2.3-678.ge911bdeb.el9.x86_64 2/102 2026-03-09T15:17:35.581 INFO:teuthology.orchestra.run.vm06.stdout:Glob pattern passed to enable, but globs are not supported for this. 2026-03-09T15:17:35.581 INFO:teuthology.orchestra.run.vm06.stdout:Invalid unit name "ceph-mgr@*.service" escaped as "ceph-mgr@\x2a.service". 2026-03-09T15:17:35.581 INFO:teuthology.orchestra.run.vm06.stdout:Removed "/etc/systemd/system/multi-user.target.wants/ceph-mgr.target". 2026-03-09T15:17:35.582 INFO:teuthology.orchestra.run.vm06.stdout:Removed "/etc/systemd/system/ceph.target.wants/ceph-mgr.target". 2026-03-09T15:17:35.582 INFO:teuthology.orchestra.run.vm06.stdout: 2026-03-09T15:17:35.582 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : ceph-mgr-2:19.2.3-678.ge911bdeb.el9.x86_64 2/102 2026-03-09T15:17:35.583 INFO:teuthology.orchestra.run.vm04.stdout: Running scriptlet: ceph-mgr-2:19.2.3-678.ge911bdeb.el9.x86_64 2/102 2026-03-09T15:17:35.583 INFO:teuthology.orchestra.run.vm04.stdout:Glob pattern passed to enable, but globs are not supported for this. 2026-03-09T15:17:35.583 INFO:teuthology.orchestra.run.vm04.stdout:Invalid unit name "ceph-mgr@*.service" escaped as "ceph-mgr@\x2a.service". 2026-03-09T15:17:35.583 INFO:teuthology.orchestra.run.vm04.stdout:Removed "/etc/systemd/system/multi-user.target.wants/ceph-mgr.target". 2026-03-09T15:17:35.583 INFO:teuthology.orchestra.run.vm04.stdout:Removed "/etc/systemd/system/ceph.target.wants/ceph-mgr.target". 2026-03-09T15:17:35.583 INFO:teuthology.orchestra.run.vm04.stdout: 2026-03-09T15:17:35.584 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : ceph-mgr-2:19.2.3-678.ge911bdeb.el9.x86_64 2/102 2026-03-09T15:17:35.595 INFO:teuthology.orchestra.run.vm06.stdout: Running scriptlet: ceph-mgr-2:19.2.3-678.ge911bdeb.el9.x86_64 2/102 2026-03-09T15:17:35.596 INFO:teuthology.orchestra.run.vm03.stdout: Running scriptlet: ceph-mgr-2:19.2.3-678.ge911bdeb.el9.x86_64 2/102 2026-03-09T15:17:35.596 INFO:teuthology.orchestra.run.vm03.stdout:Glob pattern passed to enable, but globs are not supported for this. 2026-03-09T15:17:35.596 INFO:teuthology.orchestra.run.vm03.stdout:Invalid unit name "ceph-mgr@*.service" escaped as "ceph-mgr@\x2a.service". 2026-03-09T15:17:35.596 INFO:teuthology.orchestra.run.vm03.stdout:Removed "/etc/systemd/system/multi-user.target.wants/ceph-mgr.target". 2026-03-09T15:17:35.596 INFO:teuthology.orchestra.run.vm03.stdout:Removed "/etc/systemd/system/ceph.target.wants/ceph-mgr.target". 2026-03-09T15:17:35.596 INFO:teuthology.orchestra.run.vm03.stdout: 2026-03-09T15:17:35.597 INFO:teuthology.orchestra.run.vm03.stdout: Erasing : ceph-mgr-2:19.2.3-678.ge911bdeb.el9.x86_64 2/102 2026-03-09T15:17:35.597 INFO:teuthology.orchestra.run.vm04.stdout: Running scriptlet: ceph-mgr-2:19.2.3-678.ge911bdeb.el9.x86_64 2/102 2026-03-09T15:17:35.610 INFO:teuthology.orchestra.run.vm03.stdout: Running scriptlet: ceph-mgr-2:19.2.3-678.ge911bdeb.el9.x86_64 2/102 2026-03-09T15:17:35.619 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : ceph-mgr-modules-core-2:19.2.3-678.ge911bdeb.el9 3/102 2026-03-09T15:17:35.619 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : ceph-mgr-dashboard-2:19.2.3-678.ge911bdeb.el9.no 4/102 2026-03-09T15:17:35.621 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : ceph-mgr-modules-core-2:19.2.3-678.ge911bdeb.el9 3/102 2026-03-09T15:17:35.621 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : ceph-mgr-dashboard-2:19.2.3-678.ge911bdeb.el9.no 4/102 2026-03-09T15:17:35.633 INFO:teuthology.orchestra.run.vm03.stdout: Erasing : ceph-mgr-modules-core-2:19.2.3-678.ge911bdeb.el9 3/102 2026-03-09T15:17:35.633 INFO:teuthology.orchestra.run.vm03.stdout: Erasing : ceph-mgr-dashboard-2:19.2.3-678.ge911bdeb.el9.no 4/102 2026-03-09T15:17:35.673 INFO:teuthology.orchestra.run.vm06.stdout: Running scriptlet: ceph-mgr-dashboard-2:19.2.3-678.ge911bdeb.el9.no 4/102 2026-03-09T15:17:35.675 INFO:teuthology.orchestra.run.vm04.stdout: Running scriptlet: ceph-mgr-dashboard-2:19.2.3-678.ge911bdeb.el9.no 4/102 2026-03-09T15:17:35.681 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : python3-kubernetes-1:26.1.0-3.el9.noarch 5/102 2026-03-09T15:17:35.683 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : python3-kubernetes-1:26.1.0-3.el9.noarch 5/102 2026-03-09T15:17:35.685 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : python3-requests-oauthlib-1.3.0-12.el9.noarch 6/102 2026-03-09T15:17:35.685 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : ceph-mgr-cephadm-2:19.2.3-678.ge911bdeb.el9.noar 7/102 2026-03-09T15:17:35.687 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : python3-requests-oauthlib-1.3.0-12.el9.noarch 6/102 2026-03-09T15:17:35.687 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : ceph-mgr-cephadm-2:19.2.3-678.ge911bdeb.el9.noar 7/102 2026-03-09T15:17:35.688 INFO:teuthology.orchestra.run.vm03.stdout: Running scriptlet: ceph-mgr-dashboard-2:19.2.3-678.ge911bdeb.el9.no 4/102 2026-03-09T15:17:35.696 INFO:teuthology.orchestra.run.vm06.stdout: Running scriptlet: ceph-mgr-cephadm-2:19.2.3-678.ge911bdeb.el9.noar 7/102 2026-03-09T15:17:35.697 INFO:teuthology.orchestra.run.vm03.stdout: Erasing : python3-kubernetes-1:26.1.0-3.el9.noarch 5/102 2026-03-09T15:17:35.698 INFO:teuthology.orchestra.run.vm04.stdout: Running scriptlet: ceph-mgr-cephadm-2:19.2.3-678.ge911bdeb.el9.noar 7/102 2026-03-09T15:17:35.701 INFO:teuthology.orchestra.run.vm03.stdout: Erasing : python3-requests-oauthlib-1.3.0-12.el9.noarch 6/102 2026-03-09T15:17:35.701 INFO:teuthology.orchestra.run.vm03.stdout: Erasing : ceph-mgr-cephadm-2:19.2.3-678.ge911bdeb.el9.noar 7/102 2026-03-09T15:17:35.702 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : python3-cherrypy-18.6.1-2.el9.noarch 8/102 2026-03-09T15:17:35.705 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : python3-cherrypy-18.6.1-2.el9.noarch 8/102 2026-03-09T15:17:35.706 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : python3-cheroot-10.0.1-4.el9.noarch 9/102 2026-03-09T15:17:35.709 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : python3-cheroot-10.0.1-4.el9.noarch 9/102 2026-03-09T15:17:35.712 INFO:teuthology.orchestra.run.vm03.stdout: Running scriptlet: ceph-mgr-cephadm-2:19.2.3-678.ge911bdeb.el9.noar 7/102 2026-03-09T15:17:35.714 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : python3-grpcio-tools-1.46.7-10.el9.x86_64 10/102 2026-03-09T15:17:35.717 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : python3-grpcio-tools-1.46.7-10.el9.x86_64 10/102 2026-03-09T15:17:35.718 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : python3-grpcio-1.46.7-10.el9.x86_64 11/102 2026-03-09T15:17:35.719 INFO:teuthology.orchestra.run.vm03.stdout: Erasing : python3-cherrypy-18.6.1-2.el9.noarch 8/102 2026-03-09T15:17:35.722 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : python3-grpcio-1.46.7-10.el9.x86_64 11/102 2026-03-09T15:17:35.723 INFO:teuthology.orchestra.run.vm03.stdout: Erasing : python3-cheroot-10.0.1-4.el9.noarch 9/102 2026-03-09T15:17:35.731 INFO:teuthology.orchestra.run.vm03.stdout: Erasing : python3-grpcio-tools-1.46.7-10.el9.x86_64 10/102 2026-03-09T15:17:35.735 INFO:teuthology.orchestra.run.vm03.stdout: Erasing : python3-grpcio-1.46.7-10.el9.x86_64 11/102 2026-03-09T15:17:35.738 INFO:teuthology.orchestra.run.vm06.stdout: Running scriptlet: ceph-osd-2:19.2.3-678.ge911bdeb.el9.x86_64 12/102 2026-03-09T15:17:35.738 INFO:teuthology.orchestra.run.vm06.stdout:Glob pattern passed to enable, but globs are not supported for this. 2026-03-09T15:17:35.738 INFO:teuthology.orchestra.run.vm06.stdout:Invalid unit name "ceph-osd@*.service" escaped as "ceph-osd@\x2a.service". 2026-03-09T15:17:35.738 INFO:teuthology.orchestra.run.vm06.stdout:Removed "/etc/systemd/system/multi-user.target.wants/ceph-osd.target". 2026-03-09T15:17:35.738 INFO:teuthology.orchestra.run.vm06.stdout:Removed "/etc/systemd/system/ceph.target.wants/ceph-osd.target". 2026-03-09T15:17:35.738 INFO:teuthology.orchestra.run.vm06.stdout: 2026-03-09T15:17:35.743 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : ceph-osd-2:19.2.3-678.ge911bdeb.el9.x86_64 12/102 2026-03-09T15:17:35.745 INFO:teuthology.orchestra.run.vm04.stdout: Running scriptlet: ceph-osd-2:19.2.3-678.ge911bdeb.el9.x86_64 12/102 2026-03-09T15:17:35.745 INFO:teuthology.orchestra.run.vm04.stdout:Glob pattern passed to enable, but globs are not supported for this. 2026-03-09T15:17:35.745 INFO:teuthology.orchestra.run.vm04.stdout:Invalid unit name "ceph-osd@*.service" escaped as "ceph-osd@\x2a.service". 2026-03-09T15:17:35.745 INFO:teuthology.orchestra.run.vm04.stdout:Removed "/etc/systemd/system/multi-user.target.wants/ceph-osd.target". 2026-03-09T15:17:35.745 INFO:teuthology.orchestra.run.vm04.stdout:Removed "/etc/systemd/system/ceph.target.wants/ceph-osd.target". 2026-03-09T15:17:35.745 INFO:teuthology.orchestra.run.vm04.stdout: 2026-03-09T15:17:35.751 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : ceph-osd-2:19.2.3-678.ge911bdeb.el9.x86_64 12/102 2026-03-09T15:17:35.751 INFO:teuthology.orchestra.run.vm06.stdout: Running scriptlet: ceph-osd-2:19.2.3-678.ge911bdeb.el9.x86_64 12/102 2026-03-09T15:17:35.756 INFO:teuthology.orchestra.run.vm03.stdout: Running scriptlet: ceph-osd-2:19.2.3-678.ge911bdeb.el9.x86_64 12/102 2026-03-09T15:17:35.756 INFO:teuthology.orchestra.run.vm03.stdout:Glob pattern passed to enable, but globs are not supported for this. 2026-03-09T15:17:35.756 INFO:teuthology.orchestra.run.vm03.stdout:Invalid unit name "ceph-osd@*.service" escaped as "ceph-osd@\x2a.service". 2026-03-09T15:17:35.756 INFO:teuthology.orchestra.run.vm03.stdout:Removed "/etc/systemd/system/multi-user.target.wants/ceph-osd.target". 2026-03-09T15:17:35.756 INFO:teuthology.orchestra.run.vm03.stdout:Removed "/etc/systemd/system/ceph.target.wants/ceph-osd.target". 2026-03-09T15:17:35.756 INFO:teuthology.orchestra.run.vm03.stdout: 2026-03-09T15:17:35.759 INFO:teuthology.orchestra.run.vm04.stdout: Running scriptlet: ceph-osd-2:19.2.3-678.ge911bdeb.el9.x86_64 12/102 2026-03-09T15:17:35.761 INFO:teuthology.orchestra.run.vm03.stdout: Erasing : ceph-osd-2:19.2.3-678.ge911bdeb.el9.x86_64 12/102 2026-03-09T15:17:35.767 INFO:teuthology.orchestra.run.vm06.stdout: Running scriptlet: ceph-volume-2:19.2.3-678.ge911bdeb.el9.noarch 13/102 2026-03-09T15:17:35.767 INFO:teuthology.orchestra.run.vm06.stdout:Glob pattern passed to enable, but globs are not supported for this. 2026-03-09T15:17:35.767 INFO:teuthology.orchestra.run.vm06.stdout:Invalid unit name "ceph-volume@*.service" escaped as "ceph-volume@\x2a.service". 2026-03-09T15:17:35.767 INFO:teuthology.orchestra.run.vm06.stdout: 2026-03-09T15:17:35.769 INFO:teuthology.orchestra.run.vm03.stdout: Running scriptlet: ceph-osd-2:19.2.3-678.ge911bdeb.el9.x86_64 12/102 2026-03-09T15:17:35.774 INFO:teuthology.orchestra.run.vm04.stdout: Running scriptlet: ceph-volume-2:19.2.3-678.ge911bdeb.el9.noarch 13/102 2026-03-09T15:17:35.774 INFO:teuthology.orchestra.run.vm04.stdout:Glob pattern passed to enable, but globs are not supported for this. 2026-03-09T15:17:35.774 INFO:teuthology.orchestra.run.vm04.stdout:Invalid unit name "ceph-volume@*.service" escaped as "ceph-volume@\x2a.service". 2026-03-09T15:17:35.774 INFO:teuthology.orchestra.run.vm04.stdout: 2026-03-09T15:17:35.775 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : ceph-volume-2:19.2.3-678.ge911bdeb.el9.noarch 13/102 2026-03-09T15:17:35.782 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : ceph-volume-2:19.2.3-678.ge911bdeb.el9.noarch 13/102 2026-03-09T15:17:35.784 INFO:teuthology.orchestra.run.vm06.stdout: Running scriptlet: ceph-volume-2:19.2.3-678.ge911bdeb.el9.noarch 13/102 2026-03-09T15:17:35.785 INFO:teuthology.orchestra.run.vm03.stdout: Running scriptlet: ceph-volume-2:19.2.3-678.ge911bdeb.el9.noarch 13/102 2026-03-09T15:17:35.785 INFO:teuthology.orchestra.run.vm03.stdout:Glob pattern passed to enable, but globs are not supported for this. 2026-03-09T15:17:35.785 INFO:teuthology.orchestra.run.vm03.stdout:Invalid unit name "ceph-volume@*.service" escaped as "ceph-volume@\x2a.service". 2026-03-09T15:17:35.785 INFO:teuthology.orchestra.run.vm03.stdout: 2026-03-09T15:17:35.786 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : python3-jaraco-collections-3.0.0-8.el9.noarch 14/102 2026-03-09T15:17:35.791 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : python3-jaraco-text-4.0.0-2.el9.noarch 15/102 2026-03-09T15:17:35.791 INFO:teuthology.orchestra.run.vm04.stdout: Running scriptlet: ceph-volume-2:19.2.3-678.ge911bdeb.el9.noarch 13/102 2026-03-09T15:17:35.793 INFO:teuthology.orchestra.run.vm03.stdout: Erasing : ceph-volume-2:19.2.3-678.ge911bdeb.el9.noarch 13/102 2026-03-09T15:17:35.793 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : python3-jaraco-collections-3.0.0-8.el9.noarch 14/102 2026-03-09T15:17:35.795 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : python3-jinja2-2.11.3-8.el9.noarch 16/102 2026-03-09T15:17:35.798 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : python3-jaraco-text-4.0.0-2.el9.noarch 15/102 2026-03-09T15:17:35.802 INFO:teuthology.orchestra.run.vm03.stdout: Running scriptlet: ceph-volume-2:19.2.3-678.ge911bdeb.el9.noarch 13/102 2026-03-09T15:17:35.803 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : python3-jinja2-2.11.3-8.el9.noarch 16/102 2026-03-09T15:17:35.804 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : python3-requests-2.25.1-10.el9.noarch 17/102 2026-03-09T15:17:35.805 INFO:teuthology.orchestra.run.vm03.stdout: Erasing : python3-jaraco-collections-3.0.0-8.el9.noarch 14/102 2026-03-09T15:17:35.810 INFO:teuthology.orchestra.run.vm03.stdout: Erasing : python3-jaraco-text-4.0.0-2.el9.noarch 15/102 2026-03-09T15:17:35.812 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : python3-requests-2.25.1-10.el9.noarch 17/102 2026-03-09T15:17:35.814 INFO:teuthology.orchestra.run.vm03.stdout: Erasing : python3-jinja2-2.11.3-8.el9.noarch 16/102 2026-03-09T15:17:35.816 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : python3-google-auth-1:2.45.0-1.el9.noarch 18/102 2026-03-09T15:17:35.821 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : python3-pecan-1.4.2-3.el9.noarch 19/102 2026-03-09T15:17:35.823 INFO:teuthology.orchestra.run.vm03.stdout: Erasing : python3-requests-2.25.1-10.el9.noarch 17/102 2026-03-09T15:17:35.825 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : python3-google-auth-1:2.45.0-1.el9.noarch 18/102 2026-03-09T15:17:35.830 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : python3-rsa-4.9-2.el9.noarch 20/102 2026-03-09T15:17:35.830 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : python3-pecan-1.4.2-3.el9.noarch 19/102 2026-03-09T15:17:35.835 INFO:teuthology.orchestra.run.vm03.stdout: Erasing : python3-google-auth-1:2.45.0-1.el9.noarch 18/102 2026-03-09T15:17:35.836 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : python3-pyasn1-modules-0.4.8-7.el9.noarch 21/102 2026-03-09T15:17:35.840 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : python3-rsa-4.9-2.el9.noarch 20/102 2026-03-09T15:17:35.840 INFO:teuthology.orchestra.run.vm03.stdout: Erasing : python3-pecan-1.4.2-3.el9.noarch 19/102 2026-03-09T15:17:35.847 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : python3-pyasn1-modules-0.4.8-7.el9.noarch 21/102 2026-03-09T15:17:35.850 INFO:teuthology.orchestra.run.vm03.stdout: Erasing : python3-rsa-4.9-2.el9.noarch 20/102 2026-03-09T15:17:35.856 INFO:teuthology.orchestra.run.vm03.stdout: Erasing : python3-pyasn1-modules-0.4.8-7.el9.noarch 21/102 2026-03-09T15:17:35.865 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : python3-urllib3-1.26.5-7.el9.noarch 22/102 2026-03-09T15:17:35.871 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : python3-babel-2.9.1-2.el9.noarch 23/102 2026-03-09T15:17:35.874 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : python3-jaraco-classes-3.2.1-5.el9.noarch 24/102 2026-03-09T15:17:35.877 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : python3-urllib3-1.26.5-7.el9.noarch 22/102 2026-03-09T15:17:35.882 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : python3-pyOpenSSL-21.0.0-1.el9.noarch 25/102 2026-03-09T15:17:35.883 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : python3-babel-2.9.1-2.el9.noarch 23/102 2026-03-09T15:17:35.886 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : python3-jaraco-classes-3.2.1-5.el9.noarch 24/102 2026-03-09T15:17:35.886 INFO:teuthology.orchestra.run.vm03.stdout: Erasing : python3-urllib3-1.26.5-7.el9.noarch 22/102 2026-03-09T15:17:35.893 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : python3-asyncssh-2.13.2-5.el9.noarch 26/102 2026-03-09T15:17:35.893 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : ceph-mgr-diskprediction-local-2:19.2.3-678.ge911 27/102 2026-03-09T15:17:35.895 INFO:teuthology.orchestra.run.vm03.stdout: Erasing : python3-babel-2.9.1-2.el9.noarch 23/102 2026-03-09T15:17:35.895 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : python3-pyOpenSSL-21.0.0-1.el9.noarch 25/102 2026-03-09T15:17:35.897 INFO:teuthology.orchestra.run.vm03.stdout: Erasing : python3-jaraco-classes-3.2.1-5.el9.noarch 24/102 2026-03-09T15:17:35.900 INFO:teuthology.orchestra.run.vm06.stdout: Running scriptlet: ceph-mgr-diskprediction-local-2:19.2.3-678.ge911 27/102 2026-03-09T15:17:35.906 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : python3-asyncssh-2.13.2-5.el9.noarch 26/102 2026-03-09T15:17:35.906 INFO:teuthology.orchestra.run.vm03.stdout: Erasing : python3-pyOpenSSL-21.0.0-1.el9.noarch 25/102 2026-03-09T15:17:35.906 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : ceph-mgr-diskprediction-local-2:19.2.3-678.ge911 27/102 2026-03-09T15:17:35.913 INFO:teuthology.orchestra.run.vm04.stdout: Running scriptlet: ceph-mgr-diskprediction-local-2:19.2.3-678.ge911 27/102 2026-03-09T15:17:35.916 INFO:teuthology.orchestra.run.vm03.stdout: Erasing : python3-asyncssh-2.13.2-5.el9.noarch 26/102 2026-03-09T15:17:35.917 INFO:teuthology.orchestra.run.vm03.stdout: Erasing : ceph-mgr-diskprediction-local-2:19.2.3-678.ge911 27/102 2026-03-09T15:17:35.924 INFO:teuthology.orchestra.run.vm03.stdout: Running scriptlet: ceph-mgr-diskprediction-local-2:19.2.3-678.ge911 27/102 2026-03-09T15:17:35.994 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : python3-jsonpatch-1.21-16.el9.noarch 28/102 2026-03-09T15:17:36.009 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : python3-scipy-1.9.3-2.el9.x86_64 29/102 2026-03-09T15:17:36.009 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : python3-jsonpatch-1.21-16.el9.noarch 28/102 2026-03-09T15:17:36.017 INFO:teuthology.orchestra.run.vm03.stdout: Erasing : python3-jsonpatch-1.21-16.el9.noarch 28/102 2026-03-09T15:17:36.022 INFO:teuthology.orchestra.run.vm06.stdout: Running scriptlet: libstoragemgmt-1.10.1-1.el9.x86_64 30/102 2026-03-09T15:17:36.022 INFO:teuthology.orchestra.run.vm06.stdout:Removed "/etc/systemd/system/multi-user.target.wants/libstoragemgmt.service". 2026-03-09T15:17:36.022 INFO:teuthology.orchestra.run.vm06.stdout: 2026-03-09T15:17:36.023 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : libstoragemgmt-1.10.1-1.el9.x86_64 30/102 2026-03-09T15:17:36.025 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : python3-scipy-1.9.3-2.el9.x86_64 29/102 2026-03-09T15:17:36.032 INFO:teuthology.orchestra.run.vm03.stdout: Erasing : python3-scipy-1.9.3-2.el9.x86_64 29/102 2026-03-09T15:17:36.039 INFO:teuthology.orchestra.run.vm04.stdout: Running scriptlet: libstoragemgmt-1.10.1-1.el9.x86_64 30/102 2026-03-09T15:17:36.039 INFO:teuthology.orchestra.run.vm04.stdout:Removed "/etc/systemd/system/multi-user.target.wants/libstoragemgmt.service". 2026-03-09T15:17:36.039 INFO:teuthology.orchestra.run.vm04.stdout: 2026-03-09T15:17:36.040 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : libstoragemgmt-1.10.1-1.el9.x86_64 30/102 2026-03-09T15:17:36.045 INFO:teuthology.orchestra.run.vm03.stdout: Running scriptlet: libstoragemgmt-1.10.1-1.el9.x86_64 30/102 2026-03-09T15:17:36.045 INFO:teuthology.orchestra.run.vm03.stdout:Removed "/etc/systemd/system/multi-user.target.wants/libstoragemgmt.service". 2026-03-09T15:17:36.045 INFO:teuthology.orchestra.run.vm03.stdout: 2026-03-09T15:17:36.046 INFO:teuthology.orchestra.run.vm03.stdout: Erasing : libstoragemgmt-1.10.1-1.el9.x86_64 30/102 2026-03-09T15:17:36.049 INFO:teuthology.orchestra.run.vm06.stdout: Running scriptlet: libstoragemgmt-1.10.1-1.el9.x86_64 30/102 2026-03-09T15:17:36.066 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : python3-numpy-f2py-1:1.23.5-2.el9.x86_64 31/102 2026-03-09T15:17:36.067 INFO:teuthology.orchestra.run.vm04.stdout: Running scriptlet: libstoragemgmt-1.10.1-1.el9.x86_64 30/102 2026-03-09T15:17:36.071 INFO:teuthology.orchestra.run.vm03.stdout: Running scriptlet: libstoragemgmt-1.10.1-1.el9.x86_64 30/102 2026-03-09T15:17:36.072 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : python3-cryptography-36.0.1-5.el9.x86_64 32/102 2026-03-09T15:17:36.074 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : protobuf-compiler-3.14.0-17.el9.x86_64 33/102 2026-03-09T15:17:36.076 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : python3-bcrypt-3.2.2-1.el9.x86_64 34/102 2026-03-09T15:17:36.083 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : python3-numpy-f2py-1:1.23.5-2.el9.x86_64 31/102 2026-03-09T15:17:36.086 INFO:teuthology.orchestra.run.vm03.stdout: Erasing : python3-numpy-f2py-1:1.23.5-2.el9.x86_64 31/102 2026-03-09T15:17:36.089 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : python3-cryptography-36.0.1-5.el9.x86_64 32/102 2026-03-09T15:17:36.091 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : protobuf-compiler-3.14.0-17.el9.x86_64 33/102 2026-03-09T15:17:36.092 INFO:teuthology.orchestra.run.vm03.stdout: Erasing : python3-cryptography-36.0.1-5.el9.x86_64 32/102 2026-03-09T15:17:36.094 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : python3-bcrypt-3.2.2-1.el9.x86_64 34/102 2026-03-09T15:17:36.094 INFO:teuthology.orchestra.run.vm03.stdout: Erasing : protobuf-compiler-3.14.0-17.el9.x86_64 33/102 2026-03-09T15:17:36.096 INFO:teuthology.orchestra.run.vm03.stdout: Erasing : python3-bcrypt-3.2.2-1.el9.x86_64 34/102 2026-03-09T15:17:36.097 INFO:teuthology.orchestra.run.vm06.stdout: Running scriptlet: rbd-mirror-2:19.2.3-678.ge911bdeb.el9.x86_64 35/102 2026-03-09T15:17:36.097 INFO:teuthology.orchestra.run.vm06.stdout:Glob pattern passed to enable, but globs are not supported for this. 2026-03-09T15:17:36.097 INFO:teuthology.orchestra.run.vm06.stdout:Invalid unit name "ceph-rbd-mirror@*.service" escaped as "ceph-rbd-mirror@\x2a.service". 2026-03-09T15:17:36.097 INFO:teuthology.orchestra.run.vm06.stdout:Removed "/etc/systemd/system/multi-user.target.wants/ceph-rbd-mirror.target". 2026-03-09T15:17:36.097 INFO:teuthology.orchestra.run.vm06.stdout:Removed "/etc/systemd/system/ceph.target.wants/ceph-rbd-mirror.target". 2026-03-09T15:17:36.097 INFO:teuthology.orchestra.run.vm06.stdout: 2026-03-09T15:17:36.098 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : rbd-mirror-2:19.2.3-678.ge911bdeb.el9.x86_64 35/102 2026-03-09T15:17:36.110 INFO:teuthology.orchestra.run.vm06.stdout: Running scriptlet: rbd-mirror-2:19.2.3-678.ge911bdeb.el9.x86_64 35/102 2026-03-09T15:17:36.112 INFO:teuthology.orchestra.run.vm04.stdout: Running scriptlet: rbd-mirror-2:19.2.3-678.ge911bdeb.el9.x86_64 35/102 2026-03-09T15:17:36.112 INFO:teuthology.orchestra.run.vm04.stdout:Glob pattern passed to enable, but globs are not supported for this. 2026-03-09T15:17:36.112 INFO:teuthology.orchestra.run.vm04.stdout:Invalid unit name "ceph-rbd-mirror@*.service" escaped as "ceph-rbd-mirror@\x2a.service". 2026-03-09T15:17:36.112 INFO:teuthology.orchestra.run.vm04.stdout:Removed "/etc/systemd/system/multi-user.target.wants/ceph-rbd-mirror.target". 2026-03-09T15:17:36.112 INFO:teuthology.orchestra.run.vm04.stdout:Removed "/etc/systemd/system/ceph.target.wants/ceph-rbd-mirror.target". 2026-03-09T15:17:36.112 INFO:teuthology.orchestra.run.vm04.stdout: 2026-03-09T15:17:36.113 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : rbd-mirror-2:19.2.3-678.ge911bdeb.el9.x86_64 35/102 2026-03-09T15:17:36.113 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : python3-mako-1.1.4-6.el9.noarch 36/102 2026-03-09T15:17:36.115 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : python3-jaraco-context-6.0.1-3.el9.noarch 37/102 2026-03-09T15:17:36.117 INFO:teuthology.orchestra.run.vm03.stdout: Running scriptlet: rbd-mirror-2:19.2.3-678.ge911bdeb.el9.x86_64 35/102 2026-03-09T15:17:36.117 INFO:teuthology.orchestra.run.vm03.stdout:Glob pattern passed to enable, but globs are not supported for this. 2026-03-09T15:17:36.117 INFO:teuthology.orchestra.run.vm03.stdout:Invalid unit name "ceph-rbd-mirror@*.service" escaped as "ceph-rbd-mirror@\x2a.service". 2026-03-09T15:17:36.117 INFO:teuthology.orchestra.run.vm03.stdout:Removed "/etc/systemd/system/multi-user.target.wants/ceph-rbd-mirror.target". 2026-03-09T15:17:36.117 INFO:teuthology.orchestra.run.vm03.stdout:Removed "/etc/systemd/system/ceph.target.wants/ceph-rbd-mirror.target". 2026-03-09T15:17:36.117 INFO:teuthology.orchestra.run.vm03.stdout: 2026-03-09T15:17:36.118 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : python3-portend-3.1.0-2.el9.noarch 38/102 2026-03-09T15:17:36.118 INFO:teuthology.orchestra.run.vm03.stdout: Erasing : rbd-mirror-2:19.2.3-678.ge911bdeb.el9.x86_64 35/102 2026-03-09T15:17:36.120 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : python3-tempora-5.0.0-2.el9.noarch 39/102 2026-03-09T15:17:36.124 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : python3-jaraco-functools-3.5.0-2.el9.noarch 40/102 2026-03-09T15:17:36.124 INFO:teuthology.orchestra.run.vm04.stdout: Running scriptlet: rbd-mirror-2:19.2.3-678.ge911bdeb.el9.x86_64 35/102 2026-03-09T15:17:36.128 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : python3-routes-2.5.1-5.el9.noarch 41/102 2026-03-09T15:17:36.128 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : python3-mako-1.1.4-6.el9.noarch 36/102 2026-03-09T15:17:36.129 INFO:teuthology.orchestra.run.vm03.stdout: Running scriptlet: rbd-mirror-2:19.2.3-678.ge911bdeb.el9.x86_64 35/102 2026-03-09T15:17:36.130 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : python3-jaraco-context-6.0.1-3.el9.noarch 37/102 2026-03-09T15:17:36.132 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : python3-cffi-1.14.5-5.el9.x86_64 42/102 2026-03-09T15:17:36.133 INFO:teuthology.orchestra.run.vm03.stdout: Erasing : python3-mako-1.1.4-6.el9.noarch 36/102 2026-03-09T15:17:36.133 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : python3-portend-3.1.0-2.el9.noarch 38/102 2026-03-09T15:17:36.136 INFO:teuthology.orchestra.run.vm03.stdout: Erasing : python3-jaraco-context-6.0.1-3.el9.noarch 37/102 2026-03-09T15:17:36.136 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : python3-tempora-5.0.0-2.el9.noarch 39/102 2026-03-09T15:17:36.138 INFO:teuthology.orchestra.run.vm03.stdout: Erasing : python3-portend-3.1.0-2.el9.noarch 38/102 2026-03-09T15:17:36.139 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : python3-jaraco-functools-3.5.0-2.el9.noarch 40/102 2026-03-09T15:17:36.141 INFO:teuthology.orchestra.run.vm03.stdout: Erasing : python3-tempora-5.0.0-2.el9.noarch 39/102 2026-03-09T15:17:36.143 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : python3-routes-2.5.1-5.el9.noarch 41/102 2026-03-09T15:17:36.144 INFO:teuthology.orchestra.run.vm03.stdout: Erasing : python3-jaraco-functools-3.5.0-2.el9.noarch 40/102 2026-03-09T15:17:36.147 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : python3-cffi-1.14.5-5.el9.x86_64 42/102 2026-03-09T15:17:36.148 INFO:teuthology.orchestra.run.vm03.stdout: Erasing : python3-routes-2.5.1-5.el9.noarch 41/102 2026-03-09T15:17:36.151 INFO:teuthology.orchestra.run.vm03.stdout: Erasing : python3-cffi-1.14.5-5.el9.x86_64 42/102 2026-03-09T15:17:36.178 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : python3-pycparser-2.20-6.el9.noarch 43/102 2026-03-09T15:17:36.189 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : python3-numpy-1:1.23.5-2.el9.x86_64 44/102 2026-03-09T15:17:36.191 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : flexiblas-netlib-3.0.4-9.el9.x86_64 45/102 2026-03-09T15:17:36.195 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : python3-pycparser-2.20-6.el9.noarch 43/102 2026-03-09T15:17:36.196 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : flexiblas-openblas-openmp-3.0.4-9.el9.x86_64 46/102 2026-03-09T15:17:36.197 INFO:teuthology.orchestra.run.vm03.stdout: Erasing : python3-pycparser-2.20-6.el9.noarch 43/102 2026-03-09T15:17:36.198 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : openblas-openmp-0.3.29-1.el9.x86_64 47/102 2026-03-09T15:17:36.201 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : libgfortran-11.5.0-14.el9.x86_64 48/102 2026-03-09T15:17:36.205 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : python3-libstoragemgmt-1.10.1-1.el9.x86_64 49/102 2026-03-09T15:17:36.208 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : python3-numpy-1:1.23.5-2.el9.x86_64 44/102 2026-03-09T15:17:36.209 INFO:teuthology.orchestra.run.vm03.stdout: Erasing : python3-numpy-1:1.23.5-2.el9.x86_64 44/102 2026-03-09T15:17:36.211 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : flexiblas-netlib-3.0.4-9.el9.x86_64 45/102 2026-03-09T15:17:36.212 INFO:teuthology.orchestra.run.vm03.stdout: Erasing : flexiblas-netlib-3.0.4-9.el9.x86_64 45/102 2026-03-09T15:17:36.216 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : flexiblas-openblas-openmp-3.0.4-9.el9.x86_64 46/102 2026-03-09T15:17:36.216 INFO:teuthology.orchestra.run.vm03.stdout: Erasing : flexiblas-openblas-openmp-3.0.4-9.el9.x86_64 46/102 2026-03-09T15:17:36.218 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : openblas-openmp-0.3.29-1.el9.x86_64 47/102 2026-03-09T15:17:36.218 INFO:teuthology.orchestra.run.vm03.stdout: Erasing : openblas-openmp-0.3.29-1.el9.x86_64 47/102 2026-03-09T15:17:36.221 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : libgfortran-11.5.0-14.el9.x86_64 48/102 2026-03-09T15:17:36.222 INFO:teuthology.orchestra.run.vm03.stdout: Erasing : libgfortran-11.5.0-14.el9.x86_64 48/102 2026-03-09T15:17:36.223 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : python3-libstoragemgmt-1.10.1-1.el9.x86_64 49/102 2026-03-09T15:17:36.224 INFO:teuthology.orchestra.run.vm03.stdout: Erasing : python3-libstoragemgmt-1.10.1-1.el9.x86_64 49/102 2026-03-09T15:17:36.227 INFO:teuthology.orchestra.run.vm06.stdout: Running scriptlet: ceph-immutable-object-cache-2:19.2.3-678.ge911bd 50/102 2026-03-09T15:17:36.227 INFO:teuthology.orchestra.run.vm06.stdout:Glob pattern passed to enable, but globs are not supported for this. 2026-03-09T15:17:36.227 INFO:teuthology.orchestra.run.vm06.stdout:Invalid unit name "ceph-immutable-object-cache@*.service" escaped as "ceph-immutable-object-cache@\x2a.service". 2026-03-09T15:17:36.227 INFO:teuthology.orchestra.run.vm06.stdout: 2026-03-09T15:17:36.227 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : ceph-immutable-object-cache-2:19.2.3-678.ge911bd 50/102 2026-03-09T15:17:36.235 INFO:teuthology.orchestra.run.vm06.stdout: Running scriptlet: ceph-immutable-object-cache-2:19.2.3-678.ge911bd 50/102 2026-03-09T15:17:36.237 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : openblas-0.3.29-1.el9.x86_64 51/102 2026-03-09T15:17:36.239 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : flexiblas-3.0.4-9.el9.x86_64 52/102 2026-03-09T15:17:36.241 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : python3-ply-3.11-14.el9.noarch 53/102 2026-03-09T15:17:36.243 INFO:teuthology.orchestra.run.vm04.stdout: Running scriptlet: ceph-immutable-object-cache-2:19.2.3-678.ge911bd 50/102 2026-03-09T15:17:36.243 INFO:teuthology.orchestra.run.vm04.stdout:Glob pattern passed to enable, but globs are not supported for this. 2026-03-09T15:17:36.243 INFO:teuthology.orchestra.run.vm04.stdout:Invalid unit name "ceph-immutable-object-cache@*.service" escaped as "ceph-immutable-object-cache@\x2a.service". 2026-03-09T15:17:36.243 INFO:teuthology.orchestra.run.vm04.stdout: 2026-03-09T15:17:36.243 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : ceph-immutable-object-cache-2:19.2.3-678.ge911bd 50/102 2026-03-09T15:17:36.243 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : python3-repoze-lru-0.7-16.el9.noarch 54/102 2026-03-09T15:17:36.245 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : python3-jaraco-8.2.1-3.el9.noarch 55/102 2026-03-09T15:17:36.247 INFO:teuthology.orchestra.run.vm03.stdout: Running scriptlet: ceph-immutable-object-cache-2:19.2.3-678.ge911bd 50/102 2026-03-09T15:17:36.247 INFO:teuthology.orchestra.run.vm03.stdout:Glob pattern passed to enable, but globs are not supported for this. 2026-03-09T15:17:36.247 INFO:teuthology.orchestra.run.vm03.stdout:Invalid unit name "ceph-immutable-object-cache@*.service" escaped as "ceph-immutable-object-cache@\x2a.service". 2026-03-09T15:17:36.247 INFO:teuthology.orchestra.run.vm03.stdout: 2026-03-09T15:17:36.247 INFO:teuthology.orchestra.run.vm03.stdout: Erasing : ceph-immutable-object-cache-2:19.2.3-678.ge911bd 50/102 2026-03-09T15:17:36.248 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : python3-more-itertools-8.12.0-2.el9.noarch 56/102 2026-03-09T15:17:36.250 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : python3-toml-0.10.2-6.el9.noarch 57/102 2026-03-09T15:17:36.251 INFO:teuthology.orchestra.run.vm04.stdout: Running scriptlet: ceph-immutable-object-cache-2:19.2.3-678.ge911bd 50/102 2026-03-09T15:17:36.253 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : openblas-0.3.29-1.el9.x86_64 51/102 2026-03-09T15:17:36.253 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : python3-pytz-2021.1-5.el9.noarch 58/102 2026-03-09T15:17:36.255 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : flexiblas-3.0.4-9.el9.x86_64 52/102 2026-03-09T15:17:36.255 INFO:teuthology.orchestra.run.vm03.stdout: Running scriptlet: ceph-immutable-object-cache-2:19.2.3-678.ge911bd 50/102 2026-03-09T15:17:36.257 INFO:teuthology.orchestra.run.vm03.stdout: Erasing : openblas-0.3.29-1.el9.x86_64 51/102 2026-03-09T15:17:36.257 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : python3-ply-3.11-14.el9.noarch 53/102 2026-03-09T15:17:36.258 INFO:teuthology.orchestra.run.vm03.stdout: Erasing : flexiblas-3.0.4-9.el9.x86_64 52/102 2026-03-09T15:17:36.260 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : python3-repoze-lru-0.7-16.el9.noarch 54/102 2026-03-09T15:17:36.260 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : python3-backports-tarfile-1.2.0-1.el9.noarch 59/102 2026-03-09T15:17:36.261 INFO:teuthology.orchestra.run.vm03.stdout: Erasing : python3-ply-3.11-14.el9.noarch 53/102 2026-03-09T15:17:36.262 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : python3-jaraco-8.2.1-3.el9.noarch 55/102 2026-03-09T15:17:36.264 INFO:teuthology.orchestra.run.vm03.stdout: Erasing : python3-repoze-lru-0.7-16.el9.noarch 54/102 2026-03-09T15:17:36.264 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : python3-more-itertools-8.12.0-2.el9.noarch 56/102 2026-03-09T15:17:36.264 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : python3-devel-3.9.25-3.el9.x86_64 60/102 2026-03-09T15:17:36.266 INFO:teuthology.orchestra.run.vm03.stdout: Erasing : python3-jaraco-8.2.1-3.el9.noarch 55/102 2026-03-09T15:17:36.266 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : python3-jsonpointer-2.0-4.el9.noarch 61/102 2026-03-09T15:17:36.267 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : python3-toml-0.10.2-6.el9.noarch 57/102 2026-03-09T15:17:36.268 INFO:teuthology.orchestra.run.vm03.stdout: Erasing : python3-more-itertools-8.12.0-2.el9.noarch 56/102 2026-03-09T15:17:36.269 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : python3-typing-extensions-4.15.0-1.el9.noarch 62/102 2026-03-09T15:17:36.270 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : python3-pytz-2021.1-5.el9.noarch 58/102 2026-03-09T15:17:36.271 INFO:teuthology.orchestra.run.vm03.stdout: Erasing : python3-toml-0.10.2-6.el9.noarch 57/102 2026-03-09T15:17:36.271 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : python3-idna-2.10-7.el9.1.noarch 63/102 2026-03-09T15:17:36.274 INFO:teuthology.orchestra.run.vm03.stdout: Erasing : python3-pytz-2021.1-5.el9.noarch 58/102 2026-03-09T15:17:36.276 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : python3-pysocks-1.7.1-12.el9.noarch 64/102 2026-03-09T15:17:36.277 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : python3-backports-tarfile-1.2.0-1.el9.noarch 59/102 2026-03-09T15:17:36.280 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : python3-pyasn1-0.4.8-7.el9.noarch 65/102 2026-03-09T15:17:36.281 INFO:teuthology.orchestra.run.vm03.stdout: Erasing : python3-backports-tarfile-1.2.0-1.el9.noarch 59/102 2026-03-09T15:17:36.281 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : python3-devel-3.9.25-3.el9.x86_64 60/102 2026-03-09T15:17:36.283 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : python3-jsonpointer-2.0-4.el9.noarch 61/102 2026-03-09T15:17:36.284 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : python3-logutils-0.3.5-21.el9.noarch 66/102 2026-03-09T15:17:36.285 INFO:teuthology.orchestra.run.vm03.stdout: Erasing : python3-devel-3.9.25-3.el9.x86_64 60/102 2026-03-09T15:17:36.286 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : python3-typing-extensions-4.15.0-1.el9.noarch 62/102 2026-03-09T15:17:36.287 INFO:teuthology.orchestra.run.vm03.stdout: Erasing : python3-jsonpointer-2.0-4.el9.noarch 61/102 2026-03-09T15:17:36.288 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : python3-webob-1.8.8-2.el9.noarch 67/102 2026-03-09T15:17:36.289 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : python3-idna-2.10-7.el9.1.noarch 63/102 2026-03-09T15:17:36.290 INFO:teuthology.orchestra.run.vm03.stdout: Erasing : python3-typing-extensions-4.15.0-1.el9.noarch 62/102 2026-03-09T15:17:36.292 INFO:teuthology.orchestra.run.vm03.stdout: Erasing : python3-idna-2.10-7.el9.1.noarch 63/102 2026-03-09T15:17:36.293 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : python3-pysocks-1.7.1-12.el9.noarch 64/102 2026-03-09T15:17:36.294 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : python3-cachetools-4.2.4-1.el9.noarch 68/102 2026-03-09T15:17:36.297 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : python3-chardet-4.0.0-5.el9.noarch 69/102 2026-03-09T15:17:36.297 INFO:teuthology.orchestra.run.vm03.stdout: Erasing : python3-pysocks-1.7.1-12.el9.noarch 64/102 2026-03-09T15:17:36.297 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : python3-pyasn1-0.4.8-7.el9.noarch 65/102 2026-03-09T15:17:36.300 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : python3-autocommand-2.2.2-8.el9.noarch 70/102 2026-03-09T15:17:36.301 INFO:teuthology.orchestra.run.vm03.stdout: Erasing : python3-pyasn1-0.4.8-7.el9.noarch 65/102 2026-03-09T15:17:36.302 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : python3-packaging-20.9-5.el9.noarch 71/102 2026-03-09T15:17:36.303 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : python3-logutils-0.3.5-21.el9.noarch 66/102 2026-03-09T15:17:36.306 INFO:teuthology.orchestra.run.vm03.stdout: Erasing : python3-logutils-0.3.5-21.el9.noarch 66/102 2026-03-09T15:17:36.307 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : python3-webob-1.8.8-2.el9.noarch 67/102 2026-03-09T15:17:36.308 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : grpc-data-1.46.7-10.el9.noarch 72/102 2026-03-09T15:17:36.310 INFO:teuthology.orchestra.run.vm03.stdout: Erasing : python3-webob-1.8.8-2.el9.noarch 67/102 2026-03-09T15:17:36.311 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : python3-protobuf-3.14.0-17.el9.noarch 73/102 2026-03-09T15:17:36.313 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : python3-cachetools-4.2.4-1.el9.noarch 68/102 2026-03-09T15:17:36.314 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : python3-zc-lockfile-2.0-10.el9.noarch 74/102 2026-03-09T15:17:36.315 INFO:teuthology.orchestra.run.vm03.stdout: Erasing : python3-cachetools-4.2.4-1.el9.noarch 68/102 2026-03-09T15:17:36.316 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : python3-chardet-4.0.0-5.el9.noarch 69/102 2026-03-09T15:17:36.318 INFO:teuthology.orchestra.run.vm03.stdout: Erasing : python3-chardet-4.0.0-5.el9.noarch 69/102 2026-03-09T15:17:36.320 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : python3-autocommand-2.2.2-8.el9.noarch 70/102 2026-03-09T15:17:36.322 INFO:teuthology.orchestra.run.vm03.stdout: Erasing : python3-autocommand-2.2.2-8.el9.noarch 70/102 2026-03-09T15:17:36.322 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : python3-natsort-7.1.1-5.el9.noarch 75/102 2026-03-09T15:17:36.322 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : python3-packaging-20.9-5.el9.noarch 71/102 2026-03-09T15:17:36.324 INFO:teuthology.orchestra.run.vm03.stdout: Erasing : python3-packaging-20.9-5.el9.noarch 71/102 2026-03-09T15:17:36.328 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : python3-oauthlib-3.1.1-5.el9.noarch 76/102 2026-03-09T15:17:36.328 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : grpc-data-1.46.7-10.el9.noarch 72/102 2026-03-09T15:17:36.329 INFO:teuthology.orchestra.run.vm03.stdout: Erasing : grpc-data-1.46.7-10.el9.noarch 72/102 2026-03-09T15:17:36.331 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : python3-websocket-client-1.2.3-2.el9.noarch 77/102 2026-03-09T15:17:36.331 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : python3-protobuf-3.14.0-17.el9.noarch 73/102 2026-03-09T15:17:36.333 INFO:teuthology.orchestra.run.vm03.stdout: Erasing : python3-protobuf-3.14.0-17.el9.noarch 73/102 2026-03-09T15:17:36.334 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : python3-certifi-2023.05.07-4.el9.noarch 78/102 2026-03-09T15:17:36.334 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : python3-zc-lockfile-2.0-10.el9.noarch 74/102 2026-03-09T15:17:36.335 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : ceph-grafana-dashboards-2:19.2.3-678.ge911bdeb.e 79/102 2026-03-09T15:17:36.336 INFO:teuthology.orchestra.run.vm03.stdout: Erasing : python3-zc-lockfile-2.0-10.el9.noarch 74/102 2026-03-09T15:17:36.341 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : ceph-prometheus-alerts-2:19.2.3-678.ge911bdeb.el 80/102 2026-03-09T15:17:36.342 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : python3-natsort-7.1.1-5.el9.noarch 75/102 2026-03-09T15:17:36.344 INFO:teuthology.orchestra.run.vm03.stdout: Erasing : python3-natsort-7.1.1-5.el9.noarch 75/102 2026-03-09T15:17:36.344 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : python3-werkzeug-2.0.3-3.el9.1.noarch 81/102 2026-03-09T15:17:36.348 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : python3-oauthlib-3.1.1-5.el9.noarch 76/102 2026-03-09T15:17:36.349 INFO:teuthology.orchestra.run.vm03.stdout: Erasing : python3-oauthlib-3.1.1-5.el9.noarch 76/102 2026-03-09T15:17:36.351 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : python3-websocket-client-1.2.3-2.el9.noarch 77/102 2026-03-09T15:17:36.353 INFO:teuthology.orchestra.run.vm03.stdout: Erasing : python3-websocket-client-1.2.3-2.el9.noarch 77/102 2026-03-09T15:17:36.353 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : python3-certifi-2023.05.07-4.el9.noarch 78/102 2026-03-09T15:17:36.355 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : ceph-grafana-dashboards-2:19.2.3-678.ge911bdeb.e 79/102 2026-03-09T15:17:36.355 INFO:teuthology.orchestra.run.vm03.stdout: Erasing : python3-certifi-2023.05.07-4.el9.noarch 78/102 2026-03-09T15:17:36.356 INFO:teuthology.orchestra.run.vm03.stdout: Erasing : ceph-grafana-dashboards-2:19.2.3-678.ge911bdeb.e 79/102 2026-03-09T15:17:36.361 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : ceph-prometheus-alerts-2:19.2.3-678.ge911bdeb.el 80/102 2026-03-09T15:17:36.362 INFO:teuthology.orchestra.run.vm03.stdout: Erasing : ceph-prometheus-alerts-2:19.2.3-678.ge911bdeb.el 80/102 2026-03-09T15:17:36.363 INFO:teuthology.orchestra.run.vm06.stdout: Running scriptlet: ceph-base-2:19.2.3-678.ge911bdeb.el9.x86_64 82/102 2026-03-09T15:17:36.363 INFO:teuthology.orchestra.run.vm06.stdout:Removed "/etc/systemd/system/ceph.target.wants/ceph-crash.service". 2026-03-09T15:17:36.363 INFO:teuthology.orchestra.run.vm06.stdout: 2026-03-09T15:17:36.364 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : python3-werkzeug-2.0.3-3.el9.1.noarch 81/102 2026-03-09T15:17:36.366 INFO:teuthology.orchestra.run.vm03.stdout: Erasing : python3-werkzeug-2.0.3-3.el9.1.noarch 81/102 2026-03-09T15:17:36.370 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : ceph-base-2:19.2.3-678.ge911bdeb.el9.x86_64 82/102 2026-03-09T15:17:36.384 INFO:teuthology.orchestra.run.vm04.stdout: Running scriptlet: ceph-base-2:19.2.3-678.ge911bdeb.el9.x86_64 82/102 2026-03-09T15:17:36.384 INFO:teuthology.orchestra.run.vm04.stdout:Removed "/etc/systemd/system/ceph.target.wants/ceph-crash.service". 2026-03-09T15:17:36.384 INFO:teuthology.orchestra.run.vm04.stdout: 2026-03-09T15:17:36.387 INFO:teuthology.orchestra.run.vm03.stdout: Running scriptlet: ceph-base-2:19.2.3-678.ge911bdeb.el9.x86_64 82/102 2026-03-09T15:17:36.387 INFO:teuthology.orchestra.run.vm03.stdout:Removed "/etc/systemd/system/ceph.target.wants/ceph-crash.service". 2026-03-09T15:17:36.387 INFO:teuthology.orchestra.run.vm03.stdout: 2026-03-09T15:17:36.391 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : ceph-base-2:19.2.3-678.ge911bdeb.el9.x86_64 82/102 2026-03-09T15:17:36.394 INFO:teuthology.orchestra.run.vm03.stdout: Erasing : ceph-base-2:19.2.3-678.ge911bdeb.el9.x86_64 82/102 2026-03-09T15:17:36.396 INFO:teuthology.orchestra.run.vm06.stdout: Running scriptlet: ceph-base-2:19.2.3-678.ge911bdeb.el9.x86_64 82/102 2026-03-09T15:17:36.396 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : ceph-common-2:19.2.3-678.ge911bdeb.el9.x86_64 83/102 2026-03-09T15:17:36.408 INFO:teuthology.orchestra.run.vm06.stdout: Running scriptlet: ceph-common-2:19.2.3-678.ge911bdeb.el9.x86_64 83/102 2026-03-09T15:17:36.412 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : qatzip-libs-1.3.1-1.el9.x86_64 84/102 2026-03-09T15:17:36.415 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : python3-ceph-common-2:19.2.3-678.ge911bdeb.el9.x 85/102 2026-03-09T15:17:36.416 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : python3-prettytable-0.7.2-27.el9.noarch 86/102 2026-03-09T15:17:36.417 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : ceph-selinux-2:19.2.3-678.ge911bdeb.el9.x86_64 87/102 2026-03-09T15:17:36.418 INFO:teuthology.orchestra.run.vm04.stdout: Running scriptlet: ceph-base-2:19.2.3-678.ge911bdeb.el9.x86_64 82/102 2026-03-09T15:17:36.418 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : ceph-common-2:19.2.3-678.ge911bdeb.el9.x86_64 83/102 2026-03-09T15:17:36.422 INFO:teuthology.orchestra.run.vm03.stdout: Running scriptlet: ceph-base-2:19.2.3-678.ge911bdeb.el9.x86_64 82/102 2026-03-09T15:17:36.422 INFO:teuthology.orchestra.run.vm03.stdout: Erasing : ceph-common-2:19.2.3-678.ge911bdeb.el9.x86_64 83/102 2026-03-09T15:17:36.430 INFO:teuthology.orchestra.run.vm04.stdout: Running scriptlet: ceph-common-2:19.2.3-678.ge911bdeb.el9.x86_64 83/102 2026-03-09T15:17:36.434 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : qatzip-libs-1.3.1-1.el9.x86_64 84/102 2026-03-09T15:17:36.435 INFO:teuthology.orchestra.run.vm03.stdout: Running scriptlet: ceph-common-2:19.2.3-678.ge911bdeb.el9.x86_64 83/102 2026-03-09T15:17:36.437 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : python3-ceph-common-2:19.2.3-678.ge911bdeb.el9.x 85/102 2026-03-09T15:17:36.439 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : python3-prettytable-0.7.2-27.el9.noarch 86/102 2026-03-09T15:17:36.439 INFO:teuthology.orchestra.run.vm03.stdout: Erasing : qatzip-libs-1.3.1-1.el9.x86_64 84/102 2026-03-09T15:17:36.439 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : ceph-selinux-2:19.2.3-678.ge911bdeb.el9.x86_64 87/102 2026-03-09T15:17:36.442 INFO:teuthology.orchestra.run.vm03.stdout: Erasing : python3-ceph-common-2:19.2.3-678.ge911bdeb.el9.x 85/102 2026-03-09T15:17:36.444 INFO:teuthology.orchestra.run.vm03.stdout: Erasing : python3-prettytable-0.7.2-27.el9.noarch 86/102 2026-03-09T15:17:36.444 INFO:teuthology.orchestra.run.vm03.stdout: Erasing : ceph-selinux-2:19.2.3-678.ge911bdeb.el9.x86_64 87/102 2026-03-09T15:17:41.830 INFO:teuthology.orchestra.run.vm03.stdout: Running scriptlet: ceph-selinux-2:19.2.3-678.ge911bdeb.el9.x86_64 87/102 2026-03-09T15:17:41.831 INFO:teuthology.orchestra.run.vm03.stdout:skipping the directory /sys 2026-03-09T15:17:41.831 INFO:teuthology.orchestra.run.vm03.stdout:skipping the directory /proc 2026-03-09T15:17:41.831 INFO:teuthology.orchestra.run.vm03.stdout:skipping the directory /mnt 2026-03-09T15:17:41.831 INFO:teuthology.orchestra.run.vm03.stdout:skipping the directory /var/tmp 2026-03-09T15:17:41.831 INFO:teuthology.orchestra.run.vm03.stdout:skipping the directory /home 2026-03-09T15:17:41.831 INFO:teuthology.orchestra.run.vm03.stdout:skipping the directory /root 2026-03-09T15:17:41.831 INFO:teuthology.orchestra.run.vm03.stdout:skipping the directory /tmp 2026-03-09T15:17:41.831 INFO:teuthology.orchestra.run.vm03.stdout: 2026-03-09T15:17:41.831 INFO:teuthology.orchestra.run.vm04.stdout: Running scriptlet: ceph-selinux-2:19.2.3-678.ge911bdeb.el9.x86_64 87/102 2026-03-09T15:17:41.832 INFO:teuthology.orchestra.run.vm04.stdout:skipping the directory /sys 2026-03-09T15:17:41.832 INFO:teuthology.orchestra.run.vm04.stdout:skipping the directory /proc 2026-03-09T15:17:41.832 INFO:teuthology.orchestra.run.vm04.stdout:skipping the directory /mnt 2026-03-09T15:17:41.832 INFO:teuthology.orchestra.run.vm04.stdout:skipping the directory /var/tmp 2026-03-09T15:17:41.832 INFO:teuthology.orchestra.run.vm04.stdout:skipping the directory /home 2026-03-09T15:17:41.832 INFO:teuthology.orchestra.run.vm04.stdout:skipping the directory /root 2026-03-09T15:17:41.832 INFO:teuthology.orchestra.run.vm04.stdout:skipping the directory /tmp 2026-03-09T15:17:41.832 INFO:teuthology.orchestra.run.vm04.stdout: 2026-03-09T15:17:41.840 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : qatlib-25.08.0-2.el9.x86_64 88/102 2026-03-09T15:17:41.841 INFO:teuthology.orchestra.run.vm03.stdout: Erasing : qatlib-25.08.0-2.el9.x86_64 88/102 2026-03-09T15:17:41.854 INFO:teuthology.orchestra.run.vm06.stdout: Running scriptlet: ceph-selinux-2:19.2.3-678.ge911bdeb.el9.x86_64 87/102 2026-03-09T15:17:41.854 INFO:teuthology.orchestra.run.vm06.stdout:skipping the directory /sys 2026-03-09T15:17:41.854 INFO:teuthology.orchestra.run.vm06.stdout:skipping the directory /proc 2026-03-09T15:17:41.854 INFO:teuthology.orchestra.run.vm06.stdout:skipping the directory /mnt 2026-03-09T15:17:41.854 INFO:teuthology.orchestra.run.vm06.stdout:skipping the directory /var/tmp 2026-03-09T15:17:41.854 INFO:teuthology.orchestra.run.vm06.stdout:skipping the directory /home 2026-03-09T15:17:41.854 INFO:teuthology.orchestra.run.vm06.stdout:skipping the directory /root 2026-03-09T15:17:41.854 INFO:teuthology.orchestra.run.vm06.stdout:skipping the directory /tmp 2026-03-09T15:17:41.854 INFO:teuthology.orchestra.run.vm06.stdout: 2026-03-09T15:17:41.857 INFO:teuthology.orchestra.run.vm04.stdout: Running scriptlet: qatlib-service-25.08.0-2.el9.x86_64 89/102 2026-03-09T15:17:41.857 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : qatlib-service-25.08.0-2.el9.x86_64 89/102 2026-03-09T15:17:41.858 INFO:teuthology.orchestra.run.vm03.stdout: Running scriptlet: qatlib-service-25.08.0-2.el9.x86_64 89/102 2026-03-09T15:17:41.858 INFO:teuthology.orchestra.run.vm03.stdout: Erasing : qatlib-service-25.08.0-2.el9.x86_64 89/102 2026-03-09T15:17:41.861 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : qatlib-25.08.0-2.el9.x86_64 88/102 2026-03-09T15:17:41.865 INFO:teuthology.orchestra.run.vm04.stdout: Running scriptlet: qatlib-service-25.08.0-2.el9.x86_64 89/102 2026-03-09T15:17:41.866 INFO:teuthology.orchestra.run.vm03.stdout: Running scriptlet: qatlib-service-25.08.0-2.el9.x86_64 89/102 2026-03-09T15:17:41.868 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : gperftools-libs-2.9.1-3.el9.x86_64 90/102 2026-03-09T15:17:41.868 INFO:teuthology.orchestra.run.vm03.stdout: Erasing : gperftools-libs-2.9.1-3.el9.x86_64 90/102 2026-03-09T15:17:41.870 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : libunwind-1.6.2-1.el9.x86_64 91/102 2026-03-09T15:17:41.870 INFO:teuthology.orchestra.run.vm03.stdout: Erasing : libunwind-1.6.2-1.el9.x86_64 91/102 2026-03-09T15:17:41.872 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : pciutils-3.7.0-7.el9.x86_64 92/102 2026-03-09T15:17:41.873 INFO:teuthology.orchestra.run.vm03.stdout: Erasing : pciutils-3.7.0-7.el9.x86_64 92/102 2026-03-09T15:17:41.875 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : liboath-2.6.12-1.el9.x86_64 93/102 2026-03-09T15:17:41.875 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : libradosstriper1-2:19.2.3-678.ge911bdeb.el9.x86_ 94/102 2026-03-09T15:17:41.875 INFO:teuthology.orchestra.run.vm03.stdout: Erasing : liboath-2.6.12-1.el9.x86_64 93/102 2026-03-09T15:17:41.875 INFO:teuthology.orchestra.run.vm03.stdout: Erasing : libradosstriper1-2:19.2.3-678.ge911bdeb.el9.x86_ 94/102 2026-03-09T15:17:41.880 INFO:teuthology.orchestra.run.vm06.stdout: Running scriptlet: qatlib-service-25.08.0-2.el9.x86_64 89/102 2026-03-09T15:17:41.880 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : qatlib-service-25.08.0-2.el9.x86_64 89/102 2026-03-09T15:17:41.887 INFO:teuthology.orchestra.run.vm06.stdout: Running scriptlet: qatlib-service-25.08.0-2.el9.x86_64 89/102 2026-03-09T15:17:41.888 INFO:teuthology.orchestra.run.vm03.stdout: Running scriptlet: libradosstriper1-2:19.2.3-678.ge911bdeb.el9.x86_ 94/102 2026-03-09T15:17:41.888 INFO:teuthology.orchestra.run.vm04.stdout: Running scriptlet: libradosstriper1-2:19.2.3-678.ge911bdeb.el9.x86_ 94/102 2026-03-09T15:17:41.890 INFO:teuthology.orchestra.run.vm03.stdout: Erasing : ledmon-libs-1.1.0-3.el9.x86_64 95/102 2026-03-09T15:17:41.890 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : gperftools-libs-2.9.1-3.el9.x86_64 90/102 2026-03-09T15:17:41.890 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : ledmon-libs-1.1.0-3.el9.x86_64 95/102 2026-03-09T15:17:41.892 INFO:teuthology.orchestra.run.vm03.stdout: Erasing : libquadmath-11.5.0-14.el9.x86_64 96/102 2026-03-09T15:17:41.892 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : libunwind-1.6.2-1.el9.x86_64 91/102 2026-03-09T15:17:41.892 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : libquadmath-11.5.0-14.el9.x86_64 96/102 2026-03-09T15:17:41.895 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : pciutils-3.7.0-7.el9.x86_64 92/102 2026-03-09T15:17:41.895 INFO:teuthology.orchestra.run.vm03.stdout: Erasing : python3-markupsafe-1.1.1-12.el9.x86_64 97/102 2026-03-09T15:17:41.895 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : python3-markupsafe-1.1.1-12.el9.x86_64 97/102 2026-03-09T15:17:41.896 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : liboath-2.6.12-1.el9.x86_64 93/102 2026-03-09T15:17:41.896 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : libradosstriper1-2:19.2.3-678.ge911bdeb.el9.x86_ 94/102 2026-03-09T15:17:41.898 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : protobuf-3.14.0-17.el9.x86_64 98/102 2026-03-09T15:17:41.898 INFO:teuthology.orchestra.run.vm03.stdout: Erasing : protobuf-3.14.0-17.el9.x86_64 98/102 2026-03-09T15:17:41.902 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : libconfig-1.7.2-9.el9.x86_64 99/102 2026-03-09T15:17:41.903 INFO:teuthology.orchestra.run.vm03.stdout: Erasing : libconfig-1.7.2-9.el9.x86_64 99/102 2026-03-09T15:17:41.908 INFO:teuthology.orchestra.run.vm06.stdout: Running scriptlet: libradosstriper1-2:19.2.3-678.ge911bdeb.el9.x86_ 94/102 2026-03-09T15:17:41.910 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : cryptsetup-2.8.1-3.el9.x86_64 100/102 2026-03-09T15:17:41.910 INFO:teuthology.orchestra.run.vm03.stdout: Erasing : cryptsetup-2.8.1-3.el9.x86_64 100/102 2026-03-09T15:17:41.910 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : ledmon-libs-1.1.0-3.el9.x86_64 95/102 2026-03-09T15:17:41.912 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : libquadmath-11.5.0-14.el9.x86_64 96/102 2026-03-09T15:17:41.915 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : abseil-cpp-20211102.0-4.el9.x86_64 101/102 2026-03-09T15:17:41.915 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : libcephsqlite-2:19.2.3-678.ge911bdeb.el9.x86_64 102/102 2026-03-09T15:17:41.915 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : python3-markupsafe-1.1.1-12.el9.x86_64 97/102 2026-03-09T15:17:41.915 INFO:teuthology.orchestra.run.vm03.stdout: Erasing : abseil-cpp-20211102.0-4.el9.x86_64 101/102 2026-03-09T15:17:41.915 INFO:teuthology.orchestra.run.vm03.stdout: Erasing : libcephsqlite-2:19.2.3-678.ge911bdeb.el9.x86_64 102/102 2026-03-09T15:17:41.918 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : protobuf-3.14.0-17.el9.x86_64 98/102 2026-03-09T15:17:41.923 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : libconfig-1.7.2-9.el9.x86_64 99/102 2026-03-09T15:17:41.931 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : cryptsetup-2.8.1-3.el9.x86_64 100/102 2026-03-09T15:17:41.934 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : abseil-cpp-20211102.0-4.el9.x86_64 101/102 2026-03-09T15:17:41.935 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : libcephsqlite-2:19.2.3-678.ge911bdeb.el9.x86_64 102/102 2026-03-09T15:17:42.010 INFO:teuthology.orchestra.run.vm03.stdout: Running scriptlet: libcephsqlite-2:19.2.3-678.ge911bdeb.el9.x86_64 102/102 2026-03-09T15:17:42.010 INFO:teuthology.orchestra.run.vm03.stdout: Verifying : abseil-cpp-20211102.0-4.el9.x86_64 1/102 2026-03-09T15:17:42.010 INFO:teuthology.orchestra.run.vm03.stdout: Verifying : ceph-base-2:19.2.3-678.ge911bdeb.el9.x86_64 2/102 2026-03-09T15:17:42.010 INFO:teuthology.orchestra.run.vm03.stdout: Verifying : ceph-common-2:19.2.3-678.ge911bdeb.el9.x86_64 3/102 2026-03-09T15:17:42.010 INFO:teuthology.orchestra.run.vm03.stdout: Verifying : ceph-grafana-dashboards-2:19.2.3-678.ge911bdeb.e 4/102 2026-03-09T15:17:42.010 INFO:teuthology.orchestra.run.vm03.stdout: Verifying : ceph-immutable-object-cache-2:19.2.3-678.ge911bd 5/102 2026-03-09T15:17:42.010 INFO:teuthology.orchestra.run.vm03.stdout: Verifying : ceph-mgr-2:19.2.3-678.ge911bdeb.el9.x86_64 6/102 2026-03-09T15:17:42.010 INFO:teuthology.orchestra.run.vm03.stdout: Verifying : ceph-mgr-cephadm-2:19.2.3-678.ge911bdeb.el9.noar 7/102 2026-03-09T15:17:42.010 INFO:teuthology.orchestra.run.vm03.stdout: Verifying : ceph-mgr-dashboard-2:19.2.3-678.ge911bdeb.el9.no 8/102 2026-03-09T15:17:42.010 INFO:teuthology.orchestra.run.vm03.stdout: Verifying : ceph-mgr-diskprediction-local-2:19.2.3-678.ge911 9/102 2026-03-09T15:17:42.011 INFO:teuthology.orchestra.run.vm03.stdout: Verifying : ceph-mgr-modules-core-2:19.2.3-678.ge911bdeb.el9 10/102 2026-03-09T15:17:42.011 INFO:teuthology.orchestra.run.vm03.stdout: Verifying : ceph-mgr-rook-2:19.2.3-678.ge911bdeb.el9.noarch 11/102 2026-03-09T15:17:42.011 INFO:teuthology.orchestra.run.vm03.stdout: Verifying : ceph-osd-2:19.2.3-678.ge911bdeb.el9.x86_64 12/102 2026-03-09T15:17:42.011 INFO:teuthology.orchestra.run.vm03.stdout: Verifying : ceph-prometheus-alerts-2:19.2.3-678.ge911bdeb.el 13/102 2026-03-09T15:17:42.011 INFO:teuthology.orchestra.run.vm03.stdout: Verifying : ceph-selinux-2:19.2.3-678.ge911bdeb.el9.x86_64 14/102 2026-03-09T15:17:42.011 INFO:teuthology.orchestra.run.vm03.stdout: Verifying : ceph-volume-2:19.2.3-678.ge911bdeb.el9.noarch 15/102 2026-03-09T15:17:42.011 INFO:teuthology.orchestra.run.vm03.stdout: Verifying : cryptsetup-2.8.1-3.el9.x86_64 16/102 2026-03-09T15:17:42.011 INFO:teuthology.orchestra.run.vm03.stdout: Verifying : flexiblas-3.0.4-9.el9.x86_64 17/102 2026-03-09T15:17:42.011 INFO:teuthology.orchestra.run.vm03.stdout: Verifying : flexiblas-netlib-3.0.4-9.el9.x86_64 18/102 2026-03-09T15:17:42.011 INFO:teuthology.orchestra.run.vm03.stdout: Verifying : flexiblas-openblas-openmp-3.0.4-9.el9.x86_64 19/102 2026-03-09T15:17:42.011 INFO:teuthology.orchestra.run.vm03.stdout: Verifying : gperftools-libs-2.9.1-3.el9.x86_64 20/102 2026-03-09T15:17:42.011 INFO:teuthology.orchestra.run.vm03.stdout: Verifying : grpc-data-1.46.7-10.el9.noarch 21/102 2026-03-09T15:17:42.011 INFO:teuthology.orchestra.run.vm03.stdout: Verifying : ledmon-libs-1.1.0-3.el9.x86_64 22/102 2026-03-09T15:17:42.011 INFO:teuthology.orchestra.run.vm03.stdout: Verifying : libcephsqlite-2:19.2.3-678.ge911bdeb.el9.x86_64 23/102 2026-03-09T15:17:42.011 INFO:teuthology.orchestra.run.vm03.stdout: Verifying : libconfig-1.7.2-9.el9.x86_64 24/102 2026-03-09T15:17:42.011 INFO:teuthology.orchestra.run.vm03.stdout: Verifying : libgfortran-11.5.0-14.el9.x86_64 25/102 2026-03-09T15:17:42.011 INFO:teuthology.orchestra.run.vm03.stdout: Verifying : liboath-2.6.12-1.el9.x86_64 26/102 2026-03-09T15:17:42.011 INFO:teuthology.orchestra.run.vm03.stdout: Verifying : libquadmath-11.5.0-14.el9.x86_64 27/102 2026-03-09T15:17:42.011 INFO:teuthology.orchestra.run.vm03.stdout: Verifying : libradosstriper1-2:19.2.3-678.ge911bdeb.el9.x86_ 28/102 2026-03-09T15:17:42.012 INFO:teuthology.orchestra.run.vm03.stdout: Verifying : libstoragemgmt-1.10.1-1.el9.x86_64 29/102 2026-03-09T15:17:42.012 INFO:teuthology.orchestra.run.vm03.stdout: Verifying : libunwind-1.6.2-1.el9.x86_64 30/102 2026-03-09T15:17:42.012 INFO:teuthology.orchestra.run.vm03.stdout: Verifying : openblas-0.3.29-1.el9.x86_64 31/102 2026-03-09T15:17:42.012 INFO:teuthology.orchestra.run.vm03.stdout: Verifying : openblas-openmp-0.3.29-1.el9.x86_64 32/102 2026-03-09T15:17:42.012 INFO:teuthology.orchestra.run.vm03.stdout: Verifying : pciutils-3.7.0-7.el9.x86_64 33/102 2026-03-09T15:17:42.012 INFO:teuthology.orchestra.run.vm03.stdout: Verifying : protobuf-3.14.0-17.el9.x86_64 34/102 2026-03-09T15:17:42.012 INFO:teuthology.orchestra.run.vm03.stdout: Verifying : protobuf-compiler-3.14.0-17.el9.x86_64 35/102 2026-03-09T15:17:42.012 INFO:teuthology.orchestra.run.vm03.stdout: Verifying : python3-asyncssh-2.13.2-5.el9.noarch 36/102 2026-03-09T15:17:42.012 INFO:teuthology.orchestra.run.vm03.stdout: Verifying : python3-autocommand-2.2.2-8.el9.noarch 37/102 2026-03-09T15:17:42.012 INFO:teuthology.orchestra.run.vm03.stdout: Verifying : python3-babel-2.9.1-2.el9.noarch 38/102 2026-03-09T15:17:42.012 INFO:teuthology.orchestra.run.vm03.stdout: Verifying : python3-backports-tarfile-1.2.0-1.el9.noarch 39/102 2026-03-09T15:17:42.012 INFO:teuthology.orchestra.run.vm03.stdout: Verifying : python3-bcrypt-3.2.2-1.el9.x86_64 40/102 2026-03-09T15:17:42.012 INFO:teuthology.orchestra.run.vm03.stdout: Verifying : python3-cachetools-4.2.4-1.el9.noarch 41/102 2026-03-09T15:17:42.012 INFO:teuthology.orchestra.run.vm03.stdout: Verifying : python3-ceph-common-2:19.2.3-678.ge911bdeb.el9.x 42/102 2026-03-09T15:17:42.012 INFO:teuthology.orchestra.run.vm03.stdout: Verifying : python3-certifi-2023.05.07-4.el9.noarch 43/102 2026-03-09T15:17:42.012 INFO:teuthology.orchestra.run.vm03.stdout: Verifying : python3-cffi-1.14.5-5.el9.x86_64 44/102 2026-03-09T15:17:42.012 INFO:teuthology.orchestra.run.vm03.stdout: Verifying : python3-chardet-4.0.0-5.el9.noarch 45/102 2026-03-09T15:17:42.012 INFO:teuthology.orchestra.run.vm03.stdout: Verifying : python3-cheroot-10.0.1-4.el9.noarch 46/102 2026-03-09T15:17:42.012 INFO:teuthology.orchestra.run.vm03.stdout: Verifying : python3-cherrypy-18.6.1-2.el9.noarch 47/102 2026-03-09T15:17:42.012 INFO:teuthology.orchestra.run.vm03.stdout: Verifying : python3-cryptography-36.0.1-5.el9.x86_64 48/102 2026-03-09T15:17:42.012 INFO:teuthology.orchestra.run.vm03.stdout: Verifying : python3-devel-3.9.25-3.el9.x86_64 49/102 2026-03-09T15:17:42.012 INFO:teuthology.orchestra.run.vm03.stdout: Verifying : python3-google-auth-1:2.45.0-1.el9.noarch 50/102 2026-03-09T15:17:42.012 INFO:teuthology.orchestra.run.vm03.stdout: Verifying : python3-grpcio-1.46.7-10.el9.x86_64 51/102 2026-03-09T15:17:42.012 INFO:teuthology.orchestra.run.vm03.stdout: Verifying : python3-grpcio-tools-1.46.7-10.el9.x86_64 52/102 2026-03-09T15:17:42.012 INFO:teuthology.orchestra.run.vm03.stdout: Verifying : python3-idna-2.10-7.el9.1.noarch 53/102 2026-03-09T15:17:42.013 INFO:teuthology.orchestra.run.vm03.stdout: Verifying : python3-jaraco-8.2.1-3.el9.noarch 54/102 2026-03-09T15:17:42.013 INFO:teuthology.orchestra.run.vm03.stdout: Verifying : python3-jaraco-classes-3.2.1-5.el9.noarch 55/102 2026-03-09T15:17:42.013 INFO:teuthology.orchestra.run.vm03.stdout: Verifying : python3-jaraco-collections-3.0.0-8.el9.noarch 56/102 2026-03-09T15:17:42.013 INFO:teuthology.orchestra.run.vm03.stdout: Verifying : python3-jaraco-context-6.0.1-3.el9.noarch 57/102 2026-03-09T15:17:42.013 INFO:teuthology.orchestra.run.vm03.stdout: Verifying : python3-jaraco-functools-3.5.0-2.el9.noarch 58/102 2026-03-09T15:17:42.013 INFO:teuthology.orchestra.run.vm03.stdout: Verifying : python3-jaraco-text-4.0.0-2.el9.noarch 59/102 2026-03-09T15:17:42.013 INFO:teuthology.orchestra.run.vm03.stdout: Verifying : python3-jinja2-2.11.3-8.el9.noarch 60/102 2026-03-09T15:17:42.013 INFO:teuthology.orchestra.run.vm03.stdout: Verifying : python3-jsonpatch-1.21-16.el9.noarch 61/102 2026-03-09T15:17:42.013 INFO:teuthology.orchestra.run.vm03.stdout: Verifying : python3-jsonpointer-2.0-4.el9.noarch 62/102 2026-03-09T15:17:42.013 INFO:teuthology.orchestra.run.vm03.stdout: Verifying : python3-kubernetes-1:26.1.0-3.el9.noarch 63/102 2026-03-09T15:17:42.013 INFO:teuthology.orchestra.run.vm03.stdout: Verifying : python3-libstoragemgmt-1.10.1-1.el9.x86_64 64/102 2026-03-09T15:17:42.013 INFO:teuthology.orchestra.run.vm03.stdout: Verifying : python3-logutils-0.3.5-21.el9.noarch 65/102 2026-03-09T15:17:42.013 INFO:teuthology.orchestra.run.vm03.stdout: Verifying : python3-mako-1.1.4-6.el9.noarch 66/102 2026-03-09T15:17:42.013 INFO:teuthology.orchestra.run.vm03.stdout: Verifying : python3-markupsafe-1.1.1-12.el9.x86_64 67/102 2026-03-09T15:17:42.013 INFO:teuthology.orchestra.run.vm03.stdout: Verifying : python3-more-itertools-8.12.0-2.el9.noarch 68/102 2026-03-09T15:17:42.013 INFO:teuthology.orchestra.run.vm03.stdout: Verifying : python3-natsort-7.1.1-5.el9.noarch 69/102 2026-03-09T15:17:42.013 INFO:teuthology.orchestra.run.vm03.stdout: Verifying : python3-numpy-1:1.23.5-2.el9.x86_64 70/102 2026-03-09T15:17:42.013 INFO:teuthology.orchestra.run.vm03.stdout: Verifying : python3-numpy-f2py-1:1.23.5-2.el9.x86_64 71/102 2026-03-09T15:17:42.013 INFO:teuthology.orchestra.run.vm03.stdout: Verifying : python3-oauthlib-3.1.1-5.el9.noarch 72/102 2026-03-09T15:17:42.013 INFO:teuthology.orchestra.run.vm03.stdout: Verifying : python3-packaging-20.9-5.el9.noarch 73/102 2026-03-09T15:17:42.013 INFO:teuthology.orchestra.run.vm03.stdout: Verifying : python3-pecan-1.4.2-3.el9.noarch 74/102 2026-03-09T15:17:42.013 INFO:teuthology.orchestra.run.vm03.stdout: Verifying : python3-ply-3.11-14.el9.noarch 75/102 2026-03-09T15:17:42.014 INFO:teuthology.orchestra.run.vm03.stdout: Verifying : python3-portend-3.1.0-2.el9.noarch 76/102 2026-03-09T15:17:42.014 INFO:teuthology.orchestra.run.vm03.stdout: Verifying : python3-prettytable-0.7.2-27.el9.noarch 77/102 2026-03-09T15:17:42.014 INFO:teuthology.orchestra.run.vm03.stdout: Verifying : python3-protobuf-3.14.0-17.el9.noarch 78/102 2026-03-09T15:17:42.014 INFO:teuthology.orchestra.run.vm03.stdout: Verifying : python3-pyOpenSSL-21.0.0-1.el9.noarch 79/102 2026-03-09T15:17:42.014 INFO:teuthology.orchestra.run.vm03.stdout: Verifying : python3-pyasn1-0.4.8-7.el9.noarch 80/102 2026-03-09T15:17:42.014 INFO:teuthology.orchestra.run.vm03.stdout: Verifying : python3-pyasn1-modules-0.4.8-7.el9.noarch 81/102 2026-03-09T15:17:42.014 INFO:teuthology.orchestra.run.vm03.stdout: Verifying : python3-pycparser-2.20-6.el9.noarch 82/102 2026-03-09T15:17:42.014 INFO:teuthology.orchestra.run.vm03.stdout: Verifying : python3-pysocks-1.7.1-12.el9.noarch 83/102 2026-03-09T15:17:42.014 INFO:teuthology.orchestra.run.vm03.stdout: Verifying : python3-pytz-2021.1-5.el9.noarch 84/102 2026-03-09T15:17:42.014 INFO:teuthology.orchestra.run.vm03.stdout: Verifying : python3-repoze-lru-0.7-16.el9.noarch 85/102 2026-03-09T15:17:42.014 INFO:teuthology.orchestra.run.vm03.stdout: Verifying : python3-requests-2.25.1-10.el9.noarch 86/102 2026-03-09T15:17:42.014 INFO:teuthology.orchestra.run.vm03.stdout: Verifying : python3-requests-oauthlib-1.3.0-12.el9.noarch 87/102 2026-03-09T15:17:42.014 INFO:teuthology.orchestra.run.vm03.stdout: Verifying : python3-routes-2.5.1-5.el9.noarch 88/102 2026-03-09T15:17:42.014 INFO:teuthology.orchestra.run.vm03.stdout: Verifying : python3-rsa-4.9-2.el9.noarch 89/102 2026-03-09T15:17:42.014 INFO:teuthology.orchestra.run.vm03.stdout: Verifying : python3-scipy-1.9.3-2.el9.x86_64 90/102 2026-03-09T15:17:42.014 INFO:teuthology.orchestra.run.vm03.stdout: Verifying : python3-tempora-5.0.0-2.el9.noarch 91/102 2026-03-09T15:17:42.014 INFO:teuthology.orchestra.run.vm03.stdout: Verifying : python3-toml-0.10.2-6.el9.noarch 92/102 2026-03-09T15:17:42.014 INFO:teuthology.orchestra.run.vm03.stdout: Verifying : python3-typing-extensions-4.15.0-1.el9.noarch 93/102 2026-03-09T15:17:42.014 INFO:teuthology.orchestra.run.vm03.stdout: Verifying : python3-urllib3-1.26.5-7.el9.noarch 94/102 2026-03-09T15:17:42.014 INFO:teuthology.orchestra.run.vm03.stdout: Verifying : python3-webob-1.8.8-2.el9.noarch 95/102 2026-03-09T15:17:42.014 INFO:teuthology.orchestra.run.vm03.stdout: Verifying : python3-websocket-client-1.2.3-2.el9.noarch 96/102 2026-03-09T15:17:42.014 INFO:teuthology.orchestra.run.vm03.stdout: Verifying : python3-werkzeug-2.0.3-3.el9.1.noarch 97/102 2026-03-09T15:17:42.014 INFO:teuthology.orchestra.run.vm03.stdout: Verifying : python3-zc-lockfile-2.0-10.el9.noarch 98/102 2026-03-09T15:17:42.014 INFO:teuthology.orchestra.run.vm03.stdout: Verifying : qatlib-25.08.0-2.el9.x86_64 99/102 2026-03-09T15:17:42.014 INFO:teuthology.orchestra.run.vm03.stdout: Verifying : qatlib-service-25.08.0-2.el9.x86_64 100/102 2026-03-09T15:17:42.014 INFO:teuthology.orchestra.run.vm03.stdout: Verifying : qatzip-libs-1.3.1-1.el9.x86_64 101/102 2026-03-09T15:17:42.016 INFO:teuthology.orchestra.run.vm04.stdout: Running scriptlet: libcephsqlite-2:19.2.3-678.ge911bdeb.el9.x86_64 102/102 2026-03-09T15:17:42.016 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : abseil-cpp-20211102.0-4.el9.x86_64 1/102 2026-03-09T15:17:42.016 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : ceph-base-2:19.2.3-678.ge911bdeb.el9.x86_64 2/102 2026-03-09T15:17:42.016 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : ceph-common-2:19.2.3-678.ge911bdeb.el9.x86_64 3/102 2026-03-09T15:17:42.016 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : ceph-grafana-dashboards-2:19.2.3-678.ge911bdeb.e 4/102 2026-03-09T15:17:42.016 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : ceph-immutable-object-cache-2:19.2.3-678.ge911bd 5/102 2026-03-09T15:17:42.016 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : ceph-mgr-2:19.2.3-678.ge911bdeb.el9.x86_64 6/102 2026-03-09T15:17:42.016 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : ceph-mgr-cephadm-2:19.2.3-678.ge911bdeb.el9.noar 7/102 2026-03-09T15:17:42.016 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : ceph-mgr-dashboard-2:19.2.3-678.ge911bdeb.el9.no 8/102 2026-03-09T15:17:42.016 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : ceph-mgr-diskprediction-local-2:19.2.3-678.ge911 9/102 2026-03-09T15:17:42.017 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : ceph-mgr-modules-core-2:19.2.3-678.ge911bdeb.el9 10/102 2026-03-09T15:17:42.017 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : ceph-mgr-rook-2:19.2.3-678.ge911bdeb.el9.noarch 11/102 2026-03-09T15:17:42.017 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : ceph-osd-2:19.2.3-678.ge911bdeb.el9.x86_64 12/102 2026-03-09T15:17:42.017 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : ceph-prometheus-alerts-2:19.2.3-678.ge911bdeb.el 13/102 2026-03-09T15:17:42.017 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : ceph-selinux-2:19.2.3-678.ge911bdeb.el9.x86_64 14/102 2026-03-09T15:17:42.017 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : ceph-volume-2:19.2.3-678.ge911bdeb.el9.noarch 15/102 2026-03-09T15:17:42.017 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : cryptsetup-2.8.1-3.el9.x86_64 16/102 2026-03-09T15:17:42.017 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : flexiblas-3.0.4-9.el9.x86_64 17/102 2026-03-09T15:17:42.017 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : flexiblas-netlib-3.0.4-9.el9.x86_64 18/102 2026-03-09T15:17:42.017 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : flexiblas-openblas-openmp-3.0.4-9.el9.x86_64 19/102 2026-03-09T15:17:42.017 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : gperftools-libs-2.9.1-3.el9.x86_64 20/102 2026-03-09T15:17:42.017 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : grpc-data-1.46.7-10.el9.noarch 21/102 2026-03-09T15:17:42.017 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : ledmon-libs-1.1.0-3.el9.x86_64 22/102 2026-03-09T15:17:42.017 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : libcephsqlite-2:19.2.3-678.ge911bdeb.el9.x86_64 23/102 2026-03-09T15:17:42.017 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : libconfig-1.7.2-9.el9.x86_64 24/102 2026-03-09T15:17:42.017 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : libgfortran-11.5.0-14.el9.x86_64 25/102 2026-03-09T15:17:42.017 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : liboath-2.6.12-1.el9.x86_64 26/102 2026-03-09T15:17:42.017 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : libquadmath-11.5.0-14.el9.x86_64 27/102 2026-03-09T15:17:42.017 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : libradosstriper1-2:19.2.3-678.ge911bdeb.el9.x86_ 28/102 2026-03-09T15:17:42.017 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : libstoragemgmt-1.10.1-1.el9.x86_64 29/102 2026-03-09T15:17:42.017 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : libunwind-1.6.2-1.el9.x86_64 30/102 2026-03-09T15:17:42.017 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : openblas-0.3.29-1.el9.x86_64 31/102 2026-03-09T15:17:42.017 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : openblas-openmp-0.3.29-1.el9.x86_64 32/102 2026-03-09T15:17:42.017 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : pciutils-3.7.0-7.el9.x86_64 33/102 2026-03-09T15:17:42.017 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : protobuf-3.14.0-17.el9.x86_64 34/102 2026-03-09T15:17:42.017 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : protobuf-compiler-3.14.0-17.el9.x86_64 35/102 2026-03-09T15:17:42.017 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-asyncssh-2.13.2-5.el9.noarch 36/102 2026-03-09T15:17:42.017 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-autocommand-2.2.2-8.el9.noarch 37/102 2026-03-09T15:17:42.017 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-babel-2.9.1-2.el9.noarch 38/102 2026-03-09T15:17:42.017 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-backports-tarfile-1.2.0-1.el9.noarch 39/102 2026-03-09T15:17:42.017 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-bcrypt-3.2.2-1.el9.x86_64 40/102 2026-03-09T15:17:42.017 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-cachetools-4.2.4-1.el9.noarch 41/102 2026-03-09T15:17:42.017 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-ceph-common-2:19.2.3-678.ge911bdeb.el9.x 42/102 2026-03-09T15:17:42.017 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-certifi-2023.05.07-4.el9.noarch 43/102 2026-03-09T15:17:42.017 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-cffi-1.14.5-5.el9.x86_64 44/102 2026-03-09T15:17:42.017 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-chardet-4.0.0-5.el9.noarch 45/102 2026-03-09T15:17:42.017 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-cheroot-10.0.1-4.el9.noarch 46/102 2026-03-09T15:17:42.017 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-cherrypy-18.6.1-2.el9.noarch 47/102 2026-03-09T15:17:42.017 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-cryptography-36.0.1-5.el9.x86_64 48/102 2026-03-09T15:17:42.017 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-devel-3.9.25-3.el9.x86_64 49/102 2026-03-09T15:17:42.017 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-google-auth-1:2.45.0-1.el9.noarch 50/102 2026-03-09T15:17:42.017 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-grpcio-1.46.7-10.el9.x86_64 51/102 2026-03-09T15:17:42.017 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-grpcio-tools-1.46.7-10.el9.x86_64 52/102 2026-03-09T15:17:42.017 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-idna-2.10-7.el9.1.noarch 53/102 2026-03-09T15:17:42.017 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-jaraco-8.2.1-3.el9.noarch 54/102 2026-03-09T15:17:42.017 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-jaraco-classes-3.2.1-5.el9.noarch 55/102 2026-03-09T15:17:42.017 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-jaraco-collections-3.0.0-8.el9.noarch 56/102 2026-03-09T15:17:42.017 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-jaraco-context-6.0.1-3.el9.noarch 57/102 2026-03-09T15:17:42.018 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-jaraco-functools-3.5.0-2.el9.noarch 58/102 2026-03-09T15:17:42.018 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-jaraco-text-4.0.0-2.el9.noarch 59/102 2026-03-09T15:17:42.018 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-jinja2-2.11.3-8.el9.noarch 60/102 2026-03-09T15:17:42.018 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-jsonpatch-1.21-16.el9.noarch 61/102 2026-03-09T15:17:42.018 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-jsonpointer-2.0-4.el9.noarch 62/102 2026-03-09T15:17:42.018 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-kubernetes-1:26.1.0-3.el9.noarch 63/102 2026-03-09T15:17:42.018 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-libstoragemgmt-1.10.1-1.el9.x86_64 64/102 2026-03-09T15:17:42.018 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-logutils-0.3.5-21.el9.noarch 65/102 2026-03-09T15:17:42.018 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-mako-1.1.4-6.el9.noarch 66/102 2026-03-09T15:17:42.018 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-markupsafe-1.1.1-12.el9.x86_64 67/102 2026-03-09T15:17:42.018 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-more-itertools-8.12.0-2.el9.noarch 68/102 2026-03-09T15:17:42.018 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-natsort-7.1.1-5.el9.noarch 69/102 2026-03-09T15:17:42.018 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-numpy-1:1.23.5-2.el9.x86_64 70/102 2026-03-09T15:17:42.018 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-numpy-f2py-1:1.23.5-2.el9.x86_64 71/102 2026-03-09T15:17:42.018 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-oauthlib-3.1.1-5.el9.noarch 72/102 2026-03-09T15:17:42.018 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-packaging-20.9-5.el9.noarch 73/102 2026-03-09T15:17:42.018 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-pecan-1.4.2-3.el9.noarch 74/102 2026-03-09T15:17:42.018 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-ply-3.11-14.el9.noarch 75/102 2026-03-09T15:17:42.018 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-portend-3.1.0-2.el9.noarch 76/102 2026-03-09T15:17:42.018 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-prettytable-0.7.2-27.el9.noarch 77/102 2026-03-09T15:17:42.018 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-protobuf-3.14.0-17.el9.noarch 78/102 2026-03-09T15:17:42.018 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-pyOpenSSL-21.0.0-1.el9.noarch 79/102 2026-03-09T15:17:42.018 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-pyasn1-0.4.8-7.el9.noarch 80/102 2026-03-09T15:17:42.018 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-pyasn1-modules-0.4.8-7.el9.noarch 81/102 2026-03-09T15:17:42.018 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-pycparser-2.20-6.el9.noarch 82/102 2026-03-09T15:17:42.018 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-pysocks-1.7.1-12.el9.noarch 83/102 2026-03-09T15:17:42.018 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-pytz-2021.1-5.el9.noarch 84/102 2026-03-09T15:17:42.018 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-repoze-lru-0.7-16.el9.noarch 85/102 2026-03-09T15:17:42.018 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-requests-2.25.1-10.el9.noarch 86/102 2026-03-09T15:17:42.018 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-requests-oauthlib-1.3.0-12.el9.noarch 87/102 2026-03-09T15:17:42.018 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-routes-2.5.1-5.el9.noarch 88/102 2026-03-09T15:17:42.018 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-rsa-4.9-2.el9.noarch 89/102 2026-03-09T15:17:42.018 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-scipy-1.9.3-2.el9.x86_64 90/102 2026-03-09T15:17:42.018 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-tempora-5.0.0-2.el9.noarch 91/102 2026-03-09T15:17:42.018 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-toml-0.10.2-6.el9.noarch 92/102 2026-03-09T15:17:42.018 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-typing-extensions-4.15.0-1.el9.noarch 93/102 2026-03-09T15:17:42.018 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-urllib3-1.26.5-7.el9.noarch 94/102 2026-03-09T15:17:42.018 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-webob-1.8.8-2.el9.noarch 95/102 2026-03-09T15:17:42.018 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-websocket-client-1.2.3-2.el9.noarch 96/102 2026-03-09T15:17:42.018 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-werkzeug-2.0.3-3.el9.1.noarch 97/102 2026-03-09T15:17:42.018 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-zc-lockfile-2.0-10.el9.noarch 98/102 2026-03-09T15:17:42.018 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : qatlib-25.08.0-2.el9.x86_64 99/102 2026-03-09T15:17:42.018 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : qatlib-service-25.08.0-2.el9.x86_64 100/102 2026-03-09T15:17:42.018 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : qatzip-libs-1.3.1-1.el9.x86_64 101/102 2026-03-09T15:17:42.026 INFO:teuthology.orchestra.run.vm06.stdout: Running scriptlet: libcephsqlite-2:19.2.3-678.ge911bdeb.el9.x86_64 102/102 2026-03-09T15:17:42.026 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : abseil-cpp-20211102.0-4.el9.x86_64 1/102 2026-03-09T15:17:42.026 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : ceph-base-2:19.2.3-678.ge911bdeb.el9.x86_64 2/102 2026-03-09T15:17:42.026 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : ceph-common-2:19.2.3-678.ge911bdeb.el9.x86_64 3/102 2026-03-09T15:17:42.026 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : ceph-grafana-dashboards-2:19.2.3-678.ge911bdeb.e 4/102 2026-03-09T15:17:42.026 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : ceph-immutable-object-cache-2:19.2.3-678.ge911bd 5/102 2026-03-09T15:17:42.026 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : ceph-mgr-2:19.2.3-678.ge911bdeb.el9.x86_64 6/102 2026-03-09T15:17:42.026 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : ceph-mgr-cephadm-2:19.2.3-678.ge911bdeb.el9.noar 7/102 2026-03-09T15:17:42.026 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : ceph-mgr-dashboard-2:19.2.3-678.ge911bdeb.el9.no 8/102 2026-03-09T15:17:42.026 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : ceph-mgr-diskprediction-local-2:19.2.3-678.ge911 9/102 2026-03-09T15:17:42.027 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : ceph-mgr-modules-core-2:19.2.3-678.ge911bdeb.el9 10/102 2026-03-09T15:17:42.027 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : ceph-mgr-rook-2:19.2.3-678.ge911bdeb.el9.noarch 11/102 2026-03-09T15:17:42.027 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : ceph-osd-2:19.2.3-678.ge911bdeb.el9.x86_64 12/102 2026-03-09T15:17:42.027 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : ceph-prometheus-alerts-2:19.2.3-678.ge911bdeb.el 13/102 2026-03-09T15:17:42.027 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : ceph-selinux-2:19.2.3-678.ge911bdeb.el9.x86_64 14/102 2026-03-09T15:17:42.027 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : ceph-volume-2:19.2.3-678.ge911bdeb.el9.noarch 15/102 2026-03-09T15:17:42.027 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : cryptsetup-2.8.1-3.el9.x86_64 16/102 2026-03-09T15:17:42.027 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : flexiblas-3.0.4-9.el9.x86_64 17/102 2026-03-09T15:17:42.027 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : flexiblas-netlib-3.0.4-9.el9.x86_64 18/102 2026-03-09T15:17:42.027 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : flexiblas-openblas-openmp-3.0.4-9.el9.x86_64 19/102 2026-03-09T15:17:42.027 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : gperftools-libs-2.9.1-3.el9.x86_64 20/102 2026-03-09T15:17:42.027 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : grpc-data-1.46.7-10.el9.noarch 21/102 2026-03-09T15:17:42.027 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : ledmon-libs-1.1.0-3.el9.x86_64 22/102 2026-03-09T15:17:42.027 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : libcephsqlite-2:19.2.3-678.ge911bdeb.el9.x86_64 23/102 2026-03-09T15:17:42.027 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : libconfig-1.7.2-9.el9.x86_64 24/102 2026-03-09T15:17:42.027 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : libgfortran-11.5.0-14.el9.x86_64 25/102 2026-03-09T15:17:42.027 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : liboath-2.6.12-1.el9.x86_64 26/102 2026-03-09T15:17:42.027 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : libquadmath-11.5.0-14.el9.x86_64 27/102 2026-03-09T15:17:42.027 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : libradosstriper1-2:19.2.3-678.ge911bdeb.el9.x86_ 28/102 2026-03-09T15:17:42.027 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : libstoragemgmt-1.10.1-1.el9.x86_64 29/102 2026-03-09T15:17:42.027 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : libunwind-1.6.2-1.el9.x86_64 30/102 2026-03-09T15:17:42.027 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : openblas-0.3.29-1.el9.x86_64 31/102 2026-03-09T15:17:42.028 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : openblas-openmp-0.3.29-1.el9.x86_64 32/102 2026-03-09T15:17:42.028 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : pciutils-3.7.0-7.el9.x86_64 33/102 2026-03-09T15:17:42.028 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : protobuf-3.14.0-17.el9.x86_64 34/102 2026-03-09T15:17:42.028 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : protobuf-compiler-3.14.0-17.el9.x86_64 35/102 2026-03-09T15:17:42.028 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-asyncssh-2.13.2-5.el9.noarch 36/102 2026-03-09T15:17:42.028 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-autocommand-2.2.2-8.el9.noarch 37/102 2026-03-09T15:17:42.028 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-babel-2.9.1-2.el9.noarch 38/102 2026-03-09T15:17:42.028 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-backports-tarfile-1.2.0-1.el9.noarch 39/102 2026-03-09T15:17:42.028 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-bcrypt-3.2.2-1.el9.x86_64 40/102 2026-03-09T15:17:42.028 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-cachetools-4.2.4-1.el9.noarch 41/102 2026-03-09T15:17:42.028 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-ceph-common-2:19.2.3-678.ge911bdeb.el9.x 42/102 2026-03-09T15:17:42.028 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-certifi-2023.05.07-4.el9.noarch 43/102 2026-03-09T15:17:42.028 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-cffi-1.14.5-5.el9.x86_64 44/102 2026-03-09T15:17:42.028 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-chardet-4.0.0-5.el9.noarch 45/102 2026-03-09T15:17:42.028 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-cheroot-10.0.1-4.el9.noarch 46/102 2026-03-09T15:17:42.028 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-cherrypy-18.6.1-2.el9.noarch 47/102 2026-03-09T15:17:42.028 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-cryptography-36.0.1-5.el9.x86_64 48/102 2026-03-09T15:17:42.028 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-devel-3.9.25-3.el9.x86_64 49/102 2026-03-09T15:17:42.028 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-google-auth-1:2.45.0-1.el9.noarch 50/102 2026-03-09T15:17:42.028 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-grpcio-1.46.7-10.el9.x86_64 51/102 2026-03-09T15:17:42.028 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-grpcio-tools-1.46.7-10.el9.x86_64 52/102 2026-03-09T15:17:42.028 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-idna-2.10-7.el9.1.noarch 53/102 2026-03-09T15:17:42.028 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-jaraco-8.2.1-3.el9.noarch 54/102 2026-03-09T15:17:42.028 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-jaraco-classes-3.2.1-5.el9.noarch 55/102 2026-03-09T15:17:42.028 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-jaraco-collections-3.0.0-8.el9.noarch 56/102 2026-03-09T15:17:42.028 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-jaraco-context-6.0.1-3.el9.noarch 57/102 2026-03-09T15:17:42.028 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-jaraco-functools-3.5.0-2.el9.noarch 58/102 2026-03-09T15:17:42.029 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-jaraco-text-4.0.0-2.el9.noarch 59/102 2026-03-09T15:17:42.029 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-jinja2-2.11.3-8.el9.noarch 60/102 2026-03-09T15:17:42.029 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-jsonpatch-1.21-16.el9.noarch 61/102 2026-03-09T15:17:42.029 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-jsonpointer-2.0-4.el9.noarch 62/102 2026-03-09T15:17:42.029 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-kubernetes-1:26.1.0-3.el9.noarch 63/102 2026-03-09T15:17:42.029 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-libstoragemgmt-1.10.1-1.el9.x86_64 64/102 2026-03-09T15:17:42.029 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-logutils-0.3.5-21.el9.noarch 65/102 2026-03-09T15:17:42.029 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-mako-1.1.4-6.el9.noarch 66/102 2026-03-09T15:17:42.029 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-markupsafe-1.1.1-12.el9.x86_64 67/102 2026-03-09T15:17:42.029 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-more-itertools-8.12.0-2.el9.noarch 68/102 2026-03-09T15:17:42.029 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-natsort-7.1.1-5.el9.noarch 69/102 2026-03-09T15:17:42.029 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-numpy-1:1.23.5-2.el9.x86_64 70/102 2026-03-09T15:17:42.029 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-numpy-f2py-1:1.23.5-2.el9.x86_64 71/102 2026-03-09T15:17:42.029 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-oauthlib-3.1.1-5.el9.noarch 72/102 2026-03-09T15:17:42.029 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-packaging-20.9-5.el9.noarch 73/102 2026-03-09T15:17:42.029 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-pecan-1.4.2-3.el9.noarch 74/102 2026-03-09T15:17:42.029 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-ply-3.11-14.el9.noarch 75/102 2026-03-09T15:17:42.029 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-portend-3.1.0-2.el9.noarch 76/102 2026-03-09T15:17:42.029 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-prettytable-0.7.2-27.el9.noarch 77/102 2026-03-09T15:17:42.029 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-protobuf-3.14.0-17.el9.noarch 78/102 2026-03-09T15:17:42.029 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-pyOpenSSL-21.0.0-1.el9.noarch 79/102 2026-03-09T15:17:42.029 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-pyasn1-0.4.8-7.el9.noarch 80/102 2026-03-09T15:17:42.029 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-pyasn1-modules-0.4.8-7.el9.noarch 81/102 2026-03-09T15:17:42.029 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-pycparser-2.20-6.el9.noarch 82/102 2026-03-09T15:17:42.029 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-pysocks-1.7.1-12.el9.noarch 83/102 2026-03-09T15:17:42.029 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-pytz-2021.1-5.el9.noarch 84/102 2026-03-09T15:17:42.029 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-repoze-lru-0.7-16.el9.noarch 85/102 2026-03-09T15:17:42.029 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-requests-2.25.1-10.el9.noarch 86/102 2026-03-09T15:17:42.029 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-requests-oauthlib-1.3.0-12.el9.noarch 87/102 2026-03-09T15:17:42.029 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-routes-2.5.1-5.el9.noarch 88/102 2026-03-09T15:17:42.029 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-rsa-4.9-2.el9.noarch 89/102 2026-03-09T15:17:42.029 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-scipy-1.9.3-2.el9.x86_64 90/102 2026-03-09T15:17:42.030 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-tempora-5.0.0-2.el9.noarch 91/102 2026-03-09T15:17:42.030 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-toml-0.10.2-6.el9.noarch 92/102 2026-03-09T15:17:42.030 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-typing-extensions-4.15.0-1.el9.noarch 93/102 2026-03-09T15:17:42.030 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-urllib3-1.26.5-7.el9.noarch 94/102 2026-03-09T15:17:42.030 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-webob-1.8.8-2.el9.noarch 95/102 2026-03-09T15:17:42.030 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-websocket-client-1.2.3-2.el9.noarch 96/102 2026-03-09T15:17:42.030 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-werkzeug-2.0.3-3.el9.1.noarch 97/102 2026-03-09T15:17:42.030 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-zc-lockfile-2.0-10.el9.noarch 98/102 2026-03-09T15:17:42.030 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : qatlib-25.08.0-2.el9.x86_64 99/102 2026-03-09T15:17:42.030 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : qatlib-service-25.08.0-2.el9.x86_64 100/102 2026-03-09T15:17:42.030 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : qatzip-libs-1.3.1-1.el9.x86_64 101/102 2026-03-09T15:17:42.095 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : rbd-mirror-2:19.2.3-678.ge911bdeb.el9.x86_64 102/102 2026-03-09T15:17:42.095 INFO:teuthology.orchestra.run.vm04.stdout: 2026-03-09T15:17:42.095 INFO:teuthology.orchestra.run.vm04.stdout:Removed: 2026-03-09T15:17:42.095 INFO:teuthology.orchestra.run.vm04.stdout: abseil-cpp-20211102.0-4.el9.x86_64 2026-03-09T15:17:42.095 INFO:teuthology.orchestra.run.vm04.stdout: ceph-base-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T15:17:42.095 INFO:teuthology.orchestra.run.vm04.stdout: ceph-common-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T15:17:42.095 INFO:teuthology.orchestra.run.vm04.stdout: ceph-grafana-dashboards-2:19.2.3-678.ge911bdeb.el9.noarch 2026-03-09T15:17:42.095 INFO:teuthology.orchestra.run.vm04.stdout: ceph-immutable-object-cache-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T15:17:42.095 INFO:teuthology.orchestra.run.vm04.stdout: ceph-mgr-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T15:17:42.095 INFO:teuthology.orchestra.run.vm04.stdout: ceph-mgr-cephadm-2:19.2.3-678.ge911bdeb.el9.noarch 2026-03-09T15:17:42.095 INFO:teuthology.orchestra.run.vm04.stdout: ceph-mgr-dashboard-2:19.2.3-678.ge911bdeb.el9.noarch 2026-03-09T15:17:42.095 INFO:teuthology.orchestra.run.vm04.stdout: ceph-mgr-diskprediction-local-2:19.2.3-678.ge911bdeb.el9.noarch 2026-03-09T15:17:42.095 INFO:teuthology.orchestra.run.vm04.stdout: ceph-mgr-modules-core-2:19.2.3-678.ge911bdeb.el9.noarch 2026-03-09T15:17:42.095 INFO:teuthology.orchestra.run.vm04.stdout: ceph-mgr-rook-2:19.2.3-678.ge911bdeb.el9.noarch 2026-03-09T15:17:42.095 INFO:teuthology.orchestra.run.vm04.stdout: ceph-osd-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T15:17:42.095 INFO:teuthology.orchestra.run.vm04.stdout: ceph-prometheus-alerts-2:19.2.3-678.ge911bdeb.el9.noarch 2026-03-09T15:17:42.095 INFO:teuthology.orchestra.run.vm04.stdout: ceph-selinux-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T15:17:42.095 INFO:teuthology.orchestra.run.vm04.stdout: ceph-volume-2:19.2.3-678.ge911bdeb.el9.noarch 2026-03-09T15:17:42.095 INFO:teuthology.orchestra.run.vm04.stdout: cryptsetup-2.8.1-3.el9.x86_64 2026-03-09T15:17:42.095 INFO:teuthology.orchestra.run.vm04.stdout: flexiblas-3.0.4-9.el9.x86_64 2026-03-09T15:17:42.095 INFO:teuthology.orchestra.run.vm04.stdout: flexiblas-netlib-3.0.4-9.el9.x86_64 2026-03-09T15:17:42.095 INFO:teuthology.orchestra.run.vm04.stdout: flexiblas-openblas-openmp-3.0.4-9.el9.x86_64 2026-03-09T15:17:42.095 INFO:teuthology.orchestra.run.vm04.stdout: gperftools-libs-2.9.1-3.el9.x86_64 2026-03-09T15:17:42.095 INFO:teuthology.orchestra.run.vm04.stdout: grpc-data-1.46.7-10.el9.noarch 2026-03-09T15:17:42.095 INFO:teuthology.orchestra.run.vm04.stdout: ledmon-libs-1.1.0-3.el9.x86_64 2026-03-09T15:17:42.096 INFO:teuthology.orchestra.run.vm04.stdout: libcephsqlite-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T15:17:42.096 INFO:teuthology.orchestra.run.vm04.stdout: libconfig-1.7.2-9.el9.x86_64 2026-03-09T15:17:42.096 INFO:teuthology.orchestra.run.vm04.stdout: libgfortran-11.5.0-14.el9.x86_64 2026-03-09T15:17:42.096 INFO:teuthology.orchestra.run.vm04.stdout: liboath-2.6.12-1.el9.x86_64 2026-03-09T15:17:42.096 INFO:teuthology.orchestra.run.vm04.stdout: libquadmath-11.5.0-14.el9.x86_64 2026-03-09T15:17:42.096 INFO:teuthology.orchestra.run.vm04.stdout: libradosstriper1-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T15:17:42.096 INFO:teuthology.orchestra.run.vm04.stdout: libstoragemgmt-1.10.1-1.el9.x86_64 2026-03-09T15:17:42.096 INFO:teuthology.orchestra.run.vm04.stdout: libunwind-1.6.2-1.el9.x86_64 2026-03-09T15:17:42.096 INFO:teuthology.orchestra.run.vm04.stdout: openblas-0.3.29-1.el9.x86_64 2026-03-09T15:17:42.096 INFO:teuthology.orchestra.run.vm04.stdout: openblas-openmp-0.3.29-1.el9.x86_64 2026-03-09T15:17:42.096 INFO:teuthology.orchestra.run.vm04.stdout: pciutils-3.7.0-7.el9.x86_64 2026-03-09T15:17:42.096 INFO:teuthology.orchestra.run.vm04.stdout: protobuf-3.14.0-17.el9.x86_64 2026-03-09T15:17:42.096 INFO:teuthology.orchestra.run.vm04.stdout: protobuf-compiler-3.14.0-17.el9.x86_64 2026-03-09T15:17:42.096 INFO:teuthology.orchestra.run.vm04.stdout: python3-asyncssh-2.13.2-5.el9.noarch 2026-03-09T15:17:42.096 INFO:teuthology.orchestra.run.vm04.stdout: python3-autocommand-2.2.2-8.el9.noarch 2026-03-09T15:17:42.096 INFO:teuthology.orchestra.run.vm04.stdout: python3-babel-2.9.1-2.el9.noarch 2026-03-09T15:17:42.096 INFO:teuthology.orchestra.run.vm04.stdout: python3-backports-tarfile-1.2.0-1.el9.noarch 2026-03-09T15:17:42.096 INFO:teuthology.orchestra.run.vm04.stdout: python3-bcrypt-3.2.2-1.el9.x86_64 2026-03-09T15:17:42.096 INFO:teuthology.orchestra.run.vm04.stdout: python3-cachetools-4.2.4-1.el9.noarch 2026-03-09T15:17:42.096 INFO:teuthology.orchestra.run.vm04.stdout: python3-ceph-common-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T15:17:42.096 INFO:teuthology.orchestra.run.vm04.stdout: python3-certifi-2023.05.07-4.el9.noarch 2026-03-09T15:17:42.096 INFO:teuthology.orchestra.run.vm04.stdout: python3-cffi-1.14.5-5.el9.x86_64 2026-03-09T15:17:42.096 INFO:teuthology.orchestra.run.vm04.stdout: python3-chardet-4.0.0-5.el9.noarch 2026-03-09T15:17:42.096 INFO:teuthology.orchestra.run.vm04.stdout: python3-cheroot-10.0.1-4.el9.noarch 2026-03-09T15:17:42.096 INFO:teuthology.orchestra.run.vm04.stdout: python3-cherrypy-18.6.1-2.el9.noarch 2026-03-09T15:17:42.096 INFO:teuthology.orchestra.run.vm04.stdout: python3-cryptography-36.0.1-5.el9.x86_64 2026-03-09T15:17:42.096 INFO:teuthology.orchestra.run.vm04.stdout: python3-devel-3.9.25-3.el9.x86_64 2026-03-09T15:17:42.096 INFO:teuthology.orchestra.run.vm04.stdout: python3-google-auth-1:2.45.0-1.el9.noarch 2026-03-09T15:17:42.096 INFO:teuthology.orchestra.run.vm04.stdout: python3-grpcio-1.46.7-10.el9.x86_64 2026-03-09T15:17:42.096 INFO:teuthology.orchestra.run.vm04.stdout: python3-grpcio-tools-1.46.7-10.el9.x86_64 2026-03-09T15:17:42.096 INFO:teuthology.orchestra.run.vm04.stdout: python3-idna-2.10-7.el9.1.noarch 2026-03-09T15:17:42.096 INFO:teuthology.orchestra.run.vm04.stdout: python3-jaraco-8.2.1-3.el9.noarch 2026-03-09T15:17:42.096 INFO:teuthology.orchestra.run.vm04.stdout: python3-jaraco-classes-3.2.1-5.el9.noarch 2026-03-09T15:17:42.096 INFO:teuthology.orchestra.run.vm04.stdout: python3-jaraco-collections-3.0.0-8.el9.noarch 2026-03-09T15:17:42.096 INFO:teuthology.orchestra.run.vm04.stdout: python3-jaraco-context-6.0.1-3.el9.noarch 2026-03-09T15:17:42.096 INFO:teuthology.orchestra.run.vm04.stdout: python3-jaraco-functools-3.5.0-2.el9.noarch 2026-03-09T15:17:42.096 INFO:teuthology.orchestra.run.vm04.stdout: python3-jaraco-text-4.0.0-2.el9.noarch 2026-03-09T15:17:42.096 INFO:teuthology.orchestra.run.vm04.stdout: python3-jinja2-2.11.3-8.el9.noarch 2026-03-09T15:17:42.096 INFO:teuthology.orchestra.run.vm04.stdout: python3-jsonpatch-1.21-16.el9.noarch 2026-03-09T15:17:42.096 INFO:teuthology.orchestra.run.vm04.stdout: python3-jsonpointer-2.0-4.el9.noarch 2026-03-09T15:17:42.096 INFO:teuthology.orchestra.run.vm04.stdout: python3-kubernetes-1:26.1.0-3.el9.noarch 2026-03-09T15:17:42.097 INFO:teuthology.orchestra.run.vm04.stdout: python3-libstoragemgmt-1.10.1-1.el9.x86_64 2026-03-09T15:17:42.097 INFO:teuthology.orchestra.run.vm04.stdout: python3-logutils-0.3.5-21.el9.noarch 2026-03-09T15:17:42.097 INFO:teuthology.orchestra.run.vm04.stdout: python3-mako-1.1.4-6.el9.noarch 2026-03-09T15:17:42.097 INFO:teuthology.orchestra.run.vm04.stdout: python3-markupsafe-1.1.1-12.el9.x86_64 2026-03-09T15:17:42.097 INFO:teuthology.orchestra.run.vm04.stdout: python3-more-itertools-8.12.0-2.el9.noarch 2026-03-09T15:17:42.097 INFO:teuthology.orchestra.run.vm04.stdout: python3-natsort-7.1.1-5.el9.noarch 2026-03-09T15:17:42.097 INFO:teuthology.orchestra.run.vm04.stdout: python3-numpy-1:1.23.5-2.el9.x86_64 2026-03-09T15:17:42.097 INFO:teuthology.orchestra.run.vm04.stdout: python3-numpy-f2py-1:1.23.5-2.el9.x86_64 2026-03-09T15:17:42.097 INFO:teuthology.orchestra.run.vm04.stdout: python3-oauthlib-3.1.1-5.el9.noarch 2026-03-09T15:17:42.097 INFO:teuthology.orchestra.run.vm04.stdout: python3-packaging-20.9-5.el9.noarch 2026-03-09T15:17:42.097 INFO:teuthology.orchestra.run.vm04.stdout: python3-pecan-1.4.2-3.el9.noarch 2026-03-09T15:17:42.097 INFO:teuthology.orchestra.run.vm04.stdout: python3-ply-3.11-14.el9.noarch 2026-03-09T15:17:42.097 INFO:teuthology.orchestra.run.vm04.stdout: python3-portend-3.1.0-2.el9.noarch 2026-03-09T15:17:42.097 INFO:teuthology.orchestra.run.vm04.stdout: python3-prettytable-0.7.2-27.el9.noarch 2026-03-09T15:17:42.097 INFO:teuthology.orchestra.run.vm04.stdout: python3-protobuf-3.14.0-17.el9.noarch 2026-03-09T15:17:42.097 INFO:teuthology.orchestra.run.vm04.stdout: python3-pyOpenSSL-21.0.0-1.el9.noarch 2026-03-09T15:17:42.097 INFO:teuthology.orchestra.run.vm04.stdout: python3-pyasn1-0.4.8-7.el9.noarch 2026-03-09T15:17:42.097 INFO:teuthology.orchestra.run.vm04.stdout: python3-pyasn1-modules-0.4.8-7.el9.noarch 2026-03-09T15:17:42.097 INFO:teuthology.orchestra.run.vm04.stdout: python3-pycparser-2.20-6.el9.noarch 2026-03-09T15:17:42.097 INFO:teuthology.orchestra.run.vm04.stdout: python3-pysocks-1.7.1-12.el9.noarch 2026-03-09T15:17:42.097 INFO:teuthology.orchestra.run.vm04.stdout: python3-pytz-2021.1-5.el9.noarch 2026-03-09T15:17:42.097 INFO:teuthology.orchestra.run.vm04.stdout: python3-repoze-lru-0.7-16.el9.noarch 2026-03-09T15:17:42.097 INFO:teuthology.orchestra.run.vm04.stdout: python3-requests-2.25.1-10.el9.noarch 2026-03-09T15:17:42.097 INFO:teuthology.orchestra.run.vm04.stdout: python3-requests-oauthlib-1.3.0-12.el9.noarch 2026-03-09T15:17:42.097 INFO:teuthology.orchestra.run.vm04.stdout: python3-routes-2.5.1-5.el9.noarch 2026-03-09T15:17:42.097 INFO:teuthology.orchestra.run.vm04.stdout: python3-rsa-4.9-2.el9.noarch 2026-03-09T15:17:42.097 INFO:teuthology.orchestra.run.vm04.stdout: python3-scipy-1.9.3-2.el9.x86_64 2026-03-09T15:17:42.097 INFO:teuthology.orchestra.run.vm04.stdout: python3-tempora-5.0.0-2.el9.noarch 2026-03-09T15:17:42.097 INFO:teuthology.orchestra.run.vm04.stdout: python3-toml-0.10.2-6.el9.noarch 2026-03-09T15:17:42.097 INFO:teuthology.orchestra.run.vm04.stdout: python3-typing-extensions-4.15.0-1.el9.noarch 2026-03-09T15:17:42.097 INFO:teuthology.orchestra.run.vm04.stdout: python3-urllib3-1.26.5-7.el9.noarch 2026-03-09T15:17:42.097 INFO:teuthology.orchestra.run.vm04.stdout: python3-webob-1.8.8-2.el9.noarch 2026-03-09T15:17:42.097 INFO:teuthology.orchestra.run.vm04.stdout: python3-websocket-client-1.2.3-2.el9.noarch 2026-03-09T15:17:42.097 INFO:teuthology.orchestra.run.vm04.stdout: python3-werkzeug-2.0.3-3.el9.1.noarch 2026-03-09T15:17:42.097 INFO:teuthology.orchestra.run.vm04.stdout: python3-zc-lockfile-2.0-10.el9.noarch 2026-03-09T15:17:42.097 INFO:teuthology.orchestra.run.vm04.stdout: qatlib-25.08.0-2.el9.x86_64 2026-03-09T15:17:42.097 INFO:teuthology.orchestra.run.vm04.stdout: qatlib-service-25.08.0-2.el9.x86_64 2026-03-09T15:17:42.097 INFO:teuthology.orchestra.run.vm04.stdout: qatzip-libs-1.3.1-1.el9.x86_64 2026-03-09T15:17:42.098 INFO:teuthology.orchestra.run.vm04.stdout: rbd-mirror-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T15:17:42.098 INFO:teuthology.orchestra.run.vm04.stdout: 2026-03-09T15:17:42.098 INFO:teuthology.orchestra.run.vm04.stdout:Complete! 2026-03-09T15:17:42.100 INFO:teuthology.orchestra.run.vm03.stdout: Verifying : rbd-mirror-2:19.2.3-678.ge911bdeb.el9.x86_64 102/102 2026-03-09T15:17:42.100 INFO:teuthology.orchestra.run.vm03.stdout: 2026-03-09T15:17:42.100 INFO:teuthology.orchestra.run.vm03.stdout:Removed: 2026-03-09T15:17:42.100 INFO:teuthology.orchestra.run.vm03.stdout: abseil-cpp-20211102.0-4.el9.x86_64 2026-03-09T15:17:42.100 INFO:teuthology.orchestra.run.vm03.stdout: ceph-base-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T15:17:42.100 INFO:teuthology.orchestra.run.vm03.stdout: ceph-common-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T15:17:42.100 INFO:teuthology.orchestra.run.vm03.stdout: ceph-grafana-dashboards-2:19.2.3-678.ge911bdeb.el9.noarch 2026-03-09T15:17:42.100 INFO:teuthology.orchestra.run.vm03.stdout: ceph-immutable-object-cache-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T15:17:42.100 INFO:teuthology.orchestra.run.vm03.stdout: ceph-mgr-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T15:17:42.100 INFO:teuthology.orchestra.run.vm03.stdout: ceph-mgr-cephadm-2:19.2.3-678.ge911bdeb.el9.noarch 2026-03-09T15:17:42.100 INFO:teuthology.orchestra.run.vm03.stdout: ceph-mgr-dashboard-2:19.2.3-678.ge911bdeb.el9.noarch 2026-03-09T15:17:42.100 INFO:teuthology.orchestra.run.vm03.stdout: ceph-mgr-diskprediction-local-2:19.2.3-678.ge911bdeb.el9.noarch 2026-03-09T15:17:42.100 INFO:teuthology.orchestra.run.vm03.stdout: ceph-mgr-modules-core-2:19.2.3-678.ge911bdeb.el9.noarch 2026-03-09T15:17:42.100 INFO:teuthology.orchestra.run.vm03.stdout: ceph-mgr-rook-2:19.2.3-678.ge911bdeb.el9.noarch 2026-03-09T15:17:42.100 INFO:teuthology.orchestra.run.vm03.stdout: ceph-osd-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T15:17:42.100 INFO:teuthology.orchestra.run.vm03.stdout: ceph-prometheus-alerts-2:19.2.3-678.ge911bdeb.el9.noarch 2026-03-09T15:17:42.100 INFO:teuthology.orchestra.run.vm03.stdout: ceph-selinux-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T15:17:42.100 INFO:teuthology.orchestra.run.vm03.stdout: ceph-volume-2:19.2.3-678.ge911bdeb.el9.noarch 2026-03-09T15:17:42.100 INFO:teuthology.orchestra.run.vm03.stdout: cryptsetup-2.8.1-3.el9.x86_64 2026-03-09T15:17:42.100 INFO:teuthology.orchestra.run.vm03.stdout: flexiblas-3.0.4-9.el9.x86_64 2026-03-09T15:17:42.100 INFO:teuthology.orchestra.run.vm03.stdout: flexiblas-netlib-3.0.4-9.el9.x86_64 2026-03-09T15:17:42.100 INFO:teuthology.orchestra.run.vm03.stdout: flexiblas-openblas-openmp-3.0.4-9.el9.x86_64 2026-03-09T15:17:42.100 INFO:teuthology.orchestra.run.vm03.stdout: gperftools-libs-2.9.1-3.el9.x86_64 2026-03-09T15:17:42.100 INFO:teuthology.orchestra.run.vm03.stdout: grpc-data-1.46.7-10.el9.noarch 2026-03-09T15:17:42.100 INFO:teuthology.orchestra.run.vm03.stdout: ledmon-libs-1.1.0-3.el9.x86_64 2026-03-09T15:17:42.100 INFO:teuthology.orchestra.run.vm03.stdout: libcephsqlite-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T15:17:42.100 INFO:teuthology.orchestra.run.vm03.stdout: libconfig-1.7.2-9.el9.x86_64 2026-03-09T15:17:42.100 INFO:teuthology.orchestra.run.vm03.stdout: libgfortran-11.5.0-14.el9.x86_64 2026-03-09T15:17:42.100 INFO:teuthology.orchestra.run.vm03.stdout: liboath-2.6.12-1.el9.x86_64 2026-03-09T15:17:42.100 INFO:teuthology.orchestra.run.vm03.stdout: libquadmath-11.5.0-14.el9.x86_64 2026-03-09T15:17:42.100 INFO:teuthology.orchestra.run.vm03.stdout: libradosstriper1-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T15:17:42.100 INFO:teuthology.orchestra.run.vm03.stdout: libstoragemgmt-1.10.1-1.el9.x86_64 2026-03-09T15:17:42.100 INFO:teuthology.orchestra.run.vm03.stdout: libunwind-1.6.2-1.el9.x86_64 2026-03-09T15:17:42.100 INFO:teuthology.orchestra.run.vm03.stdout: openblas-0.3.29-1.el9.x86_64 2026-03-09T15:17:42.100 INFO:teuthology.orchestra.run.vm03.stdout: openblas-openmp-0.3.29-1.el9.x86_64 2026-03-09T15:17:42.100 INFO:teuthology.orchestra.run.vm03.stdout: pciutils-3.7.0-7.el9.x86_64 2026-03-09T15:17:42.100 INFO:teuthology.orchestra.run.vm03.stdout: protobuf-3.14.0-17.el9.x86_64 2026-03-09T15:17:42.100 INFO:teuthology.orchestra.run.vm03.stdout: protobuf-compiler-3.14.0-17.el9.x86_64 2026-03-09T15:17:42.100 INFO:teuthology.orchestra.run.vm03.stdout: python3-asyncssh-2.13.2-5.el9.noarch 2026-03-09T15:17:42.100 INFO:teuthology.orchestra.run.vm03.stdout: python3-autocommand-2.2.2-8.el9.noarch 2026-03-09T15:17:42.100 INFO:teuthology.orchestra.run.vm03.stdout: python3-babel-2.9.1-2.el9.noarch 2026-03-09T15:17:42.100 INFO:teuthology.orchestra.run.vm03.stdout: python3-backports-tarfile-1.2.0-1.el9.noarch 2026-03-09T15:17:42.100 INFO:teuthology.orchestra.run.vm03.stdout: python3-bcrypt-3.2.2-1.el9.x86_64 2026-03-09T15:17:42.101 INFO:teuthology.orchestra.run.vm03.stdout: python3-cachetools-4.2.4-1.el9.noarch 2026-03-09T15:17:42.101 INFO:teuthology.orchestra.run.vm03.stdout: python3-ceph-common-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T15:17:42.101 INFO:teuthology.orchestra.run.vm03.stdout: python3-certifi-2023.05.07-4.el9.noarch 2026-03-09T15:17:42.101 INFO:teuthology.orchestra.run.vm03.stdout: python3-cffi-1.14.5-5.el9.x86_64 2026-03-09T15:17:42.101 INFO:teuthology.orchestra.run.vm03.stdout: python3-chardet-4.0.0-5.el9.noarch 2026-03-09T15:17:42.101 INFO:teuthology.orchestra.run.vm03.stdout: python3-cheroot-10.0.1-4.el9.noarch 2026-03-09T15:17:42.101 INFO:teuthology.orchestra.run.vm03.stdout: python3-cherrypy-18.6.1-2.el9.noarch 2026-03-09T15:17:42.101 INFO:teuthology.orchestra.run.vm03.stdout: python3-cryptography-36.0.1-5.el9.x86_64 2026-03-09T15:17:42.101 INFO:teuthology.orchestra.run.vm03.stdout: python3-devel-3.9.25-3.el9.x86_64 2026-03-09T15:17:42.101 INFO:teuthology.orchestra.run.vm03.stdout: python3-google-auth-1:2.45.0-1.el9.noarch 2026-03-09T15:17:42.101 INFO:teuthology.orchestra.run.vm03.stdout: python3-grpcio-1.46.7-10.el9.x86_64 2026-03-09T15:17:42.101 INFO:teuthology.orchestra.run.vm03.stdout: python3-grpcio-tools-1.46.7-10.el9.x86_64 2026-03-09T15:17:42.101 INFO:teuthology.orchestra.run.vm03.stdout: python3-idna-2.10-7.el9.1.noarch 2026-03-09T15:17:42.101 INFO:teuthology.orchestra.run.vm03.stdout: python3-jaraco-8.2.1-3.el9.noarch 2026-03-09T15:17:42.101 INFO:teuthology.orchestra.run.vm03.stdout: python3-jaraco-classes-3.2.1-5.el9.noarch 2026-03-09T15:17:42.101 INFO:teuthology.orchestra.run.vm03.stdout: python3-jaraco-collections-3.0.0-8.el9.noarch 2026-03-09T15:17:42.101 INFO:teuthology.orchestra.run.vm03.stdout: python3-jaraco-context-6.0.1-3.el9.noarch 2026-03-09T15:17:42.101 INFO:teuthology.orchestra.run.vm03.stdout: python3-jaraco-functools-3.5.0-2.el9.noarch 2026-03-09T15:17:42.101 INFO:teuthology.orchestra.run.vm03.stdout: python3-jaraco-text-4.0.0-2.el9.noarch 2026-03-09T15:17:42.101 INFO:teuthology.orchestra.run.vm03.stdout: python3-jinja2-2.11.3-8.el9.noarch 2026-03-09T15:17:42.101 INFO:teuthology.orchestra.run.vm03.stdout: python3-jsonpatch-1.21-16.el9.noarch 2026-03-09T15:17:42.101 INFO:teuthology.orchestra.run.vm03.stdout: python3-jsonpointer-2.0-4.el9.noarch 2026-03-09T15:17:42.101 INFO:teuthology.orchestra.run.vm03.stdout: python3-kubernetes-1:26.1.0-3.el9.noarch 2026-03-09T15:17:42.101 INFO:teuthology.orchestra.run.vm03.stdout: python3-libstoragemgmt-1.10.1-1.el9.x86_64 2026-03-09T15:17:42.101 INFO:teuthology.orchestra.run.vm03.stdout: python3-logutils-0.3.5-21.el9.noarch 2026-03-09T15:17:42.101 INFO:teuthology.orchestra.run.vm03.stdout: python3-mako-1.1.4-6.el9.noarch 2026-03-09T15:17:42.101 INFO:teuthology.orchestra.run.vm03.stdout: python3-markupsafe-1.1.1-12.el9.x86_64 2026-03-09T15:17:42.101 INFO:teuthology.orchestra.run.vm03.stdout: python3-more-itertools-8.12.0-2.el9.noarch 2026-03-09T15:17:42.101 INFO:teuthology.orchestra.run.vm03.stdout: python3-natsort-7.1.1-5.el9.noarch 2026-03-09T15:17:42.101 INFO:teuthology.orchestra.run.vm03.stdout: python3-numpy-1:1.23.5-2.el9.x86_64 2026-03-09T15:17:42.101 INFO:teuthology.orchestra.run.vm03.stdout: python3-numpy-f2py-1:1.23.5-2.el9.x86_64 2026-03-09T15:17:42.101 INFO:teuthology.orchestra.run.vm03.stdout: python3-oauthlib-3.1.1-5.el9.noarch 2026-03-09T15:17:42.101 INFO:teuthology.orchestra.run.vm03.stdout: python3-packaging-20.9-5.el9.noarch 2026-03-09T15:17:42.101 INFO:teuthology.orchestra.run.vm03.stdout: python3-pecan-1.4.2-3.el9.noarch 2026-03-09T15:17:42.101 INFO:teuthology.orchestra.run.vm03.stdout: python3-ply-3.11-14.el9.noarch 2026-03-09T15:17:42.101 INFO:teuthology.orchestra.run.vm03.stdout: python3-portend-3.1.0-2.el9.noarch 2026-03-09T15:17:42.101 INFO:teuthology.orchestra.run.vm03.stdout: python3-prettytable-0.7.2-27.el9.noarch 2026-03-09T15:17:42.101 INFO:teuthology.orchestra.run.vm03.stdout: python3-protobuf-3.14.0-17.el9.noarch 2026-03-09T15:17:42.101 INFO:teuthology.orchestra.run.vm03.stdout: python3-pyOpenSSL-21.0.0-1.el9.noarch 2026-03-09T15:17:42.101 INFO:teuthology.orchestra.run.vm03.stdout: python3-pyasn1-0.4.8-7.el9.noarch 2026-03-09T15:17:42.101 INFO:teuthology.orchestra.run.vm03.stdout: python3-pyasn1-modules-0.4.8-7.el9.noarch 2026-03-09T15:17:42.101 INFO:teuthology.orchestra.run.vm03.stdout: python3-pycparser-2.20-6.el9.noarch 2026-03-09T15:17:42.101 INFO:teuthology.orchestra.run.vm03.stdout: python3-pysocks-1.7.1-12.el9.noarch 2026-03-09T15:17:42.101 INFO:teuthology.orchestra.run.vm03.stdout: python3-pytz-2021.1-5.el9.noarch 2026-03-09T15:17:42.101 INFO:teuthology.orchestra.run.vm03.stdout: python3-repoze-lru-0.7-16.el9.noarch 2026-03-09T15:17:42.101 INFO:teuthology.orchestra.run.vm03.stdout: python3-requests-2.25.1-10.el9.noarch 2026-03-09T15:17:42.101 INFO:teuthology.orchestra.run.vm03.stdout: python3-requests-oauthlib-1.3.0-12.el9.noarch 2026-03-09T15:17:42.101 INFO:teuthology.orchestra.run.vm03.stdout: python3-routes-2.5.1-5.el9.noarch 2026-03-09T15:17:42.101 INFO:teuthology.orchestra.run.vm03.stdout: python3-rsa-4.9-2.el9.noarch 2026-03-09T15:17:42.101 INFO:teuthology.orchestra.run.vm03.stdout: python3-scipy-1.9.3-2.el9.x86_64 2026-03-09T15:17:42.101 INFO:teuthology.orchestra.run.vm03.stdout: python3-tempora-5.0.0-2.el9.noarch 2026-03-09T15:17:42.101 INFO:teuthology.orchestra.run.vm03.stdout: python3-toml-0.10.2-6.el9.noarch 2026-03-09T15:17:42.101 INFO:teuthology.orchestra.run.vm03.stdout: python3-typing-extensions-4.15.0-1.el9.noarch 2026-03-09T15:17:42.101 INFO:teuthology.orchestra.run.vm03.stdout: python3-urllib3-1.26.5-7.el9.noarch 2026-03-09T15:17:42.101 INFO:teuthology.orchestra.run.vm03.stdout: python3-webob-1.8.8-2.el9.noarch 2026-03-09T15:17:42.101 INFO:teuthology.orchestra.run.vm03.stdout: python3-websocket-client-1.2.3-2.el9.noarch 2026-03-09T15:17:42.101 INFO:teuthology.orchestra.run.vm03.stdout: python3-werkzeug-2.0.3-3.el9.1.noarch 2026-03-09T15:17:42.102 INFO:teuthology.orchestra.run.vm03.stdout: python3-zc-lockfile-2.0-10.el9.noarch 2026-03-09T15:17:42.102 INFO:teuthology.orchestra.run.vm03.stdout: qatlib-25.08.0-2.el9.x86_64 2026-03-09T15:17:42.102 INFO:teuthology.orchestra.run.vm03.stdout: qatlib-service-25.08.0-2.el9.x86_64 2026-03-09T15:17:42.102 INFO:teuthology.orchestra.run.vm03.stdout: qatzip-libs-1.3.1-1.el9.x86_64 2026-03-09T15:17:42.102 INFO:teuthology.orchestra.run.vm03.stdout: rbd-mirror-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T15:17:42.102 INFO:teuthology.orchestra.run.vm03.stdout: 2026-03-09T15:17:42.102 INFO:teuthology.orchestra.run.vm03.stdout:Complete! 2026-03-09T15:17:42.110 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : rbd-mirror-2:19.2.3-678.ge911bdeb.el9.x86_64 102/102 2026-03-09T15:17:42.110 INFO:teuthology.orchestra.run.vm06.stdout: 2026-03-09T15:17:42.110 INFO:teuthology.orchestra.run.vm06.stdout:Removed: 2026-03-09T15:17:42.110 INFO:teuthology.orchestra.run.vm06.stdout: abseil-cpp-20211102.0-4.el9.x86_64 2026-03-09T15:17:42.110 INFO:teuthology.orchestra.run.vm06.stdout: ceph-base-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T15:17:42.110 INFO:teuthology.orchestra.run.vm06.stdout: ceph-common-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T15:17:42.110 INFO:teuthology.orchestra.run.vm06.stdout: ceph-grafana-dashboards-2:19.2.3-678.ge911bdeb.el9.noarch 2026-03-09T15:17:42.110 INFO:teuthology.orchestra.run.vm06.stdout: ceph-immutable-object-cache-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T15:17:42.110 INFO:teuthology.orchestra.run.vm06.stdout: ceph-mgr-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T15:17:42.110 INFO:teuthology.orchestra.run.vm06.stdout: ceph-mgr-cephadm-2:19.2.3-678.ge911bdeb.el9.noarch 2026-03-09T15:17:42.110 INFO:teuthology.orchestra.run.vm06.stdout: ceph-mgr-dashboard-2:19.2.3-678.ge911bdeb.el9.noarch 2026-03-09T15:17:42.110 INFO:teuthology.orchestra.run.vm06.stdout: ceph-mgr-diskprediction-local-2:19.2.3-678.ge911bdeb.el9.noarch 2026-03-09T15:17:42.110 INFO:teuthology.orchestra.run.vm06.stdout: ceph-mgr-modules-core-2:19.2.3-678.ge911bdeb.el9.noarch 2026-03-09T15:17:42.110 INFO:teuthology.orchestra.run.vm06.stdout: ceph-mgr-rook-2:19.2.3-678.ge911bdeb.el9.noarch 2026-03-09T15:17:42.110 INFO:teuthology.orchestra.run.vm06.stdout: ceph-osd-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T15:17:42.110 INFO:teuthology.orchestra.run.vm06.stdout: ceph-prometheus-alerts-2:19.2.3-678.ge911bdeb.el9.noarch 2026-03-09T15:17:42.110 INFO:teuthology.orchestra.run.vm06.stdout: ceph-selinux-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T15:17:42.110 INFO:teuthology.orchestra.run.vm06.stdout: ceph-volume-2:19.2.3-678.ge911bdeb.el9.noarch 2026-03-09T15:17:42.111 INFO:teuthology.orchestra.run.vm06.stdout: cryptsetup-2.8.1-3.el9.x86_64 2026-03-09T15:17:42.111 INFO:teuthology.orchestra.run.vm06.stdout: flexiblas-3.0.4-9.el9.x86_64 2026-03-09T15:17:42.111 INFO:teuthology.orchestra.run.vm06.stdout: flexiblas-netlib-3.0.4-9.el9.x86_64 2026-03-09T15:17:42.111 INFO:teuthology.orchestra.run.vm06.stdout: flexiblas-openblas-openmp-3.0.4-9.el9.x86_64 2026-03-09T15:17:42.111 INFO:teuthology.orchestra.run.vm06.stdout: gperftools-libs-2.9.1-3.el9.x86_64 2026-03-09T15:17:42.111 INFO:teuthology.orchestra.run.vm06.stdout: grpc-data-1.46.7-10.el9.noarch 2026-03-09T15:17:42.111 INFO:teuthology.orchestra.run.vm06.stdout: ledmon-libs-1.1.0-3.el9.x86_64 2026-03-09T15:17:42.111 INFO:teuthology.orchestra.run.vm06.stdout: libcephsqlite-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T15:17:42.111 INFO:teuthology.orchestra.run.vm06.stdout: libconfig-1.7.2-9.el9.x86_64 2026-03-09T15:17:42.111 INFO:teuthology.orchestra.run.vm06.stdout: libgfortran-11.5.0-14.el9.x86_64 2026-03-09T15:17:42.111 INFO:teuthology.orchestra.run.vm06.stdout: liboath-2.6.12-1.el9.x86_64 2026-03-09T15:17:42.111 INFO:teuthology.orchestra.run.vm06.stdout: libquadmath-11.5.0-14.el9.x86_64 2026-03-09T15:17:42.111 INFO:teuthology.orchestra.run.vm06.stdout: libradosstriper1-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T15:17:42.111 INFO:teuthology.orchestra.run.vm06.stdout: libstoragemgmt-1.10.1-1.el9.x86_64 2026-03-09T15:17:42.111 INFO:teuthology.orchestra.run.vm06.stdout: libunwind-1.6.2-1.el9.x86_64 2026-03-09T15:17:42.111 INFO:teuthology.orchestra.run.vm06.stdout: openblas-0.3.29-1.el9.x86_64 2026-03-09T15:17:42.111 INFO:teuthology.orchestra.run.vm06.stdout: openblas-openmp-0.3.29-1.el9.x86_64 2026-03-09T15:17:42.111 INFO:teuthology.orchestra.run.vm06.stdout: pciutils-3.7.0-7.el9.x86_64 2026-03-09T15:17:42.111 INFO:teuthology.orchestra.run.vm06.stdout: protobuf-3.14.0-17.el9.x86_64 2026-03-09T15:17:42.111 INFO:teuthology.orchestra.run.vm06.stdout: protobuf-compiler-3.14.0-17.el9.x86_64 2026-03-09T15:17:42.111 INFO:teuthology.orchestra.run.vm06.stdout: python3-asyncssh-2.13.2-5.el9.noarch 2026-03-09T15:17:42.111 INFO:teuthology.orchestra.run.vm06.stdout: python3-autocommand-2.2.2-8.el9.noarch 2026-03-09T15:17:42.111 INFO:teuthology.orchestra.run.vm06.stdout: python3-babel-2.9.1-2.el9.noarch 2026-03-09T15:17:42.111 INFO:teuthology.orchestra.run.vm06.stdout: python3-backports-tarfile-1.2.0-1.el9.noarch 2026-03-09T15:17:42.111 INFO:teuthology.orchestra.run.vm06.stdout: python3-bcrypt-3.2.2-1.el9.x86_64 2026-03-09T15:17:42.111 INFO:teuthology.orchestra.run.vm06.stdout: python3-cachetools-4.2.4-1.el9.noarch 2026-03-09T15:17:42.111 INFO:teuthology.orchestra.run.vm06.stdout: python3-ceph-common-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T15:17:42.111 INFO:teuthology.orchestra.run.vm06.stdout: python3-certifi-2023.05.07-4.el9.noarch 2026-03-09T15:17:42.111 INFO:teuthology.orchestra.run.vm06.stdout: python3-cffi-1.14.5-5.el9.x86_64 2026-03-09T15:17:42.111 INFO:teuthology.orchestra.run.vm06.stdout: python3-chardet-4.0.0-5.el9.noarch 2026-03-09T15:17:42.111 INFO:teuthology.orchestra.run.vm06.stdout: python3-cheroot-10.0.1-4.el9.noarch 2026-03-09T15:17:42.111 INFO:teuthology.orchestra.run.vm06.stdout: python3-cherrypy-18.6.1-2.el9.noarch 2026-03-09T15:17:42.111 INFO:teuthology.orchestra.run.vm06.stdout: python3-cryptography-36.0.1-5.el9.x86_64 2026-03-09T15:17:42.111 INFO:teuthology.orchestra.run.vm06.stdout: python3-devel-3.9.25-3.el9.x86_64 2026-03-09T15:17:42.111 INFO:teuthology.orchestra.run.vm06.stdout: python3-google-auth-1:2.45.0-1.el9.noarch 2026-03-09T15:17:42.111 INFO:teuthology.orchestra.run.vm06.stdout: python3-grpcio-1.46.7-10.el9.x86_64 2026-03-09T15:17:42.111 INFO:teuthology.orchestra.run.vm06.stdout: python3-grpcio-tools-1.46.7-10.el9.x86_64 2026-03-09T15:17:42.111 INFO:teuthology.orchestra.run.vm06.stdout: python3-idna-2.10-7.el9.1.noarch 2026-03-09T15:17:42.111 INFO:teuthology.orchestra.run.vm06.stdout: python3-jaraco-8.2.1-3.el9.noarch 2026-03-09T15:17:42.111 INFO:teuthology.orchestra.run.vm06.stdout: python3-jaraco-classes-3.2.1-5.el9.noarch 2026-03-09T15:17:42.111 INFO:teuthology.orchestra.run.vm06.stdout: python3-jaraco-collections-3.0.0-8.el9.noarch 2026-03-09T15:17:42.111 INFO:teuthology.orchestra.run.vm06.stdout: python3-jaraco-context-6.0.1-3.el9.noarch 2026-03-09T15:17:42.111 INFO:teuthology.orchestra.run.vm06.stdout: python3-jaraco-functools-3.5.0-2.el9.noarch 2026-03-09T15:17:42.111 INFO:teuthology.orchestra.run.vm06.stdout: python3-jaraco-text-4.0.0-2.el9.noarch 2026-03-09T15:17:42.111 INFO:teuthology.orchestra.run.vm06.stdout: python3-jinja2-2.11.3-8.el9.noarch 2026-03-09T15:17:42.111 INFO:teuthology.orchestra.run.vm06.stdout: python3-jsonpatch-1.21-16.el9.noarch 2026-03-09T15:17:42.111 INFO:teuthology.orchestra.run.vm06.stdout: python3-jsonpointer-2.0-4.el9.noarch 2026-03-09T15:17:42.111 INFO:teuthology.orchestra.run.vm06.stdout: python3-kubernetes-1:26.1.0-3.el9.noarch 2026-03-09T15:17:42.111 INFO:teuthology.orchestra.run.vm06.stdout: python3-libstoragemgmt-1.10.1-1.el9.x86_64 2026-03-09T15:17:42.111 INFO:teuthology.orchestra.run.vm06.stdout: python3-logutils-0.3.5-21.el9.noarch 2026-03-09T15:17:42.111 INFO:teuthology.orchestra.run.vm06.stdout: python3-mako-1.1.4-6.el9.noarch 2026-03-09T15:17:42.111 INFO:teuthology.orchestra.run.vm06.stdout: python3-markupsafe-1.1.1-12.el9.x86_64 2026-03-09T15:17:42.111 INFO:teuthology.orchestra.run.vm06.stdout: python3-more-itertools-8.12.0-2.el9.noarch 2026-03-09T15:17:42.111 INFO:teuthology.orchestra.run.vm06.stdout: python3-natsort-7.1.1-5.el9.noarch 2026-03-09T15:17:42.111 INFO:teuthology.orchestra.run.vm06.stdout: python3-numpy-1:1.23.5-2.el9.x86_64 2026-03-09T15:17:42.111 INFO:teuthology.orchestra.run.vm06.stdout: python3-numpy-f2py-1:1.23.5-2.el9.x86_64 2026-03-09T15:17:42.111 INFO:teuthology.orchestra.run.vm06.stdout: python3-oauthlib-3.1.1-5.el9.noarch 2026-03-09T15:17:42.111 INFO:teuthology.orchestra.run.vm06.stdout: python3-packaging-20.9-5.el9.noarch 2026-03-09T15:17:42.111 INFO:teuthology.orchestra.run.vm06.stdout: python3-pecan-1.4.2-3.el9.noarch 2026-03-09T15:17:42.111 INFO:teuthology.orchestra.run.vm06.stdout: python3-ply-3.11-14.el9.noarch 2026-03-09T15:17:42.112 INFO:teuthology.orchestra.run.vm06.stdout: python3-portend-3.1.0-2.el9.noarch 2026-03-09T15:17:42.112 INFO:teuthology.orchestra.run.vm06.stdout: python3-prettytable-0.7.2-27.el9.noarch 2026-03-09T15:17:42.112 INFO:teuthology.orchestra.run.vm06.stdout: python3-protobuf-3.14.0-17.el9.noarch 2026-03-09T15:17:42.112 INFO:teuthology.orchestra.run.vm06.stdout: python3-pyOpenSSL-21.0.0-1.el9.noarch 2026-03-09T15:17:42.112 INFO:teuthology.orchestra.run.vm06.stdout: python3-pyasn1-0.4.8-7.el9.noarch 2026-03-09T15:17:42.112 INFO:teuthology.orchestra.run.vm06.stdout: python3-pyasn1-modules-0.4.8-7.el9.noarch 2026-03-09T15:17:42.112 INFO:teuthology.orchestra.run.vm06.stdout: python3-pycparser-2.20-6.el9.noarch 2026-03-09T15:17:42.112 INFO:teuthology.orchestra.run.vm06.stdout: python3-pysocks-1.7.1-12.el9.noarch 2026-03-09T15:17:42.112 INFO:teuthology.orchestra.run.vm06.stdout: python3-pytz-2021.1-5.el9.noarch 2026-03-09T15:17:42.112 INFO:teuthology.orchestra.run.vm06.stdout: python3-repoze-lru-0.7-16.el9.noarch 2026-03-09T15:17:42.112 INFO:teuthology.orchestra.run.vm06.stdout: python3-requests-2.25.1-10.el9.noarch 2026-03-09T15:17:42.112 INFO:teuthology.orchestra.run.vm06.stdout: python3-requests-oauthlib-1.3.0-12.el9.noarch 2026-03-09T15:17:42.112 INFO:teuthology.orchestra.run.vm06.stdout: python3-routes-2.5.1-5.el9.noarch 2026-03-09T15:17:42.112 INFO:teuthology.orchestra.run.vm06.stdout: python3-rsa-4.9-2.el9.noarch 2026-03-09T15:17:42.112 INFO:teuthology.orchestra.run.vm06.stdout: python3-scipy-1.9.3-2.el9.x86_64 2026-03-09T15:17:42.112 INFO:teuthology.orchestra.run.vm06.stdout: python3-tempora-5.0.0-2.el9.noarch 2026-03-09T15:17:42.112 INFO:teuthology.orchestra.run.vm06.stdout: python3-toml-0.10.2-6.el9.noarch 2026-03-09T15:17:42.112 INFO:teuthology.orchestra.run.vm06.stdout: python3-typing-extensions-4.15.0-1.el9.noarch 2026-03-09T15:17:42.112 INFO:teuthology.orchestra.run.vm06.stdout: python3-urllib3-1.26.5-7.el9.noarch 2026-03-09T15:17:42.112 INFO:teuthology.orchestra.run.vm06.stdout: python3-webob-1.8.8-2.el9.noarch 2026-03-09T15:17:42.112 INFO:teuthology.orchestra.run.vm06.stdout: python3-websocket-client-1.2.3-2.el9.noarch 2026-03-09T15:17:42.112 INFO:teuthology.orchestra.run.vm06.stdout: python3-werkzeug-2.0.3-3.el9.1.noarch 2026-03-09T15:17:42.112 INFO:teuthology.orchestra.run.vm06.stdout: python3-zc-lockfile-2.0-10.el9.noarch 2026-03-09T15:17:42.112 INFO:teuthology.orchestra.run.vm06.stdout: qatlib-25.08.0-2.el9.x86_64 2026-03-09T15:17:42.112 INFO:teuthology.orchestra.run.vm06.stdout: qatlib-service-25.08.0-2.el9.x86_64 2026-03-09T15:17:42.112 INFO:teuthology.orchestra.run.vm06.stdout: qatzip-libs-1.3.1-1.el9.x86_64 2026-03-09T15:17:42.112 INFO:teuthology.orchestra.run.vm06.stdout: rbd-mirror-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T15:17:42.112 INFO:teuthology.orchestra.run.vm06.stdout: 2026-03-09T15:17:42.112 INFO:teuthology.orchestra.run.vm06.stdout:Complete! 2026-03-09T15:17:42.293 INFO:teuthology.orchestra.run.vm04.stdout:Dependencies resolved. 2026-03-09T15:17:42.293 INFO:teuthology.orchestra.run.vm04.stdout:================================================================================ 2026-03-09T15:17:42.293 INFO:teuthology.orchestra.run.vm04.stdout: Package Arch Version Repository Size 2026-03-09T15:17:42.293 INFO:teuthology.orchestra.run.vm04.stdout:================================================================================ 2026-03-09T15:17:42.294 INFO:teuthology.orchestra.run.vm04.stdout:Removing: 2026-03-09T15:17:42.294 INFO:teuthology.orchestra.run.vm04.stdout: cephadm noarch 2:19.2.3-678.ge911bdeb.el9 @ceph-noarch 775 k 2026-03-09T15:17:42.294 INFO:teuthology.orchestra.run.vm04.stdout: 2026-03-09T15:17:42.294 INFO:teuthology.orchestra.run.vm04.stdout:Transaction Summary 2026-03-09T15:17:42.294 INFO:teuthology.orchestra.run.vm04.stdout:================================================================================ 2026-03-09T15:17:42.294 INFO:teuthology.orchestra.run.vm04.stdout:Remove 1 Package 2026-03-09T15:17:42.294 INFO:teuthology.orchestra.run.vm04.stdout: 2026-03-09T15:17:42.294 INFO:teuthology.orchestra.run.vm04.stdout:Freed space: 775 k 2026-03-09T15:17:42.294 INFO:teuthology.orchestra.run.vm04.stdout:Running transaction check 2026-03-09T15:17:42.295 INFO:teuthology.orchestra.run.vm04.stdout:Transaction check succeeded. 2026-03-09T15:17:42.295 INFO:teuthology.orchestra.run.vm04.stdout:Running transaction test 2026-03-09T15:17:42.296 INFO:teuthology.orchestra.run.vm04.stdout:Transaction test succeeded. 2026-03-09T15:17:42.297 INFO:teuthology.orchestra.run.vm04.stdout:Running transaction 2026-03-09T15:17:42.302 INFO:teuthology.orchestra.run.vm03.stdout:Dependencies resolved. 2026-03-09T15:17:42.303 INFO:teuthology.orchestra.run.vm03.stdout:================================================================================ 2026-03-09T15:17:42.303 INFO:teuthology.orchestra.run.vm03.stdout: Package Arch Version Repository Size 2026-03-09T15:17:42.303 INFO:teuthology.orchestra.run.vm03.stdout:================================================================================ 2026-03-09T15:17:42.303 INFO:teuthology.orchestra.run.vm03.stdout:Removing: 2026-03-09T15:17:42.303 INFO:teuthology.orchestra.run.vm03.stdout: cephadm noarch 2:19.2.3-678.ge911bdeb.el9 @ceph-noarch 775 k 2026-03-09T15:17:42.303 INFO:teuthology.orchestra.run.vm03.stdout: 2026-03-09T15:17:42.303 INFO:teuthology.orchestra.run.vm03.stdout:Transaction Summary 2026-03-09T15:17:42.303 INFO:teuthology.orchestra.run.vm03.stdout:================================================================================ 2026-03-09T15:17:42.303 INFO:teuthology.orchestra.run.vm03.stdout:Remove 1 Package 2026-03-09T15:17:42.303 INFO:teuthology.orchestra.run.vm03.stdout: 2026-03-09T15:17:42.303 INFO:teuthology.orchestra.run.vm03.stdout:Freed space: 775 k 2026-03-09T15:17:42.303 INFO:teuthology.orchestra.run.vm03.stdout:Running transaction check 2026-03-09T15:17:42.304 INFO:teuthology.orchestra.run.vm03.stdout:Transaction check succeeded. 2026-03-09T15:17:42.304 INFO:teuthology.orchestra.run.vm03.stdout:Running transaction test 2026-03-09T15:17:42.306 INFO:teuthology.orchestra.run.vm03.stdout:Transaction test succeeded. 2026-03-09T15:17:42.306 INFO:teuthology.orchestra.run.vm03.stdout:Running transaction 2026-03-09T15:17:42.309 INFO:teuthology.orchestra.run.vm06.stdout:Dependencies resolved. 2026-03-09T15:17:42.309 INFO:teuthology.orchestra.run.vm06.stdout:================================================================================ 2026-03-09T15:17:42.309 INFO:teuthology.orchestra.run.vm06.stdout: Package Arch Version Repository Size 2026-03-09T15:17:42.309 INFO:teuthology.orchestra.run.vm06.stdout:================================================================================ 2026-03-09T15:17:42.310 INFO:teuthology.orchestra.run.vm06.stdout:Removing: 2026-03-09T15:17:42.310 INFO:teuthology.orchestra.run.vm06.stdout: cephadm noarch 2:19.2.3-678.ge911bdeb.el9 @ceph-noarch 775 k 2026-03-09T15:17:42.310 INFO:teuthology.orchestra.run.vm06.stdout: 2026-03-09T15:17:42.310 INFO:teuthology.orchestra.run.vm06.stdout:Transaction Summary 2026-03-09T15:17:42.310 INFO:teuthology.orchestra.run.vm06.stdout:================================================================================ 2026-03-09T15:17:42.310 INFO:teuthology.orchestra.run.vm06.stdout:Remove 1 Package 2026-03-09T15:17:42.310 INFO:teuthology.orchestra.run.vm06.stdout: 2026-03-09T15:17:42.310 INFO:teuthology.orchestra.run.vm06.stdout:Freed space: 775 k 2026-03-09T15:17:42.310 INFO:teuthology.orchestra.run.vm06.stdout:Running transaction check 2026-03-09T15:17:42.311 INFO:teuthology.orchestra.run.vm06.stdout:Transaction check succeeded. 2026-03-09T15:17:42.311 INFO:teuthology.orchestra.run.vm06.stdout:Running transaction test 2026-03-09T15:17:42.312 INFO:teuthology.orchestra.run.vm04.stdout: Preparing : 1/1 2026-03-09T15:17:42.312 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : cephadm-2:19.2.3-678.ge911bdeb.el9.noarch 1/1 2026-03-09T15:17:42.313 INFO:teuthology.orchestra.run.vm06.stdout:Transaction test succeeded. 2026-03-09T15:17:42.313 INFO:teuthology.orchestra.run.vm06.stdout:Running transaction 2026-03-09T15:17:42.321 INFO:teuthology.orchestra.run.vm03.stdout: Preparing : 1/1 2026-03-09T15:17:42.321 INFO:teuthology.orchestra.run.vm03.stdout: Erasing : cephadm-2:19.2.3-678.ge911bdeb.el9.noarch 1/1 2026-03-09T15:17:42.329 INFO:teuthology.orchestra.run.vm06.stdout: Preparing : 1/1 2026-03-09T15:17:42.329 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : cephadm-2:19.2.3-678.ge911bdeb.el9.noarch 1/1 2026-03-09T15:17:42.413 INFO:teuthology.orchestra.run.vm04.stdout: Running scriptlet: cephadm-2:19.2.3-678.ge911bdeb.el9.noarch 1/1 2026-03-09T15:17:42.432 INFO:teuthology.orchestra.run.vm03.stdout: Running scriptlet: cephadm-2:19.2.3-678.ge911bdeb.el9.noarch 1/1 2026-03-09T15:17:42.435 INFO:teuthology.orchestra.run.vm06.stdout: Running scriptlet: cephadm-2:19.2.3-678.ge911bdeb.el9.noarch 1/1 2026-03-09T15:17:42.457 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : cephadm-2:19.2.3-678.ge911bdeb.el9.noarch 1/1 2026-03-09T15:17:42.457 INFO:teuthology.orchestra.run.vm04.stdout: 2026-03-09T15:17:42.457 INFO:teuthology.orchestra.run.vm04.stdout:Removed: 2026-03-09T15:17:42.457 INFO:teuthology.orchestra.run.vm04.stdout: cephadm-2:19.2.3-678.ge911bdeb.el9.noarch 2026-03-09T15:17:42.457 INFO:teuthology.orchestra.run.vm04.stdout: 2026-03-09T15:17:42.457 INFO:teuthology.orchestra.run.vm04.stdout:Complete! 2026-03-09T15:17:42.476 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : cephadm-2:19.2.3-678.ge911bdeb.el9.noarch 1/1 2026-03-09T15:17:42.477 INFO:teuthology.orchestra.run.vm06.stdout: 2026-03-09T15:17:42.477 INFO:teuthology.orchestra.run.vm06.stdout:Removed: 2026-03-09T15:17:42.477 INFO:teuthology.orchestra.run.vm06.stdout: cephadm-2:19.2.3-678.ge911bdeb.el9.noarch 2026-03-09T15:17:42.477 INFO:teuthology.orchestra.run.vm06.stdout: 2026-03-09T15:17:42.477 INFO:teuthology.orchestra.run.vm06.stdout:Complete! 2026-03-09T15:17:42.479 INFO:teuthology.orchestra.run.vm03.stdout: Verifying : cephadm-2:19.2.3-678.ge911bdeb.el9.noarch 1/1 2026-03-09T15:17:42.479 INFO:teuthology.orchestra.run.vm03.stdout: 2026-03-09T15:17:42.479 INFO:teuthology.orchestra.run.vm03.stdout:Removed: 2026-03-09T15:17:42.479 INFO:teuthology.orchestra.run.vm03.stdout: cephadm-2:19.2.3-678.ge911bdeb.el9.noarch 2026-03-09T15:17:42.479 INFO:teuthology.orchestra.run.vm03.stdout: 2026-03-09T15:17:42.479 INFO:teuthology.orchestra.run.vm03.stdout:Complete! 2026-03-09T15:17:42.631 INFO:teuthology.orchestra.run.vm04.stdout:No match for argument: ceph-immutable-object-cache 2026-03-09T15:17:42.632 INFO:teuthology.orchestra.run.vm04.stderr:No packages marked for removal. 2026-03-09T15:17:42.635 INFO:teuthology.orchestra.run.vm04.stdout:Dependencies resolved. 2026-03-09T15:17:42.636 INFO:teuthology.orchestra.run.vm04.stdout:Nothing to do. 2026-03-09T15:17:42.636 INFO:teuthology.orchestra.run.vm04.stdout:Complete! 2026-03-09T15:17:42.667 INFO:teuthology.orchestra.run.vm06.stdout:No match for argument: ceph-immutable-object-cache 2026-03-09T15:17:42.667 INFO:teuthology.orchestra.run.vm06.stderr:No packages marked for removal. 2026-03-09T15:17:42.673 INFO:teuthology.orchestra.run.vm06.stdout:Dependencies resolved. 2026-03-09T15:17:42.673 INFO:teuthology.orchestra.run.vm06.stdout:Nothing to do. 2026-03-09T15:17:42.673 INFO:teuthology.orchestra.run.vm06.stdout:Complete! 2026-03-09T15:17:42.686 INFO:teuthology.orchestra.run.vm03.stdout:No match for argument: ceph-immutable-object-cache 2026-03-09T15:17:42.686 INFO:teuthology.orchestra.run.vm03.stderr:No packages marked for removal. 2026-03-09T15:17:42.690 INFO:teuthology.orchestra.run.vm03.stdout:Dependencies resolved. 2026-03-09T15:17:42.690 INFO:teuthology.orchestra.run.vm03.stdout:Nothing to do. 2026-03-09T15:17:42.690 INFO:teuthology.orchestra.run.vm03.stdout:Complete! 2026-03-09T15:17:42.804 INFO:teuthology.orchestra.run.vm04.stdout:No match for argument: ceph-mgr 2026-03-09T15:17:42.804 INFO:teuthology.orchestra.run.vm04.stderr:No packages marked for removal. 2026-03-09T15:17:42.807 INFO:teuthology.orchestra.run.vm04.stdout:Dependencies resolved. 2026-03-09T15:17:42.807 INFO:teuthology.orchestra.run.vm04.stdout:Nothing to do. 2026-03-09T15:17:42.807 INFO:teuthology.orchestra.run.vm04.stdout:Complete! 2026-03-09T15:17:42.839 INFO:teuthology.orchestra.run.vm06.stdout:No match for argument: ceph-mgr 2026-03-09T15:17:42.839 INFO:teuthology.orchestra.run.vm06.stderr:No packages marked for removal. 2026-03-09T15:17:42.842 INFO:teuthology.orchestra.run.vm06.stdout:Dependencies resolved. 2026-03-09T15:17:42.843 INFO:teuthology.orchestra.run.vm06.stdout:Nothing to do. 2026-03-09T15:17:42.843 INFO:teuthology.orchestra.run.vm06.stdout:Complete! 2026-03-09T15:17:42.851 INFO:teuthology.orchestra.run.vm03.stdout:No match for argument: ceph-mgr 2026-03-09T15:17:42.852 INFO:teuthology.orchestra.run.vm03.stderr:No packages marked for removal. 2026-03-09T15:17:42.855 INFO:teuthology.orchestra.run.vm03.stdout:Dependencies resolved. 2026-03-09T15:17:42.855 INFO:teuthology.orchestra.run.vm03.stdout:Nothing to do. 2026-03-09T15:17:42.855 INFO:teuthology.orchestra.run.vm03.stdout:Complete! 2026-03-09T15:17:42.967 INFO:teuthology.orchestra.run.vm04.stdout:No match for argument: ceph-mgr-dashboard 2026-03-09T15:17:42.968 INFO:teuthology.orchestra.run.vm04.stderr:No packages marked for removal. 2026-03-09T15:17:42.971 INFO:teuthology.orchestra.run.vm04.stdout:Dependencies resolved. 2026-03-09T15:17:42.971 INFO:teuthology.orchestra.run.vm04.stdout:Nothing to do. 2026-03-09T15:17:42.971 INFO:teuthology.orchestra.run.vm04.stdout:Complete! 2026-03-09T15:17:43.004 INFO:teuthology.orchestra.run.vm06.stdout:No match for argument: ceph-mgr-dashboard 2026-03-09T15:17:43.005 INFO:teuthology.orchestra.run.vm06.stderr:No packages marked for removal. 2026-03-09T15:17:43.007 INFO:teuthology.orchestra.run.vm06.stdout:Dependencies resolved. 2026-03-09T15:17:43.008 INFO:teuthology.orchestra.run.vm06.stdout:Nothing to do. 2026-03-09T15:17:43.008 INFO:teuthology.orchestra.run.vm06.stdout:Complete! 2026-03-09T15:17:43.013 INFO:teuthology.orchestra.run.vm03.stdout:No match for argument: ceph-mgr-dashboard 2026-03-09T15:17:43.014 INFO:teuthology.orchestra.run.vm03.stderr:No packages marked for removal. 2026-03-09T15:17:43.017 INFO:teuthology.orchestra.run.vm03.stdout:Dependencies resolved. 2026-03-09T15:17:43.017 INFO:teuthology.orchestra.run.vm03.stdout:Nothing to do. 2026-03-09T15:17:43.017 INFO:teuthology.orchestra.run.vm03.stdout:Complete! 2026-03-09T15:17:43.128 INFO:teuthology.orchestra.run.vm04.stdout:No match for argument: ceph-mgr-diskprediction-local 2026-03-09T15:17:43.128 INFO:teuthology.orchestra.run.vm04.stderr:No packages marked for removal. 2026-03-09T15:17:43.131 INFO:teuthology.orchestra.run.vm04.stdout:Dependencies resolved. 2026-03-09T15:17:43.131 INFO:teuthology.orchestra.run.vm04.stdout:Nothing to do. 2026-03-09T15:17:43.132 INFO:teuthology.orchestra.run.vm04.stdout:Complete! 2026-03-09T15:17:43.170 INFO:teuthology.orchestra.run.vm06.stdout:No match for argument: ceph-mgr-diskprediction-local 2026-03-09T15:17:43.171 INFO:teuthology.orchestra.run.vm06.stderr:No packages marked for removal. 2026-03-09T15:17:43.174 INFO:teuthology.orchestra.run.vm06.stdout:Dependencies resolved. 2026-03-09T15:17:43.174 INFO:teuthology.orchestra.run.vm06.stdout:Nothing to do. 2026-03-09T15:17:43.174 INFO:teuthology.orchestra.run.vm06.stdout:Complete! 2026-03-09T15:17:43.177 INFO:teuthology.orchestra.run.vm03.stdout:No match for argument: ceph-mgr-diskprediction-local 2026-03-09T15:17:43.177 INFO:teuthology.orchestra.run.vm03.stderr:No packages marked for removal. 2026-03-09T15:17:43.180 INFO:teuthology.orchestra.run.vm03.stdout:Dependencies resolved. 2026-03-09T15:17:43.181 INFO:teuthology.orchestra.run.vm03.stdout:Nothing to do. 2026-03-09T15:17:43.181 INFO:teuthology.orchestra.run.vm03.stdout:Complete! 2026-03-09T15:17:43.299 INFO:teuthology.orchestra.run.vm04.stdout:No match for argument: ceph-mgr-rook 2026-03-09T15:17:43.299 INFO:teuthology.orchestra.run.vm04.stderr:No packages marked for removal. 2026-03-09T15:17:43.302 INFO:teuthology.orchestra.run.vm04.stdout:Dependencies resolved. 2026-03-09T15:17:43.303 INFO:teuthology.orchestra.run.vm04.stdout:Nothing to do. 2026-03-09T15:17:43.303 INFO:teuthology.orchestra.run.vm04.stdout:Complete! 2026-03-09T15:17:43.339 INFO:teuthology.orchestra.run.vm06.stdout:No match for argument: ceph-mgr-rook 2026-03-09T15:17:43.339 INFO:teuthology.orchestra.run.vm06.stderr:No packages marked for removal. 2026-03-09T15:17:43.342 INFO:teuthology.orchestra.run.vm06.stdout:Dependencies resolved. 2026-03-09T15:17:43.343 INFO:teuthology.orchestra.run.vm06.stdout:Nothing to do. 2026-03-09T15:17:43.343 INFO:teuthology.orchestra.run.vm06.stdout:Complete! 2026-03-09T15:17:43.345 INFO:teuthology.orchestra.run.vm03.stdout:No match for argument: ceph-mgr-rook 2026-03-09T15:17:43.346 INFO:teuthology.orchestra.run.vm03.stderr:No packages marked for removal. 2026-03-09T15:17:43.349 INFO:teuthology.orchestra.run.vm03.stdout:Dependencies resolved. 2026-03-09T15:17:43.349 INFO:teuthology.orchestra.run.vm03.stdout:Nothing to do. 2026-03-09T15:17:43.349 INFO:teuthology.orchestra.run.vm03.stdout:Complete! 2026-03-09T15:17:43.471 INFO:teuthology.orchestra.run.vm04.stdout:No match for argument: ceph-mgr-cephadm 2026-03-09T15:17:43.472 INFO:teuthology.orchestra.run.vm04.stderr:No packages marked for removal. 2026-03-09T15:17:43.475 INFO:teuthology.orchestra.run.vm04.stdout:Dependencies resolved. 2026-03-09T15:17:43.475 INFO:teuthology.orchestra.run.vm04.stdout:Nothing to do. 2026-03-09T15:17:43.475 INFO:teuthology.orchestra.run.vm04.stdout:Complete! 2026-03-09T15:17:43.520 INFO:teuthology.orchestra.run.vm06.stdout:No match for argument: ceph-mgr-cephadm 2026-03-09T15:17:43.520 INFO:teuthology.orchestra.run.vm06.stderr:No packages marked for removal. 2026-03-09T15:17:43.522 INFO:teuthology.orchestra.run.vm03.stdout:No match for argument: ceph-mgr-cephadm 2026-03-09T15:17:43.522 INFO:teuthology.orchestra.run.vm03.stderr:No packages marked for removal. 2026-03-09T15:17:43.523 INFO:teuthology.orchestra.run.vm06.stdout:Dependencies resolved. 2026-03-09T15:17:43.524 INFO:teuthology.orchestra.run.vm06.stdout:Nothing to do. 2026-03-09T15:17:43.524 INFO:teuthology.orchestra.run.vm06.stdout:Complete! 2026-03-09T15:17:43.525 INFO:teuthology.orchestra.run.vm03.stdout:Dependencies resolved. 2026-03-09T15:17:43.526 INFO:teuthology.orchestra.run.vm03.stdout:Nothing to do. 2026-03-09T15:17:43.526 INFO:teuthology.orchestra.run.vm03.stdout:Complete! 2026-03-09T15:17:43.654 INFO:teuthology.orchestra.run.vm04.stdout:Dependencies resolved. 2026-03-09T15:17:43.654 INFO:teuthology.orchestra.run.vm04.stdout:================================================================================ 2026-03-09T15:17:43.654 INFO:teuthology.orchestra.run.vm04.stdout: Package Arch Version Repository Size 2026-03-09T15:17:43.654 INFO:teuthology.orchestra.run.vm04.stdout:================================================================================ 2026-03-09T15:17:43.654 INFO:teuthology.orchestra.run.vm04.stdout:Removing: 2026-03-09T15:17:43.654 INFO:teuthology.orchestra.run.vm04.stdout: ceph-fuse x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 3.6 M 2026-03-09T15:17:43.654 INFO:teuthology.orchestra.run.vm04.stdout: 2026-03-09T15:17:43.654 INFO:teuthology.orchestra.run.vm04.stdout:Transaction Summary 2026-03-09T15:17:43.654 INFO:teuthology.orchestra.run.vm04.stdout:================================================================================ 2026-03-09T15:17:43.654 INFO:teuthology.orchestra.run.vm04.stdout:Remove 1 Package 2026-03-09T15:17:43.654 INFO:teuthology.orchestra.run.vm04.stdout: 2026-03-09T15:17:43.654 INFO:teuthology.orchestra.run.vm04.stdout:Freed space: 3.6 M 2026-03-09T15:17:43.654 INFO:teuthology.orchestra.run.vm04.stdout:Running transaction check 2026-03-09T15:17:43.656 INFO:teuthology.orchestra.run.vm04.stdout:Transaction check succeeded. 2026-03-09T15:17:43.656 INFO:teuthology.orchestra.run.vm04.stdout:Running transaction test 2026-03-09T15:17:43.665 INFO:teuthology.orchestra.run.vm04.stdout:Transaction test succeeded. 2026-03-09T15:17:43.665 INFO:teuthology.orchestra.run.vm04.stdout:Running transaction 2026-03-09T15:17:43.690 INFO:teuthology.orchestra.run.vm04.stdout: Preparing : 1/1 2026-03-09T15:17:43.702 INFO:teuthology.orchestra.run.vm06.stdout:Dependencies resolved. 2026-03-09T15:17:43.702 INFO:teuthology.orchestra.run.vm06.stdout:================================================================================ 2026-03-09T15:17:43.702 INFO:teuthology.orchestra.run.vm06.stdout: Package Arch Version Repository Size 2026-03-09T15:17:43.702 INFO:teuthology.orchestra.run.vm06.stdout:================================================================================ 2026-03-09T15:17:43.702 INFO:teuthology.orchestra.run.vm06.stdout:Removing: 2026-03-09T15:17:43.702 INFO:teuthology.orchestra.run.vm06.stdout: ceph-fuse x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 3.6 M 2026-03-09T15:17:43.702 INFO:teuthology.orchestra.run.vm06.stdout: 2026-03-09T15:17:43.702 INFO:teuthology.orchestra.run.vm06.stdout:Transaction Summary 2026-03-09T15:17:43.703 INFO:teuthology.orchestra.run.vm06.stdout:================================================================================ 2026-03-09T15:17:43.703 INFO:teuthology.orchestra.run.vm06.stdout:Remove 1 Package 2026-03-09T15:17:43.703 INFO:teuthology.orchestra.run.vm06.stdout: 2026-03-09T15:17:43.703 INFO:teuthology.orchestra.run.vm06.stdout:Freed space: 3.6 M 2026-03-09T15:17:43.703 INFO:teuthology.orchestra.run.vm06.stdout:Running transaction check 2026-03-09T15:17:43.703 INFO:teuthology.orchestra.run.vm03.stdout:Dependencies resolved. 2026-03-09T15:17:43.703 INFO:teuthology.orchestra.run.vm03.stdout:================================================================================ 2026-03-09T15:17:43.703 INFO:teuthology.orchestra.run.vm03.stdout: Package Arch Version Repository Size 2026-03-09T15:17:43.703 INFO:teuthology.orchestra.run.vm03.stdout:================================================================================ 2026-03-09T15:17:43.704 INFO:teuthology.orchestra.run.vm03.stdout:Removing: 2026-03-09T15:17:43.704 INFO:teuthology.orchestra.run.vm03.stdout: ceph-fuse x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 3.6 M 2026-03-09T15:17:43.704 INFO:teuthology.orchestra.run.vm03.stdout: 2026-03-09T15:17:43.704 INFO:teuthology.orchestra.run.vm03.stdout:Transaction Summary 2026-03-09T15:17:43.704 INFO:teuthology.orchestra.run.vm03.stdout:================================================================================ 2026-03-09T15:17:43.704 INFO:teuthology.orchestra.run.vm03.stdout:Remove 1 Package 2026-03-09T15:17:43.704 INFO:teuthology.orchestra.run.vm03.stdout: 2026-03-09T15:17:43.704 INFO:teuthology.orchestra.run.vm03.stdout:Freed space: 3.6 M 2026-03-09T15:17:43.704 INFO:teuthology.orchestra.run.vm03.stdout:Running transaction check 2026-03-09T15:17:43.704 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : ceph-fuse-2:19.2.3-678.ge911bdeb.el9.x86_64 1/1 2026-03-09T15:17:43.704 INFO:teuthology.orchestra.run.vm06.stdout:Transaction check succeeded. 2026-03-09T15:17:43.705 INFO:teuthology.orchestra.run.vm06.stdout:Running transaction test 2026-03-09T15:17:43.706 INFO:teuthology.orchestra.run.vm03.stdout:Transaction check succeeded. 2026-03-09T15:17:43.706 INFO:teuthology.orchestra.run.vm03.stdout:Running transaction test 2026-03-09T15:17:43.715 INFO:teuthology.orchestra.run.vm06.stdout:Transaction test succeeded. 2026-03-09T15:17:43.715 INFO:teuthology.orchestra.run.vm03.stdout:Transaction test succeeded. 2026-03-09T15:17:43.715 INFO:teuthology.orchestra.run.vm06.stdout:Running transaction 2026-03-09T15:17:43.716 INFO:teuthology.orchestra.run.vm03.stdout:Running transaction 2026-03-09T15:17:43.740 INFO:teuthology.orchestra.run.vm06.stdout: Preparing : 1/1 2026-03-09T15:17:43.741 INFO:teuthology.orchestra.run.vm03.stdout: Preparing : 1/1 2026-03-09T15:17:43.754 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : ceph-fuse-2:19.2.3-678.ge911bdeb.el9.x86_64 1/1 2026-03-09T15:17:43.756 INFO:teuthology.orchestra.run.vm03.stdout: Erasing : ceph-fuse-2:19.2.3-678.ge911bdeb.el9.x86_64 1/1 2026-03-09T15:17:43.766 INFO:teuthology.orchestra.run.vm04.stdout: Running scriptlet: ceph-fuse-2:19.2.3-678.ge911bdeb.el9.x86_64 1/1 2026-03-09T15:17:43.809 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : ceph-fuse-2:19.2.3-678.ge911bdeb.el9.x86_64 1/1 2026-03-09T15:17:43.809 INFO:teuthology.orchestra.run.vm04.stdout: 2026-03-09T15:17:43.809 INFO:teuthology.orchestra.run.vm04.stdout:Removed: 2026-03-09T15:17:43.809 INFO:teuthology.orchestra.run.vm04.stdout: ceph-fuse-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T15:17:43.809 INFO:teuthology.orchestra.run.vm04.stdout: 2026-03-09T15:17:43.809 INFO:teuthology.orchestra.run.vm04.stdout:Complete! 2026-03-09T15:17:43.824 INFO:teuthology.orchestra.run.vm03.stdout: Running scriptlet: ceph-fuse-2:19.2.3-678.ge911bdeb.el9.x86_64 1/1 2026-03-09T15:17:43.825 INFO:teuthology.orchestra.run.vm06.stdout: Running scriptlet: ceph-fuse-2:19.2.3-678.ge911bdeb.el9.x86_64 1/1 2026-03-09T15:17:43.865 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : ceph-fuse-2:19.2.3-678.ge911bdeb.el9.x86_64 1/1 2026-03-09T15:17:43.865 INFO:teuthology.orchestra.run.vm06.stdout: 2026-03-09T15:17:43.865 INFO:teuthology.orchestra.run.vm06.stdout:Removed: 2026-03-09T15:17:43.865 INFO:teuthology.orchestra.run.vm06.stdout: ceph-fuse-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T15:17:43.865 INFO:teuthology.orchestra.run.vm06.stdout: 2026-03-09T15:17:43.865 INFO:teuthology.orchestra.run.vm06.stdout:Complete! 2026-03-09T15:17:43.876 INFO:teuthology.orchestra.run.vm03.stdout: Verifying : ceph-fuse-2:19.2.3-678.ge911bdeb.el9.x86_64 1/1 2026-03-09T15:17:43.876 INFO:teuthology.orchestra.run.vm03.stdout: 2026-03-09T15:17:43.876 INFO:teuthology.orchestra.run.vm03.stdout:Removed: 2026-03-09T15:17:43.876 INFO:teuthology.orchestra.run.vm03.stdout: ceph-fuse-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T15:17:43.876 INFO:teuthology.orchestra.run.vm03.stdout: 2026-03-09T15:17:43.876 INFO:teuthology.orchestra.run.vm03.stdout:Complete! 2026-03-09T15:17:43.983 INFO:teuthology.orchestra.run.vm04.stdout:No match for argument: ceph-volume 2026-03-09T15:17:43.983 INFO:teuthology.orchestra.run.vm04.stderr:No packages marked for removal. 2026-03-09T15:17:43.986 INFO:teuthology.orchestra.run.vm04.stdout:Dependencies resolved. 2026-03-09T15:17:43.987 INFO:teuthology.orchestra.run.vm04.stdout:Nothing to do. 2026-03-09T15:17:43.987 INFO:teuthology.orchestra.run.vm04.stdout:Complete! 2026-03-09T15:17:44.046 INFO:teuthology.orchestra.run.vm06.stdout:No match for argument: ceph-volume 2026-03-09T15:17:44.046 INFO:teuthology.orchestra.run.vm06.stderr:No packages marked for removal. 2026-03-09T15:17:44.049 INFO:teuthology.orchestra.run.vm06.stdout:Dependencies resolved. 2026-03-09T15:17:44.050 INFO:teuthology.orchestra.run.vm06.stdout:Nothing to do. 2026-03-09T15:17:44.050 INFO:teuthology.orchestra.run.vm06.stdout:Complete! 2026-03-09T15:17:44.055 INFO:teuthology.orchestra.run.vm03.stdout:No match for argument: ceph-volume 2026-03-09T15:17:44.055 INFO:teuthology.orchestra.run.vm03.stderr:No packages marked for removal. 2026-03-09T15:17:44.058 INFO:teuthology.orchestra.run.vm03.stdout:Dependencies resolved. 2026-03-09T15:17:44.059 INFO:teuthology.orchestra.run.vm03.stdout:Nothing to do. 2026-03-09T15:17:44.059 INFO:teuthology.orchestra.run.vm03.stdout:Complete! 2026-03-09T15:17:44.162 INFO:teuthology.orchestra.run.vm04.stdout:Dependencies resolved. 2026-03-09T15:17:44.162 INFO:teuthology.orchestra.run.vm04.stdout:================================================================================ 2026-03-09T15:17:44.162 INFO:teuthology.orchestra.run.vm04.stdout: Package Arch Version Repo Size 2026-03-09T15:17:44.162 INFO:teuthology.orchestra.run.vm04.stdout:================================================================================ 2026-03-09T15:17:44.162 INFO:teuthology.orchestra.run.vm04.stdout:Removing: 2026-03-09T15:17:44.162 INFO:teuthology.orchestra.run.vm04.stdout: librados-devel x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 456 k 2026-03-09T15:17:44.162 INFO:teuthology.orchestra.run.vm04.stdout:Removing dependent packages: 2026-03-09T15:17:44.162 INFO:teuthology.orchestra.run.vm04.stdout: libcephfs-devel x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 153 k 2026-03-09T15:17:44.162 INFO:teuthology.orchestra.run.vm04.stdout: 2026-03-09T15:17:44.162 INFO:teuthology.orchestra.run.vm04.stdout:Transaction Summary 2026-03-09T15:17:44.162 INFO:teuthology.orchestra.run.vm04.stdout:================================================================================ 2026-03-09T15:17:44.162 INFO:teuthology.orchestra.run.vm04.stdout:Remove 2 Packages 2026-03-09T15:17:44.162 INFO:teuthology.orchestra.run.vm04.stdout: 2026-03-09T15:17:44.162 INFO:teuthology.orchestra.run.vm04.stdout:Freed space: 610 k 2026-03-09T15:17:44.162 INFO:teuthology.orchestra.run.vm04.stdout:Running transaction check 2026-03-09T15:17:44.164 INFO:teuthology.orchestra.run.vm04.stdout:Transaction check succeeded. 2026-03-09T15:17:44.164 INFO:teuthology.orchestra.run.vm04.stdout:Running transaction test 2026-03-09T15:17:44.174 INFO:teuthology.orchestra.run.vm04.stdout:Transaction test succeeded. 2026-03-09T15:17:44.174 INFO:teuthology.orchestra.run.vm04.stdout:Running transaction 2026-03-09T15:17:44.199 INFO:teuthology.orchestra.run.vm04.stdout: Preparing : 1/1 2026-03-09T15:17:44.200 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : libcephfs-devel-2:19.2.3-678.ge911bdeb.el9.x86_64 1/2 2026-03-09T15:17:44.213 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : librados-devel-2:19.2.3-678.ge911bdeb.el9.x86_64 2/2 2026-03-09T15:17:44.226 INFO:teuthology.orchestra.run.vm06.stdout:Dependencies resolved. 2026-03-09T15:17:44.227 INFO:teuthology.orchestra.run.vm06.stdout:================================================================================ 2026-03-09T15:17:44.227 INFO:teuthology.orchestra.run.vm06.stdout: Package Arch Version Repo Size 2026-03-09T15:17:44.227 INFO:teuthology.orchestra.run.vm06.stdout:================================================================================ 2026-03-09T15:17:44.227 INFO:teuthology.orchestra.run.vm06.stdout:Removing: 2026-03-09T15:17:44.227 INFO:teuthology.orchestra.run.vm06.stdout: librados-devel x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 456 k 2026-03-09T15:17:44.227 INFO:teuthology.orchestra.run.vm06.stdout:Removing dependent packages: 2026-03-09T15:17:44.227 INFO:teuthology.orchestra.run.vm06.stdout: libcephfs-devel x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 153 k 2026-03-09T15:17:44.227 INFO:teuthology.orchestra.run.vm06.stdout: 2026-03-09T15:17:44.227 INFO:teuthology.orchestra.run.vm06.stdout:Transaction Summary 2026-03-09T15:17:44.227 INFO:teuthology.orchestra.run.vm06.stdout:================================================================================ 2026-03-09T15:17:44.227 INFO:teuthology.orchestra.run.vm06.stdout:Remove 2 Packages 2026-03-09T15:17:44.227 INFO:teuthology.orchestra.run.vm06.stdout: 2026-03-09T15:17:44.227 INFO:teuthology.orchestra.run.vm06.stdout:Freed space: 610 k 2026-03-09T15:17:44.227 INFO:teuthology.orchestra.run.vm06.stdout:Running transaction check 2026-03-09T15:17:44.229 INFO:teuthology.orchestra.run.vm06.stdout:Transaction check succeeded. 2026-03-09T15:17:44.229 INFO:teuthology.orchestra.run.vm06.stdout:Running transaction test 2026-03-09T15:17:44.235 INFO:teuthology.orchestra.run.vm03.stdout:Dependencies resolved. 2026-03-09T15:17:44.235 INFO:teuthology.orchestra.run.vm03.stdout:================================================================================ 2026-03-09T15:17:44.235 INFO:teuthology.orchestra.run.vm03.stdout: Package Arch Version Repo Size 2026-03-09T15:17:44.235 INFO:teuthology.orchestra.run.vm03.stdout:================================================================================ 2026-03-09T15:17:44.235 INFO:teuthology.orchestra.run.vm03.stdout:Removing: 2026-03-09T15:17:44.235 INFO:teuthology.orchestra.run.vm03.stdout: librados-devel x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 456 k 2026-03-09T15:17:44.235 INFO:teuthology.orchestra.run.vm03.stdout:Removing dependent packages: 2026-03-09T15:17:44.235 INFO:teuthology.orchestra.run.vm03.stdout: libcephfs-devel x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 153 k 2026-03-09T15:17:44.235 INFO:teuthology.orchestra.run.vm03.stdout: 2026-03-09T15:17:44.235 INFO:teuthology.orchestra.run.vm03.stdout:Transaction Summary 2026-03-09T15:17:44.235 INFO:teuthology.orchestra.run.vm03.stdout:================================================================================ 2026-03-09T15:17:44.235 INFO:teuthology.orchestra.run.vm03.stdout:Remove 2 Packages 2026-03-09T15:17:44.235 INFO:teuthology.orchestra.run.vm03.stdout: 2026-03-09T15:17:44.235 INFO:teuthology.orchestra.run.vm03.stdout:Freed space: 610 k 2026-03-09T15:17:44.235 INFO:teuthology.orchestra.run.vm03.stdout:Running transaction check 2026-03-09T15:17:44.237 INFO:teuthology.orchestra.run.vm03.stdout:Transaction check succeeded. 2026-03-09T15:17:44.237 INFO:teuthology.orchestra.run.vm03.stdout:Running transaction test 2026-03-09T15:17:44.238 INFO:teuthology.orchestra.run.vm06.stdout:Transaction test succeeded. 2026-03-09T15:17:44.239 INFO:teuthology.orchestra.run.vm06.stdout:Running transaction 2026-03-09T15:17:44.248 INFO:teuthology.orchestra.run.vm03.stdout:Transaction test succeeded. 2026-03-09T15:17:44.249 INFO:teuthology.orchestra.run.vm03.stdout:Running transaction 2026-03-09T15:17:44.264 INFO:teuthology.orchestra.run.vm06.stdout: Preparing : 1/1 2026-03-09T15:17:44.266 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : libcephfs-devel-2:19.2.3-678.ge911bdeb.el9.x86_64 1/2 2026-03-09T15:17:44.273 INFO:teuthology.orchestra.run.vm03.stdout: Preparing : 1/1 2026-03-09T15:17:44.275 INFO:teuthology.orchestra.run.vm03.stdout: Erasing : libcephfs-devel-2:19.2.3-678.ge911bdeb.el9.x86_64 1/2 2026-03-09T15:17:44.277 INFO:teuthology.orchestra.run.vm04.stdout: Running scriptlet: librados-devel-2:19.2.3-678.ge911bdeb.el9.x86_64 2/2 2026-03-09T15:17:44.277 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : libcephfs-devel-2:19.2.3-678.ge911bdeb.el9.x86_64 1/2 2026-03-09T15:17:44.279 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : librados-devel-2:19.2.3-678.ge911bdeb.el9.x86_64 2/2 2026-03-09T15:17:44.288 INFO:teuthology.orchestra.run.vm03.stdout: Erasing : librados-devel-2:19.2.3-678.ge911bdeb.el9.x86_64 2/2 2026-03-09T15:17:44.319 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : librados-devel-2:19.2.3-678.ge911bdeb.el9.x86_64 2/2 2026-03-09T15:17:44.319 INFO:teuthology.orchestra.run.vm04.stdout: 2026-03-09T15:17:44.319 INFO:teuthology.orchestra.run.vm04.stdout:Removed: 2026-03-09T15:17:44.319 INFO:teuthology.orchestra.run.vm04.stdout: libcephfs-devel-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T15:17:44.319 INFO:teuthology.orchestra.run.vm04.stdout: librados-devel-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T15:17:44.319 INFO:teuthology.orchestra.run.vm04.stdout: 2026-03-09T15:17:44.319 INFO:teuthology.orchestra.run.vm04.stdout:Complete! 2026-03-09T15:17:44.338 INFO:teuthology.orchestra.run.vm06.stdout: Running scriptlet: librados-devel-2:19.2.3-678.ge911bdeb.el9.x86_64 2/2 2026-03-09T15:17:44.338 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : libcephfs-devel-2:19.2.3-678.ge911bdeb.el9.x86_64 1/2 2026-03-09T15:17:44.346 INFO:teuthology.orchestra.run.vm03.stdout: Running scriptlet: librados-devel-2:19.2.3-678.ge911bdeb.el9.x86_64 2/2 2026-03-09T15:17:44.347 INFO:teuthology.orchestra.run.vm03.stdout: Verifying : libcephfs-devel-2:19.2.3-678.ge911bdeb.el9.x86_64 1/2 2026-03-09T15:17:44.380 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : librados-devel-2:19.2.3-678.ge911bdeb.el9.x86_64 2/2 2026-03-09T15:17:44.381 INFO:teuthology.orchestra.run.vm06.stdout: 2026-03-09T15:17:44.381 INFO:teuthology.orchestra.run.vm06.stdout:Removed: 2026-03-09T15:17:44.381 INFO:teuthology.orchestra.run.vm06.stdout: libcephfs-devel-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T15:17:44.381 INFO:teuthology.orchestra.run.vm06.stdout: librados-devel-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T15:17:44.381 INFO:teuthology.orchestra.run.vm06.stdout: 2026-03-09T15:17:44.381 INFO:teuthology.orchestra.run.vm06.stdout:Complete! 2026-03-09T15:17:44.391 INFO:teuthology.orchestra.run.vm03.stdout: Verifying : librados-devel-2:19.2.3-678.ge911bdeb.el9.x86_64 2/2 2026-03-09T15:17:44.391 INFO:teuthology.orchestra.run.vm03.stdout: 2026-03-09T15:17:44.391 INFO:teuthology.orchestra.run.vm03.stdout:Removed: 2026-03-09T15:17:44.391 INFO:teuthology.orchestra.run.vm03.stdout: libcephfs-devel-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T15:17:44.391 INFO:teuthology.orchestra.run.vm03.stdout: librados-devel-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T15:17:44.391 INFO:teuthology.orchestra.run.vm03.stdout: 2026-03-09T15:17:44.391 INFO:teuthology.orchestra.run.vm03.stdout:Complete! 2026-03-09T15:17:44.510 INFO:teuthology.orchestra.run.vm04.stdout:Dependencies resolved. 2026-03-09T15:17:44.511 INFO:teuthology.orchestra.run.vm04.stdout:================================================================================ 2026-03-09T15:17:44.511 INFO:teuthology.orchestra.run.vm04.stdout: Package Arch Version Repo Size 2026-03-09T15:17:44.511 INFO:teuthology.orchestra.run.vm04.stdout:================================================================================ 2026-03-09T15:17:44.511 INFO:teuthology.orchestra.run.vm04.stdout:Removing: 2026-03-09T15:17:44.511 INFO:teuthology.orchestra.run.vm04.stdout: libcephfs2 x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 3.0 M 2026-03-09T15:17:44.511 INFO:teuthology.orchestra.run.vm04.stdout:Removing dependent packages: 2026-03-09T15:17:44.511 INFO:teuthology.orchestra.run.vm04.stdout: python3-cephfs x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 514 k 2026-03-09T15:17:44.511 INFO:teuthology.orchestra.run.vm04.stdout:Removing unused dependencies: 2026-03-09T15:17:44.511 INFO:teuthology.orchestra.run.vm04.stdout: python3-ceph-argparse x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 187 k 2026-03-09T15:17:44.511 INFO:teuthology.orchestra.run.vm04.stdout: 2026-03-09T15:17:44.511 INFO:teuthology.orchestra.run.vm04.stdout:Transaction Summary 2026-03-09T15:17:44.511 INFO:teuthology.orchestra.run.vm04.stdout:================================================================================ 2026-03-09T15:17:44.511 INFO:teuthology.orchestra.run.vm04.stdout:Remove 3 Packages 2026-03-09T15:17:44.511 INFO:teuthology.orchestra.run.vm04.stdout: 2026-03-09T15:17:44.511 INFO:teuthology.orchestra.run.vm04.stdout:Freed space: 3.7 M 2026-03-09T15:17:44.511 INFO:teuthology.orchestra.run.vm04.stdout:Running transaction check 2026-03-09T15:17:44.516 INFO:teuthology.orchestra.run.vm04.stdout:Transaction check succeeded. 2026-03-09T15:17:44.516 INFO:teuthology.orchestra.run.vm04.stdout:Running transaction test 2026-03-09T15:17:44.531 INFO:teuthology.orchestra.run.vm04.stdout:Transaction test succeeded. 2026-03-09T15:17:44.531 INFO:teuthology.orchestra.run.vm04.stdout:Running transaction 2026-03-09T15:17:44.560 INFO:teuthology.orchestra.run.vm04.stdout: Preparing : 1/1 2026-03-09T15:17:44.562 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : python3-cephfs-2:19.2.3-678.ge911bdeb.el9.x86_64 1/3 2026-03-09T15:17:44.564 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : python3-ceph-argparse-2:19.2.3-678.ge911bdeb.el9.x86 2/3 2026-03-09T15:17:44.564 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : libcephfs2-2:19.2.3-678.ge911bdeb.el9.x86_64 3/3 2026-03-09T15:17:44.574 INFO:teuthology.orchestra.run.vm03.stdout:Dependencies resolved. 2026-03-09T15:17:44.575 INFO:teuthology.orchestra.run.vm03.stdout:================================================================================ 2026-03-09T15:17:44.575 INFO:teuthology.orchestra.run.vm03.stdout: Package Arch Version Repo Size 2026-03-09T15:17:44.575 INFO:teuthology.orchestra.run.vm03.stdout:================================================================================ 2026-03-09T15:17:44.575 INFO:teuthology.orchestra.run.vm03.stdout:Removing: 2026-03-09T15:17:44.575 INFO:teuthology.orchestra.run.vm03.stdout: libcephfs2 x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 3.0 M 2026-03-09T15:17:44.575 INFO:teuthology.orchestra.run.vm03.stdout:Removing dependent packages: 2026-03-09T15:17:44.575 INFO:teuthology.orchestra.run.vm03.stdout: python3-cephfs x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 514 k 2026-03-09T15:17:44.575 INFO:teuthology.orchestra.run.vm03.stdout:Removing unused dependencies: 2026-03-09T15:17:44.575 INFO:teuthology.orchestra.run.vm03.stdout: python3-ceph-argparse x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 187 k 2026-03-09T15:17:44.575 INFO:teuthology.orchestra.run.vm03.stdout: 2026-03-09T15:17:44.575 INFO:teuthology.orchestra.run.vm03.stdout:Transaction Summary 2026-03-09T15:17:44.575 INFO:teuthology.orchestra.run.vm03.stdout:================================================================================ 2026-03-09T15:17:44.575 INFO:teuthology.orchestra.run.vm03.stdout:Remove 3 Packages 2026-03-09T15:17:44.575 INFO:teuthology.orchestra.run.vm03.stdout: 2026-03-09T15:17:44.575 INFO:teuthology.orchestra.run.vm03.stdout:Freed space: 3.7 M 2026-03-09T15:17:44.575 INFO:teuthology.orchestra.run.vm03.stdout:Running transaction check 2026-03-09T15:17:44.577 INFO:teuthology.orchestra.run.vm03.stdout:Transaction check succeeded. 2026-03-09T15:17:44.577 INFO:teuthology.orchestra.run.vm03.stdout:Running transaction test 2026-03-09T15:17:44.584 INFO:teuthology.orchestra.run.vm06.stdout:Dependencies resolved. 2026-03-09T15:17:44.584 INFO:teuthology.orchestra.run.vm06.stdout:================================================================================ 2026-03-09T15:17:44.584 INFO:teuthology.orchestra.run.vm06.stdout: Package Arch Version Repo Size 2026-03-09T15:17:44.584 INFO:teuthology.orchestra.run.vm06.stdout:================================================================================ 2026-03-09T15:17:44.584 INFO:teuthology.orchestra.run.vm06.stdout:Removing: 2026-03-09T15:17:44.584 INFO:teuthology.orchestra.run.vm06.stdout: libcephfs2 x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 3.0 M 2026-03-09T15:17:44.584 INFO:teuthology.orchestra.run.vm06.stdout:Removing dependent packages: 2026-03-09T15:17:44.584 INFO:teuthology.orchestra.run.vm06.stdout: python3-cephfs x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 514 k 2026-03-09T15:17:44.585 INFO:teuthology.orchestra.run.vm06.stdout:Removing unused dependencies: 2026-03-09T15:17:44.585 INFO:teuthology.orchestra.run.vm06.stdout: python3-ceph-argparse x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 187 k 2026-03-09T15:17:44.585 INFO:teuthology.orchestra.run.vm06.stdout: 2026-03-09T15:17:44.585 INFO:teuthology.orchestra.run.vm06.stdout:Transaction Summary 2026-03-09T15:17:44.585 INFO:teuthology.orchestra.run.vm06.stdout:================================================================================ 2026-03-09T15:17:44.585 INFO:teuthology.orchestra.run.vm06.stdout:Remove 3 Packages 2026-03-09T15:17:44.585 INFO:teuthology.orchestra.run.vm06.stdout: 2026-03-09T15:17:44.585 INFO:teuthology.orchestra.run.vm06.stdout:Freed space: 3.7 M 2026-03-09T15:17:44.585 INFO:teuthology.orchestra.run.vm06.stdout:Running transaction check 2026-03-09T15:17:44.586 INFO:teuthology.orchestra.run.vm06.stdout:Transaction check succeeded. 2026-03-09T15:17:44.586 INFO:teuthology.orchestra.run.vm06.stdout:Running transaction test 2026-03-09T15:17:44.594 INFO:teuthology.orchestra.run.vm03.stdout:Transaction test succeeded. 2026-03-09T15:17:44.594 INFO:teuthology.orchestra.run.vm03.stdout:Running transaction 2026-03-09T15:17:44.602 INFO:teuthology.orchestra.run.vm06.stdout:Transaction test succeeded. 2026-03-09T15:17:44.602 INFO:teuthology.orchestra.run.vm06.stdout:Running transaction 2026-03-09T15:17:44.626 INFO:teuthology.orchestra.run.vm03.stdout: Preparing : 1/1 2026-03-09T15:17:44.628 INFO:teuthology.orchestra.run.vm03.stdout: Erasing : python3-cephfs-2:19.2.3-678.ge911bdeb.el9.x86_64 1/3 2026-03-09T15:17:44.629 INFO:teuthology.orchestra.run.vm04.stdout: Running scriptlet: libcephfs2-2:19.2.3-678.ge911bdeb.el9.x86_64 3/3 2026-03-09T15:17:44.629 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : libcephfs2-2:19.2.3-678.ge911bdeb.el9.x86_64 1/3 2026-03-09T15:17:44.629 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-ceph-argparse-2:19.2.3-678.ge911bdeb.el9.x86 2/3 2026-03-09T15:17:44.629 INFO:teuthology.orchestra.run.vm03.stdout: Erasing : python3-ceph-argparse-2:19.2.3-678.ge911bdeb.el9.x86 2/3 2026-03-09T15:17:44.629 INFO:teuthology.orchestra.run.vm03.stdout: Erasing : libcephfs2-2:19.2.3-678.ge911bdeb.el9.x86_64 3/3 2026-03-09T15:17:44.631 INFO:teuthology.orchestra.run.vm06.stdout: Preparing : 1/1 2026-03-09T15:17:44.633 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : python3-cephfs-2:19.2.3-678.ge911bdeb.el9.x86_64 1/3 2026-03-09T15:17:44.635 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : python3-ceph-argparse-2:19.2.3-678.ge911bdeb.el9.x86 2/3 2026-03-09T15:17:44.635 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : libcephfs2-2:19.2.3-678.ge911bdeb.el9.x86_64 3/3 2026-03-09T15:17:44.668 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-cephfs-2:19.2.3-678.ge911bdeb.el9.x86_64 3/3 2026-03-09T15:17:44.669 INFO:teuthology.orchestra.run.vm04.stdout: 2026-03-09T15:17:44.669 INFO:teuthology.orchestra.run.vm04.stdout:Removed: 2026-03-09T15:17:44.669 INFO:teuthology.orchestra.run.vm04.stdout: libcephfs2-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T15:17:44.669 INFO:teuthology.orchestra.run.vm04.stdout: python3-ceph-argparse-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T15:17:44.669 INFO:teuthology.orchestra.run.vm04.stdout: python3-cephfs-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T15:17:44.669 INFO:teuthology.orchestra.run.vm04.stdout: 2026-03-09T15:17:44.669 INFO:teuthology.orchestra.run.vm04.stdout:Complete! 2026-03-09T15:17:44.690 INFO:teuthology.orchestra.run.vm03.stdout: Running scriptlet: libcephfs2-2:19.2.3-678.ge911bdeb.el9.x86_64 3/3 2026-03-09T15:17:44.690 INFO:teuthology.orchestra.run.vm03.stdout: Verifying : libcephfs2-2:19.2.3-678.ge911bdeb.el9.x86_64 1/3 2026-03-09T15:17:44.690 INFO:teuthology.orchestra.run.vm03.stdout: Verifying : python3-ceph-argparse-2:19.2.3-678.ge911bdeb.el9.x86 2/3 2026-03-09T15:17:44.692 INFO:teuthology.orchestra.run.vm06.stdout: Running scriptlet: libcephfs2-2:19.2.3-678.ge911bdeb.el9.x86_64 3/3 2026-03-09T15:17:44.692 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : libcephfs2-2:19.2.3-678.ge911bdeb.el9.x86_64 1/3 2026-03-09T15:17:44.692 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-ceph-argparse-2:19.2.3-678.ge911bdeb.el9.x86 2/3 2026-03-09T15:17:44.732 INFO:teuthology.orchestra.run.vm03.stdout: Verifying : python3-cephfs-2:19.2.3-678.ge911bdeb.el9.x86_64 3/3 2026-03-09T15:17:44.732 INFO:teuthology.orchestra.run.vm03.stdout: 2026-03-09T15:17:44.732 INFO:teuthology.orchestra.run.vm03.stdout:Removed: 2026-03-09T15:17:44.732 INFO:teuthology.orchestra.run.vm03.stdout: libcephfs2-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T15:17:44.732 INFO:teuthology.orchestra.run.vm03.stdout: python3-ceph-argparse-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T15:17:44.732 INFO:teuthology.orchestra.run.vm03.stdout: python3-cephfs-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T15:17:44.732 INFO:teuthology.orchestra.run.vm03.stdout: 2026-03-09T15:17:44.732 INFO:teuthology.orchestra.run.vm03.stdout:Complete! 2026-03-09T15:17:44.735 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-cephfs-2:19.2.3-678.ge911bdeb.el9.x86_64 3/3 2026-03-09T15:17:44.735 INFO:teuthology.orchestra.run.vm06.stdout: 2026-03-09T15:17:44.735 INFO:teuthology.orchestra.run.vm06.stdout:Removed: 2026-03-09T15:17:44.735 INFO:teuthology.orchestra.run.vm06.stdout: libcephfs2-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T15:17:44.735 INFO:teuthology.orchestra.run.vm06.stdout: python3-ceph-argparse-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T15:17:44.735 INFO:teuthology.orchestra.run.vm06.stdout: python3-cephfs-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T15:17:44.735 INFO:teuthology.orchestra.run.vm06.stdout: 2026-03-09T15:17:44.735 INFO:teuthology.orchestra.run.vm06.stdout:Complete! 2026-03-09T15:17:44.838 INFO:teuthology.orchestra.run.vm04.stdout:No match for argument: libcephfs-devel 2026-03-09T15:17:44.838 INFO:teuthology.orchestra.run.vm04.stderr:No packages marked for removal. 2026-03-09T15:17:44.841 INFO:teuthology.orchestra.run.vm04.stdout:Dependencies resolved. 2026-03-09T15:17:44.842 INFO:teuthology.orchestra.run.vm04.stdout:Nothing to do. 2026-03-09T15:17:44.842 INFO:teuthology.orchestra.run.vm04.stdout:Complete! 2026-03-09T15:17:44.904 INFO:teuthology.orchestra.run.vm03.stdout:No match for argument: libcephfs-devel 2026-03-09T15:17:44.905 INFO:teuthology.orchestra.run.vm03.stderr:No packages marked for removal. 2026-03-09T15:17:44.908 INFO:teuthology.orchestra.run.vm03.stdout:Dependencies resolved. 2026-03-09T15:17:44.909 INFO:teuthology.orchestra.run.vm03.stdout:Nothing to do. 2026-03-09T15:17:44.909 INFO:teuthology.orchestra.run.vm03.stdout:Complete! 2026-03-09T15:17:44.912 INFO:teuthology.orchestra.run.vm06.stdout:No match for argument: libcephfs-devel 2026-03-09T15:17:44.912 INFO:teuthology.orchestra.run.vm06.stderr:No packages marked for removal. 2026-03-09T15:17:44.915 INFO:teuthology.orchestra.run.vm06.stdout:Dependencies resolved. 2026-03-09T15:17:44.916 INFO:teuthology.orchestra.run.vm06.stdout:Nothing to do. 2026-03-09T15:17:44.916 INFO:teuthology.orchestra.run.vm06.stdout:Complete! 2026-03-09T15:17:45.019 INFO:teuthology.orchestra.run.vm04.stdout:Dependencies resolved. 2026-03-09T15:17:45.021 INFO:teuthology.orchestra.run.vm04.stdout:================================================================================ 2026-03-09T15:17:45.021 INFO:teuthology.orchestra.run.vm04.stdout: Package Arch Version Repository Size 2026-03-09T15:17:45.021 INFO:teuthology.orchestra.run.vm04.stdout:================================================================================ 2026-03-09T15:17:45.021 INFO:teuthology.orchestra.run.vm04.stdout:Removing: 2026-03-09T15:17:45.021 INFO:teuthology.orchestra.run.vm04.stdout: librados2 x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 12 M 2026-03-09T15:17:45.021 INFO:teuthology.orchestra.run.vm04.stdout:Removing dependent packages: 2026-03-09T15:17:45.021 INFO:teuthology.orchestra.run.vm04.stdout: python3-rados x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 1.1 M 2026-03-09T15:17:45.021 INFO:teuthology.orchestra.run.vm04.stdout: python3-rbd x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 1.1 M 2026-03-09T15:17:45.021 INFO:teuthology.orchestra.run.vm04.stdout: python3-rgw x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 265 k 2026-03-09T15:17:45.021 INFO:teuthology.orchestra.run.vm04.stdout: qemu-kvm-block-rbd x86_64 17:10.1.0-15.el9 @appstream 37 k 2026-03-09T15:17:45.021 INFO:teuthology.orchestra.run.vm04.stdout: rbd-fuse x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 227 k 2026-03-09T15:17:45.021 INFO:teuthology.orchestra.run.vm04.stdout: rbd-nbd x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 490 k 2026-03-09T15:17:45.021 INFO:teuthology.orchestra.run.vm04.stdout:Removing unused dependencies: 2026-03-09T15:17:45.021 INFO:teuthology.orchestra.run.vm04.stdout: boost-program-options x86_64 1.75.0-13.el9 @appstream 276 k 2026-03-09T15:17:45.021 INFO:teuthology.orchestra.run.vm04.stdout: libarrow x86_64 9.0.0-15.el9 @epel 18 M 2026-03-09T15:17:45.021 INFO:teuthology.orchestra.run.vm04.stdout: libarrow-doc noarch 9.0.0-15.el9 @epel 122 k 2026-03-09T15:17:45.021 INFO:teuthology.orchestra.run.vm04.stdout: libnbd x86_64 1.20.3-4.el9 @appstream 453 k 2026-03-09T15:17:45.021 INFO:teuthology.orchestra.run.vm04.stdout: libpmemobj x86_64 1.12.1-1.el9 @appstream 383 k 2026-03-09T15:17:45.021 INFO:teuthology.orchestra.run.vm04.stdout: librabbitmq x86_64 0.11.0-7.el9 @appstream 102 k 2026-03-09T15:17:45.021 INFO:teuthology.orchestra.run.vm04.stdout: librbd1 x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 13 M 2026-03-09T15:17:45.021 INFO:teuthology.orchestra.run.vm04.stdout: librdkafka x86_64 1.6.1-102.el9 @appstream 2.0 M 2026-03-09T15:17:45.021 INFO:teuthology.orchestra.run.vm04.stdout: librgw2 x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 19 M 2026-03-09T15:17:45.021 INFO:teuthology.orchestra.run.vm04.stdout: lttng-ust x86_64 2.12.0-6.el9 @appstream 1.0 M 2026-03-09T15:17:45.021 INFO:teuthology.orchestra.run.vm04.stdout: parquet-libs x86_64 9.0.0-15.el9 @epel 2.8 M 2026-03-09T15:17:45.021 INFO:teuthology.orchestra.run.vm04.stdout: re2 x86_64 1:20211101-20.el9 @epel 472 k 2026-03-09T15:17:45.021 INFO:teuthology.orchestra.run.vm04.stdout: thrift x86_64 0.15.0-4.el9 @epel 4.8 M 2026-03-09T15:17:45.021 INFO:teuthology.orchestra.run.vm04.stdout: 2026-03-09T15:17:45.021 INFO:teuthology.orchestra.run.vm04.stdout:Transaction Summary 2026-03-09T15:17:45.021 INFO:teuthology.orchestra.run.vm04.stdout:================================================================================ 2026-03-09T15:17:45.021 INFO:teuthology.orchestra.run.vm04.stdout:Remove 20 Packages 2026-03-09T15:17:45.021 INFO:teuthology.orchestra.run.vm04.stdout: 2026-03-09T15:17:45.022 INFO:teuthology.orchestra.run.vm04.stdout:Freed space: 79 M 2026-03-09T15:17:45.022 INFO:teuthology.orchestra.run.vm04.stdout:Running transaction check 2026-03-09T15:17:45.025 INFO:teuthology.orchestra.run.vm04.stdout:Transaction check succeeded. 2026-03-09T15:17:45.025 INFO:teuthology.orchestra.run.vm04.stdout:Running transaction test 2026-03-09T15:17:45.047 INFO:teuthology.orchestra.run.vm04.stdout:Transaction test succeeded. 2026-03-09T15:17:45.047 INFO:teuthology.orchestra.run.vm04.stdout:Running transaction 2026-03-09T15:17:45.087 INFO:teuthology.orchestra.run.vm04.stdout: Preparing : 1/1 2026-03-09T15:17:45.090 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : rbd-nbd-2:19.2.3-678.ge911bdeb.el9.x86_64 1/20 2026-03-09T15:17:45.090 INFO:teuthology.orchestra.run.vm03.stdout:Dependencies resolved. 2026-03-09T15:17:45.092 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : rbd-fuse-2:19.2.3-678.ge911bdeb.el9.x86_64 2/20 2026-03-09T15:17:45.092 INFO:teuthology.orchestra.run.vm03.stdout:================================================================================ 2026-03-09T15:17:45.092 INFO:teuthology.orchestra.run.vm03.stdout: Package Arch Version Repository Size 2026-03-09T15:17:45.092 INFO:teuthology.orchestra.run.vm03.stdout:================================================================================ 2026-03-09T15:17:45.092 INFO:teuthology.orchestra.run.vm03.stdout:Removing: 2026-03-09T15:17:45.092 INFO:teuthology.orchestra.run.vm03.stdout: librados2 x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 12 M 2026-03-09T15:17:45.092 INFO:teuthology.orchestra.run.vm03.stdout:Removing dependent packages: 2026-03-09T15:17:45.092 INFO:teuthology.orchestra.run.vm03.stdout: python3-rados x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 1.1 M 2026-03-09T15:17:45.092 INFO:teuthology.orchestra.run.vm03.stdout: python3-rbd x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 1.1 M 2026-03-09T15:17:45.092 INFO:teuthology.orchestra.run.vm03.stdout: python3-rgw x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 265 k 2026-03-09T15:17:45.092 INFO:teuthology.orchestra.run.vm03.stdout: qemu-kvm-block-rbd x86_64 17:10.1.0-15.el9 @appstream 37 k 2026-03-09T15:17:45.092 INFO:teuthology.orchestra.run.vm03.stdout: rbd-fuse x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 227 k 2026-03-09T15:17:45.092 INFO:teuthology.orchestra.run.vm03.stdout: rbd-nbd x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 490 k 2026-03-09T15:17:45.092 INFO:teuthology.orchestra.run.vm03.stdout:Removing unused dependencies: 2026-03-09T15:17:45.092 INFO:teuthology.orchestra.run.vm03.stdout: boost-program-options x86_64 1.75.0-13.el9 @appstream 276 k 2026-03-09T15:17:45.092 INFO:teuthology.orchestra.run.vm03.stdout: libarrow x86_64 9.0.0-15.el9 @epel 18 M 2026-03-09T15:17:45.092 INFO:teuthology.orchestra.run.vm03.stdout: libarrow-doc noarch 9.0.0-15.el9 @epel 122 k 2026-03-09T15:17:45.092 INFO:teuthology.orchestra.run.vm03.stdout: libnbd x86_64 1.20.3-4.el9 @appstream 453 k 2026-03-09T15:17:45.092 INFO:teuthology.orchestra.run.vm03.stdout: libpmemobj x86_64 1.12.1-1.el9 @appstream 383 k 2026-03-09T15:17:45.092 INFO:teuthology.orchestra.run.vm03.stdout: librabbitmq x86_64 0.11.0-7.el9 @appstream 102 k 2026-03-09T15:17:45.092 INFO:teuthology.orchestra.run.vm03.stdout: librbd1 x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 13 M 2026-03-09T15:17:45.092 INFO:teuthology.orchestra.run.vm03.stdout: librdkafka x86_64 1.6.1-102.el9 @appstream 2.0 M 2026-03-09T15:17:45.092 INFO:teuthology.orchestra.run.vm03.stdout: librgw2 x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 19 M 2026-03-09T15:17:45.093 INFO:teuthology.orchestra.run.vm03.stdout: lttng-ust x86_64 2.12.0-6.el9 @appstream 1.0 M 2026-03-09T15:17:45.093 INFO:teuthology.orchestra.run.vm03.stdout: parquet-libs x86_64 9.0.0-15.el9 @epel 2.8 M 2026-03-09T15:17:45.093 INFO:teuthology.orchestra.run.vm03.stdout: re2 x86_64 1:20211101-20.el9 @epel 472 k 2026-03-09T15:17:45.093 INFO:teuthology.orchestra.run.vm03.stdout: thrift x86_64 0.15.0-4.el9 @epel 4.8 M 2026-03-09T15:17:45.093 INFO:teuthology.orchestra.run.vm03.stdout: 2026-03-09T15:17:45.093 INFO:teuthology.orchestra.run.vm03.stdout:Transaction Summary 2026-03-09T15:17:45.093 INFO:teuthology.orchestra.run.vm03.stdout:================================================================================ 2026-03-09T15:17:45.093 INFO:teuthology.orchestra.run.vm03.stdout:Remove 20 Packages 2026-03-09T15:17:45.093 INFO:teuthology.orchestra.run.vm03.stdout: 2026-03-09T15:17:45.093 INFO:teuthology.orchestra.run.vm03.stdout:Freed space: 79 M 2026-03-09T15:17:45.093 INFO:teuthology.orchestra.run.vm03.stdout:Running transaction check 2026-03-09T15:17:45.095 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : python3-rgw-2:19.2.3-678.ge911bdeb.el9.x86_64 3/20 2026-03-09T15:17:45.095 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : librgw2-2:19.2.3-678.ge911bdeb.el9.x86_64 4/20 2026-03-09T15:17:45.096 INFO:teuthology.orchestra.run.vm03.stdout:Transaction check succeeded. 2026-03-09T15:17:45.096 INFO:teuthology.orchestra.run.vm03.stdout:Running transaction test 2026-03-09T15:17:45.102 INFO:teuthology.orchestra.run.vm06.stdout:Dependencies resolved. 2026-03-09T15:17:45.104 INFO:teuthology.orchestra.run.vm06.stdout:================================================================================ 2026-03-09T15:17:45.104 INFO:teuthology.orchestra.run.vm06.stdout: Package Arch Version Repository Size 2026-03-09T15:17:45.104 INFO:teuthology.orchestra.run.vm06.stdout:================================================================================ 2026-03-09T15:17:45.104 INFO:teuthology.orchestra.run.vm06.stdout:Removing: 2026-03-09T15:17:45.104 INFO:teuthology.orchestra.run.vm06.stdout: librados2 x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 12 M 2026-03-09T15:17:45.104 INFO:teuthology.orchestra.run.vm06.stdout:Removing dependent packages: 2026-03-09T15:17:45.104 INFO:teuthology.orchestra.run.vm06.stdout: python3-rados x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 1.1 M 2026-03-09T15:17:45.104 INFO:teuthology.orchestra.run.vm06.stdout: python3-rbd x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 1.1 M 2026-03-09T15:17:45.104 INFO:teuthology.orchestra.run.vm06.stdout: python3-rgw x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 265 k 2026-03-09T15:17:45.104 INFO:teuthology.orchestra.run.vm06.stdout: qemu-kvm-block-rbd x86_64 17:10.1.0-15.el9 @appstream 37 k 2026-03-09T15:17:45.104 INFO:teuthology.orchestra.run.vm06.stdout: rbd-fuse x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 227 k 2026-03-09T15:17:45.104 INFO:teuthology.orchestra.run.vm06.stdout: rbd-nbd x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 490 k 2026-03-09T15:17:45.104 INFO:teuthology.orchestra.run.vm06.stdout:Removing unused dependencies: 2026-03-09T15:17:45.104 INFO:teuthology.orchestra.run.vm06.stdout: boost-program-options x86_64 1.75.0-13.el9 @appstream 276 k 2026-03-09T15:17:45.104 INFO:teuthology.orchestra.run.vm06.stdout: libarrow x86_64 9.0.0-15.el9 @epel 18 M 2026-03-09T15:17:45.104 INFO:teuthology.orchestra.run.vm06.stdout: libarrow-doc noarch 9.0.0-15.el9 @epel 122 k 2026-03-09T15:17:45.104 INFO:teuthology.orchestra.run.vm06.stdout: libnbd x86_64 1.20.3-4.el9 @appstream 453 k 2026-03-09T15:17:45.104 INFO:teuthology.orchestra.run.vm06.stdout: libpmemobj x86_64 1.12.1-1.el9 @appstream 383 k 2026-03-09T15:17:45.104 INFO:teuthology.orchestra.run.vm06.stdout: librabbitmq x86_64 0.11.0-7.el9 @appstream 102 k 2026-03-09T15:17:45.104 INFO:teuthology.orchestra.run.vm06.stdout: librbd1 x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 13 M 2026-03-09T15:17:45.104 INFO:teuthology.orchestra.run.vm06.stdout: librdkafka x86_64 1.6.1-102.el9 @appstream 2.0 M 2026-03-09T15:17:45.104 INFO:teuthology.orchestra.run.vm06.stdout: librgw2 x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 19 M 2026-03-09T15:17:45.104 INFO:teuthology.orchestra.run.vm06.stdout: lttng-ust x86_64 2.12.0-6.el9 @appstream 1.0 M 2026-03-09T15:17:45.104 INFO:teuthology.orchestra.run.vm06.stdout: parquet-libs x86_64 9.0.0-15.el9 @epel 2.8 M 2026-03-09T15:17:45.104 INFO:teuthology.orchestra.run.vm06.stdout: re2 x86_64 1:20211101-20.el9 @epel 472 k 2026-03-09T15:17:45.104 INFO:teuthology.orchestra.run.vm06.stdout: thrift x86_64 0.15.0-4.el9 @epel 4.8 M 2026-03-09T15:17:45.104 INFO:teuthology.orchestra.run.vm06.stdout: 2026-03-09T15:17:45.104 INFO:teuthology.orchestra.run.vm06.stdout:Transaction Summary 2026-03-09T15:17:45.104 INFO:teuthology.orchestra.run.vm06.stdout:================================================================================ 2026-03-09T15:17:45.104 INFO:teuthology.orchestra.run.vm06.stdout:Remove 20 Packages 2026-03-09T15:17:45.104 INFO:teuthology.orchestra.run.vm06.stdout: 2026-03-09T15:17:45.105 INFO:teuthology.orchestra.run.vm06.stdout:Freed space: 79 M 2026-03-09T15:17:45.105 INFO:teuthology.orchestra.run.vm06.stdout:Running transaction check 2026-03-09T15:17:45.108 INFO:teuthology.orchestra.run.vm06.stdout:Transaction check succeeded. 2026-03-09T15:17:45.109 INFO:teuthology.orchestra.run.vm06.stdout:Running transaction test 2026-03-09T15:17:45.110 INFO:teuthology.orchestra.run.vm04.stdout: Running scriptlet: librgw2-2:19.2.3-678.ge911bdeb.el9.x86_64 4/20 2026-03-09T15:17:45.112 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : parquet-libs-9.0.0-15.el9.x86_64 5/20 2026-03-09T15:17:45.115 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : python3-rbd-2:19.2.3-678.ge911bdeb.el9.x86_64 6/20 2026-03-09T15:17:45.118 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : python3-rados-2:19.2.3-678.ge911bdeb.el9.x86_64 7/20 2026-03-09T15:17:45.120 INFO:teuthology.orchestra.run.vm03.stdout:Transaction test succeeded. 2026-03-09T15:17:45.120 INFO:teuthology.orchestra.run.vm03.stdout:Running transaction 2026-03-09T15:17:45.120 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : qemu-kvm-block-rbd-17:10.1.0-15.el9.x86_64 8/20 2026-03-09T15:17:45.123 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : libarrow-doc-9.0.0-15.el9.noarch 9/20 2026-03-09T15:17:45.123 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : librbd1-2:19.2.3-678.ge911bdeb.el9.x86_64 10/20 2026-03-09T15:17:45.134 INFO:teuthology.orchestra.run.vm06.stdout:Transaction test succeeded. 2026-03-09T15:17:45.134 INFO:teuthology.orchestra.run.vm06.stdout:Running transaction 2026-03-09T15:17:45.138 INFO:teuthology.orchestra.run.vm04.stdout: Running scriptlet: librbd1-2:19.2.3-678.ge911bdeb.el9.x86_64 10/20 2026-03-09T15:17:45.138 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : librados2-2:19.2.3-678.ge911bdeb.el9.x86_64 11/20 2026-03-09T15:17:45.138 INFO:teuthology.orchestra.run.vm04.stdout:warning: file /etc/ceph: remove failed: No such file or directory 2026-03-09T15:17:45.138 INFO:teuthology.orchestra.run.vm04.stdout: 2026-03-09T15:17:45.153 INFO:teuthology.orchestra.run.vm04.stdout: Running scriptlet: librados2-2:19.2.3-678.ge911bdeb.el9.x86_64 11/20 2026-03-09T15:17:45.155 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : libarrow-9.0.0-15.el9.x86_64 12/20 2026-03-09T15:17:45.158 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : re2-1:20211101-20.el9.x86_64 13/20 2026-03-09T15:17:45.162 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : lttng-ust-2.12.0-6.el9.x86_64 14/20 2026-03-09T15:17:45.163 INFO:teuthology.orchestra.run.vm03.stdout: Preparing : 1/1 2026-03-09T15:17:45.165 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : thrift-0.15.0-4.el9.x86_64 15/20 2026-03-09T15:17:45.166 INFO:teuthology.orchestra.run.vm03.stdout: Erasing : rbd-nbd-2:19.2.3-678.ge911bdeb.el9.x86_64 1/20 2026-03-09T15:17:45.167 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : libnbd-1.20.3-4.el9.x86_64 16/20 2026-03-09T15:17:45.168 INFO:teuthology.orchestra.run.vm03.stdout: Erasing : rbd-fuse-2:19.2.3-678.ge911bdeb.el9.x86_64 2/20 2026-03-09T15:17:45.169 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : libpmemobj-1.12.1-1.el9.x86_64 17/20 2026-03-09T15:17:45.171 INFO:teuthology.orchestra.run.vm03.stdout: Erasing : python3-rgw-2:19.2.3-678.ge911bdeb.el9.x86_64 3/20 2026-03-09T15:17:45.171 INFO:teuthology.orchestra.run.vm03.stdout: Erasing : librgw2-2:19.2.3-678.ge911bdeb.el9.x86_64 4/20 2026-03-09T15:17:45.171 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : boost-program-options-1.75.0-13.el9.x86_64 18/20 2026-03-09T15:17:45.174 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : librabbitmq-0.11.0-7.el9.x86_64 19/20 2026-03-09T15:17:45.175 INFO:teuthology.orchestra.run.vm06.stdout: Preparing : 1/1 2026-03-09T15:17:45.178 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : rbd-nbd-2:19.2.3-678.ge911bdeb.el9.x86_64 1/20 2026-03-09T15:17:45.180 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : rbd-fuse-2:19.2.3-678.ge911bdeb.el9.x86_64 2/20 2026-03-09T15:17:45.183 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : python3-rgw-2:19.2.3-678.ge911bdeb.el9.x86_64 3/20 2026-03-09T15:17:45.183 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : librgw2-2:19.2.3-678.ge911bdeb.el9.x86_64 4/20 2026-03-09T15:17:45.184 INFO:teuthology.orchestra.run.vm03.stdout: Running scriptlet: librgw2-2:19.2.3-678.ge911bdeb.el9.x86_64 4/20 2026-03-09T15:17:45.186 INFO:teuthology.orchestra.run.vm03.stdout: Erasing : parquet-libs-9.0.0-15.el9.x86_64 5/20 2026-03-09T15:17:45.187 INFO:teuthology.orchestra.run.vm03.stdout: Erasing : python3-rbd-2:19.2.3-678.ge911bdeb.el9.x86_64 6/20 2026-03-09T15:17:45.188 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : librdkafka-1.6.1-102.el9.x86_64 20/20 2026-03-09T15:17:45.189 INFO:teuthology.orchestra.run.vm03.stdout: Erasing : python3-rados-2:19.2.3-678.ge911bdeb.el9.x86_64 7/20 2026-03-09T15:17:45.190 INFO:teuthology.orchestra.run.vm03.stdout: Erasing : qemu-kvm-block-rbd-17:10.1.0-15.el9.x86_64 8/20 2026-03-09T15:17:45.193 INFO:teuthology.orchestra.run.vm03.stdout: Erasing : libarrow-doc-9.0.0-15.el9.noarch 9/20 2026-03-09T15:17:45.193 INFO:teuthology.orchestra.run.vm03.stdout: Erasing : librbd1-2:19.2.3-678.ge911bdeb.el9.x86_64 10/20 2026-03-09T15:17:45.195 INFO:teuthology.orchestra.run.vm06.stdout: Running scriptlet: librgw2-2:19.2.3-678.ge911bdeb.el9.x86_64 4/20 2026-03-09T15:17:45.197 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : parquet-libs-9.0.0-15.el9.x86_64 5/20 2026-03-09T15:17:45.198 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : python3-rbd-2:19.2.3-678.ge911bdeb.el9.x86_64 6/20 2026-03-09T15:17:45.200 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : python3-rados-2:19.2.3-678.ge911bdeb.el9.x86_64 7/20 2026-03-09T15:17:45.202 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : qemu-kvm-block-rbd-17:10.1.0-15.el9.x86_64 8/20 2026-03-09T15:17:45.204 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : libarrow-doc-9.0.0-15.el9.noarch 9/20 2026-03-09T15:17:45.204 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : librbd1-2:19.2.3-678.ge911bdeb.el9.x86_64 10/20 2026-03-09T15:17:45.206 INFO:teuthology.orchestra.run.vm03.stdout: Running scriptlet: librbd1-2:19.2.3-678.ge911bdeb.el9.x86_64 10/20 2026-03-09T15:17:45.206 INFO:teuthology.orchestra.run.vm03.stdout: Erasing : librados2-2:19.2.3-678.ge911bdeb.el9.x86_64 11/20 2026-03-09T15:17:45.206 INFO:teuthology.orchestra.run.vm03.stdout:warning: file /etc/ceph: remove failed: No such file or directory 2026-03-09T15:17:45.206 INFO:teuthology.orchestra.run.vm03.stdout: 2026-03-09T15:17:45.220 INFO:teuthology.orchestra.run.vm06.stdout: Running scriptlet: librbd1-2:19.2.3-678.ge911bdeb.el9.x86_64 10/20 2026-03-09T15:17:45.220 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : librados2-2:19.2.3-678.ge911bdeb.el9.x86_64 11/20 2026-03-09T15:17:45.220 INFO:teuthology.orchestra.run.vm06.stdout:warning: file /etc/ceph: remove failed: No such file or directory 2026-03-09T15:17:45.220 INFO:teuthology.orchestra.run.vm06.stdout: 2026-03-09T15:17:45.223 INFO:teuthology.orchestra.run.vm03.stdout: Running scriptlet: librados2-2:19.2.3-678.ge911bdeb.el9.x86_64 11/20 2026-03-09T15:17:45.226 INFO:teuthology.orchestra.run.vm03.stdout: Erasing : libarrow-9.0.0-15.el9.x86_64 12/20 2026-03-09T15:17:45.229 INFO:teuthology.orchestra.run.vm03.stdout: Erasing : re2-1:20211101-20.el9.x86_64 13/20 2026-03-09T15:17:45.232 INFO:teuthology.orchestra.run.vm03.stdout: Erasing : lttng-ust-2.12.0-6.el9.x86_64 14/20 2026-03-09T15:17:45.234 INFO:teuthology.orchestra.run.vm06.stdout: Running scriptlet: librados2-2:19.2.3-678.ge911bdeb.el9.x86_64 11/20 2026-03-09T15:17:45.234 INFO:teuthology.orchestra.run.vm03.stdout: Erasing : thrift-0.15.0-4.el9.x86_64 15/20 2026-03-09T15:17:45.236 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : libarrow-9.0.0-15.el9.x86_64 12/20 2026-03-09T15:17:45.237 INFO:teuthology.orchestra.run.vm03.stdout: Erasing : libnbd-1.20.3-4.el9.x86_64 16/20 2026-03-09T15:17:45.239 INFO:teuthology.orchestra.run.vm03.stdout: Erasing : libpmemobj-1.12.1-1.el9.x86_64 17/20 2026-03-09T15:17:45.239 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : re2-1:20211101-20.el9.x86_64 13/20 2026-03-09T15:17:45.240 INFO:teuthology.orchestra.run.vm03.stdout: Erasing : boost-program-options-1.75.0-13.el9.x86_64 18/20 2026-03-09T15:17:45.243 INFO:teuthology.orchestra.run.vm03.stdout: Erasing : librabbitmq-0.11.0-7.el9.x86_64 19/20 2026-03-09T15:17:45.243 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : lttng-ust-2.12.0-6.el9.x86_64 14/20 2026-03-09T15:17:45.245 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : thrift-0.15.0-4.el9.x86_64 15/20 2026-03-09T15:17:45.248 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : libnbd-1.20.3-4.el9.x86_64 16/20 2026-03-09T15:17:45.250 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : libpmemobj-1.12.1-1.el9.x86_64 17/20 2026-03-09T15:17:45.251 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : boost-program-options-1.75.0-13.el9.x86_64 18/20 2026-03-09T15:17:45.253 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : librabbitmq-0.11.0-7.el9.x86_64 19/20 2026-03-09T15:17:45.256 INFO:teuthology.orchestra.run.vm03.stdout: Erasing : librdkafka-1.6.1-102.el9.x86_64 20/20 2026-03-09T15:17:45.257 INFO:teuthology.orchestra.run.vm04.stdout: Running scriptlet: librdkafka-1.6.1-102.el9.x86_64 20/20 2026-03-09T15:17:45.257 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : boost-program-options-1.75.0-13.el9.x86_64 1/20 2026-03-09T15:17:45.257 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : libarrow-9.0.0-15.el9.x86_64 2/20 2026-03-09T15:17:45.257 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : libarrow-doc-9.0.0-15.el9.noarch 3/20 2026-03-09T15:17:45.257 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : libnbd-1.20.3-4.el9.x86_64 4/20 2026-03-09T15:17:45.257 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : libpmemobj-1.12.1-1.el9.x86_64 5/20 2026-03-09T15:17:45.257 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : librabbitmq-0.11.0-7.el9.x86_64 6/20 2026-03-09T15:17:45.257 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : librados2-2:19.2.3-678.ge911bdeb.el9.x86_64 7/20 2026-03-09T15:17:45.257 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : librbd1-2:19.2.3-678.ge911bdeb.el9.x86_64 8/20 2026-03-09T15:17:45.257 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : librdkafka-1.6.1-102.el9.x86_64 9/20 2026-03-09T15:17:45.257 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : librgw2-2:19.2.3-678.ge911bdeb.el9.x86_64 10/20 2026-03-09T15:17:45.257 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : lttng-ust-2.12.0-6.el9.x86_64 11/20 2026-03-09T15:17:45.258 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : parquet-libs-9.0.0-15.el9.x86_64 12/20 2026-03-09T15:17:45.258 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-rados-2:19.2.3-678.ge911bdeb.el9.x86_64 13/20 2026-03-09T15:17:45.258 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-rbd-2:19.2.3-678.ge911bdeb.el9.x86_64 14/20 2026-03-09T15:17:45.258 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-rgw-2:19.2.3-678.ge911bdeb.el9.x86_64 15/20 2026-03-09T15:17:45.258 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : qemu-kvm-block-rbd-17:10.1.0-15.el9.x86_64 16/20 2026-03-09T15:17:45.258 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : rbd-fuse-2:19.2.3-678.ge911bdeb.el9.x86_64 17/20 2026-03-09T15:17:45.258 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : rbd-nbd-2:19.2.3-678.ge911bdeb.el9.x86_64 18/20 2026-03-09T15:17:45.258 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : re2-1:20211101-20.el9.x86_64 19/20 2026-03-09T15:17:45.268 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : librdkafka-1.6.1-102.el9.x86_64 20/20 2026-03-09T15:17:45.302 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : thrift-0.15.0-4.el9.x86_64 20/20 2026-03-09T15:17:45.302 INFO:teuthology.orchestra.run.vm04.stdout: 2026-03-09T15:17:45.302 INFO:teuthology.orchestra.run.vm04.stdout:Removed: 2026-03-09T15:17:45.302 INFO:teuthology.orchestra.run.vm04.stdout: boost-program-options-1.75.0-13.el9.x86_64 2026-03-09T15:17:45.302 INFO:teuthology.orchestra.run.vm04.stdout: libarrow-9.0.0-15.el9.x86_64 2026-03-09T15:17:45.302 INFO:teuthology.orchestra.run.vm04.stdout: libarrow-doc-9.0.0-15.el9.noarch 2026-03-09T15:17:45.302 INFO:teuthology.orchestra.run.vm04.stdout: libnbd-1.20.3-4.el9.x86_64 2026-03-09T15:17:45.302 INFO:teuthology.orchestra.run.vm04.stdout: libpmemobj-1.12.1-1.el9.x86_64 2026-03-09T15:17:45.302 INFO:teuthology.orchestra.run.vm04.stdout: librabbitmq-0.11.0-7.el9.x86_64 2026-03-09T15:17:45.302 INFO:teuthology.orchestra.run.vm04.stdout: librados2-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T15:17:45.302 INFO:teuthology.orchestra.run.vm04.stdout: librbd1-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T15:17:45.302 INFO:teuthology.orchestra.run.vm04.stdout: librdkafka-1.6.1-102.el9.x86_64 2026-03-09T15:17:45.302 INFO:teuthology.orchestra.run.vm04.stdout: librgw2-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T15:17:45.302 INFO:teuthology.orchestra.run.vm04.stdout: lttng-ust-2.12.0-6.el9.x86_64 2026-03-09T15:17:45.302 INFO:teuthology.orchestra.run.vm04.stdout: parquet-libs-9.0.0-15.el9.x86_64 2026-03-09T15:17:45.302 INFO:teuthology.orchestra.run.vm04.stdout: python3-rados-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T15:17:45.302 INFO:teuthology.orchestra.run.vm04.stdout: python3-rbd-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T15:17:45.302 INFO:teuthology.orchestra.run.vm04.stdout: python3-rgw-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T15:17:45.302 INFO:teuthology.orchestra.run.vm04.stdout: qemu-kvm-block-rbd-17:10.1.0-15.el9.x86_64 2026-03-09T15:17:45.302 INFO:teuthology.orchestra.run.vm04.stdout: rbd-fuse-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T15:17:45.302 INFO:teuthology.orchestra.run.vm04.stdout: rbd-nbd-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T15:17:45.302 INFO:teuthology.orchestra.run.vm04.stdout: re2-1:20211101-20.el9.x86_64 2026-03-09T15:17:45.302 INFO:teuthology.orchestra.run.vm04.stdout: thrift-0.15.0-4.el9.x86_64 2026-03-09T15:17:45.302 INFO:teuthology.orchestra.run.vm04.stdout: 2026-03-09T15:17:45.302 INFO:teuthology.orchestra.run.vm04.stdout:Complete! 2026-03-09T15:17:45.315 INFO:teuthology.orchestra.run.vm03.stdout: Running scriptlet: librdkafka-1.6.1-102.el9.x86_64 20/20 2026-03-09T15:17:45.315 INFO:teuthology.orchestra.run.vm03.stdout: Verifying : boost-program-options-1.75.0-13.el9.x86_64 1/20 2026-03-09T15:17:45.315 INFO:teuthology.orchestra.run.vm03.stdout: Verifying : libarrow-9.0.0-15.el9.x86_64 2/20 2026-03-09T15:17:45.315 INFO:teuthology.orchestra.run.vm03.stdout: Verifying : libarrow-doc-9.0.0-15.el9.noarch 3/20 2026-03-09T15:17:45.315 INFO:teuthology.orchestra.run.vm03.stdout: Verifying : libnbd-1.20.3-4.el9.x86_64 4/20 2026-03-09T15:17:45.315 INFO:teuthology.orchestra.run.vm03.stdout: Verifying : libpmemobj-1.12.1-1.el9.x86_64 5/20 2026-03-09T15:17:45.315 INFO:teuthology.orchestra.run.vm03.stdout: Verifying : librabbitmq-0.11.0-7.el9.x86_64 6/20 2026-03-09T15:17:45.315 INFO:teuthology.orchestra.run.vm03.stdout: Verifying : librados2-2:19.2.3-678.ge911bdeb.el9.x86_64 7/20 2026-03-09T15:17:45.315 INFO:teuthology.orchestra.run.vm03.stdout: Verifying : librbd1-2:19.2.3-678.ge911bdeb.el9.x86_64 8/20 2026-03-09T15:17:45.315 INFO:teuthology.orchestra.run.vm03.stdout: Verifying : librdkafka-1.6.1-102.el9.x86_64 9/20 2026-03-09T15:17:45.315 INFO:teuthology.orchestra.run.vm03.stdout: Verifying : librgw2-2:19.2.3-678.ge911bdeb.el9.x86_64 10/20 2026-03-09T15:17:45.315 INFO:teuthology.orchestra.run.vm03.stdout: Verifying : lttng-ust-2.12.0-6.el9.x86_64 11/20 2026-03-09T15:17:45.315 INFO:teuthology.orchestra.run.vm03.stdout: Verifying : parquet-libs-9.0.0-15.el9.x86_64 12/20 2026-03-09T15:17:45.316 INFO:teuthology.orchestra.run.vm03.stdout: Verifying : python3-rados-2:19.2.3-678.ge911bdeb.el9.x86_64 13/20 2026-03-09T15:17:45.316 INFO:teuthology.orchestra.run.vm03.stdout: Verifying : python3-rbd-2:19.2.3-678.ge911bdeb.el9.x86_64 14/20 2026-03-09T15:17:45.316 INFO:teuthology.orchestra.run.vm03.stdout: Verifying : python3-rgw-2:19.2.3-678.ge911bdeb.el9.x86_64 15/20 2026-03-09T15:17:45.316 INFO:teuthology.orchestra.run.vm03.stdout: Verifying : qemu-kvm-block-rbd-17:10.1.0-15.el9.x86_64 16/20 2026-03-09T15:17:45.316 INFO:teuthology.orchestra.run.vm03.stdout: Verifying : rbd-fuse-2:19.2.3-678.ge911bdeb.el9.x86_64 17/20 2026-03-09T15:17:45.316 INFO:teuthology.orchestra.run.vm03.stdout: Verifying : rbd-nbd-2:19.2.3-678.ge911bdeb.el9.x86_64 18/20 2026-03-09T15:17:45.316 INFO:teuthology.orchestra.run.vm03.stdout: Verifying : re2-1:20211101-20.el9.x86_64 19/20 2026-03-09T15:17:45.330 INFO:teuthology.orchestra.run.vm06.stdout: Running scriptlet: librdkafka-1.6.1-102.el9.x86_64 20/20 2026-03-09T15:17:45.331 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : boost-program-options-1.75.0-13.el9.x86_64 1/20 2026-03-09T15:17:45.331 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : libarrow-9.0.0-15.el9.x86_64 2/20 2026-03-09T15:17:45.331 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : libarrow-doc-9.0.0-15.el9.noarch 3/20 2026-03-09T15:17:45.331 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : libnbd-1.20.3-4.el9.x86_64 4/20 2026-03-09T15:17:45.331 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : libpmemobj-1.12.1-1.el9.x86_64 5/20 2026-03-09T15:17:45.331 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : librabbitmq-0.11.0-7.el9.x86_64 6/20 2026-03-09T15:17:45.331 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : librados2-2:19.2.3-678.ge911bdeb.el9.x86_64 7/20 2026-03-09T15:17:45.331 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : librbd1-2:19.2.3-678.ge911bdeb.el9.x86_64 8/20 2026-03-09T15:17:45.331 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : librdkafka-1.6.1-102.el9.x86_64 9/20 2026-03-09T15:17:45.331 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : librgw2-2:19.2.3-678.ge911bdeb.el9.x86_64 10/20 2026-03-09T15:17:45.331 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : lttng-ust-2.12.0-6.el9.x86_64 11/20 2026-03-09T15:17:45.331 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : parquet-libs-9.0.0-15.el9.x86_64 12/20 2026-03-09T15:17:45.331 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-rados-2:19.2.3-678.ge911bdeb.el9.x86_64 13/20 2026-03-09T15:17:45.331 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-rbd-2:19.2.3-678.ge911bdeb.el9.x86_64 14/20 2026-03-09T15:17:45.331 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-rgw-2:19.2.3-678.ge911bdeb.el9.x86_64 15/20 2026-03-09T15:17:45.331 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : qemu-kvm-block-rbd-17:10.1.0-15.el9.x86_64 16/20 2026-03-09T15:17:45.331 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : rbd-fuse-2:19.2.3-678.ge911bdeb.el9.x86_64 17/20 2026-03-09T15:17:45.331 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : rbd-nbd-2:19.2.3-678.ge911bdeb.el9.x86_64 18/20 2026-03-09T15:17:45.331 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : re2-1:20211101-20.el9.x86_64 19/20 2026-03-09T15:17:45.361 INFO:teuthology.orchestra.run.vm03.stdout: Verifying : thrift-0.15.0-4.el9.x86_64 20/20 2026-03-09T15:17:45.361 INFO:teuthology.orchestra.run.vm03.stdout: 2026-03-09T15:17:45.361 INFO:teuthology.orchestra.run.vm03.stdout:Removed: 2026-03-09T15:17:45.361 INFO:teuthology.orchestra.run.vm03.stdout: boost-program-options-1.75.0-13.el9.x86_64 2026-03-09T15:17:45.361 INFO:teuthology.orchestra.run.vm03.stdout: libarrow-9.0.0-15.el9.x86_64 2026-03-09T15:17:45.361 INFO:teuthology.orchestra.run.vm03.stdout: libarrow-doc-9.0.0-15.el9.noarch 2026-03-09T15:17:45.361 INFO:teuthology.orchestra.run.vm03.stdout: libnbd-1.20.3-4.el9.x86_64 2026-03-09T15:17:45.361 INFO:teuthology.orchestra.run.vm03.stdout: libpmemobj-1.12.1-1.el9.x86_64 2026-03-09T15:17:45.361 INFO:teuthology.orchestra.run.vm03.stdout: librabbitmq-0.11.0-7.el9.x86_64 2026-03-09T15:17:45.361 INFO:teuthology.orchestra.run.vm03.stdout: librados2-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T15:17:45.361 INFO:teuthology.orchestra.run.vm03.stdout: librbd1-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T15:17:45.361 INFO:teuthology.orchestra.run.vm03.stdout: librdkafka-1.6.1-102.el9.x86_64 2026-03-09T15:17:45.361 INFO:teuthology.orchestra.run.vm03.stdout: librgw2-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T15:17:45.361 INFO:teuthology.orchestra.run.vm03.stdout: lttng-ust-2.12.0-6.el9.x86_64 2026-03-09T15:17:45.361 INFO:teuthology.orchestra.run.vm03.stdout: parquet-libs-9.0.0-15.el9.x86_64 2026-03-09T15:17:45.361 INFO:teuthology.orchestra.run.vm03.stdout: python3-rados-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T15:17:45.361 INFO:teuthology.orchestra.run.vm03.stdout: python3-rbd-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T15:17:45.361 INFO:teuthology.orchestra.run.vm03.stdout: python3-rgw-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T15:17:45.361 INFO:teuthology.orchestra.run.vm03.stdout: qemu-kvm-block-rbd-17:10.1.0-15.el9.x86_64 2026-03-09T15:17:45.361 INFO:teuthology.orchestra.run.vm03.stdout: rbd-fuse-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T15:17:45.361 INFO:teuthology.orchestra.run.vm03.stdout: rbd-nbd-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T15:17:45.361 INFO:teuthology.orchestra.run.vm03.stdout: re2-1:20211101-20.el9.x86_64 2026-03-09T15:17:45.361 INFO:teuthology.orchestra.run.vm03.stdout: thrift-0.15.0-4.el9.x86_64 2026-03-09T15:17:45.361 INFO:teuthology.orchestra.run.vm03.stdout: 2026-03-09T15:17:45.361 INFO:teuthology.orchestra.run.vm03.stdout:Complete! 2026-03-09T15:17:45.382 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : thrift-0.15.0-4.el9.x86_64 20/20 2026-03-09T15:17:45.382 INFO:teuthology.orchestra.run.vm06.stdout: 2026-03-09T15:17:45.382 INFO:teuthology.orchestra.run.vm06.stdout:Removed: 2026-03-09T15:17:45.382 INFO:teuthology.orchestra.run.vm06.stdout: boost-program-options-1.75.0-13.el9.x86_64 2026-03-09T15:17:45.382 INFO:teuthology.orchestra.run.vm06.stdout: libarrow-9.0.0-15.el9.x86_64 2026-03-09T15:17:45.382 INFO:teuthology.orchestra.run.vm06.stdout: libarrow-doc-9.0.0-15.el9.noarch 2026-03-09T15:17:45.382 INFO:teuthology.orchestra.run.vm06.stdout: libnbd-1.20.3-4.el9.x86_64 2026-03-09T15:17:45.382 INFO:teuthology.orchestra.run.vm06.stdout: libpmemobj-1.12.1-1.el9.x86_64 2026-03-09T15:17:45.382 INFO:teuthology.orchestra.run.vm06.stdout: librabbitmq-0.11.0-7.el9.x86_64 2026-03-09T15:17:45.382 INFO:teuthology.orchestra.run.vm06.stdout: librados2-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T15:17:45.382 INFO:teuthology.orchestra.run.vm06.stdout: librbd1-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T15:17:45.382 INFO:teuthology.orchestra.run.vm06.stdout: librdkafka-1.6.1-102.el9.x86_64 2026-03-09T15:17:45.382 INFO:teuthology.orchestra.run.vm06.stdout: librgw2-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T15:17:45.382 INFO:teuthology.orchestra.run.vm06.stdout: lttng-ust-2.12.0-6.el9.x86_64 2026-03-09T15:17:45.383 INFO:teuthology.orchestra.run.vm06.stdout: parquet-libs-9.0.0-15.el9.x86_64 2026-03-09T15:17:45.383 INFO:teuthology.orchestra.run.vm06.stdout: python3-rados-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T15:17:45.383 INFO:teuthology.orchestra.run.vm06.stdout: python3-rbd-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T15:17:45.383 INFO:teuthology.orchestra.run.vm06.stdout: python3-rgw-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T15:17:45.383 INFO:teuthology.orchestra.run.vm06.stdout: qemu-kvm-block-rbd-17:10.1.0-15.el9.x86_64 2026-03-09T15:17:45.383 INFO:teuthology.orchestra.run.vm06.stdout: rbd-fuse-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T15:17:45.383 INFO:teuthology.orchestra.run.vm06.stdout: rbd-nbd-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T15:17:45.383 INFO:teuthology.orchestra.run.vm06.stdout: re2-1:20211101-20.el9.x86_64 2026-03-09T15:17:45.383 INFO:teuthology.orchestra.run.vm06.stdout: thrift-0.15.0-4.el9.x86_64 2026-03-09T15:17:45.383 INFO:teuthology.orchestra.run.vm06.stdout: 2026-03-09T15:17:45.383 INFO:teuthology.orchestra.run.vm06.stdout:Complete! 2026-03-09T15:17:45.532 INFO:teuthology.orchestra.run.vm04.stdout:No match for argument: librbd1 2026-03-09T15:17:45.533 INFO:teuthology.orchestra.run.vm04.stderr:No packages marked for removal. 2026-03-09T15:17:45.536 INFO:teuthology.orchestra.run.vm04.stdout:Dependencies resolved. 2026-03-09T15:17:45.536 INFO:teuthology.orchestra.run.vm04.stdout:Nothing to do. 2026-03-09T15:17:45.536 INFO:teuthology.orchestra.run.vm04.stdout:Complete! 2026-03-09T15:17:45.568 INFO:teuthology.orchestra.run.vm03.stdout:No match for argument: librbd1 2026-03-09T15:17:45.568 INFO:teuthology.orchestra.run.vm03.stderr:No packages marked for removal. 2026-03-09T15:17:45.570 INFO:teuthology.orchestra.run.vm03.stdout:Dependencies resolved. 2026-03-09T15:17:45.571 INFO:teuthology.orchestra.run.vm03.stdout:Nothing to do. 2026-03-09T15:17:45.571 INFO:teuthology.orchestra.run.vm03.stdout:Complete! 2026-03-09T15:17:45.602 INFO:teuthology.orchestra.run.vm06.stdout:No match for argument: librbd1 2026-03-09T15:17:45.602 INFO:teuthology.orchestra.run.vm06.stderr:No packages marked for removal. 2026-03-09T15:17:45.604 INFO:teuthology.orchestra.run.vm06.stdout:Dependencies resolved. 2026-03-09T15:17:45.605 INFO:teuthology.orchestra.run.vm06.stdout:Nothing to do. 2026-03-09T15:17:45.605 INFO:teuthology.orchestra.run.vm06.stdout:Complete! 2026-03-09T15:17:45.721 INFO:teuthology.orchestra.run.vm04.stdout:No match for argument: python3-rados 2026-03-09T15:17:45.722 INFO:teuthology.orchestra.run.vm04.stderr:No packages marked for removal. 2026-03-09T15:17:45.723 INFO:teuthology.orchestra.run.vm04.stdout:Dependencies resolved. 2026-03-09T15:17:45.724 INFO:teuthology.orchestra.run.vm04.stdout:Nothing to do. 2026-03-09T15:17:45.724 INFO:teuthology.orchestra.run.vm04.stdout:Complete! 2026-03-09T15:17:45.755 INFO:teuthology.orchestra.run.vm03.stdout:No match for argument: python3-rados 2026-03-09T15:17:45.756 INFO:teuthology.orchestra.run.vm03.stderr:No packages marked for removal. 2026-03-09T15:17:45.758 INFO:teuthology.orchestra.run.vm03.stdout:Dependencies resolved. 2026-03-09T15:17:45.759 INFO:teuthology.orchestra.run.vm03.stdout:Nothing to do. 2026-03-09T15:17:45.759 INFO:teuthology.orchestra.run.vm03.stdout:Complete! 2026-03-09T15:17:45.787 INFO:teuthology.orchestra.run.vm06.stdout:No match for argument: python3-rados 2026-03-09T15:17:45.787 INFO:teuthology.orchestra.run.vm06.stderr:No packages marked for removal. 2026-03-09T15:17:45.789 INFO:teuthology.orchestra.run.vm06.stdout:Dependencies resolved. 2026-03-09T15:17:45.789 INFO:teuthology.orchestra.run.vm06.stdout:Nothing to do. 2026-03-09T15:17:45.789 INFO:teuthology.orchestra.run.vm06.stdout:Complete! 2026-03-09T15:17:45.884 INFO:teuthology.orchestra.run.vm04.stdout:No match for argument: python3-rgw 2026-03-09T15:17:45.885 INFO:teuthology.orchestra.run.vm04.stderr:No packages marked for removal. 2026-03-09T15:17:45.886 INFO:teuthology.orchestra.run.vm04.stdout:Dependencies resolved. 2026-03-09T15:17:45.887 INFO:teuthology.orchestra.run.vm04.stdout:Nothing to do. 2026-03-09T15:17:45.887 INFO:teuthology.orchestra.run.vm04.stdout:Complete! 2026-03-09T15:17:45.922 INFO:teuthology.orchestra.run.vm03.stdout:No match for argument: python3-rgw 2026-03-09T15:17:45.922 INFO:teuthology.orchestra.run.vm03.stderr:No packages marked for removal. 2026-03-09T15:17:45.924 INFO:teuthology.orchestra.run.vm03.stdout:Dependencies resolved. 2026-03-09T15:17:45.925 INFO:teuthology.orchestra.run.vm03.stdout:Nothing to do. 2026-03-09T15:17:45.925 INFO:teuthology.orchestra.run.vm03.stdout:Complete! 2026-03-09T15:17:45.947 INFO:teuthology.orchestra.run.vm06.stdout:No match for argument: python3-rgw 2026-03-09T15:17:45.947 INFO:teuthology.orchestra.run.vm06.stderr:No packages marked for removal. 2026-03-09T15:17:45.949 INFO:teuthology.orchestra.run.vm06.stdout:Dependencies resolved. 2026-03-09T15:17:45.949 INFO:teuthology.orchestra.run.vm06.stdout:Nothing to do. 2026-03-09T15:17:45.949 INFO:teuthology.orchestra.run.vm06.stdout:Complete! 2026-03-09T15:17:46.043 INFO:teuthology.orchestra.run.vm04.stdout:No match for argument: python3-cephfs 2026-03-09T15:17:46.043 INFO:teuthology.orchestra.run.vm04.stderr:No packages marked for removal. 2026-03-09T15:17:46.045 INFO:teuthology.orchestra.run.vm04.stdout:Dependencies resolved. 2026-03-09T15:17:46.046 INFO:teuthology.orchestra.run.vm04.stdout:Nothing to do. 2026-03-09T15:17:46.046 INFO:teuthology.orchestra.run.vm04.stdout:Complete! 2026-03-09T15:17:46.082 INFO:teuthology.orchestra.run.vm03.stdout:No match for argument: python3-cephfs 2026-03-09T15:17:46.082 INFO:teuthology.orchestra.run.vm03.stderr:No packages marked for removal. 2026-03-09T15:17:46.084 INFO:teuthology.orchestra.run.vm03.stdout:Dependencies resolved. 2026-03-09T15:17:46.084 INFO:teuthology.orchestra.run.vm03.stdout:Nothing to do. 2026-03-09T15:17:46.084 INFO:teuthology.orchestra.run.vm03.stdout:Complete! 2026-03-09T15:17:46.107 INFO:teuthology.orchestra.run.vm06.stdout:No match for argument: python3-cephfs 2026-03-09T15:17:46.108 INFO:teuthology.orchestra.run.vm06.stderr:No packages marked for removal. 2026-03-09T15:17:46.109 INFO:teuthology.orchestra.run.vm06.stdout:Dependencies resolved. 2026-03-09T15:17:46.110 INFO:teuthology.orchestra.run.vm06.stdout:Nothing to do. 2026-03-09T15:17:46.110 INFO:teuthology.orchestra.run.vm06.stdout:Complete! 2026-03-09T15:17:46.201 INFO:teuthology.orchestra.run.vm04.stdout:No match for argument: python3-rbd 2026-03-09T15:17:46.202 INFO:teuthology.orchestra.run.vm04.stderr:No packages marked for removal. 2026-03-09T15:17:46.203 INFO:teuthology.orchestra.run.vm04.stdout:Dependencies resolved. 2026-03-09T15:17:46.204 INFO:teuthology.orchestra.run.vm04.stdout:Nothing to do. 2026-03-09T15:17:46.204 INFO:teuthology.orchestra.run.vm04.stdout:Complete! 2026-03-09T15:17:46.241 INFO:teuthology.orchestra.run.vm03.stdout:No match for argument: python3-rbd 2026-03-09T15:17:46.242 INFO:teuthology.orchestra.run.vm03.stderr:No packages marked for removal. 2026-03-09T15:17:46.243 INFO:teuthology.orchestra.run.vm03.stdout:Dependencies resolved. 2026-03-09T15:17:46.244 INFO:teuthology.orchestra.run.vm03.stdout:Nothing to do. 2026-03-09T15:17:46.244 INFO:teuthology.orchestra.run.vm03.stdout:Complete! 2026-03-09T15:17:46.267 INFO:teuthology.orchestra.run.vm06.stdout:No match for argument: python3-rbd 2026-03-09T15:17:46.267 INFO:teuthology.orchestra.run.vm06.stderr:No packages marked for removal. 2026-03-09T15:17:46.269 INFO:teuthology.orchestra.run.vm06.stdout:Dependencies resolved. 2026-03-09T15:17:46.270 INFO:teuthology.orchestra.run.vm06.stdout:Nothing to do. 2026-03-09T15:17:46.270 INFO:teuthology.orchestra.run.vm06.stdout:Complete! 2026-03-09T15:17:46.361 INFO:teuthology.orchestra.run.vm04.stdout:No match for argument: rbd-fuse 2026-03-09T15:17:46.361 INFO:teuthology.orchestra.run.vm04.stderr:No packages marked for removal. 2026-03-09T15:17:46.364 INFO:teuthology.orchestra.run.vm04.stdout:Dependencies resolved. 2026-03-09T15:17:46.364 INFO:teuthology.orchestra.run.vm04.stdout:Nothing to do. 2026-03-09T15:17:46.364 INFO:teuthology.orchestra.run.vm04.stdout:Complete! 2026-03-09T15:17:46.402 INFO:teuthology.orchestra.run.vm03.stdout:No match for argument: rbd-fuse 2026-03-09T15:17:46.402 INFO:teuthology.orchestra.run.vm03.stderr:No packages marked for removal. 2026-03-09T15:17:46.404 INFO:teuthology.orchestra.run.vm03.stdout:Dependencies resolved. 2026-03-09T15:17:46.405 INFO:teuthology.orchestra.run.vm03.stdout:Nothing to do. 2026-03-09T15:17:46.405 INFO:teuthology.orchestra.run.vm03.stdout:Complete! 2026-03-09T15:17:46.428 INFO:teuthology.orchestra.run.vm06.stdout:No match for argument: rbd-fuse 2026-03-09T15:17:46.428 INFO:teuthology.orchestra.run.vm06.stderr:No packages marked for removal. 2026-03-09T15:17:46.430 INFO:teuthology.orchestra.run.vm06.stdout:Dependencies resolved. 2026-03-09T15:17:46.430 INFO:teuthology.orchestra.run.vm06.stdout:Nothing to do. 2026-03-09T15:17:46.430 INFO:teuthology.orchestra.run.vm06.stdout:Complete! 2026-03-09T15:17:46.521 INFO:teuthology.orchestra.run.vm04.stdout:No match for argument: rbd-mirror 2026-03-09T15:17:46.521 INFO:teuthology.orchestra.run.vm04.stderr:No packages marked for removal. 2026-03-09T15:17:46.523 INFO:teuthology.orchestra.run.vm04.stdout:Dependencies resolved. 2026-03-09T15:17:46.524 INFO:teuthology.orchestra.run.vm04.stdout:Nothing to do. 2026-03-09T15:17:46.524 INFO:teuthology.orchestra.run.vm04.stdout:Complete! 2026-03-09T15:17:46.563 INFO:teuthology.orchestra.run.vm03.stdout:No match for argument: rbd-mirror 2026-03-09T15:17:46.563 INFO:teuthology.orchestra.run.vm03.stderr:No packages marked for removal. 2026-03-09T15:17:46.565 INFO:teuthology.orchestra.run.vm03.stdout:Dependencies resolved. 2026-03-09T15:17:46.566 INFO:teuthology.orchestra.run.vm03.stdout:Nothing to do. 2026-03-09T15:17:46.566 INFO:teuthology.orchestra.run.vm03.stdout:Complete! 2026-03-09T15:17:46.587 INFO:teuthology.orchestra.run.vm06.stdout:No match for argument: rbd-mirror 2026-03-09T15:17:46.587 INFO:teuthology.orchestra.run.vm06.stderr:No packages marked for removal. 2026-03-09T15:17:46.589 INFO:teuthology.orchestra.run.vm06.stdout:Dependencies resolved. 2026-03-09T15:17:46.590 INFO:teuthology.orchestra.run.vm06.stdout:Nothing to do. 2026-03-09T15:17:46.590 INFO:teuthology.orchestra.run.vm06.stdout:Complete! 2026-03-09T15:17:46.681 INFO:teuthology.orchestra.run.vm04.stdout:No match for argument: rbd-nbd 2026-03-09T15:17:46.681 INFO:teuthology.orchestra.run.vm04.stderr:No packages marked for removal. 2026-03-09T15:17:46.683 INFO:teuthology.orchestra.run.vm04.stdout:Dependencies resolved. 2026-03-09T15:17:46.684 INFO:teuthology.orchestra.run.vm04.stdout:Nothing to do. 2026-03-09T15:17:46.684 INFO:teuthology.orchestra.run.vm04.stdout:Complete! 2026-03-09T15:17:46.705 DEBUG:teuthology.orchestra.run.vm04:> sudo yum clean all 2026-03-09T15:17:46.723 INFO:teuthology.orchestra.run.vm03.stdout:No match for argument: rbd-nbd 2026-03-09T15:17:46.723 INFO:teuthology.orchestra.run.vm03.stderr:No packages marked for removal. 2026-03-09T15:17:46.725 INFO:teuthology.orchestra.run.vm03.stdout:Dependencies resolved. 2026-03-09T15:17:46.725 INFO:teuthology.orchestra.run.vm03.stdout:Nothing to do. 2026-03-09T15:17:46.725 INFO:teuthology.orchestra.run.vm03.stdout:Complete! 2026-03-09T15:17:46.746 INFO:teuthology.orchestra.run.vm06.stdout:No match for argument: rbd-nbd 2026-03-09T15:17:46.746 INFO:teuthology.orchestra.run.vm06.stderr:No packages marked for removal. 2026-03-09T15:17:46.747 DEBUG:teuthology.orchestra.run.vm03:> sudo yum clean all 2026-03-09T15:17:46.748 INFO:teuthology.orchestra.run.vm06.stdout:Dependencies resolved. 2026-03-09T15:17:46.749 INFO:teuthology.orchestra.run.vm06.stdout:Nothing to do. 2026-03-09T15:17:46.749 INFO:teuthology.orchestra.run.vm06.stdout:Complete! 2026-03-09T15:17:46.769 DEBUG:teuthology.orchestra.run.vm06:> sudo yum clean all 2026-03-09T15:17:46.832 INFO:teuthology.orchestra.run.vm04.stdout:56 files removed 2026-03-09T15:17:46.853 DEBUG:teuthology.orchestra.run.vm04:> sudo rm -f /etc/yum.repos.d/ceph.repo 2026-03-09T15:17:46.869 INFO:teuthology.orchestra.run.vm03.stdout:56 files removed 2026-03-09T15:17:46.878 DEBUG:teuthology.orchestra.run.vm04:> sudo yum clean expire-cache 2026-03-09T15:17:46.889 DEBUG:teuthology.orchestra.run.vm03:> sudo rm -f /etc/yum.repos.d/ceph.repo 2026-03-09T15:17:46.890 INFO:teuthology.orchestra.run.vm06.stdout:56 files removed 2026-03-09T15:17:46.910 DEBUG:teuthology.orchestra.run.vm06:> sudo rm -f /etc/yum.repos.d/ceph.repo 2026-03-09T15:17:46.911 DEBUG:teuthology.orchestra.run.vm03:> sudo yum clean expire-cache 2026-03-09T15:17:46.931 DEBUG:teuthology.orchestra.run.vm06:> sudo yum clean expire-cache 2026-03-09T15:17:47.024 INFO:teuthology.orchestra.run.vm04.stdout:Cache was expired 2026-03-09T15:17:47.024 INFO:teuthology.orchestra.run.vm04.stdout:0 files removed 2026-03-09T15:17:47.042 DEBUG:teuthology.parallel:result is None 2026-03-09T15:17:47.057 INFO:teuthology.orchestra.run.vm03.stdout:Cache was expired 2026-03-09T15:17:47.057 INFO:teuthology.orchestra.run.vm03.stdout:0 files removed 2026-03-09T15:17:47.073 DEBUG:teuthology.parallel:result is None 2026-03-09T15:17:47.078 INFO:teuthology.orchestra.run.vm06.stdout:Cache was expired 2026-03-09T15:17:47.078 INFO:teuthology.orchestra.run.vm06.stdout:0 files removed 2026-03-09T15:17:47.094 DEBUG:teuthology.parallel:result is None 2026-03-09T15:17:47.094 INFO:teuthology.task.install:Removing ceph sources lists on ubuntu@vm03.local 2026-03-09T15:17:47.094 INFO:teuthology.task.install:Removing ceph sources lists on ubuntu@vm04.local 2026-03-09T15:17:47.094 INFO:teuthology.task.install:Removing ceph sources lists on ubuntu@vm06.local 2026-03-09T15:17:47.095 DEBUG:teuthology.orchestra.run.vm03:> sudo rm -f /etc/yum.repos.d/ceph.repo 2026-03-09T15:17:47.095 DEBUG:teuthology.orchestra.run.vm04:> sudo rm -f /etc/yum.repos.d/ceph.repo 2026-03-09T15:17:47.095 DEBUG:teuthology.orchestra.run.vm06:> sudo rm -f /etc/yum.repos.d/ceph.repo 2026-03-09T15:17:47.118 DEBUG:teuthology.orchestra.run.vm03:> sudo mv -f /etc/yum/pluginconf.d/priorities.conf.orig /etc/yum/pluginconf.d/priorities.conf 2026-03-09T15:17:47.118 DEBUG:teuthology.orchestra.run.vm04:> sudo mv -f /etc/yum/pluginconf.d/priorities.conf.orig /etc/yum/pluginconf.d/priorities.conf 2026-03-09T15:17:47.118 DEBUG:teuthology.orchestra.run.vm06:> sudo mv -f /etc/yum/pluginconf.d/priorities.conf.orig /etc/yum/pluginconf.d/priorities.conf 2026-03-09T15:17:47.182 DEBUG:teuthology.parallel:result is None 2026-03-09T15:17:47.182 DEBUG:teuthology.parallel:result is None 2026-03-09T15:17:47.182 DEBUG:teuthology.parallel:result is None 2026-03-09T15:17:47.182 DEBUG:teuthology.run_tasks:Unwinding manager clock 2026-03-09T15:17:47.184 INFO:teuthology.task.clock:Checking final clock skew... 2026-03-09T15:17:47.184 DEBUG:teuthology.orchestra.run.vm03:> PATH=/usr/bin:/usr/sbin ntpq -p || PATH=/usr/bin:/usr/sbin chronyc sources || true 2026-03-09T15:17:47.224 DEBUG:teuthology.orchestra.run.vm04:> PATH=/usr/bin:/usr/sbin ntpq -p || PATH=/usr/bin:/usr/sbin chronyc sources || true 2026-03-09T15:17:47.226 DEBUG:teuthology.orchestra.run.vm06:> PATH=/usr/bin:/usr/sbin ntpq -p || PATH=/usr/bin:/usr/sbin chronyc sources || true 2026-03-09T15:17:47.236 INFO:teuthology.orchestra.run.vm03.stderr:bash: line 1: ntpq: command not found 2026-03-09T15:17:47.238 INFO:teuthology.orchestra.run.vm04.stderr:bash: line 1: ntpq: command not found 2026-03-09T15:17:47.239 INFO:teuthology.orchestra.run.vm06.stderr:bash: line 1: ntpq: command not found 2026-03-09T15:17:47.240 INFO:teuthology.orchestra.run.vm03.stdout:MS Name/IP address Stratum Poll Reach LastRx Last sample 2026-03-09T15:17:47.240 INFO:teuthology.orchestra.run.vm03.stdout:=============================================================================== 2026-03-09T15:17:47.240 INFO:teuthology.orchestra.run.vm03.stdout:^+ formularfetischisten.de 2 6 377 47 -660us[ -660us] +/- 41ms 2026-03-09T15:17:47.240 INFO:teuthology.orchestra.run.vm03.stdout:^* s7.vonderste.in 2 6 377 47 +1018us[ +866us] +/- 19ms 2026-03-09T15:17:47.240 INFO:teuthology.orchestra.run.vm03.stdout:^+ mail2.light-speed.de 2 6 377 47 -696us[ -696us] +/- 18ms 2026-03-09T15:17:47.240 INFO:teuthology.orchestra.run.vm03.stdout:^+ bond1-1201.fsn-lf-s02.pr> 2 6 377 113 -809us[-1144us] +/- 20ms 2026-03-09T15:17:47.240 INFO:teuthology.orchestra.run.vm04.stdout:MS Name/IP address Stratum Poll Reach LastRx Last sample 2026-03-09T15:17:47.240 INFO:teuthology.orchestra.run.vm04.stdout:=============================================================================== 2026-03-09T15:17:47.240 INFO:teuthology.orchestra.run.vm04.stdout:^+ s7.vonderste.in 2 6 377 47 +915us[ +915us] +/- 19ms 2026-03-09T15:17:47.240 INFO:teuthology.orchestra.run.vm04.stdout:^* mail2.light-speed.de 2 6 377 49 -774us[-1156us] +/- 18ms 2026-03-09T15:17:47.240 INFO:teuthology.orchestra.run.vm04.stdout:^+ bond1-1201.fsn-lf-s02.pr> 2 6 377 48 -1026us[-1026us] +/- 20ms 2026-03-09T15:17:47.240 INFO:teuthology.orchestra.run.vm04.stdout:^+ formularfetischisten.de 2 6 377 47 -687us[ -687us] +/- 41ms 2026-03-09T15:17:47.242 INFO:teuthology.orchestra.run.vm06.stdout:MS Name/IP address Stratum Poll Reach LastRx Last sample 2026-03-09T15:17:47.242 INFO:teuthology.orchestra.run.vm06.stdout:=============================================================================== 2026-03-09T15:17:47.242 INFO:teuthology.orchestra.run.vm06.stdout:^+ bond1-1201.fsn-lf-s02.pr> 2 6 377 49 -841us[-1111us] +/- 20ms 2026-03-09T15:17:47.242 INFO:teuthology.orchestra.run.vm06.stdout:^+ formularfetischisten.de 2 6 377 48 -627us[ -897us] +/- 41ms 2026-03-09T15:17:47.242 INFO:teuthology.orchestra.run.vm06.stdout:^+ s7.vonderste.in 2 6 377 47 +1042us[ +772us] +/- 19ms 2026-03-09T15:17:47.242 INFO:teuthology.orchestra.run.vm06.stdout:^* mail2.light-speed.de 2 6 377 46 -618us[ -888us] +/- 18ms 2026-03-09T15:17:47.243 DEBUG:teuthology.run_tasks:Unwinding manager ansible.cephlab 2026-03-09T15:17:47.245 INFO:teuthology.task.ansible:Skipping ansible cleanup... 2026-03-09T15:17:47.245 DEBUG:teuthology.run_tasks:Unwinding manager selinux 2026-03-09T15:17:47.247 DEBUG:teuthology.run_tasks:Unwinding manager pcp 2026-03-09T15:17:47.249 DEBUG:teuthology.run_tasks:Unwinding manager internal.timer 2026-03-09T15:17:47.251 INFO:teuthology.task.internal:Duration was 574.630209 seconds 2026-03-09T15:17:47.251 DEBUG:teuthology.run_tasks:Unwinding manager internal.syslog 2026-03-09T15:17:47.253 INFO:teuthology.task.internal.syslog:Shutting down syslog monitoring... 2026-03-09T15:17:47.253 DEBUG:teuthology.orchestra.run.vm03:> sudo rm -f -- /etc/rsyslog.d/80-cephtest.conf && sudo service rsyslog restart 2026-03-09T15:17:47.282 DEBUG:teuthology.orchestra.run.vm04:> sudo rm -f -- /etc/rsyslog.d/80-cephtest.conf && sudo service rsyslog restart 2026-03-09T15:17:47.283 DEBUG:teuthology.orchestra.run.vm06:> sudo rm -f -- /etc/rsyslog.d/80-cephtest.conf && sudo service rsyslog restart 2026-03-09T15:17:47.318 INFO:teuthology.orchestra.run.vm04.stderr:Redirecting to /bin/systemctl restart rsyslog.service 2026-03-09T15:17:47.321 INFO:teuthology.orchestra.run.vm03.stderr:Redirecting to /bin/systemctl restart rsyslog.service 2026-03-09T15:17:47.321 INFO:teuthology.orchestra.run.vm06.stderr:Redirecting to /bin/systemctl restart rsyslog.service 2026-03-09T15:17:47.573 INFO:teuthology.task.internal.syslog:Checking logs for errors... 2026-03-09T15:17:47.573 DEBUG:teuthology.task.internal.syslog:Checking ubuntu@vm03.local 2026-03-09T15:17:47.573 DEBUG:teuthology.orchestra.run.vm03:> grep -E --binary-files=text '\bBUG\b|\bINFO\b|\bDEADLOCK\b' /home/ubuntu/cephtest/archive/syslog/kern.log | grep -v 'task .* blocked for more than .* seconds' | grep -v 'lockdep is turned off' | grep -v 'trying to register non-static key' | grep -v 'DEBUG: fsize' | grep -v CRON | grep -v 'BUG: bad unlock balance detected' | grep -v 'inconsistent lock state' | grep -v '*** DEADLOCK ***' | grep -v 'INFO: possible irq lock inversion dependency detected' | grep -v 'INFO: NMI handler (perf_event_nmi_handler) took too long to run' | grep -v 'INFO: recovery required on readonly' | grep -v 'ceph-create-keys: INFO' | grep -v INFO:ceph-create-keys | grep -v 'Loaded datasource DataSourceOpenStack' | grep -v 'container-storage-setup: INFO: Volume group backing root filesystem could not be determined' | grep -E -v '\bsalt-master\b|\bsalt-minion\b|\bsalt-api\b' | grep -v ceph-crash | grep -E -v '\btcmu-runner\b.*\bINFO\b' | head -n 1 2026-03-09T15:17:47.638 DEBUG:teuthology.task.internal.syslog:Checking ubuntu@vm04.local 2026-03-09T15:17:47.638 DEBUG:teuthology.orchestra.run.vm04:> grep -E --binary-files=text '\bBUG\b|\bINFO\b|\bDEADLOCK\b' /home/ubuntu/cephtest/archive/syslog/kern.log | grep -v 'task .* blocked for more than .* seconds' | grep -v 'lockdep is turned off' | grep -v 'trying to register non-static key' | grep -v 'DEBUG: fsize' | grep -v CRON | grep -v 'BUG: bad unlock balance detected' | grep -v 'inconsistent lock state' | grep -v '*** DEADLOCK ***' | grep -v 'INFO: possible irq lock inversion dependency detected' | grep -v 'INFO: NMI handler (perf_event_nmi_handler) took too long to run' | grep -v 'INFO: recovery required on readonly' | grep -v 'ceph-create-keys: INFO' | grep -v INFO:ceph-create-keys | grep -v 'Loaded datasource DataSourceOpenStack' | grep -v 'container-storage-setup: INFO: Volume group backing root filesystem could not be determined' | grep -E -v '\bsalt-master\b|\bsalt-minion\b|\bsalt-api\b' | grep -v ceph-crash | grep -E -v '\btcmu-runner\b.*\bINFO\b' | head -n 1 2026-03-09T15:17:47.664 DEBUG:teuthology.task.internal.syslog:Checking ubuntu@vm06.local 2026-03-09T15:17:47.664 DEBUG:teuthology.orchestra.run.vm06:> grep -E --binary-files=text '\bBUG\b|\bINFO\b|\bDEADLOCK\b' /home/ubuntu/cephtest/archive/syslog/kern.log | grep -v 'task .* blocked for more than .* seconds' | grep -v 'lockdep is turned off' | grep -v 'trying to register non-static key' | grep -v 'DEBUG: fsize' | grep -v CRON | grep -v 'BUG: bad unlock balance detected' | grep -v 'inconsistent lock state' | grep -v '*** DEADLOCK ***' | grep -v 'INFO: possible irq lock inversion dependency detected' | grep -v 'INFO: NMI handler (perf_event_nmi_handler) took too long to run' | grep -v 'INFO: recovery required on readonly' | grep -v 'ceph-create-keys: INFO' | grep -v INFO:ceph-create-keys | grep -v 'Loaded datasource DataSourceOpenStack' | grep -v 'container-storage-setup: INFO: Volume group backing root filesystem could not be determined' | grep -E -v '\bsalt-master\b|\bsalt-minion\b|\bsalt-api\b' | grep -v ceph-crash | grep -E -v '\btcmu-runner\b.*\bINFO\b' | head -n 1 2026-03-09T15:17:47.689 INFO:teuthology.task.internal.syslog:Gathering journactl... 2026-03-09T15:17:47.689 DEBUG:teuthology.orchestra.run.vm03:> sudo journalctl > /home/ubuntu/cephtest/archive/syslog/journalctl.log 2026-03-09T15:17:47.690 DEBUG:teuthology.orchestra.run.vm04:> sudo journalctl > /home/ubuntu/cephtest/archive/syslog/journalctl.log 2026-03-09T15:17:47.706 DEBUG:teuthology.orchestra.run.vm06:> sudo journalctl > /home/ubuntu/cephtest/archive/syslog/journalctl.log 2026-03-09T15:17:48.133 INFO:teuthology.task.internal.syslog:Compressing syslogs... 2026-03-09T15:17:48.133 DEBUG:teuthology.orchestra.run.vm03:> find /home/ubuntu/cephtest/archive/syslog -name '*.log' -print0 | sudo xargs -0 --max-args=1 --max-procs=0 --verbose --no-run-if-empty -- gzip -5 --verbose -- 2026-03-09T15:17:48.134 DEBUG:teuthology.orchestra.run.vm04:> find /home/ubuntu/cephtest/archive/syslog -name '*.log' -print0 | sudo xargs -0 --max-args=1 --max-procs=0 --verbose --no-run-if-empty -- gzip -5 --verbose -- 2026-03-09T15:17:48.136 DEBUG:teuthology.orchestra.run.vm06:> find /home/ubuntu/cephtest/archive/syslog -name '*.log' -print0 | sudo xargs -0 --max-args=1 --max-procs=0 --verbose --no-run-if-empty -- gzip -5 --verbose -- 2026-03-09T15:17:48.155 INFO:teuthology.orchestra.run.vm03.stderr:gzip -5 --verbose -- /home/ubuntu/cephtest/archive/syslog/kern.log 2026-03-09T15:17:48.155 INFO:teuthology.orchestra.run.vm03.stderr:gzip -5 --verbose -- /home/ubuntu/cephtest/archive/syslog/misc.log 2026-03-09T15:17:48.155 INFO:teuthology.orchestra.run.vm03.stderr:gzip/home/ubuntu/cephtest/archive/syslog/kern.log: -5 --verbose -- /home/ubuntu/cephtest/archive/syslog/journalctl.log 2026-03-09T15:17:48.155 INFO:teuthology.orchestra.run.vm03.stderr: 0.0% -- replaced with /home/ubuntu/cephtest/archive/syslog/kern.log.gz 2026-03-09T15:17:48.155 INFO:teuthology.orchestra.run.vm03.stderr:/home/ubuntu/cephtest/archive/syslog/misc.log: 0.0% -- replaced with /home/ubuntu/cephtest/archive/syslog/misc.log.gz 2026-03-09T15:17:48.157 INFO:teuthology.orchestra.run.vm04.stderr:gzip -5 --verbose -- /home/ubuntu/cephtest/archive/syslog/kern.log 2026-03-09T15:17:48.157 INFO:teuthology.orchestra.run.vm04.stderr:gzip -5 --verbose -- /home/ubuntu/cephtest/archive/syslog/misc.log 2026-03-09T15:17:48.157 INFO:teuthology.orchestra.run.vm04.stderr:/home/ubuntu/cephtest/archive/syslog/kern.log: 0.0% -- replaced with /home/ubuntu/cephtest/archive/syslog/kern.log.gz 2026-03-09T15:17:48.158 INFO:teuthology.orchestra.run.vm04.stderr:gzip -5 --verbose -- /home/ubuntu/cephtest/archive/syslog/journalctl.log 2026-03-09T15:17:48.158 INFO:teuthology.orchestra.run.vm04.stderr:/home/ubuntu/cephtest/archive/syslog/misc.log: 0.0% -- replaced with /home/ubuntu/cephtest/archive/syslog/misc.log.gz/home/ubuntu/cephtest/archive/syslog/journalctl.log: 2026-03-09T15:17:48.158 INFO:teuthology.orchestra.run.vm06.stderr:gzip -5 --verbose -- /home/ubuntu/cephtest/archive/syslog/kern.log 2026-03-09T15:17:48.158 INFO:teuthology.orchestra.run.vm06.stderr:gzip -5 --verbose -- /home/ubuntu/cephtest/archive/syslog/misc.log 2026-03-09T15:17:48.159 INFO:teuthology.orchestra.run.vm06.stderr:gzip -5/home/ubuntu/cephtest/archive/syslog/kern.log: --verbose -- /home/ubuntu/cephtest/archive/syslog/journalctl.log 2026-03-09T15:17:48.159 INFO:teuthology.orchestra.run.vm06.stderr: 0.0% -- replaced with /home/ubuntu/cephtest/archive/syslog/kern.log.gz 2026-03-09T15:17:48.159 INFO:teuthology.orchestra.run.vm06.stderr:/home/ubuntu/cephtest/archive/syslog/misc.log: 0.0% -- replaced with /home/ubuntu/cephtest/archive/syslog/misc.log.gz 2026-03-09T15:17:48.277 INFO:teuthology.orchestra.run.vm04.stderr: 98.3% -- replaced with /home/ubuntu/cephtest/archive/syslog/journalctl.log.gz 2026-03-09T15:17:48.283 INFO:teuthology.orchestra.run.vm03.stderr:/home/ubuntu/cephtest/archive/syslog/journalctl.log: 98.2% -- replaced with /home/ubuntu/cephtest/archive/syslog/journalctl.log.gz 2026-03-09T15:17:48.300 INFO:teuthology.orchestra.run.vm06.stderr:/home/ubuntu/cephtest/archive/syslog/journalctl.log: 98.4% -- replaced with /home/ubuntu/cephtest/archive/syslog/journalctl.log.gz 2026-03-09T15:17:48.301 DEBUG:teuthology.run_tasks:Unwinding manager internal.sudo 2026-03-09T15:17:48.304 INFO:teuthology.task.internal:Restoring /etc/sudoers... 2026-03-09T15:17:48.304 DEBUG:teuthology.orchestra.run.vm03:> sudo mv -f /etc/sudoers.orig.teuthology /etc/sudoers 2026-03-09T15:17:48.349 DEBUG:teuthology.orchestra.run.vm04:> sudo mv -f /etc/sudoers.orig.teuthology /etc/sudoers 2026-03-09T15:17:48.372 DEBUG:teuthology.orchestra.run.vm06:> sudo mv -f /etc/sudoers.orig.teuthology /etc/sudoers 2026-03-09T15:17:48.397 DEBUG:teuthology.run_tasks:Unwinding manager internal.coredump 2026-03-09T15:17:48.399 DEBUG:teuthology.orchestra.run.vm03:> sudo sysctl -w kernel.core_pattern=core && sudo bash -c 'for f in `find /home/ubuntu/cephtest/archive/coredump -type f`; do file $f | grep -q systemd-sysusers && rm $f || true ; done' && rmdir --ignore-fail-on-non-empty -- /home/ubuntu/cephtest/archive/coredump 2026-03-09T15:17:48.400 DEBUG:teuthology.orchestra.run.vm04:> sudo sysctl -w kernel.core_pattern=core && sudo bash -c 'for f in `find /home/ubuntu/cephtest/archive/coredump -type f`; do file $f | grep -q systemd-sysusers && rm $f || true ; done' && rmdir --ignore-fail-on-non-empty -- /home/ubuntu/cephtest/archive/coredump 2026-03-09T15:17:48.415 DEBUG:teuthology.orchestra.run.vm06:> sudo sysctl -w kernel.core_pattern=core && sudo bash -c 'for f in `find /home/ubuntu/cephtest/archive/coredump -type f`; do file $f | grep -q systemd-sysusers && rm $f || true ; done' && rmdir --ignore-fail-on-non-empty -- /home/ubuntu/cephtest/archive/coredump 2026-03-09T15:17:48.421 INFO:teuthology.orchestra.run.vm03.stdout:kernel.core_pattern = core 2026-03-09T15:17:48.437 INFO:teuthology.orchestra.run.vm04.stdout:kernel.core_pattern = core 2026-03-09T15:17:48.460 INFO:teuthology.orchestra.run.vm06.stdout:kernel.core_pattern = core 2026-03-09T15:17:48.473 DEBUG:teuthology.orchestra.run.vm03:> test -e /home/ubuntu/cephtest/archive/coredump 2026-03-09T15:17:48.488 DEBUG:teuthology.orchestra.run:got remote process result: 1 2026-03-09T15:17:48.488 DEBUG:teuthology.orchestra.run.vm04:> test -e /home/ubuntu/cephtest/archive/coredump 2026-03-09T15:17:48.502 DEBUG:teuthology.orchestra.run:got remote process result: 1 2026-03-09T15:17:48.502 DEBUG:teuthology.orchestra.run.vm06:> test -e /home/ubuntu/cephtest/archive/coredump 2026-03-09T15:17:48.526 DEBUG:teuthology.orchestra.run:got remote process result: 1 2026-03-09T15:17:48.527 DEBUG:teuthology.run_tasks:Unwinding manager internal.archive 2026-03-09T15:17:48.529 INFO:teuthology.task.internal:Transferring archived files... 2026-03-09T15:17:48.529 DEBUG:teuthology.misc:Transferring archived files from vm03:/home/ubuntu/cephtest/archive to /archive/kyr-2026-03-09_11:23:05-orch-squid-none-default-vps/520/remote/vm03 2026-03-09T15:17:48.529 DEBUG:teuthology.orchestra.run.vm03:> sudo tar c -f - -C /home/ubuntu/cephtest/archive -- . 2026-03-09T15:17:48.553 DEBUG:teuthology.misc:Transferring archived files from vm04:/home/ubuntu/cephtest/archive to /archive/kyr-2026-03-09_11:23:05-orch-squid-none-default-vps/520/remote/vm04 2026-03-09T15:17:48.553 DEBUG:teuthology.orchestra.run.vm04:> sudo tar c -f - -C /home/ubuntu/cephtest/archive -- . 2026-03-09T15:17:48.579 DEBUG:teuthology.misc:Transferring archived files from vm06:/home/ubuntu/cephtest/archive to /archive/kyr-2026-03-09_11:23:05-orch-squid-none-default-vps/520/remote/vm06 2026-03-09T15:17:48.579 DEBUG:teuthology.orchestra.run.vm06:> sudo tar c -f - -C /home/ubuntu/cephtest/archive -- . 2026-03-09T15:17:48.604 INFO:teuthology.task.internal:Removing archive directory... 2026-03-09T15:17:48.604 DEBUG:teuthology.orchestra.run.vm03:> rm -rf -- /home/ubuntu/cephtest/archive 2026-03-09T15:17:48.605 DEBUG:teuthology.orchestra.run.vm04:> rm -rf -- /home/ubuntu/cephtest/archive 2026-03-09T15:17:48.620 DEBUG:teuthology.orchestra.run.vm06:> rm -rf -- /home/ubuntu/cephtest/archive 2026-03-09T15:17:48.658 DEBUG:teuthology.run_tasks:Unwinding manager internal.archive_upload 2026-03-09T15:17:48.660 INFO:teuthology.task.internal:Not uploading archives. 2026-03-09T15:17:48.660 DEBUG:teuthology.run_tasks:Unwinding manager internal.base 2026-03-09T15:17:48.663 INFO:teuthology.task.internal:Tidying up after the test... 2026-03-09T15:17:48.663 DEBUG:teuthology.orchestra.run.vm03:> find /home/ubuntu/cephtest -ls ; rmdir -- /home/ubuntu/cephtest 2026-03-09T15:17:48.664 DEBUG:teuthology.orchestra.run.vm04:> find /home/ubuntu/cephtest -ls ; rmdir -- /home/ubuntu/cephtest 2026-03-09T15:17:48.674 DEBUG:teuthology.orchestra.run.vm06:> find /home/ubuntu/cephtest -ls ; rmdir -- /home/ubuntu/cephtest 2026-03-09T15:17:48.678 INFO:teuthology.orchestra.run.vm03.stdout: 8532139 0 drwxr-xr-x 2 ubuntu ubuntu 6 Mar 9 15:17 /home/ubuntu/cephtest 2026-03-09T15:17:48.688 INFO:teuthology.orchestra.run.vm04.stdout: 8532145 0 drwxr-xr-x 2 ubuntu ubuntu 6 Mar 9 15:17 /home/ubuntu/cephtest 2026-03-09T15:17:48.713 INFO:teuthology.orchestra.run.vm06.stdout: 8532142 0 drwxr-xr-x 2 ubuntu ubuntu 6 Mar 9 15:17 /home/ubuntu/cephtest 2026-03-09T15:17:48.714 DEBUG:teuthology.run_tasks:Unwinding manager console_log 2026-03-09T15:17:48.719 INFO:teuthology.run:Summary data: description: orch/cephadm/workunits/{0-distro/centos_9.stream agent/off mon_election/classic task/test_set_mon_crush_locations} duration: 574.6302094459534 flavor: default owner: kyr success: true 2026-03-09T15:17:48.719 DEBUG:teuthology.report:Pushing job info to http://localhost:8080 2026-03-09T15:17:48.740 INFO:teuthology.run:pass